var/home/core/zuul-output/0000755000175000017500000000000015117205536014532 5ustar corecorevar/home/core/zuul-output/logs/0000755000175000017500000000000015117234236015475 5ustar corecorevar/home/core/zuul-output/logs/kubelet.log0000644000000000000000006016413715117234230017703 0ustar rootrootDec 13 06:49:02 crc systemd[1]: Starting Kubernetes Kubelet... Dec 13 06:49:02 crc restorecon[4764]: Relabeled /var/lib/kubelet/config.json from system_u:object_r:unlabeled_t:s0 to system_u:object_r:container_var_lib_t:s0 Dec 13 06:49:02 crc restorecon[4764]: /var/lib/kubelet/device-plugins not reset as customized by admin to system_u:object_r:container_file_t:s0 Dec 13 06:49:02 crc restorecon[4764]: /var/lib/kubelet/device-plugins/kubelet.sock not reset as customized by admin to system_u:object_r:container_file_t:s0 Dec 13 06:49:02 crc restorecon[4764]: /var/lib/kubelet/pods/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8/volumes/kubernetes.io~configmap/nginx-conf/..2025_02_23_05_40_35.4114275528/nginx.conf not reset as customized by admin to system_u:object_r:container_file_t:s0:c15,c25 Dec 13 06:49:02 crc restorecon[4764]: /var/lib/kubelet/pods/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c15,c25 Dec 13 06:49:02 crc restorecon[4764]: /var/lib/kubelet/pods/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8/containers/networking-console-plugin/22e96971 not reset as customized by admin to system_u:object_r:container_file_t:s0:c15,c25 Dec 13 06:49:02 crc restorecon[4764]: /var/lib/kubelet/pods/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8/containers/networking-console-plugin/21c98286 not reset as customized by admin to system_u:object_r:container_file_t:s0:c15,c25 Dec 13 06:49:02 crc restorecon[4764]: /var/lib/kubelet/pods/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8/containers/networking-console-plugin/0f1869e1 not reset as customized by admin to system_u:object_r:container_file_t:s0:c15,c25 Dec 13 06:49:02 crc restorecon[4764]: /var/lib/kubelet/pods/d1b160f5dda77d281dd8e69ec8d817f9/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c215,c682 Dec 13 06:49:02 crc restorecon[4764]: /var/lib/kubelet/pods/d1b160f5dda77d281dd8e69ec8d817f9/containers/setup/46889d52 not reset as customized by admin to system_u:object_r:container_file_t:s0:c225,c458 Dec 13 06:49:02 crc restorecon[4764]: /var/lib/kubelet/pods/d1b160f5dda77d281dd8e69ec8d817f9/containers/setup/5b6a5969 not reset as customized by admin to system_u:object_r:container_file_t:s0:c666,c963 Dec 13 06:49:02 crc restorecon[4764]: /var/lib/kubelet/pods/d1b160f5dda77d281dd8e69ec8d817f9/containers/setup/6c7921f5 not reset as customized by admin to system_u:object_r:container_file_t:s0:c215,c682 Dec 13 06:49:02 crc restorecon[4764]: /var/lib/kubelet/pods/d1b160f5dda77d281dd8e69ec8d817f9/containers/kube-rbac-proxy-crio/4804f443 not reset as customized by admin to system_u:object_r:container_file_t:s0:c225,c458 Dec 13 06:49:02 crc restorecon[4764]: /var/lib/kubelet/pods/d1b160f5dda77d281dd8e69ec8d817f9/containers/kube-rbac-proxy-crio/2a46b283 not reset as customized by admin to system_u:object_r:container_file_t:s0:c225,c458 Dec 13 06:49:02 crc restorecon[4764]: /var/lib/kubelet/pods/d1b160f5dda77d281dd8e69ec8d817f9/containers/kube-rbac-proxy-crio/a6b5573e not reset as customized by admin to system_u:object_r:container_file_t:s0:c225,c458 Dec 13 06:49:02 crc restorecon[4764]: /var/lib/kubelet/pods/d1b160f5dda77d281dd8e69ec8d817f9/containers/kube-rbac-proxy-crio/4f88ee5b not reset as customized by admin to system_u:object_r:container_file_t:s0:c225,c458 Dec 13 06:49:02 crc restorecon[4764]: /var/lib/kubelet/pods/d1b160f5dda77d281dd8e69ec8d817f9/containers/kube-rbac-proxy-crio/5a4eee4b not reset as customized by admin to system_u:object_r:container_file_t:s0:c666,c963 Dec 13 06:49:02 crc restorecon[4764]: /var/lib/kubelet/pods/d1b160f5dda77d281dd8e69ec8d817f9/containers/kube-rbac-proxy-crio/cd87c521 not reset as customized by admin to system_u:object_r:container_file_t:s0:c215,c682 Dec 13 06:49:02 crc restorecon[4764]: /var/lib/kubelet/pods/c03ee662-fb2f-4fc4-a2c1-af487c19d254/volumes/kubernetes.io~configmap/service-ca-bundle not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c24 Dec 13 06:49:02 crc restorecon[4764]: /var/lib/kubelet/pods/c03ee662-fb2f-4fc4-a2c1-af487c19d254/volumes/kubernetes.io~configmap/service-ca-bundle/..2025_02_23_05_33_42.2574241751 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c24 Dec 13 06:49:02 crc restorecon[4764]: /var/lib/kubelet/pods/c03ee662-fb2f-4fc4-a2c1-af487c19d254/volumes/kubernetes.io~configmap/service-ca-bundle/..2025_02_23_05_33_42.2574241751/service-ca.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c24 Dec 13 06:49:02 crc restorecon[4764]: /var/lib/kubelet/pods/c03ee662-fb2f-4fc4-a2c1-af487c19d254/volumes/kubernetes.io~configmap/service-ca-bundle/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c24 Dec 13 06:49:02 crc restorecon[4764]: /var/lib/kubelet/pods/c03ee662-fb2f-4fc4-a2c1-af487c19d254/volumes/kubernetes.io~configmap/service-ca-bundle/service-ca.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c24 Dec 13 06:49:02 crc restorecon[4764]: /var/lib/kubelet/pods/c03ee662-fb2f-4fc4-a2c1-af487c19d254/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c24 Dec 13 06:49:02 crc restorecon[4764]: /var/lib/kubelet/pods/c03ee662-fb2f-4fc4-a2c1-af487c19d254/containers/router/38602af4 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c24 Dec 13 06:49:02 crc restorecon[4764]: /var/lib/kubelet/pods/c03ee662-fb2f-4fc4-a2c1-af487c19d254/containers/router/1483b002 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c24 Dec 13 06:49:02 crc restorecon[4764]: /var/lib/kubelet/pods/c03ee662-fb2f-4fc4-a2c1-af487c19d254/containers/router/0346718b not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c24 Dec 13 06:49:02 crc restorecon[4764]: /var/lib/kubelet/pods/c03ee662-fb2f-4fc4-a2c1-af487c19d254/containers/router/d3ed4ada not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c24 Dec 13 06:49:02 crc restorecon[4764]: /var/lib/kubelet/pods/c03ee662-fb2f-4fc4-a2c1-af487c19d254/containers/router/3bb473a5 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c24 Dec 13 06:49:02 crc restorecon[4764]: /var/lib/kubelet/pods/c03ee662-fb2f-4fc4-a2c1-af487c19d254/containers/router/8cd075a9 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c24 Dec 13 06:49:02 crc restorecon[4764]: /var/lib/kubelet/pods/c03ee662-fb2f-4fc4-a2c1-af487c19d254/containers/router/00ab4760 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c24 Dec 13 06:49:02 crc restorecon[4764]: /var/lib/kubelet/pods/c03ee662-fb2f-4fc4-a2c1-af487c19d254/containers/router/54a21c09 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c24 Dec 13 06:49:02 crc restorecon[4764]: /var/lib/kubelet/pods/37a5e44f-9a88-4405-be8a-b645485e7312/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c589,c726 Dec 13 06:49:02 crc restorecon[4764]: /var/lib/kubelet/pods/37a5e44f-9a88-4405-be8a-b645485e7312/containers/network-operator/70478888 not reset as customized by admin to system_u:object_r:container_file_t:s0:c176,c499 Dec 13 06:49:02 crc restorecon[4764]: /var/lib/kubelet/pods/37a5e44f-9a88-4405-be8a-b645485e7312/containers/network-operator/43802770 not reset as customized by admin to system_u:object_r:container_file_t:s0:c176,c499 Dec 13 06:49:02 crc restorecon[4764]: /var/lib/kubelet/pods/37a5e44f-9a88-4405-be8a-b645485e7312/containers/network-operator/955a0edc not reset as customized by admin to system_u:object_r:container_file_t:s0:c176,c499 Dec 13 06:49:02 crc restorecon[4764]: /var/lib/kubelet/pods/37a5e44f-9a88-4405-be8a-b645485e7312/containers/network-operator/bca2d009 not reset as customized by admin to system_u:object_r:container_file_t:s0:c140,c1009 Dec 13 06:49:02 crc restorecon[4764]: /var/lib/kubelet/pods/37a5e44f-9a88-4405-be8a-b645485e7312/containers/network-operator/b295f9bd not reset as customized by admin to system_u:object_r:container_file_t:s0:c589,c726 Dec 13 06:49:02 crc restorecon[4764]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/volumes/kubernetes.io~configmap/cni-binary-copy not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Dec 13 06:49:02 crc restorecon[4764]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/volumes/kubernetes.io~configmap/cni-binary-copy/..2025_02_23_05_21_22.3617465230 not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Dec 13 06:49:02 crc restorecon[4764]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/volumes/kubernetes.io~configmap/cni-binary-copy/..2025_02_23_05_21_22.3617465230/cnibincopy.sh not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Dec 13 06:49:02 crc restorecon[4764]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/volumes/kubernetes.io~configmap/cni-binary-copy/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Dec 13 06:49:02 crc restorecon[4764]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/volumes/kubernetes.io~configmap/cni-binary-copy/cnibincopy.sh not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Dec 13 06:49:02 crc restorecon[4764]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/volumes/kubernetes.io~configmap/cni-sysctl-allowlist not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Dec 13 06:49:02 crc restorecon[4764]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/volumes/kubernetes.io~configmap/cni-sysctl-allowlist/..2025_02_23_05_21_22.2050650026 not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Dec 13 06:49:02 crc restorecon[4764]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/volumes/kubernetes.io~configmap/cni-sysctl-allowlist/..2025_02_23_05_21_22.2050650026/allowlist.conf not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Dec 13 06:49:02 crc restorecon[4764]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/volumes/kubernetes.io~configmap/cni-sysctl-allowlist/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Dec 13 06:49:02 crc restorecon[4764]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/volumes/kubernetes.io~configmap/cni-sysctl-allowlist/allowlist.conf not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Dec 13 06:49:02 crc restorecon[4764]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Dec 13 06:49:02 crc restorecon[4764]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/egress-router-binary-copy/bc46ea27 not reset as customized by admin to system_u:object_r:container_file_t:s0:c203,c924 Dec 13 06:49:02 crc restorecon[4764]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/egress-router-binary-copy/5731fc1b not reset as customized by admin to system_u:object_r:container_file_t:s0:c138,c778 Dec 13 06:49:02 crc restorecon[4764]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/egress-router-binary-copy/5e1b2a3c not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Dec 13 06:49:02 crc restorecon[4764]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/cni-plugins/943f0936 not reset as customized by admin to system_u:object_r:container_file_t:s0:c203,c924 Dec 13 06:49:02 crc restorecon[4764]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/cni-plugins/3f764ee4 not reset as customized by admin to system_u:object_r:container_file_t:s0:c138,c778 Dec 13 06:49:02 crc restorecon[4764]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/cni-plugins/8695e3f9 not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Dec 13 06:49:02 crc restorecon[4764]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/bond-cni-plugin/aed7aa86 not reset as customized by admin to system_u:object_r:container_file_t:s0:c203,c924 Dec 13 06:49:02 crc restorecon[4764]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/bond-cni-plugin/c64d7448 not reset as customized by admin to system_u:object_r:container_file_t:s0:c138,c778 Dec 13 06:49:02 crc restorecon[4764]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/bond-cni-plugin/0ba16bd2 not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Dec 13 06:49:02 crc restorecon[4764]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/routeoverride-cni/207a939f not reset as customized by admin to system_u:object_r:container_file_t:s0:c203,c924 Dec 13 06:49:02 crc restorecon[4764]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/routeoverride-cni/54aa8cdb not reset as customized by admin to system_u:object_r:container_file_t:s0:c138,c778 Dec 13 06:49:02 crc restorecon[4764]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/routeoverride-cni/1f5fa595 not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Dec 13 06:49:02 crc restorecon[4764]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/whereabouts-cni-bincopy/bf9c8153 not reset as customized by admin to system_u:object_r:container_file_t:s0:c203,c924 Dec 13 06:49:02 crc restorecon[4764]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/whereabouts-cni-bincopy/47fba4ea not reset as customized by admin to system_u:object_r:container_file_t:s0:c138,c778 Dec 13 06:49:02 crc restorecon[4764]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/whereabouts-cni-bincopy/7ae55ce9 not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Dec 13 06:49:02 crc restorecon[4764]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/whereabouts-cni/7906a268 not reset as customized by admin to system_u:object_r:container_file_t:s0:c203,c924 Dec 13 06:49:02 crc restorecon[4764]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/whereabouts-cni/ce43fa69 not reset as customized by admin to system_u:object_r:container_file_t:s0:c138,c778 Dec 13 06:49:02 crc restorecon[4764]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/whereabouts-cni/7fc7ea3a not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Dec 13 06:49:02 crc restorecon[4764]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/kube-multus-additional-cni-plugins/d8c38b7d not reset as customized by admin to system_u:object_r:container_file_t:s0:c203,c924 Dec 13 06:49:02 crc restorecon[4764]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/kube-multus-additional-cni-plugins/9ef015fb not reset as customized by admin to system_u:object_r:container_file_t:s0:c138,c778 Dec 13 06:49:02 crc restorecon[4764]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/kube-multus-additional-cni-plugins/b9db6a41 not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Dec 13 06:49:02 crc restorecon[4764]: /var/lib/kubelet/pods/5b88f790-22fa-440e-b583-365168c0b23d/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c432,c991 Dec 13 06:49:02 crc restorecon[4764]: /var/lib/kubelet/pods/5b88f790-22fa-440e-b583-365168c0b23d/containers/network-metrics-daemon/b1733d79 not reset as customized by admin to system_u:object_r:container_file_t:s0:c476,c820 Dec 13 06:49:02 crc restorecon[4764]: /var/lib/kubelet/pods/5b88f790-22fa-440e-b583-365168c0b23d/containers/network-metrics-daemon/afccd338 not reset as customized by admin to system_u:object_r:container_file_t:s0:c272,c818 Dec 13 06:49:02 crc restorecon[4764]: /var/lib/kubelet/pods/5b88f790-22fa-440e-b583-365168c0b23d/containers/network-metrics-daemon/9df0a185 not reset as customized by admin to system_u:object_r:container_file_t:s0:c432,c991 Dec 13 06:49:02 crc restorecon[4764]: /var/lib/kubelet/pods/5b88f790-22fa-440e-b583-365168c0b23d/containers/kube-rbac-proxy/18938cf8 not reset as customized by admin to system_u:object_r:container_file_t:s0:c476,c820 Dec 13 06:49:02 crc restorecon[4764]: /var/lib/kubelet/pods/5b88f790-22fa-440e-b583-365168c0b23d/containers/kube-rbac-proxy/7ab4eb23 not reset as customized by admin to system_u:object_r:container_file_t:s0:c272,c818 Dec 13 06:49:02 crc restorecon[4764]: /var/lib/kubelet/pods/5b88f790-22fa-440e-b583-365168c0b23d/containers/kube-rbac-proxy/56930be6 not reset as customized by admin to system_u:object_r:container_file_t:s0:c432,c991 Dec 13 06:49:02 crc restorecon[4764]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/volumes/kubernetes.io~configmap/env-overrides not reset as customized by admin to system_u:object_r:container_file_t:s0:c440,c975 Dec 13 06:49:02 crc restorecon[4764]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/volumes/kubernetes.io~configmap/env-overrides/..2025_02_23_05_21_35.630010865 not reset as customized by admin to system_u:object_r:container_file_t:s0:c440,c975 Dec 13 06:49:02 crc restorecon[4764]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/volumes/kubernetes.io~configmap/env-overrides/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c440,c975 Dec 13 06:49:02 crc restorecon[4764]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/volumes/kubernetes.io~configmap/ovnkube-config not reset as customized by admin to system_u:object_r:container_file_t:s0:c440,c975 Dec 13 06:49:02 crc restorecon[4764]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/volumes/kubernetes.io~configmap/ovnkube-config/..2025_02_23_05_21_35.1088506337 not reset as customized by admin to system_u:object_r:container_file_t:s0:c440,c975 Dec 13 06:49:02 crc restorecon[4764]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/volumes/kubernetes.io~configmap/ovnkube-config/..2025_02_23_05_21_35.1088506337/ovnkube.conf not reset as customized by admin to system_u:object_r:container_file_t:s0:c440,c975 Dec 13 06:49:02 crc restorecon[4764]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/volumes/kubernetes.io~configmap/ovnkube-config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c440,c975 Dec 13 06:49:02 crc restorecon[4764]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/volumes/kubernetes.io~configmap/ovnkube-config/ovnkube.conf not reset as customized by admin to system_u:object_r:container_file_t:s0:c440,c975 Dec 13 06:49:02 crc restorecon[4764]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c440,c975 Dec 13 06:49:02 crc restorecon[4764]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/containers/kube-rbac-proxy/0d8e3722 not reset as customized by admin to system_u:object_r:container_file_t:s0:c89,c211 Dec 13 06:49:02 crc restorecon[4764]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/containers/kube-rbac-proxy/d22b2e76 not reset as customized by admin to system_u:object_r:container_file_t:s0:c382,c850 Dec 13 06:49:02 crc restorecon[4764]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/containers/kube-rbac-proxy/e036759f not reset as customized by admin to system_u:object_r:container_file_t:s0:c440,c975 Dec 13 06:49:02 crc restorecon[4764]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/containers/ovnkube-cluster-manager/2734c483 not reset as customized by admin to system_u:object_r:container_file_t:s0:c89,c211 Dec 13 06:49:02 crc restorecon[4764]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/containers/ovnkube-cluster-manager/57878fe7 not reset as customized by admin to system_u:object_r:container_file_t:s0:c89,c211 Dec 13 06:49:02 crc restorecon[4764]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/containers/ovnkube-cluster-manager/3f3c2e58 not reset as customized by admin to system_u:object_r:container_file_t:s0:c89,c211 Dec 13 06:49:02 crc restorecon[4764]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/containers/ovnkube-cluster-manager/375bec3e not reset as customized by admin to system_u:object_r:container_file_t:s0:c382,c850 Dec 13 06:49:02 crc restorecon[4764]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/containers/ovnkube-cluster-manager/7bc41e08 not reset as customized by admin to system_u:object_r:container_file_t:s0:c440,c975 Dec 13 06:49:02 crc restorecon[4764]: /var/lib/kubelet/pods/cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Dec 13 06:49:02 crc restorecon[4764]: /var/lib/kubelet/pods/cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d/containers/download-server/48c7a72d not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Dec 13 06:49:02 crc restorecon[4764]: /var/lib/kubelet/pods/cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d/containers/download-server/4b66701f not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Dec 13 06:49:02 crc restorecon[4764]: /var/lib/kubelet/pods/cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d/containers/download-server/a5a1c202 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Dec 13 06:49:02 crc restorecon[4764]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/volumes/kubernetes.io~configmap/ovnkube-identity-cm not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Dec 13 06:49:02 crc restorecon[4764]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/volumes/kubernetes.io~configmap/ovnkube-identity-cm/..2025_02_23_05_21_40.3350632666 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Dec 13 06:49:02 crc restorecon[4764]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/volumes/kubernetes.io~configmap/ovnkube-identity-cm/..2025_02_23_05_21_40.3350632666/additional-cert-acceptance-cond.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Dec 13 06:49:02 crc restorecon[4764]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/volumes/kubernetes.io~configmap/ovnkube-identity-cm/..2025_02_23_05_21_40.3350632666/additional-pod-admission-cond.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Dec 13 06:49:02 crc restorecon[4764]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/volumes/kubernetes.io~configmap/ovnkube-identity-cm/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Dec 13 06:49:02 crc restorecon[4764]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/volumes/kubernetes.io~configmap/ovnkube-identity-cm/additional-cert-acceptance-cond.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Dec 13 06:49:02 crc restorecon[4764]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/volumes/kubernetes.io~configmap/ovnkube-identity-cm/additional-pod-admission-cond.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Dec 13 06:49:02 crc restorecon[4764]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/volumes/kubernetes.io~configmap/env-overrides not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Dec 13 06:49:02 crc restorecon[4764]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/volumes/kubernetes.io~configmap/env-overrides/..2025_02_23_05_21_40.1388695756 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Dec 13 06:49:02 crc restorecon[4764]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/volumes/kubernetes.io~configmap/env-overrides/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Dec 13 06:49:02 crc restorecon[4764]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Dec 13 06:49:02 crc restorecon[4764]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/containers/webhook/26f3df5b not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Dec 13 06:49:02 crc restorecon[4764]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/containers/webhook/6d8fb21d not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Dec 13 06:49:02 crc restorecon[4764]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/containers/webhook/50e94777 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Dec 13 06:49:02 crc restorecon[4764]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/containers/approver/208473b3 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Dec 13 06:49:02 crc restorecon[4764]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/containers/approver/ec9e08ba not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Dec 13 06:49:02 crc restorecon[4764]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/containers/approver/3b787c39 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Dec 13 06:49:02 crc restorecon[4764]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/containers/approver/208eaed5 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Dec 13 06:49:02 crc restorecon[4764]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/containers/approver/93aa3a2b not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Dec 13 06:49:02 crc restorecon[4764]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/containers/approver/3c697968 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Dec 13 06:49:02 crc restorecon[4764]: /var/lib/kubelet/pods/3b6479f0-333b-4a96-9adf-2099afdc2447/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c20,c21 Dec 13 06:49:02 crc restorecon[4764]: /var/lib/kubelet/pods/3b6479f0-333b-4a96-9adf-2099afdc2447/containers/network-check-target-container/ba950ec9 not reset as customized by admin to system_u:object_r:container_file_t:s0:c20,c21 Dec 13 06:49:02 crc restorecon[4764]: /var/lib/kubelet/pods/3b6479f0-333b-4a96-9adf-2099afdc2447/containers/network-check-target-container/cb5cdb37 not reset as customized by admin to system_u:object_r:container_file_t:s0:c20,c21 Dec 13 06:49:02 crc restorecon[4764]: /var/lib/kubelet/pods/3b6479f0-333b-4a96-9adf-2099afdc2447/containers/network-check-target-container/f2df9827 not reset as customized by admin to system_u:object_r:container_file_t:s0:c20,c21 Dec 13 06:49:02 crc restorecon[4764]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/volumes/kubernetes.io~configmap/images not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Dec 13 06:49:02 crc restorecon[4764]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/volumes/kubernetes.io~configmap/images/..2025_02_23_05_22_30.473230615 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Dec 13 06:49:02 crc restorecon[4764]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/volumes/kubernetes.io~configmap/images/..2025_02_23_05_22_30.473230615/images.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Dec 13 06:49:02 crc restorecon[4764]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/volumes/kubernetes.io~configmap/images/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Dec 13 06:49:02 crc restorecon[4764]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/volumes/kubernetes.io~configmap/images/images.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Dec 13 06:49:02 crc restorecon[4764]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/volumes/kubernetes.io~configmap/auth-proxy-config not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Dec 13 06:49:02 crc restorecon[4764]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/volumes/kubernetes.io~configmap/auth-proxy-config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Dec 13 06:49:02 crc restorecon[4764]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/volumes/kubernetes.io~configmap/auth-proxy-config/config-file.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Dec 13 06:49:02 crc restorecon[4764]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/volumes/kubernetes.io~configmap/auth-proxy-config/..2025_02_24_06_22_02.1904938450 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Dec 13 06:49:02 crc restorecon[4764]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/volumes/kubernetes.io~configmap/auth-proxy-config/..2025_02_24_06_22_02.1904938450/config-file.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Dec 13 06:49:02 crc restorecon[4764]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Dec 13 06:49:02 crc restorecon[4764]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/containers/machine-config-operator/fedaa673 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Dec 13 06:49:02 crc restorecon[4764]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/containers/machine-config-operator/9ca2df95 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Dec 13 06:49:02 crc restorecon[4764]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/containers/machine-config-operator/b2d7460e not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Dec 13 06:49:02 crc restorecon[4764]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/containers/kube-rbac-proxy/2207853c not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Dec 13 06:49:02 crc restorecon[4764]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/containers/kube-rbac-proxy/241c1c29 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Dec 13 06:49:02 crc restorecon[4764]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/containers/kube-rbac-proxy/2d910eaf not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Dec 13 06:49:02 crc restorecon[4764]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes/kubernetes.io~configmap/etcd-ca not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Dec 13 06:49:02 crc restorecon[4764]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes/kubernetes.io~configmap/etcd-ca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Dec 13 06:49:02 crc restorecon[4764]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes/kubernetes.io~configmap/etcd-ca/..2025_02_23_05_23_49.3726007728 not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Dec 13 06:49:02 crc restorecon[4764]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes/kubernetes.io~configmap/etcd-ca/..2025_02_23_05_23_49.3726007728/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Dec 13 06:49:02 crc restorecon[4764]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes/kubernetes.io~configmap/etcd-ca/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Dec 13 06:49:02 crc restorecon[4764]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes/kubernetes.io~configmap/etcd-service-ca not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Dec 13 06:49:02 crc restorecon[4764]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes/kubernetes.io~configmap/etcd-service-ca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Dec 13 06:49:02 crc restorecon[4764]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes/kubernetes.io~configmap/etcd-service-ca/..2025_02_23_05_23_49.841175008 not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Dec 13 06:49:02 crc restorecon[4764]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes/kubernetes.io~configmap/etcd-service-ca/..2025_02_23_05_23_49.841175008/service-ca.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Dec 13 06:49:02 crc restorecon[4764]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes/kubernetes.io~configmap/etcd-service-ca/service-ca.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Dec 13 06:49:02 crc restorecon[4764]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes/kubernetes.io~configmap/config not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Dec 13 06:49:02 crc restorecon[4764]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.843437178 not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Dec 13 06:49:02 crc restorecon[4764]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.843437178/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Dec 13 06:49:02 crc restorecon[4764]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes/kubernetes.io~configmap/config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Dec 13 06:49:02 crc restorecon[4764]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes/kubernetes.io~configmap/config/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Dec 13 06:49:02 crc restorecon[4764]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Dec 13 06:49:02 crc restorecon[4764]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/containers/etcd-operator/c6c0f2e7 not reset as customized by admin to system_u:object_r:container_file_t:s0:c263,c871 Dec 13 06:49:02 crc restorecon[4764]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/containers/etcd-operator/399edc97 not reset as customized by admin to system_u:object_r:container_file_t:s0:c263,c871 Dec 13 06:49:02 crc restorecon[4764]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/containers/etcd-operator/8049f7cc not reset as customized by admin to system_u:object_r:container_file_t:s0:c263,c871 Dec 13 06:49:02 crc restorecon[4764]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/containers/etcd-operator/0cec5484 not reset as customized by admin to system_u:object_r:container_file_t:s0:c263,c871 Dec 13 06:49:02 crc restorecon[4764]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/containers/etcd-operator/312446d0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c406,c828 Dec 13 06:49:02 crc restorecon[4764]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/containers/etcd-operator/8e56a35d not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Dec 13 06:49:02 crc restorecon[4764]: /var/lib/kubelet/pods/1386a44e-36a2-460c-96d0-0359d2b6f0f5/volumes/kubernetes.io~configmap/config not reset as customized by admin to system_u:object_r:container_file_t:s0:c108,c511 Dec 13 06:49:02 crc restorecon[4764]: /var/lib/kubelet/pods/1386a44e-36a2-460c-96d0-0359d2b6f0f5/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.133159589 not reset as customized by admin to system_u:object_r:container_file_t:s0:c108,c511 Dec 13 06:49:02 crc restorecon[4764]: /var/lib/kubelet/pods/1386a44e-36a2-460c-96d0-0359d2b6f0f5/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.133159589/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c108,c511 Dec 13 06:49:02 crc restorecon[4764]: /var/lib/kubelet/pods/1386a44e-36a2-460c-96d0-0359d2b6f0f5/volumes/kubernetes.io~configmap/config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c108,c511 Dec 13 06:49:02 crc restorecon[4764]: /var/lib/kubelet/pods/1386a44e-36a2-460c-96d0-0359d2b6f0f5/volumes/kubernetes.io~configmap/config/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c108,c511 Dec 13 06:49:02 crc restorecon[4764]: /var/lib/kubelet/pods/1386a44e-36a2-460c-96d0-0359d2b6f0f5/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c108,c511 Dec 13 06:49:02 crc restorecon[4764]: /var/lib/kubelet/pods/1386a44e-36a2-460c-96d0-0359d2b6f0f5/containers/kube-controller-manager-operator/2d30ddb9 not reset as customized by admin to system_u:object_r:container_file_t:s0:c380,c909 Dec 13 06:49:02 crc restorecon[4764]: /var/lib/kubelet/pods/1386a44e-36a2-460c-96d0-0359d2b6f0f5/containers/kube-controller-manager-operator/eca8053d not reset as customized by admin to system_u:object_r:container_file_t:s0:c380,c909 Dec 13 06:49:02 crc restorecon[4764]: /var/lib/kubelet/pods/1386a44e-36a2-460c-96d0-0359d2b6f0f5/containers/kube-controller-manager-operator/c3a25c9a not reset as customized by admin to system_u:object_r:container_file_t:s0:c168,c522 Dec 13 06:49:02 crc restorecon[4764]: /var/lib/kubelet/pods/1386a44e-36a2-460c-96d0-0359d2b6f0f5/containers/kube-controller-manager-operator/b9609c22 not reset as customized by admin to system_u:object_r:container_file_t:s0:c108,c511 Dec 13 06:49:02 crc restorecon[4764]: /var/lib/kubelet/pods/96b93a3a-6083-4aea-8eab-fe1aa8245ad9/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c968,c969 Dec 13 06:49:02 crc restorecon[4764]: /var/lib/kubelet/pods/96b93a3a-6083-4aea-8eab-fe1aa8245ad9/containers/dns-operator/e8b0eca9 not reset as customized by admin to system_u:object_r:container_file_t:s0:c106,c418 Dec 13 06:49:02 crc restorecon[4764]: /var/lib/kubelet/pods/96b93a3a-6083-4aea-8eab-fe1aa8245ad9/containers/dns-operator/b36a9c3f not reset as customized by admin to system_u:object_r:container_file_t:s0:c529,c711 Dec 13 06:49:02 crc restorecon[4764]: /var/lib/kubelet/pods/96b93a3a-6083-4aea-8eab-fe1aa8245ad9/containers/dns-operator/38af7b07 not reset as customized by admin to system_u:object_r:container_file_t:s0:c968,c969 Dec 13 06:49:02 crc restorecon[4764]: /var/lib/kubelet/pods/96b93a3a-6083-4aea-8eab-fe1aa8245ad9/containers/kube-rbac-proxy/ae821620 not reset as customized by admin to system_u:object_r:container_file_t:s0:c106,c418 Dec 13 06:49:02 crc restorecon[4764]: /var/lib/kubelet/pods/96b93a3a-6083-4aea-8eab-fe1aa8245ad9/containers/kube-rbac-proxy/baa23338 not reset as customized by admin to system_u:object_r:container_file_t:s0:c529,c711 Dec 13 06:49:02 crc restorecon[4764]: /var/lib/kubelet/pods/96b93a3a-6083-4aea-8eab-fe1aa8245ad9/containers/kube-rbac-proxy/2c534809 not reset as customized by admin to system_u:object_r:container_file_t:s0:c968,c969 Dec 13 06:49:02 crc restorecon[4764]: /var/lib/kubelet/pods/496e6271-fb68-4057-954e-a0d97a4afa3f/volumes/kubernetes.io~configmap/config not reset as customized by admin to system_u:object_r:container_file_t:s0:c661,c999 Dec 13 06:49:02 crc restorecon[4764]: /var/lib/kubelet/pods/496e6271-fb68-4057-954e-a0d97a4afa3f/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.3532625537 not reset as customized by admin to system_u:object_r:container_file_t:s0:c661,c999 Dec 13 06:49:02 crc restorecon[4764]: /var/lib/kubelet/pods/496e6271-fb68-4057-954e-a0d97a4afa3f/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.3532625537/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c661,c999 Dec 13 06:49:02 crc restorecon[4764]: /var/lib/kubelet/pods/496e6271-fb68-4057-954e-a0d97a4afa3f/volumes/kubernetes.io~configmap/config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c661,c999 Dec 13 06:49:02 crc restorecon[4764]: /var/lib/kubelet/pods/496e6271-fb68-4057-954e-a0d97a4afa3f/volumes/kubernetes.io~configmap/config/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c661,c999 Dec 13 06:49:02 crc restorecon[4764]: /var/lib/kubelet/pods/496e6271-fb68-4057-954e-a0d97a4afa3f/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c661,c999 Dec 13 06:49:02 crc restorecon[4764]: /var/lib/kubelet/pods/496e6271-fb68-4057-954e-a0d97a4afa3f/containers/kube-scheduler-operator-container/59b29eae not reset as customized by admin to system_u:object_r:container_file_t:s0:c338,c381 Dec 13 06:49:02 crc restorecon[4764]: /var/lib/kubelet/pods/496e6271-fb68-4057-954e-a0d97a4afa3f/containers/kube-scheduler-operator-container/c91a8e4f not reset as customized by admin to system_u:object_r:container_file_t:s0:c338,c381 Dec 13 06:49:02 crc restorecon[4764]: /var/lib/kubelet/pods/496e6271-fb68-4057-954e-a0d97a4afa3f/containers/kube-scheduler-operator-container/4d87494a not reset as customized by admin to system_u:object_r:container_file_t:s0:c442,c857 Dec 13 06:49:02 crc restorecon[4764]: /var/lib/kubelet/pods/496e6271-fb68-4057-954e-a0d97a4afa3f/containers/kube-scheduler-operator-container/1e33ca63 not reset as customized by admin to system_u:object_r:container_file_t:s0:c661,c999 Dec 13 06:49:02 crc restorecon[4764]: /var/lib/kubelet/pods/3ab1a177-2de0-46d9-b765-d0d0649bb42e/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Dec 13 06:49:02 crc restorecon[4764]: /var/lib/kubelet/pods/3ab1a177-2de0-46d9-b765-d0d0649bb42e/containers/kube-rbac-proxy/8dea7be2 not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Dec 13 06:49:02 crc restorecon[4764]: /var/lib/kubelet/pods/3ab1a177-2de0-46d9-b765-d0d0649bb42e/containers/kube-rbac-proxy/d0b04a99 not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Dec 13 06:49:02 crc restorecon[4764]: /var/lib/kubelet/pods/3ab1a177-2de0-46d9-b765-d0d0649bb42e/containers/kube-rbac-proxy/d84f01e7 not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Dec 13 06:49:02 crc restorecon[4764]: /var/lib/kubelet/pods/3ab1a177-2de0-46d9-b765-d0d0649bb42e/containers/package-server-manager/4109059b not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Dec 13 06:49:02 crc restorecon[4764]: /var/lib/kubelet/pods/3ab1a177-2de0-46d9-b765-d0d0649bb42e/containers/package-server-manager/a7258a3e not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Dec 13 06:49:02 crc restorecon[4764]: /var/lib/kubelet/pods/3ab1a177-2de0-46d9-b765-d0d0649bb42e/containers/package-server-manager/05bdf2b6 not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Dec 13 06:49:02 crc restorecon[4764]: /var/lib/kubelet/pods/6731426b-95fe-49ff-bb5f-40441049fde2/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Dec 13 06:49:02 crc restorecon[4764]: /var/lib/kubelet/pods/6731426b-95fe-49ff-bb5f-40441049fde2/containers/control-plane-machine-set-operator/f3261b51 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Dec 13 06:49:02 crc restorecon[4764]: /var/lib/kubelet/pods/6731426b-95fe-49ff-bb5f-40441049fde2/containers/control-plane-machine-set-operator/315d045e not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Dec 13 06:49:02 crc restorecon[4764]: /var/lib/kubelet/pods/6731426b-95fe-49ff-bb5f-40441049fde2/containers/control-plane-machine-set-operator/5fdcf278 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Dec 13 06:49:02 crc restorecon[4764]: /var/lib/kubelet/pods/6731426b-95fe-49ff-bb5f-40441049fde2/containers/control-plane-machine-set-operator/d053f757 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Dec 13 06:49:02 crc restorecon[4764]: /var/lib/kubelet/pods/6731426b-95fe-49ff-bb5f-40441049fde2/containers/control-plane-machine-set-operator/c2850dc7 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Dec 13 06:49:02 crc restorecon[4764]: /var/lib/kubelet/pods/b6cd30de-2eeb-49a2-ab40-9167f4560ff5/volumes/kubernetes.io~configmap/marketplace-trusted-ca not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 13 06:49:02 crc restorecon[4764]: /var/lib/kubelet/pods/b6cd30de-2eeb-49a2-ab40-9167f4560ff5/volumes/kubernetes.io~configmap/marketplace-trusted-ca/..2025_02_23_05_22_30.2390596521 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 13 06:49:02 crc restorecon[4764]: /var/lib/kubelet/pods/b6cd30de-2eeb-49a2-ab40-9167f4560ff5/volumes/kubernetes.io~configmap/marketplace-trusted-ca/..2025_02_23_05_22_30.2390596521/tls-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 13 06:49:02 crc restorecon[4764]: /var/lib/kubelet/pods/b6cd30de-2eeb-49a2-ab40-9167f4560ff5/volumes/kubernetes.io~configmap/marketplace-trusted-ca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 13 06:49:02 crc restorecon[4764]: /var/lib/kubelet/pods/b6cd30de-2eeb-49a2-ab40-9167f4560ff5/volumes/kubernetes.io~configmap/marketplace-trusted-ca/tls-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 13 06:49:02 crc restorecon[4764]: /var/lib/kubelet/pods/b6cd30de-2eeb-49a2-ab40-9167f4560ff5/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 13 06:49:02 crc restorecon[4764]: /var/lib/kubelet/pods/b6cd30de-2eeb-49a2-ab40-9167f4560ff5/containers/marketplace-operator/fcfb0b2b not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 13 06:49:02 crc restorecon[4764]: /var/lib/kubelet/pods/b6cd30de-2eeb-49a2-ab40-9167f4560ff5/containers/marketplace-operator/c7ac9b7d not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 13 06:49:02 crc restorecon[4764]: /var/lib/kubelet/pods/b6cd30de-2eeb-49a2-ab40-9167f4560ff5/containers/marketplace-operator/fa0c0d52 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 13 06:49:02 crc restorecon[4764]: /var/lib/kubelet/pods/b6cd30de-2eeb-49a2-ab40-9167f4560ff5/containers/marketplace-operator/c609b6ba not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 13 06:49:02 crc restorecon[4764]: /var/lib/kubelet/pods/b6cd30de-2eeb-49a2-ab40-9167f4560ff5/containers/marketplace-operator/2be6c296 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 13 06:49:02 crc restorecon[4764]: /var/lib/kubelet/pods/b6cd30de-2eeb-49a2-ab40-9167f4560ff5/containers/marketplace-operator/89a32653 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 13 06:49:02 crc restorecon[4764]: /var/lib/kubelet/pods/b6cd30de-2eeb-49a2-ab40-9167f4560ff5/containers/marketplace-operator/4eb9afeb not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 13 06:49:02 crc restorecon[4764]: /var/lib/kubelet/pods/b6cd30de-2eeb-49a2-ab40-9167f4560ff5/containers/marketplace-operator/13af6efa not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 13 06:49:02 crc restorecon[4764]: /var/lib/kubelet/pods/b6312bbd-5731-4ea0-a20f-81d5a57df44a/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Dec 13 06:49:02 crc restorecon[4764]: /var/lib/kubelet/pods/b6312bbd-5731-4ea0-a20f-81d5a57df44a/containers/olm-operator/b03f9724 not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Dec 13 06:49:02 crc restorecon[4764]: /var/lib/kubelet/pods/b6312bbd-5731-4ea0-a20f-81d5a57df44a/containers/olm-operator/e3d105cc not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Dec 13 06:49:02 crc restorecon[4764]: /var/lib/kubelet/pods/b6312bbd-5731-4ea0-a20f-81d5a57df44a/containers/olm-operator/3aed4d83 not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Dec 13 06:49:02 crc restorecon[4764]: /var/lib/kubelet/pods/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c/volumes/kubernetes.io~configmap/config not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c18 Dec 13 06:49:02 crc restorecon[4764]: /var/lib/kubelet/pods/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.1906041176 not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c18 Dec 13 06:49:02 crc restorecon[4764]: /var/lib/kubelet/pods/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.1906041176/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c18 Dec 13 06:49:02 crc restorecon[4764]: /var/lib/kubelet/pods/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c/volumes/kubernetes.io~configmap/config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c18 Dec 13 06:49:02 crc restorecon[4764]: /var/lib/kubelet/pods/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c/volumes/kubernetes.io~configmap/config/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c18 Dec 13 06:49:02 crc restorecon[4764]: /var/lib/kubelet/pods/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c18 Dec 13 06:49:02 crc restorecon[4764]: /var/lib/kubelet/pods/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c/containers/kube-storage-version-migrator-operator/0765fa6e not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c18 Dec 13 06:49:02 crc restorecon[4764]: /var/lib/kubelet/pods/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c/containers/kube-storage-version-migrator-operator/2cefc627 not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c18 Dec 13 06:49:02 crc restorecon[4764]: /var/lib/kubelet/pods/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c/containers/kube-storage-version-migrator-operator/3dcc6345 not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c18 Dec 13 06:49:02 crc restorecon[4764]: /var/lib/kubelet/pods/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c/containers/kube-storage-version-migrator-operator/365af391 not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c18 Dec 13 06:49:02 crc restorecon[4764]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/volumes/kubernetes.io~empty-dir/available-featuregates not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Dec 13 06:49:02 crc restorecon[4764]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/volumes/kubernetes.io~empty-dir/available-featuregates/featureGate-SelfManagedHA-Default.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Dec 13 06:49:02 crc restorecon[4764]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/volumes/kubernetes.io~empty-dir/available-featuregates/featureGate-SelfManagedHA-TechPreviewNoUpgrade.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Dec 13 06:49:02 crc restorecon[4764]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/volumes/kubernetes.io~empty-dir/available-featuregates/featureGate-SelfManagedHA-DevPreviewNoUpgrade.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Dec 13 06:49:02 crc restorecon[4764]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/volumes/kubernetes.io~empty-dir/available-featuregates/featureGate-Hypershift-TechPreviewNoUpgrade.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Dec 13 06:49:02 crc restorecon[4764]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/volumes/kubernetes.io~empty-dir/available-featuregates/featureGate-Hypershift-DevPreviewNoUpgrade.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Dec 13 06:49:02 crc restorecon[4764]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/volumes/kubernetes.io~empty-dir/available-featuregates/featureGate-Hypershift-Default.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Dec 13 06:49:02 crc restorecon[4764]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Dec 13 06:49:02 crc restorecon[4764]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/containers/openshift-api/b1130c0f not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Dec 13 06:49:02 crc restorecon[4764]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/containers/openshift-api/236a5913 not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Dec 13 06:49:02 crc restorecon[4764]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/containers/openshift-api/b9432e26 not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Dec 13 06:49:02 crc restorecon[4764]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/containers/openshift-config-operator/5ddb0e3f not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Dec 13 06:49:02 crc restorecon[4764]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/containers/openshift-config-operator/986dc4fd not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Dec 13 06:49:02 crc restorecon[4764]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/containers/openshift-config-operator/8a23ff9a not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Dec 13 06:49:02 crc restorecon[4764]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/containers/openshift-config-operator/9728ae68 not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Dec 13 06:49:02 crc restorecon[4764]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/containers/openshift-config-operator/665f31d0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Dec 13 06:49:02 crc restorecon[4764]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes/kubernetes.io~configmap/config not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Dec 13 06:49:02 crc restorecon[4764]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.1255385357 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Dec 13 06:49:02 crc restorecon[4764]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.1255385357/operator-config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Dec 13 06:49:02 crc restorecon[4764]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes/kubernetes.io~configmap/config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Dec 13 06:49:02 crc restorecon[4764]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes/kubernetes.io~configmap/config/operator-config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Dec 13 06:49:02 crc restorecon[4764]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes/kubernetes.io~configmap/service-ca-bundle not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Dec 13 06:49:02 crc restorecon[4764]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes/kubernetes.io~configmap/service-ca-bundle/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Dec 13 06:49:02 crc restorecon[4764]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes/kubernetes.io~configmap/service-ca-bundle/..2025_02_23_05_23_57.573792656 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Dec 13 06:49:02 crc restorecon[4764]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes/kubernetes.io~configmap/service-ca-bundle/..2025_02_23_05_23_57.573792656/service-ca.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Dec 13 06:49:02 crc restorecon[4764]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes/kubernetes.io~configmap/service-ca-bundle/service-ca.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Dec 13 06:49:02 crc restorecon[4764]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes/kubernetes.io~configmap/trusted-ca-bundle not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Dec 13 06:49:02 crc restorecon[4764]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes/kubernetes.io~configmap/trusted-ca-bundle/..2025_02_23_05_22_30.3254245399 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Dec 13 06:49:02 crc restorecon[4764]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes/kubernetes.io~configmap/trusted-ca-bundle/..2025_02_23_05_22_30.3254245399/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Dec 13 06:49:02 crc restorecon[4764]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes/kubernetes.io~configmap/trusted-ca-bundle/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Dec 13 06:49:02 crc restorecon[4764]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes/kubernetes.io~configmap/trusted-ca-bundle/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Dec 13 06:49:02 crc restorecon[4764]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Dec 13 06:49:02 crc restorecon[4764]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/containers/authentication-operator/136c9b42 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Dec 13 06:49:02 crc restorecon[4764]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/containers/authentication-operator/98a1575b not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Dec 13 06:49:02 crc restorecon[4764]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/containers/authentication-operator/cac69136 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Dec 13 06:49:02 crc restorecon[4764]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/containers/authentication-operator/5deb77a7 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Dec 13 06:49:02 crc restorecon[4764]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/containers/authentication-operator/2ae53400 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Dec 13 06:49:02 crc restorecon[4764]: /var/lib/kubelet/pods/01ab3dd5-8196-46d0-ad33-122e2ca51def/volumes/kubernetes.io~configmap/config not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c16 Dec 13 06:49:02 crc restorecon[4764]: /var/lib/kubelet/pods/01ab3dd5-8196-46d0-ad33-122e2ca51def/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.3608339744 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c16 Dec 13 06:49:02 crc restorecon[4764]: /var/lib/kubelet/pods/01ab3dd5-8196-46d0-ad33-122e2ca51def/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.3608339744/operator-config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c16 Dec 13 06:49:02 crc restorecon[4764]: /var/lib/kubelet/pods/01ab3dd5-8196-46d0-ad33-122e2ca51def/volumes/kubernetes.io~configmap/config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c16 Dec 13 06:49:02 crc restorecon[4764]: /var/lib/kubelet/pods/01ab3dd5-8196-46d0-ad33-122e2ca51def/volumes/kubernetes.io~configmap/config/operator-config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c16 Dec 13 06:49:02 crc restorecon[4764]: /var/lib/kubelet/pods/01ab3dd5-8196-46d0-ad33-122e2ca51def/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c16 Dec 13 06:49:02 crc restorecon[4764]: /var/lib/kubelet/pods/01ab3dd5-8196-46d0-ad33-122e2ca51def/containers/service-ca-operator/e46f2326 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c16 Dec 13 06:49:02 crc restorecon[4764]: /var/lib/kubelet/pods/01ab3dd5-8196-46d0-ad33-122e2ca51def/containers/service-ca-operator/dc688d3c not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c16 Dec 13 06:49:02 crc restorecon[4764]: /var/lib/kubelet/pods/01ab3dd5-8196-46d0-ad33-122e2ca51def/containers/service-ca-operator/3497c3cd not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c16 Dec 13 06:49:02 crc restorecon[4764]: /var/lib/kubelet/pods/01ab3dd5-8196-46d0-ad33-122e2ca51def/containers/service-ca-operator/177eb008 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c16 Dec 13 06:49:02 crc restorecon[4764]: /var/lib/kubelet/pods/7539238d-5fe0-46ed-884e-1c3b566537ec/volumes/kubernetes.io~configmap/config not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c13 Dec 13 06:49:02 crc restorecon[4764]: /var/lib/kubelet/pods/7539238d-5fe0-46ed-884e-1c3b566537ec/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.3819292994 not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c13 Dec 13 06:49:02 crc restorecon[4764]: /var/lib/kubelet/pods/7539238d-5fe0-46ed-884e-1c3b566537ec/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.3819292994/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c13 Dec 13 06:49:02 crc restorecon[4764]: /var/lib/kubelet/pods/7539238d-5fe0-46ed-884e-1c3b566537ec/volumes/kubernetes.io~configmap/config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c13 Dec 13 06:49:02 crc restorecon[4764]: /var/lib/kubelet/pods/7539238d-5fe0-46ed-884e-1c3b566537ec/volumes/kubernetes.io~configmap/config/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c13 Dec 13 06:49:02 crc restorecon[4764]: /var/lib/kubelet/pods/7539238d-5fe0-46ed-884e-1c3b566537ec/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c13 Dec 13 06:49:02 crc restorecon[4764]: /var/lib/kubelet/pods/7539238d-5fe0-46ed-884e-1c3b566537ec/containers/openshift-apiserver-operator/af5a2afa not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c13 Dec 13 06:49:02 crc restorecon[4764]: /var/lib/kubelet/pods/7539238d-5fe0-46ed-884e-1c3b566537ec/containers/openshift-apiserver-operator/d780cb1f not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c13 Dec 13 06:49:02 crc restorecon[4764]: /var/lib/kubelet/pods/7539238d-5fe0-46ed-884e-1c3b566537ec/containers/openshift-apiserver-operator/49b0f374 not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c13 Dec 13 06:49:02 crc restorecon[4764]: /var/lib/kubelet/pods/7539238d-5fe0-46ed-884e-1c3b566537ec/containers/openshift-apiserver-operator/26fbb125 not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c13 Dec 13 06:49:02 crc restorecon[4764]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/volumes/kubernetes.io~configmap/trusted-ca not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Dec 13 06:49:02 crc restorecon[4764]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/volumes/kubernetes.io~configmap/trusted-ca/..2025_02_23_05_22_30.3244779536 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Dec 13 06:49:02 crc restorecon[4764]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/volumes/kubernetes.io~configmap/trusted-ca/..2025_02_23_05_22_30.3244779536/tls-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Dec 13 06:49:02 crc restorecon[4764]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/volumes/kubernetes.io~configmap/trusted-ca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Dec 13 06:49:02 crc restorecon[4764]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/volumes/kubernetes.io~configmap/trusted-ca/tls-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Dec 13 06:49:02 crc restorecon[4764]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Dec 13 06:49:02 crc restorecon[4764]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/containers/ingress-operator/cf14125a not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Dec 13 06:49:02 crc restorecon[4764]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/containers/ingress-operator/b7f86972 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Dec 13 06:49:02 crc restorecon[4764]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/containers/ingress-operator/e51d739c not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Dec 13 06:49:02 crc restorecon[4764]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/containers/ingress-operator/88ba6a69 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Dec 13 06:49:02 crc restorecon[4764]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/containers/ingress-operator/669a9acf not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Dec 13 06:49:02 crc restorecon[4764]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/containers/ingress-operator/5cd51231 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Dec 13 06:49:02 crc restorecon[4764]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/containers/ingress-operator/75349ec7 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Dec 13 06:49:02 crc restorecon[4764]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/containers/ingress-operator/15c26839 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Dec 13 06:49:02 crc restorecon[4764]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/containers/ingress-operator/45023dcd not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Dec 13 06:49:02 crc restorecon[4764]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/containers/ingress-operator/2bb66a50 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Dec 13 06:49:02 crc restorecon[4764]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/containers/kube-rbac-proxy/64d03bdd not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Dec 13 06:49:02 crc restorecon[4764]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/containers/kube-rbac-proxy/ab8e7ca0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Dec 13 06:49:02 crc restorecon[4764]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/containers/kube-rbac-proxy/bb9be25f not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Dec 13 06:49:02 crc restorecon[4764]: /var/lib/kubelet/pods/bf126b07-da06-4140-9a57-dfd54fc6b486/volumes/kubernetes.io~configmap/trusted-ca not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 13 06:49:02 crc restorecon[4764]: /var/lib/kubelet/pods/bf126b07-da06-4140-9a57-dfd54fc6b486/volumes/kubernetes.io~configmap/trusted-ca/..2025_02_23_05_22_30.2034221258 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 13 06:49:02 crc restorecon[4764]: /var/lib/kubelet/pods/bf126b07-da06-4140-9a57-dfd54fc6b486/volumes/kubernetes.io~configmap/trusted-ca/..2025_02_23_05_22_30.2034221258/tls-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 13 06:49:02 crc restorecon[4764]: /var/lib/kubelet/pods/bf126b07-da06-4140-9a57-dfd54fc6b486/volumes/kubernetes.io~configmap/trusted-ca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 13 06:49:02 crc restorecon[4764]: /var/lib/kubelet/pods/bf126b07-da06-4140-9a57-dfd54fc6b486/volumes/kubernetes.io~configmap/trusted-ca/tls-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 13 06:49:02 crc restorecon[4764]: /var/lib/kubelet/pods/bf126b07-da06-4140-9a57-dfd54fc6b486/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 13 06:49:02 crc restorecon[4764]: /var/lib/kubelet/pods/bf126b07-da06-4140-9a57-dfd54fc6b486/containers/cluster-image-registry-operator/9a0b61d3 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 13 06:49:02 crc restorecon[4764]: /var/lib/kubelet/pods/bf126b07-da06-4140-9a57-dfd54fc6b486/containers/cluster-image-registry-operator/d471b9d2 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 13 06:49:02 crc restorecon[4764]: /var/lib/kubelet/pods/bf126b07-da06-4140-9a57-dfd54fc6b486/containers/cluster-image-registry-operator/8cb76b8e not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 13 06:49:02 crc restorecon[4764]: /var/lib/kubelet/pods/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Dec 13 06:49:02 crc restorecon[4764]: /var/lib/kubelet/pods/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9/containers/catalog-operator/11a00840 not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Dec 13 06:49:02 crc restorecon[4764]: /var/lib/kubelet/pods/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9/containers/catalog-operator/ec355a92 not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Dec 13 06:49:02 crc restorecon[4764]: /var/lib/kubelet/pods/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9/containers/catalog-operator/992f735e not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Dec 13 06:49:02 crc restorecon[4764]: /var/lib/kubelet/pods/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b/volumes/kubernetes.io~configmap/config not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c14 Dec 13 06:49:02 crc restorecon[4764]: /var/lib/kubelet/pods/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.1782968797 not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c14 Dec 13 06:49:02 crc restorecon[4764]: /var/lib/kubelet/pods/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.1782968797/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c14 Dec 13 06:49:02 crc restorecon[4764]: /var/lib/kubelet/pods/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b/volumes/kubernetes.io~configmap/config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c14 Dec 13 06:49:02 crc restorecon[4764]: /var/lib/kubelet/pods/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b/volumes/kubernetes.io~configmap/config/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c14 Dec 13 06:49:02 crc restorecon[4764]: /var/lib/kubelet/pods/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c14 Dec 13 06:49:02 crc restorecon[4764]: /var/lib/kubelet/pods/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b/containers/openshift-controller-manager-operator/d59cdbbc not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c14 Dec 13 06:49:02 crc restorecon[4764]: /var/lib/kubelet/pods/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b/containers/openshift-controller-manager-operator/72133ff0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c14 Dec 13 06:49:02 crc restorecon[4764]: /var/lib/kubelet/pods/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b/containers/openshift-controller-manager-operator/c56c834c not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c14 Dec 13 06:49:02 crc restorecon[4764]: /var/lib/kubelet/pods/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b/containers/openshift-controller-manager-operator/d13724c7 not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c14 Dec 13 06:49:02 crc restorecon[4764]: /var/lib/kubelet/pods/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b/containers/openshift-controller-manager-operator/0a498258 not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c14 Dec 13 06:49:02 crc restorecon[4764]: /var/lib/kubelet/pods/5fe579f8-e8a6-4643-bce5-a661393c4dde/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Dec 13 06:49:02 crc restorecon[4764]: /var/lib/kubelet/pods/5fe579f8-e8a6-4643-bce5-a661393c4dde/containers/machine-config-server/fa471982 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Dec 13 06:49:02 crc restorecon[4764]: /var/lib/kubelet/pods/5fe579f8-e8a6-4643-bce5-a661393c4dde/containers/machine-config-server/fc900d92 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Dec 13 06:49:02 crc restorecon[4764]: /var/lib/kubelet/pods/5fe579f8-e8a6-4643-bce5-a661393c4dde/containers/machine-config-server/fa7d68da not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Dec 13 06:49:02 crc restorecon[4764]: /var/lib/kubelet/pods/49ef4625-1d3a-4a9f-b595-c2433d32326d/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c22 Dec 13 06:49:02 crc restorecon[4764]: /var/lib/kubelet/pods/49ef4625-1d3a-4a9f-b595-c2433d32326d/containers/migrator/4bacf9b4 not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c22 Dec 13 06:49:02 crc restorecon[4764]: /var/lib/kubelet/pods/49ef4625-1d3a-4a9f-b595-c2433d32326d/containers/migrator/424021b1 not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c22 Dec 13 06:49:02 crc restorecon[4764]: /var/lib/kubelet/pods/49ef4625-1d3a-4a9f-b595-c2433d32326d/containers/migrator/fc2e31a3 not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c22 Dec 13 06:49:02 crc restorecon[4764]: /var/lib/kubelet/pods/49ef4625-1d3a-4a9f-b595-c2433d32326d/containers/graceful-termination/f51eefac not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c22 Dec 13 06:49:02 crc restorecon[4764]: /var/lib/kubelet/pods/49ef4625-1d3a-4a9f-b595-c2433d32326d/containers/graceful-termination/c8997f2f not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c22 Dec 13 06:49:02 crc restorecon[4764]: /var/lib/kubelet/pods/49ef4625-1d3a-4a9f-b595-c2433d32326d/containers/graceful-termination/7481f599 not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c22 Dec 13 06:49:02 crc restorecon[4764]: /var/lib/kubelet/pods/25e176fe-21b4-4974-b1ed-c8b94f112a7f/volumes/kubernetes.io~configmap/signing-cabundle not reset as customized by admin to system_u:object_r:container_file_t:s0:c19,c22 Dec 13 06:49:02 crc restorecon[4764]: /var/lib/kubelet/pods/25e176fe-21b4-4974-b1ed-c8b94f112a7f/volumes/kubernetes.io~configmap/signing-cabundle/..2025_02_23_05_22_49.2255460704 not reset as customized by admin to system_u:object_r:container_file_t:s0:c19,c22 Dec 13 06:49:02 crc restorecon[4764]: /var/lib/kubelet/pods/25e176fe-21b4-4974-b1ed-c8b94f112a7f/volumes/kubernetes.io~configmap/signing-cabundle/..2025_02_23_05_22_49.2255460704/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c19,c22 Dec 13 06:49:02 crc restorecon[4764]: /var/lib/kubelet/pods/25e176fe-21b4-4974-b1ed-c8b94f112a7f/volumes/kubernetes.io~configmap/signing-cabundle/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c19,c22 Dec 13 06:49:02 crc restorecon[4764]: /var/lib/kubelet/pods/25e176fe-21b4-4974-b1ed-c8b94f112a7f/volumes/kubernetes.io~configmap/signing-cabundle/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c19,c22 Dec 13 06:49:02 crc restorecon[4764]: /var/lib/kubelet/pods/25e176fe-21b4-4974-b1ed-c8b94f112a7f/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c19,c22 Dec 13 06:49:02 crc restorecon[4764]: /var/lib/kubelet/pods/25e176fe-21b4-4974-b1ed-c8b94f112a7f/containers/service-ca-controller/fdafea19 not reset as customized by admin to system_u:object_r:container_file_t:s0:c19,c22 Dec 13 06:49:02 crc restorecon[4764]: /var/lib/kubelet/pods/25e176fe-21b4-4974-b1ed-c8b94f112a7f/containers/service-ca-controller/d0e1c571 not reset as customized by admin to system_u:object_r:container_file_t:s0:c19,c22 Dec 13 06:49:02 crc restorecon[4764]: /var/lib/kubelet/pods/25e176fe-21b4-4974-b1ed-c8b94f112a7f/containers/service-ca-controller/ee398915 not reset as customized by admin to system_u:object_r:container_file_t:s0:c19,c22 Dec 13 06:49:02 crc restorecon[4764]: /var/lib/kubelet/pods/25e176fe-21b4-4974-b1ed-c8b94f112a7f/containers/service-ca-controller/682bb6b8 not reset as customized by admin to system_u:object_r:container_file_t:s0:c19,c22 Dec 13 06:49:02 crc restorecon[4764]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c666,c920 Dec 13 06:49:02 crc restorecon[4764]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/setup/a3e67855 not reset as customized by admin to system_u:object_r:container_file_t:s0:c294,c884 Dec 13 06:49:02 crc restorecon[4764]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/setup/a989f289 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c1016 Dec 13 06:49:02 crc restorecon[4764]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/setup/915431bd not reset as customized by admin to system_u:object_r:container_file_t:s0:c666,c920 Dec 13 06:49:02 crc restorecon[4764]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd-ensure-env-vars/7796fdab not reset as customized by admin to system_u:object_r:container_file_t:s0:c294,c884 Dec 13 06:49:02 crc restorecon[4764]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd-ensure-env-vars/dcdb5f19 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c1016 Dec 13 06:49:02 crc restorecon[4764]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd-ensure-env-vars/a3aaa88c not reset as customized by admin to system_u:object_r:container_file_t:s0:c666,c920 Dec 13 06:49:02 crc restorecon[4764]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd-resources-copy/5508e3e6 not reset as customized by admin to system_u:object_r:container_file_t:s0:c294,c884 Dec 13 06:49:02 crc restorecon[4764]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd-resources-copy/160585de not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c1016 Dec 13 06:49:02 crc restorecon[4764]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd-resources-copy/e99f8da3 not reset as customized by admin to system_u:object_r:container_file_t:s0:c666,c920 Dec 13 06:49:02 crc restorecon[4764]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcdctl/8bc85570 not reset as customized by admin to system_u:object_r:container_file_t:s0:c294,c884 Dec 13 06:49:02 crc restorecon[4764]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcdctl/a5861c91 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c1016 Dec 13 06:49:02 crc restorecon[4764]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcdctl/84db1135 not reset as customized by admin to system_u:object_r:container_file_t:s0:c666,c920 Dec 13 06:49:02 crc restorecon[4764]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd/9e1a6043 not reset as customized by admin to system_u:object_r:container_file_t:s0:c294,c884 Dec 13 06:49:02 crc restorecon[4764]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd/c1aba1c2 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c1016 Dec 13 06:49:02 crc restorecon[4764]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd/d55ccd6d not reset as customized by admin to system_u:object_r:container_file_t:s0:c666,c920 Dec 13 06:49:02 crc restorecon[4764]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd-metrics/971cc9f6 not reset as customized by admin to system_u:object_r:container_file_t:s0:c294,c884 Dec 13 06:49:02 crc restorecon[4764]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd-metrics/8f2e3dcf not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c1016 Dec 13 06:49:02 crc restorecon[4764]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd-metrics/ceb35e9c not reset as customized by admin to system_u:object_r:container_file_t:s0:c666,c920 Dec 13 06:49:02 crc restorecon[4764]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd-readyz/1c192745 not reset as customized by admin to system_u:object_r:container_file_t:s0:c294,c884 Dec 13 06:49:02 crc restorecon[4764]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd-readyz/5209e501 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c1016 Dec 13 06:49:02 crc restorecon[4764]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd-readyz/f83de4df not reset as customized by admin to system_u:object_r:container_file_t:s0:c666,c920 Dec 13 06:49:02 crc restorecon[4764]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd-rev/e7b978ac not reset as customized by admin to system_u:object_r:container_file_t:s0:c294,c884 Dec 13 06:49:02 crc restorecon[4764]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd-rev/c64304a1 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c1016 Dec 13 06:49:02 crc restorecon[4764]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd-rev/5384386b not reset as customized by admin to system_u:object_r:container_file_t:s0:c666,c920 Dec 13 06:49:02 crc restorecon[4764]: /var/lib/kubelet/pods/efdd0498-1daa-4136-9a4a-3b948c2293fc/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c268,c620 Dec 13 06:49:02 crc restorecon[4764]: /var/lib/kubelet/pods/efdd0498-1daa-4136-9a4a-3b948c2293fc/containers/multus-admission-controller/cce3e3ff not reset as customized by admin to system_u:object_r:container_file_t:s0:c435,c756 Dec 13 06:49:02 crc restorecon[4764]: /var/lib/kubelet/pods/efdd0498-1daa-4136-9a4a-3b948c2293fc/containers/multus-admission-controller/8fb75465 not reset as customized by admin to system_u:object_r:container_file_t:s0:c268,c620 Dec 13 06:49:02 crc restorecon[4764]: /var/lib/kubelet/pods/efdd0498-1daa-4136-9a4a-3b948c2293fc/containers/kube-rbac-proxy/740f573e not reset as customized by admin to system_u:object_r:container_file_t:s0:c435,c756 Dec 13 06:49:02 crc restorecon[4764]: /var/lib/kubelet/pods/efdd0498-1daa-4136-9a4a-3b948c2293fc/containers/kube-rbac-proxy/32fd1134 not reset as customized by admin to system_u:object_r:container_file_t:s0:c268,c620 Dec 13 06:49:02 crc restorecon[4764]: /var/lib/kubelet/pods/20b0d48f-5fd6-431c-a545-e3c800c7b866/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c19,c24 Dec 13 06:49:02 crc restorecon[4764]: /var/lib/kubelet/pods/20b0d48f-5fd6-431c-a545-e3c800c7b866/containers/serve-healthcheck-canary/0a861bd3 not reset as customized by admin to system_u:object_r:container_file_t:s0:c19,c24 Dec 13 06:49:02 crc restorecon[4764]: /var/lib/kubelet/pods/20b0d48f-5fd6-431c-a545-e3c800c7b866/containers/serve-healthcheck-canary/80363026 not reset as customized by admin to system_u:object_r:container_file_t:s0:c19,c24 Dec 13 06:49:02 crc restorecon[4764]: /var/lib/kubelet/pods/20b0d48f-5fd6-431c-a545-e3c800c7b866/containers/serve-healthcheck-canary/bfa952a8 not reset as customized by admin to system_u:object_r:container_file_t:s0:c19,c24 Dec 13 06:49:02 crc restorecon[4764]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/volumes/kubernetes.io~configmap/auth-proxy-config not reset as customized by admin to system_u:object_r:container_file_t:s0:c129,c158 Dec 13 06:49:02 crc restorecon[4764]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/volumes/kubernetes.io~configmap/auth-proxy-config/..2025_02_23_05_33_31.2122464563 not reset as customized by admin to system_u:object_r:container_file_t:s0:c129,c158 Dec 13 06:49:02 crc restorecon[4764]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/volumes/kubernetes.io~configmap/auth-proxy-config/..2025_02_23_05_33_31.2122464563/config-file.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c129,c158 Dec 13 06:49:02 crc restorecon[4764]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/volumes/kubernetes.io~configmap/auth-proxy-config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c129,c158 Dec 13 06:49:02 crc restorecon[4764]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/volumes/kubernetes.io~configmap/auth-proxy-config/config-file.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c129,c158 Dec 13 06:49:02 crc restorecon[4764]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/volumes/kubernetes.io~configmap/config not reset as customized by admin to system_u:object_r:container_file_t:s0:c129,c158 Dec 13 06:49:02 crc restorecon[4764]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/volumes/kubernetes.io~configmap/config/..2025_02_23_05_33_31.333075221 not reset as customized by admin to system_u:object_r:container_file_t:s0:c129,c158 Dec 13 06:49:02 crc restorecon[4764]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/volumes/kubernetes.io~configmap/config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c129,c158 Dec 13 06:49:02 crc restorecon[4764]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c129,c158 Dec 13 06:49:02 crc restorecon[4764]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/containers/kube-rbac-proxy/793bf43d not reset as customized by admin to system_u:object_r:container_file_t:s0:c381,c387 Dec 13 06:49:02 crc restorecon[4764]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/containers/kube-rbac-proxy/7db1bb6e not reset as customized by admin to system_u:object_r:container_file_t:s0:c142,c438 Dec 13 06:49:02 crc restorecon[4764]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/containers/kube-rbac-proxy/4f6a0368 not reset as customized by admin to system_u:object_r:container_file_t:s0:c129,c158 Dec 13 06:49:02 crc restorecon[4764]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/containers/machine-approver-controller/c12c7d86 not reset as customized by admin to system_u:object_r:container_file_t:s0:c381,c387 Dec 13 06:49:02 crc restorecon[4764]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/containers/machine-approver-controller/36c4a773 not reset as customized by admin to system_u:object_r:container_file_t:s0:c142,c438 Dec 13 06:49:02 crc restorecon[4764]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/containers/machine-approver-controller/4c1e98ae not reset as customized by admin to system_u:object_r:container_file_t:s0:c142,c438 Dec 13 06:49:02 crc restorecon[4764]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/containers/machine-approver-controller/a4c8115c not reset as customized by admin to system_u:object_r:container_file_t:s0:c129,c158 Dec 13 06:49:02 crc restorecon[4764]: /var/lib/kubelet/pods/f4b27818a5e8e43d0dc095d08835c792/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c97,c980 Dec 13 06:49:02 crc restorecon[4764]: /var/lib/kubelet/pods/f4b27818a5e8e43d0dc095d08835c792/containers/setup/7db1802e not reset as customized by admin to system_u:object_r:container_file_t:s0:c97,c980 Dec 13 06:49:02 crc restorecon[4764]: /var/lib/kubelet/pods/f4b27818a5e8e43d0dc095d08835c792/containers/kube-apiserver/a008a7ab not reset as customized by admin to system_u:object_r:container_file_t:s0:c97,c980 Dec 13 06:49:02 crc restorecon[4764]: /var/lib/kubelet/pods/f4b27818a5e8e43d0dc095d08835c792/containers/kube-apiserver-cert-syncer/2c836bac not reset as customized by admin to system_u:object_r:container_file_t:s0:c97,c980 Dec 13 06:49:02 crc restorecon[4764]: /var/lib/kubelet/pods/f4b27818a5e8e43d0dc095d08835c792/containers/kube-apiserver-cert-regeneration-controller/0ce62299 not reset as customized by admin to system_u:object_r:container_file_t:s0:c97,c980 Dec 13 06:49:02 crc restorecon[4764]: /var/lib/kubelet/pods/f4b27818a5e8e43d0dc095d08835c792/containers/kube-apiserver-insecure-readyz/945d2457 not reset as customized by admin to system_u:object_r:container_file_t:s0:c97,c980 Dec 13 06:49:02 crc restorecon[4764]: /var/lib/kubelet/pods/f4b27818a5e8e43d0dc095d08835c792/containers/kube-apiserver-check-endpoints/7d5c1dd8 not reset as customized by admin to system_u:object_r:container_file_t:s0:c97,c980 Dec 13 06:49:02 crc restorecon[4764]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/utilities not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 13 06:49:02 crc restorecon[4764]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/utilities/copy-content not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 13 06:49:02 crc restorecon[4764]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 13 06:49:02 crc restorecon[4764]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 13 06:49:02 crc restorecon[4764]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/3scale-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 13 06:49:02 crc restorecon[4764]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/3scale-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 13 06:49:02 crc restorecon[4764]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/advanced-cluster-management not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 13 06:49:02 crc restorecon[4764]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/advanced-cluster-management/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 13 06:49:02 crc restorecon[4764]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/amq-broker-rhel8 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 13 06:49:02 crc restorecon[4764]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/amq-broker-rhel8/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 13 06:49:02 crc restorecon[4764]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/amq-online not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 13 06:49:02 crc restorecon[4764]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/amq-online/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 13 06:49:02 crc restorecon[4764]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/amq-streams not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 13 06:49:02 crc restorecon[4764]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/amq-streams/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 13 06:49:02 crc restorecon[4764]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/amq-streams-console not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 13 06:49:02 crc restorecon[4764]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/amq-streams-console/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 13 06:49:02 crc restorecon[4764]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/amq7-interconnect-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 13 06:49:02 crc restorecon[4764]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/amq7-interconnect-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 13 06:49:02 crc restorecon[4764]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ansible-automation-platform-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 13 06:49:02 crc restorecon[4764]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ansible-automation-platform-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 13 06:49:02 crc restorecon[4764]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ansible-cloud-addons-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 13 06:49:02 crc restorecon[4764]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ansible-cloud-addons-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 13 06:49:02 crc restorecon[4764]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/apicast-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 13 06:49:02 crc restorecon[4764]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/apicast-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 13 06:49:02 crc restorecon[4764]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/apicurio-registry-3 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 13 06:49:02 crc restorecon[4764]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/apicurio-registry-3/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 13 06:49:02 crc restorecon[4764]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/authorino-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 13 06:49:02 crc restorecon[4764]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/authorino-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 13 06:49:02 crc restorecon[4764]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aws-load-balancer-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 13 06:49:02 crc restorecon[4764]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aws-load-balancer-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 13 06:49:02 crc restorecon[4764]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/bamoe-businessautomation-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 13 06:49:02 crc restorecon[4764]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/bamoe-businessautomation-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 13 06:49:02 crc restorecon[4764]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/bamoe-kogito-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 13 06:49:02 crc restorecon[4764]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/bamoe-kogito-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 13 06:49:02 crc restorecon[4764]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/bpfman-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 13 06:49:02 crc restorecon[4764]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/bpfman-operator/index.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 13 06:49:02 crc restorecon[4764]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/businessautomation-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 13 06:49:02 crc restorecon[4764]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/businessautomation-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 13 06:49:02 crc restorecon[4764]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cephcsi-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 13 06:49:02 crc restorecon[4764]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cephcsi-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 13 06:49:02 crc restorecon[4764]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cincinnati-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 13 06:49:02 crc restorecon[4764]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cincinnati-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 13 06:49:02 crc restorecon[4764]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cluster-kube-descheduler-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 13 06:49:02 crc restorecon[4764]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cluster-kube-descheduler-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 13 06:49:02 crc restorecon[4764]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cluster-logging not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 13 06:49:02 crc restorecon[4764]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cluster-logging/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 13 06:49:02 crc restorecon[4764]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cluster-observability-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 13 06:49:02 crc restorecon[4764]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cluster-observability-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 13 06:49:02 crc restorecon[4764]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/compliance-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 13 06:49:02 crc restorecon[4764]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/compliance-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 13 06:49:02 crc restorecon[4764]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/container-security-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 13 06:49:02 crc restorecon[4764]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/container-security-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 13 06:49:02 crc restorecon[4764]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/costmanagement-metrics-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 13 06:49:02 crc restorecon[4764]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/costmanagement-metrics-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 13 06:49:02 crc restorecon[4764]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cryostat-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 13 06:49:02 crc restorecon[4764]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cryostat-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 13 06:49:02 crc restorecon[4764]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/datagrid not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 13 06:49:02 crc restorecon[4764]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/datagrid/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 13 06:49:02 crc restorecon[4764]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/devspaces not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 13 06:49:02 crc restorecon[4764]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/devspaces/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 13 06:49:02 crc restorecon[4764]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/devworkspace-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 13 06:49:02 crc restorecon[4764]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/devworkspace-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 13 06:49:02 crc restorecon[4764]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dpu-network-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 13 06:49:02 crc restorecon[4764]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dpu-network-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 13 06:49:02 crc restorecon[4764]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/eap not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 13 06:49:02 crc restorecon[4764]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/eap/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 13 06:49:02 crc restorecon[4764]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/elasticsearch-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 13 06:49:02 crc restorecon[4764]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/elasticsearch-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 13 06:49:02 crc restorecon[4764]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/external-dns-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 13 06:49:02 crc restorecon[4764]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/external-dns-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 13 06:49:02 crc restorecon[4764]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/fence-agents-remediation not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 13 06:49:02 crc restorecon[4764]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/fence-agents-remediation/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 13 06:49:02 crc restorecon[4764]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/file-integrity-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 13 06:49:02 crc restorecon[4764]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/file-integrity-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 13 06:49:02 crc restorecon[4764]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/fuse-apicurito not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 13 06:49:02 crc restorecon[4764]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/fuse-apicurito/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 13 06:49:02 crc restorecon[4764]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/fuse-console not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 13 06:49:02 crc restorecon[4764]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/fuse-console/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 13 06:49:02 crc restorecon[4764]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/fuse-online not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 13 06:49:02 crc restorecon[4764]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/fuse-online/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 13 06:49:02 crc restorecon[4764]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/gatekeeper-operator-product not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 13 06:49:02 crc restorecon[4764]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/gatekeeper-operator-product/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 13 06:49:02 crc restorecon[4764]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/jaeger-product not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 13 06:49:02 crc restorecon[4764]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/jaeger-product/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 13 06:49:02 crc restorecon[4764]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/jws-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 13 06:49:02 crc restorecon[4764]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/jws-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 13 06:49:02 crc restorecon[4764]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kernel-module-management not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 13 06:49:02 crc restorecon[4764]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kernel-module-management/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 13 06:49:02 crc restorecon[4764]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kernel-module-management-hub not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 13 06:49:02 crc restorecon[4764]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kernel-module-management-hub/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 13 06:49:02 crc restorecon[4764]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kiali-ossm not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 13 06:49:02 crc restorecon[4764]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kiali-ossm/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 13 06:49:02 crc restorecon[4764]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubevirt-hyperconverged not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 13 06:49:02 crc restorecon[4764]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubevirt-hyperconverged/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 13 06:49:02 crc restorecon[4764]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/logic-operator-rhel8 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 13 06:49:02 crc restorecon[4764]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/logic-operator-rhel8/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 13 06:49:02 crc restorecon[4764]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/loki-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 13 06:49:02 crc restorecon[4764]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/loki-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 13 06:49:02 crc restorecon[4764]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/lvms-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 13 06:49:02 crc restorecon[4764]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/lvms-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 13 06:49:02 crc restorecon[4764]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/machine-deletion-remediation not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 13 06:49:02 crc restorecon[4764]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/machine-deletion-remediation/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 13 06:49:02 crc restorecon[4764]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mcg-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 13 06:49:02 crc restorecon[4764]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mcg-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 13 06:49:02 crc restorecon[4764]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mta-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 13 06:49:02 crc restorecon[4764]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mta-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 13 06:49:02 crc restorecon[4764]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mtc-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 13 06:49:02 crc restorecon[4764]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mtc-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 13 06:49:02 crc restorecon[4764]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mtr-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 13 06:49:02 crc restorecon[4764]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mtr-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 13 06:49:02 crc restorecon[4764]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mtv-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 13 06:49:02 crc restorecon[4764]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mtv-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 13 06:49:02 crc restorecon[4764]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/multicluster-engine not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 13 06:49:02 crc restorecon[4764]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/multicluster-engine/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 13 06:49:02 crc restorecon[4764]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/netobserv-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 13 06:49:02 crc restorecon[4764]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/netobserv-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 13 06:49:02 crc restorecon[4764]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/node-healthcheck-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 13 06:49:02 crc restorecon[4764]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/node-healthcheck-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 13 06:49:02 crc restorecon[4764]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/node-maintenance-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 13 06:49:02 crc restorecon[4764]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/node-maintenance-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 13 06:49:02 crc restorecon[4764]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/node-observability-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 13 06:49:02 crc restorecon[4764]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/node-observability-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 13 06:49:02 crc restorecon[4764]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ocs-client-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 13 06:49:02 crc restorecon[4764]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ocs-client-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 13 06:49:02 crc restorecon[4764]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ocs-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 13 06:49:02 crc restorecon[4764]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ocs-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 13 06:49:02 crc restorecon[4764]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odf-csi-addons-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 13 06:49:02 crc restorecon[4764]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odf-csi-addons-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 13 06:49:02 crc restorecon[4764]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odf-multicluster-orchestrator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 13 06:49:02 crc restorecon[4764]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odf-multicluster-orchestrator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 13 06:49:02 crc restorecon[4764]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odf-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 13 06:49:02 crc restorecon[4764]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odf-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 13 06:49:02 crc restorecon[4764]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odf-prometheus-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 13 06:49:02 crc restorecon[4764]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odf-prometheus-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 13 06:49:02 crc restorecon[4764]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odr-cluster-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 13 06:49:02 crc restorecon[4764]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odr-cluster-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 13 06:49:02 crc restorecon[4764]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odr-hub-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 13 06:49:02 crc restorecon[4764]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odr-hub-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 13 06:49:02 crc restorecon[4764]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-cert-manager-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 13 06:49:02 crc restorecon[4764]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-cert-manager-operator/bundle-v1.15.0.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 13 06:49:02 crc restorecon[4764]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-cert-manager-operator/channel.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 13 06:49:02 crc restorecon[4764]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-cert-manager-operator/package.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 13 06:49:02 crc restorecon[4764]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-custom-metrics-autoscaler-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 13 06:49:02 crc restorecon[4764]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-custom-metrics-autoscaler-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 13 06:49:02 crc restorecon[4764]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-gitops-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 13 06:49:02 crc restorecon[4764]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-gitops-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 13 06:49:02 crc restorecon[4764]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-pipelines-operator-rh not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 13 06:49:02 crc restorecon[4764]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-pipelines-operator-rh/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 13 06:49:02 crc restorecon[4764]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-secondary-scheduler-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 13 06:49:02 crc restorecon[4764]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-secondary-scheduler-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 13 06:49:02 crc restorecon[4764]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/opentelemetry-product not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 13 06:49:02 crc restorecon[4764]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/opentelemetry-product/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 13 06:49:02 crc restorecon[4764]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/quay-bridge-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 13 06:49:02 crc restorecon[4764]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/quay-bridge-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 13 06:49:02 crc restorecon[4764]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/quay-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 13 06:49:02 crc restorecon[4764]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/quay-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 13 06:49:02 crc restorecon[4764]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/recipe not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 13 06:49:02 crc restorecon[4764]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/recipe/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 13 06:49:02 crc restorecon[4764]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/red-hat-camel-k not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 13 06:49:02 crc restorecon[4764]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/red-hat-camel-k/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 13 06:49:02 crc restorecon[4764]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/red-hat-hawtio-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 13 06:49:02 crc restorecon[4764]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/red-hat-hawtio-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 13 06:49:02 crc restorecon[4764]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/redhat-oadp-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 13 06:49:02 crc restorecon[4764]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/redhat-oadp-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 13 06:49:02 crc restorecon[4764]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rh-service-binding-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 13 06:49:02 crc restorecon[4764]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rh-service-binding-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 13 06:49:02 crc restorecon[4764]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhacs-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 13 06:49:02 crc restorecon[4764]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhacs-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 13 06:49:02 crc restorecon[4764]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhbk-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 13 06:49:02 crc restorecon[4764]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhbk-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 13 06:49:02 crc restorecon[4764]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhdh not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 13 06:49:02 crc restorecon[4764]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhdh/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 13 06:49:02 crc restorecon[4764]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhods-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 13 06:49:02 crc restorecon[4764]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhods-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 13 06:49:02 crc restorecon[4764]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhods-prometheus-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 13 06:49:02 crc restorecon[4764]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhods-prometheus-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 13 06:49:02 crc restorecon[4764]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhpam-kogito-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 13 06:49:02 crc restorecon[4764]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhpam-kogito-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 13 06:49:02 crc restorecon[4764]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhsso-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 13 06:49:02 crc restorecon[4764]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhsso-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 13 06:49:02 crc restorecon[4764]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rook-ceph-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 13 06:49:02 crc restorecon[4764]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rook-ceph-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 13 06:49:02 crc restorecon[4764]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/run-once-duration-override-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 13 06:49:02 crc restorecon[4764]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/run-once-duration-override-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 13 06:49:02 crc restorecon[4764]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sandboxed-containers-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 13 06:49:02 crc restorecon[4764]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sandboxed-containers-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 13 06:49:02 crc restorecon[4764]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/security-profiles-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 13 06:49:02 crc restorecon[4764]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/security-profiles-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 13 06:49:02 crc restorecon[4764]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/self-node-remediation not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 13 06:49:02 crc restorecon[4764]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/self-node-remediation/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 13 06:49:02 crc restorecon[4764]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/serverless-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 13 06:49:02 crc restorecon[4764]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/serverless-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 13 06:49:02 crc restorecon[4764]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/service-registry-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 13 06:49:02 crc restorecon[4764]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/service-registry-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 13 06:49:02 crc restorecon[4764]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/servicemeshoperator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 13 06:49:02 crc restorecon[4764]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/servicemeshoperator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 13 06:49:02 crc restorecon[4764]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/servicemeshoperator3 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 13 06:49:02 crc restorecon[4764]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/servicemeshoperator3/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 13 06:49:02 crc restorecon[4764]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/skupper-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 13 06:49:02 crc restorecon[4764]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/skupper-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 13 06:49:02 crc restorecon[4764]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/submariner not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 13 06:49:02 crc restorecon[4764]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/submariner/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 13 06:49:02 crc restorecon[4764]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tang-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 13 06:49:02 crc restorecon[4764]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tang-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 13 06:49:02 crc restorecon[4764]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tempo-product not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 13 06:49:02 crc restorecon[4764]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tempo-product/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 13 06:49:02 crc restorecon[4764]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/trustee-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 13 06:49:02 crc restorecon[4764]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/trustee-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 13 06:49:02 crc restorecon[4764]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/volsync-product not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 13 06:49:02 crc restorecon[4764]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/volsync-product/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 13 06:49:02 crc restorecon[4764]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/web-terminal not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 13 06:49:02 crc restorecon[4764]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/web-terminal/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 13 06:49:02 crc restorecon[4764]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/cache not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 13 06:49:02 crc restorecon[4764]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 13 06:49:02 crc restorecon[4764]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 13 06:49:02 crc restorecon[4764]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/00000-1.psg not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 13 06:49:02 crc restorecon[4764]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/00000-1.psg.pmt not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 13 06:49:02 crc restorecon[4764]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/db.pmt not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 13 06:49:02 crc restorecon[4764]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/index.pmt not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 13 06:49:02 crc restorecon[4764]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/main.pix not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 13 06:49:02 crc restorecon[4764]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/overflow.pix not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 13 06:49:02 crc restorecon[4764]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/digest not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 13 06:49:02 crc restorecon[4764]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 13 06:49:02 crc restorecon[4764]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/containers/extract-utilities/bc8d0691 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 13 06:49:02 crc restorecon[4764]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/containers/extract-utilities/6b76097a not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 13 06:49:02 crc restorecon[4764]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/containers/extract-utilities/34d1af30 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 13 06:49:02 crc restorecon[4764]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/containers/extract-content/312ba61c not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 13 06:49:02 crc restorecon[4764]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/containers/extract-content/645d5dd1 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 13 06:49:02 crc restorecon[4764]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/containers/extract-content/16e825f0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 13 06:49:02 crc restorecon[4764]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/containers/registry-server/4cf51fc9 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 13 06:49:02 crc restorecon[4764]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/containers/registry-server/2a23d348 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 13 06:49:02 crc restorecon[4764]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/containers/registry-server/075dbd49 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 13 06:49:02 crc restorecon[4764]: /var/lib/kubelet/pods/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59/volumes/kubernetes.io~configmap/serviceca not reset as customized by admin to system_u:object_r:container_file_t:s0:c842,c986 Dec 13 06:49:02 crc restorecon[4764]: /var/lib/kubelet/pods/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59/volumes/kubernetes.io~configmap/serviceca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c842,c986 Dec 13 06:49:02 crc restorecon[4764]: /var/lib/kubelet/pods/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59/volumes/kubernetes.io~configmap/serviceca/..2025_02_24_06_09_13.3521195566 not reset as customized by admin to system_u:object_r:container_file_t:s0:c842,c986 Dec 13 06:49:02 crc restorecon[4764]: /var/lib/kubelet/pods/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59/volumes/kubernetes.io~configmap/serviceca/..2025_02_24_06_09_13.3521195566/image-registry.openshift-image-registry.svc..5000 not reset as customized by admin to system_u:object_r:container_file_t:s0:c842,c986 Dec 13 06:49:02 crc restorecon[4764]: /var/lib/kubelet/pods/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59/volumes/kubernetes.io~configmap/serviceca/..2025_02_24_06_09_13.3521195566/image-registry.openshift-image-registry.svc.cluster.local..5000 not reset as customized by admin to system_u:object_r:container_file_t:s0:c842,c986 Dec 13 06:49:02 crc restorecon[4764]: /var/lib/kubelet/pods/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59/volumes/kubernetes.io~configmap/serviceca/..2025_02_24_06_09_13.3521195566/default-route-openshift-image-registry.apps-crc.testing not reset as customized by admin to system_u:object_r:container_file_t:s0:c842,c986 Dec 13 06:49:02 crc restorecon[4764]: /var/lib/kubelet/pods/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59/volumes/kubernetes.io~configmap/serviceca/default-route-openshift-image-registry.apps-crc.testing not reset as customized by admin to system_u:object_r:container_file_t:s0:c842,c986 Dec 13 06:49:02 crc restorecon[4764]: /var/lib/kubelet/pods/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59/volumes/kubernetes.io~configmap/serviceca/image-registry.openshift-image-registry.svc..5000 not reset as customized by admin to system_u:object_r:container_file_t:s0:c842,c986 Dec 13 06:49:02 crc restorecon[4764]: /var/lib/kubelet/pods/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59/volumes/kubernetes.io~configmap/serviceca/image-registry.openshift-image-registry.svc.cluster.local..5000 not reset as customized by admin to system_u:object_r:container_file_t:s0:c842,c986 Dec 13 06:49:02 crc restorecon[4764]: /var/lib/kubelet/pods/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c842,c986 Dec 13 06:49:02 crc restorecon[4764]: /var/lib/kubelet/pods/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59/containers/node-ca/dd585ddd not reset as customized by admin to system_u:object_r:container_file_t:s0:c377,c642 Dec 13 06:49:02 crc restorecon[4764]: /var/lib/kubelet/pods/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59/containers/node-ca/17ebd0ab not reset as customized by admin to system_u:object_r:container_file_t:s0:c338,c343 Dec 13 06:49:02 crc restorecon[4764]: /var/lib/kubelet/pods/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59/containers/node-ca/005579f4 not reset as customized by admin to system_u:object_r:container_file_t:s0:c842,c986 Dec 13 06:49:02 crc restorecon[4764]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/volumes/kubernetes.io~configmap/etcd-serving-ca not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Dec 13 06:49:02 crc restorecon[4764]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/volumes/kubernetes.io~configmap/etcd-serving-ca/..2025_02_23_05_23_11.449897510 not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Dec 13 06:49:02 crc restorecon[4764]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/volumes/kubernetes.io~configmap/etcd-serving-ca/..2025_02_23_05_23_11.449897510/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Dec 13 06:49:02 crc restorecon[4764]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/volumes/kubernetes.io~configmap/etcd-serving-ca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Dec 13 06:49:02 crc restorecon[4764]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/volumes/kubernetes.io~configmap/etcd-serving-ca/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Dec 13 06:49:02 crc restorecon[4764]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/volumes/kubernetes.io~configmap/trusted-ca-bundle not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Dec 13 06:49:02 crc restorecon[4764]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/volumes/kubernetes.io~configmap/trusted-ca-bundle/..2025_02_23_05_23_11.1287037894 not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Dec 13 06:49:02 crc restorecon[4764]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/volumes/kubernetes.io~configmap/trusted-ca-bundle/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Dec 13 06:49:02 crc restorecon[4764]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/volumes/kubernetes.io~configmap/audit-policies not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Dec 13 06:49:02 crc restorecon[4764]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/volumes/kubernetes.io~configmap/audit-policies/..2025_02_23_05_23_11.1301053334 not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Dec 13 06:49:02 crc restorecon[4764]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/volumes/kubernetes.io~configmap/audit-policies/..2025_02_23_05_23_11.1301053334/policy.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Dec 13 06:49:02 crc restorecon[4764]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/volumes/kubernetes.io~configmap/audit-policies/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Dec 13 06:49:02 crc restorecon[4764]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/volumes/kubernetes.io~configmap/audit-policies/policy.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Dec 13 06:49:02 crc restorecon[4764]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Dec 13 06:49:02 crc restorecon[4764]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/containers/fix-audit-permissions/bf5f3b9c not reset as customized by admin to system_u:object_r:container_file_t:s0:c49,c263 Dec 13 06:49:02 crc restorecon[4764]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/containers/fix-audit-permissions/af276eb7 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c701 Dec 13 06:49:02 crc restorecon[4764]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/containers/fix-audit-permissions/ea28e322 not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Dec 13 06:49:02 crc restorecon[4764]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/containers/oauth-apiserver/692e6683 not reset as customized by admin to system_u:object_r:container_file_t:s0:c49,c263 Dec 13 06:49:02 crc restorecon[4764]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/containers/oauth-apiserver/871746a7 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c701 Dec 13 06:49:02 crc restorecon[4764]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/containers/oauth-apiserver/4eb2e958 not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Dec 13 06:49:02 crc restorecon[4764]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/console-config not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Dec 13 06:49:02 crc restorecon[4764]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/console-config/..2025_02_24_06_09_06.2875086261 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Dec 13 06:49:02 crc restorecon[4764]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/console-config/..2025_02_24_06_09_06.2875086261/console-config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Dec 13 06:49:02 crc restorecon[4764]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/console-config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Dec 13 06:49:02 crc restorecon[4764]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/console-config/console-config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Dec 13 06:49:02 crc restorecon[4764]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/trusted-ca-bundle not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Dec 13 06:49:02 crc restorecon[4764]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/trusted-ca-bundle/..2025_02_24_06_09_06.286118152 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Dec 13 06:49:02 crc restorecon[4764]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/trusted-ca-bundle/..2025_02_24_06_09_06.286118152/tls-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Dec 13 06:49:02 crc restorecon[4764]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/trusted-ca-bundle/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Dec 13 06:49:02 crc restorecon[4764]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/trusted-ca-bundle/tls-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Dec 13 06:49:02 crc restorecon[4764]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/oauth-serving-cert not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Dec 13 06:49:02 crc restorecon[4764]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/oauth-serving-cert/..2025_02_24_06_09_06.3865795478 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Dec 13 06:49:02 crc restorecon[4764]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/oauth-serving-cert/..2025_02_24_06_09_06.3865795478/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Dec 13 06:49:02 crc restorecon[4764]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/oauth-serving-cert/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Dec 13 06:49:02 crc restorecon[4764]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/oauth-serving-cert/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Dec 13 06:49:02 crc restorecon[4764]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/service-ca not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Dec 13 06:49:02 crc restorecon[4764]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/service-ca/..2025_02_24_06_09_06.584414814 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Dec 13 06:49:02 crc restorecon[4764]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/service-ca/..2025_02_24_06_09_06.584414814/service-ca.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Dec 13 06:49:02 crc restorecon[4764]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/service-ca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Dec 13 06:49:02 crc restorecon[4764]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/service-ca/service-ca.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Dec 13 06:49:02 crc restorecon[4764]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Dec 13 06:49:02 crc restorecon[4764]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/containers/console/ca9b62da not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Dec 13 06:49:02 crc restorecon[4764]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/containers/console/0edd6fce not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Dec 13 06:49:02 crc restorecon[4764]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/config not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Dec 13 06:49:02 crc restorecon[4764]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/config/..2025_02_24_06_20_07.2406383837 not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Dec 13 06:49:02 crc restorecon[4764]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/config/..2025_02_24_06_20_07.2406383837/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Dec 13 06:49:02 crc restorecon[4764]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/config/..2025_02_24_06_20_07.2406383837/openshift-controller-manager.client-ca.configmap not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Dec 13 06:49:02 crc restorecon[4764]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/config/..2025_02_24_06_20_07.2406383837/openshift-controller-manager.openshift-global-ca.configmap not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Dec 13 06:49:02 crc restorecon[4764]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/config/..2025_02_24_06_20_07.2406383837/openshift-controller-manager.serving-cert.secret not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Dec 13 06:49:02 crc restorecon[4764]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Dec 13 06:49:02 crc restorecon[4764]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/config/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Dec 13 06:49:02 crc restorecon[4764]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/config/openshift-controller-manager.client-ca.configmap not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Dec 13 06:49:02 crc restorecon[4764]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/config/openshift-controller-manager.openshift-global-ca.configmap not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Dec 13 06:49:02 crc restorecon[4764]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/config/openshift-controller-manager.serving-cert.secret not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Dec 13 06:49:02 crc restorecon[4764]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/client-ca not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Dec 13 06:49:02 crc restorecon[4764]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/client-ca/..2025_02_24_06_20_07.1071801880 not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Dec 13 06:49:02 crc restorecon[4764]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/client-ca/..2025_02_24_06_20_07.1071801880/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Dec 13 06:49:02 crc restorecon[4764]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/client-ca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Dec 13 06:49:02 crc restorecon[4764]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/client-ca/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Dec 13 06:49:02 crc restorecon[4764]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/proxy-ca-bundles not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Dec 13 06:49:02 crc restorecon[4764]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/proxy-ca-bundles/..2025_02_24_06_20_07.2494444877 not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Dec 13 06:49:02 crc restorecon[4764]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/proxy-ca-bundles/..2025_02_24_06_20_07.2494444877/tls-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Dec 13 06:49:02 crc restorecon[4764]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/proxy-ca-bundles/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Dec 13 06:49:02 crc restorecon[4764]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/proxy-ca-bundles/tls-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Dec 13 06:49:02 crc restorecon[4764]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Dec 13 06:49:02 crc restorecon[4764]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/containers/controller-manager/89b4555f not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Dec 13 06:49:02 crc restorecon[4764]: /var/lib/kubelet/pods/87cf06ed-a83f-41a7-828d-70653580a8cb/volumes/kubernetes.io~configmap/config-volume not reset as customized by admin to system_u:object_r:container_file_t:s0:c466,c972 Dec 13 06:49:02 crc restorecon[4764]: /var/lib/kubelet/pods/87cf06ed-a83f-41a7-828d-70653580a8cb/volumes/kubernetes.io~configmap/config-volume/..2025_02_23_05_23_22.4071100442 not reset as customized by admin to system_u:object_r:container_file_t:s0:c466,c972 Dec 13 06:49:02 crc restorecon[4764]: /var/lib/kubelet/pods/87cf06ed-a83f-41a7-828d-70653580a8cb/volumes/kubernetes.io~configmap/config-volume/..2025_02_23_05_23_22.4071100442/Corefile not reset as customized by admin to system_u:object_r:container_file_t:s0:c466,c972 Dec 13 06:49:02 crc restorecon[4764]: /var/lib/kubelet/pods/87cf06ed-a83f-41a7-828d-70653580a8cb/volumes/kubernetes.io~configmap/config-volume/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c466,c972 Dec 13 06:49:02 crc restorecon[4764]: /var/lib/kubelet/pods/87cf06ed-a83f-41a7-828d-70653580a8cb/volumes/kubernetes.io~configmap/config-volume/Corefile not reset as customized by admin to system_u:object_r:container_file_t:s0:c466,c972 Dec 13 06:49:02 crc restorecon[4764]: /var/lib/kubelet/pods/87cf06ed-a83f-41a7-828d-70653580a8cb/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c466,c972 Dec 13 06:49:02 crc restorecon[4764]: /var/lib/kubelet/pods/87cf06ed-a83f-41a7-828d-70653580a8cb/containers/dns/655fcd71 not reset as customized by admin to system_u:object_r:container_file_t:s0:c457,c841 Dec 13 06:49:02 crc restorecon[4764]: /var/lib/kubelet/pods/87cf06ed-a83f-41a7-828d-70653580a8cb/containers/dns/0d43c002 not reset as customized by admin to system_u:object_r:container_file_t:s0:c55,c1022 Dec 13 06:49:02 crc restorecon[4764]: /var/lib/kubelet/pods/87cf06ed-a83f-41a7-828d-70653580a8cb/containers/dns/e68efd17 not reset as customized by admin to system_u:object_r:container_file_t:s0:c466,c972 Dec 13 06:49:02 crc restorecon[4764]: /var/lib/kubelet/pods/87cf06ed-a83f-41a7-828d-70653580a8cb/containers/kube-rbac-proxy/9acf9b65 not reset as customized by admin to system_u:object_r:container_file_t:s0:c457,c841 Dec 13 06:49:02 crc restorecon[4764]: /var/lib/kubelet/pods/87cf06ed-a83f-41a7-828d-70653580a8cb/containers/kube-rbac-proxy/5ae3ff11 not reset as customized by admin to system_u:object_r:container_file_t:s0:c55,c1022 Dec 13 06:49:02 crc restorecon[4764]: /var/lib/kubelet/pods/87cf06ed-a83f-41a7-828d-70653580a8cb/containers/kube-rbac-proxy/1e59206a not reset as customized by admin to system_u:object_r:container_file_t:s0:c466,c972 Dec 13 06:49:02 crc restorecon[4764]: /var/lib/kubelet/pods/44663579-783b-4372-86d6-acf235a62d72/containers/dns-node-resolver/27af16d1 not reset as customized by admin to system_u:object_r:container_file_t:s0:c304,c1017 Dec 13 06:49:02 crc restorecon[4764]: /var/lib/kubelet/pods/44663579-783b-4372-86d6-acf235a62d72/containers/dns-node-resolver/7918e729 not reset as customized by admin to system_u:object_r:container_file_t:s0:c853,c893 Dec 13 06:49:02 crc restorecon[4764]: /var/lib/kubelet/pods/44663579-783b-4372-86d6-acf235a62d72/containers/dns-node-resolver/5d976d0e not reset as customized by admin to system_u:object_r:container_file_t:s0:c585,c981 Dec 13 06:49:02 crc restorecon[4764]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/volumes/kubernetes.io~configmap/config not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Dec 13 06:49:02 crc restorecon[4764]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/volumes/kubernetes.io~configmap/config/..2025_02_23_05_38_56.1112187283 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Dec 13 06:49:02 crc restorecon[4764]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/volumes/kubernetes.io~configmap/config/..2025_02_23_05_38_56.1112187283/controller-config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Dec 13 06:49:02 crc restorecon[4764]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/volumes/kubernetes.io~configmap/config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Dec 13 06:49:02 crc restorecon[4764]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/volumes/kubernetes.io~configmap/config/controller-config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Dec 13 06:49:02 crc restorecon[4764]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/volumes/kubernetes.io~configmap/trusted-ca not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Dec 13 06:49:02 crc restorecon[4764]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/volumes/kubernetes.io~configmap/trusted-ca/..2025_02_23_05_38_56.2839772658 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Dec 13 06:49:02 crc restorecon[4764]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/volumes/kubernetes.io~configmap/trusted-ca/..2025_02_23_05_38_56.2839772658/tls-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Dec 13 06:49:02 crc restorecon[4764]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/volumes/kubernetes.io~configmap/trusted-ca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Dec 13 06:49:02 crc restorecon[4764]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/volumes/kubernetes.io~configmap/trusted-ca/tls-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Dec 13 06:49:02 crc restorecon[4764]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Dec 13 06:49:02 crc restorecon[4764]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/containers/console-operator/d7f55cbb not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Dec 13 06:49:02 crc restorecon[4764]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/containers/console-operator/f0812073 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Dec 13 06:49:02 crc restorecon[4764]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/containers/console-operator/1a56cbeb not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Dec 13 06:49:02 crc restorecon[4764]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/containers/console-operator/7fdd437e not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Dec 13 06:49:02 crc restorecon[4764]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/containers/console-operator/cdfb5652 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Dec 13 06:49:02 crc restorecon[4764]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/etcd-serving-ca not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Dec 13 06:49:02 crc restorecon[4764]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/etcd-serving-ca/..2025_02_24_06_17_29.3844392896 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Dec 13 06:49:02 crc restorecon[4764]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/etcd-serving-ca/..2025_02_24_06_17_29.3844392896/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Dec 13 06:49:02 crc restorecon[4764]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/etcd-serving-ca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Dec 13 06:49:02 crc restorecon[4764]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/etcd-serving-ca/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Dec 13 06:49:02 crc restorecon[4764]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/config not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Dec 13 06:49:02 crc restorecon[4764]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/config/..2025_02_24_06_17_29.848549803 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Dec 13 06:49:02 crc restorecon[4764]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/config/..2025_02_24_06_17_29.848549803/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Dec 13 06:49:02 crc restorecon[4764]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Dec 13 06:49:02 crc restorecon[4764]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/config/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Dec 13 06:49:02 crc restorecon[4764]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/audit not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Dec 13 06:49:02 crc restorecon[4764]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/audit/..2025_02_24_06_17_29.780046231 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Dec 13 06:49:02 crc restorecon[4764]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/audit/..2025_02_24_06_17_29.780046231/policy.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Dec 13 06:49:02 crc restorecon[4764]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/audit/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Dec 13 06:49:02 crc restorecon[4764]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/audit/policy.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Dec 13 06:49:02 crc restorecon[4764]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/image-import-ca not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Dec 13 06:49:02 crc restorecon[4764]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/image-import-ca/..2025_02_24_06_17_29.2926008347 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Dec 13 06:49:02 crc restorecon[4764]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/image-import-ca/..2025_02_24_06_17_29.2926008347/image-registry.openshift-image-registry.svc..5000 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Dec 13 06:49:02 crc restorecon[4764]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/image-import-ca/..2025_02_24_06_17_29.2926008347/image-registry.openshift-image-registry.svc.cluster.local..5000 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Dec 13 06:49:02 crc restorecon[4764]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/image-import-ca/..2025_02_24_06_17_29.2926008347/default-route-openshift-image-registry.apps-crc.testing not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Dec 13 06:49:02 crc restorecon[4764]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/image-import-ca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Dec 13 06:49:02 crc restorecon[4764]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/image-import-ca/image-registry.openshift-image-registry.svc..5000 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Dec 13 06:49:02 crc restorecon[4764]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/image-import-ca/image-registry.openshift-image-registry.svc.cluster.local..5000 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Dec 13 06:49:02 crc restorecon[4764]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/image-import-ca/default-route-openshift-image-registry.apps-crc.testing not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Dec 13 06:49:02 crc restorecon[4764]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/trusted-ca-bundle not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Dec 13 06:49:02 crc restorecon[4764]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/trusted-ca-bundle/..2025_02_24_06_17_29.2729721485 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Dec 13 06:49:02 crc restorecon[4764]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/trusted-ca-bundle/..2025_02_24_06_17_29.2729721485/tls-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Dec 13 06:49:02 crc restorecon[4764]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/trusted-ca-bundle/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Dec 13 06:49:02 crc restorecon[4764]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/trusted-ca-bundle/tls-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Dec 13 06:49:02 crc restorecon[4764]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Dec 13 06:49:02 crc restorecon[4764]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/containers/fix-audit-permissions/fb93119e not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Dec 13 06:49:02 crc restorecon[4764]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/containers/openshift-apiserver/f1e8fc0e not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Dec 13 06:49:02 crc restorecon[4764]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/containers/openshift-apiserver-check-endpoints/218511f3 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Dec 13 06:49:02 crc restorecon[4764]: /var/lib/kubelet/pods/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b/volumes/kubernetes.io~empty-dir/tmpfs not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Dec 13 06:49:02 crc restorecon[4764]: /var/lib/kubelet/pods/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b/volumes/kubernetes.io~empty-dir/tmpfs/k8s-webhook-server not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Dec 13 06:49:02 crc restorecon[4764]: /var/lib/kubelet/pods/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b/volumes/kubernetes.io~empty-dir/tmpfs/k8s-webhook-server/serving-certs not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Dec 13 06:49:02 crc restorecon[4764]: /var/lib/kubelet/pods/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Dec 13 06:49:02 crc restorecon[4764]: /var/lib/kubelet/pods/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b/containers/packageserver/ca8af7b3 not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Dec 13 06:49:02 crc restorecon[4764]: /var/lib/kubelet/pods/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b/containers/packageserver/72cc8a75 not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Dec 13 06:49:02 crc restorecon[4764]: /var/lib/kubelet/pods/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b/containers/packageserver/6e8a3760 not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Dec 13 06:49:02 crc restorecon[4764]: /var/lib/kubelet/pods/0b78653f-4ff9-4508-8672-245ed9b561e3/volumes/kubernetes.io~configmap/service-ca not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c6 Dec 13 06:49:02 crc restorecon[4764]: /var/lib/kubelet/pods/0b78653f-4ff9-4508-8672-245ed9b561e3/volumes/kubernetes.io~configmap/service-ca/..2025_02_23_05_27_30.557428972 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c6 Dec 13 06:49:02 crc restorecon[4764]: /var/lib/kubelet/pods/0b78653f-4ff9-4508-8672-245ed9b561e3/volumes/kubernetes.io~configmap/service-ca/..2025_02_23_05_27_30.557428972/service-ca.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c6 Dec 13 06:49:02 crc restorecon[4764]: /var/lib/kubelet/pods/0b78653f-4ff9-4508-8672-245ed9b561e3/volumes/kubernetes.io~configmap/service-ca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c6 Dec 13 06:49:02 crc restorecon[4764]: /var/lib/kubelet/pods/0b78653f-4ff9-4508-8672-245ed9b561e3/volumes/kubernetes.io~configmap/service-ca/service-ca.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c6 Dec 13 06:49:02 crc restorecon[4764]: /var/lib/kubelet/pods/0b78653f-4ff9-4508-8672-245ed9b561e3/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c6 Dec 13 06:49:02 crc restorecon[4764]: /var/lib/kubelet/pods/0b78653f-4ff9-4508-8672-245ed9b561e3/containers/cluster-version-operator/4c3455c0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c6 Dec 13 06:49:02 crc restorecon[4764]: /var/lib/kubelet/pods/0b78653f-4ff9-4508-8672-245ed9b561e3/containers/cluster-version-operator/2278acb0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c6 Dec 13 06:49:02 crc restorecon[4764]: /var/lib/kubelet/pods/0b78653f-4ff9-4508-8672-245ed9b561e3/containers/cluster-version-operator/4b453e4f not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c6 Dec 13 06:49:02 crc restorecon[4764]: /var/lib/kubelet/pods/0b78653f-4ff9-4508-8672-245ed9b561e3/containers/cluster-version-operator/3ec09bda not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c6 Dec 13 06:49:02 crc restorecon[4764]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~configmap/trusted-ca not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 13 06:49:02 crc restorecon[4764]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~configmap/trusted-ca/..2025_02_24_06_25_03.422633132 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 13 06:49:02 crc restorecon[4764]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~configmap/trusted-ca/..2025_02_24_06_25_03.422633132/anchors not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 13 06:49:02 crc restorecon[4764]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~configmap/trusted-ca/..2025_02_24_06_25_03.422633132/anchors/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 13 06:49:02 crc restorecon[4764]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~configmap/trusted-ca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 13 06:49:02 crc restorecon[4764]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~configmap/trusted-ca/anchors not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 13 06:49:02 crc restorecon[4764]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~configmap/registry-certificates not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 13 06:49:02 crc restorecon[4764]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~configmap/registry-certificates/..2025_02_24_06_25_03.3594477318 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 13 06:49:02 crc restorecon[4764]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~configmap/registry-certificates/..2025_02_24_06_25_03.3594477318/image-registry.openshift-image-registry.svc..5000 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 13 06:49:02 crc restorecon[4764]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~configmap/registry-certificates/..2025_02_24_06_25_03.3594477318/image-registry.openshift-image-registry.svc.cluster.local..5000 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 13 06:49:02 crc restorecon[4764]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~configmap/registry-certificates/..2025_02_24_06_25_03.3594477318/default-route-openshift-image-registry.apps-crc.testing not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 13 06:49:02 crc restorecon[4764]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~configmap/registry-certificates/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 13 06:49:02 crc restorecon[4764]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~configmap/registry-certificates/image-registry.openshift-image-registry.svc..5000 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 13 06:49:02 crc restorecon[4764]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~configmap/registry-certificates/image-registry.openshift-image-registry.svc.cluster.local..5000 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 13 06:49:02 crc restorecon[4764]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~configmap/registry-certificates/default-route-openshift-image-registry.apps-crc.testing not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 13 06:49:02 crc restorecon[4764]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 13 06:49:02 crc restorecon[4764]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/edk2 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 13 06:49:02 crc restorecon[4764]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/edk2/cacerts.bin not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 13 06:49:02 crc restorecon[4764]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/java not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 13 06:49:02 crc restorecon[4764]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/java/cacerts not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 13 06:49:02 crc restorecon[4764]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/openssl not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 13 06:49:02 crc restorecon[4764]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/openssl/ca-bundle.trust.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 13 06:49:02 crc restorecon[4764]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 13 06:49:02 crc restorecon[4764]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/tls-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 13 06:49:02 crc restorecon[4764]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/email-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 13 06:49:02 crc restorecon[4764]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/objsign-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 13 06:49:02 crc restorecon[4764]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 13 06:49:02 crc restorecon[4764]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/2ae6433e.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 13 06:49:02 crc restorecon[4764]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/fde84897.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 13 06:49:02 crc restorecon[4764]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/75680d2e.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 13 06:49:02 crc restorecon[4764]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/openshift-service-serving-signer_1740288168.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 13 06:49:02 crc restorecon[4764]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/facfc4fa.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 13 06:49:02 crc restorecon[4764]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/8f5a969c.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 13 06:49:02 crc restorecon[4764]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/CFCA_EV_ROOT.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 13 06:49:02 crc restorecon[4764]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9ef4a08a.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 13 06:49:02 crc restorecon[4764]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ingress-operator_1740288202.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 13 06:49:02 crc restorecon[4764]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/2f332aed.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 13 06:49:02 crc restorecon[4764]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/248c8271.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 13 06:49:02 crc restorecon[4764]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/8d10a21f.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 13 06:49:02 crc restorecon[4764]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ACCVRAIZ1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 13 06:49:02 crc restorecon[4764]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/a94d09e5.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 13 06:49:02 crc restorecon[4764]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/3c9a4d3b.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 13 06:49:02 crc restorecon[4764]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/40193066.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 13 06:49:02 crc restorecon[4764]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/AC_RAIZ_FNMT-RCM.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 13 06:49:02 crc restorecon[4764]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/cd8c0d63.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 13 06:49:02 crc restorecon[4764]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b936d1c6.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 13 06:49:02 crc restorecon[4764]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/CA_Disig_Root_R2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 13 06:49:02 crc restorecon[4764]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/4fd49c6c.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 13 06:49:02 crc restorecon[4764]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/AC_RAIZ_FNMT-RCM_SERVIDORES_SEGUROS.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 13 06:49:02 crc restorecon[4764]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b81b93f0.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 13 06:49:02 crc restorecon[4764]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/5f9a69fa.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 13 06:49:02 crc restorecon[4764]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Certigna.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 13 06:49:02 crc restorecon[4764]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b30d5fda.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 13 06:49:02 crc restorecon[4764]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ANF_Secure_Server_Root_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 13 06:49:02 crc restorecon[4764]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b433981b.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 13 06:49:02 crc restorecon[4764]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/93851c9e.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 13 06:49:02 crc restorecon[4764]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9282e51c.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 13 06:49:02 crc restorecon[4764]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/e7dd1bc4.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 13 06:49:02 crc restorecon[4764]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Actalis_Authentication_Root_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 13 06:49:02 crc restorecon[4764]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/930ac5d2.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 13 06:49:02 crc restorecon[4764]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/5f47b495.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 13 06:49:02 crc restorecon[4764]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/e113c810.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 13 06:49:02 crc restorecon[4764]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/5931b5bc.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 13 06:49:02 crc restorecon[4764]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/AffirmTrust_Commercial.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 13 06:49:02 crc restorecon[4764]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/2b349938.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 13 06:49:02 crc restorecon[4764]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/e48193cf.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 13 06:49:02 crc restorecon[4764]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/302904dd.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 13 06:49:02 crc restorecon[4764]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/a716d4ed.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 13 06:49:02 crc restorecon[4764]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/AffirmTrust_Networking.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 13 06:49:02 crc restorecon[4764]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/93bc0acc.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 13 06:49:02 crc restorecon[4764]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/86212b19.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 13 06:49:02 crc restorecon[4764]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Certigna_Root_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 13 06:49:02 crc restorecon[4764]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/AffirmTrust_Premium.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 13 06:49:02 crc restorecon[4764]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b727005e.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 13 06:49:02 crc restorecon[4764]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/dbc54cab.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 13 06:49:02 crc restorecon[4764]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/f51bb24c.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 13 06:49:02 crc restorecon[4764]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/c28a8a30.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 13 06:49:02 crc restorecon[4764]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/AffirmTrust_Premium_ECC.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 13 06:49:02 crc restorecon[4764]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9c8dfbd4.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 13 06:49:02 crc restorecon[4764]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ccc52f49.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 13 06:49:02 crc restorecon[4764]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/cb1c3204.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 13 06:49:02 crc restorecon[4764]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Amazon_Root_CA_1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 13 06:49:02 crc restorecon[4764]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ce5e74ef.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 13 06:49:02 crc restorecon[4764]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/fd08c599.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 13 06:49:02 crc restorecon[4764]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Certum_Trusted_Root_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 13 06:49:02 crc restorecon[4764]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Amazon_Root_CA_2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 13 06:49:02 crc restorecon[4764]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/6d41d539.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 13 06:49:02 crc restorecon[4764]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/fb5fa911.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 13 06:49:02 crc restorecon[4764]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/e35234b1.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 13 06:49:02 crc restorecon[4764]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Amazon_Root_CA_3.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 13 06:49:02 crc restorecon[4764]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/8cb5ee0f.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 13 06:49:02 crc restorecon[4764]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/7a7c655d.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 13 06:49:02 crc restorecon[4764]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/f8fc53da.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 13 06:49:02 crc restorecon[4764]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Amazon_Root_CA_4.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 13 06:49:02 crc restorecon[4764]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/de6d66f3.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 13 06:49:02 crc restorecon[4764]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/d41b5e2a.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 13 06:49:02 crc restorecon[4764]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/41a3f684.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 13 06:49:02 crc restorecon[4764]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/1df5a75f.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 13 06:49:02 crc restorecon[4764]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Atos_TrustedRoot_2011.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 13 06:49:02 crc restorecon[4764]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/e36a6752.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 13 06:49:02 crc restorecon[4764]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b872f2b4.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 13 06:49:02 crc restorecon[4764]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9576d26b.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 13 06:49:02 crc restorecon[4764]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/228f89db.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 13 06:49:02 crc restorecon[4764]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Atos_TrustedRoot_Root_CA_ECC_TLS_2021.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 13 06:49:02 crc restorecon[4764]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/fb717492.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 13 06:49:02 crc restorecon[4764]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/2d21b73c.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 13 06:49:02 crc restorecon[4764]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/0b1b94ef.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 13 06:49:02 crc restorecon[4764]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/595e996b.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 13 06:49:02 crc restorecon[4764]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Atos_TrustedRoot_Root_CA_RSA_TLS_2021.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 13 06:49:02 crc restorecon[4764]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9b46e03d.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 13 06:49:02 crc restorecon[4764]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/128f4b91.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 13 06:49:02 crc restorecon[4764]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Buypass_Class_3_Root_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 13 06:49:02 crc restorecon[4764]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/81f2d2b1.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 13 06:49:02 crc restorecon[4764]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Autoridad_de_Certificacion_Firmaprofesional_CIF_A62634068.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 13 06:49:02 crc restorecon[4764]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/3bde41ac.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 13 06:49:02 crc restorecon[4764]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/d16a5865.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 13 06:49:02 crc restorecon[4764]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Certum_EC-384_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 13 06:49:02 crc restorecon[4764]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/BJCA_Global_Root_CA1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 13 06:49:02 crc restorecon[4764]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/0179095f.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 13 06:49:02 crc restorecon[4764]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ffa7f1eb.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 13 06:49:02 crc restorecon[4764]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9482e63a.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 13 06:49:02 crc restorecon[4764]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/d4dae3dd.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 13 06:49:02 crc restorecon[4764]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/BJCA_Global_Root_CA2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 13 06:49:02 crc restorecon[4764]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/3e359ba6.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 13 06:49:02 crc restorecon[4764]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/7e067d03.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 13 06:49:02 crc restorecon[4764]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/95aff9e3.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 13 06:49:02 crc restorecon[4764]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/d7746a63.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 13 06:49:02 crc restorecon[4764]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Baltimore_CyberTrust_Root.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 13 06:49:02 crc restorecon[4764]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/653b494a.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 13 06:49:02 crc restorecon[4764]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/3ad48a91.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 13 06:49:02 crc restorecon[4764]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Certum_Trusted_Network_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 13 06:49:02 crc restorecon[4764]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Buypass_Class_2_Root_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 13 06:49:02 crc restorecon[4764]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/54657681.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 13 06:49:02 crc restorecon[4764]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/82223c44.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 13 06:49:02 crc restorecon[4764]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/e8de2f56.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 13 06:49:02 crc restorecon[4764]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/2d9dafe4.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 13 06:49:02 crc restorecon[4764]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/d96b65e2.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 13 06:49:02 crc restorecon[4764]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ee64a828.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 13 06:49:02 crc restorecon[4764]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/COMODO_Certification_Authority.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 13 06:49:02 crc restorecon[4764]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/40547a79.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 13 06:49:02 crc restorecon[4764]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/5a3f0ff8.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 13 06:49:02 crc restorecon[4764]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/7a780d93.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 13 06:49:02 crc restorecon[4764]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/34d996fb.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 13 06:49:02 crc restorecon[4764]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/COMODO_ECC_Certification_Authority.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 13 06:49:02 crc restorecon[4764]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/eed8c118.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 13 06:49:02 crc restorecon[4764]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/89c02a45.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 13 06:49:02 crc restorecon[4764]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Certainly_Root_R1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 13 06:49:02 crc restorecon[4764]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b1159c4c.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 13 06:49:02 crc restorecon[4764]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/COMODO_RSA_Certification_Authority.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 13 06:49:02 crc restorecon[4764]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/d6325660.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 13 06:49:02 crc restorecon[4764]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/d4c339cb.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 13 06:49:02 crc restorecon[4764]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/8312c4c1.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 13 06:49:02 crc restorecon[4764]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Certainly_Root_E1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 13 06:49:02 crc restorecon[4764]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/8508e720.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 13 06:49:02 crc restorecon[4764]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/5fdd185d.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 13 06:49:02 crc restorecon[4764]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/48bec511.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 13 06:49:02 crc restorecon[4764]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/69105f4f.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 13 06:49:02 crc restorecon[4764]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/GlobalSign.1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 13 06:49:02 crc restorecon[4764]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/0b9bc432.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 13 06:49:02 crc restorecon[4764]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Certum_Trusted_Network_CA_2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 13 06:49:02 crc restorecon[4764]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/GTS_Root_R3.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 13 06:49:02 crc restorecon[4764]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/32888f65.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 13 06:49:02 crc restorecon[4764]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/CommScope_Public_Trust_ECC_Root-01.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 13 06:49:02 crc restorecon[4764]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/6b03dec0.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 13 06:49:02 crc restorecon[4764]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/219d9499.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 13 06:49:02 crc restorecon[4764]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/CommScope_Public_Trust_ECC_Root-02.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 13 06:49:02 crc restorecon[4764]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/5acf816d.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 13 06:49:02 crc restorecon[4764]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/cbf06781.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 13 06:49:02 crc restorecon[4764]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/CommScope_Public_Trust_RSA_Root-01.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 13 06:49:02 crc restorecon[4764]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/GTS_Root_R4.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 13 06:49:02 crc restorecon[4764]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/dc99f41e.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 13 06:49:02 crc restorecon[4764]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/CommScope_Public_Trust_RSA_Root-02.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 13 06:49:02 crc restorecon[4764]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/GlobalSign.3.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 13 06:49:02 crc restorecon[4764]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/AAA_Certificate_Services.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 13 06:49:02 crc restorecon[4764]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/985c1f52.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 13 06:49:02 crc restorecon[4764]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/8794b4e3.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 13 06:49:02 crc restorecon[4764]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/D-TRUST_BR_Root_CA_1_2020.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 13 06:49:02 crc restorecon[4764]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/e7c037b4.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 13 06:49:02 crc restorecon[4764]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ef954a4e.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 13 06:49:02 crc restorecon[4764]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/D-TRUST_EV_Root_CA_1_2020.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 13 06:49:02 crc restorecon[4764]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/2add47b6.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 13 06:49:02 crc restorecon[4764]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/90c5a3c8.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 13 06:49:02 crc restorecon[4764]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/D-TRUST_Root_Class_3_CA_2_2009.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 13 06:49:02 crc restorecon[4764]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b0f3e76e.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 13 06:49:02 crc restorecon[4764]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/53a1b57a.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 13 06:49:02 crc restorecon[4764]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/D-TRUST_Root_Class_3_CA_2_EV_2009.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 13 06:49:02 crc restorecon[4764]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/GlobalSign_Root_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 13 06:49:02 crc restorecon[4764]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/DigiCert_Assured_ID_Root_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 13 06:49:02 crc restorecon[4764]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/5ad8a5d6.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 13 06:49:02 crc restorecon[4764]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/68dd7389.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 13 06:49:02 crc restorecon[4764]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/DigiCert_Assured_ID_Root_G2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 13 06:49:02 crc restorecon[4764]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9d04f354.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 13 06:49:02 crc restorecon[4764]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/8d6437c3.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 13 06:49:02 crc restorecon[4764]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/062cdee6.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 13 06:49:02 crc restorecon[4764]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/bd43e1dd.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 13 06:49:02 crc restorecon[4764]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/DigiCert_Assured_ID_Root_G3.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 13 06:49:02 crc restorecon[4764]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/7f3d5d1d.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 13 06:49:02 crc restorecon[4764]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/c491639e.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 13 06:49:02 crc restorecon[4764]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/GlobalSign_Root_E46.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 13 06:49:02 crc restorecon[4764]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/DigiCert_Global_Root_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 13 06:49:02 crc restorecon[4764]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/3513523f.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 13 06:49:02 crc restorecon[4764]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/399e7759.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 13 06:49:02 crc restorecon[4764]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/feffd413.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 13 06:49:02 crc restorecon[4764]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/d18e9066.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 13 06:49:02 crc restorecon[4764]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/DigiCert_Global_Root_G2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 13 06:49:02 crc restorecon[4764]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/607986c7.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 13 06:49:02 crc restorecon[4764]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/c90bc37d.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 13 06:49:02 crc restorecon[4764]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/1b0f7e5c.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 13 06:49:02 crc restorecon[4764]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/1e08bfd1.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 13 06:49:02 crc restorecon[4764]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/DigiCert_Global_Root_G3.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 13 06:49:02 crc restorecon[4764]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/dd8e9d41.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 13 06:49:02 crc restorecon[4764]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ed39abd0.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 13 06:49:02 crc restorecon[4764]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/a3418fda.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 13 06:49:02 crc restorecon[4764]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/bc3f2570.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 13 06:49:02 crc restorecon[4764]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/DigiCert_High_Assurance_EV_Root_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 13 06:49:02 crc restorecon[4764]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/244b5494.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 13 06:49:02 crc restorecon[4764]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/81b9768f.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 13 06:49:02 crc restorecon[4764]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/GlobalSign.2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 13 06:49:02 crc restorecon[4764]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/4be590e0.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 13 06:49:02 crc restorecon[4764]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/DigiCert_TLS_ECC_P384_Root_G5.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 13 06:49:02 crc restorecon[4764]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9846683b.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 13 06:49:02 crc restorecon[4764]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/252252d2.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 13 06:49:02 crc restorecon[4764]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/1e8e7201.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 13 06:49:02 crc restorecon[4764]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ISRG_Root_X1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 13 06:49:02 crc restorecon[4764]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/DigiCert_TLS_RSA4096_Root_G5.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 13 06:49:02 crc restorecon[4764]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/d52c538d.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 13 06:49:02 crc restorecon[4764]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/c44cc0c0.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 13 06:49:02 crc restorecon[4764]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/GlobalSign_Root_R46.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 13 06:49:02 crc restorecon[4764]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/DigiCert_Trusted_Root_G4.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 13 06:49:02 crc restorecon[4764]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/75d1b2ed.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 13 06:49:02 crc restorecon[4764]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/a2c66da8.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 13 06:49:02 crc restorecon[4764]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/GTS_Root_R2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 13 06:49:02 crc restorecon[4764]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ecccd8db.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 13 06:49:02 crc restorecon[4764]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Entrust.net_Certification_Authority__2048_.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 13 06:49:02 crc restorecon[4764]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/aee5f10d.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 13 06:49:02 crc restorecon[4764]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/3e7271e8.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 13 06:49:02 crc restorecon[4764]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b0e59380.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 13 06:49:02 crc restorecon[4764]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/4c3982f2.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 13 06:49:02 crc restorecon[4764]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Entrust_Root_Certification_Authority.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 13 06:49:02 crc restorecon[4764]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/6b99d060.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 13 06:49:02 crc restorecon[4764]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/bf64f35b.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 13 06:49:02 crc restorecon[4764]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/0a775a30.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 13 06:49:02 crc restorecon[4764]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/002c0b4f.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 13 06:49:02 crc restorecon[4764]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/cc450945.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 13 06:49:02 crc restorecon[4764]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Entrust_Root_Certification_Authority_-_EC1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 13 06:49:02 crc restorecon[4764]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/106f3e4d.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 13 06:49:02 crc restorecon[4764]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b3fb433b.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 13 06:49:02 crc restorecon[4764]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/GlobalSign.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 13 06:49:02 crc restorecon[4764]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/4042bcee.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 13 06:49:02 crc restorecon[4764]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Entrust_Root_Certification_Authority_-_G2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 13 06:49:02 crc restorecon[4764]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/02265526.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 13 06:49:02 crc restorecon[4764]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/455f1b52.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 13 06:49:02 crc restorecon[4764]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/0d69c7e1.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 13 06:49:02 crc restorecon[4764]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9f727ac7.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 13 06:49:02 crc restorecon[4764]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Entrust_Root_Certification_Authority_-_G4.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 13 06:49:02 crc restorecon[4764]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/5e98733a.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 13 06:49:02 crc restorecon[4764]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/f0cd152c.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 13 06:49:02 crc restorecon[4764]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/dc4d6a89.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 13 06:49:02 crc restorecon[4764]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/6187b673.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 13 06:49:02 crc restorecon[4764]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/FIRMAPROFESIONAL_CA_ROOT-A_WEB.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 13 06:49:02 crc restorecon[4764]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ba8887ce.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 13 06:49:02 crc restorecon[4764]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/068570d1.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 13 06:49:02 crc restorecon[4764]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/f081611a.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 13 06:49:02 crc restorecon[4764]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/48a195d8.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 13 06:49:02 crc restorecon[4764]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/GDCA_TrustAUTH_R5_ROOT.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 13 06:49:02 crc restorecon[4764]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/0f6fa695.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 13 06:49:02 crc restorecon[4764]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ab59055e.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 13 06:49:02 crc restorecon[4764]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b92fd57f.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 13 06:49:02 crc restorecon[4764]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/GLOBALTRUST_2020.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 13 06:49:02 crc restorecon[4764]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/fa5da96b.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 13 06:49:02 crc restorecon[4764]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/1ec40989.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 13 06:49:02 crc restorecon[4764]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/7719f463.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 13 06:49:02 crc restorecon[4764]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/GTS_Root_R1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 13 06:49:02 crc restorecon[4764]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/1001acf7.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 13 06:49:02 crc restorecon[4764]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/f013ecaf.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 13 06:49:02 crc restorecon[4764]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/626dceaf.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 13 06:49:02 crc restorecon[4764]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/c559d742.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 13 06:49:02 crc restorecon[4764]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/1d3472b9.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 13 06:49:02 crc restorecon[4764]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9479c8c3.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 13 06:49:02 crc restorecon[4764]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/a81e292b.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 13 06:49:02 crc restorecon[4764]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/4bfab552.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 13 06:49:02 crc restorecon[4764]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Go_Daddy_Class_2_Certification_Authority.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 13 06:49:02 crc restorecon[4764]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Sectigo_Public_Server_Authentication_Root_E46.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 13 06:49:02 crc restorecon[4764]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Go_Daddy_Root_Certificate_Authority_-_G2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 13 06:49:02 crc restorecon[4764]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/e071171e.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 13 06:49:02 crc restorecon[4764]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/57bcb2da.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 13 06:49:02 crc restorecon[4764]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/HARICA_TLS_ECC_Root_CA_2021.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 13 06:49:02 crc restorecon[4764]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ab5346f4.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 13 06:49:02 crc restorecon[4764]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/5046c355.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 13 06:49:02 crc restorecon[4764]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/HARICA_TLS_RSA_Root_CA_2021.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 13 06:49:02 crc restorecon[4764]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/865fbdf9.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 13 06:49:02 crc restorecon[4764]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/da0cfd1d.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 13 06:49:02 crc restorecon[4764]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/85cde254.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 13 06:49:02 crc restorecon[4764]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Hellenic_Academic_and_Research_Institutions_ECC_RootCA_2015.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 13 06:49:02 crc restorecon[4764]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/cbb3f32b.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 13 06:49:02 crc restorecon[4764]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/SecureSign_RootCA11.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 13 06:49:02 crc restorecon[4764]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Hellenic_Academic_and_Research_Institutions_RootCA_2015.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 13 06:49:02 crc restorecon[4764]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/5860aaa6.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 13 06:49:02 crc restorecon[4764]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/31188b5e.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 13 06:49:02 crc restorecon[4764]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/HiPKI_Root_CA_-_G1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 13 06:49:02 crc restorecon[4764]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/c7f1359b.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 13 06:49:02 crc restorecon[4764]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/5f15c80c.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 13 06:49:02 crc restorecon[4764]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Hongkong_Post_Root_CA_3.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 13 06:49:02 crc restorecon[4764]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/09789157.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 13 06:49:02 crc restorecon[4764]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ISRG_Root_X2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 13 06:49:02 crc restorecon[4764]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/18856ac4.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 13 06:49:02 crc restorecon[4764]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/1e09d511.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 13 06:49:02 crc restorecon[4764]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/IdenTrust_Commercial_Root_CA_1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 13 06:49:02 crc restorecon[4764]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/cf701eeb.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 13 06:49:02 crc restorecon[4764]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/d06393bb.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 13 06:49:02 crc restorecon[4764]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/IdenTrust_Public_Sector_Root_CA_1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 13 06:49:02 crc restorecon[4764]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/10531352.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 13 06:49:02 crc restorecon[4764]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Izenpe.com.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 13 06:49:02 crc restorecon[4764]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/SecureTrust_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 13 06:49:02 crc restorecon[4764]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b0ed035a.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 13 06:49:02 crc restorecon[4764]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Microsec_e-Szigno_Root_CA_2009.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 13 06:49:02 crc restorecon[4764]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/8160b96c.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 13 06:49:02 crc restorecon[4764]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/e8651083.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 13 06:49:02 crc restorecon[4764]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/2c63f966.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 13 06:49:02 crc restorecon[4764]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Security_Communication_RootCA2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 13 06:49:02 crc restorecon[4764]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Microsoft_ECC_Root_Certificate_Authority_2017.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 13 06:49:02 crc restorecon[4764]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/8d89cda1.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 13 06:49:02 crc restorecon[4764]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/01419da9.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 13 06:49:02 crc restorecon[4764]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/SSL.com_TLS_RSA_Root_CA_2022.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 13 06:49:02 crc restorecon[4764]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b7a5b843.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 13 06:49:02 crc restorecon[4764]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Microsoft_RSA_Root_Certificate_Authority_2017.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 13 06:49:02 crc restorecon[4764]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/bf53fb88.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 13 06:49:02 crc restorecon[4764]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9591a472.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 13 06:49:02 crc restorecon[4764]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/3afde786.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 13 06:49:02 crc restorecon[4764]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/SwissSign_Gold_CA_-_G2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 13 06:49:02 crc restorecon[4764]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/NAVER_Global_Root_Certification_Authority.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 13 06:49:02 crc restorecon[4764]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/3fb36b73.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 13 06:49:02 crc restorecon[4764]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/d39b0a2c.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 13 06:49:02 crc restorecon[4764]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/a89d74c2.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 13 06:49:02 crc restorecon[4764]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/cd58d51e.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 13 06:49:02 crc restorecon[4764]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b7db1890.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 13 06:49:02 crc restorecon[4764]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/NetLock_Arany__Class_Gold__F__tan__s__tv__ny.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 13 06:49:02 crc restorecon[4764]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/988a38cb.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 13 06:49:02 crc restorecon[4764]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/60afe812.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 13 06:49:02 crc restorecon[4764]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/f39fc864.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 13 06:49:02 crc restorecon[4764]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/5443e9e3.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 13 06:49:02 crc restorecon[4764]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/OISTE_WISeKey_Global_Root_GB_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 13 06:49:02 crc restorecon[4764]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/e73d606e.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 13 06:49:02 crc restorecon[4764]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/dfc0fe80.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 13 06:49:02 crc restorecon[4764]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b66938e9.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 13 06:49:02 crc restorecon[4764]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/1e1eab7c.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 13 06:49:02 crc restorecon[4764]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/OISTE_WISeKey_Global_Root_GC_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 13 06:49:02 crc restorecon[4764]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/773e07ad.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 13 06:49:02 crc restorecon[4764]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/3c899c73.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 13 06:49:02 crc restorecon[4764]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/d59297b8.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 13 06:49:02 crc restorecon[4764]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ddcda989.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 13 06:49:02 crc restorecon[4764]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/QuoVadis_Root_CA_1_G3.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 13 06:49:02 crc restorecon[4764]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/749e9e03.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 13 06:49:02 crc restorecon[4764]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/52b525c7.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 13 06:49:02 crc restorecon[4764]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Security_Communication_RootCA3.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 13 06:49:02 crc restorecon[4764]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/QuoVadis_Root_CA_2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 13 06:49:02 crc restorecon[4764]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/d7e8dc79.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 13 06:49:02 crc restorecon[4764]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/7a819ef2.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 13 06:49:02 crc restorecon[4764]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/08063a00.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 13 06:49:02 crc restorecon[4764]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/6b483515.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 13 06:49:02 crc restorecon[4764]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/QuoVadis_Root_CA_2_G3.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 13 06:49:02 crc restorecon[4764]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/064e0aa9.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 13 06:49:02 crc restorecon[4764]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/1f58a078.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 13 06:49:02 crc restorecon[4764]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/6f7454b3.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 13 06:49:02 crc restorecon[4764]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/7fa05551.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 13 06:49:02 crc restorecon[4764]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/QuoVadis_Root_CA_3.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 13 06:49:02 crc restorecon[4764]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/76faf6c0.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 13 06:49:02 crc restorecon[4764]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9339512a.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 13 06:49:02 crc restorecon[4764]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/f387163d.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 13 06:49:02 crc restorecon[4764]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ee37c333.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 13 06:49:02 crc restorecon[4764]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/QuoVadis_Root_CA_3_G3.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 13 06:49:02 crc restorecon[4764]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/e18bfb83.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 13 06:49:02 crc restorecon[4764]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/e442e424.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 13 06:49:02 crc restorecon[4764]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/fe8a2cd8.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 13 06:49:02 crc restorecon[4764]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/23f4c490.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 13 06:49:02 crc restorecon[4764]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/5cd81ad7.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 13 06:49:02 crc restorecon[4764]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/SSL.com_EV_Root_Certification_Authority_ECC.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 13 06:49:02 crc restorecon[4764]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/f0c70a8d.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 13 06:49:03 crc restorecon[4764]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/7892ad52.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 13 06:49:03 crc restorecon[4764]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/SZAFIR_ROOT_CA2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 13 06:49:03 crc restorecon[4764]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/4f316efb.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 13 06:49:03 crc restorecon[4764]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/SSL.com_EV_Root_Certification_Authority_RSA_R2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 13 06:49:03 crc restorecon[4764]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/06dc52d5.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 13 06:49:03 crc restorecon[4764]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/583d0756.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 13 06:49:03 crc restorecon[4764]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Sectigo_Public_Server_Authentication_Root_R46.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 13 06:49:03 crc restorecon[4764]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/SSL.com_Root_Certification_Authority_ECC.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 13 06:49:03 crc restorecon[4764]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/0bf05006.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 13 06:49:03 crc restorecon[4764]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/88950faa.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 13 06:49:03 crc restorecon[4764]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9046744a.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 13 06:49:03 crc restorecon[4764]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/3c860d51.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 13 06:49:03 crc restorecon[4764]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/SSL.com_Root_Certification_Authority_RSA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 13 06:49:03 crc restorecon[4764]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/6fa5da56.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 13 06:49:03 crc restorecon[4764]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/33ee480d.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 13 06:49:03 crc restorecon[4764]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Secure_Global_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 13 06:49:03 crc restorecon[4764]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/63a2c897.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 13 06:49:03 crc restorecon[4764]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/SSL.com_TLS_ECC_Root_CA_2022.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 13 06:49:03 crc restorecon[4764]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/bdacca6f.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 13 06:49:03 crc restorecon[4764]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ff34af3f.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 13 06:49:03 crc restorecon[4764]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/dbff3a01.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 13 06:49:03 crc restorecon[4764]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Security_Communication_ECC_RootCA1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 13 06:49:03 crc restorecon[4764]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/emSign_Root_CA_-_C1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 13 06:49:03 crc restorecon[4764]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Starfield_Class_2_Certification_Authority.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 13 06:49:03 crc restorecon[4764]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/406c9bb1.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 13 06:49:03 crc restorecon[4764]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Starfield_Root_Certificate_Authority_-_G2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 13 06:49:03 crc restorecon[4764]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/emSign_ECC_Root_CA_-_C3.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 13 06:49:03 crc restorecon[4764]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Starfield_Services_Root_Certificate_Authority_-_G2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 13 06:49:03 crc restorecon[4764]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/SwissSign_Silver_CA_-_G2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 13 06:49:03 crc restorecon[4764]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/99e1b953.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 13 06:49:03 crc restorecon[4764]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/T-TeleSec_GlobalRoot_Class_2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 13 06:49:03 crc restorecon[4764]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/vTrus_Root_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 13 06:49:03 crc restorecon[4764]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/T-TeleSec_GlobalRoot_Class_3.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 13 06:49:03 crc restorecon[4764]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/14bc7599.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 13 06:49:03 crc restorecon[4764]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/TUBITAK_Kamu_SM_SSL_Kok_Sertifikasi_-_Surum_1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 13 06:49:03 crc restorecon[4764]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/TWCA_Global_Root_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 13 06:49:03 crc restorecon[4764]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/7a3adc42.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 13 06:49:03 crc restorecon[4764]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/TWCA_Root_Certification_Authority.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 13 06:49:03 crc restorecon[4764]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/f459871d.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 13 06:49:03 crc restorecon[4764]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Telekom_Security_TLS_ECC_Root_2020.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 13 06:49:03 crc restorecon[4764]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/emSign_Root_CA_-_G1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 13 06:49:03 crc restorecon[4764]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Telekom_Security_TLS_RSA_Root_2023.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 13 06:49:03 crc restorecon[4764]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/TeliaSonera_Root_CA_v1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 13 06:49:03 crc restorecon[4764]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Telia_Root_CA_v2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 13 06:49:03 crc restorecon[4764]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/8f103249.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 13 06:49:03 crc restorecon[4764]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/f058632f.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 13 06:49:03 crc restorecon[4764]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ca-certificates.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 13 06:49:03 crc restorecon[4764]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/TrustAsia_Global_Root_CA_G3.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 13 06:49:03 crc restorecon[4764]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9bf03295.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 13 06:49:03 crc restorecon[4764]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/98aaf404.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 13 06:49:03 crc restorecon[4764]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 13 06:49:03 crc restorecon[4764]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/TrustAsia_Global_Root_CA_G4.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 13 06:49:03 crc restorecon[4764]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/1cef98f5.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 13 06:49:03 crc restorecon[4764]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/073bfcc5.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 13 06:49:03 crc restorecon[4764]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/2923b3f9.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 13 06:49:03 crc restorecon[4764]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Trustwave_Global_Certification_Authority.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 13 06:49:03 crc restorecon[4764]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/f249de83.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 13 06:49:03 crc restorecon[4764]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/edcbddb5.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 13 06:49:03 crc restorecon[4764]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/emSign_ECC_Root_CA_-_G3.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 13 06:49:03 crc restorecon[4764]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Trustwave_Global_ECC_P256_Certification_Authority.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 13 06:49:03 crc restorecon[4764]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9b5697b0.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 13 06:49:03 crc restorecon[4764]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/1ae85e5e.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 13 06:49:03 crc restorecon[4764]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b74d2bd5.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 13 06:49:03 crc restorecon[4764]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Trustwave_Global_ECC_P384_Certification_Authority.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 13 06:49:03 crc restorecon[4764]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/d887a5bb.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 13 06:49:03 crc restorecon[4764]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9aef356c.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 13 06:49:03 crc restorecon[4764]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/TunTrust_Root_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 13 06:49:03 crc restorecon[4764]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/fd64f3fc.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 13 06:49:03 crc restorecon[4764]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/e13665f9.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 13 06:49:03 crc restorecon[4764]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/UCA_Extended_Validation_Root.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 13 06:49:03 crc restorecon[4764]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/0f5dc4f3.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 13 06:49:03 crc restorecon[4764]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/da7377f6.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 13 06:49:03 crc restorecon[4764]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/UCA_Global_G2_Root.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 13 06:49:03 crc restorecon[4764]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/c01eb047.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 13 06:49:03 crc restorecon[4764]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/304d27c3.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 13 06:49:03 crc restorecon[4764]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ed858448.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 13 06:49:03 crc restorecon[4764]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/USERTrust_ECC_Certification_Authority.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 13 06:49:03 crc restorecon[4764]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/f30dd6ad.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 13 06:49:03 crc restorecon[4764]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/04f60c28.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 13 06:49:03 crc restorecon[4764]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/vTrus_ECC_Root_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 13 06:49:03 crc restorecon[4764]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/USERTrust_RSA_Certification_Authority.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 13 06:49:03 crc restorecon[4764]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/fc5a8f99.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 13 06:49:03 crc restorecon[4764]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/35105088.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 13 06:49:03 crc restorecon[4764]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ee532fd5.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 13 06:49:03 crc restorecon[4764]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/XRamp_Global_Certification_Authority.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 13 06:49:03 crc restorecon[4764]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/706f604c.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 13 06:49:03 crc restorecon[4764]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/76579174.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 13 06:49:03 crc restorecon[4764]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/certSIGN_ROOT_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 13 06:49:03 crc restorecon[4764]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/8d86cdd1.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 13 06:49:03 crc restorecon[4764]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/882de061.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 13 06:49:03 crc restorecon[4764]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/certSIGN_ROOT_CA_G2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 13 06:49:03 crc restorecon[4764]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/5f618aec.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 13 06:49:03 crc restorecon[4764]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/a9d40e02.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 13 06:49:03 crc restorecon[4764]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/e-Szigno_Root_CA_2017.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 13 06:49:03 crc restorecon[4764]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/e868b802.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 13 06:49:03 crc restorecon[4764]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/83e9984f.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 13 06:49:03 crc restorecon[4764]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ePKI_Root_Certification_Authority.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 13 06:49:03 crc restorecon[4764]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ca6e4ad9.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 13 06:49:03 crc restorecon[4764]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9d6523ce.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 13 06:49:03 crc restorecon[4764]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/4b718d9b.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 13 06:49:03 crc restorecon[4764]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/869fbf79.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 13 06:49:03 crc restorecon[4764]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 13 06:49:03 crc restorecon[4764]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/containers/registry/f8d22bdb not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 13 06:49:03 crc restorecon[4764]: /var/lib/kubelet/pods/a0128f3a-b052-44ed-a84e-c4c8aaf17c13/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c17 Dec 13 06:49:03 crc restorecon[4764]: /var/lib/kubelet/pods/a0128f3a-b052-44ed-a84e-c4c8aaf17c13/containers/cluster-samples-operator/6e8bbfac not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c17 Dec 13 06:49:03 crc restorecon[4764]: /var/lib/kubelet/pods/a0128f3a-b052-44ed-a84e-c4c8aaf17c13/containers/cluster-samples-operator/54dd7996 not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c17 Dec 13 06:49:03 crc restorecon[4764]: /var/lib/kubelet/pods/a0128f3a-b052-44ed-a84e-c4c8aaf17c13/containers/cluster-samples-operator/a4f1bb05 not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c17 Dec 13 06:49:03 crc restorecon[4764]: /var/lib/kubelet/pods/a0128f3a-b052-44ed-a84e-c4c8aaf17c13/containers/cluster-samples-operator-watch/207129da not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c17 Dec 13 06:49:03 crc restorecon[4764]: /var/lib/kubelet/pods/a0128f3a-b052-44ed-a84e-c4c8aaf17c13/containers/cluster-samples-operator-watch/c1df39e1 not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c17 Dec 13 06:49:03 crc restorecon[4764]: /var/lib/kubelet/pods/a0128f3a-b052-44ed-a84e-c4c8aaf17c13/containers/cluster-samples-operator-watch/15b8f1cd not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c17 Dec 13 06:49:03 crc restorecon[4764]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/volumes/kubernetes.io~configmap/config not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Dec 13 06:49:03 crc restorecon[4764]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/volumes/kubernetes.io~configmap/config/..2025_02_23_05_27_49.3523263858 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Dec 13 06:49:03 crc restorecon[4764]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/volumes/kubernetes.io~configmap/config/..2025_02_23_05_27_49.3523263858/config-file.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Dec 13 06:49:03 crc restorecon[4764]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/volumes/kubernetes.io~configmap/config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Dec 13 06:49:03 crc restorecon[4764]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/volumes/kubernetes.io~configmap/config/config-file.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Dec 13 06:49:03 crc restorecon[4764]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/volumes/kubernetes.io~configmap/images not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Dec 13 06:49:03 crc restorecon[4764]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/volumes/kubernetes.io~configmap/images/..2025_02_23_05_27_49.3256605594 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Dec 13 06:49:03 crc restorecon[4764]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/volumes/kubernetes.io~configmap/images/..2025_02_23_05_27_49.3256605594/images.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Dec 13 06:49:03 crc restorecon[4764]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/volumes/kubernetes.io~configmap/images/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Dec 13 06:49:03 crc restorecon[4764]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/volumes/kubernetes.io~configmap/images/images.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Dec 13 06:49:03 crc restorecon[4764]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Dec 13 06:49:03 crc restorecon[4764]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/containers/kube-rbac-proxy/77bd6913 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Dec 13 06:49:03 crc restorecon[4764]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/containers/kube-rbac-proxy/2382c1b1 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Dec 13 06:49:03 crc restorecon[4764]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/containers/kube-rbac-proxy/704ce128 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Dec 13 06:49:03 crc restorecon[4764]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/containers/machine-api-operator/70d16fe0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Dec 13 06:49:03 crc restorecon[4764]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/containers/machine-api-operator/bfb95535 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Dec 13 06:49:03 crc restorecon[4764]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/containers/machine-api-operator/57a8e8e2 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Dec 13 06:49:03 crc restorecon[4764]: /var/lib/kubelet/pods/e7e6199b-1264-4501-8953-767f51328d08/volumes/kubernetes.io~configmap/config not reset as customized by admin to system_u:object_r:container_file_t:s0:c219,c404 Dec 13 06:49:03 crc restorecon[4764]: /var/lib/kubelet/pods/e7e6199b-1264-4501-8953-767f51328d08/volumes/kubernetes.io~configmap/config/..2025_02_23_05_27_49.3413793711 not reset as customized by admin to system_u:object_r:container_file_t:s0:c219,c404 Dec 13 06:49:03 crc restorecon[4764]: /var/lib/kubelet/pods/e7e6199b-1264-4501-8953-767f51328d08/volumes/kubernetes.io~configmap/config/..2025_02_23_05_27_49.3413793711/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c219,c404 Dec 13 06:49:03 crc restorecon[4764]: /var/lib/kubelet/pods/e7e6199b-1264-4501-8953-767f51328d08/volumes/kubernetes.io~configmap/config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c219,c404 Dec 13 06:49:03 crc restorecon[4764]: /var/lib/kubelet/pods/e7e6199b-1264-4501-8953-767f51328d08/volumes/kubernetes.io~configmap/config/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c219,c404 Dec 13 06:49:03 crc restorecon[4764]: /var/lib/kubelet/pods/e7e6199b-1264-4501-8953-767f51328d08/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c219,c404 Dec 13 06:49:03 crc restorecon[4764]: /var/lib/kubelet/pods/e7e6199b-1264-4501-8953-767f51328d08/containers/kube-apiserver-operator/1b9d3e5e not reset as customized by admin to system_u:object_r:container_file_t:s0:c107,c917 Dec 13 06:49:03 crc restorecon[4764]: /var/lib/kubelet/pods/e7e6199b-1264-4501-8953-767f51328d08/containers/kube-apiserver-operator/fddb173c not reset as customized by admin to system_u:object_r:container_file_t:s0:c202,c983 Dec 13 06:49:03 crc restorecon[4764]: /var/lib/kubelet/pods/e7e6199b-1264-4501-8953-767f51328d08/containers/kube-apiserver-operator/95d3c6c4 not reset as customized by admin to system_u:object_r:container_file_t:s0:c219,c404 Dec 13 06:49:03 crc restorecon[4764]: /var/lib/kubelet/pods/9d751cbb-f2e2-430d-9754-c882a5e924a5/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c20,c21 Dec 13 06:49:03 crc restorecon[4764]: /var/lib/kubelet/pods/9d751cbb-f2e2-430d-9754-c882a5e924a5/containers/check-endpoints/bfb5fff5 not reset as customized by admin to system_u:object_r:container_file_t:s0:c20,c21 Dec 13 06:49:03 crc restorecon[4764]: /var/lib/kubelet/pods/9d751cbb-f2e2-430d-9754-c882a5e924a5/containers/check-endpoints/2aef40aa not reset as customized by admin to system_u:object_r:container_file_t:s0:c20,c21 Dec 13 06:49:03 crc restorecon[4764]: /var/lib/kubelet/pods/9d751cbb-f2e2-430d-9754-c882a5e924a5/containers/check-endpoints/c0391cad not reset as customized by admin to system_u:object_r:container_file_t:s0:c20,c21 Dec 13 06:49:03 crc restorecon[4764]: /var/lib/kubelet/pods/f614b9022728cf315e60c057852e563e/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c214,c928 Dec 13 06:49:03 crc restorecon[4764]: /var/lib/kubelet/pods/f614b9022728cf315e60c057852e563e/containers/kube-controller-manager/1119e69d not reset as customized by admin to system_u:object_r:container_file_t:s0:c776,c1007 Dec 13 06:49:03 crc restorecon[4764]: /var/lib/kubelet/pods/f614b9022728cf315e60c057852e563e/containers/kube-controller-manager/660608b4 not reset as customized by admin to system_u:object_r:container_file_t:s0:c214,c928 Dec 13 06:49:03 crc restorecon[4764]: /var/lib/kubelet/pods/f614b9022728cf315e60c057852e563e/containers/kube-controller-manager/8220bd53 not reset as customized by admin to system_u:object_r:container_file_t:s0:c214,c928 Dec 13 06:49:03 crc restorecon[4764]: /var/lib/kubelet/pods/f614b9022728cf315e60c057852e563e/containers/cluster-policy-controller/85f99d5c not reset as customized by admin to system_u:object_r:container_file_t:s0:c776,c1007 Dec 13 06:49:03 crc restorecon[4764]: /var/lib/kubelet/pods/f614b9022728cf315e60c057852e563e/containers/cluster-policy-controller/4b0225f6 not reset as customized by admin to system_u:object_r:container_file_t:s0:c214,c928 Dec 13 06:49:03 crc restorecon[4764]: /var/lib/kubelet/pods/f614b9022728cf315e60c057852e563e/containers/kube-controller-manager-cert-syncer/9c2a3394 not reset as customized by admin to system_u:object_r:container_file_t:s0:c776,c1007 Dec 13 06:49:03 crc restorecon[4764]: /var/lib/kubelet/pods/f614b9022728cf315e60c057852e563e/containers/kube-controller-manager-cert-syncer/e820b243 not reset as customized by admin to system_u:object_r:container_file_t:s0:c214,c928 Dec 13 06:49:03 crc restorecon[4764]: /var/lib/kubelet/pods/f614b9022728cf315e60c057852e563e/containers/kube-controller-manager-recovery-controller/1ca52ea0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c776,c1007 Dec 13 06:49:03 crc restorecon[4764]: /var/lib/kubelet/pods/f614b9022728cf315e60c057852e563e/containers/kube-controller-manager-recovery-controller/e6988e45 not reset as customized by admin to system_u:object_r:container_file_t:s0:c214,c928 Dec 13 06:49:03 crc restorecon[4764]: /var/lib/kubelet/pods/0b574797-001e-440a-8f4e-c0be86edad0f/volumes/kubernetes.io~configmap/mcc-auth-proxy-config not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Dec 13 06:49:03 crc restorecon[4764]: /var/lib/kubelet/pods/0b574797-001e-440a-8f4e-c0be86edad0f/volumes/kubernetes.io~configmap/mcc-auth-proxy-config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Dec 13 06:49:03 crc restorecon[4764]: /var/lib/kubelet/pods/0b574797-001e-440a-8f4e-c0be86edad0f/volumes/kubernetes.io~configmap/mcc-auth-proxy-config/config-file.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Dec 13 06:49:03 crc restorecon[4764]: /var/lib/kubelet/pods/0b574797-001e-440a-8f4e-c0be86edad0f/volumes/kubernetes.io~configmap/mcc-auth-proxy-config/..2025_02_24_06_09_21.2517297950 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Dec 13 06:49:03 crc restorecon[4764]: /var/lib/kubelet/pods/0b574797-001e-440a-8f4e-c0be86edad0f/volumes/kubernetes.io~configmap/mcc-auth-proxy-config/..2025_02_24_06_09_21.2517297950/config-file.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Dec 13 06:49:03 crc restorecon[4764]: /var/lib/kubelet/pods/0b574797-001e-440a-8f4e-c0be86edad0f/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Dec 13 06:49:03 crc restorecon[4764]: /var/lib/kubelet/pods/0b574797-001e-440a-8f4e-c0be86edad0f/containers/machine-config-controller/6655f00b not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Dec 13 06:49:03 crc restorecon[4764]: /var/lib/kubelet/pods/0b574797-001e-440a-8f4e-c0be86edad0f/containers/machine-config-controller/98bc3986 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Dec 13 06:49:03 crc restorecon[4764]: /var/lib/kubelet/pods/0b574797-001e-440a-8f4e-c0be86edad0f/containers/machine-config-controller/08e3458a not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Dec 13 06:49:03 crc restorecon[4764]: /var/lib/kubelet/pods/0b574797-001e-440a-8f4e-c0be86edad0f/containers/kube-rbac-proxy/2a191cb0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Dec 13 06:49:03 crc restorecon[4764]: /var/lib/kubelet/pods/0b574797-001e-440a-8f4e-c0be86edad0f/containers/kube-rbac-proxy/6c4eeefb not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Dec 13 06:49:03 crc restorecon[4764]: /var/lib/kubelet/pods/0b574797-001e-440a-8f4e-c0be86edad0f/containers/kube-rbac-proxy/f61a549c not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Dec 13 06:49:03 crc restorecon[4764]: /var/lib/kubelet/pods/bd23aa5c-e532-4e53-bccf-e79f130c5ae8/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c318,c553 Dec 13 06:49:03 crc restorecon[4764]: /var/lib/kubelet/pods/bd23aa5c-e532-4e53-bccf-e79f130c5ae8/containers/hostpath-provisioner/24891863 not reset as customized by admin to system_u:object_r:container_file_t:s0:c37,c572 Dec 13 06:49:03 crc restorecon[4764]: /var/lib/kubelet/pods/bd23aa5c-e532-4e53-bccf-e79f130c5ae8/containers/hostpath-provisioner/fbdfd89c not reset as customized by admin to system_u:object_r:container_file_t:s0:c318,c553 Dec 13 06:49:03 crc restorecon[4764]: /var/lib/kubelet/pods/bd23aa5c-e532-4e53-bccf-e79f130c5ae8/containers/liveness-probe/9b63b3bc not reset as customized by admin to system_u:object_r:container_file_t:s0:c37,c572 Dec 13 06:49:03 crc restorecon[4764]: /var/lib/kubelet/pods/bd23aa5c-e532-4e53-bccf-e79f130c5ae8/containers/liveness-probe/8acde6d6 not reset as customized by admin to system_u:object_r:container_file_t:s0:c318,c553 Dec 13 06:49:03 crc restorecon[4764]: /var/lib/kubelet/pods/bd23aa5c-e532-4e53-bccf-e79f130c5ae8/containers/node-driver-registrar/59ecbba3 not reset as customized by admin to system_u:object_r:container_file_t:s0:c318,c553 Dec 13 06:49:03 crc restorecon[4764]: /var/lib/kubelet/pods/bd23aa5c-e532-4e53-bccf-e79f130c5ae8/containers/csi-provisioner/685d4be3 not reset as customized by admin to system_u:object_r:container_file_t:s0:c318,c553 Dec 13 06:49:03 crc restorecon[4764]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/volumes/kubernetes.io~configmap/config not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Dec 13 06:49:03 crc restorecon[4764]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/volumes/kubernetes.io~configmap/config/..2025_02_24_06_20_07.341639300 not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Dec 13 06:49:03 crc restorecon[4764]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/volumes/kubernetes.io~configmap/config/..2025_02_24_06_20_07.341639300/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Dec 13 06:49:03 crc restorecon[4764]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/volumes/kubernetes.io~configmap/config/..2025_02_24_06_20_07.341639300/openshift-route-controller-manager.client-ca.configmap not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Dec 13 06:49:03 crc restorecon[4764]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/volumes/kubernetes.io~configmap/config/..2025_02_24_06_20_07.341639300/openshift-route-controller-manager.serving-cert.secret not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Dec 13 06:49:03 crc restorecon[4764]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/volumes/kubernetes.io~configmap/config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Dec 13 06:49:03 crc restorecon[4764]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/volumes/kubernetes.io~configmap/config/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Dec 13 06:49:03 crc restorecon[4764]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/volumes/kubernetes.io~configmap/config/openshift-route-controller-manager.client-ca.configmap not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Dec 13 06:49:03 crc restorecon[4764]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/volumes/kubernetes.io~configmap/config/openshift-route-controller-manager.serving-cert.secret not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Dec 13 06:49:03 crc restorecon[4764]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/volumes/kubernetes.io~configmap/client-ca not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Dec 13 06:49:03 crc restorecon[4764]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/volumes/kubernetes.io~configmap/client-ca/..2025_02_24_06_20_07.2950937851 not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Dec 13 06:49:03 crc restorecon[4764]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/volumes/kubernetes.io~configmap/client-ca/..2025_02_24_06_20_07.2950937851/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Dec 13 06:49:03 crc restorecon[4764]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/volumes/kubernetes.io~configmap/client-ca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Dec 13 06:49:03 crc restorecon[4764]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/volumes/kubernetes.io~configmap/client-ca/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Dec 13 06:49:03 crc restorecon[4764]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Dec 13 06:49:03 crc restorecon[4764]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/containers/route-controller-manager/feaea55e not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Dec 13 06:49:03 crc restorecon[4764]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 13 06:49:03 crc restorecon[4764]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 13 06:49:03 crc restorecon[4764]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/abinitio-runtime-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 13 06:49:03 crc restorecon[4764]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/abinitio-runtime-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 13 06:49:03 crc restorecon[4764]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/accuknox-operator-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 13 06:49:03 crc restorecon[4764]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/accuknox-operator-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 13 06:49:03 crc restorecon[4764]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aci-containers-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 13 06:49:03 crc restorecon[4764]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aci-containers-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 13 06:49:03 crc restorecon[4764]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aikit-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 13 06:49:03 crc restorecon[4764]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aikit-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 13 06:49:03 crc restorecon[4764]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/airlock-microgateway not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 13 06:49:03 crc restorecon[4764]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/airlock-microgateway/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 13 06:49:03 crc restorecon[4764]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ako-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 13 06:49:03 crc restorecon[4764]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ako-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 13 06:49:03 crc restorecon[4764]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/alloy not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 13 06:49:03 crc restorecon[4764]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/alloy/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 13 06:49:03 crc restorecon[4764]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/anchore-engine not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 13 06:49:03 crc restorecon[4764]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/anchore-engine/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 13 06:49:03 crc restorecon[4764]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/anzo-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 13 06:49:03 crc restorecon[4764]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/anzo-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 13 06:49:03 crc restorecon[4764]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/anzograph-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 13 06:49:03 crc restorecon[4764]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/anzograph-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 13 06:49:03 crc restorecon[4764]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/anzounstructured-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 13 06:49:03 crc restorecon[4764]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/anzounstructured-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 13 06:49:03 crc restorecon[4764]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/appdynamics-cloud-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 13 06:49:03 crc restorecon[4764]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/appdynamics-cloud-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 13 06:49:03 crc restorecon[4764]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/appdynamics-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 13 06:49:03 crc restorecon[4764]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/appdynamics-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 13 06:49:03 crc restorecon[4764]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aqua-operator-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 13 06:49:03 crc restorecon[4764]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aqua-operator-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 13 06:49:03 crc restorecon[4764]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cass-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 13 06:49:03 crc restorecon[4764]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cass-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 13 06:49:03 crc restorecon[4764]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ccm-node-agent-dcap-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 13 06:49:03 crc restorecon[4764]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ccm-node-agent-dcap-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 13 06:49:03 crc restorecon[4764]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ccm-node-agent-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 13 06:49:03 crc restorecon[4764]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ccm-node-agent-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 13 06:49:03 crc restorecon[4764]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cfm-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 13 06:49:03 crc restorecon[4764]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cfm-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 13 06:49:03 crc restorecon[4764]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cilium not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 13 06:49:03 crc restorecon[4764]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cilium/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 13 06:49:03 crc restorecon[4764]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cilium-enterprise not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 13 06:49:03 crc restorecon[4764]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cilium-enterprise/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 13 06:49:03 crc restorecon[4764]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cloud-native-postgresql not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 13 06:49:03 crc restorecon[4764]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cloud-native-postgresql/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 13 06:49:03 crc restorecon[4764]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cloudbees-ci not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 13 06:49:03 crc restorecon[4764]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cloudbees-ci/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 13 06:49:03 crc restorecon[4764]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cloudera-streams-messaging-kubernetes-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 13 06:49:03 crc restorecon[4764]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cloudera-streams-messaging-kubernetes-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 13 06:49:03 crc restorecon[4764]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cloudnative-pg not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 13 06:49:03 crc restorecon[4764]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cloudnative-pg/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 13 06:49:03 crc restorecon[4764]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cnfv-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 13 06:49:03 crc restorecon[4764]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cnfv-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 13 06:49:03 crc restorecon[4764]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cockroachdb-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 13 06:49:03 crc restorecon[4764]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cockroachdb-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 13 06:49:03 crc restorecon[4764]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/conjur-follower-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 13 06:49:03 crc restorecon[4764]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/conjur-follower-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 13 06:49:03 crc restorecon[4764]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/coroot-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 13 06:49:03 crc restorecon[4764]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/coroot-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 13 06:49:03 crc restorecon[4764]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/crunchy-postgres-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 13 06:49:03 crc restorecon[4764]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/crunchy-postgres-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 13 06:49:03 crc restorecon[4764]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cte-k8s-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 13 06:49:03 crc restorecon[4764]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cte-k8s-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 13 06:49:03 crc restorecon[4764]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/datadog-operator-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 13 06:49:03 crc restorecon[4764]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/datadog-operator-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 13 06:49:03 crc restorecon[4764]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dell-csm-operator-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 13 06:49:03 crc restorecon[4764]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dell-csm-operator-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 13 06:49:03 crc restorecon[4764]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/digitalai-deploy-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 13 06:49:03 crc restorecon[4764]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/digitalai-deploy-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 13 06:49:03 crc restorecon[4764]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/digitalai-release-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 13 06:49:03 crc restorecon[4764]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/digitalai-release-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 13 06:49:03 crc restorecon[4764]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dynatrace-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 13 06:49:03 crc restorecon[4764]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dynatrace-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 13 06:49:03 crc restorecon[4764]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/edb-hcp-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 13 06:49:03 crc restorecon[4764]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/edb-hcp-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 13 06:49:03 crc restorecon[4764]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/eginnovations-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 13 06:49:03 crc restorecon[4764]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/eginnovations-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 13 06:49:03 crc restorecon[4764]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/elasticsearch-eck-operator-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 13 06:49:03 crc restorecon[4764]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/elasticsearch-eck-operator-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 13 06:49:03 crc restorecon[4764]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/falcon-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 13 06:49:03 crc restorecon[4764]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/falcon-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 13 06:49:03 crc restorecon[4764]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/federatorai-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 13 06:49:03 crc restorecon[4764]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/federatorai-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 13 06:49:03 crc restorecon[4764]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/fujitsu-enterprise-postgres-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 13 06:49:03 crc restorecon[4764]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/fujitsu-enterprise-postgres-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 13 06:49:03 crc restorecon[4764]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/function-mesh not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 13 06:49:03 crc restorecon[4764]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/function-mesh/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 13 06:49:03 crc restorecon[4764]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/harness-gitops-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 13 06:49:03 crc restorecon[4764]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/harness-gitops-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 13 06:49:03 crc restorecon[4764]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hazelcast-platform-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 13 06:49:03 crc restorecon[4764]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hazelcast-platform-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 13 06:49:03 crc restorecon[4764]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hcp-terraform-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 13 06:49:03 crc restorecon[4764]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hcp-terraform-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 13 06:49:03 crc restorecon[4764]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hpe-ezmeral-csi-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 13 06:49:03 crc restorecon[4764]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hpe-ezmeral-csi-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 13 06:49:03 crc restorecon[4764]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-application-gateway-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 13 06:49:03 crc restorecon[4764]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-application-gateway-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 13 06:49:03 crc restorecon[4764]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-block-csi-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 13 06:49:03 crc restorecon[4764]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-block-csi-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 13 06:49:03 crc restorecon[4764]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-security-verify-access-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 13 06:49:03 crc restorecon[4764]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-security-verify-access-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 13 06:49:03 crc restorecon[4764]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-security-verify-directory-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 13 06:49:03 crc restorecon[4764]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-security-verify-directory-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 13 06:49:03 crc restorecon[4764]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-security-verify-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 13 06:49:03 crc restorecon[4764]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-security-verify-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 13 06:49:03 crc restorecon[4764]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/infoscale-dr-manager not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 13 06:49:03 crc restorecon[4764]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/infoscale-dr-manager/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 13 06:49:03 crc restorecon[4764]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/infoscale-licensing-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 13 06:49:03 crc restorecon[4764]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/infoscale-licensing-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 13 06:49:03 crc restorecon[4764]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/infoscale-sds-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 13 06:49:03 crc restorecon[4764]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/infoscale-sds-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 13 06:49:03 crc restorecon[4764]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/infrastructure-asset-orchestrator-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 13 06:49:03 crc restorecon[4764]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/infrastructure-asset-orchestrator-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 13 06:49:03 crc restorecon[4764]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/instana-agent-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 13 06:49:03 crc restorecon[4764]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/instana-agent-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 13 06:49:03 crc restorecon[4764]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/intel-device-plugins-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 13 06:49:03 crc restorecon[4764]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/intel-device-plugins-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 13 06:49:03 crc restorecon[4764]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/intel-kubernetes-power-manager not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 13 06:49:03 crc restorecon[4764]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/intel-kubernetes-power-manager/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 13 06:49:03 crc restorecon[4764]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/iomesh-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 13 06:49:03 crc restorecon[4764]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/iomesh-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 13 06:49:03 crc restorecon[4764]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/joget-dx-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 13 06:49:03 crc restorecon[4764]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/joget-dx-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 13 06:49:03 crc restorecon[4764]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/joget-dx8-openshift-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 13 06:49:03 crc restorecon[4764]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/joget-dx8-openshift-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 13 06:49:03 crc restorecon[4764]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/joget-dx8-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 13 06:49:03 crc restorecon[4764]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/joget-dx8-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 13 06:49:03 crc restorecon[4764]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/k8s-triliovault not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 13 06:49:03 crc restorecon[4764]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/k8s-triliovault/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 13 06:49:03 crc restorecon[4764]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-ati-updates not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 13 06:49:03 crc restorecon[4764]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-ati-updates/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 13 06:49:03 crc restorecon[4764]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-kcos-framework not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 13 06:49:03 crc restorecon[4764]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-kcos-framework/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 13 06:49:03 crc restorecon[4764]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-kcos-ingress not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 13 06:49:03 crc restorecon[4764]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-kcos-ingress/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 13 06:49:03 crc restorecon[4764]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-kcos-licensing not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 13 06:49:03 crc restorecon[4764]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-kcos-licensing/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 13 06:49:03 crc restorecon[4764]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-kcos-sso not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 13 06:49:03 crc restorecon[4764]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-kcos-sso/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 13 06:49:03 crc restorecon[4764]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-keycloak-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 13 06:49:03 crc restorecon[4764]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-keycloak-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 13 06:49:03 crc restorecon[4764]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-load-core not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 13 06:49:03 crc restorecon[4764]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-load-core/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 13 06:49:03 crc restorecon[4764]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-loadcore-agents not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 13 06:49:03 crc restorecon[4764]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-loadcore-agents/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 13 06:49:03 crc restorecon[4764]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-nats-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 13 06:49:03 crc restorecon[4764]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-nats-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 13 06:49:03 crc restorecon[4764]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-nimbusmosaic-dusim not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 13 06:49:03 crc restorecon[4764]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-nimbusmosaic-dusim/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 13 06:49:03 crc restorecon[4764]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-rest-api-browser-v1 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 13 06:49:03 crc restorecon[4764]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-rest-api-browser-v1/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 13 06:49:03 crc restorecon[4764]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-appsec not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 13 06:49:03 crc restorecon[4764]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-appsec/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 13 06:49:03 crc restorecon[4764]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-core not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 13 06:49:03 crc restorecon[4764]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-core/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 13 06:49:03 crc restorecon[4764]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-db not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 13 06:49:03 crc restorecon[4764]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-db/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 13 06:49:03 crc restorecon[4764]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-diagnostics not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 13 06:49:03 crc restorecon[4764]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-diagnostics/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 13 06:49:03 crc restorecon[4764]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-logging not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 13 06:49:03 crc restorecon[4764]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-logging/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 13 06:49:03 crc restorecon[4764]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-migration not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 13 06:49:03 crc restorecon[4764]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-migration/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 13 06:49:03 crc restorecon[4764]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-msg-broker not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 13 06:49:03 crc restorecon[4764]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-msg-broker/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 13 06:49:03 crc restorecon[4764]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-notifications not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 13 06:49:03 crc restorecon[4764]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-notifications/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 13 06:49:03 crc restorecon[4764]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-stats-dashboards not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 13 06:49:03 crc restorecon[4764]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-stats-dashboards/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 13 06:49:03 crc restorecon[4764]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-storage not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 13 06:49:03 crc restorecon[4764]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-storage/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 13 06:49:03 crc restorecon[4764]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-test-core not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 13 06:49:03 crc restorecon[4764]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-test-core/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 13 06:49:03 crc restorecon[4764]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-ui not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 13 06:49:03 crc restorecon[4764]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-ui/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 13 06:49:03 crc restorecon[4764]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-websocket-service not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 13 06:49:03 crc restorecon[4764]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-websocket-service/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 13 06:49:03 crc restorecon[4764]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kong-gateway-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 13 06:49:03 crc restorecon[4764]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kong-gateway-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 13 06:49:03 crc restorecon[4764]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubearmor-operator-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 13 06:49:03 crc restorecon[4764]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubearmor-operator-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 13 06:49:03 crc restorecon[4764]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubecost-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 13 06:49:03 crc restorecon[4764]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubecost-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 13 06:49:03 crc restorecon[4764]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubemq-operator-marketplace not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 13 06:49:03 crc restorecon[4764]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubemq-operator-marketplace/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 13 06:49:03 crc restorecon[4764]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubeturbo-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 13 06:49:03 crc restorecon[4764]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubeturbo-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 13 06:49:03 crc restorecon[4764]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/lenovo-locd-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 13 06:49:03 crc restorecon[4764]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/lenovo-locd-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 13 06:49:03 crc restorecon[4764]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/marketplace-games-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 13 06:49:03 crc restorecon[4764]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/marketplace-games-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 13 06:49:03 crc restorecon[4764]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/memcached-operator-ogaye not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 13 06:49:03 crc restorecon[4764]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/memcached-operator-ogaye/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 13 06:49:03 crc restorecon[4764]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/memory-machine-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 13 06:49:03 crc restorecon[4764]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/memory-machine-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 13 06:49:03 crc restorecon[4764]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/model-builder-for-vision-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 13 06:49:03 crc restorecon[4764]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/model-builder-for-vision-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 13 06:49:03 crc restorecon[4764]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mongodb-atlas-kubernetes not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 13 06:49:03 crc restorecon[4764]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mongodb-atlas-kubernetes/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 13 06:49:03 crc restorecon[4764]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mongodb-enterprise not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 13 06:49:03 crc restorecon[4764]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mongodb-enterprise/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 13 06:49:03 crc restorecon[4764]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/netapp-spark-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 13 06:49:03 crc restorecon[4764]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/netapp-spark-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 13 06:49:03 crc restorecon[4764]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/netscaler-adm-agent-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 13 06:49:03 crc restorecon[4764]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/netscaler-adm-agent-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 13 06:49:03 crc restorecon[4764]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/netscaler-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 13 06:49:03 crc restorecon[4764]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/netscaler-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 13 06:49:03 crc restorecon[4764]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/neuvector-certified-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 13 06:49:03 crc restorecon[4764]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/neuvector-certified-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 13 06:49:03 crc restorecon[4764]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nexus-repository-ha-operator-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 13 06:49:03 crc restorecon[4764]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nexus-repository-ha-operator-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 13 06:49:03 crc restorecon[4764]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nginx-ingress-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 13 06:49:03 crc restorecon[4764]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nginx-ingress-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 13 06:49:03 crc restorecon[4764]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pcc-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 13 06:49:03 crc restorecon[4764]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pcc-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 13 06:49:03 crc restorecon[4764]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nim-operator-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 13 06:49:03 crc restorecon[4764]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nim-operator-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 13 06:49:03 crc restorecon[4764]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nxiq-operator-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 13 06:49:03 crc restorecon[4764]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nxiq-operator-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 13 06:49:03 crc restorecon[4764]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nxrm-operator-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 13 06:49:03 crc restorecon[4764]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nxrm-operator-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 13 06:49:03 crc restorecon[4764]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odigos-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 13 06:49:03 crc restorecon[4764]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odigos-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 13 06:49:03 crc restorecon[4764]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/open-liberty-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 13 06:49:03 crc restorecon[4764]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/open-liberty-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 13 06:49:03 crc restorecon[4764]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshiftartifactoryha-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 13 06:49:03 crc restorecon[4764]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshiftartifactoryha-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 13 06:49:03 crc restorecon[4764]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshiftxray-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 13 06:49:03 crc restorecon[4764]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshiftxray-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 13 06:49:03 crc restorecon[4764]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/operator-certification-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 13 06:49:03 crc restorecon[4764]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/operator-certification-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 13 06:49:03 crc restorecon[4764]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ovms-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 13 06:49:03 crc restorecon[4764]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ovms-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 13 06:49:03 crc restorecon[4764]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pachyderm-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 13 06:49:03 crc restorecon[4764]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pachyderm-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 13 06:49:03 crc restorecon[4764]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pmem-csi-operator-os not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 13 06:49:03 crc restorecon[4764]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pmem-csi-operator-os/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 13 06:49:03 crc restorecon[4764]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/portworx-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 13 06:49:03 crc restorecon[4764]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/portworx-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 13 06:49:03 crc restorecon[4764]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/prometurbo-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 13 06:49:03 crc restorecon[4764]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/prometurbo-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 13 06:49:03 crc restorecon[4764]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pubsubplus-eventbroker-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 13 06:49:03 crc restorecon[4764]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pubsubplus-eventbroker-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 13 06:49:03 crc restorecon[4764]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/redis-enterprise-operator-cert not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 13 06:49:03 crc restorecon[4764]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/redis-enterprise-operator-cert/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 13 06:49:03 crc restorecon[4764]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/runtime-component-operator-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 13 06:49:03 crc restorecon[4764]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/runtime-component-operator-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 13 06:49:03 crc restorecon[4764]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/runtime-fabric-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 13 06:49:03 crc restorecon[4764]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/runtime-fabric-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 13 06:49:03 crc restorecon[4764]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sanstoragecsi-operator-bundle not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 13 06:49:03 crc restorecon[4764]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sanstoragecsi-operator-bundle/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 13 06:49:03 crc restorecon[4764]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/silicom-sts-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 13 06:49:03 crc restorecon[4764]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/silicom-sts-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 13 06:49:03 crc restorecon[4764]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/smilecdr-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 13 06:49:03 crc restorecon[4764]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/smilecdr-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 13 06:49:03 crc restorecon[4764]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sriov-fec not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 13 06:49:03 crc restorecon[4764]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sriov-fec/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 13 06:49:03 crc restorecon[4764]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/stackable-commons-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 13 06:49:03 crc restorecon[4764]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/stackable-commons-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 13 06:49:03 crc restorecon[4764]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/stackable-zookeeper-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 13 06:49:03 crc restorecon[4764]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/stackable-zookeeper-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 13 06:49:03 crc restorecon[4764]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/t8c-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 13 06:49:03 crc restorecon[4764]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/t8c-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 13 06:49:03 crc restorecon[4764]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/t8c-tsc-client-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 13 06:49:03 crc restorecon[4764]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/t8c-tsc-client-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 13 06:49:03 crc restorecon[4764]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tawon-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 13 06:49:03 crc restorecon[4764]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tawon-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 13 06:49:03 crc restorecon[4764]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tigera-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 13 06:49:03 crc restorecon[4764]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tigera-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 13 06:49:03 crc restorecon[4764]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/timemachine-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 13 06:49:03 crc restorecon[4764]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/timemachine-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 13 06:49:03 crc restorecon[4764]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/vault-secrets-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 13 06:49:03 crc restorecon[4764]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/vault-secrets-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 13 06:49:03 crc restorecon[4764]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/vcp-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 13 06:49:03 crc restorecon[4764]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/vcp-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 13 06:49:03 crc restorecon[4764]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/webotx-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 13 06:49:03 crc restorecon[4764]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/webotx-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 13 06:49:03 crc restorecon[4764]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/xcrypt-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 13 06:49:03 crc restorecon[4764]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/xcrypt-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 13 06:49:03 crc restorecon[4764]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/zabbix-operator-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 13 06:49:03 crc restorecon[4764]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/zabbix-operator-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 13 06:49:03 crc restorecon[4764]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/cache not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 13 06:49:03 crc restorecon[4764]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 13 06:49:03 crc restorecon[4764]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 13 06:49:03 crc restorecon[4764]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/00000-1.psg not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 13 06:49:03 crc restorecon[4764]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/00000-1.psg.pmt not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 13 06:49:03 crc restorecon[4764]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/db.pmt not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 13 06:49:03 crc restorecon[4764]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/index.pmt not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 13 06:49:03 crc restorecon[4764]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/main.pix not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 13 06:49:03 crc restorecon[4764]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/overflow.pix not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 13 06:49:03 crc restorecon[4764]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/digest not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 13 06:49:03 crc restorecon[4764]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/utilities not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 13 06:49:03 crc restorecon[4764]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/utilities/copy-content not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 13 06:49:03 crc restorecon[4764]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 13 06:49:03 crc restorecon[4764]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/containers/extract-utilities/63709497 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 13 06:49:03 crc restorecon[4764]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/containers/extract-utilities/d966b7fd not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 13 06:49:03 crc restorecon[4764]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/containers/extract-utilities/f5773757 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 13 06:49:03 crc restorecon[4764]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/containers/extract-content/81c9edb9 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 13 06:49:03 crc restorecon[4764]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/containers/extract-content/57bf57ee not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 13 06:49:03 crc restorecon[4764]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/containers/extract-content/86f5e6aa not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 13 06:49:03 crc restorecon[4764]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/containers/registry-server/0aabe31d not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 13 06:49:03 crc restorecon[4764]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/containers/registry-server/d2af85c2 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 13 06:49:03 crc restorecon[4764]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/containers/registry-server/09d157d9 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 13 06:49:03 crc restorecon[4764]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 13 06:49:03 crc restorecon[4764]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/cache not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 13 06:49:03 crc restorecon[4764]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 13 06:49:03 crc restorecon[4764]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 13 06:49:03 crc restorecon[4764]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/00000-1.psg not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 13 06:49:03 crc restorecon[4764]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/00000-1.psg.pmt not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 13 06:49:03 crc restorecon[4764]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/db.pmt not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 13 06:49:03 crc restorecon[4764]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/index.pmt not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 13 06:49:03 crc restorecon[4764]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/main.pix not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 13 06:49:03 crc restorecon[4764]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/overflow.pix not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 13 06:49:03 crc restorecon[4764]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/digest not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 13 06:49:03 crc restorecon[4764]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 13 06:49:03 crc restorecon[4764]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/3scale-community-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 13 06:49:03 crc restorecon[4764]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/3scale-community-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 13 06:49:03 crc restorecon[4764]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-acm-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 13 06:49:03 crc restorecon[4764]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-acm-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 13 06:49:03 crc restorecon[4764]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-acmpca-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 13 06:49:03 crc restorecon[4764]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-acmpca-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 13 06:49:03 crc restorecon[4764]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-apigateway-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 13 06:49:03 crc restorecon[4764]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-apigateway-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 13 06:49:03 crc restorecon[4764]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-apigatewayv2-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 13 06:49:03 crc restorecon[4764]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-apigatewayv2-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 13 06:49:03 crc restorecon[4764]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-applicationautoscaling-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 13 06:49:03 crc restorecon[4764]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-applicationautoscaling-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 13 06:49:03 crc restorecon[4764]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-athena-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 13 06:49:03 crc restorecon[4764]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-athena-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 13 06:49:03 crc restorecon[4764]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-cloudfront-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 13 06:49:03 crc restorecon[4764]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-cloudfront-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 13 06:49:03 crc restorecon[4764]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-cloudtrail-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 13 06:49:03 crc restorecon[4764]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-cloudtrail-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 13 06:49:03 crc restorecon[4764]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-cloudwatch-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 13 06:49:03 crc restorecon[4764]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-cloudwatch-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 13 06:49:03 crc restorecon[4764]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-cloudwatchlogs-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 13 06:49:03 crc restorecon[4764]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-cloudwatchlogs-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 13 06:49:03 crc restorecon[4764]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-documentdb-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 13 06:49:03 crc restorecon[4764]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-documentdb-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 13 06:49:03 crc restorecon[4764]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-dynamodb-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 13 06:49:03 crc restorecon[4764]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-dynamodb-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 13 06:49:03 crc restorecon[4764]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-ec2-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 13 06:49:03 crc restorecon[4764]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-ec2-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 13 06:49:03 crc restorecon[4764]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-ecr-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 13 06:49:03 crc restorecon[4764]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-ecr-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 13 06:49:03 crc restorecon[4764]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-ecs-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 13 06:49:03 crc restorecon[4764]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-ecs-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 13 06:49:03 crc restorecon[4764]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-efs-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 13 06:49:03 crc restorecon[4764]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-efs-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 13 06:49:03 crc restorecon[4764]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-eks-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 13 06:49:03 crc restorecon[4764]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-eks-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 13 06:49:03 crc restorecon[4764]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-elasticache-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 13 06:49:03 crc restorecon[4764]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-elasticache-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 13 06:49:03 crc restorecon[4764]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-elbv2-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 13 06:49:03 crc restorecon[4764]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-elbv2-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 13 06:49:03 crc restorecon[4764]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-emrcontainers-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 13 06:49:03 crc restorecon[4764]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-emrcontainers-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 13 06:49:03 crc restorecon[4764]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-eventbridge-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 13 06:49:03 crc restorecon[4764]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-eventbridge-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 13 06:49:03 crc restorecon[4764]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-iam-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 13 06:49:03 crc restorecon[4764]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-iam-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 13 06:49:03 crc restorecon[4764]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-kafka-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 13 06:49:03 crc restorecon[4764]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-kafka-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 13 06:49:03 crc restorecon[4764]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-keyspaces-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 13 06:49:03 crc restorecon[4764]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-keyspaces-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 13 06:49:03 crc restorecon[4764]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-kinesis-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 13 06:49:03 crc restorecon[4764]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-kinesis-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 13 06:49:03 crc restorecon[4764]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-kms-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 13 06:49:03 crc restorecon[4764]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-kms-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 13 06:49:03 crc restorecon[4764]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-lambda-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 13 06:49:03 crc restorecon[4764]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-lambda-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 13 06:49:03 crc restorecon[4764]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-memorydb-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 13 06:49:03 crc restorecon[4764]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-memorydb-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 13 06:49:03 crc restorecon[4764]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-mq-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 13 06:49:03 crc restorecon[4764]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-mq-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 13 06:49:03 crc restorecon[4764]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-networkfirewall-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 13 06:49:03 crc restorecon[4764]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-networkfirewall-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 13 06:49:03 crc restorecon[4764]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-opensearchservice-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 13 06:49:03 crc restorecon[4764]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-opensearchservice-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 13 06:49:03 crc restorecon[4764]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-organizations-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 13 06:49:03 crc restorecon[4764]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-organizations-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 13 06:49:03 crc restorecon[4764]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-pipes-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 13 06:49:03 crc restorecon[4764]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-pipes-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 13 06:49:03 crc restorecon[4764]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-prometheusservice-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 13 06:49:03 crc restorecon[4764]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-prometheusservice-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 13 06:49:03 crc restorecon[4764]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-rds-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 13 06:49:03 crc restorecon[4764]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-rds-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 13 06:49:03 crc restorecon[4764]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-recyclebin-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 13 06:49:03 crc restorecon[4764]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-recyclebin-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 13 06:49:03 crc restorecon[4764]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-route53-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 13 06:49:03 crc restorecon[4764]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-route53-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 13 06:49:03 crc restorecon[4764]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-route53resolver-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 13 06:49:03 crc restorecon[4764]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-route53resolver-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 13 06:49:03 crc restorecon[4764]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-s3-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 13 06:49:03 crc restorecon[4764]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-s3-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 13 06:49:03 crc restorecon[4764]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-sagemaker-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 13 06:49:03 crc restorecon[4764]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-sagemaker-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 13 06:49:03 crc restorecon[4764]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-secretsmanager-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 13 06:49:03 crc restorecon[4764]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-secretsmanager-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 13 06:49:03 crc restorecon[4764]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-ses-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 13 06:49:03 crc restorecon[4764]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-ses-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 13 06:49:03 crc restorecon[4764]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-sfn-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 13 06:49:03 crc restorecon[4764]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-sfn-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 13 06:49:03 crc restorecon[4764]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-sns-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 13 06:49:03 crc restorecon[4764]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-sns-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 13 06:49:03 crc restorecon[4764]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-sqs-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 13 06:49:03 crc restorecon[4764]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-sqs-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 13 06:49:03 crc restorecon[4764]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-ssm-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 13 06:49:03 crc restorecon[4764]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-ssm-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 13 06:49:03 crc restorecon[4764]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-wafv2-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 13 06:49:03 crc restorecon[4764]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-wafv2-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 13 06:49:03 crc restorecon[4764]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aerospike-kubernetes-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 13 06:49:03 crc restorecon[4764]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aerospike-kubernetes-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 13 06:49:03 crc restorecon[4764]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/airflow-helm-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 13 06:49:03 crc restorecon[4764]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/airflow-helm-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 13 06:49:03 crc restorecon[4764]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/alloydb-omni-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 13 06:49:03 crc restorecon[4764]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/alloydb-omni-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 13 06:49:03 crc restorecon[4764]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/alvearie-imaging-ingestion not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 13 06:49:03 crc restorecon[4764]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/alvearie-imaging-ingestion/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 13 06:49:03 crc restorecon[4764]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/amd-gpu-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 13 06:49:03 crc restorecon[4764]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/amd-gpu-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 13 06:49:03 crc restorecon[4764]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/analytics-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 13 06:49:03 crc restorecon[4764]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/analytics-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 13 06:49:03 crc restorecon[4764]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/annotationlab not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 13 06:49:03 crc restorecon[4764]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/annotationlab/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 13 06:49:03 crc restorecon[4764]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/apicast-community-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 13 06:49:03 crc restorecon[4764]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/apicast-community-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 13 06:49:03 crc restorecon[4764]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/apicurio-api-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 13 06:49:03 crc restorecon[4764]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/apicurio-api-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 13 06:49:03 crc restorecon[4764]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/apicurio-registry not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 13 06:49:03 crc restorecon[4764]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/apicurio-registry/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 13 06:49:03 crc restorecon[4764]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/apicurito not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 13 06:49:03 crc restorecon[4764]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/apicurito/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 13 06:49:03 crc restorecon[4764]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/apimatic-kubernetes-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 13 06:49:03 crc restorecon[4764]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/apimatic-kubernetes-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 13 06:49:03 crc restorecon[4764]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/application-services-metering-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 13 06:49:03 crc restorecon[4764]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/application-services-metering-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 13 06:49:03 crc restorecon[4764]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aqua not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 13 06:49:03 crc restorecon[4764]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aqua/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 13 06:49:03 crc restorecon[4764]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/argocd-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 13 06:49:03 crc restorecon[4764]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/argocd-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 13 06:49:03 crc restorecon[4764]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/assisted-service-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 13 06:49:03 crc restorecon[4764]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/assisted-service-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 13 06:49:03 crc restorecon[4764]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/authorino-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 13 06:49:03 crc restorecon[4764]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/authorino-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 13 06:49:03 crc restorecon[4764]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/automotive-infra not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 13 06:49:03 crc restorecon[4764]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/automotive-infra/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 13 06:49:03 crc restorecon[4764]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aws-efs-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 13 06:49:03 crc restorecon[4764]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aws-efs-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 13 06:49:03 crc restorecon[4764]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/awss3-operator-registry not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 13 06:49:03 crc restorecon[4764]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/awss3-operator-registry/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 13 06:49:03 crc restorecon[4764]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/azure-service-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 13 06:49:03 crc restorecon[4764]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/azure-service-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 13 06:49:03 crc restorecon[4764]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/beegfs-csi-driver-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 13 06:49:03 crc restorecon[4764]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/beegfs-csi-driver-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 13 06:49:03 crc restorecon[4764]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/bpfman-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 13 06:49:03 crc restorecon[4764]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/bpfman-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 13 06:49:03 crc restorecon[4764]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/camel-k not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 13 06:49:03 crc restorecon[4764]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/camel-k/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 13 06:49:03 crc restorecon[4764]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/camel-karavan-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 13 06:49:03 crc restorecon[4764]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/camel-karavan-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 13 06:49:03 crc restorecon[4764]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cass-operator-community not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 13 06:49:03 crc restorecon[4764]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cass-operator-community/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 13 06:49:03 crc restorecon[4764]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cert-manager not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 13 06:49:03 crc restorecon[4764]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cert-manager/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 13 06:49:03 crc restorecon[4764]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cert-utils-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 13 06:49:03 crc restorecon[4764]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cert-utils-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 13 06:49:03 crc restorecon[4764]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cluster-aas-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 13 06:49:03 crc restorecon[4764]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cluster-aas-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 13 06:49:03 crc restorecon[4764]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cluster-impairment-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 13 06:49:03 crc restorecon[4764]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cluster-impairment-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 13 06:49:03 crc restorecon[4764]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cluster-manager not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 13 06:49:03 crc restorecon[4764]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cluster-manager/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 13 06:49:03 crc restorecon[4764]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cockroachdb not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 13 06:49:03 crc restorecon[4764]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cockroachdb/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 13 06:49:03 crc restorecon[4764]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/codeflare-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 13 06:49:03 crc restorecon[4764]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/codeflare-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 13 06:49:03 crc restorecon[4764]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/community-kubevirt-hyperconverged not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 13 06:49:03 crc restorecon[4764]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/community-kubevirt-hyperconverged/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 13 06:49:03 crc restorecon[4764]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/community-trivy-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 13 06:49:03 crc restorecon[4764]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/community-trivy-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 13 06:49:03 crc restorecon[4764]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/community-windows-machine-config-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 13 06:49:03 crc restorecon[4764]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/community-windows-machine-config-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 13 06:49:03 crc restorecon[4764]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/customized-user-remediation not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 13 06:49:03 crc restorecon[4764]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/customized-user-remediation/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 13 06:49:03 crc restorecon[4764]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cxl-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 13 06:49:03 crc restorecon[4764]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cxl-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 13 06:49:03 crc restorecon[4764]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dapr-kubernetes-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 13 06:49:03 crc restorecon[4764]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dapr-kubernetes-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 13 06:49:03 crc restorecon[4764]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/datadog-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 13 06:49:03 crc restorecon[4764]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/datadog-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 13 06:49:03 crc restorecon[4764]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/datatrucker-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 13 06:49:03 crc restorecon[4764]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/datatrucker-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 13 06:49:03 crc restorecon[4764]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dbaas-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 13 06:49:03 crc restorecon[4764]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dbaas-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 13 06:49:03 crc restorecon[4764]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/debezium-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 13 06:49:03 crc restorecon[4764]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/debezium-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 13 06:49:03 crc restorecon[4764]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dell-csm-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 13 06:49:03 crc restorecon[4764]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dell-csm-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 13 06:49:03 crc restorecon[4764]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/deployment-validation-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 13 06:49:03 crc restorecon[4764]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/deployment-validation-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 13 06:49:03 crc restorecon[4764]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/devopsinabox not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 13 06:49:03 crc restorecon[4764]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/devopsinabox/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 13 06:49:03 crc restorecon[4764]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dns-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 13 06:49:03 crc restorecon[4764]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dns-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 13 06:49:03 crc restorecon[4764]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dynatrace-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 13 06:49:03 crc restorecon[4764]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dynatrace-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 13 06:49:03 crc restorecon[4764]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/eclipse-amlen-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 13 06:49:03 crc restorecon[4764]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/eclipse-amlen-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 13 06:49:03 crc restorecon[4764]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/eclipse-che not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 13 06:49:03 crc restorecon[4764]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/eclipse-che/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 13 06:49:03 crc restorecon[4764]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ecr-secret-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 13 06:49:03 crc restorecon[4764]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ecr-secret-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 13 06:49:03 crc restorecon[4764]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/edp-keycloak-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 13 06:49:03 crc restorecon[4764]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/edp-keycloak-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 13 06:49:03 crc restorecon[4764]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/eginnovations-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 13 06:49:03 crc restorecon[4764]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/eginnovations-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 13 06:49:03 crc restorecon[4764]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/egressip-ipam-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 13 06:49:03 crc restorecon[4764]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/egressip-ipam-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 13 06:49:03 crc restorecon[4764]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ember-csi-community-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 13 06:49:03 crc restorecon[4764]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ember-csi-community-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 13 06:49:03 crc restorecon[4764]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/etcd not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 13 06:49:03 crc restorecon[4764]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/etcd/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 13 06:49:03 crc restorecon[4764]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/eventing-kogito not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 13 06:49:03 crc restorecon[4764]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/eventing-kogito/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 13 06:49:03 crc restorecon[4764]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/external-secrets-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 13 06:49:03 crc restorecon[4764]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/external-secrets-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 13 06:49:03 crc restorecon[4764]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/falcon-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 13 06:49:03 crc restorecon[4764]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/falcon-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 13 06:49:03 crc restorecon[4764]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/fence-agents-remediation not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 13 06:49:03 crc restorecon[4764]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/fence-agents-remediation/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 13 06:49:03 crc restorecon[4764]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/flink-kubernetes-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 13 06:49:03 crc restorecon[4764]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/flink-kubernetes-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 13 06:49:03 crc restorecon[4764]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/flux not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 13 06:49:03 crc restorecon[4764]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/flux/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 13 06:49:03 crc restorecon[4764]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/k8gb not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 13 06:49:03 crc restorecon[4764]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/k8gb/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 13 06:49:03 crc restorecon[4764]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/fossul-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 13 06:49:03 crc restorecon[4764]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/fossul-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 13 06:49:03 crc restorecon[4764]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/github-arc-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 13 06:49:03 crc restorecon[4764]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/github-arc-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 13 06:49:03 crc restorecon[4764]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/gitops-primer not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 13 06:49:03 crc restorecon[4764]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/gitops-primer/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 13 06:49:03 crc restorecon[4764]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/gitwebhook-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 13 06:49:03 crc restorecon[4764]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/gitwebhook-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 13 06:49:03 crc restorecon[4764]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/global-load-balancer-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 13 06:49:03 crc restorecon[4764]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/global-load-balancer-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 13 06:49:03 crc restorecon[4764]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/grafana-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 13 06:49:03 crc restorecon[4764]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/grafana-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 13 06:49:03 crc restorecon[4764]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/group-sync-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 13 06:49:03 crc restorecon[4764]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/group-sync-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 13 06:49:03 crc restorecon[4764]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hawtio-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 13 06:49:03 crc restorecon[4764]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hawtio-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 13 06:49:03 crc restorecon[4764]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hazelcast-platform-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 13 06:49:03 crc restorecon[4764]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hazelcast-platform-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 13 06:49:03 crc restorecon[4764]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hedvig-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 13 06:49:03 crc restorecon[4764]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hedvig-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 13 06:49:03 crc restorecon[4764]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hive-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 13 06:49:03 crc restorecon[4764]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hive-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 13 06:49:03 crc restorecon[4764]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/horreum-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 13 06:49:03 crc restorecon[4764]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/horreum-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 13 06:49:03 crc restorecon[4764]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hyperfoil-bundle not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 13 06:49:03 crc restorecon[4764]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hyperfoil-bundle/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 13 06:49:03 crc restorecon[4764]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-block-csi-operator-community not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 13 06:49:03 crc restorecon[4764]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-block-csi-operator-community/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 13 06:49:03 crc restorecon[4764]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-security-verify-access-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 13 06:49:03 crc restorecon[4764]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-security-verify-access-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 13 06:49:03 crc restorecon[4764]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-spectrum-scale-csi-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 13 06:49:03 crc restorecon[4764]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-spectrum-scale-csi-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 13 06:49:03 crc restorecon[4764]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibmcloud-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 13 06:49:03 crc restorecon[4764]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibmcloud-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 13 06:49:03 crc restorecon[4764]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/infinispan not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 13 06:49:03 crc restorecon[4764]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/infinispan/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 13 06:49:03 crc restorecon[4764]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/integrity-shield-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 13 06:49:03 crc restorecon[4764]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/integrity-shield-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 13 06:49:03 crc restorecon[4764]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ipfs-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 13 06:49:03 crc restorecon[4764]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ipfs-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 13 06:49:03 crc restorecon[4764]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/istio-workspace-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 13 06:49:03 crc restorecon[4764]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/istio-workspace-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 13 06:49:03 crc restorecon[4764]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/jaeger not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 13 06:49:03 crc restorecon[4764]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/jaeger/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 13 06:49:03 crc restorecon[4764]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kaoto-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 13 06:49:03 crc restorecon[4764]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kaoto-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 13 06:49:03 crc restorecon[4764]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keda not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 13 06:49:03 crc restorecon[4764]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keda/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 13 06:49:03 crc restorecon[4764]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keepalived-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 13 06:49:03 crc restorecon[4764]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keepalived-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 13 06:49:03 crc restorecon[4764]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keycloak-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 13 06:49:03 crc restorecon[4764]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keycloak-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 13 06:49:03 crc restorecon[4764]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keycloak-permissions-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 13 06:49:03 crc restorecon[4764]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keycloak-permissions-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 13 06:49:03 crc restorecon[4764]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/klusterlet not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 13 06:49:03 crc restorecon[4764]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/klusterlet/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 13 06:49:03 crc restorecon[4764]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kogito-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 13 06:49:03 crc restorecon[4764]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kogito-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 13 06:49:03 crc restorecon[4764]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/koku-metrics-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 13 06:49:03 crc restorecon[4764]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/koku-metrics-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 13 06:49:03 crc restorecon[4764]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/konveyor-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 13 06:49:03 crc restorecon[4764]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/konveyor-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 13 06:49:03 crc restorecon[4764]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/korrel8r not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 13 06:49:03 crc restorecon[4764]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/korrel8r/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 13 06:49:03 crc restorecon[4764]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kuadrant-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 13 06:49:03 crc restorecon[4764]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kuadrant-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 13 06:49:03 crc restorecon[4764]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kube-green not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 13 06:49:03 crc restorecon[4764]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kube-green/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 13 06:49:03 crc restorecon[4764]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubecost not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 13 06:49:03 crc restorecon[4764]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubecost/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 13 06:49:03 crc restorecon[4764]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubernetes-imagepuller-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 13 06:49:03 crc restorecon[4764]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubernetes-imagepuller-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 13 06:49:03 crc restorecon[4764]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubeturbo not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 13 06:49:03 crc restorecon[4764]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubeturbo/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 13 06:49:03 crc restorecon[4764]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/l5-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 13 06:49:03 crc restorecon[4764]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/l5-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 13 06:49:03 crc restorecon[4764]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/layer7-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 13 06:49:03 crc restorecon[4764]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/layer7-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 13 06:49:03 crc restorecon[4764]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/lbconfig-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 13 06:49:03 crc restorecon[4764]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/lbconfig-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 13 06:49:03 crc restorecon[4764]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/lib-bucket-provisioner not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 13 06:49:03 crc restorecon[4764]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/lib-bucket-provisioner/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 13 06:49:03 crc restorecon[4764]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/limitador-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 13 06:49:03 crc restorecon[4764]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/limitador-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 13 06:49:03 crc restorecon[4764]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/logging-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 13 06:49:03 crc restorecon[4764]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/logging-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 13 06:49:03 crc restorecon[4764]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/loki-helm-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 13 06:49:03 crc restorecon[4764]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/loki-helm-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 13 06:49:03 crc restorecon[4764]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/loki-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 13 06:49:03 crc restorecon[4764]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/loki-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 13 06:49:03 crc restorecon[4764]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/machine-deletion-remediation not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 13 06:49:03 crc restorecon[4764]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/machine-deletion-remediation/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 13 06:49:03 crc restorecon[4764]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mariadb-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 13 06:49:03 crc restorecon[4764]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mariadb-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 13 06:49:03 crc restorecon[4764]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/marin3r not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 13 06:49:03 crc restorecon[4764]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/marin3r/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 13 06:49:03 crc restorecon[4764]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mercury-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 13 06:49:03 crc restorecon[4764]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mercury-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 13 06:49:03 crc restorecon[4764]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/microcks not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 13 06:49:03 crc restorecon[4764]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/microcks/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 13 06:49:03 crc restorecon[4764]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mongodb-atlas-kubernetes not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 13 06:49:03 crc restorecon[4764]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mongodb-atlas-kubernetes/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 13 06:49:03 crc restorecon[4764]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mongodb-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 13 06:49:03 crc restorecon[4764]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mongodb-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 13 06:49:03 crc restorecon[4764]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/move2kube-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 13 06:49:03 crc restorecon[4764]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/move2kube-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 13 06:49:03 crc restorecon[4764]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/multi-nic-cni-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 13 06:49:03 crc restorecon[4764]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/multi-nic-cni-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 13 06:49:03 crc restorecon[4764]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/multicluster-global-hub-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 13 06:49:03 crc restorecon[4764]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/multicluster-global-hub-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 13 06:49:03 crc restorecon[4764]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/multicluster-operators-subscription not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 13 06:49:03 crc restorecon[4764]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/multicluster-operators-subscription/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 13 06:49:03 crc restorecon[4764]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/must-gather-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 13 06:49:03 crc restorecon[4764]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/must-gather-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 13 06:49:03 crc restorecon[4764]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/namespace-configuration-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 13 06:49:03 crc restorecon[4764]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/namespace-configuration-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 13 06:49:03 crc restorecon[4764]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ncn-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 13 06:49:03 crc restorecon[4764]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ncn-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 13 06:49:03 crc restorecon[4764]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ndmspc-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 13 06:49:03 crc restorecon[4764]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ndmspc-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 13 06:49:03 crc restorecon[4764]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/netobserv-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 13 06:49:03 crc restorecon[4764]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/netobserv-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 13 06:49:03 crc restorecon[4764]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/neuvector-community-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 13 06:49:03 crc restorecon[4764]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/neuvector-community-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 13 06:49:03 crc restorecon[4764]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nexus-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 13 06:49:03 crc restorecon[4764]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nexus-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 13 06:49:03 crc restorecon[4764]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nexus-operator-m88i not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 13 06:49:03 crc restorecon[4764]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nexus-operator-m88i/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 13 06:49:03 crc restorecon[4764]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nfs-provisioner-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 13 06:49:03 crc restorecon[4764]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nfs-provisioner-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 13 06:49:03 crc restorecon[4764]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nlp-server not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 13 06:49:03 crc restorecon[4764]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nlp-server/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 13 06:49:03 crc restorecon[4764]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/node-discovery-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 13 06:49:03 crc restorecon[4764]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/node-discovery-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 13 06:49:03 crc restorecon[4764]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/node-healthcheck-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 13 06:49:03 crc restorecon[4764]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/node-healthcheck-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 13 06:49:03 crc restorecon[4764]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/node-maintenance-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 13 06:49:03 crc restorecon[4764]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/node-maintenance-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 13 06:49:03 crc restorecon[4764]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nsm-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 13 06:49:03 crc restorecon[4764]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nsm-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 13 06:49:03 crc restorecon[4764]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/oadp-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 13 06:49:03 crc restorecon[4764]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/oadp-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 13 06:49:03 crc restorecon[4764]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/observability-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 13 06:49:03 crc restorecon[4764]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/observability-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 13 06:49:03 crc restorecon[4764]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/oci-ccm-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 13 06:49:03 crc restorecon[4764]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/oci-ccm-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 13 06:49:03 crc restorecon[4764]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ocm-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 13 06:49:03 crc restorecon[4764]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ocm-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 13 06:49:03 crc restorecon[4764]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odoo-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 13 06:49:03 crc restorecon[4764]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odoo-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 13 06:49:03 crc restorecon[4764]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/opendatahub-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 13 06:49:03 crc restorecon[4764]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/opendatahub-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 13 06:49:03 crc restorecon[4764]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openebs not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 13 06:49:03 crc restorecon[4764]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openebs/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 13 06:49:03 crc restorecon[4764]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-nfd-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 13 06:49:03 crc restorecon[4764]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-nfd-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 13 06:49:03 crc restorecon[4764]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-node-upgrade-mutex-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 13 06:49:03 crc restorecon[4764]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-node-upgrade-mutex-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 13 06:49:03 crc restorecon[4764]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-qiskit-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 13 06:49:03 crc restorecon[4764]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-qiskit-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 13 06:49:03 crc restorecon[4764]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/opentelemetry-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 13 06:49:03 crc restorecon[4764]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/opentelemetry-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 13 06:49:03 crc restorecon[4764]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/patch-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 13 06:49:03 crc restorecon[4764]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/patch-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 13 06:49:03 crc restorecon[4764]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/patterns-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 13 06:49:03 crc restorecon[4764]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/patterns-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 13 06:49:03 crc restorecon[4764]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pcc-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 13 06:49:03 crc restorecon[4764]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pcc-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 13 06:49:03 crc restorecon[4764]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pelorus-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 13 06:49:03 crc restorecon[4764]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pelorus-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 13 06:49:03 crc restorecon[4764]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/percona-xtradb-cluster-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 13 06:49:03 crc restorecon[4764]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/percona-xtradb-cluster-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 13 06:49:03 crc restorecon[4764]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/portworx-essentials not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 13 06:49:03 crc restorecon[4764]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/portworx-essentials/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 13 06:49:03 crc restorecon[4764]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/postgresql not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 13 06:49:03 crc restorecon[4764]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/postgresql/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 13 06:49:03 crc restorecon[4764]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/proactive-node-scaling-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 13 06:49:03 crc restorecon[4764]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/proactive-node-scaling-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 13 06:49:03 crc restorecon[4764]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/project-quay not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 13 06:49:03 crc restorecon[4764]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/project-quay/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 13 06:49:03 crc restorecon[4764]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/prometheus not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 13 06:49:03 crc restorecon[4764]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/prometheus/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 13 06:49:03 crc restorecon[4764]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/prometheus-exporter-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 13 06:49:03 crc restorecon[4764]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/prometheus-exporter-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 13 06:49:03 crc restorecon[4764]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/prometurbo not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 13 06:49:03 crc restorecon[4764]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/prometurbo/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 13 06:49:03 crc restorecon[4764]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pubsubplus-eventbroker-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 13 06:49:03 crc restorecon[4764]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pubsubplus-eventbroker-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 13 06:49:03 crc restorecon[4764]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pulp-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 13 06:49:03 crc restorecon[4764]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pulp-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 13 06:49:03 crc restorecon[4764]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rabbitmq-cluster-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 13 06:49:03 crc restorecon[4764]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rabbitmq-cluster-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 13 06:49:03 crc restorecon[4764]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rabbitmq-messaging-topology-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 13 06:49:03 crc restorecon[4764]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rabbitmq-messaging-topology-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 13 06:49:03 crc restorecon[4764]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/redis-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 13 06:49:03 crc restorecon[4764]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/redis-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 13 06:49:03 crc restorecon[4764]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/reportportal-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 13 06:49:03 crc restorecon[4764]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/reportportal-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 13 06:49:03 crc restorecon[4764]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/resource-locker-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 13 06:49:03 crc restorecon[4764]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/resource-locker-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 13 06:49:03 crc restorecon[4764]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhoas-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 13 06:49:03 crc restorecon[4764]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhoas-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 13 06:49:03 crc restorecon[4764]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ripsaw not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 13 06:49:03 crc restorecon[4764]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ripsaw/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 13 06:49:03 crc restorecon[4764]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sailoperator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 13 06:49:03 crc restorecon[4764]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sailoperator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 13 06:49:03 crc restorecon[4764]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sap-commerce-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 13 06:49:03 crc restorecon[4764]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sap-commerce-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 13 06:49:03 crc restorecon[4764]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sap-data-intelligence-observer-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 13 06:49:03 crc restorecon[4764]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sap-data-intelligence-observer-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 13 06:49:03 crc restorecon[4764]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sap-hana-express-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 13 06:49:03 crc restorecon[4764]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sap-hana-express-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 13 06:49:03 crc restorecon[4764]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/seldon-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 13 06:49:03 crc restorecon[4764]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/seldon-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 13 06:49:03 crc restorecon[4764]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/self-node-remediation not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 13 06:49:03 crc restorecon[4764]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/self-node-remediation/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 13 06:49:03 crc restorecon[4764]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/service-binding-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 13 06:49:03 crc restorecon[4764]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/service-binding-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 13 06:49:03 crc restorecon[4764]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/shipwright-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 13 06:49:03 crc restorecon[4764]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/shipwright-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 13 06:49:03 crc restorecon[4764]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sigstore-helm-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 13 06:49:03 crc restorecon[4764]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sigstore-helm-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 13 06:49:03 crc restorecon[4764]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/silicom-sts-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 13 06:49:03 crc restorecon[4764]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/silicom-sts-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 13 06:49:03 crc restorecon[4764]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/skupper-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 13 06:49:03 crc restorecon[4764]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/skupper-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 13 06:49:03 crc restorecon[4764]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/snapscheduler not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 13 06:49:03 crc restorecon[4764]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/snapscheduler/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 13 06:49:03 crc restorecon[4764]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/snyk-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 13 06:49:03 crc restorecon[4764]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/snyk-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 13 06:49:03 crc restorecon[4764]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/socmmd not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 13 06:49:03 crc restorecon[4764]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/socmmd/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 13 06:49:03 crc restorecon[4764]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sonar-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 13 06:49:03 crc restorecon[4764]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sonar-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 13 06:49:03 crc restorecon[4764]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sosivio not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 13 06:49:03 crc restorecon[4764]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sosivio/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 13 06:49:03 crc restorecon[4764]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sonataflow-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 13 06:49:03 crc restorecon[4764]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sonataflow-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 13 06:49:03 crc restorecon[4764]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sosreport-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 13 06:49:03 crc restorecon[4764]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sosreport-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 13 06:49:03 crc restorecon[4764]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/spark-helm-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 13 06:49:03 crc restorecon[4764]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/spark-helm-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 13 06:49:03 crc restorecon[4764]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/special-resource-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 13 06:49:03 crc restorecon[4764]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/special-resource-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 13 06:49:03 crc restorecon[4764]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/stolostron not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 13 06:49:03 crc restorecon[4764]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/stolostron/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 13 06:49:03 crc restorecon[4764]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/stolostron-engine not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 13 06:49:03 crc restorecon[4764]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/stolostron-engine/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 13 06:49:03 crc restorecon[4764]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/strimzi-kafka-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 13 06:49:03 crc restorecon[4764]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/strimzi-kafka-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 13 06:49:03 crc restorecon[4764]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/syndesis not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 13 06:49:03 crc restorecon[4764]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/syndesis/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 13 06:49:03 crc restorecon[4764]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/t8c not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 13 06:49:03 crc restorecon[4764]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/t8c/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 13 06:49:03 crc restorecon[4764]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tagger not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 13 06:49:03 crc restorecon[4764]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tagger/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 13 06:49:03 crc restorecon[4764]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tempo-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 13 06:49:03 crc restorecon[4764]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tempo-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 13 06:49:03 crc restorecon[4764]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tf-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 13 06:49:03 crc restorecon[4764]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tf-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 13 06:49:03 crc restorecon[4764]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tidb-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 13 06:49:03 crc restorecon[4764]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tidb-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 13 06:49:03 crc restorecon[4764]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/trident-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 13 06:49:03 crc restorecon[4764]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/trident-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 13 06:49:03 crc restorecon[4764]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/trustify-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 13 06:49:03 crc restorecon[4764]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/trustify-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 13 06:49:03 crc restorecon[4764]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ucs-ci-solutions-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 13 06:49:03 crc restorecon[4764]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ucs-ci-solutions-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 13 06:49:03 crc restorecon[4764]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/universal-crossplane not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 13 06:49:03 crc restorecon[4764]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/universal-crossplane/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 13 06:49:03 crc restorecon[4764]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/varnish-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 13 06:49:03 crc restorecon[4764]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/varnish-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 13 06:49:03 crc restorecon[4764]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/vault-config-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 13 06:49:03 crc restorecon[4764]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/vault-config-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 13 06:49:03 crc restorecon[4764]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/verticadb-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 13 06:49:03 crc restorecon[4764]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/verticadb-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 13 06:49:03 crc restorecon[4764]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/volume-expander-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 13 06:49:03 crc restorecon[4764]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/volume-expander-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 13 06:49:03 crc restorecon[4764]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/wandb-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 13 06:49:03 crc restorecon[4764]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/wandb-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 13 06:49:03 crc restorecon[4764]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/windup-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 13 06:49:03 crc restorecon[4764]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/windup-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 13 06:49:03 crc restorecon[4764]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/yaks not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 13 06:49:03 crc restorecon[4764]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/yaks/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 13 06:49:03 crc restorecon[4764]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/utilities not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 13 06:49:03 crc restorecon[4764]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/utilities/copy-content not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 13 06:49:03 crc restorecon[4764]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 13 06:49:03 crc restorecon[4764]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/containers/extract-utilities/c0fe7256 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 13 06:49:03 crc restorecon[4764]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/containers/extract-utilities/c30319e4 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 13 06:49:03 crc restorecon[4764]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/containers/extract-utilities/e6b1dd45 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 13 06:49:03 crc restorecon[4764]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/containers/extract-content/2bb643f0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 13 06:49:03 crc restorecon[4764]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/containers/extract-content/920de426 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 13 06:49:03 crc restorecon[4764]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/containers/extract-content/70fa1e87 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 13 06:49:03 crc restorecon[4764]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/containers/registry-server/a1c12a2f not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 13 06:49:03 crc restorecon[4764]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/containers/registry-server/9442e6c7 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 13 06:49:03 crc restorecon[4764]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/containers/registry-server/5b45ec72 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 13 06:49:03 crc restorecon[4764]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 13 06:49:03 crc restorecon[4764]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 13 06:49:03 crc restorecon[4764]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/abot-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 13 06:49:03 crc restorecon[4764]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/abot-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 13 06:49:03 crc restorecon[4764]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aerospike-kubernetes-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 13 06:49:03 crc restorecon[4764]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aerospike-kubernetes-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 13 06:49:03 crc restorecon[4764]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aikit-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 13 06:49:03 crc restorecon[4764]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aikit-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 13 06:49:03 crc restorecon[4764]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/anzo-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 13 06:49:03 crc restorecon[4764]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/anzo-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 13 06:49:03 crc restorecon[4764]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/anzograph-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 13 06:49:03 crc restorecon[4764]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/anzograph-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 13 06:49:03 crc restorecon[4764]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/anzounstructured-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 13 06:49:03 crc restorecon[4764]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/anzounstructured-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 13 06:49:03 crc restorecon[4764]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cloudbees-ci-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 13 06:49:03 crc restorecon[4764]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cloudbees-ci-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 13 06:49:03 crc restorecon[4764]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cockroachdb-certified-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 13 06:49:03 crc restorecon[4764]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cockroachdb-certified-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 13 06:49:03 crc restorecon[4764]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/crunchy-postgres-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 13 06:49:03 crc restorecon[4764]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/crunchy-postgres-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 13 06:49:03 crc restorecon[4764]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/datadog-operator-certified-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 13 06:49:03 crc restorecon[4764]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/datadog-operator-certified-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 13 06:49:03 crc restorecon[4764]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dynatrace-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 13 06:49:03 crc restorecon[4764]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dynatrace-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 13 06:49:03 crc restorecon[4764]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/entando-k8s-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 13 06:49:03 crc restorecon[4764]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/entando-k8s-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 13 06:49:03 crc restorecon[4764]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/flux not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 13 06:49:03 crc restorecon[4764]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/flux/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 13 06:49:03 crc restorecon[4764]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/instana-agent-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 13 06:49:03 crc restorecon[4764]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/instana-agent-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 13 06:49:03 crc restorecon[4764]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/iomesh-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 13 06:49:03 crc restorecon[4764]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/iomesh-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 13 06:49:03 crc restorecon[4764]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/joget-dx-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 13 06:49:03 crc restorecon[4764]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/joget-dx-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 13 06:49:03 crc restorecon[4764]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/joget-dx8-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 13 06:49:03 crc restorecon[4764]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/joget-dx8-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 13 06:49:03 crc restorecon[4764]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/k10-kasten-operator-paygo-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 13 06:49:03 crc restorecon[4764]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/k10-kasten-operator-paygo-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 13 06:49:03 crc restorecon[4764]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/k10-kasten-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 13 06:49:03 crc restorecon[4764]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/k10-kasten-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 13 06:49:03 crc restorecon[4764]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/k10-kasten-operator-term-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 13 06:49:03 crc restorecon[4764]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/k10-kasten-operator-term-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 13 06:49:03 crc restorecon[4764]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubemq-operator-marketplace-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 13 06:49:03 crc restorecon[4764]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubemq-operator-marketplace-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 13 06:49:03 crc restorecon[4764]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubeturbo-certified-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 13 06:49:03 crc restorecon[4764]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubeturbo-certified-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 13 06:49:03 crc restorecon[4764]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/linstor-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 13 06:49:03 crc restorecon[4764]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/linstor-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 13 06:49:03 crc restorecon[4764]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/marketplace-games-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 13 06:49:03 crc restorecon[4764]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/marketplace-games-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 13 06:49:03 crc restorecon[4764]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/model-builder-for-vision-certified-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 13 06:49:03 crc restorecon[4764]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/model-builder-for-vision-certified-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 13 06:49:03 crc restorecon[4764]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/neuvector-certified-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 13 06:49:03 crc restorecon[4764]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/neuvector-certified-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 13 06:49:03 crc restorecon[4764]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ovms-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 13 06:49:03 crc restorecon[4764]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ovms-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 13 06:49:03 crc restorecon[4764]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pachyderm-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 13 06:49:03 crc restorecon[4764]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pachyderm-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 13 06:49:03 crc restorecon[4764]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/redis-enterprise-operator-cert-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 13 06:49:03 crc restorecon[4764]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/redis-enterprise-operator-cert-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 13 06:49:03 crc restorecon[4764]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/seldon-deploy-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 13 06:49:03 crc restorecon[4764]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/seldon-deploy-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 13 06:49:03 crc restorecon[4764]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/starburst-enterprise-helm-operator-paygo-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 13 06:49:03 crc restorecon[4764]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/starburst-enterprise-helm-operator-paygo-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 13 06:49:03 crc restorecon[4764]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/starburst-enterprise-helm-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 13 06:49:03 crc restorecon[4764]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/starburst-enterprise-helm-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 13 06:49:03 crc restorecon[4764]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/t8c-certified-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 13 06:49:03 crc restorecon[4764]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/t8c-certified-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 13 06:49:03 crc restorecon[4764]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/timemachine-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 13 06:49:03 crc restorecon[4764]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/timemachine-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 13 06:49:03 crc restorecon[4764]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/vfunction-server-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 13 06:49:03 crc restorecon[4764]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/vfunction-server-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 13 06:49:03 crc restorecon[4764]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/xcrypt-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 13 06:49:03 crc restorecon[4764]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/xcrypt-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 13 06:49:03 crc restorecon[4764]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/yugabyte-platform-operator-bundle-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 13 06:49:03 crc restorecon[4764]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/yugabyte-platform-operator-bundle-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 13 06:49:03 crc restorecon[4764]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/zabbix-operator-certified-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 13 06:49:03 crc restorecon[4764]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/zabbix-operator-certified-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 13 06:49:03 crc restorecon[4764]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/cache not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 13 06:49:03 crc restorecon[4764]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 13 06:49:03 crc restorecon[4764]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 13 06:49:03 crc restorecon[4764]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/00000-1.psg not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 13 06:49:03 crc restorecon[4764]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/00000-1.psg.pmt not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 13 06:49:03 crc restorecon[4764]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/db.pmt not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 13 06:49:03 crc restorecon[4764]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/index.pmt not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 13 06:49:03 crc restorecon[4764]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/main.pix not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 13 06:49:03 crc restorecon[4764]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/overflow.pix not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 13 06:49:03 crc restorecon[4764]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/digest not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 13 06:49:03 crc restorecon[4764]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/utilities not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 13 06:49:03 crc restorecon[4764]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/utilities/copy-content not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 13 06:49:03 crc restorecon[4764]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 13 06:49:03 crc restorecon[4764]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/containers/extract-utilities/3c9f3a59 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 13 06:49:03 crc restorecon[4764]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/containers/extract-utilities/1091c11b not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 13 06:49:03 crc restorecon[4764]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/containers/extract-utilities/9a6821c6 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 13 06:49:03 crc restorecon[4764]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/containers/extract-content/ec0c35e2 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 13 06:49:03 crc restorecon[4764]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/containers/extract-content/517f37e7 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 13 06:49:03 crc restorecon[4764]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/containers/extract-content/6214fe78 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 13 06:49:03 crc restorecon[4764]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/containers/registry-server/ba189c8b not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 13 06:49:03 crc restorecon[4764]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/containers/registry-server/351e4f31 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 13 06:49:03 crc restorecon[4764]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/containers/registry-server/c0f219ff not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 13 06:49:03 crc restorecon[4764]: /var/lib/kubelet/pods/3dcd261975c3d6b9a6ad6367fd4facd3/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c247,c522 Dec 13 06:49:03 crc restorecon[4764]: /var/lib/kubelet/pods/3dcd261975c3d6b9a6ad6367fd4facd3/containers/wait-for-host-port/8069f607 not reset as customized by admin to system_u:object_r:container_file_t:s0:c378,c723 Dec 13 06:49:03 crc restorecon[4764]: /var/lib/kubelet/pods/3dcd261975c3d6b9a6ad6367fd4facd3/containers/wait-for-host-port/559c3d82 not reset as customized by admin to system_u:object_r:container_file_t:s0:c133,c223 Dec 13 06:49:03 crc restorecon[4764]: /var/lib/kubelet/pods/3dcd261975c3d6b9a6ad6367fd4facd3/containers/wait-for-host-port/605ad488 not reset as customized by admin to system_u:object_r:container_file_t:s0:c247,c522 Dec 13 06:49:03 crc restorecon[4764]: /var/lib/kubelet/pods/3dcd261975c3d6b9a6ad6367fd4facd3/containers/kube-scheduler/148df488 not reset as customized by admin to system_u:object_r:container_file_t:s0:c378,c723 Dec 13 06:49:03 crc restorecon[4764]: /var/lib/kubelet/pods/3dcd261975c3d6b9a6ad6367fd4facd3/containers/kube-scheduler/3bf6dcb4 not reset as customized by admin to system_u:object_r:container_file_t:s0:c133,c223 Dec 13 06:49:03 crc restorecon[4764]: /var/lib/kubelet/pods/3dcd261975c3d6b9a6ad6367fd4facd3/containers/kube-scheduler/022a2feb not reset as customized by admin to system_u:object_r:container_file_t:s0:c247,c522 Dec 13 06:49:03 crc restorecon[4764]: /var/lib/kubelet/pods/3dcd261975c3d6b9a6ad6367fd4facd3/containers/kube-scheduler-cert-syncer/938c3924 not reset as customized by admin to system_u:object_r:container_file_t:s0:c378,c723 Dec 13 06:49:03 crc restorecon[4764]: /var/lib/kubelet/pods/3dcd261975c3d6b9a6ad6367fd4facd3/containers/kube-scheduler-cert-syncer/729fe23e not reset as customized by admin to system_u:object_r:container_file_t:s0:c133,c223 Dec 13 06:49:03 crc restorecon[4764]: /var/lib/kubelet/pods/3dcd261975c3d6b9a6ad6367fd4facd3/containers/kube-scheduler-cert-syncer/1fd5cbd4 not reset as customized by admin to system_u:object_r:container_file_t:s0:c247,c522 Dec 13 06:49:03 crc restorecon[4764]: /var/lib/kubelet/pods/3dcd261975c3d6b9a6ad6367fd4facd3/containers/kube-scheduler-recovery-controller/a96697e1 not reset as customized by admin to system_u:object_r:container_file_t:s0:c378,c723 Dec 13 06:49:03 crc restorecon[4764]: /var/lib/kubelet/pods/3dcd261975c3d6b9a6ad6367fd4facd3/containers/kube-scheduler-recovery-controller/e155ddca not reset as customized by admin to system_u:object_r:container_file_t:s0:c133,c223 Dec 13 06:49:03 crc restorecon[4764]: /var/lib/kubelet/pods/3dcd261975c3d6b9a6ad6367fd4facd3/containers/kube-scheduler-recovery-controller/10dd0e0f not reset as customized by admin to system_u:object_r:container_file_t:s0:c247,c522 Dec 13 06:49:03 crc restorecon[4764]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/v4-0-config-system-trusted-ca-bundle not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Dec 13 06:49:03 crc restorecon[4764]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/v4-0-config-system-trusted-ca-bundle/..2025_02_24_06_09_35.3018472960 not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Dec 13 06:49:03 crc restorecon[4764]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/v4-0-config-system-trusted-ca-bundle/..2025_02_24_06_09_35.3018472960/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Dec 13 06:49:03 crc restorecon[4764]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/v4-0-config-system-trusted-ca-bundle/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Dec 13 06:49:03 crc restorecon[4764]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/v4-0-config-system-trusted-ca-bundle/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Dec 13 06:49:03 crc restorecon[4764]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/audit-policies not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Dec 13 06:49:03 crc restorecon[4764]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/audit-policies/..2025_02_24_06_09_35.4262376737 not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Dec 13 06:49:03 crc restorecon[4764]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/audit-policies/..2025_02_24_06_09_35.4262376737/audit.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Dec 13 06:49:03 crc restorecon[4764]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/audit-policies/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Dec 13 06:49:03 crc restorecon[4764]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/audit-policies/audit.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Dec 13 06:49:03 crc restorecon[4764]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/v4-0-config-system-cliconfig not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Dec 13 06:49:03 crc restorecon[4764]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/v4-0-config-system-cliconfig/..2025_02_24_06_09_35.2630275752 not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Dec 13 06:49:03 crc restorecon[4764]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/v4-0-config-system-cliconfig/..2025_02_24_06_09_35.2630275752/v4-0-config-system-cliconfig not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Dec 13 06:49:03 crc restorecon[4764]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/v4-0-config-system-cliconfig/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Dec 13 06:49:03 crc restorecon[4764]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/v4-0-config-system-cliconfig/v4-0-config-system-cliconfig not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Dec 13 06:49:03 crc restorecon[4764]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/v4-0-config-system-service-ca not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Dec 13 06:49:03 crc restorecon[4764]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/v4-0-config-system-service-ca/..2025_02_24_06_09_35.2376963788 not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Dec 13 06:49:03 crc restorecon[4764]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/v4-0-config-system-service-ca/..2025_02_24_06_09_35.2376963788/service-ca.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Dec 13 06:49:03 crc restorecon[4764]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/v4-0-config-system-service-ca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Dec 13 06:49:03 crc restorecon[4764]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/v4-0-config-system-service-ca/service-ca.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Dec 13 06:49:03 crc restorecon[4764]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Dec 13 06:49:03 crc restorecon[4764]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/containers/oauth-openshift/6f2c8392 not reset as customized by admin to system_u:object_r:container_file_t:s0:c267,c588 Dec 13 06:49:03 crc restorecon[4764]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/containers/oauth-openshift/bd241ad9 not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Dec 13 06:49:03 crc restorecon[4764]: /var/lib/kubelet/plugins not reset as customized by admin to system_u:object_r:container_file_t:s0 Dec 13 06:49:03 crc restorecon[4764]: /var/lib/kubelet/plugins/csi-hostpath not reset as customized by admin to system_u:object_r:container_file_t:s0 Dec 13 06:49:03 crc restorecon[4764]: /var/lib/kubelet/plugins/csi-hostpath/csi.sock not reset as customized by admin to system_u:object_r:container_file_t:s0 Dec 13 06:49:03 crc restorecon[4764]: /var/lib/kubelet/plugins/kubernetes.io not reset as customized by admin to system_u:object_r:container_file_t:s0 Dec 13 06:49:03 crc restorecon[4764]: /var/lib/kubelet/plugins/kubernetes.io/csi not reset as customized by admin to system_u:object_r:container_file_t:s0 Dec 13 06:49:03 crc restorecon[4764]: /var/lib/kubelet/plugins/kubernetes.io/csi/kubevirt.io.hostpath-provisioner not reset as customized by admin to system_u:object_r:container_file_t:s0 Dec 13 06:49:03 crc restorecon[4764]: /var/lib/kubelet/plugins/kubernetes.io/csi/kubevirt.io.hostpath-provisioner/1f4776af88835e41c12b831b4c9fed40233456d14189815a54dbe7f892fc1983 not reset as customized by admin to system_u:object_r:container_file_t:s0 Dec 13 06:49:03 crc restorecon[4764]: /var/lib/kubelet/plugins/kubernetes.io/csi/kubevirt.io.hostpath-provisioner/1f4776af88835e41c12b831b4c9fed40233456d14189815a54dbe7f892fc1983/globalmount not reset as customized by admin to system_u:object_r:container_file_t:s0 Dec 13 06:49:03 crc restorecon[4764]: /var/lib/kubelet/plugins/kubernetes.io/csi/kubevirt.io.hostpath-provisioner/1f4776af88835e41c12b831b4c9fed40233456d14189815a54dbe7f892fc1983/vol_data.json not reset as customized by admin to system_u:object_r:container_file_t:s0 Dec 13 06:49:03 crc restorecon[4764]: /var/lib/kubelet/plugins_registry not reset as customized by admin to system_u:object_r:container_file_t:s0 Dec 13 06:49:03 crc restorecon[4764]: Relabeled /var/usrlocal/bin/kubenswrapper from system_u:object_r:bin_t:s0 to system_u:object_r:kubelet_exec_t:s0 Dec 13 06:49:03 crc kubenswrapper[4971]: Flag --container-runtime-endpoint has been deprecated, This parameter should be set via the config file specified by the Kubelet's --config flag. See https://kubernetes.io/docs/tasks/administer-cluster/kubelet-config-file/ for more information. Dec 13 06:49:03 crc kubenswrapper[4971]: Flag --minimum-container-ttl-duration has been deprecated, Use --eviction-hard or --eviction-soft instead. Will be removed in a future version. Dec 13 06:49:03 crc kubenswrapper[4971]: Flag --volume-plugin-dir has been deprecated, This parameter should be set via the config file specified by the Kubelet's --config flag. See https://kubernetes.io/docs/tasks/administer-cluster/kubelet-config-file/ for more information. Dec 13 06:49:03 crc kubenswrapper[4971]: Flag --register-with-taints has been deprecated, This parameter should be set via the config file specified by the Kubelet's --config flag. See https://kubernetes.io/docs/tasks/administer-cluster/kubelet-config-file/ for more information. Dec 13 06:49:03 crc kubenswrapper[4971]: Flag --pod-infra-container-image has been deprecated, will be removed in a future release. Image garbage collector will get sandbox image information from CRI. Dec 13 06:49:03 crc kubenswrapper[4971]: Flag --system-reserved has been deprecated, This parameter should be set via the config file specified by the Kubelet's --config flag. See https://kubernetes.io/docs/tasks/administer-cluster/kubelet-config-file/ for more information. Dec 13 06:49:03 crc kubenswrapper[4971]: I1213 06:49:03.547420 4971 server.go:211] "--pod-infra-container-image will not be pruned by the image garbage collector in kubelet and should also be set in the remote runtime" Dec 13 06:49:03 crc kubenswrapper[4971]: W1213 06:49:03.550539 4971 feature_gate.go:330] unrecognized feature gate: DNSNameResolver Dec 13 06:49:03 crc kubenswrapper[4971]: W1213 06:49:03.550551 4971 feature_gate.go:351] Setting deprecated feature gate KMSv1=true. It will be removed in a future release. Dec 13 06:49:03 crc kubenswrapper[4971]: W1213 06:49:03.550557 4971 feature_gate.go:330] unrecognized feature gate: NetworkLiveMigration Dec 13 06:49:03 crc kubenswrapper[4971]: W1213 06:49:03.550561 4971 feature_gate.go:330] unrecognized feature gate: ChunkSizeMiB Dec 13 06:49:03 crc kubenswrapper[4971]: W1213 06:49:03.550565 4971 feature_gate.go:330] unrecognized feature gate: MultiArchInstallGCP Dec 13 06:49:03 crc kubenswrapper[4971]: W1213 06:49:03.550569 4971 feature_gate.go:330] unrecognized feature gate: NetworkSegmentation Dec 13 06:49:03 crc kubenswrapper[4971]: W1213 06:49:03.550572 4971 feature_gate.go:330] unrecognized feature gate: VSphereDriverConfiguration Dec 13 06:49:03 crc kubenswrapper[4971]: W1213 06:49:03.550576 4971 feature_gate.go:330] unrecognized feature gate: AWSEFSDriverVolumeMetrics Dec 13 06:49:03 crc kubenswrapper[4971]: W1213 06:49:03.550580 4971 feature_gate.go:330] unrecognized feature gate: MachineAPIOperatorDisableMachineHealthCheckController Dec 13 06:49:03 crc kubenswrapper[4971]: W1213 06:49:03.550584 4971 feature_gate.go:330] unrecognized feature gate: ManagedBootImagesAWS Dec 13 06:49:03 crc kubenswrapper[4971]: W1213 06:49:03.550587 4971 feature_gate.go:330] unrecognized feature gate: MetricsCollectionProfiles Dec 13 06:49:03 crc kubenswrapper[4971]: W1213 06:49:03.550590 4971 feature_gate.go:330] unrecognized feature gate: VSphereMultiVCenters Dec 13 06:49:03 crc kubenswrapper[4971]: W1213 06:49:03.550594 4971 feature_gate.go:330] unrecognized feature gate: NewOLM Dec 13 06:49:03 crc kubenswrapper[4971]: W1213 06:49:03.550597 4971 feature_gate.go:330] unrecognized feature gate: OVNObservability Dec 13 06:49:03 crc kubenswrapper[4971]: W1213 06:49:03.550601 4971 feature_gate.go:330] unrecognized feature gate: SignatureStores Dec 13 06:49:03 crc kubenswrapper[4971]: W1213 06:49:03.550604 4971 feature_gate.go:330] unrecognized feature gate: SigstoreImageVerification Dec 13 06:49:03 crc kubenswrapper[4971]: W1213 06:49:03.550608 4971 feature_gate.go:330] unrecognized feature gate: MultiArchInstallAWS Dec 13 06:49:03 crc kubenswrapper[4971]: W1213 06:49:03.550615 4971 feature_gate.go:330] unrecognized feature gate: AlibabaPlatform Dec 13 06:49:03 crc kubenswrapper[4971]: W1213 06:49:03.550618 4971 feature_gate.go:330] unrecognized feature gate: PrivateHostedZoneAWS Dec 13 06:49:03 crc kubenswrapper[4971]: W1213 06:49:03.550622 4971 feature_gate.go:330] unrecognized feature gate: MixedCPUsAllocation Dec 13 06:49:03 crc kubenswrapper[4971]: W1213 06:49:03.550626 4971 feature_gate.go:330] unrecognized feature gate: VSphereControlPlaneMachineSet Dec 13 06:49:03 crc kubenswrapper[4971]: W1213 06:49:03.550630 4971 feature_gate.go:330] unrecognized feature gate: MultiArchInstallAzure Dec 13 06:49:03 crc kubenswrapper[4971]: W1213 06:49:03.550633 4971 feature_gate.go:330] unrecognized feature gate: AzureWorkloadIdentity Dec 13 06:49:03 crc kubenswrapper[4971]: W1213 06:49:03.550637 4971 feature_gate.go:330] unrecognized feature gate: GCPLabelsTags Dec 13 06:49:03 crc kubenswrapper[4971]: W1213 06:49:03.550640 4971 feature_gate.go:330] unrecognized feature gate: BareMetalLoadBalancer Dec 13 06:49:03 crc kubenswrapper[4971]: W1213 06:49:03.550643 4971 feature_gate.go:330] unrecognized feature gate: ClusterAPIInstall Dec 13 06:49:03 crc kubenswrapper[4971]: W1213 06:49:03.550647 4971 feature_gate.go:330] unrecognized feature gate: ManagedBootImages Dec 13 06:49:03 crc kubenswrapper[4971]: W1213 06:49:03.550650 4971 feature_gate.go:330] unrecognized feature gate: OpenShiftPodSecurityAdmission Dec 13 06:49:03 crc kubenswrapper[4971]: W1213 06:49:03.550653 4971 feature_gate.go:330] unrecognized feature gate: Example Dec 13 06:49:03 crc kubenswrapper[4971]: W1213 06:49:03.550657 4971 feature_gate.go:330] unrecognized feature gate: MachineConfigNodes Dec 13 06:49:03 crc kubenswrapper[4971]: W1213 06:49:03.550660 4971 feature_gate.go:330] unrecognized feature gate: MachineAPIProviderOpenStack Dec 13 06:49:03 crc kubenswrapper[4971]: W1213 06:49:03.550663 4971 feature_gate.go:330] unrecognized feature gate: CSIDriverSharedResource Dec 13 06:49:03 crc kubenswrapper[4971]: W1213 06:49:03.550667 4971 feature_gate.go:330] unrecognized feature gate: VSphereMultiNetworks Dec 13 06:49:03 crc kubenswrapper[4971]: W1213 06:49:03.550670 4971 feature_gate.go:330] unrecognized feature gate: NutanixMultiSubnets Dec 13 06:49:03 crc kubenswrapper[4971]: W1213 06:49:03.550674 4971 feature_gate.go:330] unrecognized feature gate: InsightsRuntimeExtractor Dec 13 06:49:03 crc kubenswrapper[4971]: W1213 06:49:03.550678 4971 feature_gate.go:330] unrecognized feature gate: IngressControllerDynamicConfigurationManager Dec 13 06:49:03 crc kubenswrapper[4971]: W1213 06:49:03.550681 4971 feature_gate.go:330] unrecognized feature gate: InsightsConfig Dec 13 06:49:03 crc kubenswrapper[4971]: W1213 06:49:03.550686 4971 feature_gate.go:330] unrecognized feature gate: MachineAPIMigration Dec 13 06:49:03 crc kubenswrapper[4971]: W1213 06:49:03.550690 4971 feature_gate.go:330] unrecognized feature gate: NetworkDiagnosticsConfig Dec 13 06:49:03 crc kubenswrapper[4971]: W1213 06:49:03.550693 4971 feature_gate.go:330] unrecognized feature gate: PersistentIPsForVirtualization Dec 13 06:49:03 crc kubenswrapper[4971]: W1213 06:49:03.550696 4971 feature_gate.go:330] unrecognized feature gate: ClusterMonitoringConfig Dec 13 06:49:03 crc kubenswrapper[4971]: W1213 06:49:03.550700 4971 feature_gate.go:330] unrecognized feature gate: AdminNetworkPolicy Dec 13 06:49:03 crc kubenswrapper[4971]: W1213 06:49:03.550703 4971 feature_gate.go:330] unrecognized feature gate: HardwareSpeed Dec 13 06:49:03 crc kubenswrapper[4971]: W1213 06:49:03.550707 4971 feature_gate.go:330] unrecognized feature gate: InsightsConfigAPI Dec 13 06:49:03 crc kubenswrapper[4971]: W1213 06:49:03.550710 4971 feature_gate.go:330] unrecognized feature gate: InsightsOnDemandDataGather Dec 13 06:49:03 crc kubenswrapper[4971]: W1213 06:49:03.550713 4971 feature_gate.go:330] unrecognized feature gate: NodeDisruptionPolicy Dec 13 06:49:03 crc kubenswrapper[4971]: W1213 06:49:03.550717 4971 feature_gate.go:330] unrecognized feature gate: AdditionalRoutingCapabilities Dec 13 06:49:03 crc kubenswrapper[4971]: W1213 06:49:03.550720 4971 feature_gate.go:330] unrecognized feature gate: GCPClusterHostedDNS Dec 13 06:49:03 crc kubenswrapper[4971]: W1213 06:49:03.550723 4971 feature_gate.go:330] unrecognized feature gate: ExternalOIDC Dec 13 06:49:03 crc kubenswrapper[4971]: W1213 06:49:03.550729 4971 feature_gate.go:353] Setting GA feature gate ValidatingAdmissionPolicy=true. It will be removed in a future release. Dec 13 06:49:03 crc kubenswrapper[4971]: W1213 06:49:03.550734 4971 feature_gate.go:330] unrecognized feature gate: ConsolePluginContentSecurityPolicy Dec 13 06:49:03 crc kubenswrapper[4971]: W1213 06:49:03.550738 4971 feature_gate.go:330] unrecognized feature gate: IngressControllerLBSubnetsAWS Dec 13 06:49:03 crc kubenswrapper[4971]: W1213 06:49:03.550742 4971 feature_gate.go:330] unrecognized feature gate: UpgradeStatus Dec 13 06:49:03 crc kubenswrapper[4971]: W1213 06:49:03.550745 4971 feature_gate.go:330] unrecognized feature gate: EtcdBackendQuota Dec 13 06:49:03 crc kubenswrapper[4971]: W1213 06:49:03.550749 4971 feature_gate.go:330] unrecognized feature gate: PinnedImages Dec 13 06:49:03 crc kubenswrapper[4971]: W1213 06:49:03.550753 4971 feature_gate.go:330] unrecognized feature gate: OnClusterBuild Dec 13 06:49:03 crc kubenswrapper[4971]: W1213 06:49:03.550757 4971 feature_gate.go:330] unrecognized feature gate: VSphereStaticIPs Dec 13 06:49:03 crc kubenswrapper[4971]: W1213 06:49:03.550761 4971 feature_gate.go:330] unrecognized feature gate: AutomatedEtcdBackup Dec 13 06:49:03 crc kubenswrapper[4971]: W1213 06:49:03.550764 4971 feature_gate.go:330] unrecognized feature gate: VolumeGroupSnapshot Dec 13 06:49:03 crc kubenswrapper[4971]: W1213 06:49:03.550768 4971 feature_gate.go:330] unrecognized feature gate: BootcNodeManagement Dec 13 06:49:03 crc kubenswrapper[4971]: W1213 06:49:03.550772 4971 feature_gate.go:330] unrecognized feature gate: AWSClusterHostedDNS Dec 13 06:49:03 crc kubenswrapper[4971]: W1213 06:49:03.550775 4971 feature_gate.go:330] unrecognized feature gate: SetEIPForNLBIngressController Dec 13 06:49:03 crc kubenswrapper[4971]: W1213 06:49:03.550779 4971 feature_gate.go:330] unrecognized feature gate: RouteAdvertisements Dec 13 06:49:03 crc kubenswrapper[4971]: W1213 06:49:03.550782 4971 feature_gate.go:330] unrecognized feature gate: GatewayAPI Dec 13 06:49:03 crc kubenswrapper[4971]: W1213 06:49:03.550787 4971 feature_gate.go:353] Setting GA feature gate CloudDualStackNodeIPs=true. It will be removed in a future release. Dec 13 06:49:03 crc kubenswrapper[4971]: W1213 06:49:03.550791 4971 feature_gate.go:330] unrecognized feature gate: BuildCSIVolumes Dec 13 06:49:03 crc kubenswrapper[4971]: W1213 06:49:03.550797 4971 feature_gate.go:353] Setting GA feature gate DisableKubeletCloudCredentialProviders=true. It will be removed in a future release. Dec 13 06:49:03 crc kubenswrapper[4971]: W1213 06:49:03.550801 4971 feature_gate.go:330] unrecognized feature gate: PlatformOperators Dec 13 06:49:03 crc kubenswrapper[4971]: W1213 06:49:03.550805 4971 feature_gate.go:330] unrecognized feature gate: MinimumKubeletVersion Dec 13 06:49:03 crc kubenswrapper[4971]: W1213 06:49:03.550809 4971 feature_gate.go:330] unrecognized feature gate: ImageStreamImportMode Dec 13 06:49:03 crc kubenswrapper[4971]: W1213 06:49:03.550812 4971 feature_gate.go:330] unrecognized feature gate: ClusterAPIInstallIBMCloud Dec 13 06:49:03 crc kubenswrapper[4971]: I1213 06:49:03.551183 4971 flags.go:64] FLAG: --address="0.0.0.0" Dec 13 06:49:03 crc kubenswrapper[4971]: I1213 06:49:03.551194 4971 flags.go:64] FLAG: --allowed-unsafe-sysctls="[]" Dec 13 06:49:03 crc kubenswrapper[4971]: I1213 06:49:03.551204 4971 flags.go:64] FLAG: --anonymous-auth="true" Dec 13 06:49:03 crc kubenswrapper[4971]: I1213 06:49:03.551209 4971 flags.go:64] FLAG: --application-metrics-count-limit="100" Dec 13 06:49:03 crc kubenswrapper[4971]: I1213 06:49:03.551215 4971 flags.go:64] FLAG: --authentication-token-webhook="false" Dec 13 06:49:03 crc kubenswrapper[4971]: I1213 06:49:03.551219 4971 flags.go:64] FLAG: --authentication-token-webhook-cache-ttl="2m0s" Dec 13 06:49:03 crc kubenswrapper[4971]: I1213 06:49:03.551226 4971 flags.go:64] FLAG: --authorization-mode="AlwaysAllow" Dec 13 06:49:03 crc kubenswrapper[4971]: I1213 06:49:03.551231 4971 flags.go:64] FLAG: --authorization-webhook-cache-authorized-ttl="5m0s" Dec 13 06:49:03 crc kubenswrapper[4971]: I1213 06:49:03.551235 4971 flags.go:64] FLAG: --authorization-webhook-cache-unauthorized-ttl="30s" Dec 13 06:49:03 crc kubenswrapper[4971]: I1213 06:49:03.551239 4971 flags.go:64] FLAG: --boot-id-file="/proc/sys/kernel/random/boot_id" Dec 13 06:49:03 crc kubenswrapper[4971]: I1213 06:49:03.551244 4971 flags.go:64] FLAG: --bootstrap-kubeconfig="/etc/kubernetes/kubeconfig" Dec 13 06:49:03 crc kubenswrapper[4971]: I1213 06:49:03.551248 4971 flags.go:64] FLAG: --cert-dir="/var/lib/kubelet/pki" Dec 13 06:49:03 crc kubenswrapper[4971]: I1213 06:49:03.551252 4971 flags.go:64] FLAG: --cgroup-driver="cgroupfs" Dec 13 06:49:03 crc kubenswrapper[4971]: I1213 06:49:03.551256 4971 flags.go:64] FLAG: --cgroup-root="" Dec 13 06:49:03 crc kubenswrapper[4971]: I1213 06:49:03.551260 4971 flags.go:64] FLAG: --cgroups-per-qos="true" Dec 13 06:49:03 crc kubenswrapper[4971]: I1213 06:49:03.551265 4971 flags.go:64] FLAG: --client-ca-file="" Dec 13 06:49:03 crc kubenswrapper[4971]: I1213 06:49:03.551269 4971 flags.go:64] FLAG: --cloud-config="" Dec 13 06:49:03 crc kubenswrapper[4971]: I1213 06:49:03.551273 4971 flags.go:64] FLAG: --cloud-provider="" Dec 13 06:49:03 crc kubenswrapper[4971]: I1213 06:49:03.551277 4971 flags.go:64] FLAG: --cluster-dns="[]" Dec 13 06:49:03 crc kubenswrapper[4971]: I1213 06:49:03.551284 4971 flags.go:64] FLAG: --cluster-domain="" Dec 13 06:49:03 crc kubenswrapper[4971]: I1213 06:49:03.551288 4971 flags.go:64] FLAG: --config="/etc/kubernetes/kubelet.conf" Dec 13 06:49:03 crc kubenswrapper[4971]: I1213 06:49:03.551292 4971 flags.go:64] FLAG: --config-dir="" Dec 13 06:49:03 crc kubenswrapper[4971]: I1213 06:49:03.551296 4971 flags.go:64] FLAG: --container-hints="/etc/cadvisor/container_hints.json" Dec 13 06:49:03 crc kubenswrapper[4971]: I1213 06:49:03.551300 4971 flags.go:64] FLAG: --container-log-max-files="5" Dec 13 06:49:03 crc kubenswrapper[4971]: I1213 06:49:03.551307 4971 flags.go:64] FLAG: --container-log-max-size="10Mi" Dec 13 06:49:03 crc kubenswrapper[4971]: I1213 06:49:03.551311 4971 flags.go:64] FLAG: --container-runtime-endpoint="/var/run/crio/crio.sock" Dec 13 06:49:03 crc kubenswrapper[4971]: I1213 06:49:03.551315 4971 flags.go:64] FLAG: --containerd="/run/containerd/containerd.sock" Dec 13 06:49:03 crc kubenswrapper[4971]: I1213 06:49:03.551320 4971 flags.go:64] FLAG: --containerd-namespace="k8s.io" Dec 13 06:49:03 crc kubenswrapper[4971]: I1213 06:49:03.551324 4971 flags.go:64] FLAG: --contention-profiling="false" Dec 13 06:49:03 crc kubenswrapper[4971]: I1213 06:49:03.551328 4971 flags.go:64] FLAG: --cpu-cfs-quota="true" Dec 13 06:49:03 crc kubenswrapper[4971]: I1213 06:49:03.551332 4971 flags.go:64] FLAG: --cpu-cfs-quota-period="100ms" Dec 13 06:49:03 crc kubenswrapper[4971]: I1213 06:49:03.551336 4971 flags.go:64] FLAG: --cpu-manager-policy="none" Dec 13 06:49:03 crc kubenswrapper[4971]: I1213 06:49:03.551341 4971 flags.go:64] FLAG: --cpu-manager-policy-options="" Dec 13 06:49:03 crc kubenswrapper[4971]: I1213 06:49:03.551347 4971 flags.go:64] FLAG: --cpu-manager-reconcile-period="10s" Dec 13 06:49:03 crc kubenswrapper[4971]: I1213 06:49:03.551351 4971 flags.go:64] FLAG: --enable-controller-attach-detach="true" Dec 13 06:49:03 crc kubenswrapper[4971]: I1213 06:49:03.551355 4971 flags.go:64] FLAG: --enable-debugging-handlers="true" Dec 13 06:49:03 crc kubenswrapper[4971]: I1213 06:49:03.551359 4971 flags.go:64] FLAG: --enable-load-reader="false" Dec 13 06:49:03 crc kubenswrapper[4971]: I1213 06:49:03.551363 4971 flags.go:64] FLAG: --enable-server="true" Dec 13 06:49:03 crc kubenswrapper[4971]: I1213 06:49:03.551367 4971 flags.go:64] FLAG: --enforce-node-allocatable="[pods]" Dec 13 06:49:03 crc kubenswrapper[4971]: I1213 06:49:03.551373 4971 flags.go:64] FLAG: --event-burst="100" Dec 13 06:49:03 crc kubenswrapper[4971]: I1213 06:49:03.551378 4971 flags.go:64] FLAG: --event-qps="50" Dec 13 06:49:03 crc kubenswrapper[4971]: I1213 06:49:03.551382 4971 flags.go:64] FLAG: --event-storage-age-limit="default=0" Dec 13 06:49:03 crc kubenswrapper[4971]: I1213 06:49:03.551386 4971 flags.go:64] FLAG: --event-storage-event-limit="default=0" Dec 13 06:49:03 crc kubenswrapper[4971]: I1213 06:49:03.551390 4971 flags.go:64] FLAG: --eviction-hard="" Dec 13 06:49:03 crc kubenswrapper[4971]: I1213 06:49:03.551395 4971 flags.go:64] FLAG: --eviction-max-pod-grace-period="0" Dec 13 06:49:03 crc kubenswrapper[4971]: I1213 06:49:03.551400 4971 flags.go:64] FLAG: --eviction-minimum-reclaim="" Dec 13 06:49:03 crc kubenswrapper[4971]: I1213 06:49:03.551404 4971 flags.go:64] FLAG: --eviction-pressure-transition-period="5m0s" Dec 13 06:49:03 crc kubenswrapper[4971]: I1213 06:49:03.551408 4971 flags.go:64] FLAG: --eviction-soft="" Dec 13 06:49:03 crc kubenswrapper[4971]: I1213 06:49:03.551412 4971 flags.go:64] FLAG: --eviction-soft-grace-period="" Dec 13 06:49:03 crc kubenswrapper[4971]: I1213 06:49:03.551417 4971 flags.go:64] FLAG: --exit-on-lock-contention="false" Dec 13 06:49:03 crc kubenswrapper[4971]: I1213 06:49:03.551421 4971 flags.go:64] FLAG: --experimental-allocatable-ignore-eviction="false" Dec 13 06:49:03 crc kubenswrapper[4971]: I1213 06:49:03.551426 4971 flags.go:64] FLAG: --experimental-mounter-path="" Dec 13 06:49:03 crc kubenswrapper[4971]: I1213 06:49:03.551430 4971 flags.go:64] FLAG: --fail-cgroupv1="false" Dec 13 06:49:03 crc kubenswrapper[4971]: I1213 06:49:03.551434 4971 flags.go:64] FLAG: --fail-swap-on="true" Dec 13 06:49:03 crc kubenswrapper[4971]: I1213 06:49:03.551438 4971 flags.go:64] FLAG: --feature-gates="" Dec 13 06:49:03 crc kubenswrapper[4971]: I1213 06:49:03.551443 4971 flags.go:64] FLAG: --file-check-frequency="20s" Dec 13 06:49:03 crc kubenswrapper[4971]: I1213 06:49:03.551447 4971 flags.go:64] FLAG: --global-housekeeping-interval="1m0s" Dec 13 06:49:03 crc kubenswrapper[4971]: I1213 06:49:03.551452 4971 flags.go:64] FLAG: --hairpin-mode="promiscuous-bridge" Dec 13 06:49:03 crc kubenswrapper[4971]: I1213 06:49:03.551457 4971 flags.go:64] FLAG: --healthz-bind-address="127.0.0.1" Dec 13 06:49:03 crc kubenswrapper[4971]: I1213 06:49:03.551462 4971 flags.go:64] FLAG: --healthz-port="10248" Dec 13 06:49:03 crc kubenswrapper[4971]: I1213 06:49:03.551466 4971 flags.go:64] FLAG: --help="false" Dec 13 06:49:03 crc kubenswrapper[4971]: I1213 06:49:03.551470 4971 flags.go:64] FLAG: --hostname-override="" Dec 13 06:49:03 crc kubenswrapper[4971]: I1213 06:49:03.551474 4971 flags.go:64] FLAG: --housekeeping-interval="10s" Dec 13 06:49:03 crc kubenswrapper[4971]: I1213 06:49:03.551478 4971 flags.go:64] FLAG: --http-check-frequency="20s" Dec 13 06:49:03 crc kubenswrapper[4971]: I1213 06:49:03.551482 4971 flags.go:64] FLAG: --image-credential-provider-bin-dir="" Dec 13 06:49:03 crc kubenswrapper[4971]: I1213 06:49:03.551487 4971 flags.go:64] FLAG: --image-credential-provider-config="" Dec 13 06:49:03 crc kubenswrapper[4971]: I1213 06:49:03.551491 4971 flags.go:64] FLAG: --image-gc-high-threshold="85" Dec 13 06:49:03 crc kubenswrapper[4971]: I1213 06:49:03.551495 4971 flags.go:64] FLAG: --image-gc-low-threshold="80" Dec 13 06:49:03 crc kubenswrapper[4971]: I1213 06:49:03.551500 4971 flags.go:64] FLAG: --image-service-endpoint="" Dec 13 06:49:03 crc kubenswrapper[4971]: I1213 06:49:03.551504 4971 flags.go:64] FLAG: --kernel-memcg-notification="false" Dec 13 06:49:03 crc kubenswrapper[4971]: I1213 06:49:03.551508 4971 flags.go:64] FLAG: --kube-api-burst="100" Dec 13 06:49:03 crc kubenswrapper[4971]: I1213 06:49:03.551512 4971 flags.go:64] FLAG: --kube-api-content-type="application/vnd.kubernetes.protobuf" Dec 13 06:49:03 crc kubenswrapper[4971]: I1213 06:49:03.551529 4971 flags.go:64] FLAG: --kube-api-qps="50" Dec 13 06:49:03 crc kubenswrapper[4971]: I1213 06:49:03.551534 4971 flags.go:64] FLAG: --kube-reserved="" Dec 13 06:49:03 crc kubenswrapper[4971]: I1213 06:49:03.551538 4971 flags.go:64] FLAG: --kube-reserved-cgroup="" Dec 13 06:49:03 crc kubenswrapper[4971]: I1213 06:49:03.551542 4971 flags.go:64] FLAG: --kubeconfig="/var/lib/kubelet/kubeconfig" Dec 13 06:49:03 crc kubenswrapper[4971]: I1213 06:49:03.551546 4971 flags.go:64] FLAG: --kubelet-cgroups="" Dec 13 06:49:03 crc kubenswrapper[4971]: I1213 06:49:03.551551 4971 flags.go:64] FLAG: --local-storage-capacity-isolation="true" Dec 13 06:49:03 crc kubenswrapper[4971]: I1213 06:49:03.551555 4971 flags.go:64] FLAG: --lock-file="" Dec 13 06:49:03 crc kubenswrapper[4971]: I1213 06:49:03.551559 4971 flags.go:64] FLAG: --log-cadvisor-usage="false" Dec 13 06:49:03 crc kubenswrapper[4971]: I1213 06:49:03.551563 4971 flags.go:64] FLAG: --log-flush-frequency="5s" Dec 13 06:49:03 crc kubenswrapper[4971]: I1213 06:49:03.551568 4971 flags.go:64] FLAG: --log-json-info-buffer-size="0" Dec 13 06:49:03 crc kubenswrapper[4971]: I1213 06:49:03.551575 4971 flags.go:64] FLAG: --log-json-split-stream="false" Dec 13 06:49:03 crc kubenswrapper[4971]: I1213 06:49:03.551579 4971 flags.go:64] FLAG: --log-text-info-buffer-size="0" Dec 13 06:49:03 crc kubenswrapper[4971]: I1213 06:49:03.551583 4971 flags.go:64] FLAG: --log-text-split-stream="false" Dec 13 06:49:03 crc kubenswrapper[4971]: I1213 06:49:03.551587 4971 flags.go:64] FLAG: --logging-format="text" Dec 13 06:49:03 crc kubenswrapper[4971]: I1213 06:49:03.551591 4971 flags.go:64] FLAG: --machine-id-file="/etc/machine-id,/var/lib/dbus/machine-id" Dec 13 06:49:03 crc kubenswrapper[4971]: I1213 06:49:03.551596 4971 flags.go:64] FLAG: --make-iptables-util-chains="true" Dec 13 06:49:03 crc kubenswrapper[4971]: I1213 06:49:03.551600 4971 flags.go:64] FLAG: --manifest-url="" Dec 13 06:49:03 crc kubenswrapper[4971]: I1213 06:49:03.551605 4971 flags.go:64] FLAG: --manifest-url-header="" Dec 13 06:49:03 crc kubenswrapper[4971]: I1213 06:49:03.551610 4971 flags.go:64] FLAG: --max-housekeeping-interval="15s" Dec 13 06:49:03 crc kubenswrapper[4971]: I1213 06:49:03.551615 4971 flags.go:64] FLAG: --max-open-files="1000000" Dec 13 06:49:03 crc kubenswrapper[4971]: I1213 06:49:03.551620 4971 flags.go:64] FLAG: --max-pods="110" Dec 13 06:49:03 crc kubenswrapper[4971]: I1213 06:49:03.551625 4971 flags.go:64] FLAG: --maximum-dead-containers="-1" Dec 13 06:49:03 crc kubenswrapper[4971]: I1213 06:49:03.551629 4971 flags.go:64] FLAG: --maximum-dead-containers-per-container="1" Dec 13 06:49:03 crc kubenswrapper[4971]: I1213 06:49:03.551633 4971 flags.go:64] FLAG: --memory-manager-policy="None" Dec 13 06:49:03 crc kubenswrapper[4971]: I1213 06:49:03.551637 4971 flags.go:64] FLAG: --minimum-container-ttl-duration="6m0s" Dec 13 06:49:03 crc kubenswrapper[4971]: I1213 06:49:03.551642 4971 flags.go:64] FLAG: --minimum-image-ttl-duration="2m0s" Dec 13 06:49:03 crc kubenswrapper[4971]: I1213 06:49:03.551646 4971 flags.go:64] FLAG: --node-ip="192.168.126.11" Dec 13 06:49:03 crc kubenswrapper[4971]: I1213 06:49:03.551650 4971 flags.go:64] FLAG: --node-labels="node-role.kubernetes.io/control-plane=,node-role.kubernetes.io/master=,node.openshift.io/os_id=rhcos" Dec 13 06:49:03 crc kubenswrapper[4971]: I1213 06:49:03.551660 4971 flags.go:64] FLAG: --node-status-max-images="50" Dec 13 06:49:03 crc kubenswrapper[4971]: I1213 06:49:03.551664 4971 flags.go:64] FLAG: --node-status-update-frequency="10s" Dec 13 06:49:03 crc kubenswrapper[4971]: I1213 06:49:03.551669 4971 flags.go:64] FLAG: --oom-score-adj="-999" Dec 13 06:49:03 crc kubenswrapper[4971]: I1213 06:49:03.551673 4971 flags.go:64] FLAG: --pod-cidr="" Dec 13 06:49:03 crc kubenswrapper[4971]: I1213 06:49:03.551678 4971 flags.go:64] FLAG: --pod-infra-container-image="quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:33549946e22a9ffa738fd94b1345f90921bc8f92fa6137784cb33c77ad806f9d" Dec 13 06:49:03 crc kubenswrapper[4971]: I1213 06:49:03.551684 4971 flags.go:64] FLAG: --pod-manifest-path="" Dec 13 06:49:03 crc kubenswrapper[4971]: I1213 06:49:03.551688 4971 flags.go:64] FLAG: --pod-max-pids="-1" Dec 13 06:49:03 crc kubenswrapper[4971]: I1213 06:49:03.551693 4971 flags.go:64] FLAG: --pods-per-core="0" Dec 13 06:49:03 crc kubenswrapper[4971]: I1213 06:49:03.551697 4971 flags.go:64] FLAG: --port="10250" Dec 13 06:49:03 crc kubenswrapper[4971]: I1213 06:49:03.551701 4971 flags.go:64] FLAG: --protect-kernel-defaults="false" Dec 13 06:49:03 crc kubenswrapper[4971]: I1213 06:49:03.551705 4971 flags.go:64] FLAG: --provider-id="" Dec 13 06:49:03 crc kubenswrapper[4971]: I1213 06:49:03.551709 4971 flags.go:64] FLAG: --qos-reserved="" Dec 13 06:49:03 crc kubenswrapper[4971]: I1213 06:49:03.551713 4971 flags.go:64] FLAG: --read-only-port="10255" Dec 13 06:49:03 crc kubenswrapper[4971]: I1213 06:49:03.551717 4971 flags.go:64] FLAG: --register-node="true" Dec 13 06:49:03 crc kubenswrapper[4971]: I1213 06:49:03.551722 4971 flags.go:64] FLAG: --register-schedulable="true" Dec 13 06:49:03 crc kubenswrapper[4971]: I1213 06:49:03.551726 4971 flags.go:64] FLAG: --register-with-taints="node-role.kubernetes.io/master=:NoSchedule" Dec 13 06:49:03 crc kubenswrapper[4971]: I1213 06:49:03.551733 4971 flags.go:64] FLAG: --registry-burst="10" Dec 13 06:49:03 crc kubenswrapper[4971]: I1213 06:49:03.551737 4971 flags.go:64] FLAG: --registry-qps="5" Dec 13 06:49:03 crc kubenswrapper[4971]: I1213 06:49:03.551741 4971 flags.go:64] FLAG: --reserved-cpus="" Dec 13 06:49:03 crc kubenswrapper[4971]: I1213 06:49:03.551745 4971 flags.go:64] FLAG: --reserved-memory="" Dec 13 06:49:03 crc kubenswrapper[4971]: I1213 06:49:03.551750 4971 flags.go:64] FLAG: --resolv-conf="/etc/resolv.conf" Dec 13 06:49:03 crc kubenswrapper[4971]: I1213 06:49:03.551754 4971 flags.go:64] FLAG: --root-dir="/var/lib/kubelet" Dec 13 06:49:03 crc kubenswrapper[4971]: I1213 06:49:03.551758 4971 flags.go:64] FLAG: --rotate-certificates="false" Dec 13 06:49:03 crc kubenswrapper[4971]: I1213 06:49:03.551762 4971 flags.go:64] FLAG: --rotate-server-certificates="false" Dec 13 06:49:03 crc kubenswrapper[4971]: I1213 06:49:03.551767 4971 flags.go:64] FLAG: --runonce="false" Dec 13 06:49:03 crc kubenswrapper[4971]: I1213 06:49:03.551772 4971 flags.go:64] FLAG: --runtime-cgroups="/system.slice/crio.service" Dec 13 06:49:03 crc kubenswrapper[4971]: I1213 06:49:03.551776 4971 flags.go:64] FLAG: --runtime-request-timeout="2m0s" Dec 13 06:49:03 crc kubenswrapper[4971]: I1213 06:49:03.551780 4971 flags.go:64] FLAG: --seccomp-default="false" Dec 13 06:49:03 crc kubenswrapper[4971]: I1213 06:49:03.551784 4971 flags.go:64] FLAG: --serialize-image-pulls="true" Dec 13 06:49:03 crc kubenswrapper[4971]: I1213 06:49:03.551788 4971 flags.go:64] FLAG: --storage-driver-buffer-duration="1m0s" Dec 13 06:49:03 crc kubenswrapper[4971]: I1213 06:49:03.551792 4971 flags.go:64] FLAG: --storage-driver-db="cadvisor" Dec 13 06:49:03 crc kubenswrapper[4971]: I1213 06:49:03.551796 4971 flags.go:64] FLAG: --storage-driver-host="localhost:8086" Dec 13 06:49:03 crc kubenswrapper[4971]: I1213 06:49:03.551800 4971 flags.go:64] FLAG: --storage-driver-password="root" Dec 13 06:49:03 crc kubenswrapper[4971]: I1213 06:49:03.551804 4971 flags.go:64] FLAG: --storage-driver-secure="false" Dec 13 06:49:03 crc kubenswrapper[4971]: I1213 06:49:03.551809 4971 flags.go:64] FLAG: --storage-driver-table="stats" Dec 13 06:49:03 crc kubenswrapper[4971]: I1213 06:49:03.551813 4971 flags.go:64] FLAG: --storage-driver-user="root" Dec 13 06:49:03 crc kubenswrapper[4971]: I1213 06:49:03.551816 4971 flags.go:64] FLAG: --streaming-connection-idle-timeout="4h0m0s" Dec 13 06:49:03 crc kubenswrapper[4971]: I1213 06:49:03.551821 4971 flags.go:64] FLAG: --sync-frequency="1m0s" Dec 13 06:49:03 crc kubenswrapper[4971]: I1213 06:49:03.551825 4971 flags.go:64] FLAG: --system-cgroups="" Dec 13 06:49:03 crc kubenswrapper[4971]: I1213 06:49:03.551829 4971 flags.go:64] FLAG: --system-reserved="cpu=200m,ephemeral-storage=350Mi,memory=350Mi" Dec 13 06:49:03 crc kubenswrapper[4971]: I1213 06:49:03.551837 4971 flags.go:64] FLAG: --system-reserved-cgroup="" Dec 13 06:49:03 crc kubenswrapper[4971]: I1213 06:49:03.551841 4971 flags.go:64] FLAG: --tls-cert-file="" Dec 13 06:49:03 crc kubenswrapper[4971]: I1213 06:49:03.551845 4971 flags.go:64] FLAG: --tls-cipher-suites="[]" Dec 13 06:49:03 crc kubenswrapper[4971]: I1213 06:49:03.551850 4971 flags.go:64] FLAG: --tls-min-version="" Dec 13 06:49:03 crc kubenswrapper[4971]: I1213 06:49:03.551854 4971 flags.go:64] FLAG: --tls-private-key-file="" Dec 13 06:49:03 crc kubenswrapper[4971]: I1213 06:49:03.551859 4971 flags.go:64] FLAG: --topology-manager-policy="none" Dec 13 06:49:03 crc kubenswrapper[4971]: I1213 06:49:03.551863 4971 flags.go:64] FLAG: --topology-manager-policy-options="" Dec 13 06:49:03 crc kubenswrapper[4971]: I1213 06:49:03.551867 4971 flags.go:64] FLAG: --topology-manager-scope="container" Dec 13 06:49:03 crc kubenswrapper[4971]: I1213 06:49:03.551871 4971 flags.go:64] FLAG: --v="2" Dec 13 06:49:03 crc kubenswrapper[4971]: I1213 06:49:03.551881 4971 flags.go:64] FLAG: --version="false" Dec 13 06:49:03 crc kubenswrapper[4971]: I1213 06:49:03.551886 4971 flags.go:64] FLAG: --vmodule="" Dec 13 06:49:03 crc kubenswrapper[4971]: I1213 06:49:03.551891 4971 flags.go:64] FLAG: --volume-plugin-dir="/etc/kubernetes/kubelet-plugins/volume/exec" Dec 13 06:49:03 crc kubenswrapper[4971]: I1213 06:49:03.551895 4971 flags.go:64] FLAG: --volume-stats-agg-period="1m0s" Dec 13 06:49:03 crc kubenswrapper[4971]: W1213 06:49:03.552007 4971 feature_gate.go:330] unrecognized feature gate: AdminNetworkPolicy Dec 13 06:49:03 crc kubenswrapper[4971]: W1213 06:49:03.552012 4971 feature_gate.go:330] unrecognized feature gate: BareMetalLoadBalancer Dec 13 06:49:03 crc kubenswrapper[4971]: W1213 06:49:03.552016 4971 feature_gate.go:330] unrecognized feature gate: PersistentIPsForVirtualization Dec 13 06:49:03 crc kubenswrapper[4971]: W1213 06:49:03.552021 4971 feature_gate.go:330] unrecognized feature gate: MultiArchInstallGCP Dec 13 06:49:03 crc kubenswrapper[4971]: W1213 06:49:03.552025 4971 feature_gate.go:330] unrecognized feature gate: NutanixMultiSubnets Dec 13 06:49:03 crc kubenswrapper[4971]: W1213 06:49:03.552028 4971 feature_gate.go:330] unrecognized feature gate: MixedCPUsAllocation Dec 13 06:49:03 crc kubenswrapper[4971]: W1213 06:49:03.552032 4971 feature_gate.go:330] unrecognized feature gate: NetworkDiagnosticsConfig Dec 13 06:49:03 crc kubenswrapper[4971]: W1213 06:49:03.552036 4971 feature_gate.go:330] unrecognized feature gate: UpgradeStatus Dec 13 06:49:03 crc kubenswrapper[4971]: W1213 06:49:03.552039 4971 feature_gate.go:330] unrecognized feature gate: PrivateHostedZoneAWS Dec 13 06:49:03 crc kubenswrapper[4971]: W1213 06:49:03.552043 4971 feature_gate.go:330] unrecognized feature gate: AWSClusterHostedDNS Dec 13 06:49:03 crc kubenswrapper[4971]: W1213 06:49:03.552046 4971 feature_gate.go:330] unrecognized feature gate: NetworkSegmentation Dec 13 06:49:03 crc kubenswrapper[4971]: W1213 06:49:03.552049 4971 feature_gate.go:330] unrecognized feature gate: OVNObservability Dec 13 06:49:03 crc kubenswrapper[4971]: W1213 06:49:03.552054 4971 feature_gate.go:353] Setting GA feature gate ValidatingAdmissionPolicy=true. It will be removed in a future release. Dec 13 06:49:03 crc kubenswrapper[4971]: W1213 06:49:03.552059 4971 feature_gate.go:330] unrecognized feature gate: ManagedBootImages Dec 13 06:49:03 crc kubenswrapper[4971]: W1213 06:49:03.552063 4971 feature_gate.go:330] unrecognized feature gate: PlatformOperators Dec 13 06:49:03 crc kubenswrapper[4971]: W1213 06:49:03.552066 4971 feature_gate.go:330] unrecognized feature gate: AzureWorkloadIdentity Dec 13 06:49:03 crc kubenswrapper[4971]: W1213 06:49:03.552070 4971 feature_gate.go:330] unrecognized feature gate: PinnedImages Dec 13 06:49:03 crc kubenswrapper[4971]: W1213 06:49:03.552074 4971 feature_gate.go:330] unrecognized feature gate: ClusterAPIInstall Dec 13 06:49:03 crc kubenswrapper[4971]: W1213 06:49:03.552078 4971 feature_gate.go:330] unrecognized feature gate: ExternalOIDC Dec 13 06:49:03 crc kubenswrapper[4971]: W1213 06:49:03.552082 4971 feature_gate.go:330] unrecognized feature gate: NetworkLiveMigration Dec 13 06:49:03 crc kubenswrapper[4971]: W1213 06:49:03.552086 4971 feature_gate.go:330] unrecognized feature gate: IngressControllerDynamicConfigurationManager Dec 13 06:49:03 crc kubenswrapper[4971]: W1213 06:49:03.552089 4971 feature_gate.go:330] unrecognized feature gate: NodeDisruptionPolicy Dec 13 06:49:03 crc kubenswrapper[4971]: W1213 06:49:03.552094 4971 feature_gate.go:330] unrecognized feature gate: MachineConfigNodes Dec 13 06:49:03 crc kubenswrapper[4971]: W1213 06:49:03.552098 4971 feature_gate.go:330] unrecognized feature gate: DNSNameResolver Dec 13 06:49:03 crc kubenswrapper[4971]: W1213 06:49:03.552102 4971 feature_gate.go:330] unrecognized feature gate: OnClusterBuild Dec 13 06:49:03 crc kubenswrapper[4971]: W1213 06:49:03.552107 4971 feature_gate.go:353] Setting GA feature gate DisableKubeletCloudCredentialProviders=true. It will be removed in a future release. Dec 13 06:49:03 crc kubenswrapper[4971]: W1213 06:49:03.552111 4971 feature_gate.go:330] unrecognized feature gate: Example Dec 13 06:49:03 crc kubenswrapper[4971]: W1213 06:49:03.552115 4971 feature_gate.go:330] unrecognized feature gate: SigstoreImageVerification Dec 13 06:49:03 crc kubenswrapper[4971]: W1213 06:49:03.552119 4971 feature_gate.go:330] unrecognized feature gate: IngressControllerLBSubnetsAWS Dec 13 06:49:03 crc kubenswrapper[4971]: W1213 06:49:03.552123 4971 feature_gate.go:351] Setting deprecated feature gate KMSv1=true. It will be removed in a future release. Dec 13 06:49:03 crc kubenswrapper[4971]: W1213 06:49:03.552128 4971 feature_gate.go:330] unrecognized feature gate: InsightsOnDemandDataGather Dec 13 06:49:03 crc kubenswrapper[4971]: W1213 06:49:03.552131 4971 feature_gate.go:330] unrecognized feature gate: ConsolePluginContentSecurityPolicy Dec 13 06:49:03 crc kubenswrapper[4971]: W1213 06:49:03.552135 4971 feature_gate.go:330] unrecognized feature gate: MachineAPIOperatorDisableMachineHealthCheckController Dec 13 06:49:03 crc kubenswrapper[4971]: W1213 06:49:03.552138 4971 feature_gate.go:330] unrecognized feature gate: EtcdBackendQuota Dec 13 06:49:03 crc kubenswrapper[4971]: W1213 06:49:03.552142 4971 feature_gate.go:330] unrecognized feature gate: AWSEFSDriverVolumeMetrics Dec 13 06:49:03 crc kubenswrapper[4971]: W1213 06:49:03.552145 4971 feature_gate.go:330] unrecognized feature gate: MultiArchInstallAWS Dec 13 06:49:03 crc kubenswrapper[4971]: W1213 06:49:03.552149 4971 feature_gate.go:330] unrecognized feature gate: ImageStreamImportMode Dec 13 06:49:03 crc kubenswrapper[4971]: W1213 06:49:03.552152 4971 feature_gate.go:330] unrecognized feature gate: VSphereDriverConfiguration Dec 13 06:49:03 crc kubenswrapper[4971]: W1213 06:49:03.552156 4971 feature_gate.go:330] unrecognized feature gate: VSphereMultiVCenters Dec 13 06:49:03 crc kubenswrapper[4971]: W1213 06:49:03.552159 4971 feature_gate.go:330] unrecognized feature gate: ManagedBootImagesAWS Dec 13 06:49:03 crc kubenswrapper[4971]: W1213 06:49:03.552163 4971 feature_gate.go:330] unrecognized feature gate: MetricsCollectionProfiles Dec 13 06:49:03 crc kubenswrapper[4971]: W1213 06:49:03.552167 4971 feature_gate.go:330] unrecognized feature gate: BuildCSIVolumes Dec 13 06:49:03 crc kubenswrapper[4971]: W1213 06:49:03.552170 4971 feature_gate.go:330] unrecognized feature gate: MachineAPIMigration Dec 13 06:49:03 crc kubenswrapper[4971]: W1213 06:49:03.552174 4971 feature_gate.go:330] unrecognized feature gate: HardwareSpeed Dec 13 06:49:03 crc kubenswrapper[4971]: W1213 06:49:03.552177 4971 feature_gate.go:330] unrecognized feature gate: MultiArchInstallAzure Dec 13 06:49:03 crc kubenswrapper[4971]: W1213 06:49:03.552181 4971 feature_gate.go:330] unrecognized feature gate: AlibabaPlatform Dec 13 06:49:03 crc kubenswrapper[4971]: W1213 06:49:03.552184 4971 feature_gate.go:330] unrecognized feature gate: ClusterAPIInstallIBMCloud Dec 13 06:49:03 crc kubenswrapper[4971]: W1213 06:49:03.552187 4971 feature_gate.go:330] unrecognized feature gate: VSphereControlPlaneMachineSet Dec 13 06:49:03 crc kubenswrapper[4971]: W1213 06:49:03.552191 4971 feature_gate.go:330] unrecognized feature gate: BootcNodeManagement Dec 13 06:49:03 crc kubenswrapper[4971]: W1213 06:49:03.552194 4971 feature_gate.go:330] unrecognized feature gate: VSphereMultiNetworks Dec 13 06:49:03 crc kubenswrapper[4971]: W1213 06:49:03.552198 4971 feature_gate.go:330] unrecognized feature gate: InsightsRuntimeExtractor Dec 13 06:49:03 crc kubenswrapper[4971]: W1213 06:49:03.552202 4971 feature_gate.go:330] unrecognized feature gate: VolumeGroupSnapshot Dec 13 06:49:03 crc kubenswrapper[4971]: W1213 06:49:03.552205 4971 feature_gate.go:330] unrecognized feature gate: OpenShiftPodSecurityAdmission Dec 13 06:49:03 crc kubenswrapper[4971]: W1213 06:49:03.552209 4971 feature_gate.go:330] unrecognized feature gate: ClusterMonitoringConfig Dec 13 06:49:03 crc kubenswrapper[4971]: W1213 06:49:03.552213 4971 feature_gate.go:330] unrecognized feature gate: GatewayAPI Dec 13 06:49:03 crc kubenswrapper[4971]: W1213 06:49:03.552216 4971 feature_gate.go:330] unrecognized feature gate: InsightsConfig Dec 13 06:49:03 crc kubenswrapper[4971]: W1213 06:49:03.552219 4971 feature_gate.go:330] unrecognized feature gate: AdditionalRoutingCapabilities Dec 13 06:49:03 crc kubenswrapper[4971]: W1213 06:49:03.552223 4971 feature_gate.go:330] unrecognized feature gate: InsightsConfigAPI Dec 13 06:49:03 crc kubenswrapper[4971]: W1213 06:49:03.552226 4971 feature_gate.go:330] unrecognized feature gate: SetEIPForNLBIngressController Dec 13 06:49:03 crc kubenswrapper[4971]: W1213 06:49:03.552232 4971 feature_gate.go:353] Setting GA feature gate CloudDualStackNodeIPs=true. It will be removed in a future release. Dec 13 06:49:03 crc kubenswrapper[4971]: W1213 06:49:03.552237 4971 feature_gate.go:330] unrecognized feature gate: GCPLabelsTags Dec 13 06:49:03 crc kubenswrapper[4971]: W1213 06:49:03.552241 4971 feature_gate.go:330] unrecognized feature gate: CSIDriverSharedResource Dec 13 06:49:03 crc kubenswrapper[4971]: W1213 06:49:03.552244 4971 feature_gate.go:330] unrecognized feature gate: MinimumKubeletVersion Dec 13 06:49:03 crc kubenswrapper[4971]: W1213 06:49:03.552251 4971 feature_gate.go:330] unrecognized feature gate: MachineAPIProviderOpenStack Dec 13 06:49:03 crc kubenswrapper[4971]: W1213 06:49:03.552255 4971 feature_gate.go:330] unrecognized feature gate: AutomatedEtcdBackup Dec 13 06:49:03 crc kubenswrapper[4971]: W1213 06:49:03.552258 4971 feature_gate.go:330] unrecognized feature gate: ChunkSizeMiB Dec 13 06:49:03 crc kubenswrapper[4971]: W1213 06:49:03.552262 4971 feature_gate.go:330] unrecognized feature gate: NewOLM Dec 13 06:49:03 crc kubenswrapper[4971]: W1213 06:49:03.552266 4971 feature_gate.go:330] unrecognized feature gate: GCPClusterHostedDNS Dec 13 06:49:03 crc kubenswrapper[4971]: W1213 06:49:03.552269 4971 feature_gate.go:330] unrecognized feature gate: VSphereStaticIPs Dec 13 06:49:03 crc kubenswrapper[4971]: W1213 06:49:03.552273 4971 feature_gate.go:330] unrecognized feature gate: RouteAdvertisements Dec 13 06:49:03 crc kubenswrapper[4971]: W1213 06:49:03.552277 4971 feature_gate.go:330] unrecognized feature gate: SignatureStores Dec 13 06:49:03 crc kubenswrapper[4971]: I1213 06:49:03.552283 4971 feature_gate.go:386] feature gates: {map[CloudDualStackNodeIPs:true DisableKubeletCloudCredentialProviders:true DynamicResourceAllocation:false EventedPLEG:false KMSv1:true MaxUnavailableStatefulSet:false NodeSwap:false ProcMountType:false RouteExternalCertificate:false ServiceAccountTokenNodeBinding:false TranslateStreamCloseWebsocketRequests:false UserNamespacesPodSecurityStandards:false UserNamespacesSupport:false ValidatingAdmissionPolicy:true VolumeAttributesClass:false]} Dec 13 06:49:03 crc kubenswrapper[4971]: I1213 06:49:03.563801 4971 server.go:491] "Kubelet version" kubeletVersion="v1.31.5" Dec 13 06:49:03 crc kubenswrapper[4971]: I1213 06:49:03.563883 4971 server.go:493] "Golang settings" GOGC="" GOMAXPROCS="" GOTRACEBACK="" Dec 13 06:49:03 crc kubenswrapper[4971]: W1213 06:49:03.564116 4971 feature_gate.go:330] unrecognized feature gate: NodeDisruptionPolicy Dec 13 06:49:03 crc kubenswrapper[4971]: W1213 06:49:03.564145 4971 feature_gate.go:330] unrecognized feature gate: AWSClusterHostedDNS Dec 13 06:49:03 crc kubenswrapper[4971]: W1213 06:49:03.564157 4971 feature_gate.go:330] unrecognized feature gate: SetEIPForNLBIngressController Dec 13 06:49:03 crc kubenswrapper[4971]: W1213 06:49:03.564212 4971 feature_gate.go:330] unrecognized feature gate: ChunkSizeMiB Dec 13 06:49:03 crc kubenswrapper[4971]: W1213 06:49:03.564226 4971 feature_gate.go:330] unrecognized feature gate: AWSEFSDriverVolumeMetrics Dec 13 06:49:03 crc kubenswrapper[4971]: W1213 06:49:03.564237 4971 feature_gate.go:330] unrecognized feature gate: MixedCPUsAllocation Dec 13 06:49:03 crc kubenswrapper[4971]: W1213 06:49:03.564247 4971 feature_gate.go:330] unrecognized feature gate: InsightsRuntimeExtractor Dec 13 06:49:03 crc kubenswrapper[4971]: W1213 06:49:03.564260 4971 feature_gate.go:330] unrecognized feature gate: VSphereStaticIPs Dec 13 06:49:03 crc kubenswrapper[4971]: W1213 06:49:03.564271 4971 feature_gate.go:330] unrecognized feature gate: BuildCSIVolumes Dec 13 06:49:03 crc kubenswrapper[4971]: W1213 06:49:03.564282 4971 feature_gate.go:330] unrecognized feature gate: BareMetalLoadBalancer Dec 13 06:49:03 crc kubenswrapper[4971]: W1213 06:49:03.564292 4971 feature_gate.go:330] unrecognized feature gate: VSphereMultiNetworks Dec 13 06:49:03 crc kubenswrapper[4971]: W1213 06:49:03.564303 4971 feature_gate.go:330] unrecognized feature gate: ClusterAPIInstallIBMCloud Dec 13 06:49:03 crc kubenswrapper[4971]: W1213 06:49:03.564313 4971 feature_gate.go:330] unrecognized feature gate: SignatureStores Dec 13 06:49:03 crc kubenswrapper[4971]: W1213 06:49:03.564323 4971 feature_gate.go:330] unrecognized feature gate: MachineAPIProviderOpenStack Dec 13 06:49:03 crc kubenswrapper[4971]: W1213 06:49:03.564333 4971 feature_gate.go:330] unrecognized feature gate: GatewayAPI Dec 13 06:49:03 crc kubenswrapper[4971]: W1213 06:49:03.564343 4971 feature_gate.go:330] unrecognized feature gate: HardwareSpeed Dec 13 06:49:03 crc kubenswrapper[4971]: W1213 06:49:03.564355 4971 feature_gate.go:330] unrecognized feature gate: RouteAdvertisements Dec 13 06:49:03 crc kubenswrapper[4971]: W1213 06:49:03.564364 4971 feature_gate.go:330] unrecognized feature gate: PlatformOperators Dec 13 06:49:03 crc kubenswrapper[4971]: W1213 06:49:03.564374 4971 feature_gate.go:330] unrecognized feature gate: MachineAPIOperatorDisableMachineHealthCheckController Dec 13 06:49:03 crc kubenswrapper[4971]: W1213 06:49:03.564385 4971 feature_gate.go:330] unrecognized feature gate: IngressControllerDynamicConfigurationManager Dec 13 06:49:03 crc kubenswrapper[4971]: W1213 06:49:03.564395 4971 feature_gate.go:330] unrecognized feature gate: MachineAPIMigration Dec 13 06:49:03 crc kubenswrapper[4971]: W1213 06:49:03.564405 4971 feature_gate.go:330] unrecognized feature gate: EtcdBackendQuota Dec 13 06:49:03 crc kubenswrapper[4971]: W1213 06:49:03.564454 4971 feature_gate.go:330] unrecognized feature gate: VSphereDriverConfiguration Dec 13 06:49:03 crc kubenswrapper[4971]: W1213 06:49:03.564464 4971 feature_gate.go:330] unrecognized feature gate: VSphereMultiVCenters Dec 13 06:49:03 crc kubenswrapper[4971]: W1213 06:49:03.564474 4971 feature_gate.go:330] unrecognized feature gate: OVNObservability Dec 13 06:49:03 crc kubenswrapper[4971]: W1213 06:49:03.564484 4971 feature_gate.go:330] unrecognized feature gate: OnClusterBuild Dec 13 06:49:03 crc kubenswrapper[4971]: W1213 06:49:03.564494 4971 feature_gate.go:330] unrecognized feature gate: VolumeGroupSnapshot Dec 13 06:49:03 crc kubenswrapper[4971]: W1213 06:49:03.564504 4971 feature_gate.go:330] unrecognized feature gate: NetworkLiveMigration Dec 13 06:49:03 crc kubenswrapper[4971]: W1213 06:49:03.564540 4971 feature_gate.go:330] unrecognized feature gate: ClusterAPIInstall Dec 13 06:49:03 crc kubenswrapper[4971]: W1213 06:49:03.564552 4971 feature_gate.go:330] unrecognized feature gate: AutomatedEtcdBackup Dec 13 06:49:03 crc kubenswrapper[4971]: W1213 06:49:03.564562 4971 feature_gate.go:330] unrecognized feature gate: ClusterMonitoringConfig Dec 13 06:49:03 crc kubenswrapper[4971]: W1213 06:49:03.564572 4971 feature_gate.go:330] unrecognized feature gate: InsightsConfigAPI Dec 13 06:49:03 crc kubenswrapper[4971]: W1213 06:49:03.564584 4971 feature_gate.go:330] unrecognized feature gate: OpenShiftPodSecurityAdmission Dec 13 06:49:03 crc kubenswrapper[4971]: W1213 06:49:03.564593 4971 feature_gate.go:330] unrecognized feature gate: NetworkDiagnosticsConfig Dec 13 06:49:03 crc kubenswrapper[4971]: W1213 06:49:03.564610 4971 feature_gate.go:353] Setting GA feature gate DisableKubeletCloudCredentialProviders=true. It will be removed in a future release. Dec 13 06:49:03 crc kubenswrapper[4971]: W1213 06:49:03.564626 4971 feature_gate.go:330] unrecognized feature gate: NewOLM Dec 13 06:49:03 crc kubenswrapper[4971]: W1213 06:49:03.564637 4971 feature_gate.go:330] unrecognized feature gate: InsightsConfig Dec 13 06:49:03 crc kubenswrapper[4971]: W1213 06:49:03.564649 4971 feature_gate.go:330] unrecognized feature gate: PinnedImages Dec 13 06:49:03 crc kubenswrapper[4971]: W1213 06:49:03.564661 4971 feature_gate.go:330] unrecognized feature gate: VSphereControlPlaneMachineSet Dec 13 06:49:03 crc kubenswrapper[4971]: W1213 06:49:03.564673 4971 feature_gate.go:330] unrecognized feature gate: AdminNetworkPolicy Dec 13 06:49:03 crc kubenswrapper[4971]: W1213 06:49:03.564686 4971 feature_gate.go:353] Setting GA feature gate ValidatingAdmissionPolicy=true. It will be removed in a future release. Dec 13 06:49:03 crc kubenswrapper[4971]: W1213 06:49:03.564700 4971 feature_gate.go:330] unrecognized feature gate: IngressControllerLBSubnetsAWS Dec 13 06:49:03 crc kubenswrapper[4971]: W1213 06:49:03.564711 4971 feature_gate.go:330] unrecognized feature gate: NetworkSegmentation Dec 13 06:49:03 crc kubenswrapper[4971]: W1213 06:49:03.564721 4971 feature_gate.go:330] unrecognized feature gate: MetricsCollectionProfiles Dec 13 06:49:03 crc kubenswrapper[4971]: W1213 06:49:03.564733 4971 feature_gate.go:330] unrecognized feature gate: PrivateHostedZoneAWS Dec 13 06:49:03 crc kubenswrapper[4971]: W1213 06:49:03.564746 4971 feature_gate.go:353] Setting GA feature gate CloudDualStackNodeIPs=true. It will be removed in a future release. Dec 13 06:49:03 crc kubenswrapper[4971]: W1213 06:49:03.564759 4971 feature_gate.go:330] unrecognized feature gate: ImageStreamImportMode Dec 13 06:49:03 crc kubenswrapper[4971]: W1213 06:49:03.564770 4971 feature_gate.go:330] unrecognized feature gate: DNSNameResolver Dec 13 06:49:03 crc kubenswrapper[4971]: W1213 06:49:03.564781 4971 feature_gate.go:330] unrecognized feature gate: MultiArchInstallGCP Dec 13 06:49:03 crc kubenswrapper[4971]: W1213 06:49:03.564793 4971 feature_gate.go:330] unrecognized feature gate: AlibabaPlatform Dec 13 06:49:03 crc kubenswrapper[4971]: W1213 06:49:03.564804 4971 feature_gate.go:330] unrecognized feature gate: UpgradeStatus Dec 13 06:49:03 crc kubenswrapper[4971]: W1213 06:49:03.564859 4971 feature_gate.go:351] Setting deprecated feature gate KMSv1=true. It will be removed in a future release. Dec 13 06:49:03 crc kubenswrapper[4971]: W1213 06:49:03.564872 4971 feature_gate.go:330] unrecognized feature gate: ManagedBootImagesAWS Dec 13 06:49:03 crc kubenswrapper[4971]: W1213 06:49:03.564883 4971 feature_gate.go:330] unrecognized feature gate: MultiArchInstallAzure Dec 13 06:49:03 crc kubenswrapper[4971]: W1213 06:49:03.564894 4971 feature_gate.go:330] unrecognized feature gate: MinimumKubeletVersion Dec 13 06:49:03 crc kubenswrapper[4971]: W1213 06:49:03.564906 4971 feature_gate.go:330] unrecognized feature gate: ManagedBootImages Dec 13 06:49:03 crc kubenswrapper[4971]: W1213 06:49:03.564916 4971 feature_gate.go:330] unrecognized feature gate: Example Dec 13 06:49:03 crc kubenswrapper[4971]: W1213 06:49:03.564926 4971 feature_gate.go:330] unrecognized feature gate: PersistentIPsForVirtualization Dec 13 06:49:03 crc kubenswrapper[4971]: W1213 06:49:03.564936 4971 feature_gate.go:330] unrecognized feature gate: SigstoreImageVerification Dec 13 06:49:03 crc kubenswrapper[4971]: W1213 06:49:03.564946 4971 feature_gate.go:330] unrecognized feature gate: GCPLabelsTags Dec 13 06:49:03 crc kubenswrapper[4971]: W1213 06:49:03.564957 4971 feature_gate.go:330] unrecognized feature gate: AdditionalRoutingCapabilities Dec 13 06:49:03 crc kubenswrapper[4971]: W1213 06:49:03.564967 4971 feature_gate.go:330] unrecognized feature gate: GCPClusterHostedDNS Dec 13 06:49:03 crc kubenswrapper[4971]: W1213 06:49:03.564977 4971 feature_gate.go:330] unrecognized feature gate: ExternalOIDC Dec 13 06:49:03 crc kubenswrapper[4971]: W1213 06:49:03.564987 4971 feature_gate.go:330] unrecognized feature gate: BootcNodeManagement Dec 13 06:49:03 crc kubenswrapper[4971]: W1213 06:49:03.564997 4971 feature_gate.go:330] unrecognized feature gate: MultiArchInstallAWS Dec 13 06:49:03 crc kubenswrapper[4971]: W1213 06:49:03.565007 4971 feature_gate.go:330] unrecognized feature gate: MachineConfigNodes Dec 13 06:49:03 crc kubenswrapper[4971]: W1213 06:49:03.565017 4971 feature_gate.go:330] unrecognized feature gate: CSIDriverSharedResource Dec 13 06:49:03 crc kubenswrapper[4971]: W1213 06:49:03.565026 4971 feature_gate.go:330] unrecognized feature gate: AzureWorkloadIdentity Dec 13 06:49:03 crc kubenswrapper[4971]: W1213 06:49:03.565036 4971 feature_gate.go:330] unrecognized feature gate: NutanixMultiSubnets Dec 13 06:49:03 crc kubenswrapper[4971]: W1213 06:49:03.565046 4971 feature_gate.go:330] unrecognized feature gate: InsightsOnDemandDataGather Dec 13 06:49:03 crc kubenswrapper[4971]: W1213 06:49:03.565059 4971 feature_gate.go:330] unrecognized feature gate: ConsolePluginContentSecurityPolicy Dec 13 06:49:03 crc kubenswrapper[4971]: I1213 06:49:03.565077 4971 feature_gate.go:386] feature gates: {map[CloudDualStackNodeIPs:true DisableKubeletCloudCredentialProviders:true DynamicResourceAllocation:false EventedPLEG:false KMSv1:true MaxUnavailableStatefulSet:false NodeSwap:false ProcMountType:false RouteExternalCertificate:false ServiceAccountTokenNodeBinding:false TranslateStreamCloseWebsocketRequests:false UserNamespacesPodSecurityStandards:false UserNamespacesSupport:false ValidatingAdmissionPolicy:true VolumeAttributesClass:false]} Dec 13 06:49:03 crc kubenswrapper[4971]: W1213 06:49:03.565377 4971 feature_gate.go:330] unrecognized feature gate: NetworkDiagnosticsConfig Dec 13 06:49:03 crc kubenswrapper[4971]: W1213 06:49:03.565395 4971 feature_gate.go:330] unrecognized feature gate: MachineAPIMigration Dec 13 06:49:03 crc kubenswrapper[4971]: W1213 06:49:03.565406 4971 feature_gate.go:330] unrecognized feature gate: ManagedBootImagesAWS Dec 13 06:49:03 crc kubenswrapper[4971]: W1213 06:49:03.565417 4971 feature_gate.go:330] unrecognized feature gate: MinimumKubeletVersion Dec 13 06:49:03 crc kubenswrapper[4971]: W1213 06:49:03.565431 4971 feature_gate.go:351] Setting deprecated feature gate KMSv1=true. It will be removed in a future release. Dec 13 06:49:03 crc kubenswrapper[4971]: W1213 06:49:03.565444 4971 feature_gate.go:330] unrecognized feature gate: EtcdBackendQuota Dec 13 06:49:03 crc kubenswrapper[4971]: W1213 06:49:03.565454 4971 feature_gate.go:330] unrecognized feature gate: GatewayAPI Dec 13 06:49:03 crc kubenswrapper[4971]: W1213 06:49:03.565465 4971 feature_gate.go:330] unrecognized feature gate: MultiArchInstallAzure Dec 13 06:49:03 crc kubenswrapper[4971]: W1213 06:49:03.565476 4971 feature_gate.go:330] unrecognized feature gate: AWSEFSDriverVolumeMetrics Dec 13 06:49:03 crc kubenswrapper[4971]: W1213 06:49:03.565487 4971 feature_gate.go:330] unrecognized feature gate: NetworkLiveMigration Dec 13 06:49:03 crc kubenswrapper[4971]: W1213 06:49:03.565497 4971 feature_gate.go:330] unrecognized feature gate: ConsolePluginContentSecurityPolicy Dec 13 06:49:03 crc kubenswrapper[4971]: W1213 06:49:03.565507 4971 feature_gate.go:330] unrecognized feature gate: AlibabaPlatform Dec 13 06:49:03 crc kubenswrapper[4971]: W1213 06:49:03.565576 4971 feature_gate.go:330] unrecognized feature gate: ManagedBootImages Dec 13 06:49:03 crc kubenswrapper[4971]: W1213 06:49:03.565588 4971 feature_gate.go:330] unrecognized feature gate: PlatformOperators Dec 13 06:49:03 crc kubenswrapper[4971]: W1213 06:49:03.565599 4971 feature_gate.go:330] unrecognized feature gate: IngressControllerDynamicConfigurationManager Dec 13 06:49:03 crc kubenswrapper[4971]: W1213 06:49:03.565608 4971 feature_gate.go:330] unrecognized feature gate: SigstoreImageVerification Dec 13 06:49:03 crc kubenswrapper[4971]: W1213 06:49:03.565618 4971 feature_gate.go:330] unrecognized feature gate: VSphereMultiVCenters Dec 13 06:49:03 crc kubenswrapper[4971]: W1213 06:49:03.565629 4971 feature_gate.go:330] unrecognized feature gate: MachineAPIProviderOpenStack Dec 13 06:49:03 crc kubenswrapper[4971]: W1213 06:49:03.565638 4971 feature_gate.go:330] unrecognized feature gate: AzureWorkloadIdentity Dec 13 06:49:03 crc kubenswrapper[4971]: W1213 06:49:03.565648 4971 feature_gate.go:330] unrecognized feature gate: RouteAdvertisements Dec 13 06:49:03 crc kubenswrapper[4971]: W1213 06:49:03.565657 4971 feature_gate.go:330] unrecognized feature gate: GCPClusterHostedDNS Dec 13 06:49:03 crc kubenswrapper[4971]: W1213 06:49:03.565667 4971 feature_gate.go:330] unrecognized feature gate: NutanixMultiSubnets Dec 13 06:49:03 crc kubenswrapper[4971]: W1213 06:49:03.565676 4971 feature_gate.go:330] unrecognized feature gate: MultiArchInstallAWS Dec 13 06:49:03 crc kubenswrapper[4971]: W1213 06:49:03.565685 4971 feature_gate.go:330] unrecognized feature gate: DNSNameResolver Dec 13 06:49:03 crc kubenswrapper[4971]: W1213 06:49:03.565698 4971 feature_gate.go:353] Setting GA feature gate DisableKubeletCloudCredentialProviders=true. It will be removed in a future release. Dec 13 06:49:03 crc kubenswrapper[4971]: W1213 06:49:03.565712 4971 feature_gate.go:330] unrecognized feature gate: VSphereMultiNetworks Dec 13 06:49:03 crc kubenswrapper[4971]: W1213 06:49:03.565723 4971 feature_gate.go:330] unrecognized feature gate: SetEIPForNLBIngressController Dec 13 06:49:03 crc kubenswrapper[4971]: W1213 06:49:03.565732 4971 feature_gate.go:330] unrecognized feature gate: ClusterMonitoringConfig Dec 13 06:49:03 crc kubenswrapper[4971]: W1213 06:49:03.565742 4971 feature_gate.go:330] unrecognized feature gate: BootcNodeManagement Dec 13 06:49:03 crc kubenswrapper[4971]: W1213 06:49:03.565751 4971 feature_gate.go:330] unrecognized feature gate: BareMetalLoadBalancer Dec 13 06:49:03 crc kubenswrapper[4971]: W1213 06:49:03.565764 4971 feature_gate.go:353] Setting GA feature gate ValidatingAdmissionPolicy=true. It will be removed in a future release. Dec 13 06:49:03 crc kubenswrapper[4971]: W1213 06:49:03.565775 4971 feature_gate.go:330] unrecognized feature gate: PrivateHostedZoneAWS Dec 13 06:49:03 crc kubenswrapper[4971]: W1213 06:49:03.565786 4971 feature_gate.go:330] unrecognized feature gate: VSphereControlPlaneMachineSet Dec 13 06:49:03 crc kubenswrapper[4971]: W1213 06:49:03.565799 4971 feature_gate.go:330] unrecognized feature gate: ImageStreamImportMode Dec 13 06:49:03 crc kubenswrapper[4971]: W1213 06:49:03.565811 4971 feature_gate.go:330] unrecognized feature gate: NodeDisruptionPolicy Dec 13 06:49:03 crc kubenswrapper[4971]: W1213 06:49:03.565823 4971 feature_gate.go:330] unrecognized feature gate: InsightsOnDemandDataGather Dec 13 06:49:03 crc kubenswrapper[4971]: W1213 06:49:03.565834 4971 feature_gate.go:330] unrecognized feature gate: MachineConfigNodes Dec 13 06:49:03 crc kubenswrapper[4971]: W1213 06:49:03.565848 4971 feature_gate.go:353] Setting GA feature gate CloudDualStackNodeIPs=true. It will be removed in a future release. Dec 13 06:49:03 crc kubenswrapper[4971]: W1213 06:49:03.565861 4971 feature_gate.go:330] unrecognized feature gate: Example Dec 13 06:49:03 crc kubenswrapper[4971]: W1213 06:49:03.565873 4971 feature_gate.go:330] unrecognized feature gate: PersistentIPsForVirtualization Dec 13 06:49:03 crc kubenswrapper[4971]: W1213 06:49:03.565884 4971 feature_gate.go:330] unrecognized feature gate: SignatureStores Dec 13 06:49:03 crc kubenswrapper[4971]: W1213 06:49:03.565894 4971 feature_gate.go:330] unrecognized feature gate: PinnedImages Dec 13 06:49:03 crc kubenswrapper[4971]: W1213 06:49:03.565904 4971 feature_gate.go:330] unrecognized feature gate: MixedCPUsAllocation Dec 13 06:49:03 crc kubenswrapper[4971]: W1213 06:49:03.565914 4971 feature_gate.go:330] unrecognized feature gate: AutomatedEtcdBackup Dec 13 06:49:03 crc kubenswrapper[4971]: W1213 06:49:03.565924 4971 feature_gate.go:330] unrecognized feature gate: VolumeGroupSnapshot Dec 13 06:49:03 crc kubenswrapper[4971]: W1213 06:49:03.565935 4971 feature_gate.go:330] unrecognized feature gate: VSphereStaticIPs Dec 13 06:49:03 crc kubenswrapper[4971]: W1213 06:49:03.565945 4971 feature_gate.go:330] unrecognized feature gate: MetricsCollectionProfiles Dec 13 06:49:03 crc kubenswrapper[4971]: W1213 06:49:03.565955 4971 feature_gate.go:330] unrecognized feature gate: MachineAPIOperatorDisableMachineHealthCheckController Dec 13 06:49:03 crc kubenswrapper[4971]: W1213 06:49:03.565965 4971 feature_gate.go:330] unrecognized feature gate: InsightsConfigAPI Dec 13 06:49:03 crc kubenswrapper[4971]: W1213 06:49:03.565975 4971 feature_gate.go:330] unrecognized feature gate: ChunkSizeMiB Dec 13 06:49:03 crc kubenswrapper[4971]: W1213 06:49:03.565985 4971 feature_gate.go:330] unrecognized feature gate: OnClusterBuild Dec 13 06:49:03 crc kubenswrapper[4971]: W1213 06:49:03.565995 4971 feature_gate.go:330] unrecognized feature gate: HardwareSpeed Dec 13 06:49:03 crc kubenswrapper[4971]: W1213 06:49:03.566006 4971 feature_gate.go:330] unrecognized feature gate: NetworkSegmentation Dec 13 06:49:03 crc kubenswrapper[4971]: W1213 06:49:03.566016 4971 feature_gate.go:330] unrecognized feature gate: IngressControllerLBSubnetsAWS Dec 13 06:49:03 crc kubenswrapper[4971]: W1213 06:49:03.566025 4971 feature_gate.go:330] unrecognized feature gate: GCPLabelsTags Dec 13 06:49:03 crc kubenswrapper[4971]: W1213 06:49:03.566035 4971 feature_gate.go:330] unrecognized feature gate: UpgradeStatus Dec 13 06:49:03 crc kubenswrapper[4971]: W1213 06:49:03.566045 4971 feature_gate.go:330] unrecognized feature gate: ClusterAPIInstall Dec 13 06:49:03 crc kubenswrapper[4971]: W1213 06:49:03.566055 4971 feature_gate.go:330] unrecognized feature gate: CSIDriverSharedResource Dec 13 06:49:03 crc kubenswrapper[4971]: W1213 06:49:03.566065 4971 feature_gate.go:330] unrecognized feature gate: AWSClusterHostedDNS Dec 13 06:49:03 crc kubenswrapper[4971]: W1213 06:49:03.566075 4971 feature_gate.go:330] unrecognized feature gate: InsightsRuntimeExtractor Dec 13 06:49:03 crc kubenswrapper[4971]: W1213 06:49:03.566085 4971 feature_gate.go:330] unrecognized feature gate: AdditionalRoutingCapabilities Dec 13 06:49:03 crc kubenswrapper[4971]: W1213 06:49:03.566095 4971 feature_gate.go:330] unrecognized feature gate: VSphereDriverConfiguration Dec 13 06:49:03 crc kubenswrapper[4971]: W1213 06:49:03.566105 4971 feature_gate.go:330] unrecognized feature gate: MultiArchInstallGCP Dec 13 06:49:03 crc kubenswrapper[4971]: W1213 06:49:03.566115 4971 feature_gate.go:330] unrecognized feature gate: AdminNetworkPolicy Dec 13 06:49:03 crc kubenswrapper[4971]: W1213 06:49:03.566126 4971 feature_gate.go:330] unrecognized feature gate: OVNObservability Dec 13 06:49:03 crc kubenswrapper[4971]: W1213 06:49:03.566136 4971 feature_gate.go:330] unrecognized feature gate: ExternalOIDC Dec 13 06:49:03 crc kubenswrapper[4971]: W1213 06:49:03.566146 4971 feature_gate.go:330] unrecognized feature gate: BuildCSIVolumes Dec 13 06:49:03 crc kubenswrapper[4971]: W1213 06:49:03.566156 4971 feature_gate.go:330] unrecognized feature gate: NewOLM Dec 13 06:49:03 crc kubenswrapper[4971]: W1213 06:49:03.566166 4971 feature_gate.go:330] unrecognized feature gate: ClusterAPIInstallIBMCloud Dec 13 06:49:03 crc kubenswrapper[4971]: W1213 06:49:03.566176 4971 feature_gate.go:330] unrecognized feature gate: OpenShiftPodSecurityAdmission Dec 13 06:49:03 crc kubenswrapper[4971]: W1213 06:49:03.566187 4971 feature_gate.go:330] unrecognized feature gate: InsightsConfig Dec 13 06:49:03 crc kubenswrapper[4971]: I1213 06:49:03.566203 4971 feature_gate.go:386] feature gates: {map[CloudDualStackNodeIPs:true DisableKubeletCloudCredentialProviders:true DynamicResourceAllocation:false EventedPLEG:false KMSv1:true MaxUnavailableStatefulSet:false NodeSwap:false ProcMountType:false RouteExternalCertificate:false ServiceAccountTokenNodeBinding:false TranslateStreamCloseWebsocketRequests:false UserNamespacesPodSecurityStandards:false UserNamespacesSupport:false ValidatingAdmissionPolicy:true VolumeAttributesClass:false]} Dec 13 06:49:03 crc kubenswrapper[4971]: I1213 06:49:03.566764 4971 server.go:940] "Client rotation is on, will bootstrap in background" Dec 13 06:49:03 crc kubenswrapper[4971]: I1213 06:49:03.572086 4971 bootstrap.go:85] "Current kubeconfig file contents are still valid, no bootstrap necessary" Dec 13 06:49:03 crc kubenswrapper[4971]: I1213 06:49:03.572252 4971 certificate_store.go:130] Loading cert/key pair from "/var/lib/kubelet/pki/kubelet-client-current.pem". Dec 13 06:49:03 crc kubenswrapper[4971]: I1213 06:49:03.573243 4971 server.go:997] "Starting client certificate rotation" Dec 13 06:49:03 crc kubenswrapper[4971]: I1213 06:49:03.573289 4971 certificate_manager.go:356] kubernetes.io/kube-apiserver-client-kubelet: Certificate rotation is enabled Dec 13 06:49:03 crc kubenswrapper[4971]: I1213 06:49:03.573536 4971 certificate_manager.go:356] kubernetes.io/kube-apiserver-client-kubelet: Certificate expiration is 2026-02-24 05:52:08 +0000 UTC, rotation deadline is 2025-12-23 07:00:53.932454289 +0000 UTC Dec 13 06:49:03 crc kubenswrapper[4971]: I1213 06:49:03.573641 4971 certificate_manager.go:356] kubernetes.io/kube-apiserver-client-kubelet: Waiting 240h11m50.358818248s for next certificate rotation Dec 13 06:49:03 crc kubenswrapper[4971]: I1213 06:49:03.585241 4971 dynamic_cafile_content.go:123] "Loaded a new CA Bundle and Verifier" name="client-ca-bundle::/etc/kubernetes/kubelet-ca.crt" Dec 13 06:49:03 crc kubenswrapper[4971]: I1213 06:49:03.589276 4971 dynamic_cafile_content.go:161] "Starting controller" name="client-ca-bundle::/etc/kubernetes/kubelet-ca.crt" Dec 13 06:49:03 crc kubenswrapper[4971]: I1213 06:49:03.600872 4971 log.go:25] "Validated CRI v1 runtime API" Dec 13 06:49:03 crc kubenswrapper[4971]: I1213 06:49:03.637352 4971 log.go:25] "Validated CRI v1 image API" Dec 13 06:49:03 crc kubenswrapper[4971]: I1213 06:49:03.639003 4971 server.go:1437] "Using cgroup driver setting received from the CRI runtime" cgroupDriver="systemd" Dec 13 06:49:03 crc kubenswrapper[4971]: I1213 06:49:03.642267 4971 fs.go:133] Filesystem UUIDs: map[0b076daa-c26a-46d2-b3a6-72a8dbc6e257:/dev/vda4 2025-12-13-06-43-44-00:/dev/sr0 7B77-95E7:/dev/vda2 de0497b0-db1b-465a-b278-03db02455c71:/dev/vda3] Dec 13 06:49:03 crc kubenswrapper[4971]: I1213 06:49:03.642304 4971 fs.go:134] Filesystem partitions: map[/dev/shm:{mountpoint:/dev/shm major:0 minor:22 fsType:tmpfs blockSize:0} /dev/vda3:{mountpoint:/boot major:252 minor:3 fsType:ext4 blockSize:0} /dev/vda4:{mountpoint:/var major:252 minor:4 fsType:xfs blockSize:0} /run:{mountpoint:/run major:0 minor:24 fsType:tmpfs blockSize:0} /run/user/1000:{mountpoint:/run/user/1000 major:0 minor:42 fsType:tmpfs blockSize:0} /tmp:{mountpoint:/tmp major:0 minor:30 fsType:tmpfs blockSize:0} /var/lib/etcd:{mountpoint:/var/lib/etcd major:0 minor:41 fsType:tmpfs blockSize:0}] Dec 13 06:49:03 crc kubenswrapper[4971]: I1213 06:49:03.654838 4971 manager.go:217] Machine: {Timestamp:2025-12-13 06:49:03.653760064 +0000 UTC m=+0.258169532 CPUVendorID:AuthenticAMD NumCores:12 NumPhysicalCores:1 NumSockets:12 CpuFrequency:2799998 MemoryCapacity:33654128640 SwapCapacity:0 MemoryByType:map[] NVMInfo:{MemoryModeCapacity:0 AppDirectModeCapacity:0 AvgPowerBudget:0} HugePages:[{PageSize:1048576 NumPages:0} {PageSize:2048 NumPages:0}] MachineID:21801e6708c44f15b81395eb736a7cec SystemUUID:a9af00dc-cf25-47af-9997-79577f8f3bea BootID:bd36cbc9-bbe9-45a1-a21a-ad2bd10499ee Filesystems:[{Device:/dev/shm DeviceMajor:0 DeviceMinor:22 Capacity:16827064320 Type:vfs Inodes:4108170 HasInodes:true} {Device:/run DeviceMajor:0 DeviceMinor:24 Capacity:6730825728 Type:vfs Inodes:819200 HasInodes:true} {Device:/dev/vda4 DeviceMajor:252 DeviceMinor:4 Capacity:85292941312 Type:vfs Inodes:41679680 HasInodes:true} {Device:/tmp DeviceMajor:0 DeviceMinor:30 Capacity:16827064320 Type:vfs Inodes:1048576 HasInodes:true} {Device:/dev/vda3 DeviceMajor:252 DeviceMinor:3 Capacity:366869504 Type:vfs Inodes:98304 HasInodes:true} {Device:/run/user/1000 DeviceMajor:0 DeviceMinor:42 Capacity:3365412864 Type:vfs Inodes:821634 HasInodes:true} {Device:/var/lib/etcd DeviceMajor:0 DeviceMinor:41 Capacity:1073741824 Type:vfs Inodes:4108170 HasInodes:true}] DiskMap:map[252:0:{Name:vda Major:252 Minor:0 Size:214748364800 Scheduler:none}] NetworkDevices:[{Name:br-ex MacAddress:fa:16:3e:a8:2d:dc Speed:0 Mtu:1500} {Name:br-int MacAddress:d6:39:55:2e:22:71 Speed:0 Mtu:1400} {Name:ens3 MacAddress:fa:16:3e:a8:2d:dc Speed:-1 Mtu:1500} {Name:ens7 MacAddress:fa:16:3e:14:ed:43 Speed:-1 Mtu:1500} {Name:ens7.20 MacAddress:52:54:00:df:63:94 Speed:-1 Mtu:1496} {Name:ens7.21 MacAddress:52:54:00:36:01:c5 Speed:-1 Mtu:1496} {Name:ens7.22 MacAddress:52:54:00:3f:79:a3 Speed:-1 Mtu:1496} {Name:ens7.23 MacAddress:52:54:00:a8:9a:e9 Speed:-1 Mtu:1496} {Name:eth10 MacAddress:8a:69:85:8e:a9:c2 Speed:0 Mtu:1500} {Name:ovn-k8s-mp0 MacAddress:0a:58:0a:d9:00:02 Speed:0 Mtu:1400} {Name:ovs-system MacAddress:aa:31:ab:0c:84:47 Speed:0 Mtu:1500}] Topology:[{Id:0 Memory:33654128640 HugePages:[{PageSize:1048576 NumPages:0} {PageSize:2048 NumPages:0}] Cores:[{Id:0 Threads:[0] Caches:[{Id:0 Size:32768 Type:Data Level:1} {Id:0 Size:32768 Type:Instruction Level:1} {Id:0 Size:524288 Type:Unified Level:2}] UncoreCaches:[{Id:0 Size:16777216 Type:Unified Level:3}] SocketID:0 BookID: DrawerID:} {Id:0 Threads:[1] Caches:[{Id:1 Size:32768 Type:Data Level:1} {Id:1 Size:32768 Type:Instruction Level:1} {Id:1 Size:524288 Type:Unified Level:2}] UncoreCaches:[{Id:1 Size:16777216 Type:Unified Level:3}] SocketID:1 BookID: DrawerID:} {Id:0 Threads:[10] Caches:[{Id:10 Size:32768 Type:Data Level:1} {Id:10 Size:32768 Type:Instruction Level:1} {Id:10 Size:524288 Type:Unified Level:2}] UncoreCaches:[{Id:10 Size:16777216 Type:Unified Level:3}] SocketID:10 BookID: DrawerID:} {Id:0 Threads:[11] Caches:[{Id:11 Size:32768 Type:Data Level:1} {Id:11 Size:32768 Type:Instruction Level:1} {Id:11 Size:524288 Type:Unified Level:2}] UncoreCaches:[{Id:11 Size:16777216 Type:Unified Level:3}] SocketID:11 BookID: DrawerID:} {Id:0 Threads:[2] Caches:[{Id:2 Size:32768 Type:Data Level:1} {Id:2 Size:32768 Type:Instruction Level:1} {Id:2 Size:524288 Type:Unified Level:2}] UncoreCaches:[{Id:2 Size:16777216 Type:Unified Level:3}] SocketID:2 BookID: DrawerID:} {Id:0 Threads:[3] Caches:[{Id:3 Size:32768 Type:Data Level:1} {Id:3 Size:32768 Type:Instruction Level:1} {Id:3 Size:524288 Type:Unified Level:2}] UncoreCaches:[{Id:3 Size:16777216 Type:Unified Level:3}] SocketID:3 BookID: DrawerID:} {Id:0 Threads:[4] Caches:[{Id:4 Size:32768 Type:Data Level:1} {Id:4 Size:32768 Type:Instruction Level:1} {Id:4 Size:524288 Type:Unified Level:2}] UncoreCaches:[{Id:4 Size:16777216 Type:Unified Level:3}] SocketID:4 BookID: DrawerID:} {Id:0 Threads:[5] Caches:[{Id:5 Size:32768 Type:Data Level:1} {Id:5 Size:32768 Type:Instruction Level:1} {Id:5 Size:524288 Type:Unified Level:2}] UncoreCaches:[{Id:5 Size:16777216 Type:Unified Level:3}] SocketID:5 BookID: DrawerID:} {Id:0 Threads:[6] Caches:[{Id:6 Size:32768 Type:Data Level:1} {Id:6 Size:32768 Type:Instruction Level:1} {Id:6 Size:524288 Type:Unified Level:2}] UncoreCaches:[{Id:6 Size:16777216 Type:Unified Level:3}] SocketID:6 BookID: DrawerID:} {Id:0 Threads:[7] Caches:[{Id:7 Size:32768 Type:Data Level:1} {Id:7 Size:32768 Type:Instruction Level:1} {Id:7 Size:524288 Type:Unified Level:2}] UncoreCaches:[{Id:7 Size:16777216 Type:Unified Level:3}] SocketID:7 BookID: DrawerID:} {Id:0 Threads:[8] Caches:[{Id:8 Size:32768 Type:Data Level:1} {Id:8 Size:32768 Type:Instruction Level:1} {Id:8 Size:524288 Type:Unified Level:2}] UncoreCaches:[{Id:8 Size:16777216 Type:Unified Level:3}] SocketID:8 BookID: DrawerID:} {Id:0 Threads:[9] Caches:[{Id:9 Size:32768 Type:Data Level:1} {Id:9 Size:32768 Type:Instruction Level:1} {Id:9 Size:524288 Type:Unified Level:2}] UncoreCaches:[{Id:9 Size:16777216 Type:Unified Level:3}] SocketID:9 BookID: DrawerID:}] Caches:[] Distances:[10]}] CloudProvider:Unknown InstanceType:Unknown InstanceID:None} Dec 13 06:49:03 crc kubenswrapper[4971]: I1213 06:49:03.655331 4971 manager_no_libpfm.go:29] cAdvisor is build without cgo and/or libpfm support. Perf event counters are not available. Dec 13 06:49:03 crc kubenswrapper[4971]: I1213 06:49:03.655496 4971 manager.go:233] Version: {KernelVersion:5.14.0-427.50.2.el9_4.x86_64 ContainerOsVersion:Red Hat Enterprise Linux CoreOS 418.94.202502100215-0 DockerVersion: DockerAPIVersion: CadvisorVersion: CadvisorRevision:} Dec 13 06:49:03 crc kubenswrapper[4971]: I1213 06:49:03.656471 4971 swap_util.go:113] "Swap is on" /proc/swaps contents="Filename\t\t\t\tType\t\tSize\t\tUsed\t\tPriority" Dec 13 06:49:03 crc kubenswrapper[4971]: I1213 06:49:03.657003 4971 container_manager_linux.go:267] "Container manager verified user specified cgroup-root exists" cgroupRoot=[] Dec 13 06:49:03 crc kubenswrapper[4971]: I1213 06:49:03.657557 4971 container_manager_linux.go:272] "Creating Container Manager object based on Node Config" nodeConfig={"NodeName":"crc","RuntimeCgroupsName":"/system.slice/crio.service","SystemCgroupsName":"/system.slice","KubeletCgroupsName":"","KubeletOOMScoreAdj":-999,"ContainerRuntime":"","CgroupsPerQOS":true,"CgroupRoot":"/","CgroupDriver":"systemd","KubeletRootDir":"/var/lib/kubelet","ProtectKernelDefaults":true,"KubeReservedCgroupName":"","SystemReservedCgroupName":"","ReservedSystemCPUs":{},"EnforceNodeAllocatable":{"pods":{}},"KubeReserved":null,"SystemReserved":{"cpu":"200m","ephemeral-storage":"350Mi","memory":"350Mi"},"HardEvictionThresholds":[{"Signal":"imagefs.available","Operator":"LessThan","Value":{"Quantity":null,"Percentage":0.15},"GracePeriod":0,"MinReclaim":null},{"Signal":"imagefs.inodesFree","Operator":"LessThan","Value":{"Quantity":null,"Percentage":0.05},"GracePeriod":0,"MinReclaim":null},{"Signal":"memory.available","Operator":"LessThan","Value":{"Quantity":"100Mi","Percentage":0},"GracePeriod":0,"MinReclaim":null},{"Signal":"nodefs.available","Operator":"LessThan","Value":{"Quantity":null,"Percentage":0.1},"GracePeriod":0,"MinReclaim":null},{"Signal":"nodefs.inodesFree","Operator":"LessThan","Value":{"Quantity":null,"Percentage":0.05},"GracePeriod":0,"MinReclaim":null}],"QOSReserved":{},"CPUManagerPolicy":"none","CPUManagerPolicyOptions":null,"TopologyManagerScope":"container","CPUManagerReconcilePeriod":10000000000,"ExperimentalMemoryManagerPolicy":"None","ExperimentalMemoryManagerReservedMemory":null,"PodPidsLimit":4096,"EnforceCPULimits":true,"CPUCFSQuotaPeriod":100000000,"TopologyManagerPolicy":"none","TopologyManagerPolicyOptions":null,"CgroupVersion":2} Dec 13 06:49:03 crc kubenswrapper[4971]: I1213 06:49:03.657843 4971 topology_manager.go:138] "Creating topology manager with none policy" Dec 13 06:49:03 crc kubenswrapper[4971]: I1213 06:49:03.657860 4971 container_manager_linux.go:303] "Creating device plugin manager" Dec 13 06:49:03 crc kubenswrapper[4971]: I1213 06:49:03.658073 4971 manager.go:142] "Creating Device Plugin manager" path="/var/lib/kubelet/device-plugins/kubelet.sock" Dec 13 06:49:03 crc kubenswrapper[4971]: I1213 06:49:03.658105 4971 server.go:66] "Creating device plugin registration server" version="v1beta1" socket="/var/lib/kubelet/device-plugins/kubelet.sock" Dec 13 06:49:03 crc kubenswrapper[4971]: I1213 06:49:03.658360 4971 state_mem.go:36] "Initialized new in-memory state store" Dec 13 06:49:03 crc kubenswrapper[4971]: I1213 06:49:03.658473 4971 server.go:1245] "Using root directory" path="/var/lib/kubelet" Dec 13 06:49:03 crc kubenswrapper[4971]: I1213 06:49:03.659187 4971 kubelet.go:418] "Attempting to sync node with API server" Dec 13 06:49:03 crc kubenswrapper[4971]: I1213 06:49:03.659208 4971 kubelet.go:313] "Adding static pod path" path="/etc/kubernetes/manifests" Dec 13 06:49:03 crc kubenswrapper[4971]: I1213 06:49:03.659234 4971 file.go:69] "Watching path" path="/etc/kubernetes/manifests" Dec 13 06:49:03 crc kubenswrapper[4971]: I1213 06:49:03.659249 4971 kubelet.go:324] "Adding apiserver pod source" Dec 13 06:49:03 crc kubenswrapper[4971]: I1213 06:49:03.659269 4971 apiserver.go:42] "Waiting for node sync before watching apiserver pods" Dec 13 06:49:03 crc kubenswrapper[4971]: I1213 06:49:03.661609 4971 kuberuntime_manager.go:262] "Container runtime initialized" containerRuntime="cri-o" version="1.31.5-4.rhaos4.18.gitdad78d5.el9" apiVersion="v1" Dec 13 06:49:03 crc kubenswrapper[4971]: W1213 06:49:03.661976 4971 reflector.go:561] k8s.io/client-go/informers/factory.go:160: failed to list *v1.Node: Get "https://api-int.crc.testing:6443/api/v1/nodes?fieldSelector=metadata.name%3Dcrc&limit=500&resourceVersion=0": dial tcp 38.102.83.195:6443: connect: connection refused Dec 13 06:49:03 crc kubenswrapper[4971]: I1213 06:49:03.662026 4971 certificate_store.go:130] Loading cert/key pair from "/var/lib/kubelet/pki/kubelet-server-current.pem". Dec 13 06:49:03 crc kubenswrapper[4971]: W1213 06:49:03.661979 4971 reflector.go:561] k8s.io/client-go/informers/factory.go:160: failed to list *v1.Service: Get "https://api-int.crc.testing:6443/api/v1/services?fieldSelector=spec.clusterIP%21%3DNone&limit=500&resourceVersion=0": dial tcp 38.102.83.195:6443: connect: connection refused Dec 13 06:49:03 crc kubenswrapper[4971]: E1213 06:49:03.662097 4971 reflector.go:158] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.Service: failed to list *v1.Service: Get \"https://api-int.crc.testing:6443/api/v1/services?fieldSelector=spec.clusterIP%21%3DNone&limit=500&resourceVersion=0\": dial tcp 38.102.83.195:6443: connect: connection refused" logger="UnhandledError" Dec 13 06:49:03 crc kubenswrapper[4971]: E1213 06:49:03.662102 4971 reflector.go:158] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.Node: failed to list *v1.Node: Get \"https://api-int.crc.testing:6443/api/v1/nodes?fieldSelector=metadata.name%3Dcrc&limit=500&resourceVersion=0\": dial tcp 38.102.83.195:6443: connect: connection refused" logger="UnhandledError" Dec 13 06:49:03 crc kubenswrapper[4971]: I1213 06:49:03.662831 4971 kubelet.go:854] "Not starting ClusterTrustBundle informer because we are in static kubelet mode" Dec 13 06:49:03 crc kubenswrapper[4971]: I1213 06:49:03.663394 4971 plugins.go:603] "Loaded volume plugin" pluginName="kubernetes.io/portworx-volume" Dec 13 06:49:03 crc kubenswrapper[4971]: I1213 06:49:03.663425 4971 plugins.go:603] "Loaded volume plugin" pluginName="kubernetes.io/empty-dir" Dec 13 06:49:03 crc kubenswrapper[4971]: I1213 06:49:03.663434 4971 plugins.go:603] "Loaded volume plugin" pluginName="kubernetes.io/git-repo" Dec 13 06:49:03 crc kubenswrapper[4971]: I1213 06:49:03.663441 4971 plugins.go:603] "Loaded volume plugin" pluginName="kubernetes.io/host-path" Dec 13 06:49:03 crc kubenswrapper[4971]: I1213 06:49:03.663453 4971 plugins.go:603] "Loaded volume plugin" pluginName="kubernetes.io/nfs" Dec 13 06:49:03 crc kubenswrapper[4971]: I1213 06:49:03.663460 4971 plugins.go:603] "Loaded volume plugin" pluginName="kubernetes.io/secret" Dec 13 06:49:03 crc kubenswrapper[4971]: I1213 06:49:03.663467 4971 plugins.go:603] "Loaded volume plugin" pluginName="kubernetes.io/iscsi" Dec 13 06:49:03 crc kubenswrapper[4971]: I1213 06:49:03.663488 4971 plugins.go:603] "Loaded volume plugin" pluginName="kubernetes.io/downward-api" Dec 13 06:49:03 crc kubenswrapper[4971]: I1213 06:49:03.663504 4971 plugins.go:603] "Loaded volume plugin" pluginName="kubernetes.io/fc" Dec 13 06:49:03 crc kubenswrapper[4971]: I1213 06:49:03.663530 4971 plugins.go:603] "Loaded volume plugin" pluginName="kubernetes.io/configmap" Dec 13 06:49:03 crc kubenswrapper[4971]: I1213 06:49:03.663544 4971 plugins.go:603] "Loaded volume plugin" pluginName="kubernetes.io/projected" Dec 13 06:49:03 crc kubenswrapper[4971]: I1213 06:49:03.663552 4971 plugins.go:603] "Loaded volume plugin" pluginName="kubernetes.io/local-volume" Dec 13 06:49:03 crc kubenswrapper[4971]: I1213 06:49:03.663730 4971 plugins.go:603] "Loaded volume plugin" pluginName="kubernetes.io/csi" Dec 13 06:49:03 crc kubenswrapper[4971]: I1213 06:49:03.664114 4971 server.go:1280] "Started kubelet" Dec 13 06:49:03 crc kubenswrapper[4971]: I1213 06:49:03.664338 4971 csi_plugin.go:884] Failed to contact API server when waiting for CSINode publishing: Get "https://api-int.crc.testing:6443/apis/storage.k8s.io/v1/csinodes/crc?resourceVersion=0": dial tcp 38.102.83.195:6443: connect: connection refused Dec 13 06:49:03 crc kubenswrapper[4971]: I1213 06:49:03.664462 4971 server.go:163] "Starting to listen" address="0.0.0.0" port=10250 Dec 13 06:49:03 crc kubenswrapper[4971]: I1213 06:49:03.664453 4971 ratelimit.go:55] "Setting rate limiting for endpoint" service="podresources" qps=100 burstTokens=10 Dec 13 06:49:03 crc systemd[1]: Started Kubernetes Kubelet. Dec 13 06:49:03 crc kubenswrapper[4971]: I1213 06:49:03.665968 4971 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate rotation is enabled Dec 13 06:49:03 crc kubenswrapper[4971]: I1213 06:49:03.666004 4971 fs_resource_analyzer.go:67] "Starting FS ResourceAnalyzer" Dec 13 06:49:03 crc kubenswrapper[4971]: I1213 06:49:03.666338 4971 server.go:236] "Starting to serve the podresources API" endpoint="unix:/var/lib/kubelet/pod-resources/kubelet.sock" Dec 13 06:49:03 crc kubenswrapper[4971]: E1213 06:49:03.666566 4971 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Dec 13 06:49:03 crc kubenswrapper[4971]: I1213 06:49:03.666746 4971 volume_manager.go:287] "The desired_state_of_world populator starts" Dec 13 06:49:03 crc kubenswrapper[4971]: I1213 06:49:03.666756 4971 volume_manager.go:289] "Starting Kubelet Volume Manager" Dec 13 06:49:03 crc kubenswrapper[4971]: I1213 06:49:03.666850 4971 desired_state_of_world_populator.go:146] "Desired state populator starts to run" Dec 13 06:49:03 crc kubenswrapper[4971]: I1213 06:49:03.666504 4971 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2025-11-30 11:17:37.954909698 +0000 UTC Dec 13 06:49:03 crc kubenswrapper[4971]: E1213 06:49:03.666379 4971 event.go:368] "Unable to write event (may retry after sleeping)" err="Post \"https://api-int.crc.testing:6443/api/v1/namespaces/default/events\": dial tcp 38.102.83.195:6443: connect: connection refused" event="&Event{ObjectMeta:{crc.1880b39da97e1d7c default 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Node,Namespace:,Name:crc,UID:crc,APIVersion:,ResourceVersion:,FieldPath:,},Reason:Starting,Message:Starting kubelet.,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2025-12-13 06:49:03.664086396 +0000 UTC m=+0.268495844,LastTimestamp:2025-12-13 06:49:03.664086396 +0000 UTC m=+0.268495844,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Dec 13 06:49:03 crc kubenswrapper[4971]: E1213 06:49:03.667791 4971 controller.go:145] "Failed to ensure lease exists, will retry" err="Get \"https://api-int.crc.testing:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/crc?timeout=10s\": dial tcp 38.102.83.195:6443: connect: connection refused" interval="200ms" Dec 13 06:49:03 crc kubenswrapper[4971]: I1213 06:49:03.668160 4971 factory.go:219] Registration of the containerd container factory failed: unable to create containerd client: containerd: cannot unix dial containerd api service: dial unix /run/containerd/containerd.sock: connect: no such file or directory Dec 13 06:49:03 crc kubenswrapper[4971]: I1213 06:49:03.668182 4971 factory.go:55] Registering systemd factory Dec 13 06:49:03 crc kubenswrapper[4971]: I1213 06:49:03.668214 4971 factory.go:221] Registration of the systemd container factory successfully Dec 13 06:49:03 crc kubenswrapper[4971]: I1213 06:49:03.668613 4971 factory.go:153] Registering CRI-O factory Dec 13 06:49:03 crc kubenswrapper[4971]: I1213 06:49:03.668674 4971 factory.go:221] Registration of the crio container factory successfully Dec 13 06:49:03 crc kubenswrapper[4971]: I1213 06:49:03.668710 4971 factory.go:103] Registering Raw factory Dec 13 06:49:03 crc kubenswrapper[4971]: I1213 06:49:03.668775 4971 manager.go:1196] Started watching for new ooms in manager Dec 13 06:49:03 crc kubenswrapper[4971]: W1213 06:49:03.668758 4971 reflector.go:561] k8s.io/client-go/informers/factory.go:160: failed to list *v1.CSIDriver: Get "https://api-int.crc.testing:6443/apis/storage.k8s.io/v1/csidrivers?limit=500&resourceVersion=0": dial tcp 38.102.83.195:6443: connect: connection refused Dec 13 06:49:03 crc kubenswrapper[4971]: E1213 06:49:03.668884 4971 reflector.go:158] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.CSIDriver: failed to list *v1.CSIDriver: Get \"https://api-int.crc.testing:6443/apis/storage.k8s.io/v1/csidrivers?limit=500&resourceVersion=0\": dial tcp 38.102.83.195:6443: connect: connection refused" logger="UnhandledError" Dec 13 06:49:03 crc kubenswrapper[4971]: I1213 06:49:03.670559 4971 server.go:460] "Adding debug handlers to kubelet server" Dec 13 06:49:03 crc kubenswrapper[4971]: I1213 06:49:03.671750 4971 manager.go:319] Starting recovery of all containers Dec 13 06:49:03 crc kubenswrapper[4971]: I1213 06:49:03.681224 4971 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="57a731c4-ef35-47a8-b875-bfb08a7f8011" volumeName="kubernetes.io/empty-dir/57a731c4-ef35-47a8-b875-bfb08a7f8011-utilities" seLinuxMountContext="" Dec 13 06:49:03 crc kubenswrapper[4971]: I1213 06:49:03.681303 4971 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="b11524ee-3fca-4b1b-9cdf-6da289fdbc7d" volumeName="kubernetes.io/empty-dir/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d-catalog-content" seLinuxMountContext="" Dec 13 06:49:03 crc kubenswrapper[4971]: I1213 06:49:03.681323 4971 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="1bf7eb37-55a3-4c65-b768-a94c82151e69" volumeName="kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-audit" seLinuxMountContext="" Dec 13 06:49:03 crc kubenswrapper[4971]: I1213 06:49:03.681376 4971 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="1bf7eb37-55a3-4c65-b768-a94c82151e69" volumeName="kubernetes.io/projected/1bf7eb37-55a3-4c65-b768-a94c82151e69-kube-api-access-sb6h7" seLinuxMountContext="" Dec 13 06:49:03 crc kubenswrapper[4971]: I1213 06:49:03.681391 4971 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="1bf7eb37-55a3-4c65-b768-a94c82151e69" volumeName="kubernetes.io/secret/1bf7eb37-55a3-4c65-b768-a94c82151e69-etcd-client" seLinuxMountContext="" Dec 13 06:49:03 crc kubenswrapper[4971]: I1213 06:49:03.681403 4971 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="3cb93b32-e0ae-4377-b9c8-fdb9842c6d59" volumeName="kubernetes.io/configmap/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59-serviceca" seLinuxMountContext="" Dec 13 06:49:03 crc kubenswrapper[4971]: I1213 06:49:03.681416 4971 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="09ae3b1a-e8e7-4524-b54b-61eab6f9239a" volumeName="kubernetes.io/configmap/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-audit-policies" seLinuxMountContext="" Dec 13 06:49:03 crc kubenswrapper[4971]: I1213 06:49:03.681429 4971 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="1bf7eb37-55a3-4c65-b768-a94c82151e69" volumeName="kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-image-import-ca" seLinuxMountContext="" Dec 13 06:49:03 crc kubenswrapper[4971]: I1213 06:49:03.681443 4971 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="308be0ea-9f5f-4b29-aeb1-5abd31a0b17b" volumeName="kubernetes.io/secret/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-apiservice-cert" seLinuxMountContext="" Dec 13 06:49:03 crc kubenswrapper[4971]: I1213 06:49:03.681457 4971 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="49c341d1-5089-4bc2-86a0-a5e165cfcc6b" volumeName="kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-trusted-ca-bundle" seLinuxMountContext="" Dec 13 06:49:03 crc kubenswrapper[4971]: I1213 06:49:03.681470 4971 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6509e943-70c6-444c-bc41-48a544e36fbd" volumeName="kubernetes.io/projected/6509e943-70c6-444c-bc41-48a544e36fbd-kube-api-access-6g6sz" seLinuxMountContext="" Dec 13 06:49:03 crc kubenswrapper[4971]: I1213 06:49:03.681484 4971 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="925f1c65-6136-48ba-85aa-3a3b50560753" volumeName="kubernetes.io/projected/925f1c65-6136-48ba-85aa-3a3b50560753-kube-api-access-s4n52" seLinuxMountContext="" Dec 13 06:49:03 crc kubenswrapper[4971]: I1213 06:49:03.681497 4971 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="bf126b07-da06-4140-9a57-dfd54fc6b486" volumeName="kubernetes.io/secret/bf126b07-da06-4140-9a57-dfd54fc6b486-image-registry-operator-tls" seLinuxMountContext="" Dec 13 06:49:03 crc kubenswrapper[4971]: I1213 06:49:03.682590 4971 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="01ab3dd5-8196-46d0-ad33-122e2ca51def" volumeName="kubernetes.io/configmap/01ab3dd5-8196-46d0-ad33-122e2ca51def-config" seLinuxMountContext="" Dec 13 06:49:03 crc kubenswrapper[4971]: I1213 06:49:03.682715 4971 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="22c825df-677d-4ca6-82db-3454ed06e783" volumeName="kubernetes.io/secret/22c825df-677d-4ca6-82db-3454ed06e783-machine-approver-tls" seLinuxMountContext="" Dec 13 06:49:03 crc kubenswrapper[4971]: I1213 06:49:03.682796 4971 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="31d8b7a1-420e-4252-a5b7-eebe8a111292" volumeName="kubernetes.io/projected/31d8b7a1-420e-4252-a5b7-eebe8a111292-kube-api-access-zgdk5" seLinuxMountContext="" Dec 13 06:49:03 crc kubenswrapper[4971]: I1213 06:49:03.682858 4971 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6402fda4-df10-493c-b4e5-d0569419652d" volumeName="kubernetes.io/configmap/6402fda4-df10-493c-b4e5-d0569419652d-config" seLinuxMountContext="" Dec 13 06:49:03 crc kubenswrapper[4971]: I1213 06:49:03.682947 4971 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="49ef4625-1d3a-4a9f-b595-c2433d32326d" volumeName="kubernetes.io/projected/49ef4625-1d3a-4a9f-b595-c2433d32326d-kube-api-access-pjr6v" seLinuxMountContext="" Dec 13 06:49:03 crc kubenswrapper[4971]: I1213 06:49:03.683020 4971 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="09ae3b1a-e8e7-4524-b54b-61eab6f9239a" volumeName="kubernetes.io/configmap/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-trusted-ca-bundle" seLinuxMountContext="" Dec 13 06:49:03 crc kubenswrapper[4971]: I1213 06:49:03.683084 4971 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="09efc573-dbb6-4249-bd59-9b87aba8dd28" volumeName="kubernetes.io/configmap/09efc573-dbb6-4249-bd59-9b87aba8dd28-etcd-service-ca" seLinuxMountContext="" Dec 13 06:49:03 crc kubenswrapper[4971]: I1213 06:49:03.683229 4971 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="1d611f23-29be-4491-8495-bee1670e935f" volumeName="kubernetes.io/empty-dir/1d611f23-29be-4491-8495-bee1670e935f-utilities" seLinuxMountContext="" Dec 13 06:49:03 crc kubenswrapper[4971]: I1213 06:49:03.684149 4971 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="49c341d1-5089-4bc2-86a0-a5e165cfcc6b" volumeName="kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-idp-0-file-data" seLinuxMountContext="" Dec 13 06:49:03 crc kubenswrapper[4971]: I1213 06:49:03.684205 4971 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="43509403-f426-496e-be36-56cef71462f5" volumeName="kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-service-ca" seLinuxMountContext="" Dec 13 06:49:03 crc kubenswrapper[4971]: I1213 06:49:03.684221 4971 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="49c341d1-5089-4bc2-86a0-a5e165cfcc6b" volumeName="kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-ocp-branding-template" seLinuxMountContext="" Dec 13 06:49:03 crc kubenswrapper[4971]: I1213 06:49:03.684234 4971 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="7bb08738-c794-4ee8-9972-3a62ca171029" volumeName="kubernetes.io/projected/7bb08738-c794-4ee8-9972-3a62ca171029-kube-api-access-279lb" seLinuxMountContext="" Dec 13 06:49:03 crc kubenswrapper[4971]: I1213 06:49:03.684252 4971 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="8f668bae-612b-4b75-9490-919e737c6a3b" volumeName="kubernetes.io/projected/8f668bae-612b-4b75-9490-919e737c6a3b-bound-sa-token" seLinuxMountContext="" Dec 13 06:49:03 crc kubenswrapper[4971]: I1213 06:49:03.684269 4971 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="09ae3b1a-e8e7-4524-b54b-61eab6f9239a" volumeName="kubernetes.io/secret/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-encryption-config" seLinuxMountContext="" Dec 13 06:49:03 crc kubenswrapper[4971]: I1213 06:49:03.684285 4971 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="1bf7eb37-55a3-4c65-b768-a94c82151e69" volumeName="kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-config" seLinuxMountContext="" Dec 13 06:49:03 crc kubenswrapper[4971]: I1213 06:49:03.684301 4971 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="20b0d48f-5fd6-431c-a545-e3c800c7b866" volumeName="kubernetes.io/projected/20b0d48f-5fd6-431c-a545-e3c800c7b866-kube-api-access-w9rds" seLinuxMountContext="" Dec 13 06:49:03 crc kubenswrapper[4971]: I1213 06:49:03.684314 4971 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="25e176fe-21b4-4974-b1ed-c8b94f112a7f" volumeName="kubernetes.io/configmap/25e176fe-21b4-4974-b1ed-c8b94f112a7f-signing-cabundle" seLinuxMountContext="" Dec 13 06:49:03 crc kubenswrapper[4971]: I1213 06:49:03.684352 4971 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="b6cd30de-2eeb-49a2-ab40-9167f4560ff5" volumeName="kubernetes.io/secret/b6cd30de-2eeb-49a2-ab40-9167f4560ff5-marketplace-operator-metrics" seLinuxMountContext="" Dec 13 06:49:03 crc kubenswrapper[4971]: I1213 06:49:03.684370 4971 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="e7e6199b-1264-4501-8953-767f51328d08" volumeName="kubernetes.io/projected/e7e6199b-1264-4501-8953-767f51328d08-kube-api-access" seLinuxMountContext="" Dec 13 06:49:03 crc kubenswrapper[4971]: I1213 06:49:03.684386 4971 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="5441d097-087c-4d9a-baa8-b210afa90fc9" volumeName="kubernetes.io/projected/5441d097-087c-4d9a-baa8-b210afa90fc9-kube-api-access-2d4wz" seLinuxMountContext="" Dec 13 06:49:03 crc kubenswrapper[4971]: I1213 06:49:03.684398 4971 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6509e943-70c6-444c-bc41-48a544e36fbd" volumeName="kubernetes.io/configmap/6509e943-70c6-444c-bc41-48a544e36fbd-service-ca-bundle" seLinuxMountContext="" Dec 13 06:49:03 crc kubenswrapper[4971]: I1213 06:49:03.684417 4971 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="7539238d-5fe0-46ed-884e-1c3b566537ec" volumeName="kubernetes.io/configmap/7539238d-5fe0-46ed-884e-1c3b566537ec-config" seLinuxMountContext="" Dec 13 06:49:03 crc kubenswrapper[4971]: I1213 06:49:03.684432 4971 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="b6cd30de-2eeb-49a2-ab40-9167f4560ff5" volumeName="kubernetes.io/configmap/b6cd30de-2eeb-49a2-ab40-9167f4560ff5-marketplace-trusted-ca" seLinuxMountContext="" Dec 13 06:49:03 crc kubenswrapper[4971]: I1213 06:49:03.684445 4971 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c" volumeName="kubernetes.io/projected/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c-kube-api-access-qs4fp" seLinuxMountContext="" Dec 13 06:49:03 crc kubenswrapper[4971]: I1213 06:49:03.684456 4971 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="25e176fe-21b4-4974-b1ed-c8b94f112a7f" volumeName="kubernetes.io/secret/25e176fe-21b4-4974-b1ed-c8b94f112a7f-signing-key" seLinuxMountContext="" Dec 13 06:49:03 crc kubenswrapper[4971]: I1213 06:49:03.684468 4971 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="43509403-f426-496e-be36-56cef71462f5" volumeName="kubernetes.io/secret/43509403-f426-496e-be36-56cef71462f5-console-oauth-config" seLinuxMountContext="" Dec 13 06:49:03 crc kubenswrapper[4971]: I1213 06:49:03.684481 4971 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="5441d097-087c-4d9a-baa8-b210afa90fc9" volumeName="kubernetes.io/configmap/5441d097-087c-4d9a-baa8-b210afa90fc9-config" seLinuxMountContext="" Dec 13 06:49:03 crc kubenswrapper[4971]: I1213 06:49:03.684495 4971 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="c03ee662-fb2f-4fc4-a2c1-af487c19d254" volumeName="kubernetes.io/configmap/c03ee662-fb2f-4fc4-a2c1-af487c19d254-service-ca-bundle" seLinuxMountContext="" Dec 13 06:49:03 crc kubenswrapper[4971]: I1213 06:49:03.684508 4971 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="fda69060-fa79-4696-b1a6-7980f124bf7c" volumeName="kubernetes.io/configmap/fda69060-fa79-4696-b1a6-7980f124bf7c-mcd-auth-proxy-config" seLinuxMountContext="" Dec 13 06:49:03 crc kubenswrapper[4971]: I1213 06:49:03.684545 4971 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="1386a44e-36a2-460c-96d0-0359d2b6f0f5" volumeName="kubernetes.io/configmap/1386a44e-36a2-460c-96d0-0359d2b6f0f5-config" seLinuxMountContext="" Dec 13 06:49:03 crc kubenswrapper[4971]: I1213 06:49:03.684564 4971 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="308be0ea-9f5f-4b29-aeb1-5abd31a0b17b" volumeName="kubernetes.io/projected/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-kube-api-access-6ccd8" seLinuxMountContext="" Dec 13 06:49:03 crc kubenswrapper[4971]: I1213 06:49:03.684577 4971 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="496e6271-fb68-4057-954e-a0d97a4afa3f" volumeName="kubernetes.io/projected/496e6271-fb68-4057-954e-a0d97a4afa3f-kube-api-access" seLinuxMountContext="" Dec 13 06:49:03 crc kubenswrapper[4971]: I1213 06:49:03.684590 4971 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6ea678ab-3438-413e-bfe3-290ae7725660" volumeName="kubernetes.io/configmap/6ea678ab-3438-413e-bfe3-290ae7725660-env-overrides" seLinuxMountContext="" Dec 13 06:49:03 crc kubenswrapper[4971]: I1213 06:49:03.684602 4971 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="87cf06ed-a83f-41a7-828d-70653580a8cb" volumeName="kubernetes.io/secret/87cf06ed-a83f-41a7-828d-70653580a8cb-metrics-tls" seLinuxMountContext="" Dec 13 06:49:03 crc kubenswrapper[4971]: I1213 06:49:03.684615 4971 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="c03ee662-fb2f-4fc4-a2c1-af487c19d254" volumeName="kubernetes.io/secret/c03ee662-fb2f-4fc4-a2c1-af487c19d254-stats-auth" seLinuxMountContext="" Dec 13 06:49:03 crc kubenswrapper[4971]: I1213 06:49:03.684629 4971 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="a31745f5-9847-4afe-82a5-3161cc66ca93" volumeName="kubernetes.io/secret/a31745f5-9847-4afe-82a5-3161cc66ca93-metrics-tls" seLinuxMountContext="" Dec 13 06:49:03 crc kubenswrapper[4971]: I1213 06:49:03.684642 4971 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="bf126b07-da06-4140-9a57-dfd54fc6b486" volumeName="kubernetes.io/projected/bf126b07-da06-4140-9a57-dfd54fc6b486-kube-api-access-rnphk" seLinuxMountContext="" Dec 13 06:49:03 crc kubenswrapper[4971]: I1213 06:49:03.684655 4971 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="c03ee662-fb2f-4fc4-a2c1-af487c19d254" volumeName="kubernetes.io/secret/c03ee662-fb2f-4fc4-a2c1-af487c19d254-metrics-certs" seLinuxMountContext="" Dec 13 06:49:03 crc kubenswrapper[4971]: I1213 06:49:03.684667 4971 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d" volumeName="kubernetes.io/projected/cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d-kube-api-access-x2m85" seLinuxMountContext="" Dec 13 06:49:03 crc kubenswrapper[4971]: I1213 06:49:03.684692 4971 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="0b78653f-4ff9-4508-8672-245ed9b561e3" volumeName="kubernetes.io/configmap/0b78653f-4ff9-4508-8672-245ed9b561e3-service-ca" seLinuxMountContext="" Dec 13 06:49:03 crc kubenswrapper[4971]: I1213 06:49:03.684714 4971 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6ea678ab-3438-413e-bfe3-290ae7725660" volumeName="kubernetes.io/secret/6ea678ab-3438-413e-bfe3-290ae7725660-ovn-node-metrics-cert" seLinuxMountContext="" Dec 13 06:49:03 crc kubenswrapper[4971]: I1213 06:49:03.684729 4971 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="8f668bae-612b-4b75-9490-919e737c6a3b" volumeName="kubernetes.io/projected/8f668bae-612b-4b75-9490-919e737c6a3b-registry-tls" seLinuxMountContext="" Dec 13 06:49:03 crc kubenswrapper[4971]: I1213 06:49:03.684750 4971 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="a0128f3a-b052-44ed-a84e-c4c8aaf17c13" volumeName="kubernetes.io/projected/a0128f3a-b052-44ed-a84e-c4c8aaf17c13-kube-api-access-gf66m" seLinuxMountContext="" Dec 13 06:49:03 crc kubenswrapper[4971]: I1213 06:49:03.684765 4971 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="4bb40260-dbaa-4fb0-84df-5e680505d512" volumeName="kubernetes.io/configmap/4bb40260-dbaa-4fb0-84df-5e680505d512-multus-daemon-config" seLinuxMountContext="" Dec 13 06:49:03 crc kubenswrapper[4971]: I1213 06:49:03.684781 4971 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="8cea82b4-6893-4ddc-af9f-1bb5ae425c5b" volumeName="kubernetes.io/configmap/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b-config" seLinuxMountContext="" Dec 13 06:49:03 crc kubenswrapper[4971]: I1213 06:49:03.684793 4971 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="8f668bae-612b-4b75-9490-919e737c6a3b" volumeName="kubernetes.io/projected/8f668bae-612b-4b75-9490-919e737c6a3b-kube-api-access-kfwg7" seLinuxMountContext="" Dec 13 06:49:03 crc kubenswrapper[4971]: I1213 06:49:03.684819 4971 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="925f1c65-6136-48ba-85aa-3a3b50560753" volumeName="kubernetes.io/secret/925f1c65-6136-48ba-85aa-3a3b50560753-ovn-control-plane-metrics-cert" seLinuxMountContext="" Dec 13 06:49:03 crc kubenswrapper[4971]: I1213 06:49:03.684832 4971 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="0b574797-001e-440a-8f4e-c0be86edad0f" volumeName="kubernetes.io/secret/0b574797-001e-440a-8f4e-c0be86edad0f-proxy-tls" seLinuxMountContext="" Dec 13 06:49:03 crc kubenswrapper[4971]: I1213 06:49:03.684844 4971 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="31d8b7a1-420e-4252-a5b7-eebe8a111292" volumeName="kubernetes.io/secret/31d8b7a1-420e-4252-a5b7-eebe8a111292-proxy-tls" seLinuxMountContext="" Dec 13 06:49:03 crc kubenswrapper[4971]: I1213 06:49:03.684860 4971 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="37a5e44f-9a88-4405-be8a-b645485e7312" volumeName="kubernetes.io/secret/37a5e44f-9a88-4405-be8a-b645485e7312-metrics-tls" seLinuxMountContext="" Dec 13 06:49:03 crc kubenswrapper[4971]: I1213 06:49:03.684874 4971 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="43509403-f426-496e-be36-56cef71462f5" volumeName="kubernetes.io/projected/43509403-f426-496e-be36-56cef71462f5-kube-api-access-qg5z5" seLinuxMountContext="" Dec 13 06:49:03 crc kubenswrapper[4971]: I1213 06:49:03.684896 4971 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="9d4552c7-cd75-42dd-8880-30dd377c49a4" volumeName="kubernetes.io/projected/9d4552c7-cd75-42dd-8880-30dd377c49a4-kube-api-access-pcxfs" seLinuxMountContext="" Dec 13 06:49:03 crc kubenswrapper[4971]: I1213 06:49:03.684910 4971 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="b11524ee-3fca-4b1b-9cdf-6da289fdbc7d" volumeName="kubernetes.io/empty-dir/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d-utilities" seLinuxMountContext="" Dec 13 06:49:03 crc kubenswrapper[4971]: I1213 06:49:03.684923 4971 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="b6cd30de-2eeb-49a2-ab40-9167f4560ff5" volumeName="kubernetes.io/projected/b6cd30de-2eeb-49a2-ab40-9167f4560ff5-kube-api-access-pj782" seLinuxMountContext="" Dec 13 06:49:03 crc kubenswrapper[4971]: I1213 06:49:03.684944 4971 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="bf126b07-da06-4140-9a57-dfd54fc6b486" volumeName="kubernetes.io/projected/bf126b07-da06-4140-9a57-dfd54fc6b486-bound-sa-token" seLinuxMountContext="" Dec 13 06:49:03 crc kubenswrapper[4971]: I1213 06:49:03.684956 4971 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="ef543e1b-8068-4ea3-b32a-61027b32e95d" volumeName="kubernetes.io/projected/ef543e1b-8068-4ea3-b32a-61027b32e95d-kube-api-access-s2kz5" seLinuxMountContext="" Dec 13 06:49:03 crc kubenswrapper[4971]: I1213 06:49:03.684969 4971 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="1d611f23-29be-4491-8495-bee1670e935f" volumeName="kubernetes.io/projected/1d611f23-29be-4491-8495-bee1670e935f-kube-api-access-bf2bz" seLinuxMountContext="" Dec 13 06:49:03 crc kubenswrapper[4971]: I1213 06:49:03.684981 4971 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="22c825df-677d-4ca6-82db-3454ed06e783" volumeName="kubernetes.io/configmap/22c825df-677d-4ca6-82db-3454ed06e783-config" seLinuxMountContext="" Dec 13 06:49:03 crc kubenswrapper[4971]: I1213 06:49:03.684993 4971 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="3ab1a177-2de0-46d9-b765-d0d0649bb42e" volumeName="kubernetes.io/secret/3ab1a177-2de0-46d9-b765-d0d0649bb42e-package-server-manager-serving-cert" seLinuxMountContext="" Dec 13 06:49:03 crc kubenswrapper[4971]: I1213 06:49:03.685011 4971 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="8f668bae-612b-4b75-9490-919e737c6a3b" volumeName="kubernetes.io/configmap/8f668bae-612b-4b75-9490-919e737c6a3b-registry-certificates" seLinuxMountContext="" Dec 13 06:49:03 crc kubenswrapper[4971]: I1213 06:49:03.685022 4971 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="57a731c4-ef35-47a8-b875-bfb08a7f8011" volumeName="kubernetes.io/empty-dir/57a731c4-ef35-47a8-b875-bfb08a7f8011-catalog-content" seLinuxMountContext="" Dec 13 06:49:03 crc kubenswrapper[4971]: I1213 06:49:03.685035 4971 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6ea678ab-3438-413e-bfe3-290ae7725660" volumeName="kubernetes.io/configmap/6ea678ab-3438-413e-bfe3-290ae7725660-ovnkube-config" seLinuxMountContext="" Dec 13 06:49:03 crc kubenswrapper[4971]: I1213 06:49:03.685050 4971 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="9d4552c7-cd75-42dd-8880-30dd377c49a4" volumeName="kubernetes.io/configmap/9d4552c7-cd75-42dd-8880-30dd377c49a4-trusted-ca" seLinuxMountContext="" Dec 13 06:49:03 crc kubenswrapper[4971]: I1213 06:49:03.685062 4971 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="9d4552c7-cd75-42dd-8880-30dd377c49a4" volumeName="kubernetes.io/secret/9d4552c7-cd75-42dd-8880-30dd377c49a4-serving-cert" seLinuxMountContext="" Dec 13 06:49:03 crc kubenswrapper[4971]: I1213 06:49:03.685078 4971 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="0b574797-001e-440a-8f4e-c0be86edad0f" volumeName="kubernetes.io/configmap/0b574797-001e-440a-8f4e-c0be86edad0f-mcc-auth-proxy-config" seLinuxMountContext="" Dec 13 06:49:03 crc kubenswrapper[4971]: I1213 06:49:03.685090 4971 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c" volumeName="kubernetes.io/secret/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c-serving-cert" seLinuxMountContext="" Dec 13 06:49:03 crc kubenswrapper[4971]: I1213 06:49:03.685108 4971 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="496e6271-fb68-4057-954e-a0d97a4afa3f" volumeName="kubernetes.io/configmap/496e6271-fb68-4057-954e-a0d97a4afa3f-config" seLinuxMountContext="" Dec 13 06:49:03 crc kubenswrapper[4971]: I1213 06:49:03.685138 4971 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="49c341d1-5089-4bc2-86a0-a5e165cfcc6b" volumeName="kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-template-provider-selection" seLinuxMountContext="" Dec 13 06:49:03 crc kubenswrapper[4971]: I1213 06:49:03.685156 4971 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="bf126b07-da06-4140-9a57-dfd54fc6b486" volumeName="kubernetes.io/configmap/bf126b07-da06-4140-9a57-dfd54fc6b486-trusted-ca" seLinuxMountContext="" Dec 13 06:49:03 crc kubenswrapper[4971]: I1213 06:49:03.685174 4971 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="d75a4c96-2883-4a0b-bab2-0fab2b6c0b49" volumeName="kubernetes.io/projected/d75a4c96-2883-4a0b-bab2-0fab2b6c0b49-kube-api-access-rczfb" seLinuxMountContext="" Dec 13 06:49:03 crc kubenswrapper[4971]: I1213 06:49:03.685188 4971 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="bc5039c0-ea34-426b-a2b7-fbbc87b49a6d" volumeName="kubernetes.io/projected/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d-kube-api-access-mnrrd" seLinuxMountContext="" Dec 13 06:49:03 crc kubenswrapper[4971]: I1213 06:49:03.685201 4971 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="ef543e1b-8068-4ea3-b32a-61027b32e95d" volumeName="kubernetes.io/configmap/ef543e1b-8068-4ea3-b32a-61027b32e95d-env-overrides" seLinuxMountContext="" Dec 13 06:49:03 crc kubenswrapper[4971]: I1213 06:49:03.685213 4971 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="43509403-f426-496e-be36-56cef71462f5" volumeName="kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-oauth-serving-cert" seLinuxMountContext="" Dec 13 06:49:03 crc kubenswrapper[4971]: I1213 06:49:03.685227 4971 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="49c341d1-5089-4bc2-86a0-a5e165cfcc6b" volumeName="kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-template-login" seLinuxMountContext="" Dec 13 06:49:03 crc kubenswrapper[4971]: I1213 06:49:03.685240 4971 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="5b88f790-22fa-440e-b583-365168c0b23d" volumeName="kubernetes.io/projected/5b88f790-22fa-440e-b583-365168c0b23d-kube-api-access-jkwtn" seLinuxMountContext="" Dec 13 06:49:03 crc kubenswrapper[4971]: I1213 06:49:03.685258 4971 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="b6312bbd-5731-4ea0-a20f-81d5a57df44a" volumeName="kubernetes.io/secret/b6312bbd-5731-4ea0-a20f-81d5a57df44a-srv-cert" seLinuxMountContext="" Dec 13 06:49:03 crc kubenswrapper[4971]: I1213 06:49:03.685271 4971 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="8cea82b4-6893-4ddc-af9f-1bb5ae425c5b" volumeName="kubernetes.io/projected/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b-kube-api-access-w4xd4" seLinuxMountContext="" Dec 13 06:49:03 crc kubenswrapper[4971]: I1213 06:49:03.685281 4971 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="8f668bae-612b-4b75-9490-919e737c6a3b" volumeName="kubernetes.io/secret/8f668bae-612b-4b75-9490-919e737c6a3b-installation-pull-secrets" seLinuxMountContext="" Dec 13 06:49:03 crc kubenswrapper[4971]: I1213 06:49:03.685293 4971 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="ef543e1b-8068-4ea3-b32a-61027b32e95d" volumeName="kubernetes.io/configmap/ef543e1b-8068-4ea3-b32a-61027b32e95d-ovnkube-identity-cm" seLinuxMountContext="" Dec 13 06:49:03 crc kubenswrapper[4971]: I1213 06:49:03.685317 4971 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="49c341d1-5089-4bc2-86a0-a5e165cfcc6b" volumeName="kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-service-ca" seLinuxMountContext="" Dec 13 06:49:03 crc kubenswrapper[4971]: I1213 06:49:03.685334 4971 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="4bb40260-dbaa-4fb0-84df-5e680505d512" volumeName="kubernetes.io/configmap/4bb40260-dbaa-4fb0-84df-5e680505d512-cni-binary-copy" seLinuxMountContext="" Dec 13 06:49:03 crc kubenswrapper[4971]: I1213 06:49:03.685370 4971 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="5225d0e4-402f-4861-b410-819f433b1803" volumeName="kubernetes.io/projected/5225d0e4-402f-4861-b410-819f433b1803-kube-api-access-9xfj7" seLinuxMountContext="" Dec 13 06:49:03 crc kubenswrapper[4971]: I1213 06:49:03.685382 4971 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="7583ce53-e0fe-4a16-9e4d-50516596a136" volumeName="kubernetes.io/projected/7583ce53-e0fe-4a16-9e4d-50516596a136-kube-api-access-xcphl" seLinuxMountContext="" Dec 13 06:49:03 crc kubenswrapper[4971]: I1213 06:49:03.685405 4971 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6402fda4-df10-493c-b4e5-d0569419652d" volumeName="kubernetes.io/configmap/6402fda4-df10-493c-b4e5-d0569419652d-images" seLinuxMountContext="" Dec 13 06:49:03 crc kubenswrapper[4971]: I1213 06:49:03.685418 4971 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6509e943-70c6-444c-bc41-48a544e36fbd" volumeName="kubernetes.io/secret/6509e943-70c6-444c-bc41-48a544e36fbd-serving-cert" seLinuxMountContext="" Dec 13 06:49:03 crc kubenswrapper[4971]: I1213 06:49:03.685429 4971 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="7539238d-5fe0-46ed-884e-1c3b566537ec" volumeName="kubernetes.io/secret/7539238d-5fe0-46ed-884e-1c3b566537ec-serving-cert" seLinuxMountContext="" Dec 13 06:49:03 crc kubenswrapper[4971]: I1213 06:49:03.685446 4971 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="7583ce53-e0fe-4a16-9e4d-50516596a136" volumeName="kubernetes.io/configmap/7583ce53-e0fe-4a16-9e4d-50516596a136-proxy-ca-bundles" seLinuxMountContext="" Dec 13 06:49:03 crc kubenswrapper[4971]: I1213 06:49:03.685462 4971 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="09efc573-dbb6-4249-bd59-9b87aba8dd28" volumeName="kubernetes.io/configmap/09efc573-dbb6-4249-bd59-9b87aba8dd28-config" seLinuxMountContext="" Dec 13 06:49:03 crc kubenswrapper[4971]: I1213 06:49:03.685479 4971 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="09efc573-dbb6-4249-bd59-9b87aba8dd28" volumeName="kubernetes.io/projected/09efc573-dbb6-4249-bd59-9b87aba8dd28-kube-api-access-8tdtz" seLinuxMountContext="" Dec 13 06:49:03 crc kubenswrapper[4971]: I1213 06:49:03.685492 4971 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="3cb93b32-e0ae-4377-b9c8-fdb9842c6d59" volumeName="kubernetes.io/projected/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59-kube-api-access-wxkg8" seLinuxMountContext="" Dec 13 06:49:03 crc kubenswrapper[4971]: I1213 06:49:03.685504 4971 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="496e6271-fb68-4057-954e-a0d97a4afa3f" volumeName="kubernetes.io/secret/496e6271-fb68-4057-954e-a0d97a4afa3f-serving-cert" seLinuxMountContext="" Dec 13 06:49:03 crc kubenswrapper[4971]: I1213 06:49:03.685542 4971 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="bc5039c0-ea34-426b-a2b7-fbbc87b49a6d" volumeName="kubernetes.io/secret/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d-serving-cert" seLinuxMountContext="" Dec 13 06:49:03 crc kubenswrapper[4971]: I1213 06:49:03.685559 4971 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="7583ce53-e0fe-4a16-9e4d-50516596a136" volumeName="kubernetes.io/configmap/7583ce53-e0fe-4a16-9e4d-50516596a136-config" seLinuxMountContext="" Dec 13 06:49:03 crc kubenswrapper[4971]: I1213 06:49:03.685577 4971 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="ef543e1b-8068-4ea3-b32a-61027b32e95d" volumeName="kubernetes.io/secret/ef543e1b-8068-4ea3-b32a-61027b32e95d-webhook-cert" seLinuxMountContext="" Dec 13 06:49:03 crc kubenswrapper[4971]: I1213 06:49:03.685593 4971 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="fda69060-fa79-4696-b1a6-7980f124bf7c" volumeName="kubernetes.io/projected/fda69060-fa79-4696-b1a6-7980f124bf7c-kube-api-access-xcgwh" seLinuxMountContext="" Dec 13 06:49:03 crc kubenswrapper[4971]: I1213 06:49:03.685631 4971 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="3ab1a177-2de0-46d9-b765-d0d0649bb42e" volumeName="kubernetes.io/projected/3ab1a177-2de0-46d9-b765-d0d0649bb42e-kube-api-access-4d4hj" seLinuxMountContext="" Dec 13 06:49:03 crc kubenswrapper[4971]: I1213 06:49:03.685648 4971 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="e7e6199b-1264-4501-8953-767f51328d08" volumeName="kubernetes.io/secret/e7e6199b-1264-4501-8953-767f51328d08-serving-cert" seLinuxMountContext="" Dec 13 06:49:03 crc kubenswrapper[4971]: I1213 06:49:03.685666 4971 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="f88749ec-7931-4ee7-b3fc-1ec5e11f92e9" volumeName="kubernetes.io/projected/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9-kube-api-access-dbsvg" seLinuxMountContext="" Dec 13 06:49:03 crc kubenswrapper[4971]: I1213 06:49:03.685689 4971 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="f88749ec-7931-4ee7-b3fc-1ec5e11f92e9" volumeName="kubernetes.io/secret/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9-srv-cert" seLinuxMountContext="" Dec 13 06:49:03 crc kubenswrapper[4971]: I1213 06:49:03.685707 4971 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="09ae3b1a-e8e7-4524-b54b-61eab6f9239a" volumeName="kubernetes.io/projected/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-kube-api-access-zkvpv" seLinuxMountContext="" Dec 13 06:49:03 crc kubenswrapper[4971]: I1213 06:49:03.685723 4971 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="09efc573-dbb6-4249-bd59-9b87aba8dd28" volumeName="kubernetes.io/configmap/09efc573-dbb6-4249-bd59-9b87aba8dd28-etcd-ca" seLinuxMountContext="" Dec 13 06:49:03 crc kubenswrapper[4971]: I1213 06:49:03.685743 4971 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="308be0ea-9f5f-4b29-aeb1-5abd31a0b17b" volumeName="kubernetes.io/secret/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-webhook-cert" seLinuxMountContext="" Dec 13 06:49:03 crc kubenswrapper[4971]: I1213 06:49:03.685757 4971 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="31d8b7a1-420e-4252-a5b7-eebe8a111292" volumeName="kubernetes.io/configmap/31d8b7a1-420e-4252-a5b7-eebe8a111292-images" seLinuxMountContext="" Dec 13 06:49:03 crc kubenswrapper[4971]: I1213 06:49:03.685769 4971 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="8f668bae-612b-4b75-9490-919e737c6a3b" volumeName="kubernetes.io/empty-dir/8f668bae-612b-4b75-9490-919e737c6a3b-ca-trust-extracted" seLinuxMountContext="" Dec 13 06:49:03 crc kubenswrapper[4971]: I1213 06:49:03.685783 4971 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="efdd0498-1daa-4136-9a4a-3b948c2293fc" volumeName="kubernetes.io/secret/efdd0498-1daa-4136-9a4a-3b948c2293fc-webhook-certs" seLinuxMountContext="" Dec 13 06:49:03 crc kubenswrapper[4971]: I1213 06:49:03.685795 4971 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="0b574797-001e-440a-8f4e-c0be86edad0f" volumeName="kubernetes.io/projected/0b574797-001e-440a-8f4e-c0be86edad0f-kube-api-access-lzf88" seLinuxMountContext="" Dec 13 06:49:03 crc kubenswrapper[4971]: I1213 06:49:03.685817 4971 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="49c341d1-5089-4bc2-86a0-a5e165cfcc6b" volumeName="kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-router-certs" seLinuxMountContext="" Dec 13 06:49:03 crc kubenswrapper[4971]: I1213 06:49:03.685828 4971 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="49c341d1-5089-4bc2-86a0-a5e165cfcc6b" volumeName="kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-serving-cert" seLinuxMountContext="" Dec 13 06:49:03 crc kubenswrapper[4971]: I1213 06:49:03.685840 4971 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6731426b-95fe-49ff-bb5f-40441049fde2" volumeName="kubernetes.io/projected/6731426b-95fe-49ff-bb5f-40441049fde2-kube-api-access-x7zkh" seLinuxMountContext="" Dec 13 06:49:03 crc kubenswrapper[4971]: I1213 06:49:03.685857 4971 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="7583ce53-e0fe-4a16-9e4d-50516596a136" volumeName="kubernetes.io/configmap/7583ce53-e0fe-4a16-9e4d-50516596a136-client-ca" seLinuxMountContext="" Dec 13 06:49:03 crc kubenswrapper[4971]: I1213 06:49:03.685873 4971 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6ea678ab-3438-413e-bfe3-290ae7725660" volumeName="kubernetes.io/projected/6ea678ab-3438-413e-bfe3-290ae7725660-kube-api-access-htfz6" seLinuxMountContext="" Dec 13 06:49:03 crc kubenswrapper[4971]: I1213 06:49:03.685887 4971 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="925f1c65-6136-48ba-85aa-3a3b50560753" volumeName="kubernetes.io/configmap/925f1c65-6136-48ba-85aa-3a3b50560753-env-overrides" seLinuxMountContext="" Dec 13 06:49:03 crc kubenswrapper[4971]: I1213 06:49:03.685908 4971 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="b6312bbd-5731-4ea0-a20f-81d5a57df44a" volumeName="kubernetes.io/secret/b6312bbd-5731-4ea0-a20f-81d5a57df44a-profile-collector-cert" seLinuxMountContext="" Dec 13 06:49:03 crc kubenswrapper[4971]: I1213 06:49:03.685923 4971 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="c03ee662-fb2f-4fc4-a2c1-af487c19d254" volumeName="kubernetes.io/secret/c03ee662-fb2f-4fc4-a2c1-af487c19d254-default-certificate" seLinuxMountContext="" Dec 13 06:49:03 crc kubenswrapper[4971]: I1213 06:49:03.685938 4971 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="1bf7eb37-55a3-4c65-b768-a94c82151e69" volumeName="kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-trusted-ca-bundle" seLinuxMountContext="" Dec 13 06:49:03 crc kubenswrapper[4971]: I1213 06:49:03.685950 4971 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="1bf7eb37-55a3-4c65-b768-a94c82151e69" volumeName="kubernetes.io/secret/1bf7eb37-55a3-4c65-b768-a94c82151e69-serving-cert" seLinuxMountContext="" Dec 13 06:49:03 crc kubenswrapper[4971]: I1213 06:49:03.685962 4971 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="22c825df-677d-4ca6-82db-3454ed06e783" volumeName="kubernetes.io/configmap/22c825df-677d-4ca6-82db-3454ed06e783-auth-proxy-config" seLinuxMountContext="" Dec 13 06:49:03 crc kubenswrapper[4971]: I1213 06:49:03.685981 4971 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="43509403-f426-496e-be36-56cef71462f5" volumeName="kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-trusted-ca-bundle" seLinuxMountContext="" Dec 13 06:49:03 crc kubenswrapper[4971]: I1213 06:49:03.685998 4971 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="efdd0498-1daa-4136-9a4a-3b948c2293fc" volumeName="kubernetes.io/projected/efdd0498-1daa-4136-9a4a-3b948c2293fc-kube-api-access-fqsjt" seLinuxMountContext="" Dec 13 06:49:03 crc kubenswrapper[4971]: I1213 06:49:03.686015 4971 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="7583ce53-e0fe-4a16-9e4d-50516596a136" volumeName="kubernetes.io/secret/7583ce53-e0fe-4a16-9e4d-50516596a136-serving-cert" seLinuxMountContext="" Dec 13 06:49:03 crc kubenswrapper[4971]: I1213 06:49:03.686028 4971 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="7bb08738-c794-4ee8-9972-3a62ca171029" volumeName="kubernetes.io/configmap/7bb08738-c794-4ee8-9972-3a62ca171029-cni-sysctl-allowlist" seLinuxMountContext="" Dec 13 06:49:03 crc kubenswrapper[4971]: I1213 06:49:03.686045 4971 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="87cf06ed-a83f-41a7-828d-70653580a8cb" volumeName="kubernetes.io/projected/87cf06ed-a83f-41a7-828d-70653580a8cb-kube-api-access-d6qdx" seLinuxMountContext="" Dec 13 06:49:03 crc kubenswrapper[4971]: I1213 06:49:03.686059 4971 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="a31745f5-9847-4afe-82a5-3161cc66ca93" volumeName="kubernetes.io/configmap/a31745f5-9847-4afe-82a5-3161cc66ca93-trusted-ca" seLinuxMountContext="" Dec 13 06:49:03 crc kubenswrapper[4971]: I1213 06:49:03.686077 4971 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="1386a44e-36a2-460c-96d0-0359d2b6f0f5" volumeName="kubernetes.io/projected/1386a44e-36a2-460c-96d0-0359d2b6f0f5-kube-api-access" seLinuxMountContext="" Dec 13 06:49:03 crc kubenswrapper[4971]: I1213 06:49:03.686094 4971 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="43509403-f426-496e-be36-56cef71462f5" volumeName="kubernetes.io/secret/43509403-f426-496e-be36-56cef71462f5-console-serving-cert" seLinuxMountContext="" Dec 13 06:49:03 crc kubenswrapper[4971]: I1213 06:49:03.686108 4971 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="5fe579f8-e8a6-4643-bce5-a661393c4dde" volumeName="kubernetes.io/secret/5fe579f8-e8a6-4643-bce5-a661393c4dde-certs" seLinuxMountContext="" Dec 13 06:49:03 crc kubenswrapper[4971]: I1213 06:49:03.686323 4971 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6731426b-95fe-49ff-bb5f-40441049fde2" volumeName="kubernetes.io/secret/6731426b-95fe-49ff-bb5f-40441049fde2-control-plane-machine-set-operator-tls" seLinuxMountContext="" Dec 13 06:49:03 crc kubenswrapper[4971]: I1213 06:49:03.686340 4971 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="b11524ee-3fca-4b1b-9cdf-6da289fdbc7d" volumeName="kubernetes.io/projected/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d-kube-api-access-x4zgh" seLinuxMountContext="" Dec 13 06:49:03 crc kubenswrapper[4971]: I1213 06:49:03.686352 4971 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="d75a4c96-2883-4a0b-bab2-0fab2b6c0b49" volumeName="kubernetes.io/configmap/d75a4c96-2883-4a0b-bab2-0fab2b6c0b49-iptables-alerter-script" seLinuxMountContext="" Dec 13 06:49:03 crc kubenswrapper[4971]: I1213 06:49:03.686374 4971 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="01ab3dd5-8196-46d0-ad33-122e2ca51def" volumeName="kubernetes.io/secret/01ab3dd5-8196-46d0-ad33-122e2ca51def-serving-cert" seLinuxMountContext="" Dec 13 06:49:03 crc kubenswrapper[4971]: I1213 06:49:03.686394 4971 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6402fda4-df10-493c-b4e5-d0569419652d" volumeName="kubernetes.io/secret/6402fda4-df10-493c-b4e5-d0569419652d-machine-api-operator-tls" seLinuxMountContext="" Dec 13 06:49:03 crc kubenswrapper[4971]: I1213 06:49:03.686408 4971 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="b6312bbd-5731-4ea0-a20f-81d5a57df44a" volumeName="kubernetes.io/projected/b6312bbd-5731-4ea0-a20f-81d5a57df44a-kube-api-access-249nr" seLinuxMountContext="" Dec 13 06:49:03 crc kubenswrapper[4971]: I1213 06:49:03.686426 4971 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="5225d0e4-402f-4861-b410-819f433b1803" volumeName="kubernetes.io/empty-dir/5225d0e4-402f-4861-b410-819f433b1803-utilities" seLinuxMountContext="" Dec 13 06:49:03 crc kubenswrapper[4971]: I1213 06:49:03.686439 4971 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="7bb08738-c794-4ee8-9972-3a62ca171029" volumeName="kubernetes.io/configmap/7bb08738-c794-4ee8-9972-3a62ca171029-cni-binary-copy" seLinuxMountContext="" Dec 13 06:49:03 crc kubenswrapper[4971]: I1213 06:49:03.686452 4971 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="09ae3b1a-e8e7-4524-b54b-61eab6f9239a" volumeName="kubernetes.io/configmap/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-etcd-serving-ca" seLinuxMountContext="" Dec 13 06:49:03 crc kubenswrapper[4971]: I1213 06:49:03.686472 4971 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c" volumeName="kubernetes.io/configmap/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c-config" seLinuxMountContext="" Dec 13 06:49:03 crc kubenswrapper[4971]: I1213 06:49:03.686484 4971 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="3b6479f0-333b-4a96-9adf-2099afdc2447" volumeName="kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr" seLinuxMountContext="" Dec 13 06:49:03 crc kubenswrapper[4971]: I1213 06:49:03.686496 4971 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="5225d0e4-402f-4861-b410-819f433b1803" volumeName="kubernetes.io/empty-dir/5225d0e4-402f-4861-b410-819f433b1803-catalog-content" seLinuxMountContext="" Dec 13 06:49:03 crc kubenswrapper[4971]: I1213 06:49:03.686510 4971 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="8f668bae-612b-4b75-9490-919e737c6a3b" volumeName="kubernetes.io/configmap/8f668bae-612b-4b75-9490-919e737c6a3b-trusted-ca" seLinuxMountContext="" Dec 13 06:49:03 crc kubenswrapper[4971]: I1213 06:49:03.686542 4971 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="bd23aa5c-e532-4e53-bccf-e79f130c5ae8" volumeName="kubernetes.io/projected/bd23aa5c-e532-4e53-bccf-e79f130c5ae8-kube-api-access-jhbk2" seLinuxMountContext="" Dec 13 06:49:03 crc kubenswrapper[4971]: I1213 06:49:03.686555 4971 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="01ab3dd5-8196-46d0-ad33-122e2ca51def" volumeName="kubernetes.io/projected/01ab3dd5-8196-46d0-ad33-122e2ca51def-kube-api-access-w7l8j" seLinuxMountContext="" Dec 13 06:49:03 crc kubenswrapper[4971]: I1213 06:49:03.686567 4971 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="09efc573-dbb6-4249-bd59-9b87aba8dd28" volumeName="kubernetes.io/secret/09efc573-dbb6-4249-bd59-9b87aba8dd28-etcd-client" seLinuxMountContext="" Dec 13 06:49:03 crc kubenswrapper[4971]: I1213 06:49:03.686580 4971 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="37a5e44f-9a88-4405-be8a-b645485e7312" volumeName="kubernetes.io/projected/37a5e44f-9a88-4405-be8a-b645485e7312-kube-api-access-rdwmf" seLinuxMountContext="" Dec 13 06:49:03 crc kubenswrapper[4971]: I1213 06:49:03.686593 4971 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" volumeName="kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf" seLinuxMountContext="" Dec 13 06:49:03 crc kubenswrapper[4971]: I1213 06:49:03.686605 4971 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="5441d097-087c-4d9a-baa8-b210afa90fc9" volumeName="kubernetes.io/configmap/5441d097-087c-4d9a-baa8-b210afa90fc9-client-ca" seLinuxMountContext="" Dec 13 06:49:03 crc kubenswrapper[4971]: I1213 06:49:03.686619 4971 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="5fe579f8-e8a6-4643-bce5-a661393c4dde" volumeName="kubernetes.io/secret/5fe579f8-e8a6-4643-bce5-a661393c4dde-node-bootstrap-token" seLinuxMountContext="" Dec 13 06:49:03 crc kubenswrapper[4971]: I1213 06:49:03.686631 4971 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6509e943-70c6-444c-bc41-48a544e36fbd" volumeName="kubernetes.io/configmap/6509e943-70c6-444c-bc41-48a544e36fbd-trusted-ca-bundle" seLinuxMountContext="" Dec 13 06:49:03 crc kubenswrapper[4971]: I1213 06:49:03.686644 4971 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="96b93a3a-6083-4aea-8eab-fe1aa8245ad9" volumeName="kubernetes.io/projected/96b93a3a-6083-4aea-8eab-fe1aa8245ad9-kube-api-access-nzwt7" seLinuxMountContext="" Dec 13 06:49:03 crc kubenswrapper[4971]: I1213 06:49:03.686658 4971 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="1bf7eb37-55a3-4c65-b768-a94c82151e69" volumeName="kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-etcd-serving-ca" seLinuxMountContext="" Dec 13 06:49:03 crc kubenswrapper[4971]: I1213 06:49:03.686670 4971 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="25e176fe-21b4-4974-b1ed-c8b94f112a7f" volumeName="kubernetes.io/projected/25e176fe-21b4-4974-b1ed-c8b94f112a7f-kube-api-access-d4lsv" seLinuxMountContext="" Dec 13 06:49:03 crc kubenswrapper[4971]: I1213 06:49:03.686690 4971 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="308be0ea-9f5f-4b29-aeb1-5abd31a0b17b" volumeName="kubernetes.io/empty-dir/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-tmpfs" seLinuxMountContext="" Dec 13 06:49:03 crc kubenswrapper[4971]: I1213 06:49:03.686702 4971 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="49c341d1-5089-4bc2-86a0-a5e165cfcc6b" volumeName="kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-template-error" seLinuxMountContext="" Dec 13 06:49:03 crc kubenswrapper[4971]: I1213 06:49:03.686714 4971 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="9d751cbb-f2e2-430d-9754-c882a5e924a5" volumeName="kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl" seLinuxMountContext="" Dec 13 06:49:03 crc kubenswrapper[4971]: I1213 06:49:03.686724 4971 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="c03ee662-fb2f-4fc4-a2c1-af487c19d254" volumeName="kubernetes.io/projected/c03ee662-fb2f-4fc4-a2c1-af487c19d254-kube-api-access-v47cf" seLinuxMountContext="" Dec 13 06:49:03 crc kubenswrapper[4971]: I1213 06:49:03.686737 4971 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="925f1c65-6136-48ba-85aa-3a3b50560753" volumeName="kubernetes.io/configmap/925f1c65-6136-48ba-85aa-3a3b50560753-ovnkube-config" seLinuxMountContext="" Dec 13 06:49:03 crc kubenswrapper[4971]: I1213 06:49:03.686750 4971 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="96b93a3a-6083-4aea-8eab-fe1aa8245ad9" volumeName="kubernetes.io/secret/96b93a3a-6083-4aea-8eab-fe1aa8245ad9-metrics-tls" seLinuxMountContext="" Dec 13 06:49:03 crc kubenswrapper[4971]: I1213 06:49:03.686764 4971 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="a31745f5-9847-4afe-82a5-3161cc66ca93" volumeName="kubernetes.io/projected/a31745f5-9847-4afe-82a5-3161cc66ca93-bound-sa-token" seLinuxMountContext="" Dec 13 06:49:03 crc kubenswrapper[4971]: I1213 06:49:03.686777 4971 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="09efc573-dbb6-4249-bd59-9b87aba8dd28" volumeName="kubernetes.io/secret/09efc573-dbb6-4249-bd59-9b87aba8dd28-serving-cert" seLinuxMountContext="" Dec 13 06:49:03 crc kubenswrapper[4971]: I1213 06:49:03.686790 4971 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="0b78653f-4ff9-4508-8672-245ed9b561e3" volumeName="kubernetes.io/secret/0b78653f-4ff9-4508-8672-245ed9b561e3-serving-cert" seLinuxMountContext="" Dec 13 06:49:03 crc kubenswrapper[4971]: I1213 06:49:03.686803 4971 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6402fda4-df10-493c-b4e5-d0569419652d" volumeName="kubernetes.io/projected/6402fda4-df10-493c-b4e5-d0569419652d-kube-api-access-mg5zb" seLinuxMountContext="" Dec 13 06:49:03 crc kubenswrapper[4971]: I1213 06:49:03.686816 4971 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6ea678ab-3438-413e-bfe3-290ae7725660" volumeName="kubernetes.io/configmap/6ea678ab-3438-413e-bfe3-290ae7725660-ovnkube-script-lib" seLinuxMountContext="" Dec 13 06:49:03 crc kubenswrapper[4971]: I1213 06:49:03.686837 4971 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="1386a44e-36a2-460c-96d0-0359d2b6f0f5" volumeName="kubernetes.io/secret/1386a44e-36a2-460c-96d0-0359d2b6f0f5-serving-cert" seLinuxMountContext="" Dec 13 06:49:03 crc kubenswrapper[4971]: I1213 06:49:03.686850 4971 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="22c825df-677d-4ca6-82db-3454ed06e783" volumeName="kubernetes.io/projected/22c825df-677d-4ca6-82db-3454ed06e783-kube-api-access-7c4vf" seLinuxMountContext="" Dec 13 06:49:03 crc kubenswrapper[4971]: I1213 06:49:03.686865 4971 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="31d8b7a1-420e-4252-a5b7-eebe8a111292" volumeName="kubernetes.io/configmap/31d8b7a1-420e-4252-a5b7-eebe8a111292-auth-proxy-config" seLinuxMountContext="" Dec 13 06:49:03 crc kubenswrapper[4971]: I1213 06:49:03.686877 4971 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="5fe579f8-e8a6-4643-bce5-a661393c4dde" volumeName="kubernetes.io/projected/5fe579f8-e8a6-4643-bce5-a661393c4dde-kube-api-access-fcqwp" seLinuxMountContext="" Dec 13 06:49:03 crc kubenswrapper[4971]: I1213 06:49:03.686895 4971 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="bc5039c0-ea34-426b-a2b7-fbbc87b49a6d" volumeName="kubernetes.io/empty-dir/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d-available-featuregates" seLinuxMountContext="" Dec 13 06:49:03 crc kubenswrapper[4971]: I1213 06:49:03.686907 4971 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="e7e6199b-1264-4501-8953-767f51328d08" volumeName="kubernetes.io/configmap/e7e6199b-1264-4501-8953-767f51328d08-config" seLinuxMountContext="" Dec 13 06:49:03 crc kubenswrapper[4971]: I1213 06:49:03.686937 4971 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="1d611f23-29be-4491-8495-bee1670e935f" volumeName="kubernetes.io/empty-dir/1d611f23-29be-4491-8495-bee1670e935f-catalog-content" seLinuxMountContext="" Dec 13 06:49:03 crc kubenswrapper[4971]: I1213 06:49:03.686950 4971 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="4bb40260-dbaa-4fb0-84df-5e680505d512" volumeName="kubernetes.io/projected/4bb40260-dbaa-4fb0-84df-5e680505d512-kube-api-access-2w9zh" seLinuxMountContext="" Dec 13 06:49:03 crc kubenswrapper[4971]: I1213 06:49:03.686969 4971 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="a0128f3a-b052-44ed-a84e-c4c8aaf17c13" volumeName="kubernetes.io/secret/a0128f3a-b052-44ed-a84e-c4c8aaf17c13-samples-operator-tls" seLinuxMountContext="" Dec 13 06:49:03 crc kubenswrapper[4971]: I1213 06:49:03.686985 4971 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="a31745f5-9847-4afe-82a5-3161cc66ca93" volumeName="kubernetes.io/projected/a31745f5-9847-4afe-82a5-3161cc66ca93-kube-api-access-lz9wn" seLinuxMountContext="" Dec 13 06:49:03 crc kubenswrapper[4971]: I1213 06:49:03.687017 4971 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="49c341d1-5089-4bc2-86a0-a5e165cfcc6b" volumeName="kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-session" seLinuxMountContext="" Dec 13 06:49:03 crc kubenswrapper[4971]: I1213 06:49:03.687031 4971 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="8cea82b4-6893-4ddc-af9f-1bb5ae425c5b" volumeName="kubernetes.io/secret/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b-serving-cert" seLinuxMountContext="" Dec 13 06:49:03 crc kubenswrapper[4971]: I1213 06:49:03.687043 4971 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="fda69060-fa79-4696-b1a6-7980f124bf7c" volumeName="kubernetes.io/secret/fda69060-fa79-4696-b1a6-7980f124bf7c-proxy-tls" seLinuxMountContext="" Dec 13 06:49:03 crc kubenswrapper[4971]: I1213 06:49:03.687060 4971 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="09ae3b1a-e8e7-4524-b54b-61eab6f9239a" volumeName="kubernetes.io/secret/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-serving-cert" seLinuxMountContext="" Dec 13 06:49:03 crc kubenswrapper[4971]: I1213 06:49:03.687073 4971 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="1bf7eb37-55a3-4c65-b768-a94c82151e69" volumeName="kubernetes.io/secret/1bf7eb37-55a3-4c65-b768-a94c82151e69-encryption-config" seLinuxMountContext="" Dec 13 06:49:03 crc kubenswrapper[4971]: I1213 06:49:03.687086 4971 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="20b0d48f-5fd6-431c-a545-e3c800c7b866" volumeName="kubernetes.io/secret/20b0d48f-5fd6-431c-a545-e3c800c7b866-cert" seLinuxMountContext="" Dec 13 06:49:03 crc kubenswrapper[4971]: I1213 06:49:03.687098 4971 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="49c341d1-5089-4bc2-86a0-a5e165cfcc6b" volumeName="kubernetes.io/projected/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-kube-api-access-ngvvp" seLinuxMountContext="" Dec 13 06:49:03 crc kubenswrapper[4971]: I1213 06:49:03.687111 4971 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="0b78653f-4ff9-4508-8672-245ed9b561e3" volumeName="kubernetes.io/projected/0b78653f-4ff9-4508-8672-245ed9b561e3-kube-api-access" seLinuxMountContext="" Dec 13 06:49:03 crc kubenswrapper[4971]: I1213 06:49:03.687124 4971 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="57a731c4-ef35-47a8-b875-bfb08a7f8011" volumeName="kubernetes.io/projected/57a731c4-ef35-47a8-b875-bfb08a7f8011-kube-api-access-cfbct" seLinuxMountContext="" Dec 13 06:49:03 crc kubenswrapper[4971]: I1213 06:49:03.687136 4971 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6509e943-70c6-444c-bc41-48a544e36fbd" volumeName="kubernetes.io/configmap/6509e943-70c6-444c-bc41-48a544e36fbd-config" seLinuxMountContext="" Dec 13 06:49:03 crc kubenswrapper[4971]: I1213 06:49:03.687147 4971 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="9d4552c7-cd75-42dd-8880-30dd377c49a4" volumeName="kubernetes.io/configmap/9d4552c7-cd75-42dd-8880-30dd377c49a4-config" seLinuxMountContext="" Dec 13 06:49:03 crc kubenswrapper[4971]: I1213 06:49:03.687161 4971 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="7539238d-5fe0-46ed-884e-1c3b566537ec" volumeName="kubernetes.io/projected/7539238d-5fe0-46ed-884e-1c3b566537ec-kube-api-access-tk88c" seLinuxMountContext="" Dec 13 06:49:03 crc kubenswrapper[4971]: I1213 06:49:03.687174 4971 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="8f668bae-612b-4b75-9490-919e737c6a3b" volumeName="kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" seLinuxMountContext="" Dec 13 06:49:03 crc kubenswrapper[4971]: I1213 06:49:03.691114 4971 reconstruct.go:144] "Volume is marked device as uncertain and added into the actual state" volumeName="kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" deviceMountPath="/var/lib/kubelet/plugins/kubernetes.io/csi/kubevirt.io.hostpath-provisioner/1f4776af88835e41c12b831b4c9fed40233456d14189815a54dbe7f892fc1983/globalmount" Dec 13 06:49:03 crc kubenswrapper[4971]: I1213 06:49:03.691160 4971 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="09ae3b1a-e8e7-4524-b54b-61eab6f9239a" volumeName="kubernetes.io/secret/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-etcd-client" seLinuxMountContext="" Dec 13 06:49:03 crc kubenswrapper[4971]: I1213 06:49:03.691175 4971 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="44663579-783b-4372-86d6-acf235a62d72" volumeName="kubernetes.io/projected/44663579-783b-4372-86d6-acf235a62d72-kube-api-access-vt5rc" seLinuxMountContext="" Dec 13 06:49:03 crc kubenswrapper[4971]: I1213 06:49:03.691187 4971 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="49c341d1-5089-4bc2-86a0-a5e165cfcc6b" volumeName="kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-cliconfig" seLinuxMountContext="" Dec 13 06:49:03 crc kubenswrapper[4971]: I1213 06:49:03.691199 4971 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" volumeName="kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert" seLinuxMountContext="" Dec 13 06:49:03 crc kubenswrapper[4971]: I1213 06:49:03.691210 4971 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="87cf06ed-a83f-41a7-828d-70653580a8cb" volumeName="kubernetes.io/configmap/87cf06ed-a83f-41a7-828d-70653580a8cb-config-volume" seLinuxMountContext="" Dec 13 06:49:03 crc kubenswrapper[4971]: I1213 06:49:03.691220 4971 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="f88749ec-7931-4ee7-b3fc-1ec5e11f92e9" volumeName="kubernetes.io/secret/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9-profile-collector-cert" seLinuxMountContext="" Dec 13 06:49:03 crc kubenswrapper[4971]: I1213 06:49:03.691230 4971 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="43509403-f426-496e-be36-56cef71462f5" volumeName="kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-console-config" seLinuxMountContext="" Dec 13 06:49:03 crc kubenswrapper[4971]: I1213 06:49:03.691241 4971 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="49c341d1-5089-4bc2-86a0-a5e165cfcc6b" volumeName="kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-audit-policies" seLinuxMountContext="" Dec 13 06:49:03 crc kubenswrapper[4971]: I1213 06:49:03.691251 4971 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="5441d097-087c-4d9a-baa8-b210afa90fc9" volumeName="kubernetes.io/secret/5441d097-087c-4d9a-baa8-b210afa90fc9-serving-cert" seLinuxMountContext="" Dec 13 06:49:03 crc kubenswrapper[4971]: I1213 06:49:03.691262 4971 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="5b88f790-22fa-440e-b583-365168c0b23d" volumeName="kubernetes.io/secret/5b88f790-22fa-440e-b583-365168c0b23d-metrics-certs" seLinuxMountContext="" Dec 13 06:49:03 crc kubenswrapper[4971]: I1213 06:49:03.691272 4971 reconstruct.go:97] "Volume reconstruction finished" Dec 13 06:49:03 crc kubenswrapper[4971]: I1213 06:49:03.691280 4971 reconciler.go:26] "Reconciler: start to sync state" Dec 13 06:49:03 crc kubenswrapper[4971]: I1213 06:49:03.693685 4971 manager.go:324] Recovery completed Dec 13 06:49:03 crc kubenswrapper[4971]: I1213 06:49:03.704116 4971 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Dec 13 06:49:03 crc kubenswrapper[4971]: I1213 06:49:03.705847 4971 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 13 06:49:03 crc kubenswrapper[4971]: I1213 06:49:03.705882 4971 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 13 06:49:03 crc kubenswrapper[4971]: I1213 06:49:03.705890 4971 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 13 06:49:03 crc kubenswrapper[4971]: I1213 06:49:03.709748 4971 cpu_manager.go:225] "Starting CPU manager" policy="none" Dec 13 06:49:03 crc kubenswrapper[4971]: I1213 06:49:03.709782 4971 cpu_manager.go:226] "Reconciling" reconcilePeriod="10s" Dec 13 06:49:03 crc kubenswrapper[4971]: I1213 06:49:03.709806 4971 state_mem.go:36] "Initialized new in-memory state store" Dec 13 06:49:03 crc kubenswrapper[4971]: I1213 06:49:03.715110 4971 kubelet_network_linux.go:50] "Initialized iptables rules." protocol="IPv4" Dec 13 06:49:03 crc kubenswrapper[4971]: I1213 06:49:03.716636 4971 kubelet_network_linux.go:50] "Initialized iptables rules." protocol="IPv6" Dec 13 06:49:03 crc kubenswrapper[4971]: I1213 06:49:03.767125 4971 status_manager.go:217] "Starting to sync pod status with apiserver" Dec 13 06:49:03 crc kubenswrapper[4971]: I1213 06:49:03.767195 4971 kubelet.go:2335] "Starting kubelet main sync loop" Dec 13 06:49:03 crc kubenswrapper[4971]: E1213 06:49:03.767278 4971 kubelet.go:2359] "Skipping pod synchronization" err="[container runtime status check may not have completed yet, PLEG is not healthy: pleg has yet to be successful]" Dec 13 06:49:03 crc kubenswrapper[4971]: E1213 06:49:03.767295 4971 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Dec 13 06:49:03 crc kubenswrapper[4971]: W1213 06:49:03.768625 4971 reflector.go:561] k8s.io/client-go/informers/factory.go:160: failed to list *v1.RuntimeClass: Get "https://api-int.crc.testing:6443/apis/node.k8s.io/v1/runtimeclasses?limit=500&resourceVersion=0": dial tcp 38.102.83.195:6443: connect: connection refused Dec 13 06:49:03 crc kubenswrapper[4971]: E1213 06:49:03.768701 4971 reflector.go:158] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.RuntimeClass: failed to list *v1.RuntimeClass: Get \"https://api-int.crc.testing:6443/apis/node.k8s.io/v1/runtimeclasses?limit=500&resourceVersion=0\": dial tcp 38.102.83.195:6443: connect: connection refused" logger="UnhandledError" Dec 13 06:49:03 crc kubenswrapper[4971]: I1213 06:49:03.771653 4971 policy_none.go:49] "None policy: Start" Dec 13 06:49:03 crc kubenswrapper[4971]: I1213 06:49:03.773688 4971 memory_manager.go:170] "Starting memorymanager" policy="None" Dec 13 06:49:03 crc kubenswrapper[4971]: I1213 06:49:03.773736 4971 state_mem.go:35] "Initializing new in-memory state store" Dec 13 06:49:03 crc kubenswrapper[4971]: I1213 06:49:03.838424 4971 manager.go:334] "Starting Device Plugin manager" Dec 13 06:49:03 crc kubenswrapper[4971]: I1213 06:49:03.838539 4971 manager.go:513] "Failed to read data from checkpoint" checkpoint="kubelet_internal_checkpoint" err="checkpoint is not found" Dec 13 06:49:03 crc kubenswrapper[4971]: I1213 06:49:03.838561 4971 server.go:79] "Starting device plugin registration server" Dec 13 06:49:03 crc kubenswrapper[4971]: I1213 06:49:03.839664 4971 eviction_manager.go:189] "Eviction manager: starting control loop" Dec 13 06:49:03 crc kubenswrapper[4971]: I1213 06:49:03.839709 4971 container_log_manager.go:189] "Initializing container log rotate workers" workers=1 monitorPeriod="10s" Dec 13 06:49:03 crc kubenswrapper[4971]: I1213 06:49:03.840014 4971 plugin_watcher.go:51] "Plugin Watcher Start" path="/var/lib/kubelet/plugins_registry" Dec 13 06:49:03 crc kubenswrapper[4971]: I1213 06:49:03.840125 4971 plugin_manager.go:116] "The desired_state_of_world populator (plugin watcher) starts" Dec 13 06:49:03 crc kubenswrapper[4971]: I1213 06:49:03.840142 4971 plugin_manager.go:118] "Starting Kubelet Plugin Manager" Dec 13 06:49:03 crc kubenswrapper[4971]: E1213 06:49:03.852334 4971 eviction_manager.go:285] "Eviction manager: failed to get summary stats" err="failed to get node info: node \"crc\" not found" Dec 13 06:49:03 crc kubenswrapper[4971]: I1213 06:49:03.867725 4971 kubelet.go:2421] "SyncLoop ADD" source="file" pods=["openshift-kube-controller-manager/kube-controller-manager-crc","openshift-kube-scheduler/openshift-kube-scheduler-crc","openshift-machine-config-operator/kube-rbac-proxy-crio-crc","openshift-etcd/etcd-crc","openshift-kube-apiserver/kube-apiserver-crc"] Dec 13 06:49:03 crc kubenswrapper[4971]: I1213 06:49:03.867896 4971 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Dec 13 06:49:03 crc kubenswrapper[4971]: E1213 06:49:03.868546 4971 controller.go:145] "Failed to ensure lease exists, will retry" err="Get \"https://api-int.crc.testing:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/crc?timeout=10s\": dial tcp 38.102.83.195:6443: connect: connection refused" interval="400ms" Dec 13 06:49:03 crc kubenswrapper[4971]: I1213 06:49:03.869399 4971 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 13 06:49:03 crc kubenswrapper[4971]: I1213 06:49:03.869456 4971 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 13 06:49:03 crc kubenswrapper[4971]: I1213 06:49:03.869472 4971 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 13 06:49:03 crc kubenswrapper[4971]: I1213 06:49:03.869684 4971 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Dec 13 06:49:03 crc kubenswrapper[4971]: I1213 06:49:03.869879 4971 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-controller-manager/kube-controller-manager-crc" Dec 13 06:49:03 crc kubenswrapper[4971]: I1213 06:49:03.869910 4971 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Dec 13 06:49:03 crc kubenswrapper[4971]: I1213 06:49:03.870869 4971 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 13 06:49:03 crc kubenswrapper[4971]: I1213 06:49:03.870894 4971 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 13 06:49:03 crc kubenswrapper[4971]: I1213 06:49:03.870899 4971 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 13 06:49:03 crc kubenswrapper[4971]: I1213 06:49:03.870922 4971 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 13 06:49:03 crc kubenswrapper[4971]: I1213 06:49:03.870934 4971 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 13 06:49:03 crc kubenswrapper[4971]: I1213 06:49:03.870907 4971 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 13 06:49:03 crc kubenswrapper[4971]: I1213 06:49:03.871346 4971 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Dec 13 06:49:03 crc kubenswrapper[4971]: I1213 06:49:03.871469 4971 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" Dec 13 06:49:03 crc kubenswrapper[4971]: I1213 06:49:03.871509 4971 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Dec 13 06:49:03 crc kubenswrapper[4971]: I1213 06:49:03.873161 4971 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 13 06:49:03 crc kubenswrapper[4971]: I1213 06:49:03.873202 4971 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 13 06:49:03 crc kubenswrapper[4971]: I1213 06:49:03.873210 4971 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 13 06:49:03 crc kubenswrapper[4971]: I1213 06:49:03.873214 4971 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 13 06:49:03 crc kubenswrapper[4971]: I1213 06:49:03.873229 4971 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 13 06:49:03 crc kubenswrapper[4971]: I1213 06:49:03.873242 4971 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 13 06:49:03 crc kubenswrapper[4971]: I1213 06:49:03.873358 4971 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Dec 13 06:49:03 crc kubenswrapper[4971]: I1213 06:49:03.873502 4971 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" Dec 13 06:49:03 crc kubenswrapper[4971]: I1213 06:49:03.873556 4971 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Dec 13 06:49:03 crc kubenswrapper[4971]: I1213 06:49:03.874350 4971 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 13 06:49:03 crc kubenswrapper[4971]: I1213 06:49:03.874363 4971 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 13 06:49:03 crc kubenswrapper[4971]: I1213 06:49:03.874368 4971 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 13 06:49:03 crc kubenswrapper[4971]: I1213 06:49:03.874379 4971 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 13 06:49:03 crc kubenswrapper[4971]: I1213 06:49:03.874384 4971 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 13 06:49:03 crc kubenswrapper[4971]: I1213 06:49:03.874387 4971 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 13 06:49:03 crc kubenswrapper[4971]: I1213 06:49:03.874531 4971 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Dec 13 06:49:03 crc kubenswrapper[4971]: I1213 06:49:03.874670 4971 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-etcd/etcd-crc" Dec 13 06:49:03 crc kubenswrapper[4971]: I1213 06:49:03.874717 4971 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Dec 13 06:49:03 crc kubenswrapper[4971]: I1213 06:49:03.875502 4971 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 13 06:49:03 crc kubenswrapper[4971]: I1213 06:49:03.875538 4971 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 13 06:49:03 crc kubenswrapper[4971]: I1213 06:49:03.875550 4971 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 13 06:49:03 crc kubenswrapper[4971]: I1213 06:49:03.875555 4971 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 13 06:49:03 crc kubenswrapper[4971]: I1213 06:49:03.875568 4971 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 13 06:49:03 crc kubenswrapper[4971]: I1213 06:49:03.875575 4971 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 13 06:49:03 crc kubenswrapper[4971]: I1213 06:49:03.875762 4971 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/kube-apiserver-crc" Dec 13 06:49:03 crc kubenswrapper[4971]: I1213 06:49:03.875791 4971 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Dec 13 06:49:03 crc kubenswrapper[4971]: I1213 06:49:03.876369 4971 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 13 06:49:03 crc kubenswrapper[4971]: I1213 06:49:03.876392 4971 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 13 06:49:03 crc kubenswrapper[4971]: I1213 06:49:03.876399 4971 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 13 06:49:03 crc kubenswrapper[4971]: I1213 06:49:03.893819 4971 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/f614b9022728cf315e60c057852e563e-cert-dir\") pod \"kube-controller-manager-crc\" (UID: \"f614b9022728cf315e60c057852e563e\") " pod="openshift-kube-controller-manager/kube-controller-manager-crc" Dec 13 06:49:03 crc kubenswrapper[4971]: I1213 06:49:03.893883 4971 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/f614b9022728cf315e60c057852e563e-resource-dir\") pod \"kube-controller-manager-crc\" (UID: \"f614b9022728cf315e60c057852e563e\") " pod="openshift-kube-controller-manager/kube-controller-manager-crc" Dec 13 06:49:03 crc kubenswrapper[4971]: I1213 06:49:03.940025 4971 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Dec 13 06:49:03 crc kubenswrapper[4971]: I1213 06:49:03.941323 4971 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 13 06:49:03 crc kubenswrapper[4971]: I1213 06:49:03.941557 4971 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 13 06:49:03 crc kubenswrapper[4971]: I1213 06:49:03.941579 4971 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 13 06:49:03 crc kubenswrapper[4971]: I1213 06:49:03.941601 4971 kubelet_node_status.go:76] "Attempting to register node" node="crc" Dec 13 06:49:03 crc kubenswrapper[4971]: E1213 06:49:03.944858 4971 kubelet_node_status.go:99] "Unable to register node with API server" err="Post \"https://api-int.crc.testing:6443/api/v1/nodes\": dial tcp 38.102.83.195:6443: connect: connection refused" node="crc" Dec 13 06:49:03 crc kubenswrapper[4971]: I1213 06:49:03.995252 4971 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"usr-local-bin\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-usr-local-bin\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Dec 13 06:49:03 crc kubenswrapper[4971]: I1213 06:49:03.995312 4971 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"log-dir\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-log-dir\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Dec 13 06:49:03 crc kubenswrapper[4971]: I1213 06:49:03.995339 4971 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-audit-dir\") pod \"kube-apiserver-crc\" (UID: \"f4b27818a5e8e43d0dc095d08835c792\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Dec 13 06:49:03 crc kubenswrapper[4971]: I1213 06:49:03.995433 4971 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"data-dir\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-data-dir\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Dec 13 06:49:03 crc kubenswrapper[4971]: I1213 06:49:03.995455 4971 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-resource-dir\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Dec 13 06:49:03 crc kubenswrapper[4971]: I1213 06:49:03.995503 4971 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-cert-dir\") pod \"kube-apiserver-crc\" (UID: \"f4b27818a5e8e43d0dc095d08835c792\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Dec 13 06:49:03 crc kubenswrapper[4971]: I1213 06:49:03.995578 4971 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/3dcd261975c3d6b9a6ad6367fd4facd3-cert-dir\") pod \"openshift-kube-scheduler-crc\" (UID: \"3dcd261975c3d6b9a6ad6367fd4facd3\") " pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" Dec 13 06:49:03 crc kubenswrapper[4971]: I1213 06:49:03.995639 4971 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-kube\" (UniqueName: \"kubernetes.io/host-path/d1b160f5dda77d281dd8e69ec8d817f9-etc-kube\") pod \"kube-rbac-proxy-crio-crc\" (UID: \"d1b160f5dda77d281dd8e69ec8d817f9\") " pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" Dec 13 06:49:03 crc kubenswrapper[4971]: I1213 06:49:03.995658 4971 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-lib-kubelet\" (UniqueName: \"kubernetes.io/host-path/d1b160f5dda77d281dd8e69ec8d817f9-var-lib-kubelet\") pod \"kube-rbac-proxy-crio-crc\" (UID: \"d1b160f5dda77d281dd8e69ec8d817f9\") " pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" Dec 13 06:49:03 crc kubenswrapper[4971]: I1213 06:49:03.995707 4971 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"static-pod-dir\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-static-pod-dir\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Dec 13 06:49:03 crc kubenswrapper[4971]: I1213 06:49:03.995722 4971 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-cert-dir\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Dec 13 06:49:03 crc kubenswrapper[4971]: I1213 06:49:03.995924 4971 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/f614b9022728cf315e60c057852e563e-cert-dir\") pod \"kube-controller-manager-crc\" (UID: \"f614b9022728cf315e60c057852e563e\") " pod="openshift-kube-controller-manager/kube-controller-manager-crc" Dec 13 06:49:03 crc kubenswrapper[4971]: I1213 06:49:03.996047 4971 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-resource-dir\") pod \"kube-apiserver-crc\" (UID: \"f4b27818a5e8e43d0dc095d08835c792\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Dec 13 06:49:03 crc kubenswrapper[4971]: I1213 06:49:03.996103 4971 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/f614b9022728cf315e60c057852e563e-resource-dir\") pod \"kube-controller-manager-crc\" (UID: \"f614b9022728cf315e60c057852e563e\") " pod="openshift-kube-controller-manager/kube-controller-manager-crc" Dec 13 06:49:03 crc kubenswrapper[4971]: I1213 06:49:03.996134 4971 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/3dcd261975c3d6b9a6ad6367fd4facd3-resource-dir\") pod \"openshift-kube-scheduler-crc\" (UID: \"3dcd261975c3d6b9a6ad6367fd4facd3\") " pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" Dec 13 06:49:03 crc kubenswrapper[4971]: I1213 06:49:03.996199 4971 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/f614b9022728cf315e60c057852e563e-resource-dir\") pod \"kube-controller-manager-crc\" (UID: \"f614b9022728cf315e60c057852e563e\") " pod="openshift-kube-controller-manager/kube-controller-manager-crc" Dec 13 06:49:03 crc kubenswrapper[4971]: I1213 06:49:03.996183 4971 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/f614b9022728cf315e60c057852e563e-cert-dir\") pod \"kube-controller-manager-crc\" (UID: \"f614b9022728cf315e60c057852e563e\") " pod="openshift-kube-controller-manager/kube-controller-manager-crc" Dec 13 06:49:04 crc kubenswrapper[4971]: I1213 06:49:04.097701 4971 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-audit-dir\") pod \"kube-apiserver-crc\" (UID: \"f4b27818a5e8e43d0dc095d08835c792\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Dec 13 06:49:04 crc kubenswrapper[4971]: I1213 06:49:04.097780 4971 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"data-dir\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-data-dir\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Dec 13 06:49:04 crc kubenswrapper[4971]: I1213 06:49:04.097811 4971 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"usr-local-bin\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-usr-local-bin\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Dec 13 06:49:04 crc kubenswrapper[4971]: I1213 06:49:04.097838 4971 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"log-dir\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-log-dir\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Dec 13 06:49:04 crc kubenswrapper[4971]: I1213 06:49:04.097875 4971 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-resource-dir\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Dec 13 06:49:04 crc kubenswrapper[4971]: I1213 06:49:04.097900 4971 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-cert-dir\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Dec 13 06:49:04 crc kubenswrapper[4971]: I1213 06:49:04.097904 4971 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-audit-dir\") pod \"kube-apiserver-crc\" (UID: \"f4b27818a5e8e43d0dc095d08835c792\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Dec 13 06:49:04 crc kubenswrapper[4971]: I1213 06:49:04.098029 4971 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"log-dir\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-log-dir\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Dec 13 06:49:04 crc kubenswrapper[4971]: I1213 06:49:04.098062 4971 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-resource-dir\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Dec 13 06:49:04 crc kubenswrapper[4971]: I1213 06:49:04.097943 4971 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"data-dir\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-data-dir\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Dec 13 06:49:04 crc kubenswrapper[4971]: I1213 06:49:04.098091 4971 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"usr-local-bin\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-usr-local-bin\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Dec 13 06:49:04 crc kubenswrapper[4971]: I1213 06:49:04.097926 4971 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-resource-dir\") pod \"kube-apiserver-crc\" (UID: \"f4b27818a5e8e43d0dc095d08835c792\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Dec 13 06:49:04 crc kubenswrapper[4971]: I1213 06:49:04.098030 4971 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-resource-dir\") pod \"kube-apiserver-crc\" (UID: \"f4b27818a5e8e43d0dc095d08835c792\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Dec 13 06:49:04 crc kubenswrapper[4971]: I1213 06:49:04.098139 4971 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-cert-dir\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Dec 13 06:49:04 crc kubenswrapper[4971]: I1213 06:49:04.098187 4971 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-cert-dir\") pod \"kube-apiserver-crc\" (UID: \"f4b27818a5e8e43d0dc095d08835c792\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Dec 13 06:49:04 crc kubenswrapper[4971]: I1213 06:49:04.098225 4971 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/3dcd261975c3d6b9a6ad6367fd4facd3-cert-dir\") pod \"openshift-kube-scheduler-crc\" (UID: \"3dcd261975c3d6b9a6ad6367fd4facd3\") " pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" Dec 13 06:49:04 crc kubenswrapper[4971]: I1213 06:49:04.098239 4971 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-cert-dir\") pod \"kube-apiserver-crc\" (UID: \"f4b27818a5e8e43d0dc095d08835c792\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Dec 13 06:49:04 crc kubenswrapper[4971]: I1213 06:49:04.098294 4971 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-kube\" (UniqueName: \"kubernetes.io/host-path/d1b160f5dda77d281dd8e69ec8d817f9-etc-kube\") pod \"kube-rbac-proxy-crio-crc\" (UID: \"d1b160f5dda77d281dd8e69ec8d817f9\") " pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" Dec 13 06:49:04 crc kubenswrapper[4971]: I1213 06:49:04.098327 4971 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-lib-kubelet\" (UniqueName: \"kubernetes.io/host-path/d1b160f5dda77d281dd8e69ec8d817f9-var-lib-kubelet\") pod \"kube-rbac-proxy-crio-crc\" (UID: \"d1b160f5dda77d281dd8e69ec8d817f9\") " pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" Dec 13 06:49:04 crc kubenswrapper[4971]: I1213 06:49:04.098358 4971 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"static-pod-dir\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-static-pod-dir\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Dec 13 06:49:04 crc kubenswrapper[4971]: I1213 06:49:04.098376 4971 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-kube\" (UniqueName: \"kubernetes.io/host-path/d1b160f5dda77d281dd8e69ec8d817f9-etc-kube\") pod \"kube-rbac-proxy-crio-crc\" (UID: \"d1b160f5dda77d281dd8e69ec8d817f9\") " pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" Dec 13 06:49:04 crc kubenswrapper[4971]: I1213 06:49:04.098386 4971 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/3dcd261975c3d6b9a6ad6367fd4facd3-resource-dir\") pod \"openshift-kube-scheduler-crc\" (UID: \"3dcd261975c3d6b9a6ad6367fd4facd3\") " pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" Dec 13 06:49:04 crc kubenswrapper[4971]: I1213 06:49:04.098413 4971 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"static-pod-dir\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-static-pod-dir\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Dec 13 06:49:04 crc kubenswrapper[4971]: I1213 06:49:04.098450 4971 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-lib-kubelet\" (UniqueName: \"kubernetes.io/host-path/d1b160f5dda77d281dd8e69ec8d817f9-var-lib-kubelet\") pod \"kube-rbac-proxy-crio-crc\" (UID: \"d1b160f5dda77d281dd8e69ec8d817f9\") " pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" Dec 13 06:49:04 crc kubenswrapper[4971]: I1213 06:49:04.098359 4971 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/3dcd261975c3d6b9a6ad6367fd4facd3-cert-dir\") pod \"openshift-kube-scheduler-crc\" (UID: \"3dcd261975c3d6b9a6ad6367fd4facd3\") " pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" Dec 13 06:49:04 crc kubenswrapper[4971]: I1213 06:49:04.098582 4971 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/3dcd261975c3d6b9a6ad6367fd4facd3-resource-dir\") pod \"openshift-kube-scheduler-crc\" (UID: \"3dcd261975c3d6b9a6ad6367fd4facd3\") " pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" Dec 13 06:49:04 crc kubenswrapper[4971]: I1213 06:49:04.145308 4971 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Dec 13 06:49:04 crc kubenswrapper[4971]: I1213 06:49:04.147340 4971 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 13 06:49:04 crc kubenswrapper[4971]: I1213 06:49:04.147388 4971 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 13 06:49:04 crc kubenswrapper[4971]: I1213 06:49:04.147403 4971 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 13 06:49:04 crc kubenswrapper[4971]: I1213 06:49:04.147439 4971 kubelet_node_status.go:76] "Attempting to register node" node="crc" Dec 13 06:49:04 crc kubenswrapper[4971]: E1213 06:49:04.148194 4971 kubelet_node_status.go:99] "Unable to register node with API server" err="Post \"https://api-int.crc.testing:6443/api/v1/nodes\": dial tcp 38.102.83.195:6443: connect: connection refused" node="crc" Dec 13 06:49:04 crc kubenswrapper[4971]: I1213 06:49:04.198545 4971 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-controller-manager/kube-controller-manager-crc" Dec 13 06:49:04 crc kubenswrapper[4971]: I1213 06:49:04.209055 4971 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" Dec 13 06:49:04 crc kubenswrapper[4971]: I1213 06:49:04.229404 4971 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" Dec 13 06:49:04 crc kubenswrapper[4971]: W1213 06:49:04.231105 4971 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podf614b9022728cf315e60c057852e563e.slice/crio-f29809382af792275073094cf9a003100a4cd7c121bf83f19b5b8076c4f2d05e WatchSource:0}: Error finding container f29809382af792275073094cf9a003100a4cd7c121bf83f19b5b8076c4f2d05e: Status 404 returned error can't find the container with id f29809382af792275073094cf9a003100a4cd7c121bf83f19b5b8076c4f2d05e Dec 13 06:49:04 crc kubenswrapper[4971]: W1213 06:49:04.233231 4971 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod3dcd261975c3d6b9a6ad6367fd4facd3.slice/crio-b7894499abc8805103e3408ddfc3ddef2adbab63fa3fd91e203b02b7dce036e3 WatchSource:0}: Error finding container b7894499abc8805103e3408ddfc3ddef2adbab63fa3fd91e203b02b7dce036e3: Status 404 returned error can't find the container with id b7894499abc8805103e3408ddfc3ddef2adbab63fa3fd91e203b02b7dce036e3 Dec 13 06:49:04 crc kubenswrapper[4971]: W1213 06:49:04.244009 4971 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podd1b160f5dda77d281dd8e69ec8d817f9.slice/crio-8eeab8ca1c48852d415ff1a5a0a8e77dc3decac3a1517a8c02a61eed250b8533 WatchSource:0}: Error finding container 8eeab8ca1c48852d415ff1a5a0a8e77dc3decac3a1517a8c02a61eed250b8533: Status 404 returned error can't find the container with id 8eeab8ca1c48852d415ff1a5a0a8e77dc3decac3a1517a8c02a61eed250b8533 Dec 13 06:49:04 crc kubenswrapper[4971]: I1213 06:49:04.247233 4971 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-etcd/etcd-crc" Dec 13 06:49:04 crc kubenswrapper[4971]: I1213 06:49:04.255934 4971 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/kube-apiserver-crc" Dec 13 06:49:04 crc kubenswrapper[4971]: E1213 06:49:04.269382 4971 controller.go:145] "Failed to ensure lease exists, will retry" err="Get \"https://api-int.crc.testing:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/crc?timeout=10s\": dial tcp 38.102.83.195:6443: connect: connection refused" interval="800ms" Dec 13 06:49:04 crc kubenswrapper[4971]: W1213 06:49:04.365603 4971 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod2139d3e2895fc6797b9c76a1b4c9886d.slice/crio-f95892b1e225dcc3348caafc25961476d8236fb1db29e37af88bae93428dc029 WatchSource:0}: Error finding container f95892b1e225dcc3348caafc25961476d8236fb1db29e37af88bae93428dc029: Status 404 returned error can't find the container with id f95892b1e225dcc3348caafc25961476d8236fb1db29e37af88bae93428dc029 Dec 13 06:49:04 crc kubenswrapper[4971]: W1213 06:49:04.381219 4971 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podf4b27818a5e8e43d0dc095d08835c792.slice/crio-e07b2c5918de4d0d6a8a7006b52797c13cf6c419652877d38f7e6095a871f94e WatchSource:0}: Error finding container e07b2c5918de4d0d6a8a7006b52797c13cf6c419652877d38f7e6095a871f94e: Status 404 returned error can't find the container with id e07b2c5918de4d0d6a8a7006b52797c13cf6c419652877d38f7e6095a871f94e Dec 13 06:49:04 crc kubenswrapper[4971]: I1213 06:49:04.548975 4971 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Dec 13 06:49:04 crc kubenswrapper[4971]: I1213 06:49:04.550738 4971 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 13 06:49:04 crc kubenswrapper[4971]: I1213 06:49:04.550783 4971 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 13 06:49:04 crc kubenswrapper[4971]: I1213 06:49:04.550793 4971 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 13 06:49:04 crc kubenswrapper[4971]: I1213 06:49:04.550819 4971 kubelet_node_status.go:76] "Attempting to register node" node="crc" Dec 13 06:49:04 crc kubenswrapper[4971]: E1213 06:49:04.551320 4971 kubelet_node_status.go:99] "Unable to register node with API server" err="Post \"https://api-int.crc.testing:6443/api/v1/nodes\": dial tcp 38.102.83.195:6443: connect: connection refused" node="crc" Dec 13 06:49:04 crc kubenswrapper[4971]: I1213 06:49:04.665247 4971 csi_plugin.go:884] Failed to contact API server when waiting for CSINode publishing: Get "https://api-int.crc.testing:6443/apis/storage.k8s.io/v1/csinodes/crc?resourceVersion=0": dial tcp 38.102.83.195:6443: connect: connection refused Dec 13 06:49:04 crc kubenswrapper[4971]: I1213 06:49:04.667485 4971 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2025-11-09 14:14:12.625351155 +0000 UTC Dec 13 06:49:04 crc kubenswrapper[4971]: W1213 06:49:04.678751 4971 reflector.go:561] k8s.io/client-go/informers/factory.go:160: failed to list *v1.Node: Get "https://api-int.crc.testing:6443/api/v1/nodes?fieldSelector=metadata.name%3Dcrc&limit=500&resourceVersion=0": dial tcp 38.102.83.195:6443: connect: connection refused Dec 13 06:49:04 crc kubenswrapper[4971]: E1213 06:49:04.678874 4971 reflector.go:158] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.Node: failed to list *v1.Node: Get \"https://api-int.crc.testing:6443/api/v1/nodes?fieldSelector=metadata.name%3Dcrc&limit=500&resourceVersion=0\": dial tcp 38.102.83.195:6443: connect: connection refused" logger="UnhandledError" Dec 13 06:49:04 crc kubenswrapper[4971]: I1213 06:49:04.772884 4971 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-etcd/etcd-crc" event={"ID":"2139d3e2895fc6797b9c76a1b4c9886d","Type":"ContainerStarted","Data":"f95892b1e225dcc3348caafc25961476d8236fb1db29e37af88bae93428dc029"} Dec 13 06:49:04 crc kubenswrapper[4971]: I1213 06:49:04.774167 4971 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" event={"ID":"d1b160f5dda77d281dd8e69ec8d817f9","Type":"ContainerStarted","Data":"8eeab8ca1c48852d415ff1a5a0a8e77dc3decac3a1517a8c02a61eed250b8533"} Dec 13 06:49:04 crc kubenswrapper[4971]: I1213 06:49:04.776008 4971 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" event={"ID":"3dcd261975c3d6b9a6ad6367fd4facd3","Type":"ContainerStarted","Data":"38806c430185715371ab9bf30a1af9a9d5d492abd9bb84acbdccc694f56e3a8a"} Dec 13 06:49:04 crc kubenswrapper[4971]: I1213 06:49:04.776045 4971 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" event={"ID":"3dcd261975c3d6b9a6ad6367fd4facd3","Type":"ContainerStarted","Data":"b7894499abc8805103e3408ddfc3ddef2adbab63fa3fd91e203b02b7dce036e3"} Dec 13 06:49:04 crc kubenswrapper[4971]: I1213 06:49:04.777289 4971 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" event={"ID":"f614b9022728cf315e60c057852e563e","Type":"ContainerStarted","Data":"c788de86754f0bf37ee184672fc1dcb388ccebc246d4f15be4b5da661f725f8f"} Dec 13 06:49:04 crc kubenswrapper[4971]: I1213 06:49:04.777353 4971 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" event={"ID":"f614b9022728cf315e60c057852e563e","Type":"ContainerStarted","Data":"f29809382af792275073094cf9a003100a4cd7c121bf83f19b5b8076c4f2d05e"} Dec 13 06:49:04 crc kubenswrapper[4971]: I1213 06:49:04.778399 4971 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"f4b27818a5e8e43d0dc095d08835c792","Type":"ContainerStarted","Data":"e07b2c5918de4d0d6a8a7006b52797c13cf6c419652877d38f7e6095a871f94e"} Dec 13 06:49:04 crc kubenswrapper[4971]: W1213 06:49:04.853322 4971 reflector.go:561] k8s.io/client-go/informers/factory.go:160: failed to list *v1.Service: Get "https://api-int.crc.testing:6443/api/v1/services?fieldSelector=spec.clusterIP%21%3DNone&limit=500&resourceVersion=0": dial tcp 38.102.83.195:6443: connect: connection refused Dec 13 06:49:04 crc kubenswrapper[4971]: E1213 06:49:04.853411 4971 reflector.go:158] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.Service: failed to list *v1.Service: Get \"https://api-int.crc.testing:6443/api/v1/services?fieldSelector=spec.clusterIP%21%3DNone&limit=500&resourceVersion=0\": dial tcp 38.102.83.195:6443: connect: connection refused" logger="UnhandledError" Dec 13 06:49:05 crc kubenswrapper[4971]: E1213 06:49:05.070303 4971 controller.go:145] "Failed to ensure lease exists, will retry" err="Get \"https://api-int.crc.testing:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/crc?timeout=10s\": dial tcp 38.102.83.195:6443: connect: connection refused" interval="1.6s" Dec 13 06:49:05 crc kubenswrapper[4971]: W1213 06:49:05.079255 4971 reflector.go:561] k8s.io/client-go/informers/factory.go:160: failed to list *v1.CSIDriver: Get "https://api-int.crc.testing:6443/apis/storage.k8s.io/v1/csidrivers?limit=500&resourceVersion=0": dial tcp 38.102.83.195:6443: connect: connection refused Dec 13 06:49:05 crc kubenswrapper[4971]: E1213 06:49:05.079350 4971 reflector.go:158] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.CSIDriver: failed to list *v1.CSIDriver: Get \"https://api-int.crc.testing:6443/apis/storage.k8s.io/v1/csidrivers?limit=500&resourceVersion=0\": dial tcp 38.102.83.195:6443: connect: connection refused" logger="UnhandledError" Dec 13 06:49:05 crc kubenswrapper[4971]: W1213 06:49:05.344476 4971 reflector.go:561] k8s.io/client-go/informers/factory.go:160: failed to list *v1.RuntimeClass: Get "https://api-int.crc.testing:6443/apis/node.k8s.io/v1/runtimeclasses?limit=500&resourceVersion=0": dial tcp 38.102.83.195:6443: connect: connection refused Dec 13 06:49:05 crc kubenswrapper[4971]: E1213 06:49:05.344578 4971 reflector.go:158] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.RuntimeClass: failed to list *v1.RuntimeClass: Get \"https://api-int.crc.testing:6443/apis/node.k8s.io/v1/runtimeclasses?limit=500&resourceVersion=0\": dial tcp 38.102.83.195:6443: connect: connection refused" logger="UnhandledError" Dec 13 06:49:05 crc kubenswrapper[4971]: I1213 06:49:05.351957 4971 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Dec 13 06:49:05 crc kubenswrapper[4971]: I1213 06:49:05.353440 4971 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 13 06:49:05 crc kubenswrapper[4971]: I1213 06:49:05.353500 4971 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 13 06:49:05 crc kubenswrapper[4971]: I1213 06:49:05.353529 4971 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 13 06:49:05 crc kubenswrapper[4971]: I1213 06:49:05.353561 4971 kubelet_node_status.go:76] "Attempting to register node" node="crc" Dec 13 06:49:05 crc kubenswrapper[4971]: E1213 06:49:05.354058 4971 kubelet_node_status.go:99] "Unable to register node with API server" err="Post \"https://api-int.crc.testing:6443/api/v1/nodes\": dial tcp 38.102.83.195:6443: connect: connection refused" node="crc" Dec 13 06:49:05 crc kubenswrapper[4971]: I1213 06:49:05.665840 4971 csi_plugin.go:884] Failed to contact API server when waiting for CSINode publishing: Get "https://api-int.crc.testing:6443/apis/storage.k8s.io/v1/csinodes/crc?resourceVersion=0": dial tcp 38.102.83.195:6443: connect: connection refused Dec 13 06:49:05 crc kubenswrapper[4971]: I1213 06:49:05.668533 4971 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2025-12-02 21:11:42.365475906 +0000 UTC Dec 13 06:49:05 crc kubenswrapper[4971]: I1213 06:49:05.782313 4971 generic.go:334] "Generic (PLEG): container finished" podID="3dcd261975c3d6b9a6ad6367fd4facd3" containerID="38806c430185715371ab9bf30a1af9a9d5d492abd9bb84acbdccc694f56e3a8a" exitCode=0 Dec 13 06:49:05 crc kubenswrapper[4971]: I1213 06:49:05.782389 4971 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Dec 13 06:49:05 crc kubenswrapper[4971]: I1213 06:49:05.782421 4971 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" event={"ID":"3dcd261975c3d6b9a6ad6367fd4facd3","Type":"ContainerDied","Data":"38806c430185715371ab9bf30a1af9a9d5d492abd9bb84acbdccc694f56e3a8a"} Dec 13 06:49:05 crc kubenswrapper[4971]: I1213 06:49:05.783242 4971 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 13 06:49:05 crc kubenswrapper[4971]: I1213 06:49:05.783275 4971 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 13 06:49:05 crc kubenswrapper[4971]: I1213 06:49:05.783287 4971 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 13 06:49:05 crc kubenswrapper[4971]: I1213 06:49:05.784644 4971 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" event={"ID":"f614b9022728cf315e60c057852e563e","Type":"ContainerStarted","Data":"760817511a9b4f96a68dc38d26631250eaab72e1dd8f123e653039b6b4e91f84"} Dec 13 06:49:05 crc kubenswrapper[4971]: I1213 06:49:05.784676 4971 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" event={"ID":"f614b9022728cf315e60c057852e563e","Type":"ContainerStarted","Data":"c9baa042f9dada954faa57b5454cd6b0a12a297e56ddc192c52cff50e42a16fa"} Dec 13 06:49:05 crc kubenswrapper[4971]: I1213 06:49:05.786071 4971 generic.go:334] "Generic (PLEG): container finished" podID="f4b27818a5e8e43d0dc095d08835c792" containerID="38c91cfbaecfdbd3e3b663422a91ab4294610afb6bbf9ebdb77db24e2b76c682" exitCode=0 Dec 13 06:49:05 crc kubenswrapper[4971]: I1213 06:49:05.786133 4971 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"f4b27818a5e8e43d0dc095d08835c792","Type":"ContainerDied","Data":"38c91cfbaecfdbd3e3b663422a91ab4294610afb6bbf9ebdb77db24e2b76c682"} Dec 13 06:49:05 crc kubenswrapper[4971]: I1213 06:49:05.786205 4971 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Dec 13 06:49:05 crc kubenswrapper[4971]: I1213 06:49:05.787079 4971 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 13 06:49:05 crc kubenswrapper[4971]: I1213 06:49:05.787106 4971 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 13 06:49:05 crc kubenswrapper[4971]: I1213 06:49:05.787117 4971 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 13 06:49:05 crc kubenswrapper[4971]: I1213 06:49:05.787480 4971 generic.go:334] "Generic (PLEG): container finished" podID="2139d3e2895fc6797b9c76a1b4c9886d" containerID="7d2746335c98dae3a0f43024d34cb62ebcb42a01c8f9ce18d52b29d4c45598e5" exitCode=0 Dec 13 06:49:05 crc kubenswrapper[4971]: I1213 06:49:05.787555 4971 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-etcd/etcd-crc" event={"ID":"2139d3e2895fc6797b9c76a1b4c9886d","Type":"ContainerDied","Data":"7d2746335c98dae3a0f43024d34cb62ebcb42a01c8f9ce18d52b29d4c45598e5"} Dec 13 06:49:05 crc kubenswrapper[4971]: I1213 06:49:05.787603 4971 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Dec 13 06:49:05 crc kubenswrapper[4971]: I1213 06:49:05.788182 4971 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 13 06:49:05 crc kubenswrapper[4971]: I1213 06:49:05.788202 4971 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 13 06:49:05 crc kubenswrapper[4971]: I1213 06:49:05.788211 4971 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 13 06:49:05 crc kubenswrapper[4971]: I1213 06:49:05.788428 4971 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Dec 13 06:49:05 crc kubenswrapper[4971]: I1213 06:49:05.788983 4971 generic.go:334] "Generic (PLEG): container finished" podID="d1b160f5dda77d281dd8e69ec8d817f9" containerID="f0d864149bfcbb6227e6a6fbde23fff650d688850978eae0caf04515eb21cd85" exitCode=0 Dec 13 06:49:05 crc kubenswrapper[4971]: I1213 06:49:05.789023 4971 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Dec 13 06:49:05 crc kubenswrapper[4971]: I1213 06:49:05.789156 4971 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 13 06:49:05 crc kubenswrapper[4971]: I1213 06:49:05.789170 4971 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 13 06:49:05 crc kubenswrapper[4971]: I1213 06:49:05.789178 4971 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 13 06:49:05 crc kubenswrapper[4971]: I1213 06:49:05.789019 4971 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" event={"ID":"d1b160f5dda77d281dd8e69ec8d817f9","Type":"ContainerDied","Data":"f0d864149bfcbb6227e6a6fbde23fff650d688850978eae0caf04515eb21cd85"} Dec 13 06:49:05 crc kubenswrapper[4971]: I1213 06:49:05.789637 4971 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 13 06:49:05 crc kubenswrapper[4971]: I1213 06:49:05.789660 4971 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 13 06:49:05 crc kubenswrapper[4971]: I1213 06:49:05.789668 4971 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 13 06:49:06 crc kubenswrapper[4971]: I1213 06:49:06.665940 4971 csi_plugin.go:884] Failed to contact API server when waiting for CSINode publishing: Get "https://api-int.crc.testing:6443/apis/storage.k8s.io/v1/csinodes/crc?resourceVersion=0": dial tcp 38.102.83.195:6443: connect: connection refused Dec 13 06:49:06 crc kubenswrapper[4971]: I1213 06:49:06.669170 4971 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2025-11-08 02:28:48.35197863 +0000 UTC Dec 13 06:49:06 crc kubenswrapper[4971]: E1213 06:49:06.671759 4971 controller.go:145] "Failed to ensure lease exists, will retry" err="Get \"https://api-int.crc.testing:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/crc?timeout=10s\": dial tcp 38.102.83.195:6443: connect: connection refused" interval="3.2s" Dec 13 06:49:06 crc kubenswrapper[4971]: E1213 06:49:06.726595 4971 event.go:368] "Unable to write event (may retry after sleeping)" err="Post \"https://api-int.crc.testing:6443/api/v1/namespaces/default/events\": dial tcp 38.102.83.195:6443: connect: connection refused" event="&Event{ObjectMeta:{crc.1880b39da97e1d7c default 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Node,Namespace:,Name:crc,UID:crc,APIVersion:,ResourceVersion:,FieldPath:,},Reason:Starting,Message:Starting kubelet.,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2025-12-13 06:49:03.664086396 +0000 UTC m=+0.268495844,LastTimestamp:2025-12-13 06:49:03.664086396 +0000 UTC m=+0.268495844,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Dec 13 06:49:06 crc kubenswrapper[4971]: I1213 06:49:06.793188 4971 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" event={"ID":"d1b160f5dda77d281dd8e69ec8d817f9","Type":"ContainerStarted","Data":"35b737b0d2c5c5503ba2819f34d2de0599792a77beb1fbeb561d44efd1f212c5"} Dec 13 06:49:06 crc kubenswrapper[4971]: I1213 06:49:06.793265 4971 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Dec 13 06:49:06 crc kubenswrapper[4971]: I1213 06:49:06.794226 4971 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 13 06:49:06 crc kubenswrapper[4971]: I1213 06:49:06.794270 4971 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 13 06:49:06 crc kubenswrapper[4971]: I1213 06:49:06.794282 4971 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 13 06:49:06 crc kubenswrapper[4971]: I1213 06:49:06.795193 4971 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" event={"ID":"3dcd261975c3d6b9a6ad6367fd4facd3","Type":"ContainerStarted","Data":"9e61d25add1723ac286794abfe886358e9d64455c25379f7e3057fc3dd812418"} Dec 13 06:49:06 crc kubenswrapper[4971]: I1213 06:49:06.797330 4971 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Dec 13 06:49:06 crc kubenswrapper[4971]: I1213 06:49:06.797331 4971 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" event={"ID":"f614b9022728cf315e60c057852e563e","Type":"ContainerStarted","Data":"b859be0c9feac856e9855c6a038b02b49b32fade97873c46d0bec0a32d8b5147"} Dec 13 06:49:06 crc kubenswrapper[4971]: I1213 06:49:06.797896 4971 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 13 06:49:06 crc kubenswrapper[4971]: I1213 06:49:06.797928 4971 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 13 06:49:06 crc kubenswrapper[4971]: I1213 06:49:06.797941 4971 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 13 06:49:06 crc kubenswrapper[4971]: I1213 06:49:06.798374 4971 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"f4b27818a5e8e43d0dc095d08835c792","Type":"ContainerStarted","Data":"3716afdf2bc693be80a09fc9f7727d334f435caf1d4c8ec4811b3a1f1652cdfe"} Dec 13 06:49:06 crc kubenswrapper[4971]: I1213 06:49:06.799940 4971 generic.go:334] "Generic (PLEG): container finished" podID="2139d3e2895fc6797b9c76a1b4c9886d" containerID="bc0e77d10a738128bea82122496bc42235fc1688f8907756d93e61b651902b57" exitCode=0 Dec 13 06:49:06 crc kubenswrapper[4971]: I1213 06:49:06.799981 4971 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-etcd/etcd-crc" event={"ID":"2139d3e2895fc6797b9c76a1b4c9886d","Type":"ContainerDied","Data":"bc0e77d10a738128bea82122496bc42235fc1688f8907756d93e61b651902b57"} Dec 13 06:49:06 crc kubenswrapper[4971]: I1213 06:49:06.800107 4971 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Dec 13 06:49:06 crc kubenswrapper[4971]: I1213 06:49:06.801097 4971 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 13 06:49:06 crc kubenswrapper[4971]: I1213 06:49:06.801157 4971 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 13 06:49:06 crc kubenswrapper[4971]: I1213 06:49:06.801175 4971 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 13 06:49:06 crc kubenswrapper[4971]: I1213 06:49:06.954601 4971 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Dec 13 06:49:06 crc kubenswrapper[4971]: I1213 06:49:06.955967 4971 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 13 06:49:06 crc kubenswrapper[4971]: I1213 06:49:06.956013 4971 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 13 06:49:06 crc kubenswrapper[4971]: I1213 06:49:06.956026 4971 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 13 06:49:06 crc kubenswrapper[4971]: I1213 06:49:06.956054 4971 kubelet_node_status.go:76] "Attempting to register node" node="crc" Dec 13 06:49:06 crc kubenswrapper[4971]: E1213 06:49:06.956564 4971 kubelet_node_status.go:99] "Unable to register node with API server" err="Post \"https://api-int.crc.testing:6443/api/v1/nodes\": dial tcp 38.102.83.195:6443: connect: connection refused" node="crc" Dec 13 06:49:07 crc kubenswrapper[4971]: W1213 06:49:07.208066 4971 reflector.go:561] k8s.io/client-go/informers/factory.go:160: failed to list *v1.RuntimeClass: Get "https://api-int.crc.testing:6443/apis/node.k8s.io/v1/runtimeclasses?limit=500&resourceVersion=0": dial tcp 38.102.83.195:6443: connect: connection refused Dec 13 06:49:07 crc kubenswrapper[4971]: E1213 06:49:07.208181 4971 reflector.go:158] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.RuntimeClass: failed to list *v1.RuntimeClass: Get \"https://api-int.crc.testing:6443/apis/node.k8s.io/v1/runtimeclasses?limit=500&resourceVersion=0\": dial tcp 38.102.83.195:6443: connect: connection refused" logger="UnhandledError" Dec 13 06:49:07 crc kubenswrapper[4971]: W1213 06:49:07.431578 4971 reflector.go:561] k8s.io/client-go/informers/factory.go:160: failed to list *v1.CSIDriver: Get "https://api-int.crc.testing:6443/apis/storage.k8s.io/v1/csidrivers?limit=500&resourceVersion=0": dial tcp 38.102.83.195:6443: connect: connection refused Dec 13 06:49:07 crc kubenswrapper[4971]: E1213 06:49:07.431654 4971 reflector.go:158] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.CSIDriver: failed to list *v1.CSIDriver: Get \"https://api-int.crc.testing:6443/apis/storage.k8s.io/v1/csidrivers?limit=500&resourceVersion=0\": dial tcp 38.102.83.195:6443: connect: connection refused" logger="UnhandledError" Dec 13 06:49:07 crc kubenswrapper[4971]: W1213 06:49:07.567170 4971 reflector.go:561] k8s.io/client-go/informers/factory.go:160: failed to list *v1.Service: Get "https://api-int.crc.testing:6443/api/v1/services?fieldSelector=spec.clusterIP%21%3DNone&limit=500&resourceVersion=0": dial tcp 38.102.83.195:6443: connect: connection refused Dec 13 06:49:07 crc kubenswrapper[4971]: E1213 06:49:07.567257 4971 reflector.go:158] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.Service: failed to list *v1.Service: Get \"https://api-int.crc.testing:6443/api/v1/services?fieldSelector=spec.clusterIP%21%3DNone&limit=500&resourceVersion=0\": dial tcp 38.102.83.195:6443: connect: connection refused" logger="UnhandledError" Dec 13 06:49:07 crc kubenswrapper[4971]: I1213 06:49:07.665701 4971 csi_plugin.go:884] Failed to contact API server when waiting for CSINode publishing: Get "https://api-int.crc.testing:6443/apis/storage.k8s.io/v1/csinodes/crc?resourceVersion=0": dial tcp 38.102.83.195:6443: connect: connection refused Dec 13 06:49:07 crc kubenswrapper[4971]: I1213 06:49:07.669981 4971 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2025-12-28 10:58:22.353379365 +0000 UTC Dec 13 06:49:07 crc kubenswrapper[4971]: I1213 06:49:07.670012 4971 certificate_manager.go:356] kubernetes.io/kubelet-serving: Waiting 364h9m14.683369816s for next certificate rotation Dec 13 06:49:07 crc kubenswrapper[4971]: I1213 06:49:07.704291 4971 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-kube-controller-manager/kube-controller-manager-crc" Dec 13 06:49:07 crc kubenswrapper[4971]: W1213 06:49:07.762258 4971 reflector.go:561] k8s.io/client-go/informers/factory.go:160: failed to list *v1.Node: Get "https://api-int.crc.testing:6443/api/v1/nodes?fieldSelector=metadata.name%3Dcrc&limit=500&resourceVersion=0": dial tcp 38.102.83.195:6443: connect: connection refused Dec 13 06:49:07 crc kubenswrapper[4971]: E1213 06:49:07.762387 4971 reflector.go:158] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.Node: failed to list *v1.Node: Get \"https://api-int.crc.testing:6443/api/v1/nodes?fieldSelector=metadata.name%3Dcrc&limit=500&resourceVersion=0\": dial tcp 38.102.83.195:6443: connect: connection refused" logger="UnhandledError" Dec 13 06:49:07 crc kubenswrapper[4971]: I1213 06:49:07.804617 4971 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"f4b27818a5e8e43d0dc095d08835c792","Type":"ContainerStarted","Data":"30df3abf8c79db8af52c6a918f73e1ce410ddafcf341636424c267e1eeb195a7"} Dec 13 06:49:07 crc kubenswrapper[4971]: I1213 06:49:07.804690 4971 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"f4b27818a5e8e43d0dc095d08835c792","Type":"ContainerStarted","Data":"ccdf65f6d2b1eb3647a03ba37c305c32e34addfdf79ff937c8dddaa3e5d5b5c6"} Dec 13 06:49:07 crc kubenswrapper[4971]: I1213 06:49:07.807145 4971 generic.go:334] "Generic (PLEG): container finished" podID="2139d3e2895fc6797b9c76a1b4c9886d" containerID="f821e418cba5e832da8839268bcbbd9630bdf319a212d5382bfa2c4b0b73a268" exitCode=0 Dec 13 06:49:07 crc kubenswrapper[4971]: I1213 06:49:07.807236 4971 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Dec 13 06:49:07 crc kubenswrapper[4971]: I1213 06:49:07.807236 4971 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-etcd/etcd-crc" event={"ID":"2139d3e2895fc6797b9c76a1b4c9886d","Type":"ContainerDied","Data":"f821e418cba5e832da8839268bcbbd9630bdf319a212d5382bfa2c4b0b73a268"} Dec 13 06:49:07 crc kubenswrapper[4971]: I1213 06:49:07.807973 4971 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 13 06:49:07 crc kubenswrapper[4971]: I1213 06:49:07.808022 4971 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 13 06:49:07 crc kubenswrapper[4971]: I1213 06:49:07.808047 4971 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 13 06:49:07 crc kubenswrapper[4971]: I1213 06:49:07.810251 4971 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Dec 13 06:49:07 crc kubenswrapper[4971]: I1213 06:49:07.810335 4971 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Dec 13 06:49:07 crc kubenswrapper[4971]: I1213 06:49:07.810470 4971 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" event={"ID":"3dcd261975c3d6b9a6ad6367fd4facd3","Type":"ContainerStarted","Data":"1c35f5e8d8fab6d77b9e2e0dedb195440de5d6af345cd5865abd32caa7fafd60"} Dec 13 06:49:07 crc kubenswrapper[4971]: I1213 06:49:07.810501 4971 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" event={"ID":"3dcd261975c3d6b9a6ad6367fd4facd3","Type":"ContainerStarted","Data":"7fb51d8435b4e3ce0411cae1d6c2824c552f0a33d79de01f969274516b39a5b3"} Dec 13 06:49:07 crc kubenswrapper[4971]: I1213 06:49:07.810577 4971 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Dec 13 06:49:07 crc kubenswrapper[4971]: I1213 06:49:07.812810 4971 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 13 06:49:07 crc kubenswrapper[4971]: I1213 06:49:07.812856 4971 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 13 06:49:07 crc kubenswrapper[4971]: I1213 06:49:07.812868 4971 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 13 06:49:07 crc kubenswrapper[4971]: I1213 06:49:07.815190 4971 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 13 06:49:07 crc kubenswrapper[4971]: I1213 06:49:07.815256 4971 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 13 06:49:07 crc kubenswrapper[4971]: I1213 06:49:07.815284 4971 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 13 06:49:07 crc kubenswrapper[4971]: I1213 06:49:07.815681 4971 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 13 06:49:07 crc kubenswrapper[4971]: I1213 06:49:07.815721 4971 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 13 06:49:07 crc kubenswrapper[4971]: I1213 06:49:07.815738 4971 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 13 06:49:08 crc kubenswrapper[4971]: I1213 06:49:08.824029 4971 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Dec 13 06:49:08 crc kubenswrapper[4971]: I1213 06:49:08.824080 4971 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"f4b27818a5e8e43d0dc095d08835c792","Type":"ContainerStarted","Data":"b3c1bab3d53f975b2e24582ce2771bc26d600e0dd7c30339dcfd18a071162ad4"} Dec 13 06:49:08 crc kubenswrapper[4971]: I1213 06:49:08.824155 4971 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"f4b27818a5e8e43d0dc095d08835c792","Type":"ContainerStarted","Data":"9acd07066deed483bd7fa61ee45fda09539d110717a134a56827f871186f92fc"} Dec 13 06:49:08 crc kubenswrapper[4971]: I1213 06:49:08.825217 4971 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 13 06:49:08 crc kubenswrapper[4971]: I1213 06:49:08.825250 4971 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 13 06:49:08 crc kubenswrapper[4971]: I1213 06:49:08.825261 4971 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 13 06:49:08 crc kubenswrapper[4971]: I1213 06:49:08.830658 4971 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Dec 13 06:49:08 crc kubenswrapper[4971]: I1213 06:49:08.830845 4971 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Dec 13 06:49:08 crc kubenswrapper[4971]: I1213 06:49:08.830695 4971 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-etcd/etcd-crc" event={"ID":"2139d3e2895fc6797b9c76a1b4c9886d","Type":"ContainerStarted","Data":"2c438cc74d555f7c9db4569d1c56b78b1088ed8105697d2c0f31e6c5136f3dd0"} Dec 13 06:49:08 crc kubenswrapper[4971]: I1213 06:49:08.830946 4971 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-etcd/etcd-crc" event={"ID":"2139d3e2895fc6797b9c76a1b4c9886d","Type":"ContainerStarted","Data":"eb014de8140deef097d9bd94f3b9fa4ab3558bc4449bef84b83bf62bfabf5836"} Dec 13 06:49:08 crc kubenswrapper[4971]: I1213 06:49:08.830980 4971 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" Dec 13 06:49:08 crc kubenswrapper[4971]: I1213 06:49:08.831039 4971 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-etcd/etcd-crc" event={"ID":"2139d3e2895fc6797b9c76a1b4c9886d","Type":"ContainerStarted","Data":"e457329a78a2d24337f994e8a0a1f75a2a667244aa75eb84613d1e7138364aac"} Dec 13 06:49:08 crc kubenswrapper[4971]: I1213 06:49:08.831049 4971 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-etcd/etcd-crc" event={"ID":"2139d3e2895fc6797b9c76a1b4c9886d","Type":"ContainerStarted","Data":"c4471b98854c52d5cd084d0ba3e64e5b08e9518ec3c372cce2c875a440522457"} Dec 13 06:49:08 crc kubenswrapper[4971]: I1213 06:49:08.831907 4971 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 13 06:49:08 crc kubenswrapper[4971]: I1213 06:49:08.831944 4971 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 13 06:49:08 crc kubenswrapper[4971]: I1213 06:49:08.831956 4971 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 13 06:49:08 crc kubenswrapper[4971]: I1213 06:49:08.832350 4971 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 13 06:49:08 crc kubenswrapper[4971]: I1213 06:49:08.832388 4971 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 13 06:49:08 crc kubenswrapper[4971]: I1213 06:49:08.832401 4971 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 13 06:49:09 crc kubenswrapper[4971]: I1213 06:49:09.150176 4971 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-kube-apiserver/kube-apiserver-crc" Dec 13 06:49:09 crc kubenswrapper[4971]: I1213 06:49:09.842849 4971 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-etcd/etcd-crc" event={"ID":"2139d3e2895fc6797b9c76a1b4c9886d","Type":"ContainerStarted","Data":"aa6fa6e4340dcd066252df3c56d79a3a39ed531d5eea8d1d5941d6952368d3af"} Dec 13 06:49:09 crc kubenswrapper[4971]: I1213 06:49:09.842921 4971 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Dec 13 06:49:09 crc kubenswrapper[4971]: I1213 06:49:09.843000 4971 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Dec 13 06:49:09 crc kubenswrapper[4971]: I1213 06:49:09.843745 4971 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-kube-apiserver/kube-apiserver-crc" Dec 13 06:49:09 crc kubenswrapper[4971]: I1213 06:49:09.843747 4971 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Dec 13 06:49:09 crc kubenswrapper[4971]: I1213 06:49:09.843821 4971 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 13 06:49:09 crc kubenswrapper[4971]: I1213 06:49:09.843821 4971 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 13 06:49:09 crc kubenswrapper[4971]: I1213 06:49:09.843851 4971 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 13 06:49:09 crc kubenswrapper[4971]: I1213 06:49:09.843865 4971 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 13 06:49:09 crc kubenswrapper[4971]: I1213 06:49:09.843864 4971 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 13 06:49:09 crc kubenswrapper[4971]: I1213 06:49:09.843899 4971 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 13 06:49:09 crc kubenswrapper[4971]: I1213 06:49:09.848033 4971 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 13 06:49:09 crc kubenswrapper[4971]: I1213 06:49:09.848103 4971 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 13 06:49:09 crc kubenswrapper[4971]: I1213 06:49:09.848122 4971 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 13 06:49:10 crc kubenswrapper[4971]: I1213 06:49:10.157002 4971 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Dec 13 06:49:10 crc kubenswrapper[4971]: I1213 06:49:10.158218 4971 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 13 06:49:10 crc kubenswrapper[4971]: I1213 06:49:10.158249 4971 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 13 06:49:10 crc kubenswrapper[4971]: I1213 06:49:10.158262 4971 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 13 06:49:10 crc kubenswrapper[4971]: I1213 06:49:10.158287 4971 kubelet_node_status.go:76] "Attempting to register node" node="crc" Dec 13 06:49:10 crc kubenswrapper[4971]: I1213 06:49:10.558450 4971 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-kube-apiserver/kube-apiserver-crc" Dec 13 06:49:10 crc kubenswrapper[4971]: I1213 06:49:10.844773 4971 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Dec 13 06:49:10 crc kubenswrapper[4971]: I1213 06:49:10.844849 4971 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Dec 13 06:49:10 crc kubenswrapper[4971]: I1213 06:49:10.845929 4971 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 13 06:49:10 crc kubenswrapper[4971]: I1213 06:49:10.845973 4971 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 13 06:49:10 crc kubenswrapper[4971]: I1213 06:49:10.845985 4971 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 13 06:49:10 crc kubenswrapper[4971]: I1213 06:49:10.846096 4971 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 13 06:49:10 crc kubenswrapper[4971]: I1213 06:49:10.846177 4971 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 13 06:49:10 crc kubenswrapper[4971]: I1213 06:49:10.846195 4971 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 13 06:49:11 crc kubenswrapper[4971]: I1213 06:49:11.559629 4971 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-kube-controller-manager/kube-controller-manager-crc" Dec 13 06:49:11 crc kubenswrapper[4971]: I1213 06:49:11.559849 4971 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Dec 13 06:49:11 crc kubenswrapper[4971]: I1213 06:49:11.561043 4971 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 13 06:49:11 crc kubenswrapper[4971]: I1213 06:49:11.561080 4971 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 13 06:49:11 crc kubenswrapper[4971]: I1213 06:49:11.561092 4971 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 13 06:49:11 crc kubenswrapper[4971]: I1213 06:49:11.847609 4971 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Dec 13 06:49:11 crc kubenswrapper[4971]: I1213 06:49:11.848501 4971 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 13 06:49:11 crc kubenswrapper[4971]: I1213 06:49:11.848571 4971 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 13 06:49:11 crc kubenswrapper[4971]: I1213 06:49:11.848582 4971 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 13 06:49:13 crc kubenswrapper[4971]: I1213 06:49:13.802608 4971 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-etcd/etcd-crc" Dec 13 06:49:13 crc kubenswrapper[4971]: I1213 06:49:13.802775 4971 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Dec 13 06:49:13 crc kubenswrapper[4971]: I1213 06:49:13.803749 4971 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 13 06:49:13 crc kubenswrapper[4971]: I1213 06:49:13.803793 4971 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 13 06:49:13 crc kubenswrapper[4971]: I1213 06:49:13.803804 4971 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 13 06:49:13 crc kubenswrapper[4971]: E1213 06:49:13.853457 4971 eviction_manager.go:285] "Eviction manager: failed to get summary stats" err="failed to get node info: node \"crc\" not found" Dec 13 06:49:14 crc kubenswrapper[4971]: I1213 06:49:14.381248 4971 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-kube-controller-manager/kube-controller-manager-crc" Dec 13 06:49:14 crc kubenswrapper[4971]: I1213 06:49:14.381437 4971 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Dec 13 06:49:14 crc kubenswrapper[4971]: I1213 06:49:14.384760 4971 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 13 06:49:14 crc kubenswrapper[4971]: I1213 06:49:14.384804 4971 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 13 06:49:14 crc kubenswrapper[4971]: I1213 06:49:14.384813 4971 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 13 06:49:14 crc kubenswrapper[4971]: I1213 06:49:14.402459 4971 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-etcd/etcd-crc" Dec 13 06:49:14 crc kubenswrapper[4971]: I1213 06:49:14.402674 4971 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Dec 13 06:49:14 crc kubenswrapper[4971]: I1213 06:49:14.404018 4971 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 13 06:49:14 crc kubenswrapper[4971]: I1213 06:49:14.404055 4971 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 13 06:49:14 crc kubenswrapper[4971]: I1213 06:49:14.404065 4971 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 13 06:49:14 crc kubenswrapper[4971]: I1213 06:49:14.560158 4971 patch_prober.go:28] interesting pod/kube-controller-manager-crc container/cluster-policy-controller namespace/openshift-kube-controller-manager: Startup probe status=failure output="Get \"https://192.168.126.11:10357/healthz\": context deadline exceeded (Client.Timeout exceeded while awaiting headers)" start-of-body= Dec 13 06:49:14 crc kubenswrapper[4971]: I1213 06:49:14.560240 4971 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-kube-controller-manager/kube-controller-manager-crc" podUID="f614b9022728cf315e60c057852e563e" containerName="cluster-policy-controller" probeResult="failure" output="Get \"https://192.168.126.11:10357/healthz\": context deadline exceeded (Client.Timeout exceeded while awaiting headers)" Dec 13 06:49:15 crc kubenswrapper[4971]: I1213 06:49:15.779775 4971 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-kube-controller-manager/kube-controller-manager-crc" Dec 13 06:49:15 crc kubenswrapper[4971]: I1213 06:49:15.779960 4971 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Dec 13 06:49:15 crc kubenswrapper[4971]: I1213 06:49:15.780973 4971 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 13 06:49:15 crc kubenswrapper[4971]: I1213 06:49:15.781011 4971 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 13 06:49:15 crc kubenswrapper[4971]: I1213 06:49:15.781023 4971 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 13 06:49:15 crc kubenswrapper[4971]: I1213 06:49:15.785250 4971 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-kube-controller-manager/kube-controller-manager-crc" Dec 13 06:49:15 crc kubenswrapper[4971]: I1213 06:49:15.856040 4971 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Dec 13 06:49:15 crc kubenswrapper[4971]: I1213 06:49:15.857500 4971 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 13 06:49:15 crc kubenswrapper[4971]: I1213 06:49:15.857556 4971 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 13 06:49:15 crc kubenswrapper[4971]: I1213 06:49:15.857565 4971 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 13 06:49:15 crc kubenswrapper[4971]: I1213 06:49:15.860558 4971 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-kube-controller-manager/kube-controller-manager-crc" Dec 13 06:49:16 crc kubenswrapper[4971]: I1213 06:49:16.857605 4971 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Dec 13 06:49:16 crc kubenswrapper[4971]: I1213 06:49:16.858340 4971 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 13 06:49:16 crc kubenswrapper[4971]: I1213 06:49:16.858381 4971 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 13 06:49:16 crc kubenswrapper[4971]: I1213 06:49:16.858393 4971 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 13 06:49:18 crc kubenswrapper[4971]: I1213 06:49:18.272176 4971 patch_prober.go:28] interesting pod/kube-apiserver-crc container/kube-apiserver namespace/openshift-kube-apiserver: Startup probe status=failure output="HTTP probe failed with statuscode: 403" start-of-body={"kind":"Status","apiVersion":"v1","metadata":{},"status":"Failure","message":"forbidden: User \"system:anonymous\" cannot get path \"/livez\"","reason":"Forbidden","details":{},"code":403} Dec 13 06:49:18 crc kubenswrapper[4971]: I1213 06:49:18.272236 4971 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver" probeResult="failure" output="HTTP probe failed with statuscode: 403" Dec 13 06:49:18 crc kubenswrapper[4971]: I1213 06:49:18.278374 4971 patch_prober.go:28] interesting pod/kube-apiserver-crc container/kube-apiserver namespace/openshift-kube-apiserver: Startup probe status=failure output="HTTP probe failed with statuscode: 403" start-of-body={"kind":"Status","apiVersion":"v1","metadata":{},"status":"Failure","message":"forbidden: User \"system:anonymous\" cannot get path \"/livez\"","reason":"Forbidden","details":{},"code":403} Dec 13 06:49:18 crc kubenswrapper[4971]: I1213 06:49:18.278454 4971 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver" probeResult="failure" output="HTTP probe failed with statuscode: 403" Dec 13 06:49:19 crc kubenswrapper[4971]: I1213 06:49:19.157640 4971 patch_prober.go:28] interesting pod/kube-apiserver-crc container/kube-apiserver namespace/openshift-kube-apiserver: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[+]ping ok Dec 13 06:49:19 crc kubenswrapper[4971]: [+]log ok Dec 13 06:49:19 crc kubenswrapper[4971]: [+]etcd ok Dec 13 06:49:19 crc kubenswrapper[4971]: [+]poststarthook/openshift.io-oauth-apiserver-reachable ok Dec 13 06:49:19 crc kubenswrapper[4971]: [+]poststarthook/start-apiserver-admission-initializer ok Dec 13 06:49:19 crc kubenswrapper[4971]: [+]poststarthook/quota.openshift.io-clusterquotamapping ok Dec 13 06:49:19 crc kubenswrapper[4971]: [+]poststarthook/openshift.io-api-request-count-filter ok Dec 13 06:49:19 crc kubenswrapper[4971]: [+]poststarthook/openshift.io-startkubeinformers ok Dec 13 06:49:19 crc kubenswrapper[4971]: [+]poststarthook/openshift.io-openshift-apiserver-reachable ok Dec 13 06:49:19 crc kubenswrapper[4971]: [+]poststarthook/generic-apiserver-start-informers ok Dec 13 06:49:19 crc kubenswrapper[4971]: [+]poststarthook/priority-and-fairness-config-consumer ok Dec 13 06:49:19 crc kubenswrapper[4971]: [+]poststarthook/priority-and-fairness-filter ok Dec 13 06:49:19 crc kubenswrapper[4971]: [+]poststarthook/storage-object-count-tracker-hook ok Dec 13 06:49:19 crc kubenswrapper[4971]: [+]poststarthook/start-apiextensions-informers ok Dec 13 06:49:19 crc kubenswrapper[4971]: [+]poststarthook/start-apiextensions-controllers ok Dec 13 06:49:19 crc kubenswrapper[4971]: [+]poststarthook/crd-informer-synced ok Dec 13 06:49:19 crc kubenswrapper[4971]: [+]poststarthook/start-system-namespaces-controller ok Dec 13 06:49:19 crc kubenswrapper[4971]: [+]poststarthook/start-cluster-authentication-info-controller ok Dec 13 06:49:19 crc kubenswrapper[4971]: [+]poststarthook/start-kube-apiserver-identity-lease-controller ok Dec 13 06:49:19 crc kubenswrapper[4971]: [+]poststarthook/start-kube-apiserver-identity-lease-garbage-collector ok Dec 13 06:49:19 crc kubenswrapper[4971]: [+]poststarthook/start-legacy-token-tracking-controller ok Dec 13 06:49:19 crc kubenswrapper[4971]: [+]poststarthook/start-service-ip-repair-controllers ok Dec 13 06:49:19 crc kubenswrapper[4971]: [-]poststarthook/rbac/bootstrap-roles failed: reason withheld Dec 13 06:49:19 crc kubenswrapper[4971]: [-]poststarthook/scheduling/bootstrap-system-priority-classes failed: reason withheld Dec 13 06:49:19 crc kubenswrapper[4971]: [+]poststarthook/priority-and-fairness-config-producer ok Dec 13 06:49:19 crc kubenswrapper[4971]: [+]poststarthook/bootstrap-controller ok Dec 13 06:49:19 crc kubenswrapper[4971]: [+]poststarthook/aggregator-reload-proxy-client-cert ok Dec 13 06:49:19 crc kubenswrapper[4971]: [+]poststarthook/start-kube-aggregator-informers ok Dec 13 06:49:19 crc kubenswrapper[4971]: [+]poststarthook/apiservice-status-local-available-controller ok Dec 13 06:49:19 crc kubenswrapper[4971]: [+]poststarthook/apiservice-status-remote-available-controller ok Dec 13 06:49:19 crc kubenswrapper[4971]: [+]poststarthook/apiservice-registration-controller ok Dec 13 06:49:19 crc kubenswrapper[4971]: [+]poststarthook/apiservice-wait-for-first-sync ok Dec 13 06:49:19 crc kubenswrapper[4971]: [+]poststarthook/apiservice-discovery-controller ok Dec 13 06:49:19 crc kubenswrapper[4971]: [+]poststarthook/kube-apiserver-autoregistration ok Dec 13 06:49:19 crc kubenswrapper[4971]: [+]autoregister-completion ok Dec 13 06:49:19 crc kubenswrapper[4971]: [+]poststarthook/apiservice-openapi-controller ok Dec 13 06:49:19 crc kubenswrapper[4971]: [+]poststarthook/apiservice-openapiv3-controller ok Dec 13 06:49:19 crc kubenswrapper[4971]: livez check failed Dec 13 06:49:19 crc kubenswrapper[4971]: I1213 06:49:19.157755 4971 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver" probeResult="failure" output="HTTP probe failed with statuscode: 500" Dec 13 06:49:23 crc kubenswrapper[4971]: I1213 06:49:23.249377 4971 patch_prober.go:28] interesting pod/kube-apiserver-crc container/kube-apiserver-check-endpoints namespace/openshift-kube-apiserver: Liveness probe status=failure output="Get \"https://192.168.126.11:17697/healthz\": dial tcp 192.168.126.11:17697: connect: connection refused" start-of-body= Dec 13 06:49:23 crc kubenswrapper[4971]: I1213 06:49:23.249478 4971 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-check-endpoints" probeResult="failure" output="Get \"https://192.168.126.11:17697/healthz\": dial tcp 192.168.126.11:17697: connect: connection refused" Dec 13 06:49:23 crc kubenswrapper[4971]: E1213 06:49:23.262198 4971 controller.go:145] "Failed to ensure lease exists, will retry" err="Get \"https://api-int.crc.testing:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/crc?timeout=10s\": context deadline exceeded" interval="6.4s" Dec 13 06:49:23 crc kubenswrapper[4971]: I1213 06:49:23.263963 4971 trace.go:236] Trace[1545608624]: "Reflector ListAndWatch" name:k8s.io/client-go/informers/factory.go:160 (13-Dec-2025 06:49:12.565) (total time: 10698ms): Dec 13 06:49:23 crc kubenswrapper[4971]: Trace[1545608624]: ---"Objects listed" error: 10698ms (06:49:23.263) Dec 13 06:49:23 crc kubenswrapper[4971]: Trace[1545608624]: [10.698184709s] [10.698184709s] END Dec 13 06:49:23 crc kubenswrapper[4971]: I1213 06:49:23.264001 4971 reflector.go:368] Caches populated for *v1.RuntimeClass from k8s.io/client-go/informers/factory.go:160 Dec 13 06:49:23 crc kubenswrapper[4971]: I1213 06:49:23.265237 4971 reflector.go:368] Caches populated for *v1.CSIDriver from k8s.io/client-go/informers/factory.go:160 Dec 13 06:49:23 crc kubenswrapper[4971]: E1213 06:49:23.266492 4971 kubelet_node_status.go:99] "Unable to register node with API server" err="nodes \"crc\" is forbidden: autoscaling.openshift.io/ManagedNode infra config cache not synchronized" node="crc" Dec 13 06:49:23 crc kubenswrapper[4971]: I1213 06:49:23.268972 4971 reconstruct.go:205] "DevicePaths of reconstructed volumes updated" Dec 13 06:49:23 crc kubenswrapper[4971]: I1213 06:49:23.269334 4971 trace.go:236] Trace[1611204446]: "Reflector ListAndWatch" name:k8s.io/client-go/informers/factory.go:160 (13-Dec-2025 06:49:12.560) (total time: 10708ms): Dec 13 06:49:23 crc kubenswrapper[4971]: Trace[1611204446]: ---"Objects listed" error: 10708ms (06:49:23.269) Dec 13 06:49:23 crc kubenswrapper[4971]: Trace[1611204446]: [10.708328612s] [10.708328612s] END Dec 13 06:49:23 crc kubenswrapper[4971]: I1213 06:49:23.269358 4971 reflector.go:368] Caches populated for *v1.Node from k8s.io/client-go/informers/factory.go:160 Dec 13 06:49:23 crc kubenswrapper[4971]: I1213 06:49:23.269709 4971 trace.go:236] Trace[50650043]: "Reflector ListAndWatch" name:k8s.io/client-go/informers/factory.go:160 (13-Dec-2025 06:49:11.402) (total time: 11867ms): Dec 13 06:49:23 crc kubenswrapper[4971]: Trace[50650043]: ---"Objects listed" error: 11867ms (06:49:23.269) Dec 13 06:49:23 crc kubenswrapper[4971]: Trace[50650043]: [11.867531874s] [11.867531874s] END Dec 13 06:49:23 crc kubenswrapper[4971]: I1213 06:49:23.269738 4971 reflector.go:368] Caches populated for *v1.Service from k8s.io/client-go/informers/factory.go:160 Dec 13 06:49:23 crc kubenswrapper[4971]: I1213 06:49:23.310238 4971 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-kube-controller-manager/kube-controller-manager-crc" Dec 13 06:49:23 crc kubenswrapper[4971]: I1213 06:49:23.314643 4971 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-kube-controller-manager/kube-controller-manager-crc" Dec 13 06:49:23 crc kubenswrapper[4971]: I1213 06:49:23.668250 4971 apiserver.go:52] "Watching apiserver" Dec 13 06:49:23 crc kubenswrapper[4971]: I1213 06:49:23.671168 4971 reflector.go:368] Caches populated for *v1.Pod from pkg/kubelet/config/apiserver.go:66 Dec 13 06:49:23 crc kubenswrapper[4971]: I1213 06:49:23.671505 4971 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-kube-controller-manager/kube-controller-manager-crc","openshift-network-console/networking-console-plugin-85b44fc459-gdk6g","openshift-network-diagnostics/network-check-source-55646444c4-trplf","openshift-network-diagnostics/network-check-target-xd92c","openshift-network-node-identity/network-node-identity-vrzqb","openshift-network-operator/iptables-alerter-4ln5h","openshift-network-operator/network-operator-58b4c7f79c-55gtf"] Dec 13 06:49:23 crc kubenswrapper[4971]: I1213 06:49:23.671853 4971 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" Dec 13 06:49:23 crc kubenswrapper[4971]: I1213 06:49:23.671917 4971 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 13 06:49:23 crc kubenswrapper[4971]: I1213 06:49:23.671983 4971 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 13 06:49:23 crc kubenswrapper[4971]: I1213 06:49:23.672487 4971 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-node-identity/network-node-identity-vrzqb" Dec 13 06:49:23 crc kubenswrapper[4971]: E1213 06:49:23.672502 4971 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 13 06:49:23 crc kubenswrapper[4971]: E1213 06:49:23.672566 4971 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 13 06:49:23 crc kubenswrapper[4971]: I1213 06:49:23.675247 4971 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 13 06:49:23 crc kubenswrapper[4971]: E1213 06:49:23.675314 4971 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 13 06:49:23 crc kubenswrapper[4971]: I1213 06:49:23.679130 4971 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-operator/iptables-alerter-4ln5h" Dec 13 06:49:23 crc kubenswrapper[4971]: I1213 06:49:23.679648 4971 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-node-identity"/"openshift-service-ca.crt" Dec 13 06:49:23 crc kubenswrapper[4971]: I1213 06:49:23.679849 4971 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-network-operator"/"metrics-tls" Dec 13 06:49:23 crc kubenswrapper[4971]: I1213 06:49:23.679923 4971 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-operator"/"openshift-service-ca.crt" Dec 13 06:49:23 crc kubenswrapper[4971]: I1213 06:49:23.680050 4971 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-operator"/"kube-root-ca.crt" Dec 13 06:49:23 crc kubenswrapper[4971]: I1213 06:49:23.680691 4971 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-node-identity"/"ovnkube-identity-cm" Dec 13 06:49:23 crc kubenswrapper[4971]: I1213 06:49:23.680821 4971 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-node-identity"/"env-overrides" Dec 13 06:49:23 crc kubenswrapper[4971]: I1213 06:49:23.680975 4971 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-node-identity"/"kube-root-ca.crt" Dec 13 06:49:23 crc kubenswrapper[4971]: I1213 06:49:23.681079 4971 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-network-node-identity"/"network-node-identity-cert" Dec 13 06:49:23 crc kubenswrapper[4971]: I1213 06:49:23.683962 4971 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-operator"/"iptables-alerter-script" Dec 13 06:49:23 crc kubenswrapper[4971]: I1213 06:49:23.699911 4971 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-13T06:49:23Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-operator]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-13T06:49:23Z\\\",\\\"message\\\":\\\"containers with unready status: [network-operator]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":false,\\\"restartCount\\\":5,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Dec 13 06:49:23 crc kubenswrapper[4971]: I1213 06:49:23.709638 4971 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-13T06:49:23Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-operator]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-13T06:49:23Z\\\",\\\"message\\\":\\\"containers with unready status: [network-operator]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":false,\\\"restartCount\\\":5,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Dec 13 06:49:23 crc kubenswrapper[4971]: I1213 06:49:23.721379 4971 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-13T06:49:23Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-13T06:49:23Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Dec 13 06:49:23 crc kubenswrapper[4971]: I1213 06:49:23.730599 4971 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-13T06:49:23Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [webhook approver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-13T06:49:23Z\\\",\\\"message\\\":\\\"containers with unready status: [webhook approver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":false,\\\"restartCount\\\":6,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Dec 13 06:49:23 crc kubenswrapper[4971]: I1213 06:49:23.739802 4971 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-13T06:49:23Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-13T06:49:23Z\\\",\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Dec 13 06:49:23 crc kubenswrapper[4971]: I1213 06:49:23.748273 4971 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-13T06:49:23Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-13T06:49:23Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Dec 13 06:49:23 crc kubenswrapper[4971]: I1213 06:49:23.756672 4971 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-13T06:49:23Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-13T06:49:23Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Dec 13 06:49:23 crc kubenswrapper[4971]: I1213 06:49:23.766294 4971 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"f9d31b5a-b770-4a15-95c8-35b49677c608\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T06:49:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T06:49:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T06:49:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T06:49:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T06:49:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c9baa042f9dada954faa57b5454cd6b0a12a297e56ddc192c52cff50e42a16fa\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-13T06:49:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://c788de86754f0bf37ee184672fc1dcb388ccebc246d4f15be4b5da661f725f8f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-13T06:49:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://760817511a9b4f96a68dc38d26631250eaab72e1dd8f123e653039b6b4e91f84\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-13T06:49:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b859be0c9feac856e9855c6a038b02b49b32fade97873c46d0bec0a32d8b5147\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-13T06:49:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-13T06:49:03Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Dec 13 06:49:23 crc kubenswrapper[4971]: I1213 06:49:23.767786 4971 desired_state_of_world_populator.go:154] "Finished populating initial desired state of world" Dec 13 06:49:23 crc kubenswrapper[4971]: I1213 06:49:23.770830 4971 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/6509e943-70c6-444c-bc41-48a544e36fbd-config\") pod \"6509e943-70c6-444c-bc41-48a544e36fbd\" (UID: \"6509e943-70c6-444c-bc41-48a544e36fbd\") " Dec 13 06:49:23 crc kubenswrapper[4971]: I1213 06:49:23.771006 4971 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/925f1c65-6136-48ba-85aa-3a3b50560753-env-overrides\") pod \"925f1c65-6136-48ba-85aa-3a3b50560753\" (UID: \"925f1c65-6136-48ba-85aa-3a3b50560753\") " Dec 13 06:49:23 crc kubenswrapper[4971]: I1213 06:49:23.771132 4971 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/09efc573-dbb6-4249-bd59-9b87aba8dd28-serving-cert\") pod \"09efc573-dbb6-4249-bd59-9b87aba8dd28\" (UID: \"09efc573-dbb6-4249-bd59-9b87aba8dd28\") " Dec 13 06:49:23 crc kubenswrapper[4971]: I1213 06:49:23.771602 4971 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"available-featuregates\" (UniqueName: \"kubernetes.io/empty-dir/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d-available-featuregates\") pod \"bc5039c0-ea34-426b-a2b7-fbbc87b49a6d\" (UID: \"bc5039c0-ea34-426b-a2b7-fbbc87b49a6d\") " Dec 13 06:49:23 crc kubenswrapper[4971]: I1213 06:49:23.772138 4971 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-ngvvp\" (UniqueName: \"kubernetes.io/projected/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-kube-api-access-ngvvp\") pod \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\" (UID: \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\") " Dec 13 06:49:23 crc kubenswrapper[4971]: I1213 06:49:23.772261 4971 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/01ab3dd5-8196-46d0-ad33-122e2ca51def-config\") pod \"01ab3dd5-8196-46d0-ad33-122e2ca51def\" (UID: \"01ab3dd5-8196-46d0-ad33-122e2ca51def\") " Dec 13 06:49:23 crc kubenswrapper[4971]: I1213 06:49:23.772365 4971 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"image-import-ca\" (UniqueName: \"kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-image-import-ca\") pod \"1bf7eb37-55a3-4c65-b768-a94c82151e69\" (UID: \"1bf7eb37-55a3-4c65-b768-a94c82151e69\") " Dec 13 06:49:23 crc kubenswrapper[4971]: I1213 06:49:23.771556 4971 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/09efc573-dbb6-4249-bd59-9b87aba8dd28-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "09efc573-dbb6-4249-bd59-9b87aba8dd28" (UID: "09efc573-dbb6-4249-bd59-9b87aba8dd28"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 13 06:49:23 crc kubenswrapper[4971]: I1213 06:49:23.772458 4971 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovn-node-metrics-cert\" (UniqueName: \"kubernetes.io/secret/6ea678ab-3438-413e-bfe3-290ae7725660-ovn-node-metrics-cert\") pod \"6ea678ab-3438-413e-bfe3-290ae7725660\" (UID: \"6ea678ab-3438-413e-bfe3-290ae7725660\") " Dec 13 06:49:23 crc kubenswrapper[4971]: I1213 06:49:23.771808 4971 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/925f1c65-6136-48ba-85aa-3a3b50560753-env-overrides" (OuterVolumeSpecName: "env-overrides") pod "925f1c65-6136-48ba-85aa-3a3b50560753" (UID: "925f1c65-6136-48ba-85aa-3a3b50560753"). InnerVolumeSpecName "env-overrides". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 13 06:49:23 crc kubenswrapper[4971]: I1213 06:49:23.771880 4971 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/6509e943-70c6-444c-bc41-48a544e36fbd-config" (OuterVolumeSpecName: "config") pod "6509e943-70c6-444c-bc41-48a544e36fbd" (UID: "6509e943-70c6-444c-bc41-48a544e36fbd"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 13 06:49:23 crc kubenswrapper[4971]: I1213 06:49:23.772417 4971 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-kube-api-access-ngvvp" (OuterVolumeSpecName: "kube-api-access-ngvvp") pod "49c341d1-5089-4bc2-86a0-a5e165cfcc6b" (UID: "49c341d1-5089-4bc2-86a0-a5e165cfcc6b"). InnerVolumeSpecName "kube-api-access-ngvvp". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 13 06:49:23 crc kubenswrapper[4971]: I1213 06:49:23.772640 4971 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"registry-tls\" (UniqueName: \"kubernetes.io/projected/8f668bae-612b-4b75-9490-919e737c6a3b-registry-tls\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 13 06:49:23 crc kubenswrapper[4971]: I1213 06:49:23.772675 4971 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"srv-cert\" (UniqueName: \"kubernetes.io/secret/b6312bbd-5731-4ea0-a20f-81d5a57df44a-srv-cert\") pod \"b6312bbd-5731-4ea0-a20f-81d5a57df44a\" (UID: \"b6312bbd-5731-4ea0-a20f-81d5a57df44a\") " Dec 13 06:49:23 crc kubenswrapper[4971]: I1213 06:49:23.772719 4971 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-w4xd4\" (UniqueName: \"kubernetes.io/projected/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b-kube-api-access-w4xd4\") pod \"8cea82b4-6893-4ddc-af9f-1bb5ae425c5b\" (UID: \"8cea82b4-6893-4ddc-af9f-1bb5ae425c5b\") " Dec 13 06:49:23 crc kubenswrapper[4971]: I1213 06:49:23.772741 4971 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/01ab3dd5-8196-46d0-ad33-122e2ca51def-serving-cert\") pod \"01ab3dd5-8196-46d0-ad33-122e2ca51def\" (UID: \"01ab3dd5-8196-46d0-ad33-122e2ca51def\") " Dec 13 06:49:23 crc kubenswrapper[4971]: I1213 06:49:23.772769 4971 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"samples-operator-tls\" (UniqueName: \"kubernetes.io/secret/a0128f3a-b052-44ed-a84e-c4c8aaf17c13-samples-operator-tls\") pod \"a0128f3a-b052-44ed-a84e-c4c8aaf17c13\" (UID: \"a0128f3a-b052-44ed-a84e-c4c8aaf17c13\") " Dec 13 06:49:23 crc kubenswrapper[4971]: I1213 06:49:23.772818 4971 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"webhook-certs\" (UniqueName: \"kubernetes.io/secret/efdd0498-1daa-4136-9a4a-3b948c2293fc-webhook-certs\") pod \"efdd0498-1daa-4136-9a4a-3b948c2293fc\" (UID: \"efdd0498-1daa-4136-9a4a-3b948c2293fc\") " Dec 13 06:49:23 crc kubenswrapper[4971]: I1213 06:49:23.772840 4971 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/96b93a3a-6083-4aea-8eab-fe1aa8245ad9-metrics-tls\") pod \"96b93a3a-6083-4aea-8eab-fe1aa8245ad9\" (UID: \"96b93a3a-6083-4aea-8eab-fe1aa8245ad9\") " Dec 13 06:49:23 crc kubenswrapper[4971]: I1213 06:49:23.773336 4971 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/e7e6199b-1264-4501-8953-767f51328d08-config\") pod \"e7e6199b-1264-4501-8953-767f51328d08\" (UID: \"e7e6199b-1264-4501-8953-767f51328d08\") " Dec 13 06:49:23 crc kubenswrapper[4971]: I1213 06:49:23.773006 4971 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/8f668bae-612b-4b75-9490-919e737c6a3b-registry-tls" (OuterVolumeSpecName: "registry-tls") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b"). InnerVolumeSpecName "registry-tls". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 13 06:49:23 crc kubenswrapper[4971]: I1213 06:49:23.773005 4971 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/b6312bbd-5731-4ea0-a20f-81d5a57df44a-srv-cert" (OuterVolumeSpecName: "srv-cert") pod "b6312bbd-5731-4ea0-a20f-81d5a57df44a" (UID: "b6312bbd-5731-4ea0-a20f-81d5a57df44a"). InnerVolumeSpecName "srv-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 13 06:49:23 crc kubenswrapper[4971]: I1213 06:49:23.773278 4971 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/01ab3dd5-8196-46d0-ad33-122e2ca51def-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "01ab3dd5-8196-46d0-ad33-122e2ca51def" (UID: "01ab3dd5-8196-46d0-ad33-122e2ca51def"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 13 06:49:23 crc kubenswrapper[4971]: I1213 06:49:23.773410 4971 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/efdd0498-1daa-4136-9a4a-3b948c2293fc-webhook-certs" (OuterVolumeSpecName: "webhook-certs") pod "efdd0498-1daa-4136-9a4a-3b948c2293fc" (UID: "efdd0498-1daa-4136-9a4a-3b948c2293fc"). InnerVolumeSpecName "webhook-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 13 06:49:23 crc kubenswrapper[4971]: I1213 06:49:23.773430 4971 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/a0128f3a-b052-44ed-a84e-c4c8aaf17c13-samples-operator-tls" (OuterVolumeSpecName: "samples-operator-tls") pod "a0128f3a-b052-44ed-a84e-c4c8aaf17c13" (UID: "a0128f3a-b052-44ed-a84e-c4c8aaf17c13"). InnerVolumeSpecName "samples-operator-tls". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 13 06:49:23 crc kubenswrapper[4971]: I1213 06:49:23.773428 4971 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b-kube-api-access-w4xd4" (OuterVolumeSpecName: "kube-api-access-w4xd4") pod "8cea82b4-6893-4ddc-af9f-1bb5ae425c5b" (UID: "8cea82b4-6893-4ddc-af9f-1bb5ae425c5b"). InnerVolumeSpecName "kube-api-access-w4xd4". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 13 06:49:23 crc kubenswrapper[4971]: I1213 06:49:23.773482 4971 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-fqsjt\" (UniqueName: \"kubernetes.io/projected/efdd0498-1daa-4136-9a4a-3b948c2293fc-kube-api-access-fqsjt\") pod \"efdd0498-1daa-4136-9a4a-3b948c2293fc\" (UID: \"efdd0498-1daa-4136-9a4a-3b948c2293fc\") " Dec 13 06:49:23 crc kubenswrapper[4971]: I1213 06:49:23.773537 4971 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etcd-service-ca\" (UniqueName: \"kubernetes.io/configmap/09efc573-dbb6-4249-bd59-9b87aba8dd28-etcd-service-ca\") pod \"09efc573-dbb6-4249-bd59-9b87aba8dd28\" (UID: \"09efc573-dbb6-4249-bd59-9b87aba8dd28\") " Dec 13 06:49:23 crc kubenswrapper[4971]: I1213 06:49:23.773563 4971 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"stats-auth\" (UniqueName: \"kubernetes.io/secret/c03ee662-fb2f-4fc4-a2c1-af487c19d254-stats-auth\") pod \"c03ee662-fb2f-4fc4-a2c1-af487c19d254\" (UID: \"c03ee662-fb2f-4fc4-a2c1-af487c19d254\") " Dec 13 06:49:23 crc kubenswrapper[4971]: I1213 06:49:23.773585 4971 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/a31745f5-9847-4afe-82a5-3161cc66ca93-metrics-tls\") pod \"a31745f5-9847-4afe-82a5-3161cc66ca93\" (UID: \"a31745f5-9847-4afe-82a5-3161cc66ca93\") " Dec 13 06:49:23 crc kubenswrapper[4971]: I1213 06:49:23.773689 4971 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/efdd0498-1daa-4136-9a4a-3b948c2293fc-kube-api-access-fqsjt" (OuterVolumeSpecName: "kube-api-access-fqsjt") pod "efdd0498-1daa-4136-9a4a-3b948c2293fc" (UID: "efdd0498-1daa-4136-9a4a-3b948c2293fc"). InnerVolumeSpecName "kube-api-access-fqsjt". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 13 06:49:23 crc kubenswrapper[4971]: I1213 06:49:23.773707 4971 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-x2m85\" (UniqueName: \"kubernetes.io/projected/cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d-kube-api-access-x2m85\") pod \"cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d\" (UID: \"cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d\") " Dec 13 06:49:23 crc kubenswrapper[4971]: I1213 06:49:23.773870 4971 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/c03ee662-fb2f-4fc4-a2c1-af487c19d254-stats-auth" (OuterVolumeSpecName: "stats-auth") pod "c03ee662-fb2f-4fc4-a2c1-af487c19d254" (UID: "c03ee662-fb2f-4fc4-a2c1-af487c19d254"). InnerVolumeSpecName "stats-auth". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 13 06:49:23 crc kubenswrapper[4971]: I1213 06:49:23.773892 4971 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/a31745f5-9847-4afe-82a5-3161cc66ca93-metrics-tls" (OuterVolumeSpecName: "metrics-tls") pod "a31745f5-9847-4afe-82a5-3161cc66ca93" (UID: "a31745f5-9847-4afe-82a5-3161cc66ca93"). InnerVolumeSpecName "metrics-tls". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 13 06:49:23 crc kubenswrapper[4971]: I1213 06:49:23.774014 4971 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d-kube-api-access-x2m85" (OuterVolumeSpecName: "kube-api-access-x2m85") pod "cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d" (UID: "cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d"). InnerVolumeSpecName "kube-api-access-x2m85". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 13 06:49:23 crc kubenswrapper[4971]: I1213 06:49:23.773738 4971 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"installation-pull-secrets\" (UniqueName: \"kubernetes.io/secret/8f668bae-612b-4b75-9490-919e737c6a3b-installation-pull-secrets\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 13 06:49:23 crc kubenswrapper[4971]: I1213 06:49:23.774071 4971 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/e7e6199b-1264-4501-8953-767f51328d08-config" (OuterVolumeSpecName: "config") pod "e7e6199b-1264-4501-8953-767f51328d08" (UID: "e7e6199b-1264-4501-8953-767f51328d08"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 13 06:49:23 crc kubenswrapper[4971]: I1213 06:49:23.774102 4971 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-dbsvg\" (UniqueName: \"kubernetes.io/projected/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9-kube-api-access-dbsvg\") pod \"f88749ec-7931-4ee7-b3fc-1ec5e11f92e9\" (UID: \"f88749ec-7931-4ee7-b3fc-1ec5e11f92e9\") " Dec 13 06:49:23 crc kubenswrapper[4971]: I1213 06:49:23.774129 4971 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-system-router-certs\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-router-certs\") pod \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\" (UID: \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\") " Dec 13 06:49:23 crc kubenswrapper[4971]: I1213 06:49:23.774590 4971 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/8f668bae-612b-4b75-9490-919e737c6a3b-bound-sa-token\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 13 06:49:23 crc kubenswrapper[4971]: I1213 06:49:23.774329 4971 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/8f668bae-612b-4b75-9490-919e737c6a3b-installation-pull-secrets" (OuterVolumeSpecName: "installation-pull-secrets") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b"). InnerVolumeSpecName "installation-pull-secrets". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 13 06:49:23 crc kubenswrapper[4971]: I1213 06:49:23.774346 4971 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/09efc573-dbb6-4249-bd59-9b87aba8dd28-etcd-service-ca" (OuterVolumeSpecName: "etcd-service-ca") pod "09efc573-dbb6-4249-bd59-9b87aba8dd28" (UID: "09efc573-dbb6-4249-bd59-9b87aba8dd28"). InnerVolumeSpecName "etcd-service-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 13 06:49:23 crc kubenswrapper[4971]: I1213 06:49:23.774480 4971 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/6ea678ab-3438-413e-bfe3-290ae7725660-ovn-node-metrics-cert" (OuterVolumeSpecName: "ovn-node-metrics-cert") pod "6ea678ab-3438-413e-bfe3-290ae7725660" (UID: "6ea678ab-3438-413e-bfe3-290ae7725660"). InnerVolumeSpecName "ovn-node-metrics-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 13 06:49:23 crc kubenswrapper[4971]: I1213 06:49:23.774467 4971 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9-kube-api-access-dbsvg" (OuterVolumeSpecName: "kube-api-access-dbsvg") pod "f88749ec-7931-4ee7-b3fc-1ec5e11f92e9" (UID: "f88749ec-7931-4ee7-b3fc-1ec5e11f92e9"). InnerVolumeSpecName "kube-api-access-dbsvg". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 13 06:49:23 crc kubenswrapper[4971]: I1213 06:49:23.774584 4971 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/96b93a3a-6083-4aea-8eab-fe1aa8245ad9-metrics-tls" (OuterVolumeSpecName: "metrics-tls") pod "96b93a3a-6083-4aea-8eab-fe1aa8245ad9" (UID: "96b93a3a-6083-4aea-8eab-fe1aa8245ad9"). InnerVolumeSpecName "metrics-tls". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 13 06:49:23 crc kubenswrapper[4971]: I1213 06:49:23.774607 4971 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-router-certs" (OuterVolumeSpecName: "v4-0-config-system-router-certs") pod "49c341d1-5089-4bc2-86a0-a5e165cfcc6b" (UID: "49c341d1-5089-4bc2-86a0-a5e165cfcc6b"). InnerVolumeSpecName "v4-0-config-system-router-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 13 06:49:23 crc kubenswrapper[4971]: I1213 06:49:23.774706 4971 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/01ab3dd5-8196-46d0-ad33-122e2ca51def-config" (OuterVolumeSpecName: "config") pod "01ab3dd5-8196-46d0-ad33-122e2ca51def" (UID: "01ab3dd5-8196-46d0-ad33-122e2ca51def"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 13 06:49:23 crc kubenswrapper[4971]: I1213 06:49:23.774712 4971 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-mnrrd\" (UniqueName: \"kubernetes.io/projected/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d-kube-api-access-mnrrd\") pod \"bc5039c0-ea34-426b-a2b7-fbbc87b49a6d\" (UID: \"bc5039c0-ea34-426b-a2b7-fbbc87b49a6d\") " Dec 13 06:49:23 crc kubenswrapper[4971]: I1213 06:49:23.774841 4971 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/8f668bae-612b-4b75-9490-919e737c6a3b-bound-sa-token" (OuterVolumeSpecName: "bound-sa-token") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b"). InnerVolumeSpecName "bound-sa-token". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 13 06:49:23 crc kubenswrapper[4971]: I1213 06:49:23.774863 4971 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/8f668bae-612b-4b75-9490-919e737c6a3b-trusted-ca\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 13 06:49:23 crc kubenswrapper[4971]: I1213 06:49:23.774957 4971 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-image-import-ca" (OuterVolumeSpecName: "image-import-ca") pod "1bf7eb37-55a3-4c65-b768-a94c82151e69" (UID: "1bf7eb37-55a3-4c65-b768-a94c82151e69"). InnerVolumeSpecName "image-import-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 13 06:49:23 crc kubenswrapper[4971]: I1213 06:49:23.774967 4971 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d-kube-api-access-mnrrd" (OuterVolumeSpecName: "kube-api-access-mnrrd") pod "bc5039c0-ea34-426b-a2b7-fbbc87b49a6d" (UID: "bc5039c0-ea34-426b-a2b7-fbbc87b49a6d"). InnerVolumeSpecName "kube-api-access-mnrrd". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 13 06:49:23 crc kubenswrapper[4971]: I1213 06:49:23.775380 4971 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/8f668bae-612b-4b75-9490-919e737c6a3b-trusted-ca" (OuterVolumeSpecName: "trusted-ca") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b"). InnerVolumeSpecName "trusted-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 13 06:49:23 crc kubenswrapper[4971]: I1213 06:49:23.775380 4971 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d-available-featuregates" (OuterVolumeSpecName: "available-featuregates") pod "bc5039c0-ea34-426b-a2b7-fbbc87b49a6d" (UID: "bc5039c0-ea34-426b-a2b7-fbbc87b49a6d"). InnerVolumeSpecName "available-featuregates". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 13 06:49:23 crc kubenswrapper[4971]: I1213 06:49:23.775560 4971 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 13 06:49:23 crc kubenswrapper[4971]: E1213 06:49:23.775658 4971 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-13 06:49:24.275638491 +0000 UTC m=+20.880047939 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 13 06:49:23 crc kubenswrapper[4971]: I1213 06:49:23.775916 4971 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-w9rds\" (UniqueName: \"kubernetes.io/projected/20b0d48f-5fd6-431c-a545-e3c800c7b866-kube-api-access-w9rds\") pod \"20b0d48f-5fd6-431c-a545-e3c800c7b866\" (UID: \"20b0d48f-5fd6-431c-a545-e3c800c7b866\") " Dec 13 06:49:23 crc kubenswrapper[4971]: I1213 06:49:23.776347 4971 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/20b0d48f-5fd6-431c-a545-e3c800c7b866-kube-api-access-w9rds" (OuterVolumeSpecName: "kube-api-access-w9rds") pod "20b0d48f-5fd6-431c-a545-e3c800c7b866" (UID: "20b0d48f-5fd6-431c-a545-e3c800c7b866"). InnerVolumeSpecName "kube-api-access-w9rds". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 13 06:49:23 crc kubenswrapper[4971]: I1213 06:49:23.777977 4971 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-13T06:49:23Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-13T06:49:23Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Dec 13 06:49:23 crc kubenswrapper[4971]: I1213 06:49:23.778333 4971 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"srv-cert\" (UniqueName: \"kubernetes.io/secret/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9-srv-cert\") pod \"f88749ec-7931-4ee7-b3fc-1ec5e11f92e9\" (UID: \"f88749ec-7931-4ee7-b3fc-1ec5e11f92e9\") " Dec 13 06:49:23 crc kubenswrapper[4971]: I1213 06:49:23.778452 4971 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-lzf88\" (UniqueName: \"kubernetes.io/projected/0b574797-001e-440a-8f4e-c0be86edad0f-kube-api-access-lzf88\") pod \"0b574797-001e-440a-8f4e-c0be86edad0f\" (UID: \"0b574797-001e-440a-8f4e-c0be86edad0f\") " Dec 13 06:49:23 crc kubenswrapper[4971]: I1213 06:49:23.778537 4971 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-fcqwp\" (UniqueName: \"kubernetes.io/projected/5fe579f8-e8a6-4643-bce5-a661393c4dde-kube-api-access-fcqwp\") pod \"5fe579f8-e8a6-4643-bce5-a661393c4dde\" (UID: \"5fe579f8-e8a6-4643-bce5-a661393c4dde\") " Dec 13 06:49:23 crc kubenswrapper[4971]: I1213 06:49:23.778572 4971 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"mcc-auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/0b574797-001e-440a-8f4e-c0be86edad0f-mcc-auth-proxy-config\") pod \"0b574797-001e-440a-8f4e-c0be86edad0f\" (UID: \"0b574797-001e-440a-8f4e-c0be86edad0f\") " Dec 13 06:49:23 crc kubenswrapper[4971]: I1213 06:49:23.778570 4971 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9-srv-cert" (OuterVolumeSpecName: "srv-cert") pod "f88749ec-7931-4ee7-b3fc-1ec5e11f92e9" (UID: "f88749ec-7931-4ee7-b3fc-1ec5e11f92e9"). InnerVolumeSpecName "srv-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 13 06:49:23 crc kubenswrapper[4971]: I1213 06:49:23.778624 4971 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-webhook-cert\") pod \"308be0ea-9f5f-4b29-aeb1-5abd31a0b17b\" (UID: \"308be0ea-9f5f-4b29-aeb1-5abd31a0b17b\") " Dec 13 06:49:23 crc kubenswrapper[4971]: I1213 06:49:23.778662 4971 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/bf126b07-da06-4140-9a57-dfd54fc6b486-trusted-ca\") pod \"bf126b07-da06-4140-9a57-dfd54fc6b486\" (UID: \"bf126b07-da06-4140-9a57-dfd54fc6b486\") " Dec 13 06:49:23 crc kubenswrapper[4971]: I1213 06:49:23.778712 4971 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-trusted-ca-bundle\") pod \"1bf7eb37-55a3-4c65-b768-a94c82151e69\" (UID: \"1bf7eb37-55a3-4c65-b768-a94c82151e69\") " Dec 13 06:49:23 crc kubenswrapper[4971]: I1213 06:49:23.778747 4971 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-user-template-provider-selection\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-template-provider-selection\") pod \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\" (UID: \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\") " Dec 13 06:49:23 crc kubenswrapper[4971]: I1213 06:49:23.778791 4971 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/0b574797-001e-440a-8f4e-c0be86edad0f-kube-api-access-lzf88" (OuterVolumeSpecName: "kube-api-access-lzf88") pod "0b574797-001e-440a-8f4e-c0be86edad0f" (UID: "0b574797-001e-440a-8f4e-c0be86edad0f"). InnerVolumeSpecName "kube-api-access-lzf88". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 13 06:49:23 crc kubenswrapper[4971]: I1213 06:49:23.778801 4971 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-2w9zh\" (UniqueName: \"kubernetes.io/projected/4bb40260-dbaa-4fb0-84df-5e680505d512-kube-api-access-2w9zh\") pod \"4bb40260-dbaa-4fb0-84df-5e680505d512\" (UID: \"4bb40260-dbaa-4fb0-84df-5e680505d512\") " Dec 13 06:49:23 crc kubenswrapper[4971]: I1213 06:49:23.778835 4971 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-trusted-ca-bundle\") pod \"09ae3b1a-e8e7-4524-b54b-61eab6f9239a\" (UID: \"09ae3b1a-e8e7-4524-b54b-61eab6f9239a\") " Dec 13 06:49:23 crc kubenswrapper[4971]: I1213 06:49:23.778889 4971 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"marketplace-operator-metrics\" (UniqueName: \"kubernetes.io/secret/b6cd30de-2eeb-49a2-ab40-9167f4560ff5-marketplace-operator-metrics\") pod \"b6cd30de-2eeb-49a2-ab40-9167f4560ff5\" (UID: \"b6cd30de-2eeb-49a2-ab40-9167f4560ff5\") " Dec 13 06:49:23 crc kubenswrapper[4971]: I1213 06:49:23.778966 4971 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"service-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/6509e943-70c6-444c-bc41-48a544e36fbd-service-ca-bundle\") pod \"6509e943-70c6-444c-bc41-48a544e36fbd\" (UID: \"6509e943-70c6-444c-bc41-48a544e36fbd\") " Dec 13 06:49:23 crc kubenswrapper[4971]: I1213 06:49:23.778975 4971 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/5fe579f8-e8a6-4643-bce5-a661393c4dde-kube-api-access-fcqwp" (OuterVolumeSpecName: "kube-api-access-fcqwp") pod "5fe579f8-e8a6-4643-bce5-a661393c4dde" (UID: "5fe579f8-e8a6-4643-bce5-a661393c4dde"). InnerVolumeSpecName "kube-api-access-fcqwp". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 13 06:49:23 crc kubenswrapper[4971]: I1213 06:49:23.779001 4971 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"service-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/c03ee662-fb2f-4fc4-a2c1-af487c19d254-service-ca-bundle\") pod \"c03ee662-fb2f-4fc4-a2c1-af487c19d254\" (UID: \"c03ee662-fb2f-4fc4-a2c1-af487c19d254\") " Dec 13 06:49:23 crc kubenswrapper[4971]: I1213 06:49:23.779071 4971 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"oauth-serving-cert\" (UniqueName: \"kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-oauth-serving-cert\") pod \"43509403-f426-496e-be36-56cef71462f5\" (UID: \"43509403-f426-496e-be36-56cef71462f5\") " Dec 13 06:49:23 crc kubenswrapper[4971]: I1213 06:49:23.779134 4971 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"node-bootstrap-token\" (UniqueName: \"kubernetes.io/secret/5fe579f8-e8a6-4643-bce5-a661393c4dde-node-bootstrap-token\") pod \"5fe579f8-e8a6-4643-bce5-a661393c4dde\" (UID: \"5fe579f8-e8a6-4643-bce5-a661393c4dde\") " Dec 13 06:49:23 crc kubenswrapper[4971]: I1213 06:49:23.779172 4971 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d-utilities\") pod \"b11524ee-3fca-4b1b-9cdf-6da289fdbc7d\" (UID: \"b11524ee-3fca-4b1b-9cdf-6da289fdbc7d\") " Dec 13 06:49:23 crc kubenswrapper[4971]: I1213 06:49:23.779241 4971 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-xcgwh\" (UniqueName: \"kubernetes.io/projected/fda69060-fa79-4696-b1a6-7980f124bf7c-kube-api-access-xcgwh\") pod \"fda69060-fa79-4696-b1a6-7980f124bf7c\" (UID: \"fda69060-fa79-4696-b1a6-7980f124bf7c\") " Dec 13 06:49:23 crc kubenswrapper[4971]: I1213 06:49:23.779302 4971 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"certs\" (UniqueName: \"kubernetes.io/secret/5fe579f8-e8a6-4643-bce5-a661393c4dde-certs\") pod \"5fe579f8-e8a6-4643-bce5-a661393c4dde\" (UID: \"5fe579f8-e8a6-4643-bce5-a661393c4dde\") " Dec 13 06:49:23 crc kubenswrapper[4971]: I1213 06:49:23.779339 4971 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"mcd-auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/fda69060-fa79-4696-b1a6-7980f124bf7c-mcd-auth-proxy-config\") pod \"fda69060-fa79-4696-b1a6-7980f124bf7c\" (UID: \"fda69060-fa79-4696-b1a6-7980f124bf7c\") " Dec 13 06:49:23 crc kubenswrapper[4971]: I1213 06:49:23.779414 4971 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-4d4hj\" (UniqueName: \"kubernetes.io/projected/3ab1a177-2de0-46d9-b765-d0d0649bb42e-kube-api-access-4d4hj\") pod \"3ab1a177-2de0-46d9-b765-d0d0649bb42e\" (UID: \"3ab1a177-2de0-46d9-b765-d0d0649bb42e\") " Dec 13 06:49:23 crc kubenswrapper[4971]: I1213 06:49:23.779467 4971 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/0b574797-001e-440a-8f4e-c0be86edad0f-mcc-auth-proxy-config" (OuterVolumeSpecName: "mcc-auth-proxy-config") pod "0b574797-001e-440a-8f4e-c0be86edad0f" (UID: "0b574797-001e-440a-8f4e-c0be86edad0f"). InnerVolumeSpecName "mcc-auth-proxy-config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 13 06:49:23 crc kubenswrapper[4971]: I1213 06:49:23.779480 4971 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etcd-serving-ca\" (UniqueName: \"kubernetes.io/configmap/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-etcd-serving-ca\") pod \"09ae3b1a-e8e7-4524-b54b-61eab6f9239a\" (UID: \"09ae3b1a-e8e7-4524-b54b-61eab6f9239a\") " Dec 13 06:49:23 crc kubenswrapper[4971]: I1213 06:49:23.779548 4971 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-cfbct\" (UniqueName: \"kubernetes.io/projected/57a731c4-ef35-47a8-b875-bfb08a7f8011-kube-api-access-cfbct\") pod \"57a731c4-ef35-47a8-b875-bfb08a7f8011\" (UID: \"57a731c4-ef35-47a8-b875-bfb08a7f8011\") " Dec 13 06:49:23 crc kubenswrapper[4971]: I1213 06:49:23.779592 4971 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ca-trust-extracted\" (UniqueName: \"kubernetes.io/empty-dir/8f668bae-612b-4b75-9490-919e737c6a3b-ca-trust-extracted\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 13 06:49:23 crc kubenswrapper[4971]: I1213 06:49:23.779649 4971 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/7583ce53-e0fe-4a16-9e4d-50516596a136-serving-cert\") pod \"7583ce53-e0fe-4a16-9e4d-50516596a136\" (UID: \"7583ce53-e0fe-4a16-9e4d-50516596a136\") " Dec 13 06:49:23 crc kubenswrapper[4971]: I1213 06:49:23.779686 4971 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-pjr6v\" (UniqueName: \"kubernetes.io/projected/49ef4625-1d3a-4a9f-b595-c2433d32326d-kube-api-access-pjr6v\") pod \"49ef4625-1d3a-4a9f-b595-c2433d32326d\" (UID: \"49ef4625-1d3a-4a9f-b595-c2433d32326d\") " Dec 13 06:49:23 crc kubenswrapper[4971]: I1213 06:49:23.779743 4971 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-wxkg8\" (UniqueName: \"kubernetes.io/projected/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59-kube-api-access-wxkg8\") pod \"3cb93b32-e0ae-4377-b9c8-fdb9842c6d59\" (UID: \"3cb93b32-e0ae-4377-b9c8-fdb9842c6d59\") " Dec 13 06:49:23 crc kubenswrapper[4971]: I1213 06:49:23.779797 4971 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/6402fda4-df10-493c-b4e5-d0569419652d-config\") pod \"6402fda4-df10-493c-b4e5-d0569419652d\" (UID: \"6402fda4-df10-493c-b4e5-d0569419652d\") " Dec 13 06:49:23 crc kubenswrapper[4971]: I1213 06:49:23.779832 4971 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b-config\") pod \"8cea82b4-6893-4ddc-af9f-1bb5ae425c5b\" (UID: \"8cea82b4-6893-4ddc-af9f-1bb5ae425c5b\") " Dec 13 06:49:23 crc kubenswrapper[4971]: I1213 06:49:23.779879 4971 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-pcxfs\" (UniqueName: \"kubernetes.io/projected/9d4552c7-cd75-42dd-8880-30dd377c49a4-kube-api-access-pcxfs\") pod \"9d4552c7-cd75-42dd-8880-30dd377c49a4\" (UID: \"9d4552c7-cd75-42dd-8880-30dd377c49a4\") " Dec 13 06:49:23 crc kubenswrapper[4971]: I1213 06:49:23.779917 4971 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-9xfj7\" (UniqueName: \"kubernetes.io/projected/5225d0e4-402f-4861-b410-819f433b1803-kube-api-access-9xfj7\") pod \"5225d0e4-402f-4861-b410-819f433b1803\" (UID: \"5225d0e4-402f-4861-b410-819f433b1803\") " Dec 13 06:49:23 crc kubenswrapper[4971]: I1213 06:49:23.779590 4971 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-template-provider-selection" (OuterVolumeSpecName: "v4-0-config-user-template-provider-selection") pod "49c341d1-5089-4bc2-86a0-a5e165cfcc6b" (UID: "49c341d1-5089-4bc2-86a0-a5e165cfcc6b"). InnerVolumeSpecName "v4-0-config-user-template-provider-selection". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 13 06:49:23 crc kubenswrapper[4971]: I1213 06:49:23.779636 4971 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/b6cd30de-2eeb-49a2-ab40-9167f4560ff5-marketplace-operator-metrics" (OuterVolumeSpecName: "marketplace-operator-metrics") pod "b6cd30de-2eeb-49a2-ab40-9167f4560ff5" (UID: "b6cd30de-2eeb-49a2-ab40-9167f4560ff5"). InnerVolumeSpecName "marketplace-operator-metrics". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 13 06:49:23 crc kubenswrapper[4971]: I1213 06:49:23.779746 4971 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/fda69060-fa79-4696-b1a6-7980f124bf7c-kube-api-access-xcgwh" (OuterVolumeSpecName: "kube-api-access-xcgwh") pod "fda69060-fa79-4696-b1a6-7980f124bf7c" (UID: "fda69060-fa79-4696-b1a6-7980f124bf7c"). InnerVolumeSpecName "kube-api-access-xcgwh". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 13 06:49:23 crc kubenswrapper[4971]: I1213 06:49:23.779806 4971 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/5fe579f8-e8a6-4643-bce5-a661393c4dde-certs" (OuterVolumeSpecName: "certs") pod "5fe579f8-e8a6-4643-bce5-a661393c4dde" (UID: "5fe579f8-e8a6-4643-bce5-a661393c4dde"). InnerVolumeSpecName "certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 13 06:49:23 crc kubenswrapper[4971]: I1213 06:49:23.779852 4971 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/4bb40260-dbaa-4fb0-84df-5e680505d512-kube-api-access-2w9zh" (OuterVolumeSpecName: "kube-api-access-2w9zh") pod "4bb40260-dbaa-4fb0-84df-5e680505d512" (UID: "4bb40260-dbaa-4fb0-84df-5e680505d512"). InnerVolumeSpecName "kube-api-access-2w9zh". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 13 06:49:23 crc kubenswrapper[4971]: I1213 06:49:23.779987 4971 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/925f1c65-6136-48ba-85aa-3a3b50560753-ovnkube-config\") pod \"925f1c65-6136-48ba-85aa-3a3b50560753\" (UID: \"925f1c65-6136-48ba-85aa-3a3b50560753\") " Dec 13 06:49:23 crc kubenswrapper[4971]: I1213 06:49:23.780001 4971 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/bf126b07-da06-4140-9a57-dfd54fc6b486-trusted-ca" (OuterVolumeSpecName: "trusted-ca") pod "bf126b07-da06-4140-9a57-dfd54fc6b486" (UID: "bf126b07-da06-4140-9a57-dfd54fc6b486"). InnerVolumeSpecName "trusted-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 13 06:49:23 crc kubenswrapper[4971]: I1213 06:49:23.780024 4971 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"audit\" (UniqueName: \"kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-audit\") pod \"1bf7eb37-55a3-4c65-b768-a94c82151e69\" (UID: \"1bf7eb37-55a3-4c65-b768-a94c82151e69\") " Dec 13 06:49:23 crc kubenswrapper[4971]: I1213 06:49:23.780079 4971 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-oauth-serving-cert" (OuterVolumeSpecName: "oauth-serving-cert") pod "43509403-f426-496e-be36-56cef71462f5" (UID: "43509403-f426-496e-be36-56cef71462f5"). InnerVolumeSpecName "oauth-serving-cert". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 13 06:49:23 crc kubenswrapper[4971]: I1213 06:49:23.780074 4971 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d-utilities" (OuterVolumeSpecName: "utilities") pod "b11524ee-3fca-4b1b-9cdf-6da289fdbc7d" (UID: "b11524ee-3fca-4b1b-9cdf-6da289fdbc7d"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 13 06:49:23 crc kubenswrapper[4971]: I1213 06:49:23.780087 4971 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-pj782\" (UniqueName: \"kubernetes.io/projected/b6cd30de-2eeb-49a2-ab40-9167f4560ff5-kube-api-access-pj782\") pod \"b6cd30de-2eeb-49a2-ab40-9167f4560ff5\" (UID: \"b6cd30de-2eeb-49a2-ab40-9167f4560ff5\") " Dec 13 06:49:23 crc kubenswrapper[4971]: I1213 06:49:23.780147 4971 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-system-service-ca\" (UniqueName: \"kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-service-ca\") pod \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\" (UID: \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\") " Dec 13 06:49:23 crc kubenswrapper[4971]: I1213 06:49:23.780339 4971 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"cni-binary-copy\" (UniqueName: \"kubernetes.io/configmap/7bb08738-c794-4ee8-9972-3a62ca171029-cni-binary-copy\") pod \"7bb08738-c794-4ee8-9972-3a62ca171029\" (UID: \"7bb08738-c794-4ee8-9972-3a62ca171029\") " Dec 13 06:49:23 crc kubenswrapper[4971]: I1213 06:49:23.780343 4971 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/49ef4625-1d3a-4a9f-b595-c2433d32326d-kube-api-access-pjr6v" (OuterVolumeSpecName: "kube-api-access-pjr6v") pod "49ef4625-1d3a-4a9f-b595-c2433d32326d" (UID: "49ef4625-1d3a-4a9f-b595-c2433d32326d"). InnerVolumeSpecName "kube-api-access-pjr6v". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 13 06:49:23 crc kubenswrapper[4971]: I1213 06:49:23.780369 4971 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/0b78653f-4ff9-4508-8672-245ed9b561e3-serving-cert\") pod \"0b78653f-4ff9-4508-8672-245ed9b561e3\" (UID: \"0b78653f-4ff9-4508-8672-245ed9b561e3\") " Dec 13 06:49:23 crc kubenswrapper[4971]: I1213 06:49:23.780370 4971 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/5fe579f8-e8a6-4643-bce5-a661393c4dde-node-bootstrap-token" (OuterVolumeSpecName: "node-bootstrap-token") pod "5fe579f8-e8a6-4643-bce5-a661393c4dde" (UID: "5fe579f8-e8a6-4643-bce5-a661393c4dde"). InnerVolumeSpecName "node-bootstrap-token". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 13 06:49:23 crc kubenswrapper[4971]: I1213 06:49:23.780399 4971 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etcd-serving-ca\" (UniqueName: \"kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-etcd-serving-ca\") pod \"1bf7eb37-55a3-4c65-b768-a94c82151e69\" (UID: \"1bf7eb37-55a3-4c65-b768-a94c82151e69\") " Dec 13 06:49:23 crc kubenswrapper[4971]: I1213 06:49:23.780395 4971 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-trusted-ca-bundle" (OuterVolumeSpecName: "trusted-ca-bundle") pod "09ae3b1a-e8e7-4524-b54b-61eab6f9239a" (UID: "09ae3b1a-e8e7-4524-b54b-61eab6f9239a"). InnerVolumeSpecName "trusted-ca-bundle". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 13 06:49:23 crc kubenswrapper[4971]: I1213 06:49:23.780408 4971 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/6509e943-70c6-444c-bc41-48a544e36fbd-service-ca-bundle" (OuterVolumeSpecName: "service-ca-bundle") pod "6509e943-70c6-444c-bc41-48a544e36fbd" (UID: "6509e943-70c6-444c-bc41-48a544e36fbd"). InnerVolumeSpecName "service-ca-bundle". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 13 06:49:23 crc kubenswrapper[4971]: I1213 06:49:23.780426 4971 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-system-session\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-session\") pod \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\" (UID: \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\") " Dec 13 06:49:23 crc kubenswrapper[4971]: I1213 06:49:23.780459 4971 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/0b574797-001e-440a-8f4e-c0be86edad0f-proxy-tls\") pod \"0b574797-001e-440a-8f4e-c0be86edad0f\" (UID: \"0b574797-001e-440a-8f4e-c0be86edad0f\") " Dec 13 06:49:23 crc kubenswrapper[4971]: I1213 06:49:23.780487 4971 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-nzwt7\" (UniqueName: \"kubernetes.io/projected/96b93a3a-6083-4aea-8eab-fe1aa8245ad9-kube-api-access-nzwt7\") pod \"96b93a3a-6083-4aea-8eab-fe1aa8245ad9\" (UID: \"96b93a3a-6083-4aea-8eab-fe1aa8245ad9\") " Dec 13 06:49:23 crc kubenswrapper[4971]: I1213 06:49:23.780511 4971 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-audit-policies\") pod \"09ae3b1a-e8e7-4524-b54b-61eab6f9239a\" (UID: \"09ae3b1a-e8e7-4524-b54b-61eab6f9239a\") " Dec 13 06:49:23 crc kubenswrapper[4971]: I1213 06:49:23.780566 4971 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"machine-approver-tls\" (UniqueName: \"kubernetes.io/secret/22c825df-677d-4ca6-82db-3454ed06e783-machine-approver-tls\") pod \"22c825df-677d-4ca6-82db-3454ed06e783\" (UID: \"22c825df-677d-4ca6-82db-3454ed06e783\") " Dec 13 06:49:23 crc kubenswrapper[4971]: I1213 06:49:23.780594 4971 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/496e6271-fb68-4057-954e-a0d97a4afa3f-kube-api-access\") pod \"496e6271-fb68-4057-954e-a0d97a4afa3f\" (UID: \"496e6271-fb68-4057-954e-a0d97a4afa3f\") " Dec 13 06:49:23 crc kubenswrapper[4971]: I1213 06:49:23.780605 4971 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/5225d0e4-402f-4861-b410-819f433b1803-kube-api-access-9xfj7" (OuterVolumeSpecName: "kube-api-access-9xfj7") pod "5225d0e4-402f-4861-b410-819f433b1803" (UID: "5225d0e4-402f-4861-b410-819f433b1803"). InnerVolumeSpecName "kube-api-access-9xfj7". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 13 06:49:23 crc kubenswrapper[4971]: I1213 06:49:23.780617 4971 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/9d4552c7-cd75-42dd-8880-30dd377c49a4-config\") pod \"9d4552c7-cd75-42dd-8880-30dd377c49a4\" (UID: \"9d4552c7-cd75-42dd-8880-30dd377c49a4\") " Dec 13 06:49:23 crc kubenswrapper[4971]: I1213 06:49:23.780647 4971 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/5b88f790-22fa-440e-b583-365168c0b23d-metrics-certs\") pod \"5b88f790-22fa-440e-b583-365168c0b23d\" (UID: \"5b88f790-22fa-440e-b583-365168c0b23d\") " Dec 13 06:49:23 crc kubenswrapper[4971]: I1213 06:49:23.780677 4971 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/57a731c4-ef35-47a8-b875-bfb08a7f8011-utilities\") pod \"57a731c4-ef35-47a8-b875-bfb08a7f8011\" (UID: \"57a731c4-ef35-47a8-b875-bfb08a7f8011\") " Dec 13 06:49:23 crc kubenswrapper[4971]: I1213 06:49:23.780709 4971 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d-catalog-content\") pod \"b11524ee-3fca-4b1b-9cdf-6da289fdbc7d\" (UID: \"b11524ee-3fca-4b1b-9cdf-6da289fdbc7d\") " Dec 13 06:49:23 crc kubenswrapper[4971]: I1213 06:49:23.780746 4971 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"cni-sysctl-allowlist\" (UniqueName: \"kubernetes.io/configmap/7bb08738-c794-4ee8-9972-3a62ca171029-cni-sysctl-allowlist\") pod \"7bb08738-c794-4ee8-9972-3a62ca171029\" (UID: \"7bb08738-c794-4ee8-9972-3a62ca171029\") " Dec 13 06:49:23 crc kubenswrapper[4971]: I1213 06:49:23.780772 4971 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-apiservice-cert\") pod \"308be0ea-9f5f-4b29-aeb1-5abd31a0b17b\" (UID: \"308be0ea-9f5f-4b29-aeb1-5abd31a0b17b\") " Dec 13 06:49:23 crc kubenswrapper[4971]: I1213 06:49:23.780806 4971 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/7539238d-5fe0-46ed-884e-1c3b566537ec-config\") pod \"7539238d-5fe0-46ed-884e-1c3b566537ec\" (UID: \"7539238d-5fe0-46ed-884e-1c3b566537ec\") " Dec 13 06:49:23 crc kubenswrapper[4971]: I1213 06:49:23.780831 4971 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/496e6271-fb68-4057-954e-a0d97a4afa3f-config\") pod \"496e6271-fb68-4057-954e-a0d97a4afa3f\" (UID: \"496e6271-fb68-4057-954e-a0d97a4afa3f\") " Dec 13 06:49:23 crc kubenswrapper[4971]: I1213 06:49:23.780858 4971 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-rnphk\" (UniqueName: \"kubernetes.io/projected/bf126b07-da06-4140-9a57-dfd54fc6b486-kube-api-access-rnphk\") pod \"bf126b07-da06-4140-9a57-dfd54fc6b486\" (UID: \"bf126b07-da06-4140-9a57-dfd54fc6b486\") " Dec 13 06:49:23 crc kubenswrapper[4971]: I1213 06:49:23.780883 4971 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-kfwg7\" (UniqueName: \"kubernetes.io/projected/8f668bae-612b-4b75-9490-919e737c6a3b-kube-api-access-kfwg7\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 13 06:49:23 crc kubenswrapper[4971]: I1213 06:49:23.780910 4971 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/496e6271-fb68-4057-954e-a0d97a4afa3f-serving-cert\") pod \"496e6271-fb68-4057-954e-a0d97a4afa3f\" (UID: \"496e6271-fb68-4057-954e-a0d97a4afa3f\") " Dec 13 06:49:23 crc kubenswrapper[4971]: I1213 06:49:23.780976 4971 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d-serving-cert\") pod \"bc5039c0-ea34-426b-a2b7-fbbc87b49a6d\" (UID: \"bc5039c0-ea34-426b-a2b7-fbbc87b49a6d\") " Dec 13 06:49:23 crc kubenswrapper[4971]: I1213 06:49:23.781001 4971 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-trusted-ca-bundle\") pod \"43509403-f426-496e-be36-56cef71462f5\" (UID: \"43509403-f426-496e-be36-56cef71462f5\") " Dec 13 06:49:23 crc kubenswrapper[4971]: I1213 06:49:23.781031 4971 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/5225d0e4-402f-4861-b410-819f433b1803-utilities\") pod \"5225d0e4-402f-4861-b410-819f433b1803\" (UID: \"5225d0e4-402f-4861-b410-819f433b1803\") " Dec 13 06:49:23 crc kubenswrapper[4971]: I1213 06:49:23.781173 4971 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-mg5zb\" (UniqueName: \"kubernetes.io/projected/6402fda4-df10-493c-b4e5-d0569419652d-kube-api-access-mg5zb\") pod \"6402fda4-df10-493c-b4e5-d0569419652d\" (UID: \"6402fda4-df10-493c-b4e5-d0569419652d\") " Dec 13 06:49:23 crc kubenswrapper[4971]: I1213 06:49:23.781283 4971 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c-serving-cert\") pod \"210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c\" (UID: \"210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c\") " Dec 13 06:49:23 crc kubenswrapper[4971]: I1213 06:49:23.781720 4971 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-8tdtz\" (UniqueName: \"kubernetes.io/projected/09efc573-dbb6-4249-bd59-9b87aba8dd28-kube-api-access-8tdtz\") pod \"09efc573-dbb6-4249-bd59-9b87aba8dd28\" (UID: \"09efc573-dbb6-4249-bd59-9b87aba8dd28\") " Dec 13 06:49:23 crc kubenswrapper[4971]: I1213 06:49:23.781749 4971 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-etcd-client\") pod \"09ae3b1a-e8e7-4524-b54b-61eab6f9239a\" (UID: \"09ae3b1a-e8e7-4524-b54b-61eab6f9239a\") " Dec 13 06:49:23 crc kubenswrapper[4971]: I1213 06:49:23.781774 4971 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-w7l8j\" (UniqueName: \"kubernetes.io/projected/01ab3dd5-8196-46d0-ad33-122e2ca51def-kube-api-access-w7l8j\") pod \"01ab3dd5-8196-46d0-ad33-122e2ca51def\" (UID: \"01ab3dd5-8196-46d0-ad33-122e2ca51def\") " Dec 13 06:49:23 crc kubenswrapper[4971]: I1213 06:49:23.781802 4971 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/1d611f23-29be-4491-8495-bee1670e935f-catalog-content\") pod \"1d611f23-29be-4491-8495-bee1670e935f\" (UID: \"1d611f23-29be-4491-8495-bee1670e935f\") " Dec 13 06:49:23 crc kubenswrapper[4971]: I1213 06:49:23.781872 4971 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-tk88c\" (UniqueName: \"kubernetes.io/projected/7539238d-5fe0-46ed-884e-1c3b566537ec-kube-api-access-tk88c\") pod \"7539238d-5fe0-46ed-884e-1c3b566537ec\" (UID: \"7539238d-5fe0-46ed-884e-1c3b566537ec\") " Dec 13 06:49:23 crc kubenswrapper[4971]: I1213 06:49:23.782234 4971 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"console-config\" (UniqueName: \"kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-console-config\") pod \"43509403-f426-496e-be36-56cef71462f5\" (UID: \"43509403-f426-496e-be36-56cef71462f5\") " Dec 13 06:49:23 crc kubenswrapper[4971]: I1213 06:49:23.782280 4971 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"control-plane-machine-set-operator-tls\" (UniqueName: \"kubernetes.io/secret/6731426b-95fe-49ff-bb5f-40441049fde2-control-plane-machine-set-operator-tls\") pod \"6731426b-95fe-49ff-bb5f-40441049fde2\" (UID: \"6731426b-95fe-49ff-bb5f-40441049fde2\") " Dec 13 06:49:23 crc kubenswrapper[4971]: I1213 06:49:23.782307 4971 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-x4zgh\" (UniqueName: \"kubernetes.io/projected/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d-kube-api-access-x4zgh\") pod \"b11524ee-3fca-4b1b-9cdf-6da289fdbc7d\" (UID: \"b11524ee-3fca-4b1b-9cdf-6da289fdbc7d\") " Dec 13 06:49:23 crc kubenswrapper[4971]: I1213 06:49:23.782336 4971 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c-config\") pod \"210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c\" (UID: \"210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c\") " Dec 13 06:49:23 crc kubenswrapper[4971]: I1213 06:49:23.782365 4971 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-jhbk2\" (UniqueName: \"kubernetes.io/projected/bd23aa5c-e532-4e53-bccf-e79f130c5ae8-kube-api-access-jhbk2\") pod \"bd23aa5c-e532-4e53-bccf-e79f130c5ae8\" (UID: \"bd23aa5c-e532-4e53-bccf-e79f130c5ae8\") " Dec 13 06:49:23 crc kubenswrapper[4971]: I1213 06:49:23.782429 4971 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"encryption-config\" (UniqueName: \"kubernetes.io/secret/1bf7eb37-55a3-4c65-b768-a94c82151e69-encryption-config\") pod \"1bf7eb37-55a3-4c65-b768-a94c82151e69\" (UID: \"1bf7eb37-55a3-4c65-b768-a94c82151e69\") " Dec 13 06:49:23 crc kubenswrapper[4971]: I1213 06:49:23.782459 4971 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-gf66m\" (UniqueName: \"kubernetes.io/projected/a0128f3a-b052-44ed-a84e-c4c8aaf17c13-kube-api-access-gf66m\") pod \"a0128f3a-b052-44ed-a84e-c4c8aaf17c13\" (UID: \"a0128f3a-b052-44ed-a84e-c4c8aaf17c13\") " Dec 13 06:49:23 crc kubenswrapper[4971]: I1213 06:49:23.782487 4971 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/57a731c4-ef35-47a8-b875-bfb08a7f8011-catalog-content\") pod \"57a731c4-ef35-47a8-b875-bfb08a7f8011\" (UID: \"57a731c4-ef35-47a8-b875-bfb08a7f8011\") " Dec 13 06:49:23 crc kubenswrapper[4971]: I1213 06:49:23.782534 4971 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/6509e943-70c6-444c-bc41-48a544e36fbd-serving-cert\") pod \"6509e943-70c6-444c-bc41-48a544e36fbd\" (UID: \"6509e943-70c6-444c-bc41-48a544e36fbd\") " Dec 13 06:49:23 crc kubenswrapper[4971]: I1213 06:49:23.782559 4971 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/20b0d48f-5fd6-431c-a545-e3c800c7b866-cert\") pod \"20b0d48f-5fd6-431c-a545-e3c800c7b866\" (UID: \"20b0d48f-5fd6-431c-a545-e3c800c7b866\") " Dec 13 06:49:23 crc kubenswrapper[4971]: I1213 06:49:23.782586 4971 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-vt5rc\" (UniqueName: \"kubernetes.io/projected/44663579-783b-4372-86d6-acf235a62d72-kube-api-access-vt5rc\") pod \"44663579-783b-4372-86d6-acf235a62d72\" (UID: \"44663579-783b-4372-86d6-acf235a62d72\") " Dec 13 06:49:23 crc kubenswrapper[4971]: I1213 06:49:23.780672 4971 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/b6cd30de-2eeb-49a2-ab40-9167f4560ff5-kube-api-access-pj782" (OuterVolumeSpecName: "kube-api-access-pj782") pod "b6cd30de-2eeb-49a2-ab40-9167f4560ff5" (UID: "b6cd30de-2eeb-49a2-ab40-9167f4560ff5"). InnerVolumeSpecName "kube-api-access-pj782". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 13 06:49:23 crc kubenswrapper[4971]: I1213 06:49:23.780647 4971 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/7583ce53-e0fe-4a16-9e4d-50516596a136-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "7583ce53-e0fe-4a16-9e4d-50516596a136" (UID: "7583ce53-e0fe-4a16-9e4d-50516596a136"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 13 06:49:23 crc kubenswrapper[4971]: I1213 06:49:23.780726 4971 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/c03ee662-fb2f-4fc4-a2c1-af487c19d254-service-ca-bundle" (OuterVolumeSpecName: "service-ca-bundle") pod "c03ee662-fb2f-4fc4-a2c1-af487c19d254" (UID: "c03ee662-fb2f-4fc4-a2c1-af487c19d254"). InnerVolumeSpecName "service-ca-bundle". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 13 06:49:23 crc kubenswrapper[4971]: I1213 06:49:23.781604 4971 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-etcd-serving-ca" (OuterVolumeSpecName: "etcd-serving-ca") pod "09ae3b1a-e8e7-4524-b54b-61eab6f9239a" (UID: "09ae3b1a-e8e7-4524-b54b-61eab6f9239a"). InnerVolumeSpecName "etcd-serving-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 13 06:49:23 crc kubenswrapper[4971]: I1213 06:49:23.781622 4971 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-webhook-cert" (OuterVolumeSpecName: "webhook-cert") pod "308be0ea-9f5f-4b29-aeb1-5abd31a0b17b" (UID: "308be0ea-9f5f-4b29-aeb1-5abd31a0b17b"). InnerVolumeSpecName "webhook-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 13 06:49:23 crc kubenswrapper[4971]: I1213 06:49:23.781676 4971 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/6402fda4-df10-493c-b4e5-d0569419652d-config" (OuterVolumeSpecName: "config") pod "6402fda4-df10-493c-b4e5-d0569419652d" (UID: "6402fda4-df10-493c-b4e5-d0569419652d"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 13 06:49:23 crc kubenswrapper[4971]: I1213 06:49:23.781692 4971 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/3ab1a177-2de0-46d9-b765-d0d0649bb42e-kube-api-access-4d4hj" (OuterVolumeSpecName: "kube-api-access-4d4hj") pod "3ab1a177-2de0-46d9-b765-d0d0649bb42e" (UID: "3ab1a177-2de0-46d9-b765-d0d0649bb42e"). InnerVolumeSpecName "kube-api-access-4d4hj". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 13 06:49:23 crc kubenswrapper[4971]: I1213 06:49:23.781848 4971 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/57a731c4-ef35-47a8-b875-bfb08a7f8011-kube-api-access-cfbct" (OuterVolumeSpecName: "kube-api-access-cfbct") pod "57a731c4-ef35-47a8-b875-bfb08a7f8011" (UID: "57a731c4-ef35-47a8-b875-bfb08a7f8011"). InnerVolumeSpecName "kube-api-access-cfbct". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 13 06:49:23 crc kubenswrapper[4971]: I1213 06:49:23.781953 4971 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/96b93a3a-6083-4aea-8eab-fe1aa8245ad9-kube-api-access-nzwt7" (OuterVolumeSpecName: "kube-api-access-nzwt7") pod "96b93a3a-6083-4aea-8eab-fe1aa8245ad9" (UID: "96b93a3a-6083-4aea-8eab-fe1aa8245ad9"). InnerVolumeSpecName "kube-api-access-nzwt7". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 13 06:49:23 crc kubenswrapper[4971]: I1213 06:49:23.781955 4971 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/0b78653f-4ff9-4508-8672-245ed9b561e3-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "0b78653f-4ff9-4508-8672-245ed9b561e3" (UID: "0b78653f-4ff9-4508-8672-245ed9b561e3"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 13 06:49:23 crc kubenswrapper[4971]: I1213 06:49:23.782230 4971 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/9d4552c7-cd75-42dd-8880-30dd377c49a4-kube-api-access-pcxfs" (OuterVolumeSpecName: "kube-api-access-pcxfs") pod "9d4552c7-cd75-42dd-8880-30dd377c49a4" (UID: "9d4552c7-cd75-42dd-8880-30dd377c49a4"). InnerVolumeSpecName "kube-api-access-pcxfs". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 13 06:49:23 crc kubenswrapper[4971]: I1213 06:49:23.782422 4971 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/22c825df-677d-4ca6-82db-3454ed06e783-machine-approver-tls" (OuterVolumeSpecName: "machine-approver-tls") pod "22c825df-677d-4ca6-82db-3454ed06e783" (UID: "22c825df-677d-4ca6-82db-3454ed06e783"). InnerVolumeSpecName "machine-approver-tls". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 13 06:49:23 crc kubenswrapper[4971]: I1213 06:49:23.782754 4971 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/57a731c4-ef35-47a8-b875-bfb08a7f8011-utilities" (OuterVolumeSpecName: "utilities") pod "57a731c4-ef35-47a8-b875-bfb08a7f8011" (UID: "57a731c4-ef35-47a8-b875-bfb08a7f8011"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 13 06:49:23 crc kubenswrapper[4971]: I1213 06:49:23.782763 4971 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/7bb08738-c794-4ee8-9972-3a62ca171029-cni-binary-copy" (OuterVolumeSpecName: "cni-binary-copy") pod "7bb08738-c794-4ee8-9972-3a62ca171029" (UID: "7bb08738-c794-4ee8-9972-3a62ca171029"). InnerVolumeSpecName "cni-binary-copy". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 13 06:49:23 crc kubenswrapper[4971]: I1213 06:49:23.782775 4971 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-service-ca" (OuterVolumeSpecName: "v4-0-config-system-service-ca") pod "49c341d1-5089-4bc2-86a0-a5e165cfcc6b" (UID: "49c341d1-5089-4bc2-86a0-a5e165cfcc6b"). InnerVolumeSpecName "v4-0-config-system-service-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 13 06:49:23 crc kubenswrapper[4971]: I1213 06:49:23.782877 4971 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/6731426b-95fe-49ff-bb5f-40441049fde2-control-plane-machine-set-operator-tls" (OuterVolumeSpecName: "control-plane-machine-set-operator-tls") pod "6731426b-95fe-49ff-bb5f-40441049fde2" (UID: "6731426b-95fe-49ff-bb5f-40441049fde2"). InnerVolumeSpecName "control-plane-machine-set-operator-tls". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 13 06:49:23 crc kubenswrapper[4971]: I1213 06:49:23.783073 4971 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/496e6271-fb68-4057-954e-a0d97a4afa3f-kube-api-access" (OuterVolumeSpecName: "kube-api-access") pod "496e6271-fb68-4057-954e-a0d97a4afa3f" (UID: "496e6271-fb68-4057-954e-a0d97a4afa3f"). InnerVolumeSpecName "kube-api-access". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 13 06:49:23 crc kubenswrapper[4971]: I1213 06:49:23.783315 4971 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/5225d0e4-402f-4861-b410-819f433b1803-utilities" (OuterVolumeSpecName: "utilities") pod "5225d0e4-402f-4861-b410-819f433b1803" (UID: "5225d0e4-402f-4861-b410-819f433b1803"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 13 06:49:23 crc kubenswrapper[4971]: I1213 06:49:23.784304 4971 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-session" (OuterVolumeSpecName: "v4-0-config-system-session") pod "49c341d1-5089-4bc2-86a0-a5e165cfcc6b" (UID: "49c341d1-5089-4bc2-86a0-a5e165cfcc6b"). InnerVolumeSpecName "v4-0-config-system-session". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 13 06:49:23 crc kubenswrapper[4971]: I1213 06:49:23.784345 4971 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/0b574797-001e-440a-8f4e-c0be86edad0f-proxy-tls" (OuterVolumeSpecName: "proxy-tls") pod "0b574797-001e-440a-8f4e-c0be86edad0f" (UID: "0b574797-001e-440a-8f4e-c0be86edad0f"). InnerVolumeSpecName "proxy-tls". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 13 06:49:23 crc kubenswrapper[4971]: I1213 06:49:23.784362 4971 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/6402fda4-df10-493c-b4e5-d0569419652d-kube-api-access-mg5zb" (OuterVolumeSpecName: "kube-api-access-mg5zb") pod "6402fda4-df10-493c-b4e5-d0569419652d" (UID: "6402fda4-df10-493c-b4e5-d0569419652d"). InnerVolumeSpecName "kube-api-access-mg5zb". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 13 06:49:23 crc kubenswrapper[4971]: I1213 06:49:23.784486 4971 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/01ab3dd5-8196-46d0-ad33-122e2ca51def-kube-api-access-w7l8j" (OuterVolumeSpecName: "kube-api-access-w7l8j") pod "01ab3dd5-8196-46d0-ad33-122e2ca51def" (UID: "01ab3dd5-8196-46d0-ad33-122e2ca51def"). InnerVolumeSpecName "kube-api-access-w7l8j". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 13 06:49:23 crc kubenswrapper[4971]: I1213 06:49:23.784692 4971 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-etcd-client" (OuterVolumeSpecName: "etcd-client") pod "09ae3b1a-e8e7-4524-b54b-61eab6f9239a" (UID: "09ae3b1a-e8e7-4524-b54b-61eab6f9239a"). InnerVolumeSpecName "etcd-client". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 13 06:49:23 crc kubenswrapper[4971]: I1213 06:49:23.784712 4971 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/5b88f790-22fa-440e-b583-365168c0b23d-metrics-certs" (OuterVolumeSpecName: "metrics-certs") pod "5b88f790-22fa-440e-b583-365168c0b23d" (UID: "5b88f790-22fa-440e-b583-365168c0b23d"). InnerVolumeSpecName "metrics-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 13 06:49:23 crc kubenswrapper[4971]: I1213 06:49:23.784740 4971 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b-config" (OuterVolumeSpecName: "config") pod "8cea82b4-6893-4ddc-af9f-1bb5ae425c5b" (UID: "8cea82b4-6893-4ddc-af9f-1bb5ae425c5b"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 13 06:49:23 crc kubenswrapper[4971]: I1213 06:49:23.784809 4971 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-audit-policies" (OuterVolumeSpecName: "audit-policies") pod "09ae3b1a-e8e7-4524-b54b-61eab6f9239a" (UID: "09ae3b1a-e8e7-4524-b54b-61eab6f9239a"). InnerVolumeSpecName "audit-policies". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 13 06:49:23 crc kubenswrapper[4971]: I1213 06:49:23.784853 4971 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c" (UID: "210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 13 06:49:23 crc kubenswrapper[4971]: I1213 06:49:23.784821 4971 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "bc5039c0-ea34-426b-a2b7-fbbc87b49a6d" (UID: "bc5039c0-ea34-426b-a2b7-fbbc87b49a6d"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 13 06:49:23 crc kubenswrapper[4971]: I1213 06:49:23.784999 4971 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/7539238d-5fe0-46ed-884e-1c3b566537ec-config" (OuterVolumeSpecName: "config") pod "7539238d-5fe0-46ed-884e-1c3b566537ec" (UID: "7539238d-5fe0-46ed-884e-1c3b566537ec"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 13 06:49:23 crc kubenswrapper[4971]: I1213 06:49:23.785013 4971 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/496e6271-fb68-4057-954e-a0d97a4afa3f-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "496e6271-fb68-4057-954e-a0d97a4afa3f" (UID: "496e6271-fb68-4057-954e-a0d97a4afa3f"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 13 06:49:23 crc kubenswrapper[4971]: I1213 06:49:23.785049 4971 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-console-config" (OuterVolumeSpecName: "console-config") pod "43509403-f426-496e-be36-56cef71462f5" (UID: "43509403-f426-496e-be36-56cef71462f5"). InnerVolumeSpecName "console-config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 13 06:49:23 crc kubenswrapper[4971]: I1213 06:49:23.785060 4971 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/8f668bae-612b-4b75-9490-919e737c6a3b-kube-api-access-kfwg7" (OuterVolumeSpecName: "kube-api-access-kfwg7") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b"). InnerVolumeSpecName "kube-api-access-kfwg7". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 13 06:49:23 crc kubenswrapper[4971]: I1213 06:49:23.785162 4971 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/7bb08738-c794-4ee8-9972-3a62ca171029-cni-sysctl-allowlist" (OuterVolumeSpecName: "cni-sysctl-allowlist") pod "7bb08738-c794-4ee8-9972-3a62ca171029" (UID: "7bb08738-c794-4ee8-9972-3a62ca171029"). InnerVolumeSpecName "cni-sysctl-allowlist". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 13 06:49:23 crc kubenswrapper[4971]: I1213 06:49:23.782616 4971 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"registry-certificates\" (UniqueName: \"kubernetes.io/configmap/8f668bae-612b-4b75-9490-919e737c6a3b-registry-certificates\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 13 06:49:23 crc kubenswrapper[4971]: I1213 06:49:23.785620 4971 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"default-certificate\" (UniqueName: \"kubernetes.io/secret/c03ee662-fb2f-4fc4-a2c1-af487c19d254-default-certificate\") pod \"c03ee662-fb2f-4fc4-a2c1-af487c19d254\" (UID: \"c03ee662-fb2f-4fc4-a2c1-af487c19d254\") " Dec 13 06:49:23 crc kubenswrapper[4971]: I1213 06:49:23.785669 4971 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"cni-binary-copy\" (UniqueName: \"kubernetes.io/configmap/4bb40260-dbaa-4fb0-84df-5e680505d512-cni-binary-copy\") pod \"4bb40260-dbaa-4fb0-84df-5e680505d512\" (UID: \"4bb40260-dbaa-4fb0-84df-5e680505d512\") " Dec 13 06:49:23 crc kubenswrapper[4971]: I1213 06:49:23.785695 4971 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-249nr\" (UniqueName: \"kubernetes.io/projected/b6312bbd-5731-4ea0-a20f-81d5a57df44a-kube-api-access-249nr\") pod \"b6312bbd-5731-4ea0-a20f-81d5a57df44a\" (UID: \"b6312bbd-5731-4ea0-a20f-81d5a57df44a\") " Dec 13 06:49:23 crc kubenswrapper[4971]: I1213 06:49:23.785719 4971 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"multus-daemon-config\" (UniqueName: \"kubernetes.io/configmap/4bb40260-dbaa-4fb0-84df-5e680505d512-multus-daemon-config\") pod \"4bb40260-dbaa-4fb0-84df-5e680505d512\" (UID: \"4bb40260-dbaa-4fb0-84df-5e680505d512\") " Dec 13 06:49:23 crc kubenswrapper[4971]: I1213 06:49:23.785739 4971 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-d6qdx\" (UniqueName: \"kubernetes.io/projected/87cf06ed-a83f-41a7-828d-70653580a8cb-kube-api-access-d6qdx\") pod \"87cf06ed-a83f-41a7-828d-70653580a8cb\" (UID: \"87cf06ed-a83f-41a7-828d-70653580a8cb\") " Dec 13 06:49:23 crc kubenswrapper[4971]: I1213 06:49:23.785761 4971 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/09efc573-dbb6-4249-bd59-9b87aba8dd28-etcd-client\") pod \"09efc573-dbb6-4249-bd59-9b87aba8dd28\" (UID: \"09efc573-dbb6-4249-bd59-9b87aba8dd28\") " Dec 13 06:49:23 crc kubenswrapper[4971]: I1213 06:49:23.785784 4971 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/6509e943-70c6-444c-bc41-48a544e36fbd-trusted-ca-bundle\") pod \"6509e943-70c6-444c-bc41-48a544e36fbd\" (UID: \"6509e943-70c6-444c-bc41-48a544e36fbd\") " Dec 13 06:49:23 crc kubenswrapper[4971]: I1213 06:49:23.785172 4971 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/9d4552c7-cd75-42dd-8880-30dd377c49a4-config" (OuterVolumeSpecName: "config") pod "9d4552c7-cd75-42dd-8880-30dd377c49a4" (UID: "9d4552c7-cd75-42dd-8880-30dd377c49a4"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 13 06:49:23 crc kubenswrapper[4971]: I1213 06:49:23.785402 4971 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-apiservice-cert" (OuterVolumeSpecName: "apiservice-cert") pod "308be0ea-9f5f-4b29-aeb1-5abd31a0b17b" (UID: "308be0ea-9f5f-4b29-aeb1-5abd31a0b17b"). InnerVolumeSpecName "apiservice-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 13 06:49:23 crc kubenswrapper[4971]: I1213 06:49:23.785473 4971 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/fda69060-fa79-4696-b1a6-7980f124bf7c-mcd-auth-proxy-config" (OuterVolumeSpecName: "mcd-auth-proxy-config") pod "fda69060-fa79-4696-b1a6-7980f124bf7c" (UID: "fda69060-fa79-4696-b1a6-7980f124bf7c"). InnerVolumeSpecName "mcd-auth-proxy-config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 13 06:49:23 crc kubenswrapper[4971]: I1213 06:49:23.785472 4971 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-trusted-ca-bundle" (OuterVolumeSpecName: "trusted-ca-bundle") pod "43509403-f426-496e-be36-56cef71462f5" (UID: "43509403-f426-496e-be36-56cef71462f5"). InnerVolumeSpecName "trusted-ca-bundle". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 13 06:49:23 crc kubenswrapper[4971]: I1213 06:49:23.784554 4971 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/09efc573-dbb6-4249-bd59-9b87aba8dd28-kube-api-access-8tdtz" (OuterVolumeSpecName: "kube-api-access-8tdtz") pod "09efc573-dbb6-4249-bd59-9b87aba8dd28" (UID: "09efc573-dbb6-4249-bd59-9b87aba8dd28"). InnerVolumeSpecName "kube-api-access-8tdtz". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 13 06:49:23 crc kubenswrapper[4971]: I1213 06:49:23.785506 4971 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/925f1c65-6136-48ba-85aa-3a3b50560753-ovnkube-config" (OuterVolumeSpecName: "ovnkube-config") pod "925f1c65-6136-48ba-85aa-3a3b50560753" (UID: "925f1c65-6136-48ba-85aa-3a3b50560753"). InnerVolumeSpecName "ovnkube-config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 13 06:49:23 crc kubenswrapper[4971]: I1213 06:49:23.785778 4971 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/7539238d-5fe0-46ed-884e-1c3b566537ec-kube-api-access-tk88c" (OuterVolumeSpecName: "kube-api-access-tk88c") pod "7539238d-5fe0-46ed-884e-1c3b566537ec" (UID: "7539238d-5fe0-46ed-884e-1c3b566537ec"). InnerVolumeSpecName "kube-api-access-tk88c". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 13 06:49:23 crc kubenswrapper[4971]: I1213 06:49:23.785789 4971 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-audit" (OuterVolumeSpecName: "audit") pod "1bf7eb37-55a3-4c65-b768-a94c82151e69" (UID: "1bf7eb37-55a3-4c65-b768-a94c82151e69"). InnerVolumeSpecName "audit". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 13 06:49:23 crc kubenswrapper[4971]: I1213 06:49:23.785889 4971 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/8f668bae-612b-4b75-9490-919e737c6a3b-registry-certificates" (OuterVolumeSpecName: "registry-certificates") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b"). InnerVolumeSpecName "registry-certificates". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 13 06:49:23 crc kubenswrapper[4971]: I1213 06:49:23.786285 4971 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/bf126b07-da06-4140-9a57-dfd54fc6b486-kube-api-access-rnphk" (OuterVolumeSpecName: "kube-api-access-rnphk") pod "bf126b07-da06-4140-9a57-dfd54fc6b486" (UID: "bf126b07-da06-4140-9a57-dfd54fc6b486"). InnerVolumeSpecName "kube-api-access-rnphk". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 13 06:49:23 crc kubenswrapper[4971]: I1213 06:49:23.785957 4971 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/a31745f5-9847-4afe-82a5-3161cc66ca93-bound-sa-token\") pod \"a31745f5-9847-4afe-82a5-3161cc66ca93\" (UID: \"a31745f5-9847-4afe-82a5-3161cc66ca93\") " Dec 13 06:49:23 crc kubenswrapper[4971]: I1213 06:49:23.786625 4971 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"image-registry-operator-tls\" (UniqueName: \"kubernetes.io/secret/bf126b07-da06-4140-9a57-dfd54fc6b486-image-registry-operator-tls\") pod \"bf126b07-da06-4140-9a57-dfd54fc6b486\" (UID: \"bf126b07-da06-4140-9a57-dfd54fc6b486\") " Dec 13 06:49:23 crc kubenswrapper[4971]: I1213 06:49:23.786652 4971 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"signing-key\" (UniqueName: \"kubernetes.io/secret/25e176fe-21b4-4974-b1ed-c8b94f112a7f-signing-key\") pod \"25e176fe-21b4-4974-b1ed-c8b94f112a7f\" (UID: \"25e176fe-21b4-4974-b1ed-c8b94f112a7f\") " Dec 13 06:49:23 crc kubenswrapper[4971]: I1213 06:49:23.786847 4971 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/5225d0e4-402f-4861-b410-819f433b1803-catalog-content\") pod \"5225d0e4-402f-4861-b410-819f433b1803\" (UID: \"5225d0e4-402f-4861-b410-819f433b1803\") " Dec 13 06:49:23 crc kubenswrapper[4971]: I1213 06:49:23.786899 4971 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-audit-policies\") pod \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\" (UID: \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\") " Dec 13 06:49:23 crc kubenswrapper[4971]: I1213 06:49:23.787004 4971 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/7539238d-5fe0-46ed-884e-1c3b566537ec-serving-cert\") pod \"7539238d-5fe0-46ed-884e-1c3b566537ec\" (UID: \"7539238d-5fe0-46ed-884e-1c3b566537ec\") " Dec 13 06:49:23 crc kubenswrapper[4971]: I1213 06:49:23.787038 4971 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"machine-api-operator-tls\" (UniqueName: \"kubernetes.io/secret/6402fda4-df10-493c-b4e5-d0569419652d-machine-api-operator-tls\") pod \"6402fda4-df10-493c-b4e5-d0569419652d\" (UID: \"6402fda4-df10-493c-b4e5-d0569419652d\") " Dec 13 06:49:23 crc kubenswrapper[4971]: I1213 06:49:23.787097 4971 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-d4lsv\" (UniqueName: \"kubernetes.io/projected/25e176fe-21b4-4974-b1ed-c8b94f112a7f-kube-api-access-d4lsv\") pod \"25e176fe-21b4-4974-b1ed-c8b94f112a7f\" (UID: \"25e176fe-21b4-4974-b1ed-c8b94f112a7f\") " Dec 13 06:49:23 crc kubenswrapper[4971]: I1213 06:49:23.787145 4971 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"marketplace-trusted-ca\" (UniqueName: \"kubernetes.io/configmap/b6cd30de-2eeb-49a2-ab40-9167f4560ff5-marketplace-trusted-ca\") pod \"b6cd30de-2eeb-49a2-ab40-9167f4560ff5\" (UID: \"b6cd30de-2eeb-49a2-ab40-9167f4560ff5\") " Dec 13 06:49:23 crc kubenswrapper[4971]: I1213 06:49:23.787217 4971 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-qg5z5\" (UniqueName: \"kubernetes.io/projected/43509403-f426-496e-be36-56cef71462f5-kube-api-access-qg5z5\") pod \"43509403-f426-496e-be36-56cef71462f5\" (UID: \"43509403-f426-496e-be36-56cef71462f5\") " Dec 13 06:49:23 crc kubenswrapper[4971]: I1213 06:49:23.787236 4971 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-lz9wn\" (UniqueName: \"kubernetes.io/projected/a31745f5-9847-4afe-82a5-3161cc66ca93-kube-api-access-lz9wn\") pod \"a31745f5-9847-4afe-82a5-3161cc66ca93\" (UID: \"a31745f5-9847-4afe-82a5-3161cc66ca93\") " Dec 13 06:49:23 crc kubenswrapper[4971]: I1213 06:49:23.787253 4971 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-serving-cert\") pod \"09ae3b1a-e8e7-4524-b54b-61eab6f9239a\" (UID: \"09ae3b1a-e8e7-4524-b54b-61eab6f9239a\") " Dec 13 06:49:23 crc kubenswrapper[4971]: I1213 06:49:23.787290 4971 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/5441d097-087c-4d9a-baa8-b210afa90fc9-serving-cert\") pod \"5441d097-087c-4d9a-baa8-b210afa90fc9\" (UID: \"5441d097-087c-4d9a-baa8-b210afa90fc9\") " Dec 13 06:49:23 crc kubenswrapper[4971]: I1213 06:49:23.787306 4971 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serviceca\" (UniqueName: \"kubernetes.io/configmap/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59-serviceca\") pod \"3cb93b32-e0ae-4377-b9c8-fdb9842c6d59\" (UID: \"3cb93b32-e0ae-4377-b9c8-fdb9842c6d59\") " Dec 13 06:49:23 crc kubenswrapper[4971]: I1213 06:49:23.787322 4971 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-service-ca\") pod \"43509403-f426-496e-be36-56cef71462f5\" (UID: \"43509403-f426-496e-be36-56cef71462f5\") " Dec 13 06:49:23 crc kubenswrapper[4971]: I1213 06:49:23.787338 4971 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-jkwtn\" (UniqueName: \"kubernetes.io/projected/5b88f790-22fa-440e-b583-365168c0b23d-kube-api-access-jkwtn\") pod \"5b88f790-22fa-440e-b583-365168c0b23d\" (UID: \"5b88f790-22fa-440e-b583-365168c0b23d\") " Dec 13 06:49:23 crc kubenswrapper[4971]: I1213 06:49:23.787356 4971 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"package-server-manager-serving-cert\" (UniqueName: \"kubernetes.io/secret/3ab1a177-2de0-46d9-b765-d0d0649bb42e-package-server-manager-serving-cert\") pod \"3ab1a177-2de0-46d9-b765-d0d0649bb42e\" (UID: \"3ab1a177-2de0-46d9-b765-d0d0649bb42e\") " Dec 13 06:49:23 crc kubenswrapper[4971]: I1213 06:49:23.787372 4971 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/9d4552c7-cd75-42dd-8880-30dd377c49a4-serving-cert\") pod \"9d4552c7-cd75-42dd-8880-30dd377c49a4\" (UID: \"9d4552c7-cd75-42dd-8880-30dd377c49a4\") " Dec 13 06:49:23 crc kubenswrapper[4971]: I1213 06:49:23.787390 4971 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-user-template-login\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-template-login\") pod \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\" (UID: \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\") " Dec 13 06:49:23 crc kubenswrapper[4971]: I1213 06:49:23.787407 4971 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-config\") pod \"1bf7eb37-55a3-4c65-b768-a94c82151e69\" (UID: \"1bf7eb37-55a3-4c65-b768-a94c82151e69\") " Dec 13 06:49:23 crc kubenswrapper[4971]: I1213 06:49:23.787426 4971 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-v47cf\" (UniqueName: \"kubernetes.io/projected/c03ee662-fb2f-4fc4-a2c1-af487c19d254-kube-api-access-v47cf\") pod \"c03ee662-fb2f-4fc4-a2c1-af487c19d254\" (UID: \"c03ee662-fb2f-4fc4-a2c1-af487c19d254\") " Dec 13 06:49:23 crc kubenswrapper[4971]: I1213 06:49:23.787442 4971 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/87cf06ed-a83f-41a7-828d-70653580a8cb-metrics-tls\") pod \"87cf06ed-a83f-41a7-828d-70653580a8cb\" (UID: \"87cf06ed-a83f-41a7-828d-70653580a8cb\") " Dec 13 06:49:23 crc kubenswrapper[4971]: I1213 06:49:23.787458 4971 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/1bf7eb37-55a3-4c65-b768-a94c82151e69-serving-cert\") pod \"1bf7eb37-55a3-4c65-b768-a94c82151e69\" (UID: \"1bf7eb37-55a3-4c65-b768-a94c82151e69\") " Dec 13 06:49:23 crc kubenswrapper[4971]: I1213 06:49:23.787473 4971 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/1386a44e-36a2-460c-96d0-0359d2b6f0f5-config\") pod \"1386a44e-36a2-460c-96d0-0359d2b6f0f5\" (UID: \"1386a44e-36a2-460c-96d0-0359d2b6f0f5\") " Dec 13 06:49:23 crc kubenswrapper[4971]: I1213 06:49:23.787490 4971 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/e7e6199b-1264-4501-8953-767f51328d08-serving-cert\") pod \"e7e6199b-1264-4501-8953-767f51328d08\" (UID: \"e7e6199b-1264-4501-8953-767f51328d08\") " Dec 13 06:49:23 crc kubenswrapper[4971]: I1213 06:49:23.787508 4971 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-7c4vf\" (UniqueName: \"kubernetes.io/projected/22c825df-677d-4ca6-82db-3454ed06e783-kube-api-access-7c4vf\") pod \"22c825df-677d-4ca6-82db-3454ed06e783\" (UID: \"22c825df-677d-4ca6-82db-3454ed06e783\") " Dec 13 06:49:23 crc kubenswrapper[4971]: I1213 06:49:23.787538 4971 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-279lb\" (UniqueName: \"kubernetes.io/projected/7bb08738-c794-4ee8-9972-3a62ca171029-kube-api-access-279lb\") pod \"7bb08738-c794-4ee8-9972-3a62ca171029\" (UID: \"7bb08738-c794-4ee8-9972-3a62ca171029\") " Dec 13 06:49:23 crc kubenswrapper[4971]: I1213 06:49:23.787553 4971 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-user-idp-0-file-data\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-idp-0-file-data\") pod \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\" (UID: \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\") " Dec 13 06:49:23 crc kubenswrapper[4971]: I1213 06:49:23.787570 4971 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/7583ce53-e0fe-4a16-9e4d-50516596a136-client-ca\") pod \"7583ce53-e0fe-4a16-9e4d-50516596a136\" (UID: \"7583ce53-e0fe-4a16-9e4d-50516596a136\") " Dec 13 06:49:23 crc kubenswrapper[4971]: I1213 06:49:23.787585 4971 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"tmpfs\" (UniqueName: \"kubernetes.io/empty-dir/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-tmpfs\") pod \"308be0ea-9f5f-4b29-aeb1-5abd31a0b17b\" (UID: \"308be0ea-9f5f-4b29-aeb1-5abd31a0b17b\") " Dec 13 06:49:23 crc kubenswrapper[4971]: I1213 06:49:23.787599 4971 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/87cf06ed-a83f-41a7-828d-70653580a8cb-config-volume\") pod \"87cf06ed-a83f-41a7-828d-70653580a8cb\" (UID: \"87cf06ed-a83f-41a7-828d-70653580a8cb\") " Dec 13 06:49:23 crc kubenswrapper[4971]: I1213 06:49:23.787616 4971 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-sb6h7\" (UniqueName: \"kubernetes.io/projected/1bf7eb37-55a3-4c65-b768-a94c82151e69-kube-api-access-sb6h7\") pod \"1bf7eb37-55a3-4c65-b768-a94c82151e69\" (UID: \"1bf7eb37-55a3-4c65-b768-a94c82151e69\") " Dec 13 06:49:23 crc kubenswrapper[4971]: I1213 06:49:23.787630 4971 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b-serving-cert\") pod \"8cea82b4-6893-4ddc-af9f-1bb5ae425c5b\" (UID: \"8cea82b4-6893-4ddc-af9f-1bb5ae425c5b\") " Dec 13 06:49:23 crc kubenswrapper[4971]: I1213 06:49:23.787645 4971 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/0b78653f-4ff9-4508-8672-245ed9b561e3-service-ca\") pod \"0b78653f-4ff9-4508-8672-245ed9b561e3\" (UID: \"0b78653f-4ff9-4508-8672-245ed9b561e3\") " Dec 13 06:49:23 crc kubenswrapper[4971]: I1213 06:49:23.787660 4971 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-user-template-error\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-template-error\") pod \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\" (UID: \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\") " Dec 13 06:49:23 crc kubenswrapper[4971]: I1213 06:49:23.787675 4971 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/1d611f23-29be-4491-8495-bee1670e935f-utilities\") pod \"1d611f23-29be-4491-8495-bee1670e935f\" (UID: \"1d611f23-29be-4491-8495-bee1670e935f\") " Dec 13 06:49:23 crc kubenswrapper[4971]: I1213 06:49:23.787690 4971 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-zkvpv\" (UniqueName: \"kubernetes.io/projected/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-kube-api-access-zkvpv\") pod \"09ae3b1a-e8e7-4524-b54b-61eab6f9239a\" (UID: \"09ae3b1a-e8e7-4524-b54b-61eab6f9239a\") " Dec 13 06:49:23 crc kubenswrapper[4971]: I1213 06:49:23.787705 4971 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"console-serving-cert\" (UniqueName: \"kubernetes.io/secret/43509403-f426-496e-be36-56cef71462f5-console-serving-cert\") pod \"43509403-f426-496e-be36-56cef71462f5\" (UID: \"43509403-f426-496e-be36-56cef71462f5\") " Dec 13 06:49:23 crc kubenswrapper[4971]: I1213 06:49:23.787720 4971 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/1386a44e-36a2-460c-96d0-0359d2b6f0f5-serving-cert\") pod \"1386a44e-36a2-460c-96d0-0359d2b6f0f5\" (UID: \"1386a44e-36a2-460c-96d0-0359d2b6f0f5\") " Dec 13 06:49:23 crc kubenswrapper[4971]: I1213 06:49:23.787734 4971 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/31d8b7a1-420e-4252-a5b7-eebe8a111292-auth-proxy-config\") pod \"31d8b7a1-420e-4252-a5b7-eebe8a111292\" (UID: \"31d8b7a1-420e-4252-a5b7-eebe8a111292\") " Dec 13 06:49:23 crc kubenswrapper[4971]: I1213 06:49:23.787751 4971 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/fda69060-fa79-4696-b1a6-7980f124bf7c-proxy-tls\") pod \"fda69060-fa79-4696-b1a6-7980f124bf7c\" (UID: \"fda69060-fa79-4696-b1a6-7980f124bf7c\") " Dec 13 06:49:23 crc kubenswrapper[4971]: I1213 06:49:23.787767 4971 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovn-control-plane-metrics-cert\" (UniqueName: \"kubernetes.io/secret/925f1c65-6136-48ba-85aa-3a3b50560753-ovn-control-plane-metrics-cert\") pod \"925f1c65-6136-48ba-85aa-3a3b50560753\" (UID: \"925f1c65-6136-48ba-85aa-3a3b50560753\") " Dec 13 06:49:23 crc kubenswrapper[4971]: I1213 06:49:23.787782 4971 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/7583ce53-e0fe-4a16-9e4d-50516596a136-proxy-ca-bundles\") pod \"7583ce53-e0fe-4a16-9e4d-50516596a136\" (UID: \"7583ce53-e0fe-4a16-9e4d-50516596a136\") " Dec 13 06:49:23 crc kubenswrapper[4971]: I1213 06:49:23.787798 4971 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/7583ce53-e0fe-4a16-9e4d-50516596a136-config\") pod \"7583ce53-e0fe-4a16-9e4d-50516596a136\" (UID: \"7583ce53-e0fe-4a16-9e4d-50516596a136\") " Dec 13 06:49:23 crc kubenswrapper[4971]: I1213 06:49:23.787816 4971 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-qs4fp\" (UniqueName: \"kubernetes.io/projected/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c-kube-api-access-qs4fp\") pod \"210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c\" (UID: \"210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c\") " Dec 13 06:49:23 crc kubenswrapper[4971]: I1213 06:49:23.787877 4971 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"images\" (UniqueName: \"kubernetes.io/configmap/6402fda4-df10-493c-b4e5-d0569419652d-images\") pod \"6402fda4-df10-493c-b4e5-d0569419652d\" (UID: \"6402fda4-df10-493c-b4e5-d0569419652d\") " Dec 13 06:49:23 crc kubenswrapper[4971]: I1213 06:49:23.787897 4971 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/a31745f5-9847-4afe-82a5-3161cc66ca93-trusted-ca\") pod \"a31745f5-9847-4afe-82a5-3161cc66ca93\" (UID: \"a31745f5-9847-4afe-82a5-3161cc66ca93\") " Dec 13 06:49:23 crc kubenswrapper[4971]: I1213 06:49:23.787930 4971 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-s4n52\" (UniqueName: \"kubernetes.io/projected/925f1c65-6136-48ba-85aa-3a3b50560753-kube-api-access-s4n52\") pod \"925f1c65-6136-48ba-85aa-3a3b50560753\" (UID: \"925f1c65-6136-48ba-85aa-3a3b50560753\") " Dec 13 06:49:23 crc kubenswrapper[4971]: I1213 06:49:23.787932 4971 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/6509e943-70c6-444c-bc41-48a544e36fbd-trusted-ca-bundle" (OuterVolumeSpecName: "trusted-ca-bundle") pod "6509e943-70c6-444c-bc41-48a544e36fbd" (UID: "6509e943-70c6-444c-bc41-48a544e36fbd"). InnerVolumeSpecName "trusted-ca-bundle". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 13 06:49:23 crc kubenswrapper[4971]: I1213 06:49:23.787945 4971 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/c03ee662-fb2f-4fc4-a2c1-af487c19d254-metrics-certs\") pod \"c03ee662-fb2f-4fc4-a2c1-af487c19d254\" (UID: \"c03ee662-fb2f-4fc4-a2c1-af487c19d254\") " Dec 13 06:49:23 crc kubenswrapper[4971]: I1213 06:49:23.788020 4971 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/6ea678ab-3438-413e-bfe3-290ae7725660-ovnkube-config\") pod \"6ea678ab-3438-413e-bfe3-290ae7725660\" (UID: \"6ea678ab-3438-413e-bfe3-290ae7725660\") " Dec 13 06:49:23 crc kubenswrapper[4971]: I1213 06:49:23.788036 4971 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/1386a44e-36a2-460c-96d0-0359d2b6f0f5-kube-api-access\") pod \"1386a44e-36a2-460c-96d0-0359d2b6f0f5\" (UID: \"1386a44e-36a2-460c-96d0-0359d2b6f0f5\") " Dec 13 06:49:23 crc kubenswrapper[4971]: I1213 06:49:23.788051 4971 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/5441d097-087c-4d9a-baa8-b210afa90fc9-client-ca\") pod \"5441d097-087c-4d9a-baa8-b210afa90fc9\" (UID: \"5441d097-087c-4d9a-baa8-b210afa90fc9\") " Dec 13 06:49:23 crc kubenswrapper[4971]: I1213 06:49:23.788066 4971 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/0b78653f-4ff9-4508-8672-245ed9b561e3-kube-api-access\") pod \"0b78653f-4ff9-4508-8672-245ed9b561e3\" (UID: \"0b78653f-4ff9-4508-8672-245ed9b561e3\") " Dec 13 06:49:23 crc kubenswrapper[4971]: I1213 06:49:23.788082 4971 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"profile-collector-cert\" (UniqueName: \"kubernetes.io/secret/b6312bbd-5731-4ea0-a20f-81d5a57df44a-profile-collector-cert\") pod \"b6312bbd-5731-4ea0-a20f-81d5a57df44a\" (UID: \"b6312bbd-5731-4ea0-a20f-81d5a57df44a\") " Dec 13 06:49:23 crc kubenswrapper[4971]: I1213 06:49:23.788097 4971 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-system-trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-trusted-ca-bundle\") pod \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\" (UID: \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\") " Dec 13 06:49:23 crc kubenswrapper[4971]: I1213 06:49:23.788115 4971 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-6g6sz\" (UniqueName: \"kubernetes.io/projected/6509e943-70c6-444c-bc41-48a544e36fbd-kube-api-access-6g6sz\") pod \"6509e943-70c6-444c-bc41-48a544e36fbd\" (UID: \"6509e943-70c6-444c-bc41-48a544e36fbd\") " Dec 13 06:49:23 crc kubenswrapper[4971]: I1213 06:49:23.788130 4971 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/e7e6199b-1264-4501-8953-767f51328d08-kube-api-access\") pod \"e7e6199b-1264-4501-8953-767f51328d08\" (UID: \"e7e6199b-1264-4501-8953-767f51328d08\") " Dec 13 06:49:23 crc kubenswrapper[4971]: I1213 06:49:23.788145 4971 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/9d4552c7-cd75-42dd-8880-30dd377c49a4-trusted-ca\") pod \"9d4552c7-cd75-42dd-8880-30dd377c49a4\" (UID: \"9d4552c7-cd75-42dd-8880-30dd377c49a4\") " Dec 13 06:49:23 crc kubenswrapper[4971]: I1213 06:49:23.788162 4971 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-xcphl\" (UniqueName: \"kubernetes.io/projected/7583ce53-e0fe-4a16-9e4d-50516596a136-kube-api-access-xcphl\") pod \"7583ce53-e0fe-4a16-9e4d-50516596a136\" (UID: \"7583ce53-e0fe-4a16-9e4d-50516596a136\") " Dec 13 06:49:23 crc kubenswrapper[4971]: I1213 06:49:23.788179 4971 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/09efc573-dbb6-4249-bd59-9b87aba8dd28-config\") pod \"09efc573-dbb6-4249-bd59-9b87aba8dd28\" (UID: \"09efc573-dbb6-4249-bd59-9b87aba8dd28\") " Dec 13 06:49:23 crc kubenswrapper[4971]: I1213 06:49:23.788194 4971 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-system-serving-cert\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-serving-cert\") pod \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\" (UID: \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\") " Dec 13 06:49:23 crc kubenswrapper[4971]: I1213 06:49:23.788210 4971 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"encryption-config\" (UniqueName: \"kubernetes.io/secret/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-encryption-config\") pod \"09ae3b1a-e8e7-4524-b54b-61eab6f9239a\" (UID: \"09ae3b1a-e8e7-4524-b54b-61eab6f9239a\") " Dec 13 06:49:23 crc kubenswrapper[4971]: I1213 06:49:23.788228 4971 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-bf2bz\" (UniqueName: \"kubernetes.io/projected/1d611f23-29be-4491-8495-bee1670e935f-kube-api-access-bf2bz\") pod \"1d611f23-29be-4491-8495-bee1670e935f\" (UID: \"1d611f23-29be-4491-8495-bee1670e935f\") " Dec 13 06:49:23 crc kubenswrapper[4971]: I1213 06:49:23.788243 4971 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovnkube-script-lib\" (UniqueName: \"kubernetes.io/configmap/6ea678ab-3438-413e-bfe3-290ae7725660-ovnkube-script-lib\") pod \"6ea678ab-3438-413e-bfe3-290ae7725660\" (UID: \"6ea678ab-3438-413e-bfe3-290ae7725660\") " Dec 13 06:49:23 crc kubenswrapper[4971]: I1213 06:49:23.788259 4971 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/1bf7eb37-55a3-4c65-b768-a94c82151e69-etcd-client\") pod \"1bf7eb37-55a3-4c65-b768-a94c82151e69\" (UID: \"1bf7eb37-55a3-4c65-b768-a94c82151e69\") " Dec 13 06:49:23 crc kubenswrapper[4971]: I1213 06:49:23.788275 4971 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"console-oauth-config\" (UniqueName: \"kubernetes.io/secret/43509403-f426-496e-be36-56cef71462f5-console-oauth-config\") pod \"43509403-f426-496e-be36-56cef71462f5\" (UID: \"43509403-f426-496e-be36-56cef71462f5\") " Dec 13 06:49:23 crc kubenswrapper[4971]: I1213 06:49:23.788339 4971 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-6ccd8\" (UniqueName: \"kubernetes.io/projected/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-kube-api-access-6ccd8\") pod \"308be0ea-9f5f-4b29-aeb1-5abd31a0b17b\" (UID: \"308be0ea-9f5f-4b29-aeb1-5abd31a0b17b\") " Dec 13 06:49:23 crc kubenswrapper[4971]: I1213 06:49:23.788358 4971 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/31d8b7a1-420e-4252-a5b7-eebe8a111292-proxy-tls\") pod \"31d8b7a1-420e-4252-a5b7-eebe8a111292\" (UID: \"31d8b7a1-420e-4252-a5b7-eebe8a111292\") " Dec 13 06:49:23 crc kubenswrapper[4971]: I1213 06:49:23.788376 4971 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-zgdk5\" (UniqueName: \"kubernetes.io/projected/31d8b7a1-420e-4252-a5b7-eebe8a111292-kube-api-access-zgdk5\") pod \"31d8b7a1-420e-4252-a5b7-eebe8a111292\" (UID: \"31d8b7a1-420e-4252-a5b7-eebe8a111292\") " Dec 13 06:49:23 crc kubenswrapper[4971]: I1213 06:49:23.788393 4971 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/6ea678ab-3438-413e-bfe3-290ae7725660-env-overrides\") pod \"6ea678ab-3438-413e-bfe3-290ae7725660\" (UID: \"6ea678ab-3438-413e-bfe3-290ae7725660\") " Dec 13 06:49:23 crc kubenswrapper[4971]: I1213 06:49:23.788429 4971 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"images\" (UniqueName: \"kubernetes.io/configmap/31d8b7a1-420e-4252-a5b7-eebe8a111292-images\") pod \"31d8b7a1-420e-4252-a5b7-eebe8a111292\" (UID: \"31d8b7a1-420e-4252-a5b7-eebe8a111292\") " Dec 13 06:49:23 crc kubenswrapper[4971]: I1213 06:49:23.788478 4971 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/22c825df-677d-4ca6-82db-3454ed06e783-auth-proxy-config\") pod \"22c825df-677d-4ca6-82db-3454ed06e783\" (UID: \"22c825df-677d-4ca6-82db-3454ed06e783\") " Dec 13 06:49:23 crc kubenswrapper[4971]: I1213 06:49:23.788524 4971 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"profile-collector-cert\" (UniqueName: \"kubernetes.io/secret/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9-profile-collector-cert\") pod \"f88749ec-7931-4ee7-b3fc-1ec5e11f92e9\" (UID: \"f88749ec-7931-4ee7-b3fc-1ec5e11f92e9\") " Dec 13 06:49:23 crc kubenswrapper[4971]: I1213 06:49:23.788543 4971 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/5441d097-087c-4d9a-baa8-b210afa90fc9-config\") pod \"5441d097-087c-4d9a-baa8-b210afa90fc9\" (UID: \"5441d097-087c-4d9a-baa8-b210afa90fc9\") " Dec 13 06:49:23 crc kubenswrapper[4971]: I1213 06:49:23.788560 4971 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-htfz6\" (UniqueName: \"kubernetes.io/projected/6ea678ab-3438-413e-bfe3-290ae7725660-kube-api-access-htfz6\") pod \"6ea678ab-3438-413e-bfe3-290ae7725660\" (UID: \"6ea678ab-3438-413e-bfe3-290ae7725660\") " Dec 13 06:49:23 crc kubenswrapper[4971]: I1213 06:49:23.788577 4971 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-system-ocp-branding-template\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-ocp-branding-template\") pod \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\" (UID: \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\") " Dec 13 06:49:23 crc kubenswrapper[4971]: I1213 06:49:23.788595 4971 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/22c825df-677d-4ca6-82db-3454ed06e783-config\") pod \"22c825df-677d-4ca6-82db-3454ed06e783\" (UID: \"22c825df-677d-4ca6-82db-3454ed06e783\") " Dec 13 06:49:23 crc kubenswrapper[4971]: I1213 06:49:23.788611 4971 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"signing-cabundle\" (UniqueName: \"kubernetes.io/configmap/25e176fe-21b4-4974-b1ed-c8b94f112a7f-signing-cabundle\") pod \"25e176fe-21b4-4974-b1ed-c8b94f112a7f\" (UID: \"25e176fe-21b4-4974-b1ed-c8b94f112a7f\") " Dec 13 06:49:23 crc kubenswrapper[4971]: I1213 06:49:23.788647 4971 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-2d4wz\" (UniqueName: \"kubernetes.io/projected/5441d097-087c-4d9a-baa8-b210afa90fc9-kube-api-access-2d4wz\") pod \"5441d097-087c-4d9a-baa8-b210afa90fc9\" (UID: \"5441d097-087c-4d9a-baa8-b210afa90fc9\") " Dec 13 06:49:23 crc kubenswrapper[4971]: I1213 06:49:23.788697 4971 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/bf126b07-da06-4140-9a57-dfd54fc6b486-bound-sa-token\") pod \"bf126b07-da06-4140-9a57-dfd54fc6b486\" (UID: \"bf126b07-da06-4140-9a57-dfd54fc6b486\") " Dec 13 06:49:23 crc kubenswrapper[4971]: I1213 06:49:23.788715 4971 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etcd-ca\" (UniqueName: \"kubernetes.io/configmap/09efc573-dbb6-4249-bd59-9b87aba8dd28-etcd-ca\") pod \"09efc573-dbb6-4249-bd59-9b87aba8dd28\" (UID: \"09efc573-dbb6-4249-bd59-9b87aba8dd28\") " Dec 13 06:49:23 crc kubenswrapper[4971]: I1213 06:49:23.789153 4971 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-x7zkh\" (UniqueName: \"kubernetes.io/projected/6731426b-95fe-49ff-bb5f-40441049fde2-kube-api-access-x7zkh\") pod \"6731426b-95fe-49ff-bb5f-40441049fde2\" (UID: \"6731426b-95fe-49ff-bb5f-40441049fde2\") " Dec 13 06:49:23 crc kubenswrapper[4971]: I1213 06:49:23.789173 4971 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-system-cliconfig\" (UniqueName: \"kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-cliconfig\") pod \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\" (UID: \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\") " Dec 13 06:49:23 crc kubenswrapper[4971]: I1213 06:49:23.789304 4971 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cqllr\" (UniqueName: \"kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr\") pod \"network-check-target-xd92c\" (UID: \"3b6479f0-333b-4a96-9adf-2099afdc2447\") " pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 13 06:49:23 crc kubenswrapper[4971]: I1213 06:49:23.789327 4971 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-s2kz5\" (UniqueName: \"kubernetes.io/projected/ef543e1b-8068-4ea3-b32a-61027b32e95d-kube-api-access-s2kz5\") pod \"network-node-identity-vrzqb\" (UID: \"ef543e1b-8068-4ea3-b32a-61027b32e95d\") " pod="openshift-network-node-identity/network-node-identity-vrzqb" Dec 13 06:49:23 crc kubenswrapper[4971]: I1213 06:49:23.789350 4971 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-etc-kube\" (UniqueName: \"kubernetes.io/host-path/37a5e44f-9a88-4405-be8a-b645485e7312-host-etc-kube\") pod \"network-operator-58b4c7f79c-55gtf\" (UID: \"37a5e44f-9a88-4405-be8a-b645485e7312\") " pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" Dec 13 06:49:23 crc kubenswrapper[4971]: I1213 06:49:23.789377 4971 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-slash\" (UniqueName: \"kubernetes.io/host-path/d75a4c96-2883-4a0b-bab2-0fab2b6c0b49-host-slash\") pod \"iptables-alerter-4ln5h\" (UID: \"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\") " pod="openshift-network-operator/iptables-alerter-4ln5h" Dec 13 06:49:23 crc kubenswrapper[4971]: I1213 06:49:23.789398 4971 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-rdwmf\" (UniqueName: \"kubernetes.io/projected/37a5e44f-9a88-4405-be8a-b645485e7312-kube-api-access-rdwmf\") pod \"network-operator-58b4c7f79c-55gtf\" (UID: \"37a5e44f-9a88-4405-be8a-b645485e7312\") " pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" Dec 13 06:49:23 crc kubenswrapper[4971]: I1213 06:49:23.789417 4971 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"networking-console-plugin-cert\" (UniqueName: \"kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 13 06:49:23 crc kubenswrapper[4971]: I1213 06:49:23.789437 4971 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-rczfb\" (UniqueName: \"kubernetes.io/projected/d75a4c96-2883-4a0b-bab2-0fab2b6c0b49-kube-api-access-rczfb\") pod \"iptables-alerter-4ln5h\" (UID: \"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\") " pod="openshift-network-operator/iptables-alerter-4ln5h" Dec 13 06:49:23 crc kubenswrapper[4971]: I1213 06:49:23.789455 4971 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/37a5e44f-9a88-4405-be8a-b645485e7312-metrics-tls\") pod \"network-operator-58b4c7f79c-55gtf\" (UID: \"37a5e44f-9a88-4405-be8a-b645485e7312\") " pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" Dec 13 06:49:23 crc kubenswrapper[4971]: I1213 06:49:23.789605 4971 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nginx-conf\" (UniqueName: \"kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 13 06:49:23 crc kubenswrapper[4971]: I1213 06:49:23.789624 4971 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovnkube-identity-cm\" (UniqueName: \"kubernetes.io/configmap/ef543e1b-8068-4ea3-b32a-61027b32e95d-ovnkube-identity-cm\") pod \"network-node-identity-vrzqb\" (UID: \"ef543e1b-8068-4ea3-b32a-61027b32e95d\") " pod="openshift-network-node-identity/network-node-identity-vrzqb" Dec 13 06:49:23 crc kubenswrapper[4971]: I1213 06:49:23.789640 4971 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"iptables-alerter-script\" (UniqueName: \"kubernetes.io/configmap/d75a4c96-2883-4a0b-bab2-0fab2b6c0b49-iptables-alerter-script\") pod \"iptables-alerter-4ln5h\" (UID: \"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\") " pod="openshift-network-operator/iptables-alerter-4ln5h" Dec 13 06:49:23 crc kubenswrapper[4971]: I1213 06:49:23.789878 4971 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-s2dwl\" (UniqueName: \"kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl\") pod \"network-check-source-55646444c4-trplf\" (UID: \"9d751cbb-f2e2-430d-9754-c882a5e924a5\") " pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 13 06:49:23 crc kubenswrapper[4971]: I1213 06:49:23.789897 4971 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/ef543e1b-8068-4ea3-b32a-61027b32e95d-webhook-cert\") pod \"network-node-identity-vrzqb\" (UID: \"ef543e1b-8068-4ea3-b32a-61027b32e95d\") " pod="openshift-network-node-identity/network-node-identity-vrzqb" Dec 13 06:49:23 crc kubenswrapper[4971]: I1213 06:49:23.789914 4971 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/ef543e1b-8068-4ea3-b32a-61027b32e95d-env-overrides\") pod \"network-node-identity-vrzqb\" (UID: \"ef543e1b-8068-4ea3-b32a-61027b32e95d\") " pod="openshift-network-node-identity/network-node-identity-vrzqb" Dec 13 06:49:23 crc kubenswrapper[4971]: I1213 06:49:23.789975 4971 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/6509e943-70c6-444c-bc41-48a544e36fbd-config\") on node \"crc\" DevicePath \"\"" Dec 13 06:49:23 crc kubenswrapper[4971]: I1213 06:49:23.789986 4971 reconciler_common.go:293] "Volume detached for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/925f1c65-6136-48ba-85aa-3a3b50560753-env-overrides\") on node \"crc\" DevicePath \"\"" Dec 13 06:49:23 crc kubenswrapper[4971]: I1213 06:49:23.789997 4971 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/09efc573-dbb6-4249-bd59-9b87aba8dd28-serving-cert\") on node \"crc\" DevicePath \"\"" Dec 13 06:49:23 crc kubenswrapper[4971]: I1213 06:49:23.790006 4971 reconciler_common.go:293] "Volume detached for volume \"available-featuregates\" (UniqueName: \"kubernetes.io/empty-dir/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d-available-featuregates\") on node \"crc\" DevicePath \"\"" Dec 13 06:49:23 crc kubenswrapper[4971]: I1213 06:49:23.790016 4971 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-ngvvp\" (UniqueName: \"kubernetes.io/projected/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-kube-api-access-ngvvp\") on node \"crc\" DevicePath \"\"" Dec 13 06:49:23 crc kubenswrapper[4971]: I1213 06:49:23.790024 4971 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/01ab3dd5-8196-46d0-ad33-122e2ca51def-config\") on node \"crc\" DevicePath \"\"" Dec 13 06:49:23 crc kubenswrapper[4971]: I1213 06:49:23.790033 4971 reconciler_common.go:293] "Volume detached for volume \"image-import-ca\" (UniqueName: \"kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-image-import-ca\") on node \"crc\" DevicePath \"\"" Dec 13 06:49:23 crc kubenswrapper[4971]: I1213 06:49:23.790042 4971 reconciler_common.go:293] "Volume detached for volume \"ovn-node-metrics-cert\" (UniqueName: \"kubernetes.io/secret/6ea678ab-3438-413e-bfe3-290ae7725660-ovn-node-metrics-cert\") on node \"crc\" DevicePath \"\"" Dec 13 06:49:23 crc kubenswrapper[4971]: I1213 06:49:23.790051 4971 reconciler_common.go:293] "Volume detached for volume \"registry-tls\" (UniqueName: \"kubernetes.io/projected/8f668bae-612b-4b75-9490-919e737c6a3b-registry-tls\") on node \"crc\" DevicePath \"\"" Dec 13 06:49:23 crc kubenswrapper[4971]: I1213 06:49:23.790059 4971 reconciler_common.go:293] "Volume detached for volume \"srv-cert\" (UniqueName: \"kubernetes.io/secret/b6312bbd-5731-4ea0-a20f-81d5a57df44a-srv-cert\") on node \"crc\" DevicePath \"\"" Dec 13 06:49:23 crc kubenswrapper[4971]: I1213 06:49:23.790068 4971 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-w4xd4\" (UniqueName: \"kubernetes.io/projected/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b-kube-api-access-w4xd4\") on node \"crc\" DevicePath \"\"" Dec 13 06:49:23 crc kubenswrapper[4971]: I1213 06:49:23.790077 4971 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/01ab3dd5-8196-46d0-ad33-122e2ca51def-serving-cert\") on node \"crc\" DevicePath \"\"" Dec 13 06:49:23 crc kubenswrapper[4971]: I1213 06:49:23.790086 4971 reconciler_common.go:293] "Volume detached for volume \"samples-operator-tls\" (UniqueName: \"kubernetes.io/secret/a0128f3a-b052-44ed-a84e-c4c8aaf17c13-samples-operator-tls\") on node \"crc\" DevicePath \"\"" Dec 13 06:49:23 crc kubenswrapper[4971]: I1213 06:49:23.790095 4971 reconciler_common.go:293] "Volume detached for volume \"webhook-certs\" (UniqueName: \"kubernetes.io/secret/efdd0498-1daa-4136-9a4a-3b948c2293fc-webhook-certs\") on node \"crc\" DevicePath \"\"" Dec 13 06:49:23 crc kubenswrapper[4971]: I1213 06:49:23.790103 4971 reconciler_common.go:293] "Volume detached for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/96b93a3a-6083-4aea-8eab-fe1aa8245ad9-metrics-tls\") on node \"crc\" DevicePath \"\"" Dec 13 06:49:23 crc kubenswrapper[4971]: I1213 06:49:23.790111 4971 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/e7e6199b-1264-4501-8953-767f51328d08-config\") on node \"crc\" DevicePath \"\"" Dec 13 06:49:23 crc kubenswrapper[4971]: I1213 06:49:23.790121 4971 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-fqsjt\" (UniqueName: \"kubernetes.io/projected/efdd0498-1daa-4136-9a4a-3b948c2293fc-kube-api-access-fqsjt\") on node \"crc\" DevicePath \"\"" Dec 13 06:49:23 crc kubenswrapper[4971]: I1213 06:49:23.790130 4971 reconciler_common.go:293] "Volume detached for volume \"etcd-service-ca\" (UniqueName: \"kubernetes.io/configmap/09efc573-dbb6-4249-bd59-9b87aba8dd28-etcd-service-ca\") on node \"crc\" DevicePath \"\"" Dec 13 06:49:23 crc kubenswrapper[4971]: I1213 06:49:23.790138 4971 reconciler_common.go:293] "Volume detached for volume \"stats-auth\" (UniqueName: \"kubernetes.io/secret/c03ee662-fb2f-4fc4-a2c1-af487c19d254-stats-auth\") on node \"crc\" DevicePath \"\"" Dec 13 06:49:23 crc kubenswrapper[4971]: I1213 06:49:23.790146 4971 reconciler_common.go:293] "Volume detached for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/a31745f5-9847-4afe-82a5-3161cc66ca93-metrics-tls\") on node \"crc\" DevicePath \"\"" Dec 13 06:49:23 crc kubenswrapper[4971]: I1213 06:49:23.790156 4971 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-x2m85\" (UniqueName: \"kubernetes.io/projected/cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d-kube-api-access-x2m85\") on node \"crc\" DevicePath \"\"" Dec 13 06:49:23 crc kubenswrapper[4971]: I1213 06:49:23.790165 4971 reconciler_common.go:293] "Volume detached for volume \"installation-pull-secrets\" (UniqueName: \"kubernetes.io/secret/8f668bae-612b-4b75-9490-919e737c6a3b-installation-pull-secrets\") on node \"crc\" DevicePath \"\"" Dec 13 06:49:23 crc kubenswrapper[4971]: I1213 06:49:23.790174 4971 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-dbsvg\" (UniqueName: \"kubernetes.io/projected/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9-kube-api-access-dbsvg\") on node \"crc\" DevicePath \"\"" Dec 13 06:49:23 crc kubenswrapper[4971]: I1213 06:49:23.790183 4971 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-system-router-certs\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-router-certs\") on node \"crc\" DevicePath \"\"" Dec 13 06:49:23 crc kubenswrapper[4971]: I1213 06:49:23.790192 4971 reconciler_common.go:293] "Volume detached for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/8f668bae-612b-4b75-9490-919e737c6a3b-bound-sa-token\") on node \"crc\" DevicePath \"\"" Dec 13 06:49:23 crc kubenswrapper[4971]: I1213 06:49:23.790202 4971 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-mnrrd\" (UniqueName: \"kubernetes.io/projected/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d-kube-api-access-mnrrd\") on node \"crc\" DevicePath \"\"" Dec 13 06:49:23 crc kubenswrapper[4971]: I1213 06:49:23.790212 4971 reconciler_common.go:293] "Volume detached for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/8f668bae-612b-4b75-9490-919e737c6a3b-trusted-ca\") on node \"crc\" DevicePath \"\"" Dec 13 06:49:23 crc kubenswrapper[4971]: I1213 06:49:23.790221 4971 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-w9rds\" (UniqueName: \"kubernetes.io/projected/20b0d48f-5fd6-431c-a545-e3c800c7b866-kube-api-access-w9rds\") on node \"crc\" DevicePath \"\"" Dec 13 06:49:23 crc kubenswrapper[4971]: I1213 06:49:23.790230 4971 reconciler_common.go:293] "Volume detached for volume \"srv-cert\" (UniqueName: \"kubernetes.io/secret/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9-srv-cert\") on node \"crc\" DevicePath \"\"" Dec 13 06:49:23 crc kubenswrapper[4971]: I1213 06:49:23.790239 4971 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-lzf88\" (UniqueName: \"kubernetes.io/projected/0b574797-001e-440a-8f4e-c0be86edad0f-kube-api-access-lzf88\") on node \"crc\" DevicePath \"\"" Dec 13 06:49:23 crc kubenswrapper[4971]: I1213 06:49:23.790249 4971 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-fcqwp\" (UniqueName: \"kubernetes.io/projected/5fe579f8-e8a6-4643-bce5-a661393c4dde-kube-api-access-fcqwp\") on node \"crc\" DevicePath \"\"" Dec 13 06:49:23 crc kubenswrapper[4971]: I1213 06:49:23.790258 4971 reconciler_common.go:293] "Volume detached for volume \"mcc-auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/0b574797-001e-440a-8f4e-c0be86edad0f-mcc-auth-proxy-config\") on node \"crc\" DevicePath \"\"" Dec 13 06:49:23 crc kubenswrapper[4971]: I1213 06:49:23.790267 4971 reconciler_common.go:293] "Volume detached for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-webhook-cert\") on node \"crc\" DevicePath \"\"" Dec 13 06:49:23 crc kubenswrapper[4971]: I1213 06:49:23.790275 4971 reconciler_common.go:293] "Volume detached for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/bf126b07-da06-4140-9a57-dfd54fc6b486-trusted-ca\") on node \"crc\" DevicePath \"\"" Dec 13 06:49:23 crc kubenswrapper[4971]: I1213 06:49:23.790285 4971 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-user-template-provider-selection\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-template-provider-selection\") on node \"crc\" DevicePath \"\"" Dec 13 06:49:23 crc kubenswrapper[4971]: I1213 06:49:23.790295 4971 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-2w9zh\" (UniqueName: \"kubernetes.io/projected/4bb40260-dbaa-4fb0-84df-5e680505d512-kube-api-access-2w9zh\") on node \"crc\" DevicePath \"\"" Dec 13 06:49:23 crc kubenswrapper[4971]: I1213 06:49:23.790305 4971 reconciler_common.go:293] "Volume detached for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-trusted-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 13 06:49:23 crc kubenswrapper[4971]: I1213 06:49:23.790491 4971 reconciler_common.go:293] "Volume detached for volume \"marketplace-operator-metrics\" (UniqueName: \"kubernetes.io/secret/b6cd30de-2eeb-49a2-ab40-9167f4560ff5-marketplace-operator-metrics\") on node \"crc\" DevicePath \"\"" Dec 13 06:49:23 crc kubenswrapper[4971]: I1213 06:49:23.790500 4971 reconciler_common.go:293] "Volume detached for volume \"service-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/6509e943-70c6-444c-bc41-48a544e36fbd-service-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 13 06:49:23 crc kubenswrapper[4971]: I1213 06:49:23.790509 4971 reconciler_common.go:293] "Volume detached for volume \"service-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/c03ee662-fb2f-4fc4-a2c1-af487c19d254-service-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 13 06:49:23 crc kubenswrapper[4971]: I1213 06:49:23.790535 4971 reconciler_common.go:293] "Volume detached for volume \"oauth-serving-cert\" (UniqueName: \"kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-oauth-serving-cert\") on node \"crc\" DevicePath \"\"" Dec 13 06:49:23 crc kubenswrapper[4971]: I1213 06:49:23.790544 4971 reconciler_common.go:293] "Volume detached for volume \"node-bootstrap-token\" (UniqueName: \"kubernetes.io/secret/5fe579f8-e8a6-4643-bce5-a661393c4dde-node-bootstrap-token\") on node \"crc\" DevicePath \"\"" Dec 13 06:49:23 crc kubenswrapper[4971]: I1213 06:49:23.790553 4971 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d-utilities\") on node \"crc\" DevicePath \"\"" Dec 13 06:49:23 crc kubenswrapper[4971]: I1213 06:49:23.790561 4971 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-xcgwh\" (UniqueName: \"kubernetes.io/projected/fda69060-fa79-4696-b1a6-7980f124bf7c-kube-api-access-xcgwh\") on node \"crc\" DevicePath \"\"" Dec 13 06:49:23 crc kubenswrapper[4971]: I1213 06:49:23.790570 4971 reconciler_common.go:293] "Volume detached for volume \"certs\" (UniqueName: \"kubernetes.io/secret/5fe579f8-e8a6-4643-bce5-a661393c4dde-certs\") on node \"crc\" DevicePath \"\"" Dec 13 06:49:23 crc kubenswrapper[4971]: I1213 06:49:23.790579 4971 reconciler_common.go:293] "Volume detached for volume \"mcd-auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/fda69060-fa79-4696-b1a6-7980f124bf7c-mcd-auth-proxy-config\") on node \"crc\" DevicePath \"\"" Dec 13 06:49:23 crc kubenswrapper[4971]: I1213 06:49:23.790588 4971 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-4d4hj\" (UniqueName: \"kubernetes.io/projected/3ab1a177-2de0-46d9-b765-d0d0649bb42e-kube-api-access-4d4hj\") on node \"crc\" DevicePath \"\"" Dec 13 06:49:23 crc kubenswrapper[4971]: I1213 06:49:23.790597 4971 reconciler_common.go:293] "Volume detached for volume \"etcd-serving-ca\" (UniqueName: \"kubernetes.io/configmap/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-etcd-serving-ca\") on node \"crc\" DevicePath \"\"" Dec 13 06:49:23 crc kubenswrapper[4971]: I1213 06:49:23.790607 4971 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-cfbct\" (UniqueName: \"kubernetes.io/projected/57a731c4-ef35-47a8-b875-bfb08a7f8011-kube-api-access-cfbct\") on node \"crc\" DevicePath \"\"" Dec 13 06:49:23 crc kubenswrapper[4971]: I1213 06:49:23.790616 4971 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/7583ce53-e0fe-4a16-9e4d-50516596a136-serving-cert\") on node \"crc\" DevicePath \"\"" Dec 13 06:49:23 crc kubenswrapper[4971]: I1213 06:49:23.790625 4971 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-pjr6v\" (UniqueName: \"kubernetes.io/projected/49ef4625-1d3a-4a9f-b595-c2433d32326d-kube-api-access-pjr6v\") on node \"crc\" DevicePath \"\"" Dec 13 06:49:23 crc kubenswrapper[4971]: I1213 06:49:23.790634 4971 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/6402fda4-df10-493c-b4e5-d0569419652d-config\") on node \"crc\" DevicePath \"\"" Dec 13 06:49:23 crc kubenswrapper[4971]: I1213 06:49:23.790643 4971 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b-config\") on node \"crc\" DevicePath \"\"" Dec 13 06:49:23 crc kubenswrapper[4971]: I1213 06:49:23.790651 4971 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-pcxfs\" (UniqueName: \"kubernetes.io/projected/9d4552c7-cd75-42dd-8880-30dd377c49a4-kube-api-access-pcxfs\") on node \"crc\" DevicePath \"\"" Dec 13 06:49:23 crc kubenswrapper[4971]: I1213 06:49:23.790661 4971 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-9xfj7\" (UniqueName: \"kubernetes.io/projected/5225d0e4-402f-4861-b410-819f433b1803-kube-api-access-9xfj7\") on node \"crc\" DevicePath \"\"" Dec 13 06:49:23 crc kubenswrapper[4971]: I1213 06:49:23.790670 4971 reconciler_common.go:293] "Volume detached for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/925f1c65-6136-48ba-85aa-3a3b50560753-ovnkube-config\") on node \"crc\" DevicePath \"\"" Dec 13 06:49:23 crc kubenswrapper[4971]: I1213 06:49:23.790679 4971 reconciler_common.go:293] "Volume detached for volume \"audit\" (UniqueName: \"kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-audit\") on node \"crc\" DevicePath \"\"" Dec 13 06:49:23 crc kubenswrapper[4971]: I1213 06:49:23.790688 4971 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-pj782\" (UniqueName: \"kubernetes.io/projected/b6cd30de-2eeb-49a2-ab40-9167f4560ff5-kube-api-access-pj782\") on node \"crc\" DevicePath \"\"" Dec 13 06:49:23 crc kubenswrapper[4971]: I1213 06:49:23.790696 4971 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-system-service-ca\" (UniqueName: \"kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-service-ca\") on node \"crc\" DevicePath \"\"" Dec 13 06:49:23 crc kubenswrapper[4971]: I1213 06:49:23.790706 4971 reconciler_common.go:293] "Volume detached for volume \"cni-binary-copy\" (UniqueName: \"kubernetes.io/configmap/7bb08738-c794-4ee8-9972-3a62ca171029-cni-binary-copy\") on node \"crc\" DevicePath \"\"" Dec 13 06:49:23 crc kubenswrapper[4971]: I1213 06:49:23.790716 4971 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/0b78653f-4ff9-4508-8672-245ed9b561e3-serving-cert\") on node \"crc\" DevicePath \"\"" Dec 13 06:49:23 crc kubenswrapper[4971]: I1213 06:49:23.790724 4971 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-system-session\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-session\") on node \"crc\" DevicePath \"\"" Dec 13 06:49:23 crc kubenswrapper[4971]: I1213 06:49:23.790734 4971 reconciler_common.go:293] "Volume detached for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/0b574797-001e-440a-8f4e-c0be86edad0f-proxy-tls\") on node \"crc\" DevicePath \"\"" Dec 13 06:49:23 crc kubenswrapper[4971]: I1213 06:49:23.790743 4971 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-nzwt7\" (UniqueName: \"kubernetes.io/projected/96b93a3a-6083-4aea-8eab-fe1aa8245ad9-kube-api-access-nzwt7\") on node \"crc\" DevicePath \"\"" Dec 13 06:49:23 crc kubenswrapper[4971]: I1213 06:49:23.790753 4971 reconciler_common.go:293] "Volume detached for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-audit-policies\") on node \"crc\" DevicePath \"\"" Dec 13 06:49:23 crc kubenswrapper[4971]: I1213 06:49:23.790762 4971 reconciler_common.go:293] "Volume detached for volume \"machine-approver-tls\" (UniqueName: \"kubernetes.io/secret/22c825df-677d-4ca6-82db-3454ed06e783-machine-approver-tls\") on node \"crc\" DevicePath \"\"" Dec 13 06:49:23 crc kubenswrapper[4971]: I1213 06:49:23.790771 4971 reconciler_common.go:293] "Volume detached for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/496e6271-fb68-4057-954e-a0d97a4afa3f-kube-api-access\") on node \"crc\" DevicePath \"\"" Dec 13 06:49:23 crc kubenswrapper[4971]: I1213 06:49:23.790780 4971 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/9d4552c7-cd75-42dd-8880-30dd377c49a4-config\") on node \"crc\" DevicePath \"\"" Dec 13 06:49:23 crc kubenswrapper[4971]: I1213 06:49:23.790789 4971 reconciler_common.go:293] "Volume detached for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/5b88f790-22fa-440e-b583-365168c0b23d-metrics-certs\") on node \"crc\" DevicePath \"\"" Dec 13 06:49:23 crc kubenswrapper[4971]: I1213 06:49:23.790800 4971 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/57a731c4-ef35-47a8-b875-bfb08a7f8011-utilities\") on node \"crc\" DevicePath \"\"" Dec 13 06:49:23 crc kubenswrapper[4971]: I1213 06:49:23.790809 4971 reconciler_common.go:293] "Volume detached for volume \"cni-sysctl-allowlist\" (UniqueName: \"kubernetes.io/configmap/7bb08738-c794-4ee8-9972-3a62ca171029-cni-sysctl-allowlist\") on node \"crc\" DevicePath \"\"" Dec 13 06:49:23 crc kubenswrapper[4971]: I1213 06:49:23.790820 4971 reconciler_common.go:293] "Volume detached for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-apiservice-cert\") on node \"crc\" DevicePath \"\"" Dec 13 06:49:23 crc kubenswrapper[4971]: I1213 06:49:23.790829 4971 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/7539238d-5fe0-46ed-884e-1c3b566537ec-config\") on node \"crc\" DevicePath \"\"" Dec 13 06:49:23 crc kubenswrapper[4971]: I1213 06:49:23.790839 4971 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-rnphk\" (UniqueName: \"kubernetes.io/projected/bf126b07-da06-4140-9a57-dfd54fc6b486-kube-api-access-rnphk\") on node \"crc\" DevicePath \"\"" Dec 13 06:49:23 crc kubenswrapper[4971]: I1213 06:49:23.790848 4971 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-kfwg7\" (UniqueName: \"kubernetes.io/projected/8f668bae-612b-4b75-9490-919e737c6a3b-kube-api-access-kfwg7\") on node \"crc\" DevicePath \"\"" Dec 13 06:49:23 crc kubenswrapper[4971]: I1213 06:49:23.790856 4971 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/496e6271-fb68-4057-954e-a0d97a4afa3f-serving-cert\") on node \"crc\" DevicePath \"\"" Dec 13 06:49:23 crc kubenswrapper[4971]: I1213 06:49:23.790865 4971 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d-serving-cert\") on node \"crc\" DevicePath \"\"" Dec 13 06:49:23 crc kubenswrapper[4971]: I1213 06:49:23.790874 4971 reconciler_common.go:293] "Volume detached for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-trusted-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 13 06:49:23 crc kubenswrapper[4971]: I1213 06:49:23.790882 4971 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/5225d0e4-402f-4861-b410-819f433b1803-utilities\") on node \"crc\" DevicePath \"\"" Dec 13 06:49:23 crc kubenswrapper[4971]: I1213 06:49:23.790891 4971 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-mg5zb\" (UniqueName: \"kubernetes.io/projected/6402fda4-df10-493c-b4e5-d0569419652d-kube-api-access-mg5zb\") on node \"crc\" DevicePath \"\"" Dec 13 06:49:23 crc kubenswrapper[4971]: I1213 06:49:23.790899 4971 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c-serving-cert\") on node \"crc\" DevicePath \"\"" Dec 13 06:49:23 crc kubenswrapper[4971]: I1213 06:49:23.790907 4971 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-8tdtz\" (UniqueName: \"kubernetes.io/projected/09efc573-dbb6-4249-bd59-9b87aba8dd28-kube-api-access-8tdtz\") on node \"crc\" DevicePath \"\"" Dec 13 06:49:23 crc kubenswrapper[4971]: I1213 06:49:23.790916 4971 reconciler_common.go:293] "Volume detached for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-etcd-client\") on node \"crc\" DevicePath \"\"" Dec 13 06:49:23 crc kubenswrapper[4971]: I1213 06:49:23.790926 4971 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-w7l8j\" (UniqueName: \"kubernetes.io/projected/01ab3dd5-8196-46d0-ad33-122e2ca51def-kube-api-access-w7l8j\") on node \"crc\" DevicePath \"\"" Dec 13 06:49:23 crc kubenswrapper[4971]: I1213 06:49:23.790935 4971 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-tk88c\" (UniqueName: \"kubernetes.io/projected/7539238d-5fe0-46ed-884e-1c3b566537ec-kube-api-access-tk88c\") on node \"crc\" DevicePath \"\"" Dec 13 06:49:23 crc kubenswrapper[4971]: I1213 06:49:23.790943 4971 reconciler_common.go:293] "Volume detached for volume \"console-config\" (UniqueName: \"kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-console-config\") on node \"crc\" DevicePath \"\"" Dec 13 06:49:23 crc kubenswrapper[4971]: I1213 06:49:23.790952 4971 reconciler_common.go:293] "Volume detached for volume \"control-plane-machine-set-operator-tls\" (UniqueName: \"kubernetes.io/secret/6731426b-95fe-49ff-bb5f-40441049fde2-control-plane-machine-set-operator-tls\") on node \"crc\" DevicePath \"\"" Dec 13 06:49:23 crc kubenswrapper[4971]: I1213 06:49:23.790961 4971 reconciler_common.go:293] "Volume detached for volume \"registry-certificates\" (UniqueName: \"kubernetes.io/configmap/8f668bae-612b-4b75-9490-919e737c6a3b-registry-certificates\") on node \"crc\" DevicePath \"\"" Dec 13 06:49:23 crc kubenswrapper[4971]: I1213 06:49:23.790973 4971 reconciler_common.go:293] "Volume detached for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/6509e943-70c6-444c-bc41-48a544e36fbd-trusted-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 13 06:49:23 crc kubenswrapper[4971]: I1213 06:49:23.788093 4971 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/c03ee662-fb2f-4fc4-a2c1-af487c19d254-default-certificate" (OuterVolumeSpecName: "default-certificate") pod "c03ee662-fb2f-4fc4-a2c1-af487c19d254" (UID: "c03ee662-fb2f-4fc4-a2c1-af487c19d254"). InnerVolumeSpecName "default-certificate". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 13 06:49:23 crc kubenswrapper[4971]: I1213 06:49:23.788419 4971 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/496e6271-fb68-4057-954e-a0d97a4afa3f-config" (OuterVolumeSpecName: "config") pod "496e6271-fb68-4057-954e-a0d97a4afa3f" (UID: "496e6271-fb68-4057-954e-a0d97a4afa3f"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 13 06:49:23 crc kubenswrapper[4971]: I1213 06:49:23.788437 4971 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/09efc573-dbb6-4249-bd59-9b87aba8dd28-etcd-client" (OuterVolumeSpecName: "etcd-client") pod "09efc573-dbb6-4249-bd59-9b87aba8dd28" (UID: "09efc573-dbb6-4249-bd59-9b87aba8dd28"). InnerVolumeSpecName "etcd-client". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 13 06:49:23 crc kubenswrapper[4971]: I1213 06:49:23.788449 4971 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/a31745f5-9847-4afe-82a5-3161cc66ca93-bound-sa-token" (OuterVolumeSpecName: "bound-sa-token") pod "a31745f5-9847-4afe-82a5-3161cc66ca93" (UID: "a31745f5-9847-4afe-82a5-3161cc66ca93"). InnerVolumeSpecName "bound-sa-token". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 13 06:49:23 crc kubenswrapper[4971]: I1213 06:49:23.788472 4971 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/6509e943-70c6-444c-bc41-48a544e36fbd-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "6509e943-70c6-444c-bc41-48a544e36fbd" (UID: "6509e943-70c6-444c-bc41-48a544e36fbd"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 13 06:49:23 crc kubenswrapper[4971]: I1213 06:49:23.788723 4971 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/43509403-f426-496e-be36-56cef71462f5-kube-api-access-qg5z5" (OuterVolumeSpecName: "kube-api-access-qg5z5") pod "43509403-f426-496e-be36-56cef71462f5" (UID: "43509403-f426-496e-be36-56cef71462f5"). InnerVolumeSpecName "kube-api-access-qg5z5". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 13 06:49:23 crc kubenswrapper[4971]: I1213 06:49:23.788987 4971 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/5441d097-087c-4d9a-baa8-b210afa90fc9-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "5441d097-087c-4d9a-baa8-b210afa90fc9" (UID: "5441d097-087c-4d9a-baa8-b210afa90fc9"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 13 06:49:23 crc kubenswrapper[4971]: I1213 06:49:23.789028 4971 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/a31745f5-9847-4afe-82a5-3161cc66ca93-kube-api-access-lz9wn" (OuterVolumeSpecName: "kube-api-access-lz9wn") pod "a31745f5-9847-4afe-82a5-3161cc66ca93" (UID: "a31745f5-9847-4afe-82a5-3161cc66ca93"). InnerVolumeSpecName "kube-api-access-lz9wn". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 13 06:49:23 crc kubenswrapper[4971]: I1213 06:49:23.789114 4971 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/6ea678ab-3438-413e-bfe3-290ae7725660-ovnkube-config" (OuterVolumeSpecName: "ovnkube-config") pod "6ea678ab-3438-413e-bfe3-290ae7725660" (UID: "6ea678ab-3438-413e-bfe3-290ae7725660"). InnerVolumeSpecName "ovnkube-config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 13 06:49:23 crc kubenswrapper[4971]: I1213 06:49:23.789330 4971 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/b6cd30de-2eeb-49a2-ab40-9167f4560ff5-marketplace-trusted-ca" (OuterVolumeSpecName: "marketplace-trusted-ca") pod "b6cd30de-2eeb-49a2-ab40-9167f4560ff5" (UID: "b6cd30de-2eeb-49a2-ab40-9167f4560ff5"). InnerVolumeSpecName "marketplace-trusted-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 13 06:49:23 crc kubenswrapper[4971]: I1213 06:49:23.789332 4971 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/6402fda4-df10-493c-b4e5-d0569419652d-machine-api-operator-tls" (OuterVolumeSpecName: "machine-api-operator-tls") pod "6402fda4-df10-493c-b4e5-d0569419652d" (UID: "6402fda4-df10-493c-b4e5-d0569419652d"). InnerVolumeSpecName "machine-api-operator-tls". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 13 06:49:23 crc kubenswrapper[4971]: I1213 06:49:23.789657 4971 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "09ae3b1a-e8e7-4524-b54b-61eab6f9239a" (UID: "09ae3b1a-e8e7-4524-b54b-61eab6f9239a"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 13 06:49:23 crc kubenswrapper[4971]: I1213 06:49:23.789994 4971 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/5b88f790-22fa-440e-b583-365168c0b23d-kube-api-access-jkwtn" (OuterVolumeSpecName: "kube-api-access-jkwtn") pod "5b88f790-22fa-440e-b583-365168c0b23d" (UID: "5b88f790-22fa-440e-b583-365168c0b23d"). InnerVolumeSpecName "kube-api-access-jkwtn". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 13 06:49:23 crc kubenswrapper[4971]: I1213 06:49:23.790031 4971 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/5441d097-087c-4d9a-baa8-b210afa90fc9-kube-api-access-2d4wz" (OuterVolumeSpecName: "kube-api-access-2d4wz") pod "5441d097-087c-4d9a-baa8-b210afa90fc9" (UID: "5441d097-087c-4d9a-baa8-b210afa90fc9"). InnerVolumeSpecName "kube-api-access-2d4wz". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 13 06:49:23 crc kubenswrapper[4971]: I1213 06:49:23.790602 4971 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/0b78653f-4ff9-4508-8672-245ed9b561e3-service-ca" (OuterVolumeSpecName: "service-ca") pod "0b78653f-4ff9-4508-8672-245ed9b561e3" (UID: "0b78653f-4ff9-4508-8672-245ed9b561e3"). InnerVolumeSpecName "service-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 13 06:49:23 crc kubenswrapper[4971]: I1213 06:49:23.790606 4971 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/7539238d-5fe0-46ed-884e-1c3b566537ec-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "7539238d-5fe0-46ed-884e-1c3b566537ec" (UID: "7539238d-5fe0-46ed-884e-1c3b566537ec"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 13 06:49:23 crc kubenswrapper[4971]: I1213 06:49:23.790764 4971 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-config" (OuterVolumeSpecName: "config") pod "1bf7eb37-55a3-4c65-b768-a94c82151e69" (UID: "1bf7eb37-55a3-4c65-b768-a94c82151e69"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 13 06:49:23 crc kubenswrapper[4971]: I1213 06:49:23.791387 4971 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/31d8b7a1-420e-4252-a5b7-eebe8a111292-kube-api-access-zgdk5" (OuterVolumeSpecName: "kube-api-access-zgdk5") pod "31d8b7a1-420e-4252-a5b7-eebe8a111292" (UID: "31d8b7a1-420e-4252-a5b7-eebe8a111292"). InnerVolumeSpecName "kube-api-access-zgdk5". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 13 06:49:23 crc kubenswrapper[4971]: I1213 06:49:23.791595 4971 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-kube-api-access-6ccd8" (OuterVolumeSpecName: "kube-api-access-6ccd8") pod "308be0ea-9f5f-4b29-aeb1-5abd31a0b17b" (UID: "308be0ea-9f5f-4b29-aeb1-5abd31a0b17b"). InnerVolumeSpecName "kube-api-access-6ccd8". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 13 06:49:23 crc kubenswrapper[4971]: I1213 06:49:23.791790 4971 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9-profile-collector-cert" (OuterVolumeSpecName: "profile-collector-cert") pod "f88749ec-7931-4ee7-b3fc-1ec5e11f92e9" (UID: "f88749ec-7931-4ee7-b3fc-1ec5e11f92e9"). InnerVolumeSpecName "profile-collector-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 13 06:49:23 crc kubenswrapper[4971]: I1213 06:49:23.791797 4971 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/31d8b7a1-420e-4252-a5b7-eebe8a111292-proxy-tls" (OuterVolumeSpecName: "proxy-tls") pod "31d8b7a1-420e-4252-a5b7-eebe8a111292" (UID: "31d8b7a1-420e-4252-a5b7-eebe8a111292"). InnerVolumeSpecName "proxy-tls". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 13 06:49:23 crc kubenswrapper[4971]: I1213 06:49:23.791853 4971 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/5441d097-087c-4d9a-baa8-b210afa90fc9-client-ca" (OuterVolumeSpecName: "client-ca") pod "5441d097-087c-4d9a-baa8-b210afa90fc9" (UID: "5441d097-087c-4d9a-baa8-b210afa90fc9"). InnerVolumeSpecName "client-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 13 06:49:23 crc kubenswrapper[4971]: I1213 06:49:23.791853 4971 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-kube-api-access-zkvpv" (OuterVolumeSpecName: "kube-api-access-zkvpv") pod "09ae3b1a-e8e7-4524-b54b-61eab6f9239a" (UID: "09ae3b1a-e8e7-4524-b54b-61eab6f9239a"). InnerVolumeSpecName "kube-api-access-zkvpv". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 13 06:49:23 crc kubenswrapper[4971]: I1213 06:49:23.792095 4971 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-trusted-ca-bundle" (OuterVolumeSpecName: "trusted-ca-bundle") pod "1bf7eb37-55a3-4c65-b768-a94c82151e69" (UID: "1bf7eb37-55a3-4c65-b768-a94c82151e69"). InnerVolumeSpecName "trusted-ca-bundle". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 13 06:49:23 crc kubenswrapper[4971]: I1213 06:49:23.792249 4971 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d-kube-api-access-x4zgh" (OuterVolumeSpecName: "kube-api-access-x4zgh") pod "b11524ee-3fca-4b1b-9cdf-6da289fdbc7d" (UID: "b11524ee-3fca-4b1b-9cdf-6da289fdbc7d"). InnerVolumeSpecName "kube-api-access-x4zgh". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 13 06:49:23 crc kubenswrapper[4971]: I1213 06:49:23.792341 4971 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/1bf7eb37-55a3-4c65-b768-a94c82151e69-encryption-config" (OuterVolumeSpecName: "encryption-config") pod "1bf7eb37-55a3-4c65-b768-a94c82151e69" (UID: "1bf7eb37-55a3-4c65-b768-a94c82151e69"). InnerVolumeSpecName "encryption-config". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 13 06:49:23 crc kubenswrapper[4971]: I1213 06:49:23.792362 4971 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/87cf06ed-a83f-41a7-828d-70653580a8cb-kube-api-access-d6qdx" (OuterVolumeSpecName: "kube-api-access-d6qdx") pod "87cf06ed-a83f-41a7-828d-70653580a8cb" (UID: "87cf06ed-a83f-41a7-828d-70653580a8cb"). InnerVolumeSpecName "kube-api-access-d6qdx". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 13 06:49:23 crc kubenswrapper[4971]: I1213 06:49:23.792382 4971 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/1bf7eb37-55a3-4c65-b768-a94c82151e69-etcd-client" (OuterVolumeSpecName: "etcd-client") pod "1bf7eb37-55a3-4c65-b768-a94c82151e69" (UID: "1bf7eb37-55a3-4c65-b768-a94c82151e69"). InnerVolumeSpecName "etcd-client". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 13 06:49:23 crc kubenswrapper[4971]: I1213 06:49:23.792709 4971 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/1d611f23-29be-4491-8495-bee1670e935f-kube-api-access-bf2bz" (OuterVolumeSpecName: "kube-api-access-bf2bz") pod "1d611f23-29be-4491-8495-bee1670e935f" (UID: "1d611f23-29be-4491-8495-bee1670e935f"). InnerVolumeSpecName "kube-api-access-bf2bz". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 13 06:49:23 crc kubenswrapper[4971]: I1213 06:49:23.792798 4971 swap_util.go:74] "error creating dir to test if tmpfs noswap is enabled. Assuming not supported" mount path="" error="stat /var/lib/kubelet/plugins/kubernetes.io/empty-dir: no such file or directory" Dec 13 06:49:23 crc kubenswrapper[4971]: I1213 06:49:23.792986 4971 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/4bb40260-dbaa-4fb0-84df-5e680505d512-cni-binary-copy" (OuterVolumeSpecName: "cni-binary-copy") pod "4bb40260-dbaa-4fb0-84df-5e680505d512" (UID: "4bb40260-dbaa-4fb0-84df-5e680505d512"). InnerVolumeSpecName "cni-binary-copy". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 13 06:49:23 crc kubenswrapper[4971]: I1213 06:49:23.793048 4971 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/a0128f3a-b052-44ed-a84e-c4c8aaf17c13-kube-api-access-gf66m" (OuterVolumeSpecName: "kube-api-access-gf66m") pod "a0128f3a-b052-44ed-a84e-c4c8aaf17c13" (UID: "a0128f3a-b052-44ed-a84e-c4c8aaf17c13"). InnerVolumeSpecName "kube-api-access-gf66m". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 13 06:49:23 crc kubenswrapper[4971]: I1213 06:49:23.793134 4971 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-encryption-config" (OuterVolumeSpecName: "encryption-config") pod "09ae3b1a-e8e7-4524-b54b-61eab6f9239a" (UID: "09ae3b1a-e8e7-4524-b54b-61eab6f9239a"). InnerVolumeSpecName "encryption-config". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 13 06:49:23 crc kubenswrapper[4971]: I1213 06:49:23.793408 4971 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-etcd-serving-ca" (OuterVolumeSpecName: "etcd-serving-ca") pod "1bf7eb37-55a3-4c65-b768-a94c82151e69" (UID: "1bf7eb37-55a3-4c65-b768-a94c82151e69"). InnerVolumeSpecName "etcd-serving-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 13 06:49:23 crc kubenswrapper[4971]: I1213 06:49:23.793428 4971 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59-kube-api-access-wxkg8" (OuterVolumeSpecName: "kube-api-access-wxkg8") pod "3cb93b32-e0ae-4377-b9c8-fdb9842c6d59" (UID: "3cb93b32-e0ae-4377-b9c8-fdb9842c6d59"). InnerVolumeSpecName "kube-api-access-wxkg8". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 13 06:49:23 crc kubenswrapper[4971]: I1213 06:49:23.793932 4971 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/25e176fe-21b4-4974-b1ed-c8b94f112a7f-kube-api-access-d4lsv" (OuterVolumeSpecName: "kube-api-access-d4lsv") pod "25e176fe-21b4-4974-b1ed-c8b94f112a7f" (UID: "25e176fe-21b4-4974-b1ed-c8b94f112a7f"). InnerVolumeSpecName "kube-api-access-d4lsv". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 13 06:49:23 crc kubenswrapper[4971]: I1213 06:49:23.794432 4971 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-ocp-branding-template" (OuterVolumeSpecName: "v4-0-config-system-ocp-branding-template") pod "49c341d1-5089-4bc2-86a0-a5e165cfcc6b" (UID: "49c341d1-5089-4bc2-86a0-a5e165cfcc6b"). InnerVolumeSpecName "v4-0-config-system-ocp-branding-template". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 13 06:49:23 crc kubenswrapper[4971]: I1213 06:49:23.794638 4971 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/fda69060-fa79-4696-b1a6-7980f124bf7c-proxy-tls" (OuterVolumeSpecName: "proxy-tls") pod "fda69060-fa79-4696-b1a6-7980f124bf7c" (UID: "fda69060-fa79-4696-b1a6-7980f124bf7c"). InnerVolumeSpecName "proxy-tls". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 13 06:49:23 crc kubenswrapper[4971]: E1213 06:49:23.794697 4971 configmap.go:193] Couldn't get configMap openshift-network-console/networking-console-plugin: object "openshift-network-console"/"networking-console-plugin" not registered Dec 13 06:49:23 crc kubenswrapper[4971]: I1213 06:49:23.794737 4971 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/43509403-f426-496e-be36-56cef71462f5-console-oauth-config" (OuterVolumeSpecName: "console-oauth-config") pod "43509403-f426-496e-be36-56cef71462f5" (UID: "43509403-f426-496e-be36-56cef71462f5"). InnerVolumeSpecName "console-oauth-config". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 13 06:49:23 crc kubenswrapper[4971]: E1213 06:49:23.794775 4971 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2025-12-13 06:49:24.294753872 +0000 UTC m=+20.899163340 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "nginx-conf" (UniqueName: "kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin" not registered Dec 13 06:49:23 crc kubenswrapper[4971]: I1213 06:49:23.794758 4971 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/bf126b07-da06-4140-9a57-dfd54fc6b486-bound-sa-token" (OuterVolumeSpecName: "bound-sa-token") pod "bf126b07-da06-4140-9a57-dfd54fc6b486" (UID: "bf126b07-da06-4140-9a57-dfd54fc6b486"). InnerVolumeSpecName "bound-sa-token". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 13 06:49:23 crc kubenswrapper[4971]: I1213 06:49:23.794929 4971 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/e7e6199b-1264-4501-8953-767f51328d08-kube-api-access" (OuterVolumeSpecName: "kube-api-access") pod "e7e6199b-1264-4501-8953-767f51328d08" (UID: "e7e6199b-1264-4501-8953-767f51328d08"). InnerVolumeSpecName "kube-api-access". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 13 06:49:23 crc kubenswrapper[4971]: I1213 06:49:23.795031 4971 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-service-ca" (OuterVolumeSpecName: "service-ca") pod "43509403-f426-496e-be36-56cef71462f5" (UID: "43509403-f426-496e-be36-56cef71462f5"). InnerVolumeSpecName "service-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 13 06:49:23 crc kubenswrapper[4971]: I1213 06:49:23.795051 4971 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/1386a44e-36a2-460c-96d0-0359d2b6f0f5-kube-api-access" (OuterVolumeSpecName: "kube-api-access") pod "1386a44e-36a2-460c-96d0-0359d2b6f0f5" (UID: "1386a44e-36a2-460c-96d0-0359d2b6f0f5"). InnerVolumeSpecName "kube-api-access". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 13 06:49:23 crc kubenswrapper[4971]: E1213 06:49:23.795290 4971 secret.go:188] Couldn't get secret openshift-network-console/networking-console-plugin-cert: object "openshift-network-console"/"networking-console-plugin-cert" not registered Dec 13 06:49:23 crc kubenswrapper[4971]: I1213 06:49:23.795330 4971 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/925f1c65-6136-48ba-85aa-3a3b50560753-kube-api-access-s4n52" (OuterVolumeSpecName: "kube-api-access-s4n52") pod "925f1c65-6136-48ba-85aa-3a3b50560753" (UID: "925f1c65-6136-48ba-85aa-3a3b50560753"). InnerVolumeSpecName "kube-api-access-s4n52". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 13 06:49:23 crc kubenswrapper[4971]: E1213 06:49:23.795404 4971 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2025-12-13 06:49:24.295381928 +0000 UTC m=+20.899791386 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "networking-console-plugin-cert" (UniqueName: "kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin-cert" not registered Dec 13 06:49:23 crc kubenswrapper[4971]: I1213 06:49:23.795475 4971 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/c03ee662-fb2f-4fc4-a2c1-af487c19d254-kube-api-access-v47cf" (OuterVolumeSpecName: "kube-api-access-v47cf") pod "c03ee662-fb2f-4fc4-a2c1-af487c19d254" (UID: "c03ee662-fb2f-4fc4-a2c1-af487c19d254"). InnerVolumeSpecName "kube-api-access-v47cf". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 13 06:49:23 crc kubenswrapper[4971]: I1213 06:49:23.795772 4971 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/7bb08738-c794-4ee8-9972-3a62ca171029-kube-api-access-279lb" (OuterVolumeSpecName: "kube-api-access-279lb") pod "7bb08738-c794-4ee8-9972-3a62ca171029" (UID: "7bb08738-c794-4ee8-9972-3a62ca171029"). InnerVolumeSpecName "kube-api-access-279lb". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 13 06:49:23 crc kubenswrapper[4971]: I1213 06:49:23.795954 4971 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/3ab1a177-2de0-46d9-b765-d0d0649bb42e-package-server-manager-serving-cert" (OuterVolumeSpecName: "package-server-manager-serving-cert") pod "3ab1a177-2de0-46d9-b765-d0d0649bb42e" (UID: "3ab1a177-2de0-46d9-b765-d0d0649bb42e"). InnerVolumeSpecName "package-server-manager-serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 13 06:49:23 crc kubenswrapper[4971]: I1213 06:49:23.795987 4971 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-tmpfs" (OuterVolumeSpecName: "tmpfs") pod "308be0ea-9f5f-4b29-aeb1-5abd31a0b17b" (UID: "308be0ea-9f5f-4b29-aeb1-5abd31a0b17b"). InnerVolumeSpecName "tmpfs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 13 06:49:23 crc kubenswrapper[4971]: I1213 06:49:23.796002 4971 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/22c825df-677d-4ca6-82db-3454ed06e783-kube-api-access-7c4vf" (OuterVolumeSpecName: "kube-api-access-7c4vf") pod "22c825df-677d-4ca6-82db-3454ed06e783" (UID: "22c825df-677d-4ca6-82db-3454ed06e783"). InnerVolumeSpecName "kube-api-access-7c4vf". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 13 06:49:23 crc kubenswrapper[4971]: I1213 06:49:23.790571 4971 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-13T06:49:23Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-13T06:49:23Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Dec 13 06:49:23 crc kubenswrapper[4971]: I1213 06:49:23.796115 4971 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/6509e943-70c6-444c-bc41-48a544e36fbd-kube-api-access-6g6sz" (OuterVolumeSpecName: "kube-api-access-6g6sz") pod "6509e943-70c6-444c-bc41-48a544e36fbd" (UID: "6509e943-70c6-444c-bc41-48a544e36fbd"). InnerVolumeSpecName "kube-api-access-6g6sz". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 13 06:49:23 crc kubenswrapper[4971]: I1213 06:49:23.796307 4971 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/e7e6199b-1264-4501-8953-767f51328d08-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "e7e6199b-1264-4501-8953-767f51328d08" (UID: "e7e6199b-1264-4501-8953-767f51328d08"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 13 06:49:23 crc kubenswrapper[4971]: I1213 06:49:23.796366 4971 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "8cea82b4-6893-4ddc-af9f-1bb5ae425c5b" (UID: "8cea82b4-6893-4ddc-af9f-1bb5ae425c5b"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 13 06:49:23 crc kubenswrapper[4971]: I1213 06:49:23.797063 4971 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/7583ce53-e0fe-4a16-9e4d-50516596a136-client-ca" (OuterVolumeSpecName: "client-ca") pod "7583ce53-e0fe-4a16-9e4d-50516596a136" (UID: "7583ce53-e0fe-4a16-9e4d-50516596a136"). InnerVolumeSpecName "client-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 13 06:49:23 crc kubenswrapper[4971]: I1213 06:49:23.797315 4971 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/ef543e1b-8068-4ea3-b32a-61027b32e95d-env-overrides\") pod \"network-node-identity-vrzqb\" (UID: \"ef543e1b-8068-4ea3-b32a-61027b32e95d\") " pod="openshift-network-node-identity/network-node-identity-vrzqb" Dec 13 06:49:23 crc kubenswrapper[4971]: I1213 06:49:23.797918 4971 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/7583ce53-e0fe-4a16-9e4d-50516596a136-proxy-ca-bundles" (OuterVolumeSpecName: "proxy-ca-bundles") pod "7583ce53-e0fe-4a16-9e4d-50516596a136" (UID: "7583ce53-e0fe-4a16-9e4d-50516596a136"). InnerVolumeSpecName "proxy-ca-bundles". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 13 06:49:23 crc kubenswrapper[4971]: I1213 06:49:23.798091 4971 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/5441d097-087c-4d9a-baa8-b210afa90fc9-config" (OuterVolumeSpecName: "config") pod "5441d097-087c-4d9a-baa8-b210afa90fc9" (UID: "5441d097-087c-4d9a-baa8-b210afa90fc9"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 13 06:49:23 crc kubenswrapper[4971]: I1213 06:49:23.798267 4971 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/9d4552c7-cd75-42dd-8880-30dd377c49a4-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "9d4552c7-cd75-42dd-8880-30dd377c49a4" (UID: "9d4552c7-cd75-42dd-8880-30dd377c49a4"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 13 06:49:23 crc kubenswrapper[4971]: I1213 06:49:23.797261 4971 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/7583ce53-e0fe-4a16-9e4d-50516596a136-config" (OuterVolumeSpecName: "config") pod "7583ce53-e0fe-4a16-9e4d-50516596a136" (UID: "7583ce53-e0fe-4a16-9e4d-50516596a136"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 13 06:49:23 crc kubenswrapper[4971]: I1213 06:49:23.800683 4971 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/43509403-f426-496e-be36-56cef71462f5-console-serving-cert" (OuterVolumeSpecName: "console-serving-cert") pod "43509403-f426-496e-be36-56cef71462f5" (UID: "43509403-f426-496e-be36-56cef71462f5"). InnerVolumeSpecName "console-serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 13 06:49:23 crc kubenswrapper[4971]: I1213 06:49:23.800858 4971 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/09efc573-dbb6-4249-bd59-9b87aba8dd28-etcd-ca" (OuterVolumeSpecName: "etcd-ca") pod "09efc573-dbb6-4249-bd59-9b87aba8dd28" (UID: "09efc573-dbb6-4249-bd59-9b87aba8dd28"). InnerVolumeSpecName "etcd-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 13 06:49:23 crc kubenswrapper[4971]: I1213 06:49:23.800885 4971 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/87cf06ed-a83f-41a7-828d-70653580a8cb-config-volume" (OuterVolumeSpecName: "config-volume") pod "87cf06ed-a83f-41a7-828d-70653580a8cb" (UID: "87cf06ed-a83f-41a7-828d-70653580a8cb"). InnerVolumeSpecName "config-volume". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 13 06:49:23 crc kubenswrapper[4971]: I1213 06:49:23.801009 4971 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/87cf06ed-a83f-41a7-828d-70653580a8cb-metrics-tls" (OuterVolumeSpecName: "metrics-tls") pod "87cf06ed-a83f-41a7-828d-70653580a8cb" (UID: "87cf06ed-a83f-41a7-828d-70653580a8cb"). InnerVolumeSpecName "metrics-tls". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 13 06:49:23 crc kubenswrapper[4971]: I1213 06:49:23.802401 4971 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/6402fda4-df10-493c-b4e5-d0569419652d-images" (OuterVolumeSpecName: "images") pod "6402fda4-df10-493c-b4e5-d0569419652d" (UID: "6402fda4-df10-493c-b4e5-d0569419652d"). InnerVolumeSpecName "images". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 13 06:49:23 crc kubenswrapper[4971]: E1213 06:49:23.808023 4971 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Dec 13 06:49:23 crc kubenswrapper[4971]: E1213 06:49:23.808054 4971 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Dec 13 06:49:23 crc kubenswrapper[4971]: E1213 06:49:23.808067 4971 projected.go:194] Error preparing data for projected volume kube-api-access-cqllr for pod openshift-network-diagnostics/network-check-target-xd92c: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Dec 13 06:49:23 crc kubenswrapper[4971]: E1213 06:49:23.808119 4971 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr podName:3b6479f0-333b-4a96-9adf-2099afdc2447 nodeName:}" failed. No retries permitted until 2025-12-13 06:49:24.308103316 +0000 UTC m=+20.912512764 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "kube-api-access-cqllr" (UniqueName: "kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr") pod "network-check-target-xd92c" (UID: "3b6479f0-333b-4a96-9adf-2099afdc2447") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Dec 13 06:49:23 crc kubenswrapper[4971]: I1213 06:49:23.810479 4971 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-rczfb\" (UniqueName: \"kubernetes.io/projected/d75a4c96-2883-4a0b-bab2-0fab2b6c0b49-kube-api-access-rczfb\") pod \"iptables-alerter-4ln5h\" (UID: \"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\") " pod="openshift-network-operator/iptables-alerter-4ln5h" Dec 13 06:49:23 crc kubenswrapper[4971]: I1213 06:49:23.810879 4971 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"f9d31b5a-b770-4a15-95c8-35b49677c608\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T06:49:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T06:49:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T06:49:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T06:49:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T06:49:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c9baa042f9dada954faa57b5454cd6b0a12a297e56ddc192c52cff50e42a16fa\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-13T06:49:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://c788de86754f0bf37ee184672fc1dcb388ccebc246d4f15be4b5da661f725f8f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-13T06:49:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://760817511a9b4f96a68dc38d26631250eaab72e1dd8f123e653039b6b4e91f84\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-13T06:49:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b859be0c9feac856e9855c6a038b02b49b32fade97873c46d0bec0a32d8b5147\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-13T06:49:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-13T06:49:03Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Dec 13 06:49:23 crc kubenswrapper[4971]: I1213 06:49:23.811184 4971 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/57a731c4-ef35-47a8-b875-bfb08a7f8011-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "57a731c4-ef35-47a8-b875-bfb08a7f8011" (UID: "57a731c4-ef35-47a8-b875-bfb08a7f8011"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 13 06:49:23 crc kubenswrapper[4971]: I1213 06:49:23.811922 4971 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-s2kz5\" (UniqueName: \"kubernetes.io/projected/ef543e1b-8068-4ea3-b32a-61027b32e95d-kube-api-access-s2kz5\") pod \"network-node-identity-vrzqb\" (UID: \"ef543e1b-8068-4ea3-b32a-61027b32e95d\") " pod="openshift-network-node-identity/network-node-identity-vrzqb" Dec 13 06:49:23 crc kubenswrapper[4971]: E1213 06:49:23.814486 4971 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Dec 13 06:49:23 crc kubenswrapper[4971]: E1213 06:49:23.814540 4971 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Dec 13 06:49:23 crc kubenswrapper[4971]: E1213 06:49:23.814558 4971 projected.go:194] Error preparing data for projected volume kube-api-access-s2dwl for pod openshift-network-diagnostics/network-check-source-55646444c4-trplf: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Dec 13 06:49:23 crc kubenswrapper[4971]: E1213 06:49:23.814631 4971 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl podName:9d751cbb-f2e2-430d-9754-c882a5e924a5 nodeName:}" failed. No retries permitted until 2025-12-13 06:49:24.314607723 +0000 UTC m=+20.919017171 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "kube-api-access-s2dwl" (UniqueName: "kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl") pod "network-check-source-55646444c4-trplf" (UID: "9d751cbb-f2e2-430d-9754-c882a5e924a5") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Dec 13 06:49:23 crc kubenswrapper[4971]: I1213 06:49:23.815614 4971 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/0b78653f-4ff9-4508-8672-245ed9b561e3-kube-api-access" (OuterVolumeSpecName: "kube-api-access") pod "0b78653f-4ff9-4508-8672-245ed9b561e3" (UID: "0b78653f-4ff9-4508-8672-245ed9b561e3"). InnerVolumeSpecName "kube-api-access". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 13 06:49:23 crc kubenswrapper[4971]: I1213 06:49:23.815684 4971 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/37a5e44f-9a88-4405-be8a-b645485e7312-metrics-tls\") pod \"network-operator-58b4c7f79c-55gtf\" (UID: \"37a5e44f-9a88-4405-be8a-b645485e7312\") " pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" Dec 13 06:49:23 crc kubenswrapper[4971]: I1213 06:49:23.815869 4971 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/7583ce53-e0fe-4a16-9e4d-50516596a136-kube-api-access-xcphl" (OuterVolumeSpecName: "kube-api-access-xcphl") pod "7583ce53-e0fe-4a16-9e4d-50516596a136" (UID: "7583ce53-e0fe-4a16-9e4d-50516596a136"). InnerVolumeSpecName "kube-api-access-xcphl". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 13 06:49:23 crc kubenswrapper[4971]: I1213 06:49:23.816054 4971 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/4bb40260-dbaa-4fb0-84df-5e680505d512-multus-daemon-config" (OuterVolumeSpecName: "multus-daemon-config") pod "4bb40260-dbaa-4fb0-84df-5e680505d512" (UID: "4bb40260-dbaa-4fb0-84df-5e680505d512"). InnerVolumeSpecName "multus-daemon-config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 13 06:49:23 crc kubenswrapper[4971]: I1213 06:49:23.816317 4971 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/20b0d48f-5fd6-431c-a545-e3c800c7b866-cert" (OuterVolumeSpecName: "cert") pod "20b0d48f-5fd6-431c-a545-e3c800c7b866" (UID: "20b0d48f-5fd6-431c-a545-e3c800c7b866"). InnerVolumeSpecName "cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 13 06:49:23 crc kubenswrapper[4971]: I1213 06:49:23.816900 4971 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/6ea678ab-3438-413e-bfe3-290ae7725660-kube-api-access-htfz6" (OuterVolumeSpecName: "kube-api-access-htfz6") pod "6ea678ab-3438-413e-bfe3-290ae7725660" (UID: "6ea678ab-3438-413e-bfe3-290ae7725660"). InnerVolumeSpecName "kube-api-access-htfz6". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 13 06:49:23 crc kubenswrapper[4971]: I1213 06:49:23.817026 4971 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/6ea678ab-3438-413e-bfe3-290ae7725660-ovnkube-script-lib" (OuterVolumeSpecName: "ovnkube-script-lib") pod "6ea678ab-3438-413e-bfe3-290ae7725660" (UID: "6ea678ab-3438-413e-bfe3-290ae7725660"). InnerVolumeSpecName "ovnkube-script-lib". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 13 06:49:23 crc kubenswrapper[4971]: I1213 06:49:23.817304 4971 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/31d8b7a1-420e-4252-a5b7-eebe8a111292-images" (OuterVolumeSpecName: "images") pod "31d8b7a1-420e-4252-a5b7-eebe8a111292" (UID: "31d8b7a1-420e-4252-a5b7-eebe8a111292"). InnerVolumeSpecName "images". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 13 06:49:23 crc kubenswrapper[4971]: I1213 06:49:23.817506 4971 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/6731426b-95fe-49ff-bb5f-40441049fde2-kube-api-access-x7zkh" (OuterVolumeSpecName: "kube-api-access-x7zkh") pod "6731426b-95fe-49ff-bb5f-40441049fde2" (UID: "6731426b-95fe-49ff-bb5f-40441049fde2"). InnerVolumeSpecName "kube-api-access-x7zkh". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 13 06:49:23 crc kubenswrapper[4971]: I1213 06:49:23.817587 4971 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/6ea678ab-3438-413e-bfe3-290ae7725660-env-overrides" (OuterVolumeSpecName: "env-overrides") pod "6ea678ab-3438-413e-bfe3-290ae7725660" (UID: "6ea678ab-3438-413e-bfe3-290ae7725660"). InnerVolumeSpecName "env-overrides". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 13 06:49:23 crc kubenswrapper[4971]: I1213 06:49:23.819426 4971 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-rdwmf\" (UniqueName: \"kubernetes.io/projected/37a5e44f-9a88-4405-be8a-b645485e7312-kube-api-access-rdwmf\") pod \"network-operator-58b4c7f79c-55gtf\" (UID: \"37a5e44f-9a88-4405-be8a-b645485e7312\") " pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" Dec 13 06:49:23 crc kubenswrapper[4971]: I1213 06:49:23.820586 4971 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/09efc573-dbb6-4249-bd59-9b87aba8dd28-config" (OuterVolumeSpecName: "config") pod "09efc573-dbb6-4249-bd59-9b87aba8dd28" (UID: "09efc573-dbb6-4249-bd59-9b87aba8dd28"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 13 06:49:23 crc kubenswrapper[4971]: I1213 06:49:23.821729 4971 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-13T06:49:23Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-operator]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-13T06:49:23Z\\\",\\\"message\\\":\\\"containers with unready status: [network-operator]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":false,\\\"restartCount\\\":5,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Dec 13 06:49:23 crc kubenswrapper[4971]: I1213 06:49:23.821976 4971 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/ef543e1b-8068-4ea3-b32a-61027b32e95d-webhook-cert\") pod \"network-node-identity-vrzqb\" (UID: \"ef543e1b-8068-4ea3-b32a-61027b32e95d\") " pod="openshift-network-node-identity/network-node-identity-vrzqb" Dec 13 06:49:23 crc kubenswrapper[4971]: I1213 06:49:23.822417 4971 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/bd23aa5c-e532-4e53-bccf-e79f130c5ae8-kube-api-access-jhbk2" (OuterVolumeSpecName: "kube-api-access-jhbk2") pod "bd23aa5c-e532-4e53-bccf-e79f130c5ae8" (UID: "bd23aa5c-e532-4e53-bccf-e79f130c5ae8"). InnerVolumeSpecName "kube-api-access-jhbk2". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 13 06:49:23 crc kubenswrapper[4971]: I1213 06:49:23.822742 4971 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"iptables-alerter-script\" (UniqueName: \"kubernetes.io/configmap/d75a4c96-2883-4a0b-bab2-0fab2b6c0b49-iptables-alerter-script\") pod \"iptables-alerter-4ln5h\" (UID: \"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\") " pod="openshift-network-operator/iptables-alerter-4ln5h" Dec 13 06:49:23 crc kubenswrapper[4971]: I1213 06:49:23.823039 4971 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/31d8b7a1-420e-4252-a5b7-eebe8a111292-auth-proxy-config" (OuterVolumeSpecName: "auth-proxy-config") pod "31d8b7a1-420e-4252-a5b7-eebe8a111292" (UID: "31d8b7a1-420e-4252-a5b7-eebe8a111292"). InnerVolumeSpecName "auth-proxy-config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 13 06:49:23 crc kubenswrapper[4971]: I1213 06:49:23.825437 4971 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovnkube-identity-cm\" (UniqueName: \"kubernetes.io/configmap/ef543e1b-8068-4ea3-b32a-61027b32e95d-ovnkube-identity-cm\") pod \"network-node-identity-vrzqb\" (UID: \"ef543e1b-8068-4ea3-b32a-61027b32e95d\") " pod="openshift-network-node-identity/network-node-identity-vrzqb" Dec 13 06:49:23 crc kubenswrapper[4971]: I1213 06:49:23.825970 4971 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/25e176fe-21b4-4974-b1ed-c8b94f112a7f-signing-cabundle" (OuterVolumeSpecName: "signing-cabundle") pod "25e176fe-21b4-4974-b1ed-c8b94f112a7f" (UID: "25e176fe-21b4-4974-b1ed-c8b94f112a7f"). InnerVolumeSpecName "signing-cabundle". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 13 06:49:23 crc kubenswrapper[4971]: I1213 06:49:23.826508 4971 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/44663579-783b-4372-86d6-acf235a62d72-kube-api-access-vt5rc" (OuterVolumeSpecName: "kube-api-access-vt5rc") pod "44663579-783b-4372-86d6-acf235a62d72" (UID: "44663579-783b-4372-86d6-acf235a62d72"). InnerVolumeSpecName "kube-api-access-vt5rc". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 13 06:49:23 crc kubenswrapper[4971]: I1213 06:49:23.829417 4971 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/bf126b07-da06-4140-9a57-dfd54fc6b486-image-registry-operator-tls" (OuterVolumeSpecName: "image-registry-operator-tls") pod "bf126b07-da06-4140-9a57-dfd54fc6b486" (UID: "bf126b07-da06-4140-9a57-dfd54fc6b486"). InnerVolumeSpecName "image-registry-operator-tls". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 13 06:49:23 crc kubenswrapper[4971]: I1213 06:49:23.829881 4971 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/b6312bbd-5731-4ea0-a20f-81d5a57df44a-profile-collector-cert" (OuterVolumeSpecName: "profile-collector-cert") pod "b6312bbd-5731-4ea0-a20f-81d5a57df44a" (UID: "b6312bbd-5731-4ea0-a20f-81d5a57df44a"). InnerVolumeSpecName "profile-collector-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 13 06:49:23 crc kubenswrapper[4971]: I1213 06:49:23.832468 4971 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/25e176fe-21b4-4974-b1ed-c8b94f112a7f-signing-key" (OuterVolumeSpecName: "signing-key") pod "25e176fe-21b4-4974-b1ed-c8b94f112a7f" (UID: "25e176fe-21b4-4974-b1ed-c8b94f112a7f"). InnerVolumeSpecName "signing-key". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 13 06:49:23 crc kubenswrapper[4971]: I1213 06:49:23.836739 4971 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "b11524ee-3fca-4b1b-9cdf-6da289fdbc7d" (UID: "b11524ee-3fca-4b1b-9cdf-6da289fdbc7d"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 13 06:49:23 crc kubenswrapper[4971]: I1213 06:49:23.842636 4971 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/b6312bbd-5731-4ea0-a20f-81d5a57df44a-kube-api-access-249nr" (OuterVolumeSpecName: "kube-api-access-249nr") pod "b6312bbd-5731-4ea0-a20f-81d5a57df44a" (UID: "b6312bbd-5731-4ea0-a20f-81d5a57df44a"). InnerVolumeSpecName "kube-api-access-249nr". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 13 06:49:23 crc kubenswrapper[4971]: I1213 06:49:23.842768 4971 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c-config" (OuterVolumeSpecName: "config") pod "210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c" (UID: "210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 13 06:49:23 crc kubenswrapper[4971]: I1213 06:49:23.842912 4971 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/1bf7eb37-55a3-4c65-b768-a94c82151e69-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "1bf7eb37-55a3-4c65-b768-a94c82151e69" (UID: "1bf7eb37-55a3-4c65-b768-a94c82151e69"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 13 06:49:23 crc kubenswrapper[4971]: I1213 06:49:23.846631 4971 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/1386a44e-36a2-460c-96d0-0359d2b6f0f5-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "1386a44e-36a2-460c-96d0-0359d2b6f0f5" (UID: "1386a44e-36a2-460c-96d0-0359d2b6f0f5"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 13 06:49:23 crc kubenswrapper[4971]: I1213 06:49:23.849933 4971 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/1bf7eb37-55a3-4c65-b768-a94c82151e69-kube-api-access-sb6h7" (OuterVolumeSpecName: "kube-api-access-sb6h7") pod "1bf7eb37-55a3-4c65-b768-a94c82151e69" (UID: "1bf7eb37-55a3-4c65-b768-a94c82151e69"). InnerVolumeSpecName "kube-api-access-sb6h7". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 13 06:49:23 crc kubenswrapper[4971]: I1213 06:49:23.855782 4971 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/1d611f23-29be-4491-8495-bee1670e935f-utilities" (OuterVolumeSpecName: "utilities") pod "1d611f23-29be-4491-8495-bee1670e935f" (UID: "1d611f23-29be-4491-8495-bee1670e935f"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 13 06:49:23 crc kubenswrapper[4971]: I1213 06:49:23.856013 4971 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-audit-policies" (OuterVolumeSpecName: "audit-policies") pod "49c341d1-5089-4bc2-86a0-a5e165cfcc6b" (UID: "49c341d1-5089-4bc2-86a0-a5e165cfcc6b"). InnerVolumeSpecName "audit-policies". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 13 06:49:23 crc kubenswrapper[4971]: I1213 06:49:23.858893 4971 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/1d611f23-29be-4491-8495-bee1670e935f-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "1d611f23-29be-4491-8495-bee1670e935f" (UID: "1d611f23-29be-4491-8495-bee1670e935f"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 13 06:49:23 crc kubenswrapper[4971]: I1213 06:49:23.858907 4971 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-template-error" (OuterVolumeSpecName: "v4-0-config-user-template-error") pod "49c341d1-5089-4bc2-86a0-a5e165cfcc6b" (UID: "49c341d1-5089-4bc2-86a0-a5e165cfcc6b"). InnerVolumeSpecName "v4-0-config-user-template-error". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 13 06:49:23 crc kubenswrapper[4971]: I1213 06:49:23.859168 4971 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/1386a44e-36a2-460c-96d0-0359d2b6f0f5-config" (OuterVolumeSpecName: "config") pod "1386a44e-36a2-460c-96d0-0359d2b6f0f5" (UID: "1386a44e-36a2-460c-96d0-0359d2b6f0f5"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 13 06:49:23 crc kubenswrapper[4971]: I1213 06:49:23.862140 4971 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59-serviceca" (OuterVolumeSpecName: "serviceca") pod "3cb93b32-e0ae-4377-b9c8-fdb9842c6d59" (UID: "3cb93b32-e0ae-4377-b9c8-fdb9842c6d59"). InnerVolumeSpecName "serviceca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 13 06:49:23 crc kubenswrapper[4971]: I1213 06:49:23.874499 4971 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/22c825df-677d-4ca6-82db-3454ed06e783-auth-proxy-config" (OuterVolumeSpecName: "auth-proxy-config") pod "22c825df-677d-4ca6-82db-3454ed06e783" (UID: "22c825df-677d-4ca6-82db-3454ed06e783"). InnerVolumeSpecName "auth-proxy-config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 13 06:49:23 crc kubenswrapper[4971]: I1213 06:49:23.874822 4971 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/925f1c65-6136-48ba-85aa-3a3b50560753-ovn-control-plane-metrics-cert" (OuterVolumeSpecName: "ovn-control-plane-metrics-cert") pod "925f1c65-6136-48ba-85aa-3a3b50560753" (UID: "925f1c65-6136-48ba-85aa-3a3b50560753"). InnerVolumeSpecName "ovn-control-plane-metrics-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 13 06:49:23 crc kubenswrapper[4971]: I1213 06:49:23.874859 4971 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/c03ee662-fb2f-4fc4-a2c1-af487c19d254-metrics-certs" (OuterVolumeSpecName: "metrics-certs") pod "c03ee662-fb2f-4fc4-a2c1-af487c19d254" (UID: "c03ee662-fb2f-4fc4-a2c1-af487c19d254"). InnerVolumeSpecName "metrics-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 13 06:49:23 crc kubenswrapper[4971]: I1213 06:49:23.874872 4971 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/a31745f5-9847-4afe-82a5-3161cc66ca93-trusted-ca" (OuterVolumeSpecName: "trusted-ca") pod "a31745f5-9847-4afe-82a5-3161cc66ca93" (UID: "a31745f5-9847-4afe-82a5-3161cc66ca93"). InnerVolumeSpecName "trusted-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 13 06:49:23 crc kubenswrapper[4971]: I1213 06:49:23.874946 4971 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/22c825df-677d-4ca6-82db-3454ed06e783-config" (OuterVolumeSpecName: "config") pod "22c825df-677d-4ca6-82db-3454ed06e783" (UID: "22c825df-677d-4ca6-82db-3454ed06e783"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 13 06:49:23 crc kubenswrapper[4971]: I1213 06:49:23.881058 4971 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c-kube-api-access-qs4fp" (OuterVolumeSpecName: "kube-api-access-qs4fp") pod "210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c" (UID: "210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c"). InnerVolumeSpecName "kube-api-access-qs4fp". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 13 06:49:23 crc kubenswrapper[4971]: I1213 06:49:23.881192 4971 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/9d4552c7-cd75-42dd-8880-30dd377c49a4-trusted-ca" (OuterVolumeSpecName: "trusted-ca") pod "9d4552c7-cd75-42dd-8880-30dd377c49a4" (UID: "9d4552c7-cd75-42dd-8880-30dd377c49a4"). InnerVolumeSpecName "trusted-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 13 06:49:23 crc kubenswrapper[4971]: I1213 06:49:23.881893 4971 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-serving-cert" (OuterVolumeSpecName: "v4-0-config-system-serving-cert") pod "49c341d1-5089-4bc2-86a0-a5e165cfcc6b" (UID: "49c341d1-5089-4bc2-86a0-a5e165cfcc6b"). InnerVolumeSpecName "v4-0-config-system-serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 13 06:49:23 crc kubenswrapper[4971]: I1213 06:49:23.881991 4971 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-trusted-ca-bundle" (OuterVolumeSpecName: "v4-0-config-system-trusted-ca-bundle") pod "49c341d1-5089-4bc2-86a0-a5e165cfcc6b" (UID: "49c341d1-5089-4bc2-86a0-a5e165cfcc6b"). InnerVolumeSpecName "v4-0-config-system-trusted-ca-bundle". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 13 06:49:23 crc kubenswrapper[4971]: I1213 06:49:23.882386 4971 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-13T06:49:23Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-13T06:49:23Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Dec 13 06:49:23 crc kubenswrapper[4971]: I1213 06:49:23.882628 4971 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-template-login" (OuterVolumeSpecName: "v4-0-config-user-template-login") pod "49c341d1-5089-4bc2-86a0-a5e165cfcc6b" (UID: "49c341d1-5089-4bc2-86a0-a5e165cfcc6b"). InnerVolumeSpecName "v4-0-config-user-template-login". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 13 06:49:23 crc kubenswrapper[4971]: I1213 06:49:23.882709 4971 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-idp-0-file-data" (OuterVolumeSpecName: "v4-0-config-user-idp-0-file-data") pod "49c341d1-5089-4bc2-86a0-a5e165cfcc6b" (UID: "49c341d1-5089-4bc2-86a0-a5e165cfcc6b"). InnerVolumeSpecName "v4-0-config-user-idp-0-file-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 13 06:49:23 crc kubenswrapper[4971]: I1213 06:49:23.883394 4971 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-cliconfig" (OuterVolumeSpecName: "v4-0-config-system-cliconfig") pod "49c341d1-5089-4bc2-86a0-a5e165cfcc6b" (UID: "49c341d1-5089-4bc2-86a0-a5e165cfcc6b"). InnerVolumeSpecName "v4-0-config-system-cliconfig". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 13 06:49:23 crc kubenswrapper[4971]: I1213 06:49:23.884292 4971 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/8f668bae-612b-4b75-9490-919e737c6a3b-ca-trust-extracted" (OuterVolumeSpecName: "ca-trust-extracted") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b"). InnerVolumeSpecName "ca-trust-extracted". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 13 06:49:23 crc kubenswrapper[4971]: I1213 06:49:23.891297 4971 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-slash\" (UniqueName: \"kubernetes.io/host-path/d75a4c96-2883-4a0b-bab2-0fab2b6c0b49-host-slash\") pod \"iptables-alerter-4ln5h\" (UID: \"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\") " pod="openshift-network-operator/iptables-alerter-4ln5h" Dec 13 06:49:23 crc kubenswrapper[4971]: I1213 06:49:23.891357 4971 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-etc-kube\" (UniqueName: \"kubernetes.io/host-path/37a5e44f-9a88-4405-be8a-b645485e7312-host-etc-kube\") pod \"network-operator-58b4c7f79c-55gtf\" (UID: \"37a5e44f-9a88-4405-be8a-b645485e7312\") " pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" Dec 13 06:49:23 crc kubenswrapper[4971]: I1213 06:49:23.891398 4971 reconciler_common.go:293] "Volume detached for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-trusted-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 13 06:49:23 crc kubenswrapper[4971]: I1213 06:49:23.891408 4971 reconciler_common.go:293] "Volume detached for volume \"ca-trust-extracted\" (UniqueName: \"kubernetes.io/empty-dir/8f668bae-612b-4b75-9490-919e737c6a3b-ca-trust-extracted\") on node \"crc\" DevicePath \"\"" Dec 13 06:49:23 crc kubenswrapper[4971]: I1213 06:49:23.891417 4971 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-wxkg8\" (UniqueName: \"kubernetes.io/projected/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59-kube-api-access-wxkg8\") on node \"crc\" DevicePath \"\"" Dec 13 06:49:23 crc kubenswrapper[4971]: I1213 06:49:23.891425 4971 reconciler_common.go:293] "Volume detached for volume \"etcd-serving-ca\" (UniqueName: \"kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-etcd-serving-ca\") on node \"crc\" DevicePath \"\"" Dec 13 06:49:23 crc kubenswrapper[4971]: I1213 06:49:23.891433 4971 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d-catalog-content\") on node \"crc\" DevicePath \"\"" Dec 13 06:49:23 crc kubenswrapper[4971]: I1213 06:49:23.891442 4971 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/496e6271-fb68-4057-954e-a0d97a4afa3f-config\") on node \"crc\" DevicePath \"\"" Dec 13 06:49:23 crc kubenswrapper[4971]: I1213 06:49:23.891450 4971 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/1d611f23-29be-4491-8495-bee1670e935f-catalog-content\") on node \"crc\" DevicePath \"\"" Dec 13 06:49:23 crc kubenswrapper[4971]: I1213 06:49:23.891460 4971 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c-config\") on node \"crc\" DevicePath \"\"" Dec 13 06:49:23 crc kubenswrapper[4971]: I1213 06:49:23.891468 4971 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-jhbk2\" (UniqueName: \"kubernetes.io/projected/bd23aa5c-e532-4e53-bccf-e79f130c5ae8-kube-api-access-jhbk2\") on node \"crc\" DevicePath \"\"" Dec 13 06:49:23 crc kubenswrapper[4971]: I1213 06:49:23.891475 4971 reconciler_common.go:293] "Volume detached for volume \"encryption-config\" (UniqueName: \"kubernetes.io/secret/1bf7eb37-55a3-4c65-b768-a94c82151e69-encryption-config\") on node \"crc\" DevicePath \"\"" Dec 13 06:49:23 crc kubenswrapper[4971]: I1213 06:49:23.891483 4971 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-x4zgh\" (UniqueName: \"kubernetes.io/projected/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d-kube-api-access-x4zgh\") on node \"crc\" DevicePath \"\"" Dec 13 06:49:23 crc kubenswrapper[4971]: I1213 06:49:23.891492 4971 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/6509e943-70c6-444c-bc41-48a544e36fbd-serving-cert\") on node \"crc\" DevicePath \"\"" Dec 13 06:49:23 crc kubenswrapper[4971]: I1213 06:49:23.891500 4971 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-gf66m\" (UniqueName: \"kubernetes.io/projected/a0128f3a-b052-44ed-a84e-c4c8aaf17c13-kube-api-access-gf66m\") on node \"crc\" DevicePath \"\"" Dec 13 06:49:23 crc kubenswrapper[4971]: I1213 06:49:23.891508 4971 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/57a731c4-ef35-47a8-b875-bfb08a7f8011-catalog-content\") on node \"crc\" DevicePath \"\"" Dec 13 06:49:23 crc kubenswrapper[4971]: I1213 06:49:23.891533 4971 reconciler_common.go:293] "Volume detached for volume \"cert\" (UniqueName: \"kubernetes.io/secret/20b0d48f-5fd6-431c-a545-e3c800c7b866-cert\") on node \"crc\" DevicePath \"\"" Dec 13 06:49:23 crc kubenswrapper[4971]: I1213 06:49:23.891541 4971 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-vt5rc\" (UniqueName: \"kubernetes.io/projected/44663579-783b-4372-86d6-acf235a62d72-kube-api-access-vt5rc\") on node \"crc\" DevicePath \"\"" Dec 13 06:49:23 crc kubenswrapper[4971]: I1213 06:49:23.891549 4971 reconciler_common.go:293] "Volume detached for volume \"default-certificate\" (UniqueName: \"kubernetes.io/secret/c03ee662-fb2f-4fc4-a2c1-af487c19d254-default-certificate\") on node \"crc\" DevicePath \"\"" Dec 13 06:49:23 crc kubenswrapper[4971]: I1213 06:49:23.891557 4971 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-249nr\" (UniqueName: \"kubernetes.io/projected/b6312bbd-5731-4ea0-a20f-81d5a57df44a-kube-api-access-249nr\") on node \"crc\" DevicePath \"\"" Dec 13 06:49:23 crc kubenswrapper[4971]: I1213 06:49:23.891565 4971 reconciler_common.go:293] "Volume detached for volume \"cni-binary-copy\" (UniqueName: \"kubernetes.io/configmap/4bb40260-dbaa-4fb0-84df-5e680505d512-cni-binary-copy\") on node \"crc\" DevicePath \"\"" Dec 13 06:49:23 crc kubenswrapper[4971]: I1213 06:49:23.891573 4971 reconciler_common.go:293] "Volume detached for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/09efc573-dbb6-4249-bd59-9b87aba8dd28-etcd-client\") on node \"crc\" DevicePath \"\"" Dec 13 06:49:23 crc kubenswrapper[4971]: I1213 06:49:23.891581 4971 reconciler_common.go:293] "Volume detached for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/a31745f5-9847-4afe-82a5-3161cc66ca93-bound-sa-token\") on node \"crc\" DevicePath \"\"" Dec 13 06:49:23 crc kubenswrapper[4971]: I1213 06:49:23.891589 4971 reconciler_common.go:293] "Volume detached for volume \"multus-daemon-config\" (UniqueName: \"kubernetes.io/configmap/4bb40260-dbaa-4fb0-84df-5e680505d512-multus-daemon-config\") on node \"crc\" DevicePath \"\"" Dec 13 06:49:23 crc kubenswrapper[4971]: I1213 06:49:23.891596 4971 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-d6qdx\" (UniqueName: \"kubernetes.io/projected/87cf06ed-a83f-41a7-828d-70653580a8cb-kube-api-access-d6qdx\") on node \"crc\" DevicePath \"\"" Dec 13 06:49:23 crc kubenswrapper[4971]: I1213 06:49:23.891606 4971 reconciler_common.go:293] "Volume detached for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-audit-policies\") on node \"crc\" DevicePath \"\"" Dec 13 06:49:23 crc kubenswrapper[4971]: I1213 06:49:23.891613 4971 reconciler_common.go:293] "Volume detached for volume \"image-registry-operator-tls\" (UniqueName: \"kubernetes.io/secret/bf126b07-da06-4140-9a57-dfd54fc6b486-image-registry-operator-tls\") on node \"crc\" DevicePath \"\"" Dec 13 06:49:23 crc kubenswrapper[4971]: I1213 06:49:23.891622 4971 reconciler_common.go:293] "Volume detached for volume \"signing-key\" (UniqueName: \"kubernetes.io/secret/25e176fe-21b4-4974-b1ed-c8b94f112a7f-signing-key\") on node \"crc\" DevicePath \"\"" Dec 13 06:49:23 crc kubenswrapper[4971]: I1213 06:49:23.891632 4971 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-d4lsv\" (UniqueName: \"kubernetes.io/projected/25e176fe-21b4-4974-b1ed-c8b94f112a7f-kube-api-access-d4lsv\") on node \"crc\" DevicePath \"\"" Dec 13 06:49:23 crc kubenswrapper[4971]: I1213 06:49:23.891640 4971 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/7539238d-5fe0-46ed-884e-1c3b566537ec-serving-cert\") on node \"crc\" DevicePath \"\"" Dec 13 06:49:23 crc kubenswrapper[4971]: I1213 06:49:23.891648 4971 reconciler_common.go:293] "Volume detached for volume \"machine-api-operator-tls\" (UniqueName: \"kubernetes.io/secret/6402fda4-df10-493c-b4e5-d0569419652d-machine-api-operator-tls\") on node \"crc\" DevicePath \"\"" Dec 13 06:49:23 crc kubenswrapper[4971]: I1213 06:49:23.891658 4971 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-lz9wn\" (UniqueName: \"kubernetes.io/projected/a31745f5-9847-4afe-82a5-3161cc66ca93-kube-api-access-lz9wn\") on node \"crc\" DevicePath \"\"" Dec 13 06:49:23 crc kubenswrapper[4971]: I1213 06:49:23.891666 4971 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-serving-cert\") on node \"crc\" DevicePath \"\"" Dec 13 06:49:23 crc kubenswrapper[4971]: I1213 06:49:23.891674 4971 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/5441d097-087c-4d9a-baa8-b210afa90fc9-serving-cert\") on node \"crc\" DevicePath \"\"" Dec 13 06:49:23 crc kubenswrapper[4971]: I1213 06:49:23.891681 4971 reconciler_common.go:293] "Volume detached for volume \"marketplace-trusted-ca\" (UniqueName: \"kubernetes.io/configmap/b6cd30de-2eeb-49a2-ab40-9167f4560ff5-marketplace-trusted-ca\") on node \"crc\" DevicePath \"\"" Dec 13 06:49:23 crc kubenswrapper[4971]: I1213 06:49:23.891689 4971 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-qg5z5\" (UniqueName: \"kubernetes.io/projected/43509403-f426-496e-be36-56cef71462f5-kube-api-access-qg5z5\") on node \"crc\" DevicePath \"\"" Dec 13 06:49:23 crc kubenswrapper[4971]: I1213 06:49:23.891697 4971 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-jkwtn\" (UniqueName: \"kubernetes.io/projected/5b88f790-22fa-440e-b583-365168c0b23d-kube-api-access-jkwtn\") on node \"crc\" DevicePath \"\"" Dec 13 06:49:23 crc kubenswrapper[4971]: I1213 06:49:23.891611 4971 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-slash\" (UniqueName: \"kubernetes.io/host-path/d75a4c96-2883-4a0b-bab2-0fab2b6c0b49-host-slash\") pod \"iptables-alerter-4ln5h\" (UID: \"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\") " pod="openshift-network-operator/iptables-alerter-4ln5h" Dec 13 06:49:23 crc kubenswrapper[4971]: I1213 06:49:23.891705 4971 reconciler_common.go:293] "Volume detached for volume \"serviceca\" (UniqueName: \"kubernetes.io/configmap/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59-serviceca\") on node \"crc\" DevicePath \"\"" Dec 13 06:49:23 crc kubenswrapper[4971]: I1213 06:49:23.891764 4971 reconciler_common.go:293] "Volume detached for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-service-ca\") on node \"crc\" DevicePath \"\"" Dec 13 06:49:23 crc kubenswrapper[4971]: I1213 06:49:23.891780 4971 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-user-template-login\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-template-login\") on node \"crc\" DevicePath \"\"" Dec 13 06:49:23 crc kubenswrapper[4971]: I1213 06:49:23.891793 4971 reconciler_common.go:293] "Volume detached for volume \"package-server-manager-serving-cert\" (UniqueName: \"kubernetes.io/secret/3ab1a177-2de0-46d9-b765-d0d0649bb42e-package-server-manager-serving-cert\") on node \"crc\" DevicePath \"\"" Dec 13 06:49:23 crc kubenswrapper[4971]: I1213 06:49:23.891809 4971 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/9d4552c7-cd75-42dd-8880-30dd377c49a4-serving-cert\") on node \"crc\" DevicePath \"\"" Dec 13 06:49:23 crc kubenswrapper[4971]: I1213 06:49:23.891818 4971 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-config\") on node \"crc\" DevicePath \"\"" Dec 13 06:49:23 crc kubenswrapper[4971]: I1213 06:49:23.891830 4971 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-v47cf\" (UniqueName: \"kubernetes.io/projected/c03ee662-fb2f-4fc4-a2c1-af487c19d254-kube-api-access-v47cf\") on node \"crc\" DevicePath \"\"" Dec 13 06:49:23 crc kubenswrapper[4971]: I1213 06:49:23.891843 4971 reconciler_common.go:293] "Volume detached for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/87cf06ed-a83f-41a7-828d-70653580a8cb-metrics-tls\") on node \"crc\" DevicePath \"\"" Dec 13 06:49:23 crc kubenswrapper[4971]: I1213 06:49:23.891686 4971 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-etc-kube\" (UniqueName: \"kubernetes.io/host-path/37a5e44f-9a88-4405-be8a-b645485e7312-host-etc-kube\") pod \"network-operator-58b4c7f79c-55gtf\" (UID: \"37a5e44f-9a88-4405-be8a-b645485e7312\") " pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" Dec 13 06:49:23 crc kubenswrapper[4971]: I1213 06:49:23.891856 4971 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/1bf7eb37-55a3-4c65-b768-a94c82151e69-serving-cert\") on node \"crc\" DevicePath \"\"" Dec 13 06:49:23 crc kubenswrapper[4971]: I1213 06:49:23.891890 4971 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-7c4vf\" (UniqueName: \"kubernetes.io/projected/22c825df-677d-4ca6-82db-3454ed06e783-kube-api-access-7c4vf\") on node \"crc\" DevicePath \"\"" Dec 13 06:49:23 crc kubenswrapper[4971]: I1213 06:49:23.891901 4971 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/1386a44e-36a2-460c-96d0-0359d2b6f0f5-config\") on node \"crc\" DevicePath \"\"" Dec 13 06:49:23 crc kubenswrapper[4971]: I1213 06:49:23.891909 4971 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/e7e6199b-1264-4501-8953-767f51328d08-serving-cert\") on node \"crc\" DevicePath \"\"" Dec 13 06:49:23 crc kubenswrapper[4971]: I1213 06:49:23.891919 4971 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-279lb\" (UniqueName: \"kubernetes.io/projected/7bb08738-c794-4ee8-9972-3a62ca171029-kube-api-access-279lb\") on node \"crc\" DevicePath \"\"" Dec 13 06:49:23 crc kubenswrapper[4971]: I1213 06:49:23.891927 4971 reconciler_common.go:293] "Volume detached for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/7583ce53-e0fe-4a16-9e4d-50516596a136-client-ca\") on node \"crc\" DevicePath \"\"" Dec 13 06:49:23 crc kubenswrapper[4971]: I1213 06:49:23.891934 4971 reconciler_common.go:293] "Volume detached for volume \"tmpfs\" (UniqueName: \"kubernetes.io/empty-dir/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-tmpfs\") on node \"crc\" DevicePath \"\"" Dec 13 06:49:23 crc kubenswrapper[4971]: I1213 06:49:23.891942 4971 reconciler_common.go:293] "Volume detached for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/87cf06ed-a83f-41a7-828d-70653580a8cb-config-volume\") on node \"crc\" DevicePath \"\"" Dec 13 06:49:23 crc kubenswrapper[4971]: I1213 06:49:23.891951 4971 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-user-idp-0-file-data\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-idp-0-file-data\") on node \"crc\" DevicePath \"\"" Dec 13 06:49:23 crc kubenswrapper[4971]: I1213 06:49:23.891960 4971 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-sb6h7\" (UniqueName: \"kubernetes.io/projected/1bf7eb37-55a3-4c65-b768-a94c82151e69-kube-api-access-sb6h7\") on node \"crc\" DevicePath \"\"" Dec 13 06:49:23 crc kubenswrapper[4971]: I1213 06:49:23.891970 4971 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b-serving-cert\") on node \"crc\" DevicePath \"\"" Dec 13 06:49:23 crc kubenswrapper[4971]: I1213 06:49:23.891978 4971 reconciler_common.go:293] "Volume detached for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/0b78653f-4ff9-4508-8672-245ed9b561e3-service-ca\") on node \"crc\" DevicePath \"\"" Dec 13 06:49:23 crc kubenswrapper[4971]: I1213 06:49:23.891991 4971 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-user-template-error\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-template-error\") on node \"crc\" DevicePath \"\"" Dec 13 06:49:23 crc kubenswrapper[4971]: I1213 06:49:23.891999 4971 reconciler_common.go:293] "Volume detached for volume \"console-serving-cert\" (UniqueName: \"kubernetes.io/secret/43509403-f426-496e-be36-56cef71462f5-console-serving-cert\") on node \"crc\" DevicePath \"\"" Dec 13 06:49:23 crc kubenswrapper[4971]: I1213 06:49:23.892007 4971 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/1386a44e-36a2-460c-96d0-0359d2b6f0f5-serving-cert\") on node \"crc\" DevicePath \"\"" Dec 13 06:49:23 crc kubenswrapper[4971]: I1213 06:49:23.892015 4971 reconciler_common.go:293] "Volume detached for volume \"auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/31d8b7a1-420e-4252-a5b7-eebe8a111292-auth-proxy-config\") on node \"crc\" DevicePath \"\"" Dec 13 06:49:23 crc kubenswrapper[4971]: I1213 06:49:23.892023 4971 reconciler_common.go:293] "Volume detached for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/fda69060-fa79-4696-b1a6-7980f124bf7c-proxy-tls\") on node \"crc\" DevicePath \"\"" Dec 13 06:49:23 crc kubenswrapper[4971]: I1213 06:49:23.892030 4971 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/1d611f23-29be-4491-8495-bee1670e935f-utilities\") on node \"crc\" DevicePath \"\"" Dec 13 06:49:23 crc kubenswrapper[4971]: I1213 06:49:23.892038 4971 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-zkvpv\" (UniqueName: \"kubernetes.io/projected/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-kube-api-access-zkvpv\") on node \"crc\" DevicePath \"\"" Dec 13 06:49:23 crc kubenswrapper[4971]: I1213 06:49:23.892047 4971 reconciler_common.go:293] "Volume detached for volume \"ovn-control-plane-metrics-cert\" (UniqueName: \"kubernetes.io/secret/925f1c65-6136-48ba-85aa-3a3b50560753-ovn-control-plane-metrics-cert\") on node \"crc\" DevicePath \"\"" Dec 13 06:49:23 crc kubenswrapper[4971]: I1213 06:49:23.892055 4971 reconciler_common.go:293] "Volume detached for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/7583ce53-e0fe-4a16-9e4d-50516596a136-proxy-ca-bundles\") on node \"crc\" DevicePath \"\"" Dec 13 06:49:23 crc kubenswrapper[4971]: I1213 06:49:23.892063 4971 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/7583ce53-e0fe-4a16-9e4d-50516596a136-config\") on node \"crc\" DevicePath \"\"" Dec 13 06:49:23 crc kubenswrapper[4971]: I1213 06:49:23.892071 4971 reconciler_common.go:293] "Volume detached for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/a31745f5-9847-4afe-82a5-3161cc66ca93-trusted-ca\") on node \"crc\" DevicePath \"\"" Dec 13 06:49:23 crc kubenswrapper[4971]: I1213 06:49:23.892079 4971 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-qs4fp\" (UniqueName: \"kubernetes.io/projected/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c-kube-api-access-qs4fp\") on node \"crc\" DevicePath \"\"" Dec 13 06:49:23 crc kubenswrapper[4971]: I1213 06:49:23.892088 4971 reconciler_common.go:293] "Volume detached for volume \"images\" (UniqueName: \"kubernetes.io/configmap/6402fda4-df10-493c-b4e5-d0569419652d-images\") on node \"crc\" DevicePath \"\"" Dec 13 06:49:23 crc kubenswrapper[4971]: I1213 06:49:23.892096 4971 reconciler_common.go:293] "Volume detached for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/6ea678ab-3438-413e-bfe3-290ae7725660-ovnkube-config\") on node \"crc\" DevicePath \"\"" Dec 13 06:49:23 crc kubenswrapper[4971]: I1213 06:49:23.892104 4971 reconciler_common.go:293] "Volume detached for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/1386a44e-36a2-460c-96d0-0359d2b6f0f5-kube-api-access\") on node \"crc\" DevicePath \"\"" Dec 13 06:49:23 crc kubenswrapper[4971]: I1213 06:49:23.892111 4971 reconciler_common.go:293] "Volume detached for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/5441d097-087c-4d9a-baa8-b210afa90fc9-client-ca\") on node \"crc\" DevicePath \"\"" Dec 13 06:49:23 crc kubenswrapper[4971]: I1213 06:49:23.892119 4971 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-s4n52\" (UniqueName: \"kubernetes.io/projected/925f1c65-6136-48ba-85aa-3a3b50560753-kube-api-access-s4n52\") on node \"crc\" DevicePath \"\"" Dec 13 06:49:23 crc kubenswrapper[4971]: I1213 06:49:23.892127 4971 reconciler_common.go:293] "Volume detached for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/c03ee662-fb2f-4fc4-a2c1-af487c19d254-metrics-certs\") on node \"crc\" DevicePath \"\"" Dec 13 06:49:23 crc kubenswrapper[4971]: I1213 06:49:23.892135 4971 reconciler_common.go:293] "Volume detached for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/0b78653f-4ff9-4508-8672-245ed9b561e3-kube-api-access\") on node \"crc\" DevicePath \"\"" Dec 13 06:49:23 crc kubenswrapper[4971]: I1213 06:49:23.892145 4971 reconciler_common.go:293] "Volume detached for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/e7e6199b-1264-4501-8953-767f51328d08-kube-api-access\") on node \"crc\" DevicePath \"\"" Dec 13 06:49:23 crc kubenswrapper[4971]: I1213 06:49:23.892154 4971 reconciler_common.go:293] "Volume detached for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/9d4552c7-cd75-42dd-8880-30dd377c49a4-trusted-ca\") on node \"crc\" DevicePath \"\"" Dec 13 06:49:23 crc kubenswrapper[4971]: I1213 06:49:23.892162 4971 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-xcphl\" (UniqueName: \"kubernetes.io/projected/7583ce53-e0fe-4a16-9e4d-50516596a136-kube-api-access-xcphl\") on node \"crc\" DevicePath \"\"" Dec 13 06:49:23 crc kubenswrapper[4971]: I1213 06:49:23.892171 4971 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/09efc573-dbb6-4249-bd59-9b87aba8dd28-config\") on node \"crc\" DevicePath \"\"" Dec 13 06:49:23 crc kubenswrapper[4971]: I1213 06:49:23.892180 4971 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-system-serving-cert\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-serving-cert\") on node \"crc\" DevicePath \"\"" Dec 13 06:49:23 crc kubenswrapper[4971]: I1213 06:49:23.892190 4971 reconciler_common.go:293] "Volume detached for volume \"profile-collector-cert\" (UniqueName: \"kubernetes.io/secret/b6312bbd-5731-4ea0-a20f-81d5a57df44a-profile-collector-cert\") on node \"crc\" DevicePath \"\"" Dec 13 06:49:23 crc kubenswrapper[4971]: I1213 06:49:23.892199 4971 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-system-trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-trusted-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 13 06:49:23 crc kubenswrapper[4971]: I1213 06:49:23.892208 4971 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-6g6sz\" (UniqueName: \"kubernetes.io/projected/6509e943-70c6-444c-bc41-48a544e36fbd-kube-api-access-6g6sz\") on node \"crc\" DevicePath \"\"" Dec 13 06:49:23 crc kubenswrapper[4971]: I1213 06:49:23.892217 4971 reconciler_common.go:293] "Volume detached for volume \"ovnkube-script-lib\" (UniqueName: \"kubernetes.io/configmap/6ea678ab-3438-413e-bfe3-290ae7725660-ovnkube-script-lib\") on node \"crc\" DevicePath \"\"" Dec 13 06:49:23 crc kubenswrapper[4971]: I1213 06:49:23.892225 4971 reconciler_common.go:293] "Volume detached for volume \"encryption-config\" (UniqueName: \"kubernetes.io/secret/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-encryption-config\") on node \"crc\" DevicePath \"\"" Dec 13 06:49:23 crc kubenswrapper[4971]: I1213 06:49:23.892233 4971 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-bf2bz\" (UniqueName: \"kubernetes.io/projected/1d611f23-29be-4491-8495-bee1670e935f-kube-api-access-bf2bz\") on node \"crc\" DevicePath \"\"" Dec 13 06:49:23 crc kubenswrapper[4971]: I1213 06:49:23.892251 4971 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-6ccd8\" (UniqueName: \"kubernetes.io/projected/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-kube-api-access-6ccd8\") on node \"crc\" DevicePath \"\"" Dec 13 06:49:23 crc kubenswrapper[4971]: I1213 06:49:23.892259 4971 reconciler_common.go:293] "Volume detached for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/31d8b7a1-420e-4252-a5b7-eebe8a111292-proxy-tls\") on node \"crc\" DevicePath \"\"" Dec 13 06:49:23 crc kubenswrapper[4971]: I1213 06:49:23.892266 4971 reconciler_common.go:293] "Volume detached for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/1bf7eb37-55a3-4c65-b768-a94c82151e69-etcd-client\") on node \"crc\" DevicePath \"\"" Dec 13 06:49:23 crc kubenswrapper[4971]: I1213 06:49:23.892274 4971 reconciler_common.go:293] "Volume detached for volume \"console-oauth-config\" (UniqueName: \"kubernetes.io/secret/43509403-f426-496e-be36-56cef71462f5-console-oauth-config\") on node \"crc\" DevicePath \"\"" Dec 13 06:49:23 crc kubenswrapper[4971]: I1213 06:49:23.892282 4971 reconciler_common.go:293] "Volume detached for volume \"images\" (UniqueName: \"kubernetes.io/configmap/31d8b7a1-420e-4252-a5b7-eebe8a111292-images\") on node \"crc\" DevicePath \"\"" Dec 13 06:49:23 crc kubenswrapper[4971]: I1213 06:49:23.892289 4971 reconciler_common.go:293] "Volume detached for volume \"auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/22c825df-677d-4ca6-82db-3454ed06e783-auth-proxy-config\") on node \"crc\" DevicePath \"\"" Dec 13 06:49:23 crc kubenswrapper[4971]: I1213 06:49:23.892298 4971 reconciler_common.go:293] "Volume detached for volume \"profile-collector-cert\" (UniqueName: \"kubernetes.io/secret/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9-profile-collector-cert\") on node \"crc\" DevicePath \"\"" Dec 13 06:49:23 crc kubenswrapper[4971]: I1213 06:49:23.892305 4971 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-zgdk5\" (UniqueName: \"kubernetes.io/projected/31d8b7a1-420e-4252-a5b7-eebe8a111292-kube-api-access-zgdk5\") on node \"crc\" DevicePath \"\"" Dec 13 06:49:23 crc kubenswrapper[4971]: I1213 06:49:23.892313 4971 reconciler_common.go:293] "Volume detached for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/6ea678ab-3438-413e-bfe3-290ae7725660-env-overrides\") on node \"crc\" DevicePath \"\"" Dec 13 06:49:23 crc kubenswrapper[4971]: I1213 06:49:23.892322 4971 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/5441d097-087c-4d9a-baa8-b210afa90fc9-config\") on node \"crc\" DevicePath \"\"" Dec 13 06:49:23 crc kubenswrapper[4971]: I1213 06:49:23.892329 4971 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-htfz6\" (UniqueName: \"kubernetes.io/projected/6ea678ab-3438-413e-bfe3-290ae7725660-kube-api-access-htfz6\") on node \"crc\" DevicePath \"\"" Dec 13 06:49:23 crc kubenswrapper[4971]: I1213 06:49:23.892339 4971 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-system-ocp-branding-template\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-ocp-branding-template\") on node \"crc\" DevicePath \"\"" Dec 13 06:49:23 crc kubenswrapper[4971]: I1213 06:49:23.892348 4971 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/22c825df-677d-4ca6-82db-3454ed06e783-config\") on node \"crc\" DevicePath \"\"" Dec 13 06:49:23 crc kubenswrapper[4971]: I1213 06:49:23.892356 4971 reconciler_common.go:293] "Volume detached for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/bf126b07-da06-4140-9a57-dfd54fc6b486-bound-sa-token\") on node \"crc\" DevicePath \"\"" Dec 13 06:49:23 crc kubenswrapper[4971]: I1213 06:49:23.892364 4971 reconciler_common.go:293] "Volume detached for volume \"etcd-ca\" (UniqueName: \"kubernetes.io/configmap/09efc573-dbb6-4249-bd59-9b87aba8dd28-etcd-ca\") on node \"crc\" DevicePath \"\"" Dec 13 06:49:23 crc kubenswrapper[4971]: I1213 06:49:23.892372 4971 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-x7zkh\" (UniqueName: \"kubernetes.io/projected/6731426b-95fe-49ff-bb5f-40441049fde2-kube-api-access-x7zkh\") on node \"crc\" DevicePath \"\"" Dec 13 06:49:23 crc kubenswrapper[4971]: I1213 06:49:23.892381 4971 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-system-cliconfig\" (UniqueName: \"kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-cliconfig\") on node \"crc\" DevicePath \"\"" Dec 13 06:49:23 crc kubenswrapper[4971]: I1213 06:49:23.892390 4971 reconciler_common.go:293] "Volume detached for volume \"signing-cabundle\" (UniqueName: \"kubernetes.io/configmap/25e176fe-21b4-4974-b1ed-c8b94f112a7f-signing-cabundle\") on node \"crc\" DevicePath \"\"" Dec 13 06:49:23 crc kubenswrapper[4971]: I1213 06:49:23.892401 4971 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-2d4wz\" (UniqueName: \"kubernetes.io/projected/5441d097-087c-4d9a-baa8-b210afa90fc9-kube-api-access-2d4wz\") on node \"crc\" DevicePath \"\"" Dec 13 06:49:23 crc kubenswrapper[4971]: I1213 06:49:23.892733 4971 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-13T06:49:23Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [webhook approver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-13T06:49:23Z\\\",\\\"message\\\":\\\"containers with unready status: [webhook approver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":false,\\\"restartCount\\\":6,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Dec 13 06:49:23 crc kubenswrapper[4971]: I1213 06:49:23.893090 4971 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-etcd/etcd-crc" Dec 13 06:49:23 crc kubenswrapper[4971]: I1213 06:49:23.893581 4971 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/5225d0e4-402f-4861-b410-819f433b1803-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "5225d0e4-402f-4861-b410-819f433b1803" (UID: "5225d0e4-402f-4861-b410-819f433b1803"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 13 06:49:23 crc kubenswrapper[4971]: I1213 06:49:23.904132 4971 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-13T06:49:23Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-13T06:49:23Z\\\",\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Dec 13 06:49:23 crc kubenswrapper[4971]: I1213 06:49:23.904952 4971 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-kube-apiserver_kube-apiserver-crc_f4b27818a5e8e43d0dc095d08835c792/kube-apiserver-check-endpoints/0.log" Dec 13 06:49:23 crc kubenswrapper[4971]: I1213 06:49:23.910055 4971 generic.go:334] "Generic (PLEG): container finished" podID="f4b27818a5e8e43d0dc095d08835c792" containerID="b3c1bab3d53f975b2e24582ce2771bc26d600e0dd7c30339dcfd18a071162ad4" exitCode=255 Dec 13 06:49:23 crc kubenswrapper[4971]: I1213 06:49:23.910670 4971 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"f4b27818a5e8e43d0dc095d08835c792","Type":"ContainerDied","Data":"b3c1bab3d53f975b2e24582ce2771bc26d600e0dd7c30339dcfd18a071162ad4"} Dec 13 06:49:23 crc kubenswrapper[4971]: I1213 06:49:23.912277 4971 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-etcd/etcd-crc" Dec 13 06:49:23 crc kubenswrapper[4971]: I1213 06:49:23.915059 4971 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-etcd/etcd-crc"] Dec 13 06:49:23 crc kubenswrapper[4971]: E1213 06:49:23.915792 4971 kubelet.go:1929] "Failed creating a mirror pod for" err="pods \"kube-controller-manager-crc\" already exists" pod="openshift-kube-controller-manager/kube-controller-manager-crc" Dec 13 06:49:23 crc kubenswrapper[4971]: I1213 06:49:23.919388 4971 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-13T06:49:23Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-13T06:49:23Z\\\",\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Dec 13 06:49:23 crc kubenswrapper[4971]: I1213 06:49:23.930802 4971 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-13T06:49:23Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-operator]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-13T06:49:23Z\\\",\\\"message\\\":\\\"containers with unready status: [network-operator]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":false,\\\"restartCount\\\":5,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Dec 13 06:49:23 crc kubenswrapper[4971]: I1213 06:49:23.940223 4971 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-13T06:49:23Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-13T06:49:23Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Dec 13 06:49:23 crc kubenswrapper[4971]: I1213 06:49:23.954508 4971 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-13T06:49:23Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [webhook approver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-13T06:49:23Z\\\",\\\"message\\\":\\\"containers with unready status: [webhook approver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":false,\\\"restartCount\\\":6,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Dec 13 06:49:23 crc kubenswrapper[4971]: I1213 06:49:23.966894 4971 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"f9d31b5a-b770-4a15-95c8-35b49677c608\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T06:49:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T06:49:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T06:49:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T06:49:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T06:49:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c9baa042f9dada954faa57b5454cd6b0a12a297e56ddc192c52cff50e42a16fa\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-13T06:49:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://c788de86754f0bf37ee184672fc1dcb388ccebc246d4f15be4b5da661f725f8f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-13T06:49:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://760817511a9b4f96a68dc38d26631250eaab72e1dd8f123e653039b6b4e91f84\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-13T06:49:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b859be0c9feac856e9855c6a038b02b49b32fade97873c46d0bec0a32d8b5147\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-13T06:49:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-13T06:49:03Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Dec 13 06:49:23 crc kubenswrapper[4971]: I1213 06:49:23.976449 4971 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-13T06:49:23Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-13T06:49:23Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Dec 13 06:49:23 crc kubenswrapper[4971]: I1213 06:49:23.985747 4971 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-13T06:49:23Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-13T06:49:23Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Dec 13 06:49:23 crc kubenswrapper[4971]: I1213 06:49:23.993460 4971 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" Dec 13 06:49:23 crc kubenswrapper[4971]: I1213 06:49:23.993610 4971 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/5225d0e4-402f-4861-b410-819f433b1803-catalog-content\") on node \"crc\" DevicePath \"\"" Dec 13 06:49:23 crc kubenswrapper[4971]: I1213 06:49:23.995971 4971 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-13T06:49:23Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-13T06:49:23Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Dec 13 06:49:23 crc kubenswrapper[4971]: I1213 06:49:23.999195 4971 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-node-identity/network-node-identity-vrzqb" Dec 13 06:49:24 crc kubenswrapper[4971]: I1213 06:49:24.005218 4971 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-operator/iptables-alerter-4ln5h" Dec 13 06:49:24 crc kubenswrapper[4971]: W1213 06:49:24.007074 4971 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod37a5e44f_9a88_4405_be8a_b645485e7312.slice/crio-b289fdf8d0d9704e826e4663eb01acfbb16fc7c00c3514f85d78974fe1d130c5 WatchSource:0}: Error finding container b289fdf8d0d9704e826e4663eb01acfbb16fc7c00c3514f85d78974fe1d130c5: Status 404 returned error can't find the container with id b289fdf8d0d9704e826e4663eb01acfbb16fc7c00c3514f85d78974fe1d130c5 Dec 13 06:49:24 crc kubenswrapper[4971]: I1213 06:49:24.007644 4971 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-kube-apiserver/kube-apiserver-crc"] Dec 13 06:49:24 crc kubenswrapper[4971]: I1213 06:49:24.008116 4971 scope.go:117] "RemoveContainer" containerID="b3c1bab3d53f975b2e24582ce2771bc26d600e0dd7c30339dcfd18a071162ad4" Dec 13 06:49:24 crc kubenswrapper[4971]: I1213 06:49:24.009602 4971 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-13T06:49:23Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-13T06:49:23Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Dec 13 06:49:24 crc kubenswrapper[4971]: I1213 06:49:24.026177 4971 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"f9d31b5a-b770-4a15-95c8-35b49677c608\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T06:49:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T06:49:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T06:49:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T06:49:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T06:49:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c9baa042f9dada954faa57b5454cd6b0a12a297e56ddc192c52cff50e42a16fa\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-13T06:49:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://c788de86754f0bf37ee184672fc1dcb388ccebc246d4f15be4b5da661f725f8f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-13T06:49:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://760817511a9b4f96a68dc38d26631250eaab72e1dd8f123e653039b6b4e91f84\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-13T06:49:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b859be0c9feac856e9855c6a038b02b49b32fade97873c46d0bec0a32d8b5147\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-13T06:49:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-13T06:49:03Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Dec 13 06:49:24 crc kubenswrapper[4971]: I1213 06:49:24.050320 4971 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"1ed9f82e-fe5c-4b76-a619-62e5abd88d3b\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T06:49:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T06:49:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T06:49:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T06:49:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T06:49:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e457329a78a2d24337f994e8a0a1f75a2a667244aa75eb84613d1e7138364aac\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-13T06:49:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://eb014de8140deef097d9bd94f3b9fa4ab3558bc4449bef84b83bf62bfabf5836\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-13T06:49:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://2c438cc74d555f7c9db4569d1c56b78b1088ed8105697d2c0f31e6c5136f3dd0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-13T06:49:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://aa6fa6e4340dcd066252df3c56d79a3a39ed531d5eea8d1d5941d6952368d3af\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-13T06:49:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://c4471b98854c52d5cd084d0ba3e64e5b08e9518ec3c372cce2c875a440522457\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-13T06:49:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7d2746335c98dae3a0f43024d34cb62ebcb42a01c8f9ce18d52b29d4c45598e5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://7d2746335c98dae3a0f43024d34cb62ebcb42a01c8f9ce18d52b29d4c45598e5\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-13T06:49:04Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-13T06:49:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://bc0e77d10a738128bea82122496bc42235fc1688f8907756d93e61b651902b57\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://bc0e77d10a738128bea82122496bc42235fc1688f8907756d93e61b651902b57\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-13T06:49:05Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-13T06:49:05Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://f821e418cba5e832da8839268bcbbd9630bdf319a212d5382bfa2c4b0b73a268\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://f821e418cba5e832da8839268bcbbd9630bdf319a212d5382bfa2c4b0b73a268\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-13T06:49:07Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-13T06:49:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-13T06:49:03Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Dec 13 06:49:24 crc kubenswrapper[4971]: I1213 06:49:24.062832 4971 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-13T06:49:23Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-operator]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-13T06:49:23Z\\\",\\\"message\\\":\\\"containers with unready status: [network-operator]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":false,\\\"restartCount\\\":5,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Dec 13 06:49:24 crc kubenswrapper[4971]: I1213 06:49:24.072978 4971 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-13T06:49:23Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-13T06:49:23Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Dec 13 06:49:24 crc kubenswrapper[4971]: I1213 06:49:24.082058 4971 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-13T06:49:23Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [webhook approver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-13T06:49:23Z\\\",\\\"message\\\":\\\"containers with unready status: [webhook approver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":false,\\\"restartCount\\\":6,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Dec 13 06:49:24 crc kubenswrapper[4971]: I1213 06:49:24.090958 4971 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-13T06:49:23Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-13T06:49:23Z\\\",\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Dec 13 06:49:24 crc kubenswrapper[4971]: I1213 06:49:24.155687 4971 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-kube-apiserver/kube-apiserver-crc" Dec 13 06:49:24 crc kubenswrapper[4971]: I1213 06:49:24.169880 4971 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3ad3268a-a982-428c-9c06-4aaeddd372d9\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T06:49:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T06:49:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T06:49:03Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T06:49:03Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T06:49:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3716afdf2bc693be80a09fc9f7727d334f435caf1d4c8ec4811b3a1f1652cdfe\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-13T06:49:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://30df3abf8c79db8af52c6a918f73e1ce410ddafcf341636424c267e1eeb195a7\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-13T06:49:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://ccdf65f6d2b1eb3647a03ba37c305c32e34addfdf79ff937c8dddaa3e5d5b5c6\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-13T06:49:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b3c1bab3d53f975b2e24582ce2771bc26d600e0dd7c30339dcfd18a071162ad4\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://b3c1bab3d53f975b2e24582ce2771bc26d600e0dd7c30339dcfd18a071162ad4\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-12-13T06:49:23Z\\\",\\\"message\\\":\\\"file observer\\\\nW1213 06:49:23.283403 1 builder.go:272] unable to get owner reference (falling back to namespace): pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\nI1213 06:49:23.284336 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1213 06:49:23.286978 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-981773671/tls.crt::/tmp/serving-cert-981773671/tls.key\\\\\\\"\\\\nI1213 06:49:23.775072 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1213 06:49:23.778615 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1213 06:49:23.778645 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1213 06:49:23.778680 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1213 06:49:23.778691 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1213 06:49:23.785058 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1213 06:49:23.786150 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1213 06:49:23.786196 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1213 06:49:23.786223 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1213 06:49:23.786246 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1213 06:49:23.786269 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1213 06:49:23.786291 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI1213 06:49:23.785608 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF1213 06:49:23.787563 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-13T06:49:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://9acd07066deed483bd7fa61ee45fda09539d110717a134a56827f871186f92fc\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-13T06:49:07Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://38c91cfbaecfdbd3e3b663422a91ab4294610afb6bbf9ebdb77db24e2b76c682\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://38c91cfbaecfdbd3e3b663422a91ab4294610afb6bbf9ebdb77db24e2b76c682\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-13T06:49:05Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-13T06:49:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-13T06:49:03Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Dec 13 06:49:24 crc kubenswrapper[4971]: I1213 06:49:24.182505 4971 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-13T06:49:23Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-13T06:49:23Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Dec 13 06:49:24 crc kubenswrapper[4971]: I1213 06:49:24.194426 4971 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-13T06:49:23Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-13T06:49:23Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Dec 13 06:49:24 crc kubenswrapper[4971]: I1213 06:49:24.204686 4971 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"f9d31b5a-b770-4a15-95c8-35b49677c608\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T06:49:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T06:49:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T06:49:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T06:49:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T06:49:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c9baa042f9dada954faa57b5454cd6b0a12a297e56ddc192c52cff50e42a16fa\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-13T06:49:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://c788de86754f0bf37ee184672fc1dcb388ccebc246d4f15be4b5da661f725f8f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-13T06:49:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://760817511a9b4f96a68dc38d26631250eaab72e1dd8f123e653039b6b4e91f84\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-13T06:49:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b859be0c9feac856e9855c6a038b02b49b32fade97873c46d0bec0a32d8b5147\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-13T06:49:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-13T06:49:03Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Dec 13 06:49:24 crc kubenswrapper[4971]: I1213 06:49:24.215152 4971 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-13T06:49:23Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-operator]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-13T06:49:23Z\\\",\\\"message\\\":\\\"containers with unready status: [network-operator]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":false,\\\"restartCount\\\":5,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Dec 13 06:49:24 crc kubenswrapper[4971]: I1213 06:49:24.228996 4971 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-13T06:49:23Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-13T06:49:23Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Dec 13 06:49:24 crc kubenswrapper[4971]: I1213 06:49:24.238478 4971 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-13T06:49:23Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [webhook approver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-13T06:49:23Z\\\",\\\"message\\\":\\\"containers with unready status: [webhook approver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":false,\\\"restartCount\\\":6,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Dec 13 06:49:24 crc kubenswrapper[4971]: I1213 06:49:24.248622 4971 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-13T06:49:23Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-13T06:49:23Z\\\",\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Dec 13 06:49:24 crc kubenswrapper[4971]: I1213 06:49:24.265367 4971 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"1ed9f82e-fe5c-4b76-a619-62e5abd88d3b\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T06:49:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T06:49:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T06:49:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T06:49:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T06:49:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e457329a78a2d24337f994e8a0a1f75a2a667244aa75eb84613d1e7138364aac\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-13T06:49:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://eb014de8140deef097d9bd94f3b9fa4ab3558bc4449bef84b83bf62bfabf5836\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-13T06:49:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://2c438cc74d555f7c9db4569d1c56b78b1088ed8105697d2c0f31e6c5136f3dd0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-13T06:49:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://aa6fa6e4340dcd066252df3c56d79a3a39ed531d5eea8d1d5941d6952368d3af\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-13T06:49:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://c4471b98854c52d5cd084d0ba3e64e5b08e9518ec3c372cce2c875a440522457\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-13T06:49:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7d2746335c98dae3a0f43024d34cb62ebcb42a01c8f9ce18d52b29d4c45598e5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://7d2746335c98dae3a0f43024d34cb62ebcb42a01c8f9ce18d52b29d4c45598e5\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-13T06:49:04Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-13T06:49:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://bc0e77d10a738128bea82122496bc42235fc1688f8907756d93e61b651902b57\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://bc0e77d10a738128bea82122496bc42235fc1688f8907756d93e61b651902b57\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-13T06:49:05Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-13T06:49:05Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://f821e418cba5e832da8839268bcbbd9630bdf319a212d5382bfa2c4b0b73a268\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://f821e418cba5e832da8839268bcbbd9630bdf319a212d5382bfa2c4b0b73a268\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-13T06:49:07Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-13T06:49:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-13T06:49:03Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Dec 13 06:49:24 crc kubenswrapper[4971]: I1213 06:49:24.295906 4971 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 13 06:49:24 crc kubenswrapper[4971]: I1213 06:49:24.295991 4971 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"networking-console-plugin-cert\" (UniqueName: \"kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 13 06:49:24 crc kubenswrapper[4971]: I1213 06:49:24.296013 4971 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nginx-conf\" (UniqueName: \"kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 13 06:49:24 crc kubenswrapper[4971]: E1213 06:49:24.296077 4971 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-13 06:49:25.296060887 +0000 UTC m=+21.900470335 (durationBeforeRetry 1s). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 13 06:49:24 crc kubenswrapper[4971]: E1213 06:49:24.296134 4971 secret.go:188] Couldn't get secret openshift-network-console/networking-console-plugin-cert: object "openshift-network-console"/"networking-console-plugin-cert" not registered Dec 13 06:49:24 crc kubenswrapper[4971]: E1213 06:49:24.296178 4971 configmap.go:193] Couldn't get configMap openshift-network-console/networking-console-plugin: object "openshift-network-console"/"networking-console-plugin" not registered Dec 13 06:49:24 crc kubenswrapper[4971]: E1213 06:49:24.296209 4971 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2025-12-13 06:49:25.296192431 +0000 UTC m=+21.900601879 (durationBeforeRetry 1s). Error: MountVolume.SetUp failed for volume "networking-console-plugin-cert" (UniqueName: "kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin-cert" not registered Dec 13 06:49:24 crc kubenswrapper[4971]: E1213 06:49:24.296254 4971 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2025-12-13 06:49:25.296236772 +0000 UTC m=+21.900646220 (durationBeforeRetry 1s). Error: MountVolume.SetUp failed for volume "nginx-conf" (UniqueName: "kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin" not registered Dec 13 06:49:24 crc kubenswrapper[4971]: I1213 06:49:24.396848 4971 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cqllr\" (UniqueName: \"kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr\") pod \"network-check-target-xd92c\" (UID: \"3b6479f0-333b-4a96-9adf-2099afdc2447\") " pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 13 06:49:24 crc kubenswrapper[4971]: I1213 06:49:24.396946 4971 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-s2dwl\" (UniqueName: \"kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl\") pod \"network-check-source-55646444c4-trplf\" (UID: \"9d751cbb-f2e2-430d-9754-c882a5e924a5\") " pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 13 06:49:24 crc kubenswrapper[4971]: E1213 06:49:24.397079 4971 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Dec 13 06:49:24 crc kubenswrapper[4971]: E1213 06:49:24.397132 4971 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Dec 13 06:49:24 crc kubenswrapper[4971]: E1213 06:49:24.397149 4971 projected.go:194] Error preparing data for projected volume kube-api-access-cqllr for pod openshift-network-diagnostics/network-check-target-xd92c: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Dec 13 06:49:24 crc kubenswrapper[4971]: E1213 06:49:24.397079 4971 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Dec 13 06:49:24 crc kubenswrapper[4971]: E1213 06:49:24.397230 4971 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Dec 13 06:49:24 crc kubenswrapper[4971]: E1213 06:49:24.397244 4971 projected.go:194] Error preparing data for projected volume kube-api-access-s2dwl for pod openshift-network-diagnostics/network-check-source-55646444c4-trplf: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Dec 13 06:49:24 crc kubenswrapper[4971]: E1213 06:49:24.397235 4971 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr podName:3b6479f0-333b-4a96-9adf-2099afdc2447 nodeName:}" failed. No retries permitted until 2025-12-13 06:49:25.397211884 +0000 UTC m=+22.001621342 (durationBeforeRetry 1s). Error: MountVolume.SetUp failed for volume "kube-api-access-cqllr" (UniqueName: "kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr") pod "network-check-target-xd92c" (UID: "3b6479f0-333b-4a96-9adf-2099afdc2447") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Dec 13 06:49:24 crc kubenswrapper[4971]: E1213 06:49:24.397318 4971 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl podName:9d751cbb-f2e2-430d-9754-c882a5e924a5 nodeName:}" failed. No retries permitted until 2025-12-13 06:49:25.397306876 +0000 UTC m=+22.001716324 (durationBeforeRetry 1s). Error: MountVolume.SetUp failed for volume "kube-api-access-s2dwl" (UniqueName: "kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl") pod "network-check-source-55646444c4-trplf" (UID: "9d751cbb-f2e2-430d-9754-c882a5e924a5") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Dec 13 06:49:24 crc kubenswrapper[4971]: I1213 06:49:24.915188 4971 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-kube-apiserver_kube-apiserver-crc_f4b27818a5e8e43d0dc095d08835c792/kube-apiserver-check-endpoints/0.log" Dec 13 06:49:24 crc kubenswrapper[4971]: I1213 06:49:24.917030 4971 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"f4b27818a5e8e43d0dc095d08835c792","Type":"ContainerStarted","Data":"770b73dd0c2b6f38744e2c1cd6aa186a69bb87d1cd015a603ad2e33be06a8f6b"} Dec 13 06:49:24 crc kubenswrapper[4971]: I1213 06:49:24.917294 4971 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-kube-apiserver/kube-apiserver-crc" Dec 13 06:49:24 crc kubenswrapper[4971]: I1213 06:49:24.918066 4971 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" event={"ID":"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49","Type":"ContainerStarted","Data":"be6440017966132c3a986ef8ae61032ed932e30f542d7440a9a01c75e946dd07"} Dec 13 06:49:24 crc kubenswrapper[4971]: I1213 06:49:24.920838 4971 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" event={"ID":"ef543e1b-8068-4ea3-b32a-61027b32e95d","Type":"ContainerStarted","Data":"4328cd6a8464ed468229bbfbc8265f969427916d72ebb92237dbe68efe70fa47"} Dec 13 06:49:24 crc kubenswrapper[4971]: I1213 06:49:24.921015 4971 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" event={"ID":"ef543e1b-8068-4ea3-b32a-61027b32e95d","Type":"ContainerStarted","Data":"9d715ca82d9b5a594e41f2b98da6f4277da0e4e1dec114969f901c00f6d7553d"} Dec 13 06:49:24 crc kubenswrapper[4971]: I1213 06:49:24.921132 4971 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" event={"ID":"ef543e1b-8068-4ea3-b32a-61027b32e95d","Type":"ContainerStarted","Data":"3a5bd2b1e276fdd115d9753076418bc1b059eb02aa3948e57fb6a3c8b9c40277"} Dec 13 06:49:24 crc kubenswrapper[4971]: I1213 06:49:24.921399 4971 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-kube-apiserver/kube-apiserver-crc" Dec 13 06:49:24 crc kubenswrapper[4971]: I1213 06:49:24.922236 4971 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" event={"ID":"37a5e44f-9a88-4405-be8a-b645485e7312","Type":"ContainerStarted","Data":"d0b2ecf0d5b22f23217f42f259ae7ffb6359555b22c30eaebf667eaf254b3bab"} Dec 13 06:49:24 crc kubenswrapper[4971]: I1213 06:49:24.922284 4971 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" event={"ID":"37a5e44f-9a88-4405-be8a-b645485e7312","Type":"ContainerStarted","Data":"b289fdf8d0d9704e826e4663eb01acfbb16fc7c00c3514f85d78974fe1d130c5"} Dec 13 06:49:24 crc kubenswrapper[4971]: I1213 06:49:24.931669 4971 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-13T06:49:23Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-13T06:49:23Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-13T06:49:24Z is after 2025-08-24T17:21:41Z" Dec 13 06:49:24 crc kubenswrapper[4971]: I1213 06:49:24.947108 4971 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"f9d31b5a-b770-4a15-95c8-35b49677c608\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T06:49:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T06:49:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T06:49:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T06:49:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T06:49:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c9baa042f9dada954faa57b5454cd6b0a12a297e56ddc192c52cff50e42a16fa\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-13T06:49:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://c788de86754f0bf37ee184672fc1dcb388ccebc246d4f15be4b5da661f725f8f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-13T06:49:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://760817511a9b4f96a68dc38d26631250eaab72e1dd8f123e653039b6b4e91f84\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-13T06:49:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b859be0c9feac856e9855c6a038b02b49b32fade97873c46d0bec0a32d8b5147\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-13T06:49:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-13T06:49:03Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-13T06:49:24Z is after 2025-08-24T17:21:41Z" Dec 13 06:49:24 crc kubenswrapper[4971]: I1213 06:49:24.966147 4971 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3ad3268a-a982-428c-9c06-4aaeddd372d9\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T06:49:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T06:49:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T06:49:03Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T06:49:03Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T06:49:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3716afdf2bc693be80a09fc9f7727d334f435caf1d4c8ec4811b3a1f1652cdfe\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-13T06:49:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://30df3abf8c79db8af52c6a918f73e1ce410ddafcf341636424c267e1eeb195a7\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-13T06:49:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://ccdf65f6d2b1eb3647a03ba37c305c32e34addfdf79ff937c8dddaa3e5d5b5c6\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-13T06:49:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://770b73dd0c2b6f38744e2c1cd6aa186a69bb87d1cd015a603ad2e33be06a8f6b\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://b3c1bab3d53f975b2e24582ce2771bc26d600e0dd7c30339dcfd18a071162ad4\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-12-13T06:49:23Z\\\",\\\"message\\\":\\\"file observer\\\\nW1213 06:49:23.283403 1 builder.go:272] unable to get owner reference (falling back to namespace): pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\nI1213 06:49:23.284336 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1213 06:49:23.286978 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-981773671/tls.crt::/tmp/serving-cert-981773671/tls.key\\\\\\\"\\\\nI1213 06:49:23.775072 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1213 06:49:23.778615 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1213 06:49:23.778645 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1213 06:49:23.778680 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1213 06:49:23.778691 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1213 06:49:23.785058 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1213 06:49:23.786150 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1213 06:49:23.786196 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1213 06:49:23.786223 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1213 06:49:23.786246 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1213 06:49:23.786269 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1213 06:49:23.786291 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI1213 06:49:23.785608 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF1213 06:49:23.787563 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-13T06:49:08Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-13T06:49:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://9acd07066deed483bd7fa61ee45fda09539d110717a134a56827f871186f92fc\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-13T06:49:07Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://38c91cfbaecfdbd3e3b663422a91ab4294610afb6bbf9ebdb77db24e2b76c682\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://38c91cfbaecfdbd3e3b663422a91ab4294610afb6bbf9ebdb77db24e2b76c682\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-13T06:49:05Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-13T06:49:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-13T06:49:03Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-13T06:49:24Z is after 2025-08-24T17:21:41Z" Dec 13 06:49:24 crc kubenswrapper[4971]: I1213 06:49:24.982679 4971 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-13T06:49:23Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-13T06:49:23Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-13T06:49:24Z is after 2025-08-24T17:21:41Z" Dec 13 06:49:25 crc kubenswrapper[4971]: I1213 06:49:25.010189 4971 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-13T06:49:23Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [webhook approver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-13T06:49:23Z\\\",\\\"message\\\":\\\"containers with unready status: [webhook approver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":false,\\\"restartCount\\\":6,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-13T06:49:25Z is after 2025-08-24T17:21:41Z" Dec 13 06:49:25 crc kubenswrapper[4971]: I1213 06:49:25.025477 4971 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-13T06:49:23Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-13T06:49:23Z\\\",\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-13T06:49:25Z is after 2025-08-24T17:21:41Z" Dec 13 06:49:25 crc kubenswrapper[4971]: I1213 06:49:25.049303 4971 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"1ed9f82e-fe5c-4b76-a619-62e5abd88d3b\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T06:49:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T06:49:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T06:49:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T06:49:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T06:49:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e457329a78a2d24337f994e8a0a1f75a2a667244aa75eb84613d1e7138364aac\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-13T06:49:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://eb014de8140deef097d9bd94f3b9fa4ab3558bc4449bef84b83bf62bfabf5836\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-13T06:49:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://2c438cc74d555f7c9db4569d1c56b78b1088ed8105697d2c0f31e6c5136f3dd0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-13T06:49:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://aa6fa6e4340dcd066252df3c56d79a3a39ed531d5eea8d1d5941d6952368d3af\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-13T06:49:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://c4471b98854c52d5cd084d0ba3e64e5b08e9518ec3c372cce2c875a440522457\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-13T06:49:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7d2746335c98dae3a0f43024d34cb62ebcb42a01c8f9ce18d52b29d4c45598e5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://7d2746335c98dae3a0f43024d34cb62ebcb42a01c8f9ce18d52b29d4c45598e5\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-13T06:49:04Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-13T06:49:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://bc0e77d10a738128bea82122496bc42235fc1688f8907756d93e61b651902b57\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://bc0e77d10a738128bea82122496bc42235fc1688f8907756d93e61b651902b57\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-13T06:49:05Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-13T06:49:05Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://f821e418cba5e832da8839268bcbbd9630bdf319a212d5382bfa2c4b0b73a268\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://f821e418cba5e832da8839268bcbbd9630bdf319a212d5382bfa2c4b0b73a268\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-13T06:49:07Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-13T06:49:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-13T06:49:03Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-13T06:49:25Z is after 2025-08-24T17:21:41Z" Dec 13 06:49:25 crc kubenswrapper[4971]: I1213 06:49:25.081271 4971 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-13T06:49:23Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-operator]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-13T06:49:23Z\\\",\\\"message\\\":\\\"containers with unready status: [network-operator]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":false,\\\"restartCount\\\":5,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-13T06:49:25Z is after 2025-08-24T17:21:41Z" Dec 13 06:49:25 crc kubenswrapper[4971]: I1213 06:49:25.093426 4971 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-13T06:49:23Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-13T06:49:23Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-13T06:49:25Z is after 2025-08-24T17:21:41Z" Dec 13 06:49:25 crc kubenswrapper[4971]: I1213 06:49:25.116056 4971 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3ad3268a-a982-428c-9c06-4aaeddd372d9\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T06:49:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T06:49:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T06:49:03Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T06:49:03Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T06:49:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3716afdf2bc693be80a09fc9f7727d334f435caf1d4c8ec4811b3a1f1652cdfe\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-13T06:49:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://30df3abf8c79db8af52c6a918f73e1ce410ddafcf341636424c267e1eeb195a7\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-13T06:49:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://ccdf65f6d2b1eb3647a03ba37c305c32e34addfdf79ff937c8dddaa3e5d5b5c6\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-13T06:49:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://770b73dd0c2b6f38744e2c1cd6aa186a69bb87d1cd015a603ad2e33be06a8f6b\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://b3c1bab3d53f975b2e24582ce2771bc26d600e0dd7c30339dcfd18a071162ad4\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-12-13T06:49:23Z\\\",\\\"message\\\":\\\"file observer\\\\nW1213 06:49:23.283403 1 builder.go:272] unable to get owner reference (falling back to namespace): pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\nI1213 06:49:23.284336 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1213 06:49:23.286978 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-981773671/tls.crt::/tmp/serving-cert-981773671/tls.key\\\\\\\"\\\\nI1213 06:49:23.775072 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1213 06:49:23.778615 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1213 06:49:23.778645 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1213 06:49:23.778680 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1213 06:49:23.778691 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1213 06:49:23.785058 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1213 06:49:23.786150 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1213 06:49:23.786196 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1213 06:49:23.786223 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1213 06:49:23.786246 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1213 06:49:23.786269 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1213 06:49:23.786291 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI1213 06:49:23.785608 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF1213 06:49:23.787563 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-13T06:49:08Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-13T06:49:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://9acd07066deed483bd7fa61ee45fda09539d110717a134a56827f871186f92fc\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-13T06:49:07Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://38c91cfbaecfdbd3e3b663422a91ab4294610afb6bbf9ebdb77db24e2b76c682\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://38c91cfbaecfdbd3e3b663422a91ab4294610afb6bbf9ebdb77db24e2b76c682\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-13T06:49:05Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-13T06:49:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-13T06:49:03Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-13T06:49:25Z is after 2025-08-24T17:21:41Z" Dec 13 06:49:25 crc kubenswrapper[4971]: I1213 06:49:25.139804 4971 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-13T06:49:23Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-13T06:49:23Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-13T06:49:25Z is after 2025-08-24T17:21:41Z" Dec 13 06:49:25 crc kubenswrapper[4971]: I1213 06:49:25.155906 4971 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-13T06:49:23Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-13T06:49:23Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-13T06:49:25Z is after 2025-08-24T17:21:41Z" Dec 13 06:49:25 crc kubenswrapper[4971]: I1213 06:49:25.175873 4971 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-dns/node-resolver-5rzh5"] Dec 13 06:49:25 crc kubenswrapper[4971]: I1213 06:49:25.176190 4971 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-dns/node-resolver-5rzh5" Dec 13 06:49:25 crc kubenswrapper[4971]: I1213 06:49:25.178250 4971 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-dns"/"openshift-service-ca.crt" Dec 13 06:49:25 crc kubenswrapper[4971]: I1213 06:49:25.184345 4971 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-dns"/"node-resolver-dockercfg-kz9s7" Dec 13 06:49:25 crc kubenswrapper[4971]: I1213 06:49:25.184536 4971 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-dns"/"kube-root-ca.crt" Dec 13 06:49:25 crc kubenswrapper[4971]: I1213 06:49:25.184739 4971 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"f9d31b5a-b770-4a15-95c8-35b49677c608\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T06:49:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T06:49:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T06:49:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T06:49:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T06:49:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c9baa042f9dada954faa57b5454cd6b0a12a297e56ddc192c52cff50e42a16fa\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-13T06:49:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://c788de86754f0bf37ee184672fc1dcb388ccebc246d4f15be4b5da661f725f8f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-13T06:49:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://760817511a9b4f96a68dc38d26631250eaab72e1dd8f123e653039b6b4e91f84\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-13T06:49:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b859be0c9feac856e9855c6a038b02b49b32fade97873c46d0bec0a32d8b5147\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-13T06:49:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-13T06:49:03Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-13T06:49:25Z is after 2025-08-24T17:21:41Z" Dec 13 06:49:25 crc kubenswrapper[4971]: I1213 06:49:25.202387 4971 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-dfxqn\" (UniqueName: \"kubernetes.io/projected/6bf4591d-2483-4132-9672-d0fd6c995968-kube-api-access-dfxqn\") pod \"node-resolver-5rzh5\" (UID: \"6bf4591d-2483-4132-9672-d0fd6c995968\") " pod="openshift-dns/node-resolver-5rzh5" Dec 13 06:49:25 crc kubenswrapper[4971]: I1213 06:49:25.202449 4971 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"hosts-file\" (UniqueName: \"kubernetes.io/host-path/6bf4591d-2483-4132-9672-d0fd6c995968-hosts-file\") pod \"node-resolver-5rzh5\" (UID: \"6bf4591d-2483-4132-9672-d0fd6c995968\") " pod="openshift-dns/node-resolver-5rzh5" Dec 13 06:49:25 crc kubenswrapper[4971]: I1213 06:49:25.224955 4971 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"1ed9f82e-fe5c-4b76-a619-62e5abd88d3b\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T06:49:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T06:49:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T06:49:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T06:49:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T06:49:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e457329a78a2d24337f994e8a0a1f75a2a667244aa75eb84613d1e7138364aac\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-13T06:49:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://eb014de8140deef097d9bd94f3b9fa4ab3558bc4449bef84b83bf62bfabf5836\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-13T06:49:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://2c438cc74d555f7c9db4569d1c56b78b1088ed8105697d2c0f31e6c5136f3dd0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-13T06:49:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://aa6fa6e4340dcd066252df3c56d79a3a39ed531d5eea8d1d5941d6952368d3af\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-13T06:49:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://c4471b98854c52d5cd084d0ba3e64e5b08e9518ec3c372cce2c875a440522457\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-13T06:49:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7d2746335c98dae3a0f43024d34cb62ebcb42a01c8f9ce18d52b29d4c45598e5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://7d2746335c98dae3a0f43024d34cb62ebcb42a01c8f9ce18d52b29d4c45598e5\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-13T06:49:04Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-13T06:49:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://bc0e77d10a738128bea82122496bc42235fc1688f8907756d93e61b651902b57\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://bc0e77d10a738128bea82122496bc42235fc1688f8907756d93e61b651902b57\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-13T06:49:05Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-13T06:49:05Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://f821e418cba5e832da8839268bcbbd9630bdf319a212d5382bfa2c4b0b73a268\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://f821e418cba5e832da8839268bcbbd9630bdf319a212d5382bfa2c4b0b73a268\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-13T06:49:07Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-13T06:49:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-13T06:49:03Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-13T06:49:25Z is after 2025-08-24T17:21:41Z" Dec 13 06:49:25 crc kubenswrapper[4971]: I1213 06:49:25.243798 4971 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-13T06:49:24Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-13T06:49:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-13T06:49:24Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d0b2ecf0d5b22f23217f42f259ae7ffb6359555b22c30eaebf667eaf254b3bab\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-13T06:49:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-13T06:49:25Z is after 2025-08-24T17:21:41Z" Dec 13 06:49:25 crc kubenswrapper[4971]: I1213 06:49:25.256853 4971 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-13T06:49:23Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-13T06:49:23Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-13T06:49:25Z is after 2025-08-24T17:21:41Z" Dec 13 06:49:25 crc kubenswrapper[4971]: I1213 06:49:25.268863 4971 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-13T06:49:24Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-13T06:49:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-13T06:49:24Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4328cd6a8464ed468229bbfbc8265f969427916d72ebb92237dbe68efe70fa47\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-13T06:49:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://9d715ca82d9b5a594e41f2b98da6f4277da0e4e1dec114969f901c00f6d7553d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-13T06:49:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-13T06:49:25Z is after 2025-08-24T17:21:41Z" Dec 13 06:49:25 crc kubenswrapper[4971]: I1213 06:49:25.280433 4971 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-13T06:49:23Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-13T06:49:23Z\\\",\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-13T06:49:25Z is after 2025-08-24T17:21:41Z" Dec 13 06:49:25 crc kubenswrapper[4971]: I1213 06:49:25.295964 4971 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3ad3268a-a982-428c-9c06-4aaeddd372d9\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T06:49:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T06:49:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T06:49:03Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T06:49:03Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T06:49:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3716afdf2bc693be80a09fc9f7727d334f435caf1d4c8ec4811b3a1f1652cdfe\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-13T06:49:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://30df3abf8c79db8af52c6a918f73e1ce410ddafcf341636424c267e1eeb195a7\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-13T06:49:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://ccdf65f6d2b1eb3647a03ba37c305c32e34addfdf79ff937c8dddaa3e5d5b5c6\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-13T06:49:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://770b73dd0c2b6f38744e2c1cd6aa186a69bb87d1cd015a603ad2e33be06a8f6b\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://b3c1bab3d53f975b2e24582ce2771bc26d600e0dd7c30339dcfd18a071162ad4\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-12-13T06:49:23Z\\\",\\\"message\\\":\\\"file observer\\\\nW1213 06:49:23.283403 1 builder.go:272] unable to get owner reference (falling back to namespace): pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\nI1213 06:49:23.284336 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1213 06:49:23.286978 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-981773671/tls.crt::/tmp/serving-cert-981773671/tls.key\\\\\\\"\\\\nI1213 06:49:23.775072 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1213 06:49:23.778615 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1213 06:49:23.778645 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1213 06:49:23.778680 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1213 06:49:23.778691 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1213 06:49:23.785058 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1213 06:49:23.786150 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1213 06:49:23.786196 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1213 06:49:23.786223 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1213 06:49:23.786246 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1213 06:49:23.786269 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1213 06:49:23.786291 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI1213 06:49:23.785608 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF1213 06:49:23.787563 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-13T06:49:08Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-13T06:49:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://9acd07066deed483bd7fa61ee45fda09539d110717a134a56827f871186f92fc\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-13T06:49:07Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://38c91cfbaecfdbd3e3b663422a91ab4294610afb6bbf9ebdb77db24e2b76c682\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://38c91cfbaecfdbd3e3b663422a91ab4294610afb6bbf9ebdb77db24e2b76c682\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-13T06:49:05Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-13T06:49:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-13T06:49:03Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-13T06:49:25Z is after 2025-08-24T17:21:41Z" Dec 13 06:49:25 crc kubenswrapper[4971]: I1213 06:49:25.303694 4971 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 13 06:49:25 crc kubenswrapper[4971]: I1213 06:49:25.303770 4971 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"networking-console-plugin-cert\" (UniqueName: \"kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 13 06:49:25 crc kubenswrapper[4971]: I1213 06:49:25.303794 4971 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nginx-conf\" (UniqueName: \"kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 13 06:49:25 crc kubenswrapper[4971]: I1213 06:49:25.303810 4971 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-dfxqn\" (UniqueName: \"kubernetes.io/projected/6bf4591d-2483-4132-9672-d0fd6c995968-kube-api-access-dfxqn\") pod \"node-resolver-5rzh5\" (UID: \"6bf4591d-2483-4132-9672-d0fd6c995968\") " pod="openshift-dns/node-resolver-5rzh5" Dec 13 06:49:25 crc kubenswrapper[4971]: I1213 06:49:25.303841 4971 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"hosts-file\" (UniqueName: \"kubernetes.io/host-path/6bf4591d-2483-4132-9672-d0fd6c995968-hosts-file\") pod \"node-resolver-5rzh5\" (UID: \"6bf4591d-2483-4132-9672-d0fd6c995968\") " pod="openshift-dns/node-resolver-5rzh5" Dec 13 06:49:25 crc kubenswrapper[4971]: E1213 06:49:25.303880 4971 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-13 06:49:27.303851941 +0000 UTC m=+23.908261389 (durationBeforeRetry 2s). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 13 06:49:25 crc kubenswrapper[4971]: I1213 06:49:25.303946 4971 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"hosts-file\" (UniqueName: \"kubernetes.io/host-path/6bf4591d-2483-4132-9672-d0fd6c995968-hosts-file\") pod \"node-resolver-5rzh5\" (UID: \"6bf4591d-2483-4132-9672-d0fd6c995968\") " pod="openshift-dns/node-resolver-5rzh5" Dec 13 06:49:25 crc kubenswrapper[4971]: E1213 06:49:25.303954 4971 configmap.go:193] Couldn't get configMap openshift-network-console/networking-console-plugin: object "openshift-network-console"/"networking-console-plugin" not registered Dec 13 06:49:25 crc kubenswrapper[4971]: E1213 06:49:25.304037 4971 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2025-12-13 06:49:27.304006765 +0000 UTC m=+23.908416203 (durationBeforeRetry 2s). Error: MountVolume.SetUp failed for volume "nginx-conf" (UniqueName: "kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin" not registered Dec 13 06:49:25 crc kubenswrapper[4971]: E1213 06:49:25.303950 4971 secret.go:188] Couldn't get secret openshift-network-console/networking-console-plugin-cert: object "openshift-network-console"/"networking-console-plugin-cert" not registered Dec 13 06:49:25 crc kubenswrapper[4971]: E1213 06:49:25.304115 4971 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2025-12-13 06:49:27.304097177 +0000 UTC m=+23.908506625 (durationBeforeRetry 2s). Error: MountVolume.SetUp failed for volume "networking-console-plugin-cert" (UniqueName: "kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin-cert" not registered Dec 13 06:49:25 crc kubenswrapper[4971]: I1213 06:49:25.308010 4971 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-13T06:49:23Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-13T06:49:23Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-13T06:49:25Z is after 2025-08-24T17:21:41Z" Dec 13 06:49:25 crc kubenswrapper[4971]: I1213 06:49:25.319320 4971 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-13T06:49:23Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-13T06:49:23Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-13T06:49:25Z is after 2025-08-24T17:21:41Z" Dec 13 06:49:25 crc kubenswrapper[4971]: I1213 06:49:25.325593 4971 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-dfxqn\" (UniqueName: \"kubernetes.io/projected/6bf4591d-2483-4132-9672-d0fd6c995968-kube-api-access-dfxqn\") pod \"node-resolver-5rzh5\" (UID: \"6bf4591d-2483-4132-9672-d0fd6c995968\") " pod="openshift-dns/node-resolver-5rzh5" Dec 13 06:49:25 crc kubenswrapper[4971]: I1213 06:49:25.339954 4971 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"f9d31b5a-b770-4a15-95c8-35b49677c608\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T06:49:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T06:49:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T06:49:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T06:49:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T06:49:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c9baa042f9dada954faa57b5454cd6b0a12a297e56ddc192c52cff50e42a16fa\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-13T06:49:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://c788de86754f0bf37ee184672fc1dcb388ccebc246d4f15be4b5da661f725f8f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-13T06:49:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://760817511a9b4f96a68dc38d26631250eaab72e1dd8f123e653039b6b4e91f84\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-13T06:49:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b859be0c9feac856e9855c6a038b02b49b32fade97873c46d0bec0a32d8b5147\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-13T06:49:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-13T06:49:03Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-13T06:49:25Z is after 2025-08-24T17:21:41Z" Dec 13 06:49:25 crc kubenswrapper[4971]: I1213 06:49:25.365060 4971 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"1ed9f82e-fe5c-4b76-a619-62e5abd88d3b\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T06:49:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T06:49:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T06:49:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T06:49:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T06:49:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e457329a78a2d24337f994e8a0a1f75a2a667244aa75eb84613d1e7138364aac\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-13T06:49:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://eb014de8140deef097d9bd94f3b9fa4ab3558bc4449bef84b83bf62bfabf5836\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-13T06:49:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://2c438cc74d555f7c9db4569d1c56b78b1088ed8105697d2c0f31e6c5136f3dd0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-13T06:49:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://aa6fa6e4340dcd066252df3c56d79a3a39ed531d5eea8d1d5941d6952368d3af\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-13T06:49:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://c4471b98854c52d5cd084d0ba3e64e5b08e9518ec3c372cce2c875a440522457\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-13T06:49:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7d2746335c98dae3a0f43024d34cb62ebcb42a01c8f9ce18d52b29d4c45598e5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://7d2746335c98dae3a0f43024d34cb62ebcb42a01c8f9ce18d52b29d4c45598e5\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-13T06:49:04Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-13T06:49:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://bc0e77d10a738128bea82122496bc42235fc1688f8907756d93e61b651902b57\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://bc0e77d10a738128bea82122496bc42235fc1688f8907756d93e61b651902b57\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-13T06:49:05Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-13T06:49:05Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://f821e418cba5e832da8839268bcbbd9630bdf319a212d5382bfa2c4b0b73a268\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://f821e418cba5e832da8839268bcbbd9630bdf319a212d5382bfa2c4b0b73a268\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-13T06:49:07Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-13T06:49:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-13T06:49:03Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-13T06:49:25Z is after 2025-08-24T17:21:41Z" Dec 13 06:49:25 crc kubenswrapper[4971]: I1213 06:49:25.379232 4971 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-13T06:49:24Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-13T06:49:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-13T06:49:24Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d0b2ecf0d5b22f23217f42f259ae7ffb6359555b22c30eaebf667eaf254b3bab\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-13T06:49:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-13T06:49:25Z is after 2025-08-24T17:21:41Z" Dec 13 06:49:25 crc kubenswrapper[4971]: I1213 06:49:25.394001 4971 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-13T06:49:23Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-13T06:49:23Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-13T06:49:25Z is after 2025-08-24T17:21:41Z" Dec 13 06:49:25 crc kubenswrapper[4971]: I1213 06:49:25.404599 4971 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cqllr\" (UniqueName: \"kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr\") pod \"network-check-target-xd92c\" (UID: \"3b6479f0-333b-4a96-9adf-2099afdc2447\") " pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 13 06:49:25 crc kubenswrapper[4971]: I1213 06:49:25.404670 4971 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-s2dwl\" (UniqueName: \"kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl\") pod \"network-check-source-55646444c4-trplf\" (UID: \"9d751cbb-f2e2-430d-9754-c882a5e924a5\") " pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 13 06:49:25 crc kubenswrapper[4971]: E1213 06:49:25.404792 4971 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Dec 13 06:49:25 crc kubenswrapper[4971]: E1213 06:49:25.404808 4971 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Dec 13 06:49:25 crc kubenswrapper[4971]: E1213 06:49:25.404821 4971 projected.go:194] Error preparing data for projected volume kube-api-access-s2dwl for pod openshift-network-diagnostics/network-check-source-55646444c4-trplf: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Dec 13 06:49:25 crc kubenswrapper[4971]: E1213 06:49:25.404856 4971 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Dec 13 06:49:25 crc kubenswrapper[4971]: E1213 06:49:25.404929 4971 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Dec 13 06:49:25 crc kubenswrapper[4971]: E1213 06:49:25.404948 4971 projected.go:194] Error preparing data for projected volume kube-api-access-cqllr for pod openshift-network-diagnostics/network-check-target-xd92c: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Dec 13 06:49:25 crc kubenswrapper[4971]: E1213 06:49:25.404878 4971 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl podName:9d751cbb-f2e2-430d-9754-c882a5e924a5 nodeName:}" failed. No retries permitted until 2025-12-13 06:49:27.404862424 +0000 UTC m=+24.009271872 (durationBeforeRetry 2s). Error: MountVolume.SetUp failed for volume "kube-api-access-s2dwl" (UniqueName: "kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl") pod "network-check-source-55646444c4-trplf" (UID: "9d751cbb-f2e2-430d-9754-c882a5e924a5") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Dec 13 06:49:25 crc kubenswrapper[4971]: E1213 06:49:25.405378 4971 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr podName:3b6479f0-333b-4a96-9adf-2099afdc2447 nodeName:}" failed. No retries permitted until 2025-12-13 06:49:27.405008868 +0000 UTC m=+24.009418356 (durationBeforeRetry 2s). Error: MountVolume.SetUp failed for volume "kube-api-access-cqllr" (UniqueName: "kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr") pod "network-check-target-xd92c" (UID: "3b6479f0-333b-4a96-9adf-2099afdc2447") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Dec 13 06:49:25 crc kubenswrapper[4971]: I1213 06:49:25.408870 4971 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-13T06:49:24Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-13T06:49:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-13T06:49:24Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4328cd6a8464ed468229bbfbc8265f969427916d72ebb92237dbe68efe70fa47\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-13T06:49:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://9d715ca82d9b5a594e41f2b98da6f4277da0e4e1dec114969f901c00f6d7553d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-13T06:49:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-13T06:49:25Z is after 2025-08-24T17:21:41Z" Dec 13 06:49:25 crc kubenswrapper[4971]: I1213 06:49:25.426961 4971 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-13T06:49:23Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-13T06:49:23Z\\\",\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-13T06:49:25Z is after 2025-08-24T17:21:41Z" Dec 13 06:49:25 crc kubenswrapper[4971]: I1213 06:49:25.438572 4971 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-5rzh5" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"6bf4591d-2483-4132-9672-d0fd6c995968\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T06:49:25Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T06:49:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T06:49:25Z\\\",\\\"message\\\":\\\"containers with unready status: [dns-node-resolver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T06:49:25Z\\\",\\\"message\\\":\\\"containers with unready status: [dns-node-resolver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dfxqn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-13T06:49:25Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-5rzh5\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-13T06:49:25Z is after 2025-08-24T17:21:41Z" Dec 13 06:49:25 crc kubenswrapper[4971]: I1213 06:49:25.487936 4971 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-dns/node-resolver-5rzh5" Dec 13 06:49:25 crc kubenswrapper[4971]: W1213 06:49:25.499064 4971 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod6bf4591d_2483_4132_9672_d0fd6c995968.slice/crio-73e35a618a82ce03f57fd754a1850fd3b75284a5add21fd224071e13311e3498 WatchSource:0}: Error finding container 73e35a618a82ce03f57fd754a1850fd3b75284a5add21fd224071e13311e3498: Status 404 returned error can't find the container with id 73e35a618a82ce03f57fd754a1850fd3b75284a5add21fd224071e13311e3498 Dec 13 06:49:25 crc kubenswrapper[4971]: I1213 06:49:25.767451 4971 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 13 06:49:25 crc kubenswrapper[4971]: E1213 06:49:25.767896 4971 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 13 06:49:25 crc kubenswrapper[4971]: I1213 06:49:25.767611 4971 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 13 06:49:25 crc kubenswrapper[4971]: E1213 06:49:25.767960 4971 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 13 06:49:25 crc kubenswrapper[4971]: I1213 06:49:25.767542 4971 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 13 06:49:25 crc kubenswrapper[4971]: E1213 06:49:25.768009 4971 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 13 06:49:25 crc kubenswrapper[4971]: I1213 06:49:25.771917 4971 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="01ab3dd5-8196-46d0-ad33-122e2ca51def" path="/var/lib/kubelet/pods/01ab3dd5-8196-46d0-ad33-122e2ca51def/volumes" Dec 13 06:49:25 crc kubenswrapper[4971]: I1213 06:49:25.772744 4971 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="09ae3b1a-e8e7-4524-b54b-61eab6f9239a" path="/var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/volumes" Dec 13 06:49:25 crc kubenswrapper[4971]: I1213 06:49:25.773603 4971 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="09efc573-dbb6-4249-bd59-9b87aba8dd28" path="/var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes" Dec 13 06:49:25 crc kubenswrapper[4971]: I1213 06:49:25.774320 4971 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="0b574797-001e-440a-8f4e-c0be86edad0f" path="/var/lib/kubelet/pods/0b574797-001e-440a-8f4e-c0be86edad0f/volumes" Dec 13 06:49:25 crc kubenswrapper[4971]: I1213 06:49:25.775560 4971 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="0b78653f-4ff9-4508-8672-245ed9b561e3" path="/var/lib/kubelet/pods/0b78653f-4ff9-4508-8672-245ed9b561e3/volumes" Dec 13 06:49:25 crc kubenswrapper[4971]: I1213 06:49:25.776057 4971 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="1386a44e-36a2-460c-96d0-0359d2b6f0f5" path="/var/lib/kubelet/pods/1386a44e-36a2-460c-96d0-0359d2b6f0f5/volumes" Dec 13 06:49:25 crc kubenswrapper[4971]: I1213 06:49:25.777106 4971 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="1bf7eb37-55a3-4c65-b768-a94c82151e69" path="/var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes" Dec 13 06:49:25 crc kubenswrapper[4971]: I1213 06:49:25.777699 4971 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="1d611f23-29be-4491-8495-bee1670e935f" path="/var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes" Dec 13 06:49:25 crc kubenswrapper[4971]: I1213 06:49:25.778782 4971 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="20b0d48f-5fd6-431c-a545-e3c800c7b866" path="/var/lib/kubelet/pods/20b0d48f-5fd6-431c-a545-e3c800c7b866/volumes" Dec 13 06:49:25 crc kubenswrapper[4971]: I1213 06:49:25.779331 4971 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c" path="/var/lib/kubelet/pods/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c/volumes" Dec 13 06:49:25 crc kubenswrapper[4971]: I1213 06:49:25.780662 4971 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="22c825df-677d-4ca6-82db-3454ed06e783" path="/var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/volumes" Dec 13 06:49:25 crc kubenswrapper[4971]: I1213 06:49:25.781353 4971 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="25e176fe-21b4-4974-b1ed-c8b94f112a7f" path="/var/lib/kubelet/pods/25e176fe-21b4-4974-b1ed-c8b94f112a7f/volumes" Dec 13 06:49:25 crc kubenswrapper[4971]: I1213 06:49:25.781880 4971 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="308be0ea-9f5f-4b29-aeb1-5abd31a0b17b" path="/var/lib/kubelet/pods/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b/volumes" Dec 13 06:49:25 crc kubenswrapper[4971]: I1213 06:49:25.782845 4971 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="31d8b7a1-420e-4252-a5b7-eebe8a111292" path="/var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/volumes" Dec 13 06:49:25 crc kubenswrapper[4971]: I1213 06:49:25.783478 4971 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="3ab1a177-2de0-46d9-b765-d0d0649bb42e" path="/var/lib/kubelet/pods/3ab1a177-2de0-46d9-b765-d0d0649bb42e/volumes" Dec 13 06:49:25 crc kubenswrapper[4971]: I1213 06:49:25.784446 4971 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="3cb93b32-e0ae-4377-b9c8-fdb9842c6d59" path="/var/lib/kubelet/pods/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59/volumes" Dec 13 06:49:25 crc kubenswrapper[4971]: I1213 06:49:25.785152 4971 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="43509403-f426-496e-be36-56cef71462f5" path="/var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes" Dec 13 06:49:25 crc kubenswrapper[4971]: I1213 06:49:25.785796 4971 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="44663579-783b-4372-86d6-acf235a62d72" path="/var/lib/kubelet/pods/44663579-783b-4372-86d6-acf235a62d72/volumes" Dec 13 06:49:25 crc kubenswrapper[4971]: I1213 06:49:25.786956 4971 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="496e6271-fb68-4057-954e-a0d97a4afa3f" path="/var/lib/kubelet/pods/496e6271-fb68-4057-954e-a0d97a4afa3f/volumes" Dec 13 06:49:25 crc kubenswrapper[4971]: I1213 06:49:25.787568 4971 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="49c341d1-5089-4bc2-86a0-a5e165cfcc6b" path="/var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes" Dec 13 06:49:25 crc kubenswrapper[4971]: I1213 06:49:25.788426 4971 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="49ef4625-1d3a-4a9f-b595-c2433d32326d" path="/var/lib/kubelet/pods/49ef4625-1d3a-4a9f-b595-c2433d32326d/volumes" Dec 13 06:49:25 crc kubenswrapper[4971]: I1213 06:49:25.789129 4971 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="4bb40260-dbaa-4fb0-84df-5e680505d512" path="/var/lib/kubelet/pods/4bb40260-dbaa-4fb0-84df-5e680505d512/volumes" Dec 13 06:49:25 crc kubenswrapper[4971]: I1213 06:49:25.789542 4971 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="5225d0e4-402f-4861-b410-819f433b1803" path="/var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes" Dec 13 06:49:25 crc kubenswrapper[4971]: I1213 06:49:25.790549 4971 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="5441d097-087c-4d9a-baa8-b210afa90fc9" path="/var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/volumes" Dec 13 06:49:25 crc kubenswrapper[4971]: I1213 06:49:25.790920 4971 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="57a731c4-ef35-47a8-b875-bfb08a7f8011" path="/var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes" Dec 13 06:49:25 crc kubenswrapper[4971]: I1213 06:49:25.791876 4971 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="5b88f790-22fa-440e-b583-365168c0b23d" path="/var/lib/kubelet/pods/5b88f790-22fa-440e-b583-365168c0b23d/volumes" Dec 13 06:49:25 crc kubenswrapper[4971]: I1213 06:49:25.792467 4971 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="5fe579f8-e8a6-4643-bce5-a661393c4dde" path="/var/lib/kubelet/pods/5fe579f8-e8a6-4643-bce5-a661393c4dde/volumes" Dec 13 06:49:25 crc kubenswrapper[4971]: I1213 06:49:25.793293 4971 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="6402fda4-df10-493c-b4e5-d0569419652d" path="/var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/volumes" Dec 13 06:49:25 crc kubenswrapper[4971]: I1213 06:49:25.793872 4971 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="6509e943-70c6-444c-bc41-48a544e36fbd" path="/var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes" Dec 13 06:49:25 crc kubenswrapper[4971]: I1213 06:49:25.794744 4971 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="6731426b-95fe-49ff-bb5f-40441049fde2" path="/var/lib/kubelet/pods/6731426b-95fe-49ff-bb5f-40441049fde2/volumes" Dec 13 06:49:25 crc kubenswrapper[4971]: I1213 06:49:25.795246 4971 kubelet_volumes.go:152] "Cleaned up orphaned volume subpath from pod" podUID="6ea678ab-3438-413e-bfe3-290ae7725660" path="/var/lib/kubelet/pods/6ea678ab-3438-413e-bfe3-290ae7725660/volume-subpaths/run-systemd/ovnkube-controller/6" Dec 13 06:49:25 crc kubenswrapper[4971]: I1213 06:49:25.795367 4971 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="6ea678ab-3438-413e-bfe3-290ae7725660" path="/var/lib/kubelet/pods/6ea678ab-3438-413e-bfe3-290ae7725660/volumes" Dec 13 06:49:25 crc kubenswrapper[4971]: I1213 06:49:25.797666 4971 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="7539238d-5fe0-46ed-884e-1c3b566537ec" path="/var/lib/kubelet/pods/7539238d-5fe0-46ed-884e-1c3b566537ec/volumes" Dec 13 06:49:25 crc kubenswrapper[4971]: I1213 06:49:25.798700 4971 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="7583ce53-e0fe-4a16-9e4d-50516596a136" path="/var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes" Dec 13 06:49:25 crc kubenswrapper[4971]: I1213 06:49:25.799080 4971 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="7bb08738-c794-4ee8-9972-3a62ca171029" path="/var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/volumes" Dec 13 06:49:25 crc kubenswrapper[4971]: I1213 06:49:25.800492 4971 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="87cf06ed-a83f-41a7-828d-70653580a8cb" path="/var/lib/kubelet/pods/87cf06ed-a83f-41a7-828d-70653580a8cb/volumes" Dec 13 06:49:25 crc kubenswrapper[4971]: I1213 06:49:25.801437 4971 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="8cea82b4-6893-4ddc-af9f-1bb5ae425c5b" path="/var/lib/kubelet/pods/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b/volumes" Dec 13 06:49:25 crc kubenswrapper[4971]: I1213 06:49:25.801972 4971 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="925f1c65-6136-48ba-85aa-3a3b50560753" path="/var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/volumes" Dec 13 06:49:25 crc kubenswrapper[4971]: I1213 06:49:25.802959 4971 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="96b93a3a-6083-4aea-8eab-fe1aa8245ad9" path="/var/lib/kubelet/pods/96b93a3a-6083-4aea-8eab-fe1aa8245ad9/volumes" Dec 13 06:49:25 crc kubenswrapper[4971]: I1213 06:49:25.803600 4971 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="9d4552c7-cd75-42dd-8880-30dd377c49a4" path="/var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/volumes" Dec 13 06:49:25 crc kubenswrapper[4971]: I1213 06:49:25.804473 4971 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="a0128f3a-b052-44ed-a84e-c4c8aaf17c13" path="/var/lib/kubelet/pods/a0128f3a-b052-44ed-a84e-c4c8aaf17c13/volumes" Dec 13 06:49:25 crc kubenswrapper[4971]: I1213 06:49:25.805058 4971 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="a31745f5-9847-4afe-82a5-3161cc66ca93" path="/var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/volumes" Dec 13 06:49:25 crc kubenswrapper[4971]: I1213 06:49:25.806133 4971 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="b11524ee-3fca-4b1b-9cdf-6da289fdbc7d" path="/var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes" Dec 13 06:49:25 crc kubenswrapper[4971]: I1213 06:49:25.806757 4971 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="b6312bbd-5731-4ea0-a20f-81d5a57df44a" path="/var/lib/kubelet/pods/b6312bbd-5731-4ea0-a20f-81d5a57df44a/volumes" Dec 13 06:49:25 crc kubenswrapper[4971]: I1213 06:49:25.807574 4971 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="b6cd30de-2eeb-49a2-ab40-9167f4560ff5" path="/var/lib/kubelet/pods/b6cd30de-2eeb-49a2-ab40-9167f4560ff5/volumes" Dec 13 06:49:25 crc kubenswrapper[4971]: I1213 06:49:25.808076 4971 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="bc5039c0-ea34-426b-a2b7-fbbc87b49a6d" path="/var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/volumes" Dec 13 06:49:25 crc kubenswrapper[4971]: I1213 06:49:25.808955 4971 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="bd23aa5c-e532-4e53-bccf-e79f130c5ae8" path="/var/lib/kubelet/pods/bd23aa5c-e532-4e53-bccf-e79f130c5ae8/volumes" Dec 13 06:49:25 crc kubenswrapper[4971]: I1213 06:49:25.809651 4971 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="bf126b07-da06-4140-9a57-dfd54fc6b486" path="/var/lib/kubelet/pods/bf126b07-da06-4140-9a57-dfd54fc6b486/volumes" Dec 13 06:49:25 crc kubenswrapper[4971]: I1213 06:49:25.810475 4971 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="c03ee662-fb2f-4fc4-a2c1-af487c19d254" path="/var/lib/kubelet/pods/c03ee662-fb2f-4fc4-a2c1-af487c19d254/volumes" Dec 13 06:49:25 crc kubenswrapper[4971]: I1213 06:49:25.810961 4971 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d" path="/var/lib/kubelet/pods/cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d/volumes" Dec 13 06:49:25 crc kubenswrapper[4971]: I1213 06:49:25.811864 4971 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="e7e6199b-1264-4501-8953-767f51328d08" path="/var/lib/kubelet/pods/e7e6199b-1264-4501-8953-767f51328d08/volumes" Dec 13 06:49:25 crc kubenswrapper[4971]: I1213 06:49:25.812364 4971 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="efdd0498-1daa-4136-9a4a-3b948c2293fc" path="/var/lib/kubelet/pods/efdd0498-1daa-4136-9a4a-3b948c2293fc/volumes" Dec 13 06:49:25 crc kubenswrapper[4971]: I1213 06:49:25.812902 4971 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="f88749ec-7931-4ee7-b3fc-1ec5e11f92e9" path="/var/lib/kubelet/pods/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9/volumes" Dec 13 06:49:25 crc kubenswrapper[4971]: I1213 06:49:25.813708 4971 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="fda69060-fa79-4696-b1a6-7980f124bf7c" path="/var/lib/kubelet/pods/fda69060-fa79-4696-b1a6-7980f124bf7c/volumes" Dec 13 06:49:25 crc kubenswrapper[4971]: I1213 06:49:25.925821 4971 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-dns/node-resolver-5rzh5" event={"ID":"6bf4591d-2483-4132-9672-d0fd6c995968","Type":"ContainerStarted","Data":"50e2cb14e4ffb603acb06e2281615a99f06182d345c4e70f2446fa2077033bbb"} Dec 13 06:49:25 crc kubenswrapper[4971]: I1213 06:49:25.925877 4971 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-dns/node-resolver-5rzh5" event={"ID":"6bf4591d-2483-4132-9672-d0fd6c995968","Type":"ContainerStarted","Data":"73e35a618a82ce03f57fd754a1850fd3b75284a5add21fd224071e13311e3498"} Dec 13 06:49:25 crc kubenswrapper[4971]: I1213 06:49:25.939633 4971 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-13T06:49:23Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-13T06:49:23Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-13T06:49:25Z is after 2025-08-24T17:21:41Z" Dec 13 06:49:25 crc kubenswrapper[4971]: I1213 06:49:25.954775 4971 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-13T06:49:23Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-13T06:49:23Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-13T06:49:25Z is after 2025-08-24T17:21:41Z" Dec 13 06:49:25 crc kubenswrapper[4971]: I1213 06:49:25.969282 4971 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"f9d31b5a-b770-4a15-95c8-35b49677c608\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T06:49:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T06:49:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T06:49:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T06:49:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T06:49:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c9baa042f9dada954faa57b5454cd6b0a12a297e56ddc192c52cff50e42a16fa\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-13T06:49:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://c788de86754f0bf37ee184672fc1dcb388ccebc246d4f15be4b5da661f725f8f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-13T06:49:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://760817511a9b4f96a68dc38d26631250eaab72e1dd8f123e653039b6b4e91f84\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-13T06:49:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b859be0c9feac856e9855c6a038b02b49b32fade97873c46d0bec0a32d8b5147\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-13T06:49:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-13T06:49:03Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-13T06:49:25Z is after 2025-08-24T17:21:41Z" Dec 13 06:49:25 crc kubenswrapper[4971]: I1213 06:49:25.989623 4971 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3ad3268a-a982-428c-9c06-4aaeddd372d9\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T06:49:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T06:49:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T06:49:03Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T06:49:03Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T06:49:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3716afdf2bc693be80a09fc9f7727d334f435caf1d4c8ec4811b3a1f1652cdfe\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-13T06:49:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://30df3abf8c79db8af52c6a918f73e1ce410ddafcf341636424c267e1eeb195a7\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-13T06:49:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://ccdf65f6d2b1eb3647a03ba37c305c32e34addfdf79ff937c8dddaa3e5d5b5c6\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-13T06:49:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://770b73dd0c2b6f38744e2c1cd6aa186a69bb87d1cd015a603ad2e33be06a8f6b\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://b3c1bab3d53f975b2e24582ce2771bc26d600e0dd7c30339dcfd18a071162ad4\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-12-13T06:49:23Z\\\",\\\"message\\\":\\\"file observer\\\\nW1213 06:49:23.283403 1 builder.go:272] unable to get owner reference (falling back to namespace): pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\nI1213 06:49:23.284336 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1213 06:49:23.286978 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-981773671/tls.crt::/tmp/serving-cert-981773671/tls.key\\\\\\\"\\\\nI1213 06:49:23.775072 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1213 06:49:23.778615 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1213 06:49:23.778645 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1213 06:49:23.778680 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1213 06:49:23.778691 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1213 06:49:23.785058 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1213 06:49:23.786150 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1213 06:49:23.786196 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1213 06:49:23.786223 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1213 06:49:23.786246 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1213 06:49:23.786269 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1213 06:49:23.786291 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI1213 06:49:23.785608 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF1213 06:49:23.787563 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-13T06:49:08Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-13T06:49:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://9acd07066deed483bd7fa61ee45fda09539d110717a134a56827f871186f92fc\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-13T06:49:07Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://38c91cfbaecfdbd3e3b663422a91ab4294610afb6bbf9ebdb77db24e2b76c682\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://38c91cfbaecfdbd3e3b663422a91ab4294610afb6bbf9ebdb77db24e2b76c682\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-13T06:49:05Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-13T06:49:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-13T06:49:03Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-13T06:49:25Z is after 2025-08-24T17:21:41Z" Dec 13 06:49:25 crc kubenswrapper[4971]: I1213 06:49:25.991971 4971 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-machine-config-operator/machine-config-daemon-82xjz"] Dec 13 06:49:25 crc kubenswrapper[4971]: I1213 06:49:25.992388 4971 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-multus/multus-wwvpl"] Dec 13 06:49:25 crc kubenswrapper[4971]: I1213 06:49:25.992579 4971 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-ovn-kubernetes/ovnkube-node-zlcg5"] Dec 13 06:49:25 crc kubenswrapper[4971]: I1213 06:49:25.992588 4971 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-machine-config-operator/machine-config-daemon-82xjz" Dec 13 06:49:25 crc kubenswrapper[4971]: I1213 06:49:25.992739 4971 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/multus-wwvpl" Dec 13 06:49:25 crc kubenswrapper[4971]: I1213 06:49:25.993876 4971 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-multus/multus-additional-cni-plugins-ncchz"] Dec 13 06:49:25 crc kubenswrapper[4971]: I1213 06:49:25.994865 4971 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-ovn-kubernetes/ovnkube-node-zlcg5" Dec 13 06:49:25 crc kubenswrapper[4971]: I1213 06:49:25.995271 4971 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/multus-additional-cni-plugins-ncchz" Dec 13 06:49:25 crc kubenswrapper[4971]: I1213 06:49:25.999358 4971 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-multus"/"openshift-service-ca.crt" Dec 13 06:49:25 crc kubenswrapper[4971]: I1213 06:49:25.999669 4971 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-machine-config-operator"/"kube-rbac-proxy" Dec 13 06:49:26 crc kubenswrapper[4971]: I1213 06:49:26.000493 4971 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-multus"/"kube-root-ca.crt" Dec 13 06:49:26 crc kubenswrapper[4971]: I1213 06:49:26.000574 4971 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-machine-config-operator"/"openshift-service-ca.crt" Dec 13 06:49:26 crc kubenswrapper[4971]: I1213 06:49:26.000697 4971 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-multus"/"default-cni-sysctl-allowlist" Dec 13 06:49:26 crc kubenswrapper[4971]: I1213 06:49:26.000807 4971 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-machine-config-operator"/"kube-root-ca.crt" Dec 13 06:49:26 crc kubenswrapper[4971]: I1213 06:49:26.000927 4971 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-multus"/"default-dockercfg-2q5b6" Dec 13 06:49:26 crc kubenswrapper[4971]: I1213 06:49:26.001149 4971 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ovn-kubernetes"/"env-overrides" Dec 13 06:49:26 crc kubenswrapper[4971]: I1213 06:49:26.001155 4971 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"proxy-tls" Dec 13 06:49:26 crc kubenswrapper[4971]: I1213 06:49:26.001220 4971 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ovn-kubernetes"/"openshift-service-ca.crt" Dec 13 06:49:26 crc kubenswrapper[4971]: I1213 06:49:26.001220 4971 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-multus"/"multus-daemon-config" Dec 13 06:49:26 crc kubenswrapper[4971]: I1213 06:49:26.001242 4971 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-multus"/"cni-copy-resources" Dec 13 06:49:26 crc kubenswrapper[4971]: I1213 06:49:26.001261 4971 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ovn-kubernetes"/"ovn-node-metrics-cert" Dec 13 06:49:26 crc kubenswrapper[4971]: I1213 06:49:26.001714 4971 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"machine-config-daemon-dockercfg-r5tcq" Dec 13 06:49:26 crc kubenswrapper[4971]: I1213 06:49:26.001785 4971 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ovn-kubernetes"/"kube-root-ca.crt" Dec 13 06:49:26 crc kubenswrapper[4971]: I1213 06:49:26.001815 4971 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ovn-kubernetes"/"ovn-kubernetes-node-dockercfg-pwtwl" Dec 13 06:49:26 crc kubenswrapper[4971]: I1213 06:49:26.003417 4971 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-multus"/"multus-ancillary-tools-dockercfg-vnmsz" Dec 13 06:49:26 crc kubenswrapper[4971]: I1213 06:49:26.003444 4971 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ovn-kubernetes"/"ovnkube-script-lib" Dec 13 06:49:26 crc kubenswrapper[4971]: I1213 06:49:26.003649 4971 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ovn-kubernetes"/"ovnkube-config" Dec 13 06:49:26 crc kubenswrapper[4971]: I1213 06:49:26.017644 4971 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-13T06:49:23Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-13T06:49:23Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-13T06:49:26Z is after 2025-08-24T17:21:41Z" Dec 13 06:49:26 crc kubenswrapper[4971]: I1213 06:49:26.031265 4971 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-13T06:49:24Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-13T06:49:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-13T06:49:24Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4328cd6a8464ed468229bbfbc8265f969427916d72ebb92237dbe68efe70fa47\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-13T06:49:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://9d715ca82d9b5a594e41f2b98da6f4277da0e4e1dec114969f901c00f6d7553d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-13T06:49:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-13T06:49:26Z is after 2025-08-24T17:21:41Z" Dec 13 06:49:26 crc kubenswrapper[4971]: I1213 06:49:26.041424 4971 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-13T06:49:23Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-13T06:49:23Z\\\",\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-13T06:49:26Z is after 2025-08-24T17:21:41Z" Dec 13 06:49:26 crc kubenswrapper[4971]: I1213 06:49:26.050673 4971 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-5rzh5" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"6bf4591d-2483-4132-9672-d0fd6c995968\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T06:49:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T06:49:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T06:49:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T06:49:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://50e2cb14e4ffb603acb06e2281615a99f06182d345c4e70f2446fa2077033bbb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-13T06:49:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dfxqn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-13T06:49:25Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-5rzh5\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-13T06:49:26Z is after 2025-08-24T17:21:41Z" Dec 13 06:49:26 crc kubenswrapper[4971]: I1213 06:49:26.082288 4971 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"1ed9f82e-fe5c-4b76-a619-62e5abd88d3b\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T06:49:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T06:49:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T06:49:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T06:49:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T06:49:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e457329a78a2d24337f994e8a0a1f75a2a667244aa75eb84613d1e7138364aac\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-13T06:49:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://eb014de8140deef097d9bd94f3b9fa4ab3558bc4449bef84b83bf62bfabf5836\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-13T06:49:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://2c438cc74d555f7c9db4569d1c56b78b1088ed8105697d2c0f31e6c5136f3dd0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-13T06:49:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://aa6fa6e4340dcd066252df3c56d79a3a39ed531d5eea8d1d5941d6952368d3af\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-13T06:49:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://c4471b98854c52d5cd084d0ba3e64e5b08e9518ec3c372cce2c875a440522457\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-13T06:49:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7d2746335c98dae3a0f43024d34cb62ebcb42a01c8f9ce18d52b29d4c45598e5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://7d2746335c98dae3a0f43024d34cb62ebcb42a01c8f9ce18d52b29d4c45598e5\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-13T06:49:04Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-13T06:49:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://bc0e77d10a738128bea82122496bc42235fc1688f8907756d93e61b651902b57\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://bc0e77d10a738128bea82122496bc42235fc1688f8907756d93e61b651902b57\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-13T06:49:05Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-13T06:49:05Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://f821e418cba5e832da8839268bcbbd9630bdf319a212d5382bfa2c4b0b73a268\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://f821e418cba5e832da8839268bcbbd9630bdf319a212d5382bfa2c4b0b73a268\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-13T06:49:07Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-13T06:49:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-13T06:49:03Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-13T06:49:26Z is after 2025-08-24T17:21:41Z" Dec 13 06:49:26 crc kubenswrapper[4971]: I1213 06:49:26.101816 4971 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-13T06:49:24Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-13T06:49:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-13T06:49:24Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d0b2ecf0d5b22f23217f42f259ae7ffb6359555b22c30eaebf667eaf254b3bab\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-13T06:49:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-13T06:49:26Z is after 2025-08-24T17:21:41Z" Dec 13 06:49:26 crc kubenswrapper[4971]: I1213 06:49:26.111362 4971 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-run-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/736c6b1a-93dd-4f71-81b6-1c43fcf5f556-host-run-ovn-kubernetes\") pod \"ovnkube-node-zlcg5\" (UID: \"736c6b1a-93dd-4f71-81b6-1c43fcf5f556\") " pod="openshift-ovn-kubernetes/ovnkube-node-zlcg5" Dec 13 06:49:26 crc kubenswrapper[4971]: I1213 06:49:26.111402 4971 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-kubernetes\" (UniqueName: \"kubernetes.io/host-path/dddf744b-2122-49ad-a980-105c09636e0f-etc-kubernetes\") pod \"multus-wwvpl\" (UID: \"dddf744b-2122-49ad-a980-105c09636e0f\") " pod="openshift-multus/multus-wwvpl" Dec 13 06:49:26 crc kubenswrapper[4971]: I1213 06:49:26.111427 4971 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-openvswitch\" (UniqueName: \"kubernetes.io/host-path/736c6b1a-93dd-4f71-81b6-1c43fcf5f556-etc-openvswitch\") pod \"ovnkube-node-zlcg5\" (UID: \"736c6b1a-93dd-4f71-81b6-1c43fcf5f556\") " pod="openshift-ovn-kubernetes/ovnkube-node-zlcg5" Dec 13 06:49:26 crc kubenswrapper[4971]: I1213 06:49:26.111442 4971 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-cni-bin\" (UniqueName: \"kubernetes.io/host-path/736c6b1a-93dd-4f71-81b6-1c43fcf5f556-host-cni-bin\") pod \"ovnkube-node-zlcg5\" (UID: \"736c6b1a-93dd-4f71-81b6-1c43fcf5f556\") " pod="openshift-ovn-kubernetes/ovnkube-node-zlcg5" Dec 13 06:49:26 crc kubenswrapper[4971]: I1213 06:49:26.111503 4971 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-run-k8s-cni-cncf-io\" (UniqueName: \"kubernetes.io/host-path/dddf744b-2122-49ad-a980-105c09636e0f-host-run-k8s-cni-cncf-io\") pod \"multus-wwvpl\" (UID: \"dddf744b-2122-49ad-a980-105c09636e0f\") " pod="openshift-multus/multus-wwvpl" Dec 13 06:49:26 crc kubenswrapper[4971]: I1213 06:49:26.111572 4971 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-var-lib-kubelet\" (UniqueName: \"kubernetes.io/host-path/dddf744b-2122-49ad-a980-105c09636e0f-host-var-lib-kubelet\") pod \"multus-wwvpl\" (UID: \"dddf744b-2122-49ad-a980-105c09636e0f\") " pod="openshift-multus/multus-wwvpl" Dec 13 06:49:26 crc kubenswrapper[4971]: I1213 06:49:26.111600 4971 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"log-socket\" (UniqueName: \"kubernetes.io/host-path/736c6b1a-93dd-4f71-81b6-1c43fcf5f556-log-socket\") pod \"ovnkube-node-zlcg5\" (UID: \"736c6b1a-93dd-4f71-81b6-1c43fcf5f556\") " pod="openshift-ovn-kubernetes/ovnkube-node-zlcg5" Dec 13 06:49:26 crc kubenswrapper[4971]: I1213 06:49:26.111619 4971 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run-systemd\" (UniqueName: \"kubernetes.io/host-path/736c6b1a-93dd-4f71-81b6-1c43fcf5f556-run-systemd\") pod \"ovnkube-node-zlcg5\" (UID: \"736c6b1a-93dd-4f71-81b6-1c43fcf5f556\") " pod="openshift-ovn-kubernetes/ovnkube-node-zlcg5" Dec 13 06:49:26 crc kubenswrapper[4971]: I1213 06:49:26.111642 4971 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run-ovn\" (UniqueName: \"kubernetes.io/host-path/736c6b1a-93dd-4f71-81b6-1c43fcf5f556-run-ovn\") pod \"ovnkube-node-zlcg5\" (UID: \"736c6b1a-93dd-4f71-81b6-1c43fcf5f556\") " pod="openshift-ovn-kubernetes/ovnkube-node-zlcg5" Dec 13 06:49:26 crc kubenswrapper[4971]: I1213 06:49:26.111664 4971 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"tuning-conf-dir\" (UniqueName: \"kubernetes.io/host-path/de914d21-a688-401e-a66e-861beb6336d9-tuning-conf-dir\") pod \"multus-additional-cni-plugins-ncchz\" (UID: \"de914d21-a688-401e-a66e-861beb6336d9\") " pod="openshift-multus/multus-additional-cni-plugins-ncchz" Dec 13 06:49:26 crc kubenswrapper[4971]: I1213 06:49:26.111688 4971 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-lqmjj\" (UniqueName: \"kubernetes.io/projected/de914d21-a688-401e-a66e-861beb6336d9-kube-api-access-lqmjj\") pod \"multus-additional-cni-plugins-ncchz\" (UID: \"de914d21-a688-401e-a66e-861beb6336d9\") " pod="openshift-multus/multus-additional-cni-plugins-ncchz" Dec 13 06:49:26 crc kubenswrapper[4971]: I1213 06:49:26.111755 4971 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"os-release\" (UniqueName: \"kubernetes.io/host-path/dddf744b-2122-49ad-a980-105c09636e0f-os-release\") pod \"multus-wwvpl\" (UID: \"dddf744b-2122-49ad-a980-105c09636e0f\") " pod="openshift-multus/multus-wwvpl" Dec 13 06:49:26 crc kubenswrapper[4971]: I1213 06:49:26.111799 4971 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rootfs\" (UniqueName: \"kubernetes.io/host-path/e0c4af12-edda-41d7-9b44-f87396174ef2-rootfs\") pod \"machine-config-daemon-82xjz\" (UID: \"e0c4af12-edda-41d7-9b44-f87396174ef2\") " pod="openshift-machine-config-operator/machine-config-daemon-82xjz" Dec 13 06:49:26 crc kubenswrapper[4971]: I1213 06:49:26.111822 4971 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-kubelet\" (UniqueName: \"kubernetes.io/host-path/736c6b1a-93dd-4f71-81b6-1c43fcf5f556-host-kubelet\") pod \"ovnkube-node-zlcg5\" (UID: \"736c6b1a-93dd-4f71-81b6-1c43fcf5f556\") " pod="openshift-ovn-kubernetes/ovnkube-node-zlcg5" Dec 13 06:49:26 crc kubenswrapper[4971]: I1213 06:49:26.111842 4971 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovn-node-metrics-cert\" (UniqueName: \"kubernetes.io/secret/736c6b1a-93dd-4f71-81b6-1c43fcf5f556-ovn-node-metrics-cert\") pod \"ovnkube-node-zlcg5\" (UID: \"736c6b1a-93dd-4f71-81b6-1c43fcf5f556\") " pod="openshift-ovn-kubernetes/ovnkube-node-zlcg5" Dec 13 06:49:26 crc kubenswrapper[4971]: I1213 06:49:26.111861 4971 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cni-binary-copy\" (UniqueName: \"kubernetes.io/configmap/dddf744b-2122-49ad-a980-105c09636e0f-cni-binary-copy\") pod \"multus-wwvpl\" (UID: \"dddf744b-2122-49ad-a980-105c09636e0f\") " pod="openshift-multus/multus-wwvpl" Dec 13 06:49:26 crc kubenswrapper[4971]: I1213 06:49:26.111900 4971 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"hostroot\" (UniqueName: \"kubernetes.io/host-path/dddf744b-2122-49ad-a980-105c09636e0f-hostroot\") pod \"multus-wwvpl\" (UID: \"dddf744b-2122-49ad-a980-105c09636e0f\") " pod="openshift-multus/multus-wwvpl" Dec 13 06:49:26 crc kubenswrapper[4971]: I1213 06:49:26.111928 4971 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"system-cni-dir\" (UniqueName: \"kubernetes.io/host-path/de914d21-a688-401e-a66e-861beb6336d9-system-cni-dir\") pod \"multus-additional-cni-plugins-ncchz\" (UID: \"de914d21-a688-401e-a66e-861beb6336d9\") " pod="openshift-multus/multus-additional-cni-plugins-ncchz" Dec 13 06:49:26 crc kubenswrapper[4971]: I1213 06:49:26.111960 4971 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"systemd-units\" (UniqueName: \"kubernetes.io/host-path/736c6b1a-93dd-4f71-81b6-1c43fcf5f556-systemd-units\") pod \"ovnkube-node-zlcg5\" (UID: \"736c6b1a-93dd-4f71-81b6-1c43fcf5f556\") " pod="openshift-ovn-kubernetes/ovnkube-node-zlcg5" Dec 13 06:49:26 crc kubenswrapper[4971]: I1213 06:49:26.111980 4971 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-slash\" (UniqueName: \"kubernetes.io/host-path/736c6b1a-93dd-4f71-81b6-1c43fcf5f556-host-slash\") pod \"ovnkube-node-zlcg5\" (UID: \"736c6b1a-93dd-4f71-81b6-1c43fcf5f556\") " pod="openshift-ovn-kubernetes/ovnkube-node-zlcg5" Dec 13 06:49:26 crc kubenswrapper[4971]: I1213 06:49:26.112001 4971 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"os-release\" (UniqueName: \"kubernetes.io/host-path/de914d21-a688-401e-a66e-861beb6336d9-os-release\") pod \"multus-additional-cni-plugins-ncchz\" (UID: \"de914d21-a688-401e-a66e-861beb6336d9\") " pod="openshift-multus/multus-additional-cni-plugins-ncchz" Dec 13 06:49:26 crc kubenswrapper[4971]: I1213 06:49:26.112022 4971 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cni-binary-copy\" (UniqueName: \"kubernetes.io/configmap/de914d21-a688-401e-a66e-861beb6336d9-cni-binary-copy\") pod \"multus-additional-cni-plugins-ncchz\" (UID: \"de914d21-a688-401e-a66e-861beb6336d9\") " pod="openshift-multus/multus-additional-cni-plugins-ncchz" Dec 13 06:49:26 crc kubenswrapper[4971]: I1213 06:49:26.112042 4971 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"multus-daemon-config\" (UniqueName: \"kubernetes.io/configmap/dddf744b-2122-49ad-a980-105c09636e0f-multus-daemon-config\") pod \"multus-wwvpl\" (UID: \"dddf744b-2122-49ad-a980-105c09636e0f\") " pod="openshift-multus/multus-wwvpl" Dec 13 06:49:26 crc kubenswrapper[4971]: I1213 06:49:26.112068 4971 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-cnxsq\" (UniqueName: \"kubernetes.io/projected/e0c4af12-edda-41d7-9b44-f87396174ef2-kube-api-access-cnxsq\") pod \"machine-config-daemon-82xjz\" (UID: \"e0c4af12-edda-41d7-9b44-f87396174ef2\") " pod="openshift-machine-config-operator/machine-config-daemon-82xjz" Dec 13 06:49:26 crc kubenswrapper[4971]: I1213 06:49:26.112089 4971 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/736c6b1a-93dd-4f71-81b6-1c43fcf5f556-env-overrides\") pod \"ovnkube-node-zlcg5\" (UID: \"736c6b1a-93dd-4f71-81b6-1c43fcf5f556\") " pod="openshift-ovn-kubernetes/ovnkube-node-zlcg5" Dec 13 06:49:26 crc kubenswrapper[4971]: I1213 06:49:26.112113 4971 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"node-log\" (UniqueName: \"kubernetes.io/host-path/736c6b1a-93dd-4f71-81b6-1c43fcf5f556-node-log\") pod \"ovnkube-node-zlcg5\" (UID: \"736c6b1a-93dd-4f71-81b6-1c43fcf5f556\") " pod="openshift-ovn-kubernetes/ovnkube-node-zlcg5" Dec 13 06:49:26 crc kubenswrapper[4971]: I1213 06:49:26.112136 4971 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovnkube-script-lib\" (UniqueName: \"kubernetes.io/configmap/736c6b1a-93dd-4f71-81b6-1c43fcf5f556-ovnkube-script-lib\") pod \"ovnkube-node-zlcg5\" (UID: \"736c6b1a-93dd-4f71-81b6-1c43fcf5f556\") " pod="openshift-ovn-kubernetes/ovnkube-node-zlcg5" Dec 13 06:49:26 crc kubenswrapper[4971]: I1213 06:49:26.112157 4971 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cnibin\" (UniqueName: \"kubernetes.io/host-path/de914d21-a688-401e-a66e-861beb6336d9-cnibin\") pod \"multus-additional-cni-plugins-ncchz\" (UID: \"de914d21-a688-401e-a66e-861beb6336d9\") " pod="openshift-multus/multus-additional-cni-plugins-ncchz" Dec 13 06:49:26 crc kubenswrapper[4971]: I1213 06:49:26.112179 4971 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cni-sysctl-allowlist\" (UniqueName: \"kubernetes.io/configmap/de914d21-a688-401e-a66e-861beb6336d9-cni-sysctl-allowlist\") pod \"multus-additional-cni-plugins-ncchz\" (UID: \"de914d21-a688-401e-a66e-861beb6336d9\") " pod="openshift-multus/multus-additional-cni-plugins-ncchz" Dec 13 06:49:26 crc kubenswrapper[4971]: I1213 06:49:26.112262 4971 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/e0c4af12-edda-41d7-9b44-f87396174ef2-proxy-tls\") pod \"machine-config-daemon-82xjz\" (UID: \"e0c4af12-edda-41d7-9b44-f87396174ef2\") " pod="openshift-machine-config-operator/machine-config-daemon-82xjz" Dec 13 06:49:26 crc kubenswrapper[4971]: I1213 06:49:26.112311 4971 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-cni-netd\" (UniqueName: \"kubernetes.io/host-path/736c6b1a-93dd-4f71-81b6-1c43fcf5f556-host-cni-netd\") pod \"ovnkube-node-zlcg5\" (UID: \"736c6b1a-93dd-4f71-81b6-1c43fcf5f556\") " pod="openshift-ovn-kubernetes/ovnkube-node-zlcg5" Dec 13 06:49:26 crc kubenswrapper[4971]: I1213 06:49:26.112335 4971 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"multus-socket-dir-parent\" (UniqueName: \"kubernetes.io/host-path/dddf744b-2122-49ad-a980-105c09636e0f-multus-socket-dir-parent\") pod \"multus-wwvpl\" (UID: \"dddf744b-2122-49ad-a980-105c09636e0f\") " pod="openshift-multus/multus-wwvpl" Dec 13 06:49:26 crc kubenswrapper[4971]: I1213 06:49:26.112358 4971 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-run-multus-certs\" (UniqueName: \"kubernetes.io/host-path/dddf744b-2122-49ad-a980-105c09636e0f-host-run-multus-certs\") pod \"multus-wwvpl\" (UID: \"dddf744b-2122-49ad-a980-105c09636e0f\") " pod="openshift-multus/multus-wwvpl" Dec 13 06:49:26 crc kubenswrapper[4971]: I1213 06:49:26.112391 4971 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"multus-conf-dir\" (UniqueName: \"kubernetes.io/host-path/dddf744b-2122-49ad-a980-105c09636e0f-multus-conf-dir\") pod \"multus-wwvpl\" (UID: \"dddf744b-2122-49ad-a980-105c09636e0f\") " pod="openshift-multus/multus-wwvpl" Dec 13 06:49:26 crc kubenswrapper[4971]: I1213 06:49:26.112436 4971 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"mcd-auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/e0c4af12-edda-41d7-9b44-f87396174ef2-mcd-auth-proxy-config\") pod \"machine-config-daemon-82xjz\" (UID: \"e0c4af12-edda-41d7-9b44-f87396174ef2\") " pod="openshift-machine-config-operator/machine-config-daemon-82xjz" Dec 13 06:49:26 crc kubenswrapper[4971]: I1213 06:49:26.112452 4971 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run-openvswitch\" (UniqueName: \"kubernetes.io/host-path/736c6b1a-93dd-4f71-81b6-1c43fcf5f556-run-openvswitch\") pod \"ovnkube-node-zlcg5\" (UID: \"736c6b1a-93dd-4f71-81b6-1c43fcf5f556\") " pod="openshift-ovn-kubernetes/ovnkube-node-zlcg5" Dec 13 06:49:26 crc kubenswrapper[4971]: I1213 06:49:26.112494 4971 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-run-netns\" (UniqueName: \"kubernetes.io/host-path/dddf744b-2122-49ad-a980-105c09636e0f-host-run-netns\") pod \"multus-wwvpl\" (UID: \"dddf744b-2122-49ad-a980-105c09636e0f\") " pod="openshift-multus/multus-wwvpl" Dec 13 06:49:26 crc kubenswrapper[4971]: I1213 06:49:26.112510 4971 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"system-cni-dir\" (UniqueName: \"kubernetes.io/host-path/dddf744b-2122-49ad-a980-105c09636e0f-system-cni-dir\") pod \"multus-wwvpl\" (UID: \"dddf744b-2122-49ad-a980-105c09636e0f\") " pod="openshift-multus/multus-wwvpl" Dec 13 06:49:26 crc kubenswrapper[4971]: I1213 06:49:26.112557 4971 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"multus-cni-dir\" (UniqueName: \"kubernetes.io/host-path/dddf744b-2122-49ad-a980-105c09636e0f-multus-cni-dir\") pod \"multus-wwvpl\" (UID: \"dddf744b-2122-49ad-a980-105c09636e0f\") " pod="openshift-multus/multus-wwvpl" Dec 13 06:49:26 crc kubenswrapper[4971]: I1213 06:49:26.112575 4971 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-var-lib-cni-networks-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/736c6b1a-93dd-4f71-81b6-1c43fcf5f556-host-var-lib-cni-networks-ovn-kubernetes\") pod \"ovnkube-node-zlcg5\" (UID: \"736c6b1a-93dd-4f71-81b6-1c43fcf5f556\") " pod="openshift-ovn-kubernetes/ovnkube-node-zlcg5" Dec 13 06:49:26 crc kubenswrapper[4971]: I1213 06:49:26.112627 4971 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/736c6b1a-93dd-4f71-81b6-1c43fcf5f556-ovnkube-config\") pod \"ovnkube-node-zlcg5\" (UID: \"736c6b1a-93dd-4f71-81b6-1c43fcf5f556\") " pod="openshift-ovn-kubernetes/ovnkube-node-zlcg5" Dec 13 06:49:26 crc kubenswrapper[4971]: I1213 06:49:26.112652 4971 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-run-netns\" (UniqueName: \"kubernetes.io/host-path/736c6b1a-93dd-4f71-81b6-1c43fcf5f556-host-run-netns\") pod \"ovnkube-node-zlcg5\" (UID: \"736c6b1a-93dd-4f71-81b6-1c43fcf5f556\") " pod="openshift-ovn-kubernetes/ovnkube-node-zlcg5" Dec 13 06:49:26 crc kubenswrapper[4971]: I1213 06:49:26.112666 4971 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-lib-openvswitch\" (UniqueName: \"kubernetes.io/host-path/736c6b1a-93dd-4f71-81b6-1c43fcf5f556-var-lib-openvswitch\") pod \"ovnkube-node-zlcg5\" (UID: \"736c6b1a-93dd-4f71-81b6-1c43fcf5f556\") " pod="openshift-ovn-kubernetes/ovnkube-node-zlcg5" Dec 13 06:49:26 crc kubenswrapper[4971]: I1213 06:49:26.112696 4971 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-vzmsr\" (UniqueName: \"kubernetes.io/projected/736c6b1a-93dd-4f71-81b6-1c43fcf5f556-kube-api-access-vzmsr\") pod \"ovnkube-node-zlcg5\" (UID: \"736c6b1a-93dd-4f71-81b6-1c43fcf5f556\") " pod="openshift-ovn-kubernetes/ovnkube-node-zlcg5" Dec 13 06:49:26 crc kubenswrapper[4971]: I1213 06:49:26.112714 4971 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cnibin\" (UniqueName: \"kubernetes.io/host-path/dddf744b-2122-49ad-a980-105c09636e0f-cnibin\") pod \"multus-wwvpl\" (UID: \"dddf744b-2122-49ad-a980-105c09636e0f\") " pod="openshift-multus/multus-wwvpl" Dec 13 06:49:26 crc kubenswrapper[4971]: I1213 06:49:26.112729 4971 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-var-lib-cni-bin\" (UniqueName: \"kubernetes.io/host-path/dddf744b-2122-49ad-a980-105c09636e0f-host-var-lib-cni-bin\") pod \"multus-wwvpl\" (UID: \"dddf744b-2122-49ad-a980-105c09636e0f\") " pod="openshift-multus/multus-wwvpl" Dec 13 06:49:26 crc kubenswrapper[4971]: I1213 06:49:26.112743 4971 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-var-lib-cni-multus\" (UniqueName: \"kubernetes.io/host-path/dddf744b-2122-49ad-a980-105c09636e0f-host-var-lib-cni-multus\") pod \"multus-wwvpl\" (UID: \"dddf744b-2122-49ad-a980-105c09636e0f\") " pod="openshift-multus/multus-wwvpl" Dec 13 06:49:26 crc kubenswrapper[4971]: I1213 06:49:26.112779 4971 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-czz9v\" (UniqueName: \"kubernetes.io/projected/dddf744b-2122-49ad-a980-105c09636e0f-kube-api-access-czz9v\") pod \"multus-wwvpl\" (UID: \"dddf744b-2122-49ad-a980-105c09636e0f\") " pod="openshift-multus/multus-wwvpl" Dec 13 06:49:26 crc kubenswrapper[4971]: I1213 06:49:26.164880 4971 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"1ed9f82e-fe5c-4b76-a619-62e5abd88d3b\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T06:49:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T06:49:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T06:49:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T06:49:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T06:49:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e457329a78a2d24337f994e8a0a1f75a2a667244aa75eb84613d1e7138364aac\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-13T06:49:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://eb014de8140deef097d9bd94f3b9fa4ab3558bc4449bef84b83bf62bfabf5836\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-13T06:49:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://2c438cc74d555f7c9db4569d1c56b78b1088ed8105697d2c0f31e6c5136f3dd0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-13T06:49:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://aa6fa6e4340dcd066252df3c56d79a3a39ed531d5eea8d1d5941d6952368d3af\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-13T06:49:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://c4471b98854c52d5cd084d0ba3e64e5b08e9518ec3c372cce2c875a440522457\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-13T06:49:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7d2746335c98dae3a0f43024d34cb62ebcb42a01c8f9ce18d52b29d4c45598e5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://7d2746335c98dae3a0f43024d34cb62ebcb42a01c8f9ce18d52b29d4c45598e5\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-13T06:49:04Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-13T06:49:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://bc0e77d10a738128bea82122496bc42235fc1688f8907756d93e61b651902b57\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://bc0e77d10a738128bea82122496bc42235fc1688f8907756d93e61b651902b57\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-13T06:49:05Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-13T06:49:05Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://f821e418cba5e832da8839268bcbbd9630bdf319a212d5382bfa2c4b0b73a268\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://f821e418cba5e832da8839268bcbbd9630bdf319a212d5382bfa2c4b0b73a268\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-13T06:49:07Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-13T06:49:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-13T06:49:03Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-13T06:49:26Z is after 2025-08-24T17:21:41Z" Dec 13 06:49:26 crc kubenswrapper[4971]: I1213 06:49:26.201344 4971 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-13T06:49:24Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-13T06:49:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-13T06:49:24Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4328cd6a8464ed468229bbfbc8265f969427916d72ebb92237dbe68efe70fa47\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-13T06:49:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://9d715ca82d9b5a594e41f2b98da6f4277da0e4e1dec114969f901c00f6d7553d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-13T06:49:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-13T06:49:26Z is after 2025-08-24T17:21:41Z" Dec 13 06:49:26 crc kubenswrapper[4971]: I1213 06:49:26.213912 4971 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run-systemd\" (UniqueName: \"kubernetes.io/host-path/736c6b1a-93dd-4f71-81b6-1c43fcf5f556-run-systemd\") pod \"ovnkube-node-zlcg5\" (UID: \"736c6b1a-93dd-4f71-81b6-1c43fcf5f556\") " pod="openshift-ovn-kubernetes/ovnkube-node-zlcg5" Dec 13 06:49:26 crc kubenswrapper[4971]: I1213 06:49:26.213948 4971 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run-ovn\" (UniqueName: \"kubernetes.io/host-path/736c6b1a-93dd-4f71-81b6-1c43fcf5f556-run-ovn\") pod \"ovnkube-node-zlcg5\" (UID: \"736c6b1a-93dd-4f71-81b6-1c43fcf5f556\") " pod="openshift-ovn-kubernetes/ovnkube-node-zlcg5" Dec 13 06:49:26 crc kubenswrapper[4971]: I1213 06:49:26.213966 4971 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"tuning-conf-dir\" (UniqueName: \"kubernetes.io/host-path/de914d21-a688-401e-a66e-861beb6336d9-tuning-conf-dir\") pod \"multus-additional-cni-plugins-ncchz\" (UID: \"de914d21-a688-401e-a66e-861beb6336d9\") " pod="openshift-multus/multus-additional-cni-plugins-ncchz" Dec 13 06:49:26 crc kubenswrapper[4971]: I1213 06:49:26.213988 4971 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-lqmjj\" (UniqueName: \"kubernetes.io/projected/de914d21-a688-401e-a66e-861beb6336d9-kube-api-access-lqmjj\") pod \"multus-additional-cni-plugins-ncchz\" (UID: \"de914d21-a688-401e-a66e-861beb6336d9\") " pod="openshift-multus/multus-additional-cni-plugins-ncchz" Dec 13 06:49:26 crc kubenswrapper[4971]: I1213 06:49:26.214003 4971 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"os-release\" (UniqueName: \"kubernetes.io/host-path/dddf744b-2122-49ad-a980-105c09636e0f-os-release\") pod \"multus-wwvpl\" (UID: \"dddf744b-2122-49ad-a980-105c09636e0f\") " pod="openshift-multus/multus-wwvpl" Dec 13 06:49:26 crc kubenswrapper[4971]: I1213 06:49:26.214019 4971 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rootfs\" (UniqueName: \"kubernetes.io/host-path/e0c4af12-edda-41d7-9b44-f87396174ef2-rootfs\") pod \"machine-config-daemon-82xjz\" (UID: \"e0c4af12-edda-41d7-9b44-f87396174ef2\") " pod="openshift-machine-config-operator/machine-config-daemon-82xjz" Dec 13 06:49:26 crc kubenswrapper[4971]: I1213 06:49:26.214023 4971 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run-ovn\" (UniqueName: \"kubernetes.io/host-path/736c6b1a-93dd-4f71-81b6-1c43fcf5f556-run-ovn\") pod \"ovnkube-node-zlcg5\" (UID: \"736c6b1a-93dd-4f71-81b6-1c43fcf5f556\") " pod="openshift-ovn-kubernetes/ovnkube-node-zlcg5" Dec 13 06:49:26 crc kubenswrapper[4971]: I1213 06:49:26.214058 4971 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-kubelet\" (UniqueName: \"kubernetes.io/host-path/736c6b1a-93dd-4f71-81b6-1c43fcf5f556-host-kubelet\") pod \"ovnkube-node-zlcg5\" (UID: \"736c6b1a-93dd-4f71-81b6-1c43fcf5f556\") " pod="openshift-ovn-kubernetes/ovnkube-node-zlcg5" Dec 13 06:49:26 crc kubenswrapper[4971]: I1213 06:49:26.214032 4971 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-kubelet\" (UniqueName: \"kubernetes.io/host-path/736c6b1a-93dd-4f71-81b6-1c43fcf5f556-host-kubelet\") pod \"ovnkube-node-zlcg5\" (UID: \"736c6b1a-93dd-4f71-81b6-1c43fcf5f556\") " pod="openshift-ovn-kubernetes/ovnkube-node-zlcg5" Dec 13 06:49:26 crc kubenswrapper[4971]: I1213 06:49:26.214027 4971 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run-systemd\" (UniqueName: \"kubernetes.io/host-path/736c6b1a-93dd-4f71-81b6-1c43fcf5f556-run-systemd\") pod \"ovnkube-node-zlcg5\" (UID: \"736c6b1a-93dd-4f71-81b6-1c43fcf5f556\") " pod="openshift-ovn-kubernetes/ovnkube-node-zlcg5" Dec 13 06:49:26 crc kubenswrapper[4971]: I1213 06:49:26.214097 4971 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovn-node-metrics-cert\" (UniqueName: \"kubernetes.io/secret/736c6b1a-93dd-4f71-81b6-1c43fcf5f556-ovn-node-metrics-cert\") pod \"ovnkube-node-zlcg5\" (UID: \"736c6b1a-93dd-4f71-81b6-1c43fcf5f556\") " pod="openshift-ovn-kubernetes/ovnkube-node-zlcg5" Dec 13 06:49:26 crc kubenswrapper[4971]: I1213 06:49:26.214326 4971 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rootfs\" (UniqueName: \"kubernetes.io/host-path/e0c4af12-edda-41d7-9b44-f87396174ef2-rootfs\") pod \"machine-config-daemon-82xjz\" (UID: \"e0c4af12-edda-41d7-9b44-f87396174ef2\") " pod="openshift-machine-config-operator/machine-config-daemon-82xjz" Dec 13 06:49:26 crc kubenswrapper[4971]: I1213 06:49:26.214390 4971 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"os-release\" (UniqueName: \"kubernetes.io/host-path/dddf744b-2122-49ad-a980-105c09636e0f-os-release\") pod \"multus-wwvpl\" (UID: \"dddf744b-2122-49ad-a980-105c09636e0f\") " pod="openshift-multus/multus-wwvpl" Dec 13 06:49:26 crc kubenswrapper[4971]: I1213 06:49:26.214432 4971 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cni-binary-copy\" (UniqueName: \"kubernetes.io/configmap/dddf744b-2122-49ad-a980-105c09636e0f-cni-binary-copy\") pod \"multus-wwvpl\" (UID: \"dddf744b-2122-49ad-a980-105c09636e0f\") " pod="openshift-multus/multus-wwvpl" Dec 13 06:49:26 crc kubenswrapper[4971]: I1213 06:49:26.214557 4971 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"hostroot\" (UniqueName: \"kubernetes.io/host-path/dddf744b-2122-49ad-a980-105c09636e0f-hostroot\") pod \"multus-wwvpl\" (UID: \"dddf744b-2122-49ad-a980-105c09636e0f\") " pod="openshift-multus/multus-wwvpl" Dec 13 06:49:26 crc kubenswrapper[4971]: I1213 06:49:26.214744 4971 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"system-cni-dir\" (UniqueName: \"kubernetes.io/host-path/de914d21-a688-401e-a66e-861beb6336d9-system-cni-dir\") pod \"multus-additional-cni-plugins-ncchz\" (UID: \"de914d21-a688-401e-a66e-861beb6336d9\") " pod="openshift-multus/multus-additional-cni-plugins-ncchz" Dec 13 06:49:26 crc kubenswrapper[4971]: I1213 06:49:26.214768 4971 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-slash\" (UniqueName: \"kubernetes.io/host-path/736c6b1a-93dd-4f71-81b6-1c43fcf5f556-host-slash\") pod \"ovnkube-node-zlcg5\" (UID: \"736c6b1a-93dd-4f71-81b6-1c43fcf5f556\") " pod="openshift-ovn-kubernetes/ovnkube-node-zlcg5" Dec 13 06:49:26 crc kubenswrapper[4971]: I1213 06:49:26.214789 4971 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"systemd-units\" (UniqueName: \"kubernetes.io/host-path/736c6b1a-93dd-4f71-81b6-1c43fcf5f556-systemd-units\") pod \"ovnkube-node-zlcg5\" (UID: \"736c6b1a-93dd-4f71-81b6-1c43fcf5f556\") " pod="openshift-ovn-kubernetes/ovnkube-node-zlcg5" Dec 13 06:49:26 crc kubenswrapper[4971]: I1213 06:49:26.214806 4971 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/736c6b1a-93dd-4f71-81b6-1c43fcf5f556-env-overrides\") pod \"ovnkube-node-zlcg5\" (UID: \"736c6b1a-93dd-4f71-81b6-1c43fcf5f556\") " pod="openshift-ovn-kubernetes/ovnkube-node-zlcg5" Dec 13 06:49:26 crc kubenswrapper[4971]: I1213 06:49:26.214823 4971 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"os-release\" (UniqueName: \"kubernetes.io/host-path/de914d21-a688-401e-a66e-861beb6336d9-os-release\") pod \"multus-additional-cni-plugins-ncchz\" (UID: \"de914d21-a688-401e-a66e-861beb6336d9\") " pod="openshift-multus/multus-additional-cni-plugins-ncchz" Dec 13 06:49:26 crc kubenswrapper[4971]: I1213 06:49:26.214841 4971 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cni-binary-copy\" (UniqueName: \"kubernetes.io/configmap/de914d21-a688-401e-a66e-861beb6336d9-cni-binary-copy\") pod \"multus-additional-cni-plugins-ncchz\" (UID: \"de914d21-a688-401e-a66e-861beb6336d9\") " pod="openshift-multus/multus-additional-cni-plugins-ncchz" Dec 13 06:49:26 crc kubenswrapper[4971]: I1213 06:49:26.214860 4971 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"multus-daemon-config\" (UniqueName: \"kubernetes.io/configmap/dddf744b-2122-49ad-a980-105c09636e0f-multus-daemon-config\") pod \"multus-wwvpl\" (UID: \"dddf744b-2122-49ad-a980-105c09636e0f\") " pod="openshift-multus/multus-wwvpl" Dec 13 06:49:26 crc kubenswrapper[4971]: I1213 06:49:26.214878 4971 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cnxsq\" (UniqueName: \"kubernetes.io/projected/e0c4af12-edda-41d7-9b44-f87396174ef2-kube-api-access-cnxsq\") pod \"machine-config-daemon-82xjz\" (UID: \"e0c4af12-edda-41d7-9b44-f87396174ef2\") " pod="openshift-machine-config-operator/machine-config-daemon-82xjz" Dec 13 06:49:26 crc kubenswrapper[4971]: I1213 06:49:26.214896 4971 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"node-log\" (UniqueName: \"kubernetes.io/host-path/736c6b1a-93dd-4f71-81b6-1c43fcf5f556-node-log\") pod \"ovnkube-node-zlcg5\" (UID: \"736c6b1a-93dd-4f71-81b6-1c43fcf5f556\") " pod="openshift-ovn-kubernetes/ovnkube-node-zlcg5" Dec 13 06:49:26 crc kubenswrapper[4971]: I1213 06:49:26.214917 4971 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovnkube-script-lib\" (UniqueName: \"kubernetes.io/configmap/736c6b1a-93dd-4f71-81b6-1c43fcf5f556-ovnkube-script-lib\") pod \"ovnkube-node-zlcg5\" (UID: \"736c6b1a-93dd-4f71-81b6-1c43fcf5f556\") " pod="openshift-ovn-kubernetes/ovnkube-node-zlcg5" Dec 13 06:49:26 crc kubenswrapper[4971]: I1213 06:49:26.214937 4971 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cnibin\" (UniqueName: \"kubernetes.io/host-path/de914d21-a688-401e-a66e-861beb6336d9-cnibin\") pod \"multus-additional-cni-plugins-ncchz\" (UID: \"de914d21-a688-401e-a66e-861beb6336d9\") " pod="openshift-multus/multus-additional-cni-plugins-ncchz" Dec 13 06:49:26 crc kubenswrapper[4971]: I1213 06:49:26.214955 4971 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cni-sysctl-allowlist\" (UniqueName: \"kubernetes.io/configmap/de914d21-a688-401e-a66e-861beb6336d9-cni-sysctl-allowlist\") pod \"multus-additional-cni-plugins-ncchz\" (UID: \"de914d21-a688-401e-a66e-861beb6336d9\") " pod="openshift-multus/multus-additional-cni-plugins-ncchz" Dec 13 06:49:26 crc kubenswrapper[4971]: I1213 06:49:26.214963 4971 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"system-cni-dir\" (UniqueName: \"kubernetes.io/host-path/de914d21-a688-401e-a66e-861beb6336d9-system-cni-dir\") pod \"multus-additional-cni-plugins-ncchz\" (UID: \"de914d21-a688-401e-a66e-861beb6336d9\") " pod="openshift-multus/multus-additional-cni-plugins-ncchz" Dec 13 06:49:26 crc kubenswrapper[4971]: I1213 06:49:26.214976 4971 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/e0c4af12-edda-41d7-9b44-f87396174ef2-proxy-tls\") pod \"machine-config-daemon-82xjz\" (UID: \"e0c4af12-edda-41d7-9b44-f87396174ef2\") " pod="openshift-machine-config-operator/machine-config-daemon-82xjz" Dec 13 06:49:26 crc kubenswrapper[4971]: I1213 06:49:26.214991 4971 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"hostroot\" (UniqueName: \"kubernetes.io/host-path/dddf744b-2122-49ad-a980-105c09636e0f-hostroot\") pod \"multus-wwvpl\" (UID: \"dddf744b-2122-49ad-a980-105c09636e0f\") " pod="openshift-multus/multus-wwvpl" Dec 13 06:49:26 crc kubenswrapper[4971]: I1213 06:49:26.215023 4971 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-cni-netd\" (UniqueName: \"kubernetes.io/host-path/736c6b1a-93dd-4f71-81b6-1c43fcf5f556-host-cni-netd\") pod \"ovnkube-node-zlcg5\" (UID: \"736c6b1a-93dd-4f71-81b6-1c43fcf5f556\") " pod="openshift-ovn-kubernetes/ovnkube-node-zlcg5" Dec 13 06:49:26 crc kubenswrapper[4971]: I1213 06:49:26.215045 4971 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"multus-socket-dir-parent\" (UniqueName: \"kubernetes.io/host-path/dddf744b-2122-49ad-a980-105c09636e0f-multus-socket-dir-parent\") pod \"multus-wwvpl\" (UID: \"dddf744b-2122-49ad-a980-105c09636e0f\") " pod="openshift-multus/multus-wwvpl" Dec 13 06:49:26 crc kubenswrapper[4971]: I1213 06:49:26.215084 4971 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-cni-netd\" (UniqueName: \"kubernetes.io/host-path/736c6b1a-93dd-4f71-81b6-1c43fcf5f556-host-cni-netd\") pod \"ovnkube-node-zlcg5\" (UID: \"736c6b1a-93dd-4f71-81b6-1c43fcf5f556\") " pod="openshift-ovn-kubernetes/ovnkube-node-zlcg5" Dec 13 06:49:26 crc kubenswrapper[4971]: I1213 06:49:26.215111 4971 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"tuning-conf-dir\" (UniqueName: \"kubernetes.io/host-path/de914d21-a688-401e-a66e-861beb6336d9-tuning-conf-dir\") pod \"multus-additional-cni-plugins-ncchz\" (UID: \"de914d21-a688-401e-a66e-861beb6336d9\") " pod="openshift-multus/multus-additional-cni-plugins-ncchz" Dec 13 06:49:26 crc kubenswrapper[4971]: I1213 06:49:26.215173 4971 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"multus-socket-dir-parent\" (UniqueName: \"kubernetes.io/host-path/dddf744b-2122-49ad-a980-105c09636e0f-multus-socket-dir-parent\") pod \"multus-wwvpl\" (UID: \"dddf744b-2122-49ad-a980-105c09636e0f\") " pod="openshift-multus/multus-wwvpl" Dec 13 06:49:26 crc kubenswrapper[4971]: I1213 06:49:26.215237 4971 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-slash\" (UniqueName: \"kubernetes.io/host-path/736c6b1a-93dd-4f71-81b6-1c43fcf5f556-host-slash\") pod \"ovnkube-node-zlcg5\" (UID: \"736c6b1a-93dd-4f71-81b6-1c43fcf5f556\") " pod="openshift-ovn-kubernetes/ovnkube-node-zlcg5" Dec 13 06:49:26 crc kubenswrapper[4971]: I1213 06:49:26.215262 4971 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cni-binary-copy\" (UniqueName: \"kubernetes.io/configmap/dddf744b-2122-49ad-a980-105c09636e0f-cni-binary-copy\") pod \"multus-wwvpl\" (UID: \"dddf744b-2122-49ad-a980-105c09636e0f\") " pod="openshift-multus/multus-wwvpl" Dec 13 06:49:26 crc kubenswrapper[4971]: I1213 06:49:26.215273 4971 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-run-multus-certs\" (UniqueName: \"kubernetes.io/host-path/dddf744b-2122-49ad-a980-105c09636e0f-host-run-multus-certs\") pod \"multus-wwvpl\" (UID: \"dddf744b-2122-49ad-a980-105c09636e0f\") " pod="openshift-multus/multus-wwvpl" Dec 13 06:49:26 crc kubenswrapper[4971]: I1213 06:49:26.215301 4971 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-run-multus-certs\" (UniqueName: \"kubernetes.io/host-path/dddf744b-2122-49ad-a980-105c09636e0f-host-run-multus-certs\") pod \"multus-wwvpl\" (UID: \"dddf744b-2122-49ad-a980-105c09636e0f\") " pod="openshift-multus/multus-wwvpl" Dec 13 06:49:26 crc kubenswrapper[4971]: I1213 06:49:26.215333 4971 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cnibin\" (UniqueName: \"kubernetes.io/host-path/de914d21-a688-401e-a66e-861beb6336d9-cnibin\") pod \"multus-additional-cni-plugins-ncchz\" (UID: \"de914d21-a688-401e-a66e-861beb6336d9\") " pod="openshift-multus/multus-additional-cni-plugins-ncchz" Dec 13 06:49:26 crc kubenswrapper[4971]: I1213 06:49:26.215335 4971 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"multus-conf-dir\" (UniqueName: \"kubernetes.io/host-path/dddf744b-2122-49ad-a980-105c09636e0f-multus-conf-dir\") pod \"multus-wwvpl\" (UID: \"dddf744b-2122-49ad-a980-105c09636e0f\") " pod="openshift-multus/multus-wwvpl" Dec 13 06:49:26 crc kubenswrapper[4971]: I1213 06:49:26.215351 4971 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"multus-conf-dir\" (UniqueName: \"kubernetes.io/host-path/dddf744b-2122-49ad-a980-105c09636e0f-multus-conf-dir\") pod \"multus-wwvpl\" (UID: \"dddf744b-2122-49ad-a980-105c09636e0f\") " pod="openshift-multus/multus-wwvpl" Dec 13 06:49:26 crc kubenswrapper[4971]: I1213 06:49:26.215381 4971 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"mcd-auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/e0c4af12-edda-41d7-9b44-f87396174ef2-mcd-auth-proxy-config\") pod \"machine-config-daemon-82xjz\" (UID: \"e0c4af12-edda-41d7-9b44-f87396174ef2\") " pod="openshift-machine-config-operator/machine-config-daemon-82xjz" Dec 13 06:49:26 crc kubenswrapper[4971]: I1213 06:49:26.215398 4971 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run-openvswitch\" (UniqueName: \"kubernetes.io/host-path/736c6b1a-93dd-4f71-81b6-1c43fcf5f556-run-openvswitch\") pod \"ovnkube-node-zlcg5\" (UID: \"736c6b1a-93dd-4f71-81b6-1c43fcf5f556\") " pod="openshift-ovn-kubernetes/ovnkube-node-zlcg5" Dec 13 06:49:26 crc kubenswrapper[4971]: I1213 06:49:26.215414 4971 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-run-netns\" (UniqueName: \"kubernetes.io/host-path/dddf744b-2122-49ad-a980-105c09636e0f-host-run-netns\") pod \"multus-wwvpl\" (UID: \"dddf744b-2122-49ad-a980-105c09636e0f\") " pod="openshift-multus/multus-wwvpl" Dec 13 06:49:26 crc kubenswrapper[4971]: I1213 06:49:26.215448 4971 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run-openvswitch\" (UniqueName: \"kubernetes.io/host-path/736c6b1a-93dd-4f71-81b6-1c43fcf5f556-run-openvswitch\") pod \"ovnkube-node-zlcg5\" (UID: \"736c6b1a-93dd-4f71-81b6-1c43fcf5f556\") " pod="openshift-ovn-kubernetes/ovnkube-node-zlcg5" Dec 13 06:49:26 crc kubenswrapper[4971]: I1213 06:49:26.215461 4971 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"os-release\" (UniqueName: \"kubernetes.io/host-path/de914d21-a688-401e-a66e-861beb6336d9-os-release\") pod \"multus-additional-cni-plugins-ncchz\" (UID: \"de914d21-a688-401e-a66e-861beb6336d9\") " pod="openshift-multus/multus-additional-cni-plugins-ncchz" Dec 13 06:49:26 crc kubenswrapper[4971]: I1213 06:49:26.215476 4971 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"multus-cni-dir\" (UniqueName: \"kubernetes.io/host-path/dddf744b-2122-49ad-a980-105c09636e0f-multus-cni-dir\") pod \"multus-wwvpl\" (UID: \"dddf744b-2122-49ad-a980-105c09636e0f\") " pod="openshift-multus/multus-wwvpl" Dec 13 06:49:26 crc kubenswrapper[4971]: I1213 06:49:26.215502 4971 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"system-cni-dir\" (UniqueName: \"kubernetes.io/host-path/dddf744b-2122-49ad-a980-105c09636e0f-system-cni-dir\") pod \"multus-wwvpl\" (UID: \"dddf744b-2122-49ad-a980-105c09636e0f\") " pod="openshift-multus/multus-wwvpl" Dec 13 06:49:26 crc kubenswrapper[4971]: I1213 06:49:26.215553 4971 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"multus-cni-dir\" (UniqueName: \"kubernetes.io/host-path/dddf744b-2122-49ad-a980-105c09636e0f-multus-cni-dir\") pod \"multus-wwvpl\" (UID: \"dddf744b-2122-49ad-a980-105c09636e0f\") " pod="openshift-multus/multus-wwvpl" Dec 13 06:49:26 crc kubenswrapper[4971]: I1213 06:49:26.215569 4971 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-run-netns\" (UniqueName: \"kubernetes.io/host-path/dddf744b-2122-49ad-a980-105c09636e0f-host-run-netns\") pod \"multus-wwvpl\" (UID: \"dddf744b-2122-49ad-a980-105c09636e0f\") " pod="openshift-multus/multus-wwvpl" Dec 13 06:49:26 crc kubenswrapper[4971]: I1213 06:49:26.215600 4971 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"multus-daemon-config\" (UniqueName: \"kubernetes.io/configmap/dddf744b-2122-49ad-a980-105c09636e0f-multus-daemon-config\") pod \"multus-wwvpl\" (UID: \"dddf744b-2122-49ad-a980-105c09636e0f\") " pod="openshift-multus/multus-wwvpl" Dec 13 06:49:26 crc kubenswrapper[4971]: I1213 06:49:26.215641 4971 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cni-binary-copy\" (UniqueName: \"kubernetes.io/configmap/de914d21-a688-401e-a66e-861beb6336d9-cni-binary-copy\") pod \"multus-additional-cni-plugins-ncchz\" (UID: \"de914d21-a688-401e-a66e-861beb6336d9\") " pod="openshift-multus/multus-additional-cni-plugins-ncchz" Dec 13 06:49:26 crc kubenswrapper[4971]: I1213 06:49:26.215668 4971 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"system-cni-dir\" (UniqueName: \"kubernetes.io/host-path/dddf744b-2122-49ad-a980-105c09636e0f-system-cni-dir\") pod \"multus-wwvpl\" (UID: \"dddf744b-2122-49ad-a980-105c09636e0f\") " pod="openshift-multus/multus-wwvpl" Dec 13 06:49:26 crc kubenswrapper[4971]: I1213 06:49:26.215700 4971 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"node-log\" (UniqueName: \"kubernetes.io/host-path/736c6b1a-93dd-4f71-81b6-1c43fcf5f556-node-log\") pod \"ovnkube-node-zlcg5\" (UID: \"736c6b1a-93dd-4f71-81b6-1c43fcf5f556\") " pod="openshift-ovn-kubernetes/ovnkube-node-zlcg5" Dec 13 06:49:26 crc kubenswrapper[4971]: I1213 06:49:26.215718 4971 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"systemd-units\" (UniqueName: \"kubernetes.io/host-path/736c6b1a-93dd-4f71-81b6-1c43fcf5f556-systemd-units\") pod \"ovnkube-node-zlcg5\" (UID: \"736c6b1a-93dd-4f71-81b6-1c43fcf5f556\") " pod="openshift-ovn-kubernetes/ovnkube-node-zlcg5" Dec 13 06:49:26 crc kubenswrapper[4971]: I1213 06:49:26.215719 4971 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-var-lib-cni-networks-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/736c6b1a-93dd-4f71-81b6-1c43fcf5f556-host-var-lib-cni-networks-ovn-kubernetes\") pod \"ovnkube-node-zlcg5\" (UID: \"736c6b1a-93dd-4f71-81b6-1c43fcf5f556\") " pod="openshift-ovn-kubernetes/ovnkube-node-zlcg5" Dec 13 06:49:26 crc kubenswrapper[4971]: I1213 06:49:26.215741 4971 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/736c6b1a-93dd-4f71-81b6-1c43fcf5f556-ovnkube-config\") pod \"ovnkube-node-zlcg5\" (UID: \"736c6b1a-93dd-4f71-81b6-1c43fcf5f556\") " pod="openshift-ovn-kubernetes/ovnkube-node-zlcg5" Dec 13 06:49:26 crc kubenswrapper[4971]: I1213 06:49:26.215775 4971 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-var-lib-cni-networks-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/736c6b1a-93dd-4f71-81b6-1c43fcf5f556-host-var-lib-cni-networks-ovn-kubernetes\") pod \"ovnkube-node-zlcg5\" (UID: \"736c6b1a-93dd-4f71-81b6-1c43fcf5f556\") " pod="openshift-ovn-kubernetes/ovnkube-node-zlcg5" Dec 13 06:49:26 crc kubenswrapper[4971]: I1213 06:49:26.215779 4971 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-run-netns\" (UniqueName: \"kubernetes.io/host-path/736c6b1a-93dd-4f71-81b6-1c43fcf5f556-host-run-netns\") pod \"ovnkube-node-zlcg5\" (UID: \"736c6b1a-93dd-4f71-81b6-1c43fcf5f556\") " pod="openshift-ovn-kubernetes/ovnkube-node-zlcg5" Dec 13 06:49:26 crc kubenswrapper[4971]: I1213 06:49:26.215803 4971 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-lib-openvswitch\" (UniqueName: \"kubernetes.io/host-path/736c6b1a-93dd-4f71-81b6-1c43fcf5f556-var-lib-openvswitch\") pod \"ovnkube-node-zlcg5\" (UID: \"736c6b1a-93dd-4f71-81b6-1c43fcf5f556\") " pod="openshift-ovn-kubernetes/ovnkube-node-zlcg5" Dec 13 06:49:26 crc kubenswrapper[4971]: I1213 06:49:26.215873 4971 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-lib-openvswitch\" (UniqueName: \"kubernetes.io/host-path/736c6b1a-93dd-4f71-81b6-1c43fcf5f556-var-lib-openvswitch\") pod \"ovnkube-node-zlcg5\" (UID: \"736c6b1a-93dd-4f71-81b6-1c43fcf5f556\") " pod="openshift-ovn-kubernetes/ovnkube-node-zlcg5" Dec 13 06:49:26 crc kubenswrapper[4971]: I1213 06:49:26.215881 4971 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-vzmsr\" (UniqueName: \"kubernetes.io/projected/736c6b1a-93dd-4f71-81b6-1c43fcf5f556-kube-api-access-vzmsr\") pod \"ovnkube-node-zlcg5\" (UID: \"736c6b1a-93dd-4f71-81b6-1c43fcf5f556\") " pod="openshift-ovn-kubernetes/ovnkube-node-zlcg5" Dec 13 06:49:26 crc kubenswrapper[4971]: I1213 06:49:26.215913 4971 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-run-netns\" (UniqueName: \"kubernetes.io/host-path/736c6b1a-93dd-4f71-81b6-1c43fcf5f556-host-run-netns\") pod \"ovnkube-node-zlcg5\" (UID: \"736c6b1a-93dd-4f71-81b6-1c43fcf5f556\") " pod="openshift-ovn-kubernetes/ovnkube-node-zlcg5" Dec 13 06:49:26 crc kubenswrapper[4971]: I1213 06:49:26.215946 4971 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cnibin\" (UniqueName: \"kubernetes.io/host-path/dddf744b-2122-49ad-a980-105c09636e0f-cnibin\") pod \"multus-wwvpl\" (UID: \"dddf744b-2122-49ad-a980-105c09636e0f\") " pod="openshift-multus/multus-wwvpl" Dec 13 06:49:26 crc kubenswrapper[4971]: I1213 06:49:26.215971 4971 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-var-lib-cni-bin\" (UniqueName: \"kubernetes.io/host-path/dddf744b-2122-49ad-a980-105c09636e0f-host-var-lib-cni-bin\") pod \"multus-wwvpl\" (UID: \"dddf744b-2122-49ad-a980-105c09636e0f\") " pod="openshift-multus/multus-wwvpl" Dec 13 06:49:26 crc kubenswrapper[4971]: I1213 06:49:26.216020 4971 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-var-lib-cni-multus\" (UniqueName: \"kubernetes.io/host-path/dddf744b-2122-49ad-a980-105c09636e0f-host-var-lib-cni-multus\") pod \"multus-wwvpl\" (UID: \"dddf744b-2122-49ad-a980-105c09636e0f\") " pod="openshift-multus/multus-wwvpl" Dec 13 06:49:26 crc kubenswrapper[4971]: I1213 06:49:26.216054 4971 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-czz9v\" (UniqueName: \"kubernetes.io/projected/dddf744b-2122-49ad-a980-105c09636e0f-kube-api-access-czz9v\") pod \"multus-wwvpl\" (UID: \"dddf744b-2122-49ad-a980-105c09636e0f\") " pod="openshift-multus/multus-wwvpl" Dec 13 06:49:26 crc kubenswrapper[4971]: I1213 06:49:26.216073 4971 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-openvswitch\" (UniqueName: \"kubernetes.io/host-path/736c6b1a-93dd-4f71-81b6-1c43fcf5f556-etc-openvswitch\") pod \"ovnkube-node-zlcg5\" (UID: \"736c6b1a-93dd-4f71-81b6-1c43fcf5f556\") " pod="openshift-ovn-kubernetes/ovnkube-node-zlcg5" Dec 13 06:49:26 crc kubenswrapper[4971]: I1213 06:49:26.216094 4971 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-run-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/736c6b1a-93dd-4f71-81b6-1c43fcf5f556-host-run-ovn-kubernetes\") pod \"ovnkube-node-zlcg5\" (UID: \"736c6b1a-93dd-4f71-81b6-1c43fcf5f556\") " pod="openshift-ovn-kubernetes/ovnkube-node-zlcg5" Dec 13 06:49:26 crc kubenswrapper[4971]: I1213 06:49:26.216116 4971 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-kubernetes\" (UniqueName: \"kubernetes.io/host-path/dddf744b-2122-49ad-a980-105c09636e0f-etc-kubernetes\") pod \"multus-wwvpl\" (UID: \"dddf744b-2122-49ad-a980-105c09636e0f\") " pod="openshift-multus/multus-wwvpl" Dec 13 06:49:26 crc kubenswrapper[4971]: I1213 06:49:26.216147 4971 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"log-socket\" (UniqueName: \"kubernetes.io/host-path/736c6b1a-93dd-4f71-81b6-1c43fcf5f556-log-socket\") pod \"ovnkube-node-zlcg5\" (UID: \"736c6b1a-93dd-4f71-81b6-1c43fcf5f556\") " pod="openshift-ovn-kubernetes/ovnkube-node-zlcg5" Dec 13 06:49:26 crc kubenswrapper[4971]: I1213 06:49:26.216169 4971 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-cni-bin\" (UniqueName: \"kubernetes.io/host-path/736c6b1a-93dd-4f71-81b6-1c43fcf5f556-host-cni-bin\") pod \"ovnkube-node-zlcg5\" (UID: \"736c6b1a-93dd-4f71-81b6-1c43fcf5f556\") " pod="openshift-ovn-kubernetes/ovnkube-node-zlcg5" Dec 13 06:49:26 crc kubenswrapper[4971]: I1213 06:49:26.216189 4971 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-run-k8s-cni-cncf-io\" (UniqueName: \"kubernetes.io/host-path/dddf744b-2122-49ad-a980-105c09636e0f-host-run-k8s-cni-cncf-io\") pod \"multus-wwvpl\" (UID: \"dddf744b-2122-49ad-a980-105c09636e0f\") " pod="openshift-multus/multus-wwvpl" Dec 13 06:49:26 crc kubenswrapper[4971]: I1213 06:49:26.216209 4971 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-run-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/736c6b1a-93dd-4f71-81b6-1c43fcf5f556-host-run-ovn-kubernetes\") pod \"ovnkube-node-zlcg5\" (UID: \"736c6b1a-93dd-4f71-81b6-1c43fcf5f556\") " pod="openshift-ovn-kubernetes/ovnkube-node-zlcg5" Dec 13 06:49:26 crc kubenswrapper[4971]: I1213 06:49:26.216211 4971 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-var-lib-kubelet\" (UniqueName: \"kubernetes.io/host-path/dddf744b-2122-49ad-a980-105c09636e0f-host-var-lib-kubelet\") pod \"multus-wwvpl\" (UID: \"dddf744b-2122-49ad-a980-105c09636e0f\") " pod="openshift-multus/multus-wwvpl" Dec 13 06:49:26 crc kubenswrapper[4971]: I1213 06:49:26.216245 4971 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-var-lib-kubelet\" (UniqueName: \"kubernetes.io/host-path/dddf744b-2122-49ad-a980-105c09636e0f-host-var-lib-kubelet\") pod \"multus-wwvpl\" (UID: \"dddf744b-2122-49ad-a980-105c09636e0f\") " pod="openshift-multus/multus-wwvpl" Dec 13 06:49:26 crc kubenswrapper[4971]: I1213 06:49:26.216250 4971 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-kubernetes\" (UniqueName: \"kubernetes.io/host-path/dddf744b-2122-49ad-a980-105c09636e0f-etc-kubernetes\") pod \"multus-wwvpl\" (UID: \"dddf744b-2122-49ad-a980-105c09636e0f\") " pod="openshift-multus/multus-wwvpl" Dec 13 06:49:26 crc kubenswrapper[4971]: I1213 06:49:26.216284 4971 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"log-socket\" (UniqueName: \"kubernetes.io/host-path/736c6b1a-93dd-4f71-81b6-1c43fcf5f556-log-socket\") pod \"ovnkube-node-zlcg5\" (UID: \"736c6b1a-93dd-4f71-81b6-1c43fcf5f556\") " pod="openshift-ovn-kubernetes/ovnkube-node-zlcg5" Dec 13 06:49:26 crc kubenswrapper[4971]: I1213 06:49:26.216322 4971 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-cni-bin\" (UniqueName: \"kubernetes.io/host-path/736c6b1a-93dd-4f71-81b6-1c43fcf5f556-host-cni-bin\") pod \"ovnkube-node-zlcg5\" (UID: \"736c6b1a-93dd-4f71-81b6-1c43fcf5f556\") " pod="openshift-ovn-kubernetes/ovnkube-node-zlcg5" Dec 13 06:49:26 crc kubenswrapper[4971]: I1213 06:49:26.216330 4971 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/736c6b1a-93dd-4f71-81b6-1c43fcf5f556-ovnkube-config\") pod \"ovnkube-node-zlcg5\" (UID: \"736c6b1a-93dd-4f71-81b6-1c43fcf5f556\") " pod="openshift-ovn-kubernetes/ovnkube-node-zlcg5" Dec 13 06:49:26 crc kubenswrapper[4971]: I1213 06:49:26.216349 4971 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-run-k8s-cni-cncf-io\" (UniqueName: \"kubernetes.io/host-path/dddf744b-2122-49ad-a980-105c09636e0f-host-run-k8s-cni-cncf-io\") pod \"multus-wwvpl\" (UID: \"dddf744b-2122-49ad-a980-105c09636e0f\") " pod="openshift-multus/multus-wwvpl" Dec 13 06:49:26 crc kubenswrapper[4971]: I1213 06:49:26.216381 4971 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-openvswitch\" (UniqueName: \"kubernetes.io/host-path/736c6b1a-93dd-4f71-81b6-1c43fcf5f556-etc-openvswitch\") pod \"ovnkube-node-zlcg5\" (UID: \"736c6b1a-93dd-4f71-81b6-1c43fcf5f556\") " pod="openshift-ovn-kubernetes/ovnkube-node-zlcg5" Dec 13 06:49:26 crc kubenswrapper[4971]: I1213 06:49:26.216438 4971 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-var-lib-cni-multus\" (UniqueName: \"kubernetes.io/host-path/dddf744b-2122-49ad-a980-105c09636e0f-host-var-lib-cni-multus\") pod \"multus-wwvpl\" (UID: \"dddf744b-2122-49ad-a980-105c09636e0f\") " pod="openshift-multus/multus-wwvpl" Dec 13 06:49:26 crc kubenswrapper[4971]: I1213 06:49:26.216477 4971 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovnkube-script-lib\" (UniqueName: \"kubernetes.io/configmap/736c6b1a-93dd-4f71-81b6-1c43fcf5f556-ovnkube-script-lib\") pod \"ovnkube-node-zlcg5\" (UID: \"736c6b1a-93dd-4f71-81b6-1c43fcf5f556\") " pod="openshift-ovn-kubernetes/ovnkube-node-zlcg5" Dec 13 06:49:26 crc kubenswrapper[4971]: I1213 06:49:26.216489 4971 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/736c6b1a-93dd-4f71-81b6-1c43fcf5f556-env-overrides\") pod \"ovnkube-node-zlcg5\" (UID: \"736c6b1a-93dd-4f71-81b6-1c43fcf5f556\") " pod="openshift-ovn-kubernetes/ovnkube-node-zlcg5" Dec 13 06:49:26 crc kubenswrapper[4971]: I1213 06:49:26.216505 4971 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cnibin\" (UniqueName: \"kubernetes.io/host-path/dddf744b-2122-49ad-a980-105c09636e0f-cnibin\") pod \"multus-wwvpl\" (UID: \"dddf744b-2122-49ad-a980-105c09636e0f\") " pod="openshift-multus/multus-wwvpl" Dec 13 06:49:26 crc kubenswrapper[4971]: I1213 06:49:26.216535 4971 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-var-lib-cni-bin\" (UniqueName: \"kubernetes.io/host-path/dddf744b-2122-49ad-a980-105c09636e0f-host-var-lib-cni-bin\") pod \"multus-wwvpl\" (UID: \"dddf744b-2122-49ad-a980-105c09636e0f\") " pod="openshift-multus/multus-wwvpl" Dec 13 06:49:26 crc kubenswrapper[4971]: I1213 06:49:26.216715 4971 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cni-sysctl-allowlist\" (UniqueName: \"kubernetes.io/configmap/de914d21-a688-401e-a66e-861beb6336d9-cni-sysctl-allowlist\") pod \"multus-additional-cni-plugins-ncchz\" (UID: \"de914d21-a688-401e-a66e-861beb6336d9\") " pod="openshift-multus/multus-additional-cni-plugins-ncchz" Dec 13 06:49:26 crc kubenswrapper[4971]: I1213 06:49:26.217075 4971 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"mcd-auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/e0c4af12-edda-41d7-9b44-f87396174ef2-mcd-auth-proxy-config\") pod \"machine-config-daemon-82xjz\" (UID: \"e0c4af12-edda-41d7-9b44-f87396174ef2\") " pod="openshift-machine-config-operator/machine-config-daemon-82xjz" Dec 13 06:49:26 crc kubenswrapper[4971]: I1213 06:49:26.217940 4971 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovn-node-metrics-cert\" (UniqueName: \"kubernetes.io/secret/736c6b1a-93dd-4f71-81b6-1c43fcf5f556-ovn-node-metrics-cert\") pod \"ovnkube-node-zlcg5\" (UID: \"736c6b1a-93dd-4f71-81b6-1c43fcf5f556\") " pod="openshift-ovn-kubernetes/ovnkube-node-zlcg5" Dec 13 06:49:26 crc kubenswrapper[4971]: I1213 06:49:26.228467 4971 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/e0c4af12-edda-41d7-9b44-f87396174ef2-proxy-tls\") pod \"machine-config-daemon-82xjz\" (UID: \"e0c4af12-edda-41d7-9b44-f87396174ef2\") " pod="openshift-machine-config-operator/machine-config-daemon-82xjz" Dec 13 06:49:26 crc kubenswrapper[4971]: I1213 06:49:26.241866 4971 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-5rzh5" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"6bf4591d-2483-4132-9672-d0fd6c995968\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T06:49:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T06:49:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T06:49:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T06:49:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://50e2cb14e4ffb603acb06e2281615a99f06182d345c4e70f2446fa2077033bbb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-13T06:49:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dfxqn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-13T06:49:25Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-5rzh5\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-13T06:49:26Z is after 2025-08-24T17:21:41Z" Dec 13 06:49:26 crc kubenswrapper[4971]: I1213 06:49:26.242307 4971 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-cnxsq\" (UniqueName: \"kubernetes.io/projected/e0c4af12-edda-41d7-9b44-f87396174ef2-kube-api-access-cnxsq\") pod \"machine-config-daemon-82xjz\" (UID: \"e0c4af12-edda-41d7-9b44-f87396174ef2\") " pod="openshift-machine-config-operator/machine-config-daemon-82xjz" Dec 13 06:49:26 crc kubenswrapper[4971]: I1213 06:49:26.242965 4971 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-vzmsr\" (UniqueName: \"kubernetes.io/projected/736c6b1a-93dd-4f71-81b6-1c43fcf5f556-kube-api-access-vzmsr\") pod \"ovnkube-node-zlcg5\" (UID: \"736c6b1a-93dd-4f71-81b6-1c43fcf5f556\") " pod="openshift-ovn-kubernetes/ovnkube-node-zlcg5" Dec 13 06:49:26 crc kubenswrapper[4971]: I1213 06:49:26.243079 4971 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-lqmjj\" (UniqueName: \"kubernetes.io/projected/de914d21-a688-401e-a66e-861beb6336d9-kube-api-access-lqmjj\") pod \"multus-additional-cni-plugins-ncchz\" (UID: \"de914d21-a688-401e-a66e-861beb6336d9\") " pod="openshift-multus/multus-additional-cni-plugins-ncchz" Dec 13 06:49:26 crc kubenswrapper[4971]: I1213 06:49:26.250096 4971 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-czz9v\" (UniqueName: \"kubernetes.io/projected/dddf744b-2122-49ad-a980-105c09636e0f-kube-api-access-czz9v\") pod \"multus-wwvpl\" (UID: \"dddf744b-2122-49ad-a980-105c09636e0f\") " pod="openshift-multus/multus-wwvpl" Dec 13 06:49:26 crc kubenswrapper[4971]: I1213 06:49:26.264795 4971 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3ad3268a-a982-428c-9c06-4aaeddd372d9\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T06:49:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T06:49:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T06:49:03Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T06:49:03Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T06:49:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3716afdf2bc693be80a09fc9f7727d334f435caf1d4c8ec4811b3a1f1652cdfe\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-13T06:49:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://30df3abf8c79db8af52c6a918f73e1ce410ddafcf341636424c267e1eeb195a7\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-13T06:49:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://ccdf65f6d2b1eb3647a03ba37c305c32e34addfdf79ff937c8dddaa3e5d5b5c6\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-13T06:49:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://770b73dd0c2b6f38744e2c1cd6aa186a69bb87d1cd015a603ad2e33be06a8f6b\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://b3c1bab3d53f975b2e24582ce2771bc26d600e0dd7c30339dcfd18a071162ad4\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-12-13T06:49:23Z\\\",\\\"message\\\":\\\"file observer\\\\nW1213 06:49:23.283403 1 builder.go:272] unable to get owner reference (falling back to namespace): pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\nI1213 06:49:23.284336 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1213 06:49:23.286978 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-981773671/tls.crt::/tmp/serving-cert-981773671/tls.key\\\\\\\"\\\\nI1213 06:49:23.775072 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1213 06:49:23.778615 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1213 06:49:23.778645 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1213 06:49:23.778680 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1213 06:49:23.778691 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1213 06:49:23.785058 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1213 06:49:23.786150 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1213 06:49:23.786196 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1213 06:49:23.786223 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1213 06:49:23.786246 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1213 06:49:23.786269 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1213 06:49:23.786291 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI1213 06:49:23.785608 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF1213 06:49:23.787563 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-13T06:49:08Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-13T06:49:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://9acd07066deed483bd7fa61ee45fda09539d110717a134a56827f871186f92fc\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-13T06:49:07Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://38c91cfbaecfdbd3e3b663422a91ab4294610afb6bbf9ebdb77db24e2b76c682\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://38c91cfbaecfdbd3e3b663422a91ab4294610afb6bbf9ebdb77db24e2b76c682\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-13T06:49:05Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-13T06:49:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-13T06:49:03Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-13T06:49:26Z is after 2025-08-24T17:21:41Z" Dec 13 06:49:26 crc kubenswrapper[4971]: I1213 06:49:26.276677 4971 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-13T06:49:23Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-13T06:49:23Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-13T06:49:26Z is after 2025-08-24T17:21:41Z" Dec 13 06:49:26 crc kubenswrapper[4971]: I1213 06:49:26.288387 4971 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-13T06:49:23Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-13T06:49:23Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-13T06:49:26Z is after 2025-08-24T17:21:41Z" Dec 13 06:49:26 crc kubenswrapper[4971]: I1213 06:49:26.301649 4971 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-13T06:49:24Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-13T06:49:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-13T06:49:24Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d0b2ecf0d5b22f23217f42f259ae7ffb6359555b22c30eaebf667eaf254b3bab\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-13T06:49:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-13T06:49:26Z is after 2025-08-24T17:21:41Z" Dec 13 06:49:26 crc kubenswrapper[4971]: I1213 06:49:26.305984 4971 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-machine-config-operator/machine-config-daemon-82xjz" Dec 13 06:49:26 crc kubenswrapper[4971]: I1213 06:49:26.314029 4971 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/multus-wwvpl" Dec 13 06:49:26 crc kubenswrapper[4971]: W1213 06:49:26.315682 4971 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pode0c4af12_edda_41d7_9b44_f87396174ef2.slice/crio-81902d3c633547cf11f7a8d00b514f0e8ffdba48c90ffeea5f81c919b933eb52 WatchSource:0}: Error finding container 81902d3c633547cf11f7a8d00b514f0e8ffdba48c90ffeea5f81c919b933eb52: Status 404 returned error can't find the container with id 81902d3c633547cf11f7a8d00b514f0e8ffdba48c90ffeea5f81c919b933eb52 Dec 13 06:49:26 crc kubenswrapper[4971]: I1213 06:49:26.319600 4971 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-ovn-kubernetes/ovnkube-node-zlcg5" Dec 13 06:49:26 crc kubenswrapper[4971]: I1213 06:49:26.323127 4971 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-13T06:49:23Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-13T06:49:23Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-13T06:49:26Z is after 2025-08-24T17:21:41Z" Dec 13 06:49:26 crc kubenswrapper[4971]: I1213 06:49:26.324975 4971 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/multus-additional-cni-plugins-ncchz" Dec 13 06:49:26 crc kubenswrapper[4971]: I1213 06:49:26.335601 4971 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-13T06:49:23Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-13T06:49:23Z\\\",\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-13T06:49:26Z is after 2025-08-24T17:21:41Z" Dec 13 06:49:26 crc kubenswrapper[4971]: I1213 06:49:26.346992 4971 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-82xjz" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e0c4af12-edda-41d7-9b44-f87396174ef2\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T06:49:25Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T06:49:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T06:49:25Z\\\",\\\"message\\\":\\\"containers with unready status: [machine-config-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T06:49:25Z\\\",\\\"message\\\":\\\"containers with unready status: [machine-config-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cnxsq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cnxsq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-13T06:49:25Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-82xjz\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-13T06:49:26Z is after 2025-08-24T17:21:41Z" Dec 13 06:49:26 crc kubenswrapper[4971]: I1213 06:49:26.361037 4971 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-wwvpl" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"dddf744b-2122-49ad-a980-105c09636e0f\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T06:49:25Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T06:49:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T06:49:25Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T06:49:25Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-czz9v\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-13T06:49:25Z\\\"}}\" for pod \"openshift-multus\"/\"multus-wwvpl\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-13T06:49:26Z is after 2025-08-24T17:21:41Z" Dec 13 06:49:26 crc kubenswrapper[4971]: I1213 06:49:26.378127 4971 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-zlcg5" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"736c6b1a-93dd-4f71-81b6-1c43fcf5f556\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T06:49:25Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T06:49:25Z\\\",\\\"message\\\":\\\"containers with incomplete status: [kubecfg-setup]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T06:49:25Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T06:49:25Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vzmsr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vzmsr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vzmsr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vzmsr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vzmsr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vzmsr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vzmsr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vzmsr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vzmsr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-13T06:49:25Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-zlcg5\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-13T06:49:26Z is after 2025-08-24T17:21:41Z" Dec 13 06:49:26 crc kubenswrapper[4971]: I1213 06:49:26.396870 4971 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"f9d31b5a-b770-4a15-95c8-35b49677c608\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T06:49:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T06:49:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T06:49:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T06:49:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T06:49:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c9baa042f9dada954faa57b5454cd6b0a12a297e56ddc192c52cff50e42a16fa\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-13T06:49:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://c788de86754f0bf37ee184672fc1dcb388ccebc246d4f15be4b5da661f725f8f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-13T06:49:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://760817511a9b4f96a68dc38d26631250eaab72e1dd8f123e653039b6b4e91f84\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-13T06:49:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b859be0c9feac856e9855c6a038b02b49b32fade97873c46d0bec0a32d8b5147\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-13T06:49:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-13T06:49:03Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-13T06:49:26Z is after 2025-08-24T17:21:41Z" Dec 13 06:49:26 crc kubenswrapper[4971]: I1213 06:49:26.412060 4971 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-ncchz" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"de914d21-a688-401e-a66e-861beb6336d9\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T06:49:25Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T06:49:25Z\\\",\\\"message\\\":\\\"containers with incomplete status: [egress-router-binary-copy cni-plugins bond-cni-plugin routeoverride-cni whereabouts-cni-bincopy whereabouts-cni]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T06:49:25Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T06:49:25Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lqmjj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lqmjj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lqmjj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lqmjj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lqmjj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lqmjj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lqmjj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-13T06:49:25Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-ncchz\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-13T06:49:26Z is after 2025-08-24T17:21:41Z" Dec 13 06:49:26 crc kubenswrapper[4971]: I1213 06:49:26.929742 4971 generic.go:334] "Generic (PLEG): container finished" podID="736c6b1a-93dd-4f71-81b6-1c43fcf5f556" containerID="451b22df6973ddc8fa2209a65d64d9e41b33eb231bb4adfab88cf0e5943bb246" exitCode=0 Dec 13 06:49:26 crc kubenswrapper[4971]: I1213 06:49:26.929844 4971 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-zlcg5" event={"ID":"736c6b1a-93dd-4f71-81b6-1c43fcf5f556","Type":"ContainerDied","Data":"451b22df6973ddc8fa2209a65d64d9e41b33eb231bb4adfab88cf0e5943bb246"} Dec 13 06:49:26 crc kubenswrapper[4971]: I1213 06:49:26.930145 4971 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-zlcg5" event={"ID":"736c6b1a-93dd-4f71-81b6-1c43fcf5f556","Type":"ContainerStarted","Data":"527d12e90bd8f5cf8ef2baf875cc761d6d71ff609b96d3fc24245a28d60416ab"} Dec 13 06:49:26 crc kubenswrapper[4971]: I1213 06:49:26.932543 4971 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-wwvpl" event={"ID":"dddf744b-2122-49ad-a980-105c09636e0f","Type":"ContainerStarted","Data":"4676d897c7fc668587faec563f3551a056557879cee3e865192d2bd28edf62fc"} Dec 13 06:49:26 crc kubenswrapper[4971]: I1213 06:49:26.932591 4971 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-wwvpl" event={"ID":"dddf744b-2122-49ad-a980-105c09636e0f","Type":"ContainerStarted","Data":"5a822abb9ce4a2e965eb3ba5ca96ab7c8a049834c154a27ef681208e18984182"} Dec 13 06:49:26 crc kubenswrapper[4971]: I1213 06:49:26.936813 4971 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" event={"ID":"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49","Type":"ContainerStarted","Data":"9c8a0d1db0e754462d805e82daf5f52ffd5457507a30456535ec59e096c9d092"} Dec 13 06:49:26 crc kubenswrapper[4971]: I1213 06:49:26.938448 4971 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-additional-cni-plugins-ncchz" event={"ID":"de914d21-a688-401e-a66e-861beb6336d9","Type":"ContainerStarted","Data":"0b7111af6c5a198d02bb02858837cfb4f3b20828d8d206cd18aa0ffa102d0cd4"} Dec 13 06:49:26 crc kubenswrapper[4971]: I1213 06:49:26.938477 4971 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-additional-cni-plugins-ncchz" event={"ID":"de914d21-a688-401e-a66e-861beb6336d9","Type":"ContainerStarted","Data":"1fefcab2425766e3d1895579a4fbcce2c724825c1983a52002095321099e2687"} Dec 13 06:49:26 crc kubenswrapper[4971]: I1213 06:49:26.940633 4971 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-82xjz" event={"ID":"e0c4af12-edda-41d7-9b44-f87396174ef2","Type":"ContainerStarted","Data":"0f4577d9e37f860813e221f2e208177d8d383fd04f673ce357a0a405ea29d1e6"} Dec 13 06:49:26 crc kubenswrapper[4971]: I1213 06:49:26.940658 4971 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-82xjz" event={"ID":"e0c4af12-edda-41d7-9b44-f87396174ef2","Type":"ContainerStarted","Data":"7e0d2020e62f1a277cff7047e1907dc747ed393a1f6c983e1ee258bd156c16c7"} Dec 13 06:49:26 crc kubenswrapper[4971]: I1213 06:49:26.940668 4971 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-82xjz" event={"ID":"e0c4af12-edda-41d7-9b44-f87396174ef2","Type":"ContainerStarted","Data":"81902d3c633547cf11f7a8d00b514f0e8ffdba48c90ffeea5f81c919b933eb52"} Dec 13 06:49:26 crc kubenswrapper[4971]: I1213 06:49:26.949215 4971 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-13T06:49:24Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-13T06:49:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-13T06:49:24Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4328cd6a8464ed468229bbfbc8265f969427916d72ebb92237dbe68efe70fa47\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-13T06:49:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://9d715ca82d9b5a594e41f2b98da6f4277da0e4e1dec114969f901c00f6d7553d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-13T06:49:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-13T06:49:26Z is after 2025-08-24T17:21:41Z" Dec 13 06:49:26 crc kubenswrapper[4971]: I1213 06:49:26.975017 4971 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-5rzh5" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"6bf4591d-2483-4132-9672-d0fd6c995968\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T06:49:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T06:49:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T06:49:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T06:49:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://50e2cb14e4ffb603acb06e2281615a99f06182d345c4e70f2446fa2077033bbb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-13T06:49:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dfxqn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-13T06:49:25Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-5rzh5\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-13T06:49:26Z is after 2025-08-24T17:21:41Z" Dec 13 06:49:27 crc kubenswrapper[4971]: I1213 06:49:27.018745 4971 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"1ed9f82e-fe5c-4b76-a619-62e5abd88d3b\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T06:49:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T06:49:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T06:49:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T06:49:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T06:49:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e457329a78a2d24337f994e8a0a1f75a2a667244aa75eb84613d1e7138364aac\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-13T06:49:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://eb014de8140deef097d9bd94f3b9fa4ab3558bc4449bef84b83bf62bfabf5836\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-13T06:49:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://2c438cc74d555f7c9db4569d1c56b78b1088ed8105697d2c0f31e6c5136f3dd0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-13T06:49:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://aa6fa6e4340dcd066252df3c56d79a3a39ed531d5eea8d1d5941d6952368d3af\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-13T06:49:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://c4471b98854c52d5cd084d0ba3e64e5b08e9518ec3c372cce2c875a440522457\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-13T06:49:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7d2746335c98dae3a0f43024d34cb62ebcb42a01c8f9ce18d52b29d4c45598e5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://7d2746335c98dae3a0f43024d34cb62ebcb42a01c8f9ce18d52b29d4c45598e5\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-13T06:49:04Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-13T06:49:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://bc0e77d10a738128bea82122496bc42235fc1688f8907756d93e61b651902b57\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://bc0e77d10a738128bea82122496bc42235fc1688f8907756d93e61b651902b57\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-13T06:49:05Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-13T06:49:05Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://f821e418cba5e832da8839268bcbbd9630bdf319a212d5382bfa2c4b0b73a268\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://f821e418cba5e832da8839268bcbbd9630bdf319a212d5382bfa2c4b0b73a268\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-13T06:49:07Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-13T06:49:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-13T06:49:03Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-13T06:49:27Z is after 2025-08-24T17:21:41Z" Dec 13 06:49:27 crc kubenswrapper[4971]: I1213 06:49:27.046300 4971 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-13T06:49:23Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-13T06:49:23Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-13T06:49:27Z is after 2025-08-24T17:21:41Z" Dec 13 06:49:27 crc kubenswrapper[4971]: I1213 06:49:27.115225 4971 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3ad3268a-a982-428c-9c06-4aaeddd372d9\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T06:49:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T06:49:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T06:49:03Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T06:49:03Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T06:49:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3716afdf2bc693be80a09fc9f7727d334f435caf1d4c8ec4811b3a1f1652cdfe\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-13T06:49:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://30df3abf8c79db8af52c6a918f73e1ce410ddafcf341636424c267e1eeb195a7\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-13T06:49:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://ccdf65f6d2b1eb3647a03ba37c305c32e34addfdf79ff937c8dddaa3e5d5b5c6\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-13T06:49:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://770b73dd0c2b6f38744e2c1cd6aa186a69bb87d1cd015a603ad2e33be06a8f6b\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://b3c1bab3d53f975b2e24582ce2771bc26d600e0dd7c30339dcfd18a071162ad4\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-12-13T06:49:23Z\\\",\\\"message\\\":\\\"file observer\\\\nW1213 06:49:23.283403 1 builder.go:272] unable to get owner reference (falling back to namespace): pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\nI1213 06:49:23.284336 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1213 06:49:23.286978 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-981773671/tls.crt::/tmp/serving-cert-981773671/tls.key\\\\\\\"\\\\nI1213 06:49:23.775072 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1213 06:49:23.778615 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1213 06:49:23.778645 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1213 06:49:23.778680 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1213 06:49:23.778691 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1213 06:49:23.785058 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1213 06:49:23.786150 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1213 06:49:23.786196 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1213 06:49:23.786223 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1213 06:49:23.786246 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1213 06:49:23.786269 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1213 06:49:23.786291 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI1213 06:49:23.785608 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF1213 06:49:23.787563 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-13T06:49:08Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-13T06:49:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://9acd07066deed483bd7fa61ee45fda09539d110717a134a56827f871186f92fc\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-13T06:49:07Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://38c91cfbaecfdbd3e3b663422a91ab4294610afb6bbf9ebdb77db24e2b76c682\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://38c91cfbaecfdbd3e3b663422a91ab4294610afb6bbf9ebdb77db24e2b76c682\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-13T06:49:05Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-13T06:49:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-13T06:49:03Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-13T06:49:27Z is after 2025-08-24T17:21:41Z" Dec 13 06:49:27 crc kubenswrapper[4971]: I1213 06:49:27.129280 4971 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-13T06:49:23Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-13T06:49:23Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-13T06:49:27Z is after 2025-08-24T17:21:41Z" Dec 13 06:49:27 crc kubenswrapper[4971]: I1213 06:49:27.142147 4971 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-13T06:49:23Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-13T06:49:23Z\\\",\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-13T06:49:27Z is after 2025-08-24T17:21:41Z" Dec 13 06:49:27 crc kubenswrapper[4971]: I1213 06:49:27.157017 4971 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-82xjz" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e0c4af12-edda-41d7-9b44-f87396174ef2\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T06:49:25Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T06:49:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T06:49:25Z\\\",\\\"message\\\":\\\"containers with unready status: [machine-config-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T06:49:25Z\\\",\\\"message\\\":\\\"containers with unready status: [machine-config-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cnxsq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cnxsq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-13T06:49:25Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-82xjz\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-13T06:49:27Z is after 2025-08-24T17:21:41Z" Dec 13 06:49:27 crc kubenswrapper[4971]: I1213 06:49:27.167653 4971 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-wwvpl" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"dddf744b-2122-49ad-a980-105c09636e0f\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T06:49:25Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T06:49:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T06:49:25Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T06:49:25Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-czz9v\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-13T06:49:25Z\\\"}}\" for pod \"openshift-multus\"/\"multus-wwvpl\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-13T06:49:27Z is after 2025-08-24T17:21:41Z" Dec 13 06:49:27 crc kubenswrapper[4971]: I1213 06:49:27.186374 4971 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-zlcg5" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"736c6b1a-93dd-4f71-81b6-1c43fcf5f556\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T06:49:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T06:49:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T06:49:25Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T06:49:25Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vzmsr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vzmsr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vzmsr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vzmsr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vzmsr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vzmsr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vzmsr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vzmsr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://451b22df6973ddc8fa2209a65d64d9e41b33eb231bb4adfab88cf0e5943bb246\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://451b22df6973ddc8fa2209a65d64d9e41b33eb231bb4adfab88cf0e5943bb246\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-13T06:49:26Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-13T06:49:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vzmsr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-13T06:49:25Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-zlcg5\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-13T06:49:27Z is after 2025-08-24T17:21:41Z" Dec 13 06:49:27 crc kubenswrapper[4971]: I1213 06:49:27.201817 4971 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-13T06:49:24Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-13T06:49:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-13T06:49:24Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d0b2ecf0d5b22f23217f42f259ae7ffb6359555b22c30eaebf667eaf254b3bab\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-13T06:49:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-13T06:49:27Z is after 2025-08-24T17:21:41Z" Dec 13 06:49:27 crc kubenswrapper[4971]: I1213 06:49:27.225546 4971 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-13T06:49:23Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-13T06:49:23Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-13T06:49:27Z is after 2025-08-24T17:21:41Z" Dec 13 06:49:27 crc kubenswrapper[4971]: I1213 06:49:27.254928 4971 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-ncchz" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"de914d21-a688-401e-a66e-861beb6336d9\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T06:49:25Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T06:49:25Z\\\",\\\"message\\\":\\\"containers with incomplete status: [egress-router-binary-copy cni-plugins bond-cni-plugin routeoverride-cni whereabouts-cni-bincopy whereabouts-cni]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T06:49:25Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T06:49:25Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lqmjj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lqmjj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lqmjj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lqmjj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lqmjj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lqmjj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lqmjj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-13T06:49:25Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-ncchz\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-13T06:49:27Z is after 2025-08-24T17:21:41Z" Dec 13 06:49:27 crc kubenswrapper[4971]: I1213 06:49:27.267529 4971 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"f9d31b5a-b770-4a15-95c8-35b49677c608\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T06:49:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T06:49:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T06:49:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T06:49:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T06:49:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c9baa042f9dada954faa57b5454cd6b0a12a297e56ddc192c52cff50e42a16fa\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-13T06:49:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://c788de86754f0bf37ee184672fc1dcb388ccebc246d4f15be4b5da661f725f8f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-13T06:49:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://760817511a9b4f96a68dc38d26631250eaab72e1dd8f123e653039b6b4e91f84\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-13T06:49:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b859be0c9feac856e9855c6a038b02b49b32fade97873c46d0bec0a32d8b5147\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-13T06:49:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-13T06:49:03Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-13T06:49:27Z is after 2025-08-24T17:21:41Z" Dec 13 06:49:27 crc kubenswrapper[4971]: I1213 06:49:27.279687 4971 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-13T06:49:23Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-13T06:49:23Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-13T06:49:27Z is after 2025-08-24T17:21:41Z" Dec 13 06:49:27 crc kubenswrapper[4971]: I1213 06:49:27.291639 4971 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-13T06:49:23Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-13T06:49:23Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-13T06:49:27Z is after 2025-08-24T17:21:41Z" Dec 13 06:49:27 crc kubenswrapper[4971]: I1213 06:49:27.303724 4971 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3ad3268a-a982-428c-9c06-4aaeddd372d9\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T06:49:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T06:49:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T06:49:03Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T06:49:03Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T06:49:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3716afdf2bc693be80a09fc9f7727d334f435caf1d4c8ec4811b3a1f1652cdfe\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-13T06:49:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://30df3abf8c79db8af52c6a918f73e1ce410ddafcf341636424c267e1eeb195a7\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-13T06:49:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://ccdf65f6d2b1eb3647a03ba37c305c32e34addfdf79ff937c8dddaa3e5d5b5c6\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-13T06:49:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://770b73dd0c2b6f38744e2c1cd6aa186a69bb87d1cd015a603ad2e33be06a8f6b\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://b3c1bab3d53f975b2e24582ce2771bc26d600e0dd7c30339dcfd18a071162ad4\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-12-13T06:49:23Z\\\",\\\"message\\\":\\\"file observer\\\\nW1213 06:49:23.283403 1 builder.go:272] unable to get owner reference (falling back to namespace): pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\nI1213 06:49:23.284336 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1213 06:49:23.286978 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-981773671/tls.crt::/tmp/serving-cert-981773671/tls.key\\\\\\\"\\\\nI1213 06:49:23.775072 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1213 06:49:23.778615 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1213 06:49:23.778645 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1213 06:49:23.778680 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1213 06:49:23.778691 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1213 06:49:23.785058 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1213 06:49:23.786150 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1213 06:49:23.786196 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1213 06:49:23.786223 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1213 06:49:23.786246 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1213 06:49:23.786269 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1213 06:49:23.786291 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI1213 06:49:23.785608 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF1213 06:49:23.787563 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-13T06:49:08Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-13T06:49:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://9acd07066deed483bd7fa61ee45fda09539d110717a134a56827f871186f92fc\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-13T06:49:07Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://38c91cfbaecfdbd3e3b663422a91ab4294610afb6bbf9ebdb77db24e2b76c682\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://38c91cfbaecfdbd3e3b663422a91ab4294610afb6bbf9ebdb77db24e2b76c682\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-13T06:49:05Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-13T06:49:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-13T06:49:03Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-13T06:49:27Z is after 2025-08-24T17:21:41Z" Dec 13 06:49:27 crc kubenswrapper[4971]: I1213 06:49:27.314814 4971 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-13T06:49:23Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-13T06:49:23Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-13T06:49:27Z is after 2025-08-24T17:21:41Z" Dec 13 06:49:27 crc kubenswrapper[4971]: I1213 06:49:27.324932 4971 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-13T06:49:26Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-13T06:49:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-13T06:49:26Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9c8a0d1db0e754462d805e82daf5f52ffd5457507a30456535ec59e096c9d092\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-13T06:49:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-13T06:49:27Z is after 2025-08-24T17:21:41Z" Dec 13 06:49:27 crc kubenswrapper[4971]: I1213 06:49:27.326654 4971 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 13 06:49:27 crc kubenswrapper[4971]: I1213 06:49:27.326780 4971 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"networking-console-plugin-cert\" (UniqueName: \"kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 13 06:49:27 crc kubenswrapper[4971]: I1213 06:49:27.326840 4971 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nginx-conf\" (UniqueName: \"kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 13 06:49:27 crc kubenswrapper[4971]: E1213 06:49:27.326886 4971 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-13 06:49:31.32686213 +0000 UTC m=+27.931271578 (durationBeforeRetry 4s). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 13 06:49:27 crc kubenswrapper[4971]: E1213 06:49:27.326928 4971 configmap.go:193] Couldn't get configMap openshift-network-console/networking-console-plugin: object "openshift-network-console"/"networking-console-plugin" not registered Dec 13 06:49:27 crc kubenswrapper[4971]: E1213 06:49:27.326973 4971 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2025-12-13 06:49:31.326961342 +0000 UTC m=+27.931370790 (durationBeforeRetry 4s). Error: MountVolume.SetUp failed for volume "nginx-conf" (UniqueName: "kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin" not registered Dec 13 06:49:27 crc kubenswrapper[4971]: E1213 06:49:27.326998 4971 secret.go:188] Couldn't get secret openshift-network-console/networking-console-plugin-cert: object "openshift-network-console"/"networking-console-plugin-cert" not registered Dec 13 06:49:27 crc kubenswrapper[4971]: E1213 06:49:27.327053 4971 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2025-12-13 06:49:31.327042144 +0000 UTC m=+27.931451652 (durationBeforeRetry 4s). Error: MountVolume.SetUp failed for volume "networking-console-plugin-cert" (UniqueName: "kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin-cert" not registered Dec 13 06:49:27 crc kubenswrapper[4971]: I1213 06:49:27.334551 4971 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-82xjz" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e0c4af12-edda-41d7-9b44-f87396174ef2\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T06:49:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T06:49:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T06:49:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T06:49:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://0f4577d9e37f860813e221f2e208177d8d383fd04f673ce357a0a405ea29d1e6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-13T06:49:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cnxsq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://7e0d2020e62f1a277cff7047e1907dc747ed393a1f6c983e1ee258bd156c16c7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-13T06:49:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cnxsq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-13T06:49:25Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-82xjz\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-13T06:49:27Z is after 2025-08-24T17:21:41Z" Dec 13 06:49:27 crc kubenswrapper[4971]: I1213 06:49:27.348950 4971 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-wwvpl" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"dddf744b-2122-49ad-a980-105c09636e0f\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T06:49:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T06:49:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T06:49:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T06:49:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4676d897c7fc668587faec563f3551a056557879cee3e865192d2bd28edf62fc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-13T06:49:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-czz9v\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-13T06:49:25Z\\\"}}\" for pod \"openshift-multus\"/\"multus-wwvpl\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-13T06:49:27Z is after 2025-08-24T17:21:41Z" Dec 13 06:49:27 crc kubenswrapper[4971]: I1213 06:49:27.369637 4971 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-zlcg5" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"736c6b1a-93dd-4f71-81b6-1c43fcf5f556\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T06:49:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T06:49:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T06:49:25Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T06:49:25Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vzmsr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vzmsr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vzmsr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vzmsr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vzmsr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vzmsr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vzmsr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vzmsr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://451b22df6973ddc8fa2209a65d64d9e41b33eb231bb4adfab88cf0e5943bb246\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://451b22df6973ddc8fa2209a65d64d9e41b33eb231bb4adfab88cf0e5943bb246\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-13T06:49:26Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-13T06:49:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vzmsr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-13T06:49:25Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-zlcg5\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-13T06:49:27Z is after 2025-08-24T17:21:41Z" Dec 13 06:49:27 crc kubenswrapper[4971]: I1213 06:49:27.382442 4971 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-13T06:49:24Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-13T06:49:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-13T06:49:24Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d0b2ecf0d5b22f23217f42f259ae7ffb6359555b22c30eaebf667eaf254b3bab\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-13T06:49:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-13T06:49:27Z is after 2025-08-24T17:21:41Z" Dec 13 06:49:27 crc kubenswrapper[4971]: I1213 06:49:27.396139 4971 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-ncchz" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"de914d21-a688-401e-a66e-861beb6336d9\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T06:49:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T06:49:25Z\\\",\\\"message\\\":\\\"containers with incomplete status: [egress-router-binary-copy cni-plugins bond-cni-plugin routeoverride-cni whereabouts-cni-bincopy whereabouts-cni]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T06:49:25Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T06:49:25Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lqmjj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://0b7111af6c5a198d02bb02858837cfb4f3b20828d8d206cd18aa0ffa102d0cd4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-13T06:49:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lqmjj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lqmjj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lqmjj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lqmjj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lqmjj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lqmjj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-13T06:49:25Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-ncchz\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-13T06:49:27Z is after 2025-08-24T17:21:41Z" Dec 13 06:49:27 crc kubenswrapper[4971]: I1213 06:49:27.410321 4971 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"f9d31b5a-b770-4a15-95c8-35b49677c608\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T06:49:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T06:49:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T06:49:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T06:49:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T06:49:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c9baa042f9dada954faa57b5454cd6b0a12a297e56ddc192c52cff50e42a16fa\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-13T06:49:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://c788de86754f0bf37ee184672fc1dcb388ccebc246d4f15be4b5da661f725f8f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-13T06:49:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://760817511a9b4f96a68dc38d26631250eaab72e1dd8f123e653039b6b4e91f84\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-13T06:49:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b859be0c9feac856e9855c6a038b02b49b32fade97873c46d0bec0a32d8b5147\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-13T06:49:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-13T06:49:03Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-13T06:49:27Z is after 2025-08-24T17:21:41Z" Dec 13 06:49:27 crc kubenswrapper[4971]: I1213 06:49:27.427283 4971 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cqllr\" (UniqueName: \"kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr\") pod \"network-check-target-xd92c\" (UID: \"3b6479f0-333b-4a96-9adf-2099afdc2447\") " pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 13 06:49:27 crc kubenswrapper[4971]: I1213 06:49:27.427371 4971 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-s2dwl\" (UniqueName: \"kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl\") pod \"network-check-source-55646444c4-trplf\" (UID: \"9d751cbb-f2e2-430d-9754-c882a5e924a5\") " pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 13 06:49:27 crc kubenswrapper[4971]: E1213 06:49:27.427492 4971 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Dec 13 06:49:27 crc kubenswrapper[4971]: E1213 06:49:27.427531 4971 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Dec 13 06:49:27 crc kubenswrapper[4971]: E1213 06:49:27.427543 4971 projected.go:194] Error preparing data for projected volume kube-api-access-s2dwl for pod openshift-network-diagnostics/network-check-source-55646444c4-trplf: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Dec 13 06:49:27 crc kubenswrapper[4971]: E1213 06:49:27.427584 4971 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl podName:9d751cbb-f2e2-430d-9754-c882a5e924a5 nodeName:}" failed. No retries permitted until 2025-12-13 06:49:31.427571415 +0000 UTC m=+28.031980853 (durationBeforeRetry 4s). Error: MountVolume.SetUp failed for volume "kube-api-access-s2dwl" (UniqueName: "kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl") pod "network-check-source-55646444c4-trplf" (UID: "9d751cbb-f2e2-430d-9754-c882a5e924a5") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Dec 13 06:49:27 crc kubenswrapper[4971]: E1213 06:49:27.427861 4971 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Dec 13 06:49:27 crc kubenswrapper[4971]: E1213 06:49:27.427876 4971 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Dec 13 06:49:27 crc kubenswrapper[4971]: E1213 06:49:27.427883 4971 projected.go:194] Error preparing data for projected volume kube-api-access-cqllr for pod openshift-network-diagnostics/network-check-target-xd92c: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Dec 13 06:49:27 crc kubenswrapper[4971]: E1213 06:49:27.427905 4971 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr podName:3b6479f0-333b-4a96-9adf-2099afdc2447 nodeName:}" failed. No retries permitted until 2025-12-13 06:49:31.427897953 +0000 UTC m=+28.032307391 (durationBeforeRetry 4s). Error: MountVolume.SetUp failed for volume "kube-api-access-cqllr" (UniqueName: "kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr") pod "network-check-target-xd92c" (UID: "3b6479f0-333b-4a96-9adf-2099afdc2447") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Dec 13 06:49:27 crc kubenswrapper[4971]: I1213 06:49:27.430346 4971 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"1ed9f82e-fe5c-4b76-a619-62e5abd88d3b\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T06:49:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T06:49:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T06:49:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T06:49:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T06:49:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e457329a78a2d24337f994e8a0a1f75a2a667244aa75eb84613d1e7138364aac\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-13T06:49:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://eb014de8140deef097d9bd94f3b9fa4ab3558bc4449bef84b83bf62bfabf5836\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-13T06:49:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://2c438cc74d555f7c9db4569d1c56b78b1088ed8105697d2c0f31e6c5136f3dd0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-13T06:49:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://aa6fa6e4340dcd066252df3c56d79a3a39ed531d5eea8d1d5941d6952368d3af\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-13T06:49:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://c4471b98854c52d5cd084d0ba3e64e5b08e9518ec3c372cce2c875a440522457\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-13T06:49:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7d2746335c98dae3a0f43024d34cb62ebcb42a01c8f9ce18d52b29d4c45598e5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://7d2746335c98dae3a0f43024d34cb62ebcb42a01c8f9ce18d52b29d4c45598e5\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-13T06:49:04Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-13T06:49:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://bc0e77d10a738128bea82122496bc42235fc1688f8907756d93e61b651902b57\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://bc0e77d10a738128bea82122496bc42235fc1688f8907756d93e61b651902b57\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-13T06:49:05Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-13T06:49:05Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://f821e418cba5e832da8839268bcbbd9630bdf319a212d5382bfa2c4b0b73a268\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://f821e418cba5e832da8839268bcbbd9630bdf319a212d5382bfa2c4b0b73a268\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-13T06:49:07Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-13T06:49:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-13T06:49:03Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-13T06:49:27Z is after 2025-08-24T17:21:41Z" Dec 13 06:49:27 crc kubenswrapper[4971]: I1213 06:49:27.442903 4971 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-13T06:49:24Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-13T06:49:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-13T06:49:24Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4328cd6a8464ed468229bbfbc8265f969427916d72ebb92237dbe68efe70fa47\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-13T06:49:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://9d715ca82d9b5a594e41f2b98da6f4277da0e4e1dec114969f901c00f6d7553d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-13T06:49:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-13T06:49:27Z is after 2025-08-24T17:21:41Z" Dec 13 06:49:27 crc kubenswrapper[4971]: I1213 06:49:27.453008 4971 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-5rzh5" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"6bf4591d-2483-4132-9672-d0fd6c995968\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T06:49:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T06:49:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T06:49:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T06:49:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://50e2cb14e4ffb603acb06e2281615a99f06182d345c4e70f2446fa2077033bbb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-13T06:49:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dfxqn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-13T06:49:25Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-5rzh5\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-13T06:49:27Z is after 2025-08-24T17:21:41Z" Dec 13 06:49:27 crc kubenswrapper[4971]: I1213 06:49:27.767845 4971 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 13 06:49:27 crc kubenswrapper[4971]: E1213 06:49:27.768274 4971 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 13 06:49:27 crc kubenswrapper[4971]: I1213 06:49:27.768133 4971 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 13 06:49:27 crc kubenswrapper[4971]: E1213 06:49:27.768362 4971 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 13 06:49:27 crc kubenswrapper[4971]: I1213 06:49:27.768106 4971 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 13 06:49:27 crc kubenswrapper[4971]: E1213 06:49:27.768435 4971 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 13 06:49:27 crc kubenswrapper[4971]: I1213 06:49:27.843005 4971 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-image-registry/node-ca-gkcd8"] Dec 13 06:49:27 crc kubenswrapper[4971]: I1213 06:49:27.843657 4971 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-image-registry/node-ca-gkcd8" Dec 13 06:49:27 crc kubenswrapper[4971]: I1213 06:49:27.844968 4971 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-image-registry"/"image-registry-certificates" Dec 13 06:49:27 crc kubenswrapper[4971]: I1213 06:49:27.845358 4971 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-image-registry"/"openshift-service-ca.crt" Dec 13 06:49:27 crc kubenswrapper[4971]: I1213 06:49:27.845482 4971 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-image-registry"/"kube-root-ca.crt" Dec 13 06:49:27 crc kubenswrapper[4971]: I1213 06:49:27.845936 4971 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-image-registry"/"node-ca-dockercfg-4777p" Dec 13 06:49:27 crc kubenswrapper[4971]: I1213 06:49:27.866370 4971 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"1ed9f82e-fe5c-4b76-a619-62e5abd88d3b\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T06:49:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T06:49:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T06:49:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T06:49:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T06:49:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e457329a78a2d24337f994e8a0a1f75a2a667244aa75eb84613d1e7138364aac\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-13T06:49:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://eb014de8140deef097d9bd94f3b9fa4ab3558bc4449bef84b83bf62bfabf5836\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-13T06:49:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://2c438cc74d555f7c9db4569d1c56b78b1088ed8105697d2c0f31e6c5136f3dd0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-13T06:49:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://aa6fa6e4340dcd066252df3c56d79a3a39ed531d5eea8d1d5941d6952368d3af\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-13T06:49:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://c4471b98854c52d5cd084d0ba3e64e5b08e9518ec3c372cce2c875a440522457\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-13T06:49:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7d2746335c98dae3a0f43024d34cb62ebcb42a01c8f9ce18d52b29d4c45598e5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://7d2746335c98dae3a0f43024d34cb62ebcb42a01c8f9ce18d52b29d4c45598e5\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-13T06:49:04Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-13T06:49:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://bc0e77d10a738128bea82122496bc42235fc1688f8907756d93e61b651902b57\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://bc0e77d10a738128bea82122496bc42235fc1688f8907756d93e61b651902b57\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-13T06:49:05Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-13T06:49:05Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://f821e418cba5e832da8839268bcbbd9630bdf319a212d5382bfa2c4b0b73a268\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://f821e418cba5e832da8839268bcbbd9630bdf319a212d5382bfa2c4b0b73a268\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-13T06:49:07Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-13T06:49:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-13T06:49:03Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-13T06:49:27Z is after 2025-08-24T17:21:41Z" Dec 13 06:49:27 crc kubenswrapper[4971]: I1213 06:49:27.880706 4971 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-13T06:49:24Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-13T06:49:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-13T06:49:24Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4328cd6a8464ed468229bbfbc8265f969427916d72ebb92237dbe68efe70fa47\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-13T06:49:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://9d715ca82d9b5a594e41f2b98da6f4277da0e4e1dec114969f901c00f6d7553d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-13T06:49:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-13T06:49:27Z is after 2025-08-24T17:21:41Z" Dec 13 06:49:27 crc kubenswrapper[4971]: I1213 06:49:27.891771 4971 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-5rzh5" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"6bf4591d-2483-4132-9672-d0fd6c995968\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T06:49:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T06:49:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T06:49:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T06:49:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://50e2cb14e4ffb603acb06e2281615a99f06182d345c4e70f2446fa2077033bbb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-13T06:49:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dfxqn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-13T06:49:25Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-5rzh5\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-13T06:49:27Z is after 2025-08-24T17:21:41Z" Dec 13 06:49:27 crc kubenswrapper[4971]: I1213 06:49:27.904895 4971 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3ad3268a-a982-428c-9c06-4aaeddd372d9\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T06:49:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T06:49:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T06:49:03Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T06:49:03Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T06:49:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3716afdf2bc693be80a09fc9f7727d334f435caf1d4c8ec4811b3a1f1652cdfe\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-13T06:49:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://30df3abf8c79db8af52c6a918f73e1ce410ddafcf341636424c267e1eeb195a7\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-13T06:49:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://ccdf65f6d2b1eb3647a03ba37c305c32e34addfdf79ff937c8dddaa3e5d5b5c6\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-13T06:49:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://770b73dd0c2b6f38744e2c1cd6aa186a69bb87d1cd015a603ad2e33be06a8f6b\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://b3c1bab3d53f975b2e24582ce2771bc26d600e0dd7c30339dcfd18a071162ad4\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-12-13T06:49:23Z\\\",\\\"message\\\":\\\"file observer\\\\nW1213 06:49:23.283403 1 builder.go:272] unable to get owner reference (falling back to namespace): pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\nI1213 06:49:23.284336 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1213 06:49:23.286978 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-981773671/tls.crt::/tmp/serving-cert-981773671/tls.key\\\\\\\"\\\\nI1213 06:49:23.775072 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1213 06:49:23.778615 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1213 06:49:23.778645 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1213 06:49:23.778680 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1213 06:49:23.778691 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1213 06:49:23.785058 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1213 06:49:23.786150 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1213 06:49:23.786196 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1213 06:49:23.786223 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1213 06:49:23.786246 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1213 06:49:23.786269 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1213 06:49:23.786291 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI1213 06:49:23.785608 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF1213 06:49:23.787563 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-13T06:49:08Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-13T06:49:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://9acd07066deed483bd7fa61ee45fda09539d110717a134a56827f871186f92fc\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-13T06:49:07Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://38c91cfbaecfdbd3e3b663422a91ab4294610afb6bbf9ebdb77db24e2b76c682\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://38c91cfbaecfdbd3e3b663422a91ab4294610afb6bbf9ebdb77db24e2b76c682\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-13T06:49:05Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-13T06:49:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-13T06:49:03Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-13T06:49:27Z is after 2025-08-24T17:21:41Z" Dec 13 06:49:27 crc kubenswrapper[4971]: I1213 06:49:27.918185 4971 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-13T06:49:23Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-13T06:49:23Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-13T06:49:27Z is after 2025-08-24T17:21:41Z" Dec 13 06:49:27 crc kubenswrapper[4971]: I1213 06:49:27.930907 4971 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-13T06:49:23Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-13T06:49:23Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-13T06:49:27Z is after 2025-08-24T17:21:41Z" Dec 13 06:49:27 crc kubenswrapper[4971]: I1213 06:49:27.932032 4971 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serviceca\" (UniqueName: \"kubernetes.io/configmap/13f5c327-f95e-42b4-8783-218500c17694-serviceca\") pod \"node-ca-gkcd8\" (UID: \"13f5c327-f95e-42b4-8783-218500c17694\") " pod="openshift-image-registry/node-ca-gkcd8" Dec 13 06:49:27 crc kubenswrapper[4971]: I1213 06:49:27.932161 4971 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host\" (UniqueName: \"kubernetes.io/host-path/13f5c327-f95e-42b4-8783-218500c17694-host\") pod \"node-ca-gkcd8\" (UID: \"13f5c327-f95e-42b4-8783-218500c17694\") " pod="openshift-image-registry/node-ca-gkcd8" Dec 13 06:49:27 crc kubenswrapper[4971]: I1213 06:49:27.932262 4971 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-gvlhn\" (UniqueName: \"kubernetes.io/projected/13f5c327-f95e-42b4-8783-218500c17694-kube-api-access-gvlhn\") pod \"node-ca-gkcd8\" (UID: \"13f5c327-f95e-42b4-8783-218500c17694\") " pod="openshift-image-registry/node-ca-gkcd8" Dec 13 06:49:27 crc kubenswrapper[4971]: I1213 06:49:27.944724 4971 generic.go:334] "Generic (PLEG): container finished" podID="de914d21-a688-401e-a66e-861beb6336d9" containerID="0b7111af6c5a198d02bb02858837cfb4f3b20828d8d206cd18aa0ffa102d0cd4" exitCode=0 Dec 13 06:49:27 crc kubenswrapper[4971]: I1213 06:49:27.944819 4971 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-additional-cni-plugins-ncchz" event={"ID":"de914d21-a688-401e-a66e-861beb6336d9","Type":"ContainerDied","Data":"0b7111af6c5a198d02bb02858837cfb4f3b20828d8d206cd18aa0ffa102d0cd4"} Dec 13 06:49:27 crc kubenswrapper[4971]: I1213 06:49:27.948744 4971 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-zlcg5" event={"ID":"736c6b1a-93dd-4f71-81b6-1c43fcf5f556","Type":"ContainerStarted","Data":"693a2ab434c250cb140a866e8068b156df9191414fed1d63bee919a2528ea6e5"} Dec 13 06:49:27 crc kubenswrapper[4971]: I1213 06:49:27.948782 4971 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-zlcg5" event={"ID":"736c6b1a-93dd-4f71-81b6-1c43fcf5f556","Type":"ContainerStarted","Data":"f22a4ec756119c4f791fb5105cf1f31df3e548b2b07b1f909a62802e9b6035ae"} Dec 13 06:49:27 crc kubenswrapper[4971]: I1213 06:49:27.948799 4971 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-zlcg5" event={"ID":"736c6b1a-93dd-4f71-81b6-1c43fcf5f556","Type":"ContainerStarted","Data":"325223543b3c7db417f769533559a1ca2b23a723494e1885b0fd0480b5a6399f"} Dec 13 06:49:27 crc kubenswrapper[4971]: I1213 06:49:27.948807 4971 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-zlcg5" event={"ID":"736c6b1a-93dd-4f71-81b6-1c43fcf5f556","Type":"ContainerStarted","Data":"219620f51474e60b3e1f880c0972eb612dffc4bad24462bbba2c093ae817d852"} Dec 13 06:49:27 crc kubenswrapper[4971]: I1213 06:49:27.948815 4971 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-zlcg5" event={"ID":"736c6b1a-93dd-4f71-81b6-1c43fcf5f556","Type":"ContainerStarted","Data":"f137f4b441c2402093ab430a9689a46221c4fb4bc70d99fb992f62b0a989c465"} Dec 13 06:49:27 crc kubenswrapper[4971]: I1213 06:49:27.953197 4971 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-zlcg5" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"736c6b1a-93dd-4f71-81b6-1c43fcf5f556\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T06:49:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T06:49:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T06:49:25Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T06:49:25Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vzmsr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vzmsr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vzmsr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vzmsr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vzmsr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vzmsr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vzmsr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vzmsr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://451b22df6973ddc8fa2209a65d64d9e41b33eb231bb4adfab88cf0e5943bb246\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://451b22df6973ddc8fa2209a65d64d9e41b33eb231bb4adfab88cf0e5943bb246\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-13T06:49:26Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-13T06:49:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vzmsr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-13T06:49:25Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-zlcg5\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-13T06:49:27Z is after 2025-08-24T17:21:41Z" Dec 13 06:49:27 crc kubenswrapper[4971]: I1213 06:49:27.968479 4971 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-13T06:49:24Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-13T06:49:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-13T06:49:24Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d0b2ecf0d5b22f23217f42f259ae7ffb6359555b22c30eaebf667eaf254b3bab\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-13T06:49:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-13T06:49:27Z is after 2025-08-24T17:21:41Z" Dec 13 06:49:27 crc kubenswrapper[4971]: I1213 06:49:27.981103 4971 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-13T06:49:23Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-13T06:49:23Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-13T06:49:27Z is after 2025-08-24T17:21:41Z" Dec 13 06:49:27 crc kubenswrapper[4971]: I1213 06:49:27.992438 4971 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-13T06:49:26Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-13T06:49:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-13T06:49:26Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9c8a0d1db0e754462d805e82daf5f52ffd5457507a30456535ec59e096c9d092\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-13T06:49:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-13T06:49:27Z is after 2025-08-24T17:21:41Z" Dec 13 06:49:28 crc kubenswrapper[4971]: I1213 06:49:28.006085 4971 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-82xjz" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e0c4af12-edda-41d7-9b44-f87396174ef2\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T06:49:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T06:49:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T06:49:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T06:49:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://0f4577d9e37f860813e221f2e208177d8d383fd04f673ce357a0a405ea29d1e6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-13T06:49:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cnxsq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://7e0d2020e62f1a277cff7047e1907dc747ed393a1f6c983e1ee258bd156c16c7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-13T06:49:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cnxsq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-13T06:49:25Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-82xjz\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-13T06:49:28Z is after 2025-08-24T17:21:41Z" Dec 13 06:49:28 crc kubenswrapper[4971]: I1213 06:49:28.020448 4971 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-wwvpl" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"dddf744b-2122-49ad-a980-105c09636e0f\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T06:49:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T06:49:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T06:49:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T06:49:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4676d897c7fc668587faec563f3551a056557879cee3e865192d2bd28edf62fc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-13T06:49:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-czz9v\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-13T06:49:25Z\\\"}}\" for pod \"openshift-multus\"/\"multus-wwvpl\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-13T06:49:28Z is after 2025-08-24T17:21:41Z" Dec 13 06:49:28 crc kubenswrapper[4971]: I1213 06:49:28.032988 4971 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host\" (UniqueName: \"kubernetes.io/host-path/13f5c327-f95e-42b4-8783-218500c17694-host\") pod \"node-ca-gkcd8\" (UID: \"13f5c327-f95e-42b4-8783-218500c17694\") " pod="openshift-image-registry/node-ca-gkcd8" Dec 13 06:49:28 crc kubenswrapper[4971]: I1213 06:49:28.033043 4971 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-gvlhn\" (UniqueName: \"kubernetes.io/projected/13f5c327-f95e-42b4-8783-218500c17694-kube-api-access-gvlhn\") pod \"node-ca-gkcd8\" (UID: \"13f5c327-f95e-42b4-8783-218500c17694\") " pod="openshift-image-registry/node-ca-gkcd8" Dec 13 06:49:28 crc kubenswrapper[4971]: I1213 06:49:28.033109 4971 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serviceca\" (UniqueName: \"kubernetes.io/configmap/13f5c327-f95e-42b4-8783-218500c17694-serviceca\") pod \"node-ca-gkcd8\" (UID: \"13f5c327-f95e-42b4-8783-218500c17694\") " pod="openshift-image-registry/node-ca-gkcd8" Dec 13 06:49:28 crc kubenswrapper[4971]: I1213 06:49:28.033106 4971 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host\" (UniqueName: \"kubernetes.io/host-path/13f5c327-f95e-42b4-8783-218500c17694-host\") pod \"node-ca-gkcd8\" (UID: \"13f5c327-f95e-42b4-8783-218500c17694\") " pod="openshift-image-registry/node-ca-gkcd8" Dec 13 06:49:28 crc kubenswrapper[4971]: I1213 06:49:28.033689 4971 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"f9d31b5a-b770-4a15-95c8-35b49677c608\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T06:49:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T06:49:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T06:49:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T06:49:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T06:49:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c9baa042f9dada954faa57b5454cd6b0a12a297e56ddc192c52cff50e42a16fa\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-13T06:49:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://c788de86754f0bf37ee184672fc1dcb388ccebc246d4f15be4b5da661f725f8f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-13T06:49:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://760817511a9b4f96a68dc38d26631250eaab72e1dd8f123e653039b6b4e91f84\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-13T06:49:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b859be0c9feac856e9855c6a038b02b49b32fade97873c46d0bec0a32d8b5147\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-13T06:49:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-13T06:49:03Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-13T06:49:28Z is after 2025-08-24T17:21:41Z" Dec 13 06:49:28 crc kubenswrapper[4971]: I1213 06:49:28.034055 4971 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serviceca\" (UniqueName: \"kubernetes.io/configmap/13f5c327-f95e-42b4-8783-218500c17694-serviceca\") pod \"node-ca-gkcd8\" (UID: \"13f5c327-f95e-42b4-8783-218500c17694\") " pod="openshift-image-registry/node-ca-gkcd8" Dec 13 06:49:28 crc kubenswrapper[4971]: I1213 06:49:28.056835 4971 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-gvlhn\" (UniqueName: \"kubernetes.io/projected/13f5c327-f95e-42b4-8783-218500c17694-kube-api-access-gvlhn\") pod \"node-ca-gkcd8\" (UID: \"13f5c327-f95e-42b4-8783-218500c17694\") " pod="openshift-image-registry/node-ca-gkcd8" Dec 13 06:49:28 crc kubenswrapper[4971]: I1213 06:49:28.059123 4971 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-ncchz" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"de914d21-a688-401e-a66e-861beb6336d9\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T06:49:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T06:49:25Z\\\",\\\"message\\\":\\\"containers with incomplete status: [egress-router-binary-copy cni-plugins bond-cni-plugin routeoverride-cni whereabouts-cni-bincopy whereabouts-cni]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T06:49:25Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T06:49:25Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lqmjj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://0b7111af6c5a198d02bb02858837cfb4f3b20828d8d206cd18aa0ffa102d0cd4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-13T06:49:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lqmjj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lqmjj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lqmjj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lqmjj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lqmjj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lqmjj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-13T06:49:25Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-ncchz\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-13T06:49:28Z is after 2025-08-24T17:21:41Z" Dec 13 06:49:28 crc kubenswrapper[4971]: I1213 06:49:28.071033 4971 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-gkcd8" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"13f5c327-f95e-42b4-8783-218500c17694\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T06:49:27Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T06:49:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T06:49:27Z\\\",\\\"message\\\":\\\"containers with unready status: [node-ca]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T06:49:27Z\\\",\\\"message\\\":\\\"containers with unready status: [node-ca]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gvlhn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-13T06:49:27Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-gkcd8\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-13T06:49:28Z is after 2025-08-24T17:21:41Z" Dec 13 06:49:28 crc kubenswrapper[4971]: I1213 06:49:28.085006 4971 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-ncchz" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"de914d21-a688-401e-a66e-861beb6336d9\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T06:49:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T06:49:25Z\\\",\\\"message\\\":\\\"containers with incomplete status: [cni-plugins bond-cni-plugin routeoverride-cni whereabouts-cni-bincopy whereabouts-cni]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T06:49:25Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T06:49:25Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lqmjj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://0b7111af6c5a198d02bb02858837cfb4f3b20828d8d206cd18aa0ffa102d0cd4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://0b7111af6c5a198d02bb02858837cfb4f3b20828d8d206cd18aa0ffa102d0cd4\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-13T06:49:26Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-13T06:49:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lqmjj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lqmjj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lqmjj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lqmjj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lqmjj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lqmjj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-13T06:49:25Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-ncchz\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-13T06:49:28Z is after 2025-08-24T17:21:41Z" Dec 13 06:49:28 crc kubenswrapper[4971]: I1213 06:49:28.116869 4971 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-gkcd8" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"13f5c327-f95e-42b4-8783-218500c17694\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T06:49:27Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T06:49:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T06:49:27Z\\\",\\\"message\\\":\\\"containers with unready status: [node-ca]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T06:49:27Z\\\",\\\"message\\\":\\\"containers with unready status: [node-ca]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gvlhn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-13T06:49:27Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-gkcd8\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-13T06:49:28Z is after 2025-08-24T17:21:41Z" Dec 13 06:49:28 crc kubenswrapper[4971]: I1213 06:49:28.156870 4971 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"f9d31b5a-b770-4a15-95c8-35b49677c608\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T06:49:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T06:49:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T06:49:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T06:49:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T06:49:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c9baa042f9dada954faa57b5454cd6b0a12a297e56ddc192c52cff50e42a16fa\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-13T06:49:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://c788de86754f0bf37ee184672fc1dcb388ccebc246d4f15be4b5da661f725f8f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-13T06:49:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://760817511a9b4f96a68dc38d26631250eaab72e1dd8f123e653039b6b4e91f84\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-13T06:49:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b859be0c9feac856e9855c6a038b02b49b32fade97873c46d0bec0a32d8b5147\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-13T06:49:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-13T06:49:03Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-13T06:49:28Z is after 2025-08-24T17:21:41Z" Dec 13 06:49:28 crc kubenswrapper[4971]: I1213 06:49:28.163443 4971 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-image-registry/node-ca-gkcd8" Dec 13 06:49:28 crc kubenswrapper[4971]: W1213 06:49:28.179419 4971 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod13f5c327_f95e_42b4_8783_218500c17694.slice/crio-0933af7235514acbe1d6c34149f4c316f03c10636cde60889e7d28a087a224a0 WatchSource:0}: Error finding container 0933af7235514acbe1d6c34149f4c316f03c10636cde60889e7d28a087a224a0: Status 404 returned error can't find the container with id 0933af7235514acbe1d6c34149f4c316f03c10636cde60889e7d28a087a224a0 Dec 13 06:49:28 crc kubenswrapper[4971]: I1213 06:49:28.205154 4971 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"1ed9f82e-fe5c-4b76-a619-62e5abd88d3b\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T06:49:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T06:49:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T06:49:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T06:49:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T06:49:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e457329a78a2d24337f994e8a0a1f75a2a667244aa75eb84613d1e7138364aac\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-13T06:49:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://eb014de8140deef097d9bd94f3b9fa4ab3558bc4449bef84b83bf62bfabf5836\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-13T06:49:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://2c438cc74d555f7c9db4569d1c56b78b1088ed8105697d2c0f31e6c5136f3dd0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-13T06:49:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://aa6fa6e4340dcd066252df3c56d79a3a39ed531d5eea8d1d5941d6952368d3af\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-13T06:49:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://c4471b98854c52d5cd084d0ba3e64e5b08e9518ec3c372cce2c875a440522457\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-13T06:49:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7d2746335c98dae3a0f43024d34cb62ebcb42a01c8f9ce18d52b29d4c45598e5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://7d2746335c98dae3a0f43024d34cb62ebcb42a01c8f9ce18d52b29d4c45598e5\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-13T06:49:04Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-13T06:49:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://bc0e77d10a738128bea82122496bc42235fc1688f8907756d93e61b651902b57\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://bc0e77d10a738128bea82122496bc42235fc1688f8907756d93e61b651902b57\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-13T06:49:05Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-13T06:49:05Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://f821e418cba5e832da8839268bcbbd9630bdf319a212d5382bfa2c4b0b73a268\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://f821e418cba5e832da8839268bcbbd9630bdf319a212d5382bfa2c4b0b73a268\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-13T06:49:07Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-13T06:49:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-13T06:49:03Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-13T06:49:28Z is after 2025-08-24T17:21:41Z" Dec 13 06:49:28 crc kubenswrapper[4971]: I1213 06:49:28.238046 4971 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-13T06:49:24Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-13T06:49:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-13T06:49:24Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4328cd6a8464ed468229bbfbc8265f969427916d72ebb92237dbe68efe70fa47\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-13T06:49:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://9d715ca82d9b5a594e41f2b98da6f4277da0e4e1dec114969f901c00f6d7553d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-13T06:49:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-13T06:49:28Z is after 2025-08-24T17:21:41Z" Dec 13 06:49:28 crc kubenswrapper[4971]: I1213 06:49:28.276634 4971 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-5rzh5" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"6bf4591d-2483-4132-9672-d0fd6c995968\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T06:49:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T06:49:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T06:49:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T06:49:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://50e2cb14e4ffb603acb06e2281615a99f06182d345c4e70f2446fa2077033bbb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-13T06:49:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dfxqn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-13T06:49:25Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-5rzh5\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-13T06:49:28Z is after 2025-08-24T17:21:41Z" Dec 13 06:49:28 crc kubenswrapper[4971]: I1213 06:49:28.323192 4971 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-13T06:49:23Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-13T06:49:23Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-13T06:49:28Z is after 2025-08-24T17:21:41Z" Dec 13 06:49:28 crc kubenswrapper[4971]: I1213 06:49:28.358949 4971 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-13T06:49:23Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-13T06:49:23Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-13T06:49:28Z is after 2025-08-24T17:21:41Z" Dec 13 06:49:28 crc kubenswrapper[4971]: I1213 06:49:28.401257 4971 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3ad3268a-a982-428c-9c06-4aaeddd372d9\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T06:49:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T06:49:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T06:49:03Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T06:49:03Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T06:49:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3716afdf2bc693be80a09fc9f7727d334f435caf1d4c8ec4811b3a1f1652cdfe\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-13T06:49:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://30df3abf8c79db8af52c6a918f73e1ce410ddafcf341636424c267e1eeb195a7\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-13T06:49:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://ccdf65f6d2b1eb3647a03ba37c305c32e34addfdf79ff937c8dddaa3e5d5b5c6\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-13T06:49:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://770b73dd0c2b6f38744e2c1cd6aa186a69bb87d1cd015a603ad2e33be06a8f6b\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://b3c1bab3d53f975b2e24582ce2771bc26d600e0dd7c30339dcfd18a071162ad4\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-12-13T06:49:23Z\\\",\\\"message\\\":\\\"file observer\\\\nW1213 06:49:23.283403 1 builder.go:272] unable to get owner reference (falling back to namespace): pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\nI1213 06:49:23.284336 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1213 06:49:23.286978 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-981773671/tls.crt::/tmp/serving-cert-981773671/tls.key\\\\\\\"\\\\nI1213 06:49:23.775072 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1213 06:49:23.778615 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1213 06:49:23.778645 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1213 06:49:23.778680 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1213 06:49:23.778691 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1213 06:49:23.785058 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1213 06:49:23.786150 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1213 06:49:23.786196 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1213 06:49:23.786223 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1213 06:49:23.786246 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1213 06:49:23.786269 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1213 06:49:23.786291 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI1213 06:49:23.785608 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF1213 06:49:23.787563 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-13T06:49:08Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-13T06:49:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://9acd07066deed483bd7fa61ee45fda09539d110717a134a56827f871186f92fc\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-13T06:49:07Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://38c91cfbaecfdbd3e3b663422a91ab4294610afb6bbf9ebdb77db24e2b76c682\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://38c91cfbaecfdbd3e3b663422a91ab4294610afb6bbf9ebdb77db24e2b76c682\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-13T06:49:05Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-13T06:49:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-13T06:49:03Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-13T06:49:28Z is after 2025-08-24T17:21:41Z" Dec 13 06:49:28 crc kubenswrapper[4971]: I1213 06:49:28.441713 4971 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-13T06:49:23Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-13T06:49:23Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-13T06:49:28Z is after 2025-08-24T17:21:41Z" Dec 13 06:49:28 crc kubenswrapper[4971]: I1213 06:49:28.480275 4971 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-13T06:49:26Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-13T06:49:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-13T06:49:26Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9c8a0d1db0e754462d805e82daf5f52ffd5457507a30456535ec59e096c9d092\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-13T06:49:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-13T06:49:28Z is after 2025-08-24T17:21:41Z" Dec 13 06:49:28 crc kubenswrapper[4971]: I1213 06:49:28.517679 4971 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-82xjz" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e0c4af12-edda-41d7-9b44-f87396174ef2\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T06:49:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T06:49:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T06:49:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T06:49:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://0f4577d9e37f860813e221f2e208177d8d383fd04f673ce357a0a405ea29d1e6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-13T06:49:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cnxsq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://7e0d2020e62f1a277cff7047e1907dc747ed393a1f6c983e1ee258bd156c16c7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-13T06:49:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cnxsq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-13T06:49:25Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-82xjz\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-13T06:49:28Z is after 2025-08-24T17:21:41Z" Dec 13 06:49:28 crc kubenswrapper[4971]: I1213 06:49:28.559350 4971 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-wwvpl" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"dddf744b-2122-49ad-a980-105c09636e0f\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T06:49:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T06:49:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T06:49:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T06:49:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4676d897c7fc668587faec563f3551a056557879cee3e865192d2bd28edf62fc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-13T06:49:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-czz9v\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-13T06:49:25Z\\\"}}\" for pod \"openshift-multus\"/\"multus-wwvpl\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-13T06:49:28Z is after 2025-08-24T17:21:41Z" Dec 13 06:49:28 crc kubenswrapper[4971]: I1213 06:49:28.603654 4971 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-zlcg5" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"736c6b1a-93dd-4f71-81b6-1c43fcf5f556\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T06:49:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T06:49:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T06:49:25Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T06:49:25Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vzmsr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vzmsr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vzmsr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vzmsr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vzmsr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vzmsr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vzmsr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vzmsr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://451b22df6973ddc8fa2209a65d64d9e41b33eb231bb4adfab88cf0e5943bb246\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://451b22df6973ddc8fa2209a65d64d9e41b33eb231bb4adfab88cf0e5943bb246\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-13T06:49:26Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-13T06:49:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vzmsr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-13T06:49:25Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-zlcg5\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-13T06:49:28Z is after 2025-08-24T17:21:41Z" Dec 13 06:49:28 crc kubenswrapper[4971]: I1213 06:49:28.648314 4971 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-13T06:49:24Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-13T06:49:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-13T06:49:24Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d0b2ecf0d5b22f23217f42f259ae7ffb6359555b22c30eaebf667eaf254b3bab\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-13T06:49:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-13T06:49:28Z is after 2025-08-24T17:21:41Z" Dec 13 06:49:28 crc kubenswrapper[4971]: I1213 06:49:28.953416 4971 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-image-registry/node-ca-gkcd8" event={"ID":"13f5c327-f95e-42b4-8783-218500c17694","Type":"ContainerStarted","Data":"77e846e36f1cb40d03471dc03de11b6c3da448baa407aad38093fae9a801cf92"} Dec 13 06:49:28 crc kubenswrapper[4971]: I1213 06:49:28.953462 4971 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-image-registry/node-ca-gkcd8" event={"ID":"13f5c327-f95e-42b4-8783-218500c17694","Type":"ContainerStarted","Data":"0933af7235514acbe1d6c34149f4c316f03c10636cde60889e7d28a087a224a0"} Dec 13 06:49:28 crc kubenswrapper[4971]: I1213 06:49:28.956446 4971 generic.go:334] "Generic (PLEG): container finished" podID="de914d21-a688-401e-a66e-861beb6336d9" containerID="b36777fe40e74ae1cecb7b73e204139db5279ba51cef4f1e8c4a4c491d8be403" exitCode=0 Dec 13 06:49:28 crc kubenswrapper[4971]: I1213 06:49:28.956535 4971 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-additional-cni-plugins-ncchz" event={"ID":"de914d21-a688-401e-a66e-861beb6336d9","Type":"ContainerDied","Data":"b36777fe40e74ae1cecb7b73e204139db5279ba51cef4f1e8c4a4c491d8be403"} Dec 13 06:49:28 crc kubenswrapper[4971]: I1213 06:49:28.962992 4971 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-zlcg5" event={"ID":"736c6b1a-93dd-4f71-81b6-1c43fcf5f556","Type":"ContainerStarted","Data":"f6073140d9f86929b6d4bd2960ede0617a5c9620b269bbab043bae37e0d389f5"} Dec 13 06:49:28 crc kubenswrapper[4971]: I1213 06:49:28.970087 4971 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-13T06:49:24Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-13T06:49:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-13T06:49:24Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d0b2ecf0d5b22f23217f42f259ae7ffb6359555b22c30eaebf667eaf254b3bab\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-13T06:49:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-13T06:49:28Z is after 2025-08-24T17:21:41Z" Dec 13 06:49:28 crc kubenswrapper[4971]: I1213 06:49:28.985622 4971 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-13T06:49:23Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-13T06:49:23Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-13T06:49:28Z is after 2025-08-24T17:21:41Z" Dec 13 06:49:28 crc kubenswrapper[4971]: I1213 06:49:28.998972 4971 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-13T06:49:26Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-13T06:49:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-13T06:49:26Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9c8a0d1db0e754462d805e82daf5f52ffd5457507a30456535ec59e096c9d092\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-13T06:49:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-13T06:49:28Z is after 2025-08-24T17:21:41Z" Dec 13 06:49:29 crc kubenswrapper[4971]: I1213 06:49:29.016571 4971 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-82xjz" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e0c4af12-edda-41d7-9b44-f87396174ef2\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T06:49:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T06:49:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T06:49:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T06:49:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://0f4577d9e37f860813e221f2e208177d8d383fd04f673ce357a0a405ea29d1e6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-13T06:49:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cnxsq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://7e0d2020e62f1a277cff7047e1907dc747ed393a1f6c983e1ee258bd156c16c7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-13T06:49:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cnxsq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-13T06:49:25Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-82xjz\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-13T06:49:29Z is after 2025-08-24T17:21:41Z" Dec 13 06:49:29 crc kubenswrapper[4971]: I1213 06:49:29.030110 4971 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-wwvpl" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"dddf744b-2122-49ad-a980-105c09636e0f\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T06:49:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T06:49:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T06:49:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T06:49:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4676d897c7fc668587faec563f3551a056557879cee3e865192d2bd28edf62fc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-13T06:49:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-czz9v\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-13T06:49:25Z\\\"}}\" for pod \"openshift-multus\"/\"multus-wwvpl\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-13T06:49:29Z is after 2025-08-24T17:21:41Z" Dec 13 06:49:29 crc kubenswrapper[4971]: I1213 06:49:29.051509 4971 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-zlcg5" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"736c6b1a-93dd-4f71-81b6-1c43fcf5f556\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T06:49:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T06:49:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T06:49:25Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T06:49:25Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vzmsr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vzmsr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vzmsr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vzmsr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vzmsr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vzmsr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vzmsr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vzmsr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://451b22df6973ddc8fa2209a65d64d9e41b33eb231bb4adfab88cf0e5943bb246\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://451b22df6973ddc8fa2209a65d64d9e41b33eb231bb4adfab88cf0e5943bb246\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-13T06:49:26Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-13T06:49:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vzmsr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-13T06:49:25Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-zlcg5\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-13T06:49:29Z is after 2025-08-24T17:21:41Z" Dec 13 06:49:29 crc kubenswrapper[4971]: I1213 06:49:29.063898 4971 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"f9d31b5a-b770-4a15-95c8-35b49677c608\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T06:49:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T06:49:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T06:49:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T06:49:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T06:49:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c9baa042f9dada954faa57b5454cd6b0a12a297e56ddc192c52cff50e42a16fa\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-13T06:49:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://c788de86754f0bf37ee184672fc1dcb388ccebc246d4f15be4b5da661f725f8f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-13T06:49:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://760817511a9b4f96a68dc38d26631250eaab72e1dd8f123e653039b6b4e91f84\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-13T06:49:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b859be0c9feac856e9855c6a038b02b49b32fade97873c46d0bec0a32d8b5147\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-13T06:49:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-13T06:49:03Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-13T06:49:29Z is after 2025-08-24T17:21:41Z" Dec 13 06:49:29 crc kubenswrapper[4971]: I1213 06:49:29.078305 4971 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-ncchz" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"de914d21-a688-401e-a66e-861beb6336d9\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T06:49:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T06:49:25Z\\\",\\\"message\\\":\\\"containers with incomplete status: [cni-plugins bond-cni-plugin routeoverride-cni whereabouts-cni-bincopy whereabouts-cni]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T06:49:25Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T06:49:25Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lqmjj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://0b7111af6c5a198d02bb02858837cfb4f3b20828d8d206cd18aa0ffa102d0cd4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://0b7111af6c5a198d02bb02858837cfb4f3b20828d8d206cd18aa0ffa102d0cd4\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-13T06:49:26Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-13T06:49:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lqmjj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lqmjj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lqmjj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lqmjj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lqmjj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lqmjj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-13T06:49:25Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-ncchz\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-13T06:49:29Z is after 2025-08-24T17:21:41Z" Dec 13 06:49:29 crc kubenswrapper[4971]: I1213 06:49:29.089416 4971 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-gkcd8" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"13f5c327-f95e-42b4-8783-218500c17694\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T06:49:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T06:49:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T06:49:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T06:49:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://77e846e36f1cb40d03471dc03de11b6c3da448baa407aad38093fae9a801cf92\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-13T06:49:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gvlhn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-13T06:49:27Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-gkcd8\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-13T06:49:29Z is after 2025-08-24T17:21:41Z" Dec 13 06:49:29 crc kubenswrapper[4971]: I1213 06:49:29.107591 4971 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"1ed9f82e-fe5c-4b76-a619-62e5abd88d3b\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T06:49:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T06:49:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T06:49:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T06:49:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T06:49:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e457329a78a2d24337f994e8a0a1f75a2a667244aa75eb84613d1e7138364aac\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-13T06:49:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://eb014de8140deef097d9bd94f3b9fa4ab3558bc4449bef84b83bf62bfabf5836\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-13T06:49:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://2c438cc74d555f7c9db4569d1c56b78b1088ed8105697d2c0f31e6c5136f3dd0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-13T06:49:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://aa6fa6e4340dcd066252df3c56d79a3a39ed531d5eea8d1d5941d6952368d3af\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-13T06:49:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://c4471b98854c52d5cd084d0ba3e64e5b08e9518ec3c372cce2c875a440522457\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-13T06:49:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7d2746335c98dae3a0f43024d34cb62ebcb42a01c8f9ce18d52b29d4c45598e5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://7d2746335c98dae3a0f43024d34cb62ebcb42a01c8f9ce18d52b29d4c45598e5\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-13T06:49:04Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-13T06:49:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://bc0e77d10a738128bea82122496bc42235fc1688f8907756d93e61b651902b57\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://bc0e77d10a738128bea82122496bc42235fc1688f8907756d93e61b651902b57\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-13T06:49:05Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-13T06:49:05Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://f821e418cba5e832da8839268bcbbd9630bdf319a212d5382bfa2c4b0b73a268\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://f821e418cba5e832da8839268bcbbd9630bdf319a212d5382bfa2c4b0b73a268\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-13T06:49:07Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-13T06:49:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-13T06:49:03Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-13T06:49:29Z is after 2025-08-24T17:21:41Z" Dec 13 06:49:29 crc kubenswrapper[4971]: I1213 06:49:29.120586 4971 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-13T06:49:24Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-13T06:49:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-13T06:49:24Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4328cd6a8464ed468229bbfbc8265f969427916d72ebb92237dbe68efe70fa47\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-13T06:49:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://9d715ca82d9b5a594e41f2b98da6f4277da0e4e1dec114969f901c00f6d7553d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-13T06:49:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-13T06:49:29Z is after 2025-08-24T17:21:41Z" Dec 13 06:49:29 crc kubenswrapper[4971]: I1213 06:49:29.131088 4971 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-5rzh5" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"6bf4591d-2483-4132-9672-d0fd6c995968\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T06:49:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T06:49:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T06:49:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T06:49:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://50e2cb14e4ffb603acb06e2281615a99f06182d345c4e70f2446fa2077033bbb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-13T06:49:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dfxqn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-13T06:49:25Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-5rzh5\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-13T06:49:29Z is after 2025-08-24T17:21:41Z" Dec 13 06:49:29 crc kubenswrapper[4971]: I1213 06:49:29.159141 4971 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3ad3268a-a982-428c-9c06-4aaeddd372d9\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T06:49:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T06:49:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T06:49:03Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T06:49:03Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T06:49:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3716afdf2bc693be80a09fc9f7727d334f435caf1d4c8ec4811b3a1f1652cdfe\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-13T06:49:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://30df3abf8c79db8af52c6a918f73e1ce410ddafcf341636424c267e1eeb195a7\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-13T06:49:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://ccdf65f6d2b1eb3647a03ba37c305c32e34addfdf79ff937c8dddaa3e5d5b5c6\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-13T06:49:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://770b73dd0c2b6f38744e2c1cd6aa186a69bb87d1cd015a603ad2e33be06a8f6b\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://b3c1bab3d53f975b2e24582ce2771bc26d600e0dd7c30339dcfd18a071162ad4\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-12-13T06:49:23Z\\\",\\\"message\\\":\\\"file observer\\\\nW1213 06:49:23.283403 1 builder.go:272] unable to get owner reference (falling back to namespace): pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\nI1213 06:49:23.284336 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1213 06:49:23.286978 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-981773671/tls.crt::/tmp/serving-cert-981773671/tls.key\\\\\\\"\\\\nI1213 06:49:23.775072 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1213 06:49:23.778615 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1213 06:49:23.778645 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1213 06:49:23.778680 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1213 06:49:23.778691 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1213 06:49:23.785058 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1213 06:49:23.786150 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1213 06:49:23.786196 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1213 06:49:23.786223 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1213 06:49:23.786246 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1213 06:49:23.786269 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1213 06:49:23.786291 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI1213 06:49:23.785608 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF1213 06:49:23.787563 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-13T06:49:08Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-13T06:49:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://9acd07066deed483bd7fa61ee45fda09539d110717a134a56827f871186f92fc\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-13T06:49:07Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://38c91cfbaecfdbd3e3b663422a91ab4294610afb6bbf9ebdb77db24e2b76c682\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://38c91cfbaecfdbd3e3b663422a91ab4294610afb6bbf9ebdb77db24e2b76c682\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-13T06:49:05Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-13T06:49:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-13T06:49:03Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-13T06:49:29Z is after 2025-08-24T17:21:41Z" Dec 13 06:49:29 crc kubenswrapper[4971]: I1213 06:49:29.197884 4971 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-13T06:49:23Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-13T06:49:23Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-13T06:49:29Z is after 2025-08-24T17:21:41Z" Dec 13 06:49:29 crc kubenswrapper[4971]: I1213 06:49:29.238009 4971 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-13T06:49:23Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-13T06:49:23Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-13T06:49:29Z is after 2025-08-24T17:21:41Z" Dec 13 06:49:29 crc kubenswrapper[4971]: I1213 06:49:29.287551 4971 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-ncchz" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"de914d21-a688-401e-a66e-861beb6336d9\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T06:49:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T06:49:25Z\\\",\\\"message\\\":\\\"containers with incomplete status: [bond-cni-plugin routeoverride-cni whereabouts-cni-bincopy whereabouts-cni]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T06:49:25Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T06:49:25Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lqmjj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://0b7111af6c5a198d02bb02858837cfb4f3b20828d8d206cd18aa0ffa102d0cd4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://0b7111af6c5a198d02bb02858837cfb4f3b20828d8d206cd18aa0ffa102d0cd4\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-13T06:49:26Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-13T06:49:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lqmjj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b36777fe40e74ae1cecb7b73e204139db5279ba51cef4f1e8c4a4c491d8be403\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://b36777fe40e74ae1cecb7b73e204139db5279ba51cef4f1e8c4a4c491d8be403\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-13T06:49:28Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-13T06:49:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lqmjj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lqmjj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lqmjj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lqmjj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lqmjj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-13T06:49:25Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-ncchz\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-13T06:49:29Z is after 2025-08-24T17:21:41Z" Dec 13 06:49:29 crc kubenswrapper[4971]: I1213 06:49:29.318224 4971 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-gkcd8" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"13f5c327-f95e-42b4-8783-218500c17694\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T06:49:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T06:49:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T06:49:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T06:49:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://77e846e36f1cb40d03471dc03de11b6c3da448baa407aad38093fae9a801cf92\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-13T06:49:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gvlhn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-13T06:49:27Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-gkcd8\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-13T06:49:29Z is after 2025-08-24T17:21:41Z" Dec 13 06:49:29 crc kubenswrapper[4971]: I1213 06:49:29.357053 4971 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"f9d31b5a-b770-4a15-95c8-35b49677c608\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T06:49:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T06:49:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T06:49:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T06:49:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T06:49:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c9baa042f9dada954faa57b5454cd6b0a12a297e56ddc192c52cff50e42a16fa\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-13T06:49:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://c788de86754f0bf37ee184672fc1dcb388ccebc246d4f15be4b5da661f725f8f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-13T06:49:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://760817511a9b4f96a68dc38d26631250eaab72e1dd8f123e653039b6b4e91f84\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-13T06:49:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b859be0c9feac856e9855c6a038b02b49b32fade97873c46d0bec0a32d8b5147\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-13T06:49:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-13T06:49:03Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-13T06:49:29Z is after 2025-08-24T17:21:41Z" Dec 13 06:49:29 crc kubenswrapper[4971]: I1213 06:49:29.398478 4971 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-13T06:49:24Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-13T06:49:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-13T06:49:24Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4328cd6a8464ed468229bbfbc8265f969427916d72ebb92237dbe68efe70fa47\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-13T06:49:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://9d715ca82d9b5a594e41f2b98da6f4277da0e4e1dec114969f901c00f6d7553d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-13T06:49:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-13T06:49:29Z is after 2025-08-24T17:21:41Z" Dec 13 06:49:29 crc kubenswrapper[4971]: I1213 06:49:29.435396 4971 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-5rzh5" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"6bf4591d-2483-4132-9672-d0fd6c995968\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T06:49:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T06:49:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T06:49:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T06:49:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://50e2cb14e4ffb603acb06e2281615a99f06182d345c4e70f2446fa2077033bbb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-13T06:49:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dfxqn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-13T06:49:25Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-5rzh5\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-13T06:49:29Z is after 2025-08-24T17:21:41Z" Dec 13 06:49:29 crc kubenswrapper[4971]: I1213 06:49:29.484441 4971 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"1ed9f82e-fe5c-4b76-a619-62e5abd88d3b\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T06:49:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T06:49:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T06:49:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T06:49:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T06:49:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e457329a78a2d24337f994e8a0a1f75a2a667244aa75eb84613d1e7138364aac\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-13T06:49:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://eb014de8140deef097d9bd94f3b9fa4ab3558bc4449bef84b83bf62bfabf5836\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-13T06:49:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://2c438cc74d555f7c9db4569d1c56b78b1088ed8105697d2c0f31e6c5136f3dd0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-13T06:49:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://aa6fa6e4340dcd066252df3c56d79a3a39ed531d5eea8d1d5941d6952368d3af\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-13T06:49:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://c4471b98854c52d5cd084d0ba3e64e5b08e9518ec3c372cce2c875a440522457\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-13T06:49:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7d2746335c98dae3a0f43024d34cb62ebcb42a01c8f9ce18d52b29d4c45598e5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://7d2746335c98dae3a0f43024d34cb62ebcb42a01c8f9ce18d52b29d4c45598e5\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-13T06:49:04Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-13T06:49:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://bc0e77d10a738128bea82122496bc42235fc1688f8907756d93e61b651902b57\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://bc0e77d10a738128bea82122496bc42235fc1688f8907756d93e61b651902b57\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-13T06:49:05Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-13T06:49:05Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://f821e418cba5e832da8839268bcbbd9630bdf319a212d5382bfa2c4b0b73a268\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://f821e418cba5e832da8839268bcbbd9630bdf319a212d5382bfa2c4b0b73a268\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-13T06:49:07Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-13T06:49:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-13T06:49:03Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-13T06:49:29Z is after 2025-08-24T17:21:41Z" Dec 13 06:49:29 crc kubenswrapper[4971]: I1213 06:49:29.519357 4971 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-13T06:49:23Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-13T06:49:23Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-13T06:49:29Z is after 2025-08-24T17:21:41Z" Dec 13 06:49:29 crc kubenswrapper[4971]: I1213 06:49:29.559299 4971 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3ad3268a-a982-428c-9c06-4aaeddd372d9\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T06:49:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T06:49:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T06:49:03Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T06:49:03Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T06:49:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3716afdf2bc693be80a09fc9f7727d334f435caf1d4c8ec4811b3a1f1652cdfe\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-13T06:49:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://30df3abf8c79db8af52c6a918f73e1ce410ddafcf341636424c267e1eeb195a7\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-13T06:49:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://ccdf65f6d2b1eb3647a03ba37c305c32e34addfdf79ff937c8dddaa3e5d5b5c6\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-13T06:49:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://770b73dd0c2b6f38744e2c1cd6aa186a69bb87d1cd015a603ad2e33be06a8f6b\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://b3c1bab3d53f975b2e24582ce2771bc26d600e0dd7c30339dcfd18a071162ad4\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-12-13T06:49:23Z\\\",\\\"message\\\":\\\"file observer\\\\nW1213 06:49:23.283403 1 builder.go:272] unable to get owner reference (falling back to namespace): pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\nI1213 06:49:23.284336 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1213 06:49:23.286978 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-981773671/tls.crt::/tmp/serving-cert-981773671/tls.key\\\\\\\"\\\\nI1213 06:49:23.775072 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1213 06:49:23.778615 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1213 06:49:23.778645 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1213 06:49:23.778680 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1213 06:49:23.778691 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1213 06:49:23.785058 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1213 06:49:23.786150 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1213 06:49:23.786196 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1213 06:49:23.786223 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1213 06:49:23.786246 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1213 06:49:23.786269 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1213 06:49:23.786291 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI1213 06:49:23.785608 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF1213 06:49:23.787563 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-13T06:49:08Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-13T06:49:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://9acd07066deed483bd7fa61ee45fda09539d110717a134a56827f871186f92fc\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-13T06:49:07Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://38c91cfbaecfdbd3e3b663422a91ab4294610afb6bbf9ebdb77db24e2b76c682\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://38c91cfbaecfdbd3e3b663422a91ab4294610afb6bbf9ebdb77db24e2b76c682\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-13T06:49:05Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-13T06:49:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-13T06:49:03Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-13T06:49:29Z is after 2025-08-24T17:21:41Z" Dec 13 06:49:29 crc kubenswrapper[4971]: I1213 06:49:29.599597 4971 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-13T06:49:23Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-13T06:49:23Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-13T06:49:29Z is after 2025-08-24T17:21:41Z" Dec 13 06:49:29 crc kubenswrapper[4971]: I1213 06:49:29.638332 4971 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-13T06:49:26Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-13T06:49:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-13T06:49:26Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9c8a0d1db0e754462d805e82daf5f52ffd5457507a30456535ec59e096c9d092\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-13T06:49:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-13T06:49:29Z is after 2025-08-24T17:21:41Z" Dec 13 06:49:29 crc kubenswrapper[4971]: I1213 06:49:29.667314 4971 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Dec 13 06:49:29 crc kubenswrapper[4971]: I1213 06:49:29.669395 4971 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 13 06:49:29 crc kubenswrapper[4971]: I1213 06:49:29.669613 4971 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 13 06:49:29 crc kubenswrapper[4971]: I1213 06:49:29.669712 4971 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 13 06:49:29 crc kubenswrapper[4971]: I1213 06:49:29.669911 4971 kubelet_node_status.go:76] "Attempting to register node" node="crc" Dec 13 06:49:29 crc kubenswrapper[4971]: I1213 06:49:29.678432 4971 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-82xjz" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e0c4af12-edda-41d7-9b44-f87396174ef2\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T06:49:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T06:49:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T06:49:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T06:49:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://0f4577d9e37f860813e221f2e208177d8d383fd04f673ce357a0a405ea29d1e6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-13T06:49:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cnxsq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://7e0d2020e62f1a277cff7047e1907dc747ed393a1f6c983e1ee258bd156c16c7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-13T06:49:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cnxsq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-13T06:49:25Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-82xjz\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-13T06:49:29Z is after 2025-08-24T17:21:41Z" Dec 13 06:49:29 crc kubenswrapper[4971]: I1213 06:49:29.730766 4971 kubelet_node_status.go:115] "Node was previously registered" node="crc" Dec 13 06:49:29 crc kubenswrapper[4971]: I1213 06:49:29.731077 4971 kubelet_node_status.go:79] "Successfully registered node" node="crc" Dec 13 06:49:29 crc kubenswrapper[4971]: I1213 06:49:29.732621 4971 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 13 06:49:29 crc kubenswrapper[4971]: I1213 06:49:29.732679 4971 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 13 06:49:29 crc kubenswrapper[4971]: I1213 06:49:29.732692 4971 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 13 06:49:29 crc kubenswrapper[4971]: I1213 06:49:29.732709 4971 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 13 06:49:29 crc kubenswrapper[4971]: I1213 06:49:29.732723 4971 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-13T06:49:29Z","lastTransitionTime":"2025-12-13T06:49:29Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 13 06:49:29 crc kubenswrapper[4971]: E1213 06:49:29.743024 4971 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-12-13T06:49:29Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-13T06:49:29Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-13T06:49:29Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-13T06:49:29Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-13T06:49:29Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-13T06:49:29Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-13T06:49:29Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-13T06:49:29Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"bd36cbc9-bbe9-45a1-a21a-ad2bd10499ee\\\",\\\"systemUUID\\\":\\\"a9af00dc-cf25-47af-9997-79577f8f3bea\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-13T06:49:29Z is after 2025-08-24T17:21:41Z" Dec 13 06:49:29 crc kubenswrapper[4971]: I1213 06:49:29.746293 4971 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 13 06:49:29 crc kubenswrapper[4971]: I1213 06:49:29.746336 4971 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 13 06:49:29 crc kubenswrapper[4971]: I1213 06:49:29.746347 4971 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 13 06:49:29 crc kubenswrapper[4971]: I1213 06:49:29.746364 4971 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 13 06:49:29 crc kubenswrapper[4971]: I1213 06:49:29.746377 4971 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-13T06:49:29Z","lastTransitionTime":"2025-12-13T06:49:29Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 13 06:49:29 crc kubenswrapper[4971]: I1213 06:49:29.757234 4971 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-wwvpl" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"dddf744b-2122-49ad-a980-105c09636e0f\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T06:49:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T06:49:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T06:49:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T06:49:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4676d897c7fc668587faec563f3551a056557879cee3e865192d2bd28edf62fc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-13T06:49:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-czz9v\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-13T06:49:25Z\\\"}}\" for pod \"openshift-multus\"/\"multus-wwvpl\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-13T06:49:29Z is after 2025-08-24T17:21:41Z" Dec 13 06:49:29 crc kubenswrapper[4971]: E1213 06:49:29.759583 4971 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-12-13T06:49:29Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-13T06:49:29Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-13T06:49:29Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-13T06:49:29Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-13T06:49:29Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-13T06:49:29Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-13T06:49:29Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-13T06:49:29Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"bd36cbc9-bbe9-45a1-a21a-ad2bd10499ee\\\",\\\"systemUUID\\\":\\\"a9af00dc-cf25-47af-9997-79577f8f3bea\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-13T06:49:29Z is after 2025-08-24T17:21:41Z" Dec 13 06:49:29 crc kubenswrapper[4971]: I1213 06:49:29.763182 4971 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 13 06:49:29 crc kubenswrapper[4971]: I1213 06:49:29.763215 4971 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 13 06:49:29 crc kubenswrapper[4971]: I1213 06:49:29.763223 4971 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 13 06:49:29 crc kubenswrapper[4971]: I1213 06:49:29.763238 4971 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 13 06:49:29 crc kubenswrapper[4971]: I1213 06:49:29.763249 4971 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-13T06:49:29Z","lastTransitionTime":"2025-12-13T06:49:29Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 13 06:49:29 crc kubenswrapper[4971]: I1213 06:49:29.767531 4971 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 13 06:49:29 crc kubenswrapper[4971]: I1213 06:49:29.767572 4971 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 13 06:49:29 crc kubenswrapper[4971]: I1213 06:49:29.767576 4971 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 13 06:49:29 crc kubenswrapper[4971]: E1213 06:49:29.767656 4971 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 13 06:49:29 crc kubenswrapper[4971]: E1213 06:49:29.767739 4971 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 13 06:49:29 crc kubenswrapper[4971]: E1213 06:49:29.767836 4971 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 13 06:49:29 crc kubenswrapper[4971]: E1213 06:49:29.775089 4971 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-12-13T06:49:29Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-13T06:49:29Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-13T06:49:29Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-13T06:49:29Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-13T06:49:29Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-13T06:49:29Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-13T06:49:29Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-13T06:49:29Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"bd36cbc9-bbe9-45a1-a21a-ad2bd10499ee\\\",\\\"systemUUID\\\":\\\"a9af00dc-cf25-47af-9997-79577f8f3bea\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-13T06:49:29Z is after 2025-08-24T17:21:41Z" Dec 13 06:49:29 crc kubenswrapper[4971]: I1213 06:49:29.778370 4971 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 13 06:49:29 crc kubenswrapper[4971]: I1213 06:49:29.778406 4971 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 13 06:49:29 crc kubenswrapper[4971]: I1213 06:49:29.778419 4971 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 13 06:49:29 crc kubenswrapper[4971]: I1213 06:49:29.778439 4971 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 13 06:49:29 crc kubenswrapper[4971]: I1213 06:49:29.778450 4971 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-13T06:49:29Z","lastTransitionTime":"2025-12-13T06:49:29Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 13 06:49:29 crc kubenswrapper[4971]: E1213 06:49:29.789276 4971 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-12-13T06:49:29Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-13T06:49:29Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-13T06:49:29Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-13T06:49:29Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-13T06:49:29Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-13T06:49:29Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-13T06:49:29Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-13T06:49:29Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"bd36cbc9-bbe9-45a1-a21a-ad2bd10499ee\\\",\\\"systemUUID\\\":\\\"a9af00dc-cf25-47af-9997-79577f8f3bea\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-13T06:49:29Z is after 2025-08-24T17:21:41Z" Dec 13 06:49:29 crc kubenswrapper[4971]: I1213 06:49:29.792447 4971 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 13 06:49:29 crc kubenswrapper[4971]: I1213 06:49:29.792487 4971 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 13 06:49:29 crc kubenswrapper[4971]: I1213 06:49:29.792497 4971 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 13 06:49:29 crc kubenswrapper[4971]: I1213 06:49:29.792541 4971 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 13 06:49:29 crc kubenswrapper[4971]: I1213 06:49:29.792560 4971 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-13T06:49:29Z","lastTransitionTime":"2025-12-13T06:49:29Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 13 06:49:29 crc kubenswrapper[4971]: I1213 06:49:29.804240 4971 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-zlcg5" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"736c6b1a-93dd-4f71-81b6-1c43fcf5f556\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T06:49:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T06:49:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T06:49:25Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T06:49:25Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vzmsr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vzmsr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vzmsr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vzmsr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vzmsr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vzmsr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vzmsr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vzmsr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://451b22df6973ddc8fa2209a65d64d9e41b33eb231bb4adfab88cf0e5943bb246\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://451b22df6973ddc8fa2209a65d64d9e41b33eb231bb4adfab88cf0e5943bb246\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-13T06:49:26Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-13T06:49:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vzmsr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-13T06:49:25Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-zlcg5\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-13T06:49:29Z is after 2025-08-24T17:21:41Z" Dec 13 06:49:29 crc kubenswrapper[4971]: E1213 06:49:29.806104 4971 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-12-13T06:49:29Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-13T06:49:29Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-13T06:49:29Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-13T06:49:29Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-13T06:49:29Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-13T06:49:29Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-13T06:49:29Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-13T06:49:29Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"bd36cbc9-bbe9-45a1-a21a-ad2bd10499ee\\\",\\\"systemUUID\\\":\\\"a9af00dc-cf25-47af-9997-79577f8f3bea\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-13T06:49:29Z is after 2025-08-24T17:21:41Z" Dec 13 06:49:29 crc kubenswrapper[4971]: E1213 06:49:29.806204 4971 kubelet_node_status.go:572] "Unable to update node status" err="update node status exceeds retry count" Dec 13 06:49:29 crc kubenswrapper[4971]: I1213 06:49:29.807571 4971 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 13 06:49:29 crc kubenswrapper[4971]: I1213 06:49:29.807599 4971 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 13 06:49:29 crc kubenswrapper[4971]: I1213 06:49:29.807608 4971 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 13 06:49:29 crc kubenswrapper[4971]: I1213 06:49:29.807620 4971 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 13 06:49:29 crc kubenswrapper[4971]: I1213 06:49:29.807629 4971 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-13T06:49:29Z","lastTransitionTime":"2025-12-13T06:49:29Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 13 06:49:29 crc kubenswrapper[4971]: I1213 06:49:29.840037 4971 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-13T06:49:24Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-13T06:49:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-13T06:49:24Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d0b2ecf0d5b22f23217f42f259ae7ffb6359555b22c30eaebf667eaf254b3bab\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-13T06:49:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-13T06:49:29Z is after 2025-08-24T17:21:41Z" Dec 13 06:49:29 crc kubenswrapper[4971]: I1213 06:49:29.878059 4971 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-13T06:49:23Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-13T06:49:23Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-13T06:49:29Z is after 2025-08-24T17:21:41Z" Dec 13 06:49:29 crc kubenswrapper[4971]: I1213 06:49:29.913469 4971 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 13 06:49:29 crc kubenswrapper[4971]: I1213 06:49:29.913543 4971 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 13 06:49:29 crc kubenswrapper[4971]: I1213 06:49:29.913556 4971 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 13 06:49:29 crc kubenswrapper[4971]: I1213 06:49:29.913573 4971 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 13 06:49:29 crc kubenswrapper[4971]: I1213 06:49:29.913584 4971 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-13T06:49:29Z","lastTransitionTime":"2025-12-13T06:49:29Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 13 06:49:29 crc kubenswrapper[4971]: I1213 06:49:29.967379 4971 generic.go:334] "Generic (PLEG): container finished" podID="de914d21-a688-401e-a66e-861beb6336d9" containerID="0597a06cc663e416142311a16941d98297577693a1fe27201ea42c27496cace2" exitCode=0 Dec 13 06:49:29 crc kubenswrapper[4971]: I1213 06:49:29.967421 4971 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-additional-cni-plugins-ncchz" event={"ID":"de914d21-a688-401e-a66e-861beb6336d9","Type":"ContainerDied","Data":"0597a06cc663e416142311a16941d98297577693a1fe27201ea42c27496cace2"} Dec 13 06:49:29 crc kubenswrapper[4971]: I1213 06:49:29.984949 4971 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-13T06:49:24Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-13T06:49:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-13T06:49:24Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d0b2ecf0d5b22f23217f42f259ae7ffb6359555b22c30eaebf667eaf254b3bab\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-13T06:49:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-13T06:49:29Z is after 2025-08-24T17:21:41Z" Dec 13 06:49:29 crc kubenswrapper[4971]: I1213 06:49:29.998713 4971 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-13T06:49:23Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-13T06:49:23Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-13T06:49:29Z is after 2025-08-24T17:21:41Z" Dec 13 06:49:30 crc kubenswrapper[4971]: I1213 06:49:30.011832 4971 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-13T06:49:26Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-13T06:49:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-13T06:49:26Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9c8a0d1db0e754462d805e82daf5f52ffd5457507a30456535ec59e096c9d092\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-13T06:49:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-13T06:49:30Z is after 2025-08-24T17:21:41Z" Dec 13 06:49:30 crc kubenswrapper[4971]: I1213 06:49:30.017266 4971 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 13 06:49:30 crc kubenswrapper[4971]: I1213 06:49:30.017328 4971 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 13 06:49:30 crc kubenswrapper[4971]: I1213 06:49:30.017341 4971 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 13 06:49:30 crc kubenswrapper[4971]: I1213 06:49:30.017360 4971 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 13 06:49:30 crc kubenswrapper[4971]: I1213 06:49:30.017385 4971 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-13T06:49:30Z","lastTransitionTime":"2025-12-13T06:49:30Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 13 06:49:30 crc kubenswrapper[4971]: I1213 06:49:30.037428 4971 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-82xjz" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e0c4af12-edda-41d7-9b44-f87396174ef2\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T06:49:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T06:49:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T06:49:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T06:49:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://0f4577d9e37f860813e221f2e208177d8d383fd04f673ce357a0a405ea29d1e6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-13T06:49:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cnxsq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://7e0d2020e62f1a277cff7047e1907dc747ed393a1f6c983e1ee258bd156c16c7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-13T06:49:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cnxsq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-13T06:49:25Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-82xjz\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-13T06:49:30Z is after 2025-08-24T17:21:41Z" Dec 13 06:49:30 crc kubenswrapper[4971]: I1213 06:49:30.081405 4971 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-wwvpl" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"dddf744b-2122-49ad-a980-105c09636e0f\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T06:49:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T06:49:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T06:49:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T06:49:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4676d897c7fc668587faec563f3551a056557879cee3e865192d2bd28edf62fc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-13T06:49:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-czz9v\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-13T06:49:25Z\\\"}}\" for pod \"openshift-multus\"/\"multus-wwvpl\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-13T06:49:30Z is after 2025-08-24T17:21:41Z" Dec 13 06:49:30 crc kubenswrapper[4971]: I1213 06:49:30.119382 4971 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 13 06:49:30 crc kubenswrapper[4971]: I1213 06:49:30.119426 4971 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 13 06:49:30 crc kubenswrapper[4971]: I1213 06:49:30.119435 4971 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 13 06:49:30 crc kubenswrapper[4971]: I1213 06:49:30.119448 4971 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 13 06:49:30 crc kubenswrapper[4971]: I1213 06:49:30.119456 4971 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-13T06:49:30Z","lastTransitionTime":"2025-12-13T06:49:30Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 13 06:49:30 crc kubenswrapper[4971]: I1213 06:49:30.162208 4971 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-zlcg5" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"736c6b1a-93dd-4f71-81b6-1c43fcf5f556\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T06:49:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T06:49:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T06:49:25Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T06:49:25Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vzmsr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vzmsr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vzmsr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vzmsr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vzmsr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vzmsr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vzmsr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vzmsr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://451b22df6973ddc8fa2209a65d64d9e41b33eb231bb4adfab88cf0e5943bb246\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://451b22df6973ddc8fa2209a65d64d9e41b33eb231bb4adfab88cf0e5943bb246\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-13T06:49:26Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-13T06:49:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vzmsr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-13T06:49:25Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-zlcg5\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-13T06:49:30Z is after 2025-08-24T17:21:41Z" Dec 13 06:49:30 crc kubenswrapper[4971]: I1213 06:49:30.179658 4971 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"f9d31b5a-b770-4a15-95c8-35b49677c608\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T06:49:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T06:49:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T06:49:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T06:49:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T06:49:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c9baa042f9dada954faa57b5454cd6b0a12a297e56ddc192c52cff50e42a16fa\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-13T06:49:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://c788de86754f0bf37ee184672fc1dcb388ccebc246d4f15be4b5da661f725f8f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-13T06:49:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://760817511a9b4f96a68dc38d26631250eaab72e1dd8f123e653039b6b4e91f84\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-13T06:49:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b859be0c9feac856e9855c6a038b02b49b32fade97873c46d0bec0a32d8b5147\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-13T06:49:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-13T06:49:03Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-13T06:49:30Z is after 2025-08-24T17:21:41Z" Dec 13 06:49:30 crc kubenswrapper[4971]: I1213 06:49:30.202292 4971 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-ncchz" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"de914d21-a688-401e-a66e-861beb6336d9\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T06:49:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T06:49:25Z\\\",\\\"message\\\":\\\"containers with incomplete status: [routeoverride-cni whereabouts-cni-bincopy whereabouts-cni]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T06:49:25Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T06:49:25Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lqmjj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://0b7111af6c5a198d02bb02858837cfb4f3b20828d8d206cd18aa0ffa102d0cd4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://0b7111af6c5a198d02bb02858837cfb4f3b20828d8d206cd18aa0ffa102d0cd4\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-13T06:49:26Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-13T06:49:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lqmjj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b36777fe40e74ae1cecb7b73e204139db5279ba51cef4f1e8c4a4c491d8be403\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://b36777fe40e74ae1cecb7b73e204139db5279ba51cef4f1e8c4a4c491d8be403\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-13T06:49:28Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-13T06:49:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lqmjj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://0597a06cc663e416142311a16941d98297577693a1fe27201ea42c27496cace2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://0597a06cc663e416142311a16941d98297577693a1fe27201ea42c27496cace2\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-13T06:49:29Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-13T06:49:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lqmjj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lqmjj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lqmjj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lqmjj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-13T06:49:25Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-ncchz\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-13T06:49:30Z is after 2025-08-24T17:21:41Z" Dec 13 06:49:30 crc kubenswrapper[4971]: I1213 06:49:30.221643 4971 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 13 06:49:30 crc kubenswrapper[4971]: I1213 06:49:30.221676 4971 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 13 06:49:30 crc kubenswrapper[4971]: I1213 06:49:30.221685 4971 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 13 06:49:30 crc kubenswrapper[4971]: I1213 06:49:30.221698 4971 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 13 06:49:30 crc kubenswrapper[4971]: I1213 06:49:30.221708 4971 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-13T06:49:30Z","lastTransitionTime":"2025-12-13T06:49:30Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 13 06:49:30 crc kubenswrapper[4971]: I1213 06:49:30.237600 4971 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-gkcd8" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"13f5c327-f95e-42b4-8783-218500c17694\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T06:49:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T06:49:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T06:49:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T06:49:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://77e846e36f1cb40d03471dc03de11b6c3da448baa407aad38093fae9a801cf92\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-13T06:49:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gvlhn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-13T06:49:27Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-gkcd8\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-13T06:49:30Z is after 2025-08-24T17:21:41Z" Dec 13 06:49:30 crc kubenswrapper[4971]: I1213 06:49:30.285206 4971 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"1ed9f82e-fe5c-4b76-a619-62e5abd88d3b\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T06:49:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T06:49:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T06:49:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T06:49:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T06:49:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e457329a78a2d24337f994e8a0a1f75a2a667244aa75eb84613d1e7138364aac\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-13T06:49:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://eb014de8140deef097d9bd94f3b9fa4ab3558bc4449bef84b83bf62bfabf5836\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-13T06:49:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://2c438cc74d555f7c9db4569d1c56b78b1088ed8105697d2c0f31e6c5136f3dd0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-13T06:49:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://aa6fa6e4340dcd066252df3c56d79a3a39ed531d5eea8d1d5941d6952368d3af\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-13T06:49:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://c4471b98854c52d5cd084d0ba3e64e5b08e9518ec3c372cce2c875a440522457\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-13T06:49:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7d2746335c98dae3a0f43024d34cb62ebcb42a01c8f9ce18d52b29d4c45598e5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://7d2746335c98dae3a0f43024d34cb62ebcb42a01c8f9ce18d52b29d4c45598e5\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-13T06:49:04Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-13T06:49:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://bc0e77d10a738128bea82122496bc42235fc1688f8907756d93e61b651902b57\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://bc0e77d10a738128bea82122496bc42235fc1688f8907756d93e61b651902b57\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-13T06:49:05Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-13T06:49:05Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://f821e418cba5e832da8839268bcbbd9630bdf319a212d5382bfa2c4b0b73a268\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://f821e418cba5e832da8839268bcbbd9630bdf319a212d5382bfa2c4b0b73a268\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-13T06:49:07Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-13T06:49:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-13T06:49:03Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-13T06:49:30Z is after 2025-08-24T17:21:41Z" Dec 13 06:49:30 crc kubenswrapper[4971]: I1213 06:49:30.318821 4971 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-13T06:49:24Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-13T06:49:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-13T06:49:24Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4328cd6a8464ed468229bbfbc8265f969427916d72ebb92237dbe68efe70fa47\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-13T06:49:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://9d715ca82d9b5a594e41f2b98da6f4277da0e4e1dec114969f901c00f6d7553d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-13T06:49:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-13T06:49:30Z is after 2025-08-24T17:21:41Z" Dec 13 06:49:30 crc kubenswrapper[4971]: I1213 06:49:30.324449 4971 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 13 06:49:30 crc kubenswrapper[4971]: I1213 06:49:30.324476 4971 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 13 06:49:30 crc kubenswrapper[4971]: I1213 06:49:30.324484 4971 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 13 06:49:30 crc kubenswrapper[4971]: I1213 06:49:30.324497 4971 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 13 06:49:30 crc kubenswrapper[4971]: I1213 06:49:30.324507 4971 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-13T06:49:30Z","lastTransitionTime":"2025-12-13T06:49:30Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 13 06:49:30 crc kubenswrapper[4971]: I1213 06:49:30.356690 4971 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-5rzh5" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"6bf4591d-2483-4132-9672-d0fd6c995968\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T06:49:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T06:49:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T06:49:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T06:49:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://50e2cb14e4ffb603acb06e2281615a99f06182d345c4e70f2446fa2077033bbb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-13T06:49:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dfxqn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-13T06:49:25Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-5rzh5\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-13T06:49:30Z is after 2025-08-24T17:21:41Z" Dec 13 06:49:30 crc kubenswrapper[4971]: I1213 06:49:30.399293 4971 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3ad3268a-a982-428c-9c06-4aaeddd372d9\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T06:49:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T06:49:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T06:49:03Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T06:49:03Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T06:49:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3716afdf2bc693be80a09fc9f7727d334f435caf1d4c8ec4811b3a1f1652cdfe\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-13T06:49:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://30df3abf8c79db8af52c6a918f73e1ce410ddafcf341636424c267e1eeb195a7\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-13T06:49:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://ccdf65f6d2b1eb3647a03ba37c305c32e34addfdf79ff937c8dddaa3e5d5b5c6\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-13T06:49:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://770b73dd0c2b6f38744e2c1cd6aa186a69bb87d1cd015a603ad2e33be06a8f6b\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://b3c1bab3d53f975b2e24582ce2771bc26d600e0dd7c30339dcfd18a071162ad4\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-12-13T06:49:23Z\\\",\\\"message\\\":\\\"file observer\\\\nW1213 06:49:23.283403 1 builder.go:272] unable to get owner reference (falling back to namespace): pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\nI1213 06:49:23.284336 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1213 06:49:23.286978 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-981773671/tls.crt::/tmp/serving-cert-981773671/tls.key\\\\\\\"\\\\nI1213 06:49:23.775072 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1213 06:49:23.778615 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1213 06:49:23.778645 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1213 06:49:23.778680 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1213 06:49:23.778691 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1213 06:49:23.785058 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1213 06:49:23.786150 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1213 06:49:23.786196 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1213 06:49:23.786223 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1213 06:49:23.786246 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1213 06:49:23.786269 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1213 06:49:23.786291 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI1213 06:49:23.785608 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF1213 06:49:23.787563 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-13T06:49:08Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-13T06:49:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://9acd07066deed483bd7fa61ee45fda09539d110717a134a56827f871186f92fc\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-13T06:49:07Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://38c91cfbaecfdbd3e3b663422a91ab4294610afb6bbf9ebdb77db24e2b76c682\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://38c91cfbaecfdbd3e3b663422a91ab4294610afb6bbf9ebdb77db24e2b76c682\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-13T06:49:05Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-13T06:49:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-13T06:49:03Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-13T06:49:30Z is after 2025-08-24T17:21:41Z" Dec 13 06:49:30 crc kubenswrapper[4971]: I1213 06:49:30.426916 4971 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 13 06:49:30 crc kubenswrapper[4971]: I1213 06:49:30.426967 4971 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 13 06:49:30 crc kubenswrapper[4971]: I1213 06:49:30.426978 4971 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 13 06:49:30 crc kubenswrapper[4971]: I1213 06:49:30.426991 4971 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 13 06:49:30 crc kubenswrapper[4971]: I1213 06:49:30.427004 4971 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-13T06:49:30Z","lastTransitionTime":"2025-12-13T06:49:30Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 13 06:49:30 crc kubenswrapper[4971]: I1213 06:49:30.438740 4971 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-13T06:49:23Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-13T06:49:23Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-13T06:49:30Z is after 2025-08-24T17:21:41Z" Dec 13 06:49:30 crc kubenswrapper[4971]: I1213 06:49:30.476862 4971 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-13T06:49:23Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-13T06:49:23Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-13T06:49:30Z is after 2025-08-24T17:21:41Z" Dec 13 06:49:30 crc kubenswrapper[4971]: I1213 06:49:30.528583 4971 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 13 06:49:30 crc kubenswrapper[4971]: I1213 06:49:30.528628 4971 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 13 06:49:30 crc kubenswrapper[4971]: I1213 06:49:30.528639 4971 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 13 06:49:30 crc kubenswrapper[4971]: I1213 06:49:30.528657 4971 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 13 06:49:30 crc kubenswrapper[4971]: I1213 06:49:30.528667 4971 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-13T06:49:30Z","lastTransitionTime":"2025-12-13T06:49:30Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 13 06:49:30 crc kubenswrapper[4971]: I1213 06:49:30.631376 4971 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 13 06:49:30 crc kubenswrapper[4971]: I1213 06:49:30.631423 4971 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 13 06:49:30 crc kubenswrapper[4971]: I1213 06:49:30.631431 4971 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 13 06:49:30 crc kubenswrapper[4971]: I1213 06:49:30.631447 4971 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 13 06:49:30 crc kubenswrapper[4971]: I1213 06:49:30.631457 4971 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-13T06:49:30Z","lastTransitionTime":"2025-12-13T06:49:30Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 13 06:49:30 crc kubenswrapper[4971]: I1213 06:49:30.734172 4971 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 13 06:49:30 crc kubenswrapper[4971]: I1213 06:49:30.734219 4971 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 13 06:49:30 crc kubenswrapper[4971]: I1213 06:49:30.734233 4971 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 13 06:49:30 crc kubenswrapper[4971]: I1213 06:49:30.734249 4971 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 13 06:49:30 crc kubenswrapper[4971]: I1213 06:49:30.734260 4971 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-13T06:49:30Z","lastTransitionTime":"2025-12-13T06:49:30Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 13 06:49:30 crc kubenswrapper[4971]: I1213 06:49:30.836719 4971 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 13 06:49:30 crc kubenswrapper[4971]: I1213 06:49:30.836769 4971 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 13 06:49:30 crc kubenswrapper[4971]: I1213 06:49:30.836781 4971 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 13 06:49:30 crc kubenswrapper[4971]: I1213 06:49:30.836801 4971 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 13 06:49:30 crc kubenswrapper[4971]: I1213 06:49:30.836823 4971 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-13T06:49:30Z","lastTransitionTime":"2025-12-13T06:49:30Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 13 06:49:30 crc kubenswrapper[4971]: I1213 06:49:30.939318 4971 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 13 06:49:30 crc kubenswrapper[4971]: I1213 06:49:30.939355 4971 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 13 06:49:30 crc kubenswrapper[4971]: I1213 06:49:30.939364 4971 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 13 06:49:30 crc kubenswrapper[4971]: I1213 06:49:30.939378 4971 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 13 06:49:30 crc kubenswrapper[4971]: I1213 06:49:30.939387 4971 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-13T06:49:30Z","lastTransitionTime":"2025-12-13T06:49:30Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 13 06:49:30 crc kubenswrapper[4971]: I1213 06:49:30.972177 4971 generic.go:334] "Generic (PLEG): container finished" podID="de914d21-a688-401e-a66e-861beb6336d9" containerID="3f52c2e22a75c010993ca5809a3a9d1304758d4cb6e83691ef1af15a4b514ac4" exitCode=0 Dec 13 06:49:30 crc kubenswrapper[4971]: I1213 06:49:30.972249 4971 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-additional-cni-plugins-ncchz" event={"ID":"de914d21-a688-401e-a66e-861beb6336d9","Type":"ContainerDied","Data":"3f52c2e22a75c010993ca5809a3a9d1304758d4cb6e83691ef1af15a4b514ac4"} Dec 13 06:49:30 crc kubenswrapper[4971]: I1213 06:49:30.976267 4971 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-zlcg5" event={"ID":"736c6b1a-93dd-4f71-81b6-1c43fcf5f556","Type":"ContainerStarted","Data":"868049c6a3a4ebe4036d1d22e02ba6b76bced2ee7e269bb228accc4a0636678a"} Dec 13 06:49:30 crc kubenswrapper[4971]: I1213 06:49:30.990837 4971 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3ad3268a-a982-428c-9c06-4aaeddd372d9\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T06:49:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T06:49:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T06:49:03Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T06:49:03Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T06:49:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3716afdf2bc693be80a09fc9f7727d334f435caf1d4c8ec4811b3a1f1652cdfe\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-13T06:49:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://30df3abf8c79db8af52c6a918f73e1ce410ddafcf341636424c267e1eeb195a7\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-13T06:49:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://ccdf65f6d2b1eb3647a03ba37c305c32e34addfdf79ff937c8dddaa3e5d5b5c6\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-13T06:49:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://770b73dd0c2b6f38744e2c1cd6aa186a69bb87d1cd015a603ad2e33be06a8f6b\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://b3c1bab3d53f975b2e24582ce2771bc26d600e0dd7c30339dcfd18a071162ad4\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-12-13T06:49:23Z\\\",\\\"message\\\":\\\"file observer\\\\nW1213 06:49:23.283403 1 builder.go:272] unable to get owner reference (falling back to namespace): pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\nI1213 06:49:23.284336 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1213 06:49:23.286978 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-981773671/tls.crt::/tmp/serving-cert-981773671/tls.key\\\\\\\"\\\\nI1213 06:49:23.775072 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1213 06:49:23.778615 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1213 06:49:23.778645 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1213 06:49:23.778680 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1213 06:49:23.778691 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1213 06:49:23.785058 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1213 06:49:23.786150 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1213 06:49:23.786196 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1213 06:49:23.786223 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1213 06:49:23.786246 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1213 06:49:23.786269 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1213 06:49:23.786291 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI1213 06:49:23.785608 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF1213 06:49:23.787563 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-13T06:49:08Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-13T06:49:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://9acd07066deed483bd7fa61ee45fda09539d110717a134a56827f871186f92fc\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-13T06:49:07Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://38c91cfbaecfdbd3e3b663422a91ab4294610afb6bbf9ebdb77db24e2b76c682\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://38c91cfbaecfdbd3e3b663422a91ab4294610afb6bbf9ebdb77db24e2b76c682\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-13T06:49:05Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-13T06:49:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-13T06:49:03Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-13T06:49:30Z is after 2025-08-24T17:21:41Z" Dec 13 06:49:31 crc kubenswrapper[4971]: I1213 06:49:31.001812 4971 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-13T06:49:23Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-13T06:49:23Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-13T06:49:30Z is after 2025-08-24T17:21:41Z" Dec 13 06:49:31 crc kubenswrapper[4971]: I1213 06:49:31.017328 4971 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-13T06:49:23Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-13T06:49:23Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-13T06:49:31Z is after 2025-08-24T17:21:41Z" Dec 13 06:49:31 crc kubenswrapper[4971]: I1213 06:49:31.030609 4971 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-13T06:49:24Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-13T06:49:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-13T06:49:24Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d0b2ecf0d5b22f23217f42f259ae7ffb6359555b22c30eaebf667eaf254b3bab\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-13T06:49:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-13T06:49:31Z is after 2025-08-24T17:21:41Z" Dec 13 06:49:31 crc kubenswrapper[4971]: I1213 06:49:31.041768 4971 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 13 06:49:31 crc kubenswrapper[4971]: I1213 06:49:31.041837 4971 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 13 06:49:31 crc kubenswrapper[4971]: I1213 06:49:31.041847 4971 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 13 06:49:31 crc kubenswrapper[4971]: I1213 06:49:31.041889 4971 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 13 06:49:31 crc kubenswrapper[4971]: I1213 06:49:31.041899 4971 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-13T06:49:31Z","lastTransitionTime":"2025-12-13T06:49:31Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 13 06:49:31 crc kubenswrapper[4971]: I1213 06:49:31.044908 4971 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-13T06:49:23Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-13T06:49:23Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-13T06:49:31Z is after 2025-08-24T17:21:41Z" Dec 13 06:49:31 crc kubenswrapper[4971]: I1213 06:49:31.057348 4971 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-13T06:49:26Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-13T06:49:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-13T06:49:26Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9c8a0d1db0e754462d805e82daf5f52ffd5457507a30456535ec59e096c9d092\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-13T06:49:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-13T06:49:31Z is after 2025-08-24T17:21:41Z" Dec 13 06:49:31 crc kubenswrapper[4971]: I1213 06:49:31.072383 4971 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-82xjz" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e0c4af12-edda-41d7-9b44-f87396174ef2\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T06:49:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T06:49:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T06:49:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T06:49:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://0f4577d9e37f860813e221f2e208177d8d383fd04f673ce357a0a405ea29d1e6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-13T06:49:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cnxsq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://7e0d2020e62f1a277cff7047e1907dc747ed393a1f6c983e1ee258bd156c16c7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-13T06:49:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cnxsq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-13T06:49:25Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-82xjz\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-13T06:49:31Z is after 2025-08-24T17:21:41Z" Dec 13 06:49:31 crc kubenswrapper[4971]: I1213 06:49:31.083774 4971 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-wwvpl" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"dddf744b-2122-49ad-a980-105c09636e0f\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T06:49:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T06:49:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T06:49:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T06:49:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4676d897c7fc668587faec563f3551a056557879cee3e865192d2bd28edf62fc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-13T06:49:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-czz9v\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-13T06:49:25Z\\\"}}\" for pod \"openshift-multus\"/\"multus-wwvpl\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-13T06:49:31Z is after 2025-08-24T17:21:41Z" Dec 13 06:49:31 crc kubenswrapper[4971]: I1213 06:49:31.098849 4971 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-zlcg5" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"736c6b1a-93dd-4f71-81b6-1c43fcf5f556\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T06:49:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T06:49:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T06:49:25Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T06:49:25Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vzmsr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vzmsr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vzmsr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vzmsr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vzmsr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vzmsr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vzmsr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vzmsr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://451b22df6973ddc8fa2209a65d64d9e41b33eb231bb4adfab88cf0e5943bb246\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://451b22df6973ddc8fa2209a65d64d9e41b33eb231bb4adfab88cf0e5943bb246\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-13T06:49:26Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-13T06:49:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vzmsr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-13T06:49:25Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-zlcg5\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-13T06:49:31Z is after 2025-08-24T17:21:41Z" Dec 13 06:49:31 crc kubenswrapper[4971]: I1213 06:49:31.110054 4971 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"f9d31b5a-b770-4a15-95c8-35b49677c608\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T06:49:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T06:49:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T06:49:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T06:49:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T06:49:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c9baa042f9dada954faa57b5454cd6b0a12a297e56ddc192c52cff50e42a16fa\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-13T06:49:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://c788de86754f0bf37ee184672fc1dcb388ccebc246d4f15be4b5da661f725f8f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-13T06:49:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://760817511a9b4f96a68dc38d26631250eaab72e1dd8f123e653039b6b4e91f84\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-13T06:49:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b859be0c9feac856e9855c6a038b02b49b32fade97873c46d0bec0a32d8b5147\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-13T06:49:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-13T06:49:03Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-13T06:49:31Z is after 2025-08-24T17:21:41Z" Dec 13 06:49:31 crc kubenswrapper[4971]: I1213 06:49:31.123048 4971 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-ncchz" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"de914d21-a688-401e-a66e-861beb6336d9\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T06:49:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T06:49:25Z\\\",\\\"message\\\":\\\"containers with incomplete status: [whereabouts-cni-bincopy whereabouts-cni]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T06:49:25Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T06:49:25Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lqmjj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://0b7111af6c5a198d02bb02858837cfb4f3b20828d8d206cd18aa0ffa102d0cd4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://0b7111af6c5a198d02bb02858837cfb4f3b20828d8d206cd18aa0ffa102d0cd4\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-13T06:49:26Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-13T06:49:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lqmjj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b36777fe40e74ae1cecb7b73e204139db5279ba51cef4f1e8c4a4c491d8be403\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://b36777fe40e74ae1cecb7b73e204139db5279ba51cef4f1e8c4a4c491d8be403\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-13T06:49:28Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-13T06:49:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lqmjj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://0597a06cc663e416142311a16941d98297577693a1fe27201ea42c27496cace2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://0597a06cc663e416142311a16941d98297577693a1fe27201ea42c27496cace2\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-13T06:49:29Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-13T06:49:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lqmjj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://3f52c2e22a75c010993ca5809a3a9d1304758d4cb6e83691ef1af15a4b514ac4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://3f52c2e22a75c010993ca5809a3a9d1304758d4cb6e83691ef1af15a4b514ac4\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-13T06:49:30Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-13T06:49:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lqmjj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lqmjj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lqmjj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-13T06:49:25Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-ncchz\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-13T06:49:31Z is after 2025-08-24T17:21:41Z" Dec 13 06:49:31 crc kubenswrapper[4971]: I1213 06:49:31.133740 4971 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-gkcd8" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"13f5c327-f95e-42b4-8783-218500c17694\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T06:49:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T06:49:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T06:49:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T06:49:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://77e846e36f1cb40d03471dc03de11b6c3da448baa407aad38093fae9a801cf92\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-13T06:49:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gvlhn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-13T06:49:27Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-gkcd8\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-13T06:49:31Z is after 2025-08-24T17:21:41Z" Dec 13 06:49:31 crc kubenswrapper[4971]: I1213 06:49:31.143587 4971 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 13 06:49:31 crc kubenswrapper[4971]: I1213 06:49:31.143621 4971 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 13 06:49:31 crc kubenswrapper[4971]: I1213 06:49:31.143630 4971 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 13 06:49:31 crc kubenswrapper[4971]: I1213 06:49:31.143645 4971 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 13 06:49:31 crc kubenswrapper[4971]: I1213 06:49:31.143653 4971 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-13T06:49:31Z","lastTransitionTime":"2025-12-13T06:49:31Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 13 06:49:31 crc kubenswrapper[4971]: I1213 06:49:31.154397 4971 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"1ed9f82e-fe5c-4b76-a619-62e5abd88d3b\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T06:49:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T06:49:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T06:49:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T06:49:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T06:49:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e457329a78a2d24337f994e8a0a1f75a2a667244aa75eb84613d1e7138364aac\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-13T06:49:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://eb014de8140deef097d9bd94f3b9fa4ab3558bc4449bef84b83bf62bfabf5836\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-13T06:49:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://2c438cc74d555f7c9db4569d1c56b78b1088ed8105697d2c0f31e6c5136f3dd0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-13T06:49:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://aa6fa6e4340dcd066252df3c56d79a3a39ed531d5eea8d1d5941d6952368d3af\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-13T06:49:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://c4471b98854c52d5cd084d0ba3e64e5b08e9518ec3c372cce2c875a440522457\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-13T06:49:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7d2746335c98dae3a0f43024d34cb62ebcb42a01c8f9ce18d52b29d4c45598e5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://7d2746335c98dae3a0f43024d34cb62ebcb42a01c8f9ce18d52b29d4c45598e5\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-13T06:49:04Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-13T06:49:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://bc0e77d10a738128bea82122496bc42235fc1688f8907756d93e61b651902b57\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://bc0e77d10a738128bea82122496bc42235fc1688f8907756d93e61b651902b57\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-13T06:49:05Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-13T06:49:05Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://f821e418cba5e832da8839268bcbbd9630bdf319a212d5382bfa2c4b0b73a268\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://f821e418cba5e832da8839268bcbbd9630bdf319a212d5382bfa2c4b0b73a268\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-13T06:49:07Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-13T06:49:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-13T06:49:03Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-13T06:49:31Z is after 2025-08-24T17:21:41Z" Dec 13 06:49:31 crc kubenswrapper[4971]: I1213 06:49:31.169407 4971 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-13T06:49:24Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-13T06:49:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-13T06:49:24Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4328cd6a8464ed468229bbfbc8265f969427916d72ebb92237dbe68efe70fa47\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-13T06:49:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://9d715ca82d9b5a594e41f2b98da6f4277da0e4e1dec114969f901c00f6d7553d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-13T06:49:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-13T06:49:31Z is after 2025-08-24T17:21:41Z" Dec 13 06:49:31 crc kubenswrapper[4971]: I1213 06:49:31.180098 4971 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-5rzh5" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"6bf4591d-2483-4132-9672-d0fd6c995968\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T06:49:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T06:49:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T06:49:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T06:49:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://50e2cb14e4ffb603acb06e2281615a99f06182d345c4e70f2446fa2077033bbb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-13T06:49:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dfxqn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-13T06:49:25Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-5rzh5\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-13T06:49:31Z is after 2025-08-24T17:21:41Z" Dec 13 06:49:31 crc kubenswrapper[4971]: I1213 06:49:31.245971 4971 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 13 06:49:31 crc kubenswrapper[4971]: I1213 06:49:31.246007 4971 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 13 06:49:31 crc kubenswrapper[4971]: I1213 06:49:31.246017 4971 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 13 06:49:31 crc kubenswrapper[4971]: I1213 06:49:31.246035 4971 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 13 06:49:31 crc kubenswrapper[4971]: I1213 06:49:31.246049 4971 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-13T06:49:31Z","lastTransitionTime":"2025-12-13T06:49:31Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 13 06:49:31 crc kubenswrapper[4971]: I1213 06:49:31.348154 4971 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 13 06:49:31 crc kubenswrapper[4971]: I1213 06:49:31.348200 4971 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 13 06:49:31 crc kubenswrapper[4971]: I1213 06:49:31.348212 4971 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 13 06:49:31 crc kubenswrapper[4971]: I1213 06:49:31.348228 4971 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 13 06:49:31 crc kubenswrapper[4971]: I1213 06:49:31.348239 4971 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-13T06:49:31Z","lastTransitionTime":"2025-12-13T06:49:31Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 13 06:49:31 crc kubenswrapper[4971]: I1213 06:49:31.366629 4971 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 13 06:49:31 crc kubenswrapper[4971]: I1213 06:49:31.366739 4971 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"networking-console-plugin-cert\" (UniqueName: \"kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 13 06:49:31 crc kubenswrapper[4971]: E1213 06:49:31.366787 4971 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-13 06:49:39.366768288 +0000 UTC m=+35.971177726 (durationBeforeRetry 8s). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 13 06:49:31 crc kubenswrapper[4971]: E1213 06:49:31.366816 4971 secret.go:188] Couldn't get secret openshift-network-console/networking-console-plugin-cert: object "openshift-network-console"/"networking-console-plugin-cert" not registered Dec 13 06:49:31 crc kubenswrapper[4971]: I1213 06:49:31.366826 4971 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nginx-conf\" (UniqueName: \"kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 13 06:49:31 crc kubenswrapper[4971]: E1213 06:49:31.366848 4971 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2025-12-13 06:49:39.36684036 +0000 UTC m=+35.971249808 (durationBeforeRetry 8s). Error: MountVolume.SetUp failed for volume "networking-console-plugin-cert" (UniqueName: "kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin-cert" not registered Dec 13 06:49:31 crc kubenswrapper[4971]: E1213 06:49:31.366872 4971 configmap.go:193] Couldn't get configMap openshift-network-console/networking-console-plugin: object "openshift-network-console"/"networking-console-plugin" not registered Dec 13 06:49:31 crc kubenswrapper[4971]: E1213 06:49:31.366896 4971 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2025-12-13 06:49:39.366889372 +0000 UTC m=+35.971298820 (durationBeforeRetry 8s). Error: MountVolume.SetUp failed for volume "nginx-conf" (UniqueName: "kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin" not registered Dec 13 06:49:31 crc kubenswrapper[4971]: I1213 06:49:31.451188 4971 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 13 06:49:31 crc kubenswrapper[4971]: I1213 06:49:31.451233 4971 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 13 06:49:31 crc kubenswrapper[4971]: I1213 06:49:31.451246 4971 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 13 06:49:31 crc kubenswrapper[4971]: I1213 06:49:31.451263 4971 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 13 06:49:31 crc kubenswrapper[4971]: I1213 06:49:31.451276 4971 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-13T06:49:31Z","lastTransitionTime":"2025-12-13T06:49:31Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 13 06:49:31 crc kubenswrapper[4971]: I1213 06:49:31.467736 4971 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cqllr\" (UniqueName: \"kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr\") pod \"network-check-target-xd92c\" (UID: \"3b6479f0-333b-4a96-9adf-2099afdc2447\") " pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 13 06:49:31 crc kubenswrapper[4971]: I1213 06:49:31.467806 4971 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-s2dwl\" (UniqueName: \"kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl\") pod \"network-check-source-55646444c4-trplf\" (UID: \"9d751cbb-f2e2-430d-9754-c882a5e924a5\") " pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 13 06:49:31 crc kubenswrapper[4971]: E1213 06:49:31.467945 4971 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Dec 13 06:49:31 crc kubenswrapper[4971]: E1213 06:49:31.467964 4971 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Dec 13 06:49:31 crc kubenswrapper[4971]: E1213 06:49:31.467977 4971 projected.go:194] Error preparing data for projected volume kube-api-access-s2dwl for pod openshift-network-diagnostics/network-check-source-55646444c4-trplf: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Dec 13 06:49:31 crc kubenswrapper[4971]: E1213 06:49:31.467988 4971 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Dec 13 06:49:31 crc kubenswrapper[4971]: E1213 06:49:31.468020 4971 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Dec 13 06:49:31 crc kubenswrapper[4971]: E1213 06:49:31.468034 4971 projected.go:194] Error preparing data for projected volume kube-api-access-cqllr for pod openshift-network-diagnostics/network-check-target-xd92c: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Dec 13 06:49:31 crc kubenswrapper[4971]: E1213 06:49:31.468034 4971 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl podName:9d751cbb-f2e2-430d-9754-c882a5e924a5 nodeName:}" failed. No retries permitted until 2025-12-13 06:49:39.468018817 +0000 UTC m=+36.072428265 (durationBeforeRetry 8s). Error: MountVolume.SetUp failed for volume "kube-api-access-s2dwl" (UniqueName: "kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl") pod "network-check-source-55646444c4-trplf" (UID: "9d751cbb-f2e2-430d-9754-c882a5e924a5") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Dec 13 06:49:31 crc kubenswrapper[4971]: E1213 06:49:31.468152 4971 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr podName:3b6479f0-333b-4a96-9adf-2099afdc2447 nodeName:}" failed. No retries permitted until 2025-12-13 06:49:39.46811699 +0000 UTC m=+36.072526438 (durationBeforeRetry 8s). Error: MountVolume.SetUp failed for volume "kube-api-access-cqllr" (UniqueName: "kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr") pod "network-check-target-xd92c" (UID: "3b6479f0-333b-4a96-9adf-2099afdc2447") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Dec 13 06:49:31 crc kubenswrapper[4971]: I1213 06:49:31.553687 4971 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 13 06:49:31 crc kubenswrapper[4971]: I1213 06:49:31.553736 4971 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 13 06:49:31 crc kubenswrapper[4971]: I1213 06:49:31.553749 4971 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 13 06:49:31 crc kubenswrapper[4971]: I1213 06:49:31.553768 4971 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 13 06:49:31 crc kubenswrapper[4971]: I1213 06:49:31.553780 4971 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-13T06:49:31Z","lastTransitionTime":"2025-12-13T06:49:31Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 13 06:49:31 crc kubenswrapper[4971]: I1213 06:49:31.656374 4971 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 13 06:49:31 crc kubenswrapper[4971]: I1213 06:49:31.656410 4971 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 13 06:49:31 crc kubenswrapper[4971]: I1213 06:49:31.656419 4971 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 13 06:49:31 crc kubenswrapper[4971]: I1213 06:49:31.656433 4971 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 13 06:49:31 crc kubenswrapper[4971]: I1213 06:49:31.656443 4971 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-13T06:49:31Z","lastTransitionTime":"2025-12-13T06:49:31Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 13 06:49:31 crc kubenswrapper[4971]: I1213 06:49:31.759625 4971 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 13 06:49:31 crc kubenswrapper[4971]: I1213 06:49:31.759672 4971 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 13 06:49:31 crc kubenswrapper[4971]: I1213 06:49:31.759685 4971 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 13 06:49:31 crc kubenswrapper[4971]: I1213 06:49:31.759702 4971 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 13 06:49:31 crc kubenswrapper[4971]: I1213 06:49:31.759715 4971 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-13T06:49:31Z","lastTransitionTime":"2025-12-13T06:49:31Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 13 06:49:31 crc kubenswrapper[4971]: I1213 06:49:31.768186 4971 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 13 06:49:31 crc kubenswrapper[4971]: I1213 06:49:31.768217 4971 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 13 06:49:31 crc kubenswrapper[4971]: I1213 06:49:31.768217 4971 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 13 06:49:31 crc kubenswrapper[4971]: E1213 06:49:31.768313 4971 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 13 06:49:31 crc kubenswrapper[4971]: E1213 06:49:31.768430 4971 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 13 06:49:31 crc kubenswrapper[4971]: E1213 06:49:31.768504 4971 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 13 06:49:31 crc kubenswrapper[4971]: I1213 06:49:31.861470 4971 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 13 06:49:31 crc kubenswrapper[4971]: I1213 06:49:31.861508 4971 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 13 06:49:31 crc kubenswrapper[4971]: I1213 06:49:31.861544 4971 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 13 06:49:31 crc kubenswrapper[4971]: I1213 06:49:31.861562 4971 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 13 06:49:31 crc kubenswrapper[4971]: I1213 06:49:31.861576 4971 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-13T06:49:31Z","lastTransitionTime":"2025-12-13T06:49:31Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 13 06:49:31 crc kubenswrapper[4971]: I1213 06:49:31.963578 4971 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 13 06:49:31 crc kubenswrapper[4971]: I1213 06:49:31.963626 4971 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 13 06:49:31 crc kubenswrapper[4971]: I1213 06:49:31.963637 4971 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 13 06:49:31 crc kubenswrapper[4971]: I1213 06:49:31.963652 4971 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 13 06:49:31 crc kubenswrapper[4971]: I1213 06:49:31.963663 4971 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-13T06:49:31Z","lastTransitionTime":"2025-12-13T06:49:31Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 13 06:49:31 crc kubenswrapper[4971]: I1213 06:49:31.981716 4971 generic.go:334] "Generic (PLEG): container finished" podID="de914d21-a688-401e-a66e-861beb6336d9" containerID="9e21ca708c4c4c1228b417cece51299e0792729b7c78884bd18b2dda8fe8355c" exitCode=0 Dec 13 06:49:31 crc kubenswrapper[4971]: I1213 06:49:31.981757 4971 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-additional-cni-plugins-ncchz" event={"ID":"de914d21-a688-401e-a66e-861beb6336d9","Type":"ContainerDied","Data":"9e21ca708c4c4c1228b417cece51299e0792729b7c78884bd18b2dda8fe8355c"} Dec 13 06:49:31 crc kubenswrapper[4971]: I1213 06:49:31.995963 4971 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-82xjz" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e0c4af12-edda-41d7-9b44-f87396174ef2\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T06:49:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T06:49:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T06:49:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T06:49:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://0f4577d9e37f860813e221f2e208177d8d383fd04f673ce357a0a405ea29d1e6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-13T06:49:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cnxsq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://7e0d2020e62f1a277cff7047e1907dc747ed393a1f6c983e1ee258bd156c16c7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-13T06:49:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cnxsq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-13T06:49:25Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-82xjz\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-13T06:49:31Z is after 2025-08-24T17:21:41Z" Dec 13 06:49:32 crc kubenswrapper[4971]: I1213 06:49:32.010160 4971 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-wwvpl" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"dddf744b-2122-49ad-a980-105c09636e0f\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T06:49:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T06:49:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T06:49:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T06:49:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4676d897c7fc668587faec563f3551a056557879cee3e865192d2bd28edf62fc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-13T06:49:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-czz9v\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-13T06:49:25Z\\\"}}\" for pod \"openshift-multus\"/\"multus-wwvpl\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-13T06:49:32Z is after 2025-08-24T17:21:41Z" Dec 13 06:49:32 crc kubenswrapper[4971]: I1213 06:49:32.030583 4971 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-zlcg5" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"736c6b1a-93dd-4f71-81b6-1c43fcf5f556\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T06:49:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T06:49:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T06:49:25Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T06:49:25Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vzmsr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vzmsr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vzmsr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vzmsr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vzmsr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vzmsr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vzmsr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vzmsr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://451b22df6973ddc8fa2209a65d64d9e41b33eb231bb4adfab88cf0e5943bb246\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://451b22df6973ddc8fa2209a65d64d9e41b33eb231bb4adfab88cf0e5943bb246\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-13T06:49:26Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-13T06:49:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vzmsr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-13T06:49:25Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-zlcg5\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-13T06:49:32Z is after 2025-08-24T17:21:41Z" Dec 13 06:49:32 crc kubenswrapper[4971]: I1213 06:49:32.043709 4971 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-13T06:49:24Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-13T06:49:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-13T06:49:24Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d0b2ecf0d5b22f23217f42f259ae7ffb6359555b22c30eaebf667eaf254b3bab\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-13T06:49:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-13T06:49:32Z is after 2025-08-24T17:21:41Z" Dec 13 06:49:32 crc kubenswrapper[4971]: I1213 06:49:32.055118 4971 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-13T06:49:23Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-13T06:49:23Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-13T06:49:32Z is after 2025-08-24T17:21:41Z" Dec 13 06:49:32 crc kubenswrapper[4971]: I1213 06:49:32.067963 4971 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-13T06:49:26Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-13T06:49:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-13T06:49:26Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9c8a0d1db0e754462d805e82daf5f52ffd5457507a30456535ec59e096c9d092\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-13T06:49:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-13T06:49:32Z is after 2025-08-24T17:21:41Z" Dec 13 06:49:32 crc kubenswrapper[4971]: I1213 06:49:32.068681 4971 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 13 06:49:32 crc kubenswrapper[4971]: I1213 06:49:32.068713 4971 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 13 06:49:32 crc kubenswrapper[4971]: I1213 06:49:32.068723 4971 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 13 06:49:32 crc kubenswrapper[4971]: I1213 06:49:32.068739 4971 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 13 06:49:32 crc kubenswrapper[4971]: I1213 06:49:32.068750 4971 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-13T06:49:32Z","lastTransitionTime":"2025-12-13T06:49:32Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 13 06:49:32 crc kubenswrapper[4971]: I1213 06:49:32.078064 4971 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-gkcd8" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"13f5c327-f95e-42b4-8783-218500c17694\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T06:49:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T06:49:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T06:49:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T06:49:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://77e846e36f1cb40d03471dc03de11b6c3da448baa407aad38093fae9a801cf92\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-13T06:49:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gvlhn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-13T06:49:27Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-gkcd8\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-13T06:49:32Z is after 2025-08-24T17:21:41Z" Dec 13 06:49:32 crc kubenswrapper[4971]: I1213 06:49:32.094549 4971 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"f9d31b5a-b770-4a15-95c8-35b49677c608\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T06:49:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T06:49:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T06:49:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T06:49:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T06:49:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c9baa042f9dada954faa57b5454cd6b0a12a297e56ddc192c52cff50e42a16fa\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-13T06:49:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://c788de86754f0bf37ee184672fc1dcb388ccebc246d4f15be4b5da661f725f8f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-13T06:49:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://760817511a9b4f96a68dc38d26631250eaab72e1dd8f123e653039b6b4e91f84\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-13T06:49:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b859be0c9feac856e9855c6a038b02b49b32fade97873c46d0bec0a32d8b5147\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-13T06:49:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-13T06:49:03Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-13T06:49:32Z is after 2025-08-24T17:21:41Z" Dec 13 06:49:32 crc kubenswrapper[4971]: I1213 06:49:32.111838 4971 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-ncchz" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"de914d21-a688-401e-a66e-861beb6336d9\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T06:49:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T06:49:25Z\\\",\\\"message\\\":\\\"containers with incomplete status: [whereabouts-cni]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T06:49:25Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T06:49:25Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lqmjj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://0b7111af6c5a198d02bb02858837cfb4f3b20828d8d206cd18aa0ffa102d0cd4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://0b7111af6c5a198d02bb02858837cfb4f3b20828d8d206cd18aa0ffa102d0cd4\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-13T06:49:26Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-13T06:49:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lqmjj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b36777fe40e74ae1cecb7b73e204139db5279ba51cef4f1e8c4a4c491d8be403\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://b36777fe40e74ae1cecb7b73e204139db5279ba51cef4f1e8c4a4c491d8be403\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-13T06:49:28Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-13T06:49:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lqmjj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://0597a06cc663e416142311a16941d98297577693a1fe27201ea42c27496cace2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://0597a06cc663e416142311a16941d98297577693a1fe27201ea42c27496cace2\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-13T06:49:29Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-13T06:49:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lqmjj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://3f52c2e22a75c010993ca5809a3a9d1304758d4cb6e83691ef1af15a4b514ac4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://3f52c2e22a75c010993ca5809a3a9d1304758d4cb6e83691ef1af15a4b514ac4\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-13T06:49:30Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-13T06:49:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lqmjj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://9e21ca708c4c4c1228b417cece51299e0792729b7c78884bd18b2dda8fe8355c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://9e21ca708c4c4c1228b417cece51299e0792729b7c78884bd18b2dda8fe8355c\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-13T06:49:31Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-13T06:49:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lqmjj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lqmjj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-13T06:49:25Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-ncchz\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-13T06:49:32Z is after 2025-08-24T17:21:41Z" Dec 13 06:49:32 crc kubenswrapper[4971]: I1213 06:49:32.125620 4971 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-5rzh5" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"6bf4591d-2483-4132-9672-d0fd6c995968\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T06:49:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T06:49:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T06:49:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T06:49:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://50e2cb14e4ffb603acb06e2281615a99f06182d345c4e70f2446fa2077033bbb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-13T06:49:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dfxqn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-13T06:49:25Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-5rzh5\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-13T06:49:32Z is after 2025-08-24T17:21:41Z" Dec 13 06:49:32 crc kubenswrapper[4971]: I1213 06:49:32.144129 4971 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"1ed9f82e-fe5c-4b76-a619-62e5abd88d3b\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T06:49:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T06:49:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T06:49:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T06:49:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T06:49:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e457329a78a2d24337f994e8a0a1f75a2a667244aa75eb84613d1e7138364aac\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-13T06:49:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://eb014de8140deef097d9bd94f3b9fa4ab3558bc4449bef84b83bf62bfabf5836\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-13T06:49:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://2c438cc74d555f7c9db4569d1c56b78b1088ed8105697d2c0f31e6c5136f3dd0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-13T06:49:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://aa6fa6e4340dcd066252df3c56d79a3a39ed531d5eea8d1d5941d6952368d3af\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-13T06:49:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://c4471b98854c52d5cd084d0ba3e64e5b08e9518ec3c372cce2c875a440522457\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-13T06:49:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7d2746335c98dae3a0f43024d34cb62ebcb42a01c8f9ce18d52b29d4c45598e5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://7d2746335c98dae3a0f43024d34cb62ebcb42a01c8f9ce18d52b29d4c45598e5\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-13T06:49:04Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-13T06:49:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://bc0e77d10a738128bea82122496bc42235fc1688f8907756d93e61b651902b57\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://bc0e77d10a738128bea82122496bc42235fc1688f8907756d93e61b651902b57\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-13T06:49:05Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-13T06:49:05Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://f821e418cba5e832da8839268bcbbd9630bdf319a212d5382bfa2c4b0b73a268\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://f821e418cba5e832da8839268bcbbd9630bdf319a212d5382bfa2c4b0b73a268\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-13T06:49:07Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-13T06:49:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-13T06:49:03Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-13T06:49:32Z is after 2025-08-24T17:21:41Z" Dec 13 06:49:32 crc kubenswrapper[4971]: I1213 06:49:32.156184 4971 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-13T06:49:24Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-13T06:49:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-13T06:49:24Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4328cd6a8464ed468229bbfbc8265f969427916d72ebb92237dbe68efe70fa47\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-13T06:49:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://9d715ca82d9b5a594e41f2b98da6f4277da0e4e1dec114969f901c00f6d7553d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-13T06:49:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-13T06:49:32Z is after 2025-08-24T17:21:41Z" Dec 13 06:49:32 crc kubenswrapper[4971]: I1213 06:49:32.167187 4971 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3ad3268a-a982-428c-9c06-4aaeddd372d9\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T06:49:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T06:49:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T06:49:03Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T06:49:03Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T06:49:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3716afdf2bc693be80a09fc9f7727d334f435caf1d4c8ec4811b3a1f1652cdfe\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-13T06:49:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://30df3abf8c79db8af52c6a918f73e1ce410ddafcf341636424c267e1eeb195a7\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-13T06:49:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://ccdf65f6d2b1eb3647a03ba37c305c32e34addfdf79ff937c8dddaa3e5d5b5c6\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-13T06:49:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://770b73dd0c2b6f38744e2c1cd6aa186a69bb87d1cd015a603ad2e33be06a8f6b\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://b3c1bab3d53f975b2e24582ce2771bc26d600e0dd7c30339dcfd18a071162ad4\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-12-13T06:49:23Z\\\",\\\"message\\\":\\\"file observer\\\\nW1213 06:49:23.283403 1 builder.go:272] unable to get owner reference (falling back to namespace): pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\nI1213 06:49:23.284336 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1213 06:49:23.286978 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-981773671/tls.crt::/tmp/serving-cert-981773671/tls.key\\\\\\\"\\\\nI1213 06:49:23.775072 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1213 06:49:23.778615 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1213 06:49:23.778645 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1213 06:49:23.778680 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1213 06:49:23.778691 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1213 06:49:23.785058 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1213 06:49:23.786150 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1213 06:49:23.786196 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1213 06:49:23.786223 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1213 06:49:23.786246 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1213 06:49:23.786269 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1213 06:49:23.786291 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI1213 06:49:23.785608 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF1213 06:49:23.787563 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-13T06:49:08Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-13T06:49:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://9acd07066deed483bd7fa61ee45fda09539d110717a134a56827f871186f92fc\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-13T06:49:07Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://38c91cfbaecfdbd3e3b663422a91ab4294610afb6bbf9ebdb77db24e2b76c682\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://38c91cfbaecfdbd3e3b663422a91ab4294610afb6bbf9ebdb77db24e2b76c682\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-13T06:49:05Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-13T06:49:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-13T06:49:03Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-13T06:49:32Z is after 2025-08-24T17:21:41Z" Dec 13 06:49:32 crc kubenswrapper[4971]: I1213 06:49:32.170343 4971 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 13 06:49:32 crc kubenswrapper[4971]: I1213 06:49:32.170382 4971 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 13 06:49:32 crc kubenswrapper[4971]: I1213 06:49:32.170391 4971 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 13 06:49:32 crc kubenswrapper[4971]: I1213 06:49:32.170407 4971 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 13 06:49:32 crc kubenswrapper[4971]: I1213 06:49:32.170417 4971 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-13T06:49:32Z","lastTransitionTime":"2025-12-13T06:49:32Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 13 06:49:32 crc kubenswrapper[4971]: I1213 06:49:32.177409 4971 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-13T06:49:23Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-13T06:49:23Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-13T06:49:32Z is after 2025-08-24T17:21:41Z" Dec 13 06:49:32 crc kubenswrapper[4971]: I1213 06:49:32.187852 4971 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-13T06:49:23Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-13T06:49:23Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-13T06:49:32Z is after 2025-08-24T17:21:41Z" Dec 13 06:49:32 crc kubenswrapper[4971]: I1213 06:49:32.273491 4971 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 13 06:49:32 crc kubenswrapper[4971]: I1213 06:49:32.273960 4971 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 13 06:49:32 crc kubenswrapper[4971]: I1213 06:49:32.273986 4971 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 13 06:49:32 crc kubenswrapper[4971]: I1213 06:49:32.274016 4971 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 13 06:49:32 crc kubenswrapper[4971]: I1213 06:49:32.274041 4971 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-13T06:49:32Z","lastTransitionTime":"2025-12-13T06:49:32Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 13 06:49:32 crc kubenswrapper[4971]: I1213 06:49:32.376613 4971 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 13 06:49:32 crc kubenswrapper[4971]: I1213 06:49:32.376649 4971 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 13 06:49:32 crc kubenswrapper[4971]: I1213 06:49:32.376660 4971 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 13 06:49:32 crc kubenswrapper[4971]: I1213 06:49:32.376675 4971 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 13 06:49:32 crc kubenswrapper[4971]: I1213 06:49:32.376688 4971 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-13T06:49:32Z","lastTransitionTime":"2025-12-13T06:49:32Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 13 06:49:32 crc kubenswrapper[4971]: I1213 06:49:32.479867 4971 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 13 06:49:32 crc kubenswrapper[4971]: I1213 06:49:32.479916 4971 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 13 06:49:32 crc kubenswrapper[4971]: I1213 06:49:32.479933 4971 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 13 06:49:32 crc kubenswrapper[4971]: I1213 06:49:32.479957 4971 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 13 06:49:32 crc kubenswrapper[4971]: I1213 06:49:32.479975 4971 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-13T06:49:32Z","lastTransitionTime":"2025-12-13T06:49:32Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 13 06:49:32 crc kubenswrapper[4971]: I1213 06:49:32.583823 4971 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 13 06:49:32 crc kubenswrapper[4971]: I1213 06:49:32.583897 4971 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 13 06:49:32 crc kubenswrapper[4971]: I1213 06:49:32.583918 4971 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 13 06:49:32 crc kubenswrapper[4971]: I1213 06:49:32.583944 4971 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 13 06:49:32 crc kubenswrapper[4971]: I1213 06:49:32.583964 4971 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-13T06:49:32Z","lastTransitionTime":"2025-12-13T06:49:32Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 13 06:49:32 crc kubenswrapper[4971]: I1213 06:49:32.687388 4971 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 13 06:49:32 crc kubenswrapper[4971]: I1213 06:49:32.687448 4971 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 13 06:49:32 crc kubenswrapper[4971]: I1213 06:49:32.687457 4971 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 13 06:49:32 crc kubenswrapper[4971]: I1213 06:49:32.687474 4971 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 13 06:49:32 crc kubenswrapper[4971]: I1213 06:49:32.687485 4971 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-13T06:49:32Z","lastTransitionTime":"2025-12-13T06:49:32Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 13 06:49:32 crc kubenswrapper[4971]: I1213 06:49:32.789848 4971 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 13 06:49:32 crc kubenswrapper[4971]: I1213 06:49:32.789900 4971 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 13 06:49:32 crc kubenswrapper[4971]: I1213 06:49:32.789914 4971 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 13 06:49:32 crc kubenswrapper[4971]: I1213 06:49:32.789933 4971 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 13 06:49:32 crc kubenswrapper[4971]: I1213 06:49:32.789943 4971 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-13T06:49:32Z","lastTransitionTime":"2025-12-13T06:49:32Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 13 06:49:32 crc kubenswrapper[4971]: I1213 06:49:32.892039 4971 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 13 06:49:32 crc kubenswrapper[4971]: I1213 06:49:32.892079 4971 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 13 06:49:32 crc kubenswrapper[4971]: I1213 06:49:32.892092 4971 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 13 06:49:32 crc kubenswrapper[4971]: I1213 06:49:32.892107 4971 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 13 06:49:32 crc kubenswrapper[4971]: I1213 06:49:32.892116 4971 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-13T06:49:32Z","lastTransitionTime":"2025-12-13T06:49:32Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 13 06:49:32 crc kubenswrapper[4971]: I1213 06:49:32.987681 4971 generic.go:334] "Generic (PLEG): container finished" podID="de914d21-a688-401e-a66e-861beb6336d9" containerID="74552159f12c87cdda09147692b400ccac0d4e43c1ed72fe8462c4249ad69c7b" exitCode=0 Dec 13 06:49:32 crc kubenswrapper[4971]: I1213 06:49:32.987748 4971 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-additional-cni-plugins-ncchz" event={"ID":"de914d21-a688-401e-a66e-861beb6336d9","Type":"ContainerDied","Data":"74552159f12c87cdda09147692b400ccac0d4e43c1ed72fe8462c4249ad69c7b"} Dec 13 06:49:32 crc kubenswrapper[4971]: I1213 06:49:32.995439 4971 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 13 06:49:32 crc kubenswrapper[4971]: I1213 06:49:32.995469 4971 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 13 06:49:32 crc kubenswrapper[4971]: I1213 06:49:32.995480 4971 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 13 06:49:32 crc kubenswrapper[4971]: I1213 06:49:32.995495 4971 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 13 06:49:32 crc kubenswrapper[4971]: I1213 06:49:32.995506 4971 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-13T06:49:32Z","lastTransitionTime":"2025-12-13T06:49:32Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 13 06:49:32 crc kubenswrapper[4971]: I1213 06:49:32.997244 4971 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-zlcg5" event={"ID":"736c6b1a-93dd-4f71-81b6-1c43fcf5f556","Type":"ContainerStarted","Data":"451040d83e76225457d750a4aa532aa2772decb89719e53fe001bf74aaa8e73b"} Dec 13 06:49:32 crc kubenswrapper[4971]: I1213 06:49:32.997620 4971 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-ovn-kubernetes/ovnkube-node-zlcg5" Dec 13 06:49:32 crc kubenswrapper[4971]: I1213 06:49:32.997645 4971 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-ovn-kubernetes/ovnkube-node-zlcg5" Dec 13 06:49:33 crc kubenswrapper[4971]: I1213 06:49:33.007769 4971 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3ad3268a-a982-428c-9c06-4aaeddd372d9\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T06:49:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T06:49:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T06:49:03Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T06:49:03Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T06:49:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3716afdf2bc693be80a09fc9f7727d334f435caf1d4c8ec4811b3a1f1652cdfe\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-13T06:49:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://30df3abf8c79db8af52c6a918f73e1ce410ddafcf341636424c267e1eeb195a7\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-13T06:49:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://ccdf65f6d2b1eb3647a03ba37c305c32e34addfdf79ff937c8dddaa3e5d5b5c6\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-13T06:49:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://770b73dd0c2b6f38744e2c1cd6aa186a69bb87d1cd015a603ad2e33be06a8f6b\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://b3c1bab3d53f975b2e24582ce2771bc26d600e0dd7c30339dcfd18a071162ad4\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-12-13T06:49:23Z\\\",\\\"message\\\":\\\"file observer\\\\nW1213 06:49:23.283403 1 builder.go:272] unable to get owner reference (falling back to namespace): pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\nI1213 06:49:23.284336 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1213 06:49:23.286978 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-981773671/tls.crt::/tmp/serving-cert-981773671/tls.key\\\\\\\"\\\\nI1213 06:49:23.775072 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1213 06:49:23.778615 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1213 06:49:23.778645 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1213 06:49:23.778680 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1213 06:49:23.778691 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1213 06:49:23.785058 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1213 06:49:23.786150 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1213 06:49:23.786196 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1213 06:49:23.786223 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1213 06:49:23.786246 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1213 06:49:23.786269 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1213 06:49:23.786291 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI1213 06:49:23.785608 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF1213 06:49:23.787563 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-13T06:49:08Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-13T06:49:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://9acd07066deed483bd7fa61ee45fda09539d110717a134a56827f871186f92fc\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-13T06:49:07Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://38c91cfbaecfdbd3e3b663422a91ab4294610afb6bbf9ebdb77db24e2b76c682\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://38c91cfbaecfdbd3e3b663422a91ab4294610afb6bbf9ebdb77db24e2b76c682\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-13T06:49:05Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-13T06:49:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-13T06:49:03Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-13T06:49:33Z is after 2025-08-24T17:21:41Z" Dec 13 06:49:33 crc kubenswrapper[4971]: I1213 06:49:33.022709 4971 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-13T06:49:23Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-13T06:49:23Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-13T06:49:33Z is after 2025-08-24T17:21:41Z" Dec 13 06:49:33 crc kubenswrapper[4971]: I1213 06:49:33.026329 4971 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-ovn-kubernetes/ovnkube-node-zlcg5" Dec 13 06:49:33 crc kubenswrapper[4971]: I1213 06:49:33.026555 4971 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-ovn-kubernetes/ovnkube-node-zlcg5" Dec 13 06:49:33 crc kubenswrapper[4971]: I1213 06:49:33.034036 4971 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-13T06:49:23Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-13T06:49:23Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-13T06:49:33Z is after 2025-08-24T17:21:41Z" Dec 13 06:49:33 crc kubenswrapper[4971]: I1213 06:49:33.047037 4971 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-wwvpl" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"dddf744b-2122-49ad-a980-105c09636e0f\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T06:49:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T06:49:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T06:49:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T06:49:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4676d897c7fc668587faec563f3551a056557879cee3e865192d2bd28edf62fc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-13T06:49:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-czz9v\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-13T06:49:25Z\\\"}}\" for pod \"openshift-multus\"/\"multus-wwvpl\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-13T06:49:33Z is after 2025-08-24T17:21:41Z" Dec 13 06:49:33 crc kubenswrapper[4971]: I1213 06:49:33.064457 4971 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-zlcg5" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"736c6b1a-93dd-4f71-81b6-1c43fcf5f556\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T06:49:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T06:49:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T06:49:25Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T06:49:25Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vzmsr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vzmsr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vzmsr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vzmsr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vzmsr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vzmsr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vzmsr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vzmsr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://451b22df6973ddc8fa2209a65d64d9e41b33eb231bb4adfab88cf0e5943bb246\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://451b22df6973ddc8fa2209a65d64d9e41b33eb231bb4adfab88cf0e5943bb246\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-13T06:49:26Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-13T06:49:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vzmsr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-13T06:49:25Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-zlcg5\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-13T06:49:33Z is after 2025-08-24T17:21:41Z" Dec 13 06:49:33 crc kubenswrapper[4971]: I1213 06:49:33.075874 4971 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-13T06:49:24Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-13T06:49:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-13T06:49:24Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d0b2ecf0d5b22f23217f42f259ae7ffb6359555b22c30eaebf667eaf254b3bab\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-13T06:49:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-13T06:49:33Z is after 2025-08-24T17:21:41Z" Dec 13 06:49:33 crc kubenswrapper[4971]: I1213 06:49:33.087047 4971 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-13T06:49:23Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-13T06:49:23Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-13T06:49:33Z is after 2025-08-24T17:21:41Z" Dec 13 06:49:33 crc kubenswrapper[4971]: I1213 06:49:33.097348 4971 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-13T06:49:26Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-13T06:49:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-13T06:49:26Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9c8a0d1db0e754462d805e82daf5f52ffd5457507a30456535ec59e096c9d092\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-13T06:49:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-13T06:49:33Z is after 2025-08-24T17:21:41Z" Dec 13 06:49:33 crc kubenswrapper[4971]: I1213 06:49:33.097756 4971 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 13 06:49:33 crc kubenswrapper[4971]: I1213 06:49:33.097799 4971 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 13 06:49:33 crc kubenswrapper[4971]: I1213 06:49:33.097839 4971 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 13 06:49:33 crc kubenswrapper[4971]: I1213 06:49:33.097854 4971 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 13 06:49:33 crc kubenswrapper[4971]: I1213 06:49:33.097863 4971 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-13T06:49:33Z","lastTransitionTime":"2025-12-13T06:49:33Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 13 06:49:33 crc kubenswrapper[4971]: I1213 06:49:33.108467 4971 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-82xjz" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e0c4af12-edda-41d7-9b44-f87396174ef2\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T06:49:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T06:49:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T06:49:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T06:49:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://0f4577d9e37f860813e221f2e208177d8d383fd04f673ce357a0a405ea29d1e6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-13T06:49:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cnxsq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://7e0d2020e62f1a277cff7047e1907dc747ed393a1f6c983e1ee258bd156c16c7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-13T06:49:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cnxsq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-13T06:49:25Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-82xjz\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-13T06:49:33Z is after 2025-08-24T17:21:41Z" Dec 13 06:49:33 crc kubenswrapper[4971]: I1213 06:49:33.124818 4971 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"f9d31b5a-b770-4a15-95c8-35b49677c608\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T06:49:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T06:49:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T06:49:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T06:49:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T06:49:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c9baa042f9dada954faa57b5454cd6b0a12a297e56ddc192c52cff50e42a16fa\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-13T06:49:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://c788de86754f0bf37ee184672fc1dcb388ccebc246d4f15be4b5da661f725f8f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-13T06:49:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://760817511a9b4f96a68dc38d26631250eaab72e1dd8f123e653039b6b4e91f84\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-13T06:49:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b859be0c9feac856e9855c6a038b02b49b32fade97873c46d0bec0a32d8b5147\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-13T06:49:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-13T06:49:03Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-13T06:49:33Z is after 2025-08-24T17:21:41Z" Dec 13 06:49:33 crc kubenswrapper[4971]: I1213 06:49:33.141817 4971 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-ncchz" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"de914d21-a688-401e-a66e-861beb6336d9\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T06:49:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T06:49:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T06:49:25Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T06:49:25Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lqmjj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://0b7111af6c5a198d02bb02858837cfb4f3b20828d8d206cd18aa0ffa102d0cd4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://0b7111af6c5a198d02bb02858837cfb4f3b20828d8d206cd18aa0ffa102d0cd4\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-13T06:49:26Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-13T06:49:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lqmjj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b36777fe40e74ae1cecb7b73e204139db5279ba51cef4f1e8c4a4c491d8be403\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://b36777fe40e74ae1cecb7b73e204139db5279ba51cef4f1e8c4a4c491d8be403\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-13T06:49:28Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-13T06:49:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lqmjj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://0597a06cc663e416142311a16941d98297577693a1fe27201ea42c27496cace2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://0597a06cc663e416142311a16941d98297577693a1fe27201ea42c27496cace2\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-13T06:49:29Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-13T06:49:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lqmjj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://3f52c2e22a75c010993ca5809a3a9d1304758d4cb6e83691ef1af15a4b514ac4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://3f52c2e22a75c010993ca5809a3a9d1304758d4cb6e83691ef1af15a4b514ac4\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-13T06:49:30Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-13T06:49:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lqmjj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://9e21ca708c4c4c1228b417cece51299e0792729b7c78884bd18b2dda8fe8355c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://9e21ca708c4c4c1228b417cece51299e0792729b7c78884bd18b2dda8fe8355c\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-13T06:49:31Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-13T06:49:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lqmjj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://74552159f12c87cdda09147692b400ccac0d4e43c1ed72fe8462c4249ad69c7b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://74552159f12c87cdda09147692b400ccac0d4e43c1ed72fe8462c4249ad69c7b\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-13T06:49:32Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-13T06:49:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lqmjj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-13T06:49:25Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-ncchz\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-13T06:49:33Z is after 2025-08-24T17:21:41Z" Dec 13 06:49:33 crc kubenswrapper[4971]: I1213 06:49:33.153071 4971 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-gkcd8" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"13f5c327-f95e-42b4-8783-218500c17694\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T06:49:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T06:49:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T06:49:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T06:49:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://77e846e36f1cb40d03471dc03de11b6c3da448baa407aad38093fae9a801cf92\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-13T06:49:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gvlhn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-13T06:49:27Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-gkcd8\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-13T06:49:33Z is after 2025-08-24T17:21:41Z" Dec 13 06:49:33 crc kubenswrapper[4971]: I1213 06:49:33.174726 4971 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"1ed9f82e-fe5c-4b76-a619-62e5abd88d3b\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T06:49:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T06:49:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T06:49:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T06:49:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T06:49:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e457329a78a2d24337f994e8a0a1f75a2a667244aa75eb84613d1e7138364aac\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-13T06:49:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://eb014de8140deef097d9bd94f3b9fa4ab3558bc4449bef84b83bf62bfabf5836\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-13T06:49:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://2c438cc74d555f7c9db4569d1c56b78b1088ed8105697d2c0f31e6c5136f3dd0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-13T06:49:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://aa6fa6e4340dcd066252df3c56d79a3a39ed531d5eea8d1d5941d6952368d3af\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-13T06:49:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://c4471b98854c52d5cd084d0ba3e64e5b08e9518ec3c372cce2c875a440522457\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-13T06:49:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7d2746335c98dae3a0f43024d34cb62ebcb42a01c8f9ce18d52b29d4c45598e5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://7d2746335c98dae3a0f43024d34cb62ebcb42a01c8f9ce18d52b29d4c45598e5\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-13T06:49:04Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-13T06:49:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://bc0e77d10a738128bea82122496bc42235fc1688f8907756d93e61b651902b57\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://bc0e77d10a738128bea82122496bc42235fc1688f8907756d93e61b651902b57\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-13T06:49:05Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-13T06:49:05Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://f821e418cba5e832da8839268bcbbd9630bdf319a212d5382bfa2c4b0b73a268\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://f821e418cba5e832da8839268bcbbd9630bdf319a212d5382bfa2c4b0b73a268\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-13T06:49:07Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-13T06:49:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-13T06:49:03Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-13T06:49:33Z is after 2025-08-24T17:21:41Z" Dec 13 06:49:33 crc kubenswrapper[4971]: I1213 06:49:33.186601 4971 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-13T06:49:24Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-13T06:49:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-13T06:49:24Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4328cd6a8464ed468229bbfbc8265f969427916d72ebb92237dbe68efe70fa47\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-13T06:49:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://9d715ca82d9b5a594e41f2b98da6f4277da0e4e1dec114969f901c00f6d7553d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-13T06:49:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-13T06:49:33Z is after 2025-08-24T17:21:41Z" Dec 13 06:49:33 crc kubenswrapper[4971]: I1213 06:49:33.195423 4971 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-5rzh5" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"6bf4591d-2483-4132-9672-d0fd6c995968\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T06:49:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T06:49:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T06:49:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T06:49:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://50e2cb14e4ffb603acb06e2281615a99f06182d345c4e70f2446fa2077033bbb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-13T06:49:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dfxqn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-13T06:49:25Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-5rzh5\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-13T06:49:33Z is after 2025-08-24T17:21:41Z" Dec 13 06:49:33 crc kubenswrapper[4971]: I1213 06:49:33.200089 4971 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 13 06:49:33 crc kubenswrapper[4971]: I1213 06:49:33.200121 4971 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 13 06:49:33 crc kubenswrapper[4971]: I1213 06:49:33.200132 4971 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 13 06:49:33 crc kubenswrapper[4971]: I1213 06:49:33.200147 4971 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 13 06:49:33 crc kubenswrapper[4971]: I1213 06:49:33.200158 4971 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-13T06:49:33Z","lastTransitionTime":"2025-12-13T06:49:33Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 13 06:49:33 crc kubenswrapper[4971]: I1213 06:49:33.207628 4971 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3ad3268a-a982-428c-9c06-4aaeddd372d9\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T06:49:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T06:49:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T06:49:03Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T06:49:03Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T06:49:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3716afdf2bc693be80a09fc9f7727d334f435caf1d4c8ec4811b3a1f1652cdfe\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-13T06:49:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://30df3abf8c79db8af52c6a918f73e1ce410ddafcf341636424c267e1eeb195a7\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-13T06:49:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://ccdf65f6d2b1eb3647a03ba37c305c32e34addfdf79ff937c8dddaa3e5d5b5c6\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-13T06:49:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://770b73dd0c2b6f38744e2c1cd6aa186a69bb87d1cd015a603ad2e33be06a8f6b\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://b3c1bab3d53f975b2e24582ce2771bc26d600e0dd7c30339dcfd18a071162ad4\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-12-13T06:49:23Z\\\",\\\"message\\\":\\\"file observer\\\\nW1213 06:49:23.283403 1 builder.go:272] unable to get owner reference (falling back to namespace): pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\nI1213 06:49:23.284336 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1213 06:49:23.286978 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-981773671/tls.crt::/tmp/serving-cert-981773671/tls.key\\\\\\\"\\\\nI1213 06:49:23.775072 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1213 06:49:23.778615 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1213 06:49:23.778645 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1213 06:49:23.778680 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1213 06:49:23.778691 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1213 06:49:23.785058 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1213 06:49:23.786150 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1213 06:49:23.786196 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1213 06:49:23.786223 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1213 06:49:23.786246 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1213 06:49:23.786269 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1213 06:49:23.786291 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI1213 06:49:23.785608 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF1213 06:49:23.787563 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-13T06:49:08Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-13T06:49:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://9acd07066deed483bd7fa61ee45fda09539d110717a134a56827f871186f92fc\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-13T06:49:07Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://38c91cfbaecfdbd3e3b663422a91ab4294610afb6bbf9ebdb77db24e2b76c682\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://38c91cfbaecfdbd3e3b663422a91ab4294610afb6bbf9ebdb77db24e2b76c682\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-13T06:49:05Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-13T06:49:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-13T06:49:03Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-13T06:49:33Z is after 2025-08-24T17:21:41Z" Dec 13 06:49:33 crc kubenswrapper[4971]: I1213 06:49:33.219644 4971 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-13T06:49:23Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-13T06:49:23Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-13T06:49:33Z is after 2025-08-24T17:21:41Z" Dec 13 06:49:33 crc kubenswrapper[4971]: I1213 06:49:33.232247 4971 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-13T06:49:23Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-13T06:49:23Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-13T06:49:33Z is after 2025-08-24T17:21:41Z" Dec 13 06:49:33 crc kubenswrapper[4971]: I1213 06:49:33.245554 4971 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-13T06:49:24Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-13T06:49:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-13T06:49:24Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d0b2ecf0d5b22f23217f42f259ae7ffb6359555b22c30eaebf667eaf254b3bab\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-13T06:49:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-13T06:49:33Z is after 2025-08-24T17:21:41Z" Dec 13 06:49:33 crc kubenswrapper[4971]: I1213 06:49:33.256629 4971 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-13T06:49:23Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-13T06:49:23Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-13T06:49:33Z is after 2025-08-24T17:21:41Z" Dec 13 06:49:33 crc kubenswrapper[4971]: I1213 06:49:33.269077 4971 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-13T06:49:26Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-13T06:49:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-13T06:49:26Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9c8a0d1db0e754462d805e82daf5f52ffd5457507a30456535ec59e096c9d092\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-13T06:49:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-13T06:49:33Z is after 2025-08-24T17:21:41Z" Dec 13 06:49:33 crc kubenswrapper[4971]: I1213 06:49:33.278344 4971 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-82xjz" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e0c4af12-edda-41d7-9b44-f87396174ef2\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T06:49:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T06:49:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T06:49:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T06:49:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://0f4577d9e37f860813e221f2e208177d8d383fd04f673ce357a0a405ea29d1e6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-13T06:49:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cnxsq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://7e0d2020e62f1a277cff7047e1907dc747ed393a1f6c983e1ee258bd156c16c7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-13T06:49:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cnxsq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-13T06:49:25Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-82xjz\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-13T06:49:33Z is after 2025-08-24T17:21:41Z" Dec 13 06:49:33 crc kubenswrapper[4971]: I1213 06:49:33.288183 4971 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-wwvpl" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"dddf744b-2122-49ad-a980-105c09636e0f\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T06:49:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T06:49:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T06:49:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T06:49:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4676d897c7fc668587faec563f3551a056557879cee3e865192d2bd28edf62fc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-13T06:49:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-czz9v\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-13T06:49:25Z\\\"}}\" for pod \"openshift-multus\"/\"multus-wwvpl\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-13T06:49:33Z is after 2025-08-24T17:21:41Z" Dec 13 06:49:33 crc kubenswrapper[4971]: I1213 06:49:33.302948 4971 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 13 06:49:33 crc kubenswrapper[4971]: I1213 06:49:33.302979 4971 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 13 06:49:33 crc kubenswrapper[4971]: I1213 06:49:33.302990 4971 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 13 06:49:33 crc kubenswrapper[4971]: I1213 06:49:33.303004 4971 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 13 06:49:33 crc kubenswrapper[4971]: I1213 06:49:33.303016 4971 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-13T06:49:33Z","lastTransitionTime":"2025-12-13T06:49:33Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 13 06:49:33 crc kubenswrapper[4971]: I1213 06:49:33.304211 4971 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-zlcg5" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"736c6b1a-93dd-4f71-81b6-1c43fcf5f556\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T06:49:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T06:49:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T06:49:25Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T06:49:25Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://325223543b3c7db417f769533559a1ca2b23a723494e1885b0fd0480b5a6399f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-13T06:49:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vzmsr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://f22a4ec756119c4f791fb5105cf1f31df3e548b2b07b1f909a62802e9b6035ae\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-13T06:49:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vzmsr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://f6073140d9f86929b6d4bd2960ede0617a5c9620b269bbab043bae37e0d389f5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-13T06:49:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vzmsr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://693a2ab434c250cb140a866e8068b156df9191414fed1d63bee919a2528ea6e5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-13T06:49:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vzmsr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://219620f51474e60b3e1f880c0972eb612dffc4bad24462bbba2c093ae817d852\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-13T06:49:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vzmsr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://f137f4b441c2402093ab430a9689a46221c4fb4bc70d99fb992f62b0a989c465\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-13T06:49:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vzmsr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://451040d83e76225457d750a4aa532aa2772decb89719e53fe001bf74aaa8e73b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-13T06:49:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vzmsr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://868049c6a3a4ebe4036d1d22e02ba6b76bced2ee7e269bb228accc4a0636678a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-13T06:49:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vzmsr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://451b22df6973ddc8fa2209a65d64d9e41b33eb231bb4adfab88cf0e5943bb246\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://451b22df6973ddc8fa2209a65d64d9e41b33eb231bb4adfab88cf0e5943bb246\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-13T06:49:26Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-13T06:49:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vzmsr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-13T06:49:25Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-zlcg5\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-13T06:49:33Z is after 2025-08-24T17:21:41Z" Dec 13 06:49:33 crc kubenswrapper[4971]: I1213 06:49:33.315599 4971 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"f9d31b5a-b770-4a15-95c8-35b49677c608\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T06:49:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T06:49:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T06:49:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T06:49:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T06:49:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c9baa042f9dada954faa57b5454cd6b0a12a297e56ddc192c52cff50e42a16fa\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-13T06:49:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://c788de86754f0bf37ee184672fc1dcb388ccebc246d4f15be4b5da661f725f8f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-13T06:49:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://760817511a9b4f96a68dc38d26631250eaab72e1dd8f123e653039b6b4e91f84\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-13T06:49:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b859be0c9feac856e9855c6a038b02b49b32fade97873c46d0bec0a32d8b5147\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-13T06:49:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-13T06:49:03Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-13T06:49:33Z is after 2025-08-24T17:21:41Z" Dec 13 06:49:33 crc kubenswrapper[4971]: I1213 06:49:33.328466 4971 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-ncchz" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"de914d21-a688-401e-a66e-861beb6336d9\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T06:49:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T06:49:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T06:49:25Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T06:49:25Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lqmjj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://0b7111af6c5a198d02bb02858837cfb4f3b20828d8d206cd18aa0ffa102d0cd4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://0b7111af6c5a198d02bb02858837cfb4f3b20828d8d206cd18aa0ffa102d0cd4\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-13T06:49:26Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-13T06:49:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lqmjj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b36777fe40e74ae1cecb7b73e204139db5279ba51cef4f1e8c4a4c491d8be403\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://b36777fe40e74ae1cecb7b73e204139db5279ba51cef4f1e8c4a4c491d8be403\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-13T06:49:28Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-13T06:49:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lqmjj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://0597a06cc663e416142311a16941d98297577693a1fe27201ea42c27496cace2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://0597a06cc663e416142311a16941d98297577693a1fe27201ea42c27496cace2\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-13T06:49:29Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-13T06:49:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lqmjj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://3f52c2e22a75c010993ca5809a3a9d1304758d4cb6e83691ef1af15a4b514ac4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://3f52c2e22a75c010993ca5809a3a9d1304758d4cb6e83691ef1af15a4b514ac4\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-13T06:49:30Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-13T06:49:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lqmjj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://9e21ca708c4c4c1228b417cece51299e0792729b7c78884bd18b2dda8fe8355c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://9e21ca708c4c4c1228b417cece51299e0792729b7c78884bd18b2dda8fe8355c\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-13T06:49:31Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-13T06:49:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lqmjj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://74552159f12c87cdda09147692b400ccac0d4e43c1ed72fe8462c4249ad69c7b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://74552159f12c87cdda09147692b400ccac0d4e43c1ed72fe8462c4249ad69c7b\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-13T06:49:32Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-13T06:49:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lqmjj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-13T06:49:25Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-ncchz\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-13T06:49:33Z is after 2025-08-24T17:21:41Z" Dec 13 06:49:33 crc kubenswrapper[4971]: I1213 06:49:33.339014 4971 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-gkcd8" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"13f5c327-f95e-42b4-8783-218500c17694\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T06:49:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T06:49:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T06:49:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T06:49:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://77e846e36f1cb40d03471dc03de11b6c3da448baa407aad38093fae9a801cf92\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-13T06:49:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gvlhn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-13T06:49:27Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-gkcd8\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-13T06:49:33Z is after 2025-08-24T17:21:41Z" Dec 13 06:49:33 crc kubenswrapper[4971]: I1213 06:49:33.357491 4971 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"1ed9f82e-fe5c-4b76-a619-62e5abd88d3b\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T06:49:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T06:49:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T06:49:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T06:49:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T06:49:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e457329a78a2d24337f994e8a0a1f75a2a667244aa75eb84613d1e7138364aac\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-13T06:49:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://eb014de8140deef097d9bd94f3b9fa4ab3558bc4449bef84b83bf62bfabf5836\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-13T06:49:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://2c438cc74d555f7c9db4569d1c56b78b1088ed8105697d2c0f31e6c5136f3dd0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-13T06:49:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://aa6fa6e4340dcd066252df3c56d79a3a39ed531d5eea8d1d5941d6952368d3af\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-13T06:49:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://c4471b98854c52d5cd084d0ba3e64e5b08e9518ec3c372cce2c875a440522457\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-13T06:49:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7d2746335c98dae3a0f43024d34cb62ebcb42a01c8f9ce18d52b29d4c45598e5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://7d2746335c98dae3a0f43024d34cb62ebcb42a01c8f9ce18d52b29d4c45598e5\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-13T06:49:04Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-13T06:49:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://bc0e77d10a738128bea82122496bc42235fc1688f8907756d93e61b651902b57\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://bc0e77d10a738128bea82122496bc42235fc1688f8907756d93e61b651902b57\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-13T06:49:05Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-13T06:49:05Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://f821e418cba5e832da8839268bcbbd9630bdf319a212d5382bfa2c4b0b73a268\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://f821e418cba5e832da8839268bcbbd9630bdf319a212d5382bfa2c4b0b73a268\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-13T06:49:07Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-13T06:49:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-13T06:49:03Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-13T06:49:33Z is after 2025-08-24T17:21:41Z" Dec 13 06:49:33 crc kubenswrapper[4971]: I1213 06:49:33.369578 4971 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-13T06:49:24Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-13T06:49:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-13T06:49:24Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4328cd6a8464ed468229bbfbc8265f969427916d72ebb92237dbe68efe70fa47\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-13T06:49:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://9d715ca82d9b5a594e41f2b98da6f4277da0e4e1dec114969f901c00f6d7553d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-13T06:49:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-13T06:49:33Z is after 2025-08-24T17:21:41Z" Dec 13 06:49:33 crc kubenswrapper[4971]: I1213 06:49:33.379103 4971 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-5rzh5" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"6bf4591d-2483-4132-9672-d0fd6c995968\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T06:49:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T06:49:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T06:49:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T06:49:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://50e2cb14e4ffb603acb06e2281615a99f06182d345c4e70f2446fa2077033bbb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-13T06:49:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dfxqn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-13T06:49:25Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-5rzh5\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-13T06:49:33Z is after 2025-08-24T17:21:41Z" Dec 13 06:49:33 crc kubenswrapper[4971]: I1213 06:49:33.404859 4971 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 13 06:49:33 crc kubenswrapper[4971]: I1213 06:49:33.404892 4971 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 13 06:49:33 crc kubenswrapper[4971]: I1213 06:49:33.404901 4971 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 13 06:49:33 crc kubenswrapper[4971]: I1213 06:49:33.404916 4971 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 13 06:49:33 crc kubenswrapper[4971]: I1213 06:49:33.404926 4971 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-13T06:49:33Z","lastTransitionTime":"2025-12-13T06:49:33Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 13 06:49:33 crc kubenswrapper[4971]: I1213 06:49:33.507493 4971 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 13 06:49:33 crc kubenswrapper[4971]: I1213 06:49:33.507551 4971 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 13 06:49:33 crc kubenswrapper[4971]: I1213 06:49:33.507565 4971 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 13 06:49:33 crc kubenswrapper[4971]: I1213 06:49:33.507611 4971 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 13 06:49:33 crc kubenswrapper[4971]: I1213 06:49:33.507624 4971 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-13T06:49:33Z","lastTransitionTime":"2025-12-13T06:49:33Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 13 06:49:33 crc kubenswrapper[4971]: I1213 06:49:33.609499 4971 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 13 06:49:33 crc kubenswrapper[4971]: I1213 06:49:33.609560 4971 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 13 06:49:33 crc kubenswrapper[4971]: I1213 06:49:33.609572 4971 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 13 06:49:33 crc kubenswrapper[4971]: I1213 06:49:33.609589 4971 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 13 06:49:33 crc kubenswrapper[4971]: I1213 06:49:33.609602 4971 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-13T06:49:33Z","lastTransitionTime":"2025-12-13T06:49:33Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 13 06:49:33 crc kubenswrapper[4971]: I1213 06:49:33.712367 4971 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 13 06:49:33 crc kubenswrapper[4971]: I1213 06:49:33.712411 4971 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 13 06:49:33 crc kubenswrapper[4971]: I1213 06:49:33.712421 4971 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 13 06:49:33 crc kubenswrapper[4971]: I1213 06:49:33.712437 4971 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 13 06:49:33 crc kubenswrapper[4971]: I1213 06:49:33.712448 4971 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-13T06:49:33Z","lastTransitionTime":"2025-12-13T06:49:33Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 13 06:49:33 crc kubenswrapper[4971]: I1213 06:49:33.768020 4971 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 13 06:49:33 crc kubenswrapper[4971]: E1213 06:49:33.768147 4971 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 13 06:49:33 crc kubenswrapper[4971]: I1213 06:49:33.768021 4971 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 13 06:49:33 crc kubenswrapper[4971]: I1213 06:49:33.768234 4971 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 13 06:49:33 crc kubenswrapper[4971]: E1213 06:49:33.768323 4971 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 13 06:49:33 crc kubenswrapper[4971]: E1213 06:49:33.768410 4971 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 13 06:49:33 crc kubenswrapper[4971]: I1213 06:49:33.778668 4971 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-5rzh5" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"6bf4591d-2483-4132-9672-d0fd6c995968\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T06:49:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T06:49:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T06:49:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T06:49:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://50e2cb14e4ffb603acb06e2281615a99f06182d345c4e70f2446fa2077033bbb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-13T06:49:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dfxqn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-13T06:49:25Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-5rzh5\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-13T06:49:33Z is after 2025-08-24T17:21:41Z" Dec 13 06:49:33 crc kubenswrapper[4971]: I1213 06:49:33.795082 4971 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"1ed9f82e-fe5c-4b76-a619-62e5abd88d3b\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T06:49:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T06:49:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T06:49:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T06:49:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T06:49:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e457329a78a2d24337f994e8a0a1f75a2a667244aa75eb84613d1e7138364aac\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-13T06:49:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://eb014de8140deef097d9bd94f3b9fa4ab3558bc4449bef84b83bf62bfabf5836\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-13T06:49:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://2c438cc74d555f7c9db4569d1c56b78b1088ed8105697d2c0f31e6c5136f3dd0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-13T06:49:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://aa6fa6e4340dcd066252df3c56d79a3a39ed531d5eea8d1d5941d6952368d3af\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-13T06:49:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://c4471b98854c52d5cd084d0ba3e64e5b08e9518ec3c372cce2c875a440522457\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-13T06:49:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7d2746335c98dae3a0f43024d34cb62ebcb42a01c8f9ce18d52b29d4c45598e5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://7d2746335c98dae3a0f43024d34cb62ebcb42a01c8f9ce18d52b29d4c45598e5\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-13T06:49:04Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-13T06:49:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://bc0e77d10a738128bea82122496bc42235fc1688f8907756d93e61b651902b57\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://bc0e77d10a738128bea82122496bc42235fc1688f8907756d93e61b651902b57\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-13T06:49:05Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-13T06:49:05Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://f821e418cba5e832da8839268bcbbd9630bdf319a212d5382bfa2c4b0b73a268\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://f821e418cba5e832da8839268bcbbd9630bdf319a212d5382bfa2c4b0b73a268\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-13T06:49:07Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-13T06:49:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-13T06:49:03Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-13T06:49:33Z is after 2025-08-24T17:21:41Z" Dec 13 06:49:33 crc kubenswrapper[4971]: I1213 06:49:33.807090 4971 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-13T06:49:24Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-13T06:49:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-13T06:49:24Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4328cd6a8464ed468229bbfbc8265f969427916d72ebb92237dbe68efe70fa47\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-13T06:49:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://9d715ca82d9b5a594e41f2b98da6f4277da0e4e1dec114969f901c00f6d7553d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-13T06:49:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-13T06:49:33Z is after 2025-08-24T17:21:41Z" Dec 13 06:49:33 crc kubenswrapper[4971]: I1213 06:49:33.814539 4971 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 13 06:49:33 crc kubenswrapper[4971]: I1213 06:49:33.814591 4971 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 13 06:49:33 crc kubenswrapper[4971]: I1213 06:49:33.814603 4971 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 13 06:49:33 crc kubenswrapper[4971]: I1213 06:49:33.814621 4971 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 13 06:49:33 crc kubenswrapper[4971]: I1213 06:49:33.814633 4971 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-13T06:49:33Z","lastTransitionTime":"2025-12-13T06:49:33Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 13 06:49:33 crc kubenswrapper[4971]: I1213 06:49:33.820313 4971 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3ad3268a-a982-428c-9c06-4aaeddd372d9\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T06:49:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T06:49:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T06:49:03Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T06:49:03Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T06:49:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3716afdf2bc693be80a09fc9f7727d334f435caf1d4c8ec4811b3a1f1652cdfe\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-13T06:49:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://30df3abf8c79db8af52c6a918f73e1ce410ddafcf341636424c267e1eeb195a7\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-13T06:49:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://ccdf65f6d2b1eb3647a03ba37c305c32e34addfdf79ff937c8dddaa3e5d5b5c6\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-13T06:49:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://770b73dd0c2b6f38744e2c1cd6aa186a69bb87d1cd015a603ad2e33be06a8f6b\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://b3c1bab3d53f975b2e24582ce2771bc26d600e0dd7c30339dcfd18a071162ad4\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-12-13T06:49:23Z\\\",\\\"message\\\":\\\"file observer\\\\nW1213 06:49:23.283403 1 builder.go:272] unable to get owner reference (falling back to namespace): pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\nI1213 06:49:23.284336 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1213 06:49:23.286978 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-981773671/tls.crt::/tmp/serving-cert-981773671/tls.key\\\\\\\"\\\\nI1213 06:49:23.775072 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1213 06:49:23.778615 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1213 06:49:23.778645 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1213 06:49:23.778680 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1213 06:49:23.778691 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1213 06:49:23.785058 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1213 06:49:23.786150 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1213 06:49:23.786196 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1213 06:49:23.786223 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1213 06:49:23.786246 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1213 06:49:23.786269 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1213 06:49:23.786291 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI1213 06:49:23.785608 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF1213 06:49:23.787563 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-13T06:49:08Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-13T06:49:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://9acd07066deed483bd7fa61ee45fda09539d110717a134a56827f871186f92fc\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-13T06:49:07Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://38c91cfbaecfdbd3e3b663422a91ab4294610afb6bbf9ebdb77db24e2b76c682\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://38c91cfbaecfdbd3e3b663422a91ab4294610afb6bbf9ebdb77db24e2b76c682\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-13T06:49:05Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-13T06:49:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-13T06:49:03Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-13T06:49:33Z is after 2025-08-24T17:21:41Z" Dec 13 06:49:33 crc kubenswrapper[4971]: I1213 06:49:33.832027 4971 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-13T06:49:23Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-13T06:49:23Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-13T06:49:33Z is after 2025-08-24T17:21:41Z" Dec 13 06:49:33 crc kubenswrapper[4971]: I1213 06:49:33.843090 4971 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-13T06:49:23Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-13T06:49:23Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-13T06:49:33Z is after 2025-08-24T17:21:41Z" Dec 13 06:49:33 crc kubenswrapper[4971]: I1213 06:49:33.856874 4971 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-82xjz" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e0c4af12-edda-41d7-9b44-f87396174ef2\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T06:49:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T06:49:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T06:49:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T06:49:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://0f4577d9e37f860813e221f2e208177d8d383fd04f673ce357a0a405ea29d1e6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-13T06:49:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cnxsq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://7e0d2020e62f1a277cff7047e1907dc747ed393a1f6c983e1ee258bd156c16c7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-13T06:49:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cnxsq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-13T06:49:25Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-82xjz\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-13T06:49:33Z is after 2025-08-24T17:21:41Z" Dec 13 06:49:33 crc kubenswrapper[4971]: I1213 06:49:33.873364 4971 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-wwvpl" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"dddf744b-2122-49ad-a980-105c09636e0f\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T06:49:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T06:49:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T06:49:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T06:49:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4676d897c7fc668587faec563f3551a056557879cee3e865192d2bd28edf62fc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-13T06:49:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-czz9v\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-13T06:49:25Z\\\"}}\" for pod \"openshift-multus\"/\"multus-wwvpl\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-13T06:49:33Z is after 2025-08-24T17:21:41Z" Dec 13 06:49:33 crc kubenswrapper[4971]: I1213 06:49:33.892193 4971 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-zlcg5" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"736c6b1a-93dd-4f71-81b6-1c43fcf5f556\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T06:49:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T06:49:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T06:49:25Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T06:49:25Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://325223543b3c7db417f769533559a1ca2b23a723494e1885b0fd0480b5a6399f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-13T06:49:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vzmsr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://f22a4ec756119c4f791fb5105cf1f31df3e548b2b07b1f909a62802e9b6035ae\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-13T06:49:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vzmsr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://f6073140d9f86929b6d4bd2960ede0617a5c9620b269bbab043bae37e0d389f5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-13T06:49:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vzmsr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://693a2ab434c250cb140a866e8068b156df9191414fed1d63bee919a2528ea6e5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-13T06:49:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vzmsr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://219620f51474e60b3e1f880c0972eb612dffc4bad24462bbba2c093ae817d852\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-13T06:49:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vzmsr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://f137f4b441c2402093ab430a9689a46221c4fb4bc70d99fb992f62b0a989c465\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-13T06:49:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vzmsr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://451040d83e76225457d750a4aa532aa2772decb89719e53fe001bf74aaa8e73b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-13T06:49:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vzmsr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://868049c6a3a4ebe4036d1d22e02ba6b76bced2ee7e269bb228accc4a0636678a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-13T06:49:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vzmsr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://451b22df6973ddc8fa2209a65d64d9e41b33eb231bb4adfab88cf0e5943bb246\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://451b22df6973ddc8fa2209a65d64d9e41b33eb231bb4adfab88cf0e5943bb246\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-13T06:49:26Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-13T06:49:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vzmsr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-13T06:49:25Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-zlcg5\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-13T06:49:33Z is after 2025-08-24T17:21:41Z" Dec 13 06:49:33 crc kubenswrapper[4971]: I1213 06:49:33.903628 4971 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-13T06:49:24Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-13T06:49:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-13T06:49:24Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d0b2ecf0d5b22f23217f42f259ae7ffb6359555b22c30eaebf667eaf254b3bab\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-13T06:49:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-13T06:49:33Z is after 2025-08-24T17:21:41Z" Dec 13 06:49:33 crc kubenswrapper[4971]: I1213 06:49:33.915059 4971 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-13T06:49:23Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-13T06:49:23Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-13T06:49:33Z is after 2025-08-24T17:21:41Z" Dec 13 06:49:33 crc kubenswrapper[4971]: I1213 06:49:33.916246 4971 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 13 06:49:33 crc kubenswrapper[4971]: I1213 06:49:33.916279 4971 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 13 06:49:33 crc kubenswrapper[4971]: I1213 06:49:33.916289 4971 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 13 06:49:33 crc kubenswrapper[4971]: I1213 06:49:33.916304 4971 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 13 06:49:33 crc kubenswrapper[4971]: I1213 06:49:33.916313 4971 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-13T06:49:33Z","lastTransitionTime":"2025-12-13T06:49:33Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 13 06:49:33 crc kubenswrapper[4971]: I1213 06:49:33.927081 4971 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-13T06:49:26Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-13T06:49:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-13T06:49:26Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9c8a0d1db0e754462d805e82daf5f52ffd5457507a30456535ec59e096c9d092\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-13T06:49:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-13T06:49:33Z is after 2025-08-24T17:21:41Z" Dec 13 06:49:33 crc kubenswrapper[4971]: I1213 06:49:33.937719 4971 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-gkcd8" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"13f5c327-f95e-42b4-8783-218500c17694\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T06:49:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T06:49:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T06:49:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T06:49:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://77e846e36f1cb40d03471dc03de11b6c3da448baa407aad38093fae9a801cf92\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-13T06:49:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gvlhn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-13T06:49:27Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-gkcd8\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-13T06:49:33Z is after 2025-08-24T17:21:41Z" Dec 13 06:49:33 crc kubenswrapper[4971]: I1213 06:49:33.949864 4971 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"f9d31b5a-b770-4a15-95c8-35b49677c608\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T06:49:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T06:49:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T06:49:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T06:49:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T06:49:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c9baa042f9dada954faa57b5454cd6b0a12a297e56ddc192c52cff50e42a16fa\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-13T06:49:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://c788de86754f0bf37ee184672fc1dcb388ccebc246d4f15be4b5da661f725f8f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-13T06:49:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://760817511a9b4f96a68dc38d26631250eaab72e1dd8f123e653039b6b4e91f84\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-13T06:49:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b859be0c9feac856e9855c6a038b02b49b32fade97873c46d0bec0a32d8b5147\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-13T06:49:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-13T06:49:03Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-13T06:49:33Z is after 2025-08-24T17:21:41Z" Dec 13 06:49:33 crc kubenswrapper[4971]: I1213 06:49:33.964703 4971 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-ncchz" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"de914d21-a688-401e-a66e-861beb6336d9\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T06:49:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T06:49:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T06:49:25Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T06:49:25Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lqmjj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://0b7111af6c5a198d02bb02858837cfb4f3b20828d8d206cd18aa0ffa102d0cd4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://0b7111af6c5a198d02bb02858837cfb4f3b20828d8d206cd18aa0ffa102d0cd4\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-13T06:49:26Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-13T06:49:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lqmjj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b36777fe40e74ae1cecb7b73e204139db5279ba51cef4f1e8c4a4c491d8be403\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://b36777fe40e74ae1cecb7b73e204139db5279ba51cef4f1e8c4a4c491d8be403\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-13T06:49:28Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-13T06:49:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lqmjj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://0597a06cc663e416142311a16941d98297577693a1fe27201ea42c27496cace2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://0597a06cc663e416142311a16941d98297577693a1fe27201ea42c27496cace2\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-13T06:49:29Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-13T06:49:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lqmjj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://3f52c2e22a75c010993ca5809a3a9d1304758d4cb6e83691ef1af15a4b514ac4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://3f52c2e22a75c010993ca5809a3a9d1304758d4cb6e83691ef1af15a4b514ac4\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-13T06:49:30Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-13T06:49:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lqmjj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://9e21ca708c4c4c1228b417cece51299e0792729b7c78884bd18b2dda8fe8355c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://9e21ca708c4c4c1228b417cece51299e0792729b7c78884bd18b2dda8fe8355c\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-13T06:49:31Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-13T06:49:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lqmjj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://74552159f12c87cdda09147692b400ccac0d4e43c1ed72fe8462c4249ad69c7b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://74552159f12c87cdda09147692b400ccac0d4e43c1ed72fe8462c4249ad69c7b\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-13T06:49:32Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-13T06:49:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lqmjj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-13T06:49:25Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-ncchz\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-13T06:49:33Z is after 2025-08-24T17:21:41Z" Dec 13 06:49:34 crc kubenswrapper[4971]: I1213 06:49:34.002820 4971 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-additional-cni-plugins-ncchz" event={"ID":"de914d21-a688-401e-a66e-861beb6336d9","Type":"ContainerStarted","Data":"244ffe3dc9d8cf799aff4c4d795bf41554c1e6bb6667fbb0c7df322fb5f96c68"} Dec 13 06:49:34 crc kubenswrapper[4971]: I1213 06:49:34.002903 4971 prober_manager.go:312] "Failed to trigger a manual run" probe="Readiness" Dec 13 06:49:34 crc kubenswrapper[4971]: I1213 06:49:34.018563 4971 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 13 06:49:34 crc kubenswrapper[4971]: I1213 06:49:34.018604 4971 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 13 06:49:34 crc kubenswrapper[4971]: I1213 06:49:34.018617 4971 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 13 06:49:34 crc kubenswrapper[4971]: I1213 06:49:34.018636 4971 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 13 06:49:34 crc kubenswrapper[4971]: I1213 06:49:34.018656 4971 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-13T06:49:34Z","lastTransitionTime":"2025-12-13T06:49:34Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 13 06:49:34 crc kubenswrapper[4971]: I1213 06:49:34.023392 4971 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"f9d31b5a-b770-4a15-95c8-35b49677c608\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T06:49:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T06:49:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T06:49:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T06:49:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T06:49:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c9baa042f9dada954faa57b5454cd6b0a12a297e56ddc192c52cff50e42a16fa\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-13T06:49:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://c788de86754f0bf37ee184672fc1dcb388ccebc246d4f15be4b5da661f725f8f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-13T06:49:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://760817511a9b4f96a68dc38d26631250eaab72e1dd8f123e653039b6b4e91f84\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-13T06:49:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b859be0c9feac856e9855c6a038b02b49b32fade97873c46d0bec0a32d8b5147\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-13T06:49:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-13T06:49:03Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-13T06:49:34Z is after 2025-08-24T17:21:41Z" Dec 13 06:49:34 crc kubenswrapper[4971]: I1213 06:49:34.039053 4971 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-ncchz" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"de914d21-a688-401e-a66e-861beb6336d9\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T06:49:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T06:49:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T06:49:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T06:49:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://244ffe3dc9d8cf799aff4c4d795bf41554c1e6bb6667fbb0c7df322fb5f96c68\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-13T06:49:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lqmjj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://0b7111af6c5a198d02bb02858837cfb4f3b20828d8d206cd18aa0ffa102d0cd4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://0b7111af6c5a198d02bb02858837cfb4f3b20828d8d206cd18aa0ffa102d0cd4\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-13T06:49:26Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-13T06:49:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lqmjj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b36777fe40e74ae1cecb7b73e204139db5279ba51cef4f1e8c4a4c491d8be403\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://b36777fe40e74ae1cecb7b73e204139db5279ba51cef4f1e8c4a4c491d8be403\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-13T06:49:28Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-13T06:49:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lqmjj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://0597a06cc663e416142311a16941d98297577693a1fe27201ea42c27496cace2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://0597a06cc663e416142311a16941d98297577693a1fe27201ea42c27496cace2\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-13T06:49:29Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-13T06:49:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lqmjj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://3f52c2e22a75c010993ca5809a3a9d1304758d4cb6e83691ef1af15a4b514ac4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://3f52c2e22a75c010993ca5809a3a9d1304758d4cb6e83691ef1af15a4b514ac4\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-13T06:49:30Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-13T06:49:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lqmjj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://9e21ca708c4c4c1228b417cece51299e0792729b7c78884bd18b2dda8fe8355c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://9e21ca708c4c4c1228b417cece51299e0792729b7c78884bd18b2dda8fe8355c\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-13T06:49:31Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-13T06:49:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lqmjj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://74552159f12c87cdda09147692b400ccac0d4e43c1ed72fe8462c4249ad69c7b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://74552159f12c87cdda09147692b400ccac0d4e43c1ed72fe8462c4249ad69c7b\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-13T06:49:32Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-13T06:49:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lqmjj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-13T06:49:25Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-ncchz\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-13T06:49:34Z is after 2025-08-24T17:21:41Z" Dec 13 06:49:34 crc kubenswrapper[4971]: I1213 06:49:34.050470 4971 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-gkcd8" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"13f5c327-f95e-42b4-8783-218500c17694\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T06:49:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T06:49:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T06:49:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T06:49:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://77e846e36f1cb40d03471dc03de11b6c3da448baa407aad38093fae9a801cf92\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-13T06:49:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gvlhn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-13T06:49:27Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-gkcd8\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-13T06:49:34Z is after 2025-08-24T17:21:41Z" Dec 13 06:49:34 crc kubenswrapper[4971]: I1213 06:49:34.072310 4971 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"1ed9f82e-fe5c-4b76-a619-62e5abd88d3b\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T06:49:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T06:49:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T06:49:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T06:49:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T06:49:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e457329a78a2d24337f994e8a0a1f75a2a667244aa75eb84613d1e7138364aac\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-13T06:49:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://eb014de8140deef097d9bd94f3b9fa4ab3558bc4449bef84b83bf62bfabf5836\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-13T06:49:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://2c438cc74d555f7c9db4569d1c56b78b1088ed8105697d2c0f31e6c5136f3dd0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-13T06:49:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://aa6fa6e4340dcd066252df3c56d79a3a39ed531d5eea8d1d5941d6952368d3af\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-13T06:49:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://c4471b98854c52d5cd084d0ba3e64e5b08e9518ec3c372cce2c875a440522457\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-13T06:49:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7d2746335c98dae3a0f43024d34cb62ebcb42a01c8f9ce18d52b29d4c45598e5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://7d2746335c98dae3a0f43024d34cb62ebcb42a01c8f9ce18d52b29d4c45598e5\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-13T06:49:04Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-13T06:49:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://bc0e77d10a738128bea82122496bc42235fc1688f8907756d93e61b651902b57\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://bc0e77d10a738128bea82122496bc42235fc1688f8907756d93e61b651902b57\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-13T06:49:05Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-13T06:49:05Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://f821e418cba5e832da8839268bcbbd9630bdf319a212d5382bfa2c4b0b73a268\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://f821e418cba5e832da8839268bcbbd9630bdf319a212d5382bfa2c4b0b73a268\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-13T06:49:07Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-13T06:49:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-13T06:49:03Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-13T06:49:34Z is after 2025-08-24T17:21:41Z" Dec 13 06:49:34 crc kubenswrapper[4971]: I1213 06:49:34.084848 4971 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-13T06:49:24Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-13T06:49:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-13T06:49:24Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4328cd6a8464ed468229bbfbc8265f969427916d72ebb92237dbe68efe70fa47\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-13T06:49:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://9d715ca82d9b5a594e41f2b98da6f4277da0e4e1dec114969f901c00f6d7553d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-13T06:49:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-13T06:49:34Z is after 2025-08-24T17:21:41Z" Dec 13 06:49:34 crc kubenswrapper[4971]: I1213 06:49:34.095695 4971 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-5rzh5" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"6bf4591d-2483-4132-9672-d0fd6c995968\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T06:49:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T06:49:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T06:49:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T06:49:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://50e2cb14e4ffb603acb06e2281615a99f06182d345c4e70f2446fa2077033bbb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-13T06:49:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dfxqn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-13T06:49:25Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-5rzh5\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-13T06:49:34Z is after 2025-08-24T17:21:41Z" Dec 13 06:49:34 crc kubenswrapper[4971]: I1213 06:49:34.110208 4971 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3ad3268a-a982-428c-9c06-4aaeddd372d9\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T06:49:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T06:49:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T06:49:03Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T06:49:03Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T06:49:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3716afdf2bc693be80a09fc9f7727d334f435caf1d4c8ec4811b3a1f1652cdfe\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-13T06:49:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://30df3abf8c79db8af52c6a918f73e1ce410ddafcf341636424c267e1eeb195a7\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-13T06:49:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://ccdf65f6d2b1eb3647a03ba37c305c32e34addfdf79ff937c8dddaa3e5d5b5c6\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-13T06:49:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://770b73dd0c2b6f38744e2c1cd6aa186a69bb87d1cd015a603ad2e33be06a8f6b\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://b3c1bab3d53f975b2e24582ce2771bc26d600e0dd7c30339dcfd18a071162ad4\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-12-13T06:49:23Z\\\",\\\"message\\\":\\\"file observer\\\\nW1213 06:49:23.283403 1 builder.go:272] unable to get owner reference (falling back to namespace): pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\nI1213 06:49:23.284336 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1213 06:49:23.286978 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-981773671/tls.crt::/tmp/serving-cert-981773671/tls.key\\\\\\\"\\\\nI1213 06:49:23.775072 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1213 06:49:23.778615 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1213 06:49:23.778645 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1213 06:49:23.778680 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1213 06:49:23.778691 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1213 06:49:23.785058 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1213 06:49:23.786150 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1213 06:49:23.786196 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1213 06:49:23.786223 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1213 06:49:23.786246 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1213 06:49:23.786269 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1213 06:49:23.786291 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI1213 06:49:23.785608 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF1213 06:49:23.787563 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-13T06:49:08Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-13T06:49:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://9acd07066deed483bd7fa61ee45fda09539d110717a134a56827f871186f92fc\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-13T06:49:07Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://38c91cfbaecfdbd3e3b663422a91ab4294610afb6bbf9ebdb77db24e2b76c682\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://38c91cfbaecfdbd3e3b663422a91ab4294610afb6bbf9ebdb77db24e2b76c682\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-13T06:49:05Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-13T06:49:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-13T06:49:03Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-13T06:49:34Z is after 2025-08-24T17:21:41Z" Dec 13 06:49:34 crc kubenswrapper[4971]: I1213 06:49:34.121002 4971 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 13 06:49:34 crc kubenswrapper[4971]: I1213 06:49:34.121039 4971 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 13 06:49:34 crc kubenswrapper[4971]: I1213 06:49:34.121078 4971 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 13 06:49:34 crc kubenswrapper[4971]: I1213 06:49:34.121095 4971 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 13 06:49:34 crc kubenswrapper[4971]: I1213 06:49:34.121106 4971 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-13T06:49:34Z","lastTransitionTime":"2025-12-13T06:49:34Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 13 06:49:34 crc kubenswrapper[4971]: I1213 06:49:34.122055 4971 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-13T06:49:23Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-13T06:49:23Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-13T06:49:34Z is after 2025-08-24T17:21:41Z" Dec 13 06:49:34 crc kubenswrapper[4971]: I1213 06:49:34.134384 4971 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-13T06:49:23Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-13T06:49:23Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-13T06:49:34Z is after 2025-08-24T17:21:41Z" Dec 13 06:49:34 crc kubenswrapper[4971]: I1213 06:49:34.158613 4971 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-zlcg5" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"736c6b1a-93dd-4f71-81b6-1c43fcf5f556\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T06:49:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T06:49:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T06:49:25Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T06:49:25Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://325223543b3c7db417f769533559a1ca2b23a723494e1885b0fd0480b5a6399f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-13T06:49:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vzmsr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://f22a4ec756119c4f791fb5105cf1f31df3e548b2b07b1f909a62802e9b6035ae\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-13T06:49:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vzmsr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://f6073140d9f86929b6d4bd2960ede0617a5c9620b269bbab043bae37e0d389f5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-13T06:49:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vzmsr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://693a2ab434c250cb140a866e8068b156df9191414fed1d63bee919a2528ea6e5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-13T06:49:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vzmsr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://219620f51474e60b3e1f880c0972eb612dffc4bad24462bbba2c093ae817d852\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-13T06:49:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vzmsr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://f137f4b441c2402093ab430a9689a46221c4fb4bc70d99fb992f62b0a989c465\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-13T06:49:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vzmsr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://451040d83e76225457d750a4aa532aa2772decb89719e53fe001bf74aaa8e73b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-13T06:49:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vzmsr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://868049c6a3a4ebe4036d1d22e02ba6b76bced2ee7e269bb228accc4a0636678a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-13T06:49:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vzmsr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://451b22df6973ddc8fa2209a65d64d9e41b33eb231bb4adfab88cf0e5943bb246\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://451b22df6973ddc8fa2209a65d64d9e41b33eb231bb4adfab88cf0e5943bb246\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-13T06:49:26Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-13T06:49:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vzmsr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-13T06:49:25Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-zlcg5\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-13T06:49:34Z is after 2025-08-24T17:21:41Z" Dec 13 06:49:34 crc kubenswrapper[4971]: I1213 06:49:34.174598 4971 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-13T06:49:24Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-13T06:49:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-13T06:49:24Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d0b2ecf0d5b22f23217f42f259ae7ffb6359555b22c30eaebf667eaf254b3bab\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-13T06:49:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-13T06:49:34Z is after 2025-08-24T17:21:41Z" Dec 13 06:49:34 crc kubenswrapper[4971]: I1213 06:49:34.194449 4971 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-13T06:49:23Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-13T06:49:23Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-13T06:49:34Z is after 2025-08-24T17:21:41Z" Dec 13 06:49:34 crc kubenswrapper[4971]: I1213 06:49:34.210731 4971 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-13T06:49:26Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-13T06:49:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-13T06:49:26Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9c8a0d1db0e754462d805e82daf5f52ffd5457507a30456535ec59e096c9d092\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-13T06:49:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-13T06:49:34Z is after 2025-08-24T17:21:41Z" Dec 13 06:49:34 crc kubenswrapper[4971]: I1213 06:49:34.223279 4971 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-82xjz" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e0c4af12-edda-41d7-9b44-f87396174ef2\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T06:49:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T06:49:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T06:49:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T06:49:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://0f4577d9e37f860813e221f2e208177d8d383fd04f673ce357a0a405ea29d1e6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-13T06:49:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cnxsq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://7e0d2020e62f1a277cff7047e1907dc747ed393a1f6c983e1ee258bd156c16c7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-13T06:49:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cnxsq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-13T06:49:25Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-82xjz\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-13T06:49:34Z is after 2025-08-24T17:21:41Z" Dec 13 06:49:34 crc kubenswrapper[4971]: I1213 06:49:34.228048 4971 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 13 06:49:34 crc kubenswrapper[4971]: I1213 06:49:34.228079 4971 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 13 06:49:34 crc kubenswrapper[4971]: I1213 06:49:34.228110 4971 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 13 06:49:34 crc kubenswrapper[4971]: I1213 06:49:34.228126 4971 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 13 06:49:34 crc kubenswrapper[4971]: I1213 06:49:34.228136 4971 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-13T06:49:34Z","lastTransitionTime":"2025-12-13T06:49:34Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 13 06:49:34 crc kubenswrapper[4971]: I1213 06:49:34.241227 4971 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-wwvpl" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"dddf744b-2122-49ad-a980-105c09636e0f\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T06:49:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T06:49:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T06:49:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T06:49:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4676d897c7fc668587faec563f3551a056557879cee3e865192d2bd28edf62fc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-13T06:49:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-czz9v\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-13T06:49:25Z\\\"}}\" for pod \"openshift-multus\"/\"multus-wwvpl\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-13T06:49:34Z is after 2025-08-24T17:21:41Z" Dec 13 06:49:34 crc kubenswrapper[4971]: I1213 06:49:34.330966 4971 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 13 06:49:34 crc kubenswrapper[4971]: I1213 06:49:34.331012 4971 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 13 06:49:34 crc kubenswrapper[4971]: I1213 06:49:34.331021 4971 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 13 06:49:34 crc kubenswrapper[4971]: I1213 06:49:34.331036 4971 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 13 06:49:34 crc kubenswrapper[4971]: I1213 06:49:34.331047 4971 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-13T06:49:34Z","lastTransitionTime":"2025-12-13T06:49:34Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 13 06:49:34 crc kubenswrapper[4971]: I1213 06:49:34.433833 4971 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 13 06:49:34 crc kubenswrapper[4971]: I1213 06:49:34.433872 4971 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 13 06:49:34 crc kubenswrapper[4971]: I1213 06:49:34.433882 4971 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 13 06:49:34 crc kubenswrapper[4971]: I1213 06:49:34.433898 4971 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 13 06:49:34 crc kubenswrapper[4971]: I1213 06:49:34.433908 4971 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-13T06:49:34Z","lastTransitionTime":"2025-12-13T06:49:34Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 13 06:49:34 crc kubenswrapper[4971]: I1213 06:49:34.536375 4971 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 13 06:49:34 crc kubenswrapper[4971]: I1213 06:49:34.536417 4971 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 13 06:49:34 crc kubenswrapper[4971]: I1213 06:49:34.536428 4971 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 13 06:49:34 crc kubenswrapper[4971]: I1213 06:49:34.536443 4971 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 13 06:49:34 crc kubenswrapper[4971]: I1213 06:49:34.536454 4971 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-13T06:49:34Z","lastTransitionTime":"2025-12-13T06:49:34Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 13 06:49:34 crc kubenswrapper[4971]: I1213 06:49:34.639027 4971 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 13 06:49:34 crc kubenswrapper[4971]: I1213 06:49:34.639072 4971 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 13 06:49:34 crc kubenswrapper[4971]: I1213 06:49:34.639087 4971 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 13 06:49:34 crc kubenswrapper[4971]: I1213 06:49:34.639111 4971 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 13 06:49:34 crc kubenswrapper[4971]: I1213 06:49:34.639125 4971 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-13T06:49:34Z","lastTransitionTime":"2025-12-13T06:49:34Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 13 06:49:34 crc kubenswrapper[4971]: I1213 06:49:34.741773 4971 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 13 06:49:34 crc kubenswrapper[4971]: I1213 06:49:34.741813 4971 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 13 06:49:34 crc kubenswrapper[4971]: I1213 06:49:34.741823 4971 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 13 06:49:34 crc kubenswrapper[4971]: I1213 06:49:34.741837 4971 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 13 06:49:34 crc kubenswrapper[4971]: I1213 06:49:34.741847 4971 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-13T06:49:34Z","lastTransitionTime":"2025-12-13T06:49:34Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 13 06:49:34 crc kubenswrapper[4971]: I1213 06:49:34.843890 4971 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 13 06:49:34 crc kubenswrapper[4971]: I1213 06:49:34.843931 4971 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 13 06:49:34 crc kubenswrapper[4971]: I1213 06:49:34.843941 4971 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 13 06:49:34 crc kubenswrapper[4971]: I1213 06:49:34.843957 4971 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 13 06:49:34 crc kubenswrapper[4971]: I1213 06:49:34.843967 4971 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-13T06:49:34Z","lastTransitionTime":"2025-12-13T06:49:34Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 13 06:49:34 crc kubenswrapper[4971]: I1213 06:49:34.945772 4971 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 13 06:49:34 crc kubenswrapper[4971]: I1213 06:49:34.945804 4971 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 13 06:49:34 crc kubenswrapper[4971]: I1213 06:49:34.945813 4971 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 13 06:49:34 crc kubenswrapper[4971]: I1213 06:49:34.945826 4971 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 13 06:49:34 crc kubenswrapper[4971]: I1213 06:49:34.945835 4971 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-13T06:49:34Z","lastTransitionTime":"2025-12-13T06:49:34Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 13 06:49:35 crc kubenswrapper[4971]: I1213 06:49:35.006352 4971 prober_manager.go:312] "Failed to trigger a manual run" probe="Readiness" Dec 13 06:49:35 crc kubenswrapper[4971]: I1213 06:49:35.048504 4971 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 13 06:49:35 crc kubenswrapper[4971]: I1213 06:49:35.048560 4971 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 13 06:49:35 crc kubenswrapper[4971]: I1213 06:49:35.048573 4971 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 13 06:49:35 crc kubenswrapper[4971]: I1213 06:49:35.048589 4971 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 13 06:49:35 crc kubenswrapper[4971]: I1213 06:49:35.048605 4971 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-13T06:49:35Z","lastTransitionTime":"2025-12-13T06:49:35Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 13 06:49:35 crc kubenswrapper[4971]: I1213 06:49:35.151140 4971 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 13 06:49:35 crc kubenswrapper[4971]: I1213 06:49:35.151189 4971 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 13 06:49:35 crc kubenswrapper[4971]: I1213 06:49:35.151198 4971 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 13 06:49:35 crc kubenswrapper[4971]: I1213 06:49:35.151213 4971 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 13 06:49:35 crc kubenswrapper[4971]: I1213 06:49:35.151222 4971 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-13T06:49:35Z","lastTransitionTime":"2025-12-13T06:49:35Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 13 06:49:35 crc kubenswrapper[4971]: I1213 06:49:35.254162 4971 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 13 06:49:35 crc kubenswrapper[4971]: I1213 06:49:35.254229 4971 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 13 06:49:35 crc kubenswrapper[4971]: I1213 06:49:35.254239 4971 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 13 06:49:35 crc kubenswrapper[4971]: I1213 06:49:35.254257 4971 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 13 06:49:35 crc kubenswrapper[4971]: I1213 06:49:35.254267 4971 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-13T06:49:35Z","lastTransitionTime":"2025-12-13T06:49:35Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 13 06:49:35 crc kubenswrapper[4971]: I1213 06:49:35.356838 4971 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 13 06:49:35 crc kubenswrapper[4971]: I1213 06:49:35.356911 4971 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 13 06:49:35 crc kubenswrapper[4971]: I1213 06:49:35.356928 4971 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 13 06:49:35 crc kubenswrapper[4971]: I1213 06:49:35.356948 4971 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 13 06:49:35 crc kubenswrapper[4971]: I1213 06:49:35.356960 4971 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-13T06:49:35Z","lastTransitionTime":"2025-12-13T06:49:35Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 13 06:49:35 crc kubenswrapper[4971]: I1213 06:49:35.459429 4971 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 13 06:49:35 crc kubenswrapper[4971]: I1213 06:49:35.459467 4971 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 13 06:49:35 crc kubenswrapper[4971]: I1213 06:49:35.459476 4971 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 13 06:49:35 crc kubenswrapper[4971]: I1213 06:49:35.459490 4971 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 13 06:49:35 crc kubenswrapper[4971]: I1213 06:49:35.459501 4971 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-13T06:49:35Z","lastTransitionTime":"2025-12-13T06:49:35Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 13 06:49:35 crc kubenswrapper[4971]: I1213 06:49:35.561427 4971 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 13 06:49:35 crc kubenswrapper[4971]: I1213 06:49:35.561466 4971 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 13 06:49:35 crc kubenswrapper[4971]: I1213 06:49:35.561478 4971 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 13 06:49:35 crc kubenswrapper[4971]: I1213 06:49:35.561495 4971 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 13 06:49:35 crc kubenswrapper[4971]: I1213 06:49:35.561505 4971 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-13T06:49:35Z","lastTransitionTime":"2025-12-13T06:49:35Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 13 06:49:35 crc kubenswrapper[4971]: I1213 06:49:35.663547 4971 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 13 06:49:35 crc kubenswrapper[4971]: I1213 06:49:35.663588 4971 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 13 06:49:35 crc kubenswrapper[4971]: I1213 06:49:35.663598 4971 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 13 06:49:35 crc kubenswrapper[4971]: I1213 06:49:35.663616 4971 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 13 06:49:35 crc kubenswrapper[4971]: I1213 06:49:35.663628 4971 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-13T06:49:35Z","lastTransitionTime":"2025-12-13T06:49:35Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 13 06:49:35 crc kubenswrapper[4971]: I1213 06:49:35.765917 4971 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 13 06:49:35 crc kubenswrapper[4971]: I1213 06:49:35.765951 4971 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 13 06:49:35 crc kubenswrapper[4971]: I1213 06:49:35.765961 4971 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 13 06:49:35 crc kubenswrapper[4971]: I1213 06:49:35.765975 4971 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 13 06:49:35 crc kubenswrapper[4971]: I1213 06:49:35.765985 4971 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-13T06:49:35Z","lastTransitionTime":"2025-12-13T06:49:35Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 13 06:49:35 crc kubenswrapper[4971]: I1213 06:49:35.768149 4971 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 13 06:49:35 crc kubenswrapper[4971]: I1213 06:49:35.768182 4971 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 13 06:49:35 crc kubenswrapper[4971]: E1213 06:49:35.768285 4971 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 13 06:49:35 crc kubenswrapper[4971]: I1213 06:49:35.768337 4971 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 13 06:49:35 crc kubenswrapper[4971]: E1213 06:49:35.768456 4971 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 13 06:49:35 crc kubenswrapper[4971]: E1213 06:49:35.768593 4971 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 13 06:49:35 crc kubenswrapper[4971]: I1213 06:49:35.868359 4971 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 13 06:49:35 crc kubenswrapper[4971]: I1213 06:49:35.868401 4971 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 13 06:49:35 crc kubenswrapper[4971]: I1213 06:49:35.868409 4971 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 13 06:49:35 crc kubenswrapper[4971]: I1213 06:49:35.868423 4971 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 13 06:49:35 crc kubenswrapper[4971]: I1213 06:49:35.868432 4971 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-13T06:49:35Z","lastTransitionTime":"2025-12-13T06:49:35Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 13 06:49:35 crc kubenswrapper[4971]: I1213 06:49:35.971672 4971 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 13 06:49:35 crc kubenswrapper[4971]: I1213 06:49:35.971720 4971 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 13 06:49:35 crc kubenswrapper[4971]: I1213 06:49:35.971733 4971 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 13 06:49:35 crc kubenswrapper[4971]: I1213 06:49:35.971752 4971 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 13 06:49:35 crc kubenswrapper[4971]: I1213 06:49:35.971775 4971 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-13T06:49:35Z","lastTransitionTime":"2025-12-13T06:49:35Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 13 06:49:36 crc kubenswrapper[4971]: I1213 06:49:36.011454 4971 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-zlcg5_736c6b1a-93dd-4f71-81b6-1c43fcf5f556/ovnkube-controller/0.log" Dec 13 06:49:36 crc kubenswrapper[4971]: I1213 06:49:36.014287 4971 generic.go:334] "Generic (PLEG): container finished" podID="736c6b1a-93dd-4f71-81b6-1c43fcf5f556" containerID="451040d83e76225457d750a4aa532aa2772decb89719e53fe001bf74aaa8e73b" exitCode=1 Dec 13 06:49:36 crc kubenswrapper[4971]: I1213 06:49:36.014332 4971 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-zlcg5" event={"ID":"736c6b1a-93dd-4f71-81b6-1c43fcf5f556","Type":"ContainerDied","Data":"451040d83e76225457d750a4aa532aa2772decb89719e53fe001bf74aaa8e73b"} Dec 13 06:49:36 crc kubenswrapper[4971]: I1213 06:49:36.014960 4971 scope.go:117] "RemoveContainer" containerID="451040d83e76225457d750a4aa532aa2772decb89719e53fe001bf74aaa8e73b" Dec 13 06:49:36 crc kubenswrapper[4971]: I1213 06:49:36.032038 4971 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"f9d31b5a-b770-4a15-95c8-35b49677c608\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T06:49:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T06:49:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T06:49:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T06:49:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T06:49:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c9baa042f9dada954faa57b5454cd6b0a12a297e56ddc192c52cff50e42a16fa\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-13T06:49:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://c788de86754f0bf37ee184672fc1dcb388ccebc246d4f15be4b5da661f725f8f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-13T06:49:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://760817511a9b4f96a68dc38d26631250eaab72e1dd8f123e653039b6b4e91f84\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-13T06:49:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b859be0c9feac856e9855c6a038b02b49b32fade97873c46d0bec0a32d8b5147\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-13T06:49:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-13T06:49:03Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-13T06:49:36Z is after 2025-08-24T17:21:41Z" Dec 13 06:49:36 crc kubenswrapper[4971]: I1213 06:49:36.050212 4971 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-ncchz" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"de914d21-a688-401e-a66e-861beb6336d9\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T06:49:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T06:49:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T06:49:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T06:49:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://244ffe3dc9d8cf799aff4c4d795bf41554c1e6bb6667fbb0c7df322fb5f96c68\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-13T06:49:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lqmjj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://0b7111af6c5a198d02bb02858837cfb4f3b20828d8d206cd18aa0ffa102d0cd4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://0b7111af6c5a198d02bb02858837cfb4f3b20828d8d206cd18aa0ffa102d0cd4\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-13T06:49:26Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-13T06:49:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lqmjj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b36777fe40e74ae1cecb7b73e204139db5279ba51cef4f1e8c4a4c491d8be403\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://b36777fe40e74ae1cecb7b73e204139db5279ba51cef4f1e8c4a4c491d8be403\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-13T06:49:28Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-13T06:49:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lqmjj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://0597a06cc663e416142311a16941d98297577693a1fe27201ea42c27496cace2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://0597a06cc663e416142311a16941d98297577693a1fe27201ea42c27496cace2\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-13T06:49:29Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-13T06:49:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lqmjj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://3f52c2e22a75c010993ca5809a3a9d1304758d4cb6e83691ef1af15a4b514ac4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://3f52c2e22a75c010993ca5809a3a9d1304758d4cb6e83691ef1af15a4b514ac4\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-13T06:49:30Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-13T06:49:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lqmjj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://9e21ca708c4c4c1228b417cece51299e0792729b7c78884bd18b2dda8fe8355c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://9e21ca708c4c4c1228b417cece51299e0792729b7c78884bd18b2dda8fe8355c\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-13T06:49:31Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-13T06:49:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lqmjj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://74552159f12c87cdda09147692b400ccac0d4e43c1ed72fe8462c4249ad69c7b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://74552159f12c87cdda09147692b400ccac0d4e43c1ed72fe8462c4249ad69c7b\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-13T06:49:32Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-13T06:49:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lqmjj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-13T06:49:25Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-ncchz\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-13T06:49:36Z is after 2025-08-24T17:21:41Z" Dec 13 06:49:36 crc kubenswrapper[4971]: I1213 06:49:36.065291 4971 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-gkcd8" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"13f5c327-f95e-42b4-8783-218500c17694\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T06:49:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T06:49:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T06:49:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T06:49:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://77e846e36f1cb40d03471dc03de11b6c3da448baa407aad38093fae9a801cf92\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-13T06:49:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gvlhn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-13T06:49:27Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-gkcd8\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-13T06:49:36Z is after 2025-08-24T17:21:41Z" Dec 13 06:49:36 crc kubenswrapper[4971]: I1213 06:49:36.075408 4971 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 13 06:49:36 crc kubenswrapper[4971]: I1213 06:49:36.075448 4971 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 13 06:49:36 crc kubenswrapper[4971]: I1213 06:49:36.076056 4971 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 13 06:49:36 crc kubenswrapper[4971]: I1213 06:49:36.076086 4971 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 13 06:49:36 crc kubenswrapper[4971]: I1213 06:49:36.076335 4971 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-13T06:49:36Z","lastTransitionTime":"2025-12-13T06:49:36Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 13 06:49:36 crc kubenswrapper[4971]: I1213 06:49:36.090184 4971 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"1ed9f82e-fe5c-4b76-a619-62e5abd88d3b\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T06:49:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T06:49:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T06:49:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T06:49:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T06:49:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e457329a78a2d24337f994e8a0a1f75a2a667244aa75eb84613d1e7138364aac\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-13T06:49:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://eb014de8140deef097d9bd94f3b9fa4ab3558bc4449bef84b83bf62bfabf5836\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-13T06:49:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://2c438cc74d555f7c9db4569d1c56b78b1088ed8105697d2c0f31e6c5136f3dd0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-13T06:49:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://aa6fa6e4340dcd066252df3c56d79a3a39ed531d5eea8d1d5941d6952368d3af\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-13T06:49:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://c4471b98854c52d5cd084d0ba3e64e5b08e9518ec3c372cce2c875a440522457\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-13T06:49:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7d2746335c98dae3a0f43024d34cb62ebcb42a01c8f9ce18d52b29d4c45598e5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://7d2746335c98dae3a0f43024d34cb62ebcb42a01c8f9ce18d52b29d4c45598e5\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-13T06:49:04Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-13T06:49:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://bc0e77d10a738128bea82122496bc42235fc1688f8907756d93e61b651902b57\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://bc0e77d10a738128bea82122496bc42235fc1688f8907756d93e61b651902b57\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-13T06:49:05Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-13T06:49:05Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://f821e418cba5e832da8839268bcbbd9630bdf319a212d5382bfa2c4b0b73a268\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://f821e418cba5e832da8839268bcbbd9630bdf319a212d5382bfa2c4b0b73a268\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-13T06:49:07Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-13T06:49:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-13T06:49:03Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-13T06:49:36Z is after 2025-08-24T17:21:41Z" Dec 13 06:49:36 crc kubenswrapper[4971]: I1213 06:49:36.105608 4971 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-13T06:49:24Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-13T06:49:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-13T06:49:24Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4328cd6a8464ed468229bbfbc8265f969427916d72ebb92237dbe68efe70fa47\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-13T06:49:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://9d715ca82d9b5a594e41f2b98da6f4277da0e4e1dec114969f901c00f6d7553d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-13T06:49:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-13T06:49:36Z is after 2025-08-24T17:21:41Z" Dec 13 06:49:36 crc kubenswrapper[4971]: I1213 06:49:36.116469 4971 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-5rzh5" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"6bf4591d-2483-4132-9672-d0fd6c995968\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T06:49:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T06:49:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T06:49:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T06:49:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://50e2cb14e4ffb603acb06e2281615a99f06182d345c4e70f2446fa2077033bbb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-13T06:49:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dfxqn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-13T06:49:25Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-5rzh5\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-13T06:49:36Z is after 2025-08-24T17:21:41Z" Dec 13 06:49:36 crc kubenswrapper[4971]: I1213 06:49:36.131669 4971 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3ad3268a-a982-428c-9c06-4aaeddd372d9\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T06:49:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T06:49:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T06:49:03Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T06:49:03Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T06:49:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3716afdf2bc693be80a09fc9f7727d334f435caf1d4c8ec4811b3a1f1652cdfe\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-13T06:49:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://30df3abf8c79db8af52c6a918f73e1ce410ddafcf341636424c267e1eeb195a7\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-13T06:49:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://ccdf65f6d2b1eb3647a03ba37c305c32e34addfdf79ff937c8dddaa3e5d5b5c6\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-13T06:49:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://770b73dd0c2b6f38744e2c1cd6aa186a69bb87d1cd015a603ad2e33be06a8f6b\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://b3c1bab3d53f975b2e24582ce2771bc26d600e0dd7c30339dcfd18a071162ad4\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-12-13T06:49:23Z\\\",\\\"message\\\":\\\"file observer\\\\nW1213 06:49:23.283403 1 builder.go:272] unable to get owner reference (falling back to namespace): pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\nI1213 06:49:23.284336 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1213 06:49:23.286978 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-981773671/tls.crt::/tmp/serving-cert-981773671/tls.key\\\\\\\"\\\\nI1213 06:49:23.775072 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1213 06:49:23.778615 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1213 06:49:23.778645 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1213 06:49:23.778680 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1213 06:49:23.778691 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1213 06:49:23.785058 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1213 06:49:23.786150 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1213 06:49:23.786196 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1213 06:49:23.786223 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1213 06:49:23.786246 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1213 06:49:23.786269 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1213 06:49:23.786291 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI1213 06:49:23.785608 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF1213 06:49:23.787563 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-13T06:49:08Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-13T06:49:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://9acd07066deed483bd7fa61ee45fda09539d110717a134a56827f871186f92fc\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-13T06:49:07Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://38c91cfbaecfdbd3e3b663422a91ab4294610afb6bbf9ebdb77db24e2b76c682\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://38c91cfbaecfdbd3e3b663422a91ab4294610afb6bbf9ebdb77db24e2b76c682\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-13T06:49:05Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-13T06:49:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-13T06:49:03Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-13T06:49:36Z is after 2025-08-24T17:21:41Z" Dec 13 06:49:36 crc kubenswrapper[4971]: I1213 06:49:36.146690 4971 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-13T06:49:23Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-13T06:49:23Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-13T06:49:36Z is after 2025-08-24T17:21:41Z" Dec 13 06:49:36 crc kubenswrapper[4971]: I1213 06:49:36.159842 4971 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-13T06:49:23Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-13T06:49:23Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-13T06:49:36Z is after 2025-08-24T17:21:41Z" Dec 13 06:49:36 crc kubenswrapper[4971]: I1213 06:49:36.176285 4971 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-13T06:49:24Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-13T06:49:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-13T06:49:24Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d0b2ecf0d5b22f23217f42f259ae7ffb6359555b22c30eaebf667eaf254b3bab\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-13T06:49:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-13T06:49:36Z is after 2025-08-24T17:21:41Z" Dec 13 06:49:36 crc kubenswrapper[4971]: I1213 06:49:36.179864 4971 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 13 06:49:36 crc kubenswrapper[4971]: I1213 06:49:36.179890 4971 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 13 06:49:36 crc kubenswrapper[4971]: I1213 06:49:36.179899 4971 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 13 06:49:36 crc kubenswrapper[4971]: I1213 06:49:36.179914 4971 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 13 06:49:36 crc kubenswrapper[4971]: I1213 06:49:36.179925 4971 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-13T06:49:36Z","lastTransitionTime":"2025-12-13T06:49:36Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 13 06:49:36 crc kubenswrapper[4971]: I1213 06:49:36.195256 4971 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-13T06:49:23Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-13T06:49:23Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-13T06:49:36Z is after 2025-08-24T17:21:41Z" Dec 13 06:49:36 crc kubenswrapper[4971]: I1213 06:49:36.209963 4971 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-13T06:49:26Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-13T06:49:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-13T06:49:26Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9c8a0d1db0e754462d805e82daf5f52ffd5457507a30456535ec59e096c9d092\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-13T06:49:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-13T06:49:36Z is after 2025-08-24T17:21:41Z" Dec 13 06:49:36 crc kubenswrapper[4971]: I1213 06:49:36.224501 4971 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-82xjz" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e0c4af12-edda-41d7-9b44-f87396174ef2\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T06:49:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T06:49:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T06:49:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T06:49:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://0f4577d9e37f860813e221f2e208177d8d383fd04f673ce357a0a405ea29d1e6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-13T06:49:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cnxsq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://7e0d2020e62f1a277cff7047e1907dc747ed393a1f6c983e1ee258bd156c16c7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-13T06:49:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cnxsq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-13T06:49:25Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-82xjz\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-13T06:49:36Z is after 2025-08-24T17:21:41Z" Dec 13 06:49:36 crc kubenswrapper[4971]: I1213 06:49:36.236587 4971 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-wwvpl" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"dddf744b-2122-49ad-a980-105c09636e0f\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T06:49:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T06:49:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T06:49:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T06:49:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4676d897c7fc668587faec563f3551a056557879cee3e865192d2bd28edf62fc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-13T06:49:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-czz9v\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-13T06:49:25Z\\\"}}\" for pod \"openshift-multus\"/\"multus-wwvpl\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-13T06:49:36Z is after 2025-08-24T17:21:41Z" Dec 13 06:49:36 crc kubenswrapper[4971]: I1213 06:49:36.255033 4971 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-zlcg5" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"736c6b1a-93dd-4f71-81b6-1c43fcf5f556\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T06:49:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T06:49:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T06:49:25Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T06:49:25Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://325223543b3c7db417f769533559a1ca2b23a723494e1885b0fd0480b5a6399f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-13T06:49:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vzmsr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://f22a4ec756119c4f791fb5105cf1f31df3e548b2b07b1f909a62802e9b6035ae\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-13T06:49:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vzmsr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://f6073140d9f86929b6d4bd2960ede0617a5c9620b269bbab043bae37e0d389f5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-13T06:49:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vzmsr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://693a2ab434c250cb140a866e8068b156df9191414fed1d63bee919a2528ea6e5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-13T06:49:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vzmsr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://219620f51474e60b3e1f880c0972eb612dffc4bad24462bbba2c093ae817d852\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-13T06:49:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vzmsr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://f137f4b441c2402093ab430a9689a46221c4fb4bc70d99fb992f62b0a989c465\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-13T06:49:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vzmsr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://451040d83e76225457d750a4aa532aa2772decb89719e53fe001bf74aaa8e73b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://451040d83e76225457d750a4aa532aa2772decb89719e53fe001bf74aaa8e73b\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-12-13T06:49:35Z\\\",\\\"message\\\":\\\"/pkg/client/informers/externalversions/factory.go:141\\\\nI1213 06:49:35.122568 6277 handler.go:190] Sending *v1.Namespace event handler 1 for removal\\\\nI1213 06:49:35.122615 6277 handler.go:190] Sending *v1.Namespace event handler 5 for removal\\\\nI1213 06:49:35.122659 6277 handler.go:190] Sending *v1.Node event handler 7 for removal\\\\nI1213 06:49:35.122681 6277 handler.go:190] Sending *v1.Node event handler 2 for removal\\\\nI1213 06:49:35.122723 6277 handler.go:190] Sending *v1.NetworkPolicy event handler 4 for removal\\\\nI1213 06:49:35.122771 6277 handler.go:190] Sending *v1.EgressIP event handler 8 for removal\\\\nI1213 06:49:35.122815 6277 handler.go:190] Sending *v1.EgressFirewall event handler 9 for removal\\\\nI1213 06:49:35.122852 6277 factory.go:656] Stopping watch factory\\\\nI1213 06:49:35.122935 6277 ovnkube.go:599] Stopped ovnkube\\\\nI1213 06:49:35.122983 6277 handler.go:208] Removed *v1.Namespace event handler 1\\\\nI1213 06:49:35.123009 6277 handler.go:208] Removed *v1.Namespace event handler 5\\\\nI1213 06:49:35.123033 6277 handler.go:208] Removed *v1.Node event handler 7\\\\nI1213 06:49:35.123056 6277 handler.go:208] Removed *v1.Node event handler 2\\\\nI1213 06:49:35.123079 6277 handler.go:208] Removed *v1.NetworkPolicy event handler 4\\\\nI1213 06:49:35.123102 6277 handler.go:208] Removed *v1.EgressIP event handler 8\\\\nI1213 06:49:3\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-13T06:49:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vzmsr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://868049c6a3a4ebe4036d1d22e02ba6b76bced2ee7e269bb228accc4a0636678a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-13T06:49:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vzmsr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://451b22df6973ddc8fa2209a65d64d9e41b33eb231bb4adfab88cf0e5943bb246\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://451b22df6973ddc8fa2209a65d64d9e41b33eb231bb4adfab88cf0e5943bb246\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-13T06:49:26Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-13T06:49:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vzmsr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-13T06:49:25Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-zlcg5\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-13T06:49:36Z is after 2025-08-24T17:21:41Z" Dec 13 06:49:36 crc kubenswrapper[4971]: I1213 06:49:36.282475 4971 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 13 06:49:36 crc kubenswrapper[4971]: I1213 06:49:36.282526 4971 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 13 06:49:36 crc kubenswrapper[4971]: I1213 06:49:36.282535 4971 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 13 06:49:36 crc kubenswrapper[4971]: I1213 06:49:36.282548 4971 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 13 06:49:36 crc kubenswrapper[4971]: I1213 06:49:36.282559 4971 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-13T06:49:36Z","lastTransitionTime":"2025-12-13T06:49:36Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 13 06:49:36 crc kubenswrapper[4971]: I1213 06:49:36.384755 4971 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 13 06:49:36 crc kubenswrapper[4971]: I1213 06:49:36.385633 4971 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 13 06:49:36 crc kubenswrapper[4971]: I1213 06:49:36.385716 4971 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 13 06:49:36 crc kubenswrapper[4971]: I1213 06:49:36.385751 4971 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 13 06:49:36 crc kubenswrapper[4971]: I1213 06:49:36.385788 4971 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-13T06:49:36Z","lastTransitionTime":"2025-12-13T06:49:36Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 13 06:49:36 crc kubenswrapper[4971]: I1213 06:49:36.488490 4971 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 13 06:49:36 crc kubenswrapper[4971]: I1213 06:49:36.488575 4971 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 13 06:49:36 crc kubenswrapper[4971]: I1213 06:49:36.488585 4971 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 13 06:49:36 crc kubenswrapper[4971]: I1213 06:49:36.488599 4971 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 13 06:49:36 crc kubenswrapper[4971]: I1213 06:49:36.488613 4971 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-13T06:49:36Z","lastTransitionTime":"2025-12-13T06:49:36Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 13 06:49:36 crc kubenswrapper[4971]: I1213 06:49:36.591714 4971 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 13 06:49:36 crc kubenswrapper[4971]: I1213 06:49:36.591776 4971 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 13 06:49:36 crc kubenswrapper[4971]: I1213 06:49:36.591788 4971 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 13 06:49:36 crc kubenswrapper[4971]: I1213 06:49:36.591808 4971 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 13 06:49:36 crc kubenswrapper[4971]: I1213 06:49:36.591820 4971 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-13T06:49:36Z","lastTransitionTime":"2025-12-13T06:49:36Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 13 06:49:36 crc kubenswrapper[4971]: I1213 06:49:36.694337 4971 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 13 06:49:36 crc kubenswrapper[4971]: I1213 06:49:36.694399 4971 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 13 06:49:36 crc kubenswrapper[4971]: I1213 06:49:36.694411 4971 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 13 06:49:36 crc kubenswrapper[4971]: I1213 06:49:36.694426 4971 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 13 06:49:36 crc kubenswrapper[4971]: I1213 06:49:36.694437 4971 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-13T06:49:36Z","lastTransitionTime":"2025-12-13T06:49:36Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 13 06:49:36 crc kubenswrapper[4971]: I1213 06:49:36.797164 4971 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 13 06:49:36 crc kubenswrapper[4971]: I1213 06:49:36.797214 4971 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 13 06:49:36 crc kubenswrapper[4971]: I1213 06:49:36.797230 4971 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 13 06:49:36 crc kubenswrapper[4971]: I1213 06:49:36.797254 4971 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 13 06:49:36 crc kubenswrapper[4971]: I1213 06:49:36.797271 4971 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-13T06:49:36Z","lastTransitionTime":"2025-12-13T06:49:36Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 13 06:49:36 crc kubenswrapper[4971]: I1213 06:49:36.900414 4971 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 13 06:49:36 crc kubenswrapper[4971]: I1213 06:49:36.900469 4971 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 13 06:49:36 crc kubenswrapper[4971]: I1213 06:49:36.900481 4971 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 13 06:49:36 crc kubenswrapper[4971]: I1213 06:49:36.900501 4971 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 13 06:49:36 crc kubenswrapper[4971]: I1213 06:49:36.900533 4971 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-13T06:49:36Z","lastTransitionTime":"2025-12-13T06:49:36Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 13 06:49:37 crc kubenswrapper[4971]: I1213 06:49:37.003327 4971 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 13 06:49:37 crc kubenswrapper[4971]: I1213 06:49:37.003378 4971 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 13 06:49:37 crc kubenswrapper[4971]: I1213 06:49:37.003389 4971 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 13 06:49:37 crc kubenswrapper[4971]: I1213 06:49:37.003404 4971 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 13 06:49:37 crc kubenswrapper[4971]: I1213 06:49:37.003414 4971 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-13T06:49:37Z","lastTransitionTime":"2025-12-13T06:49:37Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 13 06:49:37 crc kubenswrapper[4971]: I1213 06:49:37.020338 4971 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-zlcg5_736c6b1a-93dd-4f71-81b6-1c43fcf5f556/ovnkube-controller/0.log" Dec 13 06:49:37 crc kubenswrapper[4971]: I1213 06:49:37.023143 4971 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-zlcg5" event={"ID":"736c6b1a-93dd-4f71-81b6-1c43fcf5f556","Type":"ContainerStarted","Data":"d950a29b0f6f47ae1cc90c132adb059b5aff889a0a38de1d1296e0f083da5e51"} Dec 13 06:49:37 crc kubenswrapper[4971]: I1213 06:49:37.023281 4971 prober_manager.go:312] "Failed to trigger a manual run" probe="Readiness" Dec 13 06:49:37 crc kubenswrapper[4971]: I1213 06:49:37.037358 4971 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3ad3268a-a982-428c-9c06-4aaeddd372d9\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T06:49:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T06:49:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T06:49:03Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T06:49:03Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T06:49:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3716afdf2bc693be80a09fc9f7727d334f435caf1d4c8ec4811b3a1f1652cdfe\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-13T06:49:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://30df3abf8c79db8af52c6a918f73e1ce410ddafcf341636424c267e1eeb195a7\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-13T06:49:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://ccdf65f6d2b1eb3647a03ba37c305c32e34addfdf79ff937c8dddaa3e5d5b5c6\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-13T06:49:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://770b73dd0c2b6f38744e2c1cd6aa186a69bb87d1cd015a603ad2e33be06a8f6b\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://b3c1bab3d53f975b2e24582ce2771bc26d600e0dd7c30339dcfd18a071162ad4\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-12-13T06:49:23Z\\\",\\\"message\\\":\\\"file observer\\\\nW1213 06:49:23.283403 1 builder.go:272] unable to get owner reference (falling back to namespace): pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\nI1213 06:49:23.284336 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1213 06:49:23.286978 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-981773671/tls.crt::/tmp/serving-cert-981773671/tls.key\\\\\\\"\\\\nI1213 06:49:23.775072 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1213 06:49:23.778615 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1213 06:49:23.778645 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1213 06:49:23.778680 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1213 06:49:23.778691 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1213 06:49:23.785058 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1213 06:49:23.786150 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1213 06:49:23.786196 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1213 06:49:23.786223 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1213 06:49:23.786246 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1213 06:49:23.786269 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1213 06:49:23.786291 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI1213 06:49:23.785608 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF1213 06:49:23.787563 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-13T06:49:08Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-13T06:49:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://9acd07066deed483bd7fa61ee45fda09539d110717a134a56827f871186f92fc\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-13T06:49:07Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://38c91cfbaecfdbd3e3b663422a91ab4294610afb6bbf9ebdb77db24e2b76c682\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://38c91cfbaecfdbd3e3b663422a91ab4294610afb6bbf9ebdb77db24e2b76c682\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-13T06:49:05Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-13T06:49:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-13T06:49:03Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-13T06:49:37Z is after 2025-08-24T17:21:41Z" Dec 13 06:49:37 crc kubenswrapper[4971]: I1213 06:49:37.050828 4971 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-13T06:49:23Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-13T06:49:23Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-13T06:49:37Z is after 2025-08-24T17:21:41Z" Dec 13 06:49:37 crc kubenswrapper[4971]: I1213 06:49:37.065661 4971 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-13T06:49:23Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-13T06:49:23Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-13T06:49:37Z is after 2025-08-24T17:21:41Z" Dec 13 06:49:37 crc kubenswrapper[4971]: I1213 06:49:37.078425 4971 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-wwvpl" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"dddf744b-2122-49ad-a980-105c09636e0f\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T06:49:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T06:49:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T06:49:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T06:49:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4676d897c7fc668587faec563f3551a056557879cee3e865192d2bd28edf62fc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-13T06:49:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-czz9v\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-13T06:49:25Z\\\"}}\" for pod \"openshift-multus\"/\"multus-wwvpl\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-13T06:49:37Z is after 2025-08-24T17:21:41Z" Dec 13 06:49:37 crc kubenswrapper[4971]: I1213 06:49:37.098198 4971 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-zlcg5" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"736c6b1a-93dd-4f71-81b6-1c43fcf5f556\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T06:49:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T06:49:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T06:49:25Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T06:49:25Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://325223543b3c7db417f769533559a1ca2b23a723494e1885b0fd0480b5a6399f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-13T06:49:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vzmsr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://f22a4ec756119c4f791fb5105cf1f31df3e548b2b07b1f909a62802e9b6035ae\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-13T06:49:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vzmsr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://f6073140d9f86929b6d4bd2960ede0617a5c9620b269bbab043bae37e0d389f5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-13T06:49:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vzmsr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://693a2ab434c250cb140a866e8068b156df9191414fed1d63bee919a2528ea6e5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-13T06:49:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vzmsr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://219620f51474e60b3e1f880c0972eb612dffc4bad24462bbba2c093ae817d852\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-13T06:49:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vzmsr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://f137f4b441c2402093ab430a9689a46221c4fb4bc70d99fb992f62b0a989c465\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-13T06:49:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vzmsr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://d950a29b0f6f47ae1cc90c132adb059b5aff889a0a38de1d1296e0f083da5e51\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://451040d83e76225457d750a4aa532aa2772decb89719e53fe001bf74aaa8e73b\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-12-13T06:49:35Z\\\",\\\"message\\\":\\\"/pkg/client/informers/externalversions/factory.go:141\\\\nI1213 06:49:35.122568 6277 handler.go:190] Sending *v1.Namespace event handler 1 for removal\\\\nI1213 06:49:35.122615 6277 handler.go:190] Sending *v1.Namespace event handler 5 for removal\\\\nI1213 06:49:35.122659 6277 handler.go:190] Sending *v1.Node event handler 7 for removal\\\\nI1213 06:49:35.122681 6277 handler.go:190] Sending *v1.Node event handler 2 for removal\\\\nI1213 06:49:35.122723 6277 handler.go:190] Sending *v1.NetworkPolicy event handler 4 for removal\\\\nI1213 06:49:35.122771 6277 handler.go:190] Sending *v1.EgressIP event handler 8 for removal\\\\nI1213 06:49:35.122815 6277 handler.go:190] Sending *v1.EgressFirewall event handler 9 for removal\\\\nI1213 06:49:35.122852 6277 factory.go:656] Stopping watch factory\\\\nI1213 06:49:35.122935 6277 ovnkube.go:599] Stopped ovnkube\\\\nI1213 06:49:35.122983 6277 handler.go:208] Removed *v1.Namespace event handler 1\\\\nI1213 06:49:35.123009 6277 handler.go:208] Removed *v1.Namespace event handler 5\\\\nI1213 06:49:35.123033 6277 handler.go:208] Removed *v1.Node event handler 7\\\\nI1213 06:49:35.123056 6277 handler.go:208] Removed *v1.Node event handler 2\\\\nI1213 06:49:35.123079 6277 handler.go:208] Removed *v1.NetworkPolicy event handler 4\\\\nI1213 06:49:35.123102 6277 handler.go:208] Removed *v1.EgressIP event handler 8\\\\nI1213 06:49:3\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-13T06:49:32Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-13T06:49:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vzmsr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://868049c6a3a4ebe4036d1d22e02ba6b76bced2ee7e269bb228accc4a0636678a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-13T06:49:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vzmsr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://451b22df6973ddc8fa2209a65d64d9e41b33eb231bb4adfab88cf0e5943bb246\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://451b22df6973ddc8fa2209a65d64d9e41b33eb231bb4adfab88cf0e5943bb246\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-13T06:49:26Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-13T06:49:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vzmsr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-13T06:49:25Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-zlcg5\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-13T06:49:37Z is after 2025-08-24T17:21:41Z" Dec 13 06:49:37 crc kubenswrapper[4971]: I1213 06:49:37.105316 4971 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 13 06:49:37 crc kubenswrapper[4971]: I1213 06:49:37.105353 4971 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 13 06:49:37 crc kubenswrapper[4971]: I1213 06:49:37.105366 4971 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 13 06:49:37 crc kubenswrapper[4971]: I1213 06:49:37.105383 4971 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 13 06:49:37 crc kubenswrapper[4971]: I1213 06:49:37.105395 4971 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-13T06:49:37Z","lastTransitionTime":"2025-12-13T06:49:37Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 13 06:49:37 crc kubenswrapper[4971]: I1213 06:49:37.110613 4971 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-13T06:49:24Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-13T06:49:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-13T06:49:24Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d0b2ecf0d5b22f23217f42f259ae7ffb6359555b22c30eaebf667eaf254b3bab\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-13T06:49:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-13T06:49:37Z is after 2025-08-24T17:21:41Z" Dec 13 06:49:37 crc kubenswrapper[4971]: I1213 06:49:37.123563 4971 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-13T06:49:23Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-13T06:49:23Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-13T06:49:37Z is after 2025-08-24T17:21:41Z" Dec 13 06:49:37 crc kubenswrapper[4971]: I1213 06:49:37.132823 4971 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-13T06:49:26Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-13T06:49:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-13T06:49:26Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9c8a0d1db0e754462d805e82daf5f52ffd5457507a30456535ec59e096c9d092\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-13T06:49:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-13T06:49:37Z is after 2025-08-24T17:21:41Z" Dec 13 06:49:37 crc kubenswrapper[4971]: I1213 06:49:37.141334 4971 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-82xjz" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e0c4af12-edda-41d7-9b44-f87396174ef2\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T06:49:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T06:49:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T06:49:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T06:49:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://0f4577d9e37f860813e221f2e208177d8d383fd04f673ce357a0a405ea29d1e6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-13T06:49:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cnxsq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://7e0d2020e62f1a277cff7047e1907dc747ed393a1f6c983e1ee258bd156c16c7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-13T06:49:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cnxsq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-13T06:49:25Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-82xjz\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-13T06:49:37Z is after 2025-08-24T17:21:41Z" Dec 13 06:49:37 crc kubenswrapper[4971]: I1213 06:49:37.151231 4971 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"f9d31b5a-b770-4a15-95c8-35b49677c608\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T06:49:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T06:49:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T06:49:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T06:49:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T06:49:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c9baa042f9dada954faa57b5454cd6b0a12a297e56ddc192c52cff50e42a16fa\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-13T06:49:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://c788de86754f0bf37ee184672fc1dcb388ccebc246d4f15be4b5da661f725f8f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-13T06:49:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://760817511a9b4f96a68dc38d26631250eaab72e1dd8f123e653039b6b4e91f84\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-13T06:49:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b859be0c9feac856e9855c6a038b02b49b32fade97873c46d0bec0a32d8b5147\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-13T06:49:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-13T06:49:03Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-13T06:49:37Z is after 2025-08-24T17:21:41Z" Dec 13 06:49:37 crc kubenswrapper[4971]: I1213 06:49:37.208453 4971 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 13 06:49:37 crc kubenswrapper[4971]: I1213 06:49:37.208548 4971 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 13 06:49:37 crc kubenswrapper[4971]: I1213 06:49:37.208577 4971 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 13 06:49:37 crc kubenswrapper[4971]: I1213 06:49:37.208602 4971 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 13 06:49:37 crc kubenswrapper[4971]: I1213 06:49:37.208619 4971 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-13T06:49:37Z","lastTransitionTime":"2025-12-13T06:49:37Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 13 06:49:37 crc kubenswrapper[4971]: I1213 06:49:37.312803 4971 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 13 06:49:37 crc kubenswrapper[4971]: I1213 06:49:37.312849 4971 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 13 06:49:37 crc kubenswrapper[4971]: I1213 06:49:37.312861 4971 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 13 06:49:37 crc kubenswrapper[4971]: I1213 06:49:37.312886 4971 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 13 06:49:37 crc kubenswrapper[4971]: I1213 06:49:37.312898 4971 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-13T06:49:37Z","lastTransitionTime":"2025-12-13T06:49:37Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 13 06:49:37 crc kubenswrapper[4971]: I1213 06:49:37.372499 4971 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-ncchz" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"de914d21-a688-401e-a66e-861beb6336d9\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T06:49:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T06:49:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T06:49:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T06:49:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://244ffe3dc9d8cf799aff4c4d795bf41554c1e6bb6667fbb0c7df322fb5f96c68\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-13T06:49:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lqmjj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://0b7111af6c5a198d02bb02858837cfb4f3b20828d8d206cd18aa0ffa102d0cd4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://0b7111af6c5a198d02bb02858837cfb4f3b20828d8d206cd18aa0ffa102d0cd4\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-13T06:49:26Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-13T06:49:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lqmjj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b36777fe40e74ae1cecb7b73e204139db5279ba51cef4f1e8c4a4c491d8be403\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://b36777fe40e74ae1cecb7b73e204139db5279ba51cef4f1e8c4a4c491d8be403\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-13T06:49:28Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-13T06:49:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lqmjj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://0597a06cc663e416142311a16941d98297577693a1fe27201ea42c27496cace2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://0597a06cc663e416142311a16941d98297577693a1fe27201ea42c27496cace2\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-13T06:49:29Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-13T06:49:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lqmjj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://3f52c2e22a75c010993ca5809a3a9d1304758d4cb6e83691ef1af15a4b514ac4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://3f52c2e22a75c010993ca5809a3a9d1304758d4cb6e83691ef1af15a4b514ac4\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-13T06:49:30Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-13T06:49:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lqmjj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://9e21ca708c4c4c1228b417cece51299e0792729b7c78884bd18b2dda8fe8355c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://9e21ca708c4c4c1228b417cece51299e0792729b7c78884bd18b2dda8fe8355c\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-13T06:49:31Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-13T06:49:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lqmjj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://74552159f12c87cdda09147692b400ccac0d4e43c1ed72fe8462c4249ad69c7b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://74552159f12c87cdda09147692b400ccac0d4e43c1ed72fe8462c4249ad69c7b\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-13T06:49:32Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-13T06:49:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lqmjj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-13T06:49:25Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-ncchz\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-13T06:49:37Z is after 2025-08-24T17:21:41Z" Dec 13 06:49:37 crc kubenswrapper[4971]: I1213 06:49:37.384362 4971 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-gkcd8" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"13f5c327-f95e-42b4-8783-218500c17694\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T06:49:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T06:49:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T06:49:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T06:49:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://77e846e36f1cb40d03471dc03de11b6c3da448baa407aad38093fae9a801cf92\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-13T06:49:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gvlhn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-13T06:49:27Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-gkcd8\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-13T06:49:37Z is after 2025-08-24T17:21:41Z" Dec 13 06:49:37 crc kubenswrapper[4971]: I1213 06:49:37.404301 4971 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"1ed9f82e-fe5c-4b76-a619-62e5abd88d3b\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T06:49:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T06:49:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T06:49:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T06:49:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T06:49:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e457329a78a2d24337f994e8a0a1f75a2a667244aa75eb84613d1e7138364aac\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-13T06:49:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://eb014de8140deef097d9bd94f3b9fa4ab3558bc4449bef84b83bf62bfabf5836\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-13T06:49:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://2c438cc74d555f7c9db4569d1c56b78b1088ed8105697d2c0f31e6c5136f3dd0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-13T06:49:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://aa6fa6e4340dcd066252df3c56d79a3a39ed531d5eea8d1d5941d6952368d3af\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-13T06:49:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://c4471b98854c52d5cd084d0ba3e64e5b08e9518ec3c372cce2c875a440522457\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-13T06:49:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7d2746335c98dae3a0f43024d34cb62ebcb42a01c8f9ce18d52b29d4c45598e5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://7d2746335c98dae3a0f43024d34cb62ebcb42a01c8f9ce18d52b29d4c45598e5\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-13T06:49:04Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-13T06:49:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://bc0e77d10a738128bea82122496bc42235fc1688f8907756d93e61b651902b57\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://bc0e77d10a738128bea82122496bc42235fc1688f8907756d93e61b651902b57\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-13T06:49:05Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-13T06:49:05Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://f821e418cba5e832da8839268bcbbd9630bdf319a212d5382bfa2c4b0b73a268\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://f821e418cba5e832da8839268bcbbd9630bdf319a212d5382bfa2c4b0b73a268\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-13T06:49:07Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-13T06:49:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-13T06:49:03Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-13T06:49:37Z is after 2025-08-24T17:21:41Z" Dec 13 06:49:37 crc kubenswrapper[4971]: I1213 06:49:37.415759 4971 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 13 06:49:37 crc kubenswrapper[4971]: I1213 06:49:37.415801 4971 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 13 06:49:37 crc kubenswrapper[4971]: I1213 06:49:37.415812 4971 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 13 06:49:37 crc kubenswrapper[4971]: I1213 06:49:37.415829 4971 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 13 06:49:37 crc kubenswrapper[4971]: I1213 06:49:37.415840 4971 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-13T06:49:37Z","lastTransitionTime":"2025-12-13T06:49:37Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 13 06:49:37 crc kubenswrapper[4971]: I1213 06:49:37.418781 4971 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-13T06:49:24Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-13T06:49:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-13T06:49:24Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4328cd6a8464ed468229bbfbc8265f969427916d72ebb92237dbe68efe70fa47\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-13T06:49:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://9d715ca82d9b5a594e41f2b98da6f4277da0e4e1dec114969f901c00f6d7553d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-13T06:49:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-13T06:49:37Z is after 2025-08-24T17:21:41Z" Dec 13 06:49:37 crc kubenswrapper[4971]: I1213 06:49:37.430031 4971 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-5rzh5" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"6bf4591d-2483-4132-9672-d0fd6c995968\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T06:49:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T06:49:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T06:49:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T06:49:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://50e2cb14e4ffb603acb06e2281615a99f06182d345c4e70f2446fa2077033bbb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-13T06:49:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dfxqn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-13T06:49:25Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-5rzh5\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-13T06:49:37Z is after 2025-08-24T17:21:41Z" Dec 13 06:49:37 crc kubenswrapper[4971]: I1213 06:49:37.518369 4971 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 13 06:49:37 crc kubenswrapper[4971]: I1213 06:49:37.518443 4971 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 13 06:49:37 crc kubenswrapper[4971]: I1213 06:49:37.518455 4971 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 13 06:49:37 crc kubenswrapper[4971]: I1213 06:49:37.518477 4971 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 13 06:49:37 crc kubenswrapper[4971]: I1213 06:49:37.518927 4971 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-13T06:49:37Z","lastTransitionTime":"2025-12-13T06:49:37Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 13 06:49:37 crc kubenswrapper[4971]: I1213 06:49:37.620074 4971 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-kube-apiserver/kube-apiserver-crc" Dec 13 06:49:37 crc kubenswrapper[4971]: I1213 06:49:37.621363 4971 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 13 06:49:37 crc kubenswrapper[4971]: I1213 06:49:37.621399 4971 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 13 06:49:37 crc kubenswrapper[4971]: I1213 06:49:37.621413 4971 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 13 06:49:37 crc kubenswrapper[4971]: I1213 06:49:37.621431 4971 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 13 06:49:37 crc kubenswrapper[4971]: I1213 06:49:37.621443 4971 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-13T06:49:37Z","lastTransitionTime":"2025-12-13T06:49:37Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 13 06:49:37 crc kubenswrapper[4971]: I1213 06:49:37.640030 4971 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-82xjz" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e0c4af12-edda-41d7-9b44-f87396174ef2\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T06:49:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T06:49:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T06:49:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T06:49:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://0f4577d9e37f860813e221f2e208177d8d383fd04f673ce357a0a405ea29d1e6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-13T06:49:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cnxsq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://7e0d2020e62f1a277cff7047e1907dc747ed393a1f6c983e1ee258bd156c16c7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-13T06:49:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cnxsq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-13T06:49:25Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-82xjz\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-13T06:49:37Z is after 2025-08-24T17:21:41Z" Dec 13 06:49:37 crc kubenswrapper[4971]: I1213 06:49:37.653842 4971 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-wwvpl" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"dddf744b-2122-49ad-a980-105c09636e0f\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T06:49:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T06:49:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T06:49:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T06:49:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4676d897c7fc668587faec563f3551a056557879cee3e865192d2bd28edf62fc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-13T06:49:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-czz9v\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-13T06:49:25Z\\\"}}\" for pod \"openshift-multus\"/\"multus-wwvpl\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-13T06:49:37Z is after 2025-08-24T17:21:41Z" Dec 13 06:49:37 crc kubenswrapper[4971]: I1213 06:49:37.706826 4971 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-zlcg5" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"736c6b1a-93dd-4f71-81b6-1c43fcf5f556\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T06:49:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T06:49:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T06:49:25Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T06:49:25Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://325223543b3c7db417f769533559a1ca2b23a723494e1885b0fd0480b5a6399f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-13T06:49:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vzmsr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://f22a4ec756119c4f791fb5105cf1f31df3e548b2b07b1f909a62802e9b6035ae\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-13T06:49:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vzmsr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://f6073140d9f86929b6d4bd2960ede0617a5c9620b269bbab043bae37e0d389f5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-13T06:49:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vzmsr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://693a2ab434c250cb140a866e8068b156df9191414fed1d63bee919a2528ea6e5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-13T06:49:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vzmsr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://219620f51474e60b3e1f880c0972eb612dffc4bad24462bbba2c093ae817d852\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-13T06:49:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vzmsr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://f137f4b441c2402093ab430a9689a46221c4fb4bc70d99fb992f62b0a989c465\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-13T06:49:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vzmsr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://d950a29b0f6f47ae1cc90c132adb059b5aff889a0a38de1d1296e0f083da5e51\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://451040d83e76225457d750a4aa532aa2772decb89719e53fe001bf74aaa8e73b\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-12-13T06:49:35Z\\\",\\\"message\\\":\\\"/pkg/client/informers/externalversions/factory.go:141\\\\nI1213 06:49:35.122568 6277 handler.go:190] Sending *v1.Namespace event handler 1 for removal\\\\nI1213 06:49:35.122615 6277 handler.go:190] Sending *v1.Namespace event handler 5 for removal\\\\nI1213 06:49:35.122659 6277 handler.go:190] Sending *v1.Node event handler 7 for removal\\\\nI1213 06:49:35.122681 6277 handler.go:190] Sending *v1.Node event handler 2 for removal\\\\nI1213 06:49:35.122723 6277 handler.go:190] Sending *v1.NetworkPolicy event handler 4 for removal\\\\nI1213 06:49:35.122771 6277 handler.go:190] Sending *v1.EgressIP event handler 8 for removal\\\\nI1213 06:49:35.122815 6277 handler.go:190] Sending *v1.EgressFirewall event handler 9 for removal\\\\nI1213 06:49:35.122852 6277 factory.go:656] Stopping watch factory\\\\nI1213 06:49:35.122935 6277 ovnkube.go:599] Stopped ovnkube\\\\nI1213 06:49:35.122983 6277 handler.go:208] Removed *v1.Namespace event handler 1\\\\nI1213 06:49:35.123009 6277 handler.go:208] Removed *v1.Namespace event handler 5\\\\nI1213 06:49:35.123033 6277 handler.go:208] Removed *v1.Node event handler 7\\\\nI1213 06:49:35.123056 6277 handler.go:208] Removed *v1.Node event handler 2\\\\nI1213 06:49:35.123079 6277 handler.go:208] Removed *v1.NetworkPolicy event handler 4\\\\nI1213 06:49:35.123102 6277 handler.go:208] Removed *v1.EgressIP event handler 8\\\\nI1213 06:49:3\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-13T06:49:32Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-13T06:49:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vzmsr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://868049c6a3a4ebe4036d1d22e02ba6b76bced2ee7e269bb228accc4a0636678a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-13T06:49:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vzmsr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://451b22df6973ddc8fa2209a65d64d9e41b33eb231bb4adfab88cf0e5943bb246\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://451b22df6973ddc8fa2209a65d64d9e41b33eb231bb4adfab88cf0e5943bb246\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-13T06:49:26Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-13T06:49:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vzmsr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-13T06:49:25Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-zlcg5\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-13T06:49:37Z is after 2025-08-24T17:21:41Z" Dec 13 06:49:37 crc kubenswrapper[4971]: I1213 06:49:37.723456 4971 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 13 06:49:37 crc kubenswrapper[4971]: I1213 06:49:37.723501 4971 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 13 06:49:37 crc kubenswrapper[4971]: I1213 06:49:37.723529 4971 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 13 06:49:37 crc kubenswrapper[4971]: I1213 06:49:37.723548 4971 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 13 06:49:37 crc kubenswrapper[4971]: I1213 06:49:37.723562 4971 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-13T06:49:37Z","lastTransitionTime":"2025-12-13T06:49:37Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 13 06:49:37 crc kubenswrapper[4971]: I1213 06:49:37.731370 4971 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-13T06:49:24Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-13T06:49:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-13T06:49:24Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d0b2ecf0d5b22f23217f42f259ae7ffb6359555b22c30eaebf667eaf254b3bab\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-13T06:49:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-13T06:49:37Z is after 2025-08-24T17:21:41Z" Dec 13 06:49:37 crc kubenswrapper[4971]: I1213 06:49:37.744774 4971 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-13T06:49:23Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-13T06:49:23Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-13T06:49:37Z is after 2025-08-24T17:21:41Z" Dec 13 06:49:37 crc kubenswrapper[4971]: I1213 06:49:37.756917 4971 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-13T06:49:26Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-13T06:49:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-13T06:49:26Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9c8a0d1db0e754462d805e82daf5f52ffd5457507a30456535ec59e096c9d092\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-13T06:49:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-13T06:49:37Z is after 2025-08-24T17:21:41Z" Dec 13 06:49:37 crc kubenswrapper[4971]: I1213 06:49:37.766888 4971 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-gkcd8" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"13f5c327-f95e-42b4-8783-218500c17694\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T06:49:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T06:49:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T06:49:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T06:49:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://77e846e36f1cb40d03471dc03de11b6c3da448baa407aad38093fae9a801cf92\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-13T06:49:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gvlhn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-13T06:49:27Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-gkcd8\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-13T06:49:37Z is after 2025-08-24T17:21:41Z" Dec 13 06:49:37 crc kubenswrapper[4971]: I1213 06:49:37.767993 4971 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 13 06:49:37 crc kubenswrapper[4971]: I1213 06:49:37.768040 4971 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 13 06:49:37 crc kubenswrapper[4971]: E1213 06:49:37.768118 4971 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 13 06:49:37 crc kubenswrapper[4971]: I1213 06:49:37.768131 4971 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 13 06:49:37 crc kubenswrapper[4971]: E1213 06:49:37.768212 4971 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 13 06:49:37 crc kubenswrapper[4971]: E1213 06:49:37.768330 4971 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 13 06:49:37 crc kubenswrapper[4971]: I1213 06:49:37.779911 4971 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"f9d31b5a-b770-4a15-95c8-35b49677c608\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T06:49:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T06:49:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T06:49:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T06:49:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T06:49:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c9baa042f9dada954faa57b5454cd6b0a12a297e56ddc192c52cff50e42a16fa\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-13T06:49:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://c788de86754f0bf37ee184672fc1dcb388ccebc246d4f15be4b5da661f725f8f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-13T06:49:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://760817511a9b4f96a68dc38d26631250eaab72e1dd8f123e653039b6b4e91f84\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-13T06:49:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b859be0c9feac856e9855c6a038b02b49b32fade97873c46d0bec0a32d8b5147\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-13T06:49:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-13T06:49:03Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-13T06:49:37Z is after 2025-08-24T17:21:41Z" Dec 13 06:49:37 crc kubenswrapper[4971]: I1213 06:49:37.793215 4971 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-ncchz" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"de914d21-a688-401e-a66e-861beb6336d9\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T06:49:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T06:49:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T06:49:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T06:49:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://244ffe3dc9d8cf799aff4c4d795bf41554c1e6bb6667fbb0c7df322fb5f96c68\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-13T06:49:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lqmjj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://0b7111af6c5a198d02bb02858837cfb4f3b20828d8d206cd18aa0ffa102d0cd4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://0b7111af6c5a198d02bb02858837cfb4f3b20828d8d206cd18aa0ffa102d0cd4\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-13T06:49:26Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-13T06:49:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lqmjj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b36777fe40e74ae1cecb7b73e204139db5279ba51cef4f1e8c4a4c491d8be403\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://b36777fe40e74ae1cecb7b73e204139db5279ba51cef4f1e8c4a4c491d8be403\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-13T06:49:28Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-13T06:49:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lqmjj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://0597a06cc663e416142311a16941d98297577693a1fe27201ea42c27496cace2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://0597a06cc663e416142311a16941d98297577693a1fe27201ea42c27496cace2\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-13T06:49:29Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-13T06:49:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lqmjj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://3f52c2e22a75c010993ca5809a3a9d1304758d4cb6e83691ef1af15a4b514ac4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://3f52c2e22a75c010993ca5809a3a9d1304758d4cb6e83691ef1af15a4b514ac4\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-13T06:49:30Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-13T06:49:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lqmjj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://9e21ca708c4c4c1228b417cece51299e0792729b7c78884bd18b2dda8fe8355c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://9e21ca708c4c4c1228b417cece51299e0792729b7c78884bd18b2dda8fe8355c\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-13T06:49:31Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-13T06:49:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lqmjj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://74552159f12c87cdda09147692b400ccac0d4e43c1ed72fe8462c4249ad69c7b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://74552159f12c87cdda09147692b400ccac0d4e43c1ed72fe8462c4249ad69c7b\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-13T06:49:32Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-13T06:49:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lqmjj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-13T06:49:25Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-ncchz\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-13T06:49:37Z is after 2025-08-24T17:21:41Z" Dec 13 06:49:37 crc kubenswrapper[4971]: I1213 06:49:37.804235 4971 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-5rzh5" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"6bf4591d-2483-4132-9672-d0fd6c995968\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T06:49:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T06:49:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T06:49:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T06:49:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://50e2cb14e4ffb603acb06e2281615a99f06182d345c4e70f2446fa2077033bbb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-13T06:49:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dfxqn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-13T06:49:25Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-5rzh5\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-13T06:49:37Z is after 2025-08-24T17:21:41Z" Dec 13 06:49:37 crc kubenswrapper[4971]: I1213 06:49:37.826587 4971 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 13 06:49:37 crc kubenswrapper[4971]: I1213 06:49:37.826703 4971 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 13 06:49:37 crc kubenswrapper[4971]: I1213 06:49:37.826718 4971 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 13 06:49:37 crc kubenswrapper[4971]: I1213 06:49:37.826826 4971 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 13 06:49:37 crc kubenswrapper[4971]: I1213 06:49:37.826917 4971 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-13T06:49:37Z","lastTransitionTime":"2025-12-13T06:49:37Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 13 06:49:37 crc kubenswrapper[4971]: I1213 06:49:37.829130 4971 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"1ed9f82e-fe5c-4b76-a619-62e5abd88d3b\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T06:49:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T06:49:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T06:49:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T06:49:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T06:49:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e457329a78a2d24337f994e8a0a1f75a2a667244aa75eb84613d1e7138364aac\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-13T06:49:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://eb014de8140deef097d9bd94f3b9fa4ab3558bc4449bef84b83bf62bfabf5836\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-13T06:49:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://2c438cc74d555f7c9db4569d1c56b78b1088ed8105697d2c0f31e6c5136f3dd0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-13T06:49:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://aa6fa6e4340dcd066252df3c56d79a3a39ed531d5eea8d1d5941d6952368d3af\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-13T06:49:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://c4471b98854c52d5cd084d0ba3e64e5b08e9518ec3c372cce2c875a440522457\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-13T06:49:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7d2746335c98dae3a0f43024d34cb62ebcb42a01c8f9ce18d52b29d4c45598e5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://7d2746335c98dae3a0f43024d34cb62ebcb42a01c8f9ce18d52b29d4c45598e5\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-13T06:49:04Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-13T06:49:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://bc0e77d10a738128bea82122496bc42235fc1688f8907756d93e61b651902b57\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://bc0e77d10a738128bea82122496bc42235fc1688f8907756d93e61b651902b57\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-13T06:49:05Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-13T06:49:05Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://f821e418cba5e832da8839268bcbbd9630bdf319a212d5382bfa2c4b0b73a268\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://f821e418cba5e832da8839268bcbbd9630bdf319a212d5382bfa2c4b0b73a268\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-13T06:49:07Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-13T06:49:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-13T06:49:03Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-13T06:49:37Z is after 2025-08-24T17:21:41Z" Dec 13 06:49:37 crc kubenswrapper[4971]: I1213 06:49:37.847731 4971 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-13T06:49:24Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-13T06:49:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-13T06:49:24Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4328cd6a8464ed468229bbfbc8265f969427916d72ebb92237dbe68efe70fa47\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-13T06:49:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://9d715ca82d9b5a594e41f2b98da6f4277da0e4e1dec114969f901c00f6d7553d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-13T06:49:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-13T06:49:37Z is after 2025-08-24T17:21:41Z" Dec 13 06:49:37 crc kubenswrapper[4971]: I1213 06:49:37.865662 4971 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3ad3268a-a982-428c-9c06-4aaeddd372d9\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T06:49:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T06:49:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T06:49:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T06:49:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T06:49:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3716afdf2bc693be80a09fc9f7727d334f435caf1d4c8ec4811b3a1f1652cdfe\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-13T06:49:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://30df3abf8c79db8af52c6a918f73e1ce410ddafcf341636424c267e1eeb195a7\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-13T06:49:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://ccdf65f6d2b1eb3647a03ba37c305c32e34addfdf79ff937c8dddaa3e5d5b5c6\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-13T06:49:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://770b73dd0c2b6f38744e2c1cd6aa186a69bb87d1cd015a603ad2e33be06a8f6b\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://b3c1bab3d53f975b2e24582ce2771bc26d600e0dd7c30339dcfd18a071162ad4\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-12-13T06:49:23Z\\\",\\\"message\\\":\\\"file observer\\\\nW1213 06:49:23.283403 1 builder.go:272] unable to get owner reference (falling back to namespace): pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\nI1213 06:49:23.284336 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1213 06:49:23.286978 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-981773671/tls.crt::/tmp/serving-cert-981773671/tls.key\\\\\\\"\\\\nI1213 06:49:23.775072 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1213 06:49:23.778615 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1213 06:49:23.778645 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1213 06:49:23.778680 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1213 06:49:23.778691 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1213 06:49:23.785058 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1213 06:49:23.786150 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1213 06:49:23.786196 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1213 06:49:23.786223 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1213 06:49:23.786246 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1213 06:49:23.786269 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1213 06:49:23.786291 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI1213 06:49:23.785608 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF1213 06:49:23.787563 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-13T06:49:08Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-13T06:49:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://9acd07066deed483bd7fa61ee45fda09539d110717a134a56827f871186f92fc\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-13T06:49:07Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://38c91cfbaecfdbd3e3b663422a91ab4294610afb6bbf9ebdb77db24e2b76c682\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://38c91cfbaecfdbd3e3b663422a91ab4294610afb6bbf9ebdb77db24e2b76c682\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-13T06:49:05Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-13T06:49:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-13T06:49:03Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-13T06:49:37Z is after 2025-08-24T17:21:41Z" Dec 13 06:49:37 crc kubenswrapper[4971]: I1213 06:49:37.880146 4971 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-13T06:49:23Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-13T06:49:23Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-13T06:49:37Z is after 2025-08-24T17:21:41Z" Dec 13 06:49:37 crc kubenswrapper[4971]: I1213 06:49:37.895412 4971 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-13T06:49:23Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-13T06:49:23Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-13T06:49:37Z is after 2025-08-24T17:21:41Z" Dec 13 06:49:37 crc kubenswrapper[4971]: I1213 06:49:37.929094 4971 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 13 06:49:37 crc kubenswrapper[4971]: I1213 06:49:37.930482 4971 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 13 06:49:37 crc kubenswrapper[4971]: I1213 06:49:37.930496 4971 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 13 06:49:37 crc kubenswrapper[4971]: I1213 06:49:37.930510 4971 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 13 06:49:37 crc kubenswrapper[4971]: I1213 06:49:37.930530 4971 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-13T06:49:37Z","lastTransitionTime":"2025-12-13T06:49:37Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 13 06:49:38 crc kubenswrapper[4971]: I1213 06:49:38.028054 4971 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-zlcg5_736c6b1a-93dd-4f71-81b6-1c43fcf5f556/ovnkube-controller/1.log" Dec 13 06:49:38 crc kubenswrapper[4971]: I1213 06:49:38.028710 4971 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-zlcg5_736c6b1a-93dd-4f71-81b6-1c43fcf5f556/ovnkube-controller/0.log" Dec 13 06:49:38 crc kubenswrapper[4971]: I1213 06:49:38.031542 4971 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-zlcg5" event={"ID":"736c6b1a-93dd-4f71-81b6-1c43fcf5f556","Type":"ContainerDied","Data":"d950a29b0f6f47ae1cc90c132adb059b5aff889a0a38de1d1296e0f083da5e51"} Dec 13 06:49:38 crc kubenswrapper[4971]: I1213 06:49:38.031594 4971 scope.go:117] "RemoveContainer" containerID="451040d83e76225457d750a4aa532aa2772decb89719e53fe001bf74aaa8e73b" Dec 13 06:49:38 crc kubenswrapper[4971]: I1213 06:49:38.031509 4971 generic.go:334] "Generic (PLEG): container finished" podID="736c6b1a-93dd-4f71-81b6-1c43fcf5f556" containerID="d950a29b0f6f47ae1cc90c132adb059b5aff889a0a38de1d1296e0f083da5e51" exitCode=1 Dec 13 06:49:38 crc kubenswrapper[4971]: I1213 06:49:38.031671 4971 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 13 06:49:38 crc kubenswrapper[4971]: I1213 06:49:38.031687 4971 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 13 06:49:38 crc kubenswrapper[4971]: I1213 06:49:38.031695 4971 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 13 06:49:38 crc kubenswrapper[4971]: I1213 06:49:38.031706 4971 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 13 06:49:38 crc kubenswrapper[4971]: I1213 06:49:38.031715 4971 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-13T06:49:38Z","lastTransitionTime":"2025-12-13T06:49:38Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 13 06:49:38 crc kubenswrapper[4971]: I1213 06:49:38.032229 4971 scope.go:117] "RemoveContainer" containerID="d950a29b0f6f47ae1cc90c132adb059b5aff889a0a38de1d1296e0f083da5e51" Dec 13 06:49:38 crc kubenswrapper[4971]: E1213 06:49:38.032374 4971 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"ovnkube-controller\" with CrashLoopBackOff: \"back-off 10s restarting failed container=ovnkube-controller pod=ovnkube-node-zlcg5_openshift-ovn-kubernetes(736c6b1a-93dd-4f71-81b6-1c43fcf5f556)\"" pod="openshift-ovn-kubernetes/ovnkube-node-zlcg5" podUID="736c6b1a-93dd-4f71-81b6-1c43fcf5f556" Dec 13 06:49:38 crc kubenswrapper[4971]: I1213 06:49:38.046485 4971 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"f9d31b5a-b770-4a15-95c8-35b49677c608\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T06:49:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T06:49:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T06:49:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T06:49:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T06:49:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c9baa042f9dada954faa57b5454cd6b0a12a297e56ddc192c52cff50e42a16fa\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-13T06:49:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://c788de86754f0bf37ee184672fc1dcb388ccebc246d4f15be4b5da661f725f8f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-13T06:49:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://760817511a9b4f96a68dc38d26631250eaab72e1dd8f123e653039b6b4e91f84\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-13T06:49:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b859be0c9feac856e9855c6a038b02b49b32fade97873c46d0bec0a32d8b5147\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-13T06:49:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-13T06:49:03Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-13T06:49:38Z is after 2025-08-24T17:21:41Z" Dec 13 06:49:38 crc kubenswrapper[4971]: I1213 06:49:38.060295 4971 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-ncchz" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"de914d21-a688-401e-a66e-861beb6336d9\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T06:49:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T06:49:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T06:49:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T06:49:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://244ffe3dc9d8cf799aff4c4d795bf41554c1e6bb6667fbb0c7df322fb5f96c68\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-13T06:49:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lqmjj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://0b7111af6c5a198d02bb02858837cfb4f3b20828d8d206cd18aa0ffa102d0cd4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://0b7111af6c5a198d02bb02858837cfb4f3b20828d8d206cd18aa0ffa102d0cd4\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-13T06:49:26Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-13T06:49:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lqmjj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b36777fe40e74ae1cecb7b73e204139db5279ba51cef4f1e8c4a4c491d8be403\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://b36777fe40e74ae1cecb7b73e204139db5279ba51cef4f1e8c4a4c491d8be403\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-13T06:49:28Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-13T06:49:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lqmjj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://0597a06cc663e416142311a16941d98297577693a1fe27201ea42c27496cace2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://0597a06cc663e416142311a16941d98297577693a1fe27201ea42c27496cace2\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-13T06:49:29Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-13T06:49:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lqmjj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://3f52c2e22a75c010993ca5809a3a9d1304758d4cb6e83691ef1af15a4b514ac4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://3f52c2e22a75c010993ca5809a3a9d1304758d4cb6e83691ef1af15a4b514ac4\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-13T06:49:30Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-13T06:49:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lqmjj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://9e21ca708c4c4c1228b417cece51299e0792729b7c78884bd18b2dda8fe8355c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://9e21ca708c4c4c1228b417cece51299e0792729b7c78884bd18b2dda8fe8355c\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-13T06:49:31Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-13T06:49:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lqmjj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://74552159f12c87cdda09147692b400ccac0d4e43c1ed72fe8462c4249ad69c7b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://74552159f12c87cdda09147692b400ccac0d4e43c1ed72fe8462c4249ad69c7b\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-13T06:49:32Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-13T06:49:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lqmjj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-13T06:49:25Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-ncchz\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-13T06:49:38Z is after 2025-08-24T17:21:41Z" Dec 13 06:49:38 crc kubenswrapper[4971]: I1213 06:49:38.061354 4971 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-z572s"] Dec 13 06:49:38 crc kubenswrapper[4971]: I1213 06:49:38.061926 4971 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-z572s" Dec 13 06:49:38 crc kubenswrapper[4971]: I1213 06:49:38.064082 4971 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ovn-kubernetes"/"ovn-kubernetes-control-plane-dockercfg-gs7dd" Dec 13 06:49:38 crc kubenswrapper[4971]: I1213 06:49:38.064568 4971 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ovn-kubernetes"/"ovn-control-plane-metrics-cert" Dec 13 06:49:38 crc kubenswrapper[4971]: I1213 06:49:38.071241 4971 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-gkcd8" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"13f5c327-f95e-42b4-8783-218500c17694\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T06:49:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T06:49:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T06:49:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T06:49:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://77e846e36f1cb40d03471dc03de11b6c3da448baa407aad38093fae9a801cf92\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-13T06:49:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gvlhn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-13T06:49:27Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-gkcd8\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-13T06:49:38Z is after 2025-08-24T17:21:41Z" Dec 13 06:49:38 crc kubenswrapper[4971]: I1213 06:49:38.095326 4971 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"1ed9f82e-fe5c-4b76-a619-62e5abd88d3b\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T06:49:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T06:49:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T06:49:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T06:49:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T06:49:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e457329a78a2d24337f994e8a0a1f75a2a667244aa75eb84613d1e7138364aac\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-13T06:49:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://eb014de8140deef097d9bd94f3b9fa4ab3558bc4449bef84b83bf62bfabf5836\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-13T06:49:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://2c438cc74d555f7c9db4569d1c56b78b1088ed8105697d2c0f31e6c5136f3dd0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-13T06:49:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://aa6fa6e4340dcd066252df3c56d79a3a39ed531d5eea8d1d5941d6952368d3af\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-13T06:49:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://c4471b98854c52d5cd084d0ba3e64e5b08e9518ec3c372cce2c875a440522457\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-13T06:49:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7d2746335c98dae3a0f43024d34cb62ebcb42a01c8f9ce18d52b29d4c45598e5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://7d2746335c98dae3a0f43024d34cb62ebcb42a01c8f9ce18d52b29d4c45598e5\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-13T06:49:04Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-13T06:49:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://bc0e77d10a738128bea82122496bc42235fc1688f8907756d93e61b651902b57\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://bc0e77d10a738128bea82122496bc42235fc1688f8907756d93e61b651902b57\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-13T06:49:05Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-13T06:49:05Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://f821e418cba5e832da8839268bcbbd9630bdf319a212d5382bfa2c4b0b73a268\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://f821e418cba5e832da8839268bcbbd9630bdf319a212d5382bfa2c4b0b73a268\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-13T06:49:07Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-13T06:49:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-13T06:49:03Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-13T06:49:38Z is after 2025-08-24T17:21:41Z" Dec 13 06:49:38 crc kubenswrapper[4971]: I1213 06:49:38.109350 4971 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-13T06:49:24Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-13T06:49:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-13T06:49:24Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4328cd6a8464ed468229bbfbc8265f969427916d72ebb92237dbe68efe70fa47\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-13T06:49:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://9d715ca82d9b5a594e41f2b98da6f4277da0e4e1dec114969f901c00f6d7553d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-13T06:49:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-13T06:49:38Z is after 2025-08-24T17:21:41Z" Dec 13 06:49:38 crc kubenswrapper[4971]: I1213 06:49:38.120876 4971 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-5rzh5" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"6bf4591d-2483-4132-9672-d0fd6c995968\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T06:49:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T06:49:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T06:49:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T06:49:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://50e2cb14e4ffb603acb06e2281615a99f06182d345c4e70f2446fa2077033bbb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-13T06:49:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dfxqn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-13T06:49:25Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-5rzh5\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-13T06:49:38Z is after 2025-08-24T17:21:41Z" Dec 13 06:49:38 crc kubenswrapper[4971]: I1213 06:49:38.129785 4971 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/b4146d3d-4034-49a9-bd37-164856255137-ovnkube-config\") pod \"ovnkube-control-plane-749d76644c-z572s\" (UID: \"b4146d3d-4034-49a9-bd37-164856255137\") " pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-z572s" Dec 13 06:49:38 crc kubenswrapper[4971]: I1213 06:49:38.129850 4971 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovn-control-plane-metrics-cert\" (UniqueName: \"kubernetes.io/secret/b4146d3d-4034-49a9-bd37-164856255137-ovn-control-plane-metrics-cert\") pod \"ovnkube-control-plane-749d76644c-z572s\" (UID: \"b4146d3d-4034-49a9-bd37-164856255137\") " pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-z572s" Dec 13 06:49:38 crc kubenswrapper[4971]: I1213 06:49:38.129891 4971 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/b4146d3d-4034-49a9-bd37-164856255137-env-overrides\") pod \"ovnkube-control-plane-749d76644c-z572s\" (UID: \"b4146d3d-4034-49a9-bd37-164856255137\") " pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-z572s" Dec 13 06:49:38 crc kubenswrapper[4971]: I1213 06:49:38.129932 4971 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-j6l9b\" (UniqueName: \"kubernetes.io/projected/b4146d3d-4034-49a9-bd37-164856255137-kube-api-access-j6l9b\") pod \"ovnkube-control-plane-749d76644c-z572s\" (UID: \"b4146d3d-4034-49a9-bd37-164856255137\") " pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-z572s" Dec 13 06:49:38 crc kubenswrapper[4971]: I1213 06:49:38.135080 4971 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3ad3268a-a982-428c-9c06-4aaeddd372d9\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T06:49:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T06:49:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T06:49:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T06:49:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T06:49:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3716afdf2bc693be80a09fc9f7727d334f435caf1d4c8ec4811b3a1f1652cdfe\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-13T06:49:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://30df3abf8c79db8af52c6a918f73e1ce410ddafcf341636424c267e1eeb195a7\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-13T06:49:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://ccdf65f6d2b1eb3647a03ba37c305c32e34addfdf79ff937c8dddaa3e5d5b5c6\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-13T06:49:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://770b73dd0c2b6f38744e2c1cd6aa186a69bb87d1cd015a603ad2e33be06a8f6b\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://b3c1bab3d53f975b2e24582ce2771bc26d600e0dd7c30339dcfd18a071162ad4\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-12-13T06:49:23Z\\\",\\\"message\\\":\\\"file observer\\\\nW1213 06:49:23.283403 1 builder.go:272] unable to get owner reference (falling back to namespace): pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\nI1213 06:49:23.284336 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1213 06:49:23.286978 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-981773671/tls.crt::/tmp/serving-cert-981773671/tls.key\\\\\\\"\\\\nI1213 06:49:23.775072 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1213 06:49:23.778615 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1213 06:49:23.778645 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1213 06:49:23.778680 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1213 06:49:23.778691 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1213 06:49:23.785058 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1213 06:49:23.786150 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1213 06:49:23.786196 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1213 06:49:23.786223 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1213 06:49:23.786246 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1213 06:49:23.786269 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1213 06:49:23.786291 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI1213 06:49:23.785608 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF1213 06:49:23.787563 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-13T06:49:08Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-13T06:49:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://9acd07066deed483bd7fa61ee45fda09539d110717a134a56827f871186f92fc\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-13T06:49:07Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://38c91cfbaecfdbd3e3b663422a91ab4294610afb6bbf9ebdb77db24e2b76c682\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://38c91cfbaecfdbd3e3b663422a91ab4294610afb6bbf9ebdb77db24e2b76c682\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-13T06:49:05Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-13T06:49:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-13T06:49:03Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-13T06:49:38Z is after 2025-08-24T17:21:41Z" Dec 13 06:49:38 crc kubenswrapper[4971]: I1213 06:49:38.135267 4971 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 13 06:49:38 crc kubenswrapper[4971]: I1213 06:49:38.135289 4971 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 13 06:49:38 crc kubenswrapper[4971]: I1213 06:49:38.135300 4971 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 13 06:49:38 crc kubenswrapper[4971]: I1213 06:49:38.135316 4971 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 13 06:49:38 crc kubenswrapper[4971]: I1213 06:49:38.135337 4971 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-13T06:49:38Z","lastTransitionTime":"2025-12-13T06:49:38Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 13 06:49:38 crc kubenswrapper[4971]: I1213 06:49:38.149700 4971 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-13T06:49:23Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-13T06:49:23Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-13T06:49:38Z is after 2025-08-24T17:21:41Z" Dec 13 06:49:38 crc kubenswrapper[4971]: I1213 06:49:38.161351 4971 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-13T06:49:23Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-13T06:49:23Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-13T06:49:38Z is after 2025-08-24T17:21:41Z" Dec 13 06:49:38 crc kubenswrapper[4971]: I1213 06:49:38.183637 4971 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-13T06:49:24Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-13T06:49:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-13T06:49:24Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d0b2ecf0d5b22f23217f42f259ae7ffb6359555b22c30eaebf667eaf254b3bab\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-13T06:49:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-13T06:49:38Z is after 2025-08-24T17:21:41Z" Dec 13 06:49:38 crc kubenswrapper[4971]: I1213 06:49:38.197509 4971 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-13T06:49:23Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-13T06:49:23Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-13T06:49:38Z is after 2025-08-24T17:21:41Z" Dec 13 06:49:38 crc kubenswrapper[4971]: I1213 06:49:38.210594 4971 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-13T06:49:26Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-13T06:49:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-13T06:49:26Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9c8a0d1db0e754462d805e82daf5f52ffd5457507a30456535ec59e096c9d092\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-13T06:49:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-13T06:49:38Z is after 2025-08-24T17:21:41Z" Dec 13 06:49:38 crc kubenswrapper[4971]: I1213 06:49:38.222488 4971 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-82xjz" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e0c4af12-edda-41d7-9b44-f87396174ef2\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T06:49:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T06:49:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T06:49:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T06:49:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://0f4577d9e37f860813e221f2e208177d8d383fd04f673ce357a0a405ea29d1e6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-13T06:49:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cnxsq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://7e0d2020e62f1a277cff7047e1907dc747ed393a1f6c983e1ee258bd156c16c7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-13T06:49:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cnxsq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-13T06:49:25Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-82xjz\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-13T06:49:38Z is after 2025-08-24T17:21:41Z" Dec 13 06:49:38 crc kubenswrapper[4971]: I1213 06:49:38.231041 4971 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/b4146d3d-4034-49a9-bd37-164856255137-ovnkube-config\") pod \"ovnkube-control-plane-749d76644c-z572s\" (UID: \"b4146d3d-4034-49a9-bd37-164856255137\") " pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-z572s" Dec 13 06:49:38 crc kubenswrapper[4971]: I1213 06:49:38.231103 4971 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovn-control-plane-metrics-cert\" (UniqueName: \"kubernetes.io/secret/b4146d3d-4034-49a9-bd37-164856255137-ovn-control-plane-metrics-cert\") pod \"ovnkube-control-plane-749d76644c-z572s\" (UID: \"b4146d3d-4034-49a9-bd37-164856255137\") " pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-z572s" Dec 13 06:49:38 crc kubenswrapper[4971]: I1213 06:49:38.231133 4971 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/b4146d3d-4034-49a9-bd37-164856255137-env-overrides\") pod \"ovnkube-control-plane-749d76644c-z572s\" (UID: \"b4146d3d-4034-49a9-bd37-164856255137\") " pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-z572s" Dec 13 06:49:38 crc kubenswrapper[4971]: I1213 06:49:38.231161 4971 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-j6l9b\" (UniqueName: \"kubernetes.io/projected/b4146d3d-4034-49a9-bd37-164856255137-kube-api-access-j6l9b\") pod \"ovnkube-control-plane-749d76644c-z572s\" (UID: \"b4146d3d-4034-49a9-bd37-164856255137\") " pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-z572s" Dec 13 06:49:38 crc kubenswrapper[4971]: I1213 06:49:38.232120 4971 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/b4146d3d-4034-49a9-bd37-164856255137-ovnkube-config\") pod \"ovnkube-control-plane-749d76644c-z572s\" (UID: \"b4146d3d-4034-49a9-bd37-164856255137\") " pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-z572s" Dec 13 06:49:38 crc kubenswrapper[4971]: I1213 06:49:38.232827 4971 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/b4146d3d-4034-49a9-bd37-164856255137-env-overrides\") pod \"ovnkube-control-plane-749d76644c-z572s\" (UID: \"b4146d3d-4034-49a9-bd37-164856255137\") " pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-z572s" Dec 13 06:49:38 crc kubenswrapper[4971]: I1213 06:49:38.237638 4971 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-wwvpl" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"dddf744b-2122-49ad-a980-105c09636e0f\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T06:49:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T06:49:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T06:49:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T06:49:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4676d897c7fc668587faec563f3551a056557879cee3e865192d2bd28edf62fc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-13T06:49:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-czz9v\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-13T06:49:25Z\\\"}}\" for pod \"openshift-multus\"/\"multus-wwvpl\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-13T06:49:38Z is after 2025-08-24T17:21:41Z" Dec 13 06:49:38 crc kubenswrapper[4971]: I1213 06:49:38.238126 4971 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovn-control-plane-metrics-cert\" (UniqueName: \"kubernetes.io/secret/b4146d3d-4034-49a9-bd37-164856255137-ovn-control-plane-metrics-cert\") pod \"ovnkube-control-plane-749d76644c-z572s\" (UID: \"b4146d3d-4034-49a9-bd37-164856255137\") " pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-z572s" Dec 13 06:49:38 crc kubenswrapper[4971]: I1213 06:49:38.240836 4971 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 13 06:49:38 crc kubenswrapper[4971]: I1213 06:49:38.240895 4971 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 13 06:49:38 crc kubenswrapper[4971]: I1213 06:49:38.240912 4971 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 13 06:49:38 crc kubenswrapper[4971]: I1213 06:49:38.240938 4971 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 13 06:49:38 crc kubenswrapper[4971]: I1213 06:49:38.240955 4971 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-13T06:49:38Z","lastTransitionTime":"2025-12-13T06:49:38Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 13 06:49:38 crc kubenswrapper[4971]: I1213 06:49:38.249668 4971 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-j6l9b\" (UniqueName: \"kubernetes.io/projected/b4146d3d-4034-49a9-bd37-164856255137-kube-api-access-j6l9b\") pod \"ovnkube-control-plane-749d76644c-z572s\" (UID: \"b4146d3d-4034-49a9-bd37-164856255137\") " pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-z572s" Dec 13 06:49:38 crc kubenswrapper[4971]: I1213 06:49:38.260145 4971 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-zlcg5" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"736c6b1a-93dd-4f71-81b6-1c43fcf5f556\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T06:49:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T06:49:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T06:49:25Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T06:49:25Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://325223543b3c7db417f769533559a1ca2b23a723494e1885b0fd0480b5a6399f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-13T06:49:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vzmsr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://f22a4ec756119c4f791fb5105cf1f31df3e548b2b07b1f909a62802e9b6035ae\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-13T06:49:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vzmsr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://f6073140d9f86929b6d4bd2960ede0617a5c9620b269bbab043bae37e0d389f5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-13T06:49:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vzmsr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://693a2ab434c250cb140a866e8068b156df9191414fed1d63bee919a2528ea6e5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-13T06:49:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vzmsr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://219620f51474e60b3e1f880c0972eb612dffc4bad24462bbba2c093ae817d852\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-13T06:49:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vzmsr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://f137f4b441c2402093ab430a9689a46221c4fb4bc70d99fb992f62b0a989c465\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-13T06:49:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vzmsr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://d950a29b0f6f47ae1cc90c132adb059b5aff889a0a38de1d1296e0f083da5e51\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://451040d83e76225457d750a4aa532aa2772decb89719e53fe001bf74aaa8e73b\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-12-13T06:49:35Z\\\",\\\"message\\\":\\\"/pkg/client/informers/externalversions/factory.go:141\\\\nI1213 06:49:35.122568 6277 handler.go:190] Sending *v1.Namespace event handler 1 for removal\\\\nI1213 06:49:35.122615 6277 handler.go:190] Sending *v1.Namespace event handler 5 for removal\\\\nI1213 06:49:35.122659 6277 handler.go:190] Sending *v1.Node event handler 7 for removal\\\\nI1213 06:49:35.122681 6277 handler.go:190] Sending *v1.Node event handler 2 for removal\\\\nI1213 06:49:35.122723 6277 handler.go:190] Sending *v1.NetworkPolicy event handler 4 for removal\\\\nI1213 06:49:35.122771 6277 handler.go:190] Sending *v1.EgressIP event handler 8 for removal\\\\nI1213 06:49:35.122815 6277 handler.go:190] Sending *v1.EgressFirewall event handler 9 for removal\\\\nI1213 06:49:35.122852 6277 factory.go:656] Stopping watch factory\\\\nI1213 06:49:35.122935 6277 ovnkube.go:599] Stopped ovnkube\\\\nI1213 06:49:35.122983 6277 handler.go:208] Removed *v1.Namespace event handler 1\\\\nI1213 06:49:35.123009 6277 handler.go:208] Removed *v1.Namespace event handler 5\\\\nI1213 06:49:35.123033 6277 handler.go:208] Removed *v1.Node event handler 7\\\\nI1213 06:49:35.123056 6277 handler.go:208] Removed *v1.Node event handler 2\\\\nI1213 06:49:35.123079 6277 handler.go:208] Removed *v1.NetworkPolicy event handler 4\\\\nI1213 06:49:35.123102 6277 handler.go:208] Removed *v1.EgressIP event handler 8\\\\nI1213 06:49:3\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-13T06:49:32Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d950a29b0f6f47ae1cc90c132adb059b5aff889a0a38de1d1296e0f083da5e51\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-12-13T06:49:37Z\\\",\\\"message\\\":\\\" ovnkube.go:137] failed to run ovnkube: [failed to start network controller: failed to start default network controller: unable to create admin network policy controller, err: could not add Event Handler for anpInformer during admin network policy controller initialization, handler {0x1fcc6e0 0x1fcc3c0 0x1fcc360} was not added to shared informer because it has stopped already, failed to start node network controller: failed to start default node network controller: failed to set node crc annotations: Internal error occurred: failed calling webhook \\\\\\\"node.network-node-identity.openshift.io\\\\\\\": failed to call webhook: Post \\\\\\\"https://127.0.0.1:9743/node?timeout=10s\\\\\\\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-13T06:49:36Z is after 2025-08-24T17:21:41Z]\\\\nI1213 06:49:36.767920 6409 model_client.go:382] Update operations generated as: [{Op:update Table:Logical_Switch_Port Row:map[addresses:{GoSet:[0a:58:0a:d9:00:04 10.217.0.4]} options:{GoMap:map[iface-id-ver:3b6479f0-333b-4a96-9adf-2099afdc2447 requested-chassis:crc]} port_security:{GoSet:[0a:58:0a:d9:00:04 10.217.0.4]}] Rows:[] Columns:[] Mutations:[] Timeout:\\\\u003cnil\\\\u003e Where:[where column _uuid == {61897e97-c771-4738-8709-09636387cb00}] Until: Durable:\\\\u003cnil\\\\u003e Comment:\\\\u003cnil\\\\u003e Lock:\\\\u003cnil\\\\u003e UUID: UUIDName:}]\\\\nI1213 06:49:36.767917 6409 model_clien\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-13T06:49:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vzmsr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://868049c6a3a4ebe4036d1d22e02ba6b76bced2ee7e269bb228accc4a0636678a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-13T06:49:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vzmsr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://451b22df6973ddc8fa2209a65d64d9e41b33eb231bb4adfab88cf0e5943bb246\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://451b22df6973ddc8fa2209a65d64d9e41b33eb231bb4adfab88cf0e5943bb246\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-13T06:49:26Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-13T06:49:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vzmsr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-13T06:49:25Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-zlcg5\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-13T06:49:38Z is after 2025-08-24T17:21:41Z" Dec 13 06:49:38 crc kubenswrapper[4971]: I1213 06:49:38.273232 4971 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3ad3268a-a982-428c-9c06-4aaeddd372d9\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T06:49:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T06:49:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T06:49:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T06:49:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T06:49:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3716afdf2bc693be80a09fc9f7727d334f435caf1d4c8ec4811b3a1f1652cdfe\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-13T06:49:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://30df3abf8c79db8af52c6a918f73e1ce410ddafcf341636424c267e1eeb195a7\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-13T06:49:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://ccdf65f6d2b1eb3647a03ba37c305c32e34addfdf79ff937c8dddaa3e5d5b5c6\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-13T06:49:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://770b73dd0c2b6f38744e2c1cd6aa186a69bb87d1cd015a603ad2e33be06a8f6b\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://b3c1bab3d53f975b2e24582ce2771bc26d600e0dd7c30339dcfd18a071162ad4\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-12-13T06:49:23Z\\\",\\\"message\\\":\\\"file observer\\\\nW1213 06:49:23.283403 1 builder.go:272] unable to get owner reference (falling back to namespace): pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\nI1213 06:49:23.284336 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1213 06:49:23.286978 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-981773671/tls.crt::/tmp/serving-cert-981773671/tls.key\\\\\\\"\\\\nI1213 06:49:23.775072 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1213 06:49:23.778615 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1213 06:49:23.778645 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1213 06:49:23.778680 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1213 06:49:23.778691 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1213 06:49:23.785058 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1213 06:49:23.786150 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1213 06:49:23.786196 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1213 06:49:23.786223 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1213 06:49:23.786246 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1213 06:49:23.786269 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1213 06:49:23.786291 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI1213 06:49:23.785608 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF1213 06:49:23.787563 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-13T06:49:08Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-13T06:49:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://9acd07066deed483bd7fa61ee45fda09539d110717a134a56827f871186f92fc\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-13T06:49:07Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://38c91cfbaecfdbd3e3b663422a91ab4294610afb6bbf9ebdb77db24e2b76c682\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://38c91cfbaecfdbd3e3b663422a91ab4294610afb6bbf9ebdb77db24e2b76c682\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-13T06:49:05Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-13T06:49:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-13T06:49:03Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-13T06:49:38Z is after 2025-08-24T17:21:41Z" Dec 13 06:49:38 crc kubenswrapper[4971]: I1213 06:49:38.285783 4971 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-13T06:49:23Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-13T06:49:23Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-13T06:49:38Z is after 2025-08-24T17:21:41Z" Dec 13 06:49:38 crc kubenswrapper[4971]: I1213 06:49:38.296913 4971 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-13T06:49:23Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-13T06:49:23Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-13T06:49:38Z is after 2025-08-24T17:21:41Z" Dec 13 06:49:38 crc kubenswrapper[4971]: I1213 06:49:38.308931 4971 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-13T06:49:24Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-13T06:49:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-13T06:49:24Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d0b2ecf0d5b22f23217f42f259ae7ffb6359555b22c30eaebf667eaf254b3bab\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-13T06:49:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-13T06:49:38Z is after 2025-08-24T17:21:41Z" Dec 13 06:49:38 crc kubenswrapper[4971]: I1213 06:49:38.320195 4971 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-13T06:49:23Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-13T06:49:23Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-13T06:49:38Z is after 2025-08-24T17:21:41Z" Dec 13 06:49:38 crc kubenswrapper[4971]: I1213 06:49:38.330927 4971 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-13T06:49:26Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-13T06:49:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-13T06:49:26Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9c8a0d1db0e754462d805e82daf5f52ffd5457507a30456535ec59e096c9d092\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-13T06:49:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-13T06:49:38Z is after 2025-08-24T17:21:41Z" Dec 13 06:49:38 crc kubenswrapper[4971]: I1213 06:49:38.340134 4971 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-82xjz" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e0c4af12-edda-41d7-9b44-f87396174ef2\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T06:49:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T06:49:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T06:49:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T06:49:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://0f4577d9e37f860813e221f2e208177d8d383fd04f673ce357a0a405ea29d1e6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-13T06:49:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cnxsq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://7e0d2020e62f1a277cff7047e1907dc747ed393a1f6c983e1ee258bd156c16c7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-13T06:49:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cnxsq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-13T06:49:25Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-82xjz\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-13T06:49:38Z is after 2025-08-24T17:21:41Z" Dec 13 06:49:38 crc kubenswrapper[4971]: I1213 06:49:38.342714 4971 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 13 06:49:38 crc kubenswrapper[4971]: I1213 06:49:38.342761 4971 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 13 06:49:38 crc kubenswrapper[4971]: I1213 06:49:38.342774 4971 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 13 06:49:38 crc kubenswrapper[4971]: I1213 06:49:38.342792 4971 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 13 06:49:38 crc kubenswrapper[4971]: I1213 06:49:38.342804 4971 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-13T06:49:38Z","lastTransitionTime":"2025-12-13T06:49:38Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 13 06:49:38 crc kubenswrapper[4971]: I1213 06:49:38.353418 4971 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-wwvpl" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"dddf744b-2122-49ad-a980-105c09636e0f\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T06:49:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T06:49:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T06:49:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T06:49:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4676d897c7fc668587faec563f3551a056557879cee3e865192d2bd28edf62fc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-13T06:49:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-czz9v\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-13T06:49:25Z\\\"}}\" for pod \"openshift-multus\"/\"multus-wwvpl\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-13T06:49:38Z is after 2025-08-24T17:21:41Z" Dec 13 06:49:38 crc kubenswrapper[4971]: I1213 06:49:38.373753 4971 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-zlcg5" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"736c6b1a-93dd-4f71-81b6-1c43fcf5f556\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T06:49:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T06:49:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T06:49:25Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T06:49:25Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://325223543b3c7db417f769533559a1ca2b23a723494e1885b0fd0480b5a6399f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-13T06:49:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vzmsr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://f22a4ec756119c4f791fb5105cf1f31df3e548b2b07b1f909a62802e9b6035ae\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-13T06:49:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vzmsr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://f6073140d9f86929b6d4bd2960ede0617a5c9620b269bbab043bae37e0d389f5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-13T06:49:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vzmsr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://693a2ab434c250cb140a866e8068b156df9191414fed1d63bee919a2528ea6e5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-13T06:49:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vzmsr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://219620f51474e60b3e1f880c0972eb612dffc4bad24462bbba2c093ae817d852\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-13T06:49:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vzmsr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://f137f4b441c2402093ab430a9689a46221c4fb4bc70d99fb992f62b0a989c465\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-13T06:49:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vzmsr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://d950a29b0f6f47ae1cc90c132adb059b5aff889a0a38de1d1296e0f083da5e51\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://451040d83e76225457d750a4aa532aa2772decb89719e53fe001bf74aaa8e73b\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-12-13T06:49:35Z\\\",\\\"message\\\":\\\"/pkg/client/informers/externalversions/factory.go:141\\\\nI1213 06:49:35.122568 6277 handler.go:190] Sending *v1.Namespace event handler 1 for removal\\\\nI1213 06:49:35.122615 6277 handler.go:190] Sending *v1.Namespace event handler 5 for removal\\\\nI1213 06:49:35.122659 6277 handler.go:190] Sending *v1.Node event handler 7 for removal\\\\nI1213 06:49:35.122681 6277 handler.go:190] Sending *v1.Node event handler 2 for removal\\\\nI1213 06:49:35.122723 6277 handler.go:190] Sending *v1.NetworkPolicy event handler 4 for removal\\\\nI1213 06:49:35.122771 6277 handler.go:190] Sending *v1.EgressIP event handler 8 for removal\\\\nI1213 06:49:35.122815 6277 handler.go:190] Sending *v1.EgressFirewall event handler 9 for removal\\\\nI1213 06:49:35.122852 6277 factory.go:656] Stopping watch factory\\\\nI1213 06:49:35.122935 6277 ovnkube.go:599] Stopped ovnkube\\\\nI1213 06:49:35.122983 6277 handler.go:208] Removed *v1.Namespace event handler 1\\\\nI1213 06:49:35.123009 6277 handler.go:208] Removed *v1.Namespace event handler 5\\\\nI1213 06:49:35.123033 6277 handler.go:208] Removed *v1.Node event handler 7\\\\nI1213 06:49:35.123056 6277 handler.go:208] Removed *v1.Node event handler 2\\\\nI1213 06:49:35.123079 6277 handler.go:208] Removed *v1.NetworkPolicy event handler 4\\\\nI1213 06:49:35.123102 6277 handler.go:208] Removed *v1.EgressIP event handler 8\\\\nI1213 06:49:3\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-13T06:49:32Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d950a29b0f6f47ae1cc90c132adb059b5aff889a0a38de1d1296e0f083da5e51\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-12-13T06:49:37Z\\\",\\\"message\\\":\\\" ovnkube.go:137] failed to run ovnkube: [failed to start network controller: failed to start default network controller: unable to create admin network policy controller, err: could not add Event Handler for anpInformer during admin network policy controller initialization, handler {0x1fcc6e0 0x1fcc3c0 0x1fcc360} was not added to shared informer because it has stopped already, failed to start node network controller: failed to start default node network controller: failed to set node crc annotations: Internal error occurred: failed calling webhook \\\\\\\"node.network-node-identity.openshift.io\\\\\\\": failed to call webhook: Post \\\\\\\"https://127.0.0.1:9743/node?timeout=10s\\\\\\\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-13T06:49:36Z is after 2025-08-24T17:21:41Z]\\\\nI1213 06:49:36.767920 6409 model_client.go:382] Update operations generated as: [{Op:update Table:Logical_Switch_Port Row:map[addresses:{GoSet:[0a:58:0a:d9:00:04 10.217.0.4]} options:{GoMap:map[iface-id-ver:3b6479f0-333b-4a96-9adf-2099afdc2447 requested-chassis:crc]} port_security:{GoSet:[0a:58:0a:d9:00:04 10.217.0.4]}] Rows:[] Columns:[] Mutations:[] Timeout:\\\\u003cnil\\\\u003e Where:[where column _uuid == {61897e97-c771-4738-8709-09636387cb00}] Until: Durable:\\\\u003cnil\\\\u003e Comment:\\\\u003cnil\\\\u003e Lock:\\\\u003cnil\\\\u003e UUID: UUIDName:}]\\\\nI1213 06:49:36.767917 6409 model_clien\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-13T06:49:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vzmsr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://868049c6a3a4ebe4036d1d22e02ba6b76bced2ee7e269bb228accc4a0636678a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-13T06:49:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vzmsr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://451b22df6973ddc8fa2209a65d64d9e41b33eb231bb4adfab88cf0e5943bb246\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://451b22df6973ddc8fa2209a65d64d9e41b33eb231bb4adfab88cf0e5943bb246\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-13T06:49:26Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-13T06:49:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vzmsr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-13T06:49:25Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-zlcg5\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-13T06:49:38Z is after 2025-08-24T17:21:41Z" Dec 13 06:49:38 crc kubenswrapper[4971]: I1213 06:49:38.374682 4971 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-z572s" Dec 13 06:49:38 crc kubenswrapper[4971]: I1213 06:49:38.388802 4971 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"f9d31b5a-b770-4a15-95c8-35b49677c608\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T06:49:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T06:49:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T06:49:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T06:49:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T06:49:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c9baa042f9dada954faa57b5454cd6b0a12a297e56ddc192c52cff50e42a16fa\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-13T06:49:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://c788de86754f0bf37ee184672fc1dcb388ccebc246d4f15be4b5da661f725f8f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-13T06:49:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://760817511a9b4f96a68dc38d26631250eaab72e1dd8f123e653039b6b4e91f84\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-13T06:49:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b859be0c9feac856e9855c6a038b02b49b32fade97873c46d0bec0a32d8b5147\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-13T06:49:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-13T06:49:03Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-13T06:49:38Z is after 2025-08-24T17:21:41Z" Dec 13 06:49:38 crc kubenswrapper[4971]: I1213 06:49:38.403144 4971 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-ncchz" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"de914d21-a688-401e-a66e-861beb6336d9\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T06:49:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T06:49:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T06:49:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T06:49:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://244ffe3dc9d8cf799aff4c4d795bf41554c1e6bb6667fbb0c7df322fb5f96c68\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-13T06:49:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lqmjj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://0b7111af6c5a198d02bb02858837cfb4f3b20828d8d206cd18aa0ffa102d0cd4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://0b7111af6c5a198d02bb02858837cfb4f3b20828d8d206cd18aa0ffa102d0cd4\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-13T06:49:26Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-13T06:49:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lqmjj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b36777fe40e74ae1cecb7b73e204139db5279ba51cef4f1e8c4a4c491d8be403\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://b36777fe40e74ae1cecb7b73e204139db5279ba51cef4f1e8c4a4c491d8be403\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-13T06:49:28Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-13T06:49:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lqmjj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://0597a06cc663e416142311a16941d98297577693a1fe27201ea42c27496cace2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://0597a06cc663e416142311a16941d98297577693a1fe27201ea42c27496cace2\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-13T06:49:29Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-13T06:49:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lqmjj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://3f52c2e22a75c010993ca5809a3a9d1304758d4cb6e83691ef1af15a4b514ac4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://3f52c2e22a75c010993ca5809a3a9d1304758d4cb6e83691ef1af15a4b514ac4\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-13T06:49:30Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-13T06:49:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lqmjj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://9e21ca708c4c4c1228b417cece51299e0792729b7c78884bd18b2dda8fe8355c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://9e21ca708c4c4c1228b417cece51299e0792729b7c78884bd18b2dda8fe8355c\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-13T06:49:31Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-13T06:49:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lqmjj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://74552159f12c87cdda09147692b400ccac0d4e43c1ed72fe8462c4249ad69c7b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://74552159f12c87cdda09147692b400ccac0d4e43c1ed72fe8462c4249ad69c7b\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-13T06:49:32Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-13T06:49:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lqmjj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-13T06:49:25Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-ncchz\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-13T06:49:38Z is after 2025-08-24T17:21:41Z" Dec 13 06:49:38 crc kubenswrapper[4971]: I1213 06:49:38.414941 4971 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-gkcd8" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"13f5c327-f95e-42b4-8783-218500c17694\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T06:49:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T06:49:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T06:49:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T06:49:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://77e846e36f1cb40d03471dc03de11b6c3da448baa407aad38093fae9a801cf92\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-13T06:49:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gvlhn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-13T06:49:27Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-gkcd8\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-13T06:49:38Z is after 2025-08-24T17:21:41Z" Dec 13 06:49:38 crc kubenswrapper[4971]: I1213 06:49:38.435630 4971 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"1ed9f82e-fe5c-4b76-a619-62e5abd88d3b\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T06:49:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T06:49:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T06:49:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T06:49:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T06:49:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e457329a78a2d24337f994e8a0a1f75a2a667244aa75eb84613d1e7138364aac\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-13T06:49:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://eb014de8140deef097d9bd94f3b9fa4ab3558bc4449bef84b83bf62bfabf5836\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-13T06:49:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://2c438cc74d555f7c9db4569d1c56b78b1088ed8105697d2c0f31e6c5136f3dd0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-13T06:49:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://aa6fa6e4340dcd066252df3c56d79a3a39ed531d5eea8d1d5941d6952368d3af\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-13T06:49:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://c4471b98854c52d5cd084d0ba3e64e5b08e9518ec3c372cce2c875a440522457\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-13T06:49:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7d2746335c98dae3a0f43024d34cb62ebcb42a01c8f9ce18d52b29d4c45598e5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://7d2746335c98dae3a0f43024d34cb62ebcb42a01c8f9ce18d52b29d4c45598e5\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-13T06:49:04Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-13T06:49:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://bc0e77d10a738128bea82122496bc42235fc1688f8907756d93e61b651902b57\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://bc0e77d10a738128bea82122496bc42235fc1688f8907756d93e61b651902b57\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-13T06:49:05Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-13T06:49:05Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://f821e418cba5e832da8839268bcbbd9630bdf319a212d5382bfa2c4b0b73a268\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://f821e418cba5e832da8839268bcbbd9630bdf319a212d5382bfa2c4b0b73a268\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-13T06:49:07Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-13T06:49:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-13T06:49:03Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-13T06:49:38Z is after 2025-08-24T17:21:41Z" Dec 13 06:49:38 crc kubenswrapper[4971]: I1213 06:49:38.444997 4971 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 13 06:49:38 crc kubenswrapper[4971]: I1213 06:49:38.445038 4971 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 13 06:49:38 crc kubenswrapper[4971]: I1213 06:49:38.445050 4971 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 13 06:49:38 crc kubenswrapper[4971]: I1213 06:49:38.445070 4971 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 13 06:49:38 crc kubenswrapper[4971]: I1213 06:49:38.445084 4971 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-13T06:49:38Z","lastTransitionTime":"2025-12-13T06:49:38Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 13 06:49:38 crc kubenswrapper[4971]: I1213 06:49:38.446590 4971 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-13T06:49:24Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-13T06:49:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-13T06:49:24Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4328cd6a8464ed468229bbfbc8265f969427916d72ebb92237dbe68efe70fa47\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-13T06:49:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://9d715ca82d9b5a594e41f2b98da6f4277da0e4e1dec114969f901c00f6d7553d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-13T06:49:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-13T06:49:38Z is after 2025-08-24T17:21:41Z" Dec 13 06:49:38 crc kubenswrapper[4971]: I1213 06:49:38.456306 4971 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-5rzh5" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"6bf4591d-2483-4132-9672-d0fd6c995968\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T06:49:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T06:49:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T06:49:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T06:49:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://50e2cb14e4ffb603acb06e2281615a99f06182d345c4e70f2446fa2077033bbb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-13T06:49:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dfxqn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-13T06:49:25Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-5rzh5\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-13T06:49:38Z is after 2025-08-24T17:21:41Z" Dec 13 06:49:38 crc kubenswrapper[4971]: I1213 06:49:38.466593 4971 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-z572s" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"b4146d3d-4034-49a9-bd37-164856255137\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T06:49:38Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T06:49:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T06:49:38Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-rbac-proxy ovnkube-cluster-manager]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T06:49:38Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-rbac-proxy ovnkube-cluster-manager]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-j6l9b\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-j6l9b\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-13T06:49:38Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-z572s\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-13T06:49:38Z is after 2025-08-24T17:21:41Z" Dec 13 06:49:38 crc kubenswrapper[4971]: I1213 06:49:38.547194 4971 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 13 06:49:38 crc kubenswrapper[4971]: I1213 06:49:38.547236 4971 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 13 06:49:38 crc kubenswrapper[4971]: I1213 06:49:38.547251 4971 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 13 06:49:38 crc kubenswrapper[4971]: I1213 06:49:38.547268 4971 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 13 06:49:38 crc kubenswrapper[4971]: I1213 06:49:38.547280 4971 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-13T06:49:38Z","lastTransitionTime":"2025-12-13T06:49:38Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 13 06:49:38 crc kubenswrapper[4971]: I1213 06:49:38.649739 4971 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 13 06:49:38 crc kubenswrapper[4971]: I1213 06:49:38.649783 4971 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 13 06:49:38 crc kubenswrapper[4971]: I1213 06:49:38.649795 4971 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 13 06:49:38 crc kubenswrapper[4971]: I1213 06:49:38.649812 4971 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 13 06:49:38 crc kubenswrapper[4971]: I1213 06:49:38.649825 4971 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-13T06:49:38Z","lastTransitionTime":"2025-12-13T06:49:38Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 13 06:49:38 crc kubenswrapper[4971]: I1213 06:49:38.752380 4971 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 13 06:49:38 crc kubenswrapper[4971]: I1213 06:49:38.752416 4971 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 13 06:49:38 crc kubenswrapper[4971]: I1213 06:49:38.752425 4971 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 13 06:49:38 crc kubenswrapper[4971]: I1213 06:49:38.752440 4971 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 13 06:49:38 crc kubenswrapper[4971]: I1213 06:49:38.752450 4971 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-13T06:49:38Z","lastTransitionTime":"2025-12-13T06:49:38Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 13 06:49:38 crc kubenswrapper[4971]: I1213 06:49:38.858330 4971 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 13 06:49:38 crc kubenswrapper[4971]: I1213 06:49:38.858369 4971 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 13 06:49:38 crc kubenswrapper[4971]: I1213 06:49:38.858380 4971 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 13 06:49:38 crc kubenswrapper[4971]: I1213 06:49:38.858393 4971 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 13 06:49:38 crc kubenswrapper[4971]: I1213 06:49:38.858403 4971 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-13T06:49:38Z","lastTransitionTime":"2025-12-13T06:49:38Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 13 06:49:38 crc kubenswrapper[4971]: I1213 06:49:38.960925 4971 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 13 06:49:38 crc kubenswrapper[4971]: I1213 06:49:38.960978 4971 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 13 06:49:38 crc kubenswrapper[4971]: I1213 06:49:38.960991 4971 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 13 06:49:38 crc kubenswrapper[4971]: I1213 06:49:38.961010 4971 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 13 06:49:38 crc kubenswrapper[4971]: I1213 06:49:38.961022 4971 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-13T06:49:38Z","lastTransitionTime":"2025-12-13T06:49:38Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 13 06:49:39 crc kubenswrapper[4971]: I1213 06:49:39.035692 4971 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-z572s" event={"ID":"b4146d3d-4034-49a9-bd37-164856255137","Type":"ContainerStarted","Data":"e491a2d7bac896a2b00d3e58988fa915ee5dbb58b4e6327fbbdf7a575c18393d"} Dec 13 06:49:39 crc kubenswrapper[4971]: I1213 06:49:39.064996 4971 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 13 06:49:39 crc kubenswrapper[4971]: I1213 06:49:39.065047 4971 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 13 06:49:39 crc kubenswrapper[4971]: I1213 06:49:39.065063 4971 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 13 06:49:39 crc kubenswrapper[4971]: I1213 06:49:39.065082 4971 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 13 06:49:39 crc kubenswrapper[4971]: I1213 06:49:39.065097 4971 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-13T06:49:39Z","lastTransitionTime":"2025-12-13T06:49:39Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 13 06:49:39 crc kubenswrapper[4971]: I1213 06:49:39.169625 4971 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 13 06:49:39 crc kubenswrapper[4971]: I1213 06:49:39.169655 4971 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 13 06:49:39 crc kubenswrapper[4971]: I1213 06:49:39.169665 4971 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 13 06:49:39 crc kubenswrapper[4971]: I1213 06:49:39.169678 4971 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 13 06:49:39 crc kubenswrapper[4971]: I1213 06:49:39.169687 4971 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-13T06:49:39Z","lastTransitionTime":"2025-12-13T06:49:39Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 13 06:49:39 crc kubenswrapper[4971]: I1213 06:49:39.272407 4971 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 13 06:49:39 crc kubenswrapper[4971]: I1213 06:49:39.272457 4971 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 13 06:49:39 crc kubenswrapper[4971]: I1213 06:49:39.272471 4971 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 13 06:49:39 crc kubenswrapper[4971]: I1213 06:49:39.272490 4971 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 13 06:49:39 crc kubenswrapper[4971]: I1213 06:49:39.272502 4971 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-13T06:49:39Z","lastTransitionTime":"2025-12-13T06:49:39Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 13 06:49:39 crc kubenswrapper[4971]: I1213 06:49:39.375599 4971 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 13 06:49:39 crc kubenswrapper[4971]: I1213 06:49:39.375656 4971 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 13 06:49:39 crc kubenswrapper[4971]: I1213 06:49:39.375673 4971 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 13 06:49:39 crc kubenswrapper[4971]: I1213 06:49:39.375690 4971 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 13 06:49:39 crc kubenswrapper[4971]: I1213 06:49:39.375701 4971 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-13T06:49:39Z","lastTransitionTime":"2025-12-13T06:49:39Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 13 06:49:39 crc kubenswrapper[4971]: I1213 06:49:39.444171 4971 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 13 06:49:39 crc kubenswrapper[4971]: E1213 06:49:39.444265 4971 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-13 06:49:55.444246907 +0000 UTC m=+52.048656355 (durationBeforeRetry 16s). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 13 06:49:39 crc kubenswrapper[4971]: I1213 06:49:39.444298 4971 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"networking-console-plugin-cert\" (UniqueName: \"kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 13 06:49:39 crc kubenswrapper[4971]: I1213 06:49:39.444325 4971 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nginx-conf\" (UniqueName: \"kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 13 06:49:39 crc kubenswrapper[4971]: E1213 06:49:39.444441 4971 secret.go:188] Couldn't get secret openshift-network-console/networking-console-plugin-cert: object "openshift-network-console"/"networking-console-plugin-cert" not registered Dec 13 06:49:39 crc kubenswrapper[4971]: E1213 06:49:39.444441 4971 configmap.go:193] Couldn't get configMap openshift-network-console/networking-console-plugin: object "openshift-network-console"/"networking-console-plugin" not registered Dec 13 06:49:39 crc kubenswrapper[4971]: E1213 06:49:39.444478 4971 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2025-12-13 06:49:55.444471753 +0000 UTC m=+52.048881201 (durationBeforeRetry 16s). Error: MountVolume.SetUp failed for volume "nginx-conf" (UniqueName: "kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin" not registered Dec 13 06:49:39 crc kubenswrapper[4971]: E1213 06:49:39.444489 4971 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2025-12-13 06:49:55.444483774 +0000 UTC m=+52.048893212 (durationBeforeRetry 16s). Error: MountVolume.SetUp failed for volume "networking-console-plugin-cert" (UniqueName: "kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin-cert" not registered Dec 13 06:49:39 crc kubenswrapper[4971]: I1213 06:49:39.478956 4971 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 13 06:49:39 crc kubenswrapper[4971]: I1213 06:49:39.479004 4971 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 13 06:49:39 crc kubenswrapper[4971]: I1213 06:49:39.479019 4971 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 13 06:49:39 crc kubenswrapper[4971]: I1213 06:49:39.479037 4971 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 13 06:49:39 crc kubenswrapper[4971]: I1213 06:49:39.479048 4971 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-13T06:49:39Z","lastTransitionTime":"2025-12-13T06:49:39Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 13 06:49:39 crc kubenswrapper[4971]: I1213 06:49:39.488988 4971 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-multus/network-metrics-daemon-x98k7"] Dec 13 06:49:39 crc kubenswrapper[4971]: I1213 06:49:39.489610 4971 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-x98k7" Dec 13 06:49:39 crc kubenswrapper[4971]: E1213 06:49:39.489686 4971 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-x98k7" podUID="c7185dec-3240-42e8-9d90-6182a5e8b65c" Dec 13 06:49:39 crc kubenswrapper[4971]: I1213 06:49:39.503090 4971 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"f9d31b5a-b770-4a15-95c8-35b49677c608\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T06:49:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T06:49:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T06:49:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T06:49:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T06:49:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c9baa042f9dada954faa57b5454cd6b0a12a297e56ddc192c52cff50e42a16fa\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-13T06:49:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://c788de86754f0bf37ee184672fc1dcb388ccebc246d4f15be4b5da661f725f8f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-13T06:49:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://760817511a9b4f96a68dc38d26631250eaab72e1dd8f123e653039b6b4e91f84\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-13T06:49:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b859be0c9feac856e9855c6a038b02b49b32fade97873c46d0bec0a32d8b5147\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-13T06:49:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-13T06:49:03Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-13T06:49:39Z is after 2025-08-24T17:21:41Z" Dec 13 06:49:39 crc kubenswrapper[4971]: I1213 06:49:39.517832 4971 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-ncchz" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"de914d21-a688-401e-a66e-861beb6336d9\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T06:49:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T06:49:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T06:49:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T06:49:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://244ffe3dc9d8cf799aff4c4d795bf41554c1e6bb6667fbb0c7df322fb5f96c68\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-13T06:49:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lqmjj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://0b7111af6c5a198d02bb02858837cfb4f3b20828d8d206cd18aa0ffa102d0cd4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://0b7111af6c5a198d02bb02858837cfb4f3b20828d8d206cd18aa0ffa102d0cd4\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-13T06:49:26Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-13T06:49:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lqmjj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b36777fe40e74ae1cecb7b73e204139db5279ba51cef4f1e8c4a4c491d8be403\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://b36777fe40e74ae1cecb7b73e204139db5279ba51cef4f1e8c4a4c491d8be403\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-13T06:49:28Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-13T06:49:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lqmjj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://0597a06cc663e416142311a16941d98297577693a1fe27201ea42c27496cace2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://0597a06cc663e416142311a16941d98297577693a1fe27201ea42c27496cace2\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-13T06:49:29Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-13T06:49:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lqmjj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://3f52c2e22a75c010993ca5809a3a9d1304758d4cb6e83691ef1af15a4b514ac4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://3f52c2e22a75c010993ca5809a3a9d1304758d4cb6e83691ef1af15a4b514ac4\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-13T06:49:30Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-13T06:49:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lqmjj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://9e21ca708c4c4c1228b417cece51299e0792729b7c78884bd18b2dda8fe8355c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://9e21ca708c4c4c1228b417cece51299e0792729b7c78884bd18b2dda8fe8355c\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-13T06:49:31Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-13T06:49:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lqmjj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://74552159f12c87cdda09147692b400ccac0d4e43c1ed72fe8462c4249ad69c7b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://74552159f12c87cdda09147692b400ccac0d4e43c1ed72fe8462c4249ad69c7b\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-13T06:49:32Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-13T06:49:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lqmjj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-13T06:49:25Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-ncchz\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-13T06:49:39Z is after 2025-08-24T17:21:41Z" Dec 13 06:49:39 crc kubenswrapper[4971]: I1213 06:49:39.526746 4971 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-gkcd8" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"13f5c327-f95e-42b4-8783-218500c17694\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T06:49:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T06:49:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T06:49:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T06:49:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://77e846e36f1cb40d03471dc03de11b6c3da448baa407aad38093fae9a801cf92\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-13T06:49:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gvlhn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-13T06:49:27Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-gkcd8\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-13T06:49:39Z is after 2025-08-24T17:21:41Z" Dec 13 06:49:39 crc kubenswrapper[4971]: I1213 06:49:39.536087 4971 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-x98k7" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"c7185dec-3240-42e8-9d90-6182a5e8b65c\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T06:49:39Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T06:49:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T06:49:39Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T06:49:39Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bj2gn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bj2gn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-13T06:49:39Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-x98k7\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-13T06:49:39Z is after 2025-08-24T17:21:41Z" Dec 13 06:49:39 crc kubenswrapper[4971]: I1213 06:49:39.544810 4971 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/c7185dec-3240-42e8-9d90-6182a5e8b65c-metrics-certs\") pod \"network-metrics-daemon-x98k7\" (UID: \"c7185dec-3240-42e8-9d90-6182a5e8b65c\") " pod="openshift-multus/network-metrics-daemon-x98k7" Dec 13 06:49:39 crc kubenswrapper[4971]: I1213 06:49:39.544853 4971 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cqllr\" (UniqueName: \"kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr\") pod \"network-check-target-xd92c\" (UID: \"3b6479f0-333b-4a96-9adf-2099afdc2447\") " pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 13 06:49:39 crc kubenswrapper[4971]: I1213 06:49:39.544922 4971 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-bj2gn\" (UniqueName: \"kubernetes.io/projected/c7185dec-3240-42e8-9d90-6182a5e8b65c-kube-api-access-bj2gn\") pod \"network-metrics-daemon-x98k7\" (UID: \"c7185dec-3240-42e8-9d90-6182a5e8b65c\") " pod="openshift-multus/network-metrics-daemon-x98k7" Dec 13 06:49:39 crc kubenswrapper[4971]: I1213 06:49:39.544944 4971 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-s2dwl\" (UniqueName: \"kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl\") pod \"network-check-source-55646444c4-trplf\" (UID: \"9d751cbb-f2e2-430d-9754-c882a5e924a5\") " pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 13 06:49:39 crc kubenswrapper[4971]: E1213 06:49:39.545074 4971 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Dec 13 06:49:39 crc kubenswrapper[4971]: E1213 06:49:39.545110 4971 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Dec 13 06:49:39 crc kubenswrapper[4971]: E1213 06:49:39.545124 4971 projected.go:194] Error preparing data for projected volume kube-api-access-cqllr for pod openshift-network-diagnostics/network-check-target-xd92c: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Dec 13 06:49:39 crc kubenswrapper[4971]: E1213 06:49:39.545293 4971 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Dec 13 06:49:39 crc kubenswrapper[4971]: E1213 06:49:39.545359 4971 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Dec 13 06:49:39 crc kubenswrapper[4971]: E1213 06:49:39.545385 4971 projected.go:194] Error preparing data for projected volume kube-api-access-s2dwl for pod openshift-network-diagnostics/network-check-source-55646444c4-trplf: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Dec 13 06:49:39 crc kubenswrapper[4971]: E1213 06:49:39.545312 4971 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr podName:3b6479f0-333b-4a96-9adf-2099afdc2447 nodeName:}" failed. No retries permitted until 2025-12-13 06:49:55.545286552 +0000 UTC m=+52.149696080 (durationBeforeRetry 16s). Error: MountVolume.SetUp failed for volume "kube-api-access-cqllr" (UniqueName: "kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr") pod "network-check-target-xd92c" (UID: "3b6479f0-333b-4a96-9adf-2099afdc2447") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Dec 13 06:49:39 crc kubenswrapper[4971]: E1213 06:49:39.545506 4971 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl podName:9d751cbb-f2e2-430d-9754-c882a5e924a5 nodeName:}" failed. No retries permitted until 2025-12-13 06:49:55.545474966 +0000 UTC m=+52.149884454 (durationBeforeRetry 16s). Error: MountVolume.SetUp failed for volume "kube-api-access-s2dwl" (UniqueName: "kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl") pod "network-check-source-55646444c4-trplf" (UID: "9d751cbb-f2e2-430d-9754-c882a5e924a5") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Dec 13 06:49:39 crc kubenswrapper[4971]: I1213 06:49:39.553927 4971 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"1ed9f82e-fe5c-4b76-a619-62e5abd88d3b\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T06:49:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T06:49:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T06:49:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T06:49:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T06:49:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e457329a78a2d24337f994e8a0a1f75a2a667244aa75eb84613d1e7138364aac\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-13T06:49:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://eb014de8140deef097d9bd94f3b9fa4ab3558bc4449bef84b83bf62bfabf5836\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-13T06:49:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://2c438cc74d555f7c9db4569d1c56b78b1088ed8105697d2c0f31e6c5136f3dd0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-13T06:49:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://aa6fa6e4340dcd066252df3c56d79a3a39ed531d5eea8d1d5941d6952368d3af\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-13T06:49:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://c4471b98854c52d5cd084d0ba3e64e5b08e9518ec3c372cce2c875a440522457\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-13T06:49:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7d2746335c98dae3a0f43024d34cb62ebcb42a01c8f9ce18d52b29d4c45598e5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://7d2746335c98dae3a0f43024d34cb62ebcb42a01c8f9ce18d52b29d4c45598e5\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-13T06:49:04Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-13T06:49:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://bc0e77d10a738128bea82122496bc42235fc1688f8907756d93e61b651902b57\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://bc0e77d10a738128bea82122496bc42235fc1688f8907756d93e61b651902b57\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-13T06:49:05Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-13T06:49:05Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://f821e418cba5e832da8839268bcbbd9630bdf319a212d5382bfa2c4b0b73a268\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://f821e418cba5e832da8839268bcbbd9630bdf319a212d5382bfa2c4b0b73a268\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-13T06:49:07Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-13T06:49:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-13T06:49:03Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-13T06:49:39Z is after 2025-08-24T17:21:41Z" Dec 13 06:49:39 crc kubenswrapper[4971]: I1213 06:49:39.565489 4971 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-13T06:49:24Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-13T06:49:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-13T06:49:24Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4328cd6a8464ed468229bbfbc8265f969427916d72ebb92237dbe68efe70fa47\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-13T06:49:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://9d715ca82d9b5a594e41f2b98da6f4277da0e4e1dec114969f901c00f6d7553d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-13T06:49:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-13T06:49:39Z is after 2025-08-24T17:21:41Z" Dec 13 06:49:39 crc kubenswrapper[4971]: I1213 06:49:39.574980 4971 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-5rzh5" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"6bf4591d-2483-4132-9672-d0fd6c995968\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T06:49:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T06:49:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T06:49:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T06:49:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://50e2cb14e4ffb603acb06e2281615a99f06182d345c4e70f2446fa2077033bbb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-13T06:49:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dfxqn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-13T06:49:25Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-5rzh5\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-13T06:49:39Z is after 2025-08-24T17:21:41Z" Dec 13 06:49:39 crc kubenswrapper[4971]: I1213 06:49:39.580877 4971 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 13 06:49:39 crc kubenswrapper[4971]: I1213 06:49:39.580926 4971 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 13 06:49:39 crc kubenswrapper[4971]: I1213 06:49:39.580941 4971 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 13 06:49:39 crc kubenswrapper[4971]: I1213 06:49:39.580961 4971 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 13 06:49:39 crc kubenswrapper[4971]: I1213 06:49:39.580975 4971 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-13T06:49:39Z","lastTransitionTime":"2025-12-13T06:49:39Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 13 06:49:39 crc kubenswrapper[4971]: I1213 06:49:39.584955 4971 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-z572s" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"b4146d3d-4034-49a9-bd37-164856255137\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T06:49:38Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T06:49:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T06:49:38Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-rbac-proxy ovnkube-cluster-manager]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T06:49:38Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-rbac-proxy ovnkube-cluster-manager]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-j6l9b\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-j6l9b\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-13T06:49:38Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-z572s\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-13T06:49:39Z is after 2025-08-24T17:21:41Z" Dec 13 06:49:39 crc kubenswrapper[4971]: I1213 06:49:39.597130 4971 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3ad3268a-a982-428c-9c06-4aaeddd372d9\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T06:49:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T06:49:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T06:49:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T06:49:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T06:49:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3716afdf2bc693be80a09fc9f7727d334f435caf1d4c8ec4811b3a1f1652cdfe\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-13T06:49:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://30df3abf8c79db8af52c6a918f73e1ce410ddafcf341636424c267e1eeb195a7\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-13T06:49:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://ccdf65f6d2b1eb3647a03ba37c305c32e34addfdf79ff937c8dddaa3e5d5b5c6\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-13T06:49:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://770b73dd0c2b6f38744e2c1cd6aa186a69bb87d1cd015a603ad2e33be06a8f6b\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://b3c1bab3d53f975b2e24582ce2771bc26d600e0dd7c30339dcfd18a071162ad4\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-12-13T06:49:23Z\\\",\\\"message\\\":\\\"file observer\\\\nW1213 06:49:23.283403 1 builder.go:272] unable to get owner reference (falling back to namespace): pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\nI1213 06:49:23.284336 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1213 06:49:23.286978 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-981773671/tls.crt::/tmp/serving-cert-981773671/tls.key\\\\\\\"\\\\nI1213 06:49:23.775072 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1213 06:49:23.778615 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1213 06:49:23.778645 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1213 06:49:23.778680 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1213 06:49:23.778691 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1213 06:49:23.785058 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1213 06:49:23.786150 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1213 06:49:23.786196 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1213 06:49:23.786223 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1213 06:49:23.786246 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1213 06:49:23.786269 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1213 06:49:23.786291 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI1213 06:49:23.785608 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF1213 06:49:23.787563 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-13T06:49:08Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-13T06:49:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://9acd07066deed483bd7fa61ee45fda09539d110717a134a56827f871186f92fc\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-13T06:49:07Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://38c91cfbaecfdbd3e3b663422a91ab4294610afb6bbf9ebdb77db24e2b76c682\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://38c91cfbaecfdbd3e3b663422a91ab4294610afb6bbf9ebdb77db24e2b76c682\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-13T06:49:05Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-13T06:49:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-13T06:49:03Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-13T06:49:39Z is after 2025-08-24T17:21:41Z" Dec 13 06:49:39 crc kubenswrapper[4971]: I1213 06:49:39.608438 4971 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-13T06:49:23Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-13T06:49:23Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-13T06:49:39Z is after 2025-08-24T17:21:41Z" Dec 13 06:49:39 crc kubenswrapper[4971]: I1213 06:49:39.621535 4971 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-13T06:49:23Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-13T06:49:23Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-13T06:49:39Z is after 2025-08-24T17:21:41Z" Dec 13 06:49:39 crc kubenswrapper[4971]: I1213 06:49:39.633972 4971 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-13T06:49:24Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-13T06:49:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-13T06:49:24Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d0b2ecf0d5b22f23217f42f259ae7ffb6359555b22c30eaebf667eaf254b3bab\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-13T06:49:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-13T06:49:39Z is after 2025-08-24T17:21:41Z" Dec 13 06:49:39 crc kubenswrapper[4971]: I1213 06:49:39.645950 4971 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/c7185dec-3240-42e8-9d90-6182a5e8b65c-metrics-certs\") pod \"network-metrics-daemon-x98k7\" (UID: \"c7185dec-3240-42e8-9d90-6182a5e8b65c\") " pod="openshift-multus/network-metrics-daemon-x98k7" Dec 13 06:49:39 crc kubenswrapper[4971]: I1213 06:49:39.646022 4971 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-bj2gn\" (UniqueName: \"kubernetes.io/projected/c7185dec-3240-42e8-9d90-6182a5e8b65c-kube-api-access-bj2gn\") pod \"network-metrics-daemon-x98k7\" (UID: \"c7185dec-3240-42e8-9d90-6182a5e8b65c\") " pod="openshift-multus/network-metrics-daemon-x98k7" Dec 13 06:49:39 crc kubenswrapper[4971]: E1213 06:49:39.646130 4971 secret.go:188] Couldn't get secret openshift-multus/metrics-daemon-secret: object "openshift-multus"/"metrics-daemon-secret" not registered Dec 13 06:49:39 crc kubenswrapper[4971]: E1213 06:49:39.646188 4971 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/c7185dec-3240-42e8-9d90-6182a5e8b65c-metrics-certs podName:c7185dec-3240-42e8-9d90-6182a5e8b65c nodeName:}" failed. No retries permitted until 2025-12-13 06:49:40.146171002 +0000 UTC m=+36.750580450 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "metrics-certs" (UniqueName: "kubernetes.io/secret/c7185dec-3240-42e8-9d90-6182a5e8b65c-metrics-certs") pod "network-metrics-daemon-x98k7" (UID: "c7185dec-3240-42e8-9d90-6182a5e8b65c") : object "openshift-multus"/"metrics-daemon-secret" not registered Dec 13 06:49:39 crc kubenswrapper[4971]: I1213 06:49:39.646277 4971 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-13T06:49:23Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-13T06:49:23Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-13T06:49:39Z is after 2025-08-24T17:21:41Z" Dec 13 06:49:39 crc kubenswrapper[4971]: I1213 06:49:39.656363 4971 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-13T06:49:26Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-13T06:49:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-13T06:49:26Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9c8a0d1db0e754462d805e82daf5f52ffd5457507a30456535ec59e096c9d092\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-13T06:49:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-13T06:49:39Z is after 2025-08-24T17:21:41Z" Dec 13 06:49:39 crc kubenswrapper[4971]: I1213 06:49:39.663122 4971 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-bj2gn\" (UniqueName: \"kubernetes.io/projected/c7185dec-3240-42e8-9d90-6182a5e8b65c-kube-api-access-bj2gn\") pod \"network-metrics-daemon-x98k7\" (UID: \"c7185dec-3240-42e8-9d90-6182a5e8b65c\") " pod="openshift-multus/network-metrics-daemon-x98k7" Dec 13 06:49:39 crc kubenswrapper[4971]: I1213 06:49:39.667121 4971 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-82xjz" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e0c4af12-edda-41d7-9b44-f87396174ef2\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T06:49:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T06:49:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T06:49:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T06:49:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://0f4577d9e37f860813e221f2e208177d8d383fd04f673ce357a0a405ea29d1e6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-13T06:49:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cnxsq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://7e0d2020e62f1a277cff7047e1907dc747ed393a1f6c983e1ee258bd156c16c7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-13T06:49:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cnxsq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-13T06:49:25Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-82xjz\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-13T06:49:39Z is after 2025-08-24T17:21:41Z" Dec 13 06:49:39 crc kubenswrapper[4971]: I1213 06:49:39.679927 4971 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-wwvpl" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"dddf744b-2122-49ad-a980-105c09636e0f\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T06:49:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T06:49:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T06:49:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T06:49:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4676d897c7fc668587faec563f3551a056557879cee3e865192d2bd28edf62fc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-13T06:49:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-czz9v\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-13T06:49:25Z\\\"}}\" for pod \"openshift-multus\"/\"multus-wwvpl\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-13T06:49:39Z is after 2025-08-24T17:21:41Z" Dec 13 06:49:39 crc kubenswrapper[4971]: I1213 06:49:39.683567 4971 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 13 06:49:39 crc kubenswrapper[4971]: I1213 06:49:39.683603 4971 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 13 06:49:39 crc kubenswrapper[4971]: I1213 06:49:39.683612 4971 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 13 06:49:39 crc kubenswrapper[4971]: I1213 06:49:39.683626 4971 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 13 06:49:39 crc kubenswrapper[4971]: I1213 06:49:39.683636 4971 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-13T06:49:39Z","lastTransitionTime":"2025-12-13T06:49:39Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 13 06:49:39 crc kubenswrapper[4971]: I1213 06:49:39.697912 4971 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-zlcg5" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"736c6b1a-93dd-4f71-81b6-1c43fcf5f556\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T06:49:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T06:49:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T06:49:25Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T06:49:25Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://325223543b3c7db417f769533559a1ca2b23a723494e1885b0fd0480b5a6399f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-13T06:49:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vzmsr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://f22a4ec756119c4f791fb5105cf1f31df3e548b2b07b1f909a62802e9b6035ae\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-13T06:49:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vzmsr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://f6073140d9f86929b6d4bd2960ede0617a5c9620b269bbab043bae37e0d389f5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-13T06:49:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vzmsr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://693a2ab434c250cb140a866e8068b156df9191414fed1d63bee919a2528ea6e5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-13T06:49:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vzmsr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://219620f51474e60b3e1f880c0972eb612dffc4bad24462bbba2c093ae817d852\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-13T06:49:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vzmsr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://f137f4b441c2402093ab430a9689a46221c4fb4bc70d99fb992f62b0a989c465\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-13T06:49:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vzmsr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://d950a29b0f6f47ae1cc90c132adb059b5aff889a0a38de1d1296e0f083da5e51\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://451040d83e76225457d750a4aa532aa2772decb89719e53fe001bf74aaa8e73b\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-12-13T06:49:35Z\\\",\\\"message\\\":\\\"/pkg/client/informers/externalversions/factory.go:141\\\\nI1213 06:49:35.122568 6277 handler.go:190] Sending *v1.Namespace event handler 1 for removal\\\\nI1213 06:49:35.122615 6277 handler.go:190] Sending *v1.Namespace event handler 5 for removal\\\\nI1213 06:49:35.122659 6277 handler.go:190] Sending *v1.Node event handler 7 for removal\\\\nI1213 06:49:35.122681 6277 handler.go:190] Sending *v1.Node event handler 2 for removal\\\\nI1213 06:49:35.122723 6277 handler.go:190] Sending *v1.NetworkPolicy event handler 4 for removal\\\\nI1213 06:49:35.122771 6277 handler.go:190] Sending *v1.EgressIP event handler 8 for removal\\\\nI1213 06:49:35.122815 6277 handler.go:190] Sending *v1.EgressFirewall event handler 9 for removal\\\\nI1213 06:49:35.122852 6277 factory.go:656] Stopping watch factory\\\\nI1213 06:49:35.122935 6277 ovnkube.go:599] Stopped ovnkube\\\\nI1213 06:49:35.122983 6277 handler.go:208] Removed *v1.Namespace event handler 1\\\\nI1213 06:49:35.123009 6277 handler.go:208] Removed *v1.Namespace event handler 5\\\\nI1213 06:49:35.123033 6277 handler.go:208] Removed *v1.Node event handler 7\\\\nI1213 06:49:35.123056 6277 handler.go:208] Removed *v1.Node event handler 2\\\\nI1213 06:49:35.123079 6277 handler.go:208] Removed *v1.NetworkPolicy event handler 4\\\\nI1213 06:49:35.123102 6277 handler.go:208] Removed *v1.EgressIP event handler 8\\\\nI1213 06:49:3\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-13T06:49:32Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d950a29b0f6f47ae1cc90c132adb059b5aff889a0a38de1d1296e0f083da5e51\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-12-13T06:49:37Z\\\",\\\"message\\\":\\\" ovnkube.go:137] failed to run ovnkube: [failed to start network controller: failed to start default network controller: unable to create admin network policy controller, err: could not add Event Handler for anpInformer during admin network policy controller initialization, handler {0x1fcc6e0 0x1fcc3c0 0x1fcc360} was not added to shared informer because it has stopped already, failed to start node network controller: failed to start default node network controller: failed to set node crc annotations: Internal error occurred: failed calling webhook \\\\\\\"node.network-node-identity.openshift.io\\\\\\\": failed to call webhook: Post \\\\\\\"https://127.0.0.1:9743/node?timeout=10s\\\\\\\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-13T06:49:36Z is after 2025-08-24T17:21:41Z]\\\\nI1213 06:49:36.767920 6409 model_client.go:382] Update operations generated as: [{Op:update Table:Logical_Switch_Port Row:map[addresses:{GoSet:[0a:58:0a:d9:00:04 10.217.0.4]} options:{GoMap:map[iface-id-ver:3b6479f0-333b-4a96-9adf-2099afdc2447 requested-chassis:crc]} port_security:{GoSet:[0a:58:0a:d9:00:04 10.217.0.4]}] Rows:[] Columns:[] Mutations:[] Timeout:\\\\u003cnil\\\\u003e Where:[where column _uuid == {61897e97-c771-4738-8709-09636387cb00}] Until: Durable:\\\\u003cnil\\\\u003e Comment:\\\\u003cnil\\\\u003e Lock:\\\\u003cnil\\\\u003e UUID: UUIDName:}]\\\\nI1213 06:49:36.767917 6409 model_clien\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-13T06:49:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vzmsr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://868049c6a3a4ebe4036d1d22e02ba6b76bced2ee7e269bb228accc4a0636678a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-13T06:49:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vzmsr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://451b22df6973ddc8fa2209a65d64d9e41b33eb231bb4adfab88cf0e5943bb246\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://451b22df6973ddc8fa2209a65d64d9e41b33eb231bb4adfab88cf0e5943bb246\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-13T06:49:26Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-13T06:49:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vzmsr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-13T06:49:25Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-zlcg5\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-13T06:49:39Z is after 2025-08-24T17:21:41Z" Dec 13 06:49:39 crc kubenswrapper[4971]: I1213 06:49:39.767572 4971 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 13 06:49:39 crc kubenswrapper[4971]: I1213 06:49:39.767631 4971 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 13 06:49:39 crc kubenswrapper[4971]: I1213 06:49:39.767694 4971 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 13 06:49:39 crc kubenswrapper[4971]: E1213 06:49:39.767806 4971 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 13 06:49:39 crc kubenswrapper[4971]: E1213 06:49:39.767900 4971 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 13 06:49:39 crc kubenswrapper[4971]: E1213 06:49:39.768006 4971 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 13 06:49:39 crc kubenswrapper[4971]: I1213 06:49:39.786372 4971 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 13 06:49:39 crc kubenswrapper[4971]: I1213 06:49:39.786410 4971 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 13 06:49:39 crc kubenswrapper[4971]: I1213 06:49:39.786422 4971 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 13 06:49:39 crc kubenswrapper[4971]: I1213 06:49:39.786435 4971 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 13 06:49:39 crc kubenswrapper[4971]: I1213 06:49:39.786446 4971 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-13T06:49:39Z","lastTransitionTime":"2025-12-13T06:49:39Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 13 06:49:39 crc kubenswrapper[4971]: I1213 06:49:39.889068 4971 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 13 06:49:39 crc kubenswrapper[4971]: I1213 06:49:39.889115 4971 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 13 06:49:39 crc kubenswrapper[4971]: I1213 06:49:39.889129 4971 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 13 06:49:39 crc kubenswrapper[4971]: I1213 06:49:39.889147 4971 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 13 06:49:39 crc kubenswrapper[4971]: I1213 06:49:39.889156 4971 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-13T06:49:39Z","lastTransitionTime":"2025-12-13T06:49:39Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 13 06:49:39 crc kubenswrapper[4971]: I1213 06:49:39.991685 4971 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 13 06:49:39 crc kubenswrapper[4971]: I1213 06:49:39.991939 4971 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 13 06:49:39 crc kubenswrapper[4971]: I1213 06:49:39.992047 4971 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 13 06:49:39 crc kubenswrapper[4971]: I1213 06:49:39.992122 4971 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 13 06:49:39 crc kubenswrapper[4971]: I1213 06:49:39.992187 4971 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-13T06:49:39Z","lastTransitionTime":"2025-12-13T06:49:39Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 13 06:49:40 crc kubenswrapper[4971]: I1213 06:49:40.040800 4971 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-zlcg5_736c6b1a-93dd-4f71-81b6-1c43fcf5f556/ovnkube-controller/1.log" Dec 13 06:49:40 crc kubenswrapper[4971]: I1213 06:49:40.045974 4971 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-z572s" event={"ID":"b4146d3d-4034-49a9-bd37-164856255137","Type":"ContainerStarted","Data":"dc79a0968f85723e6ad0dc269c8da1f7472adb1bb2b87edd017ee2295bfbf89c"} Dec 13 06:49:40 crc kubenswrapper[4971]: I1213 06:49:40.046200 4971 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-z572s" event={"ID":"b4146d3d-4034-49a9-bd37-164856255137","Type":"ContainerStarted","Data":"46fd4e1070835cfbf7f35d497eff7f047473fc4048ee07812925df5ebdf55342"} Dec 13 06:49:40 crc kubenswrapper[4971]: I1213 06:49:40.060162 4971 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-13T06:49:23Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-13T06:49:23Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-13T06:49:40Z is after 2025-08-24T17:21:41Z" Dec 13 06:49:40 crc kubenswrapper[4971]: I1213 06:49:40.074759 4971 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-13T06:49:23Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-13T06:49:23Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-13T06:49:40Z is after 2025-08-24T17:21:41Z" Dec 13 06:49:40 crc kubenswrapper[4971]: I1213 06:49:40.089151 4971 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3ad3268a-a982-428c-9c06-4aaeddd372d9\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T06:49:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T06:49:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T06:49:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T06:49:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T06:49:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3716afdf2bc693be80a09fc9f7727d334f435caf1d4c8ec4811b3a1f1652cdfe\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-13T06:49:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://30df3abf8c79db8af52c6a918f73e1ce410ddafcf341636424c267e1eeb195a7\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-13T06:49:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://ccdf65f6d2b1eb3647a03ba37c305c32e34addfdf79ff937c8dddaa3e5d5b5c6\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-13T06:49:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://770b73dd0c2b6f38744e2c1cd6aa186a69bb87d1cd015a603ad2e33be06a8f6b\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://b3c1bab3d53f975b2e24582ce2771bc26d600e0dd7c30339dcfd18a071162ad4\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-12-13T06:49:23Z\\\",\\\"message\\\":\\\"file observer\\\\nW1213 06:49:23.283403 1 builder.go:272] unable to get owner reference (falling back to namespace): pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\nI1213 06:49:23.284336 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1213 06:49:23.286978 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-981773671/tls.crt::/tmp/serving-cert-981773671/tls.key\\\\\\\"\\\\nI1213 06:49:23.775072 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1213 06:49:23.778615 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1213 06:49:23.778645 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1213 06:49:23.778680 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1213 06:49:23.778691 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1213 06:49:23.785058 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1213 06:49:23.786150 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1213 06:49:23.786196 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1213 06:49:23.786223 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1213 06:49:23.786246 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1213 06:49:23.786269 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1213 06:49:23.786291 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI1213 06:49:23.785608 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF1213 06:49:23.787563 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-13T06:49:08Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-13T06:49:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://9acd07066deed483bd7fa61ee45fda09539d110717a134a56827f871186f92fc\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-13T06:49:07Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://38c91cfbaecfdbd3e3b663422a91ab4294610afb6bbf9ebdb77db24e2b76c682\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://38c91cfbaecfdbd3e3b663422a91ab4294610afb6bbf9ebdb77db24e2b76c682\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-13T06:49:05Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-13T06:49:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-13T06:49:03Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-13T06:49:40Z is after 2025-08-24T17:21:41Z" Dec 13 06:49:40 crc kubenswrapper[4971]: I1213 06:49:40.095430 4971 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 13 06:49:40 crc kubenswrapper[4971]: I1213 06:49:40.095469 4971 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 13 06:49:40 crc kubenswrapper[4971]: I1213 06:49:40.095479 4971 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 13 06:49:40 crc kubenswrapper[4971]: I1213 06:49:40.095494 4971 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 13 06:49:40 crc kubenswrapper[4971]: I1213 06:49:40.095505 4971 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-13T06:49:40Z","lastTransitionTime":"2025-12-13T06:49:40Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 13 06:49:40 crc kubenswrapper[4971]: I1213 06:49:40.104189 4971 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-13T06:49:23Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-13T06:49:23Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-13T06:49:40Z is after 2025-08-24T17:21:41Z" Dec 13 06:49:40 crc kubenswrapper[4971]: I1213 06:49:40.116999 4971 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-13T06:49:26Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-13T06:49:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-13T06:49:26Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9c8a0d1db0e754462d805e82daf5f52ffd5457507a30456535ec59e096c9d092\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-13T06:49:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-13T06:49:40Z is after 2025-08-24T17:21:41Z" Dec 13 06:49:40 crc kubenswrapper[4971]: I1213 06:49:40.128610 4971 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-82xjz" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e0c4af12-edda-41d7-9b44-f87396174ef2\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T06:49:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T06:49:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T06:49:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T06:49:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://0f4577d9e37f860813e221f2e208177d8d383fd04f673ce357a0a405ea29d1e6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-13T06:49:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cnxsq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://7e0d2020e62f1a277cff7047e1907dc747ed393a1f6c983e1ee258bd156c16c7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-13T06:49:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cnxsq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-13T06:49:25Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-82xjz\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-13T06:49:40Z is after 2025-08-24T17:21:41Z" Dec 13 06:49:40 crc kubenswrapper[4971]: I1213 06:49:40.142411 4971 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-wwvpl" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"dddf744b-2122-49ad-a980-105c09636e0f\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T06:49:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T06:49:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T06:49:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T06:49:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4676d897c7fc668587faec563f3551a056557879cee3e865192d2bd28edf62fc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-13T06:49:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-czz9v\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-13T06:49:25Z\\\"}}\" for pod \"openshift-multus\"/\"multus-wwvpl\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-13T06:49:40Z is after 2025-08-24T17:21:41Z" Dec 13 06:49:40 crc kubenswrapper[4971]: I1213 06:49:40.151831 4971 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/c7185dec-3240-42e8-9d90-6182a5e8b65c-metrics-certs\") pod \"network-metrics-daemon-x98k7\" (UID: \"c7185dec-3240-42e8-9d90-6182a5e8b65c\") " pod="openshift-multus/network-metrics-daemon-x98k7" Dec 13 06:49:40 crc kubenswrapper[4971]: E1213 06:49:40.151957 4971 secret.go:188] Couldn't get secret openshift-multus/metrics-daemon-secret: object "openshift-multus"/"metrics-daemon-secret" not registered Dec 13 06:49:40 crc kubenswrapper[4971]: E1213 06:49:40.152002 4971 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/c7185dec-3240-42e8-9d90-6182a5e8b65c-metrics-certs podName:c7185dec-3240-42e8-9d90-6182a5e8b65c nodeName:}" failed. No retries permitted until 2025-12-13 06:49:41.151988655 +0000 UTC m=+37.756398103 (durationBeforeRetry 1s). Error: MountVolume.SetUp failed for volume "metrics-certs" (UniqueName: "kubernetes.io/secret/c7185dec-3240-42e8-9d90-6182a5e8b65c-metrics-certs") pod "network-metrics-daemon-x98k7" (UID: "c7185dec-3240-42e8-9d90-6182a5e8b65c") : object "openshift-multus"/"metrics-daemon-secret" not registered Dec 13 06:49:40 crc kubenswrapper[4971]: I1213 06:49:40.163805 4971 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 13 06:49:40 crc kubenswrapper[4971]: I1213 06:49:40.163858 4971 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 13 06:49:40 crc kubenswrapper[4971]: I1213 06:49:40.163870 4971 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 13 06:49:40 crc kubenswrapper[4971]: I1213 06:49:40.163887 4971 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 13 06:49:40 crc kubenswrapper[4971]: I1213 06:49:40.163901 4971 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-13T06:49:40Z","lastTransitionTime":"2025-12-13T06:49:40Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 13 06:49:40 crc kubenswrapper[4971]: I1213 06:49:40.165436 4971 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-zlcg5" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"736c6b1a-93dd-4f71-81b6-1c43fcf5f556\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T06:49:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T06:49:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T06:49:25Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T06:49:25Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://325223543b3c7db417f769533559a1ca2b23a723494e1885b0fd0480b5a6399f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-13T06:49:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vzmsr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://f22a4ec756119c4f791fb5105cf1f31df3e548b2b07b1f909a62802e9b6035ae\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-13T06:49:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vzmsr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://f6073140d9f86929b6d4bd2960ede0617a5c9620b269bbab043bae37e0d389f5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-13T06:49:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vzmsr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://693a2ab434c250cb140a866e8068b156df9191414fed1d63bee919a2528ea6e5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-13T06:49:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vzmsr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://219620f51474e60b3e1f880c0972eb612dffc4bad24462bbba2c093ae817d852\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-13T06:49:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vzmsr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://f137f4b441c2402093ab430a9689a46221c4fb4bc70d99fb992f62b0a989c465\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-13T06:49:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vzmsr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://d950a29b0f6f47ae1cc90c132adb059b5aff889a0a38de1d1296e0f083da5e51\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://451040d83e76225457d750a4aa532aa2772decb89719e53fe001bf74aaa8e73b\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-12-13T06:49:35Z\\\",\\\"message\\\":\\\"/pkg/client/informers/externalversions/factory.go:141\\\\nI1213 06:49:35.122568 6277 handler.go:190] Sending *v1.Namespace event handler 1 for removal\\\\nI1213 06:49:35.122615 6277 handler.go:190] Sending *v1.Namespace event handler 5 for removal\\\\nI1213 06:49:35.122659 6277 handler.go:190] Sending *v1.Node event handler 7 for removal\\\\nI1213 06:49:35.122681 6277 handler.go:190] Sending *v1.Node event handler 2 for removal\\\\nI1213 06:49:35.122723 6277 handler.go:190] Sending *v1.NetworkPolicy event handler 4 for removal\\\\nI1213 06:49:35.122771 6277 handler.go:190] Sending *v1.EgressIP event handler 8 for removal\\\\nI1213 06:49:35.122815 6277 handler.go:190] Sending *v1.EgressFirewall event handler 9 for removal\\\\nI1213 06:49:35.122852 6277 factory.go:656] Stopping watch factory\\\\nI1213 06:49:35.122935 6277 ovnkube.go:599] Stopped ovnkube\\\\nI1213 06:49:35.122983 6277 handler.go:208] Removed *v1.Namespace event handler 1\\\\nI1213 06:49:35.123009 6277 handler.go:208] Removed *v1.Namespace event handler 5\\\\nI1213 06:49:35.123033 6277 handler.go:208] Removed *v1.Node event handler 7\\\\nI1213 06:49:35.123056 6277 handler.go:208] Removed *v1.Node event handler 2\\\\nI1213 06:49:35.123079 6277 handler.go:208] Removed *v1.NetworkPolicy event handler 4\\\\nI1213 06:49:35.123102 6277 handler.go:208] Removed *v1.EgressIP event handler 8\\\\nI1213 06:49:3\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-13T06:49:32Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d950a29b0f6f47ae1cc90c132adb059b5aff889a0a38de1d1296e0f083da5e51\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-12-13T06:49:37Z\\\",\\\"message\\\":\\\" ovnkube.go:137] failed to run ovnkube: [failed to start network controller: failed to start default network controller: unable to create admin network policy controller, err: could not add Event Handler for anpInformer during admin network policy controller initialization, handler {0x1fcc6e0 0x1fcc3c0 0x1fcc360} was not added to shared informer because it has stopped already, failed to start node network controller: failed to start default node network controller: failed to set node crc annotations: Internal error occurred: failed calling webhook \\\\\\\"node.network-node-identity.openshift.io\\\\\\\": failed to call webhook: Post \\\\\\\"https://127.0.0.1:9743/node?timeout=10s\\\\\\\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-13T06:49:36Z is after 2025-08-24T17:21:41Z]\\\\nI1213 06:49:36.767920 6409 model_client.go:382] Update operations generated as: [{Op:update Table:Logical_Switch_Port Row:map[addresses:{GoSet:[0a:58:0a:d9:00:04 10.217.0.4]} options:{GoMap:map[iface-id-ver:3b6479f0-333b-4a96-9adf-2099afdc2447 requested-chassis:crc]} port_security:{GoSet:[0a:58:0a:d9:00:04 10.217.0.4]}] Rows:[] Columns:[] Mutations:[] Timeout:\\\\u003cnil\\\\u003e Where:[where column _uuid == {61897e97-c771-4738-8709-09636387cb00}] Until: Durable:\\\\u003cnil\\\\u003e Comment:\\\\u003cnil\\\\u003e Lock:\\\\u003cnil\\\\u003e UUID: UUIDName:}]\\\\nI1213 06:49:36.767917 6409 model_clien\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-13T06:49:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vzmsr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://868049c6a3a4ebe4036d1d22e02ba6b76bced2ee7e269bb228accc4a0636678a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-13T06:49:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vzmsr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://451b22df6973ddc8fa2209a65d64d9e41b33eb231bb4adfab88cf0e5943bb246\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://451b22df6973ddc8fa2209a65d64d9e41b33eb231bb4adfab88cf0e5943bb246\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-13T06:49:26Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-13T06:49:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vzmsr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-13T06:49:25Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-zlcg5\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-13T06:49:40Z is after 2025-08-24T17:21:41Z" Dec 13 06:49:40 crc kubenswrapper[4971]: E1213 06:49:40.177157 4971 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-12-13T06:49:40Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-13T06:49:40Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-13T06:49:40Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-13T06:49:40Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-13T06:49:40Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-13T06:49:40Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-13T06:49:40Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-13T06:49:40Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"bd36cbc9-bbe9-45a1-a21a-ad2bd10499ee\\\",\\\"systemUUID\\\":\\\"a9af00dc-cf25-47af-9997-79577f8f3bea\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-13T06:49:40Z is after 2025-08-24T17:21:41Z" Dec 13 06:49:40 crc kubenswrapper[4971]: I1213 06:49:40.180681 4971 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 13 06:49:40 crc kubenswrapper[4971]: I1213 06:49:40.180837 4971 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 13 06:49:40 crc kubenswrapper[4971]: I1213 06:49:40.180919 4971 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 13 06:49:40 crc kubenswrapper[4971]: I1213 06:49:40.181006 4971 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 13 06:49:40 crc kubenswrapper[4971]: I1213 06:49:40.181088 4971 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-13T06:49:40Z","lastTransitionTime":"2025-12-13T06:49:40Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 13 06:49:40 crc kubenswrapper[4971]: I1213 06:49:40.182852 4971 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-13T06:49:24Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-13T06:49:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-13T06:49:24Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d0b2ecf0d5b22f23217f42f259ae7ffb6359555b22c30eaebf667eaf254b3bab\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-13T06:49:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-13T06:49:40Z is after 2025-08-24T17:21:41Z" Dec 13 06:49:40 crc kubenswrapper[4971]: E1213 06:49:40.192833 4971 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-12-13T06:49:40Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-13T06:49:40Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-13T06:49:40Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-13T06:49:40Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-13T06:49:40Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-13T06:49:40Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-13T06:49:40Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-13T06:49:40Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"bd36cbc9-bbe9-45a1-a21a-ad2bd10499ee\\\",\\\"systemUUID\\\":\\\"a9af00dc-cf25-47af-9997-79577f8f3bea\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-13T06:49:40Z is after 2025-08-24T17:21:41Z" Dec 13 06:49:40 crc kubenswrapper[4971]: I1213 06:49:40.196260 4971 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-ncchz" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"de914d21-a688-401e-a66e-861beb6336d9\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T06:49:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T06:49:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T06:49:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T06:49:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://244ffe3dc9d8cf799aff4c4d795bf41554c1e6bb6667fbb0c7df322fb5f96c68\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-13T06:49:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lqmjj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://0b7111af6c5a198d02bb02858837cfb4f3b20828d8d206cd18aa0ffa102d0cd4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://0b7111af6c5a198d02bb02858837cfb4f3b20828d8d206cd18aa0ffa102d0cd4\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-13T06:49:26Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-13T06:49:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lqmjj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b36777fe40e74ae1cecb7b73e204139db5279ba51cef4f1e8c4a4c491d8be403\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://b36777fe40e74ae1cecb7b73e204139db5279ba51cef4f1e8c4a4c491d8be403\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-13T06:49:28Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-13T06:49:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lqmjj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://0597a06cc663e416142311a16941d98297577693a1fe27201ea42c27496cace2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://0597a06cc663e416142311a16941d98297577693a1fe27201ea42c27496cace2\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-13T06:49:29Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-13T06:49:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lqmjj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://3f52c2e22a75c010993ca5809a3a9d1304758d4cb6e83691ef1af15a4b514ac4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://3f52c2e22a75c010993ca5809a3a9d1304758d4cb6e83691ef1af15a4b514ac4\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-13T06:49:30Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-13T06:49:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lqmjj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://9e21ca708c4c4c1228b417cece51299e0792729b7c78884bd18b2dda8fe8355c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://9e21ca708c4c4c1228b417cece51299e0792729b7c78884bd18b2dda8fe8355c\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-13T06:49:31Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-13T06:49:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lqmjj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://74552159f12c87cdda09147692b400ccac0d4e43c1ed72fe8462c4249ad69c7b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://74552159f12c87cdda09147692b400ccac0d4e43c1ed72fe8462c4249ad69c7b\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-13T06:49:32Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-13T06:49:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lqmjj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-13T06:49:25Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-ncchz\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-13T06:49:40Z is after 2025-08-24T17:21:41Z" Dec 13 06:49:40 crc kubenswrapper[4971]: I1213 06:49:40.197248 4971 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 13 06:49:40 crc kubenswrapper[4971]: I1213 06:49:40.197333 4971 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 13 06:49:40 crc kubenswrapper[4971]: I1213 06:49:40.197390 4971 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 13 06:49:40 crc kubenswrapper[4971]: I1213 06:49:40.197453 4971 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 13 06:49:40 crc kubenswrapper[4971]: I1213 06:49:40.197575 4971 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-13T06:49:40Z","lastTransitionTime":"2025-12-13T06:49:40Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 13 06:49:40 crc kubenswrapper[4971]: I1213 06:49:40.207405 4971 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-gkcd8" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"13f5c327-f95e-42b4-8783-218500c17694\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T06:49:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T06:49:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T06:49:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T06:49:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://77e846e36f1cb40d03471dc03de11b6c3da448baa407aad38093fae9a801cf92\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-13T06:49:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gvlhn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-13T06:49:27Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-gkcd8\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-13T06:49:40Z is after 2025-08-24T17:21:41Z" Dec 13 06:49:40 crc kubenswrapper[4971]: E1213 06:49:40.209482 4971 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-12-13T06:49:40Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-13T06:49:40Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-13T06:49:40Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-13T06:49:40Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-13T06:49:40Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-13T06:49:40Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-13T06:49:40Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-13T06:49:40Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"bd36cbc9-bbe9-45a1-a21a-ad2bd10499ee\\\",\\\"systemUUID\\\":\\\"a9af00dc-cf25-47af-9997-79577f8f3bea\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-13T06:49:40Z is after 2025-08-24T17:21:41Z" Dec 13 06:49:40 crc kubenswrapper[4971]: I1213 06:49:40.214022 4971 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 13 06:49:40 crc kubenswrapper[4971]: I1213 06:49:40.214058 4971 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 13 06:49:40 crc kubenswrapper[4971]: I1213 06:49:40.214069 4971 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 13 06:49:40 crc kubenswrapper[4971]: I1213 06:49:40.214085 4971 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 13 06:49:40 crc kubenswrapper[4971]: I1213 06:49:40.214098 4971 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-13T06:49:40Z","lastTransitionTime":"2025-12-13T06:49:40Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 13 06:49:40 crc kubenswrapper[4971]: I1213 06:49:40.219612 4971 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-x98k7" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"c7185dec-3240-42e8-9d90-6182a5e8b65c\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T06:49:39Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T06:49:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T06:49:39Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T06:49:39Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bj2gn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bj2gn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-13T06:49:39Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-x98k7\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-13T06:49:40Z is after 2025-08-24T17:21:41Z" Dec 13 06:49:40 crc kubenswrapper[4971]: E1213 06:49:40.225786 4971 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-12-13T06:49:40Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-13T06:49:40Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-13T06:49:40Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-13T06:49:40Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-13T06:49:40Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-13T06:49:40Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-13T06:49:40Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-13T06:49:40Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"bd36cbc9-bbe9-45a1-a21a-ad2bd10499ee\\\",\\\"systemUUID\\\":\\\"a9af00dc-cf25-47af-9997-79577f8f3bea\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-13T06:49:40Z is after 2025-08-24T17:21:41Z" Dec 13 06:49:40 crc kubenswrapper[4971]: I1213 06:49:40.229356 4971 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 13 06:49:40 crc kubenswrapper[4971]: I1213 06:49:40.229395 4971 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 13 06:49:40 crc kubenswrapper[4971]: I1213 06:49:40.229405 4971 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 13 06:49:40 crc kubenswrapper[4971]: I1213 06:49:40.229424 4971 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 13 06:49:40 crc kubenswrapper[4971]: I1213 06:49:40.229434 4971 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-13T06:49:40Z","lastTransitionTime":"2025-12-13T06:49:40Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 13 06:49:40 crc kubenswrapper[4971]: I1213 06:49:40.235018 4971 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"f9d31b5a-b770-4a15-95c8-35b49677c608\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T06:49:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T06:49:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T06:49:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T06:49:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T06:49:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c9baa042f9dada954faa57b5454cd6b0a12a297e56ddc192c52cff50e42a16fa\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-13T06:49:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://c788de86754f0bf37ee184672fc1dcb388ccebc246d4f15be4b5da661f725f8f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-13T06:49:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://760817511a9b4f96a68dc38d26631250eaab72e1dd8f123e653039b6b4e91f84\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-13T06:49:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b859be0c9feac856e9855c6a038b02b49b32fade97873c46d0bec0a32d8b5147\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-13T06:49:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-13T06:49:03Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-13T06:49:40Z is after 2025-08-24T17:21:41Z" Dec 13 06:49:40 crc kubenswrapper[4971]: E1213 06:49:40.242651 4971 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-12-13T06:49:40Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-13T06:49:40Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-13T06:49:40Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-13T06:49:40Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-13T06:49:40Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-13T06:49:40Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-13T06:49:40Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-13T06:49:40Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"bd36cbc9-bbe9-45a1-a21a-ad2bd10499ee\\\",\\\"systemUUID\\\":\\\"a9af00dc-cf25-47af-9997-79577f8f3bea\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-13T06:49:40Z is after 2025-08-24T17:21:41Z" Dec 13 06:49:40 crc kubenswrapper[4971]: E1213 06:49:40.242879 4971 kubelet_node_status.go:572] "Unable to update node status" err="update node status exceeds retry count" Dec 13 06:49:40 crc kubenswrapper[4971]: I1213 06:49:40.244830 4971 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 13 06:49:40 crc kubenswrapper[4971]: I1213 06:49:40.244893 4971 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 13 06:49:40 crc kubenswrapper[4971]: I1213 06:49:40.244910 4971 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 13 06:49:40 crc kubenswrapper[4971]: I1213 06:49:40.244934 4971 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 13 06:49:40 crc kubenswrapper[4971]: I1213 06:49:40.244948 4971 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-13T06:49:40Z","lastTransitionTime":"2025-12-13T06:49:40Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 13 06:49:40 crc kubenswrapper[4971]: I1213 06:49:40.260485 4971 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"1ed9f82e-fe5c-4b76-a619-62e5abd88d3b\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T06:49:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T06:49:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T06:49:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T06:49:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T06:49:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e457329a78a2d24337f994e8a0a1f75a2a667244aa75eb84613d1e7138364aac\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-13T06:49:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://eb014de8140deef097d9bd94f3b9fa4ab3558bc4449bef84b83bf62bfabf5836\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-13T06:49:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://2c438cc74d555f7c9db4569d1c56b78b1088ed8105697d2c0f31e6c5136f3dd0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-13T06:49:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://aa6fa6e4340dcd066252df3c56d79a3a39ed531d5eea8d1d5941d6952368d3af\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-13T06:49:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://c4471b98854c52d5cd084d0ba3e64e5b08e9518ec3c372cce2c875a440522457\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-13T06:49:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7d2746335c98dae3a0f43024d34cb62ebcb42a01c8f9ce18d52b29d4c45598e5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://7d2746335c98dae3a0f43024d34cb62ebcb42a01c8f9ce18d52b29d4c45598e5\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-13T06:49:04Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-13T06:49:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://bc0e77d10a738128bea82122496bc42235fc1688f8907756d93e61b651902b57\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://bc0e77d10a738128bea82122496bc42235fc1688f8907756d93e61b651902b57\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-13T06:49:05Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-13T06:49:05Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://f821e418cba5e832da8839268bcbbd9630bdf319a212d5382bfa2c4b0b73a268\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://f821e418cba5e832da8839268bcbbd9630bdf319a212d5382bfa2c4b0b73a268\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-13T06:49:07Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-13T06:49:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-13T06:49:03Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-13T06:49:40Z is after 2025-08-24T17:21:41Z" Dec 13 06:49:40 crc kubenswrapper[4971]: I1213 06:49:40.275997 4971 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-13T06:49:24Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-13T06:49:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-13T06:49:24Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4328cd6a8464ed468229bbfbc8265f969427916d72ebb92237dbe68efe70fa47\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-13T06:49:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://9d715ca82d9b5a594e41f2b98da6f4277da0e4e1dec114969f901c00f6d7553d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-13T06:49:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-13T06:49:40Z is after 2025-08-24T17:21:41Z" Dec 13 06:49:40 crc kubenswrapper[4971]: I1213 06:49:40.287261 4971 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-5rzh5" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"6bf4591d-2483-4132-9672-d0fd6c995968\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T06:49:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T06:49:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T06:49:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T06:49:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://50e2cb14e4ffb603acb06e2281615a99f06182d345c4e70f2446fa2077033bbb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-13T06:49:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dfxqn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-13T06:49:25Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-5rzh5\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-13T06:49:40Z is after 2025-08-24T17:21:41Z" Dec 13 06:49:40 crc kubenswrapper[4971]: I1213 06:49:40.301960 4971 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-z572s" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"b4146d3d-4034-49a9-bd37-164856255137\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T06:49:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T06:49:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T06:49:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T06:49:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://46fd4e1070835cfbf7f35d497eff7f047473fc4048ee07812925df5ebdf55342\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-13T06:49:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-j6l9b\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://dc79a0968f85723e6ad0dc269c8da1f7472adb1bb2b87edd017ee2295bfbf89c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-13T06:49:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-j6l9b\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-13T06:49:38Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-z572s\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-13T06:49:40Z is after 2025-08-24T17:21:41Z" Dec 13 06:49:40 crc kubenswrapper[4971]: I1213 06:49:40.346848 4971 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 13 06:49:40 crc kubenswrapper[4971]: I1213 06:49:40.346908 4971 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 13 06:49:40 crc kubenswrapper[4971]: I1213 06:49:40.346920 4971 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 13 06:49:40 crc kubenswrapper[4971]: I1213 06:49:40.346941 4971 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 13 06:49:40 crc kubenswrapper[4971]: I1213 06:49:40.346955 4971 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-13T06:49:40Z","lastTransitionTime":"2025-12-13T06:49:40Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 13 06:49:40 crc kubenswrapper[4971]: I1213 06:49:40.450061 4971 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 13 06:49:40 crc kubenswrapper[4971]: I1213 06:49:40.450127 4971 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 13 06:49:40 crc kubenswrapper[4971]: I1213 06:49:40.450139 4971 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 13 06:49:40 crc kubenswrapper[4971]: I1213 06:49:40.450163 4971 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 13 06:49:40 crc kubenswrapper[4971]: I1213 06:49:40.450178 4971 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-13T06:49:40Z","lastTransitionTime":"2025-12-13T06:49:40Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 13 06:49:40 crc kubenswrapper[4971]: I1213 06:49:40.552512 4971 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 13 06:49:40 crc kubenswrapper[4971]: I1213 06:49:40.552577 4971 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 13 06:49:40 crc kubenswrapper[4971]: I1213 06:49:40.552589 4971 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 13 06:49:40 crc kubenswrapper[4971]: I1213 06:49:40.552613 4971 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 13 06:49:40 crc kubenswrapper[4971]: I1213 06:49:40.552625 4971 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-13T06:49:40Z","lastTransitionTime":"2025-12-13T06:49:40Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 13 06:49:40 crc kubenswrapper[4971]: I1213 06:49:40.656106 4971 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 13 06:49:40 crc kubenswrapper[4971]: I1213 06:49:40.656362 4971 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 13 06:49:40 crc kubenswrapper[4971]: I1213 06:49:40.656421 4971 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 13 06:49:40 crc kubenswrapper[4971]: I1213 06:49:40.656478 4971 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 13 06:49:40 crc kubenswrapper[4971]: I1213 06:49:40.656829 4971 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-13T06:49:40Z","lastTransitionTime":"2025-12-13T06:49:40Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 13 06:49:40 crc kubenswrapper[4971]: I1213 06:49:40.758960 4971 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 13 06:49:40 crc kubenswrapper[4971]: I1213 06:49:40.758997 4971 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 13 06:49:40 crc kubenswrapper[4971]: I1213 06:49:40.759008 4971 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 13 06:49:40 crc kubenswrapper[4971]: I1213 06:49:40.759023 4971 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 13 06:49:40 crc kubenswrapper[4971]: I1213 06:49:40.759033 4971 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-13T06:49:40Z","lastTransitionTime":"2025-12-13T06:49:40Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 13 06:49:40 crc kubenswrapper[4971]: I1213 06:49:40.860870 4971 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 13 06:49:40 crc kubenswrapper[4971]: I1213 06:49:40.860919 4971 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 13 06:49:40 crc kubenswrapper[4971]: I1213 06:49:40.860934 4971 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 13 06:49:40 crc kubenswrapper[4971]: I1213 06:49:40.860955 4971 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 13 06:49:40 crc kubenswrapper[4971]: I1213 06:49:40.860970 4971 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-13T06:49:40Z","lastTransitionTime":"2025-12-13T06:49:40Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 13 06:49:40 crc kubenswrapper[4971]: I1213 06:49:40.962875 4971 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 13 06:49:40 crc kubenswrapper[4971]: I1213 06:49:40.962918 4971 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 13 06:49:40 crc kubenswrapper[4971]: I1213 06:49:40.962929 4971 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 13 06:49:40 crc kubenswrapper[4971]: I1213 06:49:40.962945 4971 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 13 06:49:40 crc kubenswrapper[4971]: I1213 06:49:40.962955 4971 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-13T06:49:40Z","lastTransitionTime":"2025-12-13T06:49:40Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 13 06:49:41 crc kubenswrapper[4971]: I1213 06:49:41.066194 4971 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 13 06:49:41 crc kubenswrapper[4971]: I1213 06:49:41.066244 4971 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 13 06:49:41 crc kubenswrapper[4971]: I1213 06:49:41.066261 4971 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 13 06:49:41 crc kubenswrapper[4971]: I1213 06:49:41.066284 4971 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 13 06:49:41 crc kubenswrapper[4971]: I1213 06:49:41.066303 4971 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-13T06:49:41Z","lastTransitionTime":"2025-12-13T06:49:41Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 13 06:49:41 crc kubenswrapper[4971]: I1213 06:49:41.162645 4971 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/c7185dec-3240-42e8-9d90-6182a5e8b65c-metrics-certs\") pod \"network-metrics-daemon-x98k7\" (UID: \"c7185dec-3240-42e8-9d90-6182a5e8b65c\") " pod="openshift-multus/network-metrics-daemon-x98k7" Dec 13 06:49:41 crc kubenswrapper[4971]: E1213 06:49:41.162919 4971 secret.go:188] Couldn't get secret openshift-multus/metrics-daemon-secret: object "openshift-multus"/"metrics-daemon-secret" not registered Dec 13 06:49:41 crc kubenswrapper[4971]: E1213 06:49:41.163010 4971 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/c7185dec-3240-42e8-9d90-6182a5e8b65c-metrics-certs podName:c7185dec-3240-42e8-9d90-6182a5e8b65c nodeName:}" failed. No retries permitted until 2025-12-13 06:49:43.162987806 +0000 UTC m=+39.767397264 (durationBeforeRetry 2s). Error: MountVolume.SetUp failed for volume "metrics-certs" (UniqueName: "kubernetes.io/secret/c7185dec-3240-42e8-9d90-6182a5e8b65c-metrics-certs") pod "network-metrics-daemon-x98k7" (UID: "c7185dec-3240-42e8-9d90-6182a5e8b65c") : object "openshift-multus"/"metrics-daemon-secret" not registered Dec 13 06:49:41 crc kubenswrapper[4971]: I1213 06:49:41.168283 4971 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 13 06:49:41 crc kubenswrapper[4971]: I1213 06:49:41.168361 4971 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 13 06:49:41 crc kubenswrapper[4971]: I1213 06:49:41.168374 4971 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 13 06:49:41 crc kubenswrapper[4971]: I1213 06:49:41.168391 4971 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 13 06:49:41 crc kubenswrapper[4971]: I1213 06:49:41.168402 4971 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-13T06:49:41Z","lastTransitionTime":"2025-12-13T06:49:41Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 13 06:49:41 crc kubenswrapper[4971]: I1213 06:49:41.271239 4971 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 13 06:49:41 crc kubenswrapper[4971]: I1213 06:49:41.271490 4971 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 13 06:49:41 crc kubenswrapper[4971]: I1213 06:49:41.271600 4971 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 13 06:49:41 crc kubenswrapper[4971]: I1213 06:49:41.271715 4971 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 13 06:49:41 crc kubenswrapper[4971]: I1213 06:49:41.271774 4971 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-13T06:49:41Z","lastTransitionTime":"2025-12-13T06:49:41Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 13 06:49:41 crc kubenswrapper[4971]: I1213 06:49:41.374447 4971 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 13 06:49:41 crc kubenswrapper[4971]: I1213 06:49:41.374554 4971 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 13 06:49:41 crc kubenswrapper[4971]: I1213 06:49:41.374568 4971 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 13 06:49:41 crc kubenswrapper[4971]: I1213 06:49:41.374585 4971 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 13 06:49:41 crc kubenswrapper[4971]: I1213 06:49:41.374598 4971 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-13T06:49:41Z","lastTransitionTime":"2025-12-13T06:49:41Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 13 06:49:41 crc kubenswrapper[4971]: I1213 06:49:41.477106 4971 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 13 06:49:41 crc kubenswrapper[4971]: I1213 06:49:41.477155 4971 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 13 06:49:41 crc kubenswrapper[4971]: I1213 06:49:41.477167 4971 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 13 06:49:41 crc kubenswrapper[4971]: I1213 06:49:41.477185 4971 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 13 06:49:41 crc kubenswrapper[4971]: I1213 06:49:41.477196 4971 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-13T06:49:41Z","lastTransitionTime":"2025-12-13T06:49:41Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 13 06:49:41 crc kubenswrapper[4971]: I1213 06:49:41.579756 4971 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 13 06:49:41 crc kubenswrapper[4971]: I1213 06:49:41.579802 4971 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 13 06:49:41 crc kubenswrapper[4971]: I1213 06:49:41.579811 4971 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 13 06:49:41 crc kubenswrapper[4971]: I1213 06:49:41.579825 4971 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 13 06:49:41 crc kubenswrapper[4971]: I1213 06:49:41.579836 4971 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-13T06:49:41Z","lastTransitionTime":"2025-12-13T06:49:41Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 13 06:49:41 crc kubenswrapper[4971]: I1213 06:49:41.682479 4971 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 13 06:49:41 crc kubenswrapper[4971]: I1213 06:49:41.682537 4971 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 13 06:49:41 crc kubenswrapper[4971]: I1213 06:49:41.682550 4971 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 13 06:49:41 crc kubenswrapper[4971]: I1213 06:49:41.682567 4971 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 13 06:49:41 crc kubenswrapper[4971]: I1213 06:49:41.682580 4971 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-13T06:49:41Z","lastTransitionTime":"2025-12-13T06:49:41Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 13 06:49:41 crc kubenswrapper[4971]: I1213 06:49:41.767433 4971 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 13 06:49:41 crc kubenswrapper[4971]: I1213 06:49:41.767477 4971 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-x98k7" Dec 13 06:49:41 crc kubenswrapper[4971]: E1213 06:49:41.767585 4971 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 13 06:49:41 crc kubenswrapper[4971]: I1213 06:49:41.767492 4971 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 13 06:49:41 crc kubenswrapper[4971]: I1213 06:49:41.767489 4971 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 13 06:49:41 crc kubenswrapper[4971]: E1213 06:49:41.767653 4971 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-x98k7" podUID="c7185dec-3240-42e8-9d90-6182a5e8b65c" Dec 13 06:49:41 crc kubenswrapper[4971]: E1213 06:49:41.767777 4971 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 13 06:49:41 crc kubenswrapper[4971]: E1213 06:49:41.767888 4971 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 13 06:49:41 crc kubenswrapper[4971]: I1213 06:49:41.784936 4971 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 13 06:49:41 crc kubenswrapper[4971]: I1213 06:49:41.784982 4971 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 13 06:49:41 crc kubenswrapper[4971]: I1213 06:49:41.784994 4971 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 13 06:49:41 crc kubenswrapper[4971]: I1213 06:49:41.785012 4971 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 13 06:49:41 crc kubenswrapper[4971]: I1213 06:49:41.785024 4971 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-13T06:49:41Z","lastTransitionTime":"2025-12-13T06:49:41Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 13 06:49:41 crc kubenswrapper[4971]: I1213 06:49:41.888044 4971 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 13 06:49:41 crc kubenswrapper[4971]: I1213 06:49:41.888278 4971 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 13 06:49:41 crc kubenswrapper[4971]: I1213 06:49:41.888356 4971 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 13 06:49:41 crc kubenswrapper[4971]: I1213 06:49:41.888442 4971 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 13 06:49:41 crc kubenswrapper[4971]: I1213 06:49:41.888581 4971 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-13T06:49:41Z","lastTransitionTime":"2025-12-13T06:49:41Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 13 06:49:41 crc kubenswrapper[4971]: I1213 06:49:41.991588 4971 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 13 06:49:41 crc kubenswrapper[4971]: I1213 06:49:41.991635 4971 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 13 06:49:41 crc kubenswrapper[4971]: I1213 06:49:41.991652 4971 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 13 06:49:41 crc kubenswrapper[4971]: I1213 06:49:41.991671 4971 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 13 06:49:41 crc kubenswrapper[4971]: I1213 06:49:41.991684 4971 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-13T06:49:41Z","lastTransitionTime":"2025-12-13T06:49:41Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 13 06:49:42 crc kubenswrapper[4971]: I1213 06:49:42.094695 4971 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 13 06:49:42 crc kubenswrapper[4971]: I1213 06:49:42.094746 4971 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 13 06:49:42 crc kubenswrapper[4971]: I1213 06:49:42.094757 4971 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 13 06:49:42 crc kubenswrapper[4971]: I1213 06:49:42.094777 4971 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 13 06:49:42 crc kubenswrapper[4971]: I1213 06:49:42.094795 4971 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-13T06:49:42Z","lastTransitionTime":"2025-12-13T06:49:42Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 13 06:49:42 crc kubenswrapper[4971]: I1213 06:49:42.197073 4971 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 13 06:49:42 crc kubenswrapper[4971]: I1213 06:49:42.197127 4971 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 13 06:49:42 crc kubenswrapper[4971]: I1213 06:49:42.197136 4971 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 13 06:49:42 crc kubenswrapper[4971]: I1213 06:49:42.197151 4971 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 13 06:49:42 crc kubenswrapper[4971]: I1213 06:49:42.197160 4971 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-13T06:49:42Z","lastTransitionTime":"2025-12-13T06:49:42Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 13 06:49:42 crc kubenswrapper[4971]: I1213 06:49:42.299841 4971 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 13 06:49:42 crc kubenswrapper[4971]: I1213 06:49:42.300080 4971 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 13 06:49:42 crc kubenswrapper[4971]: I1213 06:49:42.300176 4971 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 13 06:49:42 crc kubenswrapper[4971]: I1213 06:49:42.300295 4971 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 13 06:49:42 crc kubenswrapper[4971]: I1213 06:49:42.300381 4971 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-13T06:49:42Z","lastTransitionTime":"2025-12-13T06:49:42Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 13 06:49:42 crc kubenswrapper[4971]: I1213 06:49:42.403407 4971 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 13 06:49:42 crc kubenswrapper[4971]: I1213 06:49:42.403892 4971 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 13 06:49:42 crc kubenswrapper[4971]: I1213 06:49:42.404085 4971 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 13 06:49:42 crc kubenswrapper[4971]: I1213 06:49:42.404255 4971 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 13 06:49:42 crc kubenswrapper[4971]: I1213 06:49:42.404430 4971 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-13T06:49:42Z","lastTransitionTime":"2025-12-13T06:49:42Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 13 06:49:42 crc kubenswrapper[4971]: I1213 06:49:42.506694 4971 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 13 06:49:42 crc kubenswrapper[4971]: I1213 06:49:42.506768 4971 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 13 06:49:42 crc kubenswrapper[4971]: I1213 06:49:42.506779 4971 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 13 06:49:42 crc kubenswrapper[4971]: I1213 06:49:42.506793 4971 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 13 06:49:42 crc kubenswrapper[4971]: I1213 06:49:42.506804 4971 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-13T06:49:42Z","lastTransitionTime":"2025-12-13T06:49:42Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 13 06:49:42 crc kubenswrapper[4971]: I1213 06:49:42.609326 4971 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 13 06:49:42 crc kubenswrapper[4971]: I1213 06:49:42.609362 4971 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 13 06:49:42 crc kubenswrapper[4971]: I1213 06:49:42.609371 4971 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 13 06:49:42 crc kubenswrapper[4971]: I1213 06:49:42.609387 4971 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 13 06:49:42 crc kubenswrapper[4971]: I1213 06:49:42.609396 4971 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-13T06:49:42Z","lastTransitionTime":"2025-12-13T06:49:42Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 13 06:49:42 crc kubenswrapper[4971]: I1213 06:49:42.712328 4971 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 13 06:49:42 crc kubenswrapper[4971]: I1213 06:49:42.712368 4971 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 13 06:49:42 crc kubenswrapper[4971]: I1213 06:49:42.712382 4971 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 13 06:49:42 crc kubenswrapper[4971]: I1213 06:49:42.712404 4971 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 13 06:49:42 crc kubenswrapper[4971]: I1213 06:49:42.712417 4971 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-13T06:49:42Z","lastTransitionTime":"2025-12-13T06:49:42Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 13 06:49:42 crc kubenswrapper[4971]: I1213 06:49:42.814972 4971 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 13 06:49:42 crc kubenswrapper[4971]: I1213 06:49:42.815023 4971 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 13 06:49:42 crc kubenswrapper[4971]: I1213 06:49:42.815058 4971 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 13 06:49:42 crc kubenswrapper[4971]: I1213 06:49:42.815081 4971 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 13 06:49:42 crc kubenswrapper[4971]: I1213 06:49:42.815095 4971 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-13T06:49:42Z","lastTransitionTime":"2025-12-13T06:49:42Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 13 06:49:42 crc kubenswrapper[4971]: I1213 06:49:42.917306 4971 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 13 06:49:42 crc kubenswrapper[4971]: I1213 06:49:42.917554 4971 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 13 06:49:42 crc kubenswrapper[4971]: I1213 06:49:42.917691 4971 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 13 06:49:42 crc kubenswrapper[4971]: I1213 06:49:42.917788 4971 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 13 06:49:42 crc kubenswrapper[4971]: I1213 06:49:42.917870 4971 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-13T06:49:42Z","lastTransitionTime":"2025-12-13T06:49:42Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 13 06:49:43 crc kubenswrapper[4971]: I1213 06:49:43.020512 4971 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 13 06:49:43 crc kubenswrapper[4971]: I1213 06:49:43.020565 4971 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 13 06:49:43 crc kubenswrapper[4971]: I1213 06:49:43.020574 4971 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 13 06:49:43 crc kubenswrapper[4971]: I1213 06:49:43.020588 4971 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 13 06:49:43 crc kubenswrapper[4971]: I1213 06:49:43.020597 4971 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-13T06:49:43Z","lastTransitionTime":"2025-12-13T06:49:43Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 13 06:49:43 crc kubenswrapper[4971]: I1213 06:49:43.123000 4971 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 13 06:49:43 crc kubenswrapper[4971]: I1213 06:49:43.123261 4971 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 13 06:49:43 crc kubenswrapper[4971]: I1213 06:49:43.123331 4971 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 13 06:49:43 crc kubenswrapper[4971]: I1213 06:49:43.123393 4971 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 13 06:49:43 crc kubenswrapper[4971]: I1213 06:49:43.123450 4971 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-13T06:49:43Z","lastTransitionTime":"2025-12-13T06:49:43Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 13 06:49:43 crc kubenswrapper[4971]: I1213 06:49:43.182030 4971 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/c7185dec-3240-42e8-9d90-6182a5e8b65c-metrics-certs\") pod \"network-metrics-daemon-x98k7\" (UID: \"c7185dec-3240-42e8-9d90-6182a5e8b65c\") " pod="openshift-multus/network-metrics-daemon-x98k7" Dec 13 06:49:43 crc kubenswrapper[4971]: E1213 06:49:43.182232 4971 secret.go:188] Couldn't get secret openshift-multus/metrics-daemon-secret: object "openshift-multus"/"metrics-daemon-secret" not registered Dec 13 06:49:43 crc kubenswrapper[4971]: E1213 06:49:43.182449 4971 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/c7185dec-3240-42e8-9d90-6182a5e8b65c-metrics-certs podName:c7185dec-3240-42e8-9d90-6182a5e8b65c nodeName:}" failed. No retries permitted until 2025-12-13 06:49:47.182424307 +0000 UTC m=+43.786833765 (durationBeforeRetry 4s). Error: MountVolume.SetUp failed for volume "metrics-certs" (UniqueName: "kubernetes.io/secret/c7185dec-3240-42e8-9d90-6182a5e8b65c-metrics-certs") pod "network-metrics-daemon-x98k7" (UID: "c7185dec-3240-42e8-9d90-6182a5e8b65c") : object "openshift-multus"/"metrics-daemon-secret" not registered Dec 13 06:49:43 crc kubenswrapper[4971]: I1213 06:49:43.252929 4971 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 13 06:49:43 crc kubenswrapper[4971]: I1213 06:49:43.252968 4971 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 13 06:49:43 crc kubenswrapper[4971]: I1213 06:49:43.252979 4971 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 13 06:49:43 crc kubenswrapper[4971]: I1213 06:49:43.252996 4971 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 13 06:49:43 crc kubenswrapper[4971]: I1213 06:49:43.253007 4971 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-13T06:49:43Z","lastTransitionTime":"2025-12-13T06:49:43Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 13 06:49:43 crc kubenswrapper[4971]: I1213 06:49:43.355631 4971 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 13 06:49:43 crc kubenswrapper[4971]: I1213 06:49:43.355859 4971 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 13 06:49:43 crc kubenswrapper[4971]: I1213 06:49:43.355922 4971 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 13 06:49:43 crc kubenswrapper[4971]: I1213 06:49:43.356020 4971 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 13 06:49:43 crc kubenswrapper[4971]: I1213 06:49:43.356078 4971 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-13T06:49:43Z","lastTransitionTime":"2025-12-13T06:49:43Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 13 06:49:43 crc kubenswrapper[4971]: I1213 06:49:43.458954 4971 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 13 06:49:43 crc kubenswrapper[4971]: I1213 06:49:43.458992 4971 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 13 06:49:43 crc kubenswrapper[4971]: I1213 06:49:43.459002 4971 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 13 06:49:43 crc kubenswrapper[4971]: I1213 06:49:43.459017 4971 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 13 06:49:43 crc kubenswrapper[4971]: I1213 06:49:43.459028 4971 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-13T06:49:43Z","lastTransitionTime":"2025-12-13T06:49:43Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 13 06:49:43 crc kubenswrapper[4971]: I1213 06:49:43.561432 4971 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 13 06:49:43 crc kubenswrapper[4971]: I1213 06:49:43.561467 4971 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 13 06:49:43 crc kubenswrapper[4971]: I1213 06:49:43.561478 4971 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 13 06:49:43 crc kubenswrapper[4971]: I1213 06:49:43.561496 4971 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 13 06:49:43 crc kubenswrapper[4971]: I1213 06:49:43.561507 4971 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-13T06:49:43Z","lastTransitionTime":"2025-12-13T06:49:43Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 13 06:49:43 crc kubenswrapper[4971]: I1213 06:49:43.663663 4971 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 13 06:49:43 crc kubenswrapper[4971]: I1213 06:49:43.663695 4971 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 13 06:49:43 crc kubenswrapper[4971]: I1213 06:49:43.663703 4971 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 13 06:49:43 crc kubenswrapper[4971]: I1213 06:49:43.663719 4971 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 13 06:49:43 crc kubenswrapper[4971]: I1213 06:49:43.663728 4971 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-13T06:49:43Z","lastTransitionTime":"2025-12-13T06:49:43Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 13 06:49:43 crc kubenswrapper[4971]: I1213 06:49:43.765999 4971 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 13 06:49:43 crc kubenswrapper[4971]: I1213 06:49:43.766539 4971 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 13 06:49:43 crc kubenswrapper[4971]: I1213 06:49:43.766665 4971 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 13 06:49:43 crc kubenswrapper[4971]: I1213 06:49:43.766738 4971 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 13 06:49:43 crc kubenswrapper[4971]: I1213 06:49:43.766820 4971 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-13T06:49:43Z","lastTransitionTime":"2025-12-13T06:49:43Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 13 06:49:43 crc kubenswrapper[4971]: I1213 06:49:43.768305 4971 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 13 06:49:43 crc kubenswrapper[4971]: I1213 06:49:43.768337 4971 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-x98k7" Dec 13 06:49:43 crc kubenswrapper[4971]: I1213 06:49:43.768366 4971 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 13 06:49:43 crc kubenswrapper[4971]: E1213 06:49:43.768420 4971 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 13 06:49:43 crc kubenswrapper[4971]: I1213 06:49:43.768461 4971 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 13 06:49:43 crc kubenswrapper[4971]: E1213 06:49:43.768584 4971 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 13 06:49:43 crc kubenswrapper[4971]: E1213 06:49:43.768708 4971 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-x98k7" podUID="c7185dec-3240-42e8-9d90-6182a5e8b65c" Dec 13 06:49:43 crc kubenswrapper[4971]: E1213 06:49:43.768809 4971 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 13 06:49:43 crc kubenswrapper[4971]: I1213 06:49:43.780428 4971 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-gkcd8" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"13f5c327-f95e-42b4-8783-218500c17694\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T06:49:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T06:49:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T06:49:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T06:49:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://77e846e36f1cb40d03471dc03de11b6c3da448baa407aad38093fae9a801cf92\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-13T06:49:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gvlhn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-13T06:49:27Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-gkcd8\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-13T06:49:43Z is after 2025-08-24T17:21:41Z" Dec 13 06:49:43 crc kubenswrapper[4971]: I1213 06:49:43.791090 4971 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-x98k7" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"c7185dec-3240-42e8-9d90-6182a5e8b65c\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T06:49:39Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T06:49:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T06:49:39Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T06:49:39Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bj2gn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bj2gn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-13T06:49:39Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-x98k7\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-13T06:49:43Z is after 2025-08-24T17:21:41Z" Dec 13 06:49:43 crc kubenswrapper[4971]: I1213 06:49:43.803945 4971 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"f9d31b5a-b770-4a15-95c8-35b49677c608\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T06:49:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T06:49:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T06:49:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T06:49:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T06:49:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c9baa042f9dada954faa57b5454cd6b0a12a297e56ddc192c52cff50e42a16fa\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-13T06:49:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://c788de86754f0bf37ee184672fc1dcb388ccebc246d4f15be4b5da661f725f8f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-13T06:49:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://760817511a9b4f96a68dc38d26631250eaab72e1dd8f123e653039b6b4e91f84\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-13T06:49:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b859be0c9feac856e9855c6a038b02b49b32fade97873c46d0bec0a32d8b5147\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-13T06:49:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-13T06:49:03Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-13T06:49:43Z is after 2025-08-24T17:21:41Z" Dec 13 06:49:43 crc kubenswrapper[4971]: I1213 06:49:43.816770 4971 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-ncchz" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"de914d21-a688-401e-a66e-861beb6336d9\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T06:49:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T06:49:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T06:49:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T06:49:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://244ffe3dc9d8cf799aff4c4d795bf41554c1e6bb6667fbb0c7df322fb5f96c68\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-13T06:49:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lqmjj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://0b7111af6c5a198d02bb02858837cfb4f3b20828d8d206cd18aa0ffa102d0cd4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://0b7111af6c5a198d02bb02858837cfb4f3b20828d8d206cd18aa0ffa102d0cd4\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-13T06:49:26Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-13T06:49:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lqmjj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b36777fe40e74ae1cecb7b73e204139db5279ba51cef4f1e8c4a4c491d8be403\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://b36777fe40e74ae1cecb7b73e204139db5279ba51cef4f1e8c4a4c491d8be403\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-13T06:49:28Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-13T06:49:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lqmjj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://0597a06cc663e416142311a16941d98297577693a1fe27201ea42c27496cace2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://0597a06cc663e416142311a16941d98297577693a1fe27201ea42c27496cace2\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-13T06:49:29Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-13T06:49:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lqmjj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://3f52c2e22a75c010993ca5809a3a9d1304758d4cb6e83691ef1af15a4b514ac4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://3f52c2e22a75c010993ca5809a3a9d1304758d4cb6e83691ef1af15a4b514ac4\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-13T06:49:30Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-13T06:49:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lqmjj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://9e21ca708c4c4c1228b417cece51299e0792729b7c78884bd18b2dda8fe8355c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://9e21ca708c4c4c1228b417cece51299e0792729b7c78884bd18b2dda8fe8355c\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-13T06:49:31Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-13T06:49:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lqmjj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://74552159f12c87cdda09147692b400ccac0d4e43c1ed72fe8462c4249ad69c7b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://74552159f12c87cdda09147692b400ccac0d4e43c1ed72fe8462c4249ad69c7b\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-13T06:49:32Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-13T06:49:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lqmjj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-13T06:49:25Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-ncchz\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-13T06:49:43Z is after 2025-08-24T17:21:41Z" Dec 13 06:49:43 crc kubenswrapper[4971]: I1213 06:49:43.826744 4971 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-5rzh5" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"6bf4591d-2483-4132-9672-d0fd6c995968\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T06:49:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T06:49:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T06:49:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T06:49:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://50e2cb14e4ffb603acb06e2281615a99f06182d345c4e70f2446fa2077033bbb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-13T06:49:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dfxqn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-13T06:49:25Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-5rzh5\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-13T06:49:43Z is after 2025-08-24T17:21:41Z" Dec 13 06:49:43 crc kubenswrapper[4971]: I1213 06:49:43.835893 4971 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-z572s" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"b4146d3d-4034-49a9-bd37-164856255137\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T06:49:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T06:49:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T06:49:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T06:49:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://46fd4e1070835cfbf7f35d497eff7f047473fc4048ee07812925df5ebdf55342\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-13T06:49:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-j6l9b\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://dc79a0968f85723e6ad0dc269c8da1f7472adb1bb2b87edd017ee2295bfbf89c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-13T06:49:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-j6l9b\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-13T06:49:38Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-z572s\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-13T06:49:43Z is after 2025-08-24T17:21:41Z" Dec 13 06:49:43 crc kubenswrapper[4971]: I1213 06:49:43.859397 4971 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"1ed9f82e-fe5c-4b76-a619-62e5abd88d3b\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T06:49:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T06:49:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T06:49:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T06:49:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T06:49:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e457329a78a2d24337f994e8a0a1f75a2a667244aa75eb84613d1e7138364aac\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-13T06:49:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://eb014de8140deef097d9bd94f3b9fa4ab3558bc4449bef84b83bf62bfabf5836\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-13T06:49:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://2c438cc74d555f7c9db4569d1c56b78b1088ed8105697d2c0f31e6c5136f3dd0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-13T06:49:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://aa6fa6e4340dcd066252df3c56d79a3a39ed531d5eea8d1d5941d6952368d3af\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-13T06:49:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://c4471b98854c52d5cd084d0ba3e64e5b08e9518ec3c372cce2c875a440522457\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-13T06:49:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7d2746335c98dae3a0f43024d34cb62ebcb42a01c8f9ce18d52b29d4c45598e5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://7d2746335c98dae3a0f43024d34cb62ebcb42a01c8f9ce18d52b29d4c45598e5\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-13T06:49:04Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-13T06:49:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://bc0e77d10a738128bea82122496bc42235fc1688f8907756d93e61b651902b57\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://bc0e77d10a738128bea82122496bc42235fc1688f8907756d93e61b651902b57\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-13T06:49:05Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-13T06:49:05Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://f821e418cba5e832da8839268bcbbd9630bdf319a212d5382bfa2c4b0b73a268\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://f821e418cba5e832da8839268bcbbd9630bdf319a212d5382bfa2c4b0b73a268\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-13T06:49:07Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-13T06:49:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-13T06:49:03Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-13T06:49:43Z is after 2025-08-24T17:21:41Z" Dec 13 06:49:43 crc kubenswrapper[4971]: I1213 06:49:43.868303 4971 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 13 06:49:43 crc kubenswrapper[4971]: I1213 06:49:43.868334 4971 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 13 06:49:43 crc kubenswrapper[4971]: I1213 06:49:43.868342 4971 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 13 06:49:43 crc kubenswrapper[4971]: I1213 06:49:43.868356 4971 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 13 06:49:43 crc kubenswrapper[4971]: I1213 06:49:43.868364 4971 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-13T06:49:43Z","lastTransitionTime":"2025-12-13T06:49:43Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 13 06:49:43 crc kubenswrapper[4971]: I1213 06:49:43.874114 4971 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-13T06:49:24Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-13T06:49:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-13T06:49:24Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4328cd6a8464ed468229bbfbc8265f969427916d72ebb92237dbe68efe70fa47\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-13T06:49:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://9d715ca82d9b5a594e41f2b98da6f4277da0e4e1dec114969f901c00f6d7553d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-13T06:49:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-13T06:49:43Z is after 2025-08-24T17:21:41Z" Dec 13 06:49:43 crc kubenswrapper[4971]: I1213 06:49:43.888453 4971 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3ad3268a-a982-428c-9c06-4aaeddd372d9\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T06:49:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T06:49:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T06:49:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T06:49:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T06:49:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3716afdf2bc693be80a09fc9f7727d334f435caf1d4c8ec4811b3a1f1652cdfe\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-13T06:49:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://30df3abf8c79db8af52c6a918f73e1ce410ddafcf341636424c267e1eeb195a7\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-13T06:49:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://ccdf65f6d2b1eb3647a03ba37c305c32e34addfdf79ff937c8dddaa3e5d5b5c6\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-13T06:49:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://770b73dd0c2b6f38744e2c1cd6aa186a69bb87d1cd015a603ad2e33be06a8f6b\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://b3c1bab3d53f975b2e24582ce2771bc26d600e0dd7c30339dcfd18a071162ad4\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-12-13T06:49:23Z\\\",\\\"message\\\":\\\"file observer\\\\nW1213 06:49:23.283403 1 builder.go:272] unable to get owner reference (falling back to namespace): pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\nI1213 06:49:23.284336 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1213 06:49:23.286978 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-981773671/tls.crt::/tmp/serving-cert-981773671/tls.key\\\\\\\"\\\\nI1213 06:49:23.775072 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1213 06:49:23.778615 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1213 06:49:23.778645 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1213 06:49:23.778680 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1213 06:49:23.778691 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1213 06:49:23.785058 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1213 06:49:23.786150 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1213 06:49:23.786196 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1213 06:49:23.786223 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1213 06:49:23.786246 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1213 06:49:23.786269 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1213 06:49:23.786291 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI1213 06:49:23.785608 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF1213 06:49:23.787563 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-13T06:49:08Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-13T06:49:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://9acd07066deed483bd7fa61ee45fda09539d110717a134a56827f871186f92fc\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-13T06:49:07Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://38c91cfbaecfdbd3e3b663422a91ab4294610afb6bbf9ebdb77db24e2b76c682\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://38c91cfbaecfdbd3e3b663422a91ab4294610afb6bbf9ebdb77db24e2b76c682\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-13T06:49:05Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-13T06:49:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-13T06:49:03Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-13T06:49:43Z is after 2025-08-24T17:21:41Z" Dec 13 06:49:43 crc kubenswrapper[4971]: I1213 06:49:43.899912 4971 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-13T06:49:23Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-13T06:49:23Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-13T06:49:43Z is after 2025-08-24T17:21:41Z" Dec 13 06:49:43 crc kubenswrapper[4971]: I1213 06:49:43.910510 4971 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-13T06:49:23Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-13T06:49:23Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-13T06:49:43Z is after 2025-08-24T17:21:41Z" Dec 13 06:49:43 crc kubenswrapper[4971]: I1213 06:49:43.921764 4971 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-82xjz" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e0c4af12-edda-41d7-9b44-f87396174ef2\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T06:49:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T06:49:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T06:49:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T06:49:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://0f4577d9e37f860813e221f2e208177d8d383fd04f673ce357a0a405ea29d1e6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-13T06:49:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cnxsq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://7e0d2020e62f1a277cff7047e1907dc747ed393a1f6c983e1ee258bd156c16c7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-13T06:49:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cnxsq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-13T06:49:25Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-82xjz\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-13T06:49:43Z is after 2025-08-24T17:21:41Z" Dec 13 06:49:43 crc kubenswrapper[4971]: I1213 06:49:43.934469 4971 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-wwvpl" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"dddf744b-2122-49ad-a980-105c09636e0f\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T06:49:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T06:49:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T06:49:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T06:49:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4676d897c7fc668587faec563f3551a056557879cee3e865192d2bd28edf62fc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-13T06:49:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-czz9v\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-13T06:49:25Z\\\"}}\" for pod \"openshift-multus\"/\"multus-wwvpl\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-13T06:49:43Z is after 2025-08-24T17:21:41Z" Dec 13 06:49:43 crc kubenswrapper[4971]: I1213 06:49:43.951132 4971 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-zlcg5" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"736c6b1a-93dd-4f71-81b6-1c43fcf5f556\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T06:49:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T06:49:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T06:49:25Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T06:49:25Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://325223543b3c7db417f769533559a1ca2b23a723494e1885b0fd0480b5a6399f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-13T06:49:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vzmsr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://f22a4ec756119c4f791fb5105cf1f31df3e548b2b07b1f909a62802e9b6035ae\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-13T06:49:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vzmsr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://f6073140d9f86929b6d4bd2960ede0617a5c9620b269bbab043bae37e0d389f5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-13T06:49:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vzmsr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://693a2ab434c250cb140a866e8068b156df9191414fed1d63bee919a2528ea6e5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-13T06:49:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vzmsr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://219620f51474e60b3e1f880c0972eb612dffc4bad24462bbba2c093ae817d852\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-13T06:49:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vzmsr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://f137f4b441c2402093ab430a9689a46221c4fb4bc70d99fb992f62b0a989c465\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-13T06:49:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vzmsr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://d950a29b0f6f47ae1cc90c132adb059b5aff889a0a38de1d1296e0f083da5e51\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://451040d83e76225457d750a4aa532aa2772decb89719e53fe001bf74aaa8e73b\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-12-13T06:49:35Z\\\",\\\"message\\\":\\\"/pkg/client/informers/externalversions/factory.go:141\\\\nI1213 06:49:35.122568 6277 handler.go:190] Sending *v1.Namespace event handler 1 for removal\\\\nI1213 06:49:35.122615 6277 handler.go:190] Sending *v1.Namespace event handler 5 for removal\\\\nI1213 06:49:35.122659 6277 handler.go:190] Sending *v1.Node event handler 7 for removal\\\\nI1213 06:49:35.122681 6277 handler.go:190] Sending *v1.Node event handler 2 for removal\\\\nI1213 06:49:35.122723 6277 handler.go:190] Sending *v1.NetworkPolicy event handler 4 for removal\\\\nI1213 06:49:35.122771 6277 handler.go:190] Sending *v1.EgressIP event handler 8 for removal\\\\nI1213 06:49:35.122815 6277 handler.go:190] Sending *v1.EgressFirewall event handler 9 for removal\\\\nI1213 06:49:35.122852 6277 factory.go:656] Stopping watch factory\\\\nI1213 06:49:35.122935 6277 ovnkube.go:599] Stopped ovnkube\\\\nI1213 06:49:35.122983 6277 handler.go:208] Removed *v1.Namespace event handler 1\\\\nI1213 06:49:35.123009 6277 handler.go:208] Removed *v1.Namespace event handler 5\\\\nI1213 06:49:35.123033 6277 handler.go:208] Removed *v1.Node event handler 7\\\\nI1213 06:49:35.123056 6277 handler.go:208] Removed *v1.Node event handler 2\\\\nI1213 06:49:35.123079 6277 handler.go:208] Removed *v1.NetworkPolicy event handler 4\\\\nI1213 06:49:35.123102 6277 handler.go:208] Removed *v1.EgressIP event handler 8\\\\nI1213 06:49:3\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-13T06:49:32Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d950a29b0f6f47ae1cc90c132adb059b5aff889a0a38de1d1296e0f083da5e51\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-12-13T06:49:37Z\\\",\\\"message\\\":\\\" ovnkube.go:137] failed to run ovnkube: [failed to start network controller: failed to start default network controller: unable to create admin network policy controller, err: could not add Event Handler for anpInformer during admin network policy controller initialization, handler {0x1fcc6e0 0x1fcc3c0 0x1fcc360} was not added to shared informer because it has stopped already, failed to start node network controller: failed to start default node network controller: failed to set node crc annotations: Internal error occurred: failed calling webhook \\\\\\\"node.network-node-identity.openshift.io\\\\\\\": failed to call webhook: Post \\\\\\\"https://127.0.0.1:9743/node?timeout=10s\\\\\\\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-13T06:49:36Z is after 2025-08-24T17:21:41Z]\\\\nI1213 06:49:36.767920 6409 model_client.go:382] Update operations generated as: [{Op:update Table:Logical_Switch_Port Row:map[addresses:{GoSet:[0a:58:0a:d9:00:04 10.217.0.4]} options:{GoMap:map[iface-id-ver:3b6479f0-333b-4a96-9adf-2099afdc2447 requested-chassis:crc]} port_security:{GoSet:[0a:58:0a:d9:00:04 10.217.0.4]}] Rows:[] Columns:[] Mutations:[] Timeout:\\\\u003cnil\\\\u003e Where:[where column _uuid == {61897e97-c771-4738-8709-09636387cb00}] Until: Durable:\\\\u003cnil\\\\u003e Comment:\\\\u003cnil\\\\u003e Lock:\\\\u003cnil\\\\u003e UUID: UUIDName:}]\\\\nI1213 06:49:36.767917 6409 model_clien\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-13T06:49:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vzmsr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://868049c6a3a4ebe4036d1d22e02ba6b76bced2ee7e269bb228accc4a0636678a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-13T06:49:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vzmsr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://451b22df6973ddc8fa2209a65d64d9e41b33eb231bb4adfab88cf0e5943bb246\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://451b22df6973ddc8fa2209a65d64d9e41b33eb231bb4adfab88cf0e5943bb246\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-13T06:49:26Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-13T06:49:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vzmsr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-13T06:49:25Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-zlcg5\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-13T06:49:43Z is after 2025-08-24T17:21:41Z" Dec 13 06:49:43 crc kubenswrapper[4971]: I1213 06:49:43.963892 4971 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-13T06:49:24Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-13T06:49:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-13T06:49:24Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d0b2ecf0d5b22f23217f42f259ae7ffb6359555b22c30eaebf667eaf254b3bab\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-13T06:49:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-13T06:49:43Z is after 2025-08-24T17:21:41Z" Dec 13 06:49:43 crc kubenswrapper[4971]: I1213 06:49:43.969928 4971 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 13 06:49:43 crc kubenswrapper[4971]: I1213 06:49:43.969965 4971 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 13 06:49:43 crc kubenswrapper[4971]: I1213 06:49:43.969978 4971 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 13 06:49:43 crc kubenswrapper[4971]: I1213 06:49:43.969994 4971 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 13 06:49:43 crc kubenswrapper[4971]: I1213 06:49:43.970007 4971 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-13T06:49:43Z","lastTransitionTime":"2025-12-13T06:49:43Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 13 06:49:43 crc kubenswrapper[4971]: I1213 06:49:43.974671 4971 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-13T06:49:23Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-13T06:49:23Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-13T06:49:43Z is after 2025-08-24T17:21:41Z" Dec 13 06:49:43 crc kubenswrapper[4971]: I1213 06:49:43.986408 4971 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-13T06:49:26Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-13T06:49:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-13T06:49:26Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9c8a0d1db0e754462d805e82daf5f52ffd5457507a30456535ec59e096c9d092\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-13T06:49:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-13T06:49:43Z is after 2025-08-24T17:21:41Z" Dec 13 06:49:44 crc kubenswrapper[4971]: I1213 06:49:44.072222 4971 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 13 06:49:44 crc kubenswrapper[4971]: I1213 06:49:44.072278 4971 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 13 06:49:44 crc kubenswrapper[4971]: I1213 06:49:44.072288 4971 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 13 06:49:44 crc kubenswrapper[4971]: I1213 06:49:44.072304 4971 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 13 06:49:44 crc kubenswrapper[4971]: I1213 06:49:44.072315 4971 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-13T06:49:44Z","lastTransitionTime":"2025-12-13T06:49:44Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 13 06:49:44 crc kubenswrapper[4971]: I1213 06:49:44.174998 4971 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 13 06:49:44 crc kubenswrapper[4971]: I1213 06:49:44.175048 4971 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 13 06:49:44 crc kubenswrapper[4971]: I1213 06:49:44.175060 4971 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 13 06:49:44 crc kubenswrapper[4971]: I1213 06:49:44.175077 4971 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 13 06:49:44 crc kubenswrapper[4971]: I1213 06:49:44.175087 4971 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-13T06:49:44Z","lastTransitionTime":"2025-12-13T06:49:44Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 13 06:49:44 crc kubenswrapper[4971]: I1213 06:49:44.278243 4971 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 13 06:49:44 crc kubenswrapper[4971]: I1213 06:49:44.278317 4971 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 13 06:49:44 crc kubenswrapper[4971]: I1213 06:49:44.278337 4971 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 13 06:49:44 crc kubenswrapper[4971]: I1213 06:49:44.278362 4971 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 13 06:49:44 crc kubenswrapper[4971]: I1213 06:49:44.278384 4971 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-13T06:49:44Z","lastTransitionTime":"2025-12-13T06:49:44Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 13 06:49:44 crc kubenswrapper[4971]: I1213 06:49:44.381725 4971 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 13 06:49:44 crc kubenswrapper[4971]: I1213 06:49:44.381776 4971 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 13 06:49:44 crc kubenswrapper[4971]: I1213 06:49:44.381789 4971 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 13 06:49:44 crc kubenswrapper[4971]: I1213 06:49:44.381810 4971 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 13 06:49:44 crc kubenswrapper[4971]: I1213 06:49:44.381828 4971 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-13T06:49:44Z","lastTransitionTime":"2025-12-13T06:49:44Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 13 06:49:44 crc kubenswrapper[4971]: I1213 06:49:44.483857 4971 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 13 06:49:44 crc kubenswrapper[4971]: I1213 06:49:44.483974 4971 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 13 06:49:44 crc kubenswrapper[4971]: I1213 06:49:44.483999 4971 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 13 06:49:44 crc kubenswrapper[4971]: I1213 06:49:44.484030 4971 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 13 06:49:44 crc kubenswrapper[4971]: I1213 06:49:44.484051 4971 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-13T06:49:44Z","lastTransitionTime":"2025-12-13T06:49:44Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 13 06:49:44 crc kubenswrapper[4971]: I1213 06:49:44.586686 4971 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 13 06:49:44 crc kubenswrapper[4971]: I1213 06:49:44.586738 4971 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 13 06:49:44 crc kubenswrapper[4971]: I1213 06:49:44.586750 4971 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 13 06:49:44 crc kubenswrapper[4971]: I1213 06:49:44.586770 4971 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 13 06:49:44 crc kubenswrapper[4971]: I1213 06:49:44.586782 4971 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-13T06:49:44Z","lastTransitionTime":"2025-12-13T06:49:44Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 13 06:49:44 crc kubenswrapper[4971]: I1213 06:49:44.689199 4971 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 13 06:49:44 crc kubenswrapper[4971]: I1213 06:49:44.689254 4971 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 13 06:49:44 crc kubenswrapper[4971]: I1213 06:49:44.689268 4971 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 13 06:49:44 crc kubenswrapper[4971]: I1213 06:49:44.689286 4971 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 13 06:49:44 crc kubenswrapper[4971]: I1213 06:49:44.689303 4971 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-13T06:49:44Z","lastTransitionTime":"2025-12-13T06:49:44Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 13 06:49:44 crc kubenswrapper[4971]: I1213 06:49:44.791804 4971 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 13 06:49:44 crc kubenswrapper[4971]: I1213 06:49:44.791856 4971 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 13 06:49:44 crc kubenswrapper[4971]: I1213 06:49:44.791869 4971 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 13 06:49:44 crc kubenswrapper[4971]: I1213 06:49:44.791884 4971 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 13 06:49:44 crc kubenswrapper[4971]: I1213 06:49:44.791895 4971 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-13T06:49:44Z","lastTransitionTime":"2025-12-13T06:49:44Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 13 06:49:44 crc kubenswrapper[4971]: I1213 06:49:44.894307 4971 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 13 06:49:44 crc kubenswrapper[4971]: I1213 06:49:44.894357 4971 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 13 06:49:44 crc kubenswrapper[4971]: I1213 06:49:44.894367 4971 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 13 06:49:44 crc kubenswrapper[4971]: I1213 06:49:44.894387 4971 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 13 06:49:44 crc kubenswrapper[4971]: I1213 06:49:44.894400 4971 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-13T06:49:44Z","lastTransitionTime":"2025-12-13T06:49:44Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 13 06:49:44 crc kubenswrapper[4971]: I1213 06:49:44.979656 4971 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-ovn-kubernetes/ovnkube-node-zlcg5" Dec 13 06:49:44 crc kubenswrapper[4971]: I1213 06:49:44.980630 4971 scope.go:117] "RemoveContainer" containerID="d950a29b0f6f47ae1cc90c132adb059b5aff889a0a38de1d1296e0f083da5e51" Dec 13 06:49:44 crc kubenswrapper[4971]: E1213 06:49:44.980814 4971 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"ovnkube-controller\" with CrashLoopBackOff: \"back-off 10s restarting failed container=ovnkube-controller pod=ovnkube-node-zlcg5_openshift-ovn-kubernetes(736c6b1a-93dd-4f71-81b6-1c43fcf5f556)\"" pod="openshift-ovn-kubernetes/ovnkube-node-zlcg5" podUID="736c6b1a-93dd-4f71-81b6-1c43fcf5f556" Dec 13 06:49:44 crc kubenswrapper[4971]: I1213 06:49:44.994555 4971 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"f9d31b5a-b770-4a15-95c8-35b49677c608\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T06:49:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T06:49:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T06:49:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T06:49:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T06:49:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c9baa042f9dada954faa57b5454cd6b0a12a297e56ddc192c52cff50e42a16fa\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-13T06:49:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://c788de86754f0bf37ee184672fc1dcb388ccebc246d4f15be4b5da661f725f8f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-13T06:49:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://760817511a9b4f96a68dc38d26631250eaab72e1dd8f123e653039b6b4e91f84\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-13T06:49:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b859be0c9feac856e9855c6a038b02b49b32fade97873c46d0bec0a32d8b5147\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-13T06:49:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-13T06:49:03Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-13T06:49:44Z is after 2025-08-24T17:21:41Z" Dec 13 06:49:44 crc kubenswrapper[4971]: I1213 06:49:44.996438 4971 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 13 06:49:44 crc kubenswrapper[4971]: I1213 06:49:44.996487 4971 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 13 06:49:44 crc kubenswrapper[4971]: I1213 06:49:44.996499 4971 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 13 06:49:44 crc kubenswrapper[4971]: I1213 06:49:44.996536 4971 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 13 06:49:44 crc kubenswrapper[4971]: I1213 06:49:44.996550 4971 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-13T06:49:44Z","lastTransitionTime":"2025-12-13T06:49:44Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 13 06:49:45 crc kubenswrapper[4971]: I1213 06:49:45.011168 4971 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-ncchz" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"de914d21-a688-401e-a66e-861beb6336d9\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T06:49:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T06:49:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T06:49:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T06:49:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://244ffe3dc9d8cf799aff4c4d795bf41554c1e6bb6667fbb0c7df322fb5f96c68\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-13T06:49:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lqmjj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://0b7111af6c5a198d02bb02858837cfb4f3b20828d8d206cd18aa0ffa102d0cd4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://0b7111af6c5a198d02bb02858837cfb4f3b20828d8d206cd18aa0ffa102d0cd4\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-13T06:49:26Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-13T06:49:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lqmjj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b36777fe40e74ae1cecb7b73e204139db5279ba51cef4f1e8c4a4c491d8be403\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://b36777fe40e74ae1cecb7b73e204139db5279ba51cef4f1e8c4a4c491d8be403\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-13T06:49:28Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-13T06:49:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lqmjj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://0597a06cc663e416142311a16941d98297577693a1fe27201ea42c27496cace2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://0597a06cc663e416142311a16941d98297577693a1fe27201ea42c27496cace2\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-13T06:49:29Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-13T06:49:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lqmjj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://3f52c2e22a75c010993ca5809a3a9d1304758d4cb6e83691ef1af15a4b514ac4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://3f52c2e22a75c010993ca5809a3a9d1304758d4cb6e83691ef1af15a4b514ac4\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-13T06:49:30Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-13T06:49:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lqmjj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://9e21ca708c4c4c1228b417cece51299e0792729b7c78884bd18b2dda8fe8355c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://9e21ca708c4c4c1228b417cece51299e0792729b7c78884bd18b2dda8fe8355c\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-13T06:49:31Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-13T06:49:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lqmjj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://74552159f12c87cdda09147692b400ccac0d4e43c1ed72fe8462c4249ad69c7b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://74552159f12c87cdda09147692b400ccac0d4e43c1ed72fe8462c4249ad69c7b\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-13T06:49:32Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-13T06:49:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lqmjj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-13T06:49:25Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-ncchz\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-13T06:49:45Z is after 2025-08-24T17:21:41Z" Dec 13 06:49:45 crc kubenswrapper[4971]: I1213 06:49:45.026704 4971 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-gkcd8" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"13f5c327-f95e-42b4-8783-218500c17694\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T06:49:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T06:49:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T06:49:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T06:49:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://77e846e36f1cb40d03471dc03de11b6c3da448baa407aad38093fae9a801cf92\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-13T06:49:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gvlhn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-13T06:49:27Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-gkcd8\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-13T06:49:45Z is after 2025-08-24T17:21:41Z" Dec 13 06:49:45 crc kubenswrapper[4971]: I1213 06:49:45.037322 4971 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-x98k7" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"c7185dec-3240-42e8-9d90-6182a5e8b65c\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T06:49:39Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T06:49:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T06:49:39Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T06:49:39Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bj2gn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bj2gn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-13T06:49:39Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-x98k7\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-13T06:49:45Z is after 2025-08-24T17:21:41Z" Dec 13 06:49:45 crc kubenswrapper[4971]: I1213 06:49:45.056481 4971 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"1ed9f82e-fe5c-4b76-a619-62e5abd88d3b\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T06:49:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T06:49:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T06:49:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T06:49:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T06:49:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e457329a78a2d24337f994e8a0a1f75a2a667244aa75eb84613d1e7138364aac\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-13T06:49:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://eb014de8140deef097d9bd94f3b9fa4ab3558bc4449bef84b83bf62bfabf5836\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-13T06:49:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://2c438cc74d555f7c9db4569d1c56b78b1088ed8105697d2c0f31e6c5136f3dd0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-13T06:49:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://aa6fa6e4340dcd066252df3c56d79a3a39ed531d5eea8d1d5941d6952368d3af\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-13T06:49:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://c4471b98854c52d5cd084d0ba3e64e5b08e9518ec3c372cce2c875a440522457\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-13T06:49:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7d2746335c98dae3a0f43024d34cb62ebcb42a01c8f9ce18d52b29d4c45598e5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://7d2746335c98dae3a0f43024d34cb62ebcb42a01c8f9ce18d52b29d4c45598e5\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-13T06:49:04Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-13T06:49:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://bc0e77d10a738128bea82122496bc42235fc1688f8907756d93e61b651902b57\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://bc0e77d10a738128bea82122496bc42235fc1688f8907756d93e61b651902b57\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-13T06:49:05Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-13T06:49:05Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://f821e418cba5e832da8839268bcbbd9630bdf319a212d5382bfa2c4b0b73a268\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://f821e418cba5e832da8839268bcbbd9630bdf319a212d5382bfa2c4b0b73a268\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-13T06:49:07Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-13T06:49:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-13T06:49:03Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-13T06:49:45Z is after 2025-08-24T17:21:41Z" Dec 13 06:49:45 crc kubenswrapper[4971]: I1213 06:49:45.070355 4971 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-13T06:49:24Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-13T06:49:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-13T06:49:24Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4328cd6a8464ed468229bbfbc8265f969427916d72ebb92237dbe68efe70fa47\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-13T06:49:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://9d715ca82d9b5a594e41f2b98da6f4277da0e4e1dec114969f901c00f6d7553d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-13T06:49:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-13T06:49:45Z is after 2025-08-24T17:21:41Z" Dec 13 06:49:45 crc kubenswrapper[4971]: I1213 06:49:45.081421 4971 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-5rzh5" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"6bf4591d-2483-4132-9672-d0fd6c995968\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T06:49:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T06:49:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T06:49:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T06:49:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://50e2cb14e4ffb603acb06e2281615a99f06182d345c4e70f2446fa2077033bbb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-13T06:49:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dfxqn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-13T06:49:25Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-5rzh5\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-13T06:49:45Z is after 2025-08-24T17:21:41Z" Dec 13 06:49:45 crc kubenswrapper[4971]: I1213 06:49:45.093746 4971 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-z572s" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"b4146d3d-4034-49a9-bd37-164856255137\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T06:49:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T06:49:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T06:49:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T06:49:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://46fd4e1070835cfbf7f35d497eff7f047473fc4048ee07812925df5ebdf55342\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-13T06:49:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-j6l9b\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://dc79a0968f85723e6ad0dc269c8da1f7472adb1bb2b87edd017ee2295bfbf89c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-13T06:49:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-j6l9b\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-13T06:49:38Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-z572s\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-13T06:49:45Z is after 2025-08-24T17:21:41Z" Dec 13 06:49:45 crc kubenswrapper[4971]: I1213 06:49:45.099803 4971 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 13 06:49:45 crc kubenswrapper[4971]: I1213 06:49:45.099841 4971 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 13 06:49:45 crc kubenswrapper[4971]: I1213 06:49:45.099851 4971 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 13 06:49:45 crc kubenswrapper[4971]: I1213 06:49:45.099867 4971 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 13 06:49:45 crc kubenswrapper[4971]: I1213 06:49:45.099877 4971 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-13T06:49:45Z","lastTransitionTime":"2025-12-13T06:49:45Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 13 06:49:45 crc kubenswrapper[4971]: I1213 06:49:45.107664 4971 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3ad3268a-a982-428c-9c06-4aaeddd372d9\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T06:49:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T06:49:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T06:49:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T06:49:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T06:49:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3716afdf2bc693be80a09fc9f7727d334f435caf1d4c8ec4811b3a1f1652cdfe\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-13T06:49:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://30df3abf8c79db8af52c6a918f73e1ce410ddafcf341636424c267e1eeb195a7\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-13T06:49:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://ccdf65f6d2b1eb3647a03ba37c305c32e34addfdf79ff937c8dddaa3e5d5b5c6\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-13T06:49:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://770b73dd0c2b6f38744e2c1cd6aa186a69bb87d1cd015a603ad2e33be06a8f6b\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://b3c1bab3d53f975b2e24582ce2771bc26d600e0dd7c30339dcfd18a071162ad4\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-12-13T06:49:23Z\\\",\\\"message\\\":\\\"file observer\\\\nW1213 06:49:23.283403 1 builder.go:272] unable to get owner reference (falling back to namespace): pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\nI1213 06:49:23.284336 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1213 06:49:23.286978 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-981773671/tls.crt::/tmp/serving-cert-981773671/tls.key\\\\\\\"\\\\nI1213 06:49:23.775072 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1213 06:49:23.778615 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1213 06:49:23.778645 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1213 06:49:23.778680 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1213 06:49:23.778691 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1213 06:49:23.785058 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1213 06:49:23.786150 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1213 06:49:23.786196 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1213 06:49:23.786223 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1213 06:49:23.786246 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1213 06:49:23.786269 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1213 06:49:23.786291 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI1213 06:49:23.785608 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF1213 06:49:23.787563 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-13T06:49:08Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-13T06:49:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://9acd07066deed483bd7fa61ee45fda09539d110717a134a56827f871186f92fc\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-13T06:49:07Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://38c91cfbaecfdbd3e3b663422a91ab4294610afb6bbf9ebdb77db24e2b76c682\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://38c91cfbaecfdbd3e3b663422a91ab4294610afb6bbf9ebdb77db24e2b76c682\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-13T06:49:05Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-13T06:49:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-13T06:49:03Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-13T06:49:45Z is after 2025-08-24T17:21:41Z" Dec 13 06:49:45 crc kubenswrapper[4971]: I1213 06:49:45.120643 4971 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-13T06:49:23Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-13T06:49:23Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-13T06:49:45Z is after 2025-08-24T17:21:41Z" Dec 13 06:49:45 crc kubenswrapper[4971]: I1213 06:49:45.131711 4971 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-13T06:49:23Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-13T06:49:23Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-13T06:49:45Z is after 2025-08-24T17:21:41Z" Dec 13 06:49:45 crc kubenswrapper[4971]: I1213 06:49:45.149883 4971 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-zlcg5" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"736c6b1a-93dd-4f71-81b6-1c43fcf5f556\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T06:49:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T06:49:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T06:49:25Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T06:49:25Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://325223543b3c7db417f769533559a1ca2b23a723494e1885b0fd0480b5a6399f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-13T06:49:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vzmsr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://f22a4ec756119c4f791fb5105cf1f31df3e548b2b07b1f909a62802e9b6035ae\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-13T06:49:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vzmsr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://f6073140d9f86929b6d4bd2960ede0617a5c9620b269bbab043bae37e0d389f5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-13T06:49:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vzmsr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://693a2ab434c250cb140a866e8068b156df9191414fed1d63bee919a2528ea6e5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-13T06:49:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vzmsr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://219620f51474e60b3e1f880c0972eb612dffc4bad24462bbba2c093ae817d852\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-13T06:49:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vzmsr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://f137f4b441c2402093ab430a9689a46221c4fb4bc70d99fb992f62b0a989c465\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-13T06:49:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vzmsr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://d950a29b0f6f47ae1cc90c132adb059b5aff889a0a38de1d1296e0f083da5e51\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d950a29b0f6f47ae1cc90c132adb059b5aff889a0a38de1d1296e0f083da5e51\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-12-13T06:49:37Z\\\",\\\"message\\\":\\\" ovnkube.go:137] failed to run ovnkube: [failed to start network controller: failed to start default network controller: unable to create admin network policy controller, err: could not add Event Handler for anpInformer during admin network policy controller initialization, handler {0x1fcc6e0 0x1fcc3c0 0x1fcc360} was not added to shared informer because it has stopped already, failed to start node network controller: failed to start default node network controller: failed to set node crc annotations: Internal error occurred: failed calling webhook \\\\\\\"node.network-node-identity.openshift.io\\\\\\\": failed to call webhook: Post \\\\\\\"https://127.0.0.1:9743/node?timeout=10s\\\\\\\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-13T06:49:36Z is after 2025-08-24T17:21:41Z]\\\\nI1213 06:49:36.767920 6409 model_client.go:382] Update operations generated as: [{Op:update Table:Logical_Switch_Port Row:map[addresses:{GoSet:[0a:58:0a:d9:00:04 10.217.0.4]} options:{GoMap:map[iface-id-ver:3b6479f0-333b-4a96-9adf-2099afdc2447 requested-chassis:crc]} port_security:{GoSet:[0a:58:0a:d9:00:04 10.217.0.4]}] Rows:[] Columns:[] Mutations:[] Timeout:\\\\u003cnil\\\\u003e Where:[where column _uuid == {61897e97-c771-4738-8709-09636387cb00}] Until: Durable:\\\\u003cnil\\\\u003e Comment:\\\\u003cnil\\\\u003e Lock:\\\\u003cnil\\\\u003e UUID: UUIDName:}]\\\\nI1213 06:49:36.767917 6409 model_clien\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-13T06:49:36Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 10s restarting failed container=ovnkube-controller pod=ovnkube-node-zlcg5_openshift-ovn-kubernetes(736c6b1a-93dd-4f71-81b6-1c43fcf5f556)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vzmsr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://868049c6a3a4ebe4036d1d22e02ba6b76bced2ee7e269bb228accc4a0636678a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-13T06:49:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vzmsr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://451b22df6973ddc8fa2209a65d64d9e41b33eb231bb4adfab88cf0e5943bb246\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://451b22df6973ddc8fa2209a65d64d9e41b33eb231bb4adfab88cf0e5943bb246\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-13T06:49:26Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-13T06:49:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vzmsr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-13T06:49:25Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-zlcg5\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-13T06:49:45Z is after 2025-08-24T17:21:41Z" Dec 13 06:49:45 crc kubenswrapper[4971]: I1213 06:49:45.162561 4971 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-13T06:49:24Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-13T06:49:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-13T06:49:24Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d0b2ecf0d5b22f23217f42f259ae7ffb6359555b22c30eaebf667eaf254b3bab\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-13T06:49:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-13T06:49:45Z is after 2025-08-24T17:21:41Z" Dec 13 06:49:45 crc kubenswrapper[4971]: I1213 06:49:45.173604 4971 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-13T06:49:23Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-13T06:49:23Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-13T06:49:45Z is after 2025-08-24T17:21:41Z" Dec 13 06:49:45 crc kubenswrapper[4971]: I1213 06:49:45.183825 4971 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-13T06:49:26Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-13T06:49:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-13T06:49:26Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9c8a0d1db0e754462d805e82daf5f52ffd5457507a30456535ec59e096c9d092\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-13T06:49:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-13T06:49:45Z is after 2025-08-24T17:21:41Z" Dec 13 06:49:45 crc kubenswrapper[4971]: I1213 06:49:45.193195 4971 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-82xjz" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e0c4af12-edda-41d7-9b44-f87396174ef2\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T06:49:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T06:49:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T06:49:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T06:49:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://0f4577d9e37f860813e221f2e208177d8d383fd04f673ce357a0a405ea29d1e6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-13T06:49:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cnxsq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://7e0d2020e62f1a277cff7047e1907dc747ed393a1f6c983e1ee258bd156c16c7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-13T06:49:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cnxsq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-13T06:49:25Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-82xjz\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-13T06:49:45Z is after 2025-08-24T17:21:41Z" Dec 13 06:49:45 crc kubenswrapper[4971]: I1213 06:49:45.201641 4971 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 13 06:49:45 crc kubenswrapper[4971]: I1213 06:49:45.201679 4971 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 13 06:49:45 crc kubenswrapper[4971]: I1213 06:49:45.201690 4971 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 13 06:49:45 crc kubenswrapper[4971]: I1213 06:49:45.201703 4971 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 13 06:49:45 crc kubenswrapper[4971]: I1213 06:49:45.201715 4971 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-13T06:49:45Z","lastTransitionTime":"2025-12-13T06:49:45Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 13 06:49:45 crc kubenswrapper[4971]: I1213 06:49:45.203604 4971 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-wwvpl" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"dddf744b-2122-49ad-a980-105c09636e0f\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T06:49:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T06:49:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T06:49:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T06:49:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4676d897c7fc668587faec563f3551a056557879cee3e865192d2bd28edf62fc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-13T06:49:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-czz9v\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-13T06:49:25Z\\\"}}\" for pod \"openshift-multus\"/\"multus-wwvpl\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-13T06:49:45Z is after 2025-08-24T17:21:41Z" Dec 13 06:49:45 crc kubenswrapper[4971]: I1213 06:49:45.303800 4971 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 13 06:49:45 crc kubenswrapper[4971]: I1213 06:49:45.303837 4971 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 13 06:49:45 crc kubenswrapper[4971]: I1213 06:49:45.303848 4971 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 13 06:49:45 crc kubenswrapper[4971]: I1213 06:49:45.303863 4971 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 13 06:49:45 crc kubenswrapper[4971]: I1213 06:49:45.303874 4971 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-13T06:49:45Z","lastTransitionTime":"2025-12-13T06:49:45Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 13 06:49:45 crc kubenswrapper[4971]: I1213 06:49:45.406405 4971 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 13 06:49:45 crc kubenswrapper[4971]: I1213 06:49:45.406443 4971 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 13 06:49:45 crc kubenswrapper[4971]: I1213 06:49:45.406454 4971 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 13 06:49:45 crc kubenswrapper[4971]: I1213 06:49:45.406470 4971 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 13 06:49:45 crc kubenswrapper[4971]: I1213 06:49:45.406480 4971 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-13T06:49:45Z","lastTransitionTime":"2025-12-13T06:49:45Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 13 06:49:45 crc kubenswrapper[4971]: I1213 06:49:45.508822 4971 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 13 06:49:45 crc kubenswrapper[4971]: I1213 06:49:45.509133 4971 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 13 06:49:45 crc kubenswrapper[4971]: I1213 06:49:45.509325 4971 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 13 06:49:45 crc kubenswrapper[4971]: I1213 06:49:45.509510 4971 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 13 06:49:45 crc kubenswrapper[4971]: I1213 06:49:45.509701 4971 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-13T06:49:45Z","lastTransitionTime":"2025-12-13T06:49:45Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 13 06:49:45 crc kubenswrapper[4971]: I1213 06:49:45.613082 4971 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 13 06:49:45 crc kubenswrapper[4971]: I1213 06:49:45.613126 4971 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 13 06:49:45 crc kubenswrapper[4971]: I1213 06:49:45.613139 4971 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 13 06:49:45 crc kubenswrapper[4971]: I1213 06:49:45.613154 4971 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 13 06:49:45 crc kubenswrapper[4971]: I1213 06:49:45.613163 4971 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-13T06:49:45Z","lastTransitionTime":"2025-12-13T06:49:45Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 13 06:49:45 crc kubenswrapper[4971]: I1213 06:49:45.715990 4971 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 13 06:49:45 crc kubenswrapper[4971]: I1213 06:49:45.716054 4971 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 13 06:49:45 crc kubenswrapper[4971]: I1213 06:49:45.716069 4971 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 13 06:49:45 crc kubenswrapper[4971]: I1213 06:49:45.716092 4971 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 13 06:49:45 crc kubenswrapper[4971]: I1213 06:49:45.716109 4971 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-13T06:49:45Z","lastTransitionTime":"2025-12-13T06:49:45Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 13 06:49:45 crc kubenswrapper[4971]: I1213 06:49:45.767884 4971 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-x98k7" Dec 13 06:49:45 crc kubenswrapper[4971]: I1213 06:49:45.767944 4971 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 13 06:49:45 crc kubenswrapper[4971]: I1213 06:49:45.768041 4971 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 13 06:49:45 crc kubenswrapper[4971]: E1213 06:49:45.768035 4971 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-x98k7" podUID="c7185dec-3240-42e8-9d90-6182a5e8b65c" Dec 13 06:49:45 crc kubenswrapper[4971]: E1213 06:49:45.768170 4971 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 13 06:49:45 crc kubenswrapper[4971]: E1213 06:49:45.768294 4971 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 13 06:49:45 crc kubenswrapper[4971]: I1213 06:49:45.768331 4971 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 13 06:49:45 crc kubenswrapper[4971]: E1213 06:49:45.768607 4971 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 13 06:49:45 crc kubenswrapper[4971]: I1213 06:49:45.818879 4971 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 13 06:49:45 crc kubenswrapper[4971]: I1213 06:49:45.818928 4971 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 13 06:49:45 crc kubenswrapper[4971]: I1213 06:49:45.818940 4971 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 13 06:49:45 crc kubenswrapper[4971]: I1213 06:49:45.818961 4971 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 13 06:49:45 crc kubenswrapper[4971]: I1213 06:49:45.818974 4971 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-13T06:49:45Z","lastTransitionTime":"2025-12-13T06:49:45Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 13 06:49:45 crc kubenswrapper[4971]: I1213 06:49:45.921837 4971 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 13 06:49:45 crc kubenswrapper[4971]: I1213 06:49:45.921886 4971 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 13 06:49:45 crc kubenswrapper[4971]: I1213 06:49:45.921897 4971 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 13 06:49:45 crc kubenswrapper[4971]: I1213 06:49:45.921916 4971 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 13 06:49:45 crc kubenswrapper[4971]: I1213 06:49:45.921928 4971 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-13T06:49:45Z","lastTransitionTime":"2025-12-13T06:49:45Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 13 06:49:46 crc kubenswrapper[4971]: I1213 06:49:46.024167 4971 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 13 06:49:46 crc kubenswrapper[4971]: I1213 06:49:46.024208 4971 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 13 06:49:46 crc kubenswrapper[4971]: I1213 06:49:46.024221 4971 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 13 06:49:46 crc kubenswrapper[4971]: I1213 06:49:46.024236 4971 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 13 06:49:46 crc kubenswrapper[4971]: I1213 06:49:46.024248 4971 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-13T06:49:46Z","lastTransitionTime":"2025-12-13T06:49:46Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 13 06:49:46 crc kubenswrapper[4971]: I1213 06:49:46.127002 4971 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 13 06:49:46 crc kubenswrapper[4971]: I1213 06:49:46.127044 4971 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 13 06:49:46 crc kubenswrapper[4971]: I1213 06:49:46.127054 4971 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 13 06:49:46 crc kubenswrapper[4971]: I1213 06:49:46.127069 4971 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 13 06:49:46 crc kubenswrapper[4971]: I1213 06:49:46.127080 4971 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-13T06:49:46Z","lastTransitionTime":"2025-12-13T06:49:46Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 13 06:49:46 crc kubenswrapper[4971]: I1213 06:49:46.229992 4971 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 13 06:49:46 crc kubenswrapper[4971]: I1213 06:49:46.230026 4971 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 13 06:49:46 crc kubenswrapper[4971]: I1213 06:49:46.230035 4971 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 13 06:49:46 crc kubenswrapper[4971]: I1213 06:49:46.230051 4971 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 13 06:49:46 crc kubenswrapper[4971]: I1213 06:49:46.230059 4971 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-13T06:49:46Z","lastTransitionTime":"2025-12-13T06:49:46Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 13 06:49:46 crc kubenswrapper[4971]: I1213 06:49:46.332598 4971 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 13 06:49:46 crc kubenswrapper[4971]: I1213 06:49:46.332631 4971 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 13 06:49:46 crc kubenswrapper[4971]: I1213 06:49:46.332639 4971 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 13 06:49:46 crc kubenswrapper[4971]: I1213 06:49:46.332653 4971 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 13 06:49:46 crc kubenswrapper[4971]: I1213 06:49:46.332663 4971 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-13T06:49:46Z","lastTransitionTime":"2025-12-13T06:49:46Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 13 06:49:46 crc kubenswrapper[4971]: I1213 06:49:46.434920 4971 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 13 06:49:46 crc kubenswrapper[4971]: I1213 06:49:46.434971 4971 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 13 06:49:46 crc kubenswrapper[4971]: I1213 06:49:46.434984 4971 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 13 06:49:46 crc kubenswrapper[4971]: I1213 06:49:46.435004 4971 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 13 06:49:46 crc kubenswrapper[4971]: I1213 06:49:46.435016 4971 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-13T06:49:46Z","lastTransitionTime":"2025-12-13T06:49:46Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 13 06:49:46 crc kubenswrapper[4971]: I1213 06:49:46.537453 4971 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 13 06:49:46 crc kubenswrapper[4971]: I1213 06:49:46.537501 4971 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 13 06:49:46 crc kubenswrapper[4971]: I1213 06:49:46.537558 4971 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 13 06:49:46 crc kubenswrapper[4971]: I1213 06:49:46.537580 4971 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 13 06:49:46 crc kubenswrapper[4971]: I1213 06:49:46.537592 4971 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-13T06:49:46Z","lastTransitionTime":"2025-12-13T06:49:46Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 13 06:49:46 crc kubenswrapper[4971]: I1213 06:49:46.639715 4971 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 13 06:49:46 crc kubenswrapper[4971]: I1213 06:49:46.640417 4971 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 13 06:49:46 crc kubenswrapper[4971]: I1213 06:49:46.640436 4971 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 13 06:49:46 crc kubenswrapper[4971]: I1213 06:49:46.640452 4971 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 13 06:49:46 crc kubenswrapper[4971]: I1213 06:49:46.640461 4971 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-13T06:49:46Z","lastTransitionTime":"2025-12-13T06:49:46Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 13 06:49:46 crc kubenswrapper[4971]: I1213 06:49:46.742581 4971 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 13 06:49:46 crc kubenswrapper[4971]: I1213 06:49:46.742623 4971 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 13 06:49:46 crc kubenswrapper[4971]: I1213 06:49:46.742634 4971 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 13 06:49:46 crc kubenswrapper[4971]: I1213 06:49:46.742647 4971 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 13 06:49:46 crc kubenswrapper[4971]: I1213 06:49:46.742658 4971 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-13T06:49:46Z","lastTransitionTime":"2025-12-13T06:49:46Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 13 06:49:46 crc kubenswrapper[4971]: I1213 06:49:46.845464 4971 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 13 06:49:46 crc kubenswrapper[4971]: I1213 06:49:46.845513 4971 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 13 06:49:46 crc kubenswrapper[4971]: I1213 06:49:46.845541 4971 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 13 06:49:46 crc kubenswrapper[4971]: I1213 06:49:46.845624 4971 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 13 06:49:46 crc kubenswrapper[4971]: I1213 06:49:46.845644 4971 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-13T06:49:46Z","lastTransitionTime":"2025-12-13T06:49:46Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 13 06:49:46 crc kubenswrapper[4971]: I1213 06:49:46.948816 4971 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 13 06:49:46 crc kubenswrapper[4971]: I1213 06:49:46.948867 4971 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 13 06:49:46 crc kubenswrapper[4971]: I1213 06:49:46.948881 4971 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 13 06:49:46 crc kubenswrapper[4971]: I1213 06:49:46.948898 4971 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 13 06:49:46 crc kubenswrapper[4971]: I1213 06:49:46.948909 4971 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-13T06:49:46Z","lastTransitionTime":"2025-12-13T06:49:46Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 13 06:49:47 crc kubenswrapper[4971]: I1213 06:49:47.050877 4971 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 13 06:49:47 crc kubenswrapper[4971]: I1213 06:49:47.050913 4971 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 13 06:49:47 crc kubenswrapper[4971]: I1213 06:49:47.050923 4971 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 13 06:49:47 crc kubenswrapper[4971]: I1213 06:49:47.050938 4971 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 13 06:49:47 crc kubenswrapper[4971]: I1213 06:49:47.050947 4971 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-13T06:49:47Z","lastTransitionTime":"2025-12-13T06:49:47Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 13 06:49:47 crc kubenswrapper[4971]: I1213 06:49:47.153274 4971 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 13 06:49:47 crc kubenswrapper[4971]: I1213 06:49:47.153322 4971 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 13 06:49:47 crc kubenswrapper[4971]: I1213 06:49:47.153333 4971 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 13 06:49:47 crc kubenswrapper[4971]: I1213 06:49:47.153352 4971 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 13 06:49:47 crc kubenswrapper[4971]: I1213 06:49:47.153365 4971 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-13T06:49:47Z","lastTransitionTime":"2025-12-13T06:49:47Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 13 06:49:47 crc kubenswrapper[4971]: I1213 06:49:47.224068 4971 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/c7185dec-3240-42e8-9d90-6182a5e8b65c-metrics-certs\") pod \"network-metrics-daemon-x98k7\" (UID: \"c7185dec-3240-42e8-9d90-6182a5e8b65c\") " pod="openshift-multus/network-metrics-daemon-x98k7" Dec 13 06:49:47 crc kubenswrapper[4971]: E1213 06:49:47.224182 4971 secret.go:188] Couldn't get secret openshift-multus/metrics-daemon-secret: object "openshift-multus"/"metrics-daemon-secret" not registered Dec 13 06:49:47 crc kubenswrapper[4971]: E1213 06:49:47.224241 4971 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/c7185dec-3240-42e8-9d90-6182a5e8b65c-metrics-certs podName:c7185dec-3240-42e8-9d90-6182a5e8b65c nodeName:}" failed. No retries permitted until 2025-12-13 06:49:55.22422486 +0000 UTC m=+51.828634308 (durationBeforeRetry 8s). Error: MountVolume.SetUp failed for volume "metrics-certs" (UniqueName: "kubernetes.io/secret/c7185dec-3240-42e8-9d90-6182a5e8b65c-metrics-certs") pod "network-metrics-daemon-x98k7" (UID: "c7185dec-3240-42e8-9d90-6182a5e8b65c") : object "openshift-multus"/"metrics-daemon-secret" not registered Dec 13 06:49:47 crc kubenswrapper[4971]: I1213 06:49:47.256734 4971 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 13 06:49:47 crc kubenswrapper[4971]: I1213 06:49:47.256772 4971 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 13 06:49:47 crc kubenswrapper[4971]: I1213 06:49:47.256780 4971 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 13 06:49:47 crc kubenswrapper[4971]: I1213 06:49:47.256794 4971 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 13 06:49:47 crc kubenswrapper[4971]: I1213 06:49:47.256804 4971 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-13T06:49:47Z","lastTransitionTime":"2025-12-13T06:49:47Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 13 06:49:47 crc kubenswrapper[4971]: I1213 06:49:47.359084 4971 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 13 06:49:47 crc kubenswrapper[4971]: I1213 06:49:47.359118 4971 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 13 06:49:47 crc kubenswrapper[4971]: I1213 06:49:47.359129 4971 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 13 06:49:47 crc kubenswrapper[4971]: I1213 06:49:47.359145 4971 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 13 06:49:47 crc kubenswrapper[4971]: I1213 06:49:47.359155 4971 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-13T06:49:47Z","lastTransitionTime":"2025-12-13T06:49:47Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 13 06:49:47 crc kubenswrapper[4971]: I1213 06:49:47.461216 4971 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 13 06:49:47 crc kubenswrapper[4971]: I1213 06:49:47.461258 4971 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 13 06:49:47 crc kubenswrapper[4971]: I1213 06:49:47.461268 4971 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 13 06:49:47 crc kubenswrapper[4971]: I1213 06:49:47.461283 4971 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 13 06:49:47 crc kubenswrapper[4971]: I1213 06:49:47.461294 4971 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-13T06:49:47Z","lastTransitionTime":"2025-12-13T06:49:47Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 13 06:49:47 crc kubenswrapper[4971]: I1213 06:49:47.563098 4971 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 13 06:49:47 crc kubenswrapper[4971]: I1213 06:49:47.563140 4971 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 13 06:49:47 crc kubenswrapper[4971]: I1213 06:49:47.563151 4971 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 13 06:49:47 crc kubenswrapper[4971]: I1213 06:49:47.563165 4971 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 13 06:49:47 crc kubenswrapper[4971]: I1213 06:49:47.563177 4971 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-13T06:49:47Z","lastTransitionTime":"2025-12-13T06:49:47Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 13 06:49:47 crc kubenswrapper[4971]: I1213 06:49:47.666131 4971 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 13 06:49:47 crc kubenswrapper[4971]: I1213 06:49:47.666167 4971 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 13 06:49:47 crc kubenswrapper[4971]: I1213 06:49:47.666178 4971 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 13 06:49:47 crc kubenswrapper[4971]: I1213 06:49:47.666194 4971 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 13 06:49:47 crc kubenswrapper[4971]: I1213 06:49:47.666206 4971 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-13T06:49:47Z","lastTransitionTime":"2025-12-13T06:49:47Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 13 06:49:47 crc kubenswrapper[4971]: I1213 06:49:47.767382 4971 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-x98k7" Dec 13 06:49:47 crc kubenswrapper[4971]: I1213 06:49:47.767484 4971 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 13 06:49:47 crc kubenswrapper[4971]: I1213 06:49:47.767409 4971 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 13 06:49:47 crc kubenswrapper[4971]: E1213 06:49:47.767594 4971 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-x98k7" podUID="c7185dec-3240-42e8-9d90-6182a5e8b65c" Dec 13 06:49:47 crc kubenswrapper[4971]: I1213 06:49:47.767644 4971 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 13 06:49:47 crc kubenswrapper[4971]: E1213 06:49:47.767726 4971 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 13 06:49:47 crc kubenswrapper[4971]: E1213 06:49:47.767781 4971 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 13 06:49:47 crc kubenswrapper[4971]: E1213 06:49:47.767835 4971 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 13 06:49:47 crc kubenswrapper[4971]: I1213 06:49:47.769332 4971 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 13 06:49:47 crc kubenswrapper[4971]: I1213 06:49:47.769369 4971 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 13 06:49:47 crc kubenswrapper[4971]: I1213 06:49:47.769385 4971 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 13 06:49:47 crc kubenswrapper[4971]: I1213 06:49:47.769404 4971 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 13 06:49:47 crc kubenswrapper[4971]: I1213 06:49:47.769417 4971 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-13T06:49:47Z","lastTransitionTime":"2025-12-13T06:49:47Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 13 06:49:47 crc kubenswrapper[4971]: I1213 06:49:47.872706 4971 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 13 06:49:47 crc kubenswrapper[4971]: I1213 06:49:47.872762 4971 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 13 06:49:47 crc kubenswrapper[4971]: I1213 06:49:47.872772 4971 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 13 06:49:47 crc kubenswrapper[4971]: I1213 06:49:47.872790 4971 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 13 06:49:47 crc kubenswrapper[4971]: I1213 06:49:47.872800 4971 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-13T06:49:47Z","lastTransitionTime":"2025-12-13T06:49:47Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 13 06:49:47 crc kubenswrapper[4971]: I1213 06:49:47.975970 4971 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 13 06:49:47 crc kubenswrapper[4971]: I1213 06:49:47.976028 4971 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 13 06:49:47 crc kubenswrapper[4971]: I1213 06:49:47.976038 4971 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 13 06:49:47 crc kubenswrapper[4971]: I1213 06:49:47.976059 4971 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 13 06:49:47 crc kubenswrapper[4971]: I1213 06:49:47.976071 4971 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-13T06:49:47Z","lastTransitionTime":"2025-12-13T06:49:47Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 13 06:49:48 crc kubenswrapper[4971]: I1213 06:49:48.078063 4971 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 13 06:49:48 crc kubenswrapper[4971]: I1213 06:49:48.078117 4971 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 13 06:49:48 crc kubenswrapper[4971]: I1213 06:49:48.078129 4971 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 13 06:49:48 crc kubenswrapper[4971]: I1213 06:49:48.078145 4971 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 13 06:49:48 crc kubenswrapper[4971]: I1213 06:49:48.078155 4971 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-13T06:49:48Z","lastTransitionTime":"2025-12-13T06:49:48Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 13 06:49:48 crc kubenswrapper[4971]: I1213 06:49:48.179884 4971 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 13 06:49:48 crc kubenswrapper[4971]: I1213 06:49:48.179922 4971 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 13 06:49:48 crc kubenswrapper[4971]: I1213 06:49:48.179935 4971 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 13 06:49:48 crc kubenswrapper[4971]: I1213 06:49:48.179950 4971 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 13 06:49:48 crc kubenswrapper[4971]: I1213 06:49:48.179960 4971 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-13T06:49:48Z","lastTransitionTime":"2025-12-13T06:49:48Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 13 06:49:48 crc kubenswrapper[4971]: I1213 06:49:48.282116 4971 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 13 06:49:48 crc kubenswrapper[4971]: I1213 06:49:48.282161 4971 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 13 06:49:48 crc kubenswrapper[4971]: I1213 06:49:48.282171 4971 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 13 06:49:48 crc kubenswrapper[4971]: I1213 06:49:48.282188 4971 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 13 06:49:48 crc kubenswrapper[4971]: I1213 06:49:48.282198 4971 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-13T06:49:48Z","lastTransitionTime":"2025-12-13T06:49:48Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 13 06:49:48 crc kubenswrapper[4971]: I1213 06:49:48.384354 4971 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 13 06:49:48 crc kubenswrapper[4971]: I1213 06:49:48.384419 4971 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 13 06:49:48 crc kubenswrapper[4971]: I1213 06:49:48.384428 4971 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 13 06:49:48 crc kubenswrapper[4971]: I1213 06:49:48.384440 4971 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 13 06:49:48 crc kubenswrapper[4971]: I1213 06:49:48.384449 4971 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-13T06:49:48Z","lastTransitionTime":"2025-12-13T06:49:48Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 13 06:49:48 crc kubenswrapper[4971]: I1213 06:49:48.487034 4971 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 13 06:49:48 crc kubenswrapper[4971]: I1213 06:49:48.487080 4971 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 13 06:49:48 crc kubenswrapper[4971]: I1213 06:49:48.487091 4971 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 13 06:49:48 crc kubenswrapper[4971]: I1213 06:49:48.487107 4971 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 13 06:49:48 crc kubenswrapper[4971]: I1213 06:49:48.487118 4971 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-13T06:49:48Z","lastTransitionTime":"2025-12-13T06:49:48Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 13 06:49:48 crc kubenswrapper[4971]: I1213 06:49:48.589181 4971 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 13 06:49:48 crc kubenswrapper[4971]: I1213 06:49:48.589235 4971 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 13 06:49:48 crc kubenswrapper[4971]: I1213 06:49:48.589246 4971 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 13 06:49:48 crc kubenswrapper[4971]: I1213 06:49:48.589263 4971 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 13 06:49:48 crc kubenswrapper[4971]: I1213 06:49:48.589274 4971 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-13T06:49:48Z","lastTransitionTime":"2025-12-13T06:49:48Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 13 06:49:48 crc kubenswrapper[4971]: I1213 06:49:48.692454 4971 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 13 06:49:48 crc kubenswrapper[4971]: I1213 06:49:48.692544 4971 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 13 06:49:48 crc kubenswrapper[4971]: I1213 06:49:48.692556 4971 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 13 06:49:48 crc kubenswrapper[4971]: I1213 06:49:48.692575 4971 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 13 06:49:48 crc kubenswrapper[4971]: I1213 06:49:48.692588 4971 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-13T06:49:48Z","lastTransitionTime":"2025-12-13T06:49:48Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 13 06:49:48 crc kubenswrapper[4971]: I1213 06:49:48.795232 4971 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 13 06:49:48 crc kubenswrapper[4971]: I1213 06:49:48.795284 4971 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 13 06:49:48 crc kubenswrapper[4971]: I1213 06:49:48.795298 4971 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 13 06:49:48 crc kubenswrapper[4971]: I1213 06:49:48.795315 4971 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 13 06:49:48 crc kubenswrapper[4971]: I1213 06:49:48.795325 4971 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-13T06:49:48Z","lastTransitionTime":"2025-12-13T06:49:48Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 13 06:49:48 crc kubenswrapper[4971]: I1213 06:49:48.902905 4971 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 13 06:49:48 crc kubenswrapper[4971]: I1213 06:49:48.903472 4971 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 13 06:49:48 crc kubenswrapper[4971]: I1213 06:49:48.903485 4971 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 13 06:49:48 crc kubenswrapper[4971]: I1213 06:49:48.903504 4971 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 13 06:49:48 crc kubenswrapper[4971]: I1213 06:49:48.903536 4971 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-13T06:49:48Z","lastTransitionTime":"2025-12-13T06:49:48Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 13 06:49:49 crc kubenswrapper[4971]: I1213 06:49:49.006218 4971 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 13 06:49:49 crc kubenswrapper[4971]: I1213 06:49:49.006945 4971 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 13 06:49:49 crc kubenswrapper[4971]: I1213 06:49:49.006983 4971 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 13 06:49:49 crc kubenswrapper[4971]: I1213 06:49:49.007000 4971 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 13 06:49:49 crc kubenswrapper[4971]: I1213 06:49:49.007010 4971 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-13T06:49:49Z","lastTransitionTime":"2025-12-13T06:49:49Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 13 06:49:49 crc kubenswrapper[4971]: I1213 06:49:49.110143 4971 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 13 06:49:49 crc kubenswrapper[4971]: I1213 06:49:49.110836 4971 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 13 06:49:49 crc kubenswrapper[4971]: I1213 06:49:49.110852 4971 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 13 06:49:49 crc kubenswrapper[4971]: I1213 06:49:49.110883 4971 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 13 06:49:49 crc kubenswrapper[4971]: I1213 06:49:49.110906 4971 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-13T06:49:49Z","lastTransitionTime":"2025-12-13T06:49:49Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 13 06:49:49 crc kubenswrapper[4971]: I1213 06:49:49.213260 4971 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 13 06:49:49 crc kubenswrapper[4971]: I1213 06:49:49.213322 4971 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 13 06:49:49 crc kubenswrapper[4971]: I1213 06:49:49.213335 4971 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 13 06:49:49 crc kubenswrapper[4971]: I1213 06:49:49.213353 4971 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 13 06:49:49 crc kubenswrapper[4971]: I1213 06:49:49.213365 4971 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-13T06:49:49Z","lastTransitionTime":"2025-12-13T06:49:49Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 13 06:49:49 crc kubenswrapper[4971]: I1213 06:49:49.315999 4971 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 13 06:49:49 crc kubenswrapper[4971]: I1213 06:49:49.316049 4971 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 13 06:49:49 crc kubenswrapper[4971]: I1213 06:49:49.316060 4971 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 13 06:49:49 crc kubenswrapper[4971]: I1213 06:49:49.316076 4971 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 13 06:49:49 crc kubenswrapper[4971]: I1213 06:49:49.316088 4971 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-13T06:49:49Z","lastTransitionTime":"2025-12-13T06:49:49Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 13 06:49:49 crc kubenswrapper[4971]: I1213 06:49:49.418593 4971 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 13 06:49:49 crc kubenswrapper[4971]: I1213 06:49:49.418680 4971 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 13 06:49:49 crc kubenswrapper[4971]: I1213 06:49:49.418696 4971 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 13 06:49:49 crc kubenswrapper[4971]: I1213 06:49:49.419316 4971 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 13 06:49:49 crc kubenswrapper[4971]: I1213 06:49:49.419375 4971 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-13T06:49:49Z","lastTransitionTime":"2025-12-13T06:49:49Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 13 06:49:49 crc kubenswrapper[4971]: I1213 06:49:49.521643 4971 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 13 06:49:49 crc kubenswrapper[4971]: I1213 06:49:49.521695 4971 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 13 06:49:49 crc kubenswrapper[4971]: I1213 06:49:49.521708 4971 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 13 06:49:49 crc kubenswrapper[4971]: I1213 06:49:49.521725 4971 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 13 06:49:49 crc kubenswrapper[4971]: I1213 06:49:49.521737 4971 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-13T06:49:49Z","lastTransitionTime":"2025-12-13T06:49:49Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 13 06:49:49 crc kubenswrapper[4971]: I1213 06:49:49.624368 4971 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 13 06:49:49 crc kubenswrapper[4971]: I1213 06:49:49.624406 4971 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 13 06:49:49 crc kubenswrapper[4971]: I1213 06:49:49.624415 4971 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 13 06:49:49 crc kubenswrapper[4971]: I1213 06:49:49.624427 4971 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 13 06:49:49 crc kubenswrapper[4971]: I1213 06:49:49.624436 4971 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-13T06:49:49Z","lastTransitionTime":"2025-12-13T06:49:49Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 13 06:49:49 crc kubenswrapper[4971]: I1213 06:49:49.727338 4971 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 13 06:49:49 crc kubenswrapper[4971]: I1213 06:49:49.727395 4971 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 13 06:49:49 crc kubenswrapper[4971]: I1213 06:49:49.727410 4971 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 13 06:49:49 crc kubenswrapper[4971]: I1213 06:49:49.727426 4971 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 13 06:49:49 crc kubenswrapper[4971]: I1213 06:49:49.727438 4971 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-13T06:49:49Z","lastTransitionTime":"2025-12-13T06:49:49Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 13 06:49:49 crc kubenswrapper[4971]: I1213 06:49:49.767786 4971 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 13 06:49:49 crc kubenswrapper[4971]: E1213 06:49:49.767936 4971 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 13 06:49:49 crc kubenswrapper[4971]: I1213 06:49:49.768012 4971 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-x98k7" Dec 13 06:49:49 crc kubenswrapper[4971]: I1213 06:49:49.768054 4971 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 13 06:49:49 crc kubenswrapper[4971]: E1213 06:49:49.768105 4971 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-x98k7" podUID="c7185dec-3240-42e8-9d90-6182a5e8b65c" Dec 13 06:49:49 crc kubenswrapper[4971]: I1213 06:49:49.768075 4971 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 13 06:49:49 crc kubenswrapper[4971]: E1213 06:49:49.768227 4971 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 13 06:49:49 crc kubenswrapper[4971]: E1213 06:49:49.768342 4971 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 13 06:49:49 crc kubenswrapper[4971]: I1213 06:49:49.829821 4971 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 13 06:49:49 crc kubenswrapper[4971]: I1213 06:49:49.829867 4971 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 13 06:49:49 crc kubenswrapper[4971]: I1213 06:49:49.829876 4971 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 13 06:49:49 crc kubenswrapper[4971]: I1213 06:49:49.829893 4971 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 13 06:49:49 crc kubenswrapper[4971]: I1213 06:49:49.829904 4971 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-13T06:49:49Z","lastTransitionTime":"2025-12-13T06:49:49Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 13 06:49:49 crc kubenswrapper[4971]: I1213 06:49:49.932786 4971 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 13 06:49:49 crc kubenswrapper[4971]: I1213 06:49:49.932848 4971 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 13 06:49:49 crc kubenswrapper[4971]: I1213 06:49:49.932860 4971 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 13 06:49:49 crc kubenswrapper[4971]: I1213 06:49:49.932881 4971 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 13 06:49:49 crc kubenswrapper[4971]: I1213 06:49:49.932896 4971 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-13T06:49:49Z","lastTransitionTime":"2025-12-13T06:49:49Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 13 06:49:50 crc kubenswrapper[4971]: I1213 06:49:50.035832 4971 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 13 06:49:50 crc kubenswrapper[4971]: I1213 06:49:50.035878 4971 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 13 06:49:50 crc kubenswrapper[4971]: I1213 06:49:50.035895 4971 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 13 06:49:50 crc kubenswrapper[4971]: I1213 06:49:50.035916 4971 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 13 06:49:50 crc kubenswrapper[4971]: I1213 06:49:50.035928 4971 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-13T06:49:50Z","lastTransitionTime":"2025-12-13T06:49:50Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 13 06:49:50 crc kubenswrapper[4971]: I1213 06:49:50.138333 4971 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 13 06:49:50 crc kubenswrapper[4971]: I1213 06:49:50.138374 4971 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 13 06:49:50 crc kubenswrapper[4971]: I1213 06:49:50.138385 4971 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 13 06:49:50 crc kubenswrapper[4971]: I1213 06:49:50.138401 4971 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 13 06:49:50 crc kubenswrapper[4971]: I1213 06:49:50.138413 4971 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-13T06:49:50Z","lastTransitionTime":"2025-12-13T06:49:50Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 13 06:49:50 crc kubenswrapper[4971]: I1213 06:49:50.240331 4971 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 13 06:49:50 crc kubenswrapper[4971]: I1213 06:49:50.240365 4971 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 13 06:49:50 crc kubenswrapper[4971]: I1213 06:49:50.240373 4971 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 13 06:49:50 crc kubenswrapper[4971]: I1213 06:49:50.240387 4971 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 13 06:49:50 crc kubenswrapper[4971]: I1213 06:49:50.240396 4971 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-13T06:49:50Z","lastTransitionTime":"2025-12-13T06:49:50Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 13 06:49:50 crc kubenswrapper[4971]: I1213 06:49:50.343065 4971 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 13 06:49:50 crc kubenswrapper[4971]: I1213 06:49:50.343116 4971 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 13 06:49:50 crc kubenswrapper[4971]: I1213 06:49:50.343127 4971 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 13 06:49:50 crc kubenswrapper[4971]: I1213 06:49:50.343145 4971 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 13 06:49:50 crc kubenswrapper[4971]: I1213 06:49:50.343160 4971 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-13T06:49:50Z","lastTransitionTime":"2025-12-13T06:49:50Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 13 06:49:50 crc kubenswrapper[4971]: I1213 06:49:50.390094 4971 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 13 06:49:50 crc kubenswrapper[4971]: I1213 06:49:50.390132 4971 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 13 06:49:50 crc kubenswrapper[4971]: I1213 06:49:50.390143 4971 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 13 06:49:50 crc kubenswrapper[4971]: I1213 06:49:50.390158 4971 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 13 06:49:50 crc kubenswrapper[4971]: I1213 06:49:50.390169 4971 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-13T06:49:50Z","lastTransitionTime":"2025-12-13T06:49:50Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 13 06:49:50 crc kubenswrapper[4971]: E1213 06:49:50.401999 4971 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-12-13T06:49:50Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-13T06:49:50Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-13T06:49:50Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-13T06:49:50Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-13T06:49:50Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-13T06:49:50Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-13T06:49:50Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-13T06:49:50Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"bd36cbc9-bbe9-45a1-a21a-ad2bd10499ee\\\",\\\"systemUUID\\\":\\\"a9af00dc-cf25-47af-9997-79577f8f3bea\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-13T06:49:50Z is after 2025-08-24T17:21:41Z" Dec 13 06:49:50 crc kubenswrapper[4971]: I1213 06:49:50.405424 4971 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 13 06:49:50 crc kubenswrapper[4971]: I1213 06:49:50.405473 4971 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 13 06:49:50 crc kubenswrapper[4971]: I1213 06:49:50.405482 4971 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 13 06:49:50 crc kubenswrapper[4971]: I1213 06:49:50.405497 4971 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 13 06:49:50 crc kubenswrapper[4971]: I1213 06:49:50.405507 4971 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-13T06:49:50Z","lastTransitionTime":"2025-12-13T06:49:50Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 13 06:49:50 crc kubenswrapper[4971]: E1213 06:49:50.416824 4971 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-12-13T06:49:50Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-13T06:49:50Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-13T06:49:50Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-13T06:49:50Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-13T06:49:50Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-13T06:49:50Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-13T06:49:50Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-13T06:49:50Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"bd36cbc9-bbe9-45a1-a21a-ad2bd10499ee\\\",\\\"systemUUID\\\":\\\"a9af00dc-cf25-47af-9997-79577f8f3bea\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-13T06:49:50Z is after 2025-08-24T17:21:41Z" Dec 13 06:49:50 crc kubenswrapper[4971]: I1213 06:49:50.420291 4971 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 13 06:49:50 crc kubenswrapper[4971]: I1213 06:49:50.420417 4971 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 13 06:49:50 crc kubenswrapper[4971]: I1213 06:49:50.420434 4971 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 13 06:49:50 crc kubenswrapper[4971]: I1213 06:49:50.420462 4971 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 13 06:49:50 crc kubenswrapper[4971]: I1213 06:49:50.420477 4971 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-13T06:49:50Z","lastTransitionTime":"2025-12-13T06:49:50Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 13 06:49:50 crc kubenswrapper[4971]: E1213 06:49:50.432865 4971 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-12-13T06:49:50Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-13T06:49:50Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-13T06:49:50Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-13T06:49:50Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-13T06:49:50Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-13T06:49:50Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-13T06:49:50Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-13T06:49:50Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"bd36cbc9-bbe9-45a1-a21a-ad2bd10499ee\\\",\\\"systemUUID\\\":\\\"a9af00dc-cf25-47af-9997-79577f8f3bea\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-13T06:49:50Z is after 2025-08-24T17:21:41Z" Dec 13 06:49:50 crc kubenswrapper[4971]: I1213 06:49:50.435975 4971 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 13 06:49:50 crc kubenswrapper[4971]: I1213 06:49:50.436013 4971 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 13 06:49:50 crc kubenswrapper[4971]: I1213 06:49:50.436025 4971 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 13 06:49:50 crc kubenswrapper[4971]: I1213 06:49:50.436042 4971 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 13 06:49:50 crc kubenswrapper[4971]: I1213 06:49:50.436051 4971 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-13T06:49:50Z","lastTransitionTime":"2025-12-13T06:49:50Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 13 06:49:50 crc kubenswrapper[4971]: E1213 06:49:50.451256 4971 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-12-13T06:49:50Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-13T06:49:50Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-13T06:49:50Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-13T06:49:50Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-13T06:49:50Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-13T06:49:50Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-13T06:49:50Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-13T06:49:50Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"bd36cbc9-bbe9-45a1-a21a-ad2bd10499ee\\\",\\\"systemUUID\\\":\\\"a9af00dc-cf25-47af-9997-79577f8f3bea\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-13T06:49:50Z is after 2025-08-24T17:21:41Z" Dec 13 06:49:50 crc kubenswrapper[4971]: I1213 06:49:50.454736 4971 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 13 06:49:50 crc kubenswrapper[4971]: I1213 06:49:50.454771 4971 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 13 06:49:50 crc kubenswrapper[4971]: I1213 06:49:50.454784 4971 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 13 06:49:50 crc kubenswrapper[4971]: I1213 06:49:50.454804 4971 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 13 06:49:50 crc kubenswrapper[4971]: I1213 06:49:50.454816 4971 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-13T06:49:50Z","lastTransitionTime":"2025-12-13T06:49:50Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 13 06:49:50 crc kubenswrapper[4971]: E1213 06:49:50.465199 4971 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-12-13T06:49:50Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-13T06:49:50Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-13T06:49:50Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-13T06:49:50Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-13T06:49:50Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-13T06:49:50Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-13T06:49:50Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-13T06:49:50Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"bd36cbc9-bbe9-45a1-a21a-ad2bd10499ee\\\",\\\"systemUUID\\\":\\\"a9af00dc-cf25-47af-9997-79577f8f3bea\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-13T06:49:50Z is after 2025-08-24T17:21:41Z" Dec 13 06:49:50 crc kubenswrapper[4971]: E1213 06:49:50.465305 4971 kubelet_node_status.go:572] "Unable to update node status" err="update node status exceeds retry count" Dec 13 06:49:50 crc kubenswrapper[4971]: I1213 06:49:50.466959 4971 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 13 06:49:50 crc kubenswrapper[4971]: I1213 06:49:50.466986 4971 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 13 06:49:50 crc kubenswrapper[4971]: I1213 06:49:50.466995 4971 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 13 06:49:50 crc kubenswrapper[4971]: I1213 06:49:50.467009 4971 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 13 06:49:50 crc kubenswrapper[4971]: I1213 06:49:50.467019 4971 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-13T06:49:50Z","lastTransitionTime":"2025-12-13T06:49:50Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 13 06:49:50 crc kubenswrapper[4971]: I1213 06:49:50.569022 4971 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 13 06:49:50 crc kubenswrapper[4971]: I1213 06:49:50.569058 4971 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 13 06:49:50 crc kubenswrapper[4971]: I1213 06:49:50.569069 4971 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 13 06:49:50 crc kubenswrapper[4971]: I1213 06:49:50.569084 4971 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 13 06:49:50 crc kubenswrapper[4971]: I1213 06:49:50.569095 4971 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-13T06:49:50Z","lastTransitionTime":"2025-12-13T06:49:50Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 13 06:49:50 crc kubenswrapper[4971]: I1213 06:49:50.671781 4971 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 13 06:49:50 crc kubenswrapper[4971]: I1213 06:49:50.671826 4971 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 13 06:49:50 crc kubenswrapper[4971]: I1213 06:49:50.671836 4971 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 13 06:49:50 crc kubenswrapper[4971]: I1213 06:49:50.671851 4971 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 13 06:49:50 crc kubenswrapper[4971]: I1213 06:49:50.671862 4971 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-13T06:49:50Z","lastTransitionTime":"2025-12-13T06:49:50Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 13 06:49:50 crc kubenswrapper[4971]: I1213 06:49:50.773669 4971 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 13 06:49:50 crc kubenswrapper[4971]: I1213 06:49:50.773694 4971 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 13 06:49:50 crc kubenswrapper[4971]: I1213 06:49:50.773703 4971 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 13 06:49:50 crc kubenswrapper[4971]: I1213 06:49:50.773716 4971 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 13 06:49:50 crc kubenswrapper[4971]: I1213 06:49:50.773724 4971 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-13T06:49:50Z","lastTransitionTime":"2025-12-13T06:49:50Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 13 06:49:50 crc kubenswrapper[4971]: I1213 06:49:50.877343 4971 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 13 06:49:50 crc kubenswrapper[4971]: I1213 06:49:50.877406 4971 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 13 06:49:50 crc kubenswrapper[4971]: I1213 06:49:50.877419 4971 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 13 06:49:50 crc kubenswrapper[4971]: I1213 06:49:50.877438 4971 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 13 06:49:50 crc kubenswrapper[4971]: I1213 06:49:50.877451 4971 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-13T06:49:50Z","lastTransitionTime":"2025-12-13T06:49:50Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 13 06:49:50 crc kubenswrapper[4971]: I1213 06:49:50.980781 4971 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 13 06:49:50 crc kubenswrapper[4971]: I1213 06:49:50.980828 4971 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 13 06:49:50 crc kubenswrapper[4971]: I1213 06:49:50.980839 4971 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 13 06:49:50 crc kubenswrapper[4971]: I1213 06:49:50.980857 4971 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 13 06:49:50 crc kubenswrapper[4971]: I1213 06:49:50.980869 4971 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-13T06:49:50Z","lastTransitionTime":"2025-12-13T06:49:50Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 13 06:49:51 crc kubenswrapper[4971]: I1213 06:49:51.082485 4971 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 13 06:49:51 crc kubenswrapper[4971]: I1213 06:49:51.082578 4971 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 13 06:49:51 crc kubenswrapper[4971]: I1213 06:49:51.082591 4971 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 13 06:49:51 crc kubenswrapper[4971]: I1213 06:49:51.082608 4971 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 13 06:49:51 crc kubenswrapper[4971]: I1213 06:49:51.082619 4971 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-13T06:49:51Z","lastTransitionTime":"2025-12-13T06:49:51Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 13 06:49:51 crc kubenswrapper[4971]: I1213 06:49:51.185378 4971 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 13 06:49:51 crc kubenswrapper[4971]: I1213 06:49:51.185435 4971 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 13 06:49:51 crc kubenswrapper[4971]: I1213 06:49:51.185446 4971 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 13 06:49:51 crc kubenswrapper[4971]: I1213 06:49:51.185464 4971 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 13 06:49:51 crc kubenswrapper[4971]: I1213 06:49:51.185476 4971 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-13T06:49:51Z","lastTransitionTime":"2025-12-13T06:49:51Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 13 06:49:51 crc kubenswrapper[4971]: I1213 06:49:51.287684 4971 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 13 06:49:51 crc kubenswrapper[4971]: I1213 06:49:51.287725 4971 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 13 06:49:51 crc kubenswrapper[4971]: I1213 06:49:51.287742 4971 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 13 06:49:51 crc kubenswrapper[4971]: I1213 06:49:51.287760 4971 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 13 06:49:51 crc kubenswrapper[4971]: I1213 06:49:51.287772 4971 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-13T06:49:51Z","lastTransitionTime":"2025-12-13T06:49:51Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 13 06:49:51 crc kubenswrapper[4971]: I1213 06:49:51.390174 4971 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 13 06:49:51 crc kubenswrapper[4971]: I1213 06:49:51.390204 4971 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 13 06:49:51 crc kubenswrapper[4971]: I1213 06:49:51.390212 4971 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 13 06:49:51 crc kubenswrapper[4971]: I1213 06:49:51.390227 4971 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 13 06:49:51 crc kubenswrapper[4971]: I1213 06:49:51.390236 4971 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-13T06:49:51Z","lastTransitionTime":"2025-12-13T06:49:51Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 13 06:49:51 crc kubenswrapper[4971]: I1213 06:49:51.493122 4971 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 13 06:49:51 crc kubenswrapper[4971]: I1213 06:49:51.493165 4971 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 13 06:49:51 crc kubenswrapper[4971]: I1213 06:49:51.493174 4971 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 13 06:49:51 crc kubenswrapper[4971]: I1213 06:49:51.493190 4971 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 13 06:49:51 crc kubenswrapper[4971]: I1213 06:49:51.493206 4971 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-13T06:49:51Z","lastTransitionTime":"2025-12-13T06:49:51Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 13 06:49:51 crc kubenswrapper[4971]: I1213 06:49:51.595555 4971 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 13 06:49:51 crc kubenswrapper[4971]: I1213 06:49:51.595631 4971 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 13 06:49:51 crc kubenswrapper[4971]: I1213 06:49:51.595654 4971 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 13 06:49:51 crc kubenswrapper[4971]: I1213 06:49:51.595684 4971 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 13 06:49:51 crc kubenswrapper[4971]: I1213 06:49:51.595706 4971 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-13T06:49:51Z","lastTransitionTime":"2025-12-13T06:49:51Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 13 06:49:51 crc kubenswrapper[4971]: I1213 06:49:51.697872 4971 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 13 06:49:51 crc kubenswrapper[4971]: I1213 06:49:51.697912 4971 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 13 06:49:51 crc kubenswrapper[4971]: I1213 06:49:51.697924 4971 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 13 06:49:51 crc kubenswrapper[4971]: I1213 06:49:51.697940 4971 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 13 06:49:51 crc kubenswrapper[4971]: I1213 06:49:51.697950 4971 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-13T06:49:51Z","lastTransitionTime":"2025-12-13T06:49:51Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 13 06:49:51 crc kubenswrapper[4971]: I1213 06:49:51.768484 4971 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-x98k7" Dec 13 06:49:51 crc kubenswrapper[4971]: I1213 06:49:51.768581 4971 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 13 06:49:51 crc kubenswrapper[4971]: I1213 06:49:51.768678 4971 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 13 06:49:51 crc kubenswrapper[4971]: E1213 06:49:51.768687 4971 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-x98k7" podUID="c7185dec-3240-42e8-9d90-6182a5e8b65c" Dec 13 06:49:51 crc kubenswrapper[4971]: I1213 06:49:51.768747 4971 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 13 06:49:51 crc kubenswrapper[4971]: E1213 06:49:51.768890 4971 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 13 06:49:51 crc kubenswrapper[4971]: E1213 06:49:51.768989 4971 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 13 06:49:51 crc kubenswrapper[4971]: E1213 06:49:51.769061 4971 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 13 06:49:51 crc kubenswrapper[4971]: I1213 06:49:51.800003 4971 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 13 06:49:51 crc kubenswrapper[4971]: I1213 06:49:51.800071 4971 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 13 06:49:51 crc kubenswrapper[4971]: I1213 06:49:51.800082 4971 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 13 06:49:51 crc kubenswrapper[4971]: I1213 06:49:51.800099 4971 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 13 06:49:51 crc kubenswrapper[4971]: I1213 06:49:51.800112 4971 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-13T06:49:51Z","lastTransitionTime":"2025-12-13T06:49:51Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 13 06:49:51 crc kubenswrapper[4971]: I1213 06:49:51.902555 4971 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 13 06:49:51 crc kubenswrapper[4971]: I1213 06:49:51.902596 4971 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 13 06:49:51 crc kubenswrapper[4971]: I1213 06:49:51.902617 4971 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 13 06:49:51 crc kubenswrapper[4971]: I1213 06:49:51.902636 4971 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 13 06:49:51 crc kubenswrapper[4971]: I1213 06:49:51.902647 4971 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-13T06:49:51Z","lastTransitionTime":"2025-12-13T06:49:51Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 13 06:49:52 crc kubenswrapper[4971]: I1213 06:49:52.005592 4971 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 13 06:49:52 crc kubenswrapper[4971]: I1213 06:49:52.005642 4971 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 13 06:49:52 crc kubenswrapper[4971]: I1213 06:49:52.005653 4971 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 13 06:49:52 crc kubenswrapper[4971]: I1213 06:49:52.005671 4971 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 13 06:49:52 crc kubenswrapper[4971]: I1213 06:49:52.005687 4971 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-13T06:49:52Z","lastTransitionTime":"2025-12-13T06:49:52Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 13 06:49:52 crc kubenswrapper[4971]: I1213 06:49:52.108212 4971 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 13 06:49:52 crc kubenswrapper[4971]: I1213 06:49:52.108254 4971 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 13 06:49:52 crc kubenswrapper[4971]: I1213 06:49:52.108264 4971 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 13 06:49:52 crc kubenswrapper[4971]: I1213 06:49:52.108285 4971 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 13 06:49:52 crc kubenswrapper[4971]: I1213 06:49:52.108296 4971 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-13T06:49:52Z","lastTransitionTime":"2025-12-13T06:49:52Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 13 06:49:52 crc kubenswrapper[4971]: I1213 06:49:52.210978 4971 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 13 06:49:52 crc kubenswrapper[4971]: I1213 06:49:52.211034 4971 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 13 06:49:52 crc kubenswrapper[4971]: I1213 06:49:52.211044 4971 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 13 06:49:52 crc kubenswrapper[4971]: I1213 06:49:52.211058 4971 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 13 06:49:52 crc kubenswrapper[4971]: I1213 06:49:52.211068 4971 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-13T06:49:52Z","lastTransitionTime":"2025-12-13T06:49:52Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 13 06:49:52 crc kubenswrapper[4971]: I1213 06:49:52.313426 4971 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 13 06:49:52 crc kubenswrapper[4971]: I1213 06:49:52.313465 4971 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 13 06:49:52 crc kubenswrapper[4971]: I1213 06:49:52.313474 4971 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 13 06:49:52 crc kubenswrapper[4971]: I1213 06:49:52.313491 4971 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 13 06:49:52 crc kubenswrapper[4971]: I1213 06:49:52.313501 4971 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-13T06:49:52Z","lastTransitionTime":"2025-12-13T06:49:52Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 13 06:49:52 crc kubenswrapper[4971]: I1213 06:49:52.415620 4971 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 13 06:49:52 crc kubenswrapper[4971]: I1213 06:49:52.415673 4971 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 13 06:49:52 crc kubenswrapper[4971]: I1213 06:49:52.415688 4971 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 13 06:49:52 crc kubenswrapper[4971]: I1213 06:49:52.415705 4971 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 13 06:49:52 crc kubenswrapper[4971]: I1213 06:49:52.415718 4971 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-13T06:49:52Z","lastTransitionTime":"2025-12-13T06:49:52Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 13 06:49:52 crc kubenswrapper[4971]: I1213 06:49:52.519071 4971 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 13 06:49:52 crc kubenswrapper[4971]: I1213 06:49:52.519147 4971 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 13 06:49:52 crc kubenswrapper[4971]: I1213 06:49:52.519160 4971 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 13 06:49:52 crc kubenswrapper[4971]: I1213 06:49:52.519193 4971 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 13 06:49:52 crc kubenswrapper[4971]: I1213 06:49:52.519212 4971 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-13T06:49:52Z","lastTransitionTime":"2025-12-13T06:49:52Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 13 06:49:52 crc kubenswrapper[4971]: I1213 06:49:52.622430 4971 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 13 06:49:52 crc kubenswrapper[4971]: I1213 06:49:52.622491 4971 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 13 06:49:52 crc kubenswrapper[4971]: I1213 06:49:52.622502 4971 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 13 06:49:52 crc kubenswrapper[4971]: I1213 06:49:52.622545 4971 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 13 06:49:52 crc kubenswrapper[4971]: I1213 06:49:52.622560 4971 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-13T06:49:52Z","lastTransitionTime":"2025-12-13T06:49:52Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 13 06:49:52 crc kubenswrapper[4971]: I1213 06:49:52.725140 4971 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 13 06:49:52 crc kubenswrapper[4971]: I1213 06:49:52.725181 4971 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 13 06:49:52 crc kubenswrapper[4971]: I1213 06:49:52.725190 4971 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 13 06:49:52 crc kubenswrapper[4971]: I1213 06:49:52.725208 4971 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 13 06:49:52 crc kubenswrapper[4971]: I1213 06:49:52.725217 4971 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-13T06:49:52Z","lastTransitionTime":"2025-12-13T06:49:52Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 13 06:49:52 crc kubenswrapper[4971]: I1213 06:49:52.827691 4971 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 13 06:49:52 crc kubenswrapper[4971]: I1213 06:49:52.827743 4971 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 13 06:49:52 crc kubenswrapper[4971]: I1213 06:49:52.827755 4971 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 13 06:49:52 crc kubenswrapper[4971]: I1213 06:49:52.827772 4971 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 13 06:49:52 crc kubenswrapper[4971]: I1213 06:49:52.827785 4971 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-13T06:49:52Z","lastTransitionTime":"2025-12-13T06:49:52Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 13 06:49:52 crc kubenswrapper[4971]: I1213 06:49:52.930619 4971 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 13 06:49:52 crc kubenswrapper[4971]: I1213 06:49:52.930670 4971 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 13 06:49:52 crc kubenswrapper[4971]: I1213 06:49:52.930689 4971 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 13 06:49:52 crc kubenswrapper[4971]: I1213 06:49:52.930708 4971 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 13 06:49:52 crc kubenswrapper[4971]: I1213 06:49:52.930720 4971 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-13T06:49:52Z","lastTransitionTime":"2025-12-13T06:49:52Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 13 06:49:53 crc kubenswrapper[4971]: I1213 06:49:53.032565 4971 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 13 06:49:53 crc kubenswrapper[4971]: I1213 06:49:53.032596 4971 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 13 06:49:53 crc kubenswrapper[4971]: I1213 06:49:53.032604 4971 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 13 06:49:53 crc kubenswrapper[4971]: I1213 06:49:53.032617 4971 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 13 06:49:53 crc kubenswrapper[4971]: I1213 06:49:53.032626 4971 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-13T06:49:53Z","lastTransitionTime":"2025-12-13T06:49:53Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 13 06:49:53 crc kubenswrapper[4971]: I1213 06:49:53.135365 4971 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 13 06:49:53 crc kubenswrapper[4971]: I1213 06:49:53.135409 4971 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 13 06:49:53 crc kubenswrapper[4971]: I1213 06:49:53.135424 4971 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 13 06:49:53 crc kubenswrapper[4971]: I1213 06:49:53.135449 4971 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 13 06:49:53 crc kubenswrapper[4971]: I1213 06:49:53.135463 4971 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-13T06:49:53Z","lastTransitionTime":"2025-12-13T06:49:53Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 13 06:49:53 crc kubenswrapper[4971]: I1213 06:49:53.238456 4971 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 13 06:49:53 crc kubenswrapper[4971]: I1213 06:49:53.238899 4971 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 13 06:49:53 crc kubenswrapper[4971]: I1213 06:49:53.238999 4971 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 13 06:49:53 crc kubenswrapper[4971]: I1213 06:49:53.239086 4971 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 13 06:49:53 crc kubenswrapper[4971]: I1213 06:49:53.239148 4971 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-13T06:49:53Z","lastTransitionTime":"2025-12-13T06:49:53Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 13 06:49:53 crc kubenswrapper[4971]: I1213 06:49:53.341811 4971 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 13 06:49:53 crc kubenswrapper[4971]: I1213 06:49:53.342117 4971 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 13 06:49:53 crc kubenswrapper[4971]: I1213 06:49:53.342209 4971 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 13 06:49:53 crc kubenswrapper[4971]: I1213 06:49:53.342304 4971 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 13 06:49:53 crc kubenswrapper[4971]: I1213 06:49:53.342392 4971 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-13T06:49:53Z","lastTransitionTime":"2025-12-13T06:49:53Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 13 06:49:53 crc kubenswrapper[4971]: I1213 06:49:53.445404 4971 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 13 06:49:53 crc kubenswrapper[4971]: I1213 06:49:53.445439 4971 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 13 06:49:53 crc kubenswrapper[4971]: I1213 06:49:53.445450 4971 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 13 06:49:53 crc kubenswrapper[4971]: I1213 06:49:53.445467 4971 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 13 06:49:53 crc kubenswrapper[4971]: I1213 06:49:53.445477 4971 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-13T06:49:53Z","lastTransitionTime":"2025-12-13T06:49:53Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 13 06:49:53 crc kubenswrapper[4971]: I1213 06:49:53.547753 4971 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 13 06:49:53 crc kubenswrapper[4971]: I1213 06:49:53.547826 4971 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 13 06:49:53 crc kubenswrapper[4971]: I1213 06:49:53.547847 4971 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 13 06:49:53 crc kubenswrapper[4971]: I1213 06:49:53.547876 4971 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 13 06:49:53 crc kubenswrapper[4971]: I1213 06:49:53.547897 4971 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-13T06:49:53Z","lastTransitionTime":"2025-12-13T06:49:53Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 13 06:49:53 crc kubenswrapper[4971]: I1213 06:49:53.650359 4971 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 13 06:49:53 crc kubenswrapper[4971]: I1213 06:49:53.650404 4971 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 13 06:49:53 crc kubenswrapper[4971]: I1213 06:49:53.650414 4971 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 13 06:49:53 crc kubenswrapper[4971]: I1213 06:49:53.650431 4971 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 13 06:49:53 crc kubenswrapper[4971]: I1213 06:49:53.650439 4971 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-13T06:49:53Z","lastTransitionTime":"2025-12-13T06:49:53Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 13 06:49:53 crc kubenswrapper[4971]: I1213 06:49:53.752863 4971 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 13 06:49:53 crc kubenswrapper[4971]: I1213 06:49:53.752908 4971 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 13 06:49:53 crc kubenswrapper[4971]: I1213 06:49:53.752919 4971 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 13 06:49:53 crc kubenswrapper[4971]: I1213 06:49:53.752936 4971 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 13 06:49:53 crc kubenswrapper[4971]: I1213 06:49:53.752948 4971 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-13T06:49:53Z","lastTransitionTime":"2025-12-13T06:49:53Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 13 06:49:53 crc kubenswrapper[4971]: I1213 06:49:53.768610 4971 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 13 06:49:53 crc kubenswrapper[4971]: I1213 06:49:53.768819 4971 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 13 06:49:53 crc kubenswrapper[4971]: I1213 06:49:53.768824 4971 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 13 06:49:53 crc kubenswrapper[4971]: I1213 06:49:53.768860 4971 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-x98k7" Dec 13 06:49:53 crc kubenswrapper[4971]: E1213 06:49:53.769632 4971 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 13 06:49:53 crc kubenswrapper[4971]: E1213 06:49:53.769726 4971 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 13 06:49:53 crc kubenswrapper[4971]: E1213 06:49:53.769780 4971 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 13 06:49:53 crc kubenswrapper[4971]: E1213 06:49:53.769910 4971 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-x98k7" podUID="c7185dec-3240-42e8-9d90-6182a5e8b65c" Dec 13 06:49:53 crc kubenswrapper[4971]: I1213 06:49:53.789879 4971 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3ad3268a-a982-428c-9c06-4aaeddd372d9\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T06:49:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T06:49:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T06:49:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T06:49:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T06:49:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3716afdf2bc693be80a09fc9f7727d334f435caf1d4c8ec4811b3a1f1652cdfe\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-13T06:49:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://30df3abf8c79db8af52c6a918f73e1ce410ddafcf341636424c267e1eeb195a7\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-13T06:49:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://ccdf65f6d2b1eb3647a03ba37c305c32e34addfdf79ff937c8dddaa3e5d5b5c6\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-13T06:49:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://770b73dd0c2b6f38744e2c1cd6aa186a69bb87d1cd015a603ad2e33be06a8f6b\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://b3c1bab3d53f975b2e24582ce2771bc26d600e0dd7c30339dcfd18a071162ad4\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-12-13T06:49:23Z\\\",\\\"message\\\":\\\"file observer\\\\nW1213 06:49:23.283403 1 builder.go:272] unable to get owner reference (falling back to namespace): pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\nI1213 06:49:23.284336 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1213 06:49:23.286978 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-981773671/tls.crt::/tmp/serving-cert-981773671/tls.key\\\\\\\"\\\\nI1213 06:49:23.775072 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1213 06:49:23.778615 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1213 06:49:23.778645 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1213 06:49:23.778680 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1213 06:49:23.778691 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1213 06:49:23.785058 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1213 06:49:23.786150 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1213 06:49:23.786196 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1213 06:49:23.786223 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1213 06:49:23.786246 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1213 06:49:23.786269 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1213 06:49:23.786291 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI1213 06:49:23.785608 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF1213 06:49:23.787563 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-13T06:49:08Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-13T06:49:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://9acd07066deed483bd7fa61ee45fda09539d110717a134a56827f871186f92fc\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-13T06:49:07Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://38c91cfbaecfdbd3e3b663422a91ab4294610afb6bbf9ebdb77db24e2b76c682\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://38c91cfbaecfdbd3e3b663422a91ab4294610afb6bbf9ebdb77db24e2b76c682\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-13T06:49:05Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-13T06:49:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-13T06:49:03Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-13T06:49:53Z is after 2025-08-24T17:21:41Z" Dec 13 06:49:53 crc kubenswrapper[4971]: I1213 06:49:53.804006 4971 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-13T06:49:23Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-13T06:49:23Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-13T06:49:53Z is after 2025-08-24T17:21:41Z" Dec 13 06:49:53 crc kubenswrapper[4971]: I1213 06:49:53.813454 4971 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-13T06:49:23Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-13T06:49:23Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-13T06:49:53Z is after 2025-08-24T17:21:41Z" Dec 13 06:49:53 crc kubenswrapper[4971]: I1213 06:49:53.825861 4971 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-wwvpl" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"dddf744b-2122-49ad-a980-105c09636e0f\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T06:49:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T06:49:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T06:49:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T06:49:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4676d897c7fc668587faec563f3551a056557879cee3e865192d2bd28edf62fc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-13T06:49:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-czz9v\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-13T06:49:25Z\\\"}}\" for pod \"openshift-multus\"/\"multus-wwvpl\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-13T06:49:53Z is after 2025-08-24T17:21:41Z" Dec 13 06:49:53 crc kubenswrapper[4971]: I1213 06:49:53.849896 4971 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-zlcg5" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"736c6b1a-93dd-4f71-81b6-1c43fcf5f556\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T06:49:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T06:49:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T06:49:25Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T06:49:25Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://325223543b3c7db417f769533559a1ca2b23a723494e1885b0fd0480b5a6399f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-13T06:49:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vzmsr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://f22a4ec756119c4f791fb5105cf1f31df3e548b2b07b1f909a62802e9b6035ae\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-13T06:49:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vzmsr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://f6073140d9f86929b6d4bd2960ede0617a5c9620b269bbab043bae37e0d389f5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-13T06:49:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vzmsr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://693a2ab434c250cb140a866e8068b156df9191414fed1d63bee919a2528ea6e5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-13T06:49:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vzmsr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://219620f51474e60b3e1f880c0972eb612dffc4bad24462bbba2c093ae817d852\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-13T06:49:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vzmsr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://f137f4b441c2402093ab430a9689a46221c4fb4bc70d99fb992f62b0a989c465\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-13T06:49:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vzmsr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://d950a29b0f6f47ae1cc90c132adb059b5aff889a0a38de1d1296e0f083da5e51\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d950a29b0f6f47ae1cc90c132adb059b5aff889a0a38de1d1296e0f083da5e51\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-12-13T06:49:37Z\\\",\\\"message\\\":\\\" ovnkube.go:137] failed to run ovnkube: [failed to start network controller: failed to start default network controller: unable to create admin network policy controller, err: could not add Event Handler for anpInformer during admin network policy controller initialization, handler {0x1fcc6e0 0x1fcc3c0 0x1fcc360} was not added to shared informer because it has stopped already, failed to start node network controller: failed to start default node network controller: failed to set node crc annotations: Internal error occurred: failed calling webhook \\\\\\\"node.network-node-identity.openshift.io\\\\\\\": failed to call webhook: Post \\\\\\\"https://127.0.0.1:9743/node?timeout=10s\\\\\\\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-13T06:49:36Z is after 2025-08-24T17:21:41Z]\\\\nI1213 06:49:36.767920 6409 model_client.go:382] Update operations generated as: [{Op:update Table:Logical_Switch_Port Row:map[addresses:{GoSet:[0a:58:0a:d9:00:04 10.217.0.4]} options:{GoMap:map[iface-id-ver:3b6479f0-333b-4a96-9adf-2099afdc2447 requested-chassis:crc]} port_security:{GoSet:[0a:58:0a:d9:00:04 10.217.0.4]}] Rows:[] Columns:[] Mutations:[] Timeout:\\\\u003cnil\\\\u003e Where:[where column _uuid == {61897e97-c771-4738-8709-09636387cb00}] Until: Durable:\\\\u003cnil\\\\u003e Comment:\\\\u003cnil\\\\u003e Lock:\\\\u003cnil\\\\u003e UUID: UUIDName:}]\\\\nI1213 06:49:36.767917 6409 model_clien\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-13T06:49:36Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 10s restarting failed container=ovnkube-controller pod=ovnkube-node-zlcg5_openshift-ovn-kubernetes(736c6b1a-93dd-4f71-81b6-1c43fcf5f556)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vzmsr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://868049c6a3a4ebe4036d1d22e02ba6b76bced2ee7e269bb228accc4a0636678a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-13T06:49:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vzmsr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://451b22df6973ddc8fa2209a65d64d9e41b33eb231bb4adfab88cf0e5943bb246\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://451b22df6973ddc8fa2209a65d64d9e41b33eb231bb4adfab88cf0e5943bb246\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-13T06:49:26Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-13T06:49:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vzmsr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-13T06:49:25Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-zlcg5\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-13T06:49:53Z is after 2025-08-24T17:21:41Z" Dec 13 06:49:53 crc kubenswrapper[4971]: I1213 06:49:53.856408 4971 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 13 06:49:53 crc kubenswrapper[4971]: I1213 06:49:53.856469 4971 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 13 06:49:53 crc kubenswrapper[4971]: I1213 06:49:53.856481 4971 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 13 06:49:53 crc kubenswrapper[4971]: I1213 06:49:53.856501 4971 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 13 06:49:53 crc kubenswrapper[4971]: I1213 06:49:53.856533 4971 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-13T06:49:53Z","lastTransitionTime":"2025-12-13T06:49:53Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 13 06:49:53 crc kubenswrapper[4971]: I1213 06:49:53.868386 4971 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-13T06:49:24Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-13T06:49:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-13T06:49:24Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d0b2ecf0d5b22f23217f42f259ae7ffb6359555b22c30eaebf667eaf254b3bab\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-13T06:49:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-13T06:49:53Z is after 2025-08-24T17:21:41Z" Dec 13 06:49:53 crc kubenswrapper[4971]: I1213 06:49:53.880763 4971 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-13T06:49:23Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-13T06:49:23Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-13T06:49:53Z is after 2025-08-24T17:21:41Z" Dec 13 06:49:53 crc kubenswrapper[4971]: I1213 06:49:53.891090 4971 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-13T06:49:26Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-13T06:49:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-13T06:49:26Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9c8a0d1db0e754462d805e82daf5f52ffd5457507a30456535ec59e096c9d092\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-13T06:49:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-13T06:49:53Z is after 2025-08-24T17:21:41Z" Dec 13 06:49:53 crc kubenswrapper[4971]: I1213 06:49:53.903743 4971 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-82xjz" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e0c4af12-edda-41d7-9b44-f87396174ef2\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T06:49:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T06:49:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T06:49:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T06:49:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://0f4577d9e37f860813e221f2e208177d8d383fd04f673ce357a0a405ea29d1e6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-13T06:49:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cnxsq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://7e0d2020e62f1a277cff7047e1907dc747ed393a1f6c983e1ee258bd156c16c7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-13T06:49:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cnxsq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-13T06:49:25Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-82xjz\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-13T06:49:53Z is after 2025-08-24T17:21:41Z" Dec 13 06:49:53 crc kubenswrapper[4971]: I1213 06:49:53.914177 4971 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-x98k7" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"c7185dec-3240-42e8-9d90-6182a5e8b65c\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T06:49:39Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T06:49:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T06:49:39Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T06:49:39Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bj2gn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bj2gn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-13T06:49:39Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-x98k7\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-13T06:49:53Z is after 2025-08-24T17:21:41Z" Dec 13 06:49:53 crc kubenswrapper[4971]: I1213 06:49:53.926696 4971 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"f9d31b5a-b770-4a15-95c8-35b49677c608\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T06:49:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T06:49:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T06:49:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T06:49:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T06:49:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c9baa042f9dada954faa57b5454cd6b0a12a297e56ddc192c52cff50e42a16fa\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-13T06:49:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://c788de86754f0bf37ee184672fc1dcb388ccebc246d4f15be4b5da661f725f8f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-13T06:49:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://760817511a9b4f96a68dc38d26631250eaab72e1dd8f123e653039b6b4e91f84\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-13T06:49:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b859be0c9feac856e9855c6a038b02b49b32fade97873c46d0bec0a32d8b5147\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-13T06:49:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-13T06:49:03Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-13T06:49:53Z is after 2025-08-24T17:21:41Z" Dec 13 06:49:53 crc kubenswrapper[4971]: I1213 06:49:53.940996 4971 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-ncchz" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"de914d21-a688-401e-a66e-861beb6336d9\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T06:49:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T06:49:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T06:49:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T06:49:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://244ffe3dc9d8cf799aff4c4d795bf41554c1e6bb6667fbb0c7df322fb5f96c68\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-13T06:49:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lqmjj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://0b7111af6c5a198d02bb02858837cfb4f3b20828d8d206cd18aa0ffa102d0cd4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://0b7111af6c5a198d02bb02858837cfb4f3b20828d8d206cd18aa0ffa102d0cd4\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-13T06:49:26Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-13T06:49:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lqmjj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b36777fe40e74ae1cecb7b73e204139db5279ba51cef4f1e8c4a4c491d8be403\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://b36777fe40e74ae1cecb7b73e204139db5279ba51cef4f1e8c4a4c491d8be403\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-13T06:49:28Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-13T06:49:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lqmjj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://0597a06cc663e416142311a16941d98297577693a1fe27201ea42c27496cace2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://0597a06cc663e416142311a16941d98297577693a1fe27201ea42c27496cace2\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-13T06:49:29Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-13T06:49:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lqmjj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://3f52c2e22a75c010993ca5809a3a9d1304758d4cb6e83691ef1af15a4b514ac4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://3f52c2e22a75c010993ca5809a3a9d1304758d4cb6e83691ef1af15a4b514ac4\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-13T06:49:30Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-13T06:49:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lqmjj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://9e21ca708c4c4c1228b417cece51299e0792729b7c78884bd18b2dda8fe8355c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://9e21ca708c4c4c1228b417cece51299e0792729b7c78884bd18b2dda8fe8355c\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-13T06:49:31Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-13T06:49:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lqmjj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://74552159f12c87cdda09147692b400ccac0d4e43c1ed72fe8462c4249ad69c7b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://74552159f12c87cdda09147692b400ccac0d4e43c1ed72fe8462c4249ad69c7b\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-13T06:49:32Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-13T06:49:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lqmjj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-13T06:49:25Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-ncchz\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-13T06:49:53Z is after 2025-08-24T17:21:41Z" Dec 13 06:49:53 crc kubenswrapper[4971]: I1213 06:49:53.951263 4971 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-gkcd8" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"13f5c327-f95e-42b4-8783-218500c17694\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T06:49:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T06:49:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T06:49:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T06:49:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://77e846e36f1cb40d03471dc03de11b6c3da448baa407aad38093fae9a801cf92\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-13T06:49:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gvlhn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-13T06:49:27Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-gkcd8\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-13T06:49:53Z is after 2025-08-24T17:21:41Z" Dec 13 06:49:53 crc kubenswrapper[4971]: I1213 06:49:53.958462 4971 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 13 06:49:53 crc kubenswrapper[4971]: I1213 06:49:53.958555 4971 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 13 06:49:53 crc kubenswrapper[4971]: I1213 06:49:53.958571 4971 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 13 06:49:53 crc kubenswrapper[4971]: I1213 06:49:53.958595 4971 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 13 06:49:53 crc kubenswrapper[4971]: I1213 06:49:53.958611 4971 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-13T06:49:53Z","lastTransitionTime":"2025-12-13T06:49:53Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 13 06:49:53 crc kubenswrapper[4971]: I1213 06:49:53.966405 4971 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-z572s" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"b4146d3d-4034-49a9-bd37-164856255137\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T06:49:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T06:49:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T06:49:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T06:49:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://46fd4e1070835cfbf7f35d497eff7f047473fc4048ee07812925df5ebdf55342\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-13T06:49:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-j6l9b\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://dc79a0968f85723e6ad0dc269c8da1f7472adb1bb2b87edd017ee2295bfbf89c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-13T06:49:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-j6l9b\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-13T06:49:38Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-z572s\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-13T06:49:53Z is after 2025-08-24T17:21:41Z" Dec 13 06:49:53 crc kubenswrapper[4971]: I1213 06:49:53.986662 4971 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"1ed9f82e-fe5c-4b76-a619-62e5abd88d3b\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T06:49:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T06:49:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T06:49:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T06:49:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T06:49:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e457329a78a2d24337f994e8a0a1f75a2a667244aa75eb84613d1e7138364aac\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-13T06:49:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://eb014de8140deef097d9bd94f3b9fa4ab3558bc4449bef84b83bf62bfabf5836\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-13T06:49:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://2c438cc74d555f7c9db4569d1c56b78b1088ed8105697d2c0f31e6c5136f3dd0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-13T06:49:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://aa6fa6e4340dcd066252df3c56d79a3a39ed531d5eea8d1d5941d6952368d3af\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-13T06:49:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://c4471b98854c52d5cd084d0ba3e64e5b08e9518ec3c372cce2c875a440522457\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-13T06:49:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7d2746335c98dae3a0f43024d34cb62ebcb42a01c8f9ce18d52b29d4c45598e5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://7d2746335c98dae3a0f43024d34cb62ebcb42a01c8f9ce18d52b29d4c45598e5\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-13T06:49:04Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-13T06:49:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://bc0e77d10a738128bea82122496bc42235fc1688f8907756d93e61b651902b57\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://bc0e77d10a738128bea82122496bc42235fc1688f8907756d93e61b651902b57\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-13T06:49:05Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-13T06:49:05Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://f821e418cba5e832da8839268bcbbd9630bdf319a212d5382bfa2c4b0b73a268\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://f821e418cba5e832da8839268bcbbd9630bdf319a212d5382bfa2c4b0b73a268\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-13T06:49:07Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-13T06:49:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-13T06:49:03Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-13T06:49:53Z is after 2025-08-24T17:21:41Z" Dec 13 06:49:54 crc kubenswrapper[4971]: I1213 06:49:54.001337 4971 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-13T06:49:24Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-13T06:49:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-13T06:49:24Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4328cd6a8464ed468229bbfbc8265f969427916d72ebb92237dbe68efe70fa47\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-13T06:49:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://9d715ca82d9b5a594e41f2b98da6f4277da0e4e1dec114969f901c00f6d7553d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-13T06:49:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-13T06:49:53Z is after 2025-08-24T17:21:41Z" Dec 13 06:49:54 crc kubenswrapper[4971]: I1213 06:49:54.011629 4971 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-5rzh5" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"6bf4591d-2483-4132-9672-d0fd6c995968\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T06:49:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T06:49:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T06:49:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T06:49:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://50e2cb14e4ffb603acb06e2281615a99f06182d345c4e70f2446fa2077033bbb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-13T06:49:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dfxqn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-13T06:49:25Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-5rzh5\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-13T06:49:54Z is after 2025-08-24T17:21:41Z" Dec 13 06:49:54 crc kubenswrapper[4971]: I1213 06:49:54.061424 4971 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 13 06:49:54 crc kubenswrapper[4971]: I1213 06:49:54.061475 4971 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 13 06:49:54 crc kubenswrapper[4971]: I1213 06:49:54.061494 4971 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 13 06:49:54 crc kubenswrapper[4971]: I1213 06:49:54.061539 4971 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 13 06:49:54 crc kubenswrapper[4971]: I1213 06:49:54.061563 4971 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-13T06:49:54Z","lastTransitionTime":"2025-12-13T06:49:54Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 13 06:49:54 crc kubenswrapper[4971]: I1213 06:49:54.163383 4971 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 13 06:49:54 crc kubenswrapper[4971]: I1213 06:49:54.163411 4971 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 13 06:49:54 crc kubenswrapper[4971]: I1213 06:49:54.163419 4971 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 13 06:49:54 crc kubenswrapper[4971]: I1213 06:49:54.163432 4971 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 13 06:49:54 crc kubenswrapper[4971]: I1213 06:49:54.163441 4971 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-13T06:49:54Z","lastTransitionTime":"2025-12-13T06:49:54Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 13 06:49:54 crc kubenswrapper[4971]: I1213 06:49:54.265289 4971 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 13 06:49:54 crc kubenswrapper[4971]: I1213 06:49:54.265335 4971 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 13 06:49:54 crc kubenswrapper[4971]: I1213 06:49:54.265348 4971 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 13 06:49:54 crc kubenswrapper[4971]: I1213 06:49:54.265365 4971 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 13 06:49:54 crc kubenswrapper[4971]: I1213 06:49:54.265376 4971 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-13T06:49:54Z","lastTransitionTime":"2025-12-13T06:49:54Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 13 06:49:54 crc kubenswrapper[4971]: I1213 06:49:54.368279 4971 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 13 06:49:54 crc kubenswrapper[4971]: I1213 06:49:54.368348 4971 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 13 06:49:54 crc kubenswrapper[4971]: I1213 06:49:54.368372 4971 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 13 06:49:54 crc kubenswrapper[4971]: I1213 06:49:54.368398 4971 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 13 06:49:54 crc kubenswrapper[4971]: I1213 06:49:54.368419 4971 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-13T06:49:54Z","lastTransitionTime":"2025-12-13T06:49:54Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 13 06:49:54 crc kubenswrapper[4971]: I1213 06:49:54.471589 4971 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 13 06:49:54 crc kubenswrapper[4971]: I1213 06:49:54.471639 4971 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 13 06:49:54 crc kubenswrapper[4971]: I1213 06:49:54.471652 4971 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 13 06:49:54 crc kubenswrapper[4971]: I1213 06:49:54.471672 4971 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 13 06:49:54 crc kubenswrapper[4971]: I1213 06:49:54.471685 4971 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-13T06:49:54Z","lastTransitionTime":"2025-12-13T06:49:54Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 13 06:49:54 crc kubenswrapper[4971]: I1213 06:49:54.573900 4971 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 13 06:49:54 crc kubenswrapper[4971]: I1213 06:49:54.573936 4971 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 13 06:49:54 crc kubenswrapper[4971]: I1213 06:49:54.573944 4971 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 13 06:49:54 crc kubenswrapper[4971]: I1213 06:49:54.573962 4971 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 13 06:49:54 crc kubenswrapper[4971]: I1213 06:49:54.573971 4971 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-13T06:49:54Z","lastTransitionTime":"2025-12-13T06:49:54Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 13 06:49:54 crc kubenswrapper[4971]: I1213 06:49:54.676498 4971 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 13 06:49:54 crc kubenswrapper[4971]: I1213 06:49:54.676581 4971 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 13 06:49:54 crc kubenswrapper[4971]: I1213 06:49:54.676598 4971 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 13 06:49:54 crc kubenswrapper[4971]: I1213 06:49:54.676624 4971 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 13 06:49:54 crc kubenswrapper[4971]: I1213 06:49:54.676641 4971 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-13T06:49:54Z","lastTransitionTime":"2025-12-13T06:49:54Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 13 06:49:54 crc kubenswrapper[4971]: I1213 06:49:54.779318 4971 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 13 06:49:54 crc kubenswrapper[4971]: I1213 06:49:54.779358 4971 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 13 06:49:54 crc kubenswrapper[4971]: I1213 06:49:54.779368 4971 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 13 06:49:54 crc kubenswrapper[4971]: I1213 06:49:54.779383 4971 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 13 06:49:54 crc kubenswrapper[4971]: I1213 06:49:54.779391 4971 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-13T06:49:54Z","lastTransitionTime":"2025-12-13T06:49:54Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 13 06:49:54 crc kubenswrapper[4971]: I1213 06:49:54.881819 4971 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 13 06:49:54 crc kubenswrapper[4971]: I1213 06:49:54.881871 4971 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 13 06:49:54 crc kubenswrapper[4971]: I1213 06:49:54.881880 4971 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 13 06:49:54 crc kubenswrapper[4971]: I1213 06:49:54.881894 4971 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 13 06:49:54 crc kubenswrapper[4971]: I1213 06:49:54.881903 4971 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-13T06:49:54Z","lastTransitionTime":"2025-12-13T06:49:54Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 13 06:49:54 crc kubenswrapper[4971]: I1213 06:49:54.984306 4971 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 13 06:49:54 crc kubenswrapper[4971]: I1213 06:49:54.984351 4971 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 13 06:49:54 crc kubenswrapper[4971]: I1213 06:49:54.984360 4971 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 13 06:49:54 crc kubenswrapper[4971]: I1213 06:49:54.984377 4971 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 13 06:49:54 crc kubenswrapper[4971]: I1213 06:49:54.984386 4971 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-13T06:49:54Z","lastTransitionTime":"2025-12-13T06:49:54Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 13 06:49:55 crc kubenswrapper[4971]: I1213 06:49:55.086982 4971 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 13 06:49:55 crc kubenswrapper[4971]: I1213 06:49:55.087026 4971 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 13 06:49:55 crc kubenswrapper[4971]: I1213 06:49:55.087040 4971 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 13 06:49:55 crc kubenswrapper[4971]: I1213 06:49:55.087061 4971 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 13 06:49:55 crc kubenswrapper[4971]: I1213 06:49:55.087076 4971 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-13T06:49:55Z","lastTransitionTime":"2025-12-13T06:49:55Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 13 06:49:55 crc kubenswrapper[4971]: I1213 06:49:55.189509 4971 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 13 06:49:55 crc kubenswrapper[4971]: I1213 06:49:55.189618 4971 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 13 06:49:55 crc kubenswrapper[4971]: I1213 06:49:55.189653 4971 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 13 06:49:55 crc kubenswrapper[4971]: I1213 06:49:55.189689 4971 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 13 06:49:55 crc kubenswrapper[4971]: I1213 06:49:55.189710 4971 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-13T06:49:55Z","lastTransitionTime":"2025-12-13T06:49:55Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 13 06:49:55 crc kubenswrapper[4971]: I1213 06:49:55.292147 4971 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 13 06:49:55 crc kubenswrapper[4971]: I1213 06:49:55.292220 4971 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 13 06:49:55 crc kubenswrapper[4971]: I1213 06:49:55.292245 4971 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 13 06:49:55 crc kubenswrapper[4971]: I1213 06:49:55.292275 4971 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 13 06:49:55 crc kubenswrapper[4971]: I1213 06:49:55.292298 4971 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-13T06:49:55Z","lastTransitionTime":"2025-12-13T06:49:55Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 13 06:49:55 crc kubenswrapper[4971]: I1213 06:49:55.307840 4971 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/c7185dec-3240-42e8-9d90-6182a5e8b65c-metrics-certs\") pod \"network-metrics-daemon-x98k7\" (UID: \"c7185dec-3240-42e8-9d90-6182a5e8b65c\") " pod="openshift-multus/network-metrics-daemon-x98k7" Dec 13 06:49:55 crc kubenswrapper[4971]: E1213 06:49:55.308036 4971 secret.go:188] Couldn't get secret openshift-multus/metrics-daemon-secret: object "openshift-multus"/"metrics-daemon-secret" not registered Dec 13 06:49:55 crc kubenswrapper[4971]: E1213 06:49:55.308102 4971 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/c7185dec-3240-42e8-9d90-6182a5e8b65c-metrics-certs podName:c7185dec-3240-42e8-9d90-6182a5e8b65c nodeName:}" failed. No retries permitted until 2025-12-13 06:50:11.308082736 +0000 UTC m=+67.912492184 (durationBeforeRetry 16s). Error: MountVolume.SetUp failed for volume "metrics-certs" (UniqueName: "kubernetes.io/secret/c7185dec-3240-42e8-9d90-6182a5e8b65c-metrics-certs") pod "network-metrics-daemon-x98k7" (UID: "c7185dec-3240-42e8-9d90-6182a5e8b65c") : object "openshift-multus"/"metrics-daemon-secret" not registered Dec 13 06:49:55 crc kubenswrapper[4971]: I1213 06:49:55.394496 4971 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 13 06:49:55 crc kubenswrapper[4971]: I1213 06:49:55.394562 4971 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 13 06:49:55 crc kubenswrapper[4971]: I1213 06:49:55.394571 4971 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 13 06:49:55 crc kubenswrapper[4971]: I1213 06:49:55.394587 4971 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 13 06:49:55 crc kubenswrapper[4971]: I1213 06:49:55.394596 4971 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-13T06:49:55Z","lastTransitionTime":"2025-12-13T06:49:55Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 13 06:49:55 crc kubenswrapper[4971]: I1213 06:49:55.497853 4971 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 13 06:49:55 crc kubenswrapper[4971]: I1213 06:49:55.497906 4971 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 13 06:49:55 crc kubenswrapper[4971]: I1213 06:49:55.497939 4971 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 13 06:49:55 crc kubenswrapper[4971]: I1213 06:49:55.497959 4971 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 13 06:49:55 crc kubenswrapper[4971]: I1213 06:49:55.497970 4971 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-13T06:49:55Z","lastTransitionTime":"2025-12-13T06:49:55Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 13 06:49:55 crc kubenswrapper[4971]: I1213 06:49:55.510599 4971 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 13 06:49:55 crc kubenswrapper[4971]: E1213 06:49:55.510813 4971 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-13 06:50:27.510777357 +0000 UTC m=+84.115186815 (durationBeforeRetry 32s). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 13 06:49:55 crc kubenswrapper[4971]: I1213 06:49:55.510957 4971 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"networking-console-plugin-cert\" (UniqueName: \"kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 13 06:49:55 crc kubenswrapper[4971]: I1213 06:49:55.511013 4971 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nginx-conf\" (UniqueName: \"kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 13 06:49:55 crc kubenswrapper[4971]: E1213 06:49:55.511140 4971 secret.go:188] Couldn't get secret openshift-network-console/networking-console-plugin-cert: object "openshift-network-console"/"networking-console-plugin-cert" not registered Dec 13 06:49:55 crc kubenswrapper[4971]: E1213 06:49:55.511185 4971 configmap.go:193] Couldn't get configMap openshift-network-console/networking-console-plugin: object "openshift-network-console"/"networking-console-plugin" not registered Dec 13 06:49:55 crc kubenswrapper[4971]: E1213 06:49:55.511229 4971 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2025-12-13 06:50:27.511208448 +0000 UTC m=+84.115618046 (durationBeforeRetry 32s). Error: MountVolume.SetUp failed for volume "networking-console-plugin-cert" (UniqueName: "kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin-cert" not registered Dec 13 06:49:55 crc kubenswrapper[4971]: E1213 06:49:55.511289 4971 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2025-12-13 06:50:27.51126024 +0000 UTC m=+84.115669818 (durationBeforeRetry 32s). Error: MountVolume.SetUp failed for volume "nginx-conf" (UniqueName: "kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin" not registered Dec 13 06:49:55 crc kubenswrapper[4971]: I1213 06:49:55.600286 4971 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 13 06:49:55 crc kubenswrapper[4971]: I1213 06:49:55.600322 4971 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 13 06:49:55 crc kubenswrapper[4971]: I1213 06:49:55.600333 4971 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 13 06:49:55 crc kubenswrapper[4971]: I1213 06:49:55.600349 4971 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 13 06:49:55 crc kubenswrapper[4971]: I1213 06:49:55.600361 4971 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-13T06:49:55Z","lastTransitionTime":"2025-12-13T06:49:55Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 13 06:49:55 crc kubenswrapper[4971]: I1213 06:49:55.611915 4971 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cqllr\" (UniqueName: \"kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr\") pod \"network-check-target-xd92c\" (UID: \"3b6479f0-333b-4a96-9adf-2099afdc2447\") " pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 13 06:49:55 crc kubenswrapper[4971]: I1213 06:49:55.611977 4971 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-s2dwl\" (UniqueName: \"kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl\") pod \"network-check-source-55646444c4-trplf\" (UID: \"9d751cbb-f2e2-430d-9754-c882a5e924a5\") " pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 13 06:49:55 crc kubenswrapper[4971]: E1213 06:49:55.612096 4971 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Dec 13 06:49:55 crc kubenswrapper[4971]: E1213 06:49:55.612117 4971 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Dec 13 06:49:55 crc kubenswrapper[4971]: E1213 06:49:55.612129 4971 projected.go:194] Error preparing data for projected volume kube-api-access-s2dwl for pod openshift-network-diagnostics/network-check-source-55646444c4-trplf: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Dec 13 06:49:55 crc kubenswrapper[4971]: E1213 06:49:55.612179 4971 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl podName:9d751cbb-f2e2-430d-9754-c882a5e924a5 nodeName:}" failed. No retries permitted until 2025-12-13 06:50:27.61216375 +0000 UTC m=+84.216573198 (durationBeforeRetry 32s). Error: MountVolume.SetUp failed for volume "kube-api-access-s2dwl" (UniqueName: "kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl") pod "network-check-source-55646444c4-trplf" (UID: "9d751cbb-f2e2-430d-9754-c882a5e924a5") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Dec 13 06:49:55 crc kubenswrapper[4971]: E1213 06:49:55.612238 4971 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Dec 13 06:49:55 crc kubenswrapper[4971]: E1213 06:49:55.612280 4971 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Dec 13 06:49:55 crc kubenswrapper[4971]: E1213 06:49:55.612300 4971 projected.go:194] Error preparing data for projected volume kube-api-access-cqllr for pod openshift-network-diagnostics/network-check-target-xd92c: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Dec 13 06:49:55 crc kubenswrapper[4971]: E1213 06:49:55.612373 4971 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr podName:3b6479f0-333b-4a96-9adf-2099afdc2447 nodeName:}" failed. No retries permitted until 2025-12-13 06:50:27.612348234 +0000 UTC m=+84.216757722 (durationBeforeRetry 32s). Error: MountVolume.SetUp failed for volume "kube-api-access-cqllr" (UniqueName: "kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr") pod "network-check-target-xd92c" (UID: "3b6479f0-333b-4a96-9adf-2099afdc2447") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Dec 13 06:49:55 crc kubenswrapper[4971]: I1213 06:49:55.703152 4971 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 13 06:49:55 crc kubenswrapper[4971]: I1213 06:49:55.703194 4971 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 13 06:49:55 crc kubenswrapper[4971]: I1213 06:49:55.703204 4971 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 13 06:49:55 crc kubenswrapper[4971]: I1213 06:49:55.703217 4971 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 13 06:49:55 crc kubenswrapper[4971]: I1213 06:49:55.703226 4971 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-13T06:49:55Z","lastTransitionTime":"2025-12-13T06:49:55Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 13 06:49:55 crc kubenswrapper[4971]: I1213 06:49:55.768126 4971 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 13 06:49:55 crc kubenswrapper[4971]: I1213 06:49:55.768129 4971 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 13 06:49:55 crc kubenswrapper[4971]: I1213 06:49:55.768162 4971 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-x98k7" Dec 13 06:49:55 crc kubenswrapper[4971]: I1213 06:49:55.768472 4971 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 13 06:49:55 crc kubenswrapper[4971]: E1213 06:49:55.768614 4971 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 13 06:49:55 crc kubenswrapper[4971]: E1213 06:49:55.768704 4971 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-x98k7" podUID="c7185dec-3240-42e8-9d90-6182a5e8b65c" Dec 13 06:49:55 crc kubenswrapper[4971]: E1213 06:49:55.768879 4971 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 13 06:49:55 crc kubenswrapper[4971]: E1213 06:49:55.768997 4971 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 13 06:49:55 crc kubenswrapper[4971]: I1213 06:49:55.769593 4971 scope.go:117] "RemoveContainer" containerID="d950a29b0f6f47ae1cc90c132adb059b5aff889a0a38de1d1296e0f083da5e51" Dec 13 06:49:55 crc kubenswrapper[4971]: I1213 06:49:55.805192 4971 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 13 06:49:55 crc kubenswrapper[4971]: I1213 06:49:55.805234 4971 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 13 06:49:55 crc kubenswrapper[4971]: I1213 06:49:55.805244 4971 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 13 06:49:55 crc kubenswrapper[4971]: I1213 06:49:55.805259 4971 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 13 06:49:55 crc kubenswrapper[4971]: I1213 06:49:55.805270 4971 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-13T06:49:55Z","lastTransitionTime":"2025-12-13T06:49:55Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 13 06:49:55 crc kubenswrapper[4971]: I1213 06:49:55.907790 4971 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 13 06:49:55 crc kubenswrapper[4971]: I1213 06:49:55.908123 4971 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 13 06:49:55 crc kubenswrapper[4971]: I1213 06:49:55.908138 4971 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 13 06:49:55 crc kubenswrapper[4971]: I1213 06:49:55.908155 4971 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 13 06:49:55 crc kubenswrapper[4971]: I1213 06:49:55.908165 4971 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-13T06:49:55Z","lastTransitionTime":"2025-12-13T06:49:55Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 13 06:49:56 crc kubenswrapper[4971]: I1213 06:49:56.010010 4971 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 13 06:49:56 crc kubenswrapper[4971]: I1213 06:49:56.010062 4971 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 13 06:49:56 crc kubenswrapper[4971]: I1213 06:49:56.010083 4971 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 13 06:49:56 crc kubenswrapper[4971]: I1213 06:49:56.010107 4971 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 13 06:49:56 crc kubenswrapper[4971]: I1213 06:49:56.010122 4971 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-13T06:49:56Z","lastTransitionTime":"2025-12-13T06:49:56Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 13 06:49:56 crc kubenswrapper[4971]: I1213 06:49:56.098866 4971 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-zlcg5_736c6b1a-93dd-4f71-81b6-1c43fcf5f556/ovnkube-controller/1.log" Dec 13 06:49:56 crc kubenswrapper[4971]: I1213 06:49:56.101575 4971 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-zlcg5" event={"ID":"736c6b1a-93dd-4f71-81b6-1c43fcf5f556","Type":"ContainerStarted","Data":"0b454a87c6dd11bd8c3bb3553ee71de32ed6f185984e5057a20f2237b35c45d9"} Dec 13 06:49:56 crc kubenswrapper[4971]: I1213 06:49:56.102616 4971 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-ovn-kubernetes/ovnkube-node-zlcg5" Dec 13 06:49:56 crc kubenswrapper[4971]: I1213 06:49:56.112816 4971 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 13 06:49:56 crc kubenswrapper[4971]: I1213 06:49:56.112867 4971 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 13 06:49:56 crc kubenswrapper[4971]: I1213 06:49:56.112882 4971 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 13 06:49:56 crc kubenswrapper[4971]: I1213 06:49:56.112909 4971 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 13 06:49:56 crc kubenswrapper[4971]: I1213 06:49:56.112924 4971 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-13T06:49:56Z","lastTransitionTime":"2025-12-13T06:49:56Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 13 06:49:56 crc kubenswrapper[4971]: I1213 06:49:56.126042 4971 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-82xjz" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e0c4af12-edda-41d7-9b44-f87396174ef2\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T06:49:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T06:49:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T06:49:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T06:49:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://0f4577d9e37f860813e221f2e208177d8d383fd04f673ce357a0a405ea29d1e6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-13T06:49:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cnxsq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://7e0d2020e62f1a277cff7047e1907dc747ed393a1f6c983e1ee258bd156c16c7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-13T06:49:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cnxsq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-13T06:49:25Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-82xjz\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-13T06:49:56Z is after 2025-08-24T17:21:41Z" Dec 13 06:49:56 crc kubenswrapper[4971]: I1213 06:49:56.170780 4971 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-wwvpl" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"dddf744b-2122-49ad-a980-105c09636e0f\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T06:49:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T06:49:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T06:49:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T06:49:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4676d897c7fc668587faec563f3551a056557879cee3e865192d2bd28edf62fc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-13T06:49:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-czz9v\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-13T06:49:25Z\\\"}}\" for pod \"openshift-multus\"/\"multus-wwvpl\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-13T06:49:56Z is after 2025-08-24T17:21:41Z" Dec 13 06:49:56 crc kubenswrapper[4971]: I1213 06:49:56.201146 4971 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-zlcg5" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"736c6b1a-93dd-4f71-81b6-1c43fcf5f556\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T06:49:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T06:49:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T06:49:25Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T06:49:25Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://325223543b3c7db417f769533559a1ca2b23a723494e1885b0fd0480b5a6399f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-13T06:49:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vzmsr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://f22a4ec756119c4f791fb5105cf1f31df3e548b2b07b1f909a62802e9b6035ae\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-13T06:49:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vzmsr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://f6073140d9f86929b6d4bd2960ede0617a5c9620b269bbab043bae37e0d389f5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-13T06:49:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vzmsr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://693a2ab434c250cb140a866e8068b156df9191414fed1d63bee919a2528ea6e5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-13T06:49:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vzmsr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://219620f51474e60b3e1f880c0972eb612dffc4bad24462bbba2c093ae817d852\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-13T06:49:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vzmsr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://f137f4b441c2402093ab430a9689a46221c4fb4bc70d99fb992f62b0a989c465\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-13T06:49:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vzmsr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://0b454a87c6dd11bd8c3bb3553ee71de32ed6f185984e5057a20f2237b35c45d9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d950a29b0f6f47ae1cc90c132adb059b5aff889a0a38de1d1296e0f083da5e51\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-12-13T06:49:37Z\\\",\\\"message\\\":\\\" ovnkube.go:137] failed to run ovnkube: [failed to start network controller: failed to start default network controller: unable to create admin network policy controller, err: could not add Event Handler for anpInformer during admin network policy controller initialization, handler {0x1fcc6e0 0x1fcc3c0 0x1fcc360} was not added to shared informer because it has stopped already, failed to start node network controller: failed to start default node network controller: failed to set node crc annotations: Internal error occurred: failed calling webhook \\\\\\\"node.network-node-identity.openshift.io\\\\\\\": failed to call webhook: Post \\\\\\\"https://127.0.0.1:9743/node?timeout=10s\\\\\\\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-13T06:49:36Z is after 2025-08-24T17:21:41Z]\\\\nI1213 06:49:36.767920 6409 model_client.go:382] Update operations generated as: [{Op:update Table:Logical_Switch_Port Row:map[addresses:{GoSet:[0a:58:0a:d9:00:04 10.217.0.4]} options:{GoMap:map[iface-id-ver:3b6479f0-333b-4a96-9adf-2099afdc2447 requested-chassis:crc]} port_security:{GoSet:[0a:58:0a:d9:00:04 10.217.0.4]}] Rows:[] Columns:[] Mutations:[] Timeout:\\\\u003cnil\\\\u003e Where:[where column _uuid == {61897e97-c771-4738-8709-09636387cb00}] Until: Durable:\\\\u003cnil\\\\u003e Comment:\\\\u003cnil\\\\u003e Lock:\\\\u003cnil\\\\u003e UUID: UUIDName:}]\\\\nI1213 06:49:36.767917 6409 model_clien\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-13T06:49:36Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":2,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-13T06:49:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vzmsr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://868049c6a3a4ebe4036d1d22e02ba6b76bced2ee7e269bb228accc4a0636678a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-13T06:49:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vzmsr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://451b22df6973ddc8fa2209a65d64d9e41b33eb231bb4adfab88cf0e5943bb246\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://451b22df6973ddc8fa2209a65d64d9e41b33eb231bb4adfab88cf0e5943bb246\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-13T06:49:26Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-13T06:49:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vzmsr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-13T06:49:25Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-zlcg5\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-13T06:49:56Z is after 2025-08-24T17:21:41Z" Dec 13 06:49:56 crc kubenswrapper[4971]: I1213 06:49:56.215205 4971 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 13 06:49:56 crc kubenswrapper[4971]: I1213 06:49:56.215265 4971 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 13 06:49:56 crc kubenswrapper[4971]: I1213 06:49:56.215281 4971 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 13 06:49:56 crc kubenswrapper[4971]: I1213 06:49:56.215304 4971 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 13 06:49:56 crc kubenswrapper[4971]: I1213 06:49:56.215319 4971 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-13T06:49:56Z","lastTransitionTime":"2025-12-13T06:49:56Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 13 06:49:56 crc kubenswrapper[4971]: I1213 06:49:56.220921 4971 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-13T06:49:24Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-13T06:49:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-13T06:49:24Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d0b2ecf0d5b22f23217f42f259ae7ffb6359555b22c30eaebf667eaf254b3bab\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-13T06:49:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-13T06:49:56Z is after 2025-08-24T17:21:41Z" Dec 13 06:49:56 crc kubenswrapper[4971]: I1213 06:49:56.232993 4971 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-13T06:49:23Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-13T06:49:23Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-13T06:49:56Z is after 2025-08-24T17:21:41Z" Dec 13 06:49:56 crc kubenswrapper[4971]: I1213 06:49:56.244510 4971 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-13T06:49:26Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-13T06:49:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-13T06:49:26Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9c8a0d1db0e754462d805e82daf5f52ffd5457507a30456535ec59e096c9d092\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-13T06:49:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-13T06:49:56Z is after 2025-08-24T17:21:41Z" Dec 13 06:49:56 crc kubenswrapper[4971]: I1213 06:49:56.254899 4971 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-gkcd8" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"13f5c327-f95e-42b4-8783-218500c17694\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T06:49:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T06:49:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T06:49:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T06:49:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://77e846e36f1cb40d03471dc03de11b6c3da448baa407aad38093fae9a801cf92\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-13T06:49:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gvlhn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-13T06:49:27Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-gkcd8\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-13T06:49:56Z is after 2025-08-24T17:21:41Z" Dec 13 06:49:56 crc kubenswrapper[4971]: I1213 06:49:56.274038 4971 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-x98k7" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"c7185dec-3240-42e8-9d90-6182a5e8b65c\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T06:49:39Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T06:49:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T06:49:39Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T06:49:39Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bj2gn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bj2gn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-13T06:49:39Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-x98k7\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-13T06:49:56Z is after 2025-08-24T17:21:41Z" Dec 13 06:49:56 crc kubenswrapper[4971]: I1213 06:49:56.286746 4971 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"f9d31b5a-b770-4a15-95c8-35b49677c608\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T06:49:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T06:49:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T06:49:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T06:49:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T06:49:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c9baa042f9dada954faa57b5454cd6b0a12a297e56ddc192c52cff50e42a16fa\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-13T06:49:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://c788de86754f0bf37ee184672fc1dcb388ccebc246d4f15be4b5da661f725f8f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-13T06:49:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://760817511a9b4f96a68dc38d26631250eaab72e1dd8f123e653039b6b4e91f84\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-13T06:49:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b859be0c9feac856e9855c6a038b02b49b32fade97873c46d0bec0a32d8b5147\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-13T06:49:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-13T06:49:03Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-13T06:49:56Z is after 2025-08-24T17:21:41Z" Dec 13 06:49:56 crc kubenswrapper[4971]: I1213 06:49:56.301975 4971 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-ncchz" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"de914d21-a688-401e-a66e-861beb6336d9\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T06:49:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T06:49:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T06:49:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T06:49:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://244ffe3dc9d8cf799aff4c4d795bf41554c1e6bb6667fbb0c7df322fb5f96c68\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-13T06:49:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lqmjj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://0b7111af6c5a198d02bb02858837cfb4f3b20828d8d206cd18aa0ffa102d0cd4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://0b7111af6c5a198d02bb02858837cfb4f3b20828d8d206cd18aa0ffa102d0cd4\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-13T06:49:26Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-13T06:49:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lqmjj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b36777fe40e74ae1cecb7b73e204139db5279ba51cef4f1e8c4a4c491d8be403\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://b36777fe40e74ae1cecb7b73e204139db5279ba51cef4f1e8c4a4c491d8be403\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-13T06:49:28Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-13T06:49:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lqmjj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://0597a06cc663e416142311a16941d98297577693a1fe27201ea42c27496cace2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://0597a06cc663e416142311a16941d98297577693a1fe27201ea42c27496cace2\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-13T06:49:29Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-13T06:49:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lqmjj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://3f52c2e22a75c010993ca5809a3a9d1304758d4cb6e83691ef1af15a4b514ac4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://3f52c2e22a75c010993ca5809a3a9d1304758d4cb6e83691ef1af15a4b514ac4\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-13T06:49:30Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-13T06:49:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lqmjj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://9e21ca708c4c4c1228b417cece51299e0792729b7c78884bd18b2dda8fe8355c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://9e21ca708c4c4c1228b417cece51299e0792729b7c78884bd18b2dda8fe8355c\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-13T06:49:31Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-13T06:49:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lqmjj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://74552159f12c87cdda09147692b400ccac0d4e43c1ed72fe8462c4249ad69c7b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://74552159f12c87cdda09147692b400ccac0d4e43c1ed72fe8462c4249ad69c7b\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-13T06:49:32Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-13T06:49:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lqmjj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-13T06:49:25Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-ncchz\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-13T06:49:56Z is after 2025-08-24T17:21:41Z" Dec 13 06:49:56 crc kubenswrapper[4971]: I1213 06:49:56.312441 4971 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-5rzh5" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"6bf4591d-2483-4132-9672-d0fd6c995968\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T06:49:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T06:49:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T06:49:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T06:49:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://50e2cb14e4ffb603acb06e2281615a99f06182d345c4e70f2446fa2077033bbb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-13T06:49:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dfxqn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-13T06:49:25Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-5rzh5\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-13T06:49:56Z is after 2025-08-24T17:21:41Z" Dec 13 06:49:56 crc kubenswrapper[4971]: I1213 06:49:56.317161 4971 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 13 06:49:56 crc kubenswrapper[4971]: I1213 06:49:56.317219 4971 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 13 06:49:56 crc kubenswrapper[4971]: I1213 06:49:56.317241 4971 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 13 06:49:56 crc kubenswrapper[4971]: I1213 06:49:56.317269 4971 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 13 06:49:56 crc kubenswrapper[4971]: I1213 06:49:56.317291 4971 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-13T06:49:56Z","lastTransitionTime":"2025-12-13T06:49:56Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 13 06:49:56 crc kubenswrapper[4971]: I1213 06:49:56.328777 4971 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-z572s" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"b4146d3d-4034-49a9-bd37-164856255137\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T06:49:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T06:49:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T06:49:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T06:49:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://46fd4e1070835cfbf7f35d497eff7f047473fc4048ee07812925df5ebdf55342\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-13T06:49:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-j6l9b\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://dc79a0968f85723e6ad0dc269c8da1f7472adb1bb2b87edd017ee2295bfbf89c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-13T06:49:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-j6l9b\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-13T06:49:38Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-z572s\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-13T06:49:56Z is after 2025-08-24T17:21:41Z" Dec 13 06:49:56 crc kubenswrapper[4971]: I1213 06:49:56.356386 4971 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"1ed9f82e-fe5c-4b76-a619-62e5abd88d3b\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T06:49:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T06:49:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T06:49:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T06:49:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T06:49:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e457329a78a2d24337f994e8a0a1f75a2a667244aa75eb84613d1e7138364aac\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-13T06:49:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://eb014de8140deef097d9bd94f3b9fa4ab3558bc4449bef84b83bf62bfabf5836\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-13T06:49:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://2c438cc74d555f7c9db4569d1c56b78b1088ed8105697d2c0f31e6c5136f3dd0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-13T06:49:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://aa6fa6e4340dcd066252df3c56d79a3a39ed531d5eea8d1d5941d6952368d3af\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-13T06:49:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://c4471b98854c52d5cd084d0ba3e64e5b08e9518ec3c372cce2c875a440522457\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-13T06:49:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7d2746335c98dae3a0f43024d34cb62ebcb42a01c8f9ce18d52b29d4c45598e5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://7d2746335c98dae3a0f43024d34cb62ebcb42a01c8f9ce18d52b29d4c45598e5\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-13T06:49:04Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-13T06:49:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://bc0e77d10a738128bea82122496bc42235fc1688f8907756d93e61b651902b57\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://bc0e77d10a738128bea82122496bc42235fc1688f8907756d93e61b651902b57\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-13T06:49:05Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-13T06:49:05Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://f821e418cba5e832da8839268bcbbd9630bdf319a212d5382bfa2c4b0b73a268\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://f821e418cba5e832da8839268bcbbd9630bdf319a212d5382bfa2c4b0b73a268\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-13T06:49:07Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-13T06:49:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-13T06:49:03Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-13T06:49:56Z is after 2025-08-24T17:21:41Z" Dec 13 06:49:56 crc kubenswrapper[4971]: I1213 06:49:56.377417 4971 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-13T06:49:24Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-13T06:49:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-13T06:49:24Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4328cd6a8464ed468229bbfbc8265f969427916d72ebb92237dbe68efe70fa47\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-13T06:49:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://9d715ca82d9b5a594e41f2b98da6f4277da0e4e1dec114969f901c00f6d7553d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-13T06:49:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-13T06:49:56Z is after 2025-08-24T17:21:41Z" Dec 13 06:49:56 crc kubenswrapper[4971]: I1213 06:49:56.390793 4971 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3ad3268a-a982-428c-9c06-4aaeddd372d9\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T06:49:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T06:49:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T06:49:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T06:49:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T06:49:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3716afdf2bc693be80a09fc9f7727d334f435caf1d4c8ec4811b3a1f1652cdfe\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-13T06:49:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://30df3abf8c79db8af52c6a918f73e1ce410ddafcf341636424c267e1eeb195a7\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-13T06:49:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://ccdf65f6d2b1eb3647a03ba37c305c32e34addfdf79ff937c8dddaa3e5d5b5c6\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-13T06:49:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://770b73dd0c2b6f38744e2c1cd6aa186a69bb87d1cd015a603ad2e33be06a8f6b\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://b3c1bab3d53f975b2e24582ce2771bc26d600e0dd7c30339dcfd18a071162ad4\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-12-13T06:49:23Z\\\",\\\"message\\\":\\\"file observer\\\\nW1213 06:49:23.283403 1 builder.go:272] unable to get owner reference (falling back to namespace): pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\nI1213 06:49:23.284336 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1213 06:49:23.286978 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-981773671/tls.crt::/tmp/serving-cert-981773671/tls.key\\\\\\\"\\\\nI1213 06:49:23.775072 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1213 06:49:23.778615 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1213 06:49:23.778645 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1213 06:49:23.778680 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1213 06:49:23.778691 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1213 06:49:23.785058 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1213 06:49:23.786150 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1213 06:49:23.786196 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1213 06:49:23.786223 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1213 06:49:23.786246 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1213 06:49:23.786269 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1213 06:49:23.786291 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI1213 06:49:23.785608 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF1213 06:49:23.787563 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-13T06:49:08Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-13T06:49:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://9acd07066deed483bd7fa61ee45fda09539d110717a134a56827f871186f92fc\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-13T06:49:07Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://38c91cfbaecfdbd3e3b663422a91ab4294610afb6bbf9ebdb77db24e2b76c682\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://38c91cfbaecfdbd3e3b663422a91ab4294610afb6bbf9ebdb77db24e2b76c682\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-13T06:49:05Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-13T06:49:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-13T06:49:03Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-13T06:49:56Z is after 2025-08-24T17:21:41Z" Dec 13 06:49:56 crc kubenswrapper[4971]: I1213 06:49:56.402861 4971 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-13T06:49:23Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-13T06:49:23Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-13T06:49:56Z is after 2025-08-24T17:21:41Z" Dec 13 06:49:56 crc kubenswrapper[4971]: I1213 06:49:56.415655 4971 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-13T06:49:23Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-13T06:49:23Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-13T06:49:56Z is after 2025-08-24T17:21:41Z" Dec 13 06:49:56 crc kubenswrapper[4971]: I1213 06:49:56.420195 4971 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 13 06:49:56 crc kubenswrapper[4971]: I1213 06:49:56.420223 4971 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 13 06:49:56 crc kubenswrapper[4971]: I1213 06:49:56.420232 4971 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 13 06:49:56 crc kubenswrapper[4971]: I1213 06:49:56.420247 4971 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 13 06:49:56 crc kubenswrapper[4971]: I1213 06:49:56.420258 4971 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-13T06:49:56Z","lastTransitionTime":"2025-12-13T06:49:56Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 13 06:49:56 crc kubenswrapper[4971]: I1213 06:49:56.522225 4971 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 13 06:49:56 crc kubenswrapper[4971]: I1213 06:49:56.522271 4971 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 13 06:49:56 crc kubenswrapper[4971]: I1213 06:49:56.522282 4971 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 13 06:49:56 crc kubenswrapper[4971]: I1213 06:49:56.522299 4971 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 13 06:49:56 crc kubenswrapper[4971]: I1213 06:49:56.522311 4971 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-13T06:49:56Z","lastTransitionTime":"2025-12-13T06:49:56Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 13 06:49:56 crc kubenswrapper[4971]: I1213 06:49:56.624414 4971 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 13 06:49:56 crc kubenswrapper[4971]: I1213 06:49:56.624459 4971 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 13 06:49:56 crc kubenswrapper[4971]: I1213 06:49:56.624472 4971 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 13 06:49:56 crc kubenswrapper[4971]: I1213 06:49:56.624491 4971 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 13 06:49:56 crc kubenswrapper[4971]: I1213 06:49:56.624504 4971 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-13T06:49:56Z","lastTransitionTime":"2025-12-13T06:49:56Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 13 06:49:56 crc kubenswrapper[4971]: I1213 06:49:56.726797 4971 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 13 06:49:56 crc kubenswrapper[4971]: I1213 06:49:56.726836 4971 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 13 06:49:56 crc kubenswrapper[4971]: I1213 06:49:56.726847 4971 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 13 06:49:56 crc kubenswrapper[4971]: I1213 06:49:56.726861 4971 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 13 06:49:56 crc kubenswrapper[4971]: I1213 06:49:56.726872 4971 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-13T06:49:56Z","lastTransitionTime":"2025-12-13T06:49:56Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 13 06:49:56 crc kubenswrapper[4971]: I1213 06:49:56.829998 4971 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 13 06:49:56 crc kubenswrapper[4971]: I1213 06:49:56.830036 4971 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 13 06:49:56 crc kubenswrapper[4971]: I1213 06:49:56.830044 4971 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 13 06:49:56 crc kubenswrapper[4971]: I1213 06:49:56.830056 4971 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 13 06:49:56 crc kubenswrapper[4971]: I1213 06:49:56.830065 4971 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-13T06:49:56Z","lastTransitionTime":"2025-12-13T06:49:56Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 13 06:49:56 crc kubenswrapper[4971]: I1213 06:49:56.932645 4971 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 13 06:49:56 crc kubenswrapper[4971]: I1213 06:49:56.932722 4971 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 13 06:49:56 crc kubenswrapper[4971]: I1213 06:49:56.932745 4971 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 13 06:49:56 crc kubenswrapper[4971]: I1213 06:49:56.932774 4971 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 13 06:49:56 crc kubenswrapper[4971]: I1213 06:49:56.932795 4971 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-13T06:49:56Z","lastTransitionTime":"2025-12-13T06:49:56Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 13 06:49:57 crc kubenswrapper[4971]: I1213 06:49:57.035619 4971 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 13 06:49:57 crc kubenswrapper[4971]: I1213 06:49:57.035664 4971 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 13 06:49:57 crc kubenswrapper[4971]: I1213 06:49:57.035677 4971 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 13 06:49:57 crc kubenswrapper[4971]: I1213 06:49:57.035693 4971 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 13 06:49:57 crc kubenswrapper[4971]: I1213 06:49:57.035706 4971 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-13T06:49:57Z","lastTransitionTime":"2025-12-13T06:49:57Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 13 06:49:57 crc kubenswrapper[4971]: I1213 06:49:57.107153 4971 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-zlcg5_736c6b1a-93dd-4f71-81b6-1c43fcf5f556/ovnkube-controller/2.log" Dec 13 06:49:57 crc kubenswrapper[4971]: I1213 06:49:57.108075 4971 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-zlcg5_736c6b1a-93dd-4f71-81b6-1c43fcf5f556/ovnkube-controller/1.log" Dec 13 06:49:57 crc kubenswrapper[4971]: I1213 06:49:57.111156 4971 generic.go:334] "Generic (PLEG): container finished" podID="736c6b1a-93dd-4f71-81b6-1c43fcf5f556" containerID="0b454a87c6dd11bd8c3bb3553ee71de32ed6f185984e5057a20f2237b35c45d9" exitCode=1 Dec 13 06:49:57 crc kubenswrapper[4971]: I1213 06:49:57.111212 4971 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-zlcg5" event={"ID":"736c6b1a-93dd-4f71-81b6-1c43fcf5f556","Type":"ContainerDied","Data":"0b454a87c6dd11bd8c3bb3553ee71de32ed6f185984e5057a20f2237b35c45d9"} Dec 13 06:49:57 crc kubenswrapper[4971]: I1213 06:49:57.111259 4971 scope.go:117] "RemoveContainer" containerID="d950a29b0f6f47ae1cc90c132adb059b5aff889a0a38de1d1296e0f083da5e51" Dec 13 06:49:57 crc kubenswrapper[4971]: I1213 06:49:57.112247 4971 scope.go:117] "RemoveContainer" containerID="0b454a87c6dd11bd8c3bb3553ee71de32ed6f185984e5057a20f2237b35c45d9" Dec 13 06:49:57 crc kubenswrapper[4971]: E1213 06:49:57.112693 4971 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"ovnkube-controller\" with CrashLoopBackOff: \"back-off 20s restarting failed container=ovnkube-controller pod=ovnkube-node-zlcg5_openshift-ovn-kubernetes(736c6b1a-93dd-4f71-81b6-1c43fcf5f556)\"" pod="openshift-ovn-kubernetes/ovnkube-node-zlcg5" podUID="736c6b1a-93dd-4f71-81b6-1c43fcf5f556" Dec 13 06:49:57 crc kubenswrapper[4971]: I1213 06:49:57.137762 4971 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 13 06:49:57 crc kubenswrapper[4971]: I1213 06:49:57.137842 4971 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 13 06:49:57 crc kubenswrapper[4971]: I1213 06:49:57.137860 4971 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 13 06:49:57 crc kubenswrapper[4971]: I1213 06:49:57.138358 4971 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 13 06:49:57 crc kubenswrapper[4971]: I1213 06:49:57.138444 4971 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-13T06:49:57Z","lastTransitionTime":"2025-12-13T06:49:57Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 13 06:49:57 crc kubenswrapper[4971]: I1213 06:49:57.144672 4971 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-wwvpl" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"dddf744b-2122-49ad-a980-105c09636e0f\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T06:49:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T06:49:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T06:49:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T06:49:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4676d897c7fc668587faec563f3551a056557879cee3e865192d2bd28edf62fc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-13T06:49:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-czz9v\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-13T06:49:25Z\\\"}}\" for pod \"openshift-multus\"/\"multus-wwvpl\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-13T06:49:57Z is after 2025-08-24T17:21:41Z" Dec 13 06:49:57 crc kubenswrapper[4971]: I1213 06:49:57.169698 4971 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-zlcg5" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"736c6b1a-93dd-4f71-81b6-1c43fcf5f556\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T06:49:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T06:49:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T06:49:25Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T06:49:25Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://325223543b3c7db417f769533559a1ca2b23a723494e1885b0fd0480b5a6399f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-13T06:49:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vzmsr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://f22a4ec756119c4f791fb5105cf1f31df3e548b2b07b1f909a62802e9b6035ae\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-13T06:49:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vzmsr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://f6073140d9f86929b6d4bd2960ede0617a5c9620b269bbab043bae37e0d389f5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-13T06:49:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vzmsr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://693a2ab434c250cb140a866e8068b156df9191414fed1d63bee919a2528ea6e5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-13T06:49:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vzmsr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://219620f51474e60b3e1f880c0972eb612dffc4bad24462bbba2c093ae817d852\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-13T06:49:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vzmsr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://f137f4b441c2402093ab430a9689a46221c4fb4bc70d99fb992f62b0a989c465\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-13T06:49:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vzmsr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://0b454a87c6dd11bd8c3bb3553ee71de32ed6f185984e5057a20f2237b35c45d9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d950a29b0f6f47ae1cc90c132adb059b5aff889a0a38de1d1296e0f083da5e51\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-12-13T06:49:37Z\\\",\\\"message\\\":\\\" ovnkube.go:137] failed to run ovnkube: [failed to start network controller: failed to start default network controller: unable to create admin network policy controller, err: could not add Event Handler for anpInformer during admin network policy controller initialization, handler {0x1fcc6e0 0x1fcc3c0 0x1fcc360} was not added to shared informer because it has stopped already, failed to start node network controller: failed to start default node network controller: failed to set node crc annotations: Internal error occurred: failed calling webhook \\\\\\\"node.network-node-identity.openshift.io\\\\\\\": failed to call webhook: Post \\\\\\\"https://127.0.0.1:9743/node?timeout=10s\\\\\\\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-13T06:49:36Z is after 2025-08-24T17:21:41Z]\\\\nI1213 06:49:36.767920 6409 model_client.go:382] Update operations generated as: [{Op:update Table:Logical_Switch_Port Row:map[addresses:{GoSet:[0a:58:0a:d9:00:04 10.217.0.4]} options:{GoMap:map[iface-id-ver:3b6479f0-333b-4a96-9adf-2099afdc2447 requested-chassis:crc]} port_security:{GoSet:[0a:58:0a:d9:00:04 10.217.0.4]}] Rows:[] Columns:[] Mutations:[] Timeout:\\\\u003cnil\\\\u003e Where:[where column _uuid == {61897e97-c771-4738-8709-09636387cb00}] Until: Durable:\\\\u003cnil\\\\u003e Comment:\\\\u003cnil\\\\u003e Lock:\\\\u003cnil\\\\u003e UUID: UUIDName:}]\\\\nI1213 06:49:36.767917 6409 model_clien\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-13T06:49:36Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":2,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://0b454a87c6dd11bd8c3bb3553ee71de32ed6f185984e5057a20f2237b35c45d9\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-12-13T06:49:56Z\\\",\\\"message\\\":\\\" for pod on switch crc\\\\nI1213 06:49:56.566617 6645 metrics.go:553] Stopping metrics server at address \\\\\\\"127.0.0.1:29103\\\\\\\"\\\\nI1213 06:49:56.566625 6645 base_network_controller_pods.go:477] [default/openshift-network-console/networking-console-plugin-85b44fc459-gdk6g] creating logical port openshift-network-console_networking-console-plugin-85b44fc459-gdk6g for pod on switch crc\\\\nI1213 06:49:56.566569 6645 loadbalancer.go:304] Deleted 0 stale LBs for map[string]string{\\\\\\\"k8s.ovn.org/kind\\\\\\\":\\\\\\\"Service\\\\\\\", \\\\\\\"k8s.ovn.org/owner\\\\\\\":\\\\\\\"openshift-console-operator/metrics\\\\\\\"}\\\\nI1213 06:49:56.566666 6645 services_controller.go:360] Finished syncing service metrics on namespace openshift-console-operator for network=default : 1.804735ms\\\\nI1213 06:49:56.566669 6645 loadbalancer.go:304] Deleted 0 stale LBs for map[string]string{\\\\\\\"k8s.ovn.org/kind\\\\\\\":\\\\\\\"Service\\\\\\\", \\\\\\\"k8s.ovn.org/owner\\\\\\\":\\\\\\\"openshift-apiserver/check-endpoints\\\\\\\"}\\\\nF1213 06:49:56.566698 6645 ovnkube.go:137] failed to run ovnkube: [failed to start network controller: failed to start default network controller: unable to create admin network policy controller, err: could not add Event Handler for anpInformer during admin network policy controller initialization, handler {0x1fcc6e0 0x1fcc3c0 0x1fcc360} was not added to shared informer because it has stopped already, failed to start node network controller: failed \\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-13T06:49:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vzmsr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://868049c6a3a4ebe4036d1d22e02ba6b76bced2ee7e269bb228accc4a0636678a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-13T06:49:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vzmsr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://451b22df6973ddc8fa2209a65d64d9e41b33eb231bb4adfab88cf0e5943bb246\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://451b22df6973ddc8fa2209a65d64d9e41b33eb231bb4adfab88cf0e5943bb246\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-13T06:49:26Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-13T06:49:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vzmsr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-13T06:49:25Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-zlcg5\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-13T06:49:57Z is after 2025-08-24T17:21:41Z" Dec 13 06:49:57 crc kubenswrapper[4971]: I1213 06:49:57.184216 4971 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-13T06:49:24Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-13T06:49:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-13T06:49:24Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d0b2ecf0d5b22f23217f42f259ae7ffb6359555b22c30eaebf667eaf254b3bab\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-13T06:49:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-13T06:49:57Z is after 2025-08-24T17:21:41Z" Dec 13 06:49:57 crc kubenswrapper[4971]: I1213 06:49:57.195258 4971 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-13T06:49:23Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-13T06:49:23Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-13T06:49:57Z is after 2025-08-24T17:21:41Z" Dec 13 06:49:57 crc kubenswrapper[4971]: I1213 06:49:57.206854 4971 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-13T06:49:26Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-13T06:49:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-13T06:49:26Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9c8a0d1db0e754462d805e82daf5f52ffd5457507a30456535ec59e096c9d092\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-13T06:49:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-13T06:49:57Z is after 2025-08-24T17:21:41Z" Dec 13 06:49:57 crc kubenswrapper[4971]: I1213 06:49:57.220348 4971 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-82xjz" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e0c4af12-edda-41d7-9b44-f87396174ef2\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T06:49:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T06:49:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T06:49:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T06:49:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://0f4577d9e37f860813e221f2e208177d8d383fd04f673ce357a0a405ea29d1e6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-13T06:49:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cnxsq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://7e0d2020e62f1a277cff7047e1907dc747ed393a1f6c983e1ee258bd156c16c7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-13T06:49:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cnxsq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-13T06:49:25Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-82xjz\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-13T06:49:57Z is after 2025-08-24T17:21:41Z" Dec 13 06:49:57 crc kubenswrapper[4971]: I1213 06:49:57.231254 4971 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-x98k7" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"c7185dec-3240-42e8-9d90-6182a5e8b65c\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T06:49:39Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T06:49:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T06:49:39Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T06:49:39Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bj2gn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bj2gn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-13T06:49:39Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-x98k7\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-13T06:49:57Z is after 2025-08-24T17:21:41Z" Dec 13 06:49:57 crc kubenswrapper[4971]: I1213 06:49:57.241611 4971 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 13 06:49:57 crc kubenswrapper[4971]: I1213 06:49:57.241644 4971 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 13 06:49:57 crc kubenswrapper[4971]: I1213 06:49:57.241653 4971 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 13 06:49:57 crc kubenswrapper[4971]: I1213 06:49:57.241666 4971 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 13 06:49:57 crc kubenswrapper[4971]: I1213 06:49:57.241675 4971 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-13T06:49:57Z","lastTransitionTime":"2025-12-13T06:49:57Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 13 06:49:57 crc kubenswrapper[4971]: I1213 06:49:57.242707 4971 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"f9d31b5a-b770-4a15-95c8-35b49677c608\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T06:49:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T06:49:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T06:49:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T06:49:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T06:49:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c9baa042f9dada954faa57b5454cd6b0a12a297e56ddc192c52cff50e42a16fa\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-13T06:49:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://c788de86754f0bf37ee184672fc1dcb388ccebc246d4f15be4b5da661f725f8f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-13T06:49:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://760817511a9b4f96a68dc38d26631250eaab72e1dd8f123e653039b6b4e91f84\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-13T06:49:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b859be0c9feac856e9855c6a038b02b49b32fade97873c46d0bec0a32d8b5147\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-13T06:49:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-13T06:49:03Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-13T06:49:57Z is after 2025-08-24T17:21:41Z" Dec 13 06:49:57 crc kubenswrapper[4971]: I1213 06:49:57.256631 4971 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-ncchz" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"de914d21-a688-401e-a66e-861beb6336d9\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T06:49:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T06:49:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T06:49:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T06:49:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://244ffe3dc9d8cf799aff4c4d795bf41554c1e6bb6667fbb0c7df322fb5f96c68\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-13T06:49:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lqmjj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://0b7111af6c5a198d02bb02858837cfb4f3b20828d8d206cd18aa0ffa102d0cd4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://0b7111af6c5a198d02bb02858837cfb4f3b20828d8d206cd18aa0ffa102d0cd4\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-13T06:49:26Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-13T06:49:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lqmjj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b36777fe40e74ae1cecb7b73e204139db5279ba51cef4f1e8c4a4c491d8be403\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://b36777fe40e74ae1cecb7b73e204139db5279ba51cef4f1e8c4a4c491d8be403\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-13T06:49:28Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-13T06:49:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lqmjj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://0597a06cc663e416142311a16941d98297577693a1fe27201ea42c27496cace2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://0597a06cc663e416142311a16941d98297577693a1fe27201ea42c27496cace2\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-13T06:49:29Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-13T06:49:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lqmjj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://3f52c2e22a75c010993ca5809a3a9d1304758d4cb6e83691ef1af15a4b514ac4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://3f52c2e22a75c010993ca5809a3a9d1304758d4cb6e83691ef1af15a4b514ac4\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-13T06:49:30Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-13T06:49:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lqmjj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://9e21ca708c4c4c1228b417cece51299e0792729b7c78884bd18b2dda8fe8355c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://9e21ca708c4c4c1228b417cece51299e0792729b7c78884bd18b2dda8fe8355c\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-13T06:49:31Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-13T06:49:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lqmjj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://74552159f12c87cdda09147692b400ccac0d4e43c1ed72fe8462c4249ad69c7b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://74552159f12c87cdda09147692b400ccac0d4e43c1ed72fe8462c4249ad69c7b\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-13T06:49:32Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-13T06:49:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lqmjj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-13T06:49:25Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-ncchz\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-13T06:49:57Z is after 2025-08-24T17:21:41Z" Dec 13 06:49:57 crc kubenswrapper[4971]: I1213 06:49:57.269295 4971 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-gkcd8" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"13f5c327-f95e-42b4-8783-218500c17694\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T06:49:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T06:49:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T06:49:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T06:49:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://77e846e36f1cb40d03471dc03de11b6c3da448baa407aad38093fae9a801cf92\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-13T06:49:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gvlhn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-13T06:49:27Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-gkcd8\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-13T06:49:57Z is after 2025-08-24T17:21:41Z" Dec 13 06:49:57 crc kubenswrapper[4971]: I1213 06:49:57.280430 4971 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-z572s" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"b4146d3d-4034-49a9-bd37-164856255137\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T06:49:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T06:49:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T06:49:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T06:49:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://46fd4e1070835cfbf7f35d497eff7f047473fc4048ee07812925df5ebdf55342\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-13T06:49:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-j6l9b\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://dc79a0968f85723e6ad0dc269c8da1f7472adb1bb2b87edd017ee2295bfbf89c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-13T06:49:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-j6l9b\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-13T06:49:38Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-z572s\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-13T06:49:57Z is after 2025-08-24T17:21:41Z" Dec 13 06:49:57 crc kubenswrapper[4971]: I1213 06:49:57.297708 4971 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"1ed9f82e-fe5c-4b76-a619-62e5abd88d3b\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T06:49:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T06:49:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T06:49:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T06:49:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T06:49:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e457329a78a2d24337f994e8a0a1f75a2a667244aa75eb84613d1e7138364aac\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-13T06:49:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://eb014de8140deef097d9bd94f3b9fa4ab3558bc4449bef84b83bf62bfabf5836\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-13T06:49:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://2c438cc74d555f7c9db4569d1c56b78b1088ed8105697d2c0f31e6c5136f3dd0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-13T06:49:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://aa6fa6e4340dcd066252df3c56d79a3a39ed531d5eea8d1d5941d6952368d3af\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-13T06:49:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://c4471b98854c52d5cd084d0ba3e64e5b08e9518ec3c372cce2c875a440522457\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-13T06:49:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7d2746335c98dae3a0f43024d34cb62ebcb42a01c8f9ce18d52b29d4c45598e5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://7d2746335c98dae3a0f43024d34cb62ebcb42a01c8f9ce18d52b29d4c45598e5\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-13T06:49:04Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-13T06:49:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://bc0e77d10a738128bea82122496bc42235fc1688f8907756d93e61b651902b57\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://bc0e77d10a738128bea82122496bc42235fc1688f8907756d93e61b651902b57\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-13T06:49:05Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-13T06:49:05Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://f821e418cba5e832da8839268bcbbd9630bdf319a212d5382bfa2c4b0b73a268\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://f821e418cba5e832da8839268bcbbd9630bdf319a212d5382bfa2c4b0b73a268\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-13T06:49:07Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-13T06:49:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-13T06:49:03Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-13T06:49:57Z is after 2025-08-24T17:21:41Z" Dec 13 06:49:57 crc kubenswrapper[4971]: I1213 06:49:57.309389 4971 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-13T06:49:24Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-13T06:49:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-13T06:49:24Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4328cd6a8464ed468229bbfbc8265f969427916d72ebb92237dbe68efe70fa47\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-13T06:49:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://9d715ca82d9b5a594e41f2b98da6f4277da0e4e1dec114969f901c00f6d7553d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-13T06:49:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-13T06:49:57Z is after 2025-08-24T17:21:41Z" Dec 13 06:49:57 crc kubenswrapper[4971]: I1213 06:49:57.318926 4971 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-5rzh5" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"6bf4591d-2483-4132-9672-d0fd6c995968\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T06:49:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T06:49:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T06:49:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T06:49:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://50e2cb14e4ffb603acb06e2281615a99f06182d345c4e70f2446fa2077033bbb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-13T06:49:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dfxqn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-13T06:49:25Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-5rzh5\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-13T06:49:57Z is after 2025-08-24T17:21:41Z" Dec 13 06:49:57 crc kubenswrapper[4971]: I1213 06:49:57.333155 4971 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3ad3268a-a982-428c-9c06-4aaeddd372d9\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T06:49:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T06:49:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T06:49:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T06:49:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T06:49:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3716afdf2bc693be80a09fc9f7727d334f435caf1d4c8ec4811b3a1f1652cdfe\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-13T06:49:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://30df3abf8c79db8af52c6a918f73e1ce410ddafcf341636424c267e1eeb195a7\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-13T06:49:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://ccdf65f6d2b1eb3647a03ba37c305c32e34addfdf79ff937c8dddaa3e5d5b5c6\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-13T06:49:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://770b73dd0c2b6f38744e2c1cd6aa186a69bb87d1cd015a603ad2e33be06a8f6b\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://b3c1bab3d53f975b2e24582ce2771bc26d600e0dd7c30339dcfd18a071162ad4\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-12-13T06:49:23Z\\\",\\\"message\\\":\\\"file observer\\\\nW1213 06:49:23.283403 1 builder.go:272] unable to get owner reference (falling back to namespace): pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\nI1213 06:49:23.284336 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1213 06:49:23.286978 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-981773671/tls.crt::/tmp/serving-cert-981773671/tls.key\\\\\\\"\\\\nI1213 06:49:23.775072 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1213 06:49:23.778615 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1213 06:49:23.778645 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1213 06:49:23.778680 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1213 06:49:23.778691 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1213 06:49:23.785058 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1213 06:49:23.786150 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1213 06:49:23.786196 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1213 06:49:23.786223 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1213 06:49:23.786246 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1213 06:49:23.786269 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1213 06:49:23.786291 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI1213 06:49:23.785608 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF1213 06:49:23.787563 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-13T06:49:08Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-13T06:49:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://9acd07066deed483bd7fa61ee45fda09539d110717a134a56827f871186f92fc\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-13T06:49:07Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://38c91cfbaecfdbd3e3b663422a91ab4294610afb6bbf9ebdb77db24e2b76c682\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://38c91cfbaecfdbd3e3b663422a91ab4294610afb6bbf9ebdb77db24e2b76c682\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-13T06:49:05Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-13T06:49:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-13T06:49:03Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-13T06:49:57Z is after 2025-08-24T17:21:41Z" Dec 13 06:49:57 crc kubenswrapper[4971]: I1213 06:49:57.344096 4971 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 13 06:49:57 crc kubenswrapper[4971]: I1213 06:49:57.344138 4971 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 13 06:49:57 crc kubenswrapper[4971]: I1213 06:49:57.344149 4971 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 13 06:49:57 crc kubenswrapper[4971]: I1213 06:49:57.344168 4971 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 13 06:49:57 crc kubenswrapper[4971]: I1213 06:49:57.344182 4971 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-13T06:49:57Z","lastTransitionTime":"2025-12-13T06:49:57Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 13 06:49:57 crc kubenswrapper[4971]: I1213 06:49:57.347823 4971 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-13T06:49:23Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-13T06:49:23Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-13T06:49:57Z is after 2025-08-24T17:21:41Z" Dec 13 06:49:57 crc kubenswrapper[4971]: I1213 06:49:57.358999 4971 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-13T06:49:23Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-13T06:49:23Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-13T06:49:57Z is after 2025-08-24T17:21:41Z" Dec 13 06:49:57 crc kubenswrapper[4971]: I1213 06:49:57.395288 4971 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" Dec 13 06:49:57 crc kubenswrapper[4971]: I1213 06:49:57.405213 4971 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-kube-scheduler/openshift-kube-scheduler-crc"] Dec 13 06:49:57 crc kubenswrapper[4971]: I1213 06:49:57.407650 4971 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"f9d31b5a-b770-4a15-95c8-35b49677c608\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T06:49:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T06:49:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T06:49:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T06:49:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T06:49:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c9baa042f9dada954faa57b5454cd6b0a12a297e56ddc192c52cff50e42a16fa\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-13T06:49:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://c788de86754f0bf37ee184672fc1dcb388ccebc246d4f15be4b5da661f725f8f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-13T06:49:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://760817511a9b4f96a68dc38d26631250eaab72e1dd8f123e653039b6b4e91f84\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-13T06:49:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b859be0c9feac856e9855c6a038b02b49b32fade97873c46d0bec0a32d8b5147\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-13T06:49:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-13T06:49:03Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-13T06:49:57Z is after 2025-08-24T17:21:41Z" Dec 13 06:49:57 crc kubenswrapper[4971]: I1213 06:49:57.419391 4971 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-ncchz" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"de914d21-a688-401e-a66e-861beb6336d9\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T06:49:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T06:49:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T06:49:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T06:49:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://244ffe3dc9d8cf799aff4c4d795bf41554c1e6bb6667fbb0c7df322fb5f96c68\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-13T06:49:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lqmjj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://0b7111af6c5a198d02bb02858837cfb4f3b20828d8d206cd18aa0ffa102d0cd4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://0b7111af6c5a198d02bb02858837cfb4f3b20828d8d206cd18aa0ffa102d0cd4\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-13T06:49:26Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-13T06:49:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lqmjj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b36777fe40e74ae1cecb7b73e204139db5279ba51cef4f1e8c4a4c491d8be403\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://b36777fe40e74ae1cecb7b73e204139db5279ba51cef4f1e8c4a4c491d8be403\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-13T06:49:28Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-13T06:49:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lqmjj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://0597a06cc663e416142311a16941d98297577693a1fe27201ea42c27496cace2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://0597a06cc663e416142311a16941d98297577693a1fe27201ea42c27496cace2\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-13T06:49:29Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-13T06:49:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lqmjj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://3f52c2e22a75c010993ca5809a3a9d1304758d4cb6e83691ef1af15a4b514ac4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://3f52c2e22a75c010993ca5809a3a9d1304758d4cb6e83691ef1af15a4b514ac4\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-13T06:49:30Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-13T06:49:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lqmjj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://9e21ca708c4c4c1228b417cece51299e0792729b7c78884bd18b2dda8fe8355c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://9e21ca708c4c4c1228b417cece51299e0792729b7c78884bd18b2dda8fe8355c\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-13T06:49:31Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-13T06:49:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lqmjj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://74552159f12c87cdda09147692b400ccac0d4e43c1ed72fe8462c4249ad69c7b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://74552159f12c87cdda09147692b400ccac0d4e43c1ed72fe8462c4249ad69c7b\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-13T06:49:32Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-13T06:49:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lqmjj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-13T06:49:25Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-ncchz\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-13T06:49:57Z is after 2025-08-24T17:21:41Z" Dec 13 06:49:57 crc kubenswrapper[4971]: I1213 06:49:57.427542 4971 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-gkcd8" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"13f5c327-f95e-42b4-8783-218500c17694\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T06:49:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T06:49:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T06:49:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T06:49:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://77e846e36f1cb40d03471dc03de11b6c3da448baa407aad38093fae9a801cf92\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-13T06:49:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gvlhn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-13T06:49:27Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-gkcd8\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-13T06:49:57Z is after 2025-08-24T17:21:41Z" Dec 13 06:49:57 crc kubenswrapper[4971]: I1213 06:49:57.434769 4971 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-x98k7" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"c7185dec-3240-42e8-9d90-6182a5e8b65c\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T06:49:39Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T06:49:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T06:49:39Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T06:49:39Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bj2gn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bj2gn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-13T06:49:39Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-x98k7\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-13T06:49:57Z is after 2025-08-24T17:21:41Z" Dec 13 06:49:57 crc kubenswrapper[4971]: I1213 06:49:57.445916 4971 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 13 06:49:57 crc kubenswrapper[4971]: I1213 06:49:57.445959 4971 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 13 06:49:57 crc kubenswrapper[4971]: I1213 06:49:57.445971 4971 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 13 06:49:57 crc kubenswrapper[4971]: I1213 06:49:57.445988 4971 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 13 06:49:57 crc kubenswrapper[4971]: I1213 06:49:57.446000 4971 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-13T06:49:57Z","lastTransitionTime":"2025-12-13T06:49:57Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 13 06:49:57 crc kubenswrapper[4971]: I1213 06:49:57.452731 4971 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"1ed9f82e-fe5c-4b76-a619-62e5abd88d3b\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T06:49:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T06:49:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T06:49:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T06:49:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T06:49:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e457329a78a2d24337f994e8a0a1f75a2a667244aa75eb84613d1e7138364aac\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-13T06:49:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://eb014de8140deef097d9bd94f3b9fa4ab3558bc4449bef84b83bf62bfabf5836\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-13T06:49:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://2c438cc74d555f7c9db4569d1c56b78b1088ed8105697d2c0f31e6c5136f3dd0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-13T06:49:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://aa6fa6e4340dcd066252df3c56d79a3a39ed531d5eea8d1d5941d6952368d3af\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-13T06:49:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://c4471b98854c52d5cd084d0ba3e64e5b08e9518ec3c372cce2c875a440522457\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-13T06:49:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7d2746335c98dae3a0f43024d34cb62ebcb42a01c8f9ce18d52b29d4c45598e5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://7d2746335c98dae3a0f43024d34cb62ebcb42a01c8f9ce18d52b29d4c45598e5\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-13T06:49:04Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-13T06:49:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://bc0e77d10a738128bea82122496bc42235fc1688f8907756d93e61b651902b57\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://bc0e77d10a738128bea82122496bc42235fc1688f8907756d93e61b651902b57\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-13T06:49:05Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-13T06:49:05Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://f821e418cba5e832da8839268bcbbd9630bdf319a212d5382bfa2c4b0b73a268\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://f821e418cba5e832da8839268bcbbd9630bdf319a212d5382bfa2c4b0b73a268\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-13T06:49:07Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-13T06:49:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-13T06:49:03Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-13T06:49:57Z is after 2025-08-24T17:21:41Z" Dec 13 06:49:57 crc kubenswrapper[4971]: I1213 06:49:57.463720 4971 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-13T06:49:24Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-13T06:49:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-13T06:49:24Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4328cd6a8464ed468229bbfbc8265f969427916d72ebb92237dbe68efe70fa47\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-13T06:49:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://9d715ca82d9b5a594e41f2b98da6f4277da0e4e1dec114969f901c00f6d7553d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-13T06:49:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-13T06:49:57Z is after 2025-08-24T17:21:41Z" Dec 13 06:49:57 crc kubenswrapper[4971]: I1213 06:49:57.473417 4971 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-5rzh5" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"6bf4591d-2483-4132-9672-d0fd6c995968\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T06:49:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T06:49:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T06:49:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T06:49:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://50e2cb14e4ffb603acb06e2281615a99f06182d345c4e70f2446fa2077033bbb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-13T06:49:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dfxqn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-13T06:49:25Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-5rzh5\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-13T06:49:57Z is after 2025-08-24T17:21:41Z" Dec 13 06:49:57 crc kubenswrapper[4971]: I1213 06:49:57.484091 4971 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-z572s" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"b4146d3d-4034-49a9-bd37-164856255137\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T06:49:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T06:49:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T06:49:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T06:49:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://46fd4e1070835cfbf7f35d497eff7f047473fc4048ee07812925df5ebdf55342\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-13T06:49:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-j6l9b\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://dc79a0968f85723e6ad0dc269c8da1f7472adb1bb2b87edd017ee2295bfbf89c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-13T06:49:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-j6l9b\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-13T06:49:38Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-z572s\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-13T06:49:57Z is after 2025-08-24T17:21:41Z" Dec 13 06:49:57 crc kubenswrapper[4971]: I1213 06:49:57.496346 4971 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3ad3268a-a982-428c-9c06-4aaeddd372d9\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T06:49:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T06:49:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T06:49:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T06:49:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T06:49:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3716afdf2bc693be80a09fc9f7727d334f435caf1d4c8ec4811b3a1f1652cdfe\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-13T06:49:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://30df3abf8c79db8af52c6a918f73e1ce410ddafcf341636424c267e1eeb195a7\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-13T06:49:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://ccdf65f6d2b1eb3647a03ba37c305c32e34addfdf79ff937c8dddaa3e5d5b5c6\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-13T06:49:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://770b73dd0c2b6f38744e2c1cd6aa186a69bb87d1cd015a603ad2e33be06a8f6b\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://b3c1bab3d53f975b2e24582ce2771bc26d600e0dd7c30339dcfd18a071162ad4\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-12-13T06:49:23Z\\\",\\\"message\\\":\\\"file observer\\\\nW1213 06:49:23.283403 1 builder.go:272] unable to get owner reference (falling back to namespace): pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\nI1213 06:49:23.284336 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1213 06:49:23.286978 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-981773671/tls.crt::/tmp/serving-cert-981773671/tls.key\\\\\\\"\\\\nI1213 06:49:23.775072 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1213 06:49:23.778615 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1213 06:49:23.778645 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1213 06:49:23.778680 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1213 06:49:23.778691 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1213 06:49:23.785058 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1213 06:49:23.786150 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1213 06:49:23.786196 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1213 06:49:23.786223 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1213 06:49:23.786246 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1213 06:49:23.786269 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1213 06:49:23.786291 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI1213 06:49:23.785608 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF1213 06:49:23.787563 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-13T06:49:08Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-13T06:49:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://9acd07066deed483bd7fa61ee45fda09539d110717a134a56827f871186f92fc\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-13T06:49:07Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://38c91cfbaecfdbd3e3b663422a91ab4294610afb6bbf9ebdb77db24e2b76c682\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://38c91cfbaecfdbd3e3b663422a91ab4294610afb6bbf9ebdb77db24e2b76c682\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-13T06:49:05Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-13T06:49:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-13T06:49:03Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-13T06:49:57Z is after 2025-08-24T17:21:41Z" Dec 13 06:49:57 crc kubenswrapper[4971]: I1213 06:49:57.509481 4971 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-13T06:49:23Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-13T06:49:23Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-13T06:49:57Z is after 2025-08-24T17:21:41Z" Dec 13 06:49:57 crc kubenswrapper[4971]: I1213 06:49:57.521076 4971 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-13T06:49:23Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-13T06:49:23Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-13T06:49:57Z is after 2025-08-24T17:21:41Z" Dec 13 06:49:57 crc kubenswrapper[4971]: I1213 06:49:57.537331 4971 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-zlcg5" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"736c6b1a-93dd-4f71-81b6-1c43fcf5f556\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T06:49:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T06:49:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T06:49:25Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T06:49:25Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://325223543b3c7db417f769533559a1ca2b23a723494e1885b0fd0480b5a6399f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-13T06:49:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vzmsr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://f22a4ec756119c4f791fb5105cf1f31df3e548b2b07b1f909a62802e9b6035ae\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-13T06:49:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vzmsr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://f6073140d9f86929b6d4bd2960ede0617a5c9620b269bbab043bae37e0d389f5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-13T06:49:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vzmsr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://693a2ab434c250cb140a866e8068b156df9191414fed1d63bee919a2528ea6e5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-13T06:49:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vzmsr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://219620f51474e60b3e1f880c0972eb612dffc4bad24462bbba2c093ae817d852\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-13T06:49:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vzmsr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://f137f4b441c2402093ab430a9689a46221c4fb4bc70d99fb992f62b0a989c465\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-13T06:49:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vzmsr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://0b454a87c6dd11bd8c3bb3553ee71de32ed6f185984e5057a20f2237b35c45d9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d950a29b0f6f47ae1cc90c132adb059b5aff889a0a38de1d1296e0f083da5e51\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-12-13T06:49:37Z\\\",\\\"message\\\":\\\" ovnkube.go:137] failed to run ovnkube: [failed to start network controller: failed to start default network controller: unable to create admin network policy controller, err: could not add Event Handler for anpInformer during admin network policy controller initialization, handler {0x1fcc6e0 0x1fcc3c0 0x1fcc360} was not added to shared informer because it has stopped already, failed to start node network controller: failed to start default node network controller: failed to set node crc annotations: Internal error occurred: failed calling webhook \\\\\\\"node.network-node-identity.openshift.io\\\\\\\": failed to call webhook: Post \\\\\\\"https://127.0.0.1:9743/node?timeout=10s\\\\\\\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-13T06:49:36Z is after 2025-08-24T17:21:41Z]\\\\nI1213 06:49:36.767920 6409 model_client.go:382] Update operations generated as: [{Op:update Table:Logical_Switch_Port Row:map[addresses:{GoSet:[0a:58:0a:d9:00:04 10.217.0.4]} options:{GoMap:map[iface-id-ver:3b6479f0-333b-4a96-9adf-2099afdc2447 requested-chassis:crc]} port_security:{GoSet:[0a:58:0a:d9:00:04 10.217.0.4]}] Rows:[] Columns:[] Mutations:[] Timeout:\\\\u003cnil\\\\u003e Where:[where column _uuid == {61897e97-c771-4738-8709-09636387cb00}] Until: Durable:\\\\u003cnil\\\\u003e Comment:\\\\u003cnil\\\\u003e Lock:\\\\u003cnil\\\\u003e UUID: UUIDName:}]\\\\nI1213 06:49:36.767917 6409 model_clien\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-13T06:49:36Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":2,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://0b454a87c6dd11bd8c3bb3553ee71de32ed6f185984e5057a20f2237b35c45d9\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-12-13T06:49:56Z\\\",\\\"message\\\":\\\" for pod on switch crc\\\\nI1213 06:49:56.566617 6645 metrics.go:553] Stopping metrics server at address \\\\\\\"127.0.0.1:29103\\\\\\\"\\\\nI1213 06:49:56.566625 6645 base_network_controller_pods.go:477] [default/openshift-network-console/networking-console-plugin-85b44fc459-gdk6g] creating logical port openshift-network-console_networking-console-plugin-85b44fc459-gdk6g for pod on switch crc\\\\nI1213 06:49:56.566569 6645 loadbalancer.go:304] Deleted 0 stale LBs for map[string]string{\\\\\\\"k8s.ovn.org/kind\\\\\\\":\\\\\\\"Service\\\\\\\", \\\\\\\"k8s.ovn.org/owner\\\\\\\":\\\\\\\"openshift-console-operator/metrics\\\\\\\"}\\\\nI1213 06:49:56.566666 6645 services_controller.go:360] Finished syncing service metrics on namespace openshift-console-operator for network=default : 1.804735ms\\\\nI1213 06:49:56.566669 6645 loadbalancer.go:304] Deleted 0 stale LBs for map[string]string{\\\\\\\"k8s.ovn.org/kind\\\\\\\":\\\\\\\"Service\\\\\\\", \\\\\\\"k8s.ovn.org/owner\\\\\\\":\\\\\\\"openshift-apiserver/check-endpoints\\\\\\\"}\\\\nF1213 06:49:56.566698 6645 ovnkube.go:137] failed to run ovnkube: [failed to start network controller: failed to start default network controller: unable to create admin network policy controller, err: could not add Event Handler for anpInformer during admin network policy controller initialization, handler {0x1fcc6e0 0x1fcc3c0 0x1fcc360} was not added to shared informer because it has stopped already, failed to start node network controller: failed \\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-13T06:49:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vzmsr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://868049c6a3a4ebe4036d1d22e02ba6b76bced2ee7e269bb228accc4a0636678a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-13T06:49:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vzmsr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://451b22df6973ddc8fa2209a65d64d9e41b33eb231bb4adfab88cf0e5943bb246\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://451b22df6973ddc8fa2209a65d64d9e41b33eb231bb4adfab88cf0e5943bb246\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-13T06:49:26Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-13T06:49:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vzmsr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-13T06:49:25Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-zlcg5\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-13T06:49:57Z is after 2025-08-24T17:21:41Z" Dec 13 06:49:57 crc kubenswrapper[4971]: I1213 06:49:57.548123 4971 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 13 06:49:57 crc kubenswrapper[4971]: I1213 06:49:57.548154 4971 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 13 06:49:57 crc kubenswrapper[4971]: I1213 06:49:57.548165 4971 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 13 06:49:57 crc kubenswrapper[4971]: I1213 06:49:57.548183 4971 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 13 06:49:57 crc kubenswrapper[4971]: I1213 06:49:57.548197 4971 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-13T06:49:57Z","lastTransitionTime":"2025-12-13T06:49:57Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 13 06:49:57 crc kubenswrapper[4971]: I1213 06:49:57.550048 4971 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-13T06:49:24Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-13T06:49:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-13T06:49:24Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d0b2ecf0d5b22f23217f42f259ae7ffb6359555b22c30eaebf667eaf254b3bab\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-13T06:49:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-13T06:49:57Z is after 2025-08-24T17:21:41Z" Dec 13 06:49:57 crc kubenswrapper[4971]: I1213 06:49:57.561009 4971 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-13T06:49:23Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-13T06:49:23Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-13T06:49:57Z is after 2025-08-24T17:21:41Z" Dec 13 06:49:57 crc kubenswrapper[4971]: I1213 06:49:57.572937 4971 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-13T06:49:26Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-13T06:49:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-13T06:49:26Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9c8a0d1db0e754462d805e82daf5f52ffd5457507a30456535ec59e096c9d092\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-13T06:49:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-13T06:49:57Z is after 2025-08-24T17:21:41Z" Dec 13 06:49:57 crc kubenswrapper[4971]: I1213 06:49:57.584401 4971 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-82xjz" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e0c4af12-edda-41d7-9b44-f87396174ef2\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T06:49:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T06:49:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T06:49:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T06:49:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://0f4577d9e37f860813e221f2e208177d8d383fd04f673ce357a0a405ea29d1e6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-13T06:49:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cnxsq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://7e0d2020e62f1a277cff7047e1907dc747ed393a1f6c983e1ee258bd156c16c7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-13T06:49:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cnxsq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-13T06:49:25Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-82xjz\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-13T06:49:57Z is after 2025-08-24T17:21:41Z" Dec 13 06:49:57 crc kubenswrapper[4971]: I1213 06:49:57.595350 4971 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-wwvpl" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"dddf744b-2122-49ad-a980-105c09636e0f\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T06:49:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T06:49:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T06:49:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T06:49:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4676d897c7fc668587faec563f3551a056557879cee3e865192d2bd28edf62fc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-13T06:49:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-czz9v\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-13T06:49:25Z\\\"}}\" for pod \"openshift-multus\"/\"multus-wwvpl\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-13T06:49:57Z is after 2025-08-24T17:21:41Z" Dec 13 06:49:57 crc kubenswrapper[4971]: I1213 06:49:57.650107 4971 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 13 06:49:57 crc kubenswrapper[4971]: I1213 06:49:57.650145 4971 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 13 06:49:57 crc kubenswrapper[4971]: I1213 06:49:57.650155 4971 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 13 06:49:57 crc kubenswrapper[4971]: I1213 06:49:57.650173 4971 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 13 06:49:57 crc kubenswrapper[4971]: I1213 06:49:57.650185 4971 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-13T06:49:57Z","lastTransitionTime":"2025-12-13T06:49:57Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 13 06:49:57 crc kubenswrapper[4971]: I1213 06:49:57.752878 4971 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 13 06:49:57 crc kubenswrapper[4971]: I1213 06:49:57.752914 4971 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 13 06:49:57 crc kubenswrapper[4971]: I1213 06:49:57.752922 4971 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 13 06:49:57 crc kubenswrapper[4971]: I1213 06:49:57.752937 4971 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 13 06:49:57 crc kubenswrapper[4971]: I1213 06:49:57.752951 4971 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-13T06:49:57Z","lastTransitionTime":"2025-12-13T06:49:57Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 13 06:49:57 crc kubenswrapper[4971]: I1213 06:49:57.768429 4971 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 13 06:49:57 crc kubenswrapper[4971]: E1213 06:49:57.768605 4971 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 13 06:49:57 crc kubenswrapper[4971]: I1213 06:49:57.768441 4971 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 13 06:49:57 crc kubenswrapper[4971]: I1213 06:49:57.768429 4971 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 13 06:49:57 crc kubenswrapper[4971]: E1213 06:49:57.768680 4971 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 13 06:49:57 crc kubenswrapper[4971]: I1213 06:49:57.768441 4971 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-x98k7" Dec 13 06:49:57 crc kubenswrapper[4971]: E1213 06:49:57.768742 4971 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-x98k7" podUID="c7185dec-3240-42e8-9d90-6182a5e8b65c" Dec 13 06:49:57 crc kubenswrapper[4971]: E1213 06:49:57.768841 4971 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 13 06:49:57 crc kubenswrapper[4971]: I1213 06:49:57.855227 4971 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 13 06:49:57 crc kubenswrapper[4971]: I1213 06:49:57.855264 4971 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 13 06:49:57 crc kubenswrapper[4971]: I1213 06:49:57.855275 4971 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 13 06:49:57 crc kubenswrapper[4971]: I1213 06:49:57.855292 4971 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 13 06:49:57 crc kubenswrapper[4971]: I1213 06:49:57.855303 4971 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-13T06:49:57Z","lastTransitionTime":"2025-12-13T06:49:57Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 13 06:49:57 crc kubenswrapper[4971]: I1213 06:49:57.957457 4971 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 13 06:49:57 crc kubenswrapper[4971]: I1213 06:49:57.957539 4971 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 13 06:49:57 crc kubenswrapper[4971]: I1213 06:49:57.957548 4971 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 13 06:49:57 crc kubenswrapper[4971]: I1213 06:49:57.957563 4971 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 13 06:49:57 crc kubenswrapper[4971]: I1213 06:49:57.957572 4971 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-13T06:49:57Z","lastTransitionTime":"2025-12-13T06:49:57Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 13 06:49:58 crc kubenswrapper[4971]: I1213 06:49:58.060408 4971 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 13 06:49:58 crc kubenswrapper[4971]: I1213 06:49:58.060471 4971 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 13 06:49:58 crc kubenswrapper[4971]: I1213 06:49:58.060501 4971 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 13 06:49:58 crc kubenswrapper[4971]: I1213 06:49:58.060581 4971 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 13 06:49:58 crc kubenswrapper[4971]: I1213 06:49:58.060607 4971 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-13T06:49:58Z","lastTransitionTime":"2025-12-13T06:49:58Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 13 06:49:58 crc kubenswrapper[4971]: I1213 06:49:58.117193 4971 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-zlcg5_736c6b1a-93dd-4f71-81b6-1c43fcf5f556/ovnkube-controller/2.log" Dec 13 06:49:58 crc kubenswrapper[4971]: I1213 06:49:58.121894 4971 scope.go:117] "RemoveContainer" containerID="0b454a87c6dd11bd8c3bb3553ee71de32ed6f185984e5057a20f2237b35c45d9" Dec 13 06:49:58 crc kubenswrapper[4971]: E1213 06:49:58.122141 4971 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"ovnkube-controller\" with CrashLoopBackOff: \"back-off 20s restarting failed container=ovnkube-controller pod=ovnkube-node-zlcg5_openshift-ovn-kubernetes(736c6b1a-93dd-4f71-81b6-1c43fcf5f556)\"" pod="openshift-ovn-kubernetes/ovnkube-node-zlcg5" podUID="736c6b1a-93dd-4f71-81b6-1c43fcf5f556" Dec 13 06:49:58 crc kubenswrapper[4971]: I1213 06:49:58.141999 4971 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-13T06:49:23Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-13T06:49:23Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-13T06:49:58Z is after 2025-08-24T17:21:41Z" Dec 13 06:49:58 crc kubenswrapper[4971]: I1213 06:49:58.155416 4971 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-13T06:49:26Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-13T06:49:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-13T06:49:26Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9c8a0d1db0e754462d805e82daf5f52ffd5457507a30456535ec59e096c9d092\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-13T06:49:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-13T06:49:58Z is after 2025-08-24T17:21:41Z" Dec 13 06:49:58 crc kubenswrapper[4971]: I1213 06:49:58.163618 4971 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 13 06:49:58 crc kubenswrapper[4971]: I1213 06:49:58.163674 4971 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 13 06:49:58 crc kubenswrapper[4971]: I1213 06:49:58.163697 4971 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 13 06:49:58 crc kubenswrapper[4971]: I1213 06:49:58.163724 4971 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 13 06:49:58 crc kubenswrapper[4971]: I1213 06:49:58.163744 4971 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-13T06:49:58Z","lastTransitionTime":"2025-12-13T06:49:58Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 13 06:49:58 crc kubenswrapper[4971]: I1213 06:49:58.174097 4971 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-82xjz" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e0c4af12-edda-41d7-9b44-f87396174ef2\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T06:49:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T06:49:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T06:49:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T06:49:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://0f4577d9e37f860813e221f2e208177d8d383fd04f673ce357a0a405ea29d1e6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-13T06:49:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cnxsq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://7e0d2020e62f1a277cff7047e1907dc747ed393a1f6c983e1ee258bd156c16c7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-13T06:49:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cnxsq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-13T06:49:25Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-82xjz\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-13T06:49:58Z is after 2025-08-24T17:21:41Z" Dec 13 06:49:58 crc kubenswrapper[4971]: I1213 06:49:58.193220 4971 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-wwvpl" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"dddf744b-2122-49ad-a980-105c09636e0f\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T06:49:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T06:49:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T06:49:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T06:49:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4676d897c7fc668587faec563f3551a056557879cee3e865192d2bd28edf62fc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-13T06:49:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-czz9v\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-13T06:49:25Z\\\"}}\" for pod \"openshift-multus\"/\"multus-wwvpl\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-13T06:49:58Z is after 2025-08-24T17:21:41Z" Dec 13 06:49:58 crc kubenswrapper[4971]: I1213 06:49:58.223636 4971 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-zlcg5" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"736c6b1a-93dd-4f71-81b6-1c43fcf5f556\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T06:49:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T06:49:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T06:49:25Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T06:49:25Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://325223543b3c7db417f769533559a1ca2b23a723494e1885b0fd0480b5a6399f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-13T06:49:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vzmsr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://f22a4ec756119c4f791fb5105cf1f31df3e548b2b07b1f909a62802e9b6035ae\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-13T06:49:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vzmsr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://f6073140d9f86929b6d4bd2960ede0617a5c9620b269bbab043bae37e0d389f5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-13T06:49:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vzmsr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://693a2ab434c250cb140a866e8068b156df9191414fed1d63bee919a2528ea6e5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-13T06:49:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vzmsr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://219620f51474e60b3e1f880c0972eb612dffc4bad24462bbba2c093ae817d852\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-13T06:49:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vzmsr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://f137f4b441c2402093ab430a9689a46221c4fb4bc70d99fb992f62b0a989c465\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-13T06:49:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vzmsr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://0b454a87c6dd11bd8c3bb3553ee71de32ed6f185984e5057a20f2237b35c45d9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://0b454a87c6dd11bd8c3bb3553ee71de32ed6f185984e5057a20f2237b35c45d9\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-12-13T06:49:56Z\\\",\\\"message\\\":\\\" for pod on switch crc\\\\nI1213 06:49:56.566617 6645 metrics.go:553] Stopping metrics server at address \\\\\\\"127.0.0.1:29103\\\\\\\"\\\\nI1213 06:49:56.566625 6645 base_network_controller_pods.go:477] [default/openshift-network-console/networking-console-plugin-85b44fc459-gdk6g] creating logical port openshift-network-console_networking-console-plugin-85b44fc459-gdk6g for pod on switch crc\\\\nI1213 06:49:56.566569 6645 loadbalancer.go:304] Deleted 0 stale LBs for map[string]string{\\\\\\\"k8s.ovn.org/kind\\\\\\\":\\\\\\\"Service\\\\\\\", \\\\\\\"k8s.ovn.org/owner\\\\\\\":\\\\\\\"openshift-console-operator/metrics\\\\\\\"}\\\\nI1213 06:49:56.566666 6645 services_controller.go:360] Finished syncing service metrics on namespace openshift-console-operator for network=default : 1.804735ms\\\\nI1213 06:49:56.566669 6645 loadbalancer.go:304] Deleted 0 stale LBs for map[string]string{\\\\\\\"k8s.ovn.org/kind\\\\\\\":\\\\\\\"Service\\\\\\\", \\\\\\\"k8s.ovn.org/owner\\\\\\\":\\\\\\\"openshift-apiserver/check-endpoints\\\\\\\"}\\\\nF1213 06:49:56.566698 6645 ovnkube.go:137] failed to run ovnkube: [failed to start network controller: failed to start default network controller: unable to create admin network policy controller, err: could not add Event Handler for anpInformer during admin network policy controller initialization, handler {0x1fcc6e0 0x1fcc3c0 0x1fcc360} was not added to shared informer because it has stopped already, failed to start node network controller: failed \\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-13T06:49:55Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":2,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 20s restarting failed container=ovnkube-controller pod=ovnkube-node-zlcg5_openshift-ovn-kubernetes(736c6b1a-93dd-4f71-81b6-1c43fcf5f556)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vzmsr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://868049c6a3a4ebe4036d1d22e02ba6b76bced2ee7e269bb228accc4a0636678a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-13T06:49:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vzmsr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://451b22df6973ddc8fa2209a65d64d9e41b33eb231bb4adfab88cf0e5943bb246\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://451b22df6973ddc8fa2209a65d64d9e41b33eb231bb4adfab88cf0e5943bb246\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-13T06:49:26Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-13T06:49:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vzmsr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-13T06:49:25Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-zlcg5\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-13T06:49:58Z is after 2025-08-24T17:21:41Z" Dec 13 06:49:58 crc kubenswrapper[4971]: I1213 06:49:58.243245 4971 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-13T06:49:24Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-13T06:49:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-13T06:49:24Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d0b2ecf0d5b22f23217f42f259ae7ffb6359555b22c30eaebf667eaf254b3bab\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-13T06:49:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-13T06:49:58Z is after 2025-08-24T17:21:41Z" Dec 13 06:49:58 crc kubenswrapper[4971]: I1213 06:49:58.260828 4971 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"8ede2992-9576-4b23-83c2-24ac3b1d3caf\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T06:49:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T06:49:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T06:49:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T06:49:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T06:49:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9e61d25add1723ac286794abfe886358e9d64455c25379f7e3057fc3dd812418\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-13T06:49:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://7fb51d8435b4e3ce0411cae1d6c2824c552f0a33d79de01f969274516b39a5b3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-13T06:49:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://1c35f5e8d8fab6d77b9e2e0dedb195440de5d6af345cd5865abd32caa7fafd60\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-13T06:49:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://38806c430185715371ab9bf30a1af9a9d5d492abd9bb84acbdccc694f56e3a8a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"wait-for-host-port\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://38806c430185715371ab9bf30a1af9a9d5d492abd9bb84acbdccc694f56e3a8a\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-13T06:49:04Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-13T06:49:04Z\\\"}}}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-13T06:49:03Z\\\"}}\" for pod \"openshift-kube-scheduler\"/\"openshift-kube-scheduler-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-13T06:49:58Z is after 2025-08-24T17:21:41Z" Dec 13 06:49:58 crc kubenswrapper[4971]: I1213 06:49:58.266060 4971 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 13 06:49:58 crc kubenswrapper[4971]: I1213 06:49:58.266134 4971 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 13 06:49:58 crc kubenswrapper[4971]: I1213 06:49:58.266148 4971 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 13 06:49:58 crc kubenswrapper[4971]: I1213 06:49:58.266169 4971 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 13 06:49:58 crc kubenswrapper[4971]: I1213 06:49:58.266183 4971 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-13T06:49:58Z","lastTransitionTime":"2025-12-13T06:49:58Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 13 06:49:58 crc kubenswrapper[4971]: I1213 06:49:58.276946 4971 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-ncchz" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"de914d21-a688-401e-a66e-861beb6336d9\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T06:49:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T06:49:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T06:49:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T06:49:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://244ffe3dc9d8cf799aff4c4d795bf41554c1e6bb6667fbb0c7df322fb5f96c68\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-13T06:49:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lqmjj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://0b7111af6c5a198d02bb02858837cfb4f3b20828d8d206cd18aa0ffa102d0cd4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://0b7111af6c5a198d02bb02858837cfb4f3b20828d8d206cd18aa0ffa102d0cd4\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-13T06:49:26Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-13T06:49:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lqmjj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b36777fe40e74ae1cecb7b73e204139db5279ba51cef4f1e8c4a4c491d8be403\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://b36777fe40e74ae1cecb7b73e204139db5279ba51cef4f1e8c4a4c491d8be403\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-13T06:49:28Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-13T06:49:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lqmjj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://0597a06cc663e416142311a16941d98297577693a1fe27201ea42c27496cace2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://0597a06cc663e416142311a16941d98297577693a1fe27201ea42c27496cace2\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-13T06:49:29Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-13T06:49:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lqmjj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://3f52c2e22a75c010993ca5809a3a9d1304758d4cb6e83691ef1af15a4b514ac4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://3f52c2e22a75c010993ca5809a3a9d1304758d4cb6e83691ef1af15a4b514ac4\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-13T06:49:30Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-13T06:49:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lqmjj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://9e21ca708c4c4c1228b417cece51299e0792729b7c78884bd18b2dda8fe8355c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://9e21ca708c4c4c1228b417cece51299e0792729b7c78884bd18b2dda8fe8355c\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-13T06:49:31Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-13T06:49:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lqmjj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://74552159f12c87cdda09147692b400ccac0d4e43c1ed72fe8462c4249ad69c7b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://74552159f12c87cdda09147692b400ccac0d4e43c1ed72fe8462c4249ad69c7b\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-13T06:49:32Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-13T06:49:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lqmjj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-13T06:49:25Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-ncchz\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-13T06:49:58Z is after 2025-08-24T17:21:41Z" Dec 13 06:49:58 crc kubenswrapper[4971]: I1213 06:49:58.291200 4971 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-gkcd8" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"13f5c327-f95e-42b4-8783-218500c17694\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T06:49:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T06:49:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T06:49:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T06:49:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://77e846e36f1cb40d03471dc03de11b6c3da448baa407aad38093fae9a801cf92\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-13T06:49:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gvlhn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-13T06:49:27Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-gkcd8\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-13T06:49:58Z is after 2025-08-24T17:21:41Z" Dec 13 06:49:58 crc kubenswrapper[4971]: I1213 06:49:58.301660 4971 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-x98k7" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"c7185dec-3240-42e8-9d90-6182a5e8b65c\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T06:49:39Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T06:49:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T06:49:39Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T06:49:39Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bj2gn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bj2gn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-13T06:49:39Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-x98k7\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-13T06:49:58Z is after 2025-08-24T17:21:41Z" Dec 13 06:49:58 crc kubenswrapper[4971]: I1213 06:49:58.318792 4971 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"f9d31b5a-b770-4a15-95c8-35b49677c608\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T06:49:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T06:49:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T06:49:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T06:49:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T06:49:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c9baa042f9dada954faa57b5454cd6b0a12a297e56ddc192c52cff50e42a16fa\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-13T06:49:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://c788de86754f0bf37ee184672fc1dcb388ccebc246d4f15be4b5da661f725f8f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-13T06:49:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://760817511a9b4f96a68dc38d26631250eaab72e1dd8f123e653039b6b4e91f84\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-13T06:49:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b859be0c9feac856e9855c6a038b02b49b32fade97873c46d0bec0a32d8b5147\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-13T06:49:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-13T06:49:03Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-13T06:49:58Z is after 2025-08-24T17:21:41Z" Dec 13 06:49:58 crc kubenswrapper[4971]: I1213 06:49:58.337273 4971 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"1ed9f82e-fe5c-4b76-a619-62e5abd88d3b\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T06:49:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T06:49:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T06:49:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T06:49:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T06:49:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e457329a78a2d24337f994e8a0a1f75a2a667244aa75eb84613d1e7138364aac\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-13T06:49:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://eb014de8140deef097d9bd94f3b9fa4ab3558bc4449bef84b83bf62bfabf5836\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-13T06:49:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://2c438cc74d555f7c9db4569d1c56b78b1088ed8105697d2c0f31e6c5136f3dd0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-13T06:49:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://aa6fa6e4340dcd066252df3c56d79a3a39ed531d5eea8d1d5941d6952368d3af\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-13T06:49:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://c4471b98854c52d5cd084d0ba3e64e5b08e9518ec3c372cce2c875a440522457\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-13T06:49:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7d2746335c98dae3a0f43024d34cb62ebcb42a01c8f9ce18d52b29d4c45598e5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://7d2746335c98dae3a0f43024d34cb62ebcb42a01c8f9ce18d52b29d4c45598e5\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-13T06:49:04Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-13T06:49:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://bc0e77d10a738128bea82122496bc42235fc1688f8907756d93e61b651902b57\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://bc0e77d10a738128bea82122496bc42235fc1688f8907756d93e61b651902b57\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-13T06:49:05Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-13T06:49:05Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://f821e418cba5e832da8839268bcbbd9630bdf319a212d5382bfa2c4b0b73a268\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://f821e418cba5e832da8839268bcbbd9630bdf319a212d5382bfa2c4b0b73a268\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-13T06:49:07Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-13T06:49:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-13T06:49:03Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-13T06:49:58Z is after 2025-08-24T17:21:41Z" Dec 13 06:49:58 crc kubenswrapper[4971]: I1213 06:49:58.348655 4971 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-13T06:49:24Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-13T06:49:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-13T06:49:24Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4328cd6a8464ed468229bbfbc8265f969427916d72ebb92237dbe68efe70fa47\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-13T06:49:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://9d715ca82d9b5a594e41f2b98da6f4277da0e4e1dec114969f901c00f6d7553d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-13T06:49:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-13T06:49:58Z is after 2025-08-24T17:21:41Z" Dec 13 06:49:58 crc kubenswrapper[4971]: I1213 06:49:58.358088 4971 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-5rzh5" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"6bf4591d-2483-4132-9672-d0fd6c995968\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T06:49:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T06:49:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T06:49:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T06:49:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://50e2cb14e4ffb603acb06e2281615a99f06182d345c4e70f2446fa2077033bbb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-13T06:49:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dfxqn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-13T06:49:25Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-5rzh5\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-13T06:49:58Z is after 2025-08-24T17:21:41Z" Dec 13 06:49:58 crc kubenswrapper[4971]: I1213 06:49:58.368925 4971 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 13 06:49:58 crc kubenswrapper[4971]: I1213 06:49:58.368955 4971 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 13 06:49:58 crc kubenswrapper[4971]: I1213 06:49:58.368964 4971 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 13 06:49:58 crc kubenswrapper[4971]: I1213 06:49:58.368978 4971 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 13 06:49:58 crc kubenswrapper[4971]: I1213 06:49:58.368987 4971 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-13T06:49:58Z","lastTransitionTime":"2025-12-13T06:49:58Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 13 06:49:58 crc kubenswrapper[4971]: I1213 06:49:58.369012 4971 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-z572s" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"b4146d3d-4034-49a9-bd37-164856255137\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T06:49:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T06:49:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T06:49:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T06:49:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://46fd4e1070835cfbf7f35d497eff7f047473fc4048ee07812925df5ebdf55342\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-13T06:49:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-j6l9b\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://dc79a0968f85723e6ad0dc269c8da1f7472adb1bb2b87edd017ee2295bfbf89c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-13T06:49:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-j6l9b\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-13T06:49:38Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-z572s\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-13T06:49:58Z is after 2025-08-24T17:21:41Z" Dec 13 06:49:58 crc kubenswrapper[4971]: I1213 06:49:58.384749 4971 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-13T06:49:23Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-13T06:49:23Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-13T06:49:58Z is after 2025-08-24T17:21:41Z" Dec 13 06:49:58 crc kubenswrapper[4971]: I1213 06:49:58.397696 4971 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-13T06:49:23Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-13T06:49:23Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-13T06:49:58Z is after 2025-08-24T17:21:41Z" Dec 13 06:49:58 crc kubenswrapper[4971]: I1213 06:49:58.409927 4971 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3ad3268a-a982-428c-9c06-4aaeddd372d9\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T06:49:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T06:49:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T06:49:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T06:49:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T06:49:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3716afdf2bc693be80a09fc9f7727d334f435caf1d4c8ec4811b3a1f1652cdfe\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-13T06:49:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://30df3abf8c79db8af52c6a918f73e1ce410ddafcf341636424c267e1eeb195a7\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-13T06:49:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://ccdf65f6d2b1eb3647a03ba37c305c32e34addfdf79ff937c8dddaa3e5d5b5c6\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-13T06:49:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://770b73dd0c2b6f38744e2c1cd6aa186a69bb87d1cd015a603ad2e33be06a8f6b\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://b3c1bab3d53f975b2e24582ce2771bc26d600e0dd7c30339dcfd18a071162ad4\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-12-13T06:49:23Z\\\",\\\"message\\\":\\\"file observer\\\\nW1213 06:49:23.283403 1 builder.go:272] unable to get owner reference (falling back to namespace): pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\nI1213 06:49:23.284336 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1213 06:49:23.286978 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-981773671/tls.crt::/tmp/serving-cert-981773671/tls.key\\\\\\\"\\\\nI1213 06:49:23.775072 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1213 06:49:23.778615 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1213 06:49:23.778645 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1213 06:49:23.778680 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1213 06:49:23.778691 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1213 06:49:23.785058 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1213 06:49:23.786150 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1213 06:49:23.786196 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1213 06:49:23.786223 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1213 06:49:23.786246 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1213 06:49:23.786269 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1213 06:49:23.786291 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI1213 06:49:23.785608 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF1213 06:49:23.787563 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-13T06:49:08Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-13T06:49:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://9acd07066deed483bd7fa61ee45fda09539d110717a134a56827f871186f92fc\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-13T06:49:07Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://38c91cfbaecfdbd3e3b663422a91ab4294610afb6bbf9ebdb77db24e2b76c682\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://38c91cfbaecfdbd3e3b663422a91ab4294610afb6bbf9ebdb77db24e2b76c682\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-13T06:49:05Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-13T06:49:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-13T06:49:03Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-13T06:49:58Z is after 2025-08-24T17:21:41Z" Dec 13 06:49:58 crc kubenswrapper[4971]: I1213 06:49:58.471064 4971 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 13 06:49:58 crc kubenswrapper[4971]: I1213 06:49:58.471150 4971 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 13 06:49:58 crc kubenswrapper[4971]: I1213 06:49:58.471163 4971 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 13 06:49:58 crc kubenswrapper[4971]: I1213 06:49:58.471182 4971 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 13 06:49:58 crc kubenswrapper[4971]: I1213 06:49:58.471194 4971 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-13T06:49:58Z","lastTransitionTime":"2025-12-13T06:49:58Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 13 06:49:58 crc kubenswrapper[4971]: I1213 06:49:58.573638 4971 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 13 06:49:58 crc kubenswrapper[4971]: I1213 06:49:58.573704 4971 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 13 06:49:58 crc kubenswrapper[4971]: I1213 06:49:58.573721 4971 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 13 06:49:58 crc kubenswrapper[4971]: I1213 06:49:58.573742 4971 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 13 06:49:58 crc kubenswrapper[4971]: I1213 06:49:58.573754 4971 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-13T06:49:58Z","lastTransitionTime":"2025-12-13T06:49:58Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 13 06:49:58 crc kubenswrapper[4971]: I1213 06:49:58.676511 4971 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 13 06:49:58 crc kubenswrapper[4971]: I1213 06:49:58.676630 4971 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 13 06:49:58 crc kubenswrapper[4971]: I1213 06:49:58.676651 4971 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 13 06:49:58 crc kubenswrapper[4971]: I1213 06:49:58.676680 4971 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 13 06:49:58 crc kubenswrapper[4971]: I1213 06:49:58.676702 4971 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-13T06:49:58Z","lastTransitionTime":"2025-12-13T06:49:58Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 13 06:49:58 crc kubenswrapper[4971]: I1213 06:49:58.779647 4971 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 13 06:49:58 crc kubenswrapper[4971]: I1213 06:49:58.779683 4971 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 13 06:49:58 crc kubenswrapper[4971]: I1213 06:49:58.779694 4971 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 13 06:49:58 crc kubenswrapper[4971]: I1213 06:49:58.779710 4971 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 13 06:49:58 crc kubenswrapper[4971]: I1213 06:49:58.779723 4971 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-13T06:49:58Z","lastTransitionTime":"2025-12-13T06:49:58Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 13 06:49:58 crc kubenswrapper[4971]: I1213 06:49:58.882399 4971 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 13 06:49:58 crc kubenswrapper[4971]: I1213 06:49:58.882475 4971 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 13 06:49:58 crc kubenswrapper[4971]: I1213 06:49:58.882498 4971 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 13 06:49:58 crc kubenswrapper[4971]: I1213 06:49:58.882577 4971 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 13 06:49:58 crc kubenswrapper[4971]: I1213 06:49:58.882605 4971 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-13T06:49:58Z","lastTransitionTime":"2025-12-13T06:49:58Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 13 06:49:58 crc kubenswrapper[4971]: I1213 06:49:58.984903 4971 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 13 06:49:58 crc kubenswrapper[4971]: I1213 06:49:58.984952 4971 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 13 06:49:58 crc kubenswrapper[4971]: I1213 06:49:58.984962 4971 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 13 06:49:58 crc kubenswrapper[4971]: I1213 06:49:58.984979 4971 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 13 06:49:58 crc kubenswrapper[4971]: I1213 06:49:58.984990 4971 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-13T06:49:58Z","lastTransitionTime":"2025-12-13T06:49:58Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 13 06:49:59 crc kubenswrapper[4971]: I1213 06:49:59.087083 4971 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 13 06:49:59 crc kubenswrapper[4971]: I1213 06:49:59.087115 4971 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 13 06:49:59 crc kubenswrapper[4971]: I1213 06:49:59.087125 4971 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 13 06:49:59 crc kubenswrapper[4971]: I1213 06:49:59.087143 4971 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 13 06:49:59 crc kubenswrapper[4971]: I1213 06:49:59.087152 4971 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-13T06:49:59Z","lastTransitionTime":"2025-12-13T06:49:59Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 13 06:49:59 crc kubenswrapper[4971]: I1213 06:49:59.189477 4971 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 13 06:49:59 crc kubenswrapper[4971]: I1213 06:49:59.189511 4971 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 13 06:49:59 crc kubenswrapper[4971]: I1213 06:49:59.189576 4971 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 13 06:49:59 crc kubenswrapper[4971]: I1213 06:49:59.189599 4971 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 13 06:49:59 crc kubenswrapper[4971]: I1213 06:49:59.189613 4971 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-13T06:49:59Z","lastTransitionTime":"2025-12-13T06:49:59Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 13 06:49:59 crc kubenswrapper[4971]: I1213 06:49:59.293148 4971 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 13 06:49:59 crc kubenswrapper[4971]: I1213 06:49:59.293247 4971 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 13 06:49:59 crc kubenswrapper[4971]: I1213 06:49:59.293259 4971 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 13 06:49:59 crc kubenswrapper[4971]: I1213 06:49:59.293285 4971 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 13 06:49:59 crc kubenswrapper[4971]: I1213 06:49:59.293300 4971 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-13T06:49:59Z","lastTransitionTime":"2025-12-13T06:49:59Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 13 06:49:59 crc kubenswrapper[4971]: I1213 06:49:59.395931 4971 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 13 06:49:59 crc kubenswrapper[4971]: I1213 06:49:59.395987 4971 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 13 06:49:59 crc kubenswrapper[4971]: I1213 06:49:59.396002 4971 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 13 06:49:59 crc kubenswrapper[4971]: I1213 06:49:59.396023 4971 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 13 06:49:59 crc kubenswrapper[4971]: I1213 06:49:59.396037 4971 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-13T06:49:59Z","lastTransitionTime":"2025-12-13T06:49:59Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 13 06:49:59 crc kubenswrapper[4971]: I1213 06:49:59.500019 4971 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 13 06:49:59 crc kubenswrapper[4971]: I1213 06:49:59.500063 4971 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 13 06:49:59 crc kubenswrapper[4971]: I1213 06:49:59.500075 4971 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 13 06:49:59 crc kubenswrapper[4971]: I1213 06:49:59.500092 4971 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 13 06:49:59 crc kubenswrapper[4971]: I1213 06:49:59.500104 4971 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-13T06:49:59Z","lastTransitionTime":"2025-12-13T06:49:59Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 13 06:49:59 crc kubenswrapper[4971]: I1213 06:49:59.602876 4971 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 13 06:49:59 crc kubenswrapper[4971]: I1213 06:49:59.602913 4971 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 13 06:49:59 crc kubenswrapper[4971]: I1213 06:49:59.602924 4971 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 13 06:49:59 crc kubenswrapper[4971]: I1213 06:49:59.602942 4971 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 13 06:49:59 crc kubenswrapper[4971]: I1213 06:49:59.602954 4971 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-13T06:49:59Z","lastTransitionTime":"2025-12-13T06:49:59Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 13 06:49:59 crc kubenswrapper[4971]: I1213 06:49:59.705600 4971 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 13 06:49:59 crc kubenswrapper[4971]: I1213 06:49:59.705629 4971 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 13 06:49:59 crc kubenswrapper[4971]: I1213 06:49:59.705637 4971 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 13 06:49:59 crc kubenswrapper[4971]: I1213 06:49:59.705649 4971 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 13 06:49:59 crc kubenswrapper[4971]: I1213 06:49:59.705658 4971 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-13T06:49:59Z","lastTransitionTime":"2025-12-13T06:49:59Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 13 06:49:59 crc kubenswrapper[4971]: I1213 06:49:59.767992 4971 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 13 06:49:59 crc kubenswrapper[4971]: I1213 06:49:59.768055 4971 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-x98k7" Dec 13 06:49:59 crc kubenswrapper[4971]: E1213 06:49:59.768125 4971 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 13 06:49:59 crc kubenswrapper[4971]: I1213 06:49:59.767995 4971 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 13 06:49:59 crc kubenswrapper[4971]: E1213 06:49:59.768204 4971 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-x98k7" podUID="c7185dec-3240-42e8-9d90-6182a5e8b65c" Dec 13 06:49:59 crc kubenswrapper[4971]: I1213 06:49:59.768217 4971 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 13 06:49:59 crc kubenswrapper[4971]: E1213 06:49:59.768267 4971 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 13 06:49:59 crc kubenswrapper[4971]: E1213 06:49:59.768323 4971 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 13 06:49:59 crc kubenswrapper[4971]: I1213 06:49:59.808389 4971 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 13 06:49:59 crc kubenswrapper[4971]: I1213 06:49:59.808448 4971 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 13 06:49:59 crc kubenswrapper[4971]: I1213 06:49:59.808468 4971 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 13 06:49:59 crc kubenswrapper[4971]: I1213 06:49:59.808493 4971 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 13 06:49:59 crc kubenswrapper[4971]: I1213 06:49:59.808509 4971 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-13T06:49:59Z","lastTransitionTime":"2025-12-13T06:49:59Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 13 06:49:59 crc kubenswrapper[4971]: I1213 06:49:59.911227 4971 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 13 06:49:59 crc kubenswrapper[4971]: I1213 06:49:59.911291 4971 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 13 06:49:59 crc kubenswrapper[4971]: I1213 06:49:59.911304 4971 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 13 06:49:59 crc kubenswrapper[4971]: I1213 06:49:59.911327 4971 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 13 06:49:59 crc kubenswrapper[4971]: I1213 06:49:59.911343 4971 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-13T06:49:59Z","lastTransitionTime":"2025-12-13T06:49:59Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 13 06:50:00 crc kubenswrapper[4971]: I1213 06:50:00.013273 4971 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 13 06:50:00 crc kubenswrapper[4971]: I1213 06:50:00.013306 4971 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 13 06:50:00 crc kubenswrapper[4971]: I1213 06:50:00.013314 4971 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 13 06:50:00 crc kubenswrapper[4971]: I1213 06:50:00.013328 4971 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 13 06:50:00 crc kubenswrapper[4971]: I1213 06:50:00.013336 4971 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-13T06:50:00Z","lastTransitionTime":"2025-12-13T06:50:00Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 13 06:50:00 crc kubenswrapper[4971]: I1213 06:50:00.116262 4971 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 13 06:50:00 crc kubenswrapper[4971]: I1213 06:50:00.116313 4971 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 13 06:50:00 crc kubenswrapper[4971]: I1213 06:50:00.116322 4971 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 13 06:50:00 crc kubenswrapper[4971]: I1213 06:50:00.116336 4971 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 13 06:50:00 crc kubenswrapper[4971]: I1213 06:50:00.116345 4971 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-13T06:50:00Z","lastTransitionTime":"2025-12-13T06:50:00Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 13 06:50:00 crc kubenswrapper[4971]: I1213 06:50:00.219134 4971 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 13 06:50:00 crc kubenswrapper[4971]: I1213 06:50:00.219183 4971 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 13 06:50:00 crc kubenswrapper[4971]: I1213 06:50:00.219197 4971 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 13 06:50:00 crc kubenswrapper[4971]: I1213 06:50:00.219219 4971 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 13 06:50:00 crc kubenswrapper[4971]: I1213 06:50:00.219236 4971 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-13T06:50:00Z","lastTransitionTime":"2025-12-13T06:50:00Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 13 06:50:00 crc kubenswrapper[4971]: I1213 06:50:00.322323 4971 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 13 06:50:00 crc kubenswrapper[4971]: I1213 06:50:00.322406 4971 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 13 06:50:00 crc kubenswrapper[4971]: I1213 06:50:00.322434 4971 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 13 06:50:00 crc kubenswrapper[4971]: I1213 06:50:00.322465 4971 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 13 06:50:00 crc kubenswrapper[4971]: I1213 06:50:00.322486 4971 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-13T06:50:00Z","lastTransitionTime":"2025-12-13T06:50:00Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 13 06:50:00 crc kubenswrapper[4971]: I1213 06:50:00.425073 4971 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 13 06:50:00 crc kubenswrapper[4971]: I1213 06:50:00.425110 4971 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 13 06:50:00 crc kubenswrapper[4971]: I1213 06:50:00.425121 4971 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 13 06:50:00 crc kubenswrapper[4971]: I1213 06:50:00.425138 4971 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 13 06:50:00 crc kubenswrapper[4971]: I1213 06:50:00.425147 4971 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-13T06:50:00Z","lastTransitionTime":"2025-12-13T06:50:00Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 13 06:50:00 crc kubenswrapper[4971]: I1213 06:50:00.528134 4971 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 13 06:50:00 crc kubenswrapper[4971]: I1213 06:50:00.528225 4971 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 13 06:50:00 crc kubenswrapper[4971]: I1213 06:50:00.528241 4971 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 13 06:50:00 crc kubenswrapper[4971]: I1213 06:50:00.528265 4971 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 13 06:50:00 crc kubenswrapper[4971]: I1213 06:50:00.528279 4971 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-13T06:50:00Z","lastTransitionTime":"2025-12-13T06:50:00Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 13 06:50:00 crc kubenswrapper[4971]: I1213 06:50:00.631481 4971 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 13 06:50:00 crc kubenswrapper[4971]: I1213 06:50:00.631541 4971 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 13 06:50:00 crc kubenswrapper[4971]: I1213 06:50:00.631552 4971 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 13 06:50:00 crc kubenswrapper[4971]: I1213 06:50:00.631570 4971 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 13 06:50:00 crc kubenswrapper[4971]: I1213 06:50:00.631581 4971 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-13T06:50:00Z","lastTransitionTime":"2025-12-13T06:50:00Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 13 06:50:00 crc kubenswrapper[4971]: I1213 06:50:00.733887 4971 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 13 06:50:00 crc kubenswrapper[4971]: I1213 06:50:00.733921 4971 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 13 06:50:00 crc kubenswrapper[4971]: I1213 06:50:00.733930 4971 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 13 06:50:00 crc kubenswrapper[4971]: I1213 06:50:00.733948 4971 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 13 06:50:00 crc kubenswrapper[4971]: I1213 06:50:00.733959 4971 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-13T06:50:00Z","lastTransitionTime":"2025-12-13T06:50:00Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 13 06:50:00 crc kubenswrapper[4971]: I1213 06:50:00.817118 4971 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 13 06:50:00 crc kubenswrapper[4971]: I1213 06:50:00.817162 4971 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 13 06:50:00 crc kubenswrapper[4971]: I1213 06:50:00.817179 4971 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 13 06:50:00 crc kubenswrapper[4971]: I1213 06:50:00.817196 4971 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 13 06:50:00 crc kubenswrapper[4971]: I1213 06:50:00.817206 4971 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-13T06:50:00Z","lastTransitionTime":"2025-12-13T06:50:00Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 13 06:50:00 crc kubenswrapper[4971]: E1213 06:50:00.828992 4971 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-12-13T06:50:00Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-13T06:50:00Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-13T06:50:00Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-13T06:50:00Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-13T06:50:00Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-13T06:50:00Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-13T06:50:00Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-13T06:50:00Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"bd36cbc9-bbe9-45a1-a21a-ad2bd10499ee\\\",\\\"systemUUID\\\":\\\"a9af00dc-cf25-47af-9997-79577f8f3bea\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-13T06:50:00Z is after 2025-08-24T17:21:41Z" Dec 13 06:50:00 crc kubenswrapper[4971]: I1213 06:50:00.832376 4971 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 13 06:50:00 crc kubenswrapper[4971]: I1213 06:50:00.832415 4971 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 13 06:50:00 crc kubenswrapper[4971]: I1213 06:50:00.832424 4971 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 13 06:50:00 crc kubenswrapper[4971]: I1213 06:50:00.832442 4971 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 13 06:50:00 crc kubenswrapper[4971]: I1213 06:50:00.832452 4971 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-13T06:50:00Z","lastTransitionTime":"2025-12-13T06:50:00Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 13 06:50:00 crc kubenswrapper[4971]: E1213 06:50:00.843622 4971 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-12-13T06:50:00Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-13T06:50:00Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-13T06:50:00Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-13T06:50:00Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-13T06:50:00Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-13T06:50:00Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-13T06:50:00Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-13T06:50:00Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"bd36cbc9-bbe9-45a1-a21a-ad2bd10499ee\\\",\\\"systemUUID\\\":\\\"a9af00dc-cf25-47af-9997-79577f8f3bea\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-13T06:50:00Z is after 2025-08-24T17:21:41Z" Dec 13 06:50:00 crc kubenswrapper[4971]: I1213 06:50:00.847334 4971 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 13 06:50:00 crc kubenswrapper[4971]: I1213 06:50:00.847395 4971 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 13 06:50:00 crc kubenswrapper[4971]: I1213 06:50:00.847410 4971 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 13 06:50:00 crc kubenswrapper[4971]: I1213 06:50:00.847434 4971 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 13 06:50:00 crc kubenswrapper[4971]: I1213 06:50:00.847449 4971 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-13T06:50:00Z","lastTransitionTime":"2025-12-13T06:50:00Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 13 06:50:00 crc kubenswrapper[4971]: E1213 06:50:00.861639 4971 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-12-13T06:50:00Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-13T06:50:00Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-13T06:50:00Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-13T06:50:00Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-13T06:50:00Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-13T06:50:00Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-13T06:50:00Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-13T06:50:00Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"bd36cbc9-bbe9-45a1-a21a-ad2bd10499ee\\\",\\\"systemUUID\\\":\\\"a9af00dc-cf25-47af-9997-79577f8f3bea\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-13T06:50:00Z is after 2025-08-24T17:21:41Z" Dec 13 06:50:00 crc kubenswrapper[4971]: I1213 06:50:00.865294 4971 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 13 06:50:00 crc kubenswrapper[4971]: I1213 06:50:00.865359 4971 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 13 06:50:00 crc kubenswrapper[4971]: I1213 06:50:00.865374 4971 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 13 06:50:00 crc kubenswrapper[4971]: I1213 06:50:00.865402 4971 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 13 06:50:00 crc kubenswrapper[4971]: I1213 06:50:00.865417 4971 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-13T06:50:00Z","lastTransitionTime":"2025-12-13T06:50:00Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 13 06:50:00 crc kubenswrapper[4971]: E1213 06:50:00.878409 4971 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-12-13T06:50:00Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-13T06:50:00Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-13T06:50:00Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-13T06:50:00Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-13T06:50:00Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-13T06:50:00Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-13T06:50:00Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-13T06:50:00Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"bd36cbc9-bbe9-45a1-a21a-ad2bd10499ee\\\",\\\"systemUUID\\\":\\\"a9af00dc-cf25-47af-9997-79577f8f3bea\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-13T06:50:00Z is after 2025-08-24T17:21:41Z" Dec 13 06:50:00 crc kubenswrapper[4971]: I1213 06:50:00.882230 4971 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 13 06:50:00 crc kubenswrapper[4971]: I1213 06:50:00.882277 4971 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 13 06:50:00 crc kubenswrapper[4971]: I1213 06:50:00.882292 4971 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 13 06:50:00 crc kubenswrapper[4971]: I1213 06:50:00.882317 4971 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 13 06:50:00 crc kubenswrapper[4971]: I1213 06:50:00.882329 4971 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-13T06:50:00Z","lastTransitionTime":"2025-12-13T06:50:00Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 13 06:50:00 crc kubenswrapper[4971]: E1213 06:50:00.893738 4971 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-12-13T06:50:00Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-13T06:50:00Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-13T06:50:00Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-13T06:50:00Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-13T06:50:00Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-13T06:50:00Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-13T06:50:00Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-13T06:50:00Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"bd36cbc9-bbe9-45a1-a21a-ad2bd10499ee\\\",\\\"systemUUID\\\":\\\"a9af00dc-cf25-47af-9997-79577f8f3bea\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-13T06:50:00Z is after 2025-08-24T17:21:41Z" Dec 13 06:50:00 crc kubenswrapper[4971]: E1213 06:50:00.893855 4971 kubelet_node_status.go:572] "Unable to update node status" err="update node status exceeds retry count" Dec 13 06:50:00 crc kubenswrapper[4971]: I1213 06:50:00.895250 4971 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 13 06:50:00 crc kubenswrapper[4971]: I1213 06:50:00.895283 4971 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 13 06:50:00 crc kubenswrapper[4971]: I1213 06:50:00.895300 4971 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 13 06:50:00 crc kubenswrapper[4971]: I1213 06:50:00.895317 4971 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 13 06:50:00 crc kubenswrapper[4971]: I1213 06:50:00.895331 4971 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-13T06:50:00Z","lastTransitionTime":"2025-12-13T06:50:00Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 13 06:50:00 crc kubenswrapper[4971]: I1213 06:50:00.997628 4971 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 13 06:50:00 crc kubenswrapper[4971]: I1213 06:50:00.997676 4971 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 13 06:50:00 crc kubenswrapper[4971]: I1213 06:50:00.997688 4971 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 13 06:50:00 crc kubenswrapper[4971]: I1213 06:50:00.997706 4971 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 13 06:50:00 crc kubenswrapper[4971]: I1213 06:50:00.997718 4971 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-13T06:50:00Z","lastTransitionTime":"2025-12-13T06:50:00Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 13 06:50:01 crc kubenswrapper[4971]: I1213 06:50:01.100305 4971 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 13 06:50:01 crc kubenswrapper[4971]: I1213 06:50:01.100385 4971 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 13 06:50:01 crc kubenswrapper[4971]: I1213 06:50:01.100407 4971 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 13 06:50:01 crc kubenswrapper[4971]: I1213 06:50:01.100436 4971 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 13 06:50:01 crc kubenswrapper[4971]: I1213 06:50:01.100458 4971 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-13T06:50:01Z","lastTransitionTime":"2025-12-13T06:50:01Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 13 06:50:01 crc kubenswrapper[4971]: I1213 06:50:01.202864 4971 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 13 06:50:01 crc kubenswrapper[4971]: I1213 06:50:01.202908 4971 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 13 06:50:01 crc kubenswrapper[4971]: I1213 06:50:01.202916 4971 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 13 06:50:01 crc kubenswrapper[4971]: I1213 06:50:01.202930 4971 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 13 06:50:01 crc kubenswrapper[4971]: I1213 06:50:01.202938 4971 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-13T06:50:01Z","lastTransitionTime":"2025-12-13T06:50:01Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 13 06:50:01 crc kubenswrapper[4971]: I1213 06:50:01.306182 4971 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 13 06:50:01 crc kubenswrapper[4971]: I1213 06:50:01.306255 4971 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 13 06:50:01 crc kubenswrapper[4971]: I1213 06:50:01.306269 4971 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 13 06:50:01 crc kubenswrapper[4971]: I1213 06:50:01.306292 4971 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 13 06:50:01 crc kubenswrapper[4971]: I1213 06:50:01.306305 4971 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-13T06:50:01Z","lastTransitionTime":"2025-12-13T06:50:01Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 13 06:50:01 crc kubenswrapper[4971]: I1213 06:50:01.409188 4971 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 13 06:50:01 crc kubenswrapper[4971]: I1213 06:50:01.409237 4971 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 13 06:50:01 crc kubenswrapper[4971]: I1213 06:50:01.409253 4971 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 13 06:50:01 crc kubenswrapper[4971]: I1213 06:50:01.409276 4971 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 13 06:50:01 crc kubenswrapper[4971]: I1213 06:50:01.409293 4971 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-13T06:50:01Z","lastTransitionTime":"2025-12-13T06:50:01Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 13 06:50:01 crc kubenswrapper[4971]: I1213 06:50:01.511576 4971 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 13 06:50:01 crc kubenswrapper[4971]: I1213 06:50:01.511630 4971 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 13 06:50:01 crc kubenswrapper[4971]: I1213 06:50:01.511642 4971 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 13 06:50:01 crc kubenswrapper[4971]: I1213 06:50:01.511661 4971 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 13 06:50:01 crc kubenswrapper[4971]: I1213 06:50:01.511674 4971 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-13T06:50:01Z","lastTransitionTime":"2025-12-13T06:50:01Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 13 06:50:01 crc kubenswrapper[4971]: I1213 06:50:01.613610 4971 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 13 06:50:01 crc kubenswrapper[4971]: I1213 06:50:01.613650 4971 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 13 06:50:01 crc kubenswrapper[4971]: I1213 06:50:01.613658 4971 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 13 06:50:01 crc kubenswrapper[4971]: I1213 06:50:01.613674 4971 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 13 06:50:01 crc kubenswrapper[4971]: I1213 06:50:01.613683 4971 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-13T06:50:01Z","lastTransitionTime":"2025-12-13T06:50:01Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 13 06:50:01 crc kubenswrapper[4971]: I1213 06:50:01.716745 4971 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 13 06:50:01 crc kubenswrapper[4971]: I1213 06:50:01.716849 4971 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 13 06:50:01 crc kubenswrapper[4971]: I1213 06:50:01.716871 4971 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 13 06:50:01 crc kubenswrapper[4971]: I1213 06:50:01.716899 4971 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 13 06:50:01 crc kubenswrapper[4971]: I1213 06:50:01.716918 4971 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-13T06:50:01Z","lastTransitionTime":"2025-12-13T06:50:01Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 13 06:50:01 crc kubenswrapper[4971]: I1213 06:50:01.768284 4971 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 13 06:50:01 crc kubenswrapper[4971]: I1213 06:50:01.768358 4971 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 13 06:50:01 crc kubenswrapper[4971]: I1213 06:50:01.768396 4971 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-x98k7" Dec 13 06:50:01 crc kubenswrapper[4971]: E1213 06:50:01.768555 4971 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 13 06:50:01 crc kubenswrapper[4971]: I1213 06:50:01.768768 4971 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 13 06:50:01 crc kubenswrapper[4971]: E1213 06:50:01.768930 4971 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 13 06:50:01 crc kubenswrapper[4971]: E1213 06:50:01.769012 4971 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 13 06:50:01 crc kubenswrapper[4971]: E1213 06:50:01.769172 4971 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-x98k7" podUID="c7185dec-3240-42e8-9d90-6182a5e8b65c" Dec 13 06:50:01 crc kubenswrapper[4971]: I1213 06:50:01.821007 4971 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 13 06:50:01 crc kubenswrapper[4971]: I1213 06:50:01.821101 4971 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 13 06:50:01 crc kubenswrapper[4971]: I1213 06:50:01.821137 4971 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 13 06:50:01 crc kubenswrapper[4971]: I1213 06:50:01.821190 4971 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 13 06:50:01 crc kubenswrapper[4971]: I1213 06:50:01.821215 4971 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-13T06:50:01Z","lastTransitionTime":"2025-12-13T06:50:01Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 13 06:50:01 crc kubenswrapper[4971]: I1213 06:50:01.924479 4971 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 13 06:50:01 crc kubenswrapper[4971]: I1213 06:50:01.924571 4971 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 13 06:50:01 crc kubenswrapper[4971]: I1213 06:50:01.924584 4971 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 13 06:50:01 crc kubenswrapper[4971]: I1213 06:50:01.924611 4971 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 13 06:50:01 crc kubenswrapper[4971]: I1213 06:50:01.924637 4971 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-13T06:50:01Z","lastTransitionTime":"2025-12-13T06:50:01Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 13 06:50:02 crc kubenswrapper[4971]: I1213 06:50:02.027216 4971 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 13 06:50:02 crc kubenswrapper[4971]: I1213 06:50:02.027253 4971 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 13 06:50:02 crc kubenswrapper[4971]: I1213 06:50:02.027262 4971 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 13 06:50:02 crc kubenswrapper[4971]: I1213 06:50:02.027279 4971 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 13 06:50:02 crc kubenswrapper[4971]: I1213 06:50:02.027288 4971 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-13T06:50:02Z","lastTransitionTime":"2025-12-13T06:50:02Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 13 06:50:02 crc kubenswrapper[4971]: I1213 06:50:02.129567 4971 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 13 06:50:02 crc kubenswrapper[4971]: I1213 06:50:02.129606 4971 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 13 06:50:02 crc kubenswrapper[4971]: I1213 06:50:02.129615 4971 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 13 06:50:02 crc kubenswrapper[4971]: I1213 06:50:02.129628 4971 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 13 06:50:02 crc kubenswrapper[4971]: I1213 06:50:02.129637 4971 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-13T06:50:02Z","lastTransitionTime":"2025-12-13T06:50:02Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 13 06:50:02 crc kubenswrapper[4971]: I1213 06:50:02.232241 4971 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 13 06:50:02 crc kubenswrapper[4971]: I1213 06:50:02.232280 4971 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 13 06:50:02 crc kubenswrapper[4971]: I1213 06:50:02.232288 4971 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 13 06:50:02 crc kubenswrapper[4971]: I1213 06:50:02.232302 4971 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 13 06:50:02 crc kubenswrapper[4971]: I1213 06:50:02.232310 4971 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-13T06:50:02Z","lastTransitionTime":"2025-12-13T06:50:02Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 13 06:50:02 crc kubenswrapper[4971]: I1213 06:50:02.334759 4971 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 13 06:50:02 crc kubenswrapper[4971]: I1213 06:50:02.334794 4971 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 13 06:50:02 crc kubenswrapper[4971]: I1213 06:50:02.334802 4971 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 13 06:50:02 crc kubenswrapper[4971]: I1213 06:50:02.334820 4971 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 13 06:50:02 crc kubenswrapper[4971]: I1213 06:50:02.334830 4971 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-13T06:50:02Z","lastTransitionTime":"2025-12-13T06:50:02Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 13 06:50:02 crc kubenswrapper[4971]: I1213 06:50:02.437740 4971 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 13 06:50:02 crc kubenswrapper[4971]: I1213 06:50:02.437780 4971 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 13 06:50:02 crc kubenswrapper[4971]: I1213 06:50:02.437788 4971 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 13 06:50:02 crc kubenswrapper[4971]: I1213 06:50:02.437804 4971 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 13 06:50:02 crc kubenswrapper[4971]: I1213 06:50:02.437814 4971 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-13T06:50:02Z","lastTransitionTime":"2025-12-13T06:50:02Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 13 06:50:02 crc kubenswrapper[4971]: I1213 06:50:02.540077 4971 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 13 06:50:02 crc kubenswrapper[4971]: I1213 06:50:02.540127 4971 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 13 06:50:02 crc kubenswrapper[4971]: I1213 06:50:02.540142 4971 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 13 06:50:02 crc kubenswrapper[4971]: I1213 06:50:02.540159 4971 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 13 06:50:02 crc kubenswrapper[4971]: I1213 06:50:02.540169 4971 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-13T06:50:02Z","lastTransitionTime":"2025-12-13T06:50:02Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 13 06:50:02 crc kubenswrapper[4971]: I1213 06:50:02.642407 4971 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 13 06:50:02 crc kubenswrapper[4971]: I1213 06:50:02.642659 4971 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 13 06:50:02 crc kubenswrapper[4971]: I1213 06:50:02.642676 4971 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 13 06:50:02 crc kubenswrapper[4971]: I1213 06:50:02.642692 4971 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 13 06:50:02 crc kubenswrapper[4971]: I1213 06:50:02.642703 4971 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-13T06:50:02Z","lastTransitionTime":"2025-12-13T06:50:02Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 13 06:50:02 crc kubenswrapper[4971]: I1213 06:50:02.745304 4971 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 13 06:50:02 crc kubenswrapper[4971]: I1213 06:50:02.745392 4971 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 13 06:50:02 crc kubenswrapper[4971]: I1213 06:50:02.745407 4971 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 13 06:50:02 crc kubenswrapper[4971]: I1213 06:50:02.745424 4971 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 13 06:50:02 crc kubenswrapper[4971]: I1213 06:50:02.745439 4971 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-13T06:50:02Z","lastTransitionTime":"2025-12-13T06:50:02Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 13 06:50:02 crc kubenswrapper[4971]: I1213 06:50:02.847869 4971 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 13 06:50:02 crc kubenswrapper[4971]: I1213 06:50:02.847908 4971 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 13 06:50:02 crc kubenswrapper[4971]: I1213 06:50:02.847918 4971 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 13 06:50:02 crc kubenswrapper[4971]: I1213 06:50:02.847935 4971 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 13 06:50:02 crc kubenswrapper[4971]: I1213 06:50:02.847947 4971 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-13T06:50:02Z","lastTransitionTime":"2025-12-13T06:50:02Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 13 06:50:02 crc kubenswrapper[4971]: I1213 06:50:02.949861 4971 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 13 06:50:02 crc kubenswrapper[4971]: I1213 06:50:02.949919 4971 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 13 06:50:02 crc kubenswrapper[4971]: I1213 06:50:02.949932 4971 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 13 06:50:02 crc kubenswrapper[4971]: I1213 06:50:02.949950 4971 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 13 06:50:02 crc kubenswrapper[4971]: I1213 06:50:02.949962 4971 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-13T06:50:02Z","lastTransitionTime":"2025-12-13T06:50:02Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 13 06:50:03 crc kubenswrapper[4971]: I1213 06:50:03.053078 4971 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 13 06:50:03 crc kubenswrapper[4971]: I1213 06:50:03.053121 4971 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 13 06:50:03 crc kubenswrapper[4971]: I1213 06:50:03.053136 4971 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 13 06:50:03 crc kubenswrapper[4971]: I1213 06:50:03.053150 4971 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 13 06:50:03 crc kubenswrapper[4971]: I1213 06:50:03.053160 4971 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-13T06:50:03Z","lastTransitionTime":"2025-12-13T06:50:03Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 13 06:50:03 crc kubenswrapper[4971]: I1213 06:50:03.155868 4971 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 13 06:50:03 crc kubenswrapper[4971]: I1213 06:50:03.155907 4971 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 13 06:50:03 crc kubenswrapper[4971]: I1213 06:50:03.155915 4971 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 13 06:50:03 crc kubenswrapper[4971]: I1213 06:50:03.155929 4971 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 13 06:50:03 crc kubenswrapper[4971]: I1213 06:50:03.155939 4971 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-13T06:50:03Z","lastTransitionTime":"2025-12-13T06:50:03Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 13 06:50:03 crc kubenswrapper[4971]: I1213 06:50:03.258680 4971 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 13 06:50:03 crc kubenswrapper[4971]: I1213 06:50:03.258783 4971 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 13 06:50:03 crc kubenswrapper[4971]: I1213 06:50:03.258805 4971 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 13 06:50:03 crc kubenswrapper[4971]: I1213 06:50:03.258868 4971 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 13 06:50:03 crc kubenswrapper[4971]: I1213 06:50:03.258890 4971 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-13T06:50:03Z","lastTransitionTime":"2025-12-13T06:50:03Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 13 06:50:03 crc kubenswrapper[4971]: I1213 06:50:03.361504 4971 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 13 06:50:03 crc kubenswrapper[4971]: I1213 06:50:03.361564 4971 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 13 06:50:03 crc kubenswrapper[4971]: I1213 06:50:03.361578 4971 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 13 06:50:03 crc kubenswrapper[4971]: I1213 06:50:03.361594 4971 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 13 06:50:03 crc kubenswrapper[4971]: I1213 06:50:03.361604 4971 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-13T06:50:03Z","lastTransitionTime":"2025-12-13T06:50:03Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 13 06:50:03 crc kubenswrapper[4971]: I1213 06:50:03.465541 4971 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 13 06:50:03 crc kubenswrapper[4971]: I1213 06:50:03.465590 4971 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 13 06:50:03 crc kubenswrapper[4971]: I1213 06:50:03.465604 4971 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 13 06:50:03 crc kubenswrapper[4971]: I1213 06:50:03.465628 4971 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 13 06:50:03 crc kubenswrapper[4971]: I1213 06:50:03.465642 4971 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-13T06:50:03Z","lastTransitionTime":"2025-12-13T06:50:03Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 13 06:50:03 crc kubenswrapper[4971]: I1213 06:50:03.568542 4971 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 13 06:50:03 crc kubenswrapper[4971]: I1213 06:50:03.568575 4971 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 13 06:50:03 crc kubenswrapper[4971]: I1213 06:50:03.568582 4971 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 13 06:50:03 crc kubenswrapper[4971]: I1213 06:50:03.568595 4971 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 13 06:50:03 crc kubenswrapper[4971]: I1213 06:50:03.568604 4971 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-13T06:50:03Z","lastTransitionTime":"2025-12-13T06:50:03Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 13 06:50:03 crc kubenswrapper[4971]: I1213 06:50:03.671061 4971 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 13 06:50:03 crc kubenswrapper[4971]: I1213 06:50:03.671167 4971 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 13 06:50:03 crc kubenswrapper[4971]: I1213 06:50:03.671204 4971 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 13 06:50:03 crc kubenswrapper[4971]: I1213 06:50:03.671223 4971 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 13 06:50:03 crc kubenswrapper[4971]: I1213 06:50:03.671233 4971 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-13T06:50:03Z","lastTransitionTime":"2025-12-13T06:50:03Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 13 06:50:03 crc kubenswrapper[4971]: I1213 06:50:03.768004 4971 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 13 06:50:03 crc kubenswrapper[4971]: E1213 06:50:03.768483 4971 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 13 06:50:03 crc kubenswrapper[4971]: I1213 06:50:03.768072 4971 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 13 06:50:03 crc kubenswrapper[4971]: I1213 06:50:03.768000 4971 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-x98k7" Dec 13 06:50:03 crc kubenswrapper[4971]: E1213 06:50:03.768611 4971 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 13 06:50:03 crc kubenswrapper[4971]: I1213 06:50:03.768119 4971 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 13 06:50:03 crc kubenswrapper[4971]: E1213 06:50:03.768831 4971 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-x98k7" podUID="c7185dec-3240-42e8-9d90-6182a5e8b65c" Dec 13 06:50:03 crc kubenswrapper[4971]: E1213 06:50:03.768976 4971 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 13 06:50:03 crc kubenswrapper[4971]: I1213 06:50:03.774651 4971 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 13 06:50:03 crc kubenswrapper[4971]: I1213 06:50:03.774701 4971 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 13 06:50:03 crc kubenswrapper[4971]: I1213 06:50:03.774717 4971 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 13 06:50:03 crc kubenswrapper[4971]: I1213 06:50:03.774739 4971 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 13 06:50:03 crc kubenswrapper[4971]: I1213 06:50:03.774757 4971 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-13T06:50:03Z","lastTransitionTime":"2025-12-13T06:50:03Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 13 06:50:03 crc kubenswrapper[4971]: I1213 06:50:03.780749 4971 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"f9d31b5a-b770-4a15-95c8-35b49677c608\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T06:49:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T06:49:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T06:49:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T06:49:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T06:49:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c9baa042f9dada954faa57b5454cd6b0a12a297e56ddc192c52cff50e42a16fa\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-13T06:49:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://c788de86754f0bf37ee184672fc1dcb388ccebc246d4f15be4b5da661f725f8f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-13T06:49:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://760817511a9b4f96a68dc38d26631250eaab72e1dd8f123e653039b6b4e91f84\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-13T06:49:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b859be0c9feac856e9855c6a038b02b49b32fade97873c46d0bec0a32d8b5147\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-13T06:49:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-13T06:49:03Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-13T06:50:03Z is after 2025-08-24T17:21:41Z" Dec 13 06:50:03 crc kubenswrapper[4971]: I1213 06:50:03.792433 4971 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"8ede2992-9576-4b23-83c2-24ac3b1d3caf\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T06:49:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T06:49:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T06:49:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T06:49:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T06:49:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9e61d25add1723ac286794abfe886358e9d64455c25379f7e3057fc3dd812418\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-13T06:49:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://7fb51d8435b4e3ce0411cae1d6c2824c552f0a33d79de01f969274516b39a5b3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-13T06:49:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://1c35f5e8d8fab6d77b9e2e0dedb195440de5d6af345cd5865abd32caa7fafd60\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-13T06:49:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://38806c430185715371ab9bf30a1af9a9d5d492abd9bb84acbdccc694f56e3a8a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"wait-for-host-port\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://38806c430185715371ab9bf30a1af9a9d5d492abd9bb84acbdccc694f56e3a8a\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-13T06:49:04Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-13T06:49:04Z\\\"}}}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-13T06:49:03Z\\\"}}\" for pod \"openshift-kube-scheduler\"/\"openshift-kube-scheduler-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-13T06:50:03Z is after 2025-08-24T17:21:41Z" Dec 13 06:50:03 crc kubenswrapper[4971]: I1213 06:50:03.805994 4971 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-ncchz" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"de914d21-a688-401e-a66e-861beb6336d9\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T06:49:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T06:49:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T06:49:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T06:49:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://244ffe3dc9d8cf799aff4c4d795bf41554c1e6bb6667fbb0c7df322fb5f96c68\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-13T06:49:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lqmjj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://0b7111af6c5a198d02bb02858837cfb4f3b20828d8d206cd18aa0ffa102d0cd4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://0b7111af6c5a198d02bb02858837cfb4f3b20828d8d206cd18aa0ffa102d0cd4\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-13T06:49:26Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-13T06:49:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lqmjj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b36777fe40e74ae1cecb7b73e204139db5279ba51cef4f1e8c4a4c491d8be403\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://b36777fe40e74ae1cecb7b73e204139db5279ba51cef4f1e8c4a4c491d8be403\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-13T06:49:28Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-13T06:49:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lqmjj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://0597a06cc663e416142311a16941d98297577693a1fe27201ea42c27496cace2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://0597a06cc663e416142311a16941d98297577693a1fe27201ea42c27496cace2\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-13T06:49:29Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-13T06:49:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lqmjj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://3f52c2e22a75c010993ca5809a3a9d1304758d4cb6e83691ef1af15a4b514ac4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://3f52c2e22a75c010993ca5809a3a9d1304758d4cb6e83691ef1af15a4b514ac4\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-13T06:49:30Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-13T06:49:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lqmjj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://9e21ca708c4c4c1228b417cece51299e0792729b7c78884bd18b2dda8fe8355c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://9e21ca708c4c4c1228b417cece51299e0792729b7c78884bd18b2dda8fe8355c\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-13T06:49:31Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-13T06:49:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lqmjj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://74552159f12c87cdda09147692b400ccac0d4e43c1ed72fe8462c4249ad69c7b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://74552159f12c87cdda09147692b400ccac0d4e43c1ed72fe8462c4249ad69c7b\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-13T06:49:32Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-13T06:49:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lqmjj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-13T06:49:25Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-ncchz\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-13T06:50:03Z is after 2025-08-24T17:21:41Z" Dec 13 06:50:03 crc kubenswrapper[4971]: I1213 06:50:03.815855 4971 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-gkcd8" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"13f5c327-f95e-42b4-8783-218500c17694\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T06:49:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T06:49:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T06:49:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T06:49:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://77e846e36f1cb40d03471dc03de11b6c3da448baa407aad38093fae9a801cf92\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-13T06:49:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gvlhn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-13T06:49:27Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-gkcd8\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-13T06:50:03Z is after 2025-08-24T17:21:41Z" Dec 13 06:50:03 crc kubenswrapper[4971]: I1213 06:50:03.824894 4971 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-x98k7" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"c7185dec-3240-42e8-9d90-6182a5e8b65c\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T06:49:39Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T06:49:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T06:49:39Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T06:49:39Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bj2gn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bj2gn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-13T06:49:39Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-x98k7\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-13T06:50:03Z is after 2025-08-24T17:21:41Z" Dec 13 06:50:03 crc kubenswrapper[4971]: I1213 06:50:03.848864 4971 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"1ed9f82e-fe5c-4b76-a619-62e5abd88d3b\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T06:49:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T06:49:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T06:49:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T06:49:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T06:49:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e457329a78a2d24337f994e8a0a1f75a2a667244aa75eb84613d1e7138364aac\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-13T06:49:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://eb014de8140deef097d9bd94f3b9fa4ab3558bc4449bef84b83bf62bfabf5836\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-13T06:49:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://2c438cc74d555f7c9db4569d1c56b78b1088ed8105697d2c0f31e6c5136f3dd0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-13T06:49:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://aa6fa6e4340dcd066252df3c56d79a3a39ed531d5eea8d1d5941d6952368d3af\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-13T06:49:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://c4471b98854c52d5cd084d0ba3e64e5b08e9518ec3c372cce2c875a440522457\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-13T06:49:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7d2746335c98dae3a0f43024d34cb62ebcb42a01c8f9ce18d52b29d4c45598e5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://7d2746335c98dae3a0f43024d34cb62ebcb42a01c8f9ce18d52b29d4c45598e5\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-13T06:49:04Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-13T06:49:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://bc0e77d10a738128bea82122496bc42235fc1688f8907756d93e61b651902b57\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://bc0e77d10a738128bea82122496bc42235fc1688f8907756d93e61b651902b57\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-13T06:49:05Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-13T06:49:05Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://f821e418cba5e832da8839268bcbbd9630bdf319a212d5382bfa2c4b0b73a268\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://f821e418cba5e832da8839268bcbbd9630bdf319a212d5382bfa2c4b0b73a268\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-13T06:49:07Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-13T06:49:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-13T06:49:03Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-13T06:50:03Z is after 2025-08-24T17:21:41Z" Dec 13 06:50:03 crc kubenswrapper[4971]: I1213 06:50:03.862753 4971 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-13T06:49:24Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-13T06:49:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-13T06:49:24Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4328cd6a8464ed468229bbfbc8265f969427916d72ebb92237dbe68efe70fa47\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-13T06:49:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://9d715ca82d9b5a594e41f2b98da6f4277da0e4e1dec114969f901c00f6d7553d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-13T06:49:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-13T06:50:03Z is after 2025-08-24T17:21:41Z" Dec 13 06:50:03 crc kubenswrapper[4971]: I1213 06:50:03.873727 4971 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-5rzh5" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"6bf4591d-2483-4132-9672-d0fd6c995968\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T06:49:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T06:49:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T06:49:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T06:49:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://50e2cb14e4ffb603acb06e2281615a99f06182d345c4e70f2446fa2077033bbb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-13T06:49:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dfxqn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-13T06:49:25Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-5rzh5\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-13T06:50:03Z is after 2025-08-24T17:21:41Z" Dec 13 06:50:03 crc kubenswrapper[4971]: I1213 06:50:03.877246 4971 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 13 06:50:03 crc kubenswrapper[4971]: I1213 06:50:03.877345 4971 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 13 06:50:03 crc kubenswrapper[4971]: I1213 06:50:03.877364 4971 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 13 06:50:03 crc kubenswrapper[4971]: I1213 06:50:03.877388 4971 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 13 06:50:03 crc kubenswrapper[4971]: I1213 06:50:03.877404 4971 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-13T06:50:03Z","lastTransitionTime":"2025-12-13T06:50:03Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 13 06:50:03 crc kubenswrapper[4971]: I1213 06:50:03.884490 4971 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-z572s" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"b4146d3d-4034-49a9-bd37-164856255137\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T06:49:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T06:49:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T06:49:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T06:49:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://46fd4e1070835cfbf7f35d497eff7f047473fc4048ee07812925df5ebdf55342\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-13T06:49:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-j6l9b\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://dc79a0968f85723e6ad0dc269c8da1f7472adb1bb2b87edd017ee2295bfbf89c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-13T06:49:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-j6l9b\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-13T06:49:38Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-z572s\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-13T06:50:03Z is after 2025-08-24T17:21:41Z" Dec 13 06:50:03 crc kubenswrapper[4971]: I1213 06:50:03.896510 4971 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3ad3268a-a982-428c-9c06-4aaeddd372d9\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T06:49:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T06:49:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T06:49:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T06:49:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T06:49:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3716afdf2bc693be80a09fc9f7727d334f435caf1d4c8ec4811b3a1f1652cdfe\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-13T06:49:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://30df3abf8c79db8af52c6a918f73e1ce410ddafcf341636424c267e1eeb195a7\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-13T06:49:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://ccdf65f6d2b1eb3647a03ba37c305c32e34addfdf79ff937c8dddaa3e5d5b5c6\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-13T06:49:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://770b73dd0c2b6f38744e2c1cd6aa186a69bb87d1cd015a603ad2e33be06a8f6b\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://b3c1bab3d53f975b2e24582ce2771bc26d600e0dd7c30339dcfd18a071162ad4\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-12-13T06:49:23Z\\\",\\\"message\\\":\\\"file observer\\\\nW1213 06:49:23.283403 1 builder.go:272] unable to get owner reference (falling back to namespace): pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\nI1213 06:49:23.284336 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1213 06:49:23.286978 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-981773671/tls.crt::/tmp/serving-cert-981773671/tls.key\\\\\\\"\\\\nI1213 06:49:23.775072 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1213 06:49:23.778615 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1213 06:49:23.778645 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1213 06:49:23.778680 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1213 06:49:23.778691 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1213 06:49:23.785058 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1213 06:49:23.786150 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1213 06:49:23.786196 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1213 06:49:23.786223 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1213 06:49:23.786246 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1213 06:49:23.786269 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1213 06:49:23.786291 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI1213 06:49:23.785608 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF1213 06:49:23.787563 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-13T06:49:08Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-13T06:49:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://9acd07066deed483bd7fa61ee45fda09539d110717a134a56827f871186f92fc\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-13T06:49:07Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://38c91cfbaecfdbd3e3b663422a91ab4294610afb6bbf9ebdb77db24e2b76c682\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://38c91cfbaecfdbd3e3b663422a91ab4294610afb6bbf9ebdb77db24e2b76c682\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-13T06:49:05Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-13T06:49:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-13T06:49:03Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-13T06:50:03Z is after 2025-08-24T17:21:41Z" Dec 13 06:50:03 crc kubenswrapper[4971]: I1213 06:50:03.907065 4971 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-13T06:49:23Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-13T06:49:23Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-13T06:50:03Z is after 2025-08-24T17:21:41Z" Dec 13 06:50:03 crc kubenswrapper[4971]: I1213 06:50:03.917580 4971 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-13T06:49:23Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-13T06:49:23Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-13T06:50:03Z is after 2025-08-24T17:21:41Z" Dec 13 06:50:03 crc kubenswrapper[4971]: I1213 06:50:03.941131 4971 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-zlcg5" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"736c6b1a-93dd-4f71-81b6-1c43fcf5f556\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T06:49:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T06:49:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T06:49:25Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T06:49:25Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://325223543b3c7db417f769533559a1ca2b23a723494e1885b0fd0480b5a6399f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-13T06:49:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vzmsr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://f22a4ec756119c4f791fb5105cf1f31df3e548b2b07b1f909a62802e9b6035ae\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-13T06:49:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vzmsr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://f6073140d9f86929b6d4bd2960ede0617a5c9620b269bbab043bae37e0d389f5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-13T06:49:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vzmsr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://693a2ab434c250cb140a866e8068b156df9191414fed1d63bee919a2528ea6e5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-13T06:49:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vzmsr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://219620f51474e60b3e1f880c0972eb612dffc4bad24462bbba2c093ae817d852\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-13T06:49:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vzmsr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://f137f4b441c2402093ab430a9689a46221c4fb4bc70d99fb992f62b0a989c465\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-13T06:49:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vzmsr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://0b454a87c6dd11bd8c3bb3553ee71de32ed6f185984e5057a20f2237b35c45d9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://0b454a87c6dd11bd8c3bb3553ee71de32ed6f185984e5057a20f2237b35c45d9\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-12-13T06:49:56Z\\\",\\\"message\\\":\\\" for pod on switch crc\\\\nI1213 06:49:56.566617 6645 metrics.go:553] Stopping metrics server at address \\\\\\\"127.0.0.1:29103\\\\\\\"\\\\nI1213 06:49:56.566625 6645 base_network_controller_pods.go:477] [default/openshift-network-console/networking-console-plugin-85b44fc459-gdk6g] creating logical port openshift-network-console_networking-console-plugin-85b44fc459-gdk6g for pod on switch crc\\\\nI1213 06:49:56.566569 6645 loadbalancer.go:304] Deleted 0 stale LBs for map[string]string{\\\\\\\"k8s.ovn.org/kind\\\\\\\":\\\\\\\"Service\\\\\\\", \\\\\\\"k8s.ovn.org/owner\\\\\\\":\\\\\\\"openshift-console-operator/metrics\\\\\\\"}\\\\nI1213 06:49:56.566666 6645 services_controller.go:360] Finished syncing service metrics on namespace openshift-console-operator for network=default : 1.804735ms\\\\nI1213 06:49:56.566669 6645 loadbalancer.go:304] Deleted 0 stale LBs for map[string]string{\\\\\\\"k8s.ovn.org/kind\\\\\\\":\\\\\\\"Service\\\\\\\", \\\\\\\"k8s.ovn.org/owner\\\\\\\":\\\\\\\"openshift-apiserver/check-endpoints\\\\\\\"}\\\\nF1213 06:49:56.566698 6645 ovnkube.go:137] failed to run ovnkube: [failed to start network controller: failed to start default network controller: unable to create admin network policy controller, err: could not add Event Handler for anpInformer during admin network policy controller initialization, handler {0x1fcc6e0 0x1fcc3c0 0x1fcc360} was not added to shared informer because it has stopped already, failed to start node network controller: failed \\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-13T06:49:55Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":2,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 20s restarting failed container=ovnkube-controller pod=ovnkube-node-zlcg5_openshift-ovn-kubernetes(736c6b1a-93dd-4f71-81b6-1c43fcf5f556)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vzmsr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://868049c6a3a4ebe4036d1d22e02ba6b76bced2ee7e269bb228accc4a0636678a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-13T06:49:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vzmsr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://451b22df6973ddc8fa2209a65d64d9e41b33eb231bb4adfab88cf0e5943bb246\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://451b22df6973ddc8fa2209a65d64d9e41b33eb231bb4adfab88cf0e5943bb246\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-13T06:49:26Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-13T06:49:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vzmsr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-13T06:49:25Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-zlcg5\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-13T06:50:03Z is after 2025-08-24T17:21:41Z" Dec 13 06:50:03 crc kubenswrapper[4971]: I1213 06:50:03.956616 4971 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-13T06:49:24Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-13T06:49:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-13T06:49:24Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d0b2ecf0d5b22f23217f42f259ae7ffb6359555b22c30eaebf667eaf254b3bab\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-13T06:49:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-13T06:50:03Z is after 2025-08-24T17:21:41Z" Dec 13 06:50:03 crc kubenswrapper[4971]: I1213 06:50:03.969214 4971 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-13T06:49:23Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-13T06:49:23Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-13T06:50:03Z is after 2025-08-24T17:21:41Z" Dec 13 06:50:03 crc kubenswrapper[4971]: I1213 06:50:03.980374 4971 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 13 06:50:03 crc kubenswrapper[4971]: I1213 06:50:03.980422 4971 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 13 06:50:03 crc kubenswrapper[4971]: I1213 06:50:03.980432 4971 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 13 06:50:03 crc kubenswrapper[4971]: I1213 06:50:03.980456 4971 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 13 06:50:03 crc kubenswrapper[4971]: I1213 06:50:03.980472 4971 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-13T06:50:03Z","lastTransitionTime":"2025-12-13T06:50:03Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 13 06:50:03 crc kubenswrapper[4971]: I1213 06:50:03.981836 4971 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-13T06:49:26Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-13T06:49:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-13T06:49:26Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9c8a0d1db0e754462d805e82daf5f52ffd5457507a30456535ec59e096c9d092\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-13T06:49:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-13T06:50:03Z is after 2025-08-24T17:21:41Z" Dec 13 06:50:03 crc kubenswrapper[4971]: I1213 06:50:03.994377 4971 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-82xjz" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e0c4af12-edda-41d7-9b44-f87396174ef2\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T06:49:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T06:49:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T06:49:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T06:49:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://0f4577d9e37f860813e221f2e208177d8d383fd04f673ce357a0a405ea29d1e6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-13T06:49:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cnxsq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://7e0d2020e62f1a277cff7047e1907dc747ed393a1f6c983e1ee258bd156c16c7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-13T06:49:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cnxsq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-13T06:49:25Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-82xjz\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-13T06:50:03Z is after 2025-08-24T17:21:41Z" Dec 13 06:50:04 crc kubenswrapper[4971]: I1213 06:50:04.005815 4971 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-wwvpl" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"dddf744b-2122-49ad-a980-105c09636e0f\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T06:49:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T06:49:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T06:49:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T06:49:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4676d897c7fc668587faec563f3551a056557879cee3e865192d2bd28edf62fc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-13T06:49:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-czz9v\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-13T06:49:25Z\\\"}}\" for pod \"openshift-multus\"/\"multus-wwvpl\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-13T06:50:04Z is after 2025-08-24T17:21:41Z" Dec 13 06:50:04 crc kubenswrapper[4971]: I1213 06:50:04.082403 4971 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 13 06:50:04 crc kubenswrapper[4971]: I1213 06:50:04.082452 4971 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 13 06:50:04 crc kubenswrapper[4971]: I1213 06:50:04.082461 4971 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 13 06:50:04 crc kubenswrapper[4971]: I1213 06:50:04.082473 4971 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 13 06:50:04 crc kubenswrapper[4971]: I1213 06:50:04.082482 4971 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-13T06:50:04Z","lastTransitionTime":"2025-12-13T06:50:04Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 13 06:50:04 crc kubenswrapper[4971]: I1213 06:50:04.185657 4971 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 13 06:50:04 crc kubenswrapper[4971]: I1213 06:50:04.185812 4971 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 13 06:50:04 crc kubenswrapper[4971]: I1213 06:50:04.185900 4971 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 13 06:50:04 crc kubenswrapper[4971]: I1213 06:50:04.185979 4971 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 13 06:50:04 crc kubenswrapper[4971]: I1213 06:50:04.186013 4971 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-13T06:50:04Z","lastTransitionTime":"2025-12-13T06:50:04Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 13 06:50:04 crc kubenswrapper[4971]: I1213 06:50:04.289123 4971 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 13 06:50:04 crc kubenswrapper[4971]: I1213 06:50:04.289171 4971 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 13 06:50:04 crc kubenswrapper[4971]: I1213 06:50:04.289189 4971 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 13 06:50:04 crc kubenswrapper[4971]: I1213 06:50:04.289212 4971 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 13 06:50:04 crc kubenswrapper[4971]: I1213 06:50:04.289227 4971 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-13T06:50:04Z","lastTransitionTime":"2025-12-13T06:50:04Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 13 06:50:04 crc kubenswrapper[4971]: I1213 06:50:04.392055 4971 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 13 06:50:04 crc kubenswrapper[4971]: I1213 06:50:04.392101 4971 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 13 06:50:04 crc kubenswrapper[4971]: I1213 06:50:04.392112 4971 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 13 06:50:04 crc kubenswrapper[4971]: I1213 06:50:04.392129 4971 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 13 06:50:04 crc kubenswrapper[4971]: I1213 06:50:04.392140 4971 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-13T06:50:04Z","lastTransitionTime":"2025-12-13T06:50:04Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 13 06:50:04 crc kubenswrapper[4971]: I1213 06:50:04.495156 4971 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 13 06:50:04 crc kubenswrapper[4971]: I1213 06:50:04.495487 4971 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 13 06:50:04 crc kubenswrapper[4971]: I1213 06:50:04.495672 4971 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 13 06:50:04 crc kubenswrapper[4971]: I1213 06:50:04.495817 4971 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 13 06:50:04 crc kubenswrapper[4971]: I1213 06:50:04.495934 4971 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-13T06:50:04Z","lastTransitionTime":"2025-12-13T06:50:04Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 13 06:50:04 crc kubenswrapper[4971]: I1213 06:50:04.599357 4971 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 13 06:50:04 crc kubenswrapper[4971]: I1213 06:50:04.599392 4971 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 13 06:50:04 crc kubenswrapper[4971]: I1213 06:50:04.599401 4971 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 13 06:50:04 crc kubenswrapper[4971]: I1213 06:50:04.599414 4971 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 13 06:50:04 crc kubenswrapper[4971]: I1213 06:50:04.599424 4971 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-13T06:50:04Z","lastTransitionTime":"2025-12-13T06:50:04Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 13 06:50:04 crc kubenswrapper[4971]: I1213 06:50:04.703606 4971 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 13 06:50:04 crc kubenswrapper[4971]: I1213 06:50:04.704048 4971 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 13 06:50:04 crc kubenswrapper[4971]: I1213 06:50:04.704247 4971 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 13 06:50:04 crc kubenswrapper[4971]: I1213 06:50:04.704422 4971 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 13 06:50:04 crc kubenswrapper[4971]: I1213 06:50:04.704630 4971 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-13T06:50:04Z","lastTransitionTime":"2025-12-13T06:50:04Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 13 06:50:04 crc kubenswrapper[4971]: I1213 06:50:04.806832 4971 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 13 06:50:04 crc kubenswrapper[4971]: I1213 06:50:04.807147 4971 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 13 06:50:04 crc kubenswrapper[4971]: I1213 06:50:04.807283 4971 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 13 06:50:04 crc kubenswrapper[4971]: I1213 06:50:04.807390 4971 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 13 06:50:04 crc kubenswrapper[4971]: I1213 06:50:04.807487 4971 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-13T06:50:04Z","lastTransitionTime":"2025-12-13T06:50:04Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 13 06:50:04 crc kubenswrapper[4971]: I1213 06:50:04.909398 4971 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 13 06:50:04 crc kubenswrapper[4971]: I1213 06:50:04.909429 4971 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 13 06:50:04 crc kubenswrapper[4971]: I1213 06:50:04.909438 4971 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 13 06:50:04 crc kubenswrapper[4971]: I1213 06:50:04.909454 4971 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 13 06:50:04 crc kubenswrapper[4971]: I1213 06:50:04.909463 4971 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-13T06:50:04Z","lastTransitionTime":"2025-12-13T06:50:04Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 13 06:50:05 crc kubenswrapper[4971]: I1213 06:50:05.011630 4971 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 13 06:50:05 crc kubenswrapper[4971]: I1213 06:50:05.012129 4971 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 13 06:50:05 crc kubenswrapper[4971]: I1213 06:50:05.012197 4971 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 13 06:50:05 crc kubenswrapper[4971]: I1213 06:50:05.012270 4971 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 13 06:50:05 crc kubenswrapper[4971]: I1213 06:50:05.012332 4971 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-13T06:50:05Z","lastTransitionTime":"2025-12-13T06:50:05Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 13 06:50:05 crc kubenswrapper[4971]: I1213 06:50:05.114855 4971 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 13 06:50:05 crc kubenswrapper[4971]: I1213 06:50:05.114893 4971 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 13 06:50:05 crc kubenswrapper[4971]: I1213 06:50:05.114904 4971 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 13 06:50:05 crc kubenswrapper[4971]: I1213 06:50:05.114921 4971 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 13 06:50:05 crc kubenswrapper[4971]: I1213 06:50:05.114933 4971 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-13T06:50:05Z","lastTransitionTime":"2025-12-13T06:50:05Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 13 06:50:05 crc kubenswrapper[4971]: I1213 06:50:05.216954 4971 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 13 06:50:05 crc kubenswrapper[4971]: I1213 06:50:05.217007 4971 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 13 06:50:05 crc kubenswrapper[4971]: I1213 06:50:05.217018 4971 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 13 06:50:05 crc kubenswrapper[4971]: I1213 06:50:05.217035 4971 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 13 06:50:05 crc kubenswrapper[4971]: I1213 06:50:05.217046 4971 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-13T06:50:05Z","lastTransitionTime":"2025-12-13T06:50:05Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 13 06:50:05 crc kubenswrapper[4971]: I1213 06:50:05.319551 4971 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 13 06:50:05 crc kubenswrapper[4971]: I1213 06:50:05.319585 4971 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 13 06:50:05 crc kubenswrapper[4971]: I1213 06:50:05.319594 4971 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 13 06:50:05 crc kubenswrapper[4971]: I1213 06:50:05.319609 4971 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 13 06:50:05 crc kubenswrapper[4971]: I1213 06:50:05.319619 4971 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-13T06:50:05Z","lastTransitionTime":"2025-12-13T06:50:05Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 13 06:50:05 crc kubenswrapper[4971]: I1213 06:50:05.422333 4971 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 13 06:50:05 crc kubenswrapper[4971]: I1213 06:50:05.422392 4971 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 13 06:50:05 crc kubenswrapper[4971]: I1213 06:50:05.422413 4971 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 13 06:50:05 crc kubenswrapper[4971]: I1213 06:50:05.422442 4971 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 13 06:50:05 crc kubenswrapper[4971]: I1213 06:50:05.422464 4971 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-13T06:50:05Z","lastTransitionTime":"2025-12-13T06:50:05Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 13 06:50:05 crc kubenswrapper[4971]: I1213 06:50:05.525850 4971 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 13 06:50:05 crc kubenswrapper[4971]: I1213 06:50:05.525901 4971 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 13 06:50:05 crc kubenswrapper[4971]: I1213 06:50:05.525929 4971 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 13 06:50:05 crc kubenswrapper[4971]: I1213 06:50:05.525952 4971 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 13 06:50:05 crc kubenswrapper[4971]: I1213 06:50:05.525971 4971 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-13T06:50:05Z","lastTransitionTime":"2025-12-13T06:50:05Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 13 06:50:05 crc kubenswrapper[4971]: I1213 06:50:05.628970 4971 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 13 06:50:05 crc kubenswrapper[4971]: I1213 06:50:05.629031 4971 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 13 06:50:05 crc kubenswrapper[4971]: I1213 06:50:05.629040 4971 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 13 06:50:05 crc kubenswrapper[4971]: I1213 06:50:05.629054 4971 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 13 06:50:05 crc kubenswrapper[4971]: I1213 06:50:05.629067 4971 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-13T06:50:05Z","lastTransitionTime":"2025-12-13T06:50:05Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 13 06:50:05 crc kubenswrapper[4971]: I1213 06:50:05.732203 4971 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 13 06:50:05 crc kubenswrapper[4971]: I1213 06:50:05.732286 4971 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 13 06:50:05 crc kubenswrapper[4971]: I1213 06:50:05.732303 4971 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 13 06:50:05 crc kubenswrapper[4971]: I1213 06:50:05.732325 4971 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 13 06:50:05 crc kubenswrapper[4971]: I1213 06:50:05.732340 4971 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-13T06:50:05Z","lastTransitionTime":"2025-12-13T06:50:05Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 13 06:50:05 crc kubenswrapper[4971]: I1213 06:50:05.768040 4971 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 13 06:50:05 crc kubenswrapper[4971]: I1213 06:50:05.768084 4971 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 13 06:50:05 crc kubenswrapper[4971]: I1213 06:50:05.768141 4971 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-x98k7" Dec 13 06:50:05 crc kubenswrapper[4971]: E1213 06:50:05.768247 4971 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 13 06:50:05 crc kubenswrapper[4971]: E1213 06:50:05.768393 4971 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 13 06:50:05 crc kubenswrapper[4971]: E1213 06:50:05.768558 4971 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-x98k7" podUID="c7185dec-3240-42e8-9d90-6182a5e8b65c" Dec 13 06:50:05 crc kubenswrapper[4971]: I1213 06:50:05.768630 4971 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 13 06:50:05 crc kubenswrapper[4971]: E1213 06:50:05.768768 4971 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 13 06:50:05 crc kubenswrapper[4971]: I1213 06:50:05.834498 4971 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 13 06:50:05 crc kubenswrapper[4971]: I1213 06:50:05.834551 4971 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 13 06:50:05 crc kubenswrapper[4971]: I1213 06:50:05.834563 4971 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 13 06:50:05 crc kubenswrapper[4971]: I1213 06:50:05.834578 4971 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 13 06:50:05 crc kubenswrapper[4971]: I1213 06:50:05.834588 4971 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-13T06:50:05Z","lastTransitionTime":"2025-12-13T06:50:05Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 13 06:50:05 crc kubenswrapper[4971]: I1213 06:50:05.937917 4971 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 13 06:50:05 crc kubenswrapper[4971]: I1213 06:50:05.938044 4971 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 13 06:50:05 crc kubenswrapper[4971]: I1213 06:50:05.938120 4971 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 13 06:50:05 crc kubenswrapper[4971]: I1213 06:50:05.938150 4971 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 13 06:50:05 crc kubenswrapper[4971]: I1213 06:50:05.938212 4971 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-13T06:50:05Z","lastTransitionTime":"2025-12-13T06:50:05Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 13 06:50:06 crc kubenswrapper[4971]: I1213 06:50:06.040244 4971 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 13 06:50:06 crc kubenswrapper[4971]: I1213 06:50:06.040279 4971 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 13 06:50:06 crc kubenswrapper[4971]: I1213 06:50:06.040290 4971 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 13 06:50:06 crc kubenswrapper[4971]: I1213 06:50:06.040303 4971 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 13 06:50:06 crc kubenswrapper[4971]: I1213 06:50:06.040311 4971 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-13T06:50:06Z","lastTransitionTime":"2025-12-13T06:50:06Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 13 06:50:06 crc kubenswrapper[4971]: I1213 06:50:06.142301 4971 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 13 06:50:06 crc kubenswrapper[4971]: I1213 06:50:06.142342 4971 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 13 06:50:06 crc kubenswrapper[4971]: I1213 06:50:06.142353 4971 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 13 06:50:06 crc kubenswrapper[4971]: I1213 06:50:06.142367 4971 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 13 06:50:06 crc kubenswrapper[4971]: I1213 06:50:06.142377 4971 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-13T06:50:06Z","lastTransitionTime":"2025-12-13T06:50:06Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 13 06:50:06 crc kubenswrapper[4971]: I1213 06:50:06.244551 4971 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 13 06:50:06 crc kubenswrapper[4971]: I1213 06:50:06.244585 4971 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 13 06:50:06 crc kubenswrapper[4971]: I1213 06:50:06.244593 4971 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 13 06:50:06 crc kubenswrapper[4971]: I1213 06:50:06.244606 4971 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 13 06:50:06 crc kubenswrapper[4971]: I1213 06:50:06.244615 4971 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-13T06:50:06Z","lastTransitionTime":"2025-12-13T06:50:06Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 13 06:50:06 crc kubenswrapper[4971]: I1213 06:50:06.347311 4971 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 13 06:50:06 crc kubenswrapper[4971]: I1213 06:50:06.347344 4971 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 13 06:50:06 crc kubenswrapper[4971]: I1213 06:50:06.347353 4971 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 13 06:50:06 crc kubenswrapper[4971]: I1213 06:50:06.347366 4971 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 13 06:50:06 crc kubenswrapper[4971]: I1213 06:50:06.347375 4971 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-13T06:50:06Z","lastTransitionTime":"2025-12-13T06:50:06Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 13 06:50:06 crc kubenswrapper[4971]: I1213 06:50:06.449248 4971 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 13 06:50:06 crc kubenswrapper[4971]: I1213 06:50:06.449292 4971 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 13 06:50:06 crc kubenswrapper[4971]: I1213 06:50:06.449301 4971 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 13 06:50:06 crc kubenswrapper[4971]: I1213 06:50:06.449315 4971 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 13 06:50:06 crc kubenswrapper[4971]: I1213 06:50:06.449324 4971 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-13T06:50:06Z","lastTransitionTime":"2025-12-13T06:50:06Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 13 06:50:06 crc kubenswrapper[4971]: I1213 06:50:06.551632 4971 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 13 06:50:06 crc kubenswrapper[4971]: I1213 06:50:06.551678 4971 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 13 06:50:06 crc kubenswrapper[4971]: I1213 06:50:06.551687 4971 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 13 06:50:06 crc kubenswrapper[4971]: I1213 06:50:06.551702 4971 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 13 06:50:06 crc kubenswrapper[4971]: I1213 06:50:06.551712 4971 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-13T06:50:06Z","lastTransitionTime":"2025-12-13T06:50:06Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 13 06:50:06 crc kubenswrapper[4971]: I1213 06:50:06.653825 4971 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 13 06:50:06 crc kubenswrapper[4971]: I1213 06:50:06.653892 4971 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 13 06:50:06 crc kubenswrapper[4971]: I1213 06:50:06.653904 4971 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 13 06:50:06 crc kubenswrapper[4971]: I1213 06:50:06.653922 4971 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 13 06:50:06 crc kubenswrapper[4971]: I1213 06:50:06.653956 4971 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-13T06:50:06Z","lastTransitionTime":"2025-12-13T06:50:06Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 13 06:50:06 crc kubenswrapper[4971]: I1213 06:50:06.756632 4971 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 13 06:50:06 crc kubenswrapper[4971]: I1213 06:50:06.756681 4971 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 13 06:50:06 crc kubenswrapper[4971]: I1213 06:50:06.756696 4971 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 13 06:50:06 crc kubenswrapper[4971]: I1213 06:50:06.756713 4971 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 13 06:50:06 crc kubenswrapper[4971]: I1213 06:50:06.756724 4971 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-13T06:50:06Z","lastTransitionTime":"2025-12-13T06:50:06Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 13 06:50:06 crc kubenswrapper[4971]: I1213 06:50:06.859234 4971 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 13 06:50:06 crc kubenswrapper[4971]: I1213 06:50:06.859289 4971 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 13 06:50:06 crc kubenswrapper[4971]: I1213 06:50:06.859306 4971 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 13 06:50:06 crc kubenswrapper[4971]: I1213 06:50:06.859331 4971 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 13 06:50:06 crc kubenswrapper[4971]: I1213 06:50:06.859347 4971 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-13T06:50:06Z","lastTransitionTime":"2025-12-13T06:50:06Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 13 06:50:06 crc kubenswrapper[4971]: I1213 06:50:06.962051 4971 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 13 06:50:06 crc kubenswrapper[4971]: I1213 06:50:06.962088 4971 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 13 06:50:06 crc kubenswrapper[4971]: I1213 06:50:06.962097 4971 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 13 06:50:06 crc kubenswrapper[4971]: I1213 06:50:06.962113 4971 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 13 06:50:06 crc kubenswrapper[4971]: I1213 06:50:06.962126 4971 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-13T06:50:06Z","lastTransitionTime":"2025-12-13T06:50:06Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 13 06:50:07 crc kubenswrapper[4971]: I1213 06:50:07.064324 4971 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 13 06:50:07 crc kubenswrapper[4971]: I1213 06:50:07.064364 4971 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 13 06:50:07 crc kubenswrapper[4971]: I1213 06:50:07.064373 4971 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 13 06:50:07 crc kubenswrapper[4971]: I1213 06:50:07.064388 4971 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 13 06:50:07 crc kubenswrapper[4971]: I1213 06:50:07.064398 4971 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-13T06:50:07Z","lastTransitionTime":"2025-12-13T06:50:07Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 13 06:50:07 crc kubenswrapper[4971]: I1213 06:50:07.166828 4971 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 13 06:50:07 crc kubenswrapper[4971]: I1213 06:50:07.166872 4971 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 13 06:50:07 crc kubenswrapper[4971]: I1213 06:50:07.166884 4971 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 13 06:50:07 crc kubenswrapper[4971]: I1213 06:50:07.166902 4971 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 13 06:50:07 crc kubenswrapper[4971]: I1213 06:50:07.166915 4971 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-13T06:50:07Z","lastTransitionTime":"2025-12-13T06:50:07Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 13 06:50:07 crc kubenswrapper[4971]: I1213 06:50:07.269507 4971 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 13 06:50:07 crc kubenswrapper[4971]: I1213 06:50:07.269579 4971 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 13 06:50:07 crc kubenswrapper[4971]: I1213 06:50:07.269591 4971 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 13 06:50:07 crc kubenswrapper[4971]: I1213 06:50:07.269610 4971 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 13 06:50:07 crc kubenswrapper[4971]: I1213 06:50:07.269624 4971 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-13T06:50:07Z","lastTransitionTime":"2025-12-13T06:50:07Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 13 06:50:07 crc kubenswrapper[4971]: I1213 06:50:07.372027 4971 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 13 06:50:07 crc kubenswrapper[4971]: I1213 06:50:07.372060 4971 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 13 06:50:07 crc kubenswrapper[4971]: I1213 06:50:07.372072 4971 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 13 06:50:07 crc kubenswrapper[4971]: I1213 06:50:07.372089 4971 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 13 06:50:07 crc kubenswrapper[4971]: I1213 06:50:07.372100 4971 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-13T06:50:07Z","lastTransitionTime":"2025-12-13T06:50:07Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 13 06:50:07 crc kubenswrapper[4971]: I1213 06:50:07.474365 4971 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 13 06:50:07 crc kubenswrapper[4971]: I1213 06:50:07.474427 4971 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 13 06:50:07 crc kubenswrapper[4971]: I1213 06:50:07.474439 4971 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 13 06:50:07 crc kubenswrapper[4971]: I1213 06:50:07.474454 4971 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 13 06:50:07 crc kubenswrapper[4971]: I1213 06:50:07.474466 4971 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-13T06:50:07Z","lastTransitionTime":"2025-12-13T06:50:07Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 13 06:50:07 crc kubenswrapper[4971]: I1213 06:50:07.576640 4971 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 13 06:50:07 crc kubenswrapper[4971]: I1213 06:50:07.576671 4971 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 13 06:50:07 crc kubenswrapper[4971]: I1213 06:50:07.576679 4971 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 13 06:50:07 crc kubenswrapper[4971]: I1213 06:50:07.576692 4971 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 13 06:50:07 crc kubenswrapper[4971]: I1213 06:50:07.576700 4971 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-13T06:50:07Z","lastTransitionTime":"2025-12-13T06:50:07Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 13 06:50:07 crc kubenswrapper[4971]: I1213 06:50:07.679131 4971 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 13 06:50:07 crc kubenswrapper[4971]: I1213 06:50:07.679199 4971 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 13 06:50:07 crc kubenswrapper[4971]: I1213 06:50:07.679216 4971 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 13 06:50:07 crc kubenswrapper[4971]: I1213 06:50:07.679241 4971 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 13 06:50:07 crc kubenswrapper[4971]: I1213 06:50:07.679264 4971 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-13T06:50:07Z","lastTransitionTime":"2025-12-13T06:50:07Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 13 06:50:07 crc kubenswrapper[4971]: I1213 06:50:07.768383 4971 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-x98k7" Dec 13 06:50:07 crc kubenswrapper[4971]: I1213 06:50:07.768449 4971 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 13 06:50:07 crc kubenswrapper[4971]: I1213 06:50:07.768449 4971 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 13 06:50:07 crc kubenswrapper[4971]: I1213 06:50:07.768682 4971 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 13 06:50:07 crc kubenswrapper[4971]: E1213 06:50:07.768681 4971 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-x98k7" podUID="c7185dec-3240-42e8-9d90-6182a5e8b65c" Dec 13 06:50:07 crc kubenswrapper[4971]: E1213 06:50:07.768767 4971 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 13 06:50:07 crc kubenswrapper[4971]: E1213 06:50:07.768870 4971 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 13 06:50:07 crc kubenswrapper[4971]: E1213 06:50:07.768989 4971 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 13 06:50:07 crc kubenswrapper[4971]: I1213 06:50:07.782199 4971 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 13 06:50:07 crc kubenswrapper[4971]: I1213 06:50:07.782250 4971 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 13 06:50:07 crc kubenswrapper[4971]: I1213 06:50:07.782262 4971 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 13 06:50:07 crc kubenswrapper[4971]: I1213 06:50:07.782277 4971 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 13 06:50:07 crc kubenswrapper[4971]: I1213 06:50:07.782290 4971 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-13T06:50:07Z","lastTransitionTime":"2025-12-13T06:50:07Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 13 06:50:07 crc kubenswrapper[4971]: I1213 06:50:07.888403 4971 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 13 06:50:07 crc kubenswrapper[4971]: I1213 06:50:07.888443 4971 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 13 06:50:07 crc kubenswrapper[4971]: I1213 06:50:07.888463 4971 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 13 06:50:07 crc kubenswrapper[4971]: I1213 06:50:07.888572 4971 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 13 06:50:07 crc kubenswrapper[4971]: I1213 06:50:07.888585 4971 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-13T06:50:07Z","lastTransitionTime":"2025-12-13T06:50:07Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 13 06:50:07 crc kubenswrapper[4971]: I1213 06:50:07.990764 4971 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 13 06:50:07 crc kubenswrapper[4971]: I1213 06:50:07.990809 4971 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 13 06:50:07 crc kubenswrapper[4971]: I1213 06:50:07.990821 4971 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 13 06:50:07 crc kubenswrapper[4971]: I1213 06:50:07.990837 4971 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 13 06:50:07 crc kubenswrapper[4971]: I1213 06:50:07.990848 4971 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-13T06:50:07Z","lastTransitionTime":"2025-12-13T06:50:07Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 13 06:50:08 crc kubenswrapper[4971]: I1213 06:50:08.093238 4971 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 13 06:50:08 crc kubenswrapper[4971]: I1213 06:50:08.093263 4971 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 13 06:50:08 crc kubenswrapper[4971]: I1213 06:50:08.093271 4971 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 13 06:50:08 crc kubenswrapper[4971]: I1213 06:50:08.093284 4971 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 13 06:50:08 crc kubenswrapper[4971]: I1213 06:50:08.093293 4971 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-13T06:50:08Z","lastTransitionTime":"2025-12-13T06:50:08Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 13 06:50:08 crc kubenswrapper[4971]: I1213 06:50:08.195605 4971 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 13 06:50:08 crc kubenswrapper[4971]: I1213 06:50:08.195712 4971 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 13 06:50:08 crc kubenswrapper[4971]: I1213 06:50:08.195724 4971 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 13 06:50:08 crc kubenswrapper[4971]: I1213 06:50:08.195737 4971 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 13 06:50:08 crc kubenswrapper[4971]: I1213 06:50:08.195746 4971 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-13T06:50:08Z","lastTransitionTime":"2025-12-13T06:50:08Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 13 06:50:08 crc kubenswrapper[4971]: I1213 06:50:08.298059 4971 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 13 06:50:08 crc kubenswrapper[4971]: I1213 06:50:08.298097 4971 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 13 06:50:08 crc kubenswrapper[4971]: I1213 06:50:08.298108 4971 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 13 06:50:08 crc kubenswrapper[4971]: I1213 06:50:08.298158 4971 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 13 06:50:08 crc kubenswrapper[4971]: I1213 06:50:08.298172 4971 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-13T06:50:08Z","lastTransitionTime":"2025-12-13T06:50:08Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 13 06:50:08 crc kubenswrapper[4971]: I1213 06:50:08.400224 4971 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 13 06:50:08 crc kubenswrapper[4971]: I1213 06:50:08.400337 4971 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 13 06:50:08 crc kubenswrapper[4971]: I1213 06:50:08.400361 4971 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 13 06:50:08 crc kubenswrapper[4971]: I1213 06:50:08.400391 4971 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 13 06:50:08 crc kubenswrapper[4971]: I1213 06:50:08.400416 4971 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-13T06:50:08Z","lastTransitionTime":"2025-12-13T06:50:08Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 13 06:50:08 crc kubenswrapper[4971]: I1213 06:50:08.502038 4971 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 13 06:50:08 crc kubenswrapper[4971]: I1213 06:50:08.502076 4971 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 13 06:50:08 crc kubenswrapper[4971]: I1213 06:50:08.502088 4971 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 13 06:50:08 crc kubenswrapper[4971]: I1213 06:50:08.502104 4971 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 13 06:50:08 crc kubenswrapper[4971]: I1213 06:50:08.502114 4971 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-13T06:50:08Z","lastTransitionTime":"2025-12-13T06:50:08Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 13 06:50:08 crc kubenswrapper[4971]: I1213 06:50:08.605391 4971 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 13 06:50:08 crc kubenswrapper[4971]: I1213 06:50:08.605447 4971 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 13 06:50:08 crc kubenswrapper[4971]: I1213 06:50:08.605462 4971 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 13 06:50:08 crc kubenswrapper[4971]: I1213 06:50:08.605483 4971 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 13 06:50:08 crc kubenswrapper[4971]: I1213 06:50:08.605495 4971 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-13T06:50:08Z","lastTransitionTime":"2025-12-13T06:50:08Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 13 06:50:08 crc kubenswrapper[4971]: I1213 06:50:08.707282 4971 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 13 06:50:08 crc kubenswrapper[4971]: I1213 06:50:08.707329 4971 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 13 06:50:08 crc kubenswrapper[4971]: I1213 06:50:08.707341 4971 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 13 06:50:08 crc kubenswrapper[4971]: I1213 06:50:08.707359 4971 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 13 06:50:08 crc kubenswrapper[4971]: I1213 06:50:08.707371 4971 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-13T06:50:08Z","lastTransitionTime":"2025-12-13T06:50:08Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 13 06:50:08 crc kubenswrapper[4971]: I1213 06:50:08.768494 4971 scope.go:117] "RemoveContainer" containerID="0b454a87c6dd11bd8c3bb3553ee71de32ed6f185984e5057a20f2237b35c45d9" Dec 13 06:50:08 crc kubenswrapper[4971]: E1213 06:50:08.768685 4971 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"ovnkube-controller\" with CrashLoopBackOff: \"back-off 20s restarting failed container=ovnkube-controller pod=ovnkube-node-zlcg5_openshift-ovn-kubernetes(736c6b1a-93dd-4f71-81b6-1c43fcf5f556)\"" pod="openshift-ovn-kubernetes/ovnkube-node-zlcg5" podUID="736c6b1a-93dd-4f71-81b6-1c43fcf5f556" Dec 13 06:50:08 crc kubenswrapper[4971]: I1213 06:50:08.809818 4971 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 13 06:50:08 crc kubenswrapper[4971]: I1213 06:50:08.809847 4971 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 13 06:50:08 crc kubenswrapper[4971]: I1213 06:50:08.809856 4971 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 13 06:50:08 crc kubenswrapper[4971]: I1213 06:50:08.809869 4971 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 13 06:50:08 crc kubenswrapper[4971]: I1213 06:50:08.809877 4971 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-13T06:50:08Z","lastTransitionTime":"2025-12-13T06:50:08Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 13 06:50:08 crc kubenswrapper[4971]: I1213 06:50:08.912607 4971 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 13 06:50:08 crc kubenswrapper[4971]: I1213 06:50:08.912654 4971 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 13 06:50:08 crc kubenswrapper[4971]: I1213 06:50:08.912664 4971 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 13 06:50:08 crc kubenswrapper[4971]: I1213 06:50:08.912680 4971 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 13 06:50:08 crc kubenswrapper[4971]: I1213 06:50:08.912690 4971 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-13T06:50:08Z","lastTransitionTime":"2025-12-13T06:50:08Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 13 06:50:09 crc kubenswrapper[4971]: I1213 06:50:09.015276 4971 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 13 06:50:09 crc kubenswrapper[4971]: I1213 06:50:09.015325 4971 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 13 06:50:09 crc kubenswrapper[4971]: I1213 06:50:09.015337 4971 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 13 06:50:09 crc kubenswrapper[4971]: I1213 06:50:09.015353 4971 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 13 06:50:09 crc kubenswrapper[4971]: I1213 06:50:09.015363 4971 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-13T06:50:09Z","lastTransitionTime":"2025-12-13T06:50:09Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 13 06:50:09 crc kubenswrapper[4971]: I1213 06:50:09.118104 4971 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 13 06:50:09 crc kubenswrapper[4971]: I1213 06:50:09.118153 4971 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 13 06:50:09 crc kubenswrapper[4971]: I1213 06:50:09.118166 4971 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 13 06:50:09 crc kubenswrapper[4971]: I1213 06:50:09.118185 4971 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 13 06:50:09 crc kubenswrapper[4971]: I1213 06:50:09.118201 4971 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-13T06:50:09Z","lastTransitionTime":"2025-12-13T06:50:09Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 13 06:50:09 crc kubenswrapper[4971]: I1213 06:50:09.221137 4971 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 13 06:50:09 crc kubenswrapper[4971]: I1213 06:50:09.221182 4971 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 13 06:50:09 crc kubenswrapper[4971]: I1213 06:50:09.221194 4971 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 13 06:50:09 crc kubenswrapper[4971]: I1213 06:50:09.221211 4971 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 13 06:50:09 crc kubenswrapper[4971]: I1213 06:50:09.221223 4971 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-13T06:50:09Z","lastTransitionTime":"2025-12-13T06:50:09Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 13 06:50:09 crc kubenswrapper[4971]: I1213 06:50:09.323970 4971 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 13 06:50:09 crc kubenswrapper[4971]: I1213 06:50:09.324030 4971 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 13 06:50:09 crc kubenswrapper[4971]: I1213 06:50:09.324047 4971 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 13 06:50:09 crc kubenswrapper[4971]: I1213 06:50:09.324073 4971 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 13 06:50:09 crc kubenswrapper[4971]: I1213 06:50:09.324091 4971 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-13T06:50:09Z","lastTransitionTime":"2025-12-13T06:50:09Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 13 06:50:09 crc kubenswrapper[4971]: I1213 06:50:09.426679 4971 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 13 06:50:09 crc kubenswrapper[4971]: I1213 06:50:09.426719 4971 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 13 06:50:09 crc kubenswrapper[4971]: I1213 06:50:09.426728 4971 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 13 06:50:09 crc kubenswrapper[4971]: I1213 06:50:09.426744 4971 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 13 06:50:09 crc kubenswrapper[4971]: I1213 06:50:09.426754 4971 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-13T06:50:09Z","lastTransitionTime":"2025-12-13T06:50:09Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 13 06:50:09 crc kubenswrapper[4971]: I1213 06:50:09.528735 4971 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 13 06:50:09 crc kubenswrapper[4971]: I1213 06:50:09.528787 4971 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 13 06:50:09 crc kubenswrapper[4971]: I1213 06:50:09.528802 4971 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 13 06:50:09 crc kubenswrapper[4971]: I1213 06:50:09.528823 4971 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 13 06:50:09 crc kubenswrapper[4971]: I1213 06:50:09.528841 4971 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-13T06:50:09Z","lastTransitionTime":"2025-12-13T06:50:09Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 13 06:50:09 crc kubenswrapper[4971]: I1213 06:50:09.631583 4971 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 13 06:50:09 crc kubenswrapper[4971]: I1213 06:50:09.631627 4971 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 13 06:50:09 crc kubenswrapper[4971]: I1213 06:50:09.631638 4971 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 13 06:50:09 crc kubenswrapper[4971]: I1213 06:50:09.631656 4971 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 13 06:50:09 crc kubenswrapper[4971]: I1213 06:50:09.631669 4971 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-13T06:50:09Z","lastTransitionTime":"2025-12-13T06:50:09Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 13 06:50:09 crc kubenswrapper[4971]: I1213 06:50:09.734320 4971 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 13 06:50:09 crc kubenswrapper[4971]: I1213 06:50:09.734372 4971 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 13 06:50:09 crc kubenswrapper[4971]: I1213 06:50:09.734384 4971 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 13 06:50:09 crc kubenswrapper[4971]: I1213 06:50:09.734439 4971 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 13 06:50:09 crc kubenswrapper[4971]: I1213 06:50:09.734455 4971 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-13T06:50:09Z","lastTransitionTime":"2025-12-13T06:50:09Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 13 06:50:09 crc kubenswrapper[4971]: I1213 06:50:09.767815 4971 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-x98k7" Dec 13 06:50:09 crc kubenswrapper[4971]: I1213 06:50:09.767865 4971 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 13 06:50:09 crc kubenswrapper[4971]: I1213 06:50:09.767912 4971 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 13 06:50:09 crc kubenswrapper[4971]: I1213 06:50:09.767815 4971 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 13 06:50:09 crc kubenswrapper[4971]: E1213 06:50:09.767941 4971 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-x98k7" podUID="c7185dec-3240-42e8-9d90-6182a5e8b65c" Dec 13 06:50:09 crc kubenswrapper[4971]: E1213 06:50:09.768027 4971 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 13 06:50:09 crc kubenswrapper[4971]: E1213 06:50:09.768051 4971 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 13 06:50:09 crc kubenswrapper[4971]: E1213 06:50:09.768119 4971 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 13 06:50:09 crc kubenswrapper[4971]: I1213 06:50:09.837750 4971 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 13 06:50:09 crc kubenswrapper[4971]: I1213 06:50:09.837808 4971 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 13 06:50:09 crc kubenswrapper[4971]: I1213 06:50:09.837821 4971 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 13 06:50:09 crc kubenswrapper[4971]: I1213 06:50:09.837844 4971 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 13 06:50:09 crc kubenswrapper[4971]: I1213 06:50:09.837859 4971 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-13T06:50:09Z","lastTransitionTime":"2025-12-13T06:50:09Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 13 06:50:09 crc kubenswrapper[4971]: I1213 06:50:09.940935 4971 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 13 06:50:09 crc kubenswrapper[4971]: I1213 06:50:09.940969 4971 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 13 06:50:09 crc kubenswrapper[4971]: I1213 06:50:09.940980 4971 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 13 06:50:09 crc kubenswrapper[4971]: I1213 06:50:09.940996 4971 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 13 06:50:09 crc kubenswrapper[4971]: I1213 06:50:09.941007 4971 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-13T06:50:09Z","lastTransitionTime":"2025-12-13T06:50:09Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 13 06:50:10 crc kubenswrapper[4971]: I1213 06:50:10.042901 4971 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 13 06:50:10 crc kubenswrapper[4971]: I1213 06:50:10.042938 4971 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 13 06:50:10 crc kubenswrapper[4971]: I1213 06:50:10.042958 4971 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 13 06:50:10 crc kubenswrapper[4971]: I1213 06:50:10.042975 4971 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 13 06:50:10 crc kubenswrapper[4971]: I1213 06:50:10.042986 4971 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-13T06:50:10Z","lastTransitionTime":"2025-12-13T06:50:10Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 13 06:50:10 crc kubenswrapper[4971]: I1213 06:50:10.145088 4971 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 13 06:50:10 crc kubenswrapper[4971]: I1213 06:50:10.145131 4971 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 13 06:50:10 crc kubenswrapper[4971]: I1213 06:50:10.145144 4971 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 13 06:50:10 crc kubenswrapper[4971]: I1213 06:50:10.145161 4971 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 13 06:50:10 crc kubenswrapper[4971]: I1213 06:50:10.145172 4971 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-13T06:50:10Z","lastTransitionTime":"2025-12-13T06:50:10Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 13 06:50:10 crc kubenswrapper[4971]: I1213 06:50:10.248282 4971 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 13 06:50:10 crc kubenswrapper[4971]: I1213 06:50:10.248337 4971 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 13 06:50:10 crc kubenswrapper[4971]: I1213 06:50:10.248356 4971 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 13 06:50:10 crc kubenswrapper[4971]: I1213 06:50:10.248379 4971 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 13 06:50:10 crc kubenswrapper[4971]: I1213 06:50:10.248395 4971 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-13T06:50:10Z","lastTransitionTime":"2025-12-13T06:50:10Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 13 06:50:10 crc kubenswrapper[4971]: I1213 06:50:10.352910 4971 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 13 06:50:10 crc kubenswrapper[4971]: I1213 06:50:10.352973 4971 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 13 06:50:10 crc kubenswrapper[4971]: I1213 06:50:10.352987 4971 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 13 06:50:10 crc kubenswrapper[4971]: I1213 06:50:10.353005 4971 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 13 06:50:10 crc kubenswrapper[4971]: I1213 06:50:10.353018 4971 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-13T06:50:10Z","lastTransitionTime":"2025-12-13T06:50:10Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 13 06:50:10 crc kubenswrapper[4971]: I1213 06:50:10.455400 4971 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 13 06:50:10 crc kubenswrapper[4971]: I1213 06:50:10.455454 4971 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 13 06:50:10 crc kubenswrapper[4971]: I1213 06:50:10.455467 4971 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 13 06:50:10 crc kubenswrapper[4971]: I1213 06:50:10.455485 4971 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 13 06:50:10 crc kubenswrapper[4971]: I1213 06:50:10.455498 4971 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-13T06:50:10Z","lastTransitionTime":"2025-12-13T06:50:10Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 13 06:50:10 crc kubenswrapper[4971]: I1213 06:50:10.557289 4971 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 13 06:50:10 crc kubenswrapper[4971]: I1213 06:50:10.557341 4971 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 13 06:50:10 crc kubenswrapper[4971]: I1213 06:50:10.557352 4971 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 13 06:50:10 crc kubenswrapper[4971]: I1213 06:50:10.557375 4971 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 13 06:50:10 crc kubenswrapper[4971]: I1213 06:50:10.557389 4971 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-13T06:50:10Z","lastTransitionTime":"2025-12-13T06:50:10Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 13 06:50:10 crc kubenswrapper[4971]: I1213 06:50:10.659946 4971 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 13 06:50:10 crc kubenswrapper[4971]: I1213 06:50:10.659984 4971 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 13 06:50:10 crc kubenswrapper[4971]: I1213 06:50:10.659993 4971 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 13 06:50:10 crc kubenswrapper[4971]: I1213 06:50:10.660007 4971 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 13 06:50:10 crc kubenswrapper[4971]: I1213 06:50:10.660018 4971 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-13T06:50:10Z","lastTransitionTime":"2025-12-13T06:50:10Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 13 06:50:10 crc kubenswrapper[4971]: I1213 06:50:10.762369 4971 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 13 06:50:10 crc kubenswrapper[4971]: I1213 06:50:10.762409 4971 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 13 06:50:10 crc kubenswrapper[4971]: I1213 06:50:10.762419 4971 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 13 06:50:10 crc kubenswrapper[4971]: I1213 06:50:10.762435 4971 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 13 06:50:10 crc kubenswrapper[4971]: I1213 06:50:10.762446 4971 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-13T06:50:10Z","lastTransitionTime":"2025-12-13T06:50:10Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 13 06:50:10 crc kubenswrapper[4971]: I1213 06:50:10.865145 4971 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 13 06:50:10 crc kubenswrapper[4971]: I1213 06:50:10.865195 4971 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 13 06:50:10 crc kubenswrapper[4971]: I1213 06:50:10.865209 4971 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 13 06:50:10 crc kubenswrapper[4971]: I1213 06:50:10.865229 4971 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 13 06:50:10 crc kubenswrapper[4971]: I1213 06:50:10.865245 4971 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-13T06:50:10Z","lastTransitionTime":"2025-12-13T06:50:10Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 13 06:50:10 crc kubenswrapper[4971]: I1213 06:50:10.967560 4971 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 13 06:50:10 crc kubenswrapper[4971]: I1213 06:50:10.967598 4971 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 13 06:50:10 crc kubenswrapper[4971]: I1213 06:50:10.967607 4971 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 13 06:50:10 crc kubenswrapper[4971]: I1213 06:50:10.967621 4971 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 13 06:50:10 crc kubenswrapper[4971]: I1213 06:50:10.967632 4971 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-13T06:50:10Z","lastTransitionTime":"2025-12-13T06:50:10Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 13 06:50:11 crc kubenswrapper[4971]: I1213 06:50:11.070039 4971 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 13 06:50:11 crc kubenswrapper[4971]: I1213 06:50:11.070078 4971 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 13 06:50:11 crc kubenswrapper[4971]: I1213 06:50:11.070091 4971 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 13 06:50:11 crc kubenswrapper[4971]: I1213 06:50:11.070104 4971 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 13 06:50:11 crc kubenswrapper[4971]: I1213 06:50:11.070114 4971 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-13T06:50:11Z","lastTransitionTime":"2025-12-13T06:50:11Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 13 06:50:11 crc kubenswrapper[4971]: I1213 06:50:11.172265 4971 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 13 06:50:11 crc kubenswrapper[4971]: I1213 06:50:11.172305 4971 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 13 06:50:11 crc kubenswrapper[4971]: I1213 06:50:11.172317 4971 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 13 06:50:11 crc kubenswrapper[4971]: I1213 06:50:11.172335 4971 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 13 06:50:11 crc kubenswrapper[4971]: I1213 06:50:11.172348 4971 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-13T06:50:11Z","lastTransitionTime":"2025-12-13T06:50:11Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 13 06:50:11 crc kubenswrapper[4971]: I1213 06:50:11.252983 4971 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 13 06:50:11 crc kubenswrapper[4971]: I1213 06:50:11.253020 4971 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 13 06:50:11 crc kubenswrapper[4971]: I1213 06:50:11.253029 4971 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 13 06:50:11 crc kubenswrapper[4971]: I1213 06:50:11.253064 4971 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 13 06:50:11 crc kubenswrapper[4971]: I1213 06:50:11.253073 4971 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-13T06:50:11Z","lastTransitionTime":"2025-12-13T06:50:11Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 13 06:50:11 crc kubenswrapper[4971]: E1213 06:50:11.264716 4971 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-12-13T06:50:11Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-13T06:50:11Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-13T06:50:11Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-13T06:50:11Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-13T06:50:11Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-13T06:50:11Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-13T06:50:11Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-13T06:50:11Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"bd36cbc9-bbe9-45a1-a21a-ad2bd10499ee\\\",\\\"systemUUID\\\":\\\"a9af00dc-cf25-47af-9997-79577f8f3bea\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-13T06:50:11Z is after 2025-08-24T17:21:41Z" Dec 13 06:50:11 crc kubenswrapper[4971]: I1213 06:50:11.268004 4971 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 13 06:50:11 crc kubenswrapper[4971]: I1213 06:50:11.268048 4971 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 13 06:50:11 crc kubenswrapper[4971]: I1213 06:50:11.268058 4971 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 13 06:50:11 crc kubenswrapper[4971]: I1213 06:50:11.268078 4971 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 13 06:50:11 crc kubenswrapper[4971]: I1213 06:50:11.268089 4971 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-13T06:50:11Z","lastTransitionTime":"2025-12-13T06:50:11Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 13 06:50:11 crc kubenswrapper[4971]: E1213 06:50:11.279017 4971 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-12-13T06:50:11Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-13T06:50:11Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-13T06:50:11Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-13T06:50:11Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-13T06:50:11Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-13T06:50:11Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-13T06:50:11Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-13T06:50:11Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"bd36cbc9-bbe9-45a1-a21a-ad2bd10499ee\\\",\\\"systemUUID\\\":\\\"a9af00dc-cf25-47af-9997-79577f8f3bea\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-13T06:50:11Z is after 2025-08-24T17:21:41Z" Dec 13 06:50:11 crc kubenswrapper[4971]: I1213 06:50:11.282694 4971 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 13 06:50:11 crc kubenswrapper[4971]: I1213 06:50:11.282742 4971 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 13 06:50:11 crc kubenswrapper[4971]: I1213 06:50:11.282755 4971 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 13 06:50:11 crc kubenswrapper[4971]: I1213 06:50:11.282774 4971 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 13 06:50:11 crc kubenswrapper[4971]: I1213 06:50:11.282785 4971 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-13T06:50:11Z","lastTransitionTime":"2025-12-13T06:50:11Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 13 06:50:11 crc kubenswrapper[4971]: E1213 06:50:11.293723 4971 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-12-13T06:50:11Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-13T06:50:11Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-13T06:50:11Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-13T06:50:11Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-13T06:50:11Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-13T06:50:11Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-13T06:50:11Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-13T06:50:11Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"bd36cbc9-bbe9-45a1-a21a-ad2bd10499ee\\\",\\\"systemUUID\\\":\\\"a9af00dc-cf25-47af-9997-79577f8f3bea\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-13T06:50:11Z is after 2025-08-24T17:21:41Z" Dec 13 06:50:11 crc kubenswrapper[4971]: I1213 06:50:11.297635 4971 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 13 06:50:11 crc kubenswrapper[4971]: I1213 06:50:11.297695 4971 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 13 06:50:11 crc kubenswrapper[4971]: I1213 06:50:11.297708 4971 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 13 06:50:11 crc kubenswrapper[4971]: I1213 06:50:11.297728 4971 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 13 06:50:11 crc kubenswrapper[4971]: I1213 06:50:11.297738 4971 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-13T06:50:11Z","lastTransitionTime":"2025-12-13T06:50:11Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 13 06:50:11 crc kubenswrapper[4971]: E1213 06:50:11.309667 4971 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-12-13T06:50:11Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-13T06:50:11Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-13T06:50:11Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-13T06:50:11Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-13T06:50:11Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-13T06:50:11Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-13T06:50:11Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-13T06:50:11Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"bd36cbc9-bbe9-45a1-a21a-ad2bd10499ee\\\",\\\"systemUUID\\\":\\\"a9af00dc-cf25-47af-9997-79577f8f3bea\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-13T06:50:11Z is after 2025-08-24T17:21:41Z" Dec 13 06:50:11 crc kubenswrapper[4971]: I1213 06:50:11.314255 4971 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 13 06:50:11 crc kubenswrapper[4971]: I1213 06:50:11.314315 4971 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 13 06:50:11 crc kubenswrapper[4971]: I1213 06:50:11.314329 4971 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 13 06:50:11 crc kubenswrapper[4971]: I1213 06:50:11.314352 4971 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 13 06:50:11 crc kubenswrapper[4971]: I1213 06:50:11.314365 4971 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-13T06:50:11Z","lastTransitionTime":"2025-12-13T06:50:11Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 13 06:50:11 crc kubenswrapper[4971]: E1213 06:50:11.325759 4971 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-12-13T06:50:11Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-13T06:50:11Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-13T06:50:11Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-13T06:50:11Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-13T06:50:11Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-13T06:50:11Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-13T06:50:11Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-13T06:50:11Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"bd36cbc9-bbe9-45a1-a21a-ad2bd10499ee\\\",\\\"systemUUID\\\":\\\"a9af00dc-cf25-47af-9997-79577f8f3bea\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-13T06:50:11Z is after 2025-08-24T17:21:41Z" Dec 13 06:50:11 crc kubenswrapper[4971]: E1213 06:50:11.325945 4971 kubelet_node_status.go:572] "Unable to update node status" err="update node status exceeds retry count" Dec 13 06:50:11 crc kubenswrapper[4971]: I1213 06:50:11.327480 4971 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 13 06:50:11 crc kubenswrapper[4971]: I1213 06:50:11.327531 4971 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 13 06:50:11 crc kubenswrapper[4971]: I1213 06:50:11.327567 4971 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 13 06:50:11 crc kubenswrapper[4971]: I1213 06:50:11.327584 4971 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 13 06:50:11 crc kubenswrapper[4971]: I1213 06:50:11.327597 4971 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-13T06:50:11Z","lastTransitionTime":"2025-12-13T06:50:11Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 13 06:50:11 crc kubenswrapper[4971]: I1213 06:50:11.396373 4971 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/c7185dec-3240-42e8-9d90-6182a5e8b65c-metrics-certs\") pod \"network-metrics-daemon-x98k7\" (UID: \"c7185dec-3240-42e8-9d90-6182a5e8b65c\") " pod="openshift-multus/network-metrics-daemon-x98k7" Dec 13 06:50:11 crc kubenswrapper[4971]: E1213 06:50:11.396626 4971 secret.go:188] Couldn't get secret openshift-multus/metrics-daemon-secret: object "openshift-multus"/"metrics-daemon-secret" not registered Dec 13 06:50:11 crc kubenswrapper[4971]: E1213 06:50:11.396736 4971 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/c7185dec-3240-42e8-9d90-6182a5e8b65c-metrics-certs podName:c7185dec-3240-42e8-9d90-6182a5e8b65c nodeName:}" failed. No retries permitted until 2025-12-13 06:50:43.396714782 +0000 UTC m=+100.001124230 (durationBeforeRetry 32s). Error: MountVolume.SetUp failed for volume "metrics-certs" (UniqueName: "kubernetes.io/secret/c7185dec-3240-42e8-9d90-6182a5e8b65c-metrics-certs") pod "network-metrics-daemon-x98k7" (UID: "c7185dec-3240-42e8-9d90-6182a5e8b65c") : object "openshift-multus"/"metrics-daemon-secret" not registered Dec 13 06:50:11 crc kubenswrapper[4971]: I1213 06:50:11.430039 4971 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 13 06:50:11 crc kubenswrapper[4971]: I1213 06:50:11.430080 4971 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 13 06:50:11 crc kubenswrapper[4971]: I1213 06:50:11.430091 4971 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 13 06:50:11 crc kubenswrapper[4971]: I1213 06:50:11.430105 4971 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 13 06:50:11 crc kubenswrapper[4971]: I1213 06:50:11.430114 4971 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-13T06:50:11Z","lastTransitionTime":"2025-12-13T06:50:11Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 13 06:50:11 crc kubenswrapper[4971]: I1213 06:50:11.532719 4971 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 13 06:50:11 crc kubenswrapper[4971]: I1213 06:50:11.532762 4971 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 13 06:50:11 crc kubenswrapper[4971]: I1213 06:50:11.532774 4971 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 13 06:50:11 crc kubenswrapper[4971]: I1213 06:50:11.532795 4971 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 13 06:50:11 crc kubenswrapper[4971]: I1213 06:50:11.532806 4971 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-13T06:50:11Z","lastTransitionTime":"2025-12-13T06:50:11Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 13 06:50:11 crc kubenswrapper[4971]: I1213 06:50:11.636653 4971 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 13 06:50:11 crc kubenswrapper[4971]: I1213 06:50:11.636723 4971 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 13 06:50:11 crc kubenswrapper[4971]: I1213 06:50:11.636736 4971 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 13 06:50:11 crc kubenswrapper[4971]: I1213 06:50:11.636764 4971 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 13 06:50:11 crc kubenswrapper[4971]: I1213 06:50:11.636777 4971 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-13T06:50:11Z","lastTransitionTime":"2025-12-13T06:50:11Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 13 06:50:11 crc kubenswrapper[4971]: I1213 06:50:11.739469 4971 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 13 06:50:11 crc kubenswrapper[4971]: I1213 06:50:11.739541 4971 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 13 06:50:11 crc kubenswrapper[4971]: I1213 06:50:11.739555 4971 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 13 06:50:11 crc kubenswrapper[4971]: I1213 06:50:11.739587 4971 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 13 06:50:11 crc kubenswrapper[4971]: I1213 06:50:11.739602 4971 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-13T06:50:11Z","lastTransitionTime":"2025-12-13T06:50:11Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 13 06:50:11 crc kubenswrapper[4971]: I1213 06:50:11.767714 4971 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 13 06:50:11 crc kubenswrapper[4971]: I1213 06:50:11.767714 4971 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-x98k7" Dec 13 06:50:11 crc kubenswrapper[4971]: E1213 06:50:11.767844 4971 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 13 06:50:11 crc kubenswrapper[4971]: I1213 06:50:11.767863 4971 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 13 06:50:11 crc kubenswrapper[4971]: E1213 06:50:11.767945 4971 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-x98k7" podUID="c7185dec-3240-42e8-9d90-6182a5e8b65c" Dec 13 06:50:11 crc kubenswrapper[4971]: E1213 06:50:11.768018 4971 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 13 06:50:11 crc kubenswrapper[4971]: I1213 06:50:11.768030 4971 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 13 06:50:11 crc kubenswrapper[4971]: E1213 06:50:11.768224 4971 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 13 06:50:11 crc kubenswrapper[4971]: I1213 06:50:11.841457 4971 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 13 06:50:11 crc kubenswrapper[4971]: I1213 06:50:11.841557 4971 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 13 06:50:11 crc kubenswrapper[4971]: I1213 06:50:11.841569 4971 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 13 06:50:11 crc kubenswrapper[4971]: I1213 06:50:11.841627 4971 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 13 06:50:11 crc kubenswrapper[4971]: I1213 06:50:11.841642 4971 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-13T06:50:11Z","lastTransitionTime":"2025-12-13T06:50:11Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 13 06:50:11 crc kubenswrapper[4971]: I1213 06:50:11.944786 4971 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 13 06:50:11 crc kubenswrapper[4971]: I1213 06:50:11.944839 4971 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 13 06:50:11 crc kubenswrapper[4971]: I1213 06:50:11.944851 4971 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 13 06:50:11 crc kubenswrapper[4971]: I1213 06:50:11.944869 4971 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 13 06:50:11 crc kubenswrapper[4971]: I1213 06:50:11.944880 4971 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-13T06:50:11Z","lastTransitionTime":"2025-12-13T06:50:11Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 13 06:50:12 crc kubenswrapper[4971]: I1213 06:50:12.046760 4971 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 13 06:50:12 crc kubenswrapper[4971]: I1213 06:50:12.046798 4971 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 13 06:50:12 crc kubenswrapper[4971]: I1213 06:50:12.046806 4971 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 13 06:50:12 crc kubenswrapper[4971]: I1213 06:50:12.046820 4971 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 13 06:50:12 crc kubenswrapper[4971]: I1213 06:50:12.046832 4971 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-13T06:50:12Z","lastTransitionTime":"2025-12-13T06:50:12Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 13 06:50:12 crc kubenswrapper[4971]: I1213 06:50:12.149564 4971 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 13 06:50:12 crc kubenswrapper[4971]: I1213 06:50:12.149604 4971 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 13 06:50:12 crc kubenswrapper[4971]: I1213 06:50:12.149613 4971 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 13 06:50:12 crc kubenswrapper[4971]: I1213 06:50:12.149628 4971 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 13 06:50:12 crc kubenswrapper[4971]: I1213 06:50:12.149638 4971 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-13T06:50:12Z","lastTransitionTime":"2025-12-13T06:50:12Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 13 06:50:12 crc kubenswrapper[4971]: I1213 06:50:12.252174 4971 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 13 06:50:12 crc kubenswrapper[4971]: I1213 06:50:12.252235 4971 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 13 06:50:12 crc kubenswrapper[4971]: I1213 06:50:12.252247 4971 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 13 06:50:12 crc kubenswrapper[4971]: I1213 06:50:12.252266 4971 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 13 06:50:12 crc kubenswrapper[4971]: I1213 06:50:12.252279 4971 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-13T06:50:12Z","lastTransitionTime":"2025-12-13T06:50:12Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 13 06:50:12 crc kubenswrapper[4971]: I1213 06:50:12.354936 4971 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 13 06:50:12 crc kubenswrapper[4971]: I1213 06:50:12.354977 4971 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 13 06:50:12 crc kubenswrapper[4971]: I1213 06:50:12.354987 4971 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 13 06:50:12 crc kubenswrapper[4971]: I1213 06:50:12.355005 4971 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 13 06:50:12 crc kubenswrapper[4971]: I1213 06:50:12.355014 4971 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-13T06:50:12Z","lastTransitionTime":"2025-12-13T06:50:12Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 13 06:50:12 crc kubenswrapper[4971]: I1213 06:50:12.457795 4971 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 13 06:50:12 crc kubenswrapper[4971]: I1213 06:50:12.457858 4971 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 13 06:50:12 crc kubenswrapper[4971]: I1213 06:50:12.457882 4971 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 13 06:50:12 crc kubenswrapper[4971]: I1213 06:50:12.457915 4971 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 13 06:50:12 crc kubenswrapper[4971]: I1213 06:50:12.457937 4971 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-13T06:50:12Z","lastTransitionTime":"2025-12-13T06:50:12Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 13 06:50:12 crc kubenswrapper[4971]: I1213 06:50:12.560291 4971 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 13 06:50:12 crc kubenswrapper[4971]: I1213 06:50:12.560365 4971 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 13 06:50:12 crc kubenswrapper[4971]: I1213 06:50:12.560381 4971 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 13 06:50:12 crc kubenswrapper[4971]: I1213 06:50:12.560398 4971 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 13 06:50:12 crc kubenswrapper[4971]: I1213 06:50:12.560408 4971 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-13T06:50:12Z","lastTransitionTime":"2025-12-13T06:50:12Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 13 06:50:12 crc kubenswrapper[4971]: I1213 06:50:12.663151 4971 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 13 06:50:12 crc kubenswrapper[4971]: I1213 06:50:12.663200 4971 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 13 06:50:12 crc kubenswrapper[4971]: I1213 06:50:12.663211 4971 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 13 06:50:12 crc kubenswrapper[4971]: I1213 06:50:12.663228 4971 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 13 06:50:12 crc kubenswrapper[4971]: I1213 06:50:12.663588 4971 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-13T06:50:12Z","lastTransitionTime":"2025-12-13T06:50:12Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 13 06:50:12 crc kubenswrapper[4971]: I1213 06:50:12.766380 4971 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 13 06:50:12 crc kubenswrapper[4971]: I1213 06:50:12.766445 4971 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 13 06:50:12 crc kubenswrapper[4971]: I1213 06:50:12.766460 4971 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 13 06:50:12 crc kubenswrapper[4971]: I1213 06:50:12.766484 4971 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 13 06:50:12 crc kubenswrapper[4971]: I1213 06:50:12.766501 4971 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-13T06:50:12Z","lastTransitionTime":"2025-12-13T06:50:12Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 13 06:50:12 crc kubenswrapper[4971]: I1213 06:50:12.779226 4971 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-machine-config-operator/kube-rbac-proxy-crio-crc"] Dec 13 06:50:12 crc kubenswrapper[4971]: I1213 06:50:12.868967 4971 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 13 06:50:12 crc kubenswrapper[4971]: I1213 06:50:12.869021 4971 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 13 06:50:12 crc kubenswrapper[4971]: I1213 06:50:12.869037 4971 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 13 06:50:12 crc kubenswrapper[4971]: I1213 06:50:12.869072 4971 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 13 06:50:12 crc kubenswrapper[4971]: I1213 06:50:12.869092 4971 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-13T06:50:12Z","lastTransitionTime":"2025-12-13T06:50:12Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 13 06:50:12 crc kubenswrapper[4971]: I1213 06:50:12.971707 4971 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 13 06:50:12 crc kubenswrapper[4971]: I1213 06:50:12.971761 4971 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 13 06:50:12 crc kubenswrapper[4971]: I1213 06:50:12.971774 4971 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 13 06:50:12 crc kubenswrapper[4971]: I1213 06:50:12.971794 4971 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 13 06:50:12 crc kubenswrapper[4971]: I1213 06:50:12.971806 4971 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-13T06:50:12Z","lastTransitionTime":"2025-12-13T06:50:12Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 13 06:50:13 crc kubenswrapper[4971]: I1213 06:50:13.074800 4971 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 13 06:50:13 crc kubenswrapper[4971]: I1213 06:50:13.074871 4971 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 13 06:50:13 crc kubenswrapper[4971]: I1213 06:50:13.074891 4971 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 13 06:50:13 crc kubenswrapper[4971]: I1213 06:50:13.074917 4971 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 13 06:50:13 crc kubenswrapper[4971]: I1213 06:50:13.074935 4971 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-13T06:50:13Z","lastTransitionTime":"2025-12-13T06:50:13Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 13 06:50:13 crc kubenswrapper[4971]: I1213 06:50:13.163928 4971 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-multus_multus-wwvpl_dddf744b-2122-49ad-a980-105c09636e0f/kube-multus/0.log" Dec 13 06:50:13 crc kubenswrapper[4971]: I1213 06:50:13.164005 4971 generic.go:334] "Generic (PLEG): container finished" podID="dddf744b-2122-49ad-a980-105c09636e0f" containerID="4676d897c7fc668587faec563f3551a056557879cee3e865192d2bd28edf62fc" exitCode=1 Dec 13 06:50:13 crc kubenswrapper[4971]: I1213 06:50:13.164425 4971 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-wwvpl" event={"ID":"dddf744b-2122-49ad-a980-105c09636e0f","Type":"ContainerDied","Data":"4676d897c7fc668587faec563f3551a056557879cee3e865192d2bd28edf62fc"} Dec 13 06:50:13 crc kubenswrapper[4971]: I1213 06:50:13.164915 4971 scope.go:117] "RemoveContainer" containerID="4676d897c7fc668587faec563f3551a056557879cee3e865192d2bd28edf62fc" Dec 13 06:50:13 crc kubenswrapper[4971]: I1213 06:50:13.180024 4971 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"f9d31b5a-b770-4a15-95c8-35b49677c608\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T06:49:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T06:49:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T06:49:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T06:49:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T06:49:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c9baa042f9dada954faa57b5454cd6b0a12a297e56ddc192c52cff50e42a16fa\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-13T06:49:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://c788de86754f0bf37ee184672fc1dcb388ccebc246d4f15be4b5da661f725f8f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-13T06:49:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://760817511a9b4f96a68dc38d26631250eaab72e1dd8f123e653039b6b4e91f84\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-13T06:49:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b859be0c9feac856e9855c6a038b02b49b32fade97873c46d0bec0a32d8b5147\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-13T06:49:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-13T06:49:03Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-13T06:50:13Z is after 2025-08-24T17:21:41Z" Dec 13 06:50:13 crc kubenswrapper[4971]: I1213 06:50:13.181578 4971 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 13 06:50:13 crc kubenswrapper[4971]: I1213 06:50:13.181684 4971 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 13 06:50:13 crc kubenswrapper[4971]: I1213 06:50:13.181749 4971 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 13 06:50:13 crc kubenswrapper[4971]: I1213 06:50:13.181821 4971 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 13 06:50:13 crc kubenswrapper[4971]: I1213 06:50:13.181909 4971 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-13T06:50:13Z","lastTransitionTime":"2025-12-13T06:50:13Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 13 06:50:13 crc kubenswrapper[4971]: I1213 06:50:13.193284 4971 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"8ede2992-9576-4b23-83c2-24ac3b1d3caf\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T06:49:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T06:49:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T06:49:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T06:49:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T06:49:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9e61d25add1723ac286794abfe886358e9d64455c25379f7e3057fc3dd812418\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-13T06:49:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://7fb51d8435b4e3ce0411cae1d6c2824c552f0a33d79de01f969274516b39a5b3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-13T06:49:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://1c35f5e8d8fab6d77b9e2e0dedb195440de5d6af345cd5865abd32caa7fafd60\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-13T06:49:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://38806c430185715371ab9bf30a1af9a9d5d492abd9bb84acbdccc694f56e3a8a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"wait-for-host-port\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://38806c430185715371ab9bf30a1af9a9d5d492abd9bb84acbdccc694f56e3a8a\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-13T06:49:04Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-13T06:49:04Z\\\"}}}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-13T06:49:03Z\\\"}}\" for pod \"openshift-kube-scheduler\"/\"openshift-kube-scheduler-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-13T06:50:13Z is after 2025-08-24T17:21:41Z" Dec 13 06:50:13 crc kubenswrapper[4971]: I1213 06:50:13.206245 4971 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-ncchz" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"de914d21-a688-401e-a66e-861beb6336d9\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T06:49:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T06:49:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T06:49:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T06:49:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://244ffe3dc9d8cf799aff4c4d795bf41554c1e6bb6667fbb0c7df322fb5f96c68\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-13T06:49:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lqmjj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://0b7111af6c5a198d02bb02858837cfb4f3b20828d8d206cd18aa0ffa102d0cd4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://0b7111af6c5a198d02bb02858837cfb4f3b20828d8d206cd18aa0ffa102d0cd4\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-13T06:49:26Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-13T06:49:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lqmjj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b36777fe40e74ae1cecb7b73e204139db5279ba51cef4f1e8c4a4c491d8be403\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://b36777fe40e74ae1cecb7b73e204139db5279ba51cef4f1e8c4a4c491d8be403\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-13T06:49:28Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-13T06:49:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lqmjj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://0597a06cc663e416142311a16941d98297577693a1fe27201ea42c27496cace2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://0597a06cc663e416142311a16941d98297577693a1fe27201ea42c27496cace2\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-13T06:49:29Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-13T06:49:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lqmjj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://3f52c2e22a75c010993ca5809a3a9d1304758d4cb6e83691ef1af15a4b514ac4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://3f52c2e22a75c010993ca5809a3a9d1304758d4cb6e83691ef1af15a4b514ac4\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-13T06:49:30Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-13T06:49:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lqmjj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://9e21ca708c4c4c1228b417cece51299e0792729b7c78884bd18b2dda8fe8355c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://9e21ca708c4c4c1228b417cece51299e0792729b7c78884bd18b2dda8fe8355c\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-13T06:49:31Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-13T06:49:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lqmjj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://74552159f12c87cdda09147692b400ccac0d4e43c1ed72fe8462c4249ad69c7b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://74552159f12c87cdda09147692b400ccac0d4e43c1ed72fe8462c4249ad69c7b\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-13T06:49:32Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-13T06:49:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lqmjj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-13T06:49:25Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-ncchz\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-13T06:50:13Z is after 2025-08-24T17:21:41Z" Dec 13 06:50:13 crc kubenswrapper[4971]: I1213 06:50:13.217054 4971 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-gkcd8" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"13f5c327-f95e-42b4-8783-218500c17694\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T06:49:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T06:49:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T06:49:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T06:49:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://77e846e36f1cb40d03471dc03de11b6c3da448baa407aad38093fae9a801cf92\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-13T06:49:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gvlhn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-13T06:49:27Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-gkcd8\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-13T06:50:13Z is after 2025-08-24T17:21:41Z" Dec 13 06:50:13 crc kubenswrapper[4971]: I1213 06:50:13.227589 4971 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-x98k7" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"c7185dec-3240-42e8-9d90-6182a5e8b65c\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T06:49:39Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T06:49:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T06:49:39Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T06:49:39Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bj2gn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bj2gn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-13T06:49:39Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-x98k7\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-13T06:50:13Z is after 2025-08-24T17:21:41Z" Dec 13 06:50:13 crc kubenswrapper[4971]: I1213 06:50:13.237017 4971 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5ef36745-87bb-4be8-b941-aa8f865a5c83\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T06:49:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T06:49:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T06:49:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T06:49:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T06:49:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://35b737b0d2c5c5503ba2819f34d2de0599792a77beb1fbeb561d44efd1f212c5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-crio\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-13T06:49:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kube\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"var-lib-kubelet\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f0d864149bfcbb6227e6a6fbde23fff650d688850978eae0caf04515eb21cd85\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://f0d864149bfcbb6227e6a6fbde23fff650d688850978eae0caf04515eb21cd85\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-13T06:49:04Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-13T06:49:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var\\\",\\\"name\\\":\\\"var-lib-kubelet\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-13T06:49:03Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"kube-rbac-proxy-crio-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-13T06:50:13Z is after 2025-08-24T17:21:41Z" Dec 13 06:50:13 crc kubenswrapper[4971]: I1213 06:50:13.255583 4971 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"1ed9f82e-fe5c-4b76-a619-62e5abd88d3b\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T06:49:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T06:49:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T06:49:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T06:49:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T06:49:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e457329a78a2d24337f994e8a0a1f75a2a667244aa75eb84613d1e7138364aac\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-13T06:49:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://eb014de8140deef097d9bd94f3b9fa4ab3558bc4449bef84b83bf62bfabf5836\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-13T06:49:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://2c438cc74d555f7c9db4569d1c56b78b1088ed8105697d2c0f31e6c5136f3dd0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-13T06:49:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://aa6fa6e4340dcd066252df3c56d79a3a39ed531d5eea8d1d5941d6952368d3af\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-13T06:49:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://c4471b98854c52d5cd084d0ba3e64e5b08e9518ec3c372cce2c875a440522457\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-13T06:49:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7d2746335c98dae3a0f43024d34cb62ebcb42a01c8f9ce18d52b29d4c45598e5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://7d2746335c98dae3a0f43024d34cb62ebcb42a01c8f9ce18d52b29d4c45598e5\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-13T06:49:04Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-13T06:49:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://bc0e77d10a738128bea82122496bc42235fc1688f8907756d93e61b651902b57\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://bc0e77d10a738128bea82122496bc42235fc1688f8907756d93e61b651902b57\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-13T06:49:05Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-13T06:49:05Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://f821e418cba5e832da8839268bcbbd9630bdf319a212d5382bfa2c4b0b73a268\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://f821e418cba5e832da8839268bcbbd9630bdf319a212d5382bfa2c4b0b73a268\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-13T06:49:07Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-13T06:49:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-13T06:49:03Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-13T06:50:13Z is after 2025-08-24T17:21:41Z" Dec 13 06:50:13 crc kubenswrapper[4971]: I1213 06:50:13.268507 4971 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-13T06:49:24Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-13T06:49:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-13T06:49:24Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4328cd6a8464ed468229bbfbc8265f969427916d72ebb92237dbe68efe70fa47\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-13T06:49:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://9d715ca82d9b5a594e41f2b98da6f4277da0e4e1dec114969f901c00f6d7553d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-13T06:49:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-13T06:50:13Z is after 2025-08-24T17:21:41Z" Dec 13 06:50:13 crc kubenswrapper[4971]: I1213 06:50:13.277332 4971 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-5rzh5" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"6bf4591d-2483-4132-9672-d0fd6c995968\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T06:49:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T06:49:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T06:49:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T06:49:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://50e2cb14e4ffb603acb06e2281615a99f06182d345c4e70f2446fa2077033bbb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-13T06:49:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dfxqn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-13T06:49:25Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-5rzh5\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-13T06:50:13Z is after 2025-08-24T17:21:41Z" Dec 13 06:50:13 crc kubenswrapper[4971]: I1213 06:50:13.284221 4971 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 13 06:50:13 crc kubenswrapper[4971]: I1213 06:50:13.284260 4971 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 13 06:50:13 crc kubenswrapper[4971]: I1213 06:50:13.284271 4971 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 13 06:50:13 crc kubenswrapper[4971]: I1213 06:50:13.284288 4971 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 13 06:50:13 crc kubenswrapper[4971]: I1213 06:50:13.284296 4971 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-13T06:50:13Z","lastTransitionTime":"2025-12-13T06:50:13Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 13 06:50:13 crc kubenswrapper[4971]: I1213 06:50:13.288024 4971 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-z572s" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"b4146d3d-4034-49a9-bd37-164856255137\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T06:49:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T06:49:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T06:49:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T06:49:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://46fd4e1070835cfbf7f35d497eff7f047473fc4048ee07812925df5ebdf55342\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-13T06:49:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-j6l9b\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://dc79a0968f85723e6ad0dc269c8da1f7472adb1bb2b87edd017ee2295bfbf89c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-13T06:49:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-j6l9b\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-13T06:49:38Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-z572s\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-13T06:50:13Z is after 2025-08-24T17:21:41Z" Dec 13 06:50:13 crc kubenswrapper[4971]: I1213 06:50:13.299602 4971 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3ad3268a-a982-428c-9c06-4aaeddd372d9\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T06:49:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T06:49:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T06:49:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T06:49:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T06:49:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3716afdf2bc693be80a09fc9f7727d334f435caf1d4c8ec4811b3a1f1652cdfe\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-13T06:49:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://30df3abf8c79db8af52c6a918f73e1ce410ddafcf341636424c267e1eeb195a7\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-13T06:49:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://ccdf65f6d2b1eb3647a03ba37c305c32e34addfdf79ff937c8dddaa3e5d5b5c6\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-13T06:49:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://770b73dd0c2b6f38744e2c1cd6aa186a69bb87d1cd015a603ad2e33be06a8f6b\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://b3c1bab3d53f975b2e24582ce2771bc26d600e0dd7c30339dcfd18a071162ad4\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-12-13T06:49:23Z\\\",\\\"message\\\":\\\"file observer\\\\nW1213 06:49:23.283403 1 builder.go:272] unable to get owner reference (falling back to namespace): pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\nI1213 06:49:23.284336 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1213 06:49:23.286978 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-981773671/tls.crt::/tmp/serving-cert-981773671/tls.key\\\\\\\"\\\\nI1213 06:49:23.775072 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1213 06:49:23.778615 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1213 06:49:23.778645 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1213 06:49:23.778680 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1213 06:49:23.778691 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1213 06:49:23.785058 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1213 06:49:23.786150 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1213 06:49:23.786196 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1213 06:49:23.786223 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1213 06:49:23.786246 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1213 06:49:23.786269 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1213 06:49:23.786291 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI1213 06:49:23.785608 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF1213 06:49:23.787563 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-13T06:49:08Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-13T06:49:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://9acd07066deed483bd7fa61ee45fda09539d110717a134a56827f871186f92fc\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-13T06:49:07Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://38c91cfbaecfdbd3e3b663422a91ab4294610afb6bbf9ebdb77db24e2b76c682\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://38c91cfbaecfdbd3e3b663422a91ab4294610afb6bbf9ebdb77db24e2b76c682\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-13T06:49:05Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-13T06:49:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-13T06:49:03Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-13T06:50:13Z is after 2025-08-24T17:21:41Z" Dec 13 06:50:13 crc kubenswrapper[4971]: I1213 06:50:13.311420 4971 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-13T06:49:23Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-13T06:49:23Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-13T06:50:13Z is after 2025-08-24T17:21:41Z" Dec 13 06:50:13 crc kubenswrapper[4971]: I1213 06:50:13.323850 4971 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-13T06:49:23Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-13T06:49:23Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-13T06:50:13Z is after 2025-08-24T17:21:41Z" Dec 13 06:50:13 crc kubenswrapper[4971]: I1213 06:50:13.344769 4971 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-zlcg5" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"736c6b1a-93dd-4f71-81b6-1c43fcf5f556\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T06:49:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T06:49:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T06:49:25Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T06:49:25Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://325223543b3c7db417f769533559a1ca2b23a723494e1885b0fd0480b5a6399f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-13T06:49:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vzmsr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://f22a4ec756119c4f791fb5105cf1f31df3e548b2b07b1f909a62802e9b6035ae\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-13T06:49:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vzmsr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://f6073140d9f86929b6d4bd2960ede0617a5c9620b269bbab043bae37e0d389f5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-13T06:49:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vzmsr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://693a2ab434c250cb140a866e8068b156df9191414fed1d63bee919a2528ea6e5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-13T06:49:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vzmsr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://219620f51474e60b3e1f880c0972eb612dffc4bad24462bbba2c093ae817d852\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-13T06:49:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vzmsr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://f137f4b441c2402093ab430a9689a46221c4fb4bc70d99fb992f62b0a989c465\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-13T06:49:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vzmsr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://0b454a87c6dd11bd8c3bb3553ee71de32ed6f185984e5057a20f2237b35c45d9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://0b454a87c6dd11bd8c3bb3553ee71de32ed6f185984e5057a20f2237b35c45d9\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-12-13T06:49:56Z\\\",\\\"message\\\":\\\" for pod on switch crc\\\\nI1213 06:49:56.566617 6645 metrics.go:553] Stopping metrics server at address \\\\\\\"127.0.0.1:29103\\\\\\\"\\\\nI1213 06:49:56.566625 6645 base_network_controller_pods.go:477] [default/openshift-network-console/networking-console-plugin-85b44fc459-gdk6g] creating logical port openshift-network-console_networking-console-plugin-85b44fc459-gdk6g for pod on switch crc\\\\nI1213 06:49:56.566569 6645 loadbalancer.go:304] Deleted 0 stale LBs for map[string]string{\\\\\\\"k8s.ovn.org/kind\\\\\\\":\\\\\\\"Service\\\\\\\", \\\\\\\"k8s.ovn.org/owner\\\\\\\":\\\\\\\"openshift-console-operator/metrics\\\\\\\"}\\\\nI1213 06:49:56.566666 6645 services_controller.go:360] Finished syncing service metrics on namespace openshift-console-operator for network=default : 1.804735ms\\\\nI1213 06:49:56.566669 6645 loadbalancer.go:304] Deleted 0 stale LBs for map[string]string{\\\\\\\"k8s.ovn.org/kind\\\\\\\":\\\\\\\"Service\\\\\\\", \\\\\\\"k8s.ovn.org/owner\\\\\\\":\\\\\\\"openshift-apiserver/check-endpoints\\\\\\\"}\\\\nF1213 06:49:56.566698 6645 ovnkube.go:137] failed to run ovnkube: [failed to start network controller: failed to start default network controller: unable to create admin network policy controller, err: could not add Event Handler for anpInformer during admin network policy controller initialization, handler {0x1fcc6e0 0x1fcc3c0 0x1fcc360} was not added to shared informer because it has stopped already, failed to start node network controller: failed \\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-13T06:49:55Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":2,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 20s restarting failed container=ovnkube-controller pod=ovnkube-node-zlcg5_openshift-ovn-kubernetes(736c6b1a-93dd-4f71-81b6-1c43fcf5f556)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vzmsr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://868049c6a3a4ebe4036d1d22e02ba6b76bced2ee7e269bb228accc4a0636678a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-13T06:49:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vzmsr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://451b22df6973ddc8fa2209a65d64d9e41b33eb231bb4adfab88cf0e5943bb246\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://451b22df6973ddc8fa2209a65d64d9e41b33eb231bb4adfab88cf0e5943bb246\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-13T06:49:26Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-13T06:49:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vzmsr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-13T06:49:25Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-zlcg5\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-13T06:50:13Z is after 2025-08-24T17:21:41Z" Dec 13 06:50:13 crc kubenswrapper[4971]: I1213 06:50:13.362435 4971 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-13T06:49:24Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-13T06:49:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-13T06:49:24Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d0b2ecf0d5b22f23217f42f259ae7ffb6359555b22c30eaebf667eaf254b3bab\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-13T06:49:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-13T06:50:13Z is after 2025-08-24T17:21:41Z" Dec 13 06:50:13 crc kubenswrapper[4971]: I1213 06:50:13.375576 4971 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-13T06:49:23Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-13T06:49:23Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-13T06:50:13Z is after 2025-08-24T17:21:41Z" Dec 13 06:50:13 crc kubenswrapper[4971]: I1213 06:50:13.385636 4971 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-13T06:49:26Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-13T06:49:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-13T06:49:26Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9c8a0d1db0e754462d805e82daf5f52ffd5457507a30456535ec59e096c9d092\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-13T06:49:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-13T06:50:13Z is after 2025-08-24T17:21:41Z" Dec 13 06:50:13 crc kubenswrapper[4971]: I1213 06:50:13.386386 4971 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 13 06:50:13 crc kubenswrapper[4971]: I1213 06:50:13.386481 4971 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 13 06:50:13 crc kubenswrapper[4971]: I1213 06:50:13.386578 4971 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 13 06:50:13 crc kubenswrapper[4971]: I1213 06:50:13.386661 4971 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 13 06:50:13 crc kubenswrapper[4971]: I1213 06:50:13.386727 4971 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-13T06:50:13Z","lastTransitionTime":"2025-12-13T06:50:13Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 13 06:50:13 crc kubenswrapper[4971]: I1213 06:50:13.398695 4971 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-82xjz" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e0c4af12-edda-41d7-9b44-f87396174ef2\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T06:49:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T06:49:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T06:49:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T06:49:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://0f4577d9e37f860813e221f2e208177d8d383fd04f673ce357a0a405ea29d1e6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-13T06:49:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cnxsq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://7e0d2020e62f1a277cff7047e1907dc747ed393a1f6c983e1ee258bd156c16c7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-13T06:49:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cnxsq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-13T06:49:25Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-82xjz\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-13T06:50:13Z is after 2025-08-24T17:21:41Z" Dec 13 06:50:13 crc kubenswrapper[4971]: I1213 06:50:13.415272 4971 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-wwvpl" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"dddf744b-2122-49ad-a980-105c09636e0f\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T06:49:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T06:49:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T06:50:13Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T06:50:13Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4676d897c7fc668587faec563f3551a056557879cee3e865192d2bd28edf62fc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://4676d897c7fc668587faec563f3551a056557879cee3e865192d2bd28edf62fc\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-12-13T06:50:12Z\\\",\\\"message\\\":\\\"2025-12-13T06:49:27+00:00 [cnibincopy] Successfully copied files in /usr/src/multus-cni/rhel9/bin/ to /host/opt/cni/bin/upgrade_263f091d-d847-4fad-88d3-fbf82c65c231\\\\n2025-12-13T06:49:27+00:00 [cnibincopy] Successfully moved files in /host/opt/cni/bin/upgrade_263f091d-d847-4fad-88d3-fbf82c65c231 to /host/opt/cni/bin/\\\\n2025-12-13T06:49:27Z [verbose] multus-daemon started\\\\n2025-12-13T06:49:27Z [verbose] Readiness Indicator file check\\\\n2025-12-13T06:50:12Z [error] have you checked that your default network is ready? still waiting for readinessindicatorfile @ /host/run/multus/cni/net.d/10-ovn-kubernetes.conf. pollimmediate error: timed out waiting for the condition\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-13T06:49:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-czz9v\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-13T06:49:25Z\\\"}}\" for pod \"openshift-multus\"/\"multus-wwvpl\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-13T06:50:13Z is after 2025-08-24T17:21:41Z" Dec 13 06:50:13 crc kubenswrapper[4971]: I1213 06:50:13.489025 4971 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 13 06:50:13 crc kubenswrapper[4971]: I1213 06:50:13.489259 4971 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 13 06:50:13 crc kubenswrapper[4971]: I1213 06:50:13.489392 4971 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 13 06:50:13 crc kubenswrapper[4971]: I1213 06:50:13.489480 4971 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 13 06:50:13 crc kubenswrapper[4971]: I1213 06:50:13.489591 4971 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-13T06:50:13Z","lastTransitionTime":"2025-12-13T06:50:13Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 13 06:50:13 crc kubenswrapper[4971]: I1213 06:50:13.592134 4971 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 13 06:50:13 crc kubenswrapper[4971]: I1213 06:50:13.592308 4971 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 13 06:50:13 crc kubenswrapper[4971]: I1213 06:50:13.592397 4971 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 13 06:50:13 crc kubenswrapper[4971]: I1213 06:50:13.592605 4971 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 13 06:50:13 crc kubenswrapper[4971]: I1213 06:50:13.592772 4971 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-13T06:50:13Z","lastTransitionTime":"2025-12-13T06:50:13Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 13 06:50:13 crc kubenswrapper[4971]: I1213 06:50:13.695491 4971 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 13 06:50:13 crc kubenswrapper[4971]: I1213 06:50:13.696236 4971 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 13 06:50:13 crc kubenswrapper[4971]: I1213 06:50:13.696326 4971 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 13 06:50:13 crc kubenswrapper[4971]: I1213 06:50:13.696414 4971 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 13 06:50:13 crc kubenswrapper[4971]: I1213 06:50:13.696503 4971 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-13T06:50:13Z","lastTransitionTime":"2025-12-13T06:50:13Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 13 06:50:13 crc kubenswrapper[4971]: I1213 06:50:13.768400 4971 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 13 06:50:13 crc kubenswrapper[4971]: I1213 06:50:13.768493 4971 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 13 06:50:13 crc kubenswrapper[4971]: E1213 06:50:13.768684 4971 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 13 06:50:13 crc kubenswrapper[4971]: I1213 06:50:13.768570 4971 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 13 06:50:13 crc kubenswrapper[4971]: E1213 06:50:13.768799 4971 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 13 06:50:13 crc kubenswrapper[4971]: I1213 06:50:13.768493 4971 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-x98k7" Dec 13 06:50:13 crc kubenswrapper[4971]: E1213 06:50:13.769150 4971 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-x98k7" podUID="c7185dec-3240-42e8-9d90-6182a5e8b65c" Dec 13 06:50:13 crc kubenswrapper[4971]: E1213 06:50:13.769056 4971 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 13 06:50:13 crc kubenswrapper[4971]: I1213 06:50:13.782073 4971 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"f9d31b5a-b770-4a15-95c8-35b49677c608\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T06:49:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T06:49:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T06:49:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T06:49:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T06:49:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c9baa042f9dada954faa57b5454cd6b0a12a297e56ddc192c52cff50e42a16fa\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-13T06:49:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://c788de86754f0bf37ee184672fc1dcb388ccebc246d4f15be4b5da661f725f8f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-13T06:49:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://760817511a9b4f96a68dc38d26631250eaab72e1dd8f123e653039b6b4e91f84\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-13T06:49:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b859be0c9feac856e9855c6a038b02b49b32fade97873c46d0bec0a32d8b5147\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-13T06:49:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-13T06:49:03Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-13T06:50:13Z is after 2025-08-24T17:21:41Z" Dec 13 06:50:13 crc kubenswrapper[4971]: I1213 06:50:13.796803 4971 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"8ede2992-9576-4b23-83c2-24ac3b1d3caf\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T06:49:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T06:49:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T06:49:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T06:49:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T06:49:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9e61d25add1723ac286794abfe886358e9d64455c25379f7e3057fc3dd812418\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-13T06:49:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://7fb51d8435b4e3ce0411cae1d6c2824c552f0a33d79de01f969274516b39a5b3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-13T06:49:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://1c35f5e8d8fab6d77b9e2e0dedb195440de5d6af345cd5865abd32caa7fafd60\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-13T06:49:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://38806c430185715371ab9bf30a1af9a9d5d492abd9bb84acbdccc694f56e3a8a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"wait-for-host-port\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://38806c430185715371ab9bf30a1af9a9d5d492abd9bb84acbdccc694f56e3a8a\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-13T06:49:04Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-13T06:49:04Z\\\"}}}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-13T06:49:03Z\\\"}}\" for pod \"openshift-kube-scheduler\"/\"openshift-kube-scheduler-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-13T06:50:13Z is after 2025-08-24T17:21:41Z" Dec 13 06:50:13 crc kubenswrapper[4971]: I1213 06:50:13.799449 4971 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 13 06:50:13 crc kubenswrapper[4971]: I1213 06:50:13.799492 4971 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 13 06:50:13 crc kubenswrapper[4971]: I1213 06:50:13.799509 4971 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 13 06:50:13 crc kubenswrapper[4971]: I1213 06:50:13.799557 4971 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 13 06:50:13 crc kubenswrapper[4971]: I1213 06:50:13.799572 4971 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-13T06:50:13Z","lastTransitionTime":"2025-12-13T06:50:13Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 13 06:50:13 crc kubenswrapper[4971]: I1213 06:50:13.811031 4971 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-ncchz" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"de914d21-a688-401e-a66e-861beb6336d9\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T06:49:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T06:49:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T06:49:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T06:49:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://244ffe3dc9d8cf799aff4c4d795bf41554c1e6bb6667fbb0c7df322fb5f96c68\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-13T06:49:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lqmjj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://0b7111af6c5a198d02bb02858837cfb4f3b20828d8d206cd18aa0ffa102d0cd4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://0b7111af6c5a198d02bb02858837cfb4f3b20828d8d206cd18aa0ffa102d0cd4\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-13T06:49:26Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-13T06:49:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lqmjj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b36777fe40e74ae1cecb7b73e204139db5279ba51cef4f1e8c4a4c491d8be403\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://b36777fe40e74ae1cecb7b73e204139db5279ba51cef4f1e8c4a4c491d8be403\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-13T06:49:28Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-13T06:49:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lqmjj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://0597a06cc663e416142311a16941d98297577693a1fe27201ea42c27496cace2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://0597a06cc663e416142311a16941d98297577693a1fe27201ea42c27496cace2\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-13T06:49:29Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-13T06:49:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lqmjj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://3f52c2e22a75c010993ca5809a3a9d1304758d4cb6e83691ef1af15a4b514ac4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://3f52c2e22a75c010993ca5809a3a9d1304758d4cb6e83691ef1af15a4b514ac4\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-13T06:49:30Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-13T06:49:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lqmjj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://9e21ca708c4c4c1228b417cece51299e0792729b7c78884bd18b2dda8fe8355c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://9e21ca708c4c4c1228b417cece51299e0792729b7c78884bd18b2dda8fe8355c\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-13T06:49:31Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-13T06:49:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lqmjj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://74552159f12c87cdda09147692b400ccac0d4e43c1ed72fe8462c4249ad69c7b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://74552159f12c87cdda09147692b400ccac0d4e43c1ed72fe8462c4249ad69c7b\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-13T06:49:32Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-13T06:49:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lqmjj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-13T06:49:25Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-ncchz\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-13T06:50:13Z is after 2025-08-24T17:21:41Z" Dec 13 06:50:13 crc kubenswrapper[4971]: I1213 06:50:13.822764 4971 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-gkcd8" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"13f5c327-f95e-42b4-8783-218500c17694\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T06:49:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T06:49:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T06:49:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T06:49:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://77e846e36f1cb40d03471dc03de11b6c3da448baa407aad38093fae9a801cf92\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-13T06:49:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gvlhn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-13T06:49:27Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-gkcd8\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-13T06:50:13Z is after 2025-08-24T17:21:41Z" Dec 13 06:50:13 crc kubenswrapper[4971]: I1213 06:50:13.834061 4971 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-x98k7" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"c7185dec-3240-42e8-9d90-6182a5e8b65c\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T06:49:39Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T06:49:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T06:49:39Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T06:49:39Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bj2gn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bj2gn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-13T06:49:39Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-x98k7\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-13T06:50:13Z is after 2025-08-24T17:21:41Z" Dec 13 06:50:13 crc kubenswrapper[4971]: I1213 06:50:13.844204 4971 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5ef36745-87bb-4be8-b941-aa8f865a5c83\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T06:49:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T06:49:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T06:49:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T06:49:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T06:49:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://35b737b0d2c5c5503ba2819f34d2de0599792a77beb1fbeb561d44efd1f212c5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-crio\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-13T06:49:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kube\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"var-lib-kubelet\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f0d864149bfcbb6227e6a6fbde23fff650d688850978eae0caf04515eb21cd85\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://f0d864149bfcbb6227e6a6fbde23fff650d688850978eae0caf04515eb21cd85\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-13T06:49:04Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-13T06:49:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var\\\",\\\"name\\\":\\\"var-lib-kubelet\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-13T06:49:03Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"kube-rbac-proxy-crio-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-13T06:50:13Z is after 2025-08-24T17:21:41Z" Dec 13 06:50:13 crc kubenswrapper[4971]: I1213 06:50:13.865015 4971 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"1ed9f82e-fe5c-4b76-a619-62e5abd88d3b\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T06:49:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T06:49:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T06:49:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T06:49:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T06:49:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e457329a78a2d24337f994e8a0a1f75a2a667244aa75eb84613d1e7138364aac\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-13T06:49:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://eb014de8140deef097d9bd94f3b9fa4ab3558bc4449bef84b83bf62bfabf5836\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-13T06:49:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://2c438cc74d555f7c9db4569d1c56b78b1088ed8105697d2c0f31e6c5136f3dd0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-13T06:49:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://aa6fa6e4340dcd066252df3c56d79a3a39ed531d5eea8d1d5941d6952368d3af\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-13T06:49:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://c4471b98854c52d5cd084d0ba3e64e5b08e9518ec3c372cce2c875a440522457\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-13T06:49:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7d2746335c98dae3a0f43024d34cb62ebcb42a01c8f9ce18d52b29d4c45598e5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://7d2746335c98dae3a0f43024d34cb62ebcb42a01c8f9ce18d52b29d4c45598e5\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-13T06:49:04Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-13T06:49:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://bc0e77d10a738128bea82122496bc42235fc1688f8907756d93e61b651902b57\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://bc0e77d10a738128bea82122496bc42235fc1688f8907756d93e61b651902b57\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-13T06:49:05Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-13T06:49:05Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://f821e418cba5e832da8839268bcbbd9630bdf319a212d5382bfa2c4b0b73a268\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://f821e418cba5e832da8839268bcbbd9630bdf319a212d5382bfa2c4b0b73a268\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-13T06:49:07Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-13T06:49:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-13T06:49:03Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-13T06:50:13Z is after 2025-08-24T17:21:41Z" Dec 13 06:50:13 crc kubenswrapper[4971]: I1213 06:50:13.878413 4971 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-13T06:49:24Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-13T06:49:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-13T06:49:24Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4328cd6a8464ed468229bbfbc8265f969427916d72ebb92237dbe68efe70fa47\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-13T06:49:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://9d715ca82d9b5a594e41f2b98da6f4277da0e4e1dec114969f901c00f6d7553d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-13T06:49:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-13T06:50:13Z is after 2025-08-24T17:21:41Z" Dec 13 06:50:13 crc kubenswrapper[4971]: I1213 06:50:13.887207 4971 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-5rzh5" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"6bf4591d-2483-4132-9672-d0fd6c995968\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T06:49:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T06:49:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T06:49:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T06:49:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://50e2cb14e4ffb603acb06e2281615a99f06182d345c4e70f2446fa2077033bbb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-13T06:49:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dfxqn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-13T06:49:25Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-5rzh5\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-13T06:50:13Z is after 2025-08-24T17:21:41Z" Dec 13 06:50:13 crc kubenswrapper[4971]: I1213 06:50:13.896626 4971 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-z572s" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"b4146d3d-4034-49a9-bd37-164856255137\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T06:49:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T06:49:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T06:49:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T06:49:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://46fd4e1070835cfbf7f35d497eff7f047473fc4048ee07812925df5ebdf55342\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-13T06:49:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-j6l9b\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://dc79a0968f85723e6ad0dc269c8da1f7472adb1bb2b87edd017ee2295bfbf89c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-13T06:49:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-j6l9b\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-13T06:49:38Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-z572s\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-13T06:50:13Z is after 2025-08-24T17:21:41Z" Dec 13 06:50:13 crc kubenswrapper[4971]: I1213 06:50:13.902949 4971 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 13 06:50:13 crc kubenswrapper[4971]: I1213 06:50:13.903157 4971 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 13 06:50:13 crc kubenswrapper[4971]: I1213 06:50:13.903278 4971 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 13 06:50:13 crc kubenswrapper[4971]: I1213 06:50:13.903367 4971 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 13 06:50:13 crc kubenswrapper[4971]: I1213 06:50:13.903446 4971 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-13T06:50:13Z","lastTransitionTime":"2025-12-13T06:50:13Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 13 06:50:13 crc kubenswrapper[4971]: I1213 06:50:13.908102 4971 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3ad3268a-a982-428c-9c06-4aaeddd372d9\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T06:49:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T06:49:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T06:49:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T06:49:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T06:49:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3716afdf2bc693be80a09fc9f7727d334f435caf1d4c8ec4811b3a1f1652cdfe\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-13T06:49:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://30df3abf8c79db8af52c6a918f73e1ce410ddafcf341636424c267e1eeb195a7\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-13T06:49:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://ccdf65f6d2b1eb3647a03ba37c305c32e34addfdf79ff937c8dddaa3e5d5b5c6\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-13T06:49:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://770b73dd0c2b6f38744e2c1cd6aa186a69bb87d1cd015a603ad2e33be06a8f6b\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://b3c1bab3d53f975b2e24582ce2771bc26d600e0dd7c30339dcfd18a071162ad4\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-12-13T06:49:23Z\\\",\\\"message\\\":\\\"file observer\\\\nW1213 06:49:23.283403 1 builder.go:272] unable to get owner reference (falling back to namespace): pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\nI1213 06:49:23.284336 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1213 06:49:23.286978 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-981773671/tls.crt::/tmp/serving-cert-981773671/tls.key\\\\\\\"\\\\nI1213 06:49:23.775072 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1213 06:49:23.778615 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1213 06:49:23.778645 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1213 06:49:23.778680 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1213 06:49:23.778691 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1213 06:49:23.785058 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1213 06:49:23.786150 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1213 06:49:23.786196 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1213 06:49:23.786223 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1213 06:49:23.786246 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1213 06:49:23.786269 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1213 06:49:23.786291 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI1213 06:49:23.785608 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF1213 06:49:23.787563 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-13T06:49:08Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-13T06:49:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://9acd07066deed483bd7fa61ee45fda09539d110717a134a56827f871186f92fc\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-13T06:49:07Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://38c91cfbaecfdbd3e3b663422a91ab4294610afb6bbf9ebdb77db24e2b76c682\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://38c91cfbaecfdbd3e3b663422a91ab4294610afb6bbf9ebdb77db24e2b76c682\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-13T06:49:05Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-13T06:49:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-13T06:49:03Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-13T06:50:13Z is after 2025-08-24T17:21:41Z" Dec 13 06:50:13 crc kubenswrapper[4971]: I1213 06:50:13.918968 4971 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-13T06:49:23Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-13T06:49:23Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-13T06:50:13Z is after 2025-08-24T17:21:41Z" Dec 13 06:50:13 crc kubenswrapper[4971]: I1213 06:50:13.929037 4971 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-13T06:49:23Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-13T06:49:23Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-13T06:50:13Z is after 2025-08-24T17:21:41Z" Dec 13 06:50:13 crc kubenswrapper[4971]: I1213 06:50:13.946360 4971 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-zlcg5" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"736c6b1a-93dd-4f71-81b6-1c43fcf5f556\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T06:49:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T06:49:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T06:49:25Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T06:49:25Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://325223543b3c7db417f769533559a1ca2b23a723494e1885b0fd0480b5a6399f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-13T06:49:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vzmsr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://f22a4ec756119c4f791fb5105cf1f31df3e548b2b07b1f909a62802e9b6035ae\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-13T06:49:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vzmsr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://f6073140d9f86929b6d4bd2960ede0617a5c9620b269bbab043bae37e0d389f5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-13T06:49:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vzmsr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://693a2ab434c250cb140a866e8068b156df9191414fed1d63bee919a2528ea6e5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-13T06:49:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vzmsr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://219620f51474e60b3e1f880c0972eb612dffc4bad24462bbba2c093ae817d852\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-13T06:49:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vzmsr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://f137f4b441c2402093ab430a9689a46221c4fb4bc70d99fb992f62b0a989c465\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-13T06:49:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vzmsr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://0b454a87c6dd11bd8c3bb3553ee71de32ed6f185984e5057a20f2237b35c45d9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://0b454a87c6dd11bd8c3bb3553ee71de32ed6f185984e5057a20f2237b35c45d9\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-12-13T06:49:56Z\\\",\\\"message\\\":\\\" for pod on switch crc\\\\nI1213 06:49:56.566617 6645 metrics.go:553] Stopping metrics server at address \\\\\\\"127.0.0.1:29103\\\\\\\"\\\\nI1213 06:49:56.566625 6645 base_network_controller_pods.go:477] [default/openshift-network-console/networking-console-plugin-85b44fc459-gdk6g] creating logical port openshift-network-console_networking-console-plugin-85b44fc459-gdk6g for pod on switch crc\\\\nI1213 06:49:56.566569 6645 loadbalancer.go:304] Deleted 0 stale LBs for map[string]string{\\\\\\\"k8s.ovn.org/kind\\\\\\\":\\\\\\\"Service\\\\\\\", \\\\\\\"k8s.ovn.org/owner\\\\\\\":\\\\\\\"openshift-console-operator/metrics\\\\\\\"}\\\\nI1213 06:49:56.566666 6645 services_controller.go:360] Finished syncing service metrics on namespace openshift-console-operator for network=default : 1.804735ms\\\\nI1213 06:49:56.566669 6645 loadbalancer.go:304] Deleted 0 stale LBs for map[string]string{\\\\\\\"k8s.ovn.org/kind\\\\\\\":\\\\\\\"Service\\\\\\\", \\\\\\\"k8s.ovn.org/owner\\\\\\\":\\\\\\\"openshift-apiserver/check-endpoints\\\\\\\"}\\\\nF1213 06:49:56.566698 6645 ovnkube.go:137] failed to run ovnkube: [failed to start network controller: failed to start default network controller: unable to create admin network policy controller, err: could not add Event Handler for anpInformer during admin network policy controller initialization, handler {0x1fcc6e0 0x1fcc3c0 0x1fcc360} was not added to shared informer because it has stopped already, failed to start node network controller: failed \\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-13T06:49:55Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":2,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 20s restarting failed container=ovnkube-controller pod=ovnkube-node-zlcg5_openshift-ovn-kubernetes(736c6b1a-93dd-4f71-81b6-1c43fcf5f556)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vzmsr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://868049c6a3a4ebe4036d1d22e02ba6b76bced2ee7e269bb228accc4a0636678a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-13T06:49:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vzmsr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://451b22df6973ddc8fa2209a65d64d9e41b33eb231bb4adfab88cf0e5943bb246\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://451b22df6973ddc8fa2209a65d64d9e41b33eb231bb4adfab88cf0e5943bb246\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-13T06:49:26Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-13T06:49:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vzmsr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-13T06:49:25Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-zlcg5\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-13T06:50:13Z is after 2025-08-24T17:21:41Z" Dec 13 06:50:13 crc kubenswrapper[4971]: I1213 06:50:13.959587 4971 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-13T06:49:24Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-13T06:49:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-13T06:49:24Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d0b2ecf0d5b22f23217f42f259ae7ffb6359555b22c30eaebf667eaf254b3bab\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-13T06:49:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-13T06:50:13Z is after 2025-08-24T17:21:41Z" Dec 13 06:50:13 crc kubenswrapper[4971]: I1213 06:50:13.971049 4971 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-13T06:49:23Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-13T06:49:23Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-13T06:50:13Z is after 2025-08-24T17:21:41Z" Dec 13 06:50:13 crc kubenswrapper[4971]: I1213 06:50:13.981972 4971 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-13T06:49:26Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-13T06:49:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-13T06:49:26Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9c8a0d1db0e754462d805e82daf5f52ffd5457507a30456535ec59e096c9d092\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-13T06:49:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-13T06:50:13Z is after 2025-08-24T17:21:41Z" Dec 13 06:50:13 crc kubenswrapper[4971]: I1213 06:50:13.991996 4971 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-82xjz" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e0c4af12-edda-41d7-9b44-f87396174ef2\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T06:49:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T06:49:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T06:49:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T06:49:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://0f4577d9e37f860813e221f2e208177d8d383fd04f673ce357a0a405ea29d1e6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-13T06:49:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cnxsq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://7e0d2020e62f1a277cff7047e1907dc747ed393a1f6c983e1ee258bd156c16c7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-13T06:49:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cnxsq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-13T06:49:25Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-82xjz\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-13T06:50:13Z is after 2025-08-24T17:21:41Z" Dec 13 06:50:14 crc kubenswrapper[4971]: I1213 06:50:14.003843 4971 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-wwvpl" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"dddf744b-2122-49ad-a980-105c09636e0f\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T06:49:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T06:49:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T06:50:13Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T06:50:13Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4676d897c7fc668587faec563f3551a056557879cee3e865192d2bd28edf62fc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://4676d897c7fc668587faec563f3551a056557879cee3e865192d2bd28edf62fc\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-12-13T06:50:12Z\\\",\\\"message\\\":\\\"2025-12-13T06:49:27+00:00 [cnibincopy] Successfully copied files in /usr/src/multus-cni/rhel9/bin/ to /host/opt/cni/bin/upgrade_263f091d-d847-4fad-88d3-fbf82c65c231\\\\n2025-12-13T06:49:27+00:00 [cnibincopy] Successfully moved files in /host/opt/cni/bin/upgrade_263f091d-d847-4fad-88d3-fbf82c65c231 to /host/opt/cni/bin/\\\\n2025-12-13T06:49:27Z [verbose] multus-daemon started\\\\n2025-12-13T06:49:27Z [verbose] Readiness Indicator file check\\\\n2025-12-13T06:50:12Z [error] have you checked that your default network is ready? still waiting for readinessindicatorfile @ /host/run/multus/cni/net.d/10-ovn-kubernetes.conf. pollimmediate error: timed out waiting for the condition\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-13T06:49:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-czz9v\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-13T06:49:25Z\\\"}}\" for pod \"openshift-multus\"/\"multus-wwvpl\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-13T06:50:14Z is after 2025-08-24T17:21:41Z" Dec 13 06:50:14 crc kubenswrapper[4971]: I1213 06:50:14.005340 4971 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 13 06:50:14 crc kubenswrapper[4971]: I1213 06:50:14.005395 4971 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 13 06:50:14 crc kubenswrapper[4971]: I1213 06:50:14.005406 4971 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 13 06:50:14 crc kubenswrapper[4971]: I1213 06:50:14.005426 4971 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 13 06:50:14 crc kubenswrapper[4971]: I1213 06:50:14.005439 4971 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-13T06:50:14Z","lastTransitionTime":"2025-12-13T06:50:14Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 13 06:50:14 crc kubenswrapper[4971]: I1213 06:50:14.107001 4971 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 13 06:50:14 crc kubenswrapper[4971]: I1213 06:50:14.107046 4971 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 13 06:50:14 crc kubenswrapper[4971]: I1213 06:50:14.107059 4971 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 13 06:50:14 crc kubenswrapper[4971]: I1213 06:50:14.107076 4971 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 13 06:50:14 crc kubenswrapper[4971]: I1213 06:50:14.107086 4971 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-13T06:50:14Z","lastTransitionTime":"2025-12-13T06:50:14Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 13 06:50:14 crc kubenswrapper[4971]: I1213 06:50:14.168782 4971 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-multus_multus-wwvpl_dddf744b-2122-49ad-a980-105c09636e0f/kube-multus/0.log" Dec 13 06:50:14 crc kubenswrapper[4971]: I1213 06:50:14.168833 4971 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-wwvpl" event={"ID":"dddf744b-2122-49ad-a980-105c09636e0f","Type":"ContainerStarted","Data":"c8680aaa9e05095ac7e467c896d9e927bccff908925768ce5ed469ab89a72475"} Dec 13 06:50:14 crc kubenswrapper[4971]: I1213 06:50:14.183844 4971 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3ad3268a-a982-428c-9c06-4aaeddd372d9\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T06:49:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T06:49:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T06:49:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T06:49:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T06:49:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3716afdf2bc693be80a09fc9f7727d334f435caf1d4c8ec4811b3a1f1652cdfe\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-13T06:49:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://30df3abf8c79db8af52c6a918f73e1ce410ddafcf341636424c267e1eeb195a7\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-13T06:49:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://ccdf65f6d2b1eb3647a03ba37c305c32e34addfdf79ff937c8dddaa3e5d5b5c6\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-13T06:49:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://770b73dd0c2b6f38744e2c1cd6aa186a69bb87d1cd015a603ad2e33be06a8f6b\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://b3c1bab3d53f975b2e24582ce2771bc26d600e0dd7c30339dcfd18a071162ad4\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-12-13T06:49:23Z\\\",\\\"message\\\":\\\"file observer\\\\nW1213 06:49:23.283403 1 builder.go:272] unable to get owner reference (falling back to namespace): pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\nI1213 06:49:23.284336 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1213 06:49:23.286978 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-981773671/tls.crt::/tmp/serving-cert-981773671/tls.key\\\\\\\"\\\\nI1213 06:49:23.775072 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1213 06:49:23.778615 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1213 06:49:23.778645 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1213 06:49:23.778680 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1213 06:49:23.778691 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1213 06:49:23.785058 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1213 06:49:23.786150 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1213 06:49:23.786196 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1213 06:49:23.786223 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1213 06:49:23.786246 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1213 06:49:23.786269 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1213 06:49:23.786291 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI1213 06:49:23.785608 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF1213 06:49:23.787563 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-13T06:49:08Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-13T06:49:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://9acd07066deed483bd7fa61ee45fda09539d110717a134a56827f871186f92fc\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-13T06:49:07Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://38c91cfbaecfdbd3e3b663422a91ab4294610afb6bbf9ebdb77db24e2b76c682\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://38c91cfbaecfdbd3e3b663422a91ab4294610afb6bbf9ebdb77db24e2b76c682\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-13T06:49:05Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-13T06:49:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-13T06:49:03Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-13T06:50:14Z is after 2025-08-24T17:21:41Z" Dec 13 06:50:14 crc kubenswrapper[4971]: I1213 06:50:14.194647 4971 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-13T06:49:23Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-13T06:49:23Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-13T06:50:14Z is after 2025-08-24T17:21:41Z" Dec 13 06:50:14 crc kubenswrapper[4971]: I1213 06:50:14.205023 4971 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-13T06:49:23Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-13T06:49:23Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-13T06:50:14Z is after 2025-08-24T17:21:41Z" Dec 13 06:50:14 crc kubenswrapper[4971]: I1213 06:50:14.214397 4971 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 13 06:50:14 crc kubenswrapper[4971]: I1213 06:50:14.214443 4971 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 13 06:50:14 crc kubenswrapper[4971]: I1213 06:50:14.214455 4971 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 13 06:50:14 crc kubenswrapper[4971]: I1213 06:50:14.214475 4971 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 13 06:50:14 crc kubenswrapper[4971]: I1213 06:50:14.214489 4971 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-13T06:50:14Z","lastTransitionTime":"2025-12-13T06:50:14Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 13 06:50:14 crc kubenswrapper[4971]: I1213 06:50:14.222997 4971 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-13T06:49:24Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-13T06:49:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-13T06:49:24Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d0b2ecf0d5b22f23217f42f259ae7ffb6359555b22c30eaebf667eaf254b3bab\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-13T06:49:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-13T06:50:14Z is after 2025-08-24T17:21:41Z" Dec 13 06:50:14 crc kubenswrapper[4971]: I1213 06:50:14.235663 4971 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-13T06:49:23Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-13T06:49:23Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-13T06:50:14Z is after 2025-08-24T17:21:41Z" Dec 13 06:50:14 crc kubenswrapper[4971]: I1213 06:50:14.246450 4971 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-13T06:49:26Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-13T06:49:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-13T06:49:26Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9c8a0d1db0e754462d805e82daf5f52ffd5457507a30456535ec59e096c9d092\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-13T06:49:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-13T06:50:14Z is after 2025-08-24T17:21:41Z" Dec 13 06:50:14 crc kubenswrapper[4971]: I1213 06:50:14.255970 4971 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-82xjz" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e0c4af12-edda-41d7-9b44-f87396174ef2\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T06:49:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T06:49:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T06:49:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T06:49:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://0f4577d9e37f860813e221f2e208177d8d383fd04f673ce357a0a405ea29d1e6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-13T06:49:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cnxsq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://7e0d2020e62f1a277cff7047e1907dc747ed393a1f6c983e1ee258bd156c16c7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-13T06:49:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cnxsq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-13T06:49:25Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-82xjz\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-13T06:50:14Z is after 2025-08-24T17:21:41Z" Dec 13 06:50:14 crc kubenswrapper[4971]: I1213 06:50:14.267884 4971 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-wwvpl" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"dddf744b-2122-49ad-a980-105c09636e0f\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T06:49:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T06:49:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T06:50:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T06:50:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c8680aaa9e05095ac7e467c896d9e927bccff908925768ce5ed469ab89a72475\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://4676d897c7fc668587faec563f3551a056557879cee3e865192d2bd28edf62fc\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-12-13T06:50:12Z\\\",\\\"message\\\":\\\"2025-12-13T06:49:27+00:00 [cnibincopy] Successfully copied files in /usr/src/multus-cni/rhel9/bin/ to /host/opt/cni/bin/upgrade_263f091d-d847-4fad-88d3-fbf82c65c231\\\\n2025-12-13T06:49:27+00:00 [cnibincopy] Successfully moved files in /host/opt/cni/bin/upgrade_263f091d-d847-4fad-88d3-fbf82c65c231 to /host/opt/cni/bin/\\\\n2025-12-13T06:49:27Z [verbose] multus-daemon started\\\\n2025-12-13T06:49:27Z [verbose] Readiness Indicator file check\\\\n2025-12-13T06:50:12Z [error] have you checked that your default network is ready? still waiting for readinessindicatorfile @ /host/run/multus/cni/net.d/10-ovn-kubernetes.conf. pollimmediate error: timed out waiting for the condition\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-13T06:49:26Z\\\"}},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-13T06:50:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-czz9v\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-13T06:49:25Z\\\"}}\" for pod \"openshift-multus\"/\"multus-wwvpl\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-13T06:50:14Z is after 2025-08-24T17:21:41Z" Dec 13 06:50:14 crc kubenswrapper[4971]: I1213 06:50:14.287468 4971 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-zlcg5" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"736c6b1a-93dd-4f71-81b6-1c43fcf5f556\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T06:49:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T06:49:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T06:49:25Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T06:49:25Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://325223543b3c7db417f769533559a1ca2b23a723494e1885b0fd0480b5a6399f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-13T06:49:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vzmsr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://f22a4ec756119c4f791fb5105cf1f31df3e548b2b07b1f909a62802e9b6035ae\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-13T06:49:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vzmsr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://f6073140d9f86929b6d4bd2960ede0617a5c9620b269bbab043bae37e0d389f5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-13T06:49:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vzmsr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://693a2ab434c250cb140a866e8068b156df9191414fed1d63bee919a2528ea6e5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-13T06:49:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vzmsr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://219620f51474e60b3e1f880c0972eb612dffc4bad24462bbba2c093ae817d852\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-13T06:49:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vzmsr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://f137f4b441c2402093ab430a9689a46221c4fb4bc70d99fb992f62b0a989c465\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-13T06:49:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vzmsr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://0b454a87c6dd11bd8c3bb3553ee71de32ed6f185984e5057a20f2237b35c45d9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://0b454a87c6dd11bd8c3bb3553ee71de32ed6f185984e5057a20f2237b35c45d9\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-12-13T06:49:56Z\\\",\\\"message\\\":\\\" for pod on switch crc\\\\nI1213 06:49:56.566617 6645 metrics.go:553] Stopping metrics server at address \\\\\\\"127.0.0.1:29103\\\\\\\"\\\\nI1213 06:49:56.566625 6645 base_network_controller_pods.go:477] [default/openshift-network-console/networking-console-plugin-85b44fc459-gdk6g] creating logical port openshift-network-console_networking-console-plugin-85b44fc459-gdk6g for pod on switch crc\\\\nI1213 06:49:56.566569 6645 loadbalancer.go:304] Deleted 0 stale LBs for map[string]string{\\\\\\\"k8s.ovn.org/kind\\\\\\\":\\\\\\\"Service\\\\\\\", \\\\\\\"k8s.ovn.org/owner\\\\\\\":\\\\\\\"openshift-console-operator/metrics\\\\\\\"}\\\\nI1213 06:49:56.566666 6645 services_controller.go:360] Finished syncing service metrics on namespace openshift-console-operator for network=default : 1.804735ms\\\\nI1213 06:49:56.566669 6645 loadbalancer.go:304] Deleted 0 stale LBs for map[string]string{\\\\\\\"k8s.ovn.org/kind\\\\\\\":\\\\\\\"Service\\\\\\\", \\\\\\\"k8s.ovn.org/owner\\\\\\\":\\\\\\\"openshift-apiserver/check-endpoints\\\\\\\"}\\\\nF1213 06:49:56.566698 6645 ovnkube.go:137] failed to run ovnkube: [failed to start network controller: failed to start default network controller: unable to create admin network policy controller, err: could not add Event Handler for anpInformer during admin network policy controller initialization, handler {0x1fcc6e0 0x1fcc3c0 0x1fcc360} was not added to shared informer because it has stopped already, failed to start node network controller: failed \\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-13T06:49:55Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":2,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 20s restarting failed container=ovnkube-controller pod=ovnkube-node-zlcg5_openshift-ovn-kubernetes(736c6b1a-93dd-4f71-81b6-1c43fcf5f556)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vzmsr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://868049c6a3a4ebe4036d1d22e02ba6b76bced2ee7e269bb228accc4a0636678a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-13T06:49:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vzmsr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://451b22df6973ddc8fa2209a65d64d9e41b33eb231bb4adfab88cf0e5943bb246\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://451b22df6973ddc8fa2209a65d64d9e41b33eb231bb4adfab88cf0e5943bb246\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-13T06:49:26Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-13T06:49:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vzmsr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-13T06:49:25Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-zlcg5\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-13T06:50:14Z is after 2025-08-24T17:21:41Z" Dec 13 06:50:14 crc kubenswrapper[4971]: I1213 06:50:14.300418 4971 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"f9d31b5a-b770-4a15-95c8-35b49677c608\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T06:49:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T06:49:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T06:49:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T06:49:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T06:49:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c9baa042f9dada954faa57b5454cd6b0a12a297e56ddc192c52cff50e42a16fa\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-13T06:49:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://c788de86754f0bf37ee184672fc1dcb388ccebc246d4f15be4b5da661f725f8f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-13T06:49:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://760817511a9b4f96a68dc38d26631250eaab72e1dd8f123e653039b6b4e91f84\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-13T06:49:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b859be0c9feac856e9855c6a038b02b49b32fade97873c46d0bec0a32d8b5147\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-13T06:49:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-13T06:49:03Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-13T06:50:14Z is after 2025-08-24T17:21:41Z" Dec 13 06:50:14 crc kubenswrapper[4971]: I1213 06:50:14.311170 4971 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"8ede2992-9576-4b23-83c2-24ac3b1d3caf\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T06:49:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T06:49:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T06:49:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T06:49:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T06:49:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9e61d25add1723ac286794abfe886358e9d64455c25379f7e3057fc3dd812418\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-13T06:49:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://7fb51d8435b4e3ce0411cae1d6c2824c552f0a33d79de01f969274516b39a5b3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-13T06:49:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://1c35f5e8d8fab6d77b9e2e0dedb195440de5d6af345cd5865abd32caa7fafd60\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-13T06:49:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://38806c430185715371ab9bf30a1af9a9d5d492abd9bb84acbdccc694f56e3a8a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"wait-for-host-port\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://38806c430185715371ab9bf30a1af9a9d5d492abd9bb84acbdccc694f56e3a8a\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-13T06:49:04Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-13T06:49:04Z\\\"}}}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-13T06:49:03Z\\\"}}\" for pod \"openshift-kube-scheduler\"/\"openshift-kube-scheduler-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-13T06:50:14Z is after 2025-08-24T17:21:41Z" Dec 13 06:50:14 crc kubenswrapper[4971]: I1213 06:50:14.316456 4971 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 13 06:50:14 crc kubenswrapper[4971]: I1213 06:50:14.316504 4971 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 13 06:50:14 crc kubenswrapper[4971]: I1213 06:50:14.316541 4971 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 13 06:50:14 crc kubenswrapper[4971]: I1213 06:50:14.316560 4971 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 13 06:50:14 crc kubenswrapper[4971]: I1213 06:50:14.316573 4971 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-13T06:50:14Z","lastTransitionTime":"2025-12-13T06:50:14Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 13 06:50:14 crc kubenswrapper[4971]: I1213 06:50:14.324035 4971 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-ncchz" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"de914d21-a688-401e-a66e-861beb6336d9\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T06:49:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T06:49:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T06:49:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T06:49:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://244ffe3dc9d8cf799aff4c4d795bf41554c1e6bb6667fbb0c7df322fb5f96c68\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-13T06:49:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lqmjj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://0b7111af6c5a198d02bb02858837cfb4f3b20828d8d206cd18aa0ffa102d0cd4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://0b7111af6c5a198d02bb02858837cfb4f3b20828d8d206cd18aa0ffa102d0cd4\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-13T06:49:26Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-13T06:49:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lqmjj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b36777fe40e74ae1cecb7b73e204139db5279ba51cef4f1e8c4a4c491d8be403\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://b36777fe40e74ae1cecb7b73e204139db5279ba51cef4f1e8c4a4c491d8be403\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-13T06:49:28Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-13T06:49:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lqmjj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://0597a06cc663e416142311a16941d98297577693a1fe27201ea42c27496cace2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://0597a06cc663e416142311a16941d98297577693a1fe27201ea42c27496cace2\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-13T06:49:29Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-13T06:49:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lqmjj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://3f52c2e22a75c010993ca5809a3a9d1304758d4cb6e83691ef1af15a4b514ac4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://3f52c2e22a75c010993ca5809a3a9d1304758d4cb6e83691ef1af15a4b514ac4\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-13T06:49:30Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-13T06:49:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lqmjj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://9e21ca708c4c4c1228b417cece51299e0792729b7c78884bd18b2dda8fe8355c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://9e21ca708c4c4c1228b417cece51299e0792729b7c78884bd18b2dda8fe8355c\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-13T06:49:31Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-13T06:49:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lqmjj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://74552159f12c87cdda09147692b400ccac0d4e43c1ed72fe8462c4249ad69c7b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://74552159f12c87cdda09147692b400ccac0d4e43c1ed72fe8462c4249ad69c7b\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-13T06:49:32Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-13T06:49:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lqmjj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-13T06:49:25Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-ncchz\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-13T06:50:14Z is after 2025-08-24T17:21:41Z" Dec 13 06:50:14 crc kubenswrapper[4971]: I1213 06:50:14.333033 4971 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-gkcd8" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"13f5c327-f95e-42b4-8783-218500c17694\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T06:49:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T06:49:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T06:49:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T06:49:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://77e846e36f1cb40d03471dc03de11b6c3da448baa407aad38093fae9a801cf92\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-13T06:49:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gvlhn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-13T06:49:27Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-gkcd8\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-13T06:50:14Z is after 2025-08-24T17:21:41Z" Dec 13 06:50:14 crc kubenswrapper[4971]: I1213 06:50:14.342234 4971 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-x98k7" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"c7185dec-3240-42e8-9d90-6182a5e8b65c\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T06:49:39Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T06:49:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T06:49:39Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T06:49:39Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bj2gn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bj2gn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-13T06:49:39Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-x98k7\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-13T06:50:14Z is after 2025-08-24T17:21:41Z" Dec 13 06:50:14 crc kubenswrapper[4971]: I1213 06:50:14.350364 4971 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5ef36745-87bb-4be8-b941-aa8f865a5c83\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T06:49:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T06:49:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T06:49:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T06:49:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T06:49:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://35b737b0d2c5c5503ba2819f34d2de0599792a77beb1fbeb561d44efd1f212c5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-crio\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-13T06:49:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kube\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"var-lib-kubelet\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f0d864149bfcbb6227e6a6fbde23fff650d688850978eae0caf04515eb21cd85\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://f0d864149bfcbb6227e6a6fbde23fff650d688850978eae0caf04515eb21cd85\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-13T06:49:04Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-13T06:49:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var\\\",\\\"name\\\":\\\"var-lib-kubelet\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-13T06:49:03Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"kube-rbac-proxy-crio-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-13T06:50:14Z is after 2025-08-24T17:21:41Z" Dec 13 06:50:14 crc kubenswrapper[4971]: I1213 06:50:14.366956 4971 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"1ed9f82e-fe5c-4b76-a619-62e5abd88d3b\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T06:49:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T06:49:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T06:49:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T06:49:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T06:49:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e457329a78a2d24337f994e8a0a1f75a2a667244aa75eb84613d1e7138364aac\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-13T06:49:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://eb014de8140deef097d9bd94f3b9fa4ab3558bc4449bef84b83bf62bfabf5836\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-13T06:49:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://2c438cc74d555f7c9db4569d1c56b78b1088ed8105697d2c0f31e6c5136f3dd0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-13T06:49:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://aa6fa6e4340dcd066252df3c56d79a3a39ed531d5eea8d1d5941d6952368d3af\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-13T06:49:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://c4471b98854c52d5cd084d0ba3e64e5b08e9518ec3c372cce2c875a440522457\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-13T06:49:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7d2746335c98dae3a0f43024d34cb62ebcb42a01c8f9ce18d52b29d4c45598e5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://7d2746335c98dae3a0f43024d34cb62ebcb42a01c8f9ce18d52b29d4c45598e5\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-13T06:49:04Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-13T06:49:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://bc0e77d10a738128bea82122496bc42235fc1688f8907756d93e61b651902b57\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://bc0e77d10a738128bea82122496bc42235fc1688f8907756d93e61b651902b57\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-13T06:49:05Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-13T06:49:05Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://f821e418cba5e832da8839268bcbbd9630bdf319a212d5382bfa2c4b0b73a268\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://f821e418cba5e832da8839268bcbbd9630bdf319a212d5382bfa2c4b0b73a268\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-13T06:49:07Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-13T06:49:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-13T06:49:03Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-13T06:50:14Z is after 2025-08-24T17:21:41Z" Dec 13 06:50:14 crc kubenswrapper[4971]: I1213 06:50:14.380216 4971 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-13T06:49:24Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-13T06:49:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-13T06:49:24Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4328cd6a8464ed468229bbfbc8265f969427916d72ebb92237dbe68efe70fa47\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-13T06:49:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://9d715ca82d9b5a594e41f2b98da6f4277da0e4e1dec114969f901c00f6d7553d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-13T06:49:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-13T06:50:14Z is after 2025-08-24T17:21:41Z" Dec 13 06:50:14 crc kubenswrapper[4971]: I1213 06:50:14.390699 4971 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-5rzh5" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"6bf4591d-2483-4132-9672-d0fd6c995968\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T06:49:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T06:49:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T06:49:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T06:49:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://50e2cb14e4ffb603acb06e2281615a99f06182d345c4e70f2446fa2077033bbb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-13T06:49:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dfxqn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-13T06:49:25Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-5rzh5\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-13T06:50:14Z is after 2025-08-24T17:21:41Z" Dec 13 06:50:14 crc kubenswrapper[4971]: I1213 06:50:14.401784 4971 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-z572s" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"b4146d3d-4034-49a9-bd37-164856255137\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T06:49:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T06:49:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T06:49:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T06:49:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://46fd4e1070835cfbf7f35d497eff7f047473fc4048ee07812925df5ebdf55342\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-13T06:49:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-j6l9b\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://dc79a0968f85723e6ad0dc269c8da1f7472adb1bb2b87edd017ee2295bfbf89c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-13T06:49:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-j6l9b\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-13T06:49:38Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-z572s\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-13T06:50:14Z is after 2025-08-24T17:21:41Z" Dec 13 06:50:14 crc kubenswrapper[4971]: I1213 06:50:14.418835 4971 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 13 06:50:14 crc kubenswrapper[4971]: I1213 06:50:14.418861 4971 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 13 06:50:14 crc kubenswrapper[4971]: I1213 06:50:14.418869 4971 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 13 06:50:14 crc kubenswrapper[4971]: I1213 06:50:14.418883 4971 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 13 06:50:14 crc kubenswrapper[4971]: I1213 06:50:14.418891 4971 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-13T06:50:14Z","lastTransitionTime":"2025-12-13T06:50:14Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 13 06:50:14 crc kubenswrapper[4971]: I1213 06:50:14.521352 4971 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 13 06:50:14 crc kubenswrapper[4971]: I1213 06:50:14.521396 4971 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 13 06:50:14 crc kubenswrapper[4971]: I1213 06:50:14.521408 4971 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 13 06:50:14 crc kubenswrapper[4971]: I1213 06:50:14.521426 4971 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 13 06:50:14 crc kubenswrapper[4971]: I1213 06:50:14.521438 4971 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-13T06:50:14Z","lastTransitionTime":"2025-12-13T06:50:14Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 13 06:50:14 crc kubenswrapper[4971]: I1213 06:50:14.623690 4971 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 13 06:50:14 crc kubenswrapper[4971]: I1213 06:50:14.623731 4971 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 13 06:50:14 crc kubenswrapper[4971]: I1213 06:50:14.623743 4971 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 13 06:50:14 crc kubenswrapper[4971]: I1213 06:50:14.623760 4971 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 13 06:50:14 crc kubenswrapper[4971]: I1213 06:50:14.623772 4971 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-13T06:50:14Z","lastTransitionTime":"2025-12-13T06:50:14Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 13 06:50:14 crc kubenswrapper[4971]: I1213 06:50:14.726490 4971 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 13 06:50:14 crc kubenswrapper[4971]: I1213 06:50:14.726569 4971 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 13 06:50:14 crc kubenswrapper[4971]: I1213 06:50:14.726584 4971 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 13 06:50:14 crc kubenswrapper[4971]: I1213 06:50:14.726601 4971 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 13 06:50:14 crc kubenswrapper[4971]: I1213 06:50:14.726611 4971 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-13T06:50:14Z","lastTransitionTime":"2025-12-13T06:50:14Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 13 06:50:14 crc kubenswrapper[4971]: I1213 06:50:14.829332 4971 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 13 06:50:14 crc kubenswrapper[4971]: I1213 06:50:14.829374 4971 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 13 06:50:14 crc kubenswrapper[4971]: I1213 06:50:14.829385 4971 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 13 06:50:14 crc kubenswrapper[4971]: I1213 06:50:14.829403 4971 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 13 06:50:14 crc kubenswrapper[4971]: I1213 06:50:14.829415 4971 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-13T06:50:14Z","lastTransitionTime":"2025-12-13T06:50:14Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 13 06:50:14 crc kubenswrapper[4971]: I1213 06:50:14.931694 4971 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 13 06:50:14 crc kubenswrapper[4971]: I1213 06:50:14.931768 4971 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 13 06:50:14 crc kubenswrapper[4971]: I1213 06:50:14.931779 4971 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 13 06:50:14 crc kubenswrapper[4971]: I1213 06:50:14.931796 4971 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 13 06:50:14 crc kubenswrapper[4971]: I1213 06:50:14.931809 4971 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-13T06:50:14Z","lastTransitionTime":"2025-12-13T06:50:14Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 13 06:50:15 crc kubenswrapper[4971]: I1213 06:50:15.034141 4971 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 13 06:50:15 crc kubenswrapper[4971]: I1213 06:50:15.034187 4971 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 13 06:50:15 crc kubenswrapper[4971]: I1213 06:50:15.034196 4971 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 13 06:50:15 crc kubenswrapper[4971]: I1213 06:50:15.034212 4971 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 13 06:50:15 crc kubenswrapper[4971]: I1213 06:50:15.034222 4971 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-13T06:50:15Z","lastTransitionTime":"2025-12-13T06:50:15Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 13 06:50:15 crc kubenswrapper[4971]: I1213 06:50:15.136841 4971 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 13 06:50:15 crc kubenswrapper[4971]: I1213 06:50:15.136900 4971 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 13 06:50:15 crc kubenswrapper[4971]: I1213 06:50:15.136916 4971 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 13 06:50:15 crc kubenswrapper[4971]: I1213 06:50:15.136940 4971 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 13 06:50:15 crc kubenswrapper[4971]: I1213 06:50:15.136956 4971 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-13T06:50:15Z","lastTransitionTime":"2025-12-13T06:50:15Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 13 06:50:15 crc kubenswrapper[4971]: I1213 06:50:15.238608 4971 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 13 06:50:15 crc kubenswrapper[4971]: I1213 06:50:15.238652 4971 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 13 06:50:15 crc kubenswrapper[4971]: I1213 06:50:15.238664 4971 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 13 06:50:15 crc kubenswrapper[4971]: I1213 06:50:15.238679 4971 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 13 06:50:15 crc kubenswrapper[4971]: I1213 06:50:15.238689 4971 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-13T06:50:15Z","lastTransitionTime":"2025-12-13T06:50:15Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 13 06:50:15 crc kubenswrapper[4971]: I1213 06:50:15.340592 4971 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 13 06:50:15 crc kubenswrapper[4971]: I1213 06:50:15.340631 4971 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 13 06:50:15 crc kubenswrapper[4971]: I1213 06:50:15.340641 4971 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 13 06:50:15 crc kubenswrapper[4971]: I1213 06:50:15.340658 4971 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 13 06:50:15 crc kubenswrapper[4971]: I1213 06:50:15.340667 4971 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-13T06:50:15Z","lastTransitionTime":"2025-12-13T06:50:15Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 13 06:50:15 crc kubenswrapper[4971]: I1213 06:50:15.442598 4971 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 13 06:50:15 crc kubenswrapper[4971]: I1213 06:50:15.442635 4971 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 13 06:50:15 crc kubenswrapper[4971]: I1213 06:50:15.442646 4971 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 13 06:50:15 crc kubenswrapper[4971]: I1213 06:50:15.442663 4971 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 13 06:50:15 crc kubenswrapper[4971]: I1213 06:50:15.442672 4971 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-13T06:50:15Z","lastTransitionTime":"2025-12-13T06:50:15Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 13 06:50:15 crc kubenswrapper[4971]: I1213 06:50:15.545290 4971 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 13 06:50:15 crc kubenswrapper[4971]: I1213 06:50:15.545342 4971 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 13 06:50:15 crc kubenswrapper[4971]: I1213 06:50:15.545358 4971 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 13 06:50:15 crc kubenswrapper[4971]: I1213 06:50:15.545380 4971 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 13 06:50:15 crc kubenswrapper[4971]: I1213 06:50:15.545398 4971 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-13T06:50:15Z","lastTransitionTime":"2025-12-13T06:50:15Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 13 06:50:15 crc kubenswrapper[4971]: I1213 06:50:15.647875 4971 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 13 06:50:15 crc kubenswrapper[4971]: I1213 06:50:15.647915 4971 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 13 06:50:15 crc kubenswrapper[4971]: I1213 06:50:15.647925 4971 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 13 06:50:15 crc kubenswrapper[4971]: I1213 06:50:15.647937 4971 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 13 06:50:15 crc kubenswrapper[4971]: I1213 06:50:15.647946 4971 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-13T06:50:15Z","lastTransitionTime":"2025-12-13T06:50:15Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 13 06:50:15 crc kubenswrapper[4971]: I1213 06:50:15.749670 4971 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 13 06:50:15 crc kubenswrapper[4971]: I1213 06:50:15.749719 4971 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 13 06:50:15 crc kubenswrapper[4971]: I1213 06:50:15.749731 4971 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 13 06:50:15 crc kubenswrapper[4971]: I1213 06:50:15.749770 4971 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 13 06:50:15 crc kubenswrapper[4971]: I1213 06:50:15.749782 4971 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-13T06:50:15Z","lastTransitionTime":"2025-12-13T06:50:15Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 13 06:50:15 crc kubenswrapper[4971]: I1213 06:50:15.768578 4971 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-x98k7" Dec 13 06:50:15 crc kubenswrapper[4971]: I1213 06:50:15.768639 4971 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 13 06:50:15 crc kubenswrapper[4971]: I1213 06:50:15.768578 4971 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 13 06:50:15 crc kubenswrapper[4971]: E1213 06:50:15.768711 4971 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-x98k7" podUID="c7185dec-3240-42e8-9d90-6182a5e8b65c" Dec 13 06:50:15 crc kubenswrapper[4971]: I1213 06:50:15.768656 4971 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 13 06:50:15 crc kubenswrapper[4971]: E1213 06:50:15.768808 4971 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 13 06:50:15 crc kubenswrapper[4971]: E1213 06:50:15.768873 4971 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 13 06:50:15 crc kubenswrapper[4971]: E1213 06:50:15.768962 4971 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 13 06:50:15 crc kubenswrapper[4971]: I1213 06:50:15.851497 4971 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 13 06:50:15 crc kubenswrapper[4971]: I1213 06:50:15.851545 4971 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 13 06:50:15 crc kubenswrapper[4971]: I1213 06:50:15.851556 4971 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 13 06:50:15 crc kubenswrapper[4971]: I1213 06:50:15.851571 4971 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 13 06:50:15 crc kubenswrapper[4971]: I1213 06:50:15.851580 4971 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-13T06:50:15Z","lastTransitionTime":"2025-12-13T06:50:15Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 13 06:50:15 crc kubenswrapper[4971]: I1213 06:50:15.954223 4971 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 13 06:50:15 crc kubenswrapper[4971]: I1213 06:50:15.954266 4971 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 13 06:50:15 crc kubenswrapper[4971]: I1213 06:50:15.954276 4971 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 13 06:50:15 crc kubenswrapper[4971]: I1213 06:50:15.954291 4971 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 13 06:50:15 crc kubenswrapper[4971]: I1213 06:50:15.954301 4971 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-13T06:50:15Z","lastTransitionTime":"2025-12-13T06:50:15Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 13 06:50:16 crc kubenswrapper[4971]: I1213 06:50:16.057073 4971 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 13 06:50:16 crc kubenswrapper[4971]: I1213 06:50:16.057122 4971 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 13 06:50:16 crc kubenswrapper[4971]: I1213 06:50:16.057139 4971 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 13 06:50:16 crc kubenswrapper[4971]: I1213 06:50:16.057162 4971 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 13 06:50:16 crc kubenswrapper[4971]: I1213 06:50:16.057179 4971 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-13T06:50:16Z","lastTransitionTime":"2025-12-13T06:50:16Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 13 06:50:16 crc kubenswrapper[4971]: I1213 06:50:16.160034 4971 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 13 06:50:16 crc kubenswrapper[4971]: I1213 06:50:16.160128 4971 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 13 06:50:16 crc kubenswrapper[4971]: I1213 06:50:16.160145 4971 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 13 06:50:16 crc kubenswrapper[4971]: I1213 06:50:16.160163 4971 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 13 06:50:16 crc kubenswrapper[4971]: I1213 06:50:16.160175 4971 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-13T06:50:16Z","lastTransitionTime":"2025-12-13T06:50:16Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 13 06:50:16 crc kubenswrapper[4971]: I1213 06:50:16.262987 4971 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 13 06:50:16 crc kubenswrapper[4971]: I1213 06:50:16.263037 4971 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 13 06:50:16 crc kubenswrapper[4971]: I1213 06:50:16.263048 4971 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 13 06:50:16 crc kubenswrapper[4971]: I1213 06:50:16.263067 4971 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 13 06:50:16 crc kubenswrapper[4971]: I1213 06:50:16.263082 4971 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-13T06:50:16Z","lastTransitionTime":"2025-12-13T06:50:16Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 13 06:50:16 crc kubenswrapper[4971]: I1213 06:50:16.364965 4971 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 13 06:50:16 crc kubenswrapper[4971]: I1213 06:50:16.365004 4971 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 13 06:50:16 crc kubenswrapper[4971]: I1213 06:50:16.365013 4971 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 13 06:50:16 crc kubenswrapper[4971]: I1213 06:50:16.365029 4971 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 13 06:50:16 crc kubenswrapper[4971]: I1213 06:50:16.365038 4971 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-13T06:50:16Z","lastTransitionTime":"2025-12-13T06:50:16Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 13 06:50:16 crc kubenswrapper[4971]: I1213 06:50:16.466967 4971 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 13 06:50:16 crc kubenswrapper[4971]: I1213 06:50:16.467016 4971 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 13 06:50:16 crc kubenswrapper[4971]: I1213 06:50:16.467025 4971 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 13 06:50:16 crc kubenswrapper[4971]: I1213 06:50:16.467041 4971 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 13 06:50:16 crc kubenswrapper[4971]: I1213 06:50:16.467051 4971 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-13T06:50:16Z","lastTransitionTime":"2025-12-13T06:50:16Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 13 06:50:16 crc kubenswrapper[4971]: I1213 06:50:16.569779 4971 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 13 06:50:16 crc kubenswrapper[4971]: I1213 06:50:16.569830 4971 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 13 06:50:16 crc kubenswrapper[4971]: I1213 06:50:16.569843 4971 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 13 06:50:16 crc kubenswrapper[4971]: I1213 06:50:16.569861 4971 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 13 06:50:16 crc kubenswrapper[4971]: I1213 06:50:16.569872 4971 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-13T06:50:16Z","lastTransitionTime":"2025-12-13T06:50:16Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 13 06:50:16 crc kubenswrapper[4971]: I1213 06:50:16.672540 4971 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 13 06:50:16 crc kubenswrapper[4971]: I1213 06:50:16.672583 4971 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 13 06:50:16 crc kubenswrapper[4971]: I1213 06:50:16.672593 4971 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 13 06:50:16 crc kubenswrapper[4971]: I1213 06:50:16.672609 4971 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 13 06:50:16 crc kubenswrapper[4971]: I1213 06:50:16.672618 4971 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-13T06:50:16Z","lastTransitionTime":"2025-12-13T06:50:16Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 13 06:50:16 crc kubenswrapper[4971]: I1213 06:50:16.774643 4971 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 13 06:50:16 crc kubenswrapper[4971]: I1213 06:50:16.774682 4971 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 13 06:50:16 crc kubenswrapper[4971]: I1213 06:50:16.774691 4971 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 13 06:50:16 crc kubenswrapper[4971]: I1213 06:50:16.774708 4971 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 13 06:50:16 crc kubenswrapper[4971]: I1213 06:50:16.774718 4971 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-13T06:50:16Z","lastTransitionTime":"2025-12-13T06:50:16Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 13 06:50:16 crc kubenswrapper[4971]: I1213 06:50:16.877014 4971 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 13 06:50:16 crc kubenswrapper[4971]: I1213 06:50:16.877053 4971 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 13 06:50:16 crc kubenswrapper[4971]: I1213 06:50:16.877069 4971 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 13 06:50:16 crc kubenswrapper[4971]: I1213 06:50:16.877085 4971 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 13 06:50:16 crc kubenswrapper[4971]: I1213 06:50:16.877096 4971 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-13T06:50:16Z","lastTransitionTime":"2025-12-13T06:50:16Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 13 06:50:16 crc kubenswrapper[4971]: I1213 06:50:16.979548 4971 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 13 06:50:16 crc kubenswrapper[4971]: I1213 06:50:16.979585 4971 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 13 06:50:16 crc kubenswrapper[4971]: I1213 06:50:16.979594 4971 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 13 06:50:16 crc kubenswrapper[4971]: I1213 06:50:16.979608 4971 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 13 06:50:16 crc kubenswrapper[4971]: I1213 06:50:16.979619 4971 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-13T06:50:16Z","lastTransitionTime":"2025-12-13T06:50:16Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 13 06:50:17 crc kubenswrapper[4971]: I1213 06:50:17.081435 4971 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 13 06:50:17 crc kubenswrapper[4971]: I1213 06:50:17.081467 4971 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 13 06:50:17 crc kubenswrapper[4971]: I1213 06:50:17.081475 4971 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 13 06:50:17 crc kubenswrapper[4971]: I1213 06:50:17.081489 4971 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 13 06:50:17 crc kubenswrapper[4971]: I1213 06:50:17.081498 4971 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-13T06:50:17Z","lastTransitionTime":"2025-12-13T06:50:17Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 13 06:50:17 crc kubenswrapper[4971]: I1213 06:50:17.183426 4971 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 13 06:50:17 crc kubenswrapper[4971]: I1213 06:50:17.183479 4971 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 13 06:50:17 crc kubenswrapper[4971]: I1213 06:50:17.183487 4971 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 13 06:50:17 crc kubenswrapper[4971]: I1213 06:50:17.183503 4971 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 13 06:50:17 crc kubenswrapper[4971]: I1213 06:50:17.183529 4971 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-13T06:50:17Z","lastTransitionTime":"2025-12-13T06:50:17Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 13 06:50:17 crc kubenswrapper[4971]: I1213 06:50:17.286080 4971 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 13 06:50:17 crc kubenswrapper[4971]: I1213 06:50:17.286129 4971 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 13 06:50:17 crc kubenswrapper[4971]: I1213 06:50:17.286139 4971 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 13 06:50:17 crc kubenswrapper[4971]: I1213 06:50:17.286187 4971 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 13 06:50:17 crc kubenswrapper[4971]: I1213 06:50:17.286197 4971 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-13T06:50:17Z","lastTransitionTime":"2025-12-13T06:50:17Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 13 06:50:17 crc kubenswrapper[4971]: I1213 06:50:17.388303 4971 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 13 06:50:17 crc kubenswrapper[4971]: I1213 06:50:17.388366 4971 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 13 06:50:17 crc kubenswrapper[4971]: I1213 06:50:17.388379 4971 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 13 06:50:17 crc kubenswrapper[4971]: I1213 06:50:17.388425 4971 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 13 06:50:17 crc kubenswrapper[4971]: I1213 06:50:17.388440 4971 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-13T06:50:17Z","lastTransitionTime":"2025-12-13T06:50:17Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 13 06:50:17 crc kubenswrapper[4971]: I1213 06:50:17.490769 4971 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 13 06:50:17 crc kubenswrapper[4971]: I1213 06:50:17.490814 4971 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 13 06:50:17 crc kubenswrapper[4971]: I1213 06:50:17.490829 4971 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 13 06:50:17 crc kubenswrapper[4971]: I1213 06:50:17.490847 4971 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 13 06:50:17 crc kubenswrapper[4971]: I1213 06:50:17.490859 4971 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-13T06:50:17Z","lastTransitionTime":"2025-12-13T06:50:17Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 13 06:50:17 crc kubenswrapper[4971]: I1213 06:50:17.593141 4971 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 13 06:50:17 crc kubenswrapper[4971]: I1213 06:50:17.593187 4971 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 13 06:50:17 crc kubenswrapper[4971]: I1213 06:50:17.593198 4971 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 13 06:50:17 crc kubenswrapper[4971]: I1213 06:50:17.593213 4971 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 13 06:50:17 crc kubenswrapper[4971]: I1213 06:50:17.593222 4971 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-13T06:50:17Z","lastTransitionTime":"2025-12-13T06:50:17Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 13 06:50:17 crc kubenswrapper[4971]: I1213 06:50:17.695925 4971 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 13 06:50:17 crc kubenswrapper[4971]: I1213 06:50:17.696046 4971 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 13 06:50:17 crc kubenswrapper[4971]: I1213 06:50:17.696063 4971 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 13 06:50:17 crc kubenswrapper[4971]: I1213 06:50:17.696091 4971 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 13 06:50:17 crc kubenswrapper[4971]: I1213 06:50:17.696109 4971 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-13T06:50:17Z","lastTransitionTime":"2025-12-13T06:50:17Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 13 06:50:17 crc kubenswrapper[4971]: I1213 06:50:17.768013 4971 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-x98k7" Dec 13 06:50:17 crc kubenswrapper[4971]: I1213 06:50:17.768048 4971 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 13 06:50:17 crc kubenswrapper[4971]: I1213 06:50:17.768080 4971 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 13 06:50:17 crc kubenswrapper[4971]: I1213 06:50:17.768014 4971 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 13 06:50:17 crc kubenswrapper[4971]: E1213 06:50:17.768146 4971 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-x98k7" podUID="c7185dec-3240-42e8-9d90-6182a5e8b65c" Dec 13 06:50:17 crc kubenswrapper[4971]: E1213 06:50:17.768334 4971 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 13 06:50:17 crc kubenswrapper[4971]: E1213 06:50:17.768324 4971 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 13 06:50:17 crc kubenswrapper[4971]: E1213 06:50:17.768404 4971 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 13 06:50:17 crc kubenswrapper[4971]: I1213 06:50:17.798813 4971 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 13 06:50:17 crc kubenswrapper[4971]: I1213 06:50:17.798923 4971 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 13 06:50:17 crc kubenswrapper[4971]: I1213 06:50:17.798940 4971 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 13 06:50:17 crc kubenswrapper[4971]: I1213 06:50:17.798957 4971 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 13 06:50:17 crc kubenswrapper[4971]: I1213 06:50:17.798966 4971 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-13T06:50:17Z","lastTransitionTime":"2025-12-13T06:50:17Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 13 06:50:17 crc kubenswrapper[4971]: I1213 06:50:17.901748 4971 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 13 06:50:17 crc kubenswrapper[4971]: I1213 06:50:17.901802 4971 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 13 06:50:17 crc kubenswrapper[4971]: I1213 06:50:17.901810 4971 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 13 06:50:17 crc kubenswrapper[4971]: I1213 06:50:17.901832 4971 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 13 06:50:17 crc kubenswrapper[4971]: I1213 06:50:17.901844 4971 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-13T06:50:17Z","lastTransitionTime":"2025-12-13T06:50:17Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 13 06:50:18 crc kubenswrapper[4971]: I1213 06:50:18.004115 4971 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 13 06:50:18 crc kubenswrapper[4971]: I1213 06:50:18.004162 4971 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 13 06:50:18 crc kubenswrapper[4971]: I1213 06:50:18.004172 4971 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 13 06:50:18 crc kubenswrapper[4971]: I1213 06:50:18.004189 4971 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 13 06:50:18 crc kubenswrapper[4971]: I1213 06:50:18.004200 4971 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-13T06:50:18Z","lastTransitionTime":"2025-12-13T06:50:18Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 13 06:50:18 crc kubenswrapper[4971]: I1213 06:50:18.106383 4971 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 13 06:50:18 crc kubenswrapper[4971]: I1213 06:50:18.106434 4971 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 13 06:50:18 crc kubenswrapper[4971]: I1213 06:50:18.106451 4971 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 13 06:50:18 crc kubenswrapper[4971]: I1213 06:50:18.106470 4971 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 13 06:50:18 crc kubenswrapper[4971]: I1213 06:50:18.106482 4971 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-13T06:50:18Z","lastTransitionTime":"2025-12-13T06:50:18Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 13 06:50:18 crc kubenswrapper[4971]: I1213 06:50:18.209394 4971 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 13 06:50:18 crc kubenswrapper[4971]: I1213 06:50:18.209472 4971 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 13 06:50:18 crc kubenswrapper[4971]: I1213 06:50:18.209499 4971 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 13 06:50:18 crc kubenswrapper[4971]: I1213 06:50:18.209565 4971 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 13 06:50:18 crc kubenswrapper[4971]: I1213 06:50:18.209596 4971 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-13T06:50:18Z","lastTransitionTime":"2025-12-13T06:50:18Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 13 06:50:18 crc kubenswrapper[4971]: I1213 06:50:18.312241 4971 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 13 06:50:18 crc kubenswrapper[4971]: I1213 06:50:18.312280 4971 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 13 06:50:18 crc kubenswrapper[4971]: I1213 06:50:18.312293 4971 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 13 06:50:18 crc kubenswrapper[4971]: I1213 06:50:18.312310 4971 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 13 06:50:18 crc kubenswrapper[4971]: I1213 06:50:18.312320 4971 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-13T06:50:18Z","lastTransitionTime":"2025-12-13T06:50:18Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 13 06:50:18 crc kubenswrapper[4971]: I1213 06:50:18.414780 4971 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 13 06:50:18 crc kubenswrapper[4971]: I1213 06:50:18.414839 4971 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 13 06:50:18 crc kubenswrapper[4971]: I1213 06:50:18.414852 4971 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 13 06:50:18 crc kubenswrapper[4971]: I1213 06:50:18.414870 4971 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 13 06:50:18 crc kubenswrapper[4971]: I1213 06:50:18.414883 4971 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-13T06:50:18Z","lastTransitionTime":"2025-12-13T06:50:18Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 13 06:50:18 crc kubenswrapper[4971]: I1213 06:50:18.517570 4971 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 13 06:50:18 crc kubenswrapper[4971]: I1213 06:50:18.517625 4971 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 13 06:50:18 crc kubenswrapper[4971]: I1213 06:50:18.517638 4971 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 13 06:50:18 crc kubenswrapper[4971]: I1213 06:50:18.517656 4971 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 13 06:50:18 crc kubenswrapper[4971]: I1213 06:50:18.517667 4971 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-13T06:50:18Z","lastTransitionTime":"2025-12-13T06:50:18Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 13 06:50:18 crc kubenswrapper[4971]: I1213 06:50:18.620502 4971 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 13 06:50:18 crc kubenswrapper[4971]: I1213 06:50:18.620596 4971 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 13 06:50:18 crc kubenswrapper[4971]: I1213 06:50:18.620608 4971 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 13 06:50:18 crc kubenswrapper[4971]: I1213 06:50:18.620627 4971 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 13 06:50:18 crc kubenswrapper[4971]: I1213 06:50:18.620643 4971 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-13T06:50:18Z","lastTransitionTime":"2025-12-13T06:50:18Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 13 06:50:18 crc kubenswrapper[4971]: I1213 06:50:18.724067 4971 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 13 06:50:18 crc kubenswrapper[4971]: I1213 06:50:18.724117 4971 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 13 06:50:18 crc kubenswrapper[4971]: I1213 06:50:18.724129 4971 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 13 06:50:18 crc kubenswrapper[4971]: I1213 06:50:18.724147 4971 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 13 06:50:18 crc kubenswrapper[4971]: I1213 06:50:18.724159 4971 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-13T06:50:18Z","lastTransitionTime":"2025-12-13T06:50:18Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 13 06:50:18 crc kubenswrapper[4971]: I1213 06:50:18.826953 4971 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 13 06:50:18 crc kubenswrapper[4971]: I1213 06:50:18.826998 4971 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 13 06:50:18 crc kubenswrapper[4971]: I1213 06:50:18.827006 4971 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 13 06:50:18 crc kubenswrapper[4971]: I1213 06:50:18.827024 4971 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 13 06:50:18 crc kubenswrapper[4971]: I1213 06:50:18.827033 4971 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-13T06:50:18Z","lastTransitionTime":"2025-12-13T06:50:18Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 13 06:50:18 crc kubenswrapper[4971]: I1213 06:50:18.929073 4971 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 13 06:50:18 crc kubenswrapper[4971]: I1213 06:50:18.929160 4971 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 13 06:50:18 crc kubenswrapper[4971]: I1213 06:50:18.929176 4971 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 13 06:50:18 crc kubenswrapper[4971]: I1213 06:50:18.929201 4971 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 13 06:50:18 crc kubenswrapper[4971]: I1213 06:50:18.929225 4971 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-13T06:50:18Z","lastTransitionTime":"2025-12-13T06:50:18Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 13 06:50:19 crc kubenswrapper[4971]: I1213 06:50:19.031607 4971 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 13 06:50:19 crc kubenswrapper[4971]: I1213 06:50:19.031656 4971 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 13 06:50:19 crc kubenswrapper[4971]: I1213 06:50:19.031668 4971 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 13 06:50:19 crc kubenswrapper[4971]: I1213 06:50:19.031685 4971 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 13 06:50:19 crc kubenswrapper[4971]: I1213 06:50:19.031697 4971 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-13T06:50:19Z","lastTransitionTime":"2025-12-13T06:50:19Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 13 06:50:19 crc kubenswrapper[4971]: I1213 06:50:19.134393 4971 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 13 06:50:19 crc kubenswrapper[4971]: I1213 06:50:19.134439 4971 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 13 06:50:19 crc kubenswrapper[4971]: I1213 06:50:19.134455 4971 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 13 06:50:19 crc kubenswrapper[4971]: I1213 06:50:19.134482 4971 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 13 06:50:19 crc kubenswrapper[4971]: I1213 06:50:19.134499 4971 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-13T06:50:19Z","lastTransitionTime":"2025-12-13T06:50:19Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 13 06:50:19 crc kubenswrapper[4971]: I1213 06:50:19.236974 4971 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 13 06:50:19 crc kubenswrapper[4971]: I1213 06:50:19.237008 4971 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 13 06:50:19 crc kubenswrapper[4971]: I1213 06:50:19.237017 4971 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 13 06:50:19 crc kubenswrapper[4971]: I1213 06:50:19.237031 4971 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 13 06:50:19 crc kubenswrapper[4971]: I1213 06:50:19.237039 4971 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-13T06:50:19Z","lastTransitionTime":"2025-12-13T06:50:19Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 13 06:50:19 crc kubenswrapper[4971]: I1213 06:50:19.340111 4971 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 13 06:50:19 crc kubenswrapper[4971]: I1213 06:50:19.340189 4971 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 13 06:50:19 crc kubenswrapper[4971]: I1213 06:50:19.340208 4971 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 13 06:50:19 crc kubenswrapper[4971]: I1213 06:50:19.340232 4971 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 13 06:50:19 crc kubenswrapper[4971]: I1213 06:50:19.340247 4971 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-13T06:50:19Z","lastTransitionTime":"2025-12-13T06:50:19Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 13 06:50:19 crc kubenswrapper[4971]: I1213 06:50:19.442722 4971 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 13 06:50:19 crc kubenswrapper[4971]: I1213 06:50:19.442816 4971 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 13 06:50:19 crc kubenswrapper[4971]: I1213 06:50:19.442828 4971 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 13 06:50:19 crc kubenswrapper[4971]: I1213 06:50:19.442847 4971 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 13 06:50:19 crc kubenswrapper[4971]: I1213 06:50:19.442860 4971 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-13T06:50:19Z","lastTransitionTime":"2025-12-13T06:50:19Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 13 06:50:19 crc kubenswrapper[4971]: I1213 06:50:19.545207 4971 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 13 06:50:19 crc kubenswrapper[4971]: I1213 06:50:19.545248 4971 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 13 06:50:19 crc kubenswrapper[4971]: I1213 06:50:19.545258 4971 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 13 06:50:19 crc kubenswrapper[4971]: I1213 06:50:19.545274 4971 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 13 06:50:19 crc kubenswrapper[4971]: I1213 06:50:19.545287 4971 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-13T06:50:19Z","lastTransitionTime":"2025-12-13T06:50:19Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 13 06:50:19 crc kubenswrapper[4971]: I1213 06:50:19.647951 4971 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 13 06:50:19 crc kubenswrapper[4971]: I1213 06:50:19.647987 4971 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 13 06:50:19 crc kubenswrapper[4971]: I1213 06:50:19.647996 4971 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 13 06:50:19 crc kubenswrapper[4971]: I1213 06:50:19.648009 4971 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 13 06:50:19 crc kubenswrapper[4971]: I1213 06:50:19.648043 4971 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-13T06:50:19Z","lastTransitionTime":"2025-12-13T06:50:19Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 13 06:50:19 crc kubenswrapper[4971]: I1213 06:50:19.751184 4971 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 13 06:50:19 crc kubenswrapper[4971]: I1213 06:50:19.751273 4971 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 13 06:50:19 crc kubenswrapper[4971]: I1213 06:50:19.751302 4971 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 13 06:50:19 crc kubenswrapper[4971]: I1213 06:50:19.751330 4971 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 13 06:50:19 crc kubenswrapper[4971]: I1213 06:50:19.751346 4971 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-13T06:50:19Z","lastTransitionTime":"2025-12-13T06:50:19Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 13 06:50:19 crc kubenswrapper[4971]: I1213 06:50:19.767651 4971 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 13 06:50:19 crc kubenswrapper[4971]: I1213 06:50:19.767700 4971 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-x98k7" Dec 13 06:50:19 crc kubenswrapper[4971]: I1213 06:50:19.767670 4971 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 13 06:50:19 crc kubenswrapper[4971]: E1213 06:50:19.767802 4971 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 13 06:50:19 crc kubenswrapper[4971]: I1213 06:50:19.767825 4971 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 13 06:50:19 crc kubenswrapper[4971]: E1213 06:50:19.767909 4971 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 13 06:50:19 crc kubenswrapper[4971]: E1213 06:50:19.767981 4971 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 13 06:50:19 crc kubenswrapper[4971]: E1213 06:50:19.768041 4971 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-x98k7" podUID="c7185dec-3240-42e8-9d90-6182a5e8b65c" Dec 13 06:50:19 crc kubenswrapper[4971]: I1213 06:50:19.853455 4971 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 13 06:50:19 crc kubenswrapper[4971]: I1213 06:50:19.853491 4971 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 13 06:50:19 crc kubenswrapper[4971]: I1213 06:50:19.853501 4971 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 13 06:50:19 crc kubenswrapper[4971]: I1213 06:50:19.853532 4971 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 13 06:50:19 crc kubenswrapper[4971]: I1213 06:50:19.853542 4971 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-13T06:50:19Z","lastTransitionTime":"2025-12-13T06:50:19Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 13 06:50:19 crc kubenswrapper[4971]: I1213 06:50:19.955990 4971 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 13 06:50:19 crc kubenswrapper[4971]: I1213 06:50:19.956053 4971 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 13 06:50:19 crc kubenswrapper[4971]: I1213 06:50:19.956070 4971 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 13 06:50:19 crc kubenswrapper[4971]: I1213 06:50:19.956092 4971 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 13 06:50:19 crc kubenswrapper[4971]: I1213 06:50:19.956109 4971 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-13T06:50:19Z","lastTransitionTime":"2025-12-13T06:50:19Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 13 06:50:20 crc kubenswrapper[4971]: I1213 06:50:20.058648 4971 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 13 06:50:20 crc kubenswrapper[4971]: I1213 06:50:20.058700 4971 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 13 06:50:20 crc kubenswrapper[4971]: I1213 06:50:20.058711 4971 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 13 06:50:20 crc kubenswrapper[4971]: I1213 06:50:20.058728 4971 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 13 06:50:20 crc kubenswrapper[4971]: I1213 06:50:20.058739 4971 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-13T06:50:20Z","lastTransitionTime":"2025-12-13T06:50:20Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 13 06:50:20 crc kubenswrapper[4971]: I1213 06:50:20.160828 4971 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 13 06:50:20 crc kubenswrapper[4971]: I1213 06:50:20.160877 4971 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 13 06:50:20 crc kubenswrapper[4971]: I1213 06:50:20.160889 4971 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 13 06:50:20 crc kubenswrapper[4971]: I1213 06:50:20.160907 4971 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 13 06:50:20 crc kubenswrapper[4971]: I1213 06:50:20.160922 4971 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-13T06:50:20Z","lastTransitionTime":"2025-12-13T06:50:20Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 13 06:50:20 crc kubenswrapper[4971]: I1213 06:50:20.264124 4971 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 13 06:50:20 crc kubenswrapper[4971]: I1213 06:50:20.264186 4971 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 13 06:50:20 crc kubenswrapper[4971]: I1213 06:50:20.264203 4971 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 13 06:50:20 crc kubenswrapper[4971]: I1213 06:50:20.264226 4971 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 13 06:50:20 crc kubenswrapper[4971]: I1213 06:50:20.264244 4971 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-13T06:50:20Z","lastTransitionTime":"2025-12-13T06:50:20Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 13 06:50:20 crc kubenswrapper[4971]: I1213 06:50:20.367026 4971 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 13 06:50:20 crc kubenswrapper[4971]: I1213 06:50:20.367213 4971 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 13 06:50:20 crc kubenswrapper[4971]: I1213 06:50:20.367243 4971 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 13 06:50:20 crc kubenswrapper[4971]: I1213 06:50:20.367278 4971 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 13 06:50:20 crc kubenswrapper[4971]: I1213 06:50:20.367302 4971 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-13T06:50:20Z","lastTransitionTime":"2025-12-13T06:50:20Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 13 06:50:20 crc kubenswrapper[4971]: I1213 06:50:20.469656 4971 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 13 06:50:20 crc kubenswrapper[4971]: I1213 06:50:20.469706 4971 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 13 06:50:20 crc kubenswrapper[4971]: I1213 06:50:20.469717 4971 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 13 06:50:20 crc kubenswrapper[4971]: I1213 06:50:20.469731 4971 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 13 06:50:20 crc kubenswrapper[4971]: I1213 06:50:20.469749 4971 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-13T06:50:20Z","lastTransitionTime":"2025-12-13T06:50:20Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 13 06:50:20 crc kubenswrapper[4971]: I1213 06:50:20.572968 4971 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 13 06:50:20 crc kubenswrapper[4971]: I1213 06:50:20.573007 4971 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 13 06:50:20 crc kubenswrapper[4971]: I1213 06:50:20.573016 4971 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 13 06:50:20 crc kubenswrapper[4971]: I1213 06:50:20.573030 4971 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 13 06:50:20 crc kubenswrapper[4971]: I1213 06:50:20.573040 4971 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-13T06:50:20Z","lastTransitionTime":"2025-12-13T06:50:20Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 13 06:50:20 crc kubenswrapper[4971]: I1213 06:50:20.676172 4971 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 13 06:50:20 crc kubenswrapper[4971]: I1213 06:50:20.676225 4971 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 13 06:50:20 crc kubenswrapper[4971]: I1213 06:50:20.676236 4971 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 13 06:50:20 crc kubenswrapper[4971]: I1213 06:50:20.676253 4971 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 13 06:50:20 crc kubenswrapper[4971]: I1213 06:50:20.676266 4971 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-13T06:50:20Z","lastTransitionTime":"2025-12-13T06:50:20Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 13 06:50:20 crc kubenswrapper[4971]: I1213 06:50:20.778365 4971 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 13 06:50:20 crc kubenswrapper[4971]: I1213 06:50:20.778437 4971 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 13 06:50:20 crc kubenswrapper[4971]: I1213 06:50:20.778455 4971 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 13 06:50:20 crc kubenswrapper[4971]: I1213 06:50:20.778480 4971 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 13 06:50:20 crc kubenswrapper[4971]: I1213 06:50:20.778496 4971 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-13T06:50:20Z","lastTransitionTime":"2025-12-13T06:50:20Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 13 06:50:20 crc kubenswrapper[4971]: I1213 06:50:20.881252 4971 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 13 06:50:20 crc kubenswrapper[4971]: I1213 06:50:20.881327 4971 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 13 06:50:20 crc kubenswrapper[4971]: I1213 06:50:20.881573 4971 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 13 06:50:20 crc kubenswrapper[4971]: I1213 06:50:20.881602 4971 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 13 06:50:20 crc kubenswrapper[4971]: I1213 06:50:20.881620 4971 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-13T06:50:20Z","lastTransitionTime":"2025-12-13T06:50:20Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 13 06:50:20 crc kubenswrapper[4971]: I1213 06:50:20.984266 4971 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 13 06:50:20 crc kubenswrapper[4971]: I1213 06:50:20.984373 4971 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 13 06:50:20 crc kubenswrapper[4971]: I1213 06:50:20.984401 4971 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 13 06:50:20 crc kubenswrapper[4971]: I1213 06:50:20.984433 4971 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 13 06:50:20 crc kubenswrapper[4971]: I1213 06:50:20.984458 4971 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-13T06:50:20Z","lastTransitionTime":"2025-12-13T06:50:20Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 13 06:50:21 crc kubenswrapper[4971]: I1213 06:50:21.086469 4971 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 13 06:50:21 crc kubenswrapper[4971]: I1213 06:50:21.086550 4971 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 13 06:50:21 crc kubenswrapper[4971]: I1213 06:50:21.086561 4971 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 13 06:50:21 crc kubenswrapper[4971]: I1213 06:50:21.086576 4971 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 13 06:50:21 crc kubenswrapper[4971]: I1213 06:50:21.086587 4971 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-13T06:50:21Z","lastTransitionTime":"2025-12-13T06:50:21Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 13 06:50:21 crc kubenswrapper[4971]: I1213 06:50:21.188025 4971 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 13 06:50:21 crc kubenswrapper[4971]: I1213 06:50:21.188074 4971 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 13 06:50:21 crc kubenswrapper[4971]: I1213 06:50:21.188087 4971 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 13 06:50:21 crc kubenswrapper[4971]: I1213 06:50:21.188105 4971 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 13 06:50:21 crc kubenswrapper[4971]: I1213 06:50:21.188120 4971 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-13T06:50:21Z","lastTransitionTime":"2025-12-13T06:50:21Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 13 06:50:21 crc kubenswrapper[4971]: I1213 06:50:21.289943 4971 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 13 06:50:21 crc kubenswrapper[4971]: I1213 06:50:21.289983 4971 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 13 06:50:21 crc kubenswrapper[4971]: I1213 06:50:21.289991 4971 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 13 06:50:21 crc kubenswrapper[4971]: I1213 06:50:21.290005 4971 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 13 06:50:21 crc kubenswrapper[4971]: I1213 06:50:21.290014 4971 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-13T06:50:21Z","lastTransitionTime":"2025-12-13T06:50:21Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 13 06:50:21 crc kubenswrapper[4971]: I1213 06:50:21.392291 4971 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 13 06:50:21 crc kubenswrapper[4971]: I1213 06:50:21.392367 4971 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 13 06:50:21 crc kubenswrapper[4971]: I1213 06:50:21.392380 4971 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 13 06:50:21 crc kubenswrapper[4971]: I1213 06:50:21.392396 4971 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 13 06:50:21 crc kubenswrapper[4971]: I1213 06:50:21.392406 4971 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-13T06:50:21Z","lastTransitionTime":"2025-12-13T06:50:21Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 13 06:50:21 crc kubenswrapper[4971]: I1213 06:50:21.495271 4971 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 13 06:50:21 crc kubenswrapper[4971]: I1213 06:50:21.495325 4971 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 13 06:50:21 crc kubenswrapper[4971]: I1213 06:50:21.495336 4971 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 13 06:50:21 crc kubenswrapper[4971]: I1213 06:50:21.495351 4971 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 13 06:50:21 crc kubenswrapper[4971]: I1213 06:50:21.495361 4971 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-13T06:50:21Z","lastTransitionTime":"2025-12-13T06:50:21Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 13 06:50:21 crc kubenswrapper[4971]: I1213 06:50:21.597611 4971 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 13 06:50:21 crc kubenswrapper[4971]: I1213 06:50:21.597663 4971 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 13 06:50:21 crc kubenswrapper[4971]: I1213 06:50:21.597676 4971 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 13 06:50:21 crc kubenswrapper[4971]: I1213 06:50:21.597698 4971 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 13 06:50:21 crc kubenswrapper[4971]: I1213 06:50:21.597714 4971 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-13T06:50:21Z","lastTransitionTime":"2025-12-13T06:50:21Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 13 06:50:21 crc kubenswrapper[4971]: I1213 06:50:21.607180 4971 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 13 06:50:21 crc kubenswrapper[4971]: I1213 06:50:21.607235 4971 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 13 06:50:21 crc kubenswrapper[4971]: I1213 06:50:21.607245 4971 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 13 06:50:21 crc kubenswrapper[4971]: I1213 06:50:21.607263 4971 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 13 06:50:21 crc kubenswrapper[4971]: I1213 06:50:21.607277 4971 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-13T06:50:21Z","lastTransitionTime":"2025-12-13T06:50:21Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 13 06:50:21 crc kubenswrapper[4971]: E1213 06:50:21.624405 4971 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-12-13T06:50:21Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-13T06:50:21Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-13T06:50:21Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-13T06:50:21Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-13T06:50:21Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-13T06:50:21Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-13T06:50:21Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-13T06:50:21Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"bd36cbc9-bbe9-45a1-a21a-ad2bd10499ee\\\",\\\"systemUUID\\\":\\\"a9af00dc-cf25-47af-9997-79577f8f3bea\\\"},\\\"runtimeHandlers\\\":[{\\\"features\\\":{\\\"recursiveReadOnlyMounts\\\":true,\\\"userNamespaces\\\":false},\\\"name\\\":\\\"runc\\\"},{\\\"features\\\":{\\\"recursiveReadOnlyMounts\\\":true,\\\"userNamespaces\\\":true},\\\"name\\\":\\\"crun\\\"},{\\\"features\\\":{\\\"recursiveReadOnlyMounts\\\":true,\\\"userNamespaces\\\":true},\\\"name\\\":\\\"\\\"}]}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-13T06:50:21Z is after 2025-08-24T17:21:41Z" Dec 13 06:50:21 crc kubenswrapper[4971]: I1213 06:50:21.628585 4971 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 13 06:50:21 crc kubenswrapper[4971]: I1213 06:50:21.628631 4971 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 13 06:50:21 crc kubenswrapper[4971]: I1213 06:50:21.628644 4971 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 13 06:50:21 crc kubenswrapper[4971]: I1213 06:50:21.628664 4971 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 13 06:50:21 crc kubenswrapper[4971]: I1213 06:50:21.628677 4971 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-13T06:50:21Z","lastTransitionTime":"2025-12-13T06:50:21Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 13 06:50:21 crc kubenswrapper[4971]: E1213 06:50:21.643261 4971 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-12-13T06:50:21Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-13T06:50:21Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-13T06:50:21Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-13T06:50:21Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-13T06:50:21Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-13T06:50:21Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-13T06:50:21Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-13T06:50:21Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"bd36cbc9-bbe9-45a1-a21a-ad2bd10499ee\\\",\\\"systemUUID\\\":\\\"a9af00dc-cf25-47af-9997-79577f8f3bea\\\"},\\\"runtimeHandlers\\\":[{\\\"features\\\":{\\\"recursiveReadOnlyMounts\\\":true,\\\"userNamespaces\\\":false},\\\"name\\\":\\\"runc\\\"},{\\\"features\\\":{\\\"recursiveReadOnlyMounts\\\":true,\\\"userNamespaces\\\":true},\\\"name\\\":\\\"crun\\\"},{\\\"features\\\":{\\\"recursiveReadOnlyMounts\\\":true,\\\"userNamespaces\\\":true},\\\"name\\\":\\\"\\\"}]}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-13T06:50:21Z is after 2025-08-24T17:21:41Z" Dec 13 06:50:21 crc kubenswrapper[4971]: I1213 06:50:21.646376 4971 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 13 06:50:21 crc kubenswrapper[4971]: I1213 06:50:21.646436 4971 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 13 06:50:21 crc kubenswrapper[4971]: I1213 06:50:21.646447 4971 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 13 06:50:21 crc kubenswrapper[4971]: I1213 06:50:21.646463 4971 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 13 06:50:21 crc kubenswrapper[4971]: I1213 06:50:21.646473 4971 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-13T06:50:21Z","lastTransitionTime":"2025-12-13T06:50:21Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 13 06:50:21 crc kubenswrapper[4971]: E1213 06:50:21.659639 4971 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-12-13T06:50:21Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-13T06:50:21Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-13T06:50:21Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-13T06:50:21Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-13T06:50:21Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-13T06:50:21Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-13T06:50:21Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-13T06:50:21Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"bd36cbc9-bbe9-45a1-a21a-ad2bd10499ee\\\",\\\"systemUUID\\\":\\\"a9af00dc-cf25-47af-9997-79577f8f3bea\\\"},\\\"runtimeHandlers\\\":[{\\\"features\\\":{\\\"recursiveReadOnlyMounts\\\":true,\\\"userNamespaces\\\":false},\\\"name\\\":\\\"runc\\\"},{\\\"features\\\":{\\\"recursiveReadOnlyMounts\\\":true,\\\"userNamespaces\\\":true},\\\"name\\\":\\\"crun\\\"},{\\\"features\\\":{\\\"recursiveReadOnlyMounts\\\":true,\\\"userNamespaces\\\":true},\\\"name\\\":\\\"\\\"}]}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-13T06:50:21Z is after 2025-08-24T17:21:41Z" Dec 13 06:50:21 crc kubenswrapper[4971]: I1213 06:50:21.663132 4971 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 13 06:50:21 crc kubenswrapper[4971]: I1213 06:50:21.663169 4971 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 13 06:50:21 crc kubenswrapper[4971]: I1213 06:50:21.663181 4971 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 13 06:50:21 crc kubenswrapper[4971]: I1213 06:50:21.663199 4971 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 13 06:50:21 crc kubenswrapper[4971]: I1213 06:50:21.663210 4971 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-13T06:50:21Z","lastTransitionTime":"2025-12-13T06:50:21Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 13 06:50:21 crc kubenswrapper[4971]: E1213 06:50:21.676282 4971 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-12-13T06:50:21Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-13T06:50:21Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-13T06:50:21Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-13T06:50:21Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-13T06:50:21Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-13T06:50:21Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-13T06:50:21Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-13T06:50:21Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"bd36cbc9-bbe9-45a1-a21a-ad2bd10499ee\\\",\\\"systemUUID\\\":\\\"a9af00dc-cf25-47af-9997-79577f8f3bea\\\"},\\\"runtimeHandlers\\\":[{\\\"features\\\":{\\\"recursiveReadOnlyMounts\\\":true,\\\"userNamespaces\\\":false},\\\"name\\\":\\\"runc\\\"},{\\\"features\\\":{\\\"recursiveReadOnlyMounts\\\":true,\\\"userNamespaces\\\":true},\\\"name\\\":\\\"crun\\\"},{\\\"features\\\":{\\\"recursiveReadOnlyMounts\\\":true,\\\"userNamespaces\\\":true},\\\"name\\\":\\\"\\\"}]}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-13T06:50:21Z is after 2025-08-24T17:21:41Z" Dec 13 06:50:21 crc kubenswrapper[4971]: I1213 06:50:21.680252 4971 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 13 06:50:21 crc kubenswrapper[4971]: I1213 06:50:21.680291 4971 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 13 06:50:21 crc kubenswrapper[4971]: I1213 06:50:21.680499 4971 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 13 06:50:21 crc kubenswrapper[4971]: I1213 06:50:21.680542 4971 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 13 06:50:21 crc kubenswrapper[4971]: I1213 06:50:21.680597 4971 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-13T06:50:21Z","lastTransitionTime":"2025-12-13T06:50:21Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 13 06:50:21 crc kubenswrapper[4971]: E1213 06:50:21.697305 4971 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-12-13T06:50:21Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-13T06:50:21Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-13T06:50:21Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-13T06:50:21Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-13T06:50:21Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-13T06:50:21Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-13T06:50:21Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-13T06:50:21Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"bd36cbc9-bbe9-45a1-a21a-ad2bd10499ee\\\",\\\"systemUUID\\\":\\\"a9af00dc-cf25-47af-9997-79577f8f3bea\\\"},\\\"runtimeHandlers\\\":[{\\\"features\\\":{\\\"recursiveReadOnlyMounts\\\":true,\\\"userNamespaces\\\":false},\\\"name\\\":\\\"runc\\\"},{\\\"features\\\":{\\\"recursiveReadOnlyMounts\\\":true,\\\"userNamespaces\\\":true},\\\"name\\\":\\\"crun\\\"},{\\\"features\\\":{\\\"recursiveReadOnlyMounts\\\":true,\\\"userNamespaces\\\":true},\\\"name\\\":\\\"\\\"}]}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-13T06:50:21Z is after 2025-08-24T17:21:41Z" Dec 13 06:50:21 crc kubenswrapper[4971]: E1213 06:50:21.697553 4971 kubelet_node_status.go:572] "Unable to update node status" err="update node status exceeds retry count" Dec 13 06:50:21 crc kubenswrapper[4971]: I1213 06:50:21.700153 4971 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 13 06:50:21 crc kubenswrapper[4971]: I1213 06:50:21.700207 4971 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 13 06:50:21 crc kubenswrapper[4971]: I1213 06:50:21.700223 4971 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 13 06:50:21 crc kubenswrapper[4971]: I1213 06:50:21.700245 4971 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 13 06:50:21 crc kubenswrapper[4971]: I1213 06:50:21.700264 4971 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-13T06:50:21Z","lastTransitionTime":"2025-12-13T06:50:21Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 13 06:50:21 crc kubenswrapper[4971]: I1213 06:50:21.767809 4971 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 13 06:50:21 crc kubenswrapper[4971]: I1213 06:50:21.767861 4971 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 13 06:50:21 crc kubenswrapper[4971]: I1213 06:50:21.767910 4971 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 13 06:50:21 crc kubenswrapper[4971]: I1213 06:50:21.767836 4971 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-x98k7" Dec 13 06:50:21 crc kubenswrapper[4971]: E1213 06:50:21.767997 4971 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 13 06:50:21 crc kubenswrapper[4971]: E1213 06:50:21.768156 4971 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 13 06:50:21 crc kubenswrapper[4971]: E1213 06:50:21.768254 4971 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-x98k7" podUID="c7185dec-3240-42e8-9d90-6182a5e8b65c" Dec 13 06:50:21 crc kubenswrapper[4971]: E1213 06:50:21.768323 4971 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 13 06:50:21 crc kubenswrapper[4971]: I1213 06:50:21.802822 4971 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 13 06:50:21 crc kubenswrapper[4971]: I1213 06:50:21.802873 4971 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 13 06:50:21 crc kubenswrapper[4971]: I1213 06:50:21.802884 4971 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 13 06:50:21 crc kubenswrapper[4971]: I1213 06:50:21.802907 4971 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 13 06:50:21 crc kubenswrapper[4971]: I1213 06:50:21.802919 4971 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-13T06:50:21Z","lastTransitionTime":"2025-12-13T06:50:21Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 13 06:50:21 crc kubenswrapper[4971]: I1213 06:50:21.907245 4971 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 13 06:50:21 crc kubenswrapper[4971]: I1213 06:50:21.907298 4971 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 13 06:50:21 crc kubenswrapper[4971]: I1213 06:50:21.907309 4971 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 13 06:50:21 crc kubenswrapper[4971]: I1213 06:50:21.907328 4971 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 13 06:50:21 crc kubenswrapper[4971]: I1213 06:50:21.907342 4971 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-13T06:50:21Z","lastTransitionTime":"2025-12-13T06:50:21Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 13 06:50:22 crc kubenswrapper[4971]: I1213 06:50:22.009955 4971 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 13 06:50:22 crc kubenswrapper[4971]: I1213 06:50:22.009992 4971 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 13 06:50:22 crc kubenswrapper[4971]: I1213 06:50:22.010001 4971 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 13 06:50:22 crc kubenswrapper[4971]: I1213 06:50:22.010014 4971 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 13 06:50:22 crc kubenswrapper[4971]: I1213 06:50:22.010021 4971 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-13T06:50:22Z","lastTransitionTime":"2025-12-13T06:50:22Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 13 06:50:22 crc kubenswrapper[4971]: I1213 06:50:22.112792 4971 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 13 06:50:22 crc kubenswrapper[4971]: I1213 06:50:22.112852 4971 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 13 06:50:22 crc kubenswrapper[4971]: I1213 06:50:22.112860 4971 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 13 06:50:22 crc kubenswrapper[4971]: I1213 06:50:22.112875 4971 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 13 06:50:22 crc kubenswrapper[4971]: I1213 06:50:22.112885 4971 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-13T06:50:22Z","lastTransitionTime":"2025-12-13T06:50:22Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 13 06:50:22 crc kubenswrapper[4971]: I1213 06:50:22.215038 4971 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 13 06:50:22 crc kubenswrapper[4971]: I1213 06:50:22.215090 4971 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 13 06:50:22 crc kubenswrapper[4971]: I1213 06:50:22.215103 4971 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 13 06:50:22 crc kubenswrapper[4971]: I1213 06:50:22.215123 4971 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 13 06:50:22 crc kubenswrapper[4971]: I1213 06:50:22.215142 4971 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-13T06:50:22Z","lastTransitionTime":"2025-12-13T06:50:22Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 13 06:50:22 crc kubenswrapper[4971]: I1213 06:50:22.318409 4971 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 13 06:50:22 crc kubenswrapper[4971]: I1213 06:50:22.318466 4971 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 13 06:50:22 crc kubenswrapper[4971]: I1213 06:50:22.318475 4971 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 13 06:50:22 crc kubenswrapper[4971]: I1213 06:50:22.318488 4971 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 13 06:50:22 crc kubenswrapper[4971]: I1213 06:50:22.318496 4971 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-13T06:50:22Z","lastTransitionTime":"2025-12-13T06:50:22Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 13 06:50:22 crc kubenswrapper[4971]: I1213 06:50:22.421150 4971 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 13 06:50:22 crc kubenswrapper[4971]: I1213 06:50:22.421234 4971 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 13 06:50:22 crc kubenswrapper[4971]: I1213 06:50:22.421249 4971 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 13 06:50:22 crc kubenswrapper[4971]: I1213 06:50:22.421267 4971 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 13 06:50:22 crc kubenswrapper[4971]: I1213 06:50:22.421279 4971 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-13T06:50:22Z","lastTransitionTime":"2025-12-13T06:50:22Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 13 06:50:22 crc kubenswrapper[4971]: I1213 06:50:22.523834 4971 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 13 06:50:22 crc kubenswrapper[4971]: I1213 06:50:22.523887 4971 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 13 06:50:22 crc kubenswrapper[4971]: I1213 06:50:22.523899 4971 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 13 06:50:22 crc kubenswrapper[4971]: I1213 06:50:22.523917 4971 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 13 06:50:22 crc kubenswrapper[4971]: I1213 06:50:22.523928 4971 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-13T06:50:22Z","lastTransitionTime":"2025-12-13T06:50:22Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 13 06:50:22 crc kubenswrapper[4971]: I1213 06:50:22.626700 4971 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 13 06:50:22 crc kubenswrapper[4971]: I1213 06:50:22.626741 4971 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 13 06:50:22 crc kubenswrapper[4971]: I1213 06:50:22.626753 4971 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 13 06:50:22 crc kubenswrapper[4971]: I1213 06:50:22.626767 4971 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 13 06:50:22 crc kubenswrapper[4971]: I1213 06:50:22.626777 4971 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-13T06:50:22Z","lastTransitionTime":"2025-12-13T06:50:22Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 13 06:50:22 crc kubenswrapper[4971]: I1213 06:50:22.728712 4971 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 13 06:50:22 crc kubenswrapper[4971]: I1213 06:50:22.728755 4971 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 13 06:50:22 crc kubenswrapper[4971]: I1213 06:50:22.728770 4971 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 13 06:50:22 crc kubenswrapper[4971]: I1213 06:50:22.728787 4971 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 13 06:50:22 crc kubenswrapper[4971]: I1213 06:50:22.728798 4971 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-13T06:50:22Z","lastTransitionTime":"2025-12-13T06:50:22Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 13 06:50:22 crc kubenswrapper[4971]: I1213 06:50:22.768659 4971 scope.go:117] "RemoveContainer" containerID="0b454a87c6dd11bd8c3bb3553ee71de32ed6f185984e5057a20f2237b35c45d9" Dec 13 06:50:22 crc kubenswrapper[4971]: I1213 06:50:22.833074 4971 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 13 06:50:22 crc kubenswrapper[4971]: I1213 06:50:22.833553 4971 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 13 06:50:22 crc kubenswrapper[4971]: I1213 06:50:22.833569 4971 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 13 06:50:22 crc kubenswrapper[4971]: I1213 06:50:22.833590 4971 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 13 06:50:22 crc kubenswrapper[4971]: I1213 06:50:22.833603 4971 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-13T06:50:22Z","lastTransitionTime":"2025-12-13T06:50:22Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 13 06:50:22 crc kubenswrapper[4971]: I1213 06:50:22.936158 4971 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 13 06:50:22 crc kubenswrapper[4971]: I1213 06:50:22.936218 4971 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 13 06:50:22 crc kubenswrapper[4971]: I1213 06:50:22.936229 4971 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 13 06:50:22 crc kubenswrapper[4971]: I1213 06:50:22.936247 4971 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 13 06:50:22 crc kubenswrapper[4971]: I1213 06:50:22.936257 4971 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-13T06:50:22Z","lastTransitionTime":"2025-12-13T06:50:22Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 13 06:50:23 crc kubenswrapper[4971]: I1213 06:50:23.038059 4971 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 13 06:50:23 crc kubenswrapper[4971]: I1213 06:50:23.038140 4971 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 13 06:50:23 crc kubenswrapper[4971]: I1213 06:50:23.038156 4971 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 13 06:50:23 crc kubenswrapper[4971]: I1213 06:50:23.038177 4971 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 13 06:50:23 crc kubenswrapper[4971]: I1213 06:50:23.038196 4971 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-13T06:50:23Z","lastTransitionTime":"2025-12-13T06:50:23Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 13 06:50:23 crc kubenswrapper[4971]: I1213 06:50:23.140924 4971 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 13 06:50:23 crc kubenswrapper[4971]: I1213 06:50:23.140977 4971 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 13 06:50:23 crc kubenswrapper[4971]: I1213 06:50:23.140985 4971 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 13 06:50:23 crc kubenswrapper[4971]: I1213 06:50:23.140998 4971 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 13 06:50:23 crc kubenswrapper[4971]: I1213 06:50:23.141008 4971 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-13T06:50:23Z","lastTransitionTime":"2025-12-13T06:50:23Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 13 06:50:23 crc kubenswrapper[4971]: I1213 06:50:23.242824 4971 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 13 06:50:23 crc kubenswrapper[4971]: I1213 06:50:23.242862 4971 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 13 06:50:23 crc kubenswrapper[4971]: I1213 06:50:23.242875 4971 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 13 06:50:23 crc kubenswrapper[4971]: I1213 06:50:23.242894 4971 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 13 06:50:23 crc kubenswrapper[4971]: I1213 06:50:23.242906 4971 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-13T06:50:23Z","lastTransitionTime":"2025-12-13T06:50:23Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 13 06:50:23 crc kubenswrapper[4971]: I1213 06:50:23.345138 4971 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 13 06:50:23 crc kubenswrapper[4971]: I1213 06:50:23.345185 4971 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 13 06:50:23 crc kubenswrapper[4971]: I1213 06:50:23.345197 4971 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 13 06:50:23 crc kubenswrapper[4971]: I1213 06:50:23.345213 4971 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 13 06:50:23 crc kubenswrapper[4971]: I1213 06:50:23.345223 4971 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-13T06:50:23Z","lastTransitionTime":"2025-12-13T06:50:23Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 13 06:50:23 crc kubenswrapper[4971]: I1213 06:50:23.448891 4971 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 13 06:50:23 crc kubenswrapper[4971]: I1213 06:50:23.448920 4971 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 13 06:50:23 crc kubenswrapper[4971]: I1213 06:50:23.448929 4971 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 13 06:50:23 crc kubenswrapper[4971]: I1213 06:50:23.448946 4971 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 13 06:50:23 crc kubenswrapper[4971]: I1213 06:50:23.448955 4971 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-13T06:50:23Z","lastTransitionTime":"2025-12-13T06:50:23Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 13 06:50:23 crc kubenswrapper[4971]: I1213 06:50:23.551088 4971 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 13 06:50:23 crc kubenswrapper[4971]: I1213 06:50:23.551133 4971 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 13 06:50:23 crc kubenswrapper[4971]: I1213 06:50:23.551144 4971 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 13 06:50:23 crc kubenswrapper[4971]: I1213 06:50:23.551162 4971 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 13 06:50:23 crc kubenswrapper[4971]: I1213 06:50:23.551174 4971 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-13T06:50:23Z","lastTransitionTime":"2025-12-13T06:50:23Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 13 06:50:23 crc kubenswrapper[4971]: I1213 06:50:23.653934 4971 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 13 06:50:23 crc kubenswrapper[4971]: I1213 06:50:23.653988 4971 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 13 06:50:23 crc kubenswrapper[4971]: I1213 06:50:23.654001 4971 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 13 06:50:23 crc kubenswrapper[4971]: I1213 06:50:23.654016 4971 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 13 06:50:23 crc kubenswrapper[4971]: I1213 06:50:23.654035 4971 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-13T06:50:23Z","lastTransitionTime":"2025-12-13T06:50:23Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 13 06:50:23 crc kubenswrapper[4971]: I1213 06:50:23.756292 4971 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 13 06:50:23 crc kubenswrapper[4971]: I1213 06:50:23.756323 4971 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 13 06:50:23 crc kubenswrapper[4971]: I1213 06:50:23.756333 4971 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 13 06:50:23 crc kubenswrapper[4971]: I1213 06:50:23.756349 4971 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 13 06:50:23 crc kubenswrapper[4971]: I1213 06:50:23.756360 4971 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-13T06:50:23Z","lastTransitionTime":"2025-12-13T06:50:23Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 13 06:50:23 crc kubenswrapper[4971]: I1213 06:50:23.767760 4971 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 13 06:50:23 crc kubenswrapper[4971]: E1213 06:50:23.767868 4971 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 13 06:50:23 crc kubenswrapper[4971]: I1213 06:50:23.767983 4971 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-x98k7" Dec 13 06:50:23 crc kubenswrapper[4971]: I1213 06:50:23.768027 4971 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 13 06:50:23 crc kubenswrapper[4971]: E1213 06:50:23.768108 4971 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-x98k7" podUID="c7185dec-3240-42e8-9d90-6182a5e8b65c" Dec 13 06:50:23 crc kubenswrapper[4971]: I1213 06:50:23.768137 4971 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 13 06:50:23 crc kubenswrapper[4971]: E1213 06:50:23.768191 4971 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 13 06:50:23 crc kubenswrapper[4971]: E1213 06:50:23.768263 4971 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 13 06:50:23 crc kubenswrapper[4971]: I1213 06:50:23.788788 4971 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-13T06:49:24Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-13T06:49:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-13T06:49:24Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d0b2ecf0d5b22f23217f42f259ae7ffb6359555b22c30eaebf667eaf254b3bab\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-13T06:49:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-13T06:50:23Z is after 2025-08-24T17:21:41Z" Dec 13 06:50:23 crc kubenswrapper[4971]: I1213 06:50:23.805633 4971 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-13T06:49:23Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-13T06:49:23Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-13T06:50:23Z is after 2025-08-24T17:21:41Z" Dec 13 06:50:23 crc kubenswrapper[4971]: I1213 06:50:23.817022 4971 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-13T06:49:26Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-13T06:49:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-13T06:49:26Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9c8a0d1db0e754462d805e82daf5f52ffd5457507a30456535ec59e096c9d092\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-13T06:49:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-13T06:50:23Z is after 2025-08-24T17:21:41Z" Dec 13 06:50:23 crc kubenswrapper[4971]: I1213 06:50:23.826917 4971 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-82xjz" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e0c4af12-edda-41d7-9b44-f87396174ef2\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T06:49:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T06:49:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T06:49:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T06:49:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://0f4577d9e37f860813e221f2e208177d8d383fd04f673ce357a0a405ea29d1e6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-13T06:49:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cnxsq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://7e0d2020e62f1a277cff7047e1907dc747ed393a1f6c983e1ee258bd156c16c7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-13T06:49:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cnxsq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-13T06:49:25Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-82xjz\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-13T06:50:23Z is after 2025-08-24T17:21:41Z" Dec 13 06:50:23 crc kubenswrapper[4971]: I1213 06:50:23.839789 4971 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-wwvpl" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"dddf744b-2122-49ad-a980-105c09636e0f\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T06:49:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T06:49:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T06:50:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T06:50:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c8680aaa9e05095ac7e467c896d9e927bccff908925768ce5ed469ab89a72475\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://4676d897c7fc668587faec563f3551a056557879cee3e865192d2bd28edf62fc\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-12-13T06:50:12Z\\\",\\\"message\\\":\\\"2025-12-13T06:49:27+00:00 [cnibincopy] Successfully copied files in /usr/src/multus-cni/rhel9/bin/ to /host/opt/cni/bin/upgrade_263f091d-d847-4fad-88d3-fbf82c65c231\\\\n2025-12-13T06:49:27+00:00 [cnibincopy] Successfully moved files in /host/opt/cni/bin/upgrade_263f091d-d847-4fad-88d3-fbf82c65c231 to /host/opt/cni/bin/\\\\n2025-12-13T06:49:27Z [verbose] multus-daemon started\\\\n2025-12-13T06:49:27Z [verbose] Readiness Indicator file check\\\\n2025-12-13T06:50:12Z [error] have you checked that your default network is ready? still waiting for readinessindicatorfile @ /host/run/multus/cni/net.d/10-ovn-kubernetes.conf. pollimmediate error: timed out waiting for the condition\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-13T06:49:26Z\\\"}},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-13T06:50:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-czz9v\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-13T06:49:25Z\\\"}}\" for pod \"openshift-multus\"/\"multus-wwvpl\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-13T06:50:23Z is after 2025-08-24T17:21:41Z" Dec 13 06:50:23 crc kubenswrapper[4971]: I1213 06:50:23.858269 4971 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-zlcg5" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"736c6b1a-93dd-4f71-81b6-1c43fcf5f556\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T06:49:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T06:49:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T06:49:25Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T06:49:25Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://325223543b3c7db417f769533559a1ca2b23a723494e1885b0fd0480b5a6399f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-13T06:49:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vzmsr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://f22a4ec756119c4f791fb5105cf1f31df3e548b2b07b1f909a62802e9b6035ae\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-13T06:49:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vzmsr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://f6073140d9f86929b6d4bd2960ede0617a5c9620b269bbab043bae37e0d389f5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-13T06:49:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vzmsr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://693a2ab434c250cb140a866e8068b156df9191414fed1d63bee919a2528ea6e5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-13T06:49:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vzmsr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://219620f51474e60b3e1f880c0972eb612dffc4bad24462bbba2c093ae817d852\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-13T06:49:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vzmsr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://f137f4b441c2402093ab430a9689a46221c4fb4bc70d99fb992f62b0a989c465\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-13T06:49:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vzmsr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://0b454a87c6dd11bd8c3bb3553ee71de32ed6f185984e5057a20f2237b35c45d9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://0b454a87c6dd11bd8c3bb3553ee71de32ed6f185984e5057a20f2237b35c45d9\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-12-13T06:49:56Z\\\",\\\"message\\\":\\\" for pod on switch crc\\\\nI1213 06:49:56.566617 6645 metrics.go:553] Stopping metrics server at address \\\\\\\"127.0.0.1:29103\\\\\\\"\\\\nI1213 06:49:56.566625 6645 base_network_controller_pods.go:477] [default/openshift-network-console/networking-console-plugin-85b44fc459-gdk6g] creating logical port openshift-network-console_networking-console-plugin-85b44fc459-gdk6g for pod on switch crc\\\\nI1213 06:49:56.566569 6645 loadbalancer.go:304] Deleted 0 stale LBs for map[string]string{\\\\\\\"k8s.ovn.org/kind\\\\\\\":\\\\\\\"Service\\\\\\\", \\\\\\\"k8s.ovn.org/owner\\\\\\\":\\\\\\\"openshift-console-operator/metrics\\\\\\\"}\\\\nI1213 06:49:56.566666 6645 services_controller.go:360] Finished syncing service metrics on namespace openshift-console-operator for network=default : 1.804735ms\\\\nI1213 06:49:56.566669 6645 loadbalancer.go:304] Deleted 0 stale LBs for map[string]string{\\\\\\\"k8s.ovn.org/kind\\\\\\\":\\\\\\\"Service\\\\\\\", \\\\\\\"k8s.ovn.org/owner\\\\\\\":\\\\\\\"openshift-apiserver/check-endpoints\\\\\\\"}\\\\nF1213 06:49:56.566698 6645 ovnkube.go:137] failed to run ovnkube: [failed to start network controller: failed to start default network controller: unable to create admin network policy controller, err: could not add Event Handler for anpInformer during admin network policy controller initialization, handler {0x1fcc6e0 0x1fcc3c0 0x1fcc360} was not added to shared informer because it has stopped already, failed to start node network controller: failed \\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-13T06:49:55Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":2,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 20s restarting failed container=ovnkube-controller pod=ovnkube-node-zlcg5_openshift-ovn-kubernetes(736c6b1a-93dd-4f71-81b6-1c43fcf5f556)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vzmsr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://868049c6a3a4ebe4036d1d22e02ba6b76bced2ee7e269bb228accc4a0636678a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-13T06:49:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vzmsr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://451b22df6973ddc8fa2209a65d64d9e41b33eb231bb4adfab88cf0e5943bb246\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://451b22df6973ddc8fa2209a65d64d9e41b33eb231bb4adfab88cf0e5943bb246\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-13T06:49:26Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-13T06:49:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vzmsr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-13T06:49:25Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-zlcg5\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-13T06:50:23Z is after 2025-08-24T17:21:41Z" Dec 13 06:50:23 crc kubenswrapper[4971]: I1213 06:50:23.859576 4971 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 13 06:50:23 crc kubenswrapper[4971]: I1213 06:50:23.859620 4971 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 13 06:50:23 crc kubenswrapper[4971]: I1213 06:50:23.859631 4971 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 13 06:50:23 crc kubenswrapper[4971]: I1213 06:50:23.859658 4971 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 13 06:50:23 crc kubenswrapper[4971]: I1213 06:50:23.859670 4971 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-13T06:50:23Z","lastTransitionTime":"2025-12-13T06:50:23Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 13 06:50:23 crc kubenswrapper[4971]: I1213 06:50:23.875858 4971 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"f9d31b5a-b770-4a15-95c8-35b49677c608\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T06:49:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T06:49:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T06:49:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T06:49:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T06:49:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c9baa042f9dada954faa57b5454cd6b0a12a297e56ddc192c52cff50e42a16fa\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-13T06:49:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://c788de86754f0bf37ee184672fc1dcb388ccebc246d4f15be4b5da661f725f8f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-13T06:49:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://760817511a9b4f96a68dc38d26631250eaab72e1dd8f123e653039b6b4e91f84\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-13T06:49:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b859be0c9feac856e9855c6a038b02b49b32fade97873c46d0bec0a32d8b5147\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-13T06:49:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-13T06:49:03Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-13T06:50:23Z is after 2025-08-24T17:21:41Z" Dec 13 06:50:23 crc kubenswrapper[4971]: I1213 06:50:23.886925 4971 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"8ede2992-9576-4b23-83c2-24ac3b1d3caf\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T06:49:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T06:49:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T06:49:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T06:49:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T06:49:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9e61d25add1723ac286794abfe886358e9d64455c25379f7e3057fc3dd812418\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-13T06:49:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://7fb51d8435b4e3ce0411cae1d6c2824c552f0a33d79de01f969274516b39a5b3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-13T06:49:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://1c35f5e8d8fab6d77b9e2e0dedb195440de5d6af345cd5865abd32caa7fafd60\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-13T06:49:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://38806c430185715371ab9bf30a1af9a9d5d492abd9bb84acbdccc694f56e3a8a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"wait-for-host-port\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://38806c430185715371ab9bf30a1af9a9d5d492abd9bb84acbdccc694f56e3a8a\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-13T06:49:04Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-13T06:49:04Z\\\"}}}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-13T06:49:03Z\\\"}}\" for pod \"openshift-kube-scheduler\"/\"openshift-kube-scheduler-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-13T06:50:23Z is after 2025-08-24T17:21:41Z" Dec 13 06:50:23 crc kubenswrapper[4971]: I1213 06:50:23.901635 4971 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-ncchz" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"de914d21-a688-401e-a66e-861beb6336d9\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T06:49:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T06:49:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T06:49:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T06:49:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://244ffe3dc9d8cf799aff4c4d795bf41554c1e6bb6667fbb0c7df322fb5f96c68\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-13T06:49:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lqmjj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://0b7111af6c5a198d02bb02858837cfb4f3b20828d8d206cd18aa0ffa102d0cd4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://0b7111af6c5a198d02bb02858837cfb4f3b20828d8d206cd18aa0ffa102d0cd4\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-13T06:49:26Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-13T06:49:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lqmjj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b36777fe40e74ae1cecb7b73e204139db5279ba51cef4f1e8c4a4c491d8be403\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://b36777fe40e74ae1cecb7b73e204139db5279ba51cef4f1e8c4a4c491d8be403\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-13T06:49:28Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-13T06:49:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lqmjj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://0597a06cc663e416142311a16941d98297577693a1fe27201ea42c27496cace2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://0597a06cc663e416142311a16941d98297577693a1fe27201ea42c27496cace2\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-13T06:49:29Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-13T06:49:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lqmjj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://3f52c2e22a75c010993ca5809a3a9d1304758d4cb6e83691ef1af15a4b514ac4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://3f52c2e22a75c010993ca5809a3a9d1304758d4cb6e83691ef1af15a4b514ac4\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-13T06:49:30Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-13T06:49:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lqmjj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://9e21ca708c4c4c1228b417cece51299e0792729b7c78884bd18b2dda8fe8355c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://9e21ca708c4c4c1228b417cece51299e0792729b7c78884bd18b2dda8fe8355c\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-13T06:49:31Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-13T06:49:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lqmjj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://74552159f12c87cdda09147692b400ccac0d4e43c1ed72fe8462c4249ad69c7b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://74552159f12c87cdda09147692b400ccac0d4e43c1ed72fe8462c4249ad69c7b\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-13T06:49:32Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-13T06:49:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lqmjj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-13T06:49:25Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-ncchz\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-13T06:50:23Z is after 2025-08-24T17:21:41Z" Dec 13 06:50:23 crc kubenswrapper[4971]: I1213 06:50:23.910474 4971 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-gkcd8" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"13f5c327-f95e-42b4-8783-218500c17694\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T06:49:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T06:49:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T06:49:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T06:49:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://77e846e36f1cb40d03471dc03de11b6c3da448baa407aad38093fae9a801cf92\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-13T06:49:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gvlhn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-13T06:49:27Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-gkcd8\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-13T06:50:23Z is after 2025-08-24T17:21:41Z" Dec 13 06:50:23 crc kubenswrapper[4971]: I1213 06:50:23.919386 4971 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-x98k7" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"c7185dec-3240-42e8-9d90-6182a5e8b65c\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T06:49:39Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T06:49:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T06:49:39Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T06:49:39Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bj2gn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bj2gn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-13T06:49:39Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-x98k7\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-13T06:50:23Z is after 2025-08-24T17:21:41Z" Dec 13 06:50:23 crc kubenswrapper[4971]: I1213 06:50:23.929713 4971 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5ef36745-87bb-4be8-b941-aa8f865a5c83\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T06:49:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T06:49:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T06:49:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T06:49:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T06:49:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://35b737b0d2c5c5503ba2819f34d2de0599792a77beb1fbeb561d44efd1f212c5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-crio\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-13T06:49:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kube\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"var-lib-kubelet\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f0d864149bfcbb6227e6a6fbde23fff650d688850978eae0caf04515eb21cd85\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://f0d864149bfcbb6227e6a6fbde23fff650d688850978eae0caf04515eb21cd85\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-13T06:49:04Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-13T06:49:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var\\\",\\\"name\\\":\\\"var-lib-kubelet\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-13T06:49:03Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"kube-rbac-proxy-crio-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-13T06:50:23Z is after 2025-08-24T17:21:41Z" Dec 13 06:50:23 crc kubenswrapper[4971]: I1213 06:50:23.948129 4971 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"1ed9f82e-fe5c-4b76-a619-62e5abd88d3b\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T06:49:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T06:49:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T06:49:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T06:49:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T06:49:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e457329a78a2d24337f994e8a0a1f75a2a667244aa75eb84613d1e7138364aac\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-13T06:49:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://eb014de8140deef097d9bd94f3b9fa4ab3558bc4449bef84b83bf62bfabf5836\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-13T06:49:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://2c438cc74d555f7c9db4569d1c56b78b1088ed8105697d2c0f31e6c5136f3dd0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-13T06:49:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://aa6fa6e4340dcd066252df3c56d79a3a39ed531d5eea8d1d5941d6952368d3af\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-13T06:49:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://c4471b98854c52d5cd084d0ba3e64e5b08e9518ec3c372cce2c875a440522457\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-13T06:49:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7d2746335c98dae3a0f43024d34cb62ebcb42a01c8f9ce18d52b29d4c45598e5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://7d2746335c98dae3a0f43024d34cb62ebcb42a01c8f9ce18d52b29d4c45598e5\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-13T06:49:04Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-13T06:49:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://bc0e77d10a738128bea82122496bc42235fc1688f8907756d93e61b651902b57\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://bc0e77d10a738128bea82122496bc42235fc1688f8907756d93e61b651902b57\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-13T06:49:05Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-13T06:49:05Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://f821e418cba5e832da8839268bcbbd9630bdf319a212d5382bfa2c4b0b73a268\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://f821e418cba5e832da8839268bcbbd9630bdf319a212d5382bfa2c4b0b73a268\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-13T06:49:07Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-13T06:49:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-13T06:49:03Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-13T06:50:23Z is after 2025-08-24T17:21:41Z" Dec 13 06:50:23 crc kubenswrapper[4971]: I1213 06:50:23.961980 4971 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 13 06:50:23 crc kubenswrapper[4971]: I1213 06:50:23.962029 4971 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 13 06:50:23 crc kubenswrapper[4971]: I1213 06:50:23.962040 4971 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 13 06:50:23 crc kubenswrapper[4971]: I1213 06:50:23.962057 4971 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 13 06:50:23 crc kubenswrapper[4971]: I1213 06:50:23.962067 4971 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-13T06:50:23Z","lastTransitionTime":"2025-12-13T06:50:23Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 13 06:50:23 crc kubenswrapper[4971]: I1213 06:50:23.962702 4971 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-13T06:49:24Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-13T06:49:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-13T06:49:24Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4328cd6a8464ed468229bbfbc8265f969427916d72ebb92237dbe68efe70fa47\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-13T06:49:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://9d715ca82d9b5a594e41f2b98da6f4277da0e4e1dec114969f901c00f6d7553d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-13T06:49:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-13T06:50:23Z is after 2025-08-24T17:21:41Z" Dec 13 06:50:23 crc kubenswrapper[4971]: I1213 06:50:23.973294 4971 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-5rzh5" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"6bf4591d-2483-4132-9672-d0fd6c995968\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T06:49:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T06:49:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T06:49:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T06:49:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://50e2cb14e4ffb603acb06e2281615a99f06182d345c4e70f2446fa2077033bbb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-13T06:49:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dfxqn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-13T06:49:25Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-5rzh5\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-13T06:50:23Z is after 2025-08-24T17:21:41Z" Dec 13 06:50:23 crc kubenswrapper[4971]: I1213 06:50:23.982693 4971 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-z572s" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"b4146d3d-4034-49a9-bd37-164856255137\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T06:49:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T06:49:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T06:49:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T06:49:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://46fd4e1070835cfbf7f35d497eff7f047473fc4048ee07812925df5ebdf55342\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-13T06:49:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-j6l9b\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://dc79a0968f85723e6ad0dc269c8da1f7472adb1bb2b87edd017ee2295bfbf89c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-13T06:49:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-j6l9b\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-13T06:49:38Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-z572s\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-13T06:50:23Z is after 2025-08-24T17:21:41Z" Dec 13 06:50:23 crc kubenswrapper[4971]: I1213 06:50:23.994426 4971 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3ad3268a-a982-428c-9c06-4aaeddd372d9\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T06:49:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T06:49:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T06:49:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T06:49:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T06:49:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3716afdf2bc693be80a09fc9f7727d334f435caf1d4c8ec4811b3a1f1652cdfe\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-13T06:49:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://30df3abf8c79db8af52c6a918f73e1ce410ddafcf341636424c267e1eeb195a7\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-13T06:49:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://ccdf65f6d2b1eb3647a03ba37c305c32e34addfdf79ff937c8dddaa3e5d5b5c6\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-13T06:49:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://770b73dd0c2b6f38744e2c1cd6aa186a69bb87d1cd015a603ad2e33be06a8f6b\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://b3c1bab3d53f975b2e24582ce2771bc26d600e0dd7c30339dcfd18a071162ad4\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-12-13T06:49:23Z\\\",\\\"message\\\":\\\"file observer\\\\nW1213 06:49:23.283403 1 builder.go:272] unable to get owner reference (falling back to namespace): pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\nI1213 06:49:23.284336 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1213 06:49:23.286978 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-981773671/tls.crt::/tmp/serving-cert-981773671/tls.key\\\\\\\"\\\\nI1213 06:49:23.775072 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1213 06:49:23.778615 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1213 06:49:23.778645 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1213 06:49:23.778680 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1213 06:49:23.778691 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1213 06:49:23.785058 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1213 06:49:23.786150 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1213 06:49:23.786196 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1213 06:49:23.786223 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1213 06:49:23.786246 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1213 06:49:23.786269 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1213 06:49:23.786291 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI1213 06:49:23.785608 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF1213 06:49:23.787563 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-13T06:49:08Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-13T06:49:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://9acd07066deed483bd7fa61ee45fda09539d110717a134a56827f871186f92fc\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-13T06:49:07Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://38c91cfbaecfdbd3e3b663422a91ab4294610afb6bbf9ebdb77db24e2b76c682\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://38c91cfbaecfdbd3e3b663422a91ab4294610afb6bbf9ebdb77db24e2b76c682\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-13T06:49:05Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-13T06:49:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-13T06:49:03Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-13T06:50:23Z is after 2025-08-24T17:21:41Z" Dec 13 06:50:24 crc kubenswrapper[4971]: I1213 06:50:24.006485 4971 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-13T06:49:23Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-13T06:49:23Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-13T06:50:24Z is after 2025-08-24T17:21:41Z" Dec 13 06:50:24 crc kubenswrapper[4971]: I1213 06:50:24.019023 4971 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-13T06:49:23Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-13T06:49:23Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-13T06:50:24Z is after 2025-08-24T17:21:41Z" Dec 13 06:50:24 crc kubenswrapper[4971]: I1213 06:50:24.064733 4971 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 13 06:50:24 crc kubenswrapper[4971]: I1213 06:50:24.064775 4971 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 13 06:50:24 crc kubenswrapper[4971]: I1213 06:50:24.064786 4971 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 13 06:50:24 crc kubenswrapper[4971]: I1213 06:50:24.064804 4971 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 13 06:50:24 crc kubenswrapper[4971]: I1213 06:50:24.064816 4971 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-13T06:50:24Z","lastTransitionTime":"2025-12-13T06:50:24Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 13 06:50:24 crc kubenswrapper[4971]: I1213 06:50:24.166988 4971 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 13 06:50:24 crc kubenswrapper[4971]: I1213 06:50:24.167068 4971 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 13 06:50:24 crc kubenswrapper[4971]: I1213 06:50:24.167093 4971 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 13 06:50:24 crc kubenswrapper[4971]: I1213 06:50:24.167123 4971 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 13 06:50:24 crc kubenswrapper[4971]: I1213 06:50:24.167145 4971 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-13T06:50:24Z","lastTransitionTime":"2025-12-13T06:50:24Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 13 06:50:24 crc kubenswrapper[4971]: I1213 06:50:24.199673 4971 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-zlcg5_736c6b1a-93dd-4f71-81b6-1c43fcf5f556/ovnkube-controller/2.log" Dec 13 06:50:24 crc kubenswrapper[4971]: I1213 06:50:24.202037 4971 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-zlcg5" event={"ID":"736c6b1a-93dd-4f71-81b6-1c43fcf5f556","Type":"ContainerStarted","Data":"7dd9913a65acba73ae1f32f4398207c5e276a1e979d86f9e720e4fd761bae4a3"} Dec 13 06:50:24 crc kubenswrapper[4971]: I1213 06:50:24.202422 4971 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-ovn-kubernetes/ovnkube-node-zlcg5" Dec 13 06:50:24 crc kubenswrapper[4971]: I1213 06:50:24.212991 4971 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-13T06:49:23Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-13T06:49:23Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-13T06:50:24Z is after 2025-08-24T17:21:41Z" Dec 13 06:50:24 crc kubenswrapper[4971]: I1213 06:50:24.223846 4971 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3ad3268a-a982-428c-9c06-4aaeddd372d9\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T06:49:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T06:49:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T06:49:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T06:49:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T06:49:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3716afdf2bc693be80a09fc9f7727d334f435caf1d4c8ec4811b3a1f1652cdfe\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-13T06:49:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://30df3abf8c79db8af52c6a918f73e1ce410ddafcf341636424c267e1eeb195a7\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-13T06:49:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://ccdf65f6d2b1eb3647a03ba37c305c32e34addfdf79ff937c8dddaa3e5d5b5c6\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-13T06:49:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://770b73dd0c2b6f38744e2c1cd6aa186a69bb87d1cd015a603ad2e33be06a8f6b\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://b3c1bab3d53f975b2e24582ce2771bc26d600e0dd7c30339dcfd18a071162ad4\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-12-13T06:49:23Z\\\",\\\"message\\\":\\\"file observer\\\\nW1213 06:49:23.283403 1 builder.go:272] unable to get owner reference (falling back to namespace): pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\nI1213 06:49:23.284336 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1213 06:49:23.286978 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-981773671/tls.crt::/tmp/serving-cert-981773671/tls.key\\\\\\\"\\\\nI1213 06:49:23.775072 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1213 06:49:23.778615 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1213 06:49:23.778645 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1213 06:49:23.778680 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1213 06:49:23.778691 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1213 06:49:23.785058 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1213 06:49:23.786150 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1213 06:49:23.786196 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1213 06:49:23.786223 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1213 06:49:23.786246 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1213 06:49:23.786269 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1213 06:49:23.786291 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI1213 06:49:23.785608 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF1213 06:49:23.787563 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-13T06:49:08Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-13T06:49:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://9acd07066deed483bd7fa61ee45fda09539d110717a134a56827f871186f92fc\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-13T06:49:07Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://38c91cfbaecfdbd3e3b663422a91ab4294610afb6bbf9ebdb77db24e2b76c682\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://38c91cfbaecfdbd3e3b663422a91ab4294610afb6bbf9ebdb77db24e2b76c682\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-13T06:49:05Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-13T06:49:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-13T06:49:03Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-13T06:50:24Z is after 2025-08-24T17:21:41Z" Dec 13 06:50:24 crc kubenswrapper[4971]: I1213 06:50:24.234570 4971 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-13T06:49:23Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-13T06:49:23Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-13T06:50:24Z is after 2025-08-24T17:21:41Z" Dec 13 06:50:24 crc kubenswrapper[4971]: I1213 06:50:24.245617 4971 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-13T06:49:26Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-13T06:49:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-13T06:49:26Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9c8a0d1db0e754462d805e82daf5f52ffd5457507a30456535ec59e096c9d092\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-13T06:49:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-13T06:50:24Z is after 2025-08-24T17:21:41Z" Dec 13 06:50:24 crc kubenswrapper[4971]: I1213 06:50:24.258630 4971 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-82xjz" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e0c4af12-edda-41d7-9b44-f87396174ef2\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T06:49:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T06:49:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T06:49:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T06:49:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://0f4577d9e37f860813e221f2e208177d8d383fd04f673ce357a0a405ea29d1e6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-13T06:49:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cnxsq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://7e0d2020e62f1a277cff7047e1907dc747ed393a1f6c983e1ee258bd156c16c7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-13T06:49:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cnxsq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-13T06:49:25Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-82xjz\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-13T06:50:24Z is after 2025-08-24T17:21:41Z" Dec 13 06:50:24 crc kubenswrapper[4971]: I1213 06:50:24.269987 4971 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 13 06:50:24 crc kubenswrapper[4971]: I1213 06:50:24.270073 4971 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 13 06:50:24 crc kubenswrapper[4971]: I1213 06:50:24.270090 4971 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 13 06:50:24 crc kubenswrapper[4971]: I1213 06:50:24.270112 4971 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 13 06:50:24 crc kubenswrapper[4971]: I1213 06:50:24.270131 4971 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-13T06:50:24Z","lastTransitionTime":"2025-12-13T06:50:24Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 13 06:50:24 crc kubenswrapper[4971]: I1213 06:50:24.272007 4971 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-wwvpl" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"dddf744b-2122-49ad-a980-105c09636e0f\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T06:49:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T06:49:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T06:50:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T06:50:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c8680aaa9e05095ac7e467c896d9e927bccff908925768ce5ed469ab89a72475\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://4676d897c7fc668587faec563f3551a056557879cee3e865192d2bd28edf62fc\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-12-13T06:50:12Z\\\",\\\"message\\\":\\\"2025-12-13T06:49:27+00:00 [cnibincopy] Successfully copied files in /usr/src/multus-cni/rhel9/bin/ to /host/opt/cni/bin/upgrade_263f091d-d847-4fad-88d3-fbf82c65c231\\\\n2025-12-13T06:49:27+00:00 [cnibincopy] Successfully moved files in /host/opt/cni/bin/upgrade_263f091d-d847-4fad-88d3-fbf82c65c231 to /host/opt/cni/bin/\\\\n2025-12-13T06:49:27Z [verbose] multus-daemon started\\\\n2025-12-13T06:49:27Z [verbose] Readiness Indicator file check\\\\n2025-12-13T06:50:12Z [error] have you checked that your default network is ready? still waiting for readinessindicatorfile @ /host/run/multus/cni/net.d/10-ovn-kubernetes.conf. pollimmediate error: timed out waiting for the condition\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-13T06:49:26Z\\\"}},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-13T06:50:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-czz9v\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-13T06:49:25Z\\\"}}\" for pod \"openshift-multus\"/\"multus-wwvpl\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-13T06:50:24Z is after 2025-08-24T17:21:41Z" Dec 13 06:50:24 crc kubenswrapper[4971]: I1213 06:50:24.289618 4971 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-zlcg5" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"736c6b1a-93dd-4f71-81b6-1c43fcf5f556\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T06:49:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T06:49:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T06:49:25Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T06:49:25Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://325223543b3c7db417f769533559a1ca2b23a723494e1885b0fd0480b5a6399f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-13T06:49:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vzmsr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://f22a4ec756119c4f791fb5105cf1f31df3e548b2b07b1f909a62802e9b6035ae\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-13T06:49:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vzmsr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://f6073140d9f86929b6d4bd2960ede0617a5c9620b269bbab043bae37e0d389f5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-13T06:49:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vzmsr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://693a2ab434c250cb140a866e8068b156df9191414fed1d63bee919a2528ea6e5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-13T06:49:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vzmsr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://219620f51474e60b3e1f880c0972eb612dffc4bad24462bbba2c093ae817d852\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-13T06:49:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vzmsr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://f137f4b441c2402093ab430a9689a46221c4fb4bc70d99fb992f62b0a989c465\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-13T06:49:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vzmsr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://7dd9913a65acba73ae1f32f4398207c5e276a1e979d86f9e720e4fd761bae4a3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://0b454a87c6dd11bd8c3bb3553ee71de32ed6f185984e5057a20f2237b35c45d9\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-12-13T06:49:56Z\\\",\\\"message\\\":\\\" for pod on switch crc\\\\nI1213 06:49:56.566617 6645 metrics.go:553] Stopping metrics server at address \\\\\\\"127.0.0.1:29103\\\\\\\"\\\\nI1213 06:49:56.566625 6645 base_network_controller_pods.go:477] [default/openshift-network-console/networking-console-plugin-85b44fc459-gdk6g] creating logical port openshift-network-console_networking-console-plugin-85b44fc459-gdk6g for pod on switch crc\\\\nI1213 06:49:56.566569 6645 loadbalancer.go:304] Deleted 0 stale LBs for map[string]string{\\\\\\\"k8s.ovn.org/kind\\\\\\\":\\\\\\\"Service\\\\\\\", \\\\\\\"k8s.ovn.org/owner\\\\\\\":\\\\\\\"openshift-console-operator/metrics\\\\\\\"}\\\\nI1213 06:49:56.566666 6645 services_controller.go:360] Finished syncing service metrics on namespace openshift-console-operator for network=default : 1.804735ms\\\\nI1213 06:49:56.566669 6645 loadbalancer.go:304] Deleted 0 stale LBs for map[string]string{\\\\\\\"k8s.ovn.org/kind\\\\\\\":\\\\\\\"Service\\\\\\\", \\\\\\\"k8s.ovn.org/owner\\\\\\\":\\\\\\\"openshift-apiserver/check-endpoints\\\\\\\"}\\\\nF1213 06:49:56.566698 6645 ovnkube.go:137] failed to run ovnkube: [failed to start network controller: failed to start default network controller: unable to create admin network policy controller, err: could not add Event Handler for anpInformer during admin network policy controller initialization, handler {0x1fcc6e0 0x1fcc3c0 0x1fcc360} was not added to shared informer because it has stopped already, failed to start node network controller: failed \\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-13T06:49:55Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-13T06:50:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vzmsr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://868049c6a3a4ebe4036d1d22e02ba6b76bced2ee7e269bb228accc4a0636678a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-13T06:49:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vzmsr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://451b22df6973ddc8fa2209a65d64d9e41b33eb231bb4adfab88cf0e5943bb246\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://451b22df6973ddc8fa2209a65d64d9e41b33eb231bb4adfab88cf0e5943bb246\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-13T06:49:26Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-13T06:49:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vzmsr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-13T06:49:25Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-zlcg5\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-13T06:50:24Z is after 2025-08-24T17:21:41Z" Dec 13 06:50:24 crc kubenswrapper[4971]: I1213 06:50:24.304460 4971 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-13T06:49:24Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-13T06:49:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-13T06:49:24Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d0b2ecf0d5b22f23217f42f259ae7ffb6359555b22c30eaebf667eaf254b3bab\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-13T06:49:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-13T06:50:24Z is after 2025-08-24T17:21:41Z" Dec 13 06:50:24 crc kubenswrapper[4971]: I1213 06:50:24.316612 4971 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-13T06:49:23Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-13T06:49:23Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-13T06:50:24Z is after 2025-08-24T17:21:41Z" Dec 13 06:50:24 crc kubenswrapper[4971]: I1213 06:50:24.331556 4971 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-ncchz" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"de914d21-a688-401e-a66e-861beb6336d9\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T06:49:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T06:49:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T06:49:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T06:49:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://244ffe3dc9d8cf799aff4c4d795bf41554c1e6bb6667fbb0c7df322fb5f96c68\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-13T06:49:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lqmjj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://0b7111af6c5a198d02bb02858837cfb4f3b20828d8d206cd18aa0ffa102d0cd4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://0b7111af6c5a198d02bb02858837cfb4f3b20828d8d206cd18aa0ffa102d0cd4\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-13T06:49:26Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-13T06:49:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lqmjj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b36777fe40e74ae1cecb7b73e204139db5279ba51cef4f1e8c4a4c491d8be403\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://b36777fe40e74ae1cecb7b73e204139db5279ba51cef4f1e8c4a4c491d8be403\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-13T06:49:28Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-13T06:49:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lqmjj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://0597a06cc663e416142311a16941d98297577693a1fe27201ea42c27496cace2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://0597a06cc663e416142311a16941d98297577693a1fe27201ea42c27496cace2\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-13T06:49:29Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-13T06:49:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lqmjj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://3f52c2e22a75c010993ca5809a3a9d1304758d4cb6e83691ef1af15a4b514ac4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://3f52c2e22a75c010993ca5809a3a9d1304758d4cb6e83691ef1af15a4b514ac4\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-13T06:49:30Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-13T06:49:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lqmjj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://9e21ca708c4c4c1228b417cece51299e0792729b7c78884bd18b2dda8fe8355c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://9e21ca708c4c4c1228b417cece51299e0792729b7c78884bd18b2dda8fe8355c\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-13T06:49:31Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-13T06:49:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lqmjj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://74552159f12c87cdda09147692b400ccac0d4e43c1ed72fe8462c4249ad69c7b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://74552159f12c87cdda09147692b400ccac0d4e43c1ed72fe8462c4249ad69c7b\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-13T06:49:32Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-13T06:49:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lqmjj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-13T06:49:25Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-ncchz\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-13T06:50:24Z is after 2025-08-24T17:21:41Z" Dec 13 06:50:24 crc kubenswrapper[4971]: I1213 06:50:24.341627 4971 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-gkcd8" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"13f5c327-f95e-42b4-8783-218500c17694\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T06:49:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T06:49:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T06:49:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T06:49:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://77e846e36f1cb40d03471dc03de11b6c3da448baa407aad38093fae9a801cf92\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-13T06:49:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gvlhn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-13T06:49:27Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-gkcd8\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-13T06:50:24Z is after 2025-08-24T17:21:41Z" Dec 13 06:50:24 crc kubenswrapper[4971]: I1213 06:50:24.351736 4971 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-x98k7" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"c7185dec-3240-42e8-9d90-6182a5e8b65c\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T06:49:39Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T06:49:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T06:49:39Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T06:49:39Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bj2gn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bj2gn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-13T06:49:39Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-x98k7\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-13T06:50:24Z is after 2025-08-24T17:21:41Z" Dec 13 06:50:24 crc kubenswrapper[4971]: I1213 06:50:24.365892 4971 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"f9d31b5a-b770-4a15-95c8-35b49677c608\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T06:49:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T06:49:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T06:49:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T06:49:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T06:49:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c9baa042f9dada954faa57b5454cd6b0a12a297e56ddc192c52cff50e42a16fa\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-13T06:49:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://c788de86754f0bf37ee184672fc1dcb388ccebc246d4f15be4b5da661f725f8f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-13T06:49:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://760817511a9b4f96a68dc38d26631250eaab72e1dd8f123e653039b6b4e91f84\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-13T06:49:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b859be0c9feac856e9855c6a038b02b49b32fade97873c46d0bec0a32d8b5147\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-13T06:49:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-13T06:49:03Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-13T06:50:24Z is after 2025-08-24T17:21:41Z" Dec 13 06:50:24 crc kubenswrapper[4971]: I1213 06:50:24.372381 4971 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 13 06:50:24 crc kubenswrapper[4971]: I1213 06:50:24.372440 4971 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 13 06:50:24 crc kubenswrapper[4971]: I1213 06:50:24.372451 4971 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 13 06:50:24 crc kubenswrapper[4971]: I1213 06:50:24.372467 4971 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 13 06:50:24 crc kubenswrapper[4971]: I1213 06:50:24.372478 4971 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-13T06:50:24Z","lastTransitionTime":"2025-12-13T06:50:24Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 13 06:50:24 crc kubenswrapper[4971]: I1213 06:50:24.376841 4971 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"8ede2992-9576-4b23-83c2-24ac3b1d3caf\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T06:49:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T06:49:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T06:49:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T06:49:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T06:49:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9e61d25add1723ac286794abfe886358e9d64455c25379f7e3057fc3dd812418\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-13T06:49:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://7fb51d8435b4e3ce0411cae1d6c2824c552f0a33d79de01f969274516b39a5b3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-13T06:49:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://1c35f5e8d8fab6d77b9e2e0dedb195440de5d6af345cd5865abd32caa7fafd60\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-13T06:49:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://38806c430185715371ab9bf30a1af9a9d5d492abd9bb84acbdccc694f56e3a8a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"wait-for-host-port\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://38806c430185715371ab9bf30a1af9a9d5d492abd9bb84acbdccc694f56e3a8a\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-13T06:49:04Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-13T06:49:04Z\\\"}}}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-13T06:49:03Z\\\"}}\" for pod \"openshift-kube-scheduler\"/\"openshift-kube-scheduler-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-13T06:50:24Z is after 2025-08-24T17:21:41Z" Dec 13 06:50:24 crc kubenswrapper[4971]: I1213 06:50:24.388606 4971 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-13T06:49:24Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-13T06:49:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-13T06:49:24Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4328cd6a8464ed468229bbfbc8265f969427916d72ebb92237dbe68efe70fa47\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-13T06:49:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://9d715ca82d9b5a594e41f2b98da6f4277da0e4e1dec114969f901c00f6d7553d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-13T06:49:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-13T06:50:24Z is after 2025-08-24T17:21:41Z" Dec 13 06:50:24 crc kubenswrapper[4971]: I1213 06:50:24.399431 4971 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-5rzh5" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"6bf4591d-2483-4132-9672-d0fd6c995968\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T06:49:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T06:49:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T06:49:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T06:49:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://50e2cb14e4ffb603acb06e2281615a99f06182d345c4e70f2446fa2077033bbb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-13T06:49:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dfxqn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-13T06:49:25Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-5rzh5\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-13T06:50:24Z is after 2025-08-24T17:21:41Z" Dec 13 06:50:24 crc kubenswrapper[4971]: I1213 06:50:24.412408 4971 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-z572s" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"b4146d3d-4034-49a9-bd37-164856255137\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T06:49:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T06:49:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T06:49:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T06:49:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://46fd4e1070835cfbf7f35d497eff7f047473fc4048ee07812925df5ebdf55342\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-13T06:49:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-j6l9b\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://dc79a0968f85723e6ad0dc269c8da1f7472adb1bb2b87edd017ee2295bfbf89c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-13T06:49:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-j6l9b\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-13T06:49:38Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-z572s\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-13T06:50:24Z is after 2025-08-24T17:21:41Z" Dec 13 06:50:24 crc kubenswrapper[4971]: I1213 06:50:24.423907 4971 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5ef36745-87bb-4be8-b941-aa8f865a5c83\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T06:49:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T06:49:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T06:49:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T06:49:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T06:49:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://35b737b0d2c5c5503ba2819f34d2de0599792a77beb1fbeb561d44efd1f212c5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-crio\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-13T06:49:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kube\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"var-lib-kubelet\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f0d864149bfcbb6227e6a6fbde23fff650d688850978eae0caf04515eb21cd85\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://f0d864149bfcbb6227e6a6fbde23fff650d688850978eae0caf04515eb21cd85\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-13T06:49:04Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-13T06:49:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var\\\",\\\"name\\\":\\\"var-lib-kubelet\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-13T06:49:03Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"kube-rbac-proxy-crio-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-13T06:50:24Z is after 2025-08-24T17:21:41Z" Dec 13 06:50:24 crc kubenswrapper[4971]: I1213 06:50:24.445209 4971 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"1ed9f82e-fe5c-4b76-a619-62e5abd88d3b\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T06:49:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T06:49:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T06:49:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T06:49:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T06:49:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e457329a78a2d24337f994e8a0a1f75a2a667244aa75eb84613d1e7138364aac\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-13T06:49:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://eb014de8140deef097d9bd94f3b9fa4ab3558bc4449bef84b83bf62bfabf5836\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-13T06:49:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://2c438cc74d555f7c9db4569d1c56b78b1088ed8105697d2c0f31e6c5136f3dd0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-13T06:49:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://aa6fa6e4340dcd066252df3c56d79a3a39ed531d5eea8d1d5941d6952368d3af\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-13T06:49:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://c4471b98854c52d5cd084d0ba3e64e5b08e9518ec3c372cce2c875a440522457\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-13T06:49:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7d2746335c98dae3a0f43024d34cb62ebcb42a01c8f9ce18d52b29d4c45598e5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://7d2746335c98dae3a0f43024d34cb62ebcb42a01c8f9ce18d52b29d4c45598e5\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-13T06:49:04Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-13T06:49:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://bc0e77d10a738128bea82122496bc42235fc1688f8907756d93e61b651902b57\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://bc0e77d10a738128bea82122496bc42235fc1688f8907756d93e61b651902b57\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-13T06:49:05Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-13T06:49:05Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://f821e418cba5e832da8839268bcbbd9630bdf319a212d5382bfa2c4b0b73a268\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://f821e418cba5e832da8839268bcbbd9630bdf319a212d5382bfa2c4b0b73a268\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-13T06:49:07Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-13T06:49:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-13T06:49:03Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-13T06:50:24Z is after 2025-08-24T17:21:41Z" Dec 13 06:50:24 crc kubenswrapper[4971]: I1213 06:50:24.476612 4971 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 13 06:50:24 crc kubenswrapper[4971]: I1213 06:50:24.476656 4971 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 13 06:50:24 crc kubenswrapper[4971]: I1213 06:50:24.476669 4971 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 13 06:50:24 crc kubenswrapper[4971]: I1213 06:50:24.476695 4971 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 13 06:50:24 crc kubenswrapper[4971]: I1213 06:50:24.476708 4971 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-13T06:50:24Z","lastTransitionTime":"2025-12-13T06:50:24Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 13 06:50:24 crc kubenswrapper[4971]: I1213 06:50:24.579717 4971 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 13 06:50:24 crc kubenswrapper[4971]: I1213 06:50:24.579756 4971 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 13 06:50:24 crc kubenswrapper[4971]: I1213 06:50:24.579766 4971 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 13 06:50:24 crc kubenswrapper[4971]: I1213 06:50:24.579783 4971 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 13 06:50:24 crc kubenswrapper[4971]: I1213 06:50:24.579795 4971 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-13T06:50:24Z","lastTransitionTime":"2025-12-13T06:50:24Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 13 06:50:24 crc kubenswrapper[4971]: I1213 06:50:24.682382 4971 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 13 06:50:24 crc kubenswrapper[4971]: I1213 06:50:24.682451 4971 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 13 06:50:24 crc kubenswrapper[4971]: I1213 06:50:24.682468 4971 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 13 06:50:24 crc kubenswrapper[4971]: I1213 06:50:24.682492 4971 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 13 06:50:24 crc kubenswrapper[4971]: I1213 06:50:24.682509 4971 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-13T06:50:24Z","lastTransitionTime":"2025-12-13T06:50:24Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 13 06:50:24 crc kubenswrapper[4971]: I1213 06:50:24.786357 4971 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 13 06:50:24 crc kubenswrapper[4971]: I1213 06:50:24.786455 4971 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 13 06:50:24 crc kubenswrapper[4971]: I1213 06:50:24.786508 4971 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 13 06:50:24 crc kubenswrapper[4971]: I1213 06:50:24.786585 4971 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 13 06:50:24 crc kubenswrapper[4971]: I1213 06:50:24.786603 4971 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-13T06:50:24Z","lastTransitionTime":"2025-12-13T06:50:24Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 13 06:50:24 crc kubenswrapper[4971]: I1213 06:50:24.889128 4971 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 13 06:50:24 crc kubenswrapper[4971]: I1213 06:50:24.889157 4971 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 13 06:50:24 crc kubenswrapper[4971]: I1213 06:50:24.889166 4971 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 13 06:50:24 crc kubenswrapper[4971]: I1213 06:50:24.889195 4971 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 13 06:50:24 crc kubenswrapper[4971]: I1213 06:50:24.889205 4971 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-13T06:50:24Z","lastTransitionTime":"2025-12-13T06:50:24Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 13 06:50:24 crc kubenswrapper[4971]: I1213 06:50:24.991657 4971 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 13 06:50:24 crc kubenswrapper[4971]: I1213 06:50:24.991685 4971 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 13 06:50:24 crc kubenswrapper[4971]: I1213 06:50:24.991693 4971 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 13 06:50:24 crc kubenswrapper[4971]: I1213 06:50:24.991706 4971 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 13 06:50:24 crc kubenswrapper[4971]: I1213 06:50:24.991714 4971 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-13T06:50:24Z","lastTransitionTime":"2025-12-13T06:50:24Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 13 06:50:25 crc kubenswrapper[4971]: I1213 06:50:25.094237 4971 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 13 06:50:25 crc kubenswrapper[4971]: I1213 06:50:25.094289 4971 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 13 06:50:25 crc kubenswrapper[4971]: I1213 06:50:25.094300 4971 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 13 06:50:25 crc kubenswrapper[4971]: I1213 06:50:25.094316 4971 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 13 06:50:25 crc kubenswrapper[4971]: I1213 06:50:25.094327 4971 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-13T06:50:25Z","lastTransitionTime":"2025-12-13T06:50:25Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 13 06:50:25 crc kubenswrapper[4971]: I1213 06:50:25.196892 4971 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 13 06:50:25 crc kubenswrapper[4971]: I1213 06:50:25.196971 4971 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 13 06:50:25 crc kubenswrapper[4971]: I1213 06:50:25.196989 4971 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 13 06:50:25 crc kubenswrapper[4971]: I1213 06:50:25.197013 4971 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 13 06:50:25 crc kubenswrapper[4971]: I1213 06:50:25.197030 4971 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-13T06:50:25Z","lastTransitionTime":"2025-12-13T06:50:25Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 13 06:50:25 crc kubenswrapper[4971]: I1213 06:50:25.210115 4971 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-zlcg5_736c6b1a-93dd-4f71-81b6-1c43fcf5f556/ovnkube-controller/3.log" Dec 13 06:50:25 crc kubenswrapper[4971]: I1213 06:50:25.211512 4971 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-zlcg5_736c6b1a-93dd-4f71-81b6-1c43fcf5f556/ovnkube-controller/2.log" Dec 13 06:50:25 crc kubenswrapper[4971]: I1213 06:50:25.216824 4971 generic.go:334] "Generic (PLEG): container finished" podID="736c6b1a-93dd-4f71-81b6-1c43fcf5f556" containerID="7dd9913a65acba73ae1f32f4398207c5e276a1e979d86f9e720e4fd761bae4a3" exitCode=1 Dec 13 06:50:25 crc kubenswrapper[4971]: I1213 06:50:25.216903 4971 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-zlcg5" event={"ID":"736c6b1a-93dd-4f71-81b6-1c43fcf5f556","Type":"ContainerDied","Data":"7dd9913a65acba73ae1f32f4398207c5e276a1e979d86f9e720e4fd761bae4a3"} Dec 13 06:50:25 crc kubenswrapper[4971]: I1213 06:50:25.217003 4971 scope.go:117] "RemoveContainer" containerID="0b454a87c6dd11bd8c3bb3553ee71de32ed6f185984e5057a20f2237b35c45d9" Dec 13 06:50:25 crc kubenswrapper[4971]: I1213 06:50:25.217661 4971 scope.go:117] "RemoveContainer" containerID="7dd9913a65acba73ae1f32f4398207c5e276a1e979d86f9e720e4fd761bae4a3" Dec 13 06:50:25 crc kubenswrapper[4971]: E1213 06:50:25.217851 4971 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"ovnkube-controller\" with CrashLoopBackOff: \"back-off 40s restarting failed container=ovnkube-controller pod=ovnkube-node-zlcg5_openshift-ovn-kubernetes(736c6b1a-93dd-4f71-81b6-1c43fcf5f556)\"" pod="openshift-ovn-kubernetes/ovnkube-node-zlcg5" podUID="736c6b1a-93dd-4f71-81b6-1c43fcf5f556" Dec 13 06:50:25 crc kubenswrapper[4971]: I1213 06:50:25.240899 4971 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3ad3268a-a982-428c-9c06-4aaeddd372d9\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T06:49:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T06:49:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T06:49:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T06:49:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T06:49:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3716afdf2bc693be80a09fc9f7727d334f435caf1d4c8ec4811b3a1f1652cdfe\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-13T06:49:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://30df3abf8c79db8af52c6a918f73e1ce410ddafcf341636424c267e1eeb195a7\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-13T06:49:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://ccdf65f6d2b1eb3647a03ba37c305c32e34addfdf79ff937c8dddaa3e5d5b5c6\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-13T06:49:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://770b73dd0c2b6f38744e2c1cd6aa186a69bb87d1cd015a603ad2e33be06a8f6b\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://b3c1bab3d53f975b2e24582ce2771bc26d600e0dd7c30339dcfd18a071162ad4\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-12-13T06:49:23Z\\\",\\\"message\\\":\\\"file observer\\\\nW1213 06:49:23.283403 1 builder.go:272] unable to get owner reference (falling back to namespace): pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\nI1213 06:49:23.284336 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1213 06:49:23.286978 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-981773671/tls.crt::/tmp/serving-cert-981773671/tls.key\\\\\\\"\\\\nI1213 06:49:23.775072 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1213 06:49:23.778615 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1213 06:49:23.778645 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1213 06:49:23.778680 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1213 06:49:23.778691 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1213 06:49:23.785058 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1213 06:49:23.786150 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1213 06:49:23.786196 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1213 06:49:23.786223 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1213 06:49:23.786246 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1213 06:49:23.786269 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1213 06:49:23.786291 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI1213 06:49:23.785608 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF1213 06:49:23.787563 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-13T06:49:08Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-13T06:49:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://9acd07066deed483bd7fa61ee45fda09539d110717a134a56827f871186f92fc\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-13T06:49:07Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://38c91cfbaecfdbd3e3b663422a91ab4294610afb6bbf9ebdb77db24e2b76c682\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://38c91cfbaecfdbd3e3b663422a91ab4294610afb6bbf9ebdb77db24e2b76c682\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-13T06:49:05Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-13T06:49:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-13T06:49:03Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-13T06:50:25Z is after 2025-08-24T17:21:41Z" Dec 13 06:50:25 crc kubenswrapper[4971]: I1213 06:50:25.263338 4971 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-13T06:49:23Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-13T06:49:23Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-13T06:50:25Z is after 2025-08-24T17:21:41Z" Dec 13 06:50:25 crc kubenswrapper[4971]: I1213 06:50:25.278394 4971 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-13T06:49:23Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-13T06:49:23Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-13T06:50:25Z is after 2025-08-24T17:21:41Z" Dec 13 06:50:25 crc kubenswrapper[4971]: I1213 06:50:25.292704 4971 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-13T06:49:24Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-13T06:49:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-13T06:49:24Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d0b2ecf0d5b22f23217f42f259ae7ffb6359555b22c30eaebf667eaf254b3bab\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-13T06:49:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-13T06:50:25Z is after 2025-08-24T17:21:41Z" Dec 13 06:50:25 crc kubenswrapper[4971]: I1213 06:50:25.299901 4971 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 13 06:50:25 crc kubenswrapper[4971]: I1213 06:50:25.300056 4971 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 13 06:50:25 crc kubenswrapper[4971]: I1213 06:50:25.300074 4971 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 13 06:50:25 crc kubenswrapper[4971]: I1213 06:50:25.300123 4971 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 13 06:50:25 crc kubenswrapper[4971]: I1213 06:50:25.300135 4971 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-13T06:50:25Z","lastTransitionTime":"2025-12-13T06:50:25Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 13 06:50:25 crc kubenswrapper[4971]: I1213 06:50:25.308656 4971 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-13T06:49:23Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-13T06:49:23Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-13T06:50:25Z is after 2025-08-24T17:21:41Z" Dec 13 06:50:25 crc kubenswrapper[4971]: I1213 06:50:25.321098 4971 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-13T06:49:26Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-13T06:49:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-13T06:49:26Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9c8a0d1db0e754462d805e82daf5f52ffd5457507a30456535ec59e096c9d092\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-13T06:49:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-13T06:50:25Z is after 2025-08-24T17:21:41Z" Dec 13 06:50:25 crc kubenswrapper[4971]: I1213 06:50:25.332031 4971 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-82xjz" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e0c4af12-edda-41d7-9b44-f87396174ef2\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T06:49:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T06:49:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T06:49:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T06:49:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://0f4577d9e37f860813e221f2e208177d8d383fd04f673ce357a0a405ea29d1e6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-13T06:49:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cnxsq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://7e0d2020e62f1a277cff7047e1907dc747ed393a1f6c983e1ee258bd156c16c7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-13T06:49:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cnxsq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-13T06:49:25Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-82xjz\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-13T06:50:25Z is after 2025-08-24T17:21:41Z" Dec 13 06:50:25 crc kubenswrapper[4971]: I1213 06:50:25.344484 4971 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-wwvpl" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"dddf744b-2122-49ad-a980-105c09636e0f\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T06:49:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T06:49:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T06:50:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T06:50:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c8680aaa9e05095ac7e467c896d9e927bccff908925768ce5ed469ab89a72475\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://4676d897c7fc668587faec563f3551a056557879cee3e865192d2bd28edf62fc\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-12-13T06:50:12Z\\\",\\\"message\\\":\\\"2025-12-13T06:49:27+00:00 [cnibincopy] Successfully copied files in /usr/src/multus-cni/rhel9/bin/ to /host/opt/cni/bin/upgrade_263f091d-d847-4fad-88d3-fbf82c65c231\\\\n2025-12-13T06:49:27+00:00 [cnibincopy] Successfully moved files in /host/opt/cni/bin/upgrade_263f091d-d847-4fad-88d3-fbf82c65c231 to /host/opt/cni/bin/\\\\n2025-12-13T06:49:27Z [verbose] multus-daemon started\\\\n2025-12-13T06:49:27Z [verbose] Readiness Indicator file check\\\\n2025-12-13T06:50:12Z [error] have you checked that your default network is ready? still waiting for readinessindicatorfile @ /host/run/multus/cni/net.d/10-ovn-kubernetes.conf. pollimmediate error: timed out waiting for the condition\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-13T06:49:26Z\\\"}},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-13T06:50:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-czz9v\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-13T06:49:25Z\\\"}}\" for pod \"openshift-multus\"/\"multus-wwvpl\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-13T06:50:25Z is after 2025-08-24T17:21:41Z" Dec 13 06:50:25 crc kubenswrapper[4971]: I1213 06:50:25.362396 4971 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-zlcg5" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"736c6b1a-93dd-4f71-81b6-1c43fcf5f556\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T06:49:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T06:49:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T06:49:25Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T06:49:25Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://325223543b3c7db417f769533559a1ca2b23a723494e1885b0fd0480b5a6399f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-13T06:49:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vzmsr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://f22a4ec756119c4f791fb5105cf1f31df3e548b2b07b1f909a62802e9b6035ae\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-13T06:49:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vzmsr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://f6073140d9f86929b6d4bd2960ede0617a5c9620b269bbab043bae37e0d389f5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-13T06:49:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vzmsr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://693a2ab434c250cb140a866e8068b156df9191414fed1d63bee919a2528ea6e5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-13T06:49:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vzmsr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://219620f51474e60b3e1f880c0972eb612dffc4bad24462bbba2c093ae817d852\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-13T06:49:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vzmsr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://f137f4b441c2402093ab430a9689a46221c4fb4bc70d99fb992f62b0a989c465\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-13T06:49:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vzmsr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://7dd9913a65acba73ae1f32f4398207c5e276a1e979d86f9e720e4fd761bae4a3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://0b454a87c6dd11bd8c3bb3553ee71de32ed6f185984e5057a20f2237b35c45d9\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-12-13T06:49:56Z\\\",\\\"message\\\":\\\" for pod on switch crc\\\\nI1213 06:49:56.566617 6645 metrics.go:553] Stopping metrics server at address \\\\\\\"127.0.0.1:29103\\\\\\\"\\\\nI1213 06:49:56.566625 6645 base_network_controller_pods.go:477] [default/openshift-network-console/networking-console-plugin-85b44fc459-gdk6g] creating logical port openshift-network-console_networking-console-plugin-85b44fc459-gdk6g for pod on switch crc\\\\nI1213 06:49:56.566569 6645 loadbalancer.go:304] Deleted 0 stale LBs for map[string]string{\\\\\\\"k8s.ovn.org/kind\\\\\\\":\\\\\\\"Service\\\\\\\", \\\\\\\"k8s.ovn.org/owner\\\\\\\":\\\\\\\"openshift-console-operator/metrics\\\\\\\"}\\\\nI1213 06:49:56.566666 6645 services_controller.go:360] Finished syncing service metrics on namespace openshift-console-operator for network=default : 1.804735ms\\\\nI1213 06:49:56.566669 6645 loadbalancer.go:304] Deleted 0 stale LBs for map[string]string{\\\\\\\"k8s.ovn.org/kind\\\\\\\":\\\\\\\"Service\\\\\\\", \\\\\\\"k8s.ovn.org/owner\\\\\\\":\\\\\\\"openshift-apiserver/check-endpoints\\\\\\\"}\\\\nF1213 06:49:56.566698 6645 ovnkube.go:137] failed to run ovnkube: [failed to start network controller: failed to start default network controller: unable to create admin network policy controller, err: could not add Event Handler for anpInformer during admin network policy controller initialization, handler {0x1fcc6e0 0x1fcc3c0 0x1fcc360} was not added to shared informer because it has stopped already, failed to start node network controller: failed \\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-13T06:49:55Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://7dd9913a65acba73ae1f32f4398207c5e276a1e979d86f9e720e4fd761bae4a3\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-12-13T06:50:24Z\\\",\\\"message\\\":\\\"lse, Template:false, AddressFamily:\\\\\\\"\\\\\\\"}, Rules:[]services.LBRule{services.LBRule{Source:services.Addr{IP:\\\\\\\"10.217.4.250\\\\\\\", Port:443, Template:(*services.Template)(nil)}, Targets:[]services.Addr{}}}, Templates:services.TemplateMap(nil), Switches:[]string{}, Routers:[]string{}, Groups:[]string{\\\\\\\"clusterLBGroup\\\\\\\"}}}\\\\nI1213 06:50:24.014680 7045 services_controller.go:452] Built service openshift-machine-api/machine-api-operator-machine-webhook per-node LB for network=default: []services.LB{}\\\\nI1213 06:50:24.014686 7045 services_controller.go:453] Built service openshift-machine-api/machine-api-operator-machine-webhook template LB for network=default: []services.LB{}\\\\nI1213 06:50:24.012650 7045 obj_retry.go:303] Retry object setup: *v1.Pod openshift-kube-scheduler/openshift-kube-scheduler-crc\\\\nI1213 06:50:24.014698 7045 services_controller.go:454] Service openshift-machine-api/machine-api-operator-machine-webhook for network=default has 1 cluster-wide, 0 per-node configs, 0 template configs, making 1 (cluster) 0 (per node) and 0 (template) load balancers\\\\nF1213 06:50:24.014708 7045 ovnkube.go:137] failed to run ovnkube: [failed to start network controller: failed to start default network controller: unable to create admin network policy controller, err: could not add Event Handler for anpInformer during admin network policy controller initial\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-13T06:50:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vzmsr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://868049c6a3a4ebe4036d1d22e02ba6b76bced2ee7e269bb228accc4a0636678a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-13T06:49:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vzmsr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://451b22df6973ddc8fa2209a65d64d9e41b33eb231bb4adfab88cf0e5943bb246\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://451b22df6973ddc8fa2209a65d64d9e41b33eb231bb4adfab88cf0e5943bb246\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-13T06:49:26Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-13T06:49:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vzmsr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-13T06:49:25Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-zlcg5\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-13T06:50:25Z is after 2025-08-24T17:21:41Z" Dec 13 06:50:25 crc kubenswrapper[4971]: I1213 06:50:25.375573 4971 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"f9d31b5a-b770-4a15-95c8-35b49677c608\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T06:49:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T06:49:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T06:49:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T06:49:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T06:49:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c9baa042f9dada954faa57b5454cd6b0a12a297e56ddc192c52cff50e42a16fa\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-13T06:49:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://c788de86754f0bf37ee184672fc1dcb388ccebc246d4f15be4b5da661f725f8f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-13T06:49:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://760817511a9b4f96a68dc38d26631250eaab72e1dd8f123e653039b6b4e91f84\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-13T06:49:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b859be0c9feac856e9855c6a038b02b49b32fade97873c46d0bec0a32d8b5147\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-13T06:49:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-13T06:49:03Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-13T06:50:25Z is after 2025-08-24T17:21:41Z" Dec 13 06:50:25 crc kubenswrapper[4971]: I1213 06:50:25.388877 4971 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"8ede2992-9576-4b23-83c2-24ac3b1d3caf\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T06:49:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T06:49:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T06:49:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T06:49:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T06:49:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9e61d25add1723ac286794abfe886358e9d64455c25379f7e3057fc3dd812418\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-13T06:49:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://7fb51d8435b4e3ce0411cae1d6c2824c552f0a33d79de01f969274516b39a5b3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-13T06:49:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://1c35f5e8d8fab6d77b9e2e0dedb195440de5d6af345cd5865abd32caa7fafd60\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-13T06:49:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://38806c430185715371ab9bf30a1af9a9d5d492abd9bb84acbdccc694f56e3a8a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"wait-for-host-port\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://38806c430185715371ab9bf30a1af9a9d5d492abd9bb84acbdccc694f56e3a8a\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-13T06:49:04Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-13T06:49:04Z\\\"}}}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-13T06:49:03Z\\\"}}\" for pod \"openshift-kube-scheduler\"/\"openshift-kube-scheduler-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-13T06:50:25Z is after 2025-08-24T17:21:41Z" Dec 13 06:50:25 crc kubenswrapper[4971]: I1213 06:50:25.401495 4971 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-ncchz" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"de914d21-a688-401e-a66e-861beb6336d9\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T06:49:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T06:49:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T06:49:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T06:49:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://244ffe3dc9d8cf799aff4c4d795bf41554c1e6bb6667fbb0c7df322fb5f96c68\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-13T06:49:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lqmjj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://0b7111af6c5a198d02bb02858837cfb4f3b20828d8d206cd18aa0ffa102d0cd4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://0b7111af6c5a198d02bb02858837cfb4f3b20828d8d206cd18aa0ffa102d0cd4\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-13T06:49:26Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-13T06:49:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lqmjj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b36777fe40e74ae1cecb7b73e204139db5279ba51cef4f1e8c4a4c491d8be403\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://b36777fe40e74ae1cecb7b73e204139db5279ba51cef4f1e8c4a4c491d8be403\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-13T06:49:28Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-13T06:49:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lqmjj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://0597a06cc663e416142311a16941d98297577693a1fe27201ea42c27496cace2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://0597a06cc663e416142311a16941d98297577693a1fe27201ea42c27496cace2\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-13T06:49:29Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-13T06:49:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lqmjj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://3f52c2e22a75c010993ca5809a3a9d1304758d4cb6e83691ef1af15a4b514ac4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://3f52c2e22a75c010993ca5809a3a9d1304758d4cb6e83691ef1af15a4b514ac4\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-13T06:49:30Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-13T06:49:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lqmjj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://9e21ca708c4c4c1228b417cece51299e0792729b7c78884bd18b2dda8fe8355c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://9e21ca708c4c4c1228b417cece51299e0792729b7c78884bd18b2dda8fe8355c\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-13T06:49:31Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-13T06:49:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lqmjj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://74552159f12c87cdda09147692b400ccac0d4e43c1ed72fe8462c4249ad69c7b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://74552159f12c87cdda09147692b400ccac0d4e43c1ed72fe8462c4249ad69c7b\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-13T06:49:32Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-13T06:49:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lqmjj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-13T06:49:25Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-ncchz\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-13T06:50:25Z is after 2025-08-24T17:21:41Z" Dec 13 06:50:25 crc kubenswrapper[4971]: I1213 06:50:25.402071 4971 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 13 06:50:25 crc kubenswrapper[4971]: I1213 06:50:25.402107 4971 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 13 06:50:25 crc kubenswrapper[4971]: I1213 06:50:25.402115 4971 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 13 06:50:25 crc kubenswrapper[4971]: I1213 06:50:25.402130 4971 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 13 06:50:25 crc kubenswrapper[4971]: I1213 06:50:25.402139 4971 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-13T06:50:25Z","lastTransitionTime":"2025-12-13T06:50:25Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 13 06:50:25 crc kubenswrapper[4971]: I1213 06:50:25.412280 4971 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-gkcd8" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"13f5c327-f95e-42b4-8783-218500c17694\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T06:49:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T06:49:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T06:49:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T06:49:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://77e846e36f1cb40d03471dc03de11b6c3da448baa407aad38093fae9a801cf92\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-13T06:49:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gvlhn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-13T06:49:27Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-gkcd8\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-13T06:50:25Z is after 2025-08-24T17:21:41Z" Dec 13 06:50:25 crc kubenswrapper[4971]: I1213 06:50:25.421908 4971 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-x98k7" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"c7185dec-3240-42e8-9d90-6182a5e8b65c\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T06:49:39Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T06:49:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T06:49:39Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T06:49:39Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bj2gn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bj2gn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-13T06:49:39Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-x98k7\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-13T06:50:25Z is after 2025-08-24T17:21:41Z" Dec 13 06:50:25 crc kubenswrapper[4971]: I1213 06:50:25.431970 4971 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5ef36745-87bb-4be8-b941-aa8f865a5c83\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T06:49:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T06:49:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T06:49:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T06:49:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T06:49:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://35b737b0d2c5c5503ba2819f34d2de0599792a77beb1fbeb561d44efd1f212c5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-crio\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-13T06:49:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kube\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"var-lib-kubelet\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f0d864149bfcbb6227e6a6fbde23fff650d688850978eae0caf04515eb21cd85\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://f0d864149bfcbb6227e6a6fbde23fff650d688850978eae0caf04515eb21cd85\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-13T06:49:04Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-13T06:49:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var\\\",\\\"name\\\":\\\"var-lib-kubelet\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-13T06:49:03Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"kube-rbac-proxy-crio-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-13T06:50:25Z is after 2025-08-24T17:21:41Z" Dec 13 06:50:25 crc kubenswrapper[4971]: I1213 06:50:25.457086 4971 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"1ed9f82e-fe5c-4b76-a619-62e5abd88d3b\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T06:49:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T06:49:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T06:49:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T06:49:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T06:49:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e457329a78a2d24337f994e8a0a1f75a2a667244aa75eb84613d1e7138364aac\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-13T06:49:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://eb014de8140deef097d9bd94f3b9fa4ab3558bc4449bef84b83bf62bfabf5836\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-13T06:49:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://2c438cc74d555f7c9db4569d1c56b78b1088ed8105697d2c0f31e6c5136f3dd0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-13T06:49:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://aa6fa6e4340dcd066252df3c56d79a3a39ed531d5eea8d1d5941d6952368d3af\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-13T06:49:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://c4471b98854c52d5cd084d0ba3e64e5b08e9518ec3c372cce2c875a440522457\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-13T06:49:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7d2746335c98dae3a0f43024d34cb62ebcb42a01c8f9ce18d52b29d4c45598e5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://7d2746335c98dae3a0f43024d34cb62ebcb42a01c8f9ce18d52b29d4c45598e5\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-13T06:49:04Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-13T06:49:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://bc0e77d10a738128bea82122496bc42235fc1688f8907756d93e61b651902b57\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://bc0e77d10a738128bea82122496bc42235fc1688f8907756d93e61b651902b57\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-13T06:49:05Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-13T06:49:05Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://f821e418cba5e832da8839268bcbbd9630bdf319a212d5382bfa2c4b0b73a268\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://f821e418cba5e832da8839268bcbbd9630bdf319a212d5382bfa2c4b0b73a268\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-13T06:49:07Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-13T06:49:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-13T06:49:03Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-13T06:50:25Z is after 2025-08-24T17:21:41Z" Dec 13 06:50:25 crc kubenswrapper[4971]: I1213 06:50:25.470590 4971 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-13T06:49:24Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-13T06:49:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-13T06:49:24Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4328cd6a8464ed468229bbfbc8265f969427916d72ebb92237dbe68efe70fa47\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-13T06:49:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://9d715ca82d9b5a594e41f2b98da6f4277da0e4e1dec114969f901c00f6d7553d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-13T06:49:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-13T06:50:25Z is after 2025-08-24T17:21:41Z" Dec 13 06:50:25 crc kubenswrapper[4971]: I1213 06:50:25.481544 4971 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-5rzh5" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"6bf4591d-2483-4132-9672-d0fd6c995968\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T06:49:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T06:49:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T06:49:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T06:49:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://50e2cb14e4ffb603acb06e2281615a99f06182d345c4e70f2446fa2077033bbb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-13T06:49:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dfxqn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-13T06:49:25Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-5rzh5\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-13T06:50:25Z is after 2025-08-24T17:21:41Z" Dec 13 06:50:25 crc kubenswrapper[4971]: I1213 06:50:25.493290 4971 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-z572s" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"b4146d3d-4034-49a9-bd37-164856255137\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T06:49:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T06:49:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T06:49:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T06:49:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://46fd4e1070835cfbf7f35d497eff7f047473fc4048ee07812925df5ebdf55342\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-13T06:49:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-j6l9b\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://dc79a0968f85723e6ad0dc269c8da1f7472adb1bb2b87edd017ee2295bfbf89c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-13T06:49:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-j6l9b\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-13T06:49:38Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-z572s\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-13T06:50:25Z is after 2025-08-24T17:21:41Z" Dec 13 06:50:25 crc kubenswrapper[4971]: I1213 06:50:25.504053 4971 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 13 06:50:25 crc kubenswrapper[4971]: I1213 06:50:25.504087 4971 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 13 06:50:25 crc kubenswrapper[4971]: I1213 06:50:25.504096 4971 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 13 06:50:25 crc kubenswrapper[4971]: I1213 06:50:25.504110 4971 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 13 06:50:25 crc kubenswrapper[4971]: I1213 06:50:25.504119 4971 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-13T06:50:25Z","lastTransitionTime":"2025-12-13T06:50:25Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 13 06:50:25 crc kubenswrapper[4971]: I1213 06:50:25.607013 4971 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 13 06:50:25 crc kubenswrapper[4971]: I1213 06:50:25.607104 4971 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 13 06:50:25 crc kubenswrapper[4971]: I1213 06:50:25.607130 4971 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 13 06:50:25 crc kubenswrapper[4971]: I1213 06:50:25.607160 4971 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 13 06:50:25 crc kubenswrapper[4971]: I1213 06:50:25.607178 4971 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-13T06:50:25Z","lastTransitionTime":"2025-12-13T06:50:25Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 13 06:50:25 crc kubenswrapper[4971]: I1213 06:50:25.709405 4971 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 13 06:50:25 crc kubenswrapper[4971]: I1213 06:50:25.709445 4971 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 13 06:50:25 crc kubenswrapper[4971]: I1213 06:50:25.709456 4971 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 13 06:50:25 crc kubenswrapper[4971]: I1213 06:50:25.709472 4971 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 13 06:50:25 crc kubenswrapper[4971]: I1213 06:50:25.709483 4971 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-13T06:50:25Z","lastTransitionTime":"2025-12-13T06:50:25Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 13 06:50:25 crc kubenswrapper[4971]: I1213 06:50:25.767870 4971 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-x98k7" Dec 13 06:50:25 crc kubenswrapper[4971]: I1213 06:50:25.767945 4971 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 13 06:50:25 crc kubenswrapper[4971]: E1213 06:50:25.768040 4971 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-x98k7" podUID="c7185dec-3240-42e8-9d90-6182a5e8b65c" Dec 13 06:50:25 crc kubenswrapper[4971]: I1213 06:50:25.768116 4971 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 13 06:50:25 crc kubenswrapper[4971]: I1213 06:50:25.768140 4971 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 13 06:50:25 crc kubenswrapper[4971]: E1213 06:50:25.768220 4971 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 13 06:50:25 crc kubenswrapper[4971]: E1213 06:50:25.768265 4971 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 13 06:50:25 crc kubenswrapper[4971]: E1213 06:50:25.768336 4971 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 13 06:50:25 crc kubenswrapper[4971]: I1213 06:50:25.811785 4971 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 13 06:50:25 crc kubenswrapper[4971]: I1213 06:50:25.811830 4971 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 13 06:50:25 crc kubenswrapper[4971]: I1213 06:50:25.811841 4971 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 13 06:50:25 crc kubenswrapper[4971]: I1213 06:50:25.811856 4971 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 13 06:50:25 crc kubenswrapper[4971]: I1213 06:50:25.811910 4971 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-13T06:50:25Z","lastTransitionTime":"2025-12-13T06:50:25Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 13 06:50:25 crc kubenswrapper[4971]: I1213 06:50:25.913981 4971 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 13 06:50:25 crc kubenswrapper[4971]: I1213 06:50:25.914018 4971 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 13 06:50:25 crc kubenswrapper[4971]: I1213 06:50:25.914028 4971 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 13 06:50:25 crc kubenswrapper[4971]: I1213 06:50:25.914045 4971 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 13 06:50:25 crc kubenswrapper[4971]: I1213 06:50:25.914057 4971 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-13T06:50:25Z","lastTransitionTime":"2025-12-13T06:50:25Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 13 06:50:26 crc kubenswrapper[4971]: I1213 06:50:26.016130 4971 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 13 06:50:26 crc kubenswrapper[4971]: I1213 06:50:26.017053 4971 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 13 06:50:26 crc kubenswrapper[4971]: I1213 06:50:26.017071 4971 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 13 06:50:26 crc kubenswrapper[4971]: I1213 06:50:26.017091 4971 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 13 06:50:26 crc kubenswrapper[4971]: I1213 06:50:26.017103 4971 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-13T06:50:26Z","lastTransitionTime":"2025-12-13T06:50:26Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 13 06:50:26 crc kubenswrapper[4971]: I1213 06:50:26.120103 4971 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 13 06:50:26 crc kubenswrapper[4971]: I1213 06:50:26.120145 4971 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 13 06:50:26 crc kubenswrapper[4971]: I1213 06:50:26.120154 4971 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 13 06:50:26 crc kubenswrapper[4971]: I1213 06:50:26.120169 4971 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 13 06:50:26 crc kubenswrapper[4971]: I1213 06:50:26.120180 4971 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-13T06:50:26Z","lastTransitionTime":"2025-12-13T06:50:26Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 13 06:50:26 crc kubenswrapper[4971]: I1213 06:50:26.221979 4971 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-zlcg5_736c6b1a-93dd-4f71-81b6-1c43fcf5f556/ovnkube-controller/3.log" Dec 13 06:50:26 crc kubenswrapper[4971]: I1213 06:50:26.222069 4971 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 13 06:50:26 crc kubenswrapper[4971]: I1213 06:50:26.222105 4971 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 13 06:50:26 crc kubenswrapper[4971]: I1213 06:50:26.222115 4971 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 13 06:50:26 crc kubenswrapper[4971]: I1213 06:50:26.222134 4971 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 13 06:50:26 crc kubenswrapper[4971]: I1213 06:50:26.222145 4971 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-13T06:50:26Z","lastTransitionTime":"2025-12-13T06:50:26Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 13 06:50:26 crc kubenswrapper[4971]: I1213 06:50:26.226294 4971 scope.go:117] "RemoveContainer" containerID="7dd9913a65acba73ae1f32f4398207c5e276a1e979d86f9e720e4fd761bae4a3" Dec 13 06:50:26 crc kubenswrapper[4971]: E1213 06:50:26.226482 4971 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"ovnkube-controller\" with CrashLoopBackOff: \"back-off 40s restarting failed container=ovnkube-controller pod=ovnkube-node-zlcg5_openshift-ovn-kubernetes(736c6b1a-93dd-4f71-81b6-1c43fcf5f556)\"" pod="openshift-ovn-kubernetes/ovnkube-node-zlcg5" podUID="736c6b1a-93dd-4f71-81b6-1c43fcf5f556" Dec 13 06:50:26 crc kubenswrapper[4971]: I1213 06:50:26.240218 4971 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-wwvpl" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"dddf744b-2122-49ad-a980-105c09636e0f\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T06:49:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T06:49:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T06:50:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T06:50:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c8680aaa9e05095ac7e467c896d9e927bccff908925768ce5ed469ab89a72475\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://4676d897c7fc668587faec563f3551a056557879cee3e865192d2bd28edf62fc\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-12-13T06:50:12Z\\\",\\\"message\\\":\\\"2025-12-13T06:49:27+00:00 [cnibincopy] Successfully copied files in /usr/src/multus-cni/rhel9/bin/ to /host/opt/cni/bin/upgrade_263f091d-d847-4fad-88d3-fbf82c65c231\\\\n2025-12-13T06:49:27+00:00 [cnibincopy] Successfully moved files in /host/opt/cni/bin/upgrade_263f091d-d847-4fad-88d3-fbf82c65c231 to /host/opt/cni/bin/\\\\n2025-12-13T06:49:27Z [verbose] multus-daemon started\\\\n2025-12-13T06:49:27Z [verbose] Readiness Indicator file check\\\\n2025-12-13T06:50:12Z [error] have you checked that your default network is ready? still waiting for readinessindicatorfile @ /host/run/multus/cni/net.d/10-ovn-kubernetes.conf. pollimmediate error: timed out waiting for the condition\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-13T06:49:26Z\\\"}},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-13T06:50:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-czz9v\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-13T06:49:25Z\\\"}}\" for pod \"openshift-multus\"/\"multus-wwvpl\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-13T06:50:26Z is after 2025-08-24T17:21:41Z" Dec 13 06:50:26 crc kubenswrapper[4971]: I1213 06:50:26.256876 4971 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-zlcg5" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"736c6b1a-93dd-4f71-81b6-1c43fcf5f556\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T06:49:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T06:49:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T06:49:25Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T06:49:25Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://325223543b3c7db417f769533559a1ca2b23a723494e1885b0fd0480b5a6399f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-13T06:49:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vzmsr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://f22a4ec756119c4f791fb5105cf1f31df3e548b2b07b1f909a62802e9b6035ae\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-13T06:49:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vzmsr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://f6073140d9f86929b6d4bd2960ede0617a5c9620b269bbab043bae37e0d389f5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-13T06:49:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vzmsr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://693a2ab434c250cb140a866e8068b156df9191414fed1d63bee919a2528ea6e5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-13T06:49:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vzmsr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://219620f51474e60b3e1f880c0972eb612dffc4bad24462bbba2c093ae817d852\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-13T06:49:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vzmsr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://f137f4b441c2402093ab430a9689a46221c4fb4bc70d99fb992f62b0a989c465\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-13T06:49:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vzmsr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://7dd9913a65acba73ae1f32f4398207c5e276a1e979d86f9e720e4fd761bae4a3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://7dd9913a65acba73ae1f32f4398207c5e276a1e979d86f9e720e4fd761bae4a3\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-12-13T06:50:24Z\\\",\\\"message\\\":\\\"lse, Template:false, AddressFamily:\\\\\\\"\\\\\\\"}, Rules:[]services.LBRule{services.LBRule{Source:services.Addr{IP:\\\\\\\"10.217.4.250\\\\\\\", Port:443, Template:(*services.Template)(nil)}, Targets:[]services.Addr{}}}, Templates:services.TemplateMap(nil), Switches:[]string{}, Routers:[]string{}, Groups:[]string{\\\\\\\"clusterLBGroup\\\\\\\"}}}\\\\nI1213 06:50:24.014680 7045 services_controller.go:452] Built service openshift-machine-api/machine-api-operator-machine-webhook per-node LB for network=default: []services.LB{}\\\\nI1213 06:50:24.014686 7045 services_controller.go:453] Built service openshift-machine-api/machine-api-operator-machine-webhook template LB for network=default: []services.LB{}\\\\nI1213 06:50:24.012650 7045 obj_retry.go:303] Retry object setup: *v1.Pod openshift-kube-scheduler/openshift-kube-scheduler-crc\\\\nI1213 06:50:24.014698 7045 services_controller.go:454] Service openshift-machine-api/machine-api-operator-machine-webhook for network=default has 1 cluster-wide, 0 per-node configs, 0 template configs, making 1 (cluster) 0 (per node) and 0 (template) load balancers\\\\nF1213 06:50:24.014708 7045 ovnkube.go:137] failed to run ovnkube: [failed to start network controller: failed to start default network controller: unable to create admin network policy controller, err: could not add Event Handler for anpInformer during admin network policy controller initial\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-13T06:50:23Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 40s restarting failed container=ovnkube-controller pod=ovnkube-node-zlcg5_openshift-ovn-kubernetes(736c6b1a-93dd-4f71-81b6-1c43fcf5f556)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vzmsr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://868049c6a3a4ebe4036d1d22e02ba6b76bced2ee7e269bb228accc4a0636678a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-13T06:49:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vzmsr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://451b22df6973ddc8fa2209a65d64d9e41b33eb231bb4adfab88cf0e5943bb246\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://451b22df6973ddc8fa2209a65d64d9e41b33eb231bb4adfab88cf0e5943bb246\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-13T06:49:26Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-13T06:49:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vzmsr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-13T06:49:25Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-zlcg5\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-13T06:50:26Z is after 2025-08-24T17:21:41Z" Dec 13 06:50:26 crc kubenswrapper[4971]: I1213 06:50:26.270505 4971 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-13T06:49:24Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-13T06:49:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-13T06:49:24Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d0b2ecf0d5b22f23217f42f259ae7ffb6359555b22c30eaebf667eaf254b3bab\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-13T06:49:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-13T06:50:26Z is after 2025-08-24T17:21:41Z" Dec 13 06:50:26 crc kubenswrapper[4971]: I1213 06:50:26.283361 4971 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-13T06:49:23Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-13T06:49:23Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-13T06:50:26Z is after 2025-08-24T17:21:41Z" Dec 13 06:50:26 crc kubenswrapper[4971]: I1213 06:50:26.293011 4971 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-13T06:49:26Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-13T06:49:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-13T06:49:26Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9c8a0d1db0e754462d805e82daf5f52ffd5457507a30456535ec59e096c9d092\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-13T06:49:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-13T06:50:26Z is after 2025-08-24T17:21:41Z" Dec 13 06:50:26 crc kubenswrapper[4971]: I1213 06:50:26.304374 4971 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-82xjz" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e0c4af12-edda-41d7-9b44-f87396174ef2\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T06:49:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T06:49:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T06:49:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T06:49:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://0f4577d9e37f860813e221f2e208177d8d383fd04f673ce357a0a405ea29d1e6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-13T06:49:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cnxsq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://7e0d2020e62f1a277cff7047e1907dc747ed393a1f6c983e1ee258bd156c16c7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-13T06:49:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cnxsq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-13T06:49:25Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-82xjz\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-13T06:50:26Z is after 2025-08-24T17:21:41Z" Dec 13 06:50:26 crc kubenswrapper[4971]: I1213 06:50:26.313476 4971 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-x98k7" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"c7185dec-3240-42e8-9d90-6182a5e8b65c\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T06:49:39Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T06:49:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T06:49:39Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T06:49:39Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bj2gn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bj2gn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-13T06:49:39Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-x98k7\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-13T06:50:26Z is after 2025-08-24T17:21:41Z" Dec 13 06:50:26 crc kubenswrapper[4971]: I1213 06:50:26.323718 4971 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"f9d31b5a-b770-4a15-95c8-35b49677c608\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T06:49:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T06:49:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T06:49:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T06:49:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T06:49:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c9baa042f9dada954faa57b5454cd6b0a12a297e56ddc192c52cff50e42a16fa\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-13T06:49:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://c788de86754f0bf37ee184672fc1dcb388ccebc246d4f15be4b5da661f725f8f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-13T06:49:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://760817511a9b4f96a68dc38d26631250eaab72e1dd8f123e653039b6b4e91f84\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-13T06:49:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b859be0c9feac856e9855c6a038b02b49b32fade97873c46d0bec0a32d8b5147\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-13T06:49:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-13T06:49:03Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-13T06:50:26Z is after 2025-08-24T17:21:41Z" Dec 13 06:50:26 crc kubenswrapper[4971]: I1213 06:50:26.324819 4971 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 13 06:50:26 crc kubenswrapper[4971]: I1213 06:50:26.324879 4971 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 13 06:50:26 crc kubenswrapper[4971]: I1213 06:50:26.324901 4971 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 13 06:50:26 crc kubenswrapper[4971]: I1213 06:50:26.324931 4971 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 13 06:50:26 crc kubenswrapper[4971]: I1213 06:50:26.324954 4971 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-13T06:50:26Z","lastTransitionTime":"2025-12-13T06:50:26Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 13 06:50:26 crc kubenswrapper[4971]: I1213 06:50:26.333853 4971 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"8ede2992-9576-4b23-83c2-24ac3b1d3caf\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T06:49:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T06:49:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T06:49:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T06:49:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T06:49:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9e61d25add1723ac286794abfe886358e9d64455c25379f7e3057fc3dd812418\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-13T06:49:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://7fb51d8435b4e3ce0411cae1d6c2824c552f0a33d79de01f969274516b39a5b3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-13T06:49:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://1c35f5e8d8fab6d77b9e2e0dedb195440de5d6af345cd5865abd32caa7fafd60\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-13T06:49:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://38806c430185715371ab9bf30a1af9a9d5d492abd9bb84acbdccc694f56e3a8a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"wait-for-host-port\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://38806c430185715371ab9bf30a1af9a9d5d492abd9bb84acbdccc694f56e3a8a\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-13T06:49:04Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-13T06:49:04Z\\\"}}}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-13T06:49:03Z\\\"}}\" for pod \"openshift-kube-scheduler\"/\"openshift-kube-scheduler-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-13T06:50:26Z is after 2025-08-24T17:21:41Z" Dec 13 06:50:26 crc kubenswrapper[4971]: I1213 06:50:26.347506 4971 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-ncchz" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"de914d21-a688-401e-a66e-861beb6336d9\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T06:49:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T06:49:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T06:49:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T06:49:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://244ffe3dc9d8cf799aff4c4d795bf41554c1e6bb6667fbb0c7df322fb5f96c68\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-13T06:49:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lqmjj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://0b7111af6c5a198d02bb02858837cfb4f3b20828d8d206cd18aa0ffa102d0cd4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://0b7111af6c5a198d02bb02858837cfb4f3b20828d8d206cd18aa0ffa102d0cd4\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-13T06:49:26Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-13T06:49:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lqmjj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b36777fe40e74ae1cecb7b73e204139db5279ba51cef4f1e8c4a4c491d8be403\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://b36777fe40e74ae1cecb7b73e204139db5279ba51cef4f1e8c4a4c491d8be403\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-13T06:49:28Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-13T06:49:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lqmjj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://0597a06cc663e416142311a16941d98297577693a1fe27201ea42c27496cace2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://0597a06cc663e416142311a16941d98297577693a1fe27201ea42c27496cace2\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-13T06:49:29Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-13T06:49:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lqmjj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://3f52c2e22a75c010993ca5809a3a9d1304758d4cb6e83691ef1af15a4b514ac4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://3f52c2e22a75c010993ca5809a3a9d1304758d4cb6e83691ef1af15a4b514ac4\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-13T06:49:30Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-13T06:49:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lqmjj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://9e21ca708c4c4c1228b417cece51299e0792729b7c78884bd18b2dda8fe8355c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://9e21ca708c4c4c1228b417cece51299e0792729b7c78884bd18b2dda8fe8355c\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-13T06:49:31Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-13T06:49:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lqmjj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://74552159f12c87cdda09147692b400ccac0d4e43c1ed72fe8462c4249ad69c7b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://74552159f12c87cdda09147692b400ccac0d4e43c1ed72fe8462c4249ad69c7b\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-13T06:49:32Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-13T06:49:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lqmjj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-13T06:49:25Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-ncchz\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-13T06:50:26Z is after 2025-08-24T17:21:41Z" Dec 13 06:50:26 crc kubenswrapper[4971]: I1213 06:50:26.358863 4971 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-gkcd8" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"13f5c327-f95e-42b4-8783-218500c17694\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T06:49:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T06:49:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T06:49:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T06:49:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://77e846e36f1cb40d03471dc03de11b6c3da448baa407aad38093fae9a801cf92\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-13T06:49:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gvlhn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-13T06:49:27Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-gkcd8\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-13T06:50:26Z is after 2025-08-24T17:21:41Z" Dec 13 06:50:26 crc kubenswrapper[4971]: I1213 06:50:26.370401 4971 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-z572s" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"b4146d3d-4034-49a9-bd37-164856255137\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T06:49:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T06:49:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T06:49:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T06:49:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://46fd4e1070835cfbf7f35d497eff7f047473fc4048ee07812925df5ebdf55342\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-13T06:49:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-j6l9b\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://dc79a0968f85723e6ad0dc269c8da1f7472adb1bb2b87edd017ee2295bfbf89c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-13T06:49:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-j6l9b\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-13T06:49:38Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-z572s\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-13T06:50:26Z is after 2025-08-24T17:21:41Z" Dec 13 06:50:26 crc kubenswrapper[4971]: I1213 06:50:26.382487 4971 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5ef36745-87bb-4be8-b941-aa8f865a5c83\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T06:49:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T06:49:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T06:49:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T06:49:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T06:49:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://35b737b0d2c5c5503ba2819f34d2de0599792a77beb1fbeb561d44efd1f212c5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-crio\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-13T06:49:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kube\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"var-lib-kubelet\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f0d864149bfcbb6227e6a6fbde23fff650d688850978eae0caf04515eb21cd85\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://f0d864149bfcbb6227e6a6fbde23fff650d688850978eae0caf04515eb21cd85\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-13T06:49:04Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-13T06:49:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var\\\",\\\"name\\\":\\\"var-lib-kubelet\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-13T06:49:03Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"kube-rbac-proxy-crio-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-13T06:50:26Z is after 2025-08-24T17:21:41Z" Dec 13 06:50:26 crc kubenswrapper[4971]: I1213 06:50:26.401246 4971 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"1ed9f82e-fe5c-4b76-a619-62e5abd88d3b\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T06:49:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T06:49:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T06:49:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T06:49:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T06:49:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e457329a78a2d24337f994e8a0a1f75a2a667244aa75eb84613d1e7138364aac\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-13T06:49:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://eb014de8140deef097d9bd94f3b9fa4ab3558bc4449bef84b83bf62bfabf5836\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-13T06:49:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://2c438cc74d555f7c9db4569d1c56b78b1088ed8105697d2c0f31e6c5136f3dd0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-13T06:49:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://aa6fa6e4340dcd066252df3c56d79a3a39ed531d5eea8d1d5941d6952368d3af\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-13T06:49:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://c4471b98854c52d5cd084d0ba3e64e5b08e9518ec3c372cce2c875a440522457\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-13T06:49:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7d2746335c98dae3a0f43024d34cb62ebcb42a01c8f9ce18d52b29d4c45598e5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://7d2746335c98dae3a0f43024d34cb62ebcb42a01c8f9ce18d52b29d4c45598e5\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-13T06:49:04Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-13T06:49:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://bc0e77d10a738128bea82122496bc42235fc1688f8907756d93e61b651902b57\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://bc0e77d10a738128bea82122496bc42235fc1688f8907756d93e61b651902b57\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-13T06:49:05Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-13T06:49:05Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://f821e418cba5e832da8839268bcbbd9630bdf319a212d5382bfa2c4b0b73a268\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://f821e418cba5e832da8839268bcbbd9630bdf319a212d5382bfa2c4b0b73a268\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-13T06:49:07Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-13T06:49:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-13T06:49:03Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-13T06:50:26Z is after 2025-08-24T17:21:41Z" Dec 13 06:50:26 crc kubenswrapper[4971]: I1213 06:50:26.417022 4971 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-13T06:49:24Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-13T06:49:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-13T06:49:24Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4328cd6a8464ed468229bbfbc8265f969427916d72ebb92237dbe68efe70fa47\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-13T06:49:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://9d715ca82d9b5a594e41f2b98da6f4277da0e4e1dec114969f901c00f6d7553d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-13T06:49:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-13T06:50:26Z is after 2025-08-24T17:21:41Z" Dec 13 06:50:26 crc kubenswrapper[4971]: I1213 06:50:26.426684 4971 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-5rzh5" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"6bf4591d-2483-4132-9672-d0fd6c995968\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T06:49:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T06:49:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T06:49:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T06:49:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://50e2cb14e4ffb603acb06e2281615a99f06182d345c4e70f2446fa2077033bbb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-13T06:49:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dfxqn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-13T06:49:25Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-5rzh5\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-13T06:50:26Z is after 2025-08-24T17:21:41Z" Dec 13 06:50:26 crc kubenswrapper[4971]: I1213 06:50:26.427425 4971 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 13 06:50:26 crc kubenswrapper[4971]: I1213 06:50:26.427453 4971 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 13 06:50:26 crc kubenswrapper[4971]: I1213 06:50:26.427461 4971 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 13 06:50:26 crc kubenswrapper[4971]: I1213 06:50:26.427478 4971 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 13 06:50:26 crc kubenswrapper[4971]: I1213 06:50:26.427488 4971 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-13T06:50:26Z","lastTransitionTime":"2025-12-13T06:50:26Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 13 06:50:26 crc kubenswrapper[4971]: I1213 06:50:26.442438 4971 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3ad3268a-a982-428c-9c06-4aaeddd372d9\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T06:49:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T06:49:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T06:49:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T06:49:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T06:49:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3716afdf2bc693be80a09fc9f7727d334f435caf1d4c8ec4811b3a1f1652cdfe\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-13T06:49:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://30df3abf8c79db8af52c6a918f73e1ce410ddafcf341636424c267e1eeb195a7\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-13T06:49:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://ccdf65f6d2b1eb3647a03ba37c305c32e34addfdf79ff937c8dddaa3e5d5b5c6\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-13T06:49:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://770b73dd0c2b6f38744e2c1cd6aa186a69bb87d1cd015a603ad2e33be06a8f6b\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://b3c1bab3d53f975b2e24582ce2771bc26d600e0dd7c30339dcfd18a071162ad4\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-12-13T06:49:23Z\\\",\\\"message\\\":\\\"file observer\\\\nW1213 06:49:23.283403 1 builder.go:272] unable to get owner reference (falling back to namespace): pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\nI1213 06:49:23.284336 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1213 06:49:23.286978 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-981773671/tls.crt::/tmp/serving-cert-981773671/tls.key\\\\\\\"\\\\nI1213 06:49:23.775072 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1213 06:49:23.778615 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1213 06:49:23.778645 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1213 06:49:23.778680 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1213 06:49:23.778691 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1213 06:49:23.785058 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1213 06:49:23.786150 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1213 06:49:23.786196 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1213 06:49:23.786223 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1213 06:49:23.786246 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1213 06:49:23.786269 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1213 06:49:23.786291 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI1213 06:49:23.785608 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF1213 06:49:23.787563 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-13T06:49:08Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-13T06:49:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://9acd07066deed483bd7fa61ee45fda09539d110717a134a56827f871186f92fc\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-13T06:49:07Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://38c91cfbaecfdbd3e3b663422a91ab4294610afb6bbf9ebdb77db24e2b76c682\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://38c91cfbaecfdbd3e3b663422a91ab4294610afb6bbf9ebdb77db24e2b76c682\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-13T06:49:05Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-13T06:49:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-13T06:49:03Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-13T06:50:26Z is after 2025-08-24T17:21:41Z" Dec 13 06:50:26 crc kubenswrapper[4971]: I1213 06:50:26.456469 4971 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-13T06:49:23Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-13T06:49:23Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-13T06:50:26Z is after 2025-08-24T17:21:41Z" Dec 13 06:50:26 crc kubenswrapper[4971]: I1213 06:50:26.468357 4971 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-13T06:49:23Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-13T06:49:23Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-13T06:50:26Z is after 2025-08-24T17:21:41Z" Dec 13 06:50:26 crc kubenswrapper[4971]: I1213 06:50:26.530354 4971 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 13 06:50:26 crc kubenswrapper[4971]: I1213 06:50:26.530697 4971 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 13 06:50:26 crc kubenswrapper[4971]: I1213 06:50:26.530706 4971 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 13 06:50:26 crc kubenswrapper[4971]: I1213 06:50:26.530723 4971 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 13 06:50:26 crc kubenswrapper[4971]: I1213 06:50:26.530733 4971 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-13T06:50:26Z","lastTransitionTime":"2025-12-13T06:50:26Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 13 06:50:26 crc kubenswrapper[4971]: I1213 06:50:26.632399 4971 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 13 06:50:26 crc kubenswrapper[4971]: I1213 06:50:26.632445 4971 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 13 06:50:26 crc kubenswrapper[4971]: I1213 06:50:26.632463 4971 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 13 06:50:26 crc kubenswrapper[4971]: I1213 06:50:26.632484 4971 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 13 06:50:26 crc kubenswrapper[4971]: I1213 06:50:26.632497 4971 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-13T06:50:26Z","lastTransitionTime":"2025-12-13T06:50:26Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 13 06:50:26 crc kubenswrapper[4971]: I1213 06:50:26.734276 4971 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 13 06:50:26 crc kubenswrapper[4971]: I1213 06:50:26.734319 4971 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 13 06:50:26 crc kubenswrapper[4971]: I1213 06:50:26.734329 4971 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 13 06:50:26 crc kubenswrapper[4971]: I1213 06:50:26.734345 4971 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 13 06:50:26 crc kubenswrapper[4971]: I1213 06:50:26.734355 4971 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-13T06:50:26Z","lastTransitionTime":"2025-12-13T06:50:26Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 13 06:50:26 crc kubenswrapper[4971]: I1213 06:50:26.836245 4971 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 13 06:50:26 crc kubenswrapper[4971]: I1213 06:50:26.836314 4971 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 13 06:50:26 crc kubenswrapper[4971]: I1213 06:50:26.836326 4971 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 13 06:50:26 crc kubenswrapper[4971]: I1213 06:50:26.836342 4971 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 13 06:50:26 crc kubenswrapper[4971]: I1213 06:50:26.836354 4971 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-13T06:50:26Z","lastTransitionTime":"2025-12-13T06:50:26Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 13 06:50:26 crc kubenswrapper[4971]: I1213 06:50:26.939184 4971 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 13 06:50:26 crc kubenswrapper[4971]: I1213 06:50:26.939254 4971 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 13 06:50:26 crc kubenswrapper[4971]: I1213 06:50:26.939277 4971 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 13 06:50:26 crc kubenswrapper[4971]: I1213 06:50:26.939307 4971 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 13 06:50:26 crc kubenswrapper[4971]: I1213 06:50:26.939328 4971 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-13T06:50:26Z","lastTransitionTime":"2025-12-13T06:50:26Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 13 06:50:27 crc kubenswrapper[4971]: I1213 06:50:27.041585 4971 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 13 06:50:27 crc kubenswrapper[4971]: I1213 06:50:27.041626 4971 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 13 06:50:27 crc kubenswrapper[4971]: I1213 06:50:27.041638 4971 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 13 06:50:27 crc kubenswrapper[4971]: I1213 06:50:27.041655 4971 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 13 06:50:27 crc kubenswrapper[4971]: I1213 06:50:27.041668 4971 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-13T06:50:27Z","lastTransitionTime":"2025-12-13T06:50:27Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 13 06:50:27 crc kubenswrapper[4971]: I1213 06:50:27.144445 4971 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 13 06:50:27 crc kubenswrapper[4971]: I1213 06:50:27.144481 4971 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 13 06:50:27 crc kubenswrapper[4971]: I1213 06:50:27.144490 4971 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 13 06:50:27 crc kubenswrapper[4971]: I1213 06:50:27.144504 4971 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 13 06:50:27 crc kubenswrapper[4971]: I1213 06:50:27.144533 4971 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-13T06:50:27Z","lastTransitionTime":"2025-12-13T06:50:27Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 13 06:50:27 crc kubenswrapper[4971]: I1213 06:50:27.246931 4971 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 13 06:50:27 crc kubenswrapper[4971]: I1213 06:50:27.246976 4971 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 13 06:50:27 crc kubenswrapper[4971]: I1213 06:50:27.246987 4971 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 13 06:50:27 crc kubenswrapper[4971]: I1213 06:50:27.247027 4971 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 13 06:50:27 crc kubenswrapper[4971]: I1213 06:50:27.247040 4971 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-13T06:50:27Z","lastTransitionTime":"2025-12-13T06:50:27Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 13 06:50:27 crc kubenswrapper[4971]: I1213 06:50:27.348963 4971 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 13 06:50:27 crc kubenswrapper[4971]: I1213 06:50:27.348998 4971 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 13 06:50:27 crc kubenswrapper[4971]: I1213 06:50:27.349009 4971 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 13 06:50:27 crc kubenswrapper[4971]: I1213 06:50:27.349028 4971 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 13 06:50:27 crc kubenswrapper[4971]: I1213 06:50:27.349040 4971 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-13T06:50:27Z","lastTransitionTime":"2025-12-13T06:50:27Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 13 06:50:27 crc kubenswrapper[4971]: I1213 06:50:27.451722 4971 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 13 06:50:27 crc kubenswrapper[4971]: I1213 06:50:27.451751 4971 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 13 06:50:27 crc kubenswrapper[4971]: I1213 06:50:27.451759 4971 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 13 06:50:27 crc kubenswrapper[4971]: I1213 06:50:27.451774 4971 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 13 06:50:27 crc kubenswrapper[4971]: I1213 06:50:27.451783 4971 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-13T06:50:27Z","lastTransitionTime":"2025-12-13T06:50:27Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 13 06:50:27 crc kubenswrapper[4971]: I1213 06:50:27.544580 4971 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 13 06:50:27 crc kubenswrapper[4971]: I1213 06:50:27.544750 4971 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"networking-console-plugin-cert\" (UniqueName: \"kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 13 06:50:27 crc kubenswrapper[4971]: I1213 06:50:27.544781 4971 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nginx-conf\" (UniqueName: \"kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 13 06:50:27 crc kubenswrapper[4971]: E1213 06:50:27.544911 4971 configmap.go:193] Couldn't get configMap openshift-network-console/networking-console-plugin: object "openshift-network-console"/"networking-console-plugin" not registered Dec 13 06:50:27 crc kubenswrapper[4971]: E1213 06:50:27.544995 4971 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2025-12-13 06:51:31.544977739 +0000 UTC m=+148.149387187 (durationBeforeRetry 1m4s). Error: MountVolume.SetUp failed for volume "nginx-conf" (UniqueName: "kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin" not registered Dec 13 06:50:27 crc kubenswrapper[4971]: E1213 06:50:27.545137 4971 secret.go:188] Couldn't get secret openshift-network-console/networking-console-plugin-cert: object "openshift-network-console"/"networking-console-plugin-cert" not registered Dec 13 06:50:27 crc kubenswrapper[4971]: E1213 06:50:27.545176 4971 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2025-12-13 06:51:31.545166766 +0000 UTC m=+148.149576214 (durationBeforeRetry 1m4s). Error: MountVolume.SetUp failed for volume "networking-console-plugin-cert" (UniqueName: "kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin-cert" not registered Dec 13 06:50:27 crc kubenswrapper[4971]: E1213 06:50:27.545227 4971 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-13 06:51:31.545215957 +0000 UTC m=+148.149625405 (durationBeforeRetry 1m4s). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 13 06:50:27 crc kubenswrapper[4971]: I1213 06:50:27.554136 4971 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 13 06:50:27 crc kubenswrapper[4971]: I1213 06:50:27.554189 4971 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 13 06:50:27 crc kubenswrapper[4971]: I1213 06:50:27.554198 4971 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 13 06:50:27 crc kubenswrapper[4971]: I1213 06:50:27.554214 4971 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 13 06:50:27 crc kubenswrapper[4971]: I1213 06:50:27.554223 4971 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-13T06:50:27Z","lastTransitionTime":"2025-12-13T06:50:27Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 13 06:50:27 crc kubenswrapper[4971]: I1213 06:50:27.645592 4971 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cqllr\" (UniqueName: \"kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr\") pod \"network-check-target-xd92c\" (UID: \"3b6479f0-333b-4a96-9adf-2099afdc2447\") " pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 13 06:50:27 crc kubenswrapper[4971]: I1213 06:50:27.645702 4971 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-s2dwl\" (UniqueName: \"kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl\") pod \"network-check-source-55646444c4-trplf\" (UID: \"9d751cbb-f2e2-430d-9754-c882a5e924a5\") " pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 13 06:50:27 crc kubenswrapper[4971]: E1213 06:50:27.645775 4971 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Dec 13 06:50:27 crc kubenswrapper[4971]: E1213 06:50:27.645806 4971 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Dec 13 06:50:27 crc kubenswrapper[4971]: E1213 06:50:27.645819 4971 projected.go:194] Error preparing data for projected volume kube-api-access-cqllr for pod openshift-network-diagnostics/network-check-target-xd92c: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Dec 13 06:50:27 crc kubenswrapper[4971]: E1213 06:50:27.645860 4971 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Dec 13 06:50:27 crc kubenswrapper[4971]: E1213 06:50:27.645878 4971 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr podName:3b6479f0-333b-4a96-9adf-2099afdc2447 nodeName:}" failed. No retries permitted until 2025-12-13 06:51:31.645862519 +0000 UTC m=+148.250271967 (durationBeforeRetry 1m4s). Error: MountVolume.SetUp failed for volume "kube-api-access-cqllr" (UniqueName: "kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr") pod "network-check-target-xd92c" (UID: "3b6479f0-333b-4a96-9adf-2099afdc2447") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Dec 13 06:50:27 crc kubenswrapper[4971]: E1213 06:50:27.645883 4971 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Dec 13 06:50:27 crc kubenswrapper[4971]: E1213 06:50:27.645905 4971 projected.go:194] Error preparing data for projected volume kube-api-access-s2dwl for pod openshift-network-diagnostics/network-check-source-55646444c4-trplf: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Dec 13 06:50:27 crc kubenswrapper[4971]: E1213 06:50:27.645999 4971 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl podName:9d751cbb-f2e2-430d-9754-c882a5e924a5 nodeName:}" failed. No retries permitted until 2025-12-13 06:51:31.645978373 +0000 UTC m=+148.250387831 (durationBeforeRetry 1m4s). Error: MountVolume.SetUp failed for volume "kube-api-access-s2dwl" (UniqueName: "kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl") pod "network-check-source-55646444c4-trplf" (UID: "9d751cbb-f2e2-430d-9754-c882a5e924a5") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Dec 13 06:50:27 crc kubenswrapper[4971]: I1213 06:50:27.656877 4971 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 13 06:50:27 crc kubenswrapper[4971]: I1213 06:50:27.656946 4971 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 13 06:50:27 crc kubenswrapper[4971]: I1213 06:50:27.656962 4971 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 13 06:50:27 crc kubenswrapper[4971]: I1213 06:50:27.656987 4971 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 13 06:50:27 crc kubenswrapper[4971]: I1213 06:50:27.657001 4971 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-13T06:50:27Z","lastTransitionTime":"2025-12-13T06:50:27Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 13 06:50:27 crc kubenswrapper[4971]: I1213 06:50:27.759404 4971 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 13 06:50:27 crc kubenswrapper[4971]: I1213 06:50:27.759487 4971 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 13 06:50:27 crc kubenswrapper[4971]: I1213 06:50:27.759502 4971 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 13 06:50:27 crc kubenswrapper[4971]: I1213 06:50:27.759533 4971 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 13 06:50:27 crc kubenswrapper[4971]: I1213 06:50:27.759542 4971 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-13T06:50:27Z","lastTransitionTime":"2025-12-13T06:50:27Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 13 06:50:27 crc kubenswrapper[4971]: I1213 06:50:27.768004 4971 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-x98k7" Dec 13 06:50:27 crc kubenswrapper[4971]: I1213 06:50:27.768029 4971 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 13 06:50:27 crc kubenswrapper[4971]: I1213 06:50:27.768040 4971 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 13 06:50:27 crc kubenswrapper[4971]: E1213 06:50:27.768134 4971 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-x98k7" podUID="c7185dec-3240-42e8-9d90-6182a5e8b65c" Dec 13 06:50:27 crc kubenswrapper[4971]: I1213 06:50:27.768175 4971 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 13 06:50:27 crc kubenswrapper[4971]: E1213 06:50:27.768281 4971 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 13 06:50:27 crc kubenswrapper[4971]: E1213 06:50:27.768344 4971 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 13 06:50:27 crc kubenswrapper[4971]: E1213 06:50:27.768427 4971 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 13 06:50:27 crc kubenswrapper[4971]: I1213 06:50:27.862667 4971 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 13 06:50:27 crc kubenswrapper[4971]: I1213 06:50:27.862706 4971 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 13 06:50:27 crc kubenswrapper[4971]: I1213 06:50:27.862717 4971 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 13 06:50:27 crc kubenswrapper[4971]: I1213 06:50:27.862734 4971 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 13 06:50:27 crc kubenswrapper[4971]: I1213 06:50:27.862745 4971 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-13T06:50:27Z","lastTransitionTime":"2025-12-13T06:50:27Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 13 06:50:27 crc kubenswrapper[4971]: I1213 06:50:27.965340 4971 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 13 06:50:27 crc kubenswrapper[4971]: I1213 06:50:27.965377 4971 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 13 06:50:27 crc kubenswrapper[4971]: I1213 06:50:27.965385 4971 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 13 06:50:27 crc kubenswrapper[4971]: I1213 06:50:27.965399 4971 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 13 06:50:27 crc kubenswrapper[4971]: I1213 06:50:27.965407 4971 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-13T06:50:27Z","lastTransitionTime":"2025-12-13T06:50:27Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 13 06:50:28 crc kubenswrapper[4971]: I1213 06:50:28.067368 4971 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 13 06:50:28 crc kubenswrapper[4971]: I1213 06:50:28.067416 4971 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 13 06:50:28 crc kubenswrapper[4971]: I1213 06:50:28.067427 4971 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 13 06:50:28 crc kubenswrapper[4971]: I1213 06:50:28.067478 4971 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 13 06:50:28 crc kubenswrapper[4971]: I1213 06:50:28.067491 4971 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-13T06:50:28Z","lastTransitionTime":"2025-12-13T06:50:28Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 13 06:50:28 crc kubenswrapper[4971]: I1213 06:50:28.169470 4971 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 13 06:50:28 crc kubenswrapper[4971]: I1213 06:50:28.169506 4971 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 13 06:50:28 crc kubenswrapper[4971]: I1213 06:50:28.169530 4971 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 13 06:50:28 crc kubenswrapper[4971]: I1213 06:50:28.169546 4971 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 13 06:50:28 crc kubenswrapper[4971]: I1213 06:50:28.169555 4971 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-13T06:50:28Z","lastTransitionTime":"2025-12-13T06:50:28Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 13 06:50:28 crc kubenswrapper[4971]: I1213 06:50:28.272230 4971 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 13 06:50:28 crc kubenswrapper[4971]: I1213 06:50:28.272305 4971 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 13 06:50:28 crc kubenswrapper[4971]: I1213 06:50:28.272317 4971 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 13 06:50:28 crc kubenswrapper[4971]: I1213 06:50:28.272336 4971 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 13 06:50:28 crc kubenswrapper[4971]: I1213 06:50:28.272349 4971 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-13T06:50:28Z","lastTransitionTime":"2025-12-13T06:50:28Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 13 06:50:28 crc kubenswrapper[4971]: I1213 06:50:28.374498 4971 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 13 06:50:28 crc kubenswrapper[4971]: I1213 06:50:28.374559 4971 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 13 06:50:28 crc kubenswrapper[4971]: I1213 06:50:28.374572 4971 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 13 06:50:28 crc kubenswrapper[4971]: I1213 06:50:28.374588 4971 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 13 06:50:28 crc kubenswrapper[4971]: I1213 06:50:28.374599 4971 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-13T06:50:28Z","lastTransitionTime":"2025-12-13T06:50:28Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 13 06:50:28 crc kubenswrapper[4971]: I1213 06:50:28.477471 4971 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 13 06:50:28 crc kubenswrapper[4971]: I1213 06:50:28.477532 4971 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 13 06:50:28 crc kubenswrapper[4971]: I1213 06:50:28.477544 4971 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 13 06:50:28 crc kubenswrapper[4971]: I1213 06:50:28.477561 4971 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 13 06:50:28 crc kubenswrapper[4971]: I1213 06:50:28.477572 4971 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-13T06:50:28Z","lastTransitionTime":"2025-12-13T06:50:28Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 13 06:50:28 crc kubenswrapper[4971]: I1213 06:50:28.580572 4971 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 13 06:50:28 crc kubenswrapper[4971]: I1213 06:50:28.580628 4971 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 13 06:50:28 crc kubenswrapper[4971]: I1213 06:50:28.580640 4971 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 13 06:50:28 crc kubenswrapper[4971]: I1213 06:50:28.580655 4971 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 13 06:50:28 crc kubenswrapper[4971]: I1213 06:50:28.580665 4971 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-13T06:50:28Z","lastTransitionTime":"2025-12-13T06:50:28Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 13 06:50:28 crc kubenswrapper[4971]: I1213 06:50:28.683426 4971 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 13 06:50:28 crc kubenswrapper[4971]: I1213 06:50:28.683469 4971 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 13 06:50:28 crc kubenswrapper[4971]: I1213 06:50:28.683481 4971 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 13 06:50:28 crc kubenswrapper[4971]: I1213 06:50:28.683498 4971 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 13 06:50:28 crc kubenswrapper[4971]: I1213 06:50:28.683510 4971 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-13T06:50:28Z","lastTransitionTime":"2025-12-13T06:50:28Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 13 06:50:28 crc kubenswrapper[4971]: I1213 06:50:28.785955 4971 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 13 06:50:28 crc kubenswrapper[4971]: I1213 06:50:28.786000 4971 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 13 06:50:28 crc kubenswrapper[4971]: I1213 06:50:28.786014 4971 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 13 06:50:28 crc kubenswrapper[4971]: I1213 06:50:28.786033 4971 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 13 06:50:28 crc kubenswrapper[4971]: I1213 06:50:28.786045 4971 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-13T06:50:28Z","lastTransitionTime":"2025-12-13T06:50:28Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 13 06:50:28 crc kubenswrapper[4971]: I1213 06:50:28.888091 4971 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 13 06:50:28 crc kubenswrapper[4971]: I1213 06:50:28.888171 4971 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 13 06:50:28 crc kubenswrapper[4971]: I1213 06:50:28.888184 4971 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 13 06:50:28 crc kubenswrapper[4971]: I1213 06:50:28.888203 4971 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 13 06:50:28 crc kubenswrapper[4971]: I1213 06:50:28.888214 4971 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-13T06:50:28Z","lastTransitionTime":"2025-12-13T06:50:28Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 13 06:50:28 crc kubenswrapper[4971]: I1213 06:50:28.990316 4971 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 13 06:50:28 crc kubenswrapper[4971]: I1213 06:50:28.990355 4971 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 13 06:50:28 crc kubenswrapper[4971]: I1213 06:50:28.990366 4971 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 13 06:50:28 crc kubenswrapper[4971]: I1213 06:50:28.990382 4971 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 13 06:50:28 crc kubenswrapper[4971]: I1213 06:50:28.990393 4971 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-13T06:50:28Z","lastTransitionTime":"2025-12-13T06:50:28Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 13 06:50:29 crc kubenswrapper[4971]: I1213 06:50:29.092149 4971 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 13 06:50:29 crc kubenswrapper[4971]: I1213 06:50:29.092183 4971 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 13 06:50:29 crc kubenswrapper[4971]: I1213 06:50:29.092200 4971 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 13 06:50:29 crc kubenswrapper[4971]: I1213 06:50:29.092215 4971 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 13 06:50:29 crc kubenswrapper[4971]: I1213 06:50:29.092225 4971 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-13T06:50:29Z","lastTransitionTime":"2025-12-13T06:50:29Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 13 06:50:29 crc kubenswrapper[4971]: I1213 06:50:29.195595 4971 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 13 06:50:29 crc kubenswrapper[4971]: I1213 06:50:29.195637 4971 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 13 06:50:29 crc kubenswrapper[4971]: I1213 06:50:29.195648 4971 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 13 06:50:29 crc kubenswrapper[4971]: I1213 06:50:29.195666 4971 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 13 06:50:29 crc kubenswrapper[4971]: I1213 06:50:29.195676 4971 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-13T06:50:29Z","lastTransitionTime":"2025-12-13T06:50:29Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 13 06:50:29 crc kubenswrapper[4971]: I1213 06:50:29.298595 4971 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 13 06:50:29 crc kubenswrapper[4971]: I1213 06:50:29.298629 4971 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 13 06:50:29 crc kubenswrapper[4971]: I1213 06:50:29.298639 4971 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 13 06:50:29 crc kubenswrapper[4971]: I1213 06:50:29.298654 4971 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 13 06:50:29 crc kubenswrapper[4971]: I1213 06:50:29.298664 4971 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-13T06:50:29Z","lastTransitionTime":"2025-12-13T06:50:29Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 13 06:50:29 crc kubenswrapper[4971]: I1213 06:50:29.401217 4971 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 13 06:50:29 crc kubenswrapper[4971]: I1213 06:50:29.401258 4971 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 13 06:50:29 crc kubenswrapper[4971]: I1213 06:50:29.401268 4971 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 13 06:50:29 crc kubenswrapper[4971]: I1213 06:50:29.401281 4971 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 13 06:50:29 crc kubenswrapper[4971]: I1213 06:50:29.401290 4971 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-13T06:50:29Z","lastTransitionTime":"2025-12-13T06:50:29Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 13 06:50:29 crc kubenswrapper[4971]: I1213 06:50:29.503533 4971 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 13 06:50:29 crc kubenswrapper[4971]: I1213 06:50:29.503991 4971 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 13 06:50:29 crc kubenswrapper[4971]: I1213 06:50:29.504074 4971 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 13 06:50:29 crc kubenswrapper[4971]: I1213 06:50:29.504217 4971 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 13 06:50:29 crc kubenswrapper[4971]: I1213 06:50:29.504300 4971 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-13T06:50:29Z","lastTransitionTime":"2025-12-13T06:50:29Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 13 06:50:29 crc kubenswrapper[4971]: I1213 06:50:29.607280 4971 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 13 06:50:29 crc kubenswrapper[4971]: I1213 06:50:29.607339 4971 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 13 06:50:29 crc kubenswrapper[4971]: I1213 06:50:29.607355 4971 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 13 06:50:29 crc kubenswrapper[4971]: I1213 06:50:29.607376 4971 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 13 06:50:29 crc kubenswrapper[4971]: I1213 06:50:29.607391 4971 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-13T06:50:29Z","lastTransitionTime":"2025-12-13T06:50:29Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 13 06:50:29 crc kubenswrapper[4971]: I1213 06:50:29.710024 4971 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 13 06:50:29 crc kubenswrapper[4971]: I1213 06:50:29.710072 4971 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 13 06:50:29 crc kubenswrapper[4971]: I1213 06:50:29.710083 4971 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 13 06:50:29 crc kubenswrapper[4971]: I1213 06:50:29.710103 4971 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 13 06:50:29 crc kubenswrapper[4971]: I1213 06:50:29.710115 4971 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-13T06:50:29Z","lastTransitionTime":"2025-12-13T06:50:29Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 13 06:50:29 crc kubenswrapper[4971]: I1213 06:50:29.767834 4971 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-x98k7" Dec 13 06:50:29 crc kubenswrapper[4971]: I1213 06:50:29.767884 4971 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 13 06:50:29 crc kubenswrapper[4971]: E1213 06:50:29.768036 4971 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-x98k7" podUID="c7185dec-3240-42e8-9d90-6182a5e8b65c" Dec 13 06:50:29 crc kubenswrapper[4971]: I1213 06:50:29.768056 4971 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 13 06:50:29 crc kubenswrapper[4971]: I1213 06:50:29.768107 4971 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 13 06:50:29 crc kubenswrapper[4971]: E1213 06:50:29.768196 4971 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 13 06:50:29 crc kubenswrapper[4971]: E1213 06:50:29.768269 4971 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 13 06:50:29 crc kubenswrapper[4971]: E1213 06:50:29.768344 4971 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 13 06:50:29 crc kubenswrapper[4971]: I1213 06:50:29.817686 4971 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 13 06:50:29 crc kubenswrapper[4971]: I1213 06:50:29.817746 4971 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 13 06:50:29 crc kubenswrapper[4971]: I1213 06:50:29.817761 4971 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 13 06:50:29 crc kubenswrapper[4971]: I1213 06:50:29.817784 4971 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 13 06:50:29 crc kubenswrapper[4971]: I1213 06:50:29.817806 4971 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-13T06:50:29Z","lastTransitionTime":"2025-12-13T06:50:29Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 13 06:50:29 crc kubenswrapper[4971]: I1213 06:50:29.920743 4971 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 13 06:50:29 crc kubenswrapper[4971]: I1213 06:50:29.920781 4971 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 13 06:50:29 crc kubenswrapper[4971]: I1213 06:50:29.920810 4971 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 13 06:50:29 crc kubenswrapper[4971]: I1213 06:50:29.920828 4971 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 13 06:50:29 crc kubenswrapper[4971]: I1213 06:50:29.920839 4971 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-13T06:50:29Z","lastTransitionTime":"2025-12-13T06:50:29Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 13 06:50:30 crc kubenswrapper[4971]: I1213 06:50:30.023113 4971 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 13 06:50:30 crc kubenswrapper[4971]: I1213 06:50:30.023144 4971 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 13 06:50:30 crc kubenswrapper[4971]: I1213 06:50:30.023153 4971 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 13 06:50:30 crc kubenswrapper[4971]: I1213 06:50:30.023166 4971 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 13 06:50:30 crc kubenswrapper[4971]: I1213 06:50:30.023174 4971 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-13T06:50:30Z","lastTransitionTime":"2025-12-13T06:50:30Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 13 06:50:30 crc kubenswrapper[4971]: I1213 06:50:30.125430 4971 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 13 06:50:30 crc kubenswrapper[4971]: I1213 06:50:30.125495 4971 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 13 06:50:30 crc kubenswrapper[4971]: I1213 06:50:30.125509 4971 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 13 06:50:30 crc kubenswrapper[4971]: I1213 06:50:30.125579 4971 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 13 06:50:30 crc kubenswrapper[4971]: I1213 06:50:30.125592 4971 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-13T06:50:30Z","lastTransitionTime":"2025-12-13T06:50:30Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 13 06:50:30 crc kubenswrapper[4971]: I1213 06:50:30.228314 4971 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 13 06:50:30 crc kubenswrapper[4971]: I1213 06:50:30.228359 4971 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 13 06:50:30 crc kubenswrapper[4971]: I1213 06:50:30.228378 4971 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 13 06:50:30 crc kubenswrapper[4971]: I1213 06:50:30.228397 4971 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 13 06:50:30 crc kubenswrapper[4971]: I1213 06:50:30.228408 4971 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-13T06:50:30Z","lastTransitionTime":"2025-12-13T06:50:30Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 13 06:50:30 crc kubenswrapper[4971]: I1213 06:50:30.330296 4971 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 13 06:50:30 crc kubenswrapper[4971]: I1213 06:50:30.330330 4971 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 13 06:50:30 crc kubenswrapper[4971]: I1213 06:50:30.330341 4971 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 13 06:50:30 crc kubenswrapper[4971]: I1213 06:50:30.330354 4971 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 13 06:50:30 crc kubenswrapper[4971]: I1213 06:50:30.330373 4971 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-13T06:50:30Z","lastTransitionTime":"2025-12-13T06:50:30Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 13 06:50:30 crc kubenswrapper[4971]: I1213 06:50:30.432340 4971 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 13 06:50:30 crc kubenswrapper[4971]: I1213 06:50:30.432393 4971 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 13 06:50:30 crc kubenswrapper[4971]: I1213 06:50:30.432406 4971 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 13 06:50:30 crc kubenswrapper[4971]: I1213 06:50:30.432423 4971 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 13 06:50:30 crc kubenswrapper[4971]: I1213 06:50:30.432435 4971 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-13T06:50:30Z","lastTransitionTime":"2025-12-13T06:50:30Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 13 06:50:30 crc kubenswrapper[4971]: I1213 06:50:30.534988 4971 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 13 06:50:30 crc kubenswrapper[4971]: I1213 06:50:30.535035 4971 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 13 06:50:30 crc kubenswrapper[4971]: I1213 06:50:30.535047 4971 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 13 06:50:30 crc kubenswrapper[4971]: I1213 06:50:30.535067 4971 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 13 06:50:30 crc kubenswrapper[4971]: I1213 06:50:30.535079 4971 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-13T06:50:30Z","lastTransitionTime":"2025-12-13T06:50:30Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 13 06:50:30 crc kubenswrapper[4971]: I1213 06:50:30.637145 4971 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 13 06:50:30 crc kubenswrapper[4971]: I1213 06:50:30.637179 4971 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 13 06:50:30 crc kubenswrapper[4971]: I1213 06:50:30.637188 4971 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 13 06:50:30 crc kubenswrapper[4971]: I1213 06:50:30.637201 4971 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 13 06:50:30 crc kubenswrapper[4971]: I1213 06:50:30.637211 4971 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-13T06:50:30Z","lastTransitionTime":"2025-12-13T06:50:30Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 13 06:50:30 crc kubenswrapper[4971]: I1213 06:50:30.739176 4971 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 13 06:50:30 crc kubenswrapper[4971]: I1213 06:50:30.739214 4971 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 13 06:50:30 crc kubenswrapper[4971]: I1213 06:50:30.739225 4971 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 13 06:50:30 crc kubenswrapper[4971]: I1213 06:50:30.739239 4971 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 13 06:50:30 crc kubenswrapper[4971]: I1213 06:50:30.739249 4971 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-13T06:50:30Z","lastTransitionTime":"2025-12-13T06:50:30Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 13 06:50:30 crc kubenswrapper[4971]: I1213 06:50:30.840969 4971 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 13 06:50:30 crc kubenswrapper[4971]: I1213 06:50:30.841015 4971 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 13 06:50:30 crc kubenswrapper[4971]: I1213 06:50:30.841030 4971 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 13 06:50:30 crc kubenswrapper[4971]: I1213 06:50:30.841051 4971 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 13 06:50:30 crc kubenswrapper[4971]: I1213 06:50:30.841067 4971 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-13T06:50:30Z","lastTransitionTime":"2025-12-13T06:50:30Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 13 06:50:30 crc kubenswrapper[4971]: I1213 06:50:30.943572 4971 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 13 06:50:30 crc kubenswrapper[4971]: I1213 06:50:30.943621 4971 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 13 06:50:30 crc kubenswrapper[4971]: I1213 06:50:30.943631 4971 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 13 06:50:30 crc kubenswrapper[4971]: I1213 06:50:30.943648 4971 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 13 06:50:30 crc kubenswrapper[4971]: I1213 06:50:30.943660 4971 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-13T06:50:30Z","lastTransitionTime":"2025-12-13T06:50:30Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 13 06:50:31 crc kubenswrapper[4971]: I1213 06:50:31.047010 4971 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 13 06:50:31 crc kubenswrapper[4971]: I1213 06:50:31.047048 4971 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 13 06:50:31 crc kubenswrapper[4971]: I1213 06:50:31.047057 4971 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 13 06:50:31 crc kubenswrapper[4971]: I1213 06:50:31.047072 4971 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 13 06:50:31 crc kubenswrapper[4971]: I1213 06:50:31.047083 4971 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-13T06:50:31Z","lastTransitionTime":"2025-12-13T06:50:31Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 13 06:50:31 crc kubenswrapper[4971]: I1213 06:50:31.149581 4971 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 13 06:50:31 crc kubenswrapper[4971]: I1213 06:50:31.149687 4971 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 13 06:50:31 crc kubenswrapper[4971]: I1213 06:50:31.149703 4971 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 13 06:50:31 crc kubenswrapper[4971]: I1213 06:50:31.149722 4971 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 13 06:50:31 crc kubenswrapper[4971]: I1213 06:50:31.149733 4971 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-13T06:50:31Z","lastTransitionTime":"2025-12-13T06:50:31Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 13 06:50:31 crc kubenswrapper[4971]: I1213 06:50:31.252027 4971 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 13 06:50:31 crc kubenswrapper[4971]: I1213 06:50:31.252068 4971 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 13 06:50:31 crc kubenswrapper[4971]: I1213 06:50:31.252078 4971 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 13 06:50:31 crc kubenswrapper[4971]: I1213 06:50:31.252092 4971 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 13 06:50:31 crc kubenswrapper[4971]: I1213 06:50:31.252101 4971 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-13T06:50:31Z","lastTransitionTime":"2025-12-13T06:50:31Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 13 06:50:31 crc kubenswrapper[4971]: I1213 06:50:31.353799 4971 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 13 06:50:31 crc kubenswrapper[4971]: I1213 06:50:31.353872 4971 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 13 06:50:31 crc kubenswrapper[4971]: I1213 06:50:31.353885 4971 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 13 06:50:31 crc kubenswrapper[4971]: I1213 06:50:31.354096 4971 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 13 06:50:31 crc kubenswrapper[4971]: I1213 06:50:31.354107 4971 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-13T06:50:31Z","lastTransitionTime":"2025-12-13T06:50:31Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 13 06:50:31 crc kubenswrapper[4971]: I1213 06:50:31.456329 4971 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 13 06:50:31 crc kubenswrapper[4971]: I1213 06:50:31.456369 4971 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 13 06:50:31 crc kubenswrapper[4971]: I1213 06:50:31.456378 4971 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 13 06:50:31 crc kubenswrapper[4971]: I1213 06:50:31.456396 4971 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 13 06:50:31 crc kubenswrapper[4971]: I1213 06:50:31.456407 4971 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-13T06:50:31Z","lastTransitionTime":"2025-12-13T06:50:31Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 13 06:50:31 crc kubenswrapper[4971]: I1213 06:50:31.558825 4971 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 13 06:50:31 crc kubenswrapper[4971]: I1213 06:50:31.558857 4971 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 13 06:50:31 crc kubenswrapper[4971]: I1213 06:50:31.558866 4971 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 13 06:50:31 crc kubenswrapper[4971]: I1213 06:50:31.558879 4971 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 13 06:50:31 crc kubenswrapper[4971]: I1213 06:50:31.558888 4971 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-13T06:50:31Z","lastTransitionTime":"2025-12-13T06:50:31Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 13 06:50:31 crc kubenswrapper[4971]: I1213 06:50:31.661261 4971 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 13 06:50:31 crc kubenswrapper[4971]: I1213 06:50:31.661303 4971 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 13 06:50:31 crc kubenswrapper[4971]: I1213 06:50:31.661315 4971 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 13 06:50:31 crc kubenswrapper[4971]: I1213 06:50:31.661333 4971 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 13 06:50:31 crc kubenswrapper[4971]: I1213 06:50:31.661343 4971 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-13T06:50:31Z","lastTransitionTime":"2025-12-13T06:50:31Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 13 06:50:31 crc kubenswrapper[4971]: I1213 06:50:31.763417 4971 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 13 06:50:31 crc kubenswrapper[4971]: I1213 06:50:31.764120 4971 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 13 06:50:31 crc kubenswrapper[4971]: I1213 06:50:31.764142 4971 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 13 06:50:31 crc kubenswrapper[4971]: I1213 06:50:31.764164 4971 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 13 06:50:31 crc kubenswrapper[4971]: I1213 06:50:31.764178 4971 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-13T06:50:31Z","lastTransitionTime":"2025-12-13T06:50:31Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 13 06:50:31 crc kubenswrapper[4971]: I1213 06:50:31.767753 4971 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 13 06:50:31 crc kubenswrapper[4971]: I1213 06:50:31.767802 4971 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-x98k7" Dec 13 06:50:31 crc kubenswrapper[4971]: E1213 06:50:31.767897 4971 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 13 06:50:31 crc kubenswrapper[4971]: I1213 06:50:31.767763 4971 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 13 06:50:31 crc kubenswrapper[4971]: E1213 06:50:31.768008 4971 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-x98k7" podUID="c7185dec-3240-42e8-9d90-6182a5e8b65c" Dec 13 06:50:31 crc kubenswrapper[4971]: I1213 06:50:31.768041 4971 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 13 06:50:31 crc kubenswrapper[4971]: E1213 06:50:31.768092 4971 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 13 06:50:31 crc kubenswrapper[4971]: E1213 06:50:31.768140 4971 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 13 06:50:31 crc kubenswrapper[4971]: I1213 06:50:31.866674 4971 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 13 06:50:31 crc kubenswrapper[4971]: I1213 06:50:31.866726 4971 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 13 06:50:31 crc kubenswrapper[4971]: I1213 06:50:31.866737 4971 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 13 06:50:31 crc kubenswrapper[4971]: I1213 06:50:31.866757 4971 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 13 06:50:31 crc kubenswrapper[4971]: I1213 06:50:31.866769 4971 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-13T06:50:31Z","lastTransitionTime":"2025-12-13T06:50:31Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 13 06:50:31 crc kubenswrapper[4971]: I1213 06:50:31.969051 4971 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 13 06:50:31 crc kubenswrapper[4971]: I1213 06:50:31.969086 4971 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 13 06:50:31 crc kubenswrapper[4971]: I1213 06:50:31.969097 4971 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 13 06:50:31 crc kubenswrapper[4971]: I1213 06:50:31.969112 4971 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 13 06:50:31 crc kubenswrapper[4971]: I1213 06:50:31.969123 4971 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-13T06:50:31Z","lastTransitionTime":"2025-12-13T06:50:31Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 13 06:50:32 crc kubenswrapper[4971]: I1213 06:50:32.047795 4971 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 13 06:50:32 crc kubenswrapper[4971]: I1213 06:50:32.047840 4971 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 13 06:50:32 crc kubenswrapper[4971]: I1213 06:50:32.047851 4971 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 13 06:50:32 crc kubenswrapper[4971]: I1213 06:50:32.047869 4971 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 13 06:50:32 crc kubenswrapper[4971]: I1213 06:50:32.047884 4971 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-13T06:50:32Z","lastTransitionTime":"2025-12-13T06:50:32Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 13 06:50:32 crc kubenswrapper[4971]: E1213 06:50:32.065198 4971 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-12-13T06:50:32Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-13T06:50:32Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-13T06:50:32Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-13T06:50:32Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-13T06:50:32Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-13T06:50:32Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-13T06:50:32Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-13T06:50:32Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"bd36cbc9-bbe9-45a1-a21a-ad2bd10499ee\\\",\\\"systemUUID\\\":\\\"a9af00dc-cf25-47af-9997-79577f8f3bea\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-13T06:50:32Z is after 2025-08-24T17:21:41Z" Dec 13 06:50:32 crc kubenswrapper[4971]: I1213 06:50:32.069226 4971 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 13 06:50:32 crc kubenswrapper[4971]: I1213 06:50:32.069278 4971 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 13 06:50:32 crc kubenswrapper[4971]: I1213 06:50:32.069290 4971 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 13 06:50:32 crc kubenswrapper[4971]: I1213 06:50:32.069308 4971 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 13 06:50:32 crc kubenswrapper[4971]: I1213 06:50:32.069321 4971 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-13T06:50:32Z","lastTransitionTime":"2025-12-13T06:50:32Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 13 06:50:32 crc kubenswrapper[4971]: E1213 06:50:32.088541 4971 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-12-13T06:50:32Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-13T06:50:32Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-13T06:50:32Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-13T06:50:32Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-13T06:50:32Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-13T06:50:32Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-13T06:50:32Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-13T06:50:32Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"bd36cbc9-bbe9-45a1-a21a-ad2bd10499ee\\\",\\\"systemUUID\\\":\\\"a9af00dc-cf25-47af-9997-79577f8f3bea\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-13T06:50:32Z is after 2025-08-24T17:21:41Z" Dec 13 06:50:32 crc kubenswrapper[4971]: I1213 06:50:32.091695 4971 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 13 06:50:32 crc kubenswrapper[4971]: I1213 06:50:32.091719 4971 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 13 06:50:32 crc kubenswrapper[4971]: I1213 06:50:32.091728 4971 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 13 06:50:32 crc kubenswrapper[4971]: I1213 06:50:32.091741 4971 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 13 06:50:32 crc kubenswrapper[4971]: I1213 06:50:32.091749 4971 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-13T06:50:32Z","lastTransitionTime":"2025-12-13T06:50:32Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 13 06:50:32 crc kubenswrapper[4971]: E1213 06:50:32.102233 4971 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-12-13T06:50:32Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-13T06:50:32Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-13T06:50:32Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-13T06:50:32Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-13T06:50:32Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-13T06:50:32Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-13T06:50:32Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-13T06:50:32Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"bd36cbc9-bbe9-45a1-a21a-ad2bd10499ee\\\",\\\"systemUUID\\\":\\\"a9af00dc-cf25-47af-9997-79577f8f3bea\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-13T06:50:32Z is after 2025-08-24T17:21:41Z" Dec 13 06:50:32 crc kubenswrapper[4971]: I1213 06:50:32.105603 4971 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 13 06:50:32 crc kubenswrapper[4971]: I1213 06:50:32.105636 4971 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 13 06:50:32 crc kubenswrapper[4971]: I1213 06:50:32.105646 4971 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 13 06:50:32 crc kubenswrapper[4971]: I1213 06:50:32.105662 4971 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 13 06:50:32 crc kubenswrapper[4971]: I1213 06:50:32.105673 4971 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-13T06:50:32Z","lastTransitionTime":"2025-12-13T06:50:32Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 13 06:50:32 crc kubenswrapper[4971]: E1213 06:50:32.117069 4971 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-12-13T06:50:32Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-13T06:50:32Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-13T06:50:32Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-13T06:50:32Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-13T06:50:32Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-13T06:50:32Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-13T06:50:32Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-13T06:50:32Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"bd36cbc9-bbe9-45a1-a21a-ad2bd10499ee\\\",\\\"systemUUID\\\":\\\"a9af00dc-cf25-47af-9997-79577f8f3bea\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-13T06:50:32Z is after 2025-08-24T17:21:41Z" Dec 13 06:50:32 crc kubenswrapper[4971]: I1213 06:50:32.120033 4971 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 13 06:50:32 crc kubenswrapper[4971]: I1213 06:50:32.120065 4971 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 13 06:50:32 crc kubenswrapper[4971]: I1213 06:50:32.120077 4971 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 13 06:50:32 crc kubenswrapper[4971]: I1213 06:50:32.120093 4971 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 13 06:50:32 crc kubenswrapper[4971]: I1213 06:50:32.120104 4971 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-13T06:50:32Z","lastTransitionTime":"2025-12-13T06:50:32Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 13 06:50:32 crc kubenswrapper[4971]: E1213 06:50:32.130116 4971 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-12-13T06:50:32Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-13T06:50:32Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-13T06:50:32Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-13T06:50:32Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-13T06:50:32Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-13T06:50:32Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-13T06:50:32Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-13T06:50:32Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"bd36cbc9-bbe9-45a1-a21a-ad2bd10499ee\\\",\\\"systemUUID\\\":\\\"a9af00dc-cf25-47af-9997-79577f8f3bea\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-13T06:50:32Z is after 2025-08-24T17:21:41Z" Dec 13 06:50:32 crc kubenswrapper[4971]: E1213 06:50:32.130226 4971 kubelet_node_status.go:572] "Unable to update node status" err="update node status exceeds retry count" Dec 13 06:50:32 crc kubenswrapper[4971]: I1213 06:50:32.131602 4971 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 13 06:50:32 crc kubenswrapper[4971]: I1213 06:50:32.131628 4971 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 13 06:50:32 crc kubenswrapper[4971]: I1213 06:50:32.131636 4971 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 13 06:50:32 crc kubenswrapper[4971]: I1213 06:50:32.131649 4971 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 13 06:50:32 crc kubenswrapper[4971]: I1213 06:50:32.131658 4971 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-13T06:50:32Z","lastTransitionTime":"2025-12-13T06:50:32Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 13 06:50:32 crc kubenswrapper[4971]: I1213 06:50:32.234330 4971 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 13 06:50:32 crc kubenswrapper[4971]: I1213 06:50:32.234364 4971 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 13 06:50:32 crc kubenswrapper[4971]: I1213 06:50:32.234375 4971 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 13 06:50:32 crc kubenswrapper[4971]: I1213 06:50:32.234389 4971 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 13 06:50:32 crc kubenswrapper[4971]: I1213 06:50:32.234398 4971 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-13T06:50:32Z","lastTransitionTime":"2025-12-13T06:50:32Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 13 06:50:32 crc kubenswrapper[4971]: I1213 06:50:32.336648 4971 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 13 06:50:32 crc kubenswrapper[4971]: I1213 06:50:32.336680 4971 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 13 06:50:32 crc kubenswrapper[4971]: I1213 06:50:32.336758 4971 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 13 06:50:32 crc kubenswrapper[4971]: I1213 06:50:32.336783 4971 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 13 06:50:32 crc kubenswrapper[4971]: I1213 06:50:32.336795 4971 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-13T06:50:32Z","lastTransitionTime":"2025-12-13T06:50:32Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 13 06:50:32 crc kubenswrapper[4971]: I1213 06:50:32.439276 4971 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 13 06:50:32 crc kubenswrapper[4971]: I1213 06:50:32.439314 4971 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 13 06:50:32 crc kubenswrapper[4971]: I1213 06:50:32.439326 4971 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 13 06:50:32 crc kubenswrapper[4971]: I1213 06:50:32.439344 4971 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 13 06:50:32 crc kubenswrapper[4971]: I1213 06:50:32.439356 4971 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-13T06:50:32Z","lastTransitionTime":"2025-12-13T06:50:32Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 13 06:50:32 crc kubenswrapper[4971]: I1213 06:50:32.541944 4971 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 13 06:50:32 crc kubenswrapper[4971]: I1213 06:50:32.541991 4971 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 13 06:50:32 crc kubenswrapper[4971]: I1213 06:50:32.542008 4971 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 13 06:50:32 crc kubenswrapper[4971]: I1213 06:50:32.542031 4971 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 13 06:50:32 crc kubenswrapper[4971]: I1213 06:50:32.542047 4971 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-13T06:50:32Z","lastTransitionTime":"2025-12-13T06:50:32Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 13 06:50:32 crc kubenswrapper[4971]: I1213 06:50:32.644363 4971 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 13 06:50:32 crc kubenswrapper[4971]: I1213 06:50:32.644402 4971 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 13 06:50:32 crc kubenswrapper[4971]: I1213 06:50:32.644433 4971 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 13 06:50:32 crc kubenswrapper[4971]: I1213 06:50:32.644446 4971 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 13 06:50:32 crc kubenswrapper[4971]: I1213 06:50:32.644456 4971 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-13T06:50:32Z","lastTransitionTime":"2025-12-13T06:50:32Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 13 06:50:32 crc kubenswrapper[4971]: I1213 06:50:32.747029 4971 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 13 06:50:32 crc kubenswrapper[4971]: I1213 06:50:32.747062 4971 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 13 06:50:32 crc kubenswrapper[4971]: I1213 06:50:32.747073 4971 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 13 06:50:32 crc kubenswrapper[4971]: I1213 06:50:32.747089 4971 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 13 06:50:32 crc kubenswrapper[4971]: I1213 06:50:32.747100 4971 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-13T06:50:32Z","lastTransitionTime":"2025-12-13T06:50:32Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 13 06:50:32 crc kubenswrapper[4971]: I1213 06:50:32.849939 4971 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 13 06:50:32 crc kubenswrapper[4971]: I1213 06:50:32.849980 4971 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 13 06:50:32 crc kubenswrapper[4971]: I1213 06:50:32.849989 4971 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 13 06:50:32 crc kubenswrapper[4971]: I1213 06:50:32.850004 4971 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 13 06:50:32 crc kubenswrapper[4971]: I1213 06:50:32.850013 4971 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-13T06:50:32Z","lastTransitionTime":"2025-12-13T06:50:32Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 13 06:50:32 crc kubenswrapper[4971]: I1213 06:50:32.952257 4971 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 13 06:50:32 crc kubenswrapper[4971]: I1213 06:50:32.952302 4971 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 13 06:50:32 crc kubenswrapper[4971]: I1213 06:50:32.952313 4971 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 13 06:50:32 crc kubenswrapper[4971]: I1213 06:50:32.952362 4971 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 13 06:50:32 crc kubenswrapper[4971]: I1213 06:50:32.952375 4971 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-13T06:50:32Z","lastTransitionTime":"2025-12-13T06:50:32Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 13 06:50:33 crc kubenswrapper[4971]: I1213 06:50:33.054555 4971 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 13 06:50:33 crc kubenswrapper[4971]: I1213 06:50:33.054601 4971 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 13 06:50:33 crc kubenswrapper[4971]: I1213 06:50:33.054610 4971 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 13 06:50:33 crc kubenswrapper[4971]: I1213 06:50:33.054626 4971 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 13 06:50:33 crc kubenswrapper[4971]: I1213 06:50:33.054635 4971 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-13T06:50:33Z","lastTransitionTime":"2025-12-13T06:50:33Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 13 06:50:33 crc kubenswrapper[4971]: I1213 06:50:33.157163 4971 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 13 06:50:33 crc kubenswrapper[4971]: I1213 06:50:33.157213 4971 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 13 06:50:33 crc kubenswrapper[4971]: I1213 06:50:33.157223 4971 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 13 06:50:33 crc kubenswrapper[4971]: I1213 06:50:33.157240 4971 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 13 06:50:33 crc kubenswrapper[4971]: I1213 06:50:33.157252 4971 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-13T06:50:33Z","lastTransitionTime":"2025-12-13T06:50:33Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 13 06:50:33 crc kubenswrapper[4971]: I1213 06:50:33.259533 4971 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 13 06:50:33 crc kubenswrapper[4971]: I1213 06:50:33.259571 4971 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 13 06:50:33 crc kubenswrapper[4971]: I1213 06:50:33.259580 4971 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 13 06:50:33 crc kubenswrapper[4971]: I1213 06:50:33.259597 4971 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 13 06:50:33 crc kubenswrapper[4971]: I1213 06:50:33.259608 4971 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-13T06:50:33Z","lastTransitionTime":"2025-12-13T06:50:33Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 13 06:50:33 crc kubenswrapper[4971]: I1213 06:50:33.362127 4971 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 13 06:50:33 crc kubenswrapper[4971]: I1213 06:50:33.362193 4971 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 13 06:50:33 crc kubenswrapper[4971]: I1213 06:50:33.362209 4971 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 13 06:50:33 crc kubenswrapper[4971]: I1213 06:50:33.362232 4971 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 13 06:50:33 crc kubenswrapper[4971]: I1213 06:50:33.362248 4971 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-13T06:50:33Z","lastTransitionTime":"2025-12-13T06:50:33Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 13 06:50:33 crc kubenswrapper[4971]: I1213 06:50:33.464623 4971 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 13 06:50:33 crc kubenswrapper[4971]: I1213 06:50:33.464675 4971 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 13 06:50:33 crc kubenswrapper[4971]: I1213 06:50:33.464688 4971 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 13 06:50:33 crc kubenswrapper[4971]: I1213 06:50:33.464707 4971 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 13 06:50:33 crc kubenswrapper[4971]: I1213 06:50:33.464720 4971 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-13T06:50:33Z","lastTransitionTime":"2025-12-13T06:50:33Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 13 06:50:33 crc kubenswrapper[4971]: I1213 06:50:33.567437 4971 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 13 06:50:33 crc kubenswrapper[4971]: I1213 06:50:33.567473 4971 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 13 06:50:33 crc kubenswrapper[4971]: I1213 06:50:33.567488 4971 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 13 06:50:33 crc kubenswrapper[4971]: I1213 06:50:33.567508 4971 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 13 06:50:33 crc kubenswrapper[4971]: I1213 06:50:33.567539 4971 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-13T06:50:33Z","lastTransitionTime":"2025-12-13T06:50:33Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 13 06:50:33 crc kubenswrapper[4971]: I1213 06:50:33.670158 4971 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 13 06:50:33 crc kubenswrapper[4971]: I1213 06:50:33.670226 4971 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 13 06:50:33 crc kubenswrapper[4971]: I1213 06:50:33.670246 4971 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 13 06:50:33 crc kubenswrapper[4971]: I1213 06:50:33.670262 4971 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 13 06:50:33 crc kubenswrapper[4971]: I1213 06:50:33.670271 4971 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-13T06:50:33Z","lastTransitionTime":"2025-12-13T06:50:33Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 13 06:50:33 crc kubenswrapper[4971]: I1213 06:50:33.768317 4971 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 13 06:50:33 crc kubenswrapper[4971]: I1213 06:50:33.768406 4971 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 13 06:50:33 crc kubenswrapper[4971]: I1213 06:50:33.768446 4971 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-x98k7" Dec 13 06:50:33 crc kubenswrapper[4971]: E1213 06:50:33.768673 4971 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 13 06:50:33 crc kubenswrapper[4971]: I1213 06:50:33.768954 4971 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 13 06:50:33 crc kubenswrapper[4971]: E1213 06:50:33.768947 4971 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 13 06:50:33 crc kubenswrapper[4971]: E1213 06:50:33.769092 4971 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-x98k7" podUID="c7185dec-3240-42e8-9d90-6182a5e8b65c" Dec 13 06:50:33 crc kubenswrapper[4971]: E1213 06:50:33.769336 4971 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 13 06:50:33 crc kubenswrapper[4971]: I1213 06:50:33.772903 4971 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 13 06:50:33 crc kubenswrapper[4971]: I1213 06:50:33.772944 4971 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 13 06:50:33 crc kubenswrapper[4971]: I1213 06:50:33.772962 4971 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 13 06:50:33 crc kubenswrapper[4971]: I1213 06:50:33.772984 4971 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 13 06:50:33 crc kubenswrapper[4971]: I1213 06:50:33.773000 4971 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-13T06:50:33Z","lastTransitionTime":"2025-12-13T06:50:33Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 13 06:50:33 crc kubenswrapper[4971]: I1213 06:50:33.786376 4971 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"f9d31b5a-b770-4a15-95c8-35b49677c608\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T06:49:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T06:49:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T06:49:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T06:49:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T06:49:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c9baa042f9dada954faa57b5454cd6b0a12a297e56ddc192c52cff50e42a16fa\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-13T06:49:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://c788de86754f0bf37ee184672fc1dcb388ccebc246d4f15be4b5da661f725f8f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-13T06:49:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://760817511a9b4f96a68dc38d26631250eaab72e1dd8f123e653039b6b4e91f84\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-13T06:49:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b859be0c9feac856e9855c6a038b02b49b32fade97873c46d0bec0a32d8b5147\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-13T06:49:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-13T06:49:03Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-13T06:50:33Z is after 2025-08-24T17:21:41Z" Dec 13 06:50:33 crc kubenswrapper[4971]: I1213 06:50:33.800701 4971 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"8ede2992-9576-4b23-83c2-24ac3b1d3caf\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T06:49:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T06:49:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T06:49:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T06:49:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T06:49:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9e61d25add1723ac286794abfe886358e9d64455c25379f7e3057fc3dd812418\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-13T06:49:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://7fb51d8435b4e3ce0411cae1d6c2824c552f0a33d79de01f969274516b39a5b3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-13T06:49:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://1c35f5e8d8fab6d77b9e2e0dedb195440de5d6af345cd5865abd32caa7fafd60\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-13T06:49:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://38806c430185715371ab9bf30a1af9a9d5d492abd9bb84acbdccc694f56e3a8a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"wait-for-host-port\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://38806c430185715371ab9bf30a1af9a9d5d492abd9bb84acbdccc694f56e3a8a\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-13T06:49:04Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-13T06:49:04Z\\\"}}}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-13T06:49:03Z\\\"}}\" for pod \"openshift-kube-scheduler\"/\"openshift-kube-scheduler-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-13T06:50:33Z is after 2025-08-24T17:21:41Z" Dec 13 06:50:33 crc kubenswrapper[4971]: I1213 06:50:33.816241 4971 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-ncchz" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"de914d21-a688-401e-a66e-861beb6336d9\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T06:49:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T06:49:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T06:49:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T06:49:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://244ffe3dc9d8cf799aff4c4d795bf41554c1e6bb6667fbb0c7df322fb5f96c68\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-13T06:49:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lqmjj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://0b7111af6c5a198d02bb02858837cfb4f3b20828d8d206cd18aa0ffa102d0cd4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://0b7111af6c5a198d02bb02858837cfb4f3b20828d8d206cd18aa0ffa102d0cd4\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-13T06:49:26Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-13T06:49:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lqmjj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b36777fe40e74ae1cecb7b73e204139db5279ba51cef4f1e8c4a4c491d8be403\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://b36777fe40e74ae1cecb7b73e204139db5279ba51cef4f1e8c4a4c491d8be403\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-13T06:49:28Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-13T06:49:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lqmjj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://0597a06cc663e416142311a16941d98297577693a1fe27201ea42c27496cace2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://0597a06cc663e416142311a16941d98297577693a1fe27201ea42c27496cace2\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-13T06:49:29Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-13T06:49:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lqmjj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://3f52c2e22a75c010993ca5809a3a9d1304758d4cb6e83691ef1af15a4b514ac4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://3f52c2e22a75c010993ca5809a3a9d1304758d4cb6e83691ef1af15a4b514ac4\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-13T06:49:30Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-13T06:49:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lqmjj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://9e21ca708c4c4c1228b417cece51299e0792729b7c78884bd18b2dda8fe8355c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://9e21ca708c4c4c1228b417cece51299e0792729b7c78884bd18b2dda8fe8355c\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-13T06:49:31Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-13T06:49:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lqmjj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://74552159f12c87cdda09147692b400ccac0d4e43c1ed72fe8462c4249ad69c7b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://74552159f12c87cdda09147692b400ccac0d4e43c1ed72fe8462c4249ad69c7b\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-13T06:49:32Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-13T06:49:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lqmjj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-13T06:49:25Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-ncchz\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-13T06:50:33Z is after 2025-08-24T17:21:41Z" Dec 13 06:50:33 crc kubenswrapper[4971]: I1213 06:50:33.827919 4971 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-gkcd8" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"13f5c327-f95e-42b4-8783-218500c17694\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T06:49:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T06:49:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T06:49:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T06:49:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://77e846e36f1cb40d03471dc03de11b6c3da448baa407aad38093fae9a801cf92\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-13T06:49:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gvlhn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-13T06:49:27Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-gkcd8\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-13T06:50:33Z is after 2025-08-24T17:21:41Z" Dec 13 06:50:33 crc kubenswrapper[4971]: I1213 06:50:33.837453 4971 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-x98k7" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"c7185dec-3240-42e8-9d90-6182a5e8b65c\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T06:49:39Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T06:49:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T06:49:39Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T06:49:39Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bj2gn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bj2gn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-13T06:49:39Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-x98k7\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-13T06:50:33Z is after 2025-08-24T17:21:41Z" Dec 13 06:50:33 crc kubenswrapper[4971]: I1213 06:50:33.847154 4971 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5ef36745-87bb-4be8-b941-aa8f865a5c83\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T06:49:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T06:49:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T06:49:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T06:49:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T06:49:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://35b737b0d2c5c5503ba2819f34d2de0599792a77beb1fbeb561d44efd1f212c5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-crio\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-13T06:49:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kube\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"var-lib-kubelet\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f0d864149bfcbb6227e6a6fbde23fff650d688850978eae0caf04515eb21cd85\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://f0d864149bfcbb6227e6a6fbde23fff650d688850978eae0caf04515eb21cd85\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-13T06:49:04Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-13T06:49:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var\\\",\\\"name\\\":\\\"var-lib-kubelet\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-13T06:49:03Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"kube-rbac-proxy-crio-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-13T06:50:33Z is after 2025-08-24T17:21:41Z" Dec 13 06:50:33 crc kubenswrapper[4971]: I1213 06:50:33.871093 4971 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"1ed9f82e-fe5c-4b76-a619-62e5abd88d3b\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T06:49:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T06:49:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T06:49:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T06:49:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T06:49:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e457329a78a2d24337f994e8a0a1f75a2a667244aa75eb84613d1e7138364aac\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-13T06:49:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://eb014de8140deef097d9bd94f3b9fa4ab3558bc4449bef84b83bf62bfabf5836\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-13T06:49:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://2c438cc74d555f7c9db4569d1c56b78b1088ed8105697d2c0f31e6c5136f3dd0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-13T06:49:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://aa6fa6e4340dcd066252df3c56d79a3a39ed531d5eea8d1d5941d6952368d3af\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-13T06:49:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://c4471b98854c52d5cd084d0ba3e64e5b08e9518ec3c372cce2c875a440522457\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-13T06:49:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7d2746335c98dae3a0f43024d34cb62ebcb42a01c8f9ce18d52b29d4c45598e5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://7d2746335c98dae3a0f43024d34cb62ebcb42a01c8f9ce18d52b29d4c45598e5\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-13T06:49:04Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-13T06:49:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://bc0e77d10a738128bea82122496bc42235fc1688f8907756d93e61b651902b57\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://bc0e77d10a738128bea82122496bc42235fc1688f8907756d93e61b651902b57\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-13T06:49:05Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-13T06:49:05Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://f821e418cba5e832da8839268bcbbd9630bdf319a212d5382bfa2c4b0b73a268\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://f821e418cba5e832da8839268bcbbd9630bdf319a212d5382bfa2c4b0b73a268\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-13T06:49:07Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-13T06:49:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-13T06:49:03Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-13T06:50:33Z is after 2025-08-24T17:21:41Z" Dec 13 06:50:33 crc kubenswrapper[4971]: I1213 06:50:33.874741 4971 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 13 06:50:33 crc kubenswrapper[4971]: I1213 06:50:33.874779 4971 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 13 06:50:33 crc kubenswrapper[4971]: I1213 06:50:33.874790 4971 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 13 06:50:33 crc kubenswrapper[4971]: I1213 06:50:33.874841 4971 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 13 06:50:33 crc kubenswrapper[4971]: I1213 06:50:33.874855 4971 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-13T06:50:33Z","lastTransitionTime":"2025-12-13T06:50:33Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 13 06:50:33 crc kubenswrapper[4971]: I1213 06:50:33.883150 4971 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-13T06:49:24Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-13T06:49:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-13T06:49:24Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4328cd6a8464ed468229bbfbc8265f969427916d72ebb92237dbe68efe70fa47\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-13T06:49:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://9d715ca82d9b5a594e41f2b98da6f4277da0e4e1dec114969f901c00f6d7553d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-13T06:49:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-13T06:50:33Z is after 2025-08-24T17:21:41Z" Dec 13 06:50:33 crc kubenswrapper[4971]: I1213 06:50:33.893837 4971 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-5rzh5" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"6bf4591d-2483-4132-9672-d0fd6c995968\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T06:49:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T06:49:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T06:49:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T06:49:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://50e2cb14e4ffb603acb06e2281615a99f06182d345c4e70f2446fa2077033bbb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-13T06:49:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dfxqn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-13T06:49:25Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-5rzh5\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-13T06:50:33Z is after 2025-08-24T17:21:41Z" Dec 13 06:50:33 crc kubenswrapper[4971]: I1213 06:50:33.905510 4971 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-z572s" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"b4146d3d-4034-49a9-bd37-164856255137\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T06:49:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T06:49:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T06:49:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T06:49:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://46fd4e1070835cfbf7f35d497eff7f047473fc4048ee07812925df5ebdf55342\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-13T06:49:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-j6l9b\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://dc79a0968f85723e6ad0dc269c8da1f7472adb1bb2b87edd017ee2295bfbf89c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-13T06:49:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-j6l9b\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-13T06:49:38Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-z572s\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-13T06:50:33Z is after 2025-08-24T17:21:41Z" Dec 13 06:50:33 crc kubenswrapper[4971]: I1213 06:50:33.918070 4971 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3ad3268a-a982-428c-9c06-4aaeddd372d9\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T06:49:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T06:49:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T06:49:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T06:49:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T06:49:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3716afdf2bc693be80a09fc9f7727d334f435caf1d4c8ec4811b3a1f1652cdfe\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-13T06:49:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://30df3abf8c79db8af52c6a918f73e1ce410ddafcf341636424c267e1eeb195a7\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-13T06:49:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://ccdf65f6d2b1eb3647a03ba37c305c32e34addfdf79ff937c8dddaa3e5d5b5c6\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-13T06:49:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://770b73dd0c2b6f38744e2c1cd6aa186a69bb87d1cd015a603ad2e33be06a8f6b\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://b3c1bab3d53f975b2e24582ce2771bc26d600e0dd7c30339dcfd18a071162ad4\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-12-13T06:49:23Z\\\",\\\"message\\\":\\\"file observer\\\\nW1213 06:49:23.283403 1 builder.go:272] unable to get owner reference (falling back to namespace): pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\nI1213 06:49:23.284336 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1213 06:49:23.286978 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-981773671/tls.crt::/tmp/serving-cert-981773671/tls.key\\\\\\\"\\\\nI1213 06:49:23.775072 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1213 06:49:23.778615 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1213 06:49:23.778645 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1213 06:49:23.778680 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1213 06:49:23.778691 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1213 06:49:23.785058 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1213 06:49:23.786150 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1213 06:49:23.786196 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1213 06:49:23.786223 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1213 06:49:23.786246 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1213 06:49:23.786269 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1213 06:49:23.786291 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI1213 06:49:23.785608 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF1213 06:49:23.787563 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-13T06:49:08Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-13T06:49:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://9acd07066deed483bd7fa61ee45fda09539d110717a134a56827f871186f92fc\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-13T06:49:07Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://38c91cfbaecfdbd3e3b663422a91ab4294610afb6bbf9ebdb77db24e2b76c682\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://38c91cfbaecfdbd3e3b663422a91ab4294610afb6bbf9ebdb77db24e2b76c682\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-13T06:49:05Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-13T06:49:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-13T06:49:03Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-13T06:50:33Z is after 2025-08-24T17:21:41Z" Dec 13 06:50:33 crc kubenswrapper[4971]: I1213 06:50:33.929542 4971 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-13T06:49:23Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-13T06:49:23Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-13T06:50:33Z is after 2025-08-24T17:21:41Z" Dec 13 06:50:33 crc kubenswrapper[4971]: I1213 06:50:33.940384 4971 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-13T06:49:23Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-13T06:49:23Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-13T06:50:33Z is after 2025-08-24T17:21:41Z" Dec 13 06:50:33 crc kubenswrapper[4971]: I1213 06:50:33.952453 4971 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-13T06:49:24Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-13T06:49:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-13T06:49:24Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d0b2ecf0d5b22f23217f42f259ae7ffb6359555b22c30eaebf667eaf254b3bab\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-13T06:49:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-13T06:50:33Z is after 2025-08-24T17:21:41Z" Dec 13 06:50:33 crc kubenswrapper[4971]: I1213 06:50:33.964572 4971 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-13T06:49:23Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-13T06:49:23Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-13T06:50:33Z is after 2025-08-24T17:21:41Z" Dec 13 06:50:33 crc kubenswrapper[4971]: I1213 06:50:33.976205 4971 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-13T06:49:26Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-13T06:49:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-13T06:49:26Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9c8a0d1db0e754462d805e82daf5f52ffd5457507a30456535ec59e096c9d092\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-13T06:49:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-13T06:50:33Z is after 2025-08-24T17:21:41Z" Dec 13 06:50:33 crc kubenswrapper[4971]: I1213 06:50:33.977726 4971 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 13 06:50:33 crc kubenswrapper[4971]: I1213 06:50:33.977768 4971 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 13 06:50:33 crc kubenswrapper[4971]: I1213 06:50:33.977779 4971 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 13 06:50:33 crc kubenswrapper[4971]: I1213 06:50:33.977796 4971 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 13 06:50:33 crc kubenswrapper[4971]: I1213 06:50:33.977806 4971 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-13T06:50:33Z","lastTransitionTime":"2025-12-13T06:50:33Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 13 06:50:33 crc kubenswrapper[4971]: I1213 06:50:33.994978 4971 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-82xjz" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e0c4af12-edda-41d7-9b44-f87396174ef2\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T06:49:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T06:49:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T06:49:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T06:49:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://0f4577d9e37f860813e221f2e208177d8d383fd04f673ce357a0a405ea29d1e6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-13T06:49:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cnxsq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://7e0d2020e62f1a277cff7047e1907dc747ed393a1f6c983e1ee258bd156c16c7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-13T06:49:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cnxsq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-13T06:49:25Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-82xjz\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-13T06:50:33Z is after 2025-08-24T17:21:41Z" Dec 13 06:50:34 crc kubenswrapper[4971]: I1213 06:50:34.007296 4971 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-wwvpl" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"dddf744b-2122-49ad-a980-105c09636e0f\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T06:49:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T06:49:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T06:50:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T06:50:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c8680aaa9e05095ac7e467c896d9e927bccff908925768ce5ed469ab89a72475\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://4676d897c7fc668587faec563f3551a056557879cee3e865192d2bd28edf62fc\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-12-13T06:50:12Z\\\",\\\"message\\\":\\\"2025-12-13T06:49:27+00:00 [cnibincopy] Successfully copied files in /usr/src/multus-cni/rhel9/bin/ to /host/opt/cni/bin/upgrade_263f091d-d847-4fad-88d3-fbf82c65c231\\\\n2025-12-13T06:49:27+00:00 [cnibincopy] Successfully moved files in /host/opt/cni/bin/upgrade_263f091d-d847-4fad-88d3-fbf82c65c231 to /host/opt/cni/bin/\\\\n2025-12-13T06:49:27Z [verbose] multus-daemon started\\\\n2025-12-13T06:49:27Z [verbose] Readiness Indicator file check\\\\n2025-12-13T06:50:12Z [error] have you checked that your default network is ready? still waiting for readinessindicatorfile @ /host/run/multus/cni/net.d/10-ovn-kubernetes.conf. pollimmediate error: timed out waiting for the condition\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-13T06:49:26Z\\\"}},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-13T06:50:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-czz9v\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-13T06:49:25Z\\\"}}\" for pod \"openshift-multus\"/\"multus-wwvpl\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-13T06:50:34Z is after 2025-08-24T17:21:41Z" Dec 13 06:50:34 crc kubenswrapper[4971]: I1213 06:50:34.031365 4971 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-zlcg5" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"736c6b1a-93dd-4f71-81b6-1c43fcf5f556\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T06:49:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T06:49:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T06:49:25Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T06:49:25Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://325223543b3c7db417f769533559a1ca2b23a723494e1885b0fd0480b5a6399f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-13T06:49:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vzmsr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://f22a4ec756119c4f791fb5105cf1f31df3e548b2b07b1f909a62802e9b6035ae\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-13T06:49:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vzmsr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://f6073140d9f86929b6d4bd2960ede0617a5c9620b269bbab043bae37e0d389f5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-13T06:49:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vzmsr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://693a2ab434c250cb140a866e8068b156df9191414fed1d63bee919a2528ea6e5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-13T06:49:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vzmsr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://219620f51474e60b3e1f880c0972eb612dffc4bad24462bbba2c093ae817d852\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-13T06:49:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vzmsr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://f137f4b441c2402093ab430a9689a46221c4fb4bc70d99fb992f62b0a989c465\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-13T06:49:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vzmsr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://7dd9913a65acba73ae1f32f4398207c5e276a1e979d86f9e720e4fd761bae4a3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://7dd9913a65acba73ae1f32f4398207c5e276a1e979d86f9e720e4fd761bae4a3\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-12-13T06:50:24Z\\\",\\\"message\\\":\\\"lse, Template:false, AddressFamily:\\\\\\\"\\\\\\\"}, Rules:[]services.LBRule{services.LBRule{Source:services.Addr{IP:\\\\\\\"10.217.4.250\\\\\\\", Port:443, Template:(*services.Template)(nil)}, Targets:[]services.Addr{}}}, Templates:services.TemplateMap(nil), Switches:[]string{}, Routers:[]string{}, Groups:[]string{\\\\\\\"clusterLBGroup\\\\\\\"}}}\\\\nI1213 06:50:24.014680 7045 services_controller.go:452] Built service openshift-machine-api/machine-api-operator-machine-webhook per-node LB for network=default: []services.LB{}\\\\nI1213 06:50:24.014686 7045 services_controller.go:453] Built service openshift-machine-api/machine-api-operator-machine-webhook template LB for network=default: []services.LB{}\\\\nI1213 06:50:24.012650 7045 obj_retry.go:303] Retry object setup: *v1.Pod openshift-kube-scheduler/openshift-kube-scheduler-crc\\\\nI1213 06:50:24.014698 7045 services_controller.go:454] Service openshift-machine-api/machine-api-operator-machine-webhook for network=default has 1 cluster-wide, 0 per-node configs, 0 template configs, making 1 (cluster) 0 (per node) and 0 (template) load balancers\\\\nF1213 06:50:24.014708 7045 ovnkube.go:137] failed to run ovnkube: [failed to start network controller: failed to start default network controller: unable to create admin network policy controller, err: could not add Event Handler for anpInformer during admin network policy controller initial\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-13T06:50:23Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 40s restarting failed container=ovnkube-controller pod=ovnkube-node-zlcg5_openshift-ovn-kubernetes(736c6b1a-93dd-4f71-81b6-1c43fcf5f556)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vzmsr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://868049c6a3a4ebe4036d1d22e02ba6b76bced2ee7e269bb228accc4a0636678a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-13T06:49:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vzmsr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://451b22df6973ddc8fa2209a65d64d9e41b33eb231bb4adfab88cf0e5943bb246\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://451b22df6973ddc8fa2209a65d64d9e41b33eb231bb4adfab88cf0e5943bb246\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-13T06:49:26Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-13T06:49:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vzmsr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-13T06:49:25Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-zlcg5\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-13T06:50:34Z is after 2025-08-24T17:21:41Z" Dec 13 06:50:34 crc kubenswrapper[4971]: I1213 06:50:34.079919 4971 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 13 06:50:34 crc kubenswrapper[4971]: I1213 06:50:34.079954 4971 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 13 06:50:34 crc kubenswrapper[4971]: I1213 06:50:34.079964 4971 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 13 06:50:34 crc kubenswrapper[4971]: I1213 06:50:34.079978 4971 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 13 06:50:34 crc kubenswrapper[4971]: I1213 06:50:34.079986 4971 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-13T06:50:34Z","lastTransitionTime":"2025-12-13T06:50:34Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 13 06:50:34 crc kubenswrapper[4971]: I1213 06:50:34.182123 4971 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 13 06:50:34 crc kubenswrapper[4971]: I1213 06:50:34.182155 4971 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 13 06:50:34 crc kubenswrapper[4971]: I1213 06:50:34.182165 4971 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 13 06:50:34 crc kubenswrapper[4971]: I1213 06:50:34.182180 4971 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 13 06:50:34 crc kubenswrapper[4971]: I1213 06:50:34.182192 4971 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-13T06:50:34Z","lastTransitionTime":"2025-12-13T06:50:34Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 13 06:50:34 crc kubenswrapper[4971]: I1213 06:50:34.284858 4971 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 13 06:50:34 crc kubenswrapper[4971]: I1213 06:50:34.284915 4971 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 13 06:50:34 crc kubenswrapper[4971]: I1213 06:50:34.284937 4971 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 13 06:50:34 crc kubenswrapper[4971]: I1213 06:50:34.284966 4971 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 13 06:50:34 crc kubenswrapper[4971]: I1213 06:50:34.284988 4971 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-13T06:50:34Z","lastTransitionTime":"2025-12-13T06:50:34Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 13 06:50:34 crc kubenswrapper[4971]: I1213 06:50:34.387721 4971 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 13 06:50:34 crc kubenswrapper[4971]: I1213 06:50:34.387769 4971 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 13 06:50:34 crc kubenswrapper[4971]: I1213 06:50:34.387783 4971 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 13 06:50:34 crc kubenswrapper[4971]: I1213 06:50:34.387801 4971 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 13 06:50:34 crc kubenswrapper[4971]: I1213 06:50:34.387813 4971 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-13T06:50:34Z","lastTransitionTime":"2025-12-13T06:50:34Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 13 06:50:34 crc kubenswrapper[4971]: I1213 06:50:34.490637 4971 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 13 06:50:34 crc kubenswrapper[4971]: I1213 06:50:34.490677 4971 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 13 06:50:34 crc kubenswrapper[4971]: I1213 06:50:34.490689 4971 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 13 06:50:34 crc kubenswrapper[4971]: I1213 06:50:34.490706 4971 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 13 06:50:34 crc kubenswrapper[4971]: I1213 06:50:34.490718 4971 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-13T06:50:34Z","lastTransitionTime":"2025-12-13T06:50:34Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 13 06:50:34 crc kubenswrapper[4971]: I1213 06:50:34.593439 4971 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 13 06:50:34 crc kubenswrapper[4971]: I1213 06:50:34.593491 4971 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 13 06:50:34 crc kubenswrapper[4971]: I1213 06:50:34.593499 4971 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 13 06:50:34 crc kubenswrapper[4971]: I1213 06:50:34.593533 4971 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 13 06:50:34 crc kubenswrapper[4971]: I1213 06:50:34.593545 4971 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-13T06:50:34Z","lastTransitionTime":"2025-12-13T06:50:34Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 13 06:50:34 crc kubenswrapper[4971]: I1213 06:50:34.700926 4971 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 13 06:50:34 crc kubenswrapper[4971]: I1213 06:50:34.701442 4971 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 13 06:50:34 crc kubenswrapper[4971]: I1213 06:50:34.701546 4971 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 13 06:50:34 crc kubenswrapper[4971]: I1213 06:50:34.701644 4971 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 13 06:50:34 crc kubenswrapper[4971]: I1213 06:50:34.701733 4971 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-13T06:50:34Z","lastTransitionTime":"2025-12-13T06:50:34Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 13 06:50:34 crc kubenswrapper[4971]: I1213 06:50:34.804381 4971 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 13 06:50:34 crc kubenswrapper[4971]: I1213 06:50:34.804419 4971 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 13 06:50:34 crc kubenswrapper[4971]: I1213 06:50:34.804430 4971 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 13 06:50:34 crc kubenswrapper[4971]: I1213 06:50:34.804446 4971 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 13 06:50:34 crc kubenswrapper[4971]: I1213 06:50:34.804463 4971 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-13T06:50:34Z","lastTransitionTime":"2025-12-13T06:50:34Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 13 06:50:34 crc kubenswrapper[4971]: I1213 06:50:34.906723 4971 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 13 06:50:34 crc kubenswrapper[4971]: I1213 06:50:34.906755 4971 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 13 06:50:34 crc kubenswrapper[4971]: I1213 06:50:34.906766 4971 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 13 06:50:34 crc kubenswrapper[4971]: I1213 06:50:34.906778 4971 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 13 06:50:34 crc kubenswrapper[4971]: I1213 06:50:34.906787 4971 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-13T06:50:34Z","lastTransitionTime":"2025-12-13T06:50:34Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 13 06:50:35 crc kubenswrapper[4971]: I1213 06:50:35.015393 4971 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 13 06:50:35 crc kubenswrapper[4971]: I1213 06:50:35.015443 4971 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 13 06:50:35 crc kubenswrapper[4971]: I1213 06:50:35.015453 4971 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 13 06:50:35 crc kubenswrapper[4971]: I1213 06:50:35.015467 4971 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 13 06:50:35 crc kubenswrapper[4971]: I1213 06:50:35.015477 4971 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-13T06:50:35Z","lastTransitionTime":"2025-12-13T06:50:35Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 13 06:50:35 crc kubenswrapper[4971]: I1213 06:50:35.118535 4971 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 13 06:50:35 crc kubenswrapper[4971]: I1213 06:50:35.118580 4971 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 13 06:50:35 crc kubenswrapper[4971]: I1213 06:50:35.118594 4971 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 13 06:50:35 crc kubenswrapper[4971]: I1213 06:50:35.118613 4971 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 13 06:50:35 crc kubenswrapper[4971]: I1213 06:50:35.118627 4971 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-13T06:50:35Z","lastTransitionTime":"2025-12-13T06:50:35Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 13 06:50:35 crc kubenswrapper[4971]: I1213 06:50:35.220944 4971 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 13 06:50:35 crc kubenswrapper[4971]: I1213 06:50:35.220984 4971 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 13 06:50:35 crc kubenswrapper[4971]: I1213 06:50:35.220998 4971 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 13 06:50:35 crc kubenswrapper[4971]: I1213 06:50:35.221016 4971 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 13 06:50:35 crc kubenswrapper[4971]: I1213 06:50:35.221024 4971 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-13T06:50:35Z","lastTransitionTime":"2025-12-13T06:50:35Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 13 06:50:35 crc kubenswrapper[4971]: I1213 06:50:35.323596 4971 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 13 06:50:35 crc kubenswrapper[4971]: I1213 06:50:35.323641 4971 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 13 06:50:35 crc kubenswrapper[4971]: I1213 06:50:35.323652 4971 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 13 06:50:35 crc kubenswrapper[4971]: I1213 06:50:35.323669 4971 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 13 06:50:35 crc kubenswrapper[4971]: I1213 06:50:35.323683 4971 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-13T06:50:35Z","lastTransitionTime":"2025-12-13T06:50:35Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 13 06:50:35 crc kubenswrapper[4971]: I1213 06:50:35.426660 4971 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 13 06:50:35 crc kubenswrapper[4971]: I1213 06:50:35.426729 4971 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 13 06:50:35 crc kubenswrapper[4971]: I1213 06:50:35.426741 4971 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 13 06:50:35 crc kubenswrapper[4971]: I1213 06:50:35.426759 4971 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 13 06:50:35 crc kubenswrapper[4971]: I1213 06:50:35.426771 4971 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-13T06:50:35Z","lastTransitionTime":"2025-12-13T06:50:35Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 13 06:50:35 crc kubenswrapper[4971]: I1213 06:50:35.529082 4971 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 13 06:50:35 crc kubenswrapper[4971]: I1213 06:50:35.529115 4971 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 13 06:50:35 crc kubenswrapper[4971]: I1213 06:50:35.529124 4971 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 13 06:50:35 crc kubenswrapper[4971]: I1213 06:50:35.529156 4971 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 13 06:50:35 crc kubenswrapper[4971]: I1213 06:50:35.529166 4971 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-13T06:50:35Z","lastTransitionTime":"2025-12-13T06:50:35Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 13 06:50:35 crc kubenswrapper[4971]: I1213 06:50:35.632017 4971 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 13 06:50:35 crc kubenswrapper[4971]: I1213 06:50:35.632051 4971 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 13 06:50:35 crc kubenswrapper[4971]: I1213 06:50:35.632060 4971 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 13 06:50:35 crc kubenswrapper[4971]: I1213 06:50:35.632074 4971 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 13 06:50:35 crc kubenswrapper[4971]: I1213 06:50:35.632084 4971 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-13T06:50:35Z","lastTransitionTime":"2025-12-13T06:50:35Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 13 06:50:35 crc kubenswrapper[4971]: I1213 06:50:35.734598 4971 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 13 06:50:35 crc kubenswrapper[4971]: I1213 06:50:35.734631 4971 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 13 06:50:35 crc kubenswrapper[4971]: I1213 06:50:35.734643 4971 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 13 06:50:35 crc kubenswrapper[4971]: I1213 06:50:35.734657 4971 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 13 06:50:35 crc kubenswrapper[4971]: I1213 06:50:35.734666 4971 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-13T06:50:35Z","lastTransitionTime":"2025-12-13T06:50:35Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 13 06:50:35 crc kubenswrapper[4971]: I1213 06:50:35.768102 4971 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-x98k7" Dec 13 06:50:35 crc kubenswrapper[4971]: I1213 06:50:35.768165 4971 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 13 06:50:35 crc kubenswrapper[4971]: I1213 06:50:35.768130 4971 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 13 06:50:35 crc kubenswrapper[4971]: E1213 06:50:35.768213 4971 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-x98k7" podUID="c7185dec-3240-42e8-9d90-6182a5e8b65c" Dec 13 06:50:35 crc kubenswrapper[4971]: I1213 06:50:35.768103 4971 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 13 06:50:35 crc kubenswrapper[4971]: E1213 06:50:35.768311 4971 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 13 06:50:35 crc kubenswrapper[4971]: E1213 06:50:35.768372 4971 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 13 06:50:35 crc kubenswrapper[4971]: E1213 06:50:35.768783 4971 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 13 06:50:35 crc kubenswrapper[4971]: I1213 06:50:35.838080 4971 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 13 06:50:35 crc kubenswrapper[4971]: I1213 06:50:35.838115 4971 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 13 06:50:35 crc kubenswrapper[4971]: I1213 06:50:35.838126 4971 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 13 06:50:35 crc kubenswrapper[4971]: I1213 06:50:35.838143 4971 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 13 06:50:35 crc kubenswrapper[4971]: I1213 06:50:35.838154 4971 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-13T06:50:35Z","lastTransitionTime":"2025-12-13T06:50:35Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 13 06:50:35 crc kubenswrapper[4971]: I1213 06:50:35.941366 4971 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 13 06:50:35 crc kubenswrapper[4971]: I1213 06:50:35.941423 4971 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 13 06:50:35 crc kubenswrapper[4971]: I1213 06:50:35.941445 4971 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 13 06:50:35 crc kubenswrapper[4971]: I1213 06:50:35.941475 4971 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 13 06:50:35 crc kubenswrapper[4971]: I1213 06:50:35.941496 4971 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-13T06:50:35Z","lastTransitionTime":"2025-12-13T06:50:35Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 13 06:50:36 crc kubenswrapper[4971]: I1213 06:50:36.043341 4971 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 13 06:50:36 crc kubenswrapper[4971]: I1213 06:50:36.043381 4971 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 13 06:50:36 crc kubenswrapper[4971]: I1213 06:50:36.043393 4971 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 13 06:50:36 crc kubenswrapper[4971]: I1213 06:50:36.043413 4971 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 13 06:50:36 crc kubenswrapper[4971]: I1213 06:50:36.043428 4971 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-13T06:50:36Z","lastTransitionTime":"2025-12-13T06:50:36Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 13 06:50:36 crc kubenswrapper[4971]: I1213 06:50:36.145442 4971 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 13 06:50:36 crc kubenswrapper[4971]: I1213 06:50:36.145482 4971 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 13 06:50:36 crc kubenswrapper[4971]: I1213 06:50:36.145509 4971 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 13 06:50:36 crc kubenswrapper[4971]: I1213 06:50:36.145536 4971 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 13 06:50:36 crc kubenswrapper[4971]: I1213 06:50:36.145557 4971 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-13T06:50:36Z","lastTransitionTime":"2025-12-13T06:50:36Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 13 06:50:36 crc kubenswrapper[4971]: I1213 06:50:36.248487 4971 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 13 06:50:36 crc kubenswrapper[4971]: I1213 06:50:36.248534 4971 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 13 06:50:36 crc kubenswrapper[4971]: I1213 06:50:36.248542 4971 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 13 06:50:36 crc kubenswrapper[4971]: I1213 06:50:36.248557 4971 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 13 06:50:36 crc kubenswrapper[4971]: I1213 06:50:36.248567 4971 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-13T06:50:36Z","lastTransitionTime":"2025-12-13T06:50:36Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 13 06:50:36 crc kubenswrapper[4971]: I1213 06:50:36.350601 4971 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 13 06:50:36 crc kubenswrapper[4971]: I1213 06:50:36.350661 4971 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 13 06:50:36 crc kubenswrapper[4971]: I1213 06:50:36.350673 4971 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 13 06:50:36 crc kubenswrapper[4971]: I1213 06:50:36.350691 4971 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 13 06:50:36 crc kubenswrapper[4971]: I1213 06:50:36.350705 4971 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-13T06:50:36Z","lastTransitionTime":"2025-12-13T06:50:36Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 13 06:50:36 crc kubenswrapper[4971]: I1213 06:50:36.453949 4971 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 13 06:50:36 crc kubenswrapper[4971]: I1213 06:50:36.454025 4971 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 13 06:50:36 crc kubenswrapper[4971]: I1213 06:50:36.454045 4971 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 13 06:50:36 crc kubenswrapper[4971]: I1213 06:50:36.454075 4971 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 13 06:50:36 crc kubenswrapper[4971]: I1213 06:50:36.454091 4971 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-13T06:50:36Z","lastTransitionTime":"2025-12-13T06:50:36Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 13 06:50:36 crc kubenswrapper[4971]: I1213 06:50:36.557424 4971 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 13 06:50:36 crc kubenswrapper[4971]: I1213 06:50:36.557487 4971 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 13 06:50:36 crc kubenswrapper[4971]: I1213 06:50:36.557500 4971 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 13 06:50:36 crc kubenswrapper[4971]: I1213 06:50:36.557561 4971 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 13 06:50:36 crc kubenswrapper[4971]: I1213 06:50:36.557576 4971 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-13T06:50:36Z","lastTransitionTime":"2025-12-13T06:50:36Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 13 06:50:36 crc kubenswrapper[4971]: I1213 06:50:36.660368 4971 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 13 06:50:36 crc kubenswrapper[4971]: I1213 06:50:36.660438 4971 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 13 06:50:36 crc kubenswrapper[4971]: I1213 06:50:36.660451 4971 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 13 06:50:36 crc kubenswrapper[4971]: I1213 06:50:36.660479 4971 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 13 06:50:36 crc kubenswrapper[4971]: I1213 06:50:36.660499 4971 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-13T06:50:36Z","lastTransitionTime":"2025-12-13T06:50:36Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 13 06:50:36 crc kubenswrapper[4971]: I1213 06:50:36.764088 4971 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 13 06:50:36 crc kubenswrapper[4971]: I1213 06:50:36.764160 4971 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 13 06:50:36 crc kubenswrapper[4971]: I1213 06:50:36.764173 4971 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 13 06:50:36 crc kubenswrapper[4971]: I1213 06:50:36.764198 4971 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 13 06:50:36 crc kubenswrapper[4971]: I1213 06:50:36.764210 4971 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-13T06:50:36Z","lastTransitionTime":"2025-12-13T06:50:36Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 13 06:50:36 crc kubenswrapper[4971]: I1213 06:50:36.867214 4971 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 13 06:50:36 crc kubenswrapper[4971]: I1213 06:50:36.867279 4971 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 13 06:50:36 crc kubenswrapper[4971]: I1213 06:50:36.867298 4971 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 13 06:50:36 crc kubenswrapper[4971]: I1213 06:50:36.867323 4971 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 13 06:50:36 crc kubenswrapper[4971]: I1213 06:50:36.867339 4971 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-13T06:50:36Z","lastTransitionTime":"2025-12-13T06:50:36Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 13 06:50:36 crc kubenswrapper[4971]: I1213 06:50:36.970751 4971 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 13 06:50:36 crc kubenswrapper[4971]: I1213 06:50:36.970822 4971 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 13 06:50:36 crc kubenswrapper[4971]: I1213 06:50:36.970837 4971 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 13 06:50:36 crc kubenswrapper[4971]: I1213 06:50:36.970862 4971 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 13 06:50:36 crc kubenswrapper[4971]: I1213 06:50:36.970875 4971 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-13T06:50:36Z","lastTransitionTime":"2025-12-13T06:50:36Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 13 06:50:37 crc kubenswrapper[4971]: I1213 06:50:37.073505 4971 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 13 06:50:37 crc kubenswrapper[4971]: I1213 06:50:37.073602 4971 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 13 06:50:37 crc kubenswrapper[4971]: I1213 06:50:37.073621 4971 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 13 06:50:37 crc kubenswrapper[4971]: I1213 06:50:37.073659 4971 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 13 06:50:37 crc kubenswrapper[4971]: I1213 06:50:37.073676 4971 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-13T06:50:37Z","lastTransitionTime":"2025-12-13T06:50:37Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 13 06:50:37 crc kubenswrapper[4971]: I1213 06:50:37.177080 4971 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 13 06:50:37 crc kubenswrapper[4971]: I1213 06:50:37.177152 4971 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 13 06:50:37 crc kubenswrapper[4971]: I1213 06:50:37.177168 4971 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 13 06:50:37 crc kubenswrapper[4971]: I1213 06:50:37.177198 4971 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 13 06:50:37 crc kubenswrapper[4971]: I1213 06:50:37.177215 4971 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-13T06:50:37Z","lastTransitionTime":"2025-12-13T06:50:37Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 13 06:50:37 crc kubenswrapper[4971]: I1213 06:50:37.280693 4971 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 13 06:50:37 crc kubenswrapper[4971]: I1213 06:50:37.280736 4971 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 13 06:50:37 crc kubenswrapper[4971]: I1213 06:50:37.280745 4971 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 13 06:50:37 crc kubenswrapper[4971]: I1213 06:50:37.280763 4971 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 13 06:50:37 crc kubenswrapper[4971]: I1213 06:50:37.280777 4971 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-13T06:50:37Z","lastTransitionTime":"2025-12-13T06:50:37Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 13 06:50:37 crc kubenswrapper[4971]: I1213 06:50:37.383465 4971 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 13 06:50:37 crc kubenswrapper[4971]: I1213 06:50:37.383578 4971 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 13 06:50:37 crc kubenswrapper[4971]: I1213 06:50:37.383594 4971 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 13 06:50:37 crc kubenswrapper[4971]: I1213 06:50:37.383615 4971 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 13 06:50:37 crc kubenswrapper[4971]: I1213 06:50:37.383628 4971 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-13T06:50:37Z","lastTransitionTime":"2025-12-13T06:50:37Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 13 06:50:37 crc kubenswrapper[4971]: I1213 06:50:37.486465 4971 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 13 06:50:37 crc kubenswrapper[4971]: I1213 06:50:37.486505 4971 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 13 06:50:37 crc kubenswrapper[4971]: I1213 06:50:37.486550 4971 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 13 06:50:37 crc kubenswrapper[4971]: I1213 06:50:37.486568 4971 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 13 06:50:37 crc kubenswrapper[4971]: I1213 06:50:37.486579 4971 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-13T06:50:37Z","lastTransitionTime":"2025-12-13T06:50:37Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 13 06:50:37 crc kubenswrapper[4971]: I1213 06:50:37.590437 4971 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 13 06:50:37 crc kubenswrapper[4971]: I1213 06:50:37.590483 4971 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 13 06:50:37 crc kubenswrapper[4971]: I1213 06:50:37.590505 4971 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 13 06:50:37 crc kubenswrapper[4971]: I1213 06:50:37.590558 4971 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 13 06:50:37 crc kubenswrapper[4971]: I1213 06:50:37.590571 4971 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-13T06:50:37Z","lastTransitionTime":"2025-12-13T06:50:37Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 13 06:50:37 crc kubenswrapper[4971]: I1213 06:50:37.693089 4971 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 13 06:50:37 crc kubenswrapper[4971]: I1213 06:50:37.693156 4971 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 13 06:50:37 crc kubenswrapper[4971]: I1213 06:50:37.693171 4971 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 13 06:50:37 crc kubenswrapper[4971]: I1213 06:50:37.693187 4971 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 13 06:50:37 crc kubenswrapper[4971]: I1213 06:50:37.693196 4971 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-13T06:50:37Z","lastTransitionTime":"2025-12-13T06:50:37Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 13 06:50:37 crc kubenswrapper[4971]: I1213 06:50:37.767598 4971 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-x98k7" Dec 13 06:50:37 crc kubenswrapper[4971]: I1213 06:50:37.767925 4971 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 13 06:50:37 crc kubenswrapper[4971]: I1213 06:50:37.767983 4971 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 13 06:50:37 crc kubenswrapper[4971]: E1213 06:50:37.768038 4971 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-x98k7" podUID="c7185dec-3240-42e8-9d90-6182a5e8b65c" Dec 13 06:50:37 crc kubenswrapper[4971]: I1213 06:50:37.768094 4971 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 13 06:50:37 crc kubenswrapper[4971]: E1213 06:50:37.768162 4971 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 13 06:50:37 crc kubenswrapper[4971]: E1213 06:50:37.768227 4971 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 13 06:50:37 crc kubenswrapper[4971]: E1213 06:50:37.768291 4971 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 13 06:50:37 crc kubenswrapper[4971]: I1213 06:50:37.768368 4971 scope.go:117] "RemoveContainer" containerID="7dd9913a65acba73ae1f32f4398207c5e276a1e979d86f9e720e4fd761bae4a3" Dec 13 06:50:37 crc kubenswrapper[4971]: E1213 06:50:37.768595 4971 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"ovnkube-controller\" with CrashLoopBackOff: \"back-off 40s restarting failed container=ovnkube-controller pod=ovnkube-node-zlcg5_openshift-ovn-kubernetes(736c6b1a-93dd-4f71-81b6-1c43fcf5f556)\"" pod="openshift-ovn-kubernetes/ovnkube-node-zlcg5" podUID="736c6b1a-93dd-4f71-81b6-1c43fcf5f556" Dec 13 06:50:37 crc kubenswrapper[4971]: I1213 06:50:37.795892 4971 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 13 06:50:37 crc kubenswrapper[4971]: I1213 06:50:37.795932 4971 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 13 06:50:37 crc kubenswrapper[4971]: I1213 06:50:37.795946 4971 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 13 06:50:37 crc kubenswrapper[4971]: I1213 06:50:37.795964 4971 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 13 06:50:37 crc kubenswrapper[4971]: I1213 06:50:37.795978 4971 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-13T06:50:37Z","lastTransitionTime":"2025-12-13T06:50:37Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 13 06:50:37 crc kubenswrapper[4971]: I1213 06:50:37.898021 4971 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 13 06:50:37 crc kubenswrapper[4971]: I1213 06:50:37.898079 4971 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 13 06:50:37 crc kubenswrapper[4971]: I1213 06:50:37.898089 4971 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 13 06:50:37 crc kubenswrapper[4971]: I1213 06:50:37.898106 4971 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 13 06:50:37 crc kubenswrapper[4971]: I1213 06:50:37.898115 4971 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-13T06:50:37Z","lastTransitionTime":"2025-12-13T06:50:37Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 13 06:50:38 crc kubenswrapper[4971]: I1213 06:50:38.001707 4971 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 13 06:50:38 crc kubenswrapper[4971]: I1213 06:50:38.001755 4971 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 13 06:50:38 crc kubenswrapper[4971]: I1213 06:50:38.001767 4971 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 13 06:50:38 crc kubenswrapper[4971]: I1213 06:50:38.001787 4971 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 13 06:50:38 crc kubenswrapper[4971]: I1213 06:50:38.001800 4971 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-13T06:50:38Z","lastTransitionTime":"2025-12-13T06:50:38Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 13 06:50:38 crc kubenswrapper[4971]: I1213 06:50:38.105657 4971 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 13 06:50:38 crc kubenswrapper[4971]: I1213 06:50:38.105735 4971 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 13 06:50:38 crc kubenswrapper[4971]: I1213 06:50:38.105756 4971 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 13 06:50:38 crc kubenswrapper[4971]: I1213 06:50:38.105788 4971 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 13 06:50:38 crc kubenswrapper[4971]: I1213 06:50:38.105804 4971 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-13T06:50:38Z","lastTransitionTime":"2025-12-13T06:50:38Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 13 06:50:38 crc kubenswrapper[4971]: I1213 06:50:38.208447 4971 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 13 06:50:38 crc kubenswrapper[4971]: I1213 06:50:38.208499 4971 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 13 06:50:38 crc kubenswrapper[4971]: I1213 06:50:38.208511 4971 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 13 06:50:38 crc kubenswrapper[4971]: I1213 06:50:38.208548 4971 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 13 06:50:38 crc kubenswrapper[4971]: I1213 06:50:38.208561 4971 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-13T06:50:38Z","lastTransitionTime":"2025-12-13T06:50:38Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 13 06:50:38 crc kubenswrapper[4971]: I1213 06:50:38.311629 4971 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 13 06:50:38 crc kubenswrapper[4971]: I1213 06:50:38.311693 4971 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 13 06:50:38 crc kubenswrapper[4971]: I1213 06:50:38.311705 4971 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 13 06:50:38 crc kubenswrapper[4971]: I1213 06:50:38.311729 4971 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 13 06:50:38 crc kubenswrapper[4971]: I1213 06:50:38.311743 4971 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-13T06:50:38Z","lastTransitionTime":"2025-12-13T06:50:38Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 13 06:50:38 crc kubenswrapper[4971]: I1213 06:50:38.415608 4971 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 13 06:50:38 crc kubenswrapper[4971]: I1213 06:50:38.415655 4971 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 13 06:50:38 crc kubenswrapper[4971]: I1213 06:50:38.415668 4971 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 13 06:50:38 crc kubenswrapper[4971]: I1213 06:50:38.415687 4971 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 13 06:50:38 crc kubenswrapper[4971]: I1213 06:50:38.415701 4971 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-13T06:50:38Z","lastTransitionTime":"2025-12-13T06:50:38Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 13 06:50:38 crc kubenswrapper[4971]: I1213 06:50:38.519394 4971 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 13 06:50:38 crc kubenswrapper[4971]: I1213 06:50:38.519439 4971 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 13 06:50:38 crc kubenswrapper[4971]: I1213 06:50:38.519452 4971 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 13 06:50:38 crc kubenswrapper[4971]: I1213 06:50:38.519470 4971 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 13 06:50:38 crc kubenswrapper[4971]: I1213 06:50:38.519482 4971 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-13T06:50:38Z","lastTransitionTime":"2025-12-13T06:50:38Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 13 06:50:38 crc kubenswrapper[4971]: I1213 06:50:38.622114 4971 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 13 06:50:38 crc kubenswrapper[4971]: I1213 06:50:38.622177 4971 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 13 06:50:38 crc kubenswrapper[4971]: I1213 06:50:38.622193 4971 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 13 06:50:38 crc kubenswrapper[4971]: I1213 06:50:38.622217 4971 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 13 06:50:38 crc kubenswrapper[4971]: I1213 06:50:38.622236 4971 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-13T06:50:38Z","lastTransitionTime":"2025-12-13T06:50:38Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 13 06:50:38 crc kubenswrapper[4971]: I1213 06:50:38.724263 4971 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 13 06:50:38 crc kubenswrapper[4971]: I1213 06:50:38.724307 4971 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 13 06:50:38 crc kubenswrapper[4971]: I1213 06:50:38.724318 4971 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 13 06:50:38 crc kubenswrapper[4971]: I1213 06:50:38.724335 4971 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 13 06:50:38 crc kubenswrapper[4971]: I1213 06:50:38.724347 4971 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-13T06:50:38Z","lastTransitionTime":"2025-12-13T06:50:38Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 13 06:50:38 crc kubenswrapper[4971]: I1213 06:50:38.826896 4971 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 13 06:50:38 crc kubenswrapper[4971]: I1213 06:50:38.826968 4971 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 13 06:50:38 crc kubenswrapper[4971]: I1213 06:50:38.826979 4971 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 13 06:50:38 crc kubenswrapper[4971]: I1213 06:50:38.826999 4971 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 13 06:50:38 crc kubenswrapper[4971]: I1213 06:50:38.827012 4971 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-13T06:50:38Z","lastTransitionTime":"2025-12-13T06:50:38Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 13 06:50:38 crc kubenswrapper[4971]: I1213 06:50:38.929339 4971 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 13 06:50:38 crc kubenswrapper[4971]: I1213 06:50:38.929414 4971 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 13 06:50:38 crc kubenswrapper[4971]: I1213 06:50:38.929424 4971 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 13 06:50:38 crc kubenswrapper[4971]: I1213 06:50:38.929439 4971 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 13 06:50:38 crc kubenswrapper[4971]: I1213 06:50:38.929448 4971 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-13T06:50:38Z","lastTransitionTime":"2025-12-13T06:50:38Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 13 06:50:39 crc kubenswrapper[4971]: I1213 06:50:39.031964 4971 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 13 06:50:39 crc kubenswrapper[4971]: I1213 06:50:39.032288 4971 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 13 06:50:39 crc kubenswrapper[4971]: I1213 06:50:39.032298 4971 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 13 06:50:39 crc kubenswrapper[4971]: I1213 06:50:39.032312 4971 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 13 06:50:39 crc kubenswrapper[4971]: I1213 06:50:39.032321 4971 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-13T06:50:39Z","lastTransitionTime":"2025-12-13T06:50:39Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 13 06:50:39 crc kubenswrapper[4971]: I1213 06:50:39.135323 4971 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 13 06:50:39 crc kubenswrapper[4971]: I1213 06:50:39.135411 4971 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 13 06:50:39 crc kubenswrapper[4971]: I1213 06:50:39.135430 4971 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 13 06:50:39 crc kubenswrapper[4971]: I1213 06:50:39.135455 4971 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 13 06:50:39 crc kubenswrapper[4971]: I1213 06:50:39.135470 4971 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-13T06:50:39Z","lastTransitionTime":"2025-12-13T06:50:39Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 13 06:50:39 crc kubenswrapper[4971]: I1213 06:50:39.239441 4971 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 13 06:50:39 crc kubenswrapper[4971]: I1213 06:50:39.239493 4971 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 13 06:50:39 crc kubenswrapper[4971]: I1213 06:50:39.239509 4971 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 13 06:50:39 crc kubenswrapper[4971]: I1213 06:50:39.239555 4971 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 13 06:50:39 crc kubenswrapper[4971]: I1213 06:50:39.239568 4971 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-13T06:50:39Z","lastTransitionTime":"2025-12-13T06:50:39Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 13 06:50:39 crc kubenswrapper[4971]: I1213 06:50:39.342231 4971 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 13 06:50:39 crc kubenswrapper[4971]: I1213 06:50:39.342309 4971 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 13 06:50:39 crc kubenswrapper[4971]: I1213 06:50:39.342323 4971 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 13 06:50:39 crc kubenswrapper[4971]: I1213 06:50:39.342345 4971 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 13 06:50:39 crc kubenswrapper[4971]: I1213 06:50:39.342365 4971 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-13T06:50:39Z","lastTransitionTime":"2025-12-13T06:50:39Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 13 06:50:39 crc kubenswrapper[4971]: I1213 06:50:39.444062 4971 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 13 06:50:39 crc kubenswrapper[4971]: I1213 06:50:39.444105 4971 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 13 06:50:39 crc kubenswrapper[4971]: I1213 06:50:39.444119 4971 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 13 06:50:39 crc kubenswrapper[4971]: I1213 06:50:39.444137 4971 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 13 06:50:39 crc kubenswrapper[4971]: I1213 06:50:39.444148 4971 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-13T06:50:39Z","lastTransitionTime":"2025-12-13T06:50:39Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 13 06:50:39 crc kubenswrapper[4971]: I1213 06:50:39.546885 4971 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 13 06:50:39 crc kubenswrapper[4971]: I1213 06:50:39.546932 4971 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 13 06:50:39 crc kubenswrapper[4971]: I1213 06:50:39.546945 4971 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 13 06:50:39 crc kubenswrapper[4971]: I1213 06:50:39.546961 4971 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 13 06:50:39 crc kubenswrapper[4971]: I1213 06:50:39.546973 4971 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-13T06:50:39Z","lastTransitionTime":"2025-12-13T06:50:39Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 13 06:50:39 crc kubenswrapper[4971]: I1213 06:50:39.649838 4971 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 13 06:50:39 crc kubenswrapper[4971]: I1213 06:50:39.649882 4971 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 13 06:50:39 crc kubenswrapper[4971]: I1213 06:50:39.649894 4971 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 13 06:50:39 crc kubenswrapper[4971]: I1213 06:50:39.649909 4971 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 13 06:50:39 crc kubenswrapper[4971]: I1213 06:50:39.649920 4971 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-13T06:50:39Z","lastTransitionTime":"2025-12-13T06:50:39Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 13 06:50:39 crc kubenswrapper[4971]: I1213 06:50:39.752645 4971 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 13 06:50:39 crc kubenswrapper[4971]: I1213 06:50:39.752719 4971 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 13 06:50:39 crc kubenswrapper[4971]: I1213 06:50:39.752742 4971 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 13 06:50:39 crc kubenswrapper[4971]: I1213 06:50:39.752770 4971 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 13 06:50:39 crc kubenswrapper[4971]: I1213 06:50:39.752792 4971 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-13T06:50:39Z","lastTransitionTime":"2025-12-13T06:50:39Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 13 06:50:39 crc kubenswrapper[4971]: I1213 06:50:39.768218 4971 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 13 06:50:39 crc kubenswrapper[4971]: I1213 06:50:39.768280 4971 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 13 06:50:39 crc kubenswrapper[4971]: I1213 06:50:39.768434 4971 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 13 06:50:39 crc kubenswrapper[4971]: E1213 06:50:39.768427 4971 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 13 06:50:39 crc kubenswrapper[4971]: E1213 06:50:39.768717 4971 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 13 06:50:39 crc kubenswrapper[4971]: I1213 06:50:39.769298 4971 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-x98k7" Dec 13 06:50:39 crc kubenswrapper[4971]: E1213 06:50:39.769415 4971 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-x98k7" podUID="c7185dec-3240-42e8-9d90-6182a5e8b65c" Dec 13 06:50:39 crc kubenswrapper[4971]: E1213 06:50:39.768971 4971 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 13 06:50:39 crc kubenswrapper[4971]: I1213 06:50:39.856450 4971 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 13 06:50:39 crc kubenswrapper[4971]: I1213 06:50:39.856490 4971 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 13 06:50:39 crc kubenswrapper[4971]: I1213 06:50:39.856501 4971 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 13 06:50:39 crc kubenswrapper[4971]: I1213 06:50:39.856529 4971 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 13 06:50:39 crc kubenswrapper[4971]: I1213 06:50:39.856540 4971 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-13T06:50:39Z","lastTransitionTime":"2025-12-13T06:50:39Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 13 06:50:39 crc kubenswrapper[4971]: I1213 06:50:39.959155 4971 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 13 06:50:39 crc kubenswrapper[4971]: I1213 06:50:39.959208 4971 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 13 06:50:39 crc kubenswrapper[4971]: I1213 06:50:39.959220 4971 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 13 06:50:39 crc kubenswrapper[4971]: I1213 06:50:39.959237 4971 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 13 06:50:39 crc kubenswrapper[4971]: I1213 06:50:39.959249 4971 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-13T06:50:39Z","lastTransitionTime":"2025-12-13T06:50:39Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 13 06:50:40 crc kubenswrapper[4971]: I1213 06:50:40.063579 4971 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 13 06:50:40 crc kubenswrapper[4971]: I1213 06:50:40.063636 4971 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 13 06:50:40 crc kubenswrapper[4971]: I1213 06:50:40.063654 4971 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 13 06:50:40 crc kubenswrapper[4971]: I1213 06:50:40.063676 4971 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 13 06:50:40 crc kubenswrapper[4971]: I1213 06:50:40.063691 4971 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-13T06:50:40Z","lastTransitionTime":"2025-12-13T06:50:40Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 13 06:50:40 crc kubenswrapper[4971]: I1213 06:50:40.166813 4971 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 13 06:50:40 crc kubenswrapper[4971]: I1213 06:50:40.166875 4971 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 13 06:50:40 crc kubenswrapper[4971]: I1213 06:50:40.166892 4971 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 13 06:50:40 crc kubenswrapper[4971]: I1213 06:50:40.166917 4971 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 13 06:50:40 crc kubenswrapper[4971]: I1213 06:50:40.166935 4971 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-13T06:50:40Z","lastTransitionTime":"2025-12-13T06:50:40Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 13 06:50:40 crc kubenswrapper[4971]: I1213 06:50:40.268688 4971 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 13 06:50:40 crc kubenswrapper[4971]: I1213 06:50:40.268742 4971 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 13 06:50:40 crc kubenswrapper[4971]: I1213 06:50:40.268789 4971 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 13 06:50:40 crc kubenswrapper[4971]: I1213 06:50:40.268808 4971 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 13 06:50:40 crc kubenswrapper[4971]: I1213 06:50:40.268817 4971 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-13T06:50:40Z","lastTransitionTime":"2025-12-13T06:50:40Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 13 06:50:40 crc kubenswrapper[4971]: I1213 06:50:40.372179 4971 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 13 06:50:40 crc kubenswrapper[4971]: I1213 06:50:40.372237 4971 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 13 06:50:40 crc kubenswrapper[4971]: I1213 06:50:40.372250 4971 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 13 06:50:40 crc kubenswrapper[4971]: I1213 06:50:40.372270 4971 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 13 06:50:40 crc kubenswrapper[4971]: I1213 06:50:40.372283 4971 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-13T06:50:40Z","lastTransitionTime":"2025-12-13T06:50:40Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 13 06:50:40 crc kubenswrapper[4971]: I1213 06:50:40.475949 4971 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 13 06:50:40 crc kubenswrapper[4971]: I1213 06:50:40.476004 4971 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 13 06:50:40 crc kubenswrapper[4971]: I1213 06:50:40.476017 4971 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 13 06:50:40 crc kubenswrapper[4971]: I1213 06:50:40.476046 4971 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 13 06:50:40 crc kubenswrapper[4971]: I1213 06:50:40.476065 4971 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-13T06:50:40Z","lastTransitionTime":"2025-12-13T06:50:40Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 13 06:50:40 crc kubenswrapper[4971]: I1213 06:50:40.578980 4971 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 13 06:50:40 crc kubenswrapper[4971]: I1213 06:50:40.579047 4971 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 13 06:50:40 crc kubenswrapper[4971]: I1213 06:50:40.579057 4971 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 13 06:50:40 crc kubenswrapper[4971]: I1213 06:50:40.579077 4971 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 13 06:50:40 crc kubenswrapper[4971]: I1213 06:50:40.579088 4971 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-13T06:50:40Z","lastTransitionTime":"2025-12-13T06:50:40Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 13 06:50:40 crc kubenswrapper[4971]: I1213 06:50:40.682738 4971 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 13 06:50:40 crc kubenswrapper[4971]: I1213 06:50:40.682797 4971 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 13 06:50:40 crc kubenswrapper[4971]: I1213 06:50:40.682810 4971 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 13 06:50:40 crc kubenswrapper[4971]: I1213 06:50:40.682831 4971 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 13 06:50:40 crc kubenswrapper[4971]: I1213 06:50:40.682845 4971 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-13T06:50:40Z","lastTransitionTime":"2025-12-13T06:50:40Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 13 06:50:40 crc kubenswrapper[4971]: I1213 06:50:40.785878 4971 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 13 06:50:40 crc kubenswrapper[4971]: I1213 06:50:40.785921 4971 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 13 06:50:40 crc kubenswrapper[4971]: I1213 06:50:40.785933 4971 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 13 06:50:40 crc kubenswrapper[4971]: I1213 06:50:40.785950 4971 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 13 06:50:40 crc kubenswrapper[4971]: I1213 06:50:40.785963 4971 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-13T06:50:40Z","lastTransitionTime":"2025-12-13T06:50:40Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 13 06:50:40 crc kubenswrapper[4971]: I1213 06:50:40.889020 4971 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 13 06:50:40 crc kubenswrapper[4971]: I1213 06:50:40.889065 4971 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 13 06:50:40 crc kubenswrapper[4971]: I1213 06:50:40.889073 4971 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 13 06:50:40 crc kubenswrapper[4971]: I1213 06:50:40.889089 4971 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 13 06:50:40 crc kubenswrapper[4971]: I1213 06:50:40.889099 4971 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-13T06:50:40Z","lastTransitionTime":"2025-12-13T06:50:40Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 13 06:50:40 crc kubenswrapper[4971]: I1213 06:50:40.990956 4971 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 13 06:50:40 crc kubenswrapper[4971]: I1213 06:50:40.991000 4971 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 13 06:50:40 crc kubenswrapper[4971]: I1213 06:50:40.991011 4971 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 13 06:50:40 crc kubenswrapper[4971]: I1213 06:50:40.991028 4971 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 13 06:50:40 crc kubenswrapper[4971]: I1213 06:50:40.991036 4971 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-13T06:50:40Z","lastTransitionTime":"2025-12-13T06:50:40Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 13 06:50:41 crc kubenswrapper[4971]: I1213 06:50:41.093398 4971 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 13 06:50:41 crc kubenswrapper[4971]: I1213 06:50:41.093431 4971 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 13 06:50:41 crc kubenswrapper[4971]: I1213 06:50:41.093440 4971 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 13 06:50:41 crc kubenswrapper[4971]: I1213 06:50:41.093455 4971 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 13 06:50:41 crc kubenswrapper[4971]: I1213 06:50:41.093464 4971 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-13T06:50:41Z","lastTransitionTime":"2025-12-13T06:50:41Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 13 06:50:41 crc kubenswrapper[4971]: I1213 06:50:41.195740 4971 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 13 06:50:41 crc kubenswrapper[4971]: I1213 06:50:41.195773 4971 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 13 06:50:41 crc kubenswrapper[4971]: I1213 06:50:41.195782 4971 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 13 06:50:41 crc kubenswrapper[4971]: I1213 06:50:41.195796 4971 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 13 06:50:41 crc kubenswrapper[4971]: I1213 06:50:41.195805 4971 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-13T06:50:41Z","lastTransitionTime":"2025-12-13T06:50:41Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 13 06:50:41 crc kubenswrapper[4971]: I1213 06:50:41.298307 4971 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 13 06:50:41 crc kubenswrapper[4971]: I1213 06:50:41.298358 4971 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 13 06:50:41 crc kubenswrapper[4971]: I1213 06:50:41.298374 4971 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 13 06:50:41 crc kubenswrapper[4971]: I1213 06:50:41.298403 4971 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 13 06:50:41 crc kubenswrapper[4971]: I1213 06:50:41.298418 4971 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-13T06:50:41Z","lastTransitionTime":"2025-12-13T06:50:41Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 13 06:50:41 crc kubenswrapper[4971]: I1213 06:50:41.400850 4971 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 13 06:50:41 crc kubenswrapper[4971]: I1213 06:50:41.400951 4971 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 13 06:50:41 crc kubenswrapper[4971]: I1213 06:50:41.400966 4971 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 13 06:50:41 crc kubenswrapper[4971]: I1213 06:50:41.400990 4971 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 13 06:50:41 crc kubenswrapper[4971]: I1213 06:50:41.401006 4971 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-13T06:50:41Z","lastTransitionTime":"2025-12-13T06:50:41Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 13 06:50:41 crc kubenswrapper[4971]: I1213 06:50:41.502690 4971 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 13 06:50:41 crc kubenswrapper[4971]: I1213 06:50:41.502769 4971 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 13 06:50:41 crc kubenswrapper[4971]: I1213 06:50:41.502778 4971 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 13 06:50:41 crc kubenswrapper[4971]: I1213 06:50:41.502795 4971 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 13 06:50:41 crc kubenswrapper[4971]: I1213 06:50:41.502804 4971 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-13T06:50:41Z","lastTransitionTime":"2025-12-13T06:50:41Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 13 06:50:41 crc kubenswrapper[4971]: I1213 06:50:41.604887 4971 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 13 06:50:41 crc kubenswrapper[4971]: I1213 06:50:41.604967 4971 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 13 06:50:41 crc kubenswrapper[4971]: I1213 06:50:41.604976 4971 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 13 06:50:41 crc kubenswrapper[4971]: I1213 06:50:41.604994 4971 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 13 06:50:41 crc kubenswrapper[4971]: I1213 06:50:41.605007 4971 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-13T06:50:41Z","lastTransitionTime":"2025-12-13T06:50:41Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 13 06:50:41 crc kubenswrapper[4971]: I1213 06:50:41.707492 4971 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 13 06:50:41 crc kubenswrapper[4971]: I1213 06:50:41.707569 4971 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 13 06:50:41 crc kubenswrapper[4971]: I1213 06:50:41.707581 4971 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 13 06:50:41 crc kubenswrapper[4971]: I1213 06:50:41.707597 4971 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 13 06:50:41 crc kubenswrapper[4971]: I1213 06:50:41.707609 4971 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-13T06:50:41Z","lastTransitionTime":"2025-12-13T06:50:41Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 13 06:50:41 crc kubenswrapper[4971]: I1213 06:50:41.768274 4971 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-x98k7" Dec 13 06:50:41 crc kubenswrapper[4971]: E1213 06:50:41.768417 4971 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-x98k7" podUID="c7185dec-3240-42e8-9d90-6182a5e8b65c" Dec 13 06:50:41 crc kubenswrapper[4971]: I1213 06:50:41.768437 4971 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 13 06:50:41 crc kubenswrapper[4971]: I1213 06:50:41.768547 4971 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 13 06:50:41 crc kubenswrapper[4971]: E1213 06:50:41.768611 4971 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 13 06:50:41 crc kubenswrapper[4971]: I1213 06:50:41.768620 4971 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 13 06:50:41 crc kubenswrapper[4971]: E1213 06:50:41.768713 4971 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 13 06:50:41 crc kubenswrapper[4971]: E1213 06:50:41.768818 4971 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 13 06:50:41 crc kubenswrapper[4971]: I1213 06:50:41.810551 4971 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 13 06:50:41 crc kubenswrapper[4971]: I1213 06:50:41.810606 4971 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 13 06:50:41 crc kubenswrapper[4971]: I1213 06:50:41.810629 4971 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 13 06:50:41 crc kubenswrapper[4971]: I1213 06:50:41.810651 4971 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 13 06:50:41 crc kubenswrapper[4971]: I1213 06:50:41.810666 4971 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-13T06:50:41Z","lastTransitionTime":"2025-12-13T06:50:41Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 13 06:50:41 crc kubenswrapper[4971]: I1213 06:50:41.913830 4971 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 13 06:50:41 crc kubenswrapper[4971]: I1213 06:50:41.913884 4971 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 13 06:50:41 crc kubenswrapper[4971]: I1213 06:50:41.913895 4971 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 13 06:50:41 crc kubenswrapper[4971]: I1213 06:50:41.913913 4971 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 13 06:50:41 crc kubenswrapper[4971]: I1213 06:50:41.913926 4971 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-13T06:50:41Z","lastTransitionTime":"2025-12-13T06:50:41Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 13 06:50:42 crc kubenswrapper[4971]: I1213 06:50:42.017276 4971 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 13 06:50:42 crc kubenswrapper[4971]: I1213 06:50:42.017338 4971 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 13 06:50:42 crc kubenswrapper[4971]: I1213 06:50:42.017351 4971 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 13 06:50:42 crc kubenswrapper[4971]: I1213 06:50:42.017371 4971 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 13 06:50:42 crc kubenswrapper[4971]: I1213 06:50:42.017386 4971 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-13T06:50:42Z","lastTransitionTime":"2025-12-13T06:50:42Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 13 06:50:42 crc kubenswrapper[4971]: I1213 06:50:42.120153 4971 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 13 06:50:42 crc kubenswrapper[4971]: I1213 06:50:42.120185 4971 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 13 06:50:42 crc kubenswrapper[4971]: I1213 06:50:42.120194 4971 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 13 06:50:42 crc kubenswrapper[4971]: I1213 06:50:42.120207 4971 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 13 06:50:42 crc kubenswrapper[4971]: I1213 06:50:42.120215 4971 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-13T06:50:42Z","lastTransitionTime":"2025-12-13T06:50:42Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 13 06:50:42 crc kubenswrapper[4971]: I1213 06:50:42.223054 4971 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 13 06:50:42 crc kubenswrapper[4971]: I1213 06:50:42.223135 4971 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 13 06:50:42 crc kubenswrapper[4971]: I1213 06:50:42.223153 4971 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 13 06:50:42 crc kubenswrapper[4971]: I1213 06:50:42.223168 4971 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 13 06:50:42 crc kubenswrapper[4971]: I1213 06:50:42.223180 4971 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-13T06:50:42Z","lastTransitionTime":"2025-12-13T06:50:42Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 13 06:50:42 crc kubenswrapper[4971]: I1213 06:50:42.325371 4971 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 13 06:50:42 crc kubenswrapper[4971]: I1213 06:50:42.325490 4971 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 13 06:50:42 crc kubenswrapper[4971]: I1213 06:50:42.325506 4971 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 13 06:50:42 crc kubenswrapper[4971]: I1213 06:50:42.325551 4971 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 13 06:50:42 crc kubenswrapper[4971]: I1213 06:50:42.325564 4971 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-13T06:50:42Z","lastTransitionTime":"2025-12-13T06:50:42Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 13 06:50:42 crc kubenswrapper[4971]: I1213 06:50:42.326834 4971 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 13 06:50:42 crc kubenswrapper[4971]: I1213 06:50:42.326870 4971 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 13 06:50:42 crc kubenswrapper[4971]: I1213 06:50:42.326880 4971 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 13 06:50:42 crc kubenswrapper[4971]: I1213 06:50:42.326898 4971 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 13 06:50:42 crc kubenswrapper[4971]: I1213 06:50:42.326916 4971 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-13T06:50:42Z","lastTransitionTime":"2025-12-13T06:50:42Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 13 06:50:42 crc kubenswrapper[4971]: E1213 06:50:42.344625 4971 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-12-13T06:50:42Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-13T06:50:42Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-13T06:50:42Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-13T06:50:42Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-13T06:50:42Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-13T06:50:42Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-13T06:50:42Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-13T06:50:42Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"bd36cbc9-bbe9-45a1-a21a-ad2bd10499ee\\\",\\\"systemUUID\\\":\\\"a9af00dc-cf25-47af-9997-79577f8f3bea\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-13T06:50:42Z is after 2025-08-24T17:21:41Z" Dec 13 06:50:42 crc kubenswrapper[4971]: I1213 06:50:42.348363 4971 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 13 06:50:42 crc kubenswrapper[4971]: I1213 06:50:42.348389 4971 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 13 06:50:42 crc kubenswrapper[4971]: I1213 06:50:42.348399 4971 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 13 06:50:42 crc kubenswrapper[4971]: I1213 06:50:42.348415 4971 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 13 06:50:42 crc kubenswrapper[4971]: I1213 06:50:42.348426 4971 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-13T06:50:42Z","lastTransitionTime":"2025-12-13T06:50:42Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 13 06:50:42 crc kubenswrapper[4971]: E1213 06:50:42.361329 4971 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-12-13T06:50:42Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-13T06:50:42Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-13T06:50:42Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-13T06:50:42Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-13T06:50:42Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-13T06:50:42Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-13T06:50:42Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-13T06:50:42Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"bd36cbc9-bbe9-45a1-a21a-ad2bd10499ee\\\",\\\"systemUUID\\\":\\\"a9af00dc-cf25-47af-9997-79577f8f3bea\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-13T06:50:42Z is after 2025-08-24T17:21:41Z" Dec 13 06:50:42 crc kubenswrapper[4971]: I1213 06:50:42.364487 4971 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 13 06:50:42 crc kubenswrapper[4971]: I1213 06:50:42.364549 4971 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 13 06:50:42 crc kubenswrapper[4971]: I1213 06:50:42.364559 4971 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 13 06:50:42 crc kubenswrapper[4971]: I1213 06:50:42.364593 4971 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 13 06:50:42 crc kubenswrapper[4971]: I1213 06:50:42.364605 4971 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-13T06:50:42Z","lastTransitionTime":"2025-12-13T06:50:42Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 13 06:50:42 crc kubenswrapper[4971]: E1213 06:50:42.375878 4971 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-12-13T06:50:42Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-13T06:50:42Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-13T06:50:42Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-13T06:50:42Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-13T06:50:42Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-13T06:50:42Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-13T06:50:42Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-13T06:50:42Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"bd36cbc9-bbe9-45a1-a21a-ad2bd10499ee\\\",\\\"systemUUID\\\":\\\"a9af00dc-cf25-47af-9997-79577f8f3bea\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-13T06:50:42Z is after 2025-08-24T17:21:41Z" Dec 13 06:50:42 crc kubenswrapper[4971]: I1213 06:50:42.378865 4971 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 13 06:50:42 crc kubenswrapper[4971]: I1213 06:50:42.378928 4971 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 13 06:50:42 crc kubenswrapper[4971]: I1213 06:50:42.378942 4971 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 13 06:50:42 crc kubenswrapper[4971]: I1213 06:50:42.378959 4971 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 13 06:50:42 crc kubenswrapper[4971]: I1213 06:50:42.378971 4971 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-13T06:50:42Z","lastTransitionTime":"2025-12-13T06:50:42Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 13 06:50:42 crc kubenswrapper[4971]: E1213 06:50:42.389988 4971 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-12-13T06:50:42Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-13T06:50:42Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-13T06:50:42Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-13T06:50:42Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-13T06:50:42Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-13T06:50:42Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-13T06:50:42Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-13T06:50:42Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"bd36cbc9-bbe9-45a1-a21a-ad2bd10499ee\\\",\\\"systemUUID\\\":\\\"a9af00dc-cf25-47af-9997-79577f8f3bea\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-13T06:50:42Z is after 2025-08-24T17:21:41Z" Dec 13 06:50:42 crc kubenswrapper[4971]: I1213 06:50:42.393195 4971 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 13 06:50:42 crc kubenswrapper[4971]: I1213 06:50:42.393229 4971 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 13 06:50:42 crc kubenswrapper[4971]: I1213 06:50:42.393239 4971 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 13 06:50:42 crc kubenswrapper[4971]: I1213 06:50:42.393254 4971 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 13 06:50:42 crc kubenswrapper[4971]: I1213 06:50:42.393264 4971 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-13T06:50:42Z","lastTransitionTime":"2025-12-13T06:50:42Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 13 06:50:42 crc kubenswrapper[4971]: E1213 06:50:42.403913 4971 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-12-13T06:50:42Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-13T06:50:42Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-13T06:50:42Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-13T06:50:42Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-13T06:50:42Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-13T06:50:42Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-13T06:50:42Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-13T06:50:42Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"bd36cbc9-bbe9-45a1-a21a-ad2bd10499ee\\\",\\\"systemUUID\\\":\\\"a9af00dc-cf25-47af-9997-79577f8f3bea\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-13T06:50:42Z is after 2025-08-24T17:21:41Z" Dec 13 06:50:42 crc kubenswrapper[4971]: E1213 06:50:42.404025 4971 kubelet_node_status.go:572] "Unable to update node status" err="update node status exceeds retry count" Dec 13 06:50:42 crc kubenswrapper[4971]: I1213 06:50:42.427177 4971 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 13 06:50:42 crc kubenswrapper[4971]: I1213 06:50:42.427215 4971 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 13 06:50:42 crc kubenswrapper[4971]: I1213 06:50:42.427230 4971 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 13 06:50:42 crc kubenswrapper[4971]: I1213 06:50:42.427250 4971 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 13 06:50:42 crc kubenswrapper[4971]: I1213 06:50:42.427264 4971 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-13T06:50:42Z","lastTransitionTime":"2025-12-13T06:50:42Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 13 06:50:42 crc kubenswrapper[4971]: I1213 06:50:42.529673 4971 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 13 06:50:42 crc kubenswrapper[4971]: I1213 06:50:42.529712 4971 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 13 06:50:42 crc kubenswrapper[4971]: I1213 06:50:42.529723 4971 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 13 06:50:42 crc kubenswrapper[4971]: I1213 06:50:42.529740 4971 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 13 06:50:42 crc kubenswrapper[4971]: I1213 06:50:42.529751 4971 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-13T06:50:42Z","lastTransitionTime":"2025-12-13T06:50:42Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 13 06:50:42 crc kubenswrapper[4971]: I1213 06:50:42.631575 4971 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 13 06:50:42 crc kubenswrapper[4971]: I1213 06:50:42.631609 4971 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 13 06:50:42 crc kubenswrapper[4971]: I1213 06:50:42.631617 4971 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 13 06:50:42 crc kubenswrapper[4971]: I1213 06:50:42.631630 4971 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 13 06:50:42 crc kubenswrapper[4971]: I1213 06:50:42.631639 4971 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-13T06:50:42Z","lastTransitionTime":"2025-12-13T06:50:42Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 13 06:50:42 crc kubenswrapper[4971]: I1213 06:50:42.733939 4971 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 13 06:50:42 crc kubenswrapper[4971]: I1213 06:50:42.733983 4971 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 13 06:50:42 crc kubenswrapper[4971]: I1213 06:50:42.733993 4971 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 13 06:50:42 crc kubenswrapper[4971]: I1213 06:50:42.734009 4971 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 13 06:50:42 crc kubenswrapper[4971]: I1213 06:50:42.734019 4971 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-13T06:50:42Z","lastTransitionTime":"2025-12-13T06:50:42Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 13 06:50:42 crc kubenswrapper[4971]: I1213 06:50:42.836960 4971 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 13 06:50:42 crc kubenswrapper[4971]: I1213 06:50:42.837002 4971 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 13 06:50:42 crc kubenswrapper[4971]: I1213 06:50:42.837014 4971 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 13 06:50:42 crc kubenswrapper[4971]: I1213 06:50:42.837030 4971 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 13 06:50:42 crc kubenswrapper[4971]: I1213 06:50:42.837041 4971 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-13T06:50:42Z","lastTransitionTime":"2025-12-13T06:50:42Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 13 06:50:42 crc kubenswrapper[4971]: I1213 06:50:42.939917 4971 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 13 06:50:42 crc kubenswrapper[4971]: I1213 06:50:42.939966 4971 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 13 06:50:42 crc kubenswrapper[4971]: I1213 06:50:42.939975 4971 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 13 06:50:42 crc kubenswrapper[4971]: I1213 06:50:42.939992 4971 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 13 06:50:42 crc kubenswrapper[4971]: I1213 06:50:42.940001 4971 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-13T06:50:42Z","lastTransitionTime":"2025-12-13T06:50:42Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 13 06:50:43 crc kubenswrapper[4971]: I1213 06:50:43.042706 4971 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 13 06:50:43 crc kubenswrapper[4971]: I1213 06:50:43.042785 4971 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 13 06:50:43 crc kubenswrapper[4971]: I1213 06:50:43.042802 4971 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 13 06:50:43 crc kubenswrapper[4971]: I1213 06:50:43.042822 4971 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 13 06:50:43 crc kubenswrapper[4971]: I1213 06:50:43.042834 4971 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-13T06:50:43Z","lastTransitionTime":"2025-12-13T06:50:43Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 13 06:50:43 crc kubenswrapper[4971]: I1213 06:50:43.145396 4971 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 13 06:50:43 crc kubenswrapper[4971]: I1213 06:50:43.145469 4971 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 13 06:50:43 crc kubenswrapper[4971]: I1213 06:50:43.145482 4971 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 13 06:50:43 crc kubenswrapper[4971]: I1213 06:50:43.145497 4971 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 13 06:50:43 crc kubenswrapper[4971]: I1213 06:50:43.145548 4971 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-13T06:50:43Z","lastTransitionTime":"2025-12-13T06:50:43Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 13 06:50:43 crc kubenswrapper[4971]: I1213 06:50:43.247683 4971 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 13 06:50:43 crc kubenswrapper[4971]: I1213 06:50:43.247716 4971 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 13 06:50:43 crc kubenswrapper[4971]: I1213 06:50:43.247725 4971 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 13 06:50:43 crc kubenswrapper[4971]: I1213 06:50:43.247739 4971 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 13 06:50:43 crc kubenswrapper[4971]: I1213 06:50:43.247748 4971 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-13T06:50:43Z","lastTransitionTime":"2025-12-13T06:50:43Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 13 06:50:43 crc kubenswrapper[4971]: I1213 06:50:43.349635 4971 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 13 06:50:43 crc kubenswrapper[4971]: I1213 06:50:43.349678 4971 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 13 06:50:43 crc kubenswrapper[4971]: I1213 06:50:43.349693 4971 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 13 06:50:43 crc kubenswrapper[4971]: I1213 06:50:43.349711 4971 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 13 06:50:43 crc kubenswrapper[4971]: I1213 06:50:43.349722 4971 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-13T06:50:43Z","lastTransitionTime":"2025-12-13T06:50:43Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 13 06:50:43 crc kubenswrapper[4971]: I1213 06:50:43.409606 4971 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/c7185dec-3240-42e8-9d90-6182a5e8b65c-metrics-certs\") pod \"network-metrics-daemon-x98k7\" (UID: \"c7185dec-3240-42e8-9d90-6182a5e8b65c\") " pod="openshift-multus/network-metrics-daemon-x98k7" Dec 13 06:50:43 crc kubenswrapper[4971]: E1213 06:50:43.409765 4971 secret.go:188] Couldn't get secret openshift-multus/metrics-daemon-secret: object "openshift-multus"/"metrics-daemon-secret" not registered Dec 13 06:50:43 crc kubenswrapper[4971]: E1213 06:50:43.409846 4971 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/c7185dec-3240-42e8-9d90-6182a5e8b65c-metrics-certs podName:c7185dec-3240-42e8-9d90-6182a5e8b65c nodeName:}" failed. No retries permitted until 2025-12-13 06:51:47.409829563 +0000 UTC m=+164.014239011 (durationBeforeRetry 1m4s). Error: MountVolume.SetUp failed for volume "metrics-certs" (UniqueName: "kubernetes.io/secret/c7185dec-3240-42e8-9d90-6182a5e8b65c-metrics-certs") pod "network-metrics-daemon-x98k7" (UID: "c7185dec-3240-42e8-9d90-6182a5e8b65c") : object "openshift-multus"/"metrics-daemon-secret" not registered Dec 13 06:50:43 crc kubenswrapper[4971]: I1213 06:50:43.451941 4971 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 13 06:50:43 crc kubenswrapper[4971]: I1213 06:50:43.451978 4971 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 13 06:50:43 crc kubenswrapper[4971]: I1213 06:50:43.451993 4971 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 13 06:50:43 crc kubenswrapper[4971]: I1213 06:50:43.452009 4971 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 13 06:50:43 crc kubenswrapper[4971]: I1213 06:50:43.452018 4971 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-13T06:50:43Z","lastTransitionTime":"2025-12-13T06:50:43Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 13 06:50:43 crc kubenswrapper[4971]: I1213 06:50:43.555449 4971 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 13 06:50:43 crc kubenswrapper[4971]: I1213 06:50:43.555572 4971 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 13 06:50:43 crc kubenswrapper[4971]: I1213 06:50:43.555601 4971 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 13 06:50:43 crc kubenswrapper[4971]: I1213 06:50:43.555633 4971 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 13 06:50:43 crc kubenswrapper[4971]: I1213 06:50:43.555654 4971 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-13T06:50:43Z","lastTransitionTime":"2025-12-13T06:50:43Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 13 06:50:43 crc kubenswrapper[4971]: I1213 06:50:43.658263 4971 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 13 06:50:43 crc kubenswrapper[4971]: I1213 06:50:43.658315 4971 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 13 06:50:43 crc kubenswrapper[4971]: I1213 06:50:43.658322 4971 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 13 06:50:43 crc kubenswrapper[4971]: I1213 06:50:43.658339 4971 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 13 06:50:43 crc kubenswrapper[4971]: I1213 06:50:43.658348 4971 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-13T06:50:43Z","lastTransitionTime":"2025-12-13T06:50:43Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 13 06:50:43 crc kubenswrapper[4971]: I1213 06:50:43.760776 4971 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 13 06:50:43 crc kubenswrapper[4971]: I1213 06:50:43.760809 4971 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 13 06:50:43 crc kubenswrapper[4971]: I1213 06:50:43.760820 4971 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 13 06:50:43 crc kubenswrapper[4971]: I1213 06:50:43.760838 4971 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 13 06:50:43 crc kubenswrapper[4971]: I1213 06:50:43.760850 4971 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-13T06:50:43Z","lastTransitionTime":"2025-12-13T06:50:43Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 13 06:50:43 crc kubenswrapper[4971]: I1213 06:50:43.768162 4971 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-x98k7" Dec 13 06:50:43 crc kubenswrapper[4971]: I1213 06:50:43.768207 4971 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 13 06:50:43 crc kubenswrapper[4971]: I1213 06:50:43.768223 4971 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 13 06:50:43 crc kubenswrapper[4971]: E1213 06:50:43.768270 4971 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-x98k7" podUID="c7185dec-3240-42e8-9d90-6182a5e8b65c" Dec 13 06:50:43 crc kubenswrapper[4971]: I1213 06:50:43.768169 4971 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 13 06:50:43 crc kubenswrapper[4971]: E1213 06:50:43.768431 4971 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 13 06:50:43 crc kubenswrapper[4971]: E1213 06:50:43.768494 4971 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 13 06:50:43 crc kubenswrapper[4971]: E1213 06:50:43.768592 4971 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 13 06:50:43 crc kubenswrapper[4971]: I1213 06:50:43.779569 4971 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-13T06:49:23Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-13T06:49:23Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-13T06:50:43Z is after 2025-08-24T17:21:41Z" Dec 13 06:50:43 crc kubenswrapper[4971]: I1213 06:50:43.789801 4971 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-13T06:49:26Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-13T06:49:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-13T06:49:26Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9c8a0d1db0e754462d805e82daf5f52ffd5457507a30456535ec59e096c9d092\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-13T06:49:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-13T06:50:43Z is after 2025-08-24T17:21:41Z" Dec 13 06:50:43 crc kubenswrapper[4971]: I1213 06:50:43.799220 4971 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-82xjz" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e0c4af12-edda-41d7-9b44-f87396174ef2\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T06:49:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T06:49:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T06:49:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T06:49:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://0f4577d9e37f860813e221f2e208177d8d383fd04f673ce357a0a405ea29d1e6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-13T06:49:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cnxsq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://7e0d2020e62f1a277cff7047e1907dc747ed393a1f6c983e1ee258bd156c16c7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-13T06:49:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cnxsq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-13T06:49:25Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-82xjz\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-13T06:50:43Z is after 2025-08-24T17:21:41Z" Dec 13 06:50:43 crc kubenswrapper[4971]: I1213 06:50:43.811671 4971 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-wwvpl" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"dddf744b-2122-49ad-a980-105c09636e0f\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T06:49:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T06:49:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T06:50:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T06:50:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c8680aaa9e05095ac7e467c896d9e927bccff908925768ce5ed469ab89a72475\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://4676d897c7fc668587faec563f3551a056557879cee3e865192d2bd28edf62fc\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-12-13T06:50:12Z\\\",\\\"message\\\":\\\"2025-12-13T06:49:27+00:00 [cnibincopy] Successfully copied files in /usr/src/multus-cni/rhel9/bin/ to /host/opt/cni/bin/upgrade_263f091d-d847-4fad-88d3-fbf82c65c231\\\\n2025-12-13T06:49:27+00:00 [cnibincopy] Successfully moved files in /host/opt/cni/bin/upgrade_263f091d-d847-4fad-88d3-fbf82c65c231 to /host/opt/cni/bin/\\\\n2025-12-13T06:49:27Z [verbose] multus-daemon started\\\\n2025-12-13T06:49:27Z [verbose] Readiness Indicator file check\\\\n2025-12-13T06:50:12Z [error] have you checked that your default network is ready? still waiting for readinessindicatorfile @ /host/run/multus/cni/net.d/10-ovn-kubernetes.conf. pollimmediate error: timed out waiting for the condition\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-13T06:49:26Z\\\"}},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-13T06:50:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-czz9v\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-13T06:49:25Z\\\"}}\" for pod \"openshift-multus\"/\"multus-wwvpl\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-13T06:50:43Z is after 2025-08-24T17:21:41Z" Dec 13 06:50:43 crc kubenswrapper[4971]: I1213 06:50:43.834230 4971 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-zlcg5" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"736c6b1a-93dd-4f71-81b6-1c43fcf5f556\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T06:49:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T06:49:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T06:49:25Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T06:49:25Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://325223543b3c7db417f769533559a1ca2b23a723494e1885b0fd0480b5a6399f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-13T06:49:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vzmsr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://f22a4ec756119c4f791fb5105cf1f31df3e548b2b07b1f909a62802e9b6035ae\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-13T06:49:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vzmsr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://f6073140d9f86929b6d4bd2960ede0617a5c9620b269bbab043bae37e0d389f5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-13T06:49:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vzmsr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://693a2ab434c250cb140a866e8068b156df9191414fed1d63bee919a2528ea6e5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-13T06:49:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vzmsr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://219620f51474e60b3e1f880c0972eb612dffc4bad24462bbba2c093ae817d852\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-13T06:49:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vzmsr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://f137f4b441c2402093ab430a9689a46221c4fb4bc70d99fb992f62b0a989c465\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-13T06:49:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vzmsr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://7dd9913a65acba73ae1f32f4398207c5e276a1e979d86f9e720e4fd761bae4a3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://7dd9913a65acba73ae1f32f4398207c5e276a1e979d86f9e720e4fd761bae4a3\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-12-13T06:50:24Z\\\",\\\"message\\\":\\\"lse, Template:false, AddressFamily:\\\\\\\"\\\\\\\"}, Rules:[]services.LBRule{services.LBRule{Source:services.Addr{IP:\\\\\\\"10.217.4.250\\\\\\\", Port:443, Template:(*services.Template)(nil)}, Targets:[]services.Addr{}}}, Templates:services.TemplateMap(nil), Switches:[]string{}, Routers:[]string{}, Groups:[]string{\\\\\\\"clusterLBGroup\\\\\\\"}}}\\\\nI1213 06:50:24.014680 7045 services_controller.go:452] Built service openshift-machine-api/machine-api-operator-machine-webhook per-node LB for network=default: []services.LB{}\\\\nI1213 06:50:24.014686 7045 services_controller.go:453] Built service openshift-machine-api/machine-api-operator-machine-webhook template LB for network=default: []services.LB{}\\\\nI1213 06:50:24.012650 7045 obj_retry.go:303] Retry object setup: *v1.Pod openshift-kube-scheduler/openshift-kube-scheduler-crc\\\\nI1213 06:50:24.014698 7045 services_controller.go:454] Service openshift-machine-api/machine-api-operator-machine-webhook for network=default has 1 cluster-wide, 0 per-node configs, 0 template configs, making 1 (cluster) 0 (per node) and 0 (template) load balancers\\\\nF1213 06:50:24.014708 7045 ovnkube.go:137] failed to run ovnkube: [failed to start network controller: failed to start default network controller: unable to create admin network policy controller, err: could not add Event Handler for anpInformer during admin network policy controller initial\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-13T06:50:23Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 40s restarting failed container=ovnkube-controller pod=ovnkube-node-zlcg5_openshift-ovn-kubernetes(736c6b1a-93dd-4f71-81b6-1c43fcf5f556)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vzmsr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://868049c6a3a4ebe4036d1d22e02ba6b76bced2ee7e269bb228accc4a0636678a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-13T06:49:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vzmsr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://451b22df6973ddc8fa2209a65d64d9e41b33eb231bb4adfab88cf0e5943bb246\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://451b22df6973ddc8fa2209a65d64d9e41b33eb231bb4adfab88cf0e5943bb246\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-13T06:49:26Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-13T06:49:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vzmsr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-13T06:49:25Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-zlcg5\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-13T06:50:43Z is after 2025-08-24T17:21:41Z" Dec 13 06:50:43 crc kubenswrapper[4971]: I1213 06:50:43.848032 4971 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-13T06:49:24Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-13T06:49:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-13T06:49:24Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d0b2ecf0d5b22f23217f42f259ae7ffb6359555b22c30eaebf667eaf254b3bab\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-13T06:49:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-13T06:50:43Z is after 2025-08-24T17:21:41Z" Dec 13 06:50:43 crc kubenswrapper[4971]: I1213 06:50:43.861349 4971 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"8ede2992-9576-4b23-83c2-24ac3b1d3caf\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T06:49:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T06:49:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T06:49:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T06:49:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T06:49:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9e61d25add1723ac286794abfe886358e9d64455c25379f7e3057fc3dd812418\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-13T06:49:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://7fb51d8435b4e3ce0411cae1d6c2824c552f0a33d79de01f969274516b39a5b3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-13T06:49:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://1c35f5e8d8fab6d77b9e2e0dedb195440de5d6af345cd5865abd32caa7fafd60\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-13T06:49:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://38806c430185715371ab9bf30a1af9a9d5d492abd9bb84acbdccc694f56e3a8a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"wait-for-host-port\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://38806c430185715371ab9bf30a1af9a9d5d492abd9bb84acbdccc694f56e3a8a\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-13T06:49:04Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-13T06:49:04Z\\\"}}}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-13T06:49:03Z\\\"}}\" for pod \"openshift-kube-scheduler\"/\"openshift-kube-scheduler-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-13T06:50:43Z is after 2025-08-24T17:21:41Z" Dec 13 06:50:43 crc kubenswrapper[4971]: I1213 06:50:43.863240 4971 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 13 06:50:43 crc kubenswrapper[4971]: I1213 06:50:43.863282 4971 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 13 06:50:43 crc kubenswrapper[4971]: I1213 06:50:43.863299 4971 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 13 06:50:43 crc kubenswrapper[4971]: I1213 06:50:43.863323 4971 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 13 06:50:43 crc kubenswrapper[4971]: I1213 06:50:43.863345 4971 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-13T06:50:43Z","lastTransitionTime":"2025-12-13T06:50:43Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 13 06:50:43 crc kubenswrapper[4971]: I1213 06:50:43.877111 4971 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-ncchz" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"de914d21-a688-401e-a66e-861beb6336d9\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T06:49:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T06:49:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T06:49:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T06:49:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://244ffe3dc9d8cf799aff4c4d795bf41554c1e6bb6667fbb0c7df322fb5f96c68\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-13T06:49:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lqmjj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://0b7111af6c5a198d02bb02858837cfb4f3b20828d8d206cd18aa0ffa102d0cd4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://0b7111af6c5a198d02bb02858837cfb4f3b20828d8d206cd18aa0ffa102d0cd4\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-13T06:49:26Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-13T06:49:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lqmjj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b36777fe40e74ae1cecb7b73e204139db5279ba51cef4f1e8c4a4c491d8be403\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://b36777fe40e74ae1cecb7b73e204139db5279ba51cef4f1e8c4a4c491d8be403\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-13T06:49:28Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-13T06:49:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lqmjj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://0597a06cc663e416142311a16941d98297577693a1fe27201ea42c27496cace2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://0597a06cc663e416142311a16941d98297577693a1fe27201ea42c27496cace2\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-13T06:49:29Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-13T06:49:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lqmjj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://3f52c2e22a75c010993ca5809a3a9d1304758d4cb6e83691ef1af15a4b514ac4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://3f52c2e22a75c010993ca5809a3a9d1304758d4cb6e83691ef1af15a4b514ac4\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-13T06:49:30Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-13T06:49:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lqmjj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://9e21ca708c4c4c1228b417cece51299e0792729b7c78884bd18b2dda8fe8355c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://9e21ca708c4c4c1228b417cece51299e0792729b7c78884bd18b2dda8fe8355c\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-13T06:49:31Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-13T06:49:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lqmjj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://74552159f12c87cdda09147692b400ccac0d4e43c1ed72fe8462c4249ad69c7b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://74552159f12c87cdda09147692b400ccac0d4e43c1ed72fe8462c4249ad69c7b\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-13T06:49:32Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-13T06:49:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lqmjj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-13T06:49:25Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-ncchz\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-13T06:50:43Z is after 2025-08-24T17:21:41Z" Dec 13 06:50:43 crc kubenswrapper[4971]: I1213 06:50:43.890182 4971 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-gkcd8" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"13f5c327-f95e-42b4-8783-218500c17694\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T06:49:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T06:49:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T06:49:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T06:49:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://77e846e36f1cb40d03471dc03de11b6c3da448baa407aad38093fae9a801cf92\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-13T06:49:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gvlhn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-13T06:49:27Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-gkcd8\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-13T06:50:43Z is after 2025-08-24T17:21:41Z" Dec 13 06:50:43 crc kubenswrapper[4971]: I1213 06:50:43.902739 4971 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-x98k7" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"c7185dec-3240-42e8-9d90-6182a5e8b65c\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T06:49:39Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T06:49:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T06:49:39Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T06:49:39Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bj2gn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bj2gn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-13T06:49:39Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-x98k7\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-13T06:50:43Z is after 2025-08-24T17:21:41Z" Dec 13 06:50:43 crc kubenswrapper[4971]: I1213 06:50:43.917330 4971 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"f9d31b5a-b770-4a15-95c8-35b49677c608\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T06:49:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T06:49:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T06:49:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T06:49:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T06:49:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c9baa042f9dada954faa57b5454cd6b0a12a297e56ddc192c52cff50e42a16fa\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-13T06:49:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://c788de86754f0bf37ee184672fc1dcb388ccebc246d4f15be4b5da661f725f8f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-13T06:49:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://760817511a9b4f96a68dc38d26631250eaab72e1dd8f123e653039b6b4e91f84\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-13T06:49:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b859be0c9feac856e9855c6a038b02b49b32fade97873c46d0bec0a32d8b5147\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-13T06:49:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-13T06:49:03Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-13T06:50:43Z is after 2025-08-24T17:21:41Z" Dec 13 06:50:43 crc kubenswrapper[4971]: I1213 06:50:43.938607 4971 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"1ed9f82e-fe5c-4b76-a619-62e5abd88d3b\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T06:49:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T06:49:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T06:49:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T06:49:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T06:49:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e457329a78a2d24337f994e8a0a1f75a2a667244aa75eb84613d1e7138364aac\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-13T06:49:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://eb014de8140deef097d9bd94f3b9fa4ab3558bc4449bef84b83bf62bfabf5836\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-13T06:49:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://2c438cc74d555f7c9db4569d1c56b78b1088ed8105697d2c0f31e6c5136f3dd0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-13T06:49:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://aa6fa6e4340dcd066252df3c56d79a3a39ed531d5eea8d1d5941d6952368d3af\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-13T06:49:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://c4471b98854c52d5cd084d0ba3e64e5b08e9518ec3c372cce2c875a440522457\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-13T06:49:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7d2746335c98dae3a0f43024d34cb62ebcb42a01c8f9ce18d52b29d4c45598e5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://7d2746335c98dae3a0f43024d34cb62ebcb42a01c8f9ce18d52b29d4c45598e5\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-13T06:49:04Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-13T06:49:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://bc0e77d10a738128bea82122496bc42235fc1688f8907756d93e61b651902b57\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://bc0e77d10a738128bea82122496bc42235fc1688f8907756d93e61b651902b57\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-13T06:49:05Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-13T06:49:05Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://f821e418cba5e832da8839268bcbbd9630bdf319a212d5382bfa2c4b0b73a268\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://f821e418cba5e832da8839268bcbbd9630bdf319a212d5382bfa2c4b0b73a268\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-13T06:49:07Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-13T06:49:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-13T06:49:03Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-13T06:50:43Z is after 2025-08-24T17:21:41Z" Dec 13 06:50:43 crc kubenswrapper[4971]: I1213 06:50:43.951659 4971 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-13T06:49:24Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-13T06:49:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-13T06:49:24Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4328cd6a8464ed468229bbfbc8265f969427916d72ebb92237dbe68efe70fa47\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-13T06:49:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://9d715ca82d9b5a594e41f2b98da6f4277da0e4e1dec114969f901c00f6d7553d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-13T06:49:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-13T06:50:43Z is after 2025-08-24T17:21:41Z" Dec 13 06:50:43 crc kubenswrapper[4971]: I1213 06:50:43.964764 4971 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-5rzh5" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"6bf4591d-2483-4132-9672-d0fd6c995968\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T06:49:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T06:49:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T06:49:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T06:49:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://50e2cb14e4ffb603acb06e2281615a99f06182d345c4e70f2446fa2077033bbb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-13T06:49:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dfxqn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-13T06:49:25Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-5rzh5\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-13T06:50:43Z is after 2025-08-24T17:21:41Z" Dec 13 06:50:43 crc kubenswrapper[4971]: I1213 06:50:43.966481 4971 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 13 06:50:43 crc kubenswrapper[4971]: I1213 06:50:43.966574 4971 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 13 06:50:43 crc kubenswrapper[4971]: I1213 06:50:43.966591 4971 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 13 06:50:43 crc kubenswrapper[4971]: I1213 06:50:43.966616 4971 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 13 06:50:43 crc kubenswrapper[4971]: I1213 06:50:43.966635 4971 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-13T06:50:43Z","lastTransitionTime":"2025-12-13T06:50:43Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 13 06:50:43 crc kubenswrapper[4971]: I1213 06:50:43.977070 4971 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-z572s" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"b4146d3d-4034-49a9-bd37-164856255137\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T06:49:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T06:49:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T06:49:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T06:49:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://46fd4e1070835cfbf7f35d497eff7f047473fc4048ee07812925df5ebdf55342\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-13T06:49:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-j6l9b\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://dc79a0968f85723e6ad0dc269c8da1f7472adb1bb2b87edd017ee2295bfbf89c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-13T06:49:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-j6l9b\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-13T06:49:38Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-z572s\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-13T06:50:43Z is after 2025-08-24T17:21:41Z" Dec 13 06:50:43 crc kubenswrapper[4971]: I1213 06:50:43.987862 4971 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5ef36745-87bb-4be8-b941-aa8f865a5c83\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T06:49:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T06:49:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T06:49:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T06:49:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T06:49:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://35b737b0d2c5c5503ba2819f34d2de0599792a77beb1fbeb561d44efd1f212c5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-crio\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-13T06:49:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kube\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"var-lib-kubelet\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f0d864149bfcbb6227e6a6fbde23fff650d688850978eae0caf04515eb21cd85\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://f0d864149bfcbb6227e6a6fbde23fff650d688850978eae0caf04515eb21cd85\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-13T06:49:04Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-13T06:49:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var\\\",\\\"name\\\":\\\"var-lib-kubelet\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-13T06:49:03Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"kube-rbac-proxy-crio-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-13T06:50:43Z is after 2025-08-24T17:21:41Z" Dec 13 06:50:44 crc kubenswrapper[4971]: I1213 06:50:44.001565 4971 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-13T06:49:23Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-13T06:49:23Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-13T06:50:43Z is after 2025-08-24T17:21:41Z" Dec 13 06:50:44 crc kubenswrapper[4971]: I1213 06:50:44.013843 4971 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-13T06:49:23Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-13T06:49:23Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-13T06:50:44Z is after 2025-08-24T17:21:41Z" Dec 13 06:50:44 crc kubenswrapper[4971]: I1213 06:50:44.028405 4971 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3ad3268a-a982-428c-9c06-4aaeddd372d9\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T06:49:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T06:49:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T06:49:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T06:49:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T06:49:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3716afdf2bc693be80a09fc9f7727d334f435caf1d4c8ec4811b3a1f1652cdfe\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-13T06:49:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://30df3abf8c79db8af52c6a918f73e1ce410ddafcf341636424c267e1eeb195a7\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-13T06:49:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://ccdf65f6d2b1eb3647a03ba37c305c32e34addfdf79ff937c8dddaa3e5d5b5c6\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-13T06:49:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://770b73dd0c2b6f38744e2c1cd6aa186a69bb87d1cd015a603ad2e33be06a8f6b\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://b3c1bab3d53f975b2e24582ce2771bc26d600e0dd7c30339dcfd18a071162ad4\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-12-13T06:49:23Z\\\",\\\"message\\\":\\\"file observer\\\\nW1213 06:49:23.283403 1 builder.go:272] unable to get owner reference (falling back to namespace): pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\nI1213 06:49:23.284336 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1213 06:49:23.286978 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-981773671/tls.crt::/tmp/serving-cert-981773671/tls.key\\\\\\\"\\\\nI1213 06:49:23.775072 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1213 06:49:23.778615 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1213 06:49:23.778645 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1213 06:49:23.778680 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1213 06:49:23.778691 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1213 06:49:23.785058 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1213 06:49:23.786150 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1213 06:49:23.786196 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1213 06:49:23.786223 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1213 06:49:23.786246 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1213 06:49:23.786269 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1213 06:49:23.786291 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI1213 06:49:23.785608 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF1213 06:49:23.787563 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-13T06:49:08Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-13T06:49:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://9acd07066deed483bd7fa61ee45fda09539d110717a134a56827f871186f92fc\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-13T06:49:07Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://38c91cfbaecfdbd3e3b663422a91ab4294610afb6bbf9ebdb77db24e2b76c682\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://38c91cfbaecfdbd3e3b663422a91ab4294610afb6bbf9ebdb77db24e2b76c682\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-13T06:49:05Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-13T06:49:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-13T06:49:03Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-13T06:50:44Z is after 2025-08-24T17:21:41Z" Dec 13 06:50:44 crc kubenswrapper[4971]: I1213 06:50:44.069245 4971 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 13 06:50:44 crc kubenswrapper[4971]: I1213 06:50:44.069282 4971 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 13 06:50:44 crc kubenswrapper[4971]: I1213 06:50:44.069291 4971 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 13 06:50:44 crc kubenswrapper[4971]: I1213 06:50:44.069306 4971 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 13 06:50:44 crc kubenswrapper[4971]: I1213 06:50:44.069317 4971 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-13T06:50:44Z","lastTransitionTime":"2025-12-13T06:50:44Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 13 06:50:44 crc kubenswrapper[4971]: I1213 06:50:44.171608 4971 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 13 06:50:44 crc kubenswrapper[4971]: I1213 06:50:44.171655 4971 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 13 06:50:44 crc kubenswrapper[4971]: I1213 06:50:44.171666 4971 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 13 06:50:44 crc kubenswrapper[4971]: I1213 06:50:44.171684 4971 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 13 06:50:44 crc kubenswrapper[4971]: I1213 06:50:44.171695 4971 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-13T06:50:44Z","lastTransitionTime":"2025-12-13T06:50:44Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 13 06:50:44 crc kubenswrapper[4971]: I1213 06:50:44.273099 4971 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 13 06:50:44 crc kubenswrapper[4971]: I1213 06:50:44.273142 4971 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 13 06:50:44 crc kubenswrapper[4971]: I1213 06:50:44.273152 4971 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 13 06:50:44 crc kubenswrapper[4971]: I1213 06:50:44.273168 4971 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 13 06:50:44 crc kubenswrapper[4971]: I1213 06:50:44.273180 4971 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-13T06:50:44Z","lastTransitionTime":"2025-12-13T06:50:44Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 13 06:50:44 crc kubenswrapper[4971]: I1213 06:50:44.375850 4971 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 13 06:50:44 crc kubenswrapper[4971]: I1213 06:50:44.375949 4971 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 13 06:50:44 crc kubenswrapper[4971]: I1213 06:50:44.375968 4971 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 13 06:50:44 crc kubenswrapper[4971]: I1213 06:50:44.375996 4971 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 13 06:50:44 crc kubenswrapper[4971]: I1213 06:50:44.376011 4971 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-13T06:50:44Z","lastTransitionTime":"2025-12-13T06:50:44Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 13 06:50:44 crc kubenswrapper[4971]: I1213 06:50:44.478209 4971 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 13 06:50:44 crc kubenswrapper[4971]: I1213 06:50:44.478252 4971 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 13 06:50:44 crc kubenswrapper[4971]: I1213 06:50:44.478264 4971 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 13 06:50:44 crc kubenswrapper[4971]: I1213 06:50:44.478282 4971 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 13 06:50:44 crc kubenswrapper[4971]: I1213 06:50:44.478293 4971 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-13T06:50:44Z","lastTransitionTime":"2025-12-13T06:50:44Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 13 06:50:44 crc kubenswrapper[4971]: I1213 06:50:44.581270 4971 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 13 06:50:44 crc kubenswrapper[4971]: I1213 06:50:44.581317 4971 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 13 06:50:44 crc kubenswrapper[4971]: I1213 06:50:44.581351 4971 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 13 06:50:44 crc kubenswrapper[4971]: I1213 06:50:44.581370 4971 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 13 06:50:44 crc kubenswrapper[4971]: I1213 06:50:44.581382 4971 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-13T06:50:44Z","lastTransitionTime":"2025-12-13T06:50:44Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 13 06:50:44 crc kubenswrapper[4971]: I1213 06:50:44.684707 4971 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 13 06:50:44 crc kubenswrapper[4971]: I1213 06:50:44.684766 4971 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 13 06:50:44 crc kubenswrapper[4971]: I1213 06:50:44.684786 4971 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 13 06:50:44 crc kubenswrapper[4971]: I1213 06:50:44.684811 4971 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 13 06:50:44 crc kubenswrapper[4971]: I1213 06:50:44.684825 4971 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-13T06:50:44Z","lastTransitionTime":"2025-12-13T06:50:44Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 13 06:50:44 crc kubenswrapper[4971]: I1213 06:50:44.787153 4971 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 13 06:50:44 crc kubenswrapper[4971]: I1213 06:50:44.787192 4971 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 13 06:50:44 crc kubenswrapper[4971]: I1213 06:50:44.787202 4971 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 13 06:50:44 crc kubenswrapper[4971]: I1213 06:50:44.787218 4971 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 13 06:50:44 crc kubenswrapper[4971]: I1213 06:50:44.787229 4971 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-13T06:50:44Z","lastTransitionTime":"2025-12-13T06:50:44Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 13 06:50:44 crc kubenswrapper[4971]: I1213 06:50:44.889937 4971 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 13 06:50:44 crc kubenswrapper[4971]: I1213 06:50:44.889979 4971 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 13 06:50:44 crc kubenswrapper[4971]: I1213 06:50:44.889990 4971 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 13 06:50:44 crc kubenswrapper[4971]: I1213 06:50:44.890004 4971 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 13 06:50:44 crc kubenswrapper[4971]: I1213 06:50:44.890015 4971 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-13T06:50:44Z","lastTransitionTime":"2025-12-13T06:50:44Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 13 06:50:44 crc kubenswrapper[4971]: I1213 06:50:44.993239 4971 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 13 06:50:44 crc kubenswrapper[4971]: I1213 06:50:44.993273 4971 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 13 06:50:44 crc kubenswrapper[4971]: I1213 06:50:44.993285 4971 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 13 06:50:44 crc kubenswrapper[4971]: I1213 06:50:44.993302 4971 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 13 06:50:44 crc kubenswrapper[4971]: I1213 06:50:44.993314 4971 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-13T06:50:44Z","lastTransitionTime":"2025-12-13T06:50:44Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 13 06:50:45 crc kubenswrapper[4971]: I1213 06:50:45.095544 4971 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 13 06:50:45 crc kubenswrapper[4971]: I1213 06:50:45.095592 4971 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 13 06:50:45 crc kubenswrapper[4971]: I1213 06:50:45.095604 4971 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 13 06:50:45 crc kubenswrapper[4971]: I1213 06:50:45.095622 4971 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 13 06:50:45 crc kubenswrapper[4971]: I1213 06:50:45.095633 4971 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-13T06:50:45Z","lastTransitionTime":"2025-12-13T06:50:45Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 13 06:50:45 crc kubenswrapper[4971]: I1213 06:50:45.199035 4971 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 13 06:50:45 crc kubenswrapper[4971]: I1213 06:50:45.199109 4971 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 13 06:50:45 crc kubenswrapper[4971]: I1213 06:50:45.199124 4971 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 13 06:50:45 crc kubenswrapper[4971]: I1213 06:50:45.199142 4971 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 13 06:50:45 crc kubenswrapper[4971]: I1213 06:50:45.199155 4971 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-13T06:50:45Z","lastTransitionTime":"2025-12-13T06:50:45Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 13 06:50:45 crc kubenswrapper[4971]: I1213 06:50:45.305477 4971 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 13 06:50:45 crc kubenswrapper[4971]: I1213 06:50:45.305556 4971 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 13 06:50:45 crc kubenswrapper[4971]: I1213 06:50:45.305572 4971 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 13 06:50:45 crc kubenswrapper[4971]: I1213 06:50:45.305592 4971 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 13 06:50:45 crc kubenswrapper[4971]: I1213 06:50:45.305614 4971 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-13T06:50:45Z","lastTransitionTime":"2025-12-13T06:50:45Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 13 06:50:45 crc kubenswrapper[4971]: I1213 06:50:45.408568 4971 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 13 06:50:45 crc kubenswrapper[4971]: I1213 06:50:45.408619 4971 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 13 06:50:45 crc kubenswrapper[4971]: I1213 06:50:45.408630 4971 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 13 06:50:45 crc kubenswrapper[4971]: I1213 06:50:45.408646 4971 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 13 06:50:45 crc kubenswrapper[4971]: I1213 06:50:45.408656 4971 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-13T06:50:45Z","lastTransitionTime":"2025-12-13T06:50:45Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 13 06:50:45 crc kubenswrapper[4971]: I1213 06:50:45.511114 4971 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 13 06:50:45 crc kubenswrapper[4971]: I1213 06:50:45.511152 4971 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 13 06:50:45 crc kubenswrapper[4971]: I1213 06:50:45.511163 4971 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 13 06:50:45 crc kubenswrapper[4971]: I1213 06:50:45.511181 4971 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 13 06:50:45 crc kubenswrapper[4971]: I1213 06:50:45.511194 4971 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-13T06:50:45Z","lastTransitionTime":"2025-12-13T06:50:45Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 13 06:50:45 crc kubenswrapper[4971]: I1213 06:50:45.614102 4971 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 13 06:50:45 crc kubenswrapper[4971]: I1213 06:50:45.614157 4971 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 13 06:50:45 crc kubenswrapper[4971]: I1213 06:50:45.614169 4971 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 13 06:50:45 crc kubenswrapper[4971]: I1213 06:50:45.614187 4971 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 13 06:50:45 crc kubenswrapper[4971]: I1213 06:50:45.614198 4971 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-13T06:50:45Z","lastTransitionTime":"2025-12-13T06:50:45Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 13 06:50:45 crc kubenswrapper[4971]: I1213 06:50:45.716784 4971 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 13 06:50:45 crc kubenswrapper[4971]: I1213 06:50:45.716828 4971 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 13 06:50:45 crc kubenswrapper[4971]: I1213 06:50:45.716836 4971 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 13 06:50:45 crc kubenswrapper[4971]: I1213 06:50:45.716856 4971 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 13 06:50:45 crc kubenswrapper[4971]: I1213 06:50:45.716866 4971 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-13T06:50:45Z","lastTransitionTime":"2025-12-13T06:50:45Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 13 06:50:45 crc kubenswrapper[4971]: I1213 06:50:45.767806 4971 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-x98k7" Dec 13 06:50:45 crc kubenswrapper[4971]: I1213 06:50:45.767912 4971 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 13 06:50:45 crc kubenswrapper[4971]: E1213 06:50:45.768398 4971 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 13 06:50:45 crc kubenswrapper[4971]: I1213 06:50:45.767964 4971 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 13 06:50:45 crc kubenswrapper[4971]: E1213 06:50:45.768489 4971 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 13 06:50:45 crc kubenswrapper[4971]: E1213 06:50:45.768247 4971 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-x98k7" podUID="c7185dec-3240-42e8-9d90-6182a5e8b65c" Dec 13 06:50:45 crc kubenswrapper[4971]: I1213 06:50:45.767926 4971 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 13 06:50:45 crc kubenswrapper[4971]: E1213 06:50:45.768596 4971 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 13 06:50:45 crc kubenswrapper[4971]: I1213 06:50:45.819163 4971 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 13 06:50:45 crc kubenswrapper[4971]: I1213 06:50:45.819442 4971 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 13 06:50:45 crc kubenswrapper[4971]: I1213 06:50:45.819565 4971 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 13 06:50:45 crc kubenswrapper[4971]: I1213 06:50:45.819658 4971 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 13 06:50:45 crc kubenswrapper[4971]: I1213 06:50:45.819719 4971 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-13T06:50:45Z","lastTransitionTime":"2025-12-13T06:50:45Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 13 06:50:45 crc kubenswrapper[4971]: I1213 06:50:45.921734 4971 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 13 06:50:45 crc kubenswrapper[4971]: I1213 06:50:45.921774 4971 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 13 06:50:45 crc kubenswrapper[4971]: I1213 06:50:45.921783 4971 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 13 06:50:45 crc kubenswrapper[4971]: I1213 06:50:45.921800 4971 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 13 06:50:45 crc kubenswrapper[4971]: I1213 06:50:45.921808 4971 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-13T06:50:45Z","lastTransitionTime":"2025-12-13T06:50:45Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 13 06:50:46 crc kubenswrapper[4971]: I1213 06:50:46.024294 4971 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 13 06:50:46 crc kubenswrapper[4971]: I1213 06:50:46.024342 4971 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 13 06:50:46 crc kubenswrapper[4971]: I1213 06:50:46.024353 4971 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 13 06:50:46 crc kubenswrapper[4971]: I1213 06:50:46.024371 4971 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 13 06:50:46 crc kubenswrapper[4971]: I1213 06:50:46.024381 4971 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-13T06:50:46Z","lastTransitionTime":"2025-12-13T06:50:46Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 13 06:50:46 crc kubenswrapper[4971]: I1213 06:50:46.127600 4971 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 13 06:50:46 crc kubenswrapper[4971]: I1213 06:50:46.127670 4971 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 13 06:50:46 crc kubenswrapper[4971]: I1213 06:50:46.127683 4971 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 13 06:50:46 crc kubenswrapper[4971]: I1213 06:50:46.127721 4971 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 13 06:50:46 crc kubenswrapper[4971]: I1213 06:50:46.127741 4971 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-13T06:50:46Z","lastTransitionTime":"2025-12-13T06:50:46Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 13 06:50:46 crc kubenswrapper[4971]: I1213 06:50:46.231327 4971 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 13 06:50:46 crc kubenswrapper[4971]: I1213 06:50:46.231385 4971 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 13 06:50:46 crc kubenswrapper[4971]: I1213 06:50:46.231396 4971 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 13 06:50:46 crc kubenswrapper[4971]: I1213 06:50:46.231416 4971 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 13 06:50:46 crc kubenswrapper[4971]: I1213 06:50:46.231428 4971 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-13T06:50:46Z","lastTransitionTime":"2025-12-13T06:50:46Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 13 06:50:46 crc kubenswrapper[4971]: I1213 06:50:46.334144 4971 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 13 06:50:46 crc kubenswrapper[4971]: I1213 06:50:46.334204 4971 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 13 06:50:46 crc kubenswrapper[4971]: I1213 06:50:46.334222 4971 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 13 06:50:46 crc kubenswrapper[4971]: I1213 06:50:46.334248 4971 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 13 06:50:46 crc kubenswrapper[4971]: I1213 06:50:46.334265 4971 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-13T06:50:46Z","lastTransitionTime":"2025-12-13T06:50:46Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 13 06:50:46 crc kubenswrapper[4971]: I1213 06:50:46.438020 4971 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 13 06:50:46 crc kubenswrapper[4971]: I1213 06:50:46.438268 4971 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 13 06:50:46 crc kubenswrapper[4971]: I1213 06:50:46.438362 4971 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 13 06:50:46 crc kubenswrapper[4971]: I1213 06:50:46.438453 4971 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 13 06:50:46 crc kubenswrapper[4971]: I1213 06:50:46.438551 4971 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-13T06:50:46Z","lastTransitionTime":"2025-12-13T06:50:46Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 13 06:50:46 crc kubenswrapper[4971]: I1213 06:50:46.542095 4971 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 13 06:50:46 crc kubenswrapper[4971]: I1213 06:50:46.542160 4971 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 13 06:50:46 crc kubenswrapper[4971]: I1213 06:50:46.542173 4971 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 13 06:50:46 crc kubenswrapper[4971]: I1213 06:50:46.542189 4971 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 13 06:50:46 crc kubenswrapper[4971]: I1213 06:50:46.542204 4971 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-13T06:50:46Z","lastTransitionTime":"2025-12-13T06:50:46Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 13 06:50:46 crc kubenswrapper[4971]: I1213 06:50:46.644920 4971 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 13 06:50:46 crc kubenswrapper[4971]: I1213 06:50:46.644966 4971 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 13 06:50:46 crc kubenswrapper[4971]: I1213 06:50:46.644980 4971 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 13 06:50:46 crc kubenswrapper[4971]: I1213 06:50:46.644997 4971 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 13 06:50:46 crc kubenswrapper[4971]: I1213 06:50:46.645010 4971 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-13T06:50:46Z","lastTransitionTime":"2025-12-13T06:50:46Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 13 06:50:46 crc kubenswrapper[4971]: I1213 06:50:46.747335 4971 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 13 06:50:46 crc kubenswrapper[4971]: I1213 06:50:46.747368 4971 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 13 06:50:46 crc kubenswrapper[4971]: I1213 06:50:46.747380 4971 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 13 06:50:46 crc kubenswrapper[4971]: I1213 06:50:46.747398 4971 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 13 06:50:46 crc kubenswrapper[4971]: I1213 06:50:46.747410 4971 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-13T06:50:46Z","lastTransitionTime":"2025-12-13T06:50:46Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 13 06:50:46 crc kubenswrapper[4971]: I1213 06:50:46.849834 4971 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 13 06:50:46 crc kubenswrapper[4971]: I1213 06:50:46.849914 4971 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 13 06:50:46 crc kubenswrapper[4971]: I1213 06:50:46.849925 4971 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 13 06:50:46 crc kubenswrapper[4971]: I1213 06:50:46.849972 4971 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 13 06:50:46 crc kubenswrapper[4971]: I1213 06:50:46.849984 4971 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-13T06:50:46Z","lastTransitionTime":"2025-12-13T06:50:46Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 13 06:50:46 crc kubenswrapper[4971]: I1213 06:50:46.952622 4971 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 13 06:50:46 crc kubenswrapper[4971]: I1213 06:50:46.952846 4971 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 13 06:50:46 crc kubenswrapper[4971]: I1213 06:50:46.952858 4971 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 13 06:50:46 crc kubenswrapper[4971]: I1213 06:50:46.952875 4971 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 13 06:50:46 crc kubenswrapper[4971]: I1213 06:50:46.952890 4971 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-13T06:50:46Z","lastTransitionTime":"2025-12-13T06:50:46Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 13 06:50:47 crc kubenswrapper[4971]: I1213 06:50:47.054959 4971 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 13 06:50:47 crc kubenswrapper[4971]: I1213 06:50:47.055019 4971 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 13 06:50:47 crc kubenswrapper[4971]: I1213 06:50:47.055041 4971 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 13 06:50:47 crc kubenswrapper[4971]: I1213 06:50:47.055071 4971 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 13 06:50:47 crc kubenswrapper[4971]: I1213 06:50:47.055092 4971 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-13T06:50:47Z","lastTransitionTime":"2025-12-13T06:50:47Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 13 06:50:47 crc kubenswrapper[4971]: I1213 06:50:47.157398 4971 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 13 06:50:47 crc kubenswrapper[4971]: I1213 06:50:47.157457 4971 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 13 06:50:47 crc kubenswrapper[4971]: I1213 06:50:47.157467 4971 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 13 06:50:47 crc kubenswrapper[4971]: I1213 06:50:47.157484 4971 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 13 06:50:47 crc kubenswrapper[4971]: I1213 06:50:47.157494 4971 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-13T06:50:47Z","lastTransitionTime":"2025-12-13T06:50:47Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 13 06:50:47 crc kubenswrapper[4971]: I1213 06:50:47.260756 4971 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 13 06:50:47 crc kubenswrapper[4971]: I1213 06:50:47.260805 4971 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 13 06:50:47 crc kubenswrapper[4971]: I1213 06:50:47.260815 4971 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 13 06:50:47 crc kubenswrapper[4971]: I1213 06:50:47.260833 4971 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 13 06:50:47 crc kubenswrapper[4971]: I1213 06:50:47.260851 4971 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-13T06:50:47Z","lastTransitionTime":"2025-12-13T06:50:47Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 13 06:50:47 crc kubenswrapper[4971]: I1213 06:50:47.363339 4971 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 13 06:50:47 crc kubenswrapper[4971]: I1213 06:50:47.363383 4971 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 13 06:50:47 crc kubenswrapper[4971]: I1213 06:50:47.363392 4971 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 13 06:50:47 crc kubenswrapper[4971]: I1213 06:50:47.363408 4971 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 13 06:50:47 crc kubenswrapper[4971]: I1213 06:50:47.363419 4971 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-13T06:50:47Z","lastTransitionTime":"2025-12-13T06:50:47Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 13 06:50:47 crc kubenswrapper[4971]: I1213 06:50:47.466665 4971 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 13 06:50:47 crc kubenswrapper[4971]: I1213 06:50:47.466703 4971 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 13 06:50:47 crc kubenswrapper[4971]: I1213 06:50:47.466714 4971 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 13 06:50:47 crc kubenswrapper[4971]: I1213 06:50:47.466742 4971 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 13 06:50:47 crc kubenswrapper[4971]: I1213 06:50:47.466755 4971 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-13T06:50:47Z","lastTransitionTime":"2025-12-13T06:50:47Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 13 06:50:47 crc kubenswrapper[4971]: I1213 06:50:47.569285 4971 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 13 06:50:47 crc kubenswrapper[4971]: I1213 06:50:47.569363 4971 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 13 06:50:47 crc kubenswrapper[4971]: I1213 06:50:47.569376 4971 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 13 06:50:47 crc kubenswrapper[4971]: I1213 06:50:47.569396 4971 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 13 06:50:47 crc kubenswrapper[4971]: I1213 06:50:47.569410 4971 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-13T06:50:47Z","lastTransitionTime":"2025-12-13T06:50:47Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 13 06:50:47 crc kubenswrapper[4971]: I1213 06:50:47.671456 4971 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 13 06:50:47 crc kubenswrapper[4971]: I1213 06:50:47.672025 4971 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 13 06:50:47 crc kubenswrapper[4971]: I1213 06:50:47.672096 4971 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 13 06:50:47 crc kubenswrapper[4971]: I1213 06:50:47.672169 4971 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 13 06:50:47 crc kubenswrapper[4971]: I1213 06:50:47.672243 4971 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-13T06:50:47Z","lastTransitionTime":"2025-12-13T06:50:47Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 13 06:50:47 crc kubenswrapper[4971]: I1213 06:50:47.768616 4971 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-x98k7" Dec 13 06:50:47 crc kubenswrapper[4971]: E1213 06:50:47.768760 4971 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-x98k7" podUID="c7185dec-3240-42e8-9d90-6182a5e8b65c" Dec 13 06:50:47 crc kubenswrapper[4971]: I1213 06:50:47.768625 4971 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 13 06:50:47 crc kubenswrapper[4971]: E1213 06:50:47.769060 4971 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 13 06:50:47 crc kubenswrapper[4971]: I1213 06:50:47.769161 4971 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 13 06:50:47 crc kubenswrapper[4971]: E1213 06:50:47.769315 4971 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 13 06:50:47 crc kubenswrapper[4971]: I1213 06:50:47.769201 4971 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 13 06:50:47 crc kubenswrapper[4971]: E1213 06:50:47.769561 4971 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 13 06:50:47 crc kubenswrapper[4971]: I1213 06:50:47.774110 4971 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 13 06:50:47 crc kubenswrapper[4971]: I1213 06:50:47.774155 4971 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 13 06:50:47 crc kubenswrapper[4971]: I1213 06:50:47.774167 4971 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 13 06:50:47 crc kubenswrapper[4971]: I1213 06:50:47.774181 4971 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 13 06:50:47 crc kubenswrapper[4971]: I1213 06:50:47.774192 4971 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-13T06:50:47Z","lastTransitionTime":"2025-12-13T06:50:47Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 13 06:50:47 crc kubenswrapper[4971]: I1213 06:50:47.876232 4971 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 13 06:50:47 crc kubenswrapper[4971]: I1213 06:50:47.876541 4971 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 13 06:50:47 crc kubenswrapper[4971]: I1213 06:50:47.876639 4971 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 13 06:50:47 crc kubenswrapper[4971]: I1213 06:50:47.876730 4971 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 13 06:50:47 crc kubenswrapper[4971]: I1213 06:50:47.876809 4971 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-13T06:50:47Z","lastTransitionTime":"2025-12-13T06:50:47Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 13 06:50:47 crc kubenswrapper[4971]: I1213 06:50:47.978758 4971 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 13 06:50:47 crc kubenswrapper[4971]: I1213 06:50:47.978798 4971 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 13 06:50:47 crc kubenswrapper[4971]: I1213 06:50:47.978813 4971 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 13 06:50:47 crc kubenswrapper[4971]: I1213 06:50:47.978829 4971 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 13 06:50:47 crc kubenswrapper[4971]: I1213 06:50:47.978840 4971 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-13T06:50:47Z","lastTransitionTime":"2025-12-13T06:50:47Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 13 06:50:48 crc kubenswrapper[4971]: I1213 06:50:48.081416 4971 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 13 06:50:48 crc kubenswrapper[4971]: I1213 06:50:48.081477 4971 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 13 06:50:48 crc kubenswrapper[4971]: I1213 06:50:48.081491 4971 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 13 06:50:48 crc kubenswrapper[4971]: I1213 06:50:48.081508 4971 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 13 06:50:48 crc kubenswrapper[4971]: I1213 06:50:48.081544 4971 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-13T06:50:48Z","lastTransitionTime":"2025-12-13T06:50:48Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 13 06:50:48 crc kubenswrapper[4971]: I1213 06:50:48.183672 4971 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 13 06:50:48 crc kubenswrapper[4971]: I1213 06:50:48.183726 4971 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 13 06:50:48 crc kubenswrapper[4971]: I1213 06:50:48.183737 4971 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 13 06:50:48 crc kubenswrapper[4971]: I1213 06:50:48.183751 4971 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 13 06:50:48 crc kubenswrapper[4971]: I1213 06:50:48.183760 4971 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-13T06:50:48Z","lastTransitionTime":"2025-12-13T06:50:48Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 13 06:50:48 crc kubenswrapper[4971]: I1213 06:50:48.286022 4971 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 13 06:50:48 crc kubenswrapper[4971]: I1213 06:50:48.286084 4971 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 13 06:50:48 crc kubenswrapper[4971]: I1213 06:50:48.286095 4971 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 13 06:50:48 crc kubenswrapper[4971]: I1213 06:50:48.286109 4971 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 13 06:50:48 crc kubenswrapper[4971]: I1213 06:50:48.286117 4971 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-13T06:50:48Z","lastTransitionTime":"2025-12-13T06:50:48Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 13 06:50:48 crc kubenswrapper[4971]: I1213 06:50:48.387822 4971 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 13 06:50:48 crc kubenswrapper[4971]: I1213 06:50:48.387862 4971 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 13 06:50:48 crc kubenswrapper[4971]: I1213 06:50:48.387871 4971 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 13 06:50:48 crc kubenswrapper[4971]: I1213 06:50:48.387886 4971 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 13 06:50:48 crc kubenswrapper[4971]: I1213 06:50:48.387905 4971 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-13T06:50:48Z","lastTransitionTime":"2025-12-13T06:50:48Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 13 06:50:48 crc kubenswrapper[4971]: I1213 06:50:48.490119 4971 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 13 06:50:48 crc kubenswrapper[4971]: I1213 06:50:48.490219 4971 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 13 06:50:48 crc kubenswrapper[4971]: I1213 06:50:48.490231 4971 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 13 06:50:48 crc kubenswrapper[4971]: I1213 06:50:48.490252 4971 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 13 06:50:48 crc kubenswrapper[4971]: I1213 06:50:48.490264 4971 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-13T06:50:48Z","lastTransitionTime":"2025-12-13T06:50:48Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 13 06:50:48 crc kubenswrapper[4971]: I1213 06:50:48.593748 4971 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 13 06:50:48 crc kubenswrapper[4971]: I1213 06:50:48.593781 4971 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 13 06:50:48 crc kubenswrapper[4971]: I1213 06:50:48.593793 4971 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 13 06:50:48 crc kubenswrapper[4971]: I1213 06:50:48.593808 4971 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 13 06:50:48 crc kubenswrapper[4971]: I1213 06:50:48.593816 4971 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-13T06:50:48Z","lastTransitionTime":"2025-12-13T06:50:48Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 13 06:50:48 crc kubenswrapper[4971]: I1213 06:50:48.696126 4971 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 13 06:50:48 crc kubenswrapper[4971]: I1213 06:50:48.696181 4971 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 13 06:50:48 crc kubenswrapper[4971]: I1213 06:50:48.696194 4971 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 13 06:50:48 crc kubenswrapper[4971]: I1213 06:50:48.696212 4971 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 13 06:50:48 crc kubenswrapper[4971]: I1213 06:50:48.696222 4971 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-13T06:50:48Z","lastTransitionTime":"2025-12-13T06:50:48Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 13 06:50:48 crc kubenswrapper[4971]: I1213 06:50:48.798873 4971 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 13 06:50:48 crc kubenswrapper[4971]: I1213 06:50:48.799140 4971 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 13 06:50:48 crc kubenswrapper[4971]: I1213 06:50:48.799218 4971 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 13 06:50:48 crc kubenswrapper[4971]: I1213 06:50:48.799290 4971 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 13 06:50:48 crc kubenswrapper[4971]: I1213 06:50:48.799348 4971 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-13T06:50:48Z","lastTransitionTime":"2025-12-13T06:50:48Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 13 06:50:48 crc kubenswrapper[4971]: I1213 06:50:48.901900 4971 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 13 06:50:48 crc kubenswrapper[4971]: I1213 06:50:48.901942 4971 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 13 06:50:48 crc kubenswrapper[4971]: I1213 06:50:48.901957 4971 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 13 06:50:48 crc kubenswrapper[4971]: I1213 06:50:48.901972 4971 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 13 06:50:48 crc kubenswrapper[4971]: I1213 06:50:48.901981 4971 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-13T06:50:48Z","lastTransitionTime":"2025-12-13T06:50:48Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 13 06:50:49 crc kubenswrapper[4971]: I1213 06:50:49.004368 4971 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 13 06:50:49 crc kubenswrapper[4971]: I1213 06:50:49.004409 4971 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 13 06:50:49 crc kubenswrapper[4971]: I1213 06:50:49.004418 4971 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 13 06:50:49 crc kubenswrapper[4971]: I1213 06:50:49.004435 4971 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 13 06:50:49 crc kubenswrapper[4971]: I1213 06:50:49.004443 4971 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-13T06:50:49Z","lastTransitionTime":"2025-12-13T06:50:49Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 13 06:50:49 crc kubenswrapper[4971]: I1213 06:50:49.106801 4971 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 13 06:50:49 crc kubenswrapper[4971]: I1213 06:50:49.107062 4971 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 13 06:50:49 crc kubenswrapper[4971]: I1213 06:50:49.107134 4971 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 13 06:50:49 crc kubenswrapper[4971]: I1213 06:50:49.107205 4971 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 13 06:50:49 crc kubenswrapper[4971]: I1213 06:50:49.107275 4971 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-13T06:50:49Z","lastTransitionTime":"2025-12-13T06:50:49Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 13 06:50:49 crc kubenswrapper[4971]: I1213 06:50:49.209296 4971 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 13 06:50:49 crc kubenswrapper[4971]: I1213 06:50:49.209598 4971 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 13 06:50:49 crc kubenswrapper[4971]: I1213 06:50:49.209691 4971 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 13 06:50:49 crc kubenswrapper[4971]: I1213 06:50:49.209766 4971 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 13 06:50:49 crc kubenswrapper[4971]: I1213 06:50:49.209854 4971 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-13T06:50:49Z","lastTransitionTime":"2025-12-13T06:50:49Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 13 06:50:49 crc kubenswrapper[4971]: I1213 06:50:49.311844 4971 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 13 06:50:49 crc kubenswrapper[4971]: I1213 06:50:49.311879 4971 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 13 06:50:49 crc kubenswrapper[4971]: I1213 06:50:49.311887 4971 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 13 06:50:49 crc kubenswrapper[4971]: I1213 06:50:49.311902 4971 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 13 06:50:49 crc kubenswrapper[4971]: I1213 06:50:49.311912 4971 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-13T06:50:49Z","lastTransitionTime":"2025-12-13T06:50:49Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 13 06:50:49 crc kubenswrapper[4971]: I1213 06:50:49.414440 4971 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 13 06:50:49 crc kubenswrapper[4971]: I1213 06:50:49.414494 4971 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 13 06:50:49 crc kubenswrapper[4971]: I1213 06:50:49.414505 4971 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 13 06:50:49 crc kubenswrapper[4971]: I1213 06:50:49.414546 4971 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 13 06:50:49 crc kubenswrapper[4971]: I1213 06:50:49.414560 4971 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-13T06:50:49Z","lastTransitionTime":"2025-12-13T06:50:49Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 13 06:50:49 crc kubenswrapper[4971]: I1213 06:50:49.517224 4971 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 13 06:50:49 crc kubenswrapper[4971]: I1213 06:50:49.517459 4971 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 13 06:50:49 crc kubenswrapper[4971]: I1213 06:50:49.517566 4971 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 13 06:50:49 crc kubenswrapper[4971]: I1213 06:50:49.517706 4971 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 13 06:50:49 crc kubenswrapper[4971]: I1213 06:50:49.517806 4971 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-13T06:50:49Z","lastTransitionTime":"2025-12-13T06:50:49Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 13 06:50:49 crc kubenswrapper[4971]: I1213 06:50:49.620623 4971 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 13 06:50:49 crc kubenswrapper[4971]: I1213 06:50:49.620674 4971 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 13 06:50:49 crc kubenswrapper[4971]: I1213 06:50:49.620685 4971 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 13 06:50:49 crc kubenswrapper[4971]: I1213 06:50:49.620705 4971 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 13 06:50:49 crc kubenswrapper[4971]: I1213 06:50:49.620716 4971 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-13T06:50:49Z","lastTransitionTime":"2025-12-13T06:50:49Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 13 06:50:49 crc kubenswrapper[4971]: I1213 06:50:49.724391 4971 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 13 06:50:49 crc kubenswrapper[4971]: I1213 06:50:49.724434 4971 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 13 06:50:49 crc kubenswrapper[4971]: I1213 06:50:49.724446 4971 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 13 06:50:49 crc kubenswrapper[4971]: I1213 06:50:49.724464 4971 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 13 06:50:49 crc kubenswrapper[4971]: I1213 06:50:49.724475 4971 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-13T06:50:49Z","lastTransitionTime":"2025-12-13T06:50:49Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 13 06:50:49 crc kubenswrapper[4971]: I1213 06:50:49.768164 4971 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 13 06:50:49 crc kubenswrapper[4971]: E1213 06:50:49.768312 4971 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 13 06:50:49 crc kubenswrapper[4971]: I1213 06:50:49.768561 4971 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 13 06:50:49 crc kubenswrapper[4971]: E1213 06:50:49.768632 4971 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 13 06:50:49 crc kubenswrapper[4971]: I1213 06:50:49.768758 4971 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-x98k7" Dec 13 06:50:49 crc kubenswrapper[4971]: I1213 06:50:49.768790 4971 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 13 06:50:49 crc kubenswrapper[4971]: E1213 06:50:49.768864 4971 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-x98k7" podUID="c7185dec-3240-42e8-9d90-6182a5e8b65c" Dec 13 06:50:49 crc kubenswrapper[4971]: E1213 06:50:49.769019 4971 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 13 06:50:49 crc kubenswrapper[4971]: I1213 06:50:49.827742 4971 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 13 06:50:49 crc kubenswrapper[4971]: I1213 06:50:49.827784 4971 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 13 06:50:49 crc kubenswrapper[4971]: I1213 06:50:49.827798 4971 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 13 06:50:49 crc kubenswrapper[4971]: I1213 06:50:49.827815 4971 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 13 06:50:49 crc kubenswrapper[4971]: I1213 06:50:49.827827 4971 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-13T06:50:49Z","lastTransitionTime":"2025-12-13T06:50:49Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 13 06:50:49 crc kubenswrapper[4971]: I1213 06:50:49.929628 4971 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 13 06:50:49 crc kubenswrapper[4971]: I1213 06:50:49.929689 4971 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 13 06:50:49 crc kubenswrapper[4971]: I1213 06:50:49.929713 4971 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 13 06:50:49 crc kubenswrapper[4971]: I1213 06:50:49.929740 4971 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 13 06:50:49 crc kubenswrapper[4971]: I1213 06:50:49.929763 4971 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-13T06:50:49Z","lastTransitionTime":"2025-12-13T06:50:49Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 13 06:50:50 crc kubenswrapper[4971]: I1213 06:50:50.031777 4971 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 13 06:50:50 crc kubenswrapper[4971]: I1213 06:50:50.031806 4971 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 13 06:50:50 crc kubenswrapper[4971]: I1213 06:50:50.031817 4971 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 13 06:50:50 crc kubenswrapper[4971]: I1213 06:50:50.031833 4971 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 13 06:50:50 crc kubenswrapper[4971]: I1213 06:50:50.031845 4971 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-13T06:50:50Z","lastTransitionTime":"2025-12-13T06:50:50Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 13 06:50:50 crc kubenswrapper[4971]: I1213 06:50:50.135934 4971 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 13 06:50:50 crc kubenswrapper[4971]: I1213 06:50:50.135969 4971 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 13 06:50:50 crc kubenswrapper[4971]: I1213 06:50:50.135978 4971 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 13 06:50:50 crc kubenswrapper[4971]: I1213 06:50:50.135997 4971 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 13 06:50:50 crc kubenswrapper[4971]: I1213 06:50:50.136007 4971 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-13T06:50:50Z","lastTransitionTime":"2025-12-13T06:50:50Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 13 06:50:50 crc kubenswrapper[4971]: I1213 06:50:50.238848 4971 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 13 06:50:50 crc kubenswrapper[4971]: I1213 06:50:50.238929 4971 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 13 06:50:50 crc kubenswrapper[4971]: I1213 06:50:50.238939 4971 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 13 06:50:50 crc kubenswrapper[4971]: I1213 06:50:50.238953 4971 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 13 06:50:50 crc kubenswrapper[4971]: I1213 06:50:50.238962 4971 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-13T06:50:50Z","lastTransitionTime":"2025-12-13T06:50:50Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 13 06:50:50 crc kubenswrapper[4971]: I1213 06:50:50.341306 4971 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 13 06:50:50 crc kubenswrapper[4971]: I1213 06:50:50.341344 4971 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 13 06:50:50 crc kubenswrapper[4971]: I1213 06:50:50.341353 4971 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 13 06:50:50 crc kubenswrapper[4971]: I1213 06:50:50.341367 4971 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 13 06:50:50 crc kubenswrapper[4971]: I1213 06:50:50.341377 4971 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-13T06:50:50Z","lastTransitionTime":"2025-12-13T06:50:50Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 13 06:50:50 crc kubenswrapper[4971]: I1213 06:50:50.443365 4971 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 13 06:50:50 crc kubenswrapper[4971]: I1213 06:50:50.443407 4971 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 13 06:50:50 crc kubenswrapper[4971]: I1213 06:50:50.443418 4971 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 13 06:50:50 crc kubenswrapper[4971]: I1213 06:50:50.443434 4971 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 13 06:50:50 crc kubenswrapper[4971]: I1213 06:50:50.443444 4971 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-13T06:50:50Z","lastTransitionTime":"2025-12-13T06:50:50Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 13 06:50:50 crc kubenswrapper[4971]: I1213 06:50:50.545977 4971 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 13 06:50:50 crc kubenswrapper[4971]: I1213 06:50:50.546028 4971 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 13 06:50:50 crc kubenswrapper[4971]: I1213 06:50:50.546045 4971 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 13 06:50:50 crc kubenswrapper[4971]: I1213 06:50:50.546069 4971 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 13 06:50:50 crc kubenswrapper[4971]: I1213 06:50:50.546081 4971 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-13T06:50:50Z","lastTransitionTime":"2025-12-13T06:50:50Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 13 06:50:50 crc kubenswrapper[4971]: I1213 06:50:50.648126 4971 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 13 06:50:50 crc kubenswrapper[4971]: I1213 06:50:50.648168 4971 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 13 06:50:50 crc kubenswrapper[4971]: I1213 06:50:50.648180 4971 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 13 06:50:50 crc kubenswrapper[4971]: I1213 06:50:50.648194 4971 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 13 06:50:50 crc kubenswrapper[4971]: I1213 06:50:50.648205 4971 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-13T06:50:50Z","lastTransitionTime":"2025-12-13T06:50:50Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 13 06:50:50 crc kubenswrapper[4971]: I1213 06:50:50.750362 4971 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 13 06:50:50 crc kubenswrapper[4971]: I1213 06:50:50.750397 4971 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 13 06:50:50 crc kubenswrapper[4971]: I1213 06:50:50.750407 4971 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 13 06:50:50 crc kubenswrapper[4971]: I1213 06:50:50.750424 4971 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 13 06:50:50 crc kubenswrapper[4971]: I1213 06:50:50.750434 4971 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-13T06:50:50Z","lastTransitionTime":"2025-12-13T06:50:50Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 13 06:50:50 crc kubenswrapper[4971]: I1213 06:50:50.768972 4971 scope.go:117] "RemoveContainer" containerID="7dd9913a65acba73ae1f32f4398207c5e276a1e979d86f9e720e4fd761bae4a3" Dec 13 06:50:50 crc kubenswrapper[4971]: E1213 06:50:50.769150 4971 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"ovnkube-controller\" with CrashLoopBackOff: \"back-off 40s restarting failed container=ovnkube-controller pod=ovnkube-node-zlcg5_openshift-ovn-kubernetes(736c6b1a-93dd-4f71-81b6-1c43fcf5f556)\"" pod="openshift-ovn-kubernetes/ovnkube-node-zlcg5" podUID="736c6b1a-93dd-4f71-81b6-1c43fcf5f556" Dec 13 06:50:50 crc kubenswrapper[4971]: I1213 06:50:50.852279 4971 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 13 06:50:50 crc kubenswrapper[4971]: I1213 06:50:50.852324 4971 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 13 06:50:50 crc kubenswrapper[4971]: I1213 06:50:50.852342 4971 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 13 06:50:50 crc kubenswrapper[4971]: I1213 06:50:50.852356 4971 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 13 06:50:50 crc kubenswrapper[4971]: I1213 06:50:50.852364 4971 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-13T06:50:50Z","lastTransitionTime":"2025-12-13T06:50:50Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 13 06:50:50 crc kubenswrapper[4971]: I1213 06:50:50.954436 4971 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 13 06:50:50 crc kubenswrapper[4971]: I1213 06:50:50.954471 4971 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 13 06:50:50 crc kubenswrapper[4971]: I1213 06:50:50.954480 4971 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 13 06:50:50 crc kubenswrapper[4971]: I1213 06:50:50.954497 4971 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 13 06:50:50 crc kubenswrapper[4971]: I1213 06:50:50.954530 4971 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-13T06:50:50Z","lastTransitionTime":"2025-12-13T06:50:50Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 13 06:50:51 crc kubenswrapper[4971]: I1213 06:50:51.057073 4971 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 13 06:50:51 crc kubenswrapper[4971]: I1213 06:50:51.057125 4971 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 13 06:50:51 crc kubenswrapper[4971]: I1213 06:50:51.057137 4971 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 13 06:50:51 crc kubenswrapper[4971]: I1213 06:50:51.057154 4971 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 13 06:50:51 crc kubenswrapper[4971]: I1213 06:50:51.057165 4971 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-13T06:50:51Z","lastTransitionTime":"2025-12-13T06:50:51Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 13 06:50:51 crc kubenswrapper[4971]: I1213 06:50:51.159824 4971 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 13 06:50:51 crc kubenswrapper[4971]: I1213 06:50:51.159877 4971 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 13 06:50:51 crc kubenswrapper[4971]: I1213 06:50:51.159894 4971 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 13 06:50:51 crc kubenswrapper[4971]: I1213 06:50:51.159913 4971 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 13 06:50:51 crc kubenswrapper[4971]: I1213 06:50:51.159932 4971 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-13T06:50:51Z","lastTransitionTime":"2025-12-13T06:50:51Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 13 06:50:51 crc kubenswrapper[4971]: I1213 06:50:51.262349 4971 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 13 06:50:51 crc kubenswrapper[4971]: I1213 06:50:51.262385 4971 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 13 06:50:51 crc kubenswrapper[4971]: I1213 06:50:51.262397 4971 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 13 06:50:51 crc kubenswrapper[4971]: I1213 06:50:51.262413 4971 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 13 06:50:51 crc kubenswrapper[4971]: I1213 06:50:51.262423 4971 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-13T06:50:51Z","lastTransitionTime":"2025-12-13T06:50:51Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 13 06:50:51 crc kubenswrapper[4971]: I1213 06:50:51.365056 4971 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 13 06:50:51 crc kubenswrapper[4971]: I1213 06:50:51.365109 4971 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 13 06:50:51 crc kubenswrapper[4971]: I1213 06:50:51.365127 4971 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 13 06:50:51 crc kubenswrapper[4971]: I1213 06:50:51.365146 4971 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 13 06:50:51 crc kubenswrapper[4971]: I1213 06:50:51.365156 4971 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-13T06:50:51Z","lastTransitionTime":"2025-12-13T06:50:51Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 13 06:50:51 crc kubenswrapper[4971]: I1213 06:50:51.467069 4971 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 13 06:50:51 crc kubenswrapper[4971]: I1213 06:50:51.467119 4971 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 13 06:50:51 crc kubenswrapper[4971]: I1213 06:50:51.467130 4971 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 13 06:50:51 crc kubenswrapper[4971]: I1213 06:50:51.467145 4971 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 13 06:50:51 crc kubenswrapper[4971]: I1213 06:50:51.467155 4971 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-13T06:50:51Z","lastTransitionTime":"2025-12-13T06:50:51Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 13 06:50:51 crc kubenswrapper[4971]: I1213 06:50:51.570873 4971 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 13 06:50:51 crc kubenswrapper[4971]: I1213 06:50:51.570915 4971 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 13 06:50:51 crc kubenswrapper[4971]: I1213 06:50:51.570923 4971 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 13 06:50:51 crc kubenswrapper[4971]: I1213 06:50:51.570941 4971 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 13 06:50:51 crc kubenswrapper[4971]: I1213 06:50:51.570951 4971 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-13T06:50:51Z","lastTransitionTime":"2025-12-13T06:50:51Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 13 06:50:51 crc kubenswrapper[4971]: I1213 06:50:51.673280 4971 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 13 06:50:51 crc kubenswrapper[4971]: I1213 06:50:51.673318 4971 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 13 06:50:51 crc kubenswrapper[4971]: I1213 06:50:51.673328 4971 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 13 06:50:51 crc kubenswrapper[4971]: I1213 06:50:51.673342 4971 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 13 06:50:51 crc kubenswrapper[4971]: I1213 06:50:51.673351 4971 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-13T06:50:51Z","lastTransitionTime":"2025-12-13T06:50:51Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 13 06:50:51 crc kubenswrapper[4971]: I1213 06:50:51.768016 4971 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 13 06:50:51 crc kubenswrapper[4971]: I1213 06:50:51.768108 4971 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 13 06:50:51 crc kubenswrapper[4971]: E1213 06:50:51.768138 4971 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 13 06:50:51 crc kubenswrapper[4971]: I1213 06:50:51.768025 4971 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-x98k7" Dec 13 06:50:51 crc kubenswrapper[4971]: E1213 06:50:51.768244 4971 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 13 06:50:51 crc kubenswrapper[4971]: I1213 06:50:51.768041 4971 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 13 06:50:51 crc kubenswrapper[4971]: E1213 06:50:51.768373 4971 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-x98k7" podUID="c7185dec-3240-42e8-9d90-6182a5e8b65c" Dec 13 06:50:51 crc kubenswrapper[4971]: E1213 06:50:51.768473 4971 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 13 06:50:51 crc kubenswrapper[4971]: I1213 06:50:51.775008 4971 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 13 06:50:51 crc kubenswrapper[4971]: I1213 06:50:51.775057 4971 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 13 06:50:51 crc kubenswrapper[4971]: I1213 06:50:51.775082 4971 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 13 06:50:51 crc kubenswrapper[4971]: I1213 06:50:51.775096 4971 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 13 06:50:51 crc kubenswrapper[4971]: I1213 06:50:51.775108 4971 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-13T06:50:51Z","lastTransitionTime":"2025-12-13T06:50:51Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 13 06:50:51 crc kubenswrapper[4971]: I1213 06:50:51.877416 4971 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 13 06:50:51 crc kubenswrapper[4971]: I1213 06:50:51.877506 4971 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 13 06:50:51 crc kubenswrapper[4971]: I1213 06:50:51.877546 4971 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 13 06:50:51 crc kubenswrapper[4971]: I1213 06:50:51.877570 4971 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 13 06:50:51 crc kubenswrapper[4971]: I1213 06:50:51.877592 4971 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-13T06:50:51Z","lastTransitionTime":"2025-12-13T06:50:51Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 13 06:50:51 crc kubenswrapper[4971]: I1213 06:50:51.980816 4971 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 13 06:50:51 crc kubenswrapper[4971]: I1213 06:50:51.980865 4971 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 13 06:50:51 crc kubenswrapper[4971]: I1213 06:50:51.980876 4971 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 13 06:50:51 crc kubenswrapper[4971]: I1213 06:50:51.980899 4971 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 13 06:50:51 crc kubenswrapper[4971]: I1213 06:50:51.980914 4971 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-13T06:50:51Z","lastTransitionTime":"2025-12-13T06:50:51Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 13 06:50:52 crc kubenswrapper[4971]: I1213 06:50:52.083306 4971 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 13 06:50:52 crc kubenswrapper[4971]: I1213 06:50:52.083340 4971 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 13 06:50:52 crc kubenswrapper[4971]: I1213 06:50:52.083356 4971 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 13 06:50:52 crc kubenswrapper[4971]: I1213 06:50:52.083371 4971 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 13 06:50:52 crc kubenswrapper[4971]: I1213 06:50:52.083382 4971 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-13T06:50:52Z","lastTransitionTime":"2025-12-13T06:50:52Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 13 06:50:52 crc kubenswrapper[4971]: I1213 06:50:52.185898 4971 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 13 06:50:52 crc kubenswrapper[4971]: I1213 06:50:52.186204 4971 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 13 06:50:52 crc kubenswrapper[4971]: I1213 06:50:52.186276 4971 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 13 06:50:52 crc kubenswrapper[4971]: I1213 06:50:52.186351 4971 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 13 06:50:52 crc kubenswrapper[4971]: I1213 06:50:52.186421 4971 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-13T06:50:52Z","lastTransitionTime":"2025-12-13T06:50:52Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 13 06:50:52 crc kubenswrapper[4971]: I1213 06:50:52.288710 4971 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 13 06:50:52 crc kubenswrapper[4971]: I1213 06:50:52.288750 4971 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 13 06:50:52 crc kubenswrapper[4971]: I1213 06:50:52.288764 4971 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 13 06:50:52 crc kubenswrapper[4971]: I1213 06:50:52.288779 4971 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 13 06:50:52 crc kubenswrapper[4971]: I1213 06:50:52.288790 4971 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-13T06:50:52Z","lastTransitionTime":"2025-12-13T06:50:52Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 13 06:50:52 crc kubenswrapper[4971]: I1213 06:50:52.391071 4971 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 13 06:50:52 crc kubenswrapper[4971]: I1213 06:50:52.391113 4971 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 13 06:50:52 crc kubenswrapper[4971]: I1213 06:50:52.391123 4971 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 13 06:50:52 crc kubenswrapper[4971]: I1213 06:50:52.391138 4971 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 13 06:50:52 crc kubenswrapper[4971]: I1213 06:50:52.391147 4971 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-13T06:50:52Z","lastTransitionTime":"2025-12-13T06:50:52Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 13 06:50:52 crc kubenswrapper[4971]: I1213 06:50:52.493473 4971 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 13 06:50:52 crc kubenswrapper[4971]: I1213 06:50:52.493540 4971 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 13 06:50:52 crc kubenswrapper[4971]: I1213 06:50:52.493555 4971 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 13 06:50:52 crc kubenswrapper[4971]: I1213 06:50:52.493574 4971 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 13 06:50:52 crc kubenswrapper[4971]: I1213 06:50:52.493584 4971 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-13T06:50:52Z","lastTransitionTime":"2025-12-13T06:50:52Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 13 06:50:52 crc kubenswrapper[4971]: I1213 06:50:52.535406 4971 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 13 06:50:52 crc kubenswrapper[4971]: I1213 06:50:52.535695 4971 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 13 06:50:52 crc kubenswrapper[4971]: I1213 06:50:52.535791 4971 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 13 06:50:52 crc kubenswrapper[4971]: I1213 06:50:52.535924 4971 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 13 06:50:52 crc kubenswrapper[4971]: I1213 06:50:52.536013 4971 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-13T06:50:52Z","lastTransitionTime":"2025-12-13T06:50:52Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 13 06:50:52 crc kubenswrapper[4971]: I1213 06:50:52.584304 4971 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-cluster-version/cluster-version-operator-5c965bbfc6-jtmld"] Dec 13 06:50:52 crc kubenswrapper[4971]: I1213 06:50:52.584695 4971 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-jtmld" Dec 13 06:50:52 crc kubenswrapper[4971]: I1213 06:50:52.586413 4971 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-cluster-version"/"kube-root-ca.crt" Dec 13 06:50:52 crc kubenswrapper[4971]: I1213 06:50:52.587002 4971 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-cluster-version"/"openshift-service-ca.crt" Dec 13 06:50:52 crc kubenswrapper[4971]: I1213 06:50:52.587321 4971 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-cluster-version"/"cluster-version-operator-serving-cert" Dec 13 06:50:52 crc kubenswrapper[4971]: I1213 06:50:52.590732 4971 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-cluster-version"/"default-dockercfg-gxtc4" Dec 13 06:50:52 crc kubenswrapper[4971]: I1213 06:50:52.605037 4971 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" podStartSLOduration=40.605007604 podStartE2EDuration="40.605007604s" podCreationTimestamp="2025-12-13 06:50:12 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-13 06:50:52.604968702 +0000 UTC m=+109.209378200" watchObservedRunningTime="2025-12-13 06:50:52.605007604 +0000 UTC m=+109.209417052" Dec 13 06:50:52 crc kubenswrapper[4971]: I1213 06:50:52.646501 4971 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-etcd/etcd-crc" podStartSLOduration=89.646484783 podStartE2EDuration="1m29.646484783s" podCreationTimestamp="2025-12-13 06:49:23 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-13 06:50:52.630283373 +0000 UTC m=+109.234692821" watchObservedRunningTime="2025-12-13 06:50:52.646484783 +0000 UTC m=+109.250894221" Dec 13 06:50:52 crc kubenswrapper[4971]: I1213 06:50:52.662558 4971 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-ssl-certs\" (UniqueName: \"kubernetes.io/host-path/2f42c147-5c5e-415c-9b62-faa800eb8a1d-etc-ssl-certs\") pod \"cluster-version-operator-5c965bbfc6-jtmld\" (UID: \"2f42c147-5c5e-415c-9b62-faa800eb8a1d\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-jtmld" Dec 13 06:50:52 crc kubenswrapper[4971]: I1213 06:50:52.662625 4971 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-cvo-updatepayloads\" (UniqueName: \"kubernetes.io/host-path/2f42c147-5c5e-415c-9b62-faa800eb8a1d-etc-cvo-updatepayloads\") pod \"cluster-version-operator-5c965bbfc6-jtmld\" (UID: \"2f42c147-5c5e-415c-9b62-faa800eb8a1d\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-jtmld" Dec 13 06:50:52 crc kubenswrapper[4971]: I1213 06:50:52.662658 4971 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/2f42c147-5c5e-415c-9b62-faa800eb8a1d-service-ca\") pod \"cluster-version-operator-5c965bbfc6-jtmld\" (UID: \"2f42c147-5c5e-415c-9b62-faa800eb8a1d\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-jtmld" Dec 13 06:50:52 crc kubenswrapper[4971]: I1213 06:50:52.662694 4971 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/2f42c147-5c5e-415c-9b62-faa800eb8a1d-serving-cert\") pod \"cluster-version-operator-5c965bbfc6-jtmld\" (UID: \"2f42c147-5c5e-415c-9b62-faa800eb8a1d\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-jtmld" Dec 13 06:50:52 crc kubenswrapper[4971]: I1213 06:50:52.662728 4971 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/2f42c147-5c5e-415c-9b62-faa800eb8a1d-kube-api-access\") pod \"cluster-version-operator-5c965bbfc6-jtmld\" (UID: \"2f42c147-5c5e-415c-9b62-faa800eb8a1d\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-jtmld" Dec 13 06:50:52 crc kubenswrapper[4971]: I1213 06:50:52.671391 4971 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-dns/node-resolver-5rzh5" podStartSLOduration=87.67137021 podStartE2EDuration="1m27.67137021s" podCreationTimestamp="2025-12-13 06:49:25 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-13 06:50:52.659378872 +0000 UTC m=+109.263788320" watchObservedRunningTime="2025-12-13 06:50:52.67137021 +0000 UTC m=+109.275779658" Dec 13 06:50:52 crc kubenswrapper[4971]: I1213 06:50:52.686344 4971 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-kube-apiserver/kube-apiserver-crc" podStartSLOduration=88.686321408 podStartE2EDuration="1m28.686321408s" podCreationTimestamp="2025-12-13 06:49:24 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-13 06:50:52.686231075 +0000 UTC m=+109.290640543" watchObservedRunningTime="2025-12-13 06:50:52.686321408 +0000 UTC m=+109.290730856" Dec 13 06:50:52 crc kubenswrapper[4971]: I1213 06:50:52.686502 4971 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-z572s" podStartSLOduration=87.686496534 podStartE2EDuration="1m27.686496534s" podCreationTimestamp="2025-12-13 06:49:25 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-13 06:50:52.672001121 +0000 UTC m=+109.276410569" watchObservedRunningTime="2025-12-13 06:50:52.686496534 +0000 UTC m=+109.290905982" Dec 13 06:50:52 crc kubenswrapper[4971]: I1213 06:50:52.763502 4971 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/2f42c147-5c5e-415c-9b62-faa800eb8a1d-service-ca\") pod \"cluster-version-operator-5c965bbfc6-jtmld\" (UID: \"2f42c147-5c5e-415c-9b62-faa800eb8a1d\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-jtmld" Dec 13 06:50:52 crc kubenswrapper[4971]: I1213 06:50:52.763575 4971 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/2f42c147-5c5e-415c-9b62-faa800eb8a1d-serving-cert\") pod \"cluster-version-operator-5c965bbfc6-jtmld\" (UID: \"2f42c147-5c5e-415c-9b62-faa800eb8a1d\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-jtmld" Dec 13 06:50:52 crc kubenswrapper[4971]: I1213 06:50:52.763596 4971 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/2f42c147-5c5e-415c-9b62-faa800eb8a1d-kube-api-access\") pod \"cluster-version-operator-5c965bbfc6-jtmld\" (UID: \"2f42c147-5c5e-415c-9b62-faa800eb8a1d\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-jtmld" Dec 13 06:50:52 crc kubenswrapper[4971]: I1213 06:50:52.763660 4971 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-ssl-certs\" (UniqueName: \"kubernetes.io/host-path/2f42c147-5c5e-415c-9b62-faa800eb8a1d-etc-ssl-certs\") pod \"cluster-version-operator-5c965bbfc6-jtmld\" (UID: \"2f42c147-5c5e-415c-9b62-faa800eb8a1d\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-jtmld" Dec 13 06:50:52 crc kubenswrapper[4971]: I1213 06:50:52.763690 4971 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-cvo-updatepayloads\" (UniqueName: \"kubernetes.io/host-path/2f42c147-5c5e-415c-9b62-faa800eb8a1d-etc-cvo-updatepayloads\") pod \"cluster-version-operator-5c965bbfc6-jtmld\" (UID: \"2f42c147-5c5e-415c-9b62-faa800eb8a1d\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-jtmld" Dec 13 06:50:52 crc kubenswrapper[4971]: I1213 06:50:52.763737 4971 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-cvo-updatepayloads\" (UniqueName: \"kubernetes.io/host-path/2f42c147-5c5e-415c-9b62-faa800eb8a1d-etc-cvo-updatepayloads\") pod \"cluster-version-operator-5c965bbfc6-jtmld\" (UID: \"2f42c147-5c5e-415c-9b62-faa800eb8a1d\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-jtmld" Dec 13 06:50:52 crc kubenswrapper[4971]: I1213 06:50:52.763781 4971 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-ssl-certs\" (UniqueName: \"kubernetes.io/host-path/2f42c147-5c5e-415c-9b62-faa800eb8a1d-etc-ssl-certs\") pod \"cluster-version-operator-5c965bbfc6-jtmld\" (UID: \"2f42c147-5c5e-415c-9b62-faa800eb8a1d\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-jtmld" Dec 13 06:50:52 crc kubenswrapper[4971]: I1213 06:50:52.764343 4971 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/2f42c147-5c5e-415c-9b62-faa800eb8a1d-service-ca\") pod \"cluster-version-operator-5c965bbfc6-jtmld\" (UID: \"2f42c147-5c5e-415c-9b62-faa800eb8a1d\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-jtmld" Dec 13 06:50:52 crc kubenswrapper[4971]: I1213 06:50:52.769022 4971 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-machine-config-operator/machine-config-daemon-82xjz" podStartSLOduration=87.769000079 podStartE2EDuration="1m27.769000079s" podCreationTimestamp="2025-12-13 06:49:25 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-13 06:50:52.76872337 +0000 UTC m=+109.373132828" watchObservedRunningTime="2025-12-13 06:50:52.769000079 +0000 UTC m=+109.373409527" Dec 13 06:50:52 crc kubenswrapper[4971]: I1213 06:50:52.773705 4971 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/2f42c147-5c5e-415c-9b62-faa800eb8a1d-serving-cert\") pod \"cluster-version-operator-5c965bbfc6-jtmld\" (UID: \"2f42c147-5c5e-415c-9b62-faa800eb8a1d\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-jtmld" Dec 13 06:50:52 crc kubenswrapper[4971]: I1213 06:50:52.779664 4971 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/2f42c147-5c5e-415c-9b62-faa800eb8a1d-kube-api-access\") pod \"cluster-version-operator-5c965bbfc6-jtmld\" (UID: \"2f42c147-5c5e-415c-9b62-faa800eb8a1d\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-jtmld" Dec 13 06:50:52 crc kubenswrapper[4971]: I1213 06:50:52.802857 4971 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-multus/multus-wwvpl" podStartSLOduration=87.80283884 podStartE2EDuration="1m27.80283884s" podCreationTimestamp="2025-12-13 06:49:25 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-13 06:50:52.781701661 +0000 UTC m=+109.386111109" watchObservedRunningTime="2025-12-13 06:50:52.80283884 +0000 UTC m=+109.407248288" Dec 13 06:50:52 crc kubenswrapper[4971]: I1213 06:50:52.826603 4971 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-kube-controller-manager/kube-controller-manager-crc" podStartSLOduration=89.826582587 podStartE2EDuration="1m29.826582587s" podCreationTimestamp="2025-12-13 06:49:23 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-13 06:50:52.82638721 +0000 UTC m=+109.430796658" watchObservedRunningTime="2025-12-13 06:50:52.826582587 +0000 UTC m=+109.430992035" Dec 13 06:50:52 crc kubenswrapper[4971]: I1213 06:50:52.843258 4971 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" podStartSLOduration=55.843237733 podStartE2EDuration="55.843237733s" podCreationTimestamp="2025-12-13 06:49:57 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-13 06:50:52.842754786 +0000 UTC m=+109.447164254" watchObservedRunningTime="2025-12-13 06:50:52.843237733 +0000 UTC m=+109.447647181" Dec 13 06:50:52 crc kubenswrapper[4971]: I1213 06:50:52.896667 4971 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-multus/multus-additional-cni-plugins-ncchz" podStartSLOduration=87.896632729 podStartE2EDuration="1m27.896632729s" podCreationTimestamp="2025-12-13 06:49:25 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-13 06:50:52.872370664 +0000 UTC m=+109.476780142" watchObservedRunningTime="2025-12-13 06:50:52.896632729 +0000 UTC m=+109.501042177" Dec 13 06:50:52 crc kubenswrapper[4971]: I1213 06:50:52.897336 4971 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-jtmld" Dec 13 06:50:52 crc kubenswrapper[4971]: I1213 06:50:52.897551 4971 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-image-registry/node-ca-gkcd8" podStartSLOduration=87.89754535 podStartE2EDuration="1m27.89754535s" podCreationTimestamp="2025-12-13 06:49:25 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-13 06:50:52.896277157 +0000 UTC m=+109.500686615" watchObservedRunningTime="2025-12-13 06:50:52.89754535 +0000 UTC m=+109.501954798" Dec 13 06:50:52 crc kubenswrapper[4971]: W1213 06:50:52.912564 4971 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod2f42c147_5c5e_415c_9b62_faa800eb8a1d.slice/crio-fbb5559900dda2fcaa6514c83523e4de313d645eefd9604326474da0cd2999b4 WatchSource:0}: Error finding container fbb5559900dda2fcaa6514c83523e4de313d645eefd9604326474da0cd2999b4: Status 404 returned error can't find the container with id fbb5559900dda2fcaa6514c83523e4de313d645eefd9604326474da0cd2999b4 Dec 13 06:50:53 crc kubenswrapper[4971]: I1213 06:50:53.304266 4971 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-jtmld" event={"ID":"2f42c147-5c5e-415c-9b62-faa800eb8a1d","Type":"ContainerStarted","Data":"00f236bb2dc65a5710643d8d01b4bfbb39e6699f30ef07ea9c55128df4ba0d74"} Dec 13 06:50:53 crc kubenswrapper[4971]: I1213 06:50:53.304333 4971 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-jtmld" event={"ID":"2f42c147-5c5e-415c-9b62-faa800eb8a1d","Type":"ContainerStarted","Data":"fbb5559900dda2fcaa6514c83523e4de313d645eefd9604326474da0cd2999b4"} Dec 13 06:50:53 crc kubenswrapper[4971]: I1213 06:50:53.323309 4971 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-jtmld" podStartSLOduration=88.323291785 podStartE2EDuration="1m28.323291785s" podCreationTimestamp="2025-12-13 06:49:25 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-13 06:50:53.323200552 +0000 UTC m=+109.927610020" watchObservedRunningTime="2025-12-13 06:50:53.323291785 +0000 UTC m=+109.927701233" Dec 13 06:50:53 crc kubenswrapper[4971]: I1213 06:50:53.767593 4971 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 13 06:50:53 crc kubenswrapper[4971]: I1213 06:50:53.767643 4971 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 13 06:50:53 crc kubenswrapper[4971]: I1213 06:50:53.767594 4971 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-x98k7" Dec 13 06:50:53 crc kubenswrapper[4971]: I1213 06:50:53.767651 4971 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 13 06:50:53 crc kubenswrapper[4971]: E1213 06:50:53.769062 4971 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 13 06:50:53 crc kubenswrapper[4971]: E1213 06:50:53.769236 4971 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-x98k7" podUID="c7185dec-3240-42e8-9d90-6182a5e8b65c" Dec 13 06:50:53 crc kubenswrapper[4971]: E1213 06:50:53.769321 4971 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 13 06:50:53 crc kubenswrapper[4971]: E1213 06:50:53.769387 4971 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 13 06:50:55 crc kubenswrapper[4971]: I1213 06:50:55.768614 4971 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 13 06:50:55 crc kubenswrapper[4971]: E1213 06:50:55.768739 4971 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 13 06:50:55 crc kubenswrapper[4971]: I1213 06:50:55.768824 4971 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 13 06:50:55 crc kubenswrapper[4971]: E1213 06:50:55.768989 4971 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 13 06:50:55 crc kubenswrapper[4971]: I1213 06:50:55.769077 4971 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-x98k7" Dec 13 06:50:55 crc kubenswrapper[4971]: E1213 06:50:55.769309 4971 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-x98k7" podUID="c7185dec-3240-42e8-9d90-6182a5e8b65c" Dec 13 06:50:55 crc kubenswrapper[4971]: I1213 06:50:55.769359 4971 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 13 06:50:55 crc kubenswrapper[4971]: E1213 06:50:55.769438 4971 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 13 06:50:57 crc kubenswrapper[4971]: I1213 06:50:57.767404 4971 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-x98k7" Dec 13 06:50:57 crc kubenswrapper[4971]: I1213 06:50:57.767447 4971 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 13 06:50:57 crc kubenswrapper[4971]: I1213 06:50:57.767451 4971 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 13 06:50:57 crc kubenswrapper[4971]: E1213 06:50:57.767573 4971 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-x98k7" podUID="c7185dec-3240-42e8-9d90-6182a5e8b65c" Dec 13 06:50:57 crc kubenswrapper[4971]: I1213 06:50:57.767604 4971 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 13 06:50:57 crc kubenswrapper[4971]: E1213 06:50:57.767715 4971 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 13 06:50:57 crc kubenswrapper[4971]: E1213 06:50:57.767759 4971 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 13 06:50:57 crc kubenswrapper[4971]: E1213 06:50:57.767811 4971 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 13 06:50:59 crc kubenswrapper[4971]: I1213 06:50:59.321411 4971 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-multus_multus-wwvpl_dddf744b-2122-49ad-a980-105c09636e0f/kube-multus/1.log" Dec 13 06:50:59 crc kubenswrapper[4971]: I1213 06:50:59.321835 4971 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-multus_multus-wwvpl_dddf744b-2122-49ad-a980-105c09636e0f/kube-multus/0.log" Dec 13 06:50:59 crc kubenswrapper[4971]: I1213 06:50:59.321881 4971 generic.go:334] "Generic (PLEG): container finished" podID="dddf744b-2122-49ad-a980-105c09636e0f" containerID="c8680aaa9e05095ac7e467c896d9e927bccff908925768ce5ed469ab89a72475" exitCode=1 Dec 13 06:50:59 crc kubenswrapper[4971]: I1213 06:50:59.321911 4971 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-wwvpl" event={"ID":"dddf744b-2122-49ad-a980-105c09636e0f","Type":"ContainerDied","Data":"c8680aaa9e05095ac7e467c896d9e927bccff908925768ce5ed469ab89a72475"} Dec 13 06:50:59 crc kubenswrapper[4971]: I1213 06:50:59.321947 4971 scope.go:117] "RemoveContainer" containerID="4676d897c7fc668587faec563f3551a056557879cee3e865192d2bd28edf62fc" Dec 13 06:50:59 crc kubenswrapper[4971]: I1213 06:50:59.322330 4971 scope.go:117] "RemoveContainer" containerID="c8680aaa9e05095ac7e467c896d9e927bccff908925768ce5ed469ab89a72475" Dec 13 06:50:59 crc kubenswrapper[4971]: E1213 06:50:59.322491 4971 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"kube-multus\" with CrashLoopBackOff: \"back-off 10s restarting failed container=kube-multus pod=multus-wwvpl_openshift-multus(dddf744b-2122-49ad-a980-105c09636e0f)\"" pod="openshift-multus/multus-wwvpl" podUID="dddf744b-2122-49ad-a980-105c09636e0f" Dec 13 06:50:59 crc kubenswrapper[4971]: I1213 06:50:59.767468 4971 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 13 06:50:59 crc kubenswrapper[4971]: I1213 06:50:59.767530 4971 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-x98k7" Dec 13 06:50:59 crc kubenswrapper[4971]: I1213 06:50:59.767504 4971 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 13 06:50:59 crc kubenswrapper[4971]: I1213 06:50:59.767643 4971 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 13 06:50:59 crc kubenswrapper[4971]: E1213 06:50:59.767643 4971 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 13 06:50:59 crc kubenswrapper[4971]: E1213 06:50:59.767742 4971 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 13 06:50:59 crc kubenswrapper[4971]: E1213 06:50:59.767922 4971 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 13 06:50:59 crc kubenswrapper[4971]: E1213 06:50:59.768005 4971 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-x98k7" podUID="c7185dec-3240-42e8-9d90-6182a5e8b65c" Dec 13 06:51:00 crc kubenswrapper[4971]: I1213 06:51:00.331292 4971 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-multus_multus-wwvpl_dddf744b-2122-49ad-a980-105c09636e0f/kube-multus/1.log" Dec 13 06:51:01 crc kubenswrapper[4971]: I1213 06:51:01.768412 4971 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-x98k7" Dec 13 06:51:01 crc kubenswrapper[4971]: I1213 06:51:01.768699 4971 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 13 06:51:01 crc kubenswrapper[4971]: I1213 06:51:01.768870 4971 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 13 06:51:01 crc kubenswrapper[4971]: I1213 06:51:01.768894 4971 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 13 06:51:01 crc kubenswrapper[4971]: E1213 06:51:01.769927 4971 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-x98k7" podUID="c7185dec-3240-42e8-9d90-6182a5e8b65c" Dec 13 06:51:01 crc kubenswrapper[4971]: E1213 06:51:01.770046 4971 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 13 06:51:01 crc kubenswrapper[4971]: E1213 06:51:01.770133 4971 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 13 06:51:01 crc kubenswrapper[4971]: E1213 06:51:01.770186 4971 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 13 06:51:02 crc kubenswrapper[4971]: I1213 06:51:02.768671 4971 scope.go:117] "RemoveContainer" containerID="7dd9913a65acba73ae1f32f4398207c5e276a1e979d86f9e720e4fd761bae4a3" Dec 13 06:51:02 crc kubenswrapper[4971]: E1213 06:51:02.768829 4971 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"ovnkube-controller\" with CrashLoopBackOff: \"back-off 40s restarting failed container=ovnkube-controller pod=ovnkube-node-zlcg5_openshift-ovn-kubernetes(736c6b1a-93dd-4f71-81b6-1c43fcf5f556)\"" pod="openshift-ovn-kubernetes/ovnkube-node-zlcg5" podUID="736c6b1a-93dd-4f71-81b6-1c43fcf5f556" Dec 13 06:51:03 crc kubenswrapper[4971]: E1213 06:51:03.677422 4971 kubelet_node_status.go:497] "Node not becoming ready in time after startup" Dec 13 06:51:03 crc kubenswrapper[4971]: I1213 06:51:03.768435 4971 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-x98k7" Dec 13 06:51:03 crc kubenswrapper[4971]: I1213 06:51:03.768557 4971 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 13 06:51:03 crc kubenswrapper[4971]: I1213 06:51:03.768619 4971 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 13 06:51:03 crc kubenswrapper[4971]: E1213 06:51:03.769721 4971 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-x98k7" podUID="c7185dec-3240-42e8-9d90-6182a5e8b65c" Dec 13 06:51:03 crc kubenswrapper[4971]: E1213 06:51:03.769814 4971 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 13 06:51:03 crc kubenswrapper[4971]: E1213 06:51:03.769936 4971 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 13 06:51:03 crc kubenswrapper[4971]: I1213 06:51:03.770228 4971 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 13 06:51:03 crc kubenswrapper[4971]: E1213 06:51:03.770446 4971 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 13 06:51:03 crc kubenswrapper[4971]: E1213 06:51:03.870128 4971 kubelet.go:2916] "Container runtime network not ready" networkReady="NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" Dec 13 06:51:05 crc kubenswrapper[4971]: I1213 06:51:05.768174 4971 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 13 06:51:05 crc kubenswrapper[4971]: E1213 06:51:05.768278 4971 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 13 06:51:05 crc kubenswrapper[4971]: I1213 06:51:05.768420 4971 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 13 06:51:05 crc kubenswrapper[4971]: E1213 06:51:05.768462 4971 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 13 06:51:05 crc kubenswrapper[4971]: I1213 06:51:05.768580 4971 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 13 06:51:05 crc kubenswrapper[4971]: E1213 06:51:05.768643 4971 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 13 06:51:05 crc kubenswrapper[4971]: I1213 06:51:05.768736 4971 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-x98k7" Dec 13 06:51:05 crc kubenswrapper[4971]: E1213 06:51:05.768808 4971 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-x98k7" podUID="c7185dec-3240-42e8-9d90-6182a5e8b65c" Dec 13 06:51:07 crc kubenswrapper[4971]: I1213 06:51:07.767653 4971 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 13 06:51:07 crc kubenswrapper[4971]: I1213 06:51:07.767663 4971 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 13 06:51:07 crc kubenswrapper[4971]: E1213 06:51:07.767833 4971 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 13 06:51:07 crc kubenswrapper[4971]: I1213 06:51:07.767967 4971 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 13 06:51:07 crc kubenswrapper[4971]: I1213 06:51:07.767679 4971 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-x98k7" Dec 13 06:51:07 crc kubenswrapper[4971]: E1213 06:51:07.768046 4971 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 13 06:51:07 crc kubenswrapper[4971]: E1213 06:51:07.768119 4971 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 13 06:51:07 crc kubenswrapper[4971]: E1213 06:51:07.768297 4971 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-x98k7" podUID="c7185dec-3240-42e8-9d90-6182a5e8b65c" Dec 13 06:51:08 crc kubenswrapper[4971]: E1213 06:51:08.872626 4971 kubelet.go:2916] "Container runtime network not ready" networkReady="NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" Dec 13 06:51:09 crc kubenswrapper[4971]: I1213 06:51:09.767402 4971 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 13 06:51:09 crc kubenswrapper[4971]: I1213 06:51:09.767448 4971 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-x98k7" Dec 13 06:51:09 crc kubenswrapper[4971]: I1213 06:51:09.767471 4971 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 13 06:51:09 crc kubenswrapper[4971]: I1213 06:51:09.767548 4971 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 13 06:51:09 crc kubenswrapper[4971]: E1213 06:51:09.767554 4971 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 13 06:51:09 crc kubenswrapper[4971]: E1213 06:51:09.767625 4971 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 13 06:51:09 crc kubenswrapper[4971]: E1213 06:51:09.767811 4971 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 13 06:51:09 crc kubenswrapper[4971]: E1213 06:51:09.767904 4971 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-x98k7" podUID="c7185dec-3240-42e8-9d90-6182a5e8b65c" Dec 13 06:51:11 crc kubenswrapper[4971]: I1213 06:51:11.767407 4971 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-x98k7" Dec 13 06:51:11 crc kubenswrapper[4971]: E1213 06:51:11.767908 4971 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-x98k7" podUID="c7185dec-3240-42e8-9d90-6182a5e8b65c" Dec 13 06:51:11 crc kubenswrapper[4971]: I1213 06:51:11.767413 4971 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 13 06:51:11 crc kubenswrapper[4971]: E1213 06:51:11.767991 4971 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 13 06:51:11 crc kubenswrapper[4971]: I1213 06:51:11.767435 4971 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 13 06:51:11 crc kubenswrapper[4971]: E1213 06:51:11.768045 4971 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 13 06:51:11 crc kubenswrapper[4971]: I1213 06:51:11.767566 4971 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 13 06:51:11 crc kubenswrapper[4971]: E1213 06:51:11.768106 4971 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 13 06:51:13 crc kubenswrapper[4971]: I1213 06:51:13.768253 4971 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 13 06:51:13 crc kubenswrapper[4971]: I1213 06:51:13.768321 4971 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-x98k7" Dec 13 06:51:13 crc kubenswrapper[4971]: E1213 06:51:13.769553 4971 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 13 06:51:13 crc kubenswrapper[4971]: I1213 06:51:13.769621 4971 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 13 06:51:13 crc kubenswrapper[4971]: I1213 06:51:13.769678 4971 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 13 06:51:13 crc kubenswrapper[4971]: E1213 06:51:13.769817 4971 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-x98k7" podUID="c7185dec-3240-42e8-9d90-6182a5e8b65c" Dec 13 06:51:13 crc kubenswrapper[4971]: E1213 06:51:13.770012 4971 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 13 06:51:13 crc kubenswrapper[4971]: E1213 06:51:13.770162 4971 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 13 06:51:13 crc kubenswrapper[4971]: E1213 06:51:13.873542 4971 kubelet.go:2916] "Container runtime network not ready" networkReady="NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" Dec 13 06:51:14 crc kubenswrapper[4971]: I1213 06:51:14.768898 4971 scope.go:117] "RemoveContainer" containerID="c8680aaa9e05095ac7e467c896d9e927bccff908925768ce5ed469ab89a72475" Dec 13 06:51:15 crc kubenswrapper[4971]: I1213 06:51:15.373416 4971 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-multus_multus-wwvpl_dddf744b-2122-49ad-a980-105c09636e0f/kube-multus/1.log" Dec 13 06:51:15 crc kubenswrapper[4971]: I1213 06:51:15.373458 4971 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-wwvpl" event={"ID":"dddf744b-2122-49ad-a980-105c09636e0f","Type":"ContainerStarted","Data":"b02826c7d90e9a7a2ed0fa1dd785f2690d152ec880d5eee474ddd4d2d89c8f75"} Dec 13 06:51:15 crc kubenswrapper[4971]: I1213 06:51:15.767629 4971 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 13 06:51:15 crc kubenswrapper[4971]: I1213 06:51:15.767691 4971 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 13 06:51:15 crc kubenswrapper[4971]: I1213 06:51:15.767705 4971 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 13 06:51:15 crc kubenswrapper[4971]: E1213 06:51:15.767764 4971 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 13 06:51:15 crc kubenswrapper[4971]: I1213 06:51:15.767780 4971 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-x98k7" Dec 13 06:51:15 crc kubenswrapper[4971]: E1213 06:51:15.767879 4971 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 13 06:51:15 crc kubenswrapper[4971]: E1213 06:51:15.767955 4971 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-x98k7" podUID="c7185dec-3240-42e8-9d90-6182a5e8b65c" Dec 13 06:51:15 crc kubenswrapper[4971]: E1213 06:51:15.768044 4971 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 13 06:51:17 crc kubenswrapper[4971]: I1213 06:51:17.767953 4971 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-x98k7" Dec 13 06:51:17 crc kubenswrapper[4971]: I1213 06:51:17.768016 4971 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 13 06:51:17 crc kubenswrapper[4971]: E1213 06:51:17.768138 4971 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-x98k7" podUID="c7185dec-3240-42e8-9d90-6182a5e8b65c" Dec 13 06:51:17 crc kubenswrapper[4971]: I1213 06:51:17.767973 4971 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 13 06:51:17 crc kubenswrapper[4971]: I1213 06:51:17.768812 4971 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 13 06:51:17 crc kubenswrapper[4971]: E1213 06:51:17.768918 4971 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 13 06:51:17 crc kubenswrapper[4971]: E1213 06:51:17.769018 4971 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 13 06:51:17 crc kubenswrapper[4971]: E1213 06:51:17.769171 4971 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 13 06:51:17 crc kubenswrapper[4971]: I1213 06:51:17.769577 4971 scope.go:117] "RemoveContainer" containerID="7dd9913a65acba73ae1f32f4398207c5e276a1e979d86f9e720e4fd761bae4a3" Dec 13 06:51:18 crc kubenswrapper[4971]: I1213 06:51:18.385290 4971 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-zlcg5_736c6b1a-93dd-4f71-81b6-1c43fcf5f556/ovnkube-controller/3.log" Dec 13 06:51:18 crc kubenswrapper[4971]: I1213 06:51:18.387510 4971 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-zlcg5" event={"ID":"736c6b1a-93dd-4f71-81b6-1c43fcf5f556","Type":"ContainerStarted","Data":"6afaff46af01dcc99e08b76967fe4181e615cd044389b9b8cea1581505eef247"} Dec 13 06:51:18 crc kubenswrapper[4971]: I1213 06:51:18.387933 4971 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-ovn-kubernetes/ovnkube-node-zlcg5" Dec 13 06:51:18 crc kubenswrapper[4971]: I1213 06:51:18.427264 4971 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-ovn-kubernetes/ovnkube-node-zlcg5" podStartSLOduration=113.427242322 podStartE2EDuration="1m53.427242322s" podCreationTimestamp="2025-12-13 06:49:25 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-13 06:51:18.422191794 +0000 UTC m=+135.026601263" watchObservedRunningTime="2025-12-13 06:51:18.427242322 +0000 UTC m=+135.031651780" Dec 13 06:51:18 crc kubenswrapper[4971]: I1213 06:51:18.534284 4971 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-multus/network-metrics-daemon-x98k7"] Dec 13 06:51:18 crc kubenswrapper[4971]: I1213 06:51:18.534414 4971 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-x98k7" Dec 13 06:51:18 crc kubenswrapper[4971]: E1213 06:51:18.534576 4971 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-x98k7" podUID="c7185dec-3240-42e8-9d90-6182a5e8b65c" Dec 13 06:51:18 crc kubenswrapper[4971]: E1213 06:51:18.876108 4971 kubelet.go:2916] "Container runtime network not ready" networkReady="NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" Dec 13 06:51:19 crc kubenswrapper[4971]: I1213 06:51:19.768219 4971 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 13 06:51:19 crc kubenswrapper[4971]: E1213 06:51:19.768350 4971 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 13 06:51:19 crc kubenswrapper[4971]: I1213 06:51:19.768362 4971 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 13 06:51:19 crc kubenswrapper[4971]: I1213 06:51:19.768536 4971 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-x98k7" Dec 13 06:51:19 crc kubenswrapper[4971]: E1213 06:51:19.768617 4971 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-x98k7" podUID="c7185dec-3240-42e8-9d90-6182a5e8b65c" Dec 13 06:51:19 crc kubenswrapper[4971]: I1213 06:51:19.768628 4971 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 13 06:51:19 crc kubenswrapper[4971]: E1213 06:51:19.768827 4971 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 13 06:51:19 crc kubenswrapper[4971]: E1213 06:51:19.768879 4971 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 13 06:51:21 crc kubenswrapper[4971]: I1213 06:51:21.768158 4971 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-x98k7" Dec 13 06:51:21 crc kubenswrapper[4971]: I1213 06:51:21.768205 4971 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 13 06:51:21 crc kubenswrapper[4971]: I1213 06:51:21.768183 4971 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 13 06:51:21 crc kubenswrapper[4971]: I1213 06:51:21.768300 4971 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 13 06:51:21 crc kubenswrapper[4971]: E1213 06:51:21.768388 4971 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-x98k7" podUID="c7185dec-3240-42e8-9d90-6182a5e8b65c" Dec 13 06:51:21 crc kubenswrapper[4971]: E1213 06:51:21.768583 4971 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 13 06:51:21 crc kubenswrapper[4971]: E1213 06:51:21.768639 4971 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 13 06:51:21 crc kubenswrapper[4971]: E1213 06:51:21.768812 4971 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 13 06:51:23 crc kubenswrapper[4971]: I1213 06:51:23.768205 4971 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-x98k7" Dec 13 06:51:23 crc kubenswrapper[4971]: E1213 06:51:23.769692 4971 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-x98k7" podUID="c7185dec-3240-42e8-9d90-6182a5e8b65c" Dec 13 06:51:23 crc kubenswrapper[4971]: I1213 06:51:23.769884 4971 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 13 06:51:23 crc kubenswrapper[4971]: I1213 06:51:23.769927 4971 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 13 06:51:23 crc kubenswrapper[4971]: E1213 06:51:23.769985 4971 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 13 06:51:23 crc kubenswrapper[4971]: I1213 06:51:23.770116 4971 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 13 06:51:23 crc kubenswrapper[4971]: E1213 06:51:23.770174 4971 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 13 06:51:23 crc kubenswrapper[4971]: E1213 06:51:23.770315 4971 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 13 06:51:25 crc kubenswrapper[4971]: I1213 06:51:25.768121 4971 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 13 06:51:25 crc kubenswrapper[4971]: I1213 06:51:25.768249 4971 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 13 06:51:25 crc kubenswrapper[4971]: I1213 06:51:25.768249 4971 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 13 06:51:25 crc kubenswrapper[4971]: I1213 06:51:25.768287 4971 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-x98k7" Dec 13 06:51:25 crc kubenswrapper[4971]: I1213 06:51:25.770630 4971 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-console"/"networking-console-plugin" Dec 13 06:51:25 crc kubenswrapper[4971]: I1213 06:51:25.770690 4971 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-diagnostics"/"kube-root-ca.crt" Dec 13 06:51:25 crc kubenswrapper[4971]: I1213 06:51:25.770736 4971 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-multus"/"metrics-daemon-sa-dockercfg-d427c" Dec 13 06:51:25 crc kubenswrapper[4971]: I1213 06:51:25.770745 4971 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-diagnostics"/"openshift-service-ca.crt" Dec 13 06:51:25 crc kubenswrapper[4971]: I1213 06:51:25.771229 4971 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-network-console"/"networking-console-plugin-cert" Dec 13 06:51:25 crc kubenswrapper[4971]: I1213 06:51:25.771276 4971 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-multus"/"metrics-daemon-secret" Dec 13 06:51:31 crc kubenswrapper[4971]: I1213 06:51:31.626319 4971 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 13 06:51:31 crc kubenswrapper[4971]: E1213 06:51:31.626485 4971 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-13 06:53:33.626467988 +0000 UTC m=+270.230877436 (durationBeforeRetry 2m2s). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 13 06:51:31 crc kubenswrapper[4971]: I1213 06:51:31.627024 4971 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"networking-console-plugin-cert\" (UniqueName: \"kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 13 06:51:31 crc kubenswrapper[4971]: I1213 06:51:31.627049 4971 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nginx-conf\" (UniqueName: \"kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 13 06:51:31 crc kubenswrapper[4971]: I1213 06:51:31.628114 4971 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"nginx-conf\" (UniqueName: \"kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 13 06:51:31 crc kubenswrapper[4971]: I1213 06:51:31.633089 4971 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"networking-console-plugin-cert\" (UniqueName: \"kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 13 06:51:31 crc kubenswrapper[4971]: I1213 06:51:31.727570 4971 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-s2dwl\" (UniqueName: \"kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl\") pod \"network-check-source-55646444c4-trplf\" (UID: \"9d751cbb-f2e2-430d-9754-c882a5e924a5\") " pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 13 06:51:31 crc kubenswrapper[4971]: I1213 06:51:31.727626 4971 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cqllr\" (UniqueName: \"kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr\") pod \"network-check-target-xd92c\" (UID: \"3b6479f0-333b-4a96-9adf-2099afdc2447\") " pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 13 06:51:31 crc kubenswrapper[4971]: I1213 06:51:31.730470 4971 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-cqllr\" (UniqueName: \"kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr\") pod \"network-check-target-xd92c\" (UID: \"3b6479f0-333b-4a96-9adf-2099afdc2447\") " pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 13 06:51:31 crc kubenswrapper[4971]: I1213 06:51:31.731036 4971 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-s2dwl\" (UniqueName: \"kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl\") pod \"network-check-source-55646444c4-trplf\" (UID: \"9d751cbb-f2e2-430d-9754-c882a5e924a5\") " pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 13 06:51:31 crc kubenswrapper[4971]: I1213 06:51:31.780495 4971 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 13 06:51:31 crc kubenswrapper[4971]: I1213 06:51:31.788086 4971 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 13 06:51:31 crc kubenswrapper[4971]: I1213 06:51:31.800559 4971 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 13 06:51:32 crc kubenswrapper[4971]: W1213 06:51:32.234690 4971 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod3b6479f0_333b_4a96_9adf_2099afdc2447.slice/crio-9abc52505a915f5a06be85412de0badbff7c84f43fb05cc4ad085ba4f0197459 WatchSource:0}: Error finding container 9abc52505a915f5a06be85412de0badbff7c84f43fb05cc4ad085ba4f0197459: Status 404 returned error can't find the container with id 9abc52505a915f5a06be85412de0badbff7c84f43fb05cc4ad085ba4f0197459 Dec 13 06:51:32 crc kubenswrapper[4971]: I1213 06:51:32.431741 4971 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" event={"ID":"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8","Type":"ContainerStarted","Data":"1f869d783c903dcda97c77b00ece76671afe591d7369b6a78c3655113192bcb0"} Dec 13 06:51:32 crc kubenswrapper[4971]: I1213 06:51:32.431799 4971 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" event={"ID":"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8","Type":"ContainerStarted","Data":"e54705ac752b890cc42d54e9fcf908e878f0a8295dc104234fc9bee1b53c1684"} Dec 13 06:51:32 crc kubenswrapper[4971]: I1213 06:51:32.434308 4971 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" event={"ID":"3b6479f0-333b-4a96-9adf-2099afdc2447","Type":"ContainerStarted","Data":"40655b0344d7c4fc4ee5a259d99e3cc21ddc5cf881e4ddb0eb6c6aca1e0c4bf2"} Dec 13 06:51:32 crc kubenswrapper[4971]: I1213 06:51:32.434360 4971 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" event={"ID":"3b6479f0-333b-4a96-9adf-2099afdc2447","Type":"ContainerStarted","Data":"9abc52505a915f5a06be85412de0badbff7c84f43fb05cc4ad085ba4f0197459"} Dec 13 06:51:32 crc kubenswrapper[4971]: I1213 06:51:32.434537 4971 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 13 06:51:32 crc kubenswrapper[4971]: I1213 06:51:32.436342 4971 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" event={"ID":"9d751cbb-f2e2-430d-9754-c882a5e924a5","Type":"ContainerStarted","Data":"2a6ded911fdd4bfe620abdb113ba9edcb49233c17fd16f10cbe0c3d503943e24"} Dec 13 06:51:32 crc kubenswrapper[4971]: I1213 06:51:32.436368 4971 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" event={"ID":"9d751cbb-f2e2-430d-9754-c882a5e924a5","Type":"ContainerStarted","Data":"efb0807e0bfc7d0af3c0b07b0c757189801f41397baf87241918b132c2f4db8c"} Dec 13 06:51:33 crc kubenswrapper[4971]: I1213 06:51:33.184320 4971 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeReady" Dec 13 06:51:33 crc kubenswrapper[4971]: I1213 06:51:33.221117 4971 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-cluster-machine-approver/machine-approver-56656f9798-s75fl"] Dec 13 06:51:33 crc kubenswrapper[4971]: I1213 06:51:33.221679 4971 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-cluster-machine-approver/machine-approver-56656f9798-s75fl" Dec 13 06:51:33 crc kubenswrapper[4971]: I1213 06:51:33.225161 4971 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-cluster-machine-approver"/"openshift-service-ca.crt" Dec 13 06:51:33 crc kubenswrapper[4971]: I1213 06:51:33.225205 4971 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-cluster-machine-approver"/"kube-rbac-proxy" Dec 13 06:51:33 crc kubenswrapper[4971]: I1213 06:51:33.225205 4971 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-cluster-machine-approver"/"machine-approver-sa-dockercfg-nl2j4" Dec 13 06:51:33 crc kubenswrapper[4971]: I1213 06:51:33.225210 4971 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-cluster-machine-approver"/"machine-approver-tls" Dec 13 06:51:33 crc kubenswrapper[4971]: I1213 06:51:33.225862 4971 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-cluster-machine-approver"/"machine-approver-config" Dec 13 06:51:33 crc kubenswrapper[4971]: I1213 06:51:33.225952 4971 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-cluster-machine-approver"/"kube-root-ca.crt" Dec 13 06:51:33 crc kubenswrapper[4971]: I1213 06:51:33.226203 4971 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-oauth-apiserver/apiserver-7bbb656c7d-mb2q7"] Dec 13 06:51:33 crc kubenswrapper[4971]: I1213 06:51:33.226735 4971 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-mb2q7" Dec 13 06:51:33 crc kubenswrapper[4971]: I1213 06:51:33.229929 4971 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-machine-api/machine-api-operator-5694c8668f-t24dm"] Dec 13 06:51:33 crc kubenswrapper[4971]: I1213 06:51:33.244336 4971 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-oauth-apiserver"/"etcd-serving-ca" Dec 13 06:51:33 crc kubenswrapper[4971]: I1213 06:51:33.244588 4971 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-oauth-apiserver"/"etcd-client" Dec 13 06:51:33 crc kubenswrapper[4971]: I1213 06:51:33.244753 4971 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-oauth-apiserver"/"oauth-apiserver-sa-dockercfg-6r2bq" Dec 13 06:51:33 crc kubenswrapper[4971]: I1213 06:51:33.244923 4971 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-oauth-apiserver"/"openshift-service-ca.crt" Dec 13 06:51:33 crc kubenswrapper[4971]: I1213 06:51:33.245050 4971 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-oauth-apiserver"/"trusted-ca-bundle" Dec 13 06:51:33 crc kubenswrapper[4971]: I1213 06:51:33.245176 4971 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-oauth-apiserver"/"kube-root-ca.crt" Dec 13 06:51:33 crc kubenswrapper[4971]: I1213 06:51:33.245338 4971 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-oauth-apiserver"/"audit-1" Dec 13 06:51:33 crc kubenswrapper[4971]: I1213 06:51:33.245545 4971 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-oauth-apiserver"/"serving-cert" Dec 13 06:51:33 crc kubenswrapper[4971]: I1213 06:51:33.245702 4971 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-oauth-apiserver"/"encryption-config-1" Dec 13 06:51:33 crc kubenswrapper[4971]: I1213 06:51:33.245939 4971 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-machine-api/machine-api-operator-5694c8668f-t24dm" Dec 13 06:51:33 crc kubenswrapper[4971]: I1213 06:51:33.253434 4971 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-7ddbb"] Dec 13 06:51:33 crc kubenswrapper[4971]: I1213 06:51:33.262886 4971 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-api"/"machine-api-operator-dockercfg-mfbb7" Dec 13 06:51:33 crc kubenswrapper[4971]: I1213 06:51:33.263046 4971 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-machine-api"/"kube-rbac-proxy" Dec 13 06:51:33 crc kubenswrapper[4971]: I1213 06:51:33.263131 4971 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-machine-api"/"machine-api-operator-images" Dec 13 06:51:33 crc kubenswrapper[4971]: I1213 06:51:33.263253 4971 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-machine-api"/"kube-root-ca.crt" Dec 13 06:51:33 crc kubenswrapper[4971]: I1213 06:51:33.263302 4971 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-api"/"machine-api-operator-tls" Dec 13 06:51:33 crc kubenswrapper[4971]: I1213 06:51:33.263415 4971 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-machine-api"/"openshift-service-ca.crt" Dec 13 06:51:33 crc kubenswrapper[4971]: I1213 06:51:33.263551 4971 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-7ddbb" Dec 13 06:51:33 crc kubenswrapper[4971]: I1213 06:51:33.263706 4971 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-mlbwt"] Dec 13 06:51:33 crc kubenswrapper[4971]: I1213 06:51:33.265238 4971 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-mlbwt" Dec 13 06:51:33 crc kubenswrapper[4971]: I1213 06:51:33.265733 4971 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-apiserver/apiserver-76f77b778f-2n4cq"] Dec 13 06:51:33 crc kubenswrapper[4971]: I1213 06:51:33.266287 4971 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-apiserver/apiserver-76f77b778f-2n4cq" Dec 13 06:51:33 crc kubenswrapper[4971]: I1213 06:51:33.266833 4971 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-zqt2t"] Dec 13 06:51:33 crc kubenswrapper[4971]: I1213 06:51:33.267252 4971 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-zqt2t" Dec 13 06:51:33 crc kubenswrapper[4971]: I1213 06:51:33.268144 4971 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver-operator"/"openshift-apiserver-operator-config" Dec 13 06:51:33 crc kubenswrapper[4971]: I1213 06:51:33.268331 4971 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-apiserver-operator"/"openshift-apiserver-operator-dockercfg-xtcjv" Dec 13 06:51:33 crc kubenswrapper[4971]: I1213 06:51:33.268637 4971 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-cluster-samples-operator"/"samples-operator-tls" Dec 13 06:51:33 crc kubenswrapper[4971]: I1213 06:51:33.268822 4971 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-apiserver-operator"/"openshift-apiserver-operator-serving-cert" Dec 13 06:51:33 crc kubenswrapper[4971]: I1213 06:51:33.268935 4971 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver-operator"/"kube-root-ca.crt" Dec 13 06:51:33 crc kubenswrapper[4971]: I1213 06:51:33.269020 4971 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver-operator"/"openshift-service-ca.crt" Dec 13 06:51:33 crc kubenswrapper[4971]: I1213 06:51:33.269105 4971 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-cluster-samples-operator"/"cluster-samples-operator-dockercfg-xpp9w" Dec 13 06:51:33 crc kubenswrapper[4971]: I1213 06:51:33.269616 4971 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-apiserver"/"openshift-apiserver-sa-dockercfg-djjff" Dec 13 06:51:33 crc kubenswrapper[4971]: I1213 06:51:33.269850 4971 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-apiserver"/"serving-cert" Dec 13 06:51:33 crc kubenswrapper[4971]: I1213 06:51:33.270011 4971 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-image-registry"/"cluster-image-registry-operator-dockercfg-m4qtx" Dec 13 06:51:33 crc kubenswrapper[4971]: I1213 06:51:33.270214 4971 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-image-registry"/"image-registry-operator-tls" Dec 13 06:51:33 crc kubenswrapper[4971]: I1213 06:51:33.270372 4971 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-apiserver"/"encryption-config-1" Dec 13 06:51:33 crc kubenswrapper[4971]: I1213 06:51:33.270586 4971 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-apiserver"/"etcd-client" Dec 13 06:51:33 crc kubenswrapper[4971]: I1213 06:51:33.274467 4971 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-controller-manager/controller-manager-879f6c89f-xtxr8"] Dec 13 06:51:33 crc kubenswrapper[4971]: I1213 06:51:33.275001 4971 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-console/console-f9d7485db-5fd9b"] Dec 13 06:51:33 crc kubenswrapper[4971]: I1213 06:51:33.275420 4971 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-console/console-f9d7485db-5fd9b" Dec 13 06:51:33 crc kubenswrapper[4971]: I1213 06:51:33.275833 4971 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-controller-manager/controller-manager-879f6c89f-xtxr8" Dec 13 06:51:33 crc kubenswrapper[4971]: I1213 06:51:33.277170 4971 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-authentication/oauth-openshift-558db77b4-nzzbm"] Dec 13 06:51:33 crc kubenswrapper[4971]: I1213 06:51:33.277723 4971 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-authentication/oauth-openshift-558db77b4-nzzbm" Dec 13 06:51:33 crc kubenswrapper[4971]: I1213 06:51:33.280046 4971 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-cluster-samples-operator"/"kube-root-ca.crt" Dec 13 06:51:33 crc kubenswrapper[4971]: I1213 06:51:33.280315 4971 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver"/"audit-1" Dec 13 06:51:33 crc kubenswrapper[4971]: I1213 06:51:33.280494 4971 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-cluster-samples-operator"/"openshift-service-ca.crt" Dec 13 06:51:33 crc kubenswrapper[4971]: I1213 06:51:33.282752 4971 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-config-operator/openshift-config-operator-7777fb866f-vmdsp"] Dec 13 06:51:33 crc kubenswrapper[4971]: I1213 06:51:33.283462 4971 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-config-operator/openshift-config-operator-7777fb866f-vmdsp" Dec 13 06:51:33 crc kubenswrapper[4971]: I1213 06:51:33.283724 4971 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console"/"openshift-service-ca.crt" Dec 13 06:51:33 crc kubenswrapper[4971]: I1213 06:51:33.285278 4971 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver"/"etcd-serving-ca" Dec 13 06:51:33 crc kubenswrapper[4971]: I1213 06:51:33.285494 4971 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver"/"image-import-ca" Dec 13 06:51:33 crc kubenswrapper[4971]: I1213 06:51:33.285699 4971 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver"/"openshift-service-ca.crt" Dec 13 06:51:33 crc kubenswrapper[4971]: I1213 06:51:33.285714 4971 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver"/"config" Dec 13 06:51:33 crc kubenswrapper[4971]: I1213 06:51:33.285970 4971 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"config" Dec 13 06:51:33 crc kubenswrapper[4971]: I1213 06:51:33.287269 4971 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-route-controller-manager/route-controller-manager-6576b87f9c-9d86n"] Dec 13 06:51:33 crc kubenswrapper[4971]: I1213 06:51:33.287866 4971 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-9d86n" Dec 13 06:51:33 crc kubenswrapper[4971]: I1213 06:51:33.289572 4971 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-console-operator/console-operator-58897d9998-kdfcd"] Dec 13 06:51:33 crc kubenswrapper[4971]: I1213 06:51:33.290139 4971 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-console-operator/console-operator-58897d9998-kdfcd" Dec 13 06:51:33 crc kubenswrapper[4971]: I1213 06:51:33.291021 4971 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-console/downloads-7954f5f757-dqk9z"] Dec 13 06:51:33 crc kubenswrapper[4971]: I1213 06:51:33.291552 4971 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-console/downloads-7954f5f757-dqk9z" Dec 13 06:51:33 crc kubenswrapper[4971]: I1213 06:51:33.293155 4971 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-system-session" Dec 13 06:51:33 crc kubenswrapper[4971]: I1213 06:51:33.293468 4971 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver"/"kube-root-ca.crt" Dec 13 06:51:33 crc kubenswrapper[4971]: I1213 06:51:33.293884 4971 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-console"/"console-oauth-config" Dec 13 06:51:33 crc kubenswrapper[4971]: I1213 06:51:33.294052 4971 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"openshift-service-ca.crt" Dec 13 06:51:33 crc kubenswrapper[4971]: I1213 06:51:33.294796 4971 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console-operator"/"openshift-service-ca.crt" Dec 13 06:51:33 crc kubenswrapper[4971]: I1213 06:51:33.295738 4971 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console"/"console-config" Dec 13 06:51:33 crc kubenswrapper[4971]: I1213 06:51:33.296422 4971 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-route-controller-manager"/"client-ca" Dec 13 06:51:33 crc kubenswrapper[4971]: I1213 06:51:33.296572 4971 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"oauth-openshift-dockercfg-znhcc" Dec 13 06:51:33 crc kubenswrapper[4971]: I1213 06:51:33.296708 4971 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-system-serving-cert" Dec 13 06:51:33 crc kubenswrapper[4971]: I1213 06:51:33.297585 4971 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-user-idp-0-file-data" Dec 13 06:51:33 crc kubenswrapper[4971]: I1213 06:51:33.297740 4971 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console"/"service-ca" Dec 13 06:51:33 crc kubenswrapper[4971]: I1213 06:51:33.297880 4971 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-controller-manager"/"serving-cert" Dec 13 06:51:33 crc kubenswrapper[4971]: I1213 06:51:33.298025 4971 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-console"/"console-serving-cert" Dec 13 06:51:33 crc kubenswrapper[4971]: I1213 06:51:33.298289 4971 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console"/"kube-root-ca.crt" Dec 13 06:51:33 crc kubenswrapper[4971]: I1213 06:51:33.298394 4971 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication"/"openshift-service-ca.crt" Dec 13 06:51:33 crc kubenswrapper[4971]: I1213 06:51:33.298505 4971 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication"/"audit" Dec 13 06:51:33 crc kubenswrapper[4971]: I1213 06:51:33.298627 4971 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"client-ca" Dec 13 06:51:33 crc kubenswrapper[4971]: I1213 06:51:33.298720 4971 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-user-template-provider-selection" Dec 13 06:51:33 crc kubenswrapper[4971]: I1213 06:51:33.298733 4971 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-route-controller-manager"/"kube-root-ca.crt" Dec 13 06:51:33 crc kubenswrapper[4971]: I1213 06:51:33.298824 4971 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-controller-manager"/"openshift-controller-manager-sa-dockercfg-msq4c" Dec 13 06:51:33 crc kubenswrapper[4971]: I1213 06:51:33.298840 4971 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-route-controller-manager"/"openshift-service-ca.crt" Dec 13 06:51:33 crc kubenswrapper[4971]: I1213 06:51:33.298909 4971 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-route-controller-manager"/"config" Dec 13 06:51:33 crc kubenswrapper[4971]: I1213 06:51:33.298956 4971 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console"/"oauth-serving-cert" Dec 13 06:51:33 crc kubenswrapper[4971]: I1213 06:51:33.299012 4971 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console-operator"/"console-operator-config" Dec 13 06:51:33 crc kubenswrapper[4971]: I1213 06:51:33.298913 4971 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-user-template-error" Dec 13 06:51:33 crc kubenswrapper[4971]: I1213 06:51:33.299153 4971 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-route-controller-manager"/"route-controller-manager-sa-dockercfg-h2zr2" Dec 13 06:51:33 crc kubenswrapper[4971]: I1213 06:51:33.299258 4971 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-system-router-certs" Dec 13 06:51:33 crc kubenswrapper[4971]: I1213 06:51:33.299357 4971 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication"/"v4-0-config-system-cliconfig" Dec 13 06:51:33 crc kubenswrapper[4971]: I1213 06:51:33.299452 4971 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication"/"v4-0-config-system-service-ca" Dec 13 06:51:33 crc kubenswrapper[4971]: I1213 06:51:33.299586 4971 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-config-operator"/"openshift-service-ca.crt" Dec 13 06:51:33 crc kubenswrapper[4971]: I1213 06:51:33.299743 4971 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-console-operator"/"console-operator-dockercfg-4xjcr" Dec 13 06:51:33 crc kubenswrapper[4971]: I1213 06:51:33.299797 4971 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication"/"kube-root-ca.crt" Dec 13 06:51:33 crc kubenswrapper[4971]: I1213 06:51:33.299830 4971 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-console"/"default-dockercfg-chnjx" Dec 13 06:51:33 crc kubenswrapper[4971]: I1213 06:51:33.299907 4971 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-config-operator"/"config-operator-serving-cert" Dec 13 06:51:33 crc kubenswrapper[4971]: I1213 06:51:33.299997 4971 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-route-controller-manager"/"serving-cert" Dec 13 06:51:33 crc kubenswrapper[4971]: I1213 06:51:33.300133 4971 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console-operator"/"kube-root-ca.crt" Dec 13 06:51:33 crc kubenswrapper[4971]: I1213 06:51:33.300185 4971 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"kube-root-ca.crt" Dec 13 06:51:33 crc kubenswrapper[4971]: I1213 06:51:33.300299 4971 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-config-operator"/"kube-root-ca.crt" Dec 13 06:51:33 crc kubenswrapper[4971]: I1213 06:51:33.300410 4971 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-console-operator"/"serving-cert" Dec 13 06:51:33 crc kubenswrapper[4971]: I1213 06:51:33.300465 4971 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-config-operator"/"openshift-config-operator-dockercfg-7pc5z" Dec 13 06:51:33 crc kubenswrapper[4971]: I1213 06:51:33.304276 4971 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-gtrwg"] Dec 13 06:51:33 crc kubenswrapper[4971]: I1213 06:51:33.304875 4971 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-gtrwg" Dec 13 06:51:33 crc kubenswrapper[4971]: I1213 06:51:33.309401 4971 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-console"/"console-dockercfg-f62pw" Dec 13 06:51:33 crc kubenswrapper[4971]: I1213 06:51:33.333925 4971 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-image-registry/image-registry-697d97f7c8-mdd8q"] Dec 13 06:51:33 crc kubenswrapper[4971]: I1213 06:51:33.354483 4971 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager-operator"/"openshift-service-ca.crt" Dec 13 06:51:33 crc kubenswrapper[4971]: I1213 06:51:33.354649 4971 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager-operator"/"openshift-controller-manager-operator-config" Dec 13 06:51:33 crc kubenswrapper[4971]: I1213 06:51:33.355833 4971 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-controller-manager-operator"/"openshift-controller-manager-operator-serving-cert" Dec 13 06:51:33 crc kubenswrapper[4971]: I1213 06:51:33.357102 4971 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-dns-operator/dns-operator-744455d44c-b999k"] Dec 13 06:51:33 crc kubenswrapper[4971]: I1213 06:51:33.357689 4971 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-etcd-operator/etcd-operator-b45778765-6xfwq"] Dec 13 06:51:33 crc kubenswrapper[4971]: I1213 06:51:33.357792 4971 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console"/"trusted-ca-bundle" Dec 13 06:51:33 crc kubenswrapper[4971]: I1213 06:51:33.357914 4971 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-image-registry/image-registry-697d97f7c8-mdd8q" Dec 13 06:51:33 crc kubenswrapper[4971]: I1213 06:51:33.358083 4971 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-ingress-operator/ingress-operator-5b745b69d9-pkmb5"] Dec 13 06:51:33 crc kubenswrapper[4971]: I1213 06:51:33.358194 4971 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/bec6e53b-3998-4402-b77b-87ff79f31827-trusted-ca-bundle\") pod \"apiserver-76f77b778f-2n4cq\" (UID: \"bec6e53b-3998-4402-b77b-87ff79f31827\") " pod="openshift-apiserver/apiserver-76f77b778f-2n4cq" Dec 13 06:51:33 crc kubenswrapper[4971]: I1213 06:51:33.358216 4971 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/97219ee9-4c52-49aa-903d-044a079d4ecb-config\") pod \"openshift-apiserver-operator-796bbdcf4f-7ddbb\" (UID: \"97219ee9-4c52-49aa-903d-044a079d4ecb\") " pod="openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-7ddbb" Dec 13 06:51:33 crc kubenswrapper[4971]: I1213 06:51:33.358235 4971 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/8e7aefba-ae2e-4069-892e-544c85b9af15-audit-policies\") pod \"apiserver-7bbb656c7d-mb2q7\" (UID: \"8e7aefba-ae2e-4069-892e-544c85b9af15\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-mb2q7" Dec 13 06:51:33 crc kubenswrapper[4971]: I1213 06:51:33.358250 4971 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/8e7aefba-ae2e-4069-892e-544c85b9af15-serving-cert\") pod \"apiserver-7bbb656c7d-mb2q7\" (UID: \"8e7aefba-ae2e-4069-892e-544c85b9af15\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-mb2q7" Dec 13 06:51:33 crc kubenswrapper[4971]: I1213 06:51:33.358267 4971 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-user-template-error\" (UniqueName: \"kubernetes.io/secret/23e16ced-5b54-4a51-a1bc-4d11c907c90a-v4-0-config-user-template-error\") pod \"oauth-openshift-558db77b4-nzzbm\" (UID: \"23e16ced-5b54-4a51-a1bc-4d11c907c90a\") " pod="openshift-authentication/oauth-openshift-558db77b4-nzzbm" Dec 13 06:51:33 crc kubenswrapper[4971]: I1213 06:51:33.358286 4971 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/c45a314c-77da-43fc-b2ae-9de350540c2b-config\") pod \"machine-api-operator-5694c8668f-t24dm\" (UID: \"c45a314c-77da-43fc-b2ae-9de350540c2b\") " pod="openshift-machine-api/machine-api-operator-5694c8668f-t24dm" Dec 13 06:51:33 crc kubenswrapper[4971]: I1213 06:51:33.358300 4971 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-system-ocp-branding-template\" (UniqueName: \"kubernetes.io/secret/23e16ced-5b54-4a51-a1bc-4d11c907c90a-v4-0-config-system-ocp-branding-template\") pod \"oauth-openshift-558db77b4-nzzbm\" (UID: \"23e16ced-5b54-4a51-a1bc-4d11c907c90a\") " pod="openshift-authentication/oauth-openshift-558db77b4-nzzbm" Dec 13 06:51:33 crc kubenswrapper[4971]: I1213 06:51:33.358317 4971 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-system-trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/23e16ced-5b54-4a51-a1bc-4d11c907c90a-v4-0-config-system-trusted-ca-bundle\") pod \"oauth-openshift-558db77b4-nzzbm\" (UID: \"23e16ced-5b54-4a51-a1bc-4d11c907c90a\") " pod="openshift-authentication/oauth-openshift-558db77b4-nzzbm" Dec 13 06:51:33 crc kubenswrapper[4971]: I1213 06:51:33.358332 4971 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"console-serving-cert\" (UniqueName: \"kubernetes.io/secret/9f0db656-7a94-415b-b737-30ed342817ee-console-serving-cert\") pod \"console-f9d7485db-5fd9b\" (UID: \"9f0db656-7a94-415b-b737-30ed342817ee\") " pod="openshift-console/console-f9d7485db-5fd9b" Dec 13 06:51:33 crc kubenswrapper[4971]: I1213 06:51:33.358347 4971 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/bec6e53b-3998-4402-b77b-87ff79f31827-etcd-client\") pod \"apiserver-76f77b778f-2n4cq\" (UID: \"bec6e53b-3998-4402-b77b-87ff79f31827\") " pod="openshift-apiserver/apiserver-76f77b778f-2n4cq" Dec 13 06:51:33 crc kubenswrapper[4971]: I1213 06:51:33.358360 4971 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-krxlt\" (UniqueName: \"kubernetes.io/projected/23e16ced-5b54-4a51-a1bc-4d11c907c90a-kube-api-access-krxlt\") pod \"oauth-openshift-558db77b4-nzzbm\" (UID: \"23e16ced-5b54-4a51-a1bc-4d11c907c90a\") " pod="openshift-authentication/oauth-openshift-558db77b4-nzzbm" Dec 13 06:51:33 crc kubenswrapper[4971]: I1213 06:51:33.358376 4971 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/8294e250-74ea-497c-b6ea-08b957184eec-trusted-ca\") pod \"cluster-image-registry-operator-dc59b4c8b-zqt2t\" (UID: \"8294e250-74ea-497c-b6ea-08b957184eec\") " pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-zqt2t" Dec 13 06:51:33 crc kubenswrapper[4971]: I1213 06:51:33.358391 4971 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/7148e573-e2f5-4fc5-b30c-4e777516509e-config\") pod \"controller-manager-879f6c89f-xtxr8\" (UID: \"7148e573-e2f5-4fc5-b30c-4e777516509e\") " pod="openshift-controller-manager/controller-manager-879f6c89f-xtxr8" Dec 13 06:51:33 crc kubenswrapper[4971]: I1213 06:51:33.358406 4971 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-user-template-login\" (UniqueName: \"kubernetes.io/secret/23e16ced-5b54-4a51-a1bc-4d11c907c90a-v4-0-config-user-template-login\") pod \"oauth-openshift-558db77b4-nzzbm\" (UID: \"23e16ced-5b54-4a51-a1bc-4d11c907c90a\") " pod="openshift-authentication/oauth-openshift-558db77b4-nzzbm" Dec 13 06:51:33 crc kubenswrapper[4971]: I1213 06:51:33.358428 4971 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-znjpx\" (UniqueName: \"kubernetes.io/projected/97219ee9-4c52-49aa-903d-044a079d4ecb-kube-api-access-znjpx\") pod \"openshift-apiserver-operator-796bbdcf4f-7ddbb\" (UID: \"97219ee9-4c52-49aa-903d-044a079d4ecb\") " pod="openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-7ddbb" Dec 13 06:51:33 crc kubenswrapper[4971]: I1213 06:51:33.358443 4971 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/8e7aefba-ae2e-4069-892e-544c85b9af15-trusted-ca-bundle\") pod \"apiserver-7bbb656c7d-mb2q7\" (UID: \"8e7aefba-ae2e-4069-892e-544c85b9af15\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-mb2q7" Dec 13 06:51:33 crc kubenswrapper[4971]: I1213 06:51:33.358457 4971 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"machine-api-operator-tls\" (UniqueName: \"kubernetes.io/secret/c45a314c-77da-43fc-b2ae-9de350540c2b-machine-api-operator-tls\") pod \"machine-api-operator-5694c8668f-t24dm\" (UID: \"c45a314c-77da-43fc-b2ae-9de350540c2b\") " pod="openshift-machine-api/machine-api-operator-5694c8668f-t24dm" Dec 13 06:51:33 crc kubenswrapper[4971]: I1213 06:51:33.358472 4971 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/23e16ced-5b54-4a51-a1bc-4d11c907c90a-audit-dir\") pod \"oauth-openshift-558db77b4-nzzbm\" (UID: \"23e16ced-5b54-4a51-a1bc-4d11c907c90a\") " pod="openshift-authentication/oauth-openshift-558db77b4-nzzbm" Dec 13 06:51:33 crc kubenswrapper[4971]: I1213 06:51:33.358485 4971 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-user-idp-0-file-data\" (UniqueName: \"kubernetes.io/secret/23e16ced-5b54-4a51-a1bc-4d11c907c90a-v4-0-config-user-idp-0-file-data\") pod \"oauth-openshift-558db77b4-nzzbm\" (UID: \"23e16ced-5b54-4a51-a1bc-4d11c907c90a\") " pod="openshift-authentication/oauth-openshift-558db77b4-nzzbm" Dec 13 06:51:33 crc kubenswrapper[4971]: I1213 06:51:33.358499 4971 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-68rgh\" (UniqueName: \"kubernetes.io/projected/9f0db656-7a94-415b-b737-30ed342817ee-kube-api-access-68rgh\") pod \"console-f9d7485db-5fd9b\" (UID: \"9f0db656-7a94-415b-b737-30ed342817ee\") " pod="openshift-console/console-f9d7485db-5fd9b" Dec 13 06:51:33 crc kubenswrapper[4971]: I1213 06:51:33.358517 4971 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"image-import-ca\" (UniqueName: \"kubernetes.io/configmap/bec6e53b-3998-4402-b77b-87ff79f31827-image-import-ca\") pod \"apiserver-76f77b778f-2n4cq\" (UID: \"bec6e53b-3998-4402-b77b-87ff79f31827\") " pod="openshift-apiserver/apiserver-76f77b778f-2n4cq" Dec 13 06:51:33 crc kubenswrapper[4971]: I1213 06:51:33.358552 4971 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"available-featuregates\" (UniqueName: \"kubernetes.io/empty-dir/874901b0-9650-4029-8d21-74441f40e201-available-featuregates\") pod \"openshift-config-operator-7777fb866f-vmdsp\" (UID: \"874901b0-9650-4029-8d21-74441f40e201\") " pod="openshift-config-operator/openshift-config-operator-7777fb866f-vmdsp" Dec 13 06:51:33 crc kubenswrapper[4971]: I1213 06:51:33.358574 4971 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"images\" (UniqueName: \"kubernetes.io/configmap/c45a314c-77da-43fc-b2ae-9de350540c2b-images\") pod \"machine-api-operator-5694c8668f-t24dm\" (UID: \"c45a314c-77da-43fc-b2ae-9de350540c2b\") " pod="openshift-machine-api/machine-api-operator-5694c8668f-t24dm" Dec 13 06:51:33 crc kubenswrapper[4971]: I1213 06:51:33.358588 4971 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/9f0db656-7a94-415b-b737-30ed342817ee-service-ca\") pod \"console-f9d7485db-5fd9b\" (UID: \"9f0db656-7a94-415b-b737-30ed342817ee\") " pod="openshift-console/console-f9d7485db-5fd9b" Dec 13 06:51:33 crc kubenswrapper[4971]: I1213 06:51:33.358604 4971 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/8294e250-74ea-497c-b6ea-08b957184eec-bound-sa-token\") pod \"cluster-image-registry-operator-dc59b4c8b-zqt2t\" (UID: \"8294e250-74ea-497c-b6ea-08b957184eec\") " pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-zqt2t" Dec 13 06:51:33 crc kubenswrapper[4971]: I1213 06:51:33.358618 4971 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/23e16ced-5b54-4a51-a1bc-4d11c907c90a-audit-policies\") pod \"oauth-openshift-558db77b4-nzzbm\" (UID: \"23e16ced-5b54-4a51-a1bc-4d11c907c90a\") " pod="openshift-authentication/oauth-openshift-558db77b4-nzzbm" Dec 13 06:51:33 crc kubenswrapper[4971]: I1213 06:51:33.358625 4971 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-ingress-operator/ingress-operator-5b745b69d9-pkmb5" Dec 13 06:51:33 crc kubenswrapper[4971]: I1213 06:51:33.358926 4971 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-controller-manager-operator"/"openshift-controller-manager-operator-dockercfg-vw8fw" Dec 13 06:51:33 crc kubenswrapper[4971]: I1213 06:51:33.359002 4971 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-dns-operator/dns-operator-744455d44c-b999k" Dec 13 06:51:33 crc kubenswrapper[4971]: I1213 06:51:33.359265 4971 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-etcd-operator/etcd-operator-b45778765-6xfwq" Dec 13 06:51:33 crc kubenswrapper[4971]: I1213 06:51:33.358634 4971 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"samples-operator-tls\" (UniqueName: \"kubernetes.io/secret/817cbb9d-ed5b-4955-8b6a-4b33155b744c-samples-operator-tls\") pod \"cluster-samples-operator-665b6dd947-mlbwt\" (UID: \"817cbb9d-ed5b-4955-8b6a-4b33155b744c\") " pod="openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-mlbwt" Dec 13 06:51:33 crc kubenswrapper[4971]: I1213 06:51:33.359494 4971 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etcd-serving-ca\" (UniqueName: \"kubernetes.io/configmap/8e7aefba-ae2e-4069-892e-544c85b9af15-etcd-serving-ca\") pod \"apiserver-7bbb656c7d-mb2q7\" (UID: \"8e7aefba-ae2e-4069-892e-544c85b9af15\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-mb2q7" Dec 13 06:51:33 crc kubenswrapper[4971]: I1213 06:51:33.359542 4971 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-system-serving-cert\" (UniqueName: \"kubernetes.io/secret/23e16ced-5b54-4a51-a1bc-4d11c907c90a-v4-0-config-system-serving-cert\") pod \"oauth-openshift-558db77b4-nzzbm\" (UID: \"23e16ced-5b54-4a51-a1bc-4d11c907c90a\") " pod="openshift-authentication/oauth-openshift-558db77b4-nzzbm" Dec 13 06:51:33 crc kubenswrapper[4971]: I1213 06:51:33.359585 4971 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/7148e573-e2f5-4fc5-b30c-4e777516509e-serving-cert\") pod \"controller-manager-879f6c89f-xtxr8\" (UID: \"7148e573-e2f5-4fc5-b30c-4e777516509e\") " pod="openshift-controller-manager/controller-manager-879f6c89f-xtxr8" Dec 13 06:51:33 crc kubenswrapper[4971]: I1213 06:51:33.359601 4971 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver"/"trusted-ca-bundle" Dec 13 06:51:33 crc kubenswrapper[4971]: I1213 06:51:33.359611 4971 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/874901b0-9650-4029-8d21-74441f40e201-serving-cert\") pod \"openshift-config-operator-7777fb866f-vmdsp\" (UID: \"874901b0-9650-4029-8d21-74441f40e201\") " pod="openshift-config-operator/openshift-config-operator-7777fb866f-vmdsp" Dec 13 06:51:33 crc kubenswrapper[4971]: I1213 06:51:33.359764 4971 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-dlhrl\" (UniqueName: \"kubernetes.io/projected/874901b0-9650-4029-8d21-74441f40e201-kube-api-access-dlhrl\") pod \"openshift-config-operator-7777fb866f-vmdsp\" (UID: \"874901b0-9650-4029-8d21-74441f40e201\") " pod="openshift-config-operator/openshift-config-operator-7777fb866f-vmdsp" Dec 13 06:51:33 crc kubenswrapper[4971]: I1213 06:51:33.359790 4971 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"console-oauth-config\" (UniqueName: \"kubernetes.io/secret/9f0db656-7a94-415b-b737-30ed342817ee-console-oauth-config\") pod \"console-f9d7485db-5fd9b\" (UID: \"9f0db656-7a94-415b-b737-30ed342817ee\") " pod="openshift-console/console-f9d7485db-5fd9b" Dec 13 06:51:33 crc kubenswrapper[4971]: I1213 06:51:33.359809 4971 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-kjkp6\" (UniqueName: \"kubernetes.io/projected/817cbb9d-ed5b-4955-8b6a-4b33155b744c-kube-api-access-kjkp6\") pod \"cluster-samples-operator-665b6dd947-mlbwt\" (UID: \"817cbb9d-ed5b-4955-8b6a-4b33155b744c\") " pod="openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-mlbwt" Dec 13 06:51:33 crc kubenswrapper[4971]: I1213 06:51:33.359827 4971 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/88850e0f-df47-4650-be7e-d14a83ce05f4-config\") pod \"machine-approver-56656f9798-s75fl\" (UID: \"88850e0f-df47-4650-be7e-d14a83ce05f4\") " pod="openshift-cluster-machine-approver/machine-approver-56656f9798-s75fl" Dec 13 06:51:33 crc kubenswrapper[4971]: I1213 06:51:33.359843 4971 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/97219ee9-4c52-49aa-903d-044a079d4ecb-serving-cert\") pod \"openshift-apiserver-operator-796bbdcf4f-7ddbb\" (UID: \"97219ee9-4c52-49aa-903d-044a079d4ecb\") " pod="openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-7ddbb" Dec 13 06:51:33 crc kubenswrapper[4971]: I1213 06:51:33.359858 4971 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/7148e573-e2f5-4fc5-b30c-4e777516509e-proxy-ca-bundles\") pod \"controller-manager-879f6c89f-xtxr8\" (UID: \"7148e573-e2f5-4fc5-b30c-4e777516509e\") " pod="openshift-controller-manager/controller-manager-879f6c89f-xtxr8" Dec 13 06:51:33 crc kubenswrapper[4971]: I1213 06:51:33.359877 4971 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"node-pullsecrets\" (UniqueName: \"kubernetes.io/host-path/bec6e53b-3998-4402-b77b-87ff79f31827-node-pullsecrets\") pod \"apiserver-76f77b778f-2n4cq\" (UID: \"bec6e53b-3998-4402-b77b-87ff79f31827\") " pod="openshift-apiserver/apiserver-76f77b778f-2n4cq" Dec 13 06:51:33 crc kubenswrapper[4971]: I1213 06:51:33.359892 4971 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/bec6e53b-3998-4402-b77b-87ff79f31827-serving-cert\") pod \"apiserver-76f77b778f-2n4cq\" (UID: \"bec6e53b-3998-4402-b77b-87ff79f31827\") " pod="openshift-apiserver/apiserver-76f77b778f-2n4cq" Dec 13 06:51:33 crc kubenswrapper[4971]: I1213 06:51:33.359906 4971 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-8m6vz\" (UniqueName: \"kubernetes.io/projected/88850e0f-df47-4650-be7e-d14a83ce05f4-kube-api-access-8m6vz\") pod \"machine-approver-56656f9798-s75fl\" (UID: \"88850e0f-df47-4650-be7e-d14a83ce05f4\") " pod="openshift-cluster-machine-approver/machine-approver-56656f9798-s75fl" Dec 13 06:51:33 crc kubenswrapper[4971]: I1213 06:51:33.359924 4971 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-system-service-ca\" (UniqueName: \"kubernetes.io/configmap/23e16ced-5b54-4a51-a1bc-4d11c907c90a-v4-0-config-system-service-ca\") pod \"oauth-openshift-558db77b4-nzzbm\" (UID: \"23e16ced-5b54-4a51-a1bc-4d11c907c90a\") " pod="openshift-authentication/oauth-openshift-558db77b4-nzzbm" Dec 13 06:51:33 crc kubenswrapper[4971]: I1213 06:51:33.359940 4971 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-zvw59\" (UniqueName: \"kubernetes.io/projected/bec6e53b-3998-4402-b77b-87ff79f31827-kube-api-access-zvw59\") pod \"apiserver-76f77b778f-2n4cq\" (UID: \"bec6e53b-3998-4402-b77b-87ff79f31827\") " pod="openshift-apiserver/apiserver-76f77b778f-2n4cq" Dec 13 06:51:33 crc kubenswrapper[4971]: I1213 06:51:33.359961 4971 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/bec6e53b-3998-4402-b77b-87ff79f31827-audit-dir\") pod \"apiserver-76f77b778f-2n4cq\" (UID: \"bec6e53b-3998-4402-b77b-87ff79f31827\") " pod="openshift-apiserver/apiserver-76f77b778f-2n4cq" Dec 13 06:51:33 crc kubenswrapper[4971]: I1213 06:51:33.359975 4971 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"image-registry-operator-tls\" (UniqueName: \"kubernetes.io/secret/8294e250-74ea-497c-b6ea-08b957184eec-image-registry-operator-tls\") pod \"cluster-image-registry-operator-dc59b4c8b-zqt2t\" (UID: \"8294e250-74ea-497c-b6ea-08b957184eec\") " pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-zqt2t" Dec 13 06:51:33 crc kubenswrapper[4971]: I1213 06:51:33.359992 4971 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-tjjmd\" (UniqueName: \"kubernetes.io/projected/8e7aefba-ae2e-4069-892e-544c85b9af15-kube-api-access-tjjmd\") pod \"apiserver-7bbb656c7d-mb2q7\" (UID: \"8e7aefba-ae2e-4069-892e-544c85b9af15\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-mb2q7" Dec 13 06:51:33 crc kubenswrapper[4971]: I1213 06:51:33.360008 4971 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-lmhvx\" (UniqueName: \"kubernetes.io/projected/7148e573-e2f5-4fc5-b30c-4e777516509e-kube-api-access-lmhvx\") pod \"controller-manager-879f6c89f-xtxr8\" (UID: \"7148e573-e2f5-4fc5-b30c-4e777516509e\") " pod="openshift-controller-manager/controller-manager-879f6c89f-xtxr8" Dec 13 06:51:33 crc kubenswrapper[4971]: I1213 06:51:33.360022 4971 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-87jr5\" (UniqueName: \"kubernetes.io/projected/c45a314c-77da-43fc-b2ae-9de350540c2b-kube-api-access-87jr5\") pod \"machine-api-operator-5694c8668f-t24dm\" (UID: \"c45a314c-77da-43fc-b2ae-9de350540c2b\") " pod="openshift-machine-api/machine-api-operator-5694c8668f-t24dm" Dec 13 06:51:33 crc kubenswrapper[4971]: I1213 06:51:33.360038 4971 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"encryption-config\" (UniqueName: \"kubernetes.io/secret/bec6e53b-3998-4402-b77b-87ff79f31827-encryption-config\") pod \"apiserver-76f77b778f-2n4cq\" (UID: \"bec6e53b-3998-4402-b77b-87ff79f31827\") " pod="openshift-apiserver/apiserver-76f77b778f-2n4cq" Dec 13 06:51:33 crc kubenswrapper[4971]: I1213 06:51:33.360054 4971 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-tb7rw\" (UniqueName: \"kubernetes.io/projected/8294e250-74ea-497c-b6ea-08b957184eec-kube-api-access-tb7rw\") pod \"cluster-image-registry-operator-dc59b4c8b-zqt2t\" (UID: \"8294e250-74ea-497c-b6ea-08b957184eec\") " pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-zqt2t" Dec 13 06:51:33 crc kubenswrapper[4971]: I1213 06:51:33.360068 4971 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"audit\" (UniqueName: \"kubernetes.io/configmap/bec6e53b-3998-4402-b77b-87ff79f31827-audit\") pod \"apiserver-76f77b778f-2n4cq\" (UID: \"bec6e53b-3998-4402-b77b-87ff79f31827\") " pod="openshift-apiserver/apiserver-76f77b778f-2n4cq" Dec 13 06:51:33 crc kubenswrapper[4971]: I1213 06:51:33.360083 4971 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"machine-approver-tls\" (UniqueName: \"kubernetes.io/secret/88850e0f-df47-4650-be7e-d14a83ce05f4-machine-approver-tls\") pod \"machine-approver-56656f9798-s75fl\" (UID: \"88850e0f-df47-4650-be7e-d14a83ce05f4\") " pod="openshift-cluster-machine-approver/machine-approver-56656f9798-s75fl" Dec 13 06:51:33 crc kubenswrapper[4971]: I1213 06:51:33.360098 4971 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/8e7aefba-ae2e-4069-892e-544c85b9af15-etcd-client\") pod \"apiserver-7bbb656c7d-mb2q7\" (UID: \"8e7aefba-ae2e-4069-892e-544c85b9af15\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-mb2q7" Dec 13 06:51:33 crc kubenswrapper[4971]: I1213 06:51:33.360111 4971 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/9f0db656-7a94-415b-b737-30ed342817ee-trusted-ca-bundle\") pod \"console-f9d7485db-5fd9b\" (UID: \"9f0db656-7a94-415b-b737-30ed342817ee\") " pod="openshift-console/console-f9d7485db-5fd9b" Dec 13 06:51:33 crc kubenswrapper[4971]: I1213 06:51:33.360124 4971 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"oauth-serving-cert\" (UniqueName: \"kubernetes.io/configmap/9f0db656-7a94-415b-b737-30ed342817ee-oauth-serving-cert\") pod \"console-f9d7485db-5fd9b\" (UID: \"9f0db656-7a94-415b-b737-30ed342817ee\") " pod="openshift-console/console-f9d7485db-5fd9b" Dec 13 06:51:33 crc kubenswrapper[4971]: I1213 06:51:33.360141 4971 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/88850e0f-df47-4650-be7e-d14a83ce05f4-auth-proxy-config\") pod \"machine-approver-56656f9798-s75fl\" (UID: \"88850e0f-df47-4650-be7e-d14a83ce05f4\") " pod="openshift-cluster-machine-approver/machine-approver-56656f9798-s75fl" Dec 13 06:51:33 crc kubenswrapper[4971]: I1213 06:51:33.360158 4971 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/7148e573-e2f5-4fc5-b30c-4e777516509e-client-ca\") pod \"controller-manager-879f6c89f-xtxr8\" (UID: \"7148e573-e2f5-4fc5-b30c-4e777516509e\") " pod="openshift-controller-manager/controller-manager-879f6c89f-xtxr8" Dec 13 06:51:33 crc kubenswrapper[4971]: I1213 06:51:33.360172 4971 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-system-session\" (UniqueName: \"kubernetes.io/secret/23e16ced-5b54-4a51-a1bc-4d11c907c90a-v4-0-config-system-session\") pod \"oauth-openshift-558db77b4-nzzbm\" (UID: \"23e16ced-5b54-4a51-a1bc-4d11c907c90a\") " pod="openshift-authentication/oauth-openshift-558db77b4-nzzbm" Dec 13 06:51:33 crc kubenswrapper[4971]: I1213 06:51:33.360188 4971 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-system-router-certs\" (UniqueName: \"kubernetes.io/secret/23e16ced-5b54-4a51-a1bc-4d11c907c90a-v4-0-config-system-router-certs\") pod \"oauth-openshift-558db77b4-nzzbm\" (UID: \"23e16ced-5b54-4a51-a1bc-4d11c907c90a\") " pod="openshift-authentication/oauth-openshift-558db77b4-nzzbm" Dec 13 06:51:33 crc kubenswrapper[4971]: I1213 06:51:33.360208 4971 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/8e7aefba-ae2e-4069-892e-544c85b9af15-audit-dir\") pod \"apiserver-7bbb656c7d-mb2q7\" (UID: \"8e7aefba-ae2e-4069-892e-544c85b9af15\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-mb2q7" Dec 13 06:51:33 crc kubenswrapper[4971]: I1213 06:51:33.360224 4971 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-system-cliconfig\" (UniqueName: \"kubernetes.io/configmap/23e16ced-5b54-4a51-a1bc-4d11c907c90a-v4-0-config-system-cliconfig\") pod \"oauth-openshift-558db77b4-nzzbm\" (UID: \"23e16ced-5b54-4a51-a1bc-4d11c907c90a\") " pod="openshift-authentication/oauth-openshift-558db77b4-nzzbm" Dec 13 06:51:33 crc kubenswrapper[4971]: I1213 06:51:33.360245 4971 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-user-template-provider-selection\" (UniqueName: \"kubernetes.io/secret/23e16ced-5b54-4a51-a1bc-4d11c907c90a-v4-0-config-user-template-provider-selection\") pod \"oauth-openshift-558db77b4-nzzbm\" (UID: \"23e16ced-5b54-4a51-a1bc-4d11c907c90a\") " pod="openshift-authentication/oauth-openshift-558db77b4-nzzbm" Dec 13 06:51:33 crc kubenswrapper[4971]: I1213 06:51:33.360262 4971 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"encryption-config\" (UniqueName: \"kubernetes.io/secret/8e7aefba-ae2e-4069-892e-544c85b9af15-encryption-config\") pod \"apiserver-7bbb656c7d-mb2q7\" (UID: \"8e7aefba-ae2e-4069-892e-544c85b9af15\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-mb2q7" Dec 13 06:51:33 crc kubenswrapper[4971]: I1213 06:51:33.360277 4971 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etcd-serving-ca\" (UniqueName: \"kubernetes.io/configmap/bec6e53b-3998-4402-b77b-87ff79f31827-etcd-serving-ca\") pod \"apiserver-76f77b778f-2n4cq\" (UID: \"bec6e53b-3998-4402-b77b-87ff79f31827\") " pod="openshift-apiserver/apiserver-76f77b778f-2n4cq" Dec 13 06:51:33 crc kubenswrapper[4971]: I1213 06:51:33.360302 4971 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"console-config\" (UniqueName: \"kubernetes.io/configmap/9f0db656-7a94-415b-b737-30ed342817ee-console-config\") pod \"console-f9d7485db-5fd9b\" (UID: \"9f0db656-7a94-415b-b737-30ed342817ee\") " pod="openshift-console/console-f9d7485db-5fd9b" Dec 13 06:51:33 crc kubenswrapper[4971]: I1213 06:51:33.360321 4971 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/bec6e53b-3998-4402-b77b-87ff79f31827-config\") pod \"apiserver-76f77b778f-2n4cq\" (UID: \"bec6e53b-3998-4402-b77b-87ff79f31827\") " pod="openshift-apiserver/apiserver-76f77b778f-2n4cq" Dec 13 06:51:33 crc kubenswrapper[4971]: I1213 06:51:33.362387 4971 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-machine-config-operator/machine-config-controller-84d6567774-kf7sp"] Dec 13 06:51:33 crc kubenswrapper[4971]: I1213 06:51:33.363141 4971 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-machine-config-operator/machine-config-controller-84d6567774-kf7sp" Dec 13 06:51:33 crc kubenswrapper[4971]: I1213 06:51:33.363275 4971 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-multus/multus-admission-controller-857f4d67dd-z68ks"] Dec 13 06:51:33 crc kubenswrapper[4971]: I1213 06:51:33.366254 4971 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-machine-api/control-plane-machine-set-operator-78cbb6b69f-bgdbk"] Dec 13 06:51:33 crc kubenswrapper[4971]: I1213 06:51:33.366499 4971 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager-operator"/"kube-root-ca.crt" Dec 13 06:51:33 crc kubenswrapper[4971]: I1213 06:51:33.366724 4971 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-machine-api/control-plane-machine-set-operator-78cbb6b69f-bgdbk" Dec 13 06:51:33 crc kubenswrapper[4971]: I1213 06:51:33.366994 4971 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/multus-admission-controller-857f4d67dd-z68ks" Dec 13 06:51:33 crc kubenswrapper[4971]: I1213 06:51:33.367102 4971 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console-operator"/"trusted-ca" Dec 13 06:51:33 crc kubenswrapper[4971]: I1213 06:51:33.367682 4971 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication"/"v4-0-config-system-trusted-ca-bundle" Dec 13 06:51:33 crc kubenswrapper[4971]: I1213 06:51:33.367784 4971 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-g8vfp"] Dec 13 06:51:33 crc kubenswrapper[4971]: I1213 06:51:33.368427 4971 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"openshift-global-ca" Dec 13 06:51:33 crc kubenswrapper[4971]: I1213 06:51:33.368554 4971 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/marketplace-operator-79b997595-j9zqf"] Dec 13 06:51:33 crc kubenswrapper[4971]: I1213 06:51:33.369057 4971 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/marketplace-operator-79b997595-j9zqf" Dec 13 06:51:33 crc kubenswrapper[4971]: I1213 06:51:33.369387 4971 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-g8vfp" Dec 13 06:51:33 crc kubenswrapper[4971]: I1213 06:51:33.371840 4971 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-user-template-login" Dec 13 06:51:33 crc kubenswrapper[4971]: I1213 06:51:33.375669 4971 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-oauth-apiserver/apiserver-7bbb656c7d-mb2q7"] Dec 13 06:51:33 crc kubenswrapper[4971]: I1213 06:51:33.376104 4971 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ingress-operator"/"openshift-service-ca.crt" Dec 13 06:51:33 crc kubenswrapper[4971]: I1213 06:51:33.376184 4971 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-image-registry"/"registry-dockercfg-kzzsd" Dec 13 06:51:33 crc kubenswrapper[4971]: I1213 06:51:33.376343 4971 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-image-registry"/"image-registry-tls" Dec 13 06:51:33 crc kubenswrapper[4971]: I1213 06:51:33.376464 4971 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-service-ca/service-ca-9c57cc56f-q99rh"] Dec 13 06:51:33 crc kubenswrapper[4971]: I1213 06:51:33.376928 4971 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-service-ca/service-ca-9c57cc56f-q99rh" Dec 13 06:51:33 crc kubenswrapper[4971]: I1213 06:51:33.377486 4971 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-image-registry"/"trusted-ca" Dec 13 06:51:33 crc kubenswrapper[4971]: I1213 06:51:33.377751 4971 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-image-registry"/"installation-pull-secrets" Dec 13 06:51:33 crc kubenswrapper[4971]: I1213 06:51:33.378015 4971 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-dns-operator"/"openshift-service-ca.crt" Dec 13 06:51:33 crc kubenswrapper[4971]: I1213 06:51:33.379472 4971 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-system-ocp-branding-template" Dec 13 06:51:33 crc kubenswrapper[4971]: I1213 06:51:33.380296 4971 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-machine-config-operator/machine-config-operator-74547568cd-4cn2z"] Dec 13 06:51:33 crc kubenswrapper[4971]: I1213 06:51:33.381017 4971 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-machine-config-operator/machine-config-operator-74547568cd-4cn2z" Dec 13 06:51:33 crc kubenswrapper[4971]: I1213 06:51:33.381217 4971 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-xfvfg"] Dec 13 06:51:33 crc kubenswrapper[4971]: I1213 06:51:33.381871 4971 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-xfvfg" Dec 13 06:51:33 crc kubenswrapper[4971]: I1213 06:51:33.382645 4971 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-ingress/router-default-5444994796-mzhp6"] Dec 13 06:51:33 crc kubenswrapper[4971]: I1213 06:51:33.384142 4971 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-ingress/router-default-5444994796-mzhp6" Dec 13 06:51:33 crc kubenswrapper[4971]: I1213 06:51:33.387897 4971 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-fz6wk"] Dec 13 06:51:33 crc kubenswrapper[4971]: I1213 06:51:33.388300 4971 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-fz6wk" Dec 13 06:51:33 crc kubenswrapper[4971]: I1213 06:51:33.400165 4971 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-cwwf9"] Dec 13 06:51:33 crc kubenswrapper[4971]: I1213 06:51:33.401833 4971 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-6blxl"] Dec 13 06:51:33 crc kubenswrapper[4971]: I1213 06:51:33.403049 4971 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-6blxl" Dec 13 06:51:33 crc kubenswrapper[4971]: I1213 06:51:33.403490 4971 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-cwwf9" Dec 13 06:51:33 crc kubenswrapper[4971]: I1213 06:51:33.406921 4971 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ingress-operator"/"kube-root-ca.crt" Dec 13 06:51:33 crc kubenswrapper[4971]: I1213 06:51:33.423927 4971 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ingress-operator"/"trusted-ca" Dec 13 06:51:33 crc kubenswrapper[4971]: I1213 06:51:33.425166 4971 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-pxjqw"] Dec 13 06:51:33 crc kubenswrapper[4971]: I1213 06:51:33.425863 4971 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-pxjqw" Dec 13 06:51:33 crc kubenswrapper[4971]: I1213 06:51:33.428291 4971 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-operator-lifecycle-manager/catalog-operator-68c6474976-l4rzp"] Dec 13 06:51:33 crc kubenswrapper[4971]: I1213 06:51:33.429001 4971 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-service-ca-operator/service-ca-operator-777779d784-drq5g"] Dec 13 06:51:33 crc kubenswrapper[4971]: I1213 06:51:33.429437 4971 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-l4rzp" Dec 13 06:51:33 crc kubenswrapper[4971]: I1213 06:51:33.429475 4971 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-service-ca-operator/service-ca-operator-777779d784-drq5g" Dec 13 06:51:33 crc kubenswrapper[4971]: I1213 06:51:33.432451 4971 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-dns-operator"/"dns-operator-dockercfg-9mqw5" Dec 13 06:51:33 crc kubenswrapper[4971]: I1213 06:51:33.433777 4971 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29426805-fxxrz"] Dec 13 06:51:33 crc kubenswrapper[4971]: I1213 06:51:33.434583 4971 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29426805-fxxrz" Dec 13 06:51:33 crc kubenswrapper[4971]: I1213 06:51:33.436685 4971 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-kube-storage-version-migrator/migrator-59844c95c7-z2p8z"] Dec 13 06:51:33 crc kubenswrapper[4971]: I1213 06:51:33.437196 4971 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-storage-version-migrator/migrator-59844c95c7-z2p8z" Dec 13 06:51:33 crc kubenswrapper[4971]: I1213 06:51:33.438215 4971 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-authentication-operator/authentication-operator-69f744f599-wjdd5"] Dec 13 06:51:33 crc kubenswrapper[4971]: I1213 06:51:33.443168 4971 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-lnrkm"] Dec 13 06:51:33 crc kubenswrapper[4971]: I1213 06:51:33.443382 4971 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-authentication-operator/authentication-operator-69f744f599-wjdd5" Dec 13 06:51:33 crc kubenswrapper[4971]: I1213 06:51:33.444207 4971 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-machine-api/machine-api-operator-5694c8668f-t24dm"] Dec 13 06:51:33 crc kubenswrapper[4971]: I1213 06:51:33.444329 4971 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-lnrkm" Dec 13 06:51:33 crc kubenswrapper[4971]: I1213 06:51:33.444608 4971 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-dns-operator"/"metrics-tls" Dec 13 06:51:33 crc kubenswrapper[4971]: I1213 06:51:33.450284 4971 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-apiserver/apiserver-76f77b778f-2n4cq"] Dec 13 06:51:33 crc kubenswrapper[4971]: I1213 06:51:33.450332 4971 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-authentication/oauth-openshift-558db77b4-nzzbm"] Dec 13 06:51:33 crc kubenswrapper[4971]: I1213 06:51:33.450969 4971 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["hostpath-provisioner/csi-hostpathplugin-7nhhk"] Dec 13 06:51:33 crc kubenswrapper[4971]: I1213 06:51:33.451892 4971 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="hostpath-provisioner/csi-hostpathplugin-7nhhk" Dec 13 06:51:33 crc kubenswrapper[4971]: I1213 06:51:33.451960 4971 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-ingress-operator/ingress-operator-5b745b69d9-pkmb5"] Dec 13 06:51:33 crc kubenswrapper[4971]: I1213 06:51:33.457107 4971 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-route-controller-manager/route-controller-manager-6576b87f9c-9d86n"] Dec 13 06:51:33 crc kubenswrapper[4971]: I1213 06:51:33.459145 4971 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-mlbwt"] Dec 13 06:51:33 crc kubenswrapper[4971]: I1213 06:51:33.460937 4971 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"console-config\" (UniqueName: \"kubernetes.io/configmap/9f0db656-7a94-415b-b737-30ed342817ee-console-config\") pod \"console-f9d7485db-5fd9b\" (UID: \"9f0db656-7a94-415b-b737-30ed342817ee\") " pod="openshift-console/console-f9d7485db-5fd9b" Dec 13 06:51:33 crc kubenswrapper[4971]: I1213 06:51:33.460970 4971 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"webhook-certs\" (UniqueName: \"kubernetes.io/secret/32c8b95e-38c8-43fb-b1e4-86ac9a27c2f0-webhook-certs\") pod \"multus-admission-controller-857f4d67dd-z68ks\" (UID: \"32c8b95e-38c8-43fb-b1e4-86ac9a27c2f0\") " pod="openshift-multus/multus-admission-controller-857f4d67dd-z68ks" Dec 13 06:51:33 crc kubenswrapper[4971]: I1213 06:51:33.460991 4971 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-pbjnf\" (UniqueName: \"kubernetes.io/projected/56c62b80-0a33-4156-b52b-cb54c98a4343-kube-api-access-pbjnf\") pod \"etcd-operator-b45778765-6xfwq\" (UID: \"56c62b80-0a33-4156-b52b-cb54c98a4343\") " pod="openshift-etcd-operator/etcd-operator-b45778765-6xfwq" Dec 13 06:51:33 crc kubenswrapper[4971]: I1213 06:51:33.461047 4971 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etcd-serving-ca\" (UniqueName: \"kubernetes.io/configmap/bec6e53b-3998-4402-b77b-87ff79f31827-etcd-serving-ca\") pod \"apiserver-76f77b778f-2n4cq\" (UID: \"bec6e53b-3998-4402-b77b-87ff79f31827\") " pod="openshift-apiserver/apiserver-76f77b778f-2n4cq" Dec 13 06:51:33 crc kubenswrapper[4971]: I1213 06:51:33.461065 4971 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/bec6e53b-3998-4402-b77b-87ff79f31827-config\") pod \"apiserver-76f77b778f-2n4cq\" (UID: \"bec6e53b-3998-4402-b77b-87ff79f31827\") " pod="openshift-apiserver/apiserver-76f77b778f-2n4cq" Dec 13 06:51:33 crc kubenswrapper[4971]: I1213 06:51:33.461082 4971 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/bec6e53b-3998-4402-b77b-87ff79f31827-trusted-ca-bundle\") pod \"apiserver-76f77b778f-2n4cq\" (UID: \"bec6e53b-3998-4402-b77b-87ff79f31827\") " pod="openshift-apiserver/apiserver-76f77b778f-2n4cq" Dec 13 06:51:33 crc kubenswrapper[4971]: I1213 06:51:33.461114 4971 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/97219ee9-4c52-49aa-903d-044a079d4ecb-config\") pod \"openshift-apiserver-operator-796bbdcf4f-7ddbb\" (UID: \"97219ee9-4c52-49aa-903d-044a079d4ecb\") " pod="openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-7ddbb" Dec 13 06:51:33 crc kubenswrapper[4971]: I1213 06:51:33.461562 4971 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/8e7aefba-ae2e-4069-892e-544c85b9af15-audit-policies\") pod \"apiserver-7bbb656c7d-mb2q7\" (UID: \"8e7aefba-ae2e-4069-892e-544c85b9af15\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-mb2q7" Dec 13 06:51:33 crc kubenswrapper[4971]: I1213 06:51:33.461586 4971 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/8e7aefba-ae2e-4069-892e-544c85b9af15-serving-cert\") pod \"apiserver-7bbb656c7d-mb2q7\" (UID: \"8e7aefba-ae2e-4069-892e-544c85b9af15\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-mb2q7" Dec 13 06:51:33 crc kubenswrapper[4971]: I1213 06:51:33.461604 4971 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-user-template-error\" (UniqueName: \"kubernetes.io/secret/23e16ced-5b54-4a51-a1bc-4d11c907c90a-v4-0-config-user-template-error\") pod \"oauth-openshift-558db77b4-nzzbm\" (UID: \"23e16ced-5b54-4a51-a1bc-4d11c907c90a\") " pod="openshift-authentication/oauth-openshift-558db77b4-nzzbm" Dec 13 06:51:33 crc kubenswrapper[4971]: I1213 06:51:33.461632 4971 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-tllvh\" (UniqueName: \"kubernetes.io/projected/f14c8a94-8589-4059-bddf-329452355fab-kube-api-access-tllvh\") pod \"marketplace-operator-79b997595-j9zqf\" (UID: \"f14c8a94-8589-4059-bddf-329452355fab\") " pod="openshift-marketplace/marketplace-operator-79b997595-j9zqf" Dec 13 06:51:33 crc kubenswrapper[4971]: I1213 06:51:33.461650 4971 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/c45a314c-77da-43fc-b2ae-9de350540c2b-config\") pod \"machine-api-operator-5694c8668f-t24dm\" (UID: \"c45a314c-77da-43fc-b2ae-9de350540c2b\") " pod="openshift-machine-api/machine-api-operator-5694c8668f-t24dm" Dec 13 06:51:33 crc kubenswrapper[4971]: I1213 06:51:33.461668 4971 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-system-ocp-branding-template\" (UniqueName: \"kubernetes.io/secret/23e16ced-5b54-4a51-a1bc-4d11c907c90a-v4-0-config-system-ocp-branding-template\") pod \"oauth-openshift-558db77b4-nzzbm\" (UID: \"23e16ced-5b54-4a51-a1bc-4d11c907c90a\") " pod="openshift-authentication/oauth-openshift-558db77b4-nzzbm" Dec 13 06:51:33 crc kubenswrapper[4971]: I1213 06:51:33.461704 4971 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-system-trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/23e16ced-5b54-4a51-a1bc-4d11c907c90a-v4-0-config-system-trusted-ca-bundle\") pod \"oauth-openshift-558db77b4-nzzbm\" (UID: \"23e16ced-5b54-4a51-a1bc-4d11c907c90a\") " pod="openshift-authentication/oauth-openshift-558db77b4-nzzbm" Dec 13 06:51:33 crc kubenswrapper[4971]: I1213 06:51:33.461723 4971 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-2sjfn\" (UniqueName: \"kubernetes.io/projected/32c8b95e-38c8-43fb-b1e4-86ac9a27c2f0-kube-api-access-2sjfn\") pod \"multus-admission-controller-857f4d67dd-z68ks\" (UID: \"32c8b95e-38c8-43fb-b1e4-86ac9a27c2f0\") " pod="openshift-multus/multus-admission-controller-857f4d67dd-z68ks" Dec 13 06:51:33 crc kubenswrapper[4971]: I1213 06:51:33.461740 4971 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/bb0ead7b-eba0-4d16-9363-df54caee0a32-config\") pod \"openshift-controller-manager-operator-756b6f6bc6-gtrwg\" (UID: \"bb0ead7b-eba0-4d16-9363-df54caee0a32\") " pod="openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-gtrwg" Dec 13 06:51:33 crc kubenswrapper[4971]: I1213 06:51:33.461772 4971 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"console-serving-cert\" (UniqueName: \"kubernetes.io/secret/9f0db656-7a94-415b-b737-30ed342817ee-console-serving-cert\") pod \"console-f9d7485db-5fd9b\" (UID: \"9f0db656-7a94-415b-b737-30ed342817ee\") " pod="openshift-console/console-f9d7485db-5fd9b" Dec 13 06:51:33 crc kubenswrapper[4971]: I1213 06:51:33.461788 4971 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/bec6e53b-3998-4402-b77b-87ff79f31827-etcd-client\") pod \"apiserver-76f77b778f-2n4cq\" (UID: \"bec6e53b-3998-4402-b77b-87ff79f31827\") " pod="openshift-apiserver/apiserver-76f77b778f-2n4cq" Dec 13 06:51:33 crc kubenswrapper[4971]: I1213 06:51:33.461805 4971 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-krxlt\" (UniqueName: \"kubernetes.io/projected/23e16ced-5b54-4a51-a1bc-4d11c907c90a-kube-api-access-krxlt\") pod \"oauth-openshift-558db77b4-nzzbm\" (UID: \"23e16ced-5b54-4a51-a1bc-4d11c907c90a\") " pod="openshift-authentication/oauth-openshift-558db77b4-nzzbm" Dec 13 06:51:33 crc kubenswrapper[4971]: I1213 06:51:33.461822 4971 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/8294e250-74ea-497c-b6ea-08b957184eec-trusted-ca\") pod \"cluster-image-registry-operator-dc59b4c8b-zqt2t\" (UID: \"8294e250-74ea-497c-b6ea-08b957184eec\") " pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-zqt2t" Dec 13 06:51:33 crc kubenswrapper[4971]: I1213 06:51:33.461870 4971 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-fs9ph\" (UniqueName: \"kubernetes.io/projected/2f87c67a-fcf1-46e3-9bbd-9bf03ee55da3-kube-api-access-fs9ph\") pod \"machine-config-controller-84d6567774-kf7sp\" (UID: \"2f87c67a-fcf1-46e3-9bbd-9bf03ee55da3\") " pod="openshift-machine-config-operator/machine-config-controller-84d6567774-kf7sp" Dec 13 06:51:33 crc kubenswrapper[4971]: I1213 06:51:33.461891 4971 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/7148e573-e2f5-4fc5-b30c-4e777516509e-config\") pod \"controller-manager-879f6c89f-xtxr8\" (UID: \"7148e573-e2f5-4fc5-b30c-4e777516509e\") " pod="openshift-controller-manager/controller-manager-879f6c89f-xtxr8" Dec 13 06:51:33 crc kubenswrapper[4971]: I1213 06:51:33.461907 4971 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-user-template-login\" (UniqueName: \"kubernetes.io/secret/23e16ced-5b54-4a51-a1bc-4d11c907c90a-v4-0-config-user-template-login\") pod \"oauth-openshift-558db77b4-nzzbm\" (UID: \"23e16ced-5b54-4a51-a1bc-4d11c907c90a\") " pod="openshift-authentication/oauth-openshift-558db77b4-nzzbm" Dec 13 06:51:33 crc kubenswrapper[4971]: I1213 06:51:33.461923 4971 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/8e7aefba-ae2e-4069-892e-544c85b9af15-trusted-ca-bundle\") pod \"apiserver-7bbb656c7d-mb2q7\" (UID: \"8e7aefba-ae2e-4069-892e-544c85b9af15\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-mb2q7" Dec 13 06:51:33 crc kubenswrapper[4971]: I1213 06:51:33.462026 4971 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"machine-api-operator-tls\" (UniqueName: \"kubernetes.io/secret/c45a314c-77da-43fc-b2ae-9de350540c2b-machine-api-operator-tls\") pod \"machine-api-operator-5694c8668f-t24dm\" (UID: \"c45a314c-77da-43fc-b2ae-9de350540c2b\") " pod="openshift-machine-api/machine-api-operator-5694c8668f-t24dm" Dec 13 06:51:33 crc kubenswrapper[4971]: I1213 06:51:33.462070 4971 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/23e16ced-5b54-4a51-a1bc-4d11c907c90a-audit-dir\") pod \"oauth-openshift-558db77b4-nzzbm\" (UID: \"23e16ced-5b54-4a51-a1bc-4d11c907c90a\") " pod="openshift-authentication/oauth-openshift-558db77b4-nzzbm" Dec 13 06:51:33 crc kubenswrapper[4971]: I1213 06:51:33.462646 4971 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-gtrwg"] Dec 13 06:51:33 crc kubenswrapper[4971]: I1213 06:51:33.462705 4971 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-console/console-f9d7485db-5fd9b"] Dec 13 06:51:33 crc kubenswrapper[4971]: I1213 06:51:33.463681 4971 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-user-idp-0-file-data\" (UniqueName: \"kubernetes.io/secret/23e16ced-5b54-4a51-a1bc-4d11c907c90a-v4-0-config-user-idp-0-file-data\") pod \"oauth-openshift-558db77b4-nzzbm\" (UID: \"23e16ced-5b54-4a51-a1bc-4d11c907c90a\") " pod="openshift-authentication/oauth-openshift-558db77b4-nzzbm" Dec 13 06:51:33 crc kubenswrapper[4971]: I1213 06:51:33.464856 4971 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/7148e573-e2f5-4fc5-b30c-4e777516509e-config\") pod \"controller-manager-879f6c89f-xtxr8\" (UID: \"7148e573-e2f5-4fc5-b30c-4e777516509e\") " pod="openshift-controller-manager/controller-manager-879f6c89f-xtxr8" Dec 13 06:51:33 crc kubenswrapper[4971]: I1213 06:51:33.465107 4971 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-dns-operator"/"kube-root-ca.crt" Dec 13 06:51:33 crc kubenswrapper[4971]: I1213 06:51:33.465182 4971 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/bec6e53b-3998-4402-b77b-87ff79f31827-trusted-ca-bundle\") pod \"apiserver-76f77b778f-2n4cq\" (UID: \"bec6e53b-3998-4402-b77b-87ff79f31827\") " pod="openshift-apiserver/apiserver-76f77b778f-2n4cq" Dec 13 06:51:33 crc kubenswrapper[4971]: I1213 06:51:33.465577 4971 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"console-config\" (UniqueName: \"kubernetes.io/configmap/9f0db656-7a94-415b-b737-30ed342817ee-console-config\") pod \"console-f9d7485db-5fd9b\" (UID: \"9f0db656-7a94-415b-b737-30ed342817ee\") " pod="openshift-console/console-f9d7485db-5fd9b" Dec 13 06:51:33 crc kubenswrapper[4971]: I1213 06:51:33.466013 4971 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/8e7aefba-ae2e-4069-892e-544c85b9af15-audit-policies\") pod \"apiserver-7bbb656c7d-mb2q7\" (UID: \"8e7aefba-ae2e-4069-892e-544c85b9af15\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-mb2q7" Dec 13 06:51:33 crc kubenswrapper[4971]: I1213 06:51:33.466426 4971 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-system-trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/23e16ced-5b54-4a51-a1bc-4d11c907c90a-v4-0-config-system-trusted-ca-bundle\") pod \"oauth-openshift-558db77b4-nzzbm\" (UID: \"23e16ced-5b54-4a51-a1bc-4d11c907c90a\") " pod="openshift-authentication/oauth-openshift-558db77b4-nzzbm" Dec 13 06:51:33 crc kubenswrapper[4971]: I1213 06:51:33.467317 4971 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/c45a314c-77da-43fc-b2ae-9de350540c2b-config\") pod \"machine-api-operator-5694c8668f-t24dm\" (UID: \"c45a314c-77da-43fc-b2ae-9de350540c2b\") " pod="openshift-machine-api/machine-api-operator-5694c8668f-t24dm" Dec 13 06:51:33 crc kubenswrapper[4971]: I1213 06:51:33.467504 4971 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/23e16ced-5b54-4a51-a1bc-4d11c907c90a-audit-dir\") pod \"oauth-openshift-558db77b4-nzzbm\" (UID: \"23e16ced-5b54-4a51-a1bc-4d11c907c90a\") " pod="openshift-authentication/oauth-openshift-558db77b4-nzzbm" Dec 13 06:51:33 crc kubenswrapper[4971]: I1213 06:51:33.467783 4971 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/97219ee9-4c52-49aa-903d-044a079d4ecb-config\") pod \"openshift-apiserver-operator-796bbdcf4f-7ddbb\" (UID: \"97219ee9-4c52-49aa-903d-044a079d4ecb\") " pod="openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-7ddbb" Dec 13 06:51:33 crc kubenswrapper[4971]: I1213 06:51:33.467856 4971 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/8e7aefba-ae2e-4069-892e-544c85b9af15-trusted-ca-bundle\") pod \"apiserver-7bbb656c7d-mb2q7\" (UID: \"8e7aefba-ae2e-4069-892e-544c85b9af15\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-mb2q7" Dec 13 06:51:33 crc kubenswrapper[4971]: I1213 06:51:33.467863 4971 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-68rgh\" (UniqueName: \"kubernetes.io/projected/9f0db656-7a94-415b-b737-30ed342817ee-kube-api-access-68rgh\") pod \"console-f9d7485db-5fd9b\" (UID: \"9f0db656-7a94-415b-b737-30ed342817ee\") " pod="openshift-console/console-f9d7485db-5fd9b" Dec 13 06:51:33 crc kubenswrapper[4971]: I1213 06:51:33.467931 4971 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-znjpx\" (UniqueName: \"kubernetes.io/projected/97219ee9-4c52-49aa-903d-044a079d4ecb-kube-api-access-znjpx\") pod \"openshift-apiserver-operator-796bbdcf4f-7ddbb\" (UID: \"97219ee9-4c52-49aa-903d-044a079d4ecb\") " pod="openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-7ddbb" Dec 13 06:51:33 crc kubenswrapper[4971]: I1213 06:51:33.467971 4971 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/56c62b80-0a33-4156-b52b-cb54c98a4343-serving-cert\") pod \"etcd-operator-b45778765-6xfwq\" (UID: \"56c62b80-0a33-4156-b52b-cb54c98a4343\") " pod="openshift-etcd-operator/etcd-operator-b45778765-6xfwq" Dec 13 06:51:33 crc kubenswrapper[4971]: I1213 06:51:33.468030 4971 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"images\" (UniqueName: \"kubernetes.io/configmap/c45a314c-77da-43fc-b2ae-9de350540c2b-images\") pod \"machine-api-operator-5694c8668f-t24dm\" (UID: \"c45a314c-77da-43fc-b2ae-9de350540c2b\") " pod="openshift-machine-api/machine-api-operator-5694c8668f-t24dm" Dec 13 06:51:33 crc kubenswrapper[4971]: I1213 06:51:33.468057 4971 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/9f0db656-7a94-415b-b737-30ed342817ee-service-ca\") pod \"console-f9d7485db-5fd9b\" (UID: \"9f0db656-7a94-415b-b737-30ed342817ee\") " pod="openshift-console/console-f9d7485db-5fd9b" Dec 13 06:51:33 crc kubenswrapper[4971]: I1213 06:51:33.468080 4971 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"image-import-ca\" (UniqueName: \"kubernetes.io/configmap/bec6e53b-3998-4402-b77b-87ff79f31827-image-import-ca\") pod \"apiserver-76f77b778f-2n4cq\" (UID: \"bec6e53b-3998-4402-b77b-87ff79f31827\") " pod="openshift-apiserver/apiserver-76f77b778f-2n4cq" Dec 13 06:51:33 crc kubenswrapper[4971]: I1213 06:51:33.468121 4971 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/bec6e53b-3998-4402-b77b-87ff79f31827-config\") pod \"apiserver-76f77b778f-2n4cq\" (UID: \"bec6e53b-3998-4402-b77b-87ff79f31827\") " pod="openshift-apiserver/apiserver-76f77b778f-2n4cq" Dec 13 06:51:33 crc kubenswrapper[4971]: I1213 06:51:33.468150 4971 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"available-featuregates\" (UniqueName: \"kubernetes.io/empty-dir/874901b0-9650-4029-8d21-74441f40e201-available-featuregates\") pod \"openshift-config-operator-7777fb866f-vmdsp\" (UID: \"874901b0-9650-4029-8d21-74441f40e201\") " pod="openshift-config-operator/openshift-config-operator-7777fb866f-vmdsp" Dec 13 06:51:33 crc kubenswrapper[4971]: I1213 06:51:33.468404 4971 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/8294e250-74ea-497c-b6ea-08b957184eec-trusted-ca\") pod \"cluster-image-registry-operator-dc59b4c8b-zqt2t\" (UID: \"8294e250-74ea-497c-b6ea-08b957184eec\") " pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-zqt2t" Dec 13 06:51:33 crc kubenswrapper[4971]: I1213 06:51:33.468619 4971 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/8294e250-74ea-497c-b6ea-08b957184eec-bound-sa-token\") pod \"cluster-image-registry-operator-dc59b4c8b-zqt2t\" (UID: \"8294e250-74ea-497c-b6ea-08b957184eec\") " pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-zqt2t" Dec 13 06:51:33 crc kubenswrapper[4971]: I1213 06:51:33.468668 4971 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/23e16ced-5b54-4a51-a1bc-4d11c907c90a-audit-policies\") pod \"oauth-openshift-558db77b4-nzzbm\" (UID: \"23e16ced-5b54-4a51-a1bc-4d11c907c90a\") " pod="openshift-authentication/oauth-openshift-558db77b4-nzzbm" Dec 13 06:51:33 crc kubenswrapper[4971]: I1213 06:51:33.468699 4971 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"signing-cabundle\" (UniqueName: \"kubernetes.io/configmap/dcdf01b4-83e9-4d28-821a-e4321ce6b92c-signing-cabundle\") pod \"service-ca-9c57cc56f-q99rh\" (UID: \"dcdf01b4-83e9-4d28-821a-e4321ce6b92c\") " pod="openshift-service-ca/service-ca-9c57cc56f-q99rh" Dec 13 06:51:33 crc kubenswrapper[4971]: I1213 06:51:33.468793 4971 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"available-featuregates\" (UniqueName: \"kubernetes.io/empty-dir/874901b0-9650-4029-8d21-74441f40e201-available-featuregates\") pod \"openshift-config-operator-7777fb866f-vmdsp\" (UID: \"874901b0-9650-4029-8d21-74441f40e201\") " pod="openshift-config-operator/openshift-config-operator-7777fb866f-vmdsp" Dec 13 06:51:33 crc kubenswrapper[4971]: I1213 06:51:33.469008 4971 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/9f0db656-7a94-415b-b737-30ed342817ee-service-ca\") pod \"console-f9d7485db-5fd9b\" (UID: \"9f0db656-7a94-415b-b737-30ed342817ee\") " pod="openshift-console/console-f9d7485db-5fd9b" Dec 13 06:51:33 crc kubenswrapper[4971]: I1213 06:51:33.469047 4971 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etcd-serving-ca\" (UniqueName: \"kubernetes.io/configmap/bec6e53b-3998-4402-b77b-87ff79f31827-etcd-serving-ca\") pod \"apiserver-76f77b778f-2n4cq\" (UID: \"bec6e53b-3998-4402-b77b-87ff79f31827\") " pod="openshift-apiserver/apiserver-76f77b778f-2n4cq" Dec 13 06:51:33 crc kubenswrapper[4971]: I1213 06:51:33.469608 4971 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"images\" (UniqueName: \"kubernetes.io/configmap/c45a314c-77da-43fc-b2ae-9de350540c2b-images\") pod \"machine-api-operator-5694c8668f-t24dm\" (UID: \"c45a314c-77da-43fc-b2ae-9de350540c2b\") " pod="openshift-machine-api/machine-api-operator-5694c8668f-t24dm" Dec 13 06:51:33 crc kubenswrapper[4971]: I1213 06:51:33.469658 4971 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"image-import-ca\" (UniqueName: \"kubernetes.io/configmap/bec6e53b-3998-4402-b77b-87ff79f31827-image-import-ca\") pod \"apiserver-76f77b778f-2n4cq\" (UID: \"bec6e53b-3998-4402-b77b-87ff79f31827\") " pod="openshift-apiserver/apiserver-76f77b778f-2n4cq" Dec 13 06:51:33 crc kubenswrapper[4971]: I1213 06:51:33.469722 4971 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/23e16ced-5b54-4a51-a1bc-4d11c907c90a-audit-policies\") pod \"oauth-openshift-558db77b4-nzzbm\" (UID: \"23e16ced-5b54-4a51-a1bc-4d11c907c90a\") " pod="openshift-authentication/oauth-openshift-558db77b4-nzzbm" Dec 13 06:51:33 crc kubenswrapper[4971]: I1213 06:51:33.469733 4971 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"samples-operator-tls\" (UniqueName: \"kubernetes.io/secret/817cbb9d-ed5b-4955-8b6a-4b33155b744c-samples-operator-tls\") pod \"cluster-samples-operator-665b6dd947-mlbwt\" (UID: \"817cbb9d-ed5b-4955-8b6a-4b33155b744c\") " pod="openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-mlbwt" Dec 13 06:51:33 crc kubenswrapper[4971]: I1213 06:51:33.469855 4971 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etcd-serving-ca\" (UniqueName: \"kubernetes.io/configmap/8e7aefba-ae2e-4069-892e-544c85b9af15-etcd-serving-ca\") pod \"apiserver-7bbb656c7d-mb2q7\" (UID: \"8e7aefba-ae2e-4069-892e-544c85b9af15\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-mb2q7" Dec 13 06:51:33 crc kubenswrapper[4971]: I1213 06:51:33.470396 4971 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-system-serving-cert\" (UniqueName: \"kubernetes.io/secret/23e16ced-5b54-4a51-a1bc-4d11c907c90a-v4-0-config-system-serving-cert\") pod \"oauth-openshift-558db77b4-nzzbm\" (UID: \"23e16ced-5b54-4a51-a1bc-4d11c907c90a\") " pod="openshift-authentication/oauth-openshift-558db77b4-nzzbm" Dec 13 06:51:33 crc kubenswrapper[4971]: I1213 06:51:33.470990 4971 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etcd-serving-ca\" (UniqueName: \"kubernetes.io/configmap/8e7aefba-ae2e-4069-892e-544c85b9af15-etcd-serving-ca\") pod \"apiserver-7bbb656c7d-mb2q7\" (UID: \"8e7aefba-ae2e-4069-892e-544c85b9af15\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-mb2q7" Dec 13 06:51:33 crc kubenswrapper[4971]: I1213 06:51:33.471144 4971 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/7148e573-e2f5-4fc5-b30c-4e777516509e-serving-cert\") pod \"controller-manager-879f6c89f-xtxr8\" (UID: \"7148e573-e2f5-4fc5-b30c-4e777516509e\") " pod="openshift-controller-manager/controller-manager-879f6c89f-xtxr8" Dec 13 06:51:33 crc kubenswrapper[4971]: I1213 06:51:33.471182 4971 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/874901b0-9650-4029-8d21-74441f40e201-serving-cert\") pod \"openshift-config-operator-7777fb866f-vmdsp\" (UID: \"874901b0-9650-4029-8d21-74441f40e201\") " pod="openshift-config-operator/openshift-config-operator-7777fb866f-vmdsp" Dec 13 06:51:33 crc kubenswrapper[4971]: I1213 06:51:33.471292 4971 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-dlhrl\" (UniqueName: \"kubernetes.io/projected/874901b0-9650-4029-8d21-74441f40e201-kube-api-access-dlhrl\") pod \"openshift-config-operator-7777fb866f-vmdsp\" (UID: \"874901b0-9650-4029-8d21-74441f40e201\") " pod="openshift-config-operator/openshift-config-operator-7777fb866f-vmdsp" Dec 13 06:51:33 crc kubenswrapper[4971]: I1213 06:51:33.471573 4971 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-kjkp6\" (UniqueName: \"kubernetes.io/projected/817cbb9d-ed5b-4955-8b6a-4b33155b744c-kube-api-access-kjkp6\") pod \"cluster-samples-operator-665b6dd947-mlbwt\" (UID: \"817cbb9d-ed5b-4955-8b6a-4b33155b744c\") " pod="openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-mlbwt" Dec 13 06:51:33 crc kubenswrapper[4971]: I1213 06:51:33.471607 4971 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"console-oauth-config\" (UniqueName: \"kubernetes.io/secret/9f0db656-7a94-415b-b737-30ed342817ee-console-oauth-config\") pod \"console-f9d7485db-5fd9b\" (UID: \"9f0db656-7a94-415b-b737-30ed342817ee\") " pod="openshift-console/console-f9d7485db-5fd9b" Dec 13 06:51:33 crc kubenswrapper[4971]: I1213 06:51:33.471631 4971 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/88850e0f-df47-4650-be7e-d14a83ce05f4-config\") pod \"machine-approver-56656f9798-s75fl\" (UID: \"88850e0f-df47-4650-be7e-d14a83ce05f4\") " pod="openshift-cluster-machine-approver/machine-approver-56656f9798-s75fl" Dec 13 06:51:33 crc kubenswrapper[4971]: I1213 06:51:33.471650 4971 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/97219ee9-4c52-49aa-903d-044a079d4ecb-serving-cert\") pod \"openshift-apiserver-operator-796bbdcf4f-7ddbb\" (UID: \"97219ee9-4c52-49aa-903d-044a079d4ecb\") " pod="openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-7ddbb" Dec 13 06:51:33 crc kubenswrapper[4971]: I1213 06:51:33.471675 4971 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/7148e573-e2f5-4fc5-b30c-4e777516509e-proxy-ca-bundles\") pod \"controller-manager-879f6c89f-xtxr8\" (UID: \"7148e573-e2f5-4fc5-b30c-4e777516509e\") " pod="openshift-controller-manager/controller-manager-879f6c89f-xtxr8" Dec 13 06:51:33 crc kubenswrapper[4971]: I1213 06:51:33.471700 4971 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"node-pullsecrets\" (UniqueName: \"kubernetes.io/host-path/bec6e53b-3998-4402-b77b-87ff79f31827-node-pullsecrets\") pod \"apiserver-76f77b778f-2n4cq\" (UID: \"bec6e53b-3998-4402-b77b-87ff79f31827\") " pod="openshift-apiserver/apiserver-76f77b778f-2n4cq" Dec 13 06:51:33 crc kubenswrapper[4971]: I1213 06:51:33.471723 4971 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/bec6e53b-3998-4402-b77b-87ff79f31827-serving-cert\") pod \"apiserver-76f77b778f-2n4cq\" (UID: \"bec6e53b-3998-4402-b77b-87ff79f31827\") " pod="openshift-apiserver/apiserver-76f77b778f-2n4cq" Dec 13 06:51:33 crc kubenswrapper[4971]: I1213 06:51:33.471744 4971 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-8m6vz\" (UniqueName: \"kubernetes.io/projected/88850e0f-df47-4650-be7e-d14a83ce05f4-kube-api-access-8m6vz\") pod \"machine-approver-56656f9798-s75fl\" (UID: \"88850e0f-df47-4650-be7e-d14a83ce05f4\") " pod="openshift-cluster-machine-approver/machine-approver-56656f9798-s75fl" Dec 13 06:51:33 crc kubenswrapper[4971]: I1213 06:51:33.471836 4971 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-system-service-ca\" (UniqueName: \"kubernetes.io/configmap/23e16ced-5b54-4a51-a1bc-4d11c907c90a-v4-0-config-system-service-ca\") pod \"oauth-openshift-558db77b4-nzzbm\" (UID: \"23e16ced-5b54-4a51-a1bc-4d11c907c90a\") " pod="openshift-authentication/oauth-openshift-558db77b4-nzzbm" Dec 13 06:51:33 crc kubenswrapper[4971]: I1213 06:51:33.471865 4971 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/bec6e53b-3998-4402-b77b-87ff79f31827-etcd-client\") pod \"apiserver-76f77b778f-2n4cq\" (UID: \"bec6e53b-3998-4402-b77b-87ff79f31827\") " pod="openshift-apiserver/apiserver-76f77b778f-2n4cq" Dec 13 06:51:33 crc kubenswrapper[4971]: I1213 06:51:33.471873 4971 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/56c62b80-0a33-4156-b52b-cb54c98a4343-etcd-client\") pod \"etcd-operator-b45778765-6xfwq\" (UID: \"56c62b80-0a33-4156-b52b-cb54c98a4343\") " pod="openshift-etcd-operator/etcd-operator-b45778765-6xfwq" Dec 13 06:51:33 crc kubenswrapper[4971]: I1213 06:51:33.471939 4971 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etcd-ca\" (UniqueName: \"kubernetes.io/configmap/56c62b80-0a33-4156-b52b-cb54c98a4343-etcd-ca\") pod \"etcd-operator-b45778765-6xfwq\" (UID: \"56c62b80-0a33-4156-b52b-cb54c98a4343\") " pod="openshift-etcd-operator/etcd-operator-b45778765-6xfwq" Dec 13 06:51:33 crc kubenswrapper[4971]: I1213 06:51:33.471976 4971 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-zvw59\" (UniqueName: \"kubernetes.io/projected/bec6e53b-3998-4402-b77b-87ff79f31827-kube-api-access-zvw59\") pod \"apiserver-76f77b778f-2n4cq\" (UID: \"bec6e53b-3998-4402-b77b-87ff79f31827\") " pod="openshift-apiserver/apiserver-76f77b778f-2n4cq" Dec 13 06:51:33 crc kubenswrapper[4971]: I1213 06:51:33.472002 4971 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"mcc-auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/2f87c67a-fcf1-46e3-9bbd-9bf03ee55da3-mcc-auth-proxy-config\") pod \"machine-config-controller-84d6567774-kf7sp\" (UID: \"2f87c67a-fcf1-46e3-9bbd-9bf03ee55da3\") " pod="openshift-machine-config-operator/machine-config-controller-84d6567774-kf7sp" Dec 13 06:51:33 crc kubenswrapper[4971]: I1213 06:51:33.472029 4971 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/bec6e53b-3998-4402-b77b-87ff79f31827-audit-dir\") pod \"apiserver-76f77b778f-2n4cq\" (UID: \"bec6e53b-3998-4402-b77b-87ff79f31827\") " pod="openshift-apiserver/apiserver-76f77b778f-2n4cq" Dec 13 06:51:33 crc kubenswrapper[4971]: I1213 06:51:33.472053 4971 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"image-registry-operator-tls\" (UniqueName: \"kubernetes.io/secret/8294e250-74ea-497c-b6ea-08b957184eec-image-registry-operator-tls\") pod \"cluster-image-registry-operator-dc59b4c8b-zqt2t\" (UID: \"8294e250-74ea-497c-b6ea-08b957184eec\") " pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-zqt2t" Dec 13 06:51:33 crc kubenswrapper[4971]: I1213 06:51:33.472078 4971 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-tjjmd\" (UniqueName: \"kubernetes.io/projected/8e7aefba-ae2e-4069-892e-544c85b9af15-kube-api-access-tjjmd\") pod \"apiserver-7bbb656c7d-mb2q7\" (UID: \"8e7aefba-ae2e-4069-892e-544c85b9af15\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-mb2q7" Dec 13 06:51:33 crc kubenswrapper[4971]: I1213 06:51:33.472105 4971 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-lmhvx\" (UniqueName: \"kubernetes.io/projected/7148e573-e2f5-4fc5-b30c-4e777516509e-kube-api-access-lmhvx\") pod \"controller-manager-879f6c89f-xtxr8\" (UID: \"7148e573-e2f5-4fc5-b30c-4e777516509e\") " pod="openshift-controller-manager/controller-manager-879f6c89f-xtxr8" Dec 13 06:51:33 crc kubenswrapper[4971]: I1213 06:51:33.472128 4971 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-87jr5\" (UniqueName: \"kubernetes.io/projected/c45a314c-77da-43fc-b2ae-9de350540c2b-kube-api-access-87jr5\") pod \"machine-api-operator-5694c8668f-t24dm\" (UID: \"c45a314c-77da-43fc-b2ae-9de350540c2b\") " pod="openshift-machine-api/machine-api-operator-5694c8668f-t24dm" Dec 13 06:51:33 crc kubenswrapper[4971]: I1213 06:51:33.472156 4971 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/bb0ead7b-eba0-4d16-9363-df54caee0a32-serving-cert\") pod \"openshift-controller-manager-operator-756b6f6bc6-gtrwg\" (UID: \"bb0ead7b-eba0-4d16-9363-df54caee0a32\") " pod="openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-gtrwg" Dec 13 06:51:33 crc kubenswrapper[4971]: I1213 06:51:33.472178 4971 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"audit\" (UniqueName: \"kubernetes.io/configmap/bec6e53b-3998-4402-b77b-87ff79f31827-audit\") pod \"apiserver-76f77b778f-2n4cq\" (UID: \"bec6e53b-3998-4402-b77b-87ff79f31827\") " pod="openshift-apiserver/apiserver-76f77b778f-2n4cq" Dec 13 06:51:33 crc kubenswrapper[4971]: I1213 06:51:33.472198 4971 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"encryption-config\" (UniqueName: \"kubernetes.io/secret/bec6e53b-3998-4402-b77b-87ff79f31827-encryption-config\") pod \"apiserver-76f77b778f-2n4cq\" (UID: \"bec6e53b-3998-4402-b77b-87ff79f31827\") " pod="openshift-apiserver/apiserver-76f77b778f-2n4cq" Dec 13 06:51:33 crc kubenswrapper[4971]: I1213 06:51:33.472222 4971 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-tb7rw\" (UniqueName: \"kubernetes.io/projected/8294e250-74ea-497c-b6ea-08b957184eec-kube-api-access-tb7rw\") pod \"cluster-image-registry-operator-dc59b4c8b-zqt2t\" (UID: \"8294e250-74ea-497c-b6ea-08b957184eec\") " pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-zqt2t" Dec 13 06:51:33 crc kubenswrapper[4971]: I1213 06:51:33.472245 4971 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/9f0db656-7a94-415b-b737-30ed342817ee-trusted-ca-bundle\") pod \"console-f9d7485db-5fd9b\" (UID: \"9f0db656-7a94-415b-b737-30ed342817ee\") " pod="openshift-console/console-f9d7485db-5fd9b" Dec 13 06:51:33 crc kubenswrapper[4971]: I1213 06:51:33.472266 4971 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"oauth-serving-cert\" (UniqueName: \"kubernetes.io/configmap/9f0db656-7a94-415b-b737-30ed342817ee-oauth-serving-cert\") pod \"console-f9d7485db-5fd9b\" (UID: \"9f0db656-7a94-415b-b737-30ed342817ee\") " pod="openshift-console/console-f9d7485db-5fd9b" Dec 13 06:51:33 crc kubenswrapper[4971]: I1213 06:51:33.472289 4971 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-vbcwz\" (UniqueName: \"kubernetes.io/projected/dcdf01b4-83e9-4d28-821a-e4321ce6b92c-kube-api-access-vbcwz\") pod \"service-ca-9c57cc56f-q99rh\" (UID: \"dcdf01b4-83e9-4d28-821a-e4321ce6b92c\") " pod="openshift-service-ca/service-ca-9c57cc56f-q99rh" Dec 13 06:51:33 crc kubenswrapper[4971]: I1213 06:51:33.472310 4971 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/56c62b80-0a33-4156-b52b-cb54c98a4343-config\") pod \"etcd-operator-b45778765-6xfwq\" (UID: \"56c62b80-0a33-4156-b52b-cb54c98a4343\") " pod="openshift-etcd-operator/etcd-operator-b45778765-6xfwq" Dec 13 06:51:33 crc kubenswrapper[4971]: I1213 06:51:33.472333 4971 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"machine-approver-tls\" (UniqueName: \"kubernetes.io/secret/88850e0f-df47-4650-be7e-d14a83ce05f4-machine-approver-tls\") pod \"machine-approver-56656f9798-s75fl\" (UID: \"88850e0f-df47-4650-be7e-d14a83ce05f4\") " pod="openshift-cluster-machine-approver/machine-approver-56656f9798-s75fl" Dec 13 06:51:33 crc kubenswrapper[4971]: I1213 06:51:33.472348 4971 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/8e7aefba-ae2e-4069-892e-544c85b9af15-etcd-client\") pod \"apiserver-7bbb656c7d-mb2q7\" (UID: \"8e7aefba-ae2e-4069-892e-544c85b9af15\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-mb2q7" Dec 13 06:51:33 crc kubenswrapper[4971]: I1213 06:51:33.472430 4971 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-96z5g\" (UniqueName: \"kubernetes.io/projected/bb0ead7b-eba0-4d16-9363-df54caee0a32-kube-api-access-96z5g\") pod \"openshift-controller-manager-operator-756b6f6bc6-gtrwg\" (UID: \"bb0ead7b-eba0-4d16-9363-df54caee0a32\") " pod="openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-gtrwg" Dec 13 06:51:33 crc kubenswrapper[4971]: I1213 06:51:33.472453 4971 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/88850e0f-df47-4650-be7e-d14a83ce05f4-auth-proxy-config\") pod \"machine-approver-56656f9798-s75fl\" (UID: \"88850e0f-df47-4650-be7e-d14a83ce05f4\") " pod="openshift-cluster-machine-approver/machine-approver-56656f9798-s75fl" Dec 13 06:51:33 crc kubenswrapper[4971]: I1213 06:51:33.472470 4971 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/2f87c67a-fcf1-46e3-9bbd-9bf03ee55da3-proxy-tls\") pod \"machine-config-controller-84d6567774-kf7sp\" (UID: \"2f87c67a-fcf1-46e3-9bbd-9bf03ee55da3\") " pod="openshift-machine-config-operator/machine-config-controller-84d6567774-kf7sp" Dec 13 06:51:33 crc kubenswrapper[4971]: I1213 06:51:33.472488 4971 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-system-router-certs\" (UniqueName: \"kubernetes.io/secret/23e16ced-5b54-4a51-a1bc-4d11c907c90a-v4-0-config-system-router-certs\") pod \"oauth-openshift-558db77b4-nzzbm\" (UID: \"23e16ced-5b54-4a51-a1bc-4d11c907c90a\") " pod="openshift-authentication/oauth-openshift-558db77b4-nzzbm" Dec 13 06:51:33 crc kubenswrapper[4971]: I1213 06:51:33.472506 4971 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/7148e573-e2f5-4fc5-b30c-4e777516509e-client-ca\") pod \"controller-manager-879f6c89f-xtxr8\" (UID: \"7148e573-e2f5-4fc5-b30c-4e777516509e\") " pod="openshift-controller-manager/controller-manager-879f6c89f-xtxr8" Dec 13 06:51:33 crc kubenswrapper[4971]: I1213 06:51:33.472544 4971 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-system-session\" (UniqueName: \"kubernetes.io/secret/23e16ced-5b54-4a51-a1bc-4d11c907c90a-v4-0-config-system-session\") pod \"oauth-openshift-558db77b4-nzzbm\" (UID: \"23e16ced-5b54-4a51-a1bc-4d11c907c90a\") " pod="openshift-authentication/oauth-openshift-558db77b4-nzzbm" Dec 13 06:51:33 crc kubenswrapper[4971]: I1213 06:51:33.472571 4971 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/8e7aefba-ae2e-4069-892e-544c85b9af15-audit-dir\") pod \"apiserver-7bbb656c7d-mb2q7\" (UID: \"8e7aefba-ae2e-4069-892e-544c85b9af15\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-mb2q7" Dec 13 06:51:33 crc kubenswrapper[4971]: I1213 06:51:33.472590 4971 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-system-cliconfig\" (UniqueName: \"kubernetes.io/configmap/23e16ced-5b54-4a51-a1bc-4d11c907c90a-v4-0-config-system-cliconfig\") pod \"oauth-openshift-558db77b4-nzzbm\" (UID: \"23e16ced-5b54-4a51-a1bc-4d11c907c90a\") " pod="openshift-authentication/oauth-openshift-558db77b4-nzzbm" Dec 13 06:51:33 crc kubenswrapper[4971]: I1213 06:51:33.472608 4971 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"signing-key\" (UniqueName: \"kubernetes.io/secret/dcdf01b4-83e9-4d28-821a-e4321ce6b92c-signing-key\") pod \"service-ca-9c57cc56f-q99rh\" (UID: \"dcdf01b4-83e9-4d28-821a-e4321ce6b92c\") " pod="openshift-service-ca/service-ca-9c57cc56f-q99rh" Dec 13 06:51:33 crc kubenswrapper[4971]: I1213 06:51:33.472625 4971 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"marketplace-trusted-ca\" (UniqueName: \"kubernetes.io/configmap/f14c8a94-8589-4059-bddf-329452355fab-marketplace-trusted-ca\") pod \"marketplace-operator-79b997595-j9zqf\" (UID: \"f14c8a94-8589-4059-bddf-329452355fab\") " pod="openshift-marketplace/marketplace-operator-79b997595-j9zqf" Dec 13 06:51:33 crc kubenswrapper[4971]: I1213 06:51:33.472647 4971 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"marketplace-operator-metrics\" (UniqueName: \"kubernetes.io/secret/f14c8a94-8589-4059-bddf-329452355fab-marketplace-operator-metrics\") pod \"marketplace-operator-79b997595-j9zqf\" (UID: \"f14c8a94-8589-4059-bddf-329452355fab\") " pod="openshift-marketplace/marketplace-operator-79b997595-j9zqf" Dec 13 06:51:33 crc kubenswrapper[4971]: I1213 06:51:33.472667 4971 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-user-template-provider-selection\" (UniqueName: \"kubernetes.io/secret/23e16ced-5b54-4a51-a1bc-4d11c907c90a-v4-0-config-user-template-provider-selection\") pod \"oauth-openshift-558db77b4-nzzbm\" (UID: \"23e16ced-5b54-4a51-a1bc-4d11c907c90a\") " pod="openshift-authentication/oauth-openshift-558db77b4-nzzbm" Dec 13 06:51:33 crc kubenswrapper[4971]: I1213 06:51:33.472688 4971 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"encryption-config\" (UniqueName: \"kubernetes.io/secret/8e7aefba-ae2e-4069-892e-544c85b9af15-encryption-config\") pod \"apiserver-7bbb656c7d-mb2q7\" (UID: \"8e7aefba-ae2e-4069-892e-544c85b9af15\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-mb2q7" Dec 13 06:51:33 crc kubenswrapper[4971]: I1213 06:51:33.472704 4971 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etcd-service-ca\" (UniqueName: \"kubernetes.io/configmap/56c62b80-0a33-4156-b52b-cb54c98a4343-etcd-service-ca\") pod \"etcd-operator-b45778765-6xfwq\" (UID: \"56c62b80-0a33-4156-b52b-cb54c98a4343\") " pod="openshift-etcd-operator/etcd-operator-b45778765-6xfwq" Dec 13 06:51:33 crc kubenswrapper[4971]: I1213 06:51:33.472881 4971 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/bec6e53b-3998-4402-b77b-87ff79f31827-audit-dir\") pod \"apiserver-76f77b778f-2n4cq\" (UID: \"bec6e53b-3998-4402-b77b-87ff79f31827\") " pod="openshift-apiserver/apiserver-76f77b778f-2n4cq" Dec 13 06:51:33 crc kubenswrapper[4971]: I1213 06:51:33.473666 4971 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-system-ocp-branding-template\" (UniqueName: \"kubernetes.io/secret/23e16ced-5b54-4a51-a1bc-4d11c907c90a-v4-0-config-system-ocp-branding-template\") pod \"oauth-openshift-558db77b4-nzzbm\" (UID: \"23e16ced-5b54-4a51-a1bc-4d11c907c90a\") " pod="openshift-authentication/oauth-openshift-558db77b4-nzzbm" Dec 13 06:51:33 crc kubenswrapper[4971]: I1213 06:51:33.473961 4971 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-system-serving-cert\" (UniqueName: \"kubernetes.io/secret/23e16ced-5b54-4a51-a1bc-4d11c907c90a-v4-0-config-system-serving-cert\") pod \"oauth-openshift-558db77b4-nzzbm\" (UID: \"23e16ced-5b54-4a51-a1bc-4d11c907c90a\") " pod="openshift-authentication/oauth-openshift-558db77b4-nzzbm" Dec 13 06:51:33 crc kubenswrapper[4971]: I1213 06:51:33.474442 4971 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-system-cliconfig\" (UniqueName: \"kubernetes.io/configmap/23e16ced-5b54-4a51-a1bc-4d11c907c90a-v4-0-config-system-cliconfig\") pod \"oauth-openshift-558db77b4-nzzbm\" (UID: \"23e16ced-5b54-4a51-a1bc-4d11c907c90a\") " pod="openshift-authentication/oauth-openshift-558db77b4-nzzbm" Dec 13 06:51:33 crc kubenswrapper[4971]: I1213 06:51:33.474468 4971 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/88850e0f-df47-4650-be7e-d14a83ce05f4-auth-proxy-config\") pod \"machine-approver-56656f9798-s75fl\" (UID: \"88850e0f-df47-4650-be7e-d14a83ce05f4\") " pod="openshift-cluster-machine-approver/machine-approver-56656f9798-s75fl" Dec 13 06:51:33 crc kubenswrapper[4971]: I1213 06:51:33.475214 4971 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"samples-operator-tls\" (UniqueName: \"kubernetes.io/secret/817cbb9d-ed5b-4955-8b6a-4b33155b744c-samples-operator-tls\") pod \"cluster-samples-operator-665b6dd947-mlbwt\" (UID: \"817cbb9d-ed5b-4955-8b6a-4b33155b744c\") " pod="openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-mlbwt" Dec 13 06:51:33 crc kubenswrapper[4971]: I1213 06:51:33.475490 4971 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"node-pullsecrets\" (UniqueName: \"kubernetes.io/host-path/bec6e53b-3998-4402-b77b-87ff79f31827-node-pullsecrets\") pod \"apiserver-76f77b778f-2n4cq\" (UID: \"bec6e53b-3998-4402-b77b-87ff79f31827\") " pod="openshift-apiserver/apiserver-76f77b778f-2n4cq" Dec 13 06:51:33 crc kubenswrapper[4971]: I1213 06:51:33.475994 4971 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/88850e0f-df47-4650-be7e-d14a83ce05f4-config\") pod \"machine-approver-56656f9798-s75fl\" (UID: \"88850e0f-df47-4650-be7e-d14a83ce05f4\") " pod="openshift-cluster-machine-approver/machine-approver-56656f9798-s75fl" Dec 13 06:51:33 crc kubenswrapper[4971]: I1213 06:51:33.477958 4971 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/8e7aefba-ae2e-4069-892e-544c85b9af15-audit-dir\") pod \"apiserver-7bbb656c7d-mb2q7\" (UID: \"8e7aefba-ae2e-4069-892e-544c85b9af15\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-mb2q7" Dec 13 06:51:33 crc kubenswrapper[4971]: I1213 06:51:33.478435 4971 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-system-service-ca\" (UniqueName: \"kubernetes.io/configmap/23e16ced-5b54-4a51-a1bc-4d11c907c90a-v4-0-config-system-service-ca\") pod \"oauth-openshift-558db77b4-nzzbm\" (UID: \"23e16ced-5b54-4a51-a1bc-4d11c907c90a\") " pod="openshift-authentication/oauth-openshift-558db77b4-nzzbm" Dec 13 06:51:33 crc kubenswrapper[4971]: I1213 06:51:33.478958 4971 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/97219ee9-4c52-49aa-903d-044a079d4ecb-serving-cert\") pod \"openshift-apiserver-operator-796bbdcf4f-7ddbb\" (UID: \"97219ee9-4c52-49aa-903d-044a079d4ecb\") " pod="openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-7ddbb" Dec 13 06:51:33 crc kubenswrapper[4971]: I1213 06:51:33.479239 4971 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-user-idp-0-file-data\" (UniqueName: \"kubernetes.io/secret/23e16ced-5b54-4a51-a1bc-4d11c907c90a-v4-0-config-user-idp-0-file-data\") pod \"oauth-openshift-558db77b4-nzzbm\" (UID: \"23e16ced-5b54-4a51-a1bc-4d11c907c90a\") " pod="openshift-authentication/oauth-openshift-558db77b4-nzzbm" Dec 13 06:51:33 crc kubenswrapper[4971]: I1213 06:51:33.479354 4971 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/7148e573-e2f5-4fc5-b30c-4e777516509e-proxy-ca-bundles\") pod \"controller-manager-879f6c89f-xtxr8\" (UID: \"7148e573-e2f5-4fc5-b30c-4e777516509e\") " pod="openshift-controller-manager/controller-manager-879f6c89f-xtxr8" Dec 13 06:51:33 crc kubenswrapper[4971]: I1213 06:51:33.479691 4971 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"machine-api-operator-tls\" (UniqueName: \"kubernetes.io/secret/c45a314c-77da-43fc-b2ae-9de350540c2b-machine-api-operator-tls\") pod \"machine-api-operator-5694c8668f-t24dm\" (UID: \"c45a314c-77da-43fc-b2ae-9de350540c2b\") " pod="openshift-machine-api/machine-api-operator-5694c8668f-t24dm" Dec 13 06:51:33 crc kubenswrapper[4971]: I1213 06:51:33.479913 4971 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/7148e573-e2f5-4fc5-b30c-4e777516509e-client-ca\") pod \"controller-manager-879f6c89f-xtxr8\" (UID: \"7148e573-e2f5-4fc5-b30c-4e777516509e\") " pod="openshift-controller-manager/controller-manager-879f6c89f-xtxr8" Dec 13 06:51:33 crc kubenswrapper[4971]: I1213 06:51:33.480019 4971 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-user-template-login\" (UniqueName: \"kubernetes.io/secret/23e16ced-5b54-4a51-a1bc-4d11c907c90a-v4-0-config-user-template-login\") pod \"oauth-openshift-558db77b4-nzzbm\" (UID: \"23e16ced-5b54-4a51-a1bc-4d11c907c90a\") " pod="openshift-authentication/oauth-openshift-558db77b4-nzzbm" Dec 13 06:51:33 crc kubenswrapper[4971]: I1213 06:51:33.480140 4971 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"console-oauth-config\" (UniqueName: \"kubernetes.io/secret/9f0db656-7a94-415b-b737-30ed342817ee-console-oauth-config\") pod \"console-f9d7485db-5fd9b\" (UID: \"9f0db656-7a94-415b-b737-30ed342817ee\") " pod="openshift-console/console-f9d7485db-5fd9b" Dec 13 06:51:33 crc kubenswrapper[4971]: I1213 06:51:33.480294 4971 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/8e7aefba-ae2e-4069-892e-544c85b9af15-etcd-client\") pod \"apiserver-7bbb656c7d-mb2q7\" (UID: \"8e7aefba-ae2e-4069-892e-544c85b9af15\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-mb2q7" Dec 13 06:51:33 crc kubenswrapper[4971]: I1213 06:51:33.480422 4971 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-multus/multus-admission-controller-857f4d67dd-z68ks"] Dec 13 06:51:33 crc kubenswrapper[4971]: I1213 06:51:33.480558 4971 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"oauth-serving-cert\" (UniqueName: \"kubernetes.io/configmap/9f0db656-7a94-415b-b737-30ed342817ee-oauth-serving-cert\") pod \"console-f9d7485db-5fd9b\" (UID: \"9f0db656-7a94-415b-b737-30ed342817ee\") " pod="openshift-console/console-f9d7485db-5fd9b" Dec 13 06:51:33 crc kubenswrapper[4971]: I1213 06:51:33.480673 4971 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/9f0db656-7a94-415b-b737-30ed342817ee-trusted-ca-bundle\") pod \"console-f9d7485db-5fd9b\" (UID: \"9f0db656-7a94-415b-b737-30ed342817ee\") " pod="openshift-console/console-f9d7485db-5fd9b" Dec 13 06:51:33 crc kubenswrapper[4971]: I1213 06:51:33.481201 4971 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"audit\" (UniqueName: \"kubernetes.io/configmap/bec6e53b-3998-4402-b77b-87ff79f31827-audit\") pod \"apiserver-76f77b778f-2n4cq\" (UID: \"bec6e53b-3998-4402-b77b-87ff79f31827\") " pod="openshift-apiserver/apiserver-76f77b778f-2n4cq" Dec 13 06:51:33 crc kubenswrapper[4971]: I1213 06:51:33.482387 4971 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-7ddbb"] Dec 13 06:51:33 crc kubenswrapper[4971]: I1213 06:51:33.483027 4971 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-system-router-certs\" (UniqueName: \"kubernetes.io/secret/23e16ced-5b54-4a51-a1bc-4d11c907c90a-v4-0-config-system-router-certs\") pod \"oauth-openshift-558db77b4-nzzbm\" (UID: \"23e16ced-5b54-4a51-a1bc-4d11c907c90a\") " pod="openshift-authentication/oauth-openshift-558db77b4-nzzbm" Dec 13 06:51:33 crc kubenswrapper[4971]: I1213 06:51:33.483270 4971 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/7148e573-e2f5-4fc5-b30c-4e777516509e-serving-cert\") pod \"controller-manager-879f6c89f-xtxr8\" (UID: \"7148e573-e2f5-4fc5-b30c-4e777516509e\") " pod="openshift-controller-manager/controller-manager-879f6c89f-xtxr8" Dec 13 06:51:33 crc kubenswrapper[4971]: I1213 06:51:33.483573 4971 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-user-template-provider-selection\" (UniqueName: \"kubernetes.io/secret/23e16ced-5b54-4a51-a1bc-4d11c907c90a-v4-0-config-user-template-provider-selection\") pod \"oauth-openshift-558db77b4-nzzbm\" (UID: \"23e16ced-5b54-4a51-a1bc-4d11c907c90a\") " pod="openshift-authentication/oauth-openshift-558db77b4-nzzbm" Dec 13 06:51:33 crc kubenswrapper[4971]: I1213 06:51:33.483577 4971 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/874901b0-9650-4029-8d21-74441f40e201-serving-cert\") pod \"openshift-config-operator-7777fb866f-vmdsp\" (UID: \"874901b0-9650-4029-8d21-74441f40e201\") " pod="openshift-config-operator/openshift-config-operator-7777fb866f-vmdsp" Dec 13 06:51:33 crc kubenswrapper[4971]: I1213 06:51:33.484723 4971 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/bec6e53b-3998-4402-b77b-87ff79f31827-serving-cert\") pod \"apiserver-76f77b778f-2n4cq\" (UID: \"bec6e53b-3998-4402-b77b-87ff79f31827\") " pod="openshift-apiserver/apiserver-76f77b778f-2n4cq" Dec 13 06:51:33 crc kubenswrapper[4971]: I1213 06:51:33.485352 4971 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"console-serving-cert\" (UniqueName: \"kubernetes.io/secret/9f0db656-7a94-415b-b737-30ed342817ee-console-serving-cert\") pod \"console-f9d7485db-5fd9b\" (UID: \"9f0db656-7a94-415b-b737-30ed342817ee\") " pod="openshift-console/console-f9d7485db-5fd9b" Dec 13 06:51:33 crc kubenswrapper[4971]: I1213 06:51:33.485489 4971 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"encryption-config\" (UniqueName: \"kubernetes.io/secret/8e7aefba-ae2e-4069-892e-544c85b9af15-encryption-config\") pod \"apiserver-7bbb656c7d-mb2q7\" (UID: \"8e7aefba-ae2e-4069-892e-544c85b9af15\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-mb2q7" Dec 13 06:51:33 crc kubenswrapper[4971]: I1213 06:51:33.485748 4971 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"machine-approver-tls\" (UniqueName: \"kubernetes.io/secret/88850e0f-df47-4650-be7e-d14a83ce05f4-machine-approver-tls\") pod \"machine-approver-56656f9798-s75fl\" (UID: \"88850e0f-df47-4650-be7e-d14a83ce05f4\") " pod="openshift-cluster-machine-approver/machine-approver-56656f9798-s75fl" Dec 13 06:51:33 crc kubenswrapper[4971]: I1213 06:51:33.485744 4971 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-user-template-error\" (UniqueName: \"kubernetes.io/secret/23e16ced-5b54-4a51-a1bc-4d11c907c90a-v4-0-config-user-template-error\") pod \"oauth-openshift-558db77b4-nzzbm\" (UID: \"23e16ced-5b54-4a51-a1bc-4d11c907c90a\") " pod="openshift-authentication/oauth-openshift-558db77b4-nzzbm" Dec 13 06:51:33 crc kubenswrapper[4971]: I1213 06:51:33.487194 4971 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"image-registry-operator-tls\" (UniqueName: \"kubernetes.io/secret/8294e250-74ea-497c-b6ea-08b957184eec-image-registry-operator-tls\") pod \"cluster-image-registry-operator-dc59b4c8b-zqt2t\" (UID: \"8294e250-74ea-497c-b6ea-08b957184eec\") " pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-zqt2t" Dec 13 06:51:33 crc kubenswrapper[4971]: I1213 06:51:33.487718 4971 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ingress-operator"/"ingress-operator-dockercfg-7lnqk" Dec 13 06:51:33 crc kubenswrapper[4971]: I1213 06:51:33.489651 4971 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-machine-api/control-plane-machine-set-operator-78cbb6b69f-bgdbk"] Dec 13 06:51:33 crc kubenswrapper[4971]: I1213 06:51:33.489988 4971 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"encryption-config\" (UniqueName: \"kubernetes.io/secret/bec6e53b-3998-4402-b77b-87ff79f31827-encryption-config\") pod \"apiserver-76f77b778f-2n4cq\" (UID: \"bec6e53b-3998-4402-b77b-87ff79f31827\") " pod="openshift-apiserver/apiserver-76f77b778f-2n4cq" Dec 13 06:51:33 crc kubenswrapper[4971]: I1213 06:51:33.491068 4971 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-console/downloads-7954f5f757-dqk9z"] Dec 13 06:51:33 crc kubenswrapper[4971]: I1213 06:51:33.492286 4971 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-system-session\" (UniqueName: \"kubernetes.io/secret/23e16ced-5b54-4a51-a1bc-4d11c907c90a-v4-0-config-system-session\") pod \"oauth-openshift-558db77b4-nzzbm\" (UID: \"23e16ced-5b54-4a51-a1bc-4d11c907c90a\") " pod="openshift-authentication/oauth-openshift-558db77b4-nzzbm" Dec 13 06:51:33 crc kubenswrapper[4971]: I1213 06:51:33.493794 4971 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-controller-manager/controller-manager-879f6c89f-xtxr8"] Dec 13 06:51:33 crc kubenswrapper[4971]: I1213 06:51:33.493836 4971 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-cwwf9"] Dec 13 06:51:33 crc kubenswrapper[4971]: I1213 06:51:33.495661 4971 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-machine-config-operator/machine-config-operator-74547568cd-4cn2z"] Dec 13 06:51:33 crc kubenswrapper[4971]: I1213 06:51:33.496509 4971 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-zqt2t"] Dec 13 06:51:33 crc kubenswrapper[4971]: I1213 06:51:33.497538 4971 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-xfvfg"] Dec 13 06:51:33 crc kubenswrapper[4971]: I1213 06:51:33.498986 4971 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-image-registry/image-registry-697d97f7c8-mdd8q"] Dec 13 06:51:33 crc kubenswrapper[4971]: I1213 06:51:33.499331 4971 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/8e7aefba-ae2e-4069-892e-544c85b9af15-serving-cert\") pod \"apiserver-7bbb656c7d-mb2q7\" (UID: \"8e7aefba-ae2e-4069-892e-544c85b9af15\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-mb2q7" Dec 13 06:51:33 crc kubenswrapper[4971]: I1213 06:51:33.500292 4971 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-machine-config-operator/machine-config-controller-84d6567774-kf7sp"] Dec 13 06:51:33 crc kubenswrapper[4971]: I1213 06:51:33.501658 4971 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-dns-operator/dns-operator-744455d44c-b999k"] Dec 13 06:51:33 crc kubenswrapper[4971]: I1213 06:51:33.503358 4971 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-config-operator/openshift-config-operator-7777fb866f-vmdsp"] Dec 13 06:51:33 crc kubenswrapper[4971]: I1213 06:51:33.503409 4971 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-etcd-operator"/"etcd-service-ca-bundle" Dec 13 06:51:33 crc kubenswrapper[4971]: I1213 06:51:33.504907 4971 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/marketplace-operator-79b997595-j9zqf"] Dec 13 06:51:33 crc kubenswrapper[4971]: I1213 06:51:33.508353 4971 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-etcd-operator/etcd-operator-b45778765-6xfwq"] Dec 13 06:51:33 crc kubenswrapper[4971]: I1213 06:51:33.510286 4971 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-lnrkm"] Dec 13 06:51:33 crc kubenswrapper[4971]: I1213 06:51:33.512611 4971 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-pxjqw"] Dec 13 06:51:33 crc kubenswrapper[4971]: I1213 06:51:33.514346 4971 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-6blxl"] Dec 13 06:51:33 crc kubenswrapper[4971]: I1213 06:51:33.517679 4971 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-dns/dns-default-dntmz"] Dec 13 06:51:33 crc kubenswrapper[4971]: I1213 06:51:33.518654 4971 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-dns/dns-default-dntmz" Dec 13 06:51:33 crc kubenswrapper[4971]: I1213 06:51:33.521947 4971 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-machine-config-operator/machine-config-server-xdxjz"] Dec 13 06:51:33 crc kubenswrapper[4971]: I1213 06:51:33.525213 4971 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ingress-operator"/"metrics-tls" Dec 13 06:51:33 crc kubenswrapper[4971]: I1213 06:51:33.532008 4971 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-console-operator/console-operator-58897d9998-kdfcd"] Dec 13 06:51:33 crc kubenswrapper[4971]: I1213 06:51:33.532046 4971 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/catalog-operator-68c6474976-l4rzp"] Dec 13 06:51:33 crc kubenswrapper[4971]: I1213 06:51:33.532158 4971 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-machine-config-operator/machine-config-server-xdxjz" Dec 13 06:51:33 crc kubenswrapper[4971]: I1213 06:51:33.534295 4971 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-storage-version-migrator/migrator-59844c95c7-z2p8z"] Dec 13 06:51:33 crc kubenswrapper[4971]: I1213 06:51:33.537667 4971 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29426805-fxxrz"] Dec 13 06:51:33 crc kubenswrapper[4971]: I1213 06:51:33.539232 4971 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-service-ca/service-ca-9c57cc56f-q99rh"] Dec 13 06:51:33 crc kubenswrapper[4971]: I1213 06:51:33.540989 4971 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-fz6wk"] Dec 13 06:51:33 crc kubenswrapper[4971]: I1213 06:51:33.542878 4971 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["hostpath-provisioner/csi-hostpathplugin-7nhhk"] Dec 13 06:51:33 crc kubenswrapper[4971]: I1213 06:51:33.544788 4971 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-g8vfp"] Dec 13 06:51:33 crc kubenswrapper[4971]: I1213 06:51:33.545011 4971 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-etcd-operator"/"etcd-operator-dockercfg-r9srn" Dec 13 06:51:33 crc kubenswrapper[4971]: I1213 06:51:33.546453 4971 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-dns/dns-default-dntmz"] Dec 13 06:51:33 crc kubenswrapper[4971]: I1213 06:51:33.547737 4971 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-service-ca-operator/service-ca-operator-777779d784-drq5g"] Dec 13 06:51:33 crc kubenswrapper[4971]: I1213 06:51:33.549187 4971 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-authentication-operator/authentication-operator-69f744f599-wjdd5"] Dec 13 06:51:33 crc kubenswrapper[4971]: I1213 06:51:33.550253 4971 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-ingress-canary/ingress-canary-ffhlq"] Dec 13 06:51:33 crc kubenswrapper[4971]: I1213 06:51:33.551018 4971 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-ingress-canary/ingress-canary-ffhlq" Dec 13 06:51:33 crc kubenswrapper[4971]: I1213 06:51:33.551376 4971 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-ingress-canary/ingress-canary-ffhlq"] Dec 13 06:51:33 crc kubenswrapper[4971]: I1213 06:51:33.564103 4971 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-etcd-operator"/"etcd-operator-serving-cert" Dec 13 06:51:33 crc kubenswrapper[4971]: I1213 06:51:33.573599 4971 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"signing-cabundle\" (UniqueName: \"kubernetes.io/configmap/dcdf01b4-83e9-4d28-821a-e4321ce6b92c-signing-cabundle\") pod \"service-ca-9c57cc56f-q99rh\" (UID: \"dcdf01b4-83e9-4d28-821a-e4321ce6b92c\") " pod="openshift-service-ca/service-ca-9c57cc56f-q99rh" Dec 13 06:51:33 crc kubenswrapper[4971]: I1213 06:51:33.573676 4971 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/56c62b80-0a33-4156-b52b-cb54c98a4343-etcd-client\") pod \"etcd-operator-b45778765-6xfwq\" (UID: \"56c62b80-0a33-4156-b52b-cb54c98a4343\") " pod="openshift-etcd-operator/etcd-operator-b45778765-6xfwq" Dec 13 06:51:33 crc kubenswrapper[4971]: I1213 06:51:33.575253 4971 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"mcc-auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/2f87c67a-fcf1-46e3-9bbd-9bf03ee55da3-mcc-auth-proxy-config\") pod \"machine-config-controller-84d6567774-kf7sp\" (UID: \"2f87c67a-fcf1-46e3-9bbd-9bf03ee55da3\") " pod="openshift-machine-config-operator/machine-config-controller-84d6567774-kf7sp" Dec 13 06:51:33 crc kubenswrapper[4971]: I1213 06:51:33.575337 4971 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etcd-ca\" (UniqueName: \"kubernetes.io/configmap/56c62b80-0a33-4156-b52b-cb54c98a4343-etcd-ca\") pod \"etcd-operator-b45778765-6xfwq\" (UID: \"56c62b80-0a33-4156-b52b-cb54c98a4343\") " pod="openshift-etcd-operator/etcd-operator-b45778765-6xfwq" Dec 13 06:51:33 crc kubenswrapper[4971]: I1213 06:51:33.575556 4971 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/bb0ead7b-eba0-4d16-9363-df54caee0a32-serving-cert\") pod \"openshift-controller-manager-operator-756b6f6bc6-gtrwg\" (UID: \"bb0ead7b-eba0-4d16-9363-df54caee0a32\") " pod="openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-gtrwg" Dec 13 06:51:33 crc kubenswrapper[4971]: I1213 06:51:33.575711 4971 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-vbcwz\" (UniqueName: \"kubernetes.io/projected/dcdf01b4-83e9-4d28-821a-e4321ce6b92c-kube-api-access-vbcwz\") pod \"service-ca-9c57cc56f-q99rh\" (UID: \"dcdf01b4-83e9-4d28-821a-e4321ce6b92c\") " pod="openshift-service-ca/service-ca-9c57cc56f-q99rh" Dec 13 06:51:33 crc kubenswrapper[4971]: I1213 06:51:33.575790 4971 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/56c62b80-0a33-4156-b52b-cb54c98a4343-config\") pod \"etcd-operator-b45778765-6xfwq\" (UID: \"56c62b80-0a33-4156-b52b-cb54c98a4343\") " pod="openshift-etcd-operator/etcd-operator-b45778765-6xfwq" Dec 13 06:51:33 crc kubenswrapper[4971]: I1213 06:51:33.575881 4971 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/2f87c67a-fcf1-46e3-9bbd-9bf03ee55da3-proxy-tls\") pod \"machine-config-controller-84d6567774-kf7sp\" (UID: \"2f87c67a-fcf1-46e3-9bbd-9bf03ee55da3\") " pod="openshift-machine-config-operator/machine-config-controller-84d6567774-kf7sp" Dec 13 06:51:33 crc kubenswrapper[4971]: I1213 06:51:33.576156 4971 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"mcc-auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/2f87c67a-fcf1-46e3-9bbd-9bf03ee55da3-mcc-auth-proxy-config\") pod \"machine-config-controller-84d6567774-kf7sp\" (UID: \"2f87c67a-fcf1-46e3-9bbd-9bf03ee55da3\") " pod="openshift-machine-config-operator/machine-config-controller-84d6567774-kf7sp" Dec 13 06:51:33 crc kubenswrapper[4971]: I1213 06:51:33.576382 4971 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-96z5g\" (UniqueName: \"kubernetes.io/projected/bb0ead7b-eba0-4d16-9363-df54caee0a32-kube-api-access-96z5g\") pod \"openshift-controller-manager-operator-756b6f6bc6-gtrwg\" (UID: \"bb0ead7b-eba0-4d16-9363-df54caee0a32\") " pod="openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-gtrwg" Dec 13 06:51:33 crc kubenswrapper[4971]: I1213 06:51:33.576570 4971 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"signing-key\" (UniqueName: \"kubernetes.io/secret/dcdf01b4-83e9-4d28-821a-e4321ce6b92c-signing-key\") pod \"service-ca-9c57cc56f-q99rh\" (UID: \"dcdf01b4-83e9-4d28-821a-e4321ce6b92c\") " pod="openshift-service-ca/service-ca-9c57cc56f-q99rh" Dec 13 06:51:33 crc kubenswrapper[4971]: I1213 06:51:33.576687 4971 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"marketplace-trusted-ca\" (UniqueName: \"kubernetes.io/configmap/f14c8a94-8589-4059-bddf-329452355fab-marketplace-trusted-ca\") pod \"marketplace-operator-79b997595-j9zqf\" (UID: \"f14c8a94-8589-4059-bddf-329452355fab\") " pod="openshift-marketplace/marketplace-operator-79b997595-j9zqf" Dec 13 06:51:33 crc kubenswrapper[4971]: I1213 06:51:33.576786 4971 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"marketplace-operator-metrics\" (UniqueName: \"kubernetes.io/secret/f14c8a94-8589-4059-bddf-329452355fab-marketplace-operator-metrics\") pod \"marketplace-operator-79b997595-j9zqf\" (UID: \"f14c8a94-8589-4059-bddf-329452355fab\") " pod="openshift-marketplace/marketplace-operator-79b997595-j9zqf" Dec 13 06:51:33 crc kubenswrapper[4971]: I1213 06:51:33.576939 4971 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etcd-service-ca\" (UniqueName: \"kubernetes.io/configmap/56c62b80-0a33-4156-b52b-cb54c98a4343-etcd-service-ca\") pod \"etcd-operator-b45778765-6xfwq\" (UID: \"56c62b80-0a33-4156-b52b-cb54c98a4343\") " pod="openshift-etcd-operator/etcd-operator-b45778765-6xfwq" Dec 13 06:51:33 crc kubenswrapper[4971]: I1213 06:51:33.577019 4971 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"webhook-certs\" (UniqueName: \"kubernetes.io/secret/32c8b95e-38c8-43fb-b1e4-86ac9a27c2f0-webhook-certs\") pod \"multus-admission-controller-857f4d67dd-z68ks\" (UID: \"32c8b95e-38c8-43fb-b1e4-86ac9a27c2f0\") " pod="openshift-multus/multus-admission-controller-857f4d67dd-z68ks" Dec 13 06:51:33 crc kubenswrapper[4971]: I1213 06:51:33.577095 4971 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-pbjnf\" (UniqueName: \"kubernetes.io/projected/56c62b80-0a33-4156-b52b-cb54c98a4343-kube-api-access-pbjnf\") pod \"etcd-operator-b45778765-6xfwq\" (UID: \"56c62b80-0a33-4156-b52b-cb54c98a4343\") " pod="openshift-etcd-operator/etcd-operator-b45778765-6xfwq" Dec 13 06:51:33 crc kubenswrapper[4971]: I1213 06:51:33.577181 4971 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-tllvh\" (UniqueName: \"kubernetes.io/projected/f14c8a94-8589-4059-bddf-329452355fab-kube-api-access-tllvh\") pod \"marketplace-operator-79b997595-j9zqf\" (UID: \"f14c8a94-8589-4059-bddf-329452355fab\") " pod="openshift-marketplace/marketplace-operator-79b997595-j9zqf" Dec 13 06:51:33 crc kubenswrapper[4971]: I1213 06:51:33.577308 4971 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-2sjfn\" (UniqueName: \"kubernetes.io/projected/32c8b95e-38c8-43fb-b1e4-86ac9a27c2f0-kube-api-access-2sjfn\") pod \"multus-admission-controller-857f4d67dd-z68ks\" (UID: \"32c8b95e-38c8-43fb-b1e4-86ac9a27c2f0\") " pod="openshift-multus/multus-admission-controller-857f4d67dd-z68ks" Dec 13 06:51:33 crc kubenswrapper[4971]: I1213 06:51:33.577395 4971 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/bb0ead7b-eba0-4d16-9363-df54caee0a32-config\") pod \"openshift-controller-manager-operator-756b6f6bc6-gtrwg\" (UID: \"bb0ead7b-eba0-4d16-9363-df54caee0a32\") " pod="openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-gtrwg" Dec 13 06:51:33 crc kubenswrapper[4971]: I1213 06:51:33.577492 4971 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-fs9ph\" (UniqueName: \"kubernetes.io/projected/2f87c67a-fcf1-46e3-9bbd-9bf03ee55da3-kube-api-access-fs9ph\") pod \"machine-config-controller-84d6567774-kf7sp\" (UID: \"2f87c67a-fcf1-46e3-9bbd-9bf03ee55da3\") " pod="openshift-machine-config-operator/machine-config-controller-84d6567774-kf7sp" Dec 13 06:51:33 crc kubenswrapper[4971]: I1213 06:51:33.577682 4971 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/56c62b80-0a33-4156-b52b-cb54c98a4343-serving-cert\") pod \"etcd-operator-b45778765-6xfwq\" (UID: \"56c62b80-0a33-4156-b52b-cb54c98a4343\") " pod="openshift-etcd-operator/etcd-operator-b45778765-6xfwq" Dec 13 06:51:33 crc kubenswrapper[4971]: I1213 06:51:33.577837 4971 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etcd-service-ca\" (UniqueName: \"kubernetes.io/configmap/56c62b80-0a33-4156-b52b-cb54c98a4343-etcd-service-ca\") pod \"etcd-operator-b45778765-6xfwq\" (UID: \"56c62b80-0a33-4156-b52b-cb54c98a4343\") " pod="openshift-etcd-operator/etcd-operator-b45778765-6xfwq" Dec 13 06:51:33 crc kubenswrapper[4971]: I1213 06:51:33.578612 4971 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/bb0ead7b-eba0-4d16-9363-df54caee0a32-serving-cert\") pod \"openshift-controller-manager-operator-756b6f6bc6-gtrwg\" (UID: \"bb0ead7b-eba0-4d16-9363-df54caee0a32\") " pod="openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-gtrwg" Dec 13 06:51:33 crc kubenswrapper[4971]: I1213 06:51:33.580157 4971 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/bb0ead7b-eba0-4d16-9363-df54caee0a32-config\") pod \"openshift-controller-manager-operator-756b6f6bc6-gtrwg\" (UID: \"bb0ead7b-eba0-4d16-9363-df54caee0a32\") " pod="openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-gtrwg" Dec 13 06:51:33 crc kubenswrapper[4971]: I1213 06:51:33.580954 4971 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/56c62b80-0a33-4156-b52b-cb54c98a4343-serving-cert\") pod \"etcd-operator-b45778765-6xfwq\" (UID: \"56c62b80-0a33-4156-b52b-cb54c98a4343\") " pod="openshift-etcd-operator/etcd-operator-b45778765-6xfwq" Dec 13 06:51:33 crc kubenswrapper[4971]: I1213 06:51:33.583998 4971 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-etcd-operator"/"etcd-client" Dec 13 06:51:33 crc kubenswrapper[4971]: I1213 06:51:33.590167 4971 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/56c62b80-0a33-4156-b52b-cb54c98a4343-etcd-client\") pod \"etcd-operator-b45778765-6xfwq\" (UID: \"56c62b80-0a33-4156-b52b-cb54c98a4343\") " pod="openshift-etcd-operator/etcd-operator-b45778765-6xfwq" Dec 13 06:51:33 crc kubenswrapper[4971]: I1213 06:51:33.603708 4971 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-etcd-operator"/"kube-root-ca.crt" Dec 13 06:51:33 crc kubenswrapper[4971]: I1213 06:51:33.623443 4971 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-etcd-operator"/"openshift-service-ca.crt" Dec 13 06:51:33 crc kubenswrapper[4971]: I1213 06:51:33.644789 4971 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-etcd-operator"/"etcd-operator-config" Dec 13 06:51:33 crc kubenswrapper[4971]: I1213 06:51:33.647041 4971 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/56c62b80-0a33-4156-b52b-cb54c98a4343-config\") pod \"etcd-operator-b45778765-6xfwq\" (UID: \"56c62b80-0a33-4156-b52b-cb54c98a4343\") " pod="openshift-etcd-operator/etcd-operator-b45778765-6xfwq" Dec 13 06:51:33 crc kubenswrapper[4971]: I1213 06:51:33.663279 4971 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-etcd-operator"/"etcd-ca-bundle" Dec 13 06:51:33 crc kubenswrapper[4971]: I1213 06:51:33.666447 4971 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etcd-ca\" (UniqueName: \"kubernetes.io/configmap/56c62b80-0a33-4156-b52b-cb54c98a4343-etcd-ca\") pod \"etcd-operator-b45778765-6xfwq\" (UID: \"56c62b80-0a33-4156-b52b-cb54c98a4343\") " pod="openshift-etcd-operator/etcd-operator-b45778765-6xfwq" Dec 13 06:51:33 crc kubenswrapper[4971]: I1213 06:51:33.683704 4971 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"machine-config-controller-dockercfg-c2lfx" Dec 13 06:51:33 crc kubenswrapper[4971]: I1213 06:51:33.704182 4971 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"mcc-proxy-tls" Dec 13 06:51:33 crc kubenswrapper[4971]: I1213 06:51:33.718666 4971 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/2f87c67a-fcf1-46e3-9bbd-9bf03ee55da3-proxy-tls\") pod \"machine-config-controller-84d6567774-kf7sp\" (UID: \"2f87c67a-fcf1-46e3-9bbd-9bf03ee55da3\") " pod="openshift-machine-config-operator/machine-config-controller-84d6567774-kf7sp" Dec 13 06:51:33 crc kubenswrapper[4971]: I1213 06:51:33.724787 4971 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-api"/"control-plane-machine-set-operator-tls" Dec 13 06:51:33 crc kubenswrapper[4971]: I1213 06:51:33.743940 4971 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-multus"/"multus-admission-controller-secret" Dec 13 06:51:33 crc kubenswrapper[4971]: I1213 06:51:33.751826 4971 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"webhook-certs\" (UniqueName: \"kubernetes.io/secret/32c8b95e-38c8-43fb-b1e4-86ac9a27c2f0-webhook-certs\") pod \"multus-admission-controller-857f4d67dd-z68ks\" (UID: \"32c8b95e-38c8-43fb-b1e4-86ac9a27c2f0\") " pod="openshift-multus/multus-admission-controller-857f4d67dd-z68ks" Dec 13 06:51:33 crc kubenswrapper[4971]: I1213 06:51:33.764661 4971 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-api"/"control-plane-machine-set-operator-dockercfg-k9rxt" Dec 13 06:51:33 crc kubenswrapper[4971]: I1213 06:51:33.783994 4971 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-multus"/"multus-ac-dockercfg-9lkdf" Dec 13 06:51:33 crc kubenswrapper[4971]: I1213 06:51:33.804210 4971 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-marketplace"/"openshift-service-ca.crt" Dec 13 06:51:33 crc kubenswrapper[4971]: I1213 06:51:33.825791 4971 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"marketplace-operator-metrics" Dec 13 06:51:33 crc kubenswrapper[4971]: I1213 06:51:33.831400 4971 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"marketplace-operator-metrics\" (UniqueName: \"kubernetes.io/secret/f14c8a94-8589-4059-bddf-329452355fab-marketplace-operator-metrics\") pod \"marketplace-operator-79b997595-j9zqf\" (UID: \"f14c8a94-8589-4059-bddf-329452355fab\") " pod="openshift-marketplace/marketplace-operator-79b997595-j9zqf" Dec 13 06:51:33 crc kubenswrapper[4971]: I1213 06:51:33.844001 4971 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"marketplace-operator-dockercfg-5nsgg" Dec 13 06:51:33 crc kubenswrapper[4971]: I1213 06:51:33.864759 4971 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"olm-operator-serviceaccount-dockercfg-rq7zk" Dec 13 06:51:33 crc kubenswrapper[4971]: I1213 06:51:33.888427 4971 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-marketplace"/"marketplace-trusted-ca" Dec 13 06:51:33 crc kubenswrapper[4971]: I1213 06:51:33.898857 4971 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"marketplace-trusted-ca\" (UniqueName: \"kubernetes.io/configmap/f14c8a94-8589-4059-bddf-329452355fab-marketplace-trusted-ca\") pod \"marketplace-operator-79b997595-j9zqf\" (UID: \"f14c8a94-8589-4059-bddf-329452355fab\") " pod="openshift-marketplace/marketplace-operator-79b997595-j9zqf" Dec 13 06:51:33 crc kubenswrapper[4971]: I1213 06:51:33.904097 4971 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-marketplace"/"kube-root-ca.crt" Dec 13 06:51:33 crc kubenswrapper[4971]: I1213 06:51:33.924053 4971 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-operator-lifecycle-manager"/"kube-root-ca.crt" Dec 13 06:51:33 crc kubenswrapper[4971]: I1213 06:51:33.944407 4971 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"package-server-manager-serving-cert" Dec 13 06:51:33 crc kubenswrapper[4971]: I1213 06:51:33.964471 4971 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-operator-lifecycle-manager"/"openshift-service-ca.crt" Dec 13 06:51:33 crc kubenswrapper[4971]: I1213 06:51:33.984654 4971 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-service-ca"/"openshift-service-ca.crt" Dec 13 06:51:34 crc kubenswrapper[4971]: I1213 06:51:34.003954 4971 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-service-ca"/"service-ca-dockercfg-pn86c" Dec 13 06:51:34 crc kubenswrapper[4971]: I1213 06:51:34.024150 4971 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-service-ca"/"signing-key" Dec 13 06:51:34 crc kubenswrapper[4971]: I1213 06:51:34.031791 4971 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"signing-key\" (UniqueName: \"kubernetes.io/secret/dcdf01b4-83e9-4d28-821a-e4321ce6b92c-signing-key\") pod \"service-ca-9c57cc56f-q99rh\" (UID: \"dcdf01b4-83e9-4d28-821a-e4321ce6b92c\") " pod="openshift-service-ca/service-ca-9c57cc56f-q99rh" Dec 13 06:51:34 crc kubenswrapper[4971]: I1213 06:51:34.043771 4971 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-service-ca"/"signing-cabundle" Dec 13 06:51:34 crc kubenswrapper[4971]: I1213 06:51:34.045458 4971 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"signing-cabundle\" (UniqueName: \"kubernetes.io/configmap/dcdf01b4-83e9-4d28-821a-e4321ce6b92c-signing-cabundle\") pod \"service-ca-9c57cc56f-q99rh\" (UID: \"dcdf01b4-83e9-4d28-821a-e4321ce6b92c\") " pod="openshift-service-ca/service-ca-9c57cc56f-q99rh" Dec 13 06:51:34 crc kubenswrapper[4971]: I1213 06:51:34.064742 4971 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-service-ca"/"kube-root-ca.crt" Dec 13 06:51:34 crc kubenswrapper[4971]: I1213 06:51:34.103579 4971 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-machine-config-operator"/"machine-config-operator-images" Dec 13 06:51:34 crc kubenswrapper[4971]: I1213 06:51:34.124387 4971 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"machine-config-operator-dockercfg-98p87" Dec 13 06:51:34 crc kubenswrapper[4971]: I1213 06:51:34.143757 4971 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"mco-proxy-tls" Dec 13 06:51:34 crc kubenswrapper[4971]: I1213 06:51:34.163851 4971 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-apiserver-operator"/"kube-apiserver-operator-config" Dec 13 06:51:34 crc kubenswrapper[4971]: I1213 06:51:34.184797 4971 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-apiserver-operator"/"kube-apiserver-operator-dockercfg-x57mr" Dec 13 06:51:34 crc kubenswrapper[4971]: I1213 06:51:34.204337 4971 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-apiserver-operator"/"kube-apiserver-operator-serving-cert" Dec 13 06:51:34 crc kubenswrapper[4971]: I1213 06:51:34.223968 4971 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-apiserver-operator"/"kube-root-ca.crt" Dec 13 06:51:34 crc kubenswrapper[4971]: I1213 06:51:34.243769 4971 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ingress"/"kube-root-ca.crt" Dec 13 06:51:34 crc kubenswrapper[4971]: I1213 06:51:34.264218 4971 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ingress"/"router-dockercfg-zdk86" Dec 13 06:51:34 crc kubenswrapper[4971]: I1213 06:51:34.284387 4971 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ingress"/"router-certs-default" Dec 13 06:51:34 crc kubenswrapper[4971]: I1213 06:51:34.304089 4971 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ingress"/"router-stats-default" Dec 13 06:51:34 crc kubenswrapper[4971]: I1213 06:51:34.324389 4971 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ingress"/"router-metrics-certs-default" Dec 13 06:51:34 crc kubenswrapper[4971]: I1213 06:51:34.344240 4971 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ingress"/"openshift-service-ca.crt" Dec 13 06:51:34 crc kubenswrapper[4971]: I1213 06:51:34.363403 4971 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ingress"/"service-ca-bundle" Dec 13 06:51:34 crc kubenswrapper[4971]: I1213 06:51:34.402556 4971 request.go:700] Waited for 1.014040228s due to client-side throttling, not priority and fairness, request: GET:https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-scheduler-operator/configmaps?fieldSelector=metadata.name%3Dopenshift-kube-scheduler-operator-config&limit=500&resourceVersion=0 Dec 13 06:51:34 crc kubenswrapper[4971]: I1213 06:51:34.404210 4971 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-scheduler-operator"/"openshift-kube-scheduler-operator-config" Dec 13 06:51:34 crc kubenswrapper[4971]: I1213 06:51:34.423592 4971 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-scheduler-operator"/"openshift-kube-scheduler-operator-dockercfg-qt55r" Dec 13 06:51:34 crc kubenswrapper[4971]: I1213 06:51:34.444048 4971 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-scheduler-operator"/"kube-scheduler-operator-serving-cert" Dec 13 06:51:34 crc kubenswrapper[4971]: I1213 06:51:34.463841 4971 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-scheduler-operator"/"kube-root-ca.crt" Dec 13 06:51:34 crc kubenswrapper[4971]: I1213 06:51:34.483710 4971 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-storage-version-migrator-operator"/"openshift-service-ca.crt" Dec 13 06:51:34 crc kubenswrapper[4971]: I1213 06:51:34.503858 4971 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-storage-version-migrator-operator"/"serving-cert" Dec 13 06:51:34 crc kubenswrapper[4971]: I1213 06:51:34.524442 4971 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-storage-version-migrator-operator"/"kube-storage-version-migrator-operator-dockercfg-2bh8d" Dec 13 06:51:34 crc kubenswrapper[4971]: I1213 06:51:34.544135 4971 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-storage-version-migrator-operator"/"config" Dec 13 06:51:34 crc kubenswrapper[4971]: I1213 06:51:34.564308 4971 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-storage-version-migrator-operator"/"kube-root-ca.crt" Dec 13 06:51:34 crc kubenswrapper[4971]: I1213 06:51:34.583696 4971 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-controller-manager-operator"/"kube-controller-manager-operator-serving-cert" Dec 13 06:51:34 crc kubenswrapper[4971]: I1213 06:51:34.604376 4971 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-controller-manager-operator"/"kube-root-ca.crt" Dec 13 06:51:34 crc kubenswrapper[4971]: I1213 06:51:34.624149 4971 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-controller-manager-operator"/"kube-controller-manager-operator-dockercfg-gkqpw" Dec 13 06:51:34 crc kubenswrapper[4971]: I1213 06:51:34.644441 4971 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-controller-manager-operator"/"kube-controller-manager-operator-config" Dec 13 06:51:34 crc kubenswrapper[4971]: I1213 06:51:34.663458 4971 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"pprof-cert" Dec 13 06:51:34 crc kubenswrapper[4971]: I1213 06:51:34.684308 4971 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"olm-operator-serving-cert" Dec 13 06:51:34 crc kubenswrapper[4971]: I1213 06:51:34.704176 4971 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"catalog-operator-serving-cert" Dec 13 06:51:34 crc kubenswrapper[4971]: I1213 06:51:34.723847 4971 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-service-ca-operator"/"serving-cert" Dec 13 06:51:34 crc kubenswrapper[4971]: I1213 06:51:34.744090 4971 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-service-ca-operator"/"kube-root-ca.crt" Dec 13 06:51:34 crc kubenswrapper[4971]: I1213 06:51:34.766174 4971 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-service-ca-operator"/"service-ca-operator-dockercfg-rg9jl" Dec 13 06:51:34 crc kubenswrapper[4971]: I1213 06:51:34.783774 4971 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-service-ca-operator"/"openshift-service-ca.crt" Dec 13 06:51:34 crc kubenswrapper[4971]: I1213 06:51:34.804342 4971 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-service-ca-operator"/"service-ca-operator-config" Dec 13 06:51:34 crc kubenswrapper[4971]: I1213 06:51:34.823719 4971 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"collect-profiles-dockercfg-kzf4t" Dec 13 06:51:34 crc kubenswrapper[4971]: I1213 06:51:34.844701 4971 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-operator-lifecycle-manager"/"collect-profiles-config" Dec 13 06:51:34 crc kubenswrapper[4971]: I1213 06:51:34.863848 4971 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-storage-version-migrator"/"kube-root-ca.crt" Dec 13 06:51:34 crc kubenswrapper[4971]: I1213 06:51:34.884281 4971 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-storage-version-migrator"/"kube-storage-version-migrator-sa-dockercfg-5xfcg" Dec 13 06:51:34 crc kubenswrapper[4971]: I1213 06:51:34.903448 4971 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-storage-version-migrator"/"openshift-service-ca.crt" Dec 13 06:51:34 crc kubenswrapper[4971]: I1213 06:51:34.924221 4971 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication-operator"/"authentication-operator-config" Dec 13 06:51:34 crc kubenswrapper[4971]: I1213 06:51:34.951629 4971 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication-operator"/"trusted-ca-bundle" Dec 13 06:51:34 crc kubenswrapper[4971]: I1213 06:51:34.963860 4971 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication-operator"/"service-ca-bundle" Dec 13 06:51:34 crc kubenswrapper[4971]: I1213 06:51:34.983925 4971 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication-operator"/"authentication-operator-dockercfg-mz9bj" Dec 13 06:51:35 crc kubenswrapper[4971]: I1213 06:51:35.004100 4971 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication-operator"/"serving-cert" Dec 13 06:51:35 crc kubenswrapper[4971]: I1213 06:51:35.023755 4971 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication-operator"/"openshift-service-ca.crt" Dec 13 06:51:35 crc kubenswrapper[4971]: I1213 06:51:35.044603 4971 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"packageserver-service-cert" Dec 13 06:51:35 crc kubenswrapper[4971]: I1213 06:51:35.064735 4971 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication-operator"/"kube-root-ca.crt" Dec 13 06:51:35 crc kubenswrapper[4971]: I1213 06:51:35.084810 4971 reflector.go:368] Caches populated for *v1.ConfigMap from object-"hostpath-provisioner"/"kube-root-ca.crt" Dec 13 06:51:35 crc kubenswrapper[4971]: I1213 06:51:35.103877 4971 reflector.go:368] Caches populated for *v1.ConfigMap from object-"hostpath-provisioner"/"openshift-service-ca.crt" Dec 13 06:51:35 crc kubenswrapper[4971]: I1213 06:51:35.123436 4971 reflector.go:368] Caches populated for *v1.Secret from object-"hostpath-provisioner"/"csi-hostpath-provisioner-sa-dockercfg-qd74k" Dec 13 06:51:35 crc kubenswrapper[4971]: I1213 06:51:35.158978 4971 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-krxlt\" (UniqueName: \"kubernetes.io/projected/23e16ced-5b54-4a51-a1bc-4d11c907c90a-kube-api-access-krxlt\") pod \"oauth-openshift-558db77b4-nzzbm\" (UID: \"23e16ced-5b54-4a51-a1bc-4d11c907c90a\") " pod="openshift-authentication/oauth-openshift-558db77b4-nzzbm" Dec 13 06:51:35 crc kubenswrapper[4971]: I1213 06:51:35.176500 4971 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-68rgh\" (UniqueName: \"kubernetes.io/projected/9f0db656-7a94-415b-b737-30ed342817ee-kube-api-access-68rgh\") pod \"console-f9d7485db-5fd9b\" (UID: \"9f0db656-7a94-415b-b737-30ed342817ee\") " pod="openshift-console/console-f9d7485db-5fd9b" Dec 13 06:51:35 crc kubenswrapper[4971]: I1213 06:51:35.196716 4971 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-znjpx\" (UniqueName: \"kubernetes.io/projected/97219ee9-4c52-49aa-903d-044a079d4ecb-kube-api-access-znjpx\") pod \"openshift-apiserver-operator-796bbdcf4f-7ddbb\" (UID: \"97219ee9-4c52-49aa-903d-044a079d4ecb\") " pod="openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-7ddbb" Dec 13 06:51:35 crc kubenswrapper[4971]: I1213 06:51:35.219609 4971 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/8294e250-74ea-497c-b6ea-08b957184eec-bound-sa-token\") pod \"cluster-image-registry-operator-dc59b4c8b-zqt2t\" (UID: \"8294e250-74ea-497c-b6ea-08b957184eec\") " pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-zqt2t" Dec 13 06:51:35 crc kubenswrapper[4971]: I1213 06:51:35.225813 4971 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-console/console-f9d7485db-5fd9b" Dec 13 06:51:35 crc kubenswrapper[4971]: I1213 06:51:35.239293 4971 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-dlhrl\" (UniqueName: \"kubernetes.io/projected/874901b0-9650-4029-8d21-74441f40e201-kube-api-access-dlhrl\") pod \"openshift-config-operator-7777fb866f-vmdsp\" (UID: \"874901b0-9650-4029-8d21-74441f40e201\") " pod="openshift-config-operator/openshift-config-operator-7777fb866f-vmdsp" Dec 13 06:51:35 crc kubenswrapper[4971]: I1213 06:51:35.245535 4971 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-authentication/oauth-openshift-558db77b4-nzzbm" Dec 13 06:51:35 crc kubenswrapper[4971]: I1213 06:51:35.258210 4971 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-config-operator/openshift-config-operator-7777fb866f-vmdsp" Dec 13 06:51:35 crc kubenswrapper[4971]: I1213 06:51:35.258378 4971 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-kjkp6\" (UniqueName: \"kubernetes.io/projected/817cbb9d-ed5b-4955-8b6a-4b33155b744c-kube-api-access-kjkp6\") pod \"cluster-samples-operator-665b6dd947-mlbwt\" (UID: \"817cbb9d-ed5b-4955-8b6a-4b33155b744c\") " pod="openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-mlbwt" Dec 13 06:51:35 crc kubenswrapper[4971]: I1213 06:51:35.277681 4971 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-zvw59\" (UniqueName: \"kubernetes.io/projected/bec6e53b-3998-4402-b77b-87ff79f31827-kube-api-access-zvw59\") pod \"apiserver-76f77b778f-2n4cq\" (UID: \"bec6e53b-3998-4402-b77b-87ff79f31827\") " pod="openshift-apiserver/apiserver-76f77b778f-2n4cq" Dec 13 06:51:35 crc kubenswrapper[4971]: I1213 06:51:35.301359 4971 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-8m6vz\" (UniqueName: \"kubernetes.io/projected/88850e0f-df47-4650-be7e-d14a83ce05f4-kube-api-access-8m6vz\") pod \"machine-approver-56656f9798-s75fl\" (UID: \"88850e0f-df47-4650-be7e-d14a83ce05f4\") " pod="openshift-cluster-machine-approver/machine-approver-56656f9798-s75fl" Dec 13 06:51:35 crc kubenswrapper[4971]: I1213 06:51:35.324021 4971 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-tb7rw\" (UniqueName: \"kubernetes.io/projected/8294e250-74ea-497c-b6ea-08b957184eec-kube-api-access-tb7rw\") pod \"cluster-image-registry-operator-dc59b4c8b-zqt2t\" (UID: \"8294e250-74ea-497c-b6ea-08b957184eec\") " pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-zqt2t" Dec 13 06:51:35 crc kubenswrapper[4971]: I1213 06:51:35.337067 4971 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-87jr5\" (UniqueName: \"kubernetes.io/projected/c45a314c-77da-43fc-b2ae-9de350540c2b-kube-api-access-87jr5\") pod \"machine-api-operator-5694c8668f-t24dm\" (UID: \"c45a314c-77da-43fc-b2ae-9de350540c2b\") " pod="openshift-machine-api/machine-api-operator-5694c8668f-t24dm" Dec 13 06:51:35 crc kubenswrapper[4971]: I1213 06:51:35.361881 4971 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-tjjmd\" (UniqueName: \"kubernetes.io/projected/8e7aefba-ae2e-4069-892e-544c85b9af15-kube-api-access-tjjmd\") pod \"apiserver-7bbb656c7d-mb2q7\" (UID: \"8e7aefba-ae2e-4069-892e-544c85b9af15\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-mb2q7" Dec 13 06:51:35 crc kubenswrapper[4971]: I1213 06:51:35.362894 4971 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-cluster-machine-approver/machine-approver-56656f9798-s75fl" Dec 13 06:51:35 crc kubenswrapper[4971]: I1213 06:51:35.381794 4971 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-console/console-f9d7485db-5fd9b"] Dec 13 06:51:35 crc kubenswrapper[4971]: I1213 06:51:35.382056 4971 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-mb2q7" Dec 13 06:51:35 crc kubenswrapper[4971]: I1213 06:51:35.382441 4971 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-lmhvx\" (UniqueName: \"kubernetes.io/projected/7148e573-e2f5-4fc5-b30c-4e777516509e-kube-api-access-lmhvx\") pod \"controller-manager-879f6c89f-xtxr8\" (UID: \"7148e573-e2f5-4fc5-b30c-4e777516509e\") " pod="openshift-controller-manager/controller-manager-879f6c89f-xtxr8" Dec 13 06:51:35 crc kubenswrapper[4971]: I1213 06:51:35.386921 4971 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-dns"/"dns-dockercfg-jwfmh" Dec 13 06:51:35 crc kubenswrapper[4971]: I1213 06:51:35.402788 4971 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-machine-api/machine-api-operator-5694c8668f-t24dm" Dec 13 06:51:35 crc kubenswrapper[4971]: I1213 06:51:35.403300 4971 request.go:700] Waited for 1.883738944s due to client-side throttling, not priority and fairness, request: GET:https://api-int.crc.testing:6443/api/v1/namespaces/openshift-dns/configmaps?fieldSelector=metadata.name%3Ddns-default&limit=500&resourceVersion=0 Dec 13 06:51:35 crc kubenswrapper[4971]: I1213 06:51:35.404694 4971 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-dns"/"dns-default" Dec 13 06:51:35 crc kubenswrapper[4971]: I1213 06:51:35.416272 4971 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-7ddbb" Dec 13 06:51:35 crc kubenswrapper[4971]: I1213 06:51:35.424890 4971 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-dns"/"dns-default-metrics-tls" Dec 13 06:51:35 crc kubenswrapper[4971]: I1213 06:51:35.432361 4971 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-authentication/oauth-openshift-558db77b4-nzzbm"] Dec 13 06:51:35 crc kubenswrapper[4971]: W1213 06:51:35.439952 4971 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod23e16ced_5b54_4a51_a1bc_4d11c907c90a.slice/crio-eae50d9db612b30a2aa3d28acec408f9df99cb8757dc54079b0da3f4449bb56b WatchSource:0}: Error finding container eae50d9db612b30a2aa3d28acec408f9df99cb8757dc54079b0da3f4449bb56b: Status 404 returned error can't find the container with id eae50d9db612b30a2aa3d28acec408f9df99cb8757dc54079b0da3f4449bb56b Dec 13 06:51:35 crc kubenswrapper[4971]: I1213 06:51:35.444282 4971 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"machine-config-server-dockercfg-qx5rd" Dec 13 06:51:35 crc kubenswrapper[4971]: I1213 06:51:35.461965 4971 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-console/console-f9d7485db-5fd9b" event={"ID":"9f0db656-7a94-415b-b737-30ed342817ee","Type":"ContainerStarted","Data":"13f95dbc7e9f1041c783b48af992f4b000a5028dcacf9bc06266384582807d80"} Dec 13 06:51:35 crc kubenswrapper[4971]: I1213 06:51:35.462070 4971 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-config-operator/openshift-config-operator-7777fb866f-vmdsp"] Dec 13 06:51:35 crc kubenswrapper[4971]: I1213 06:51:35.462964 4971 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-mlbwt" Dec 13 06:51:35 crc kubenswrapper[4971]: I1213 06:51:35.464753 4971 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"machine-config-server-tls" Dec 13 06:51:35 crc kubenswrapper[4971]: I1213 06:51:35.466139 4971 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-cluster-machine-approver/machine-approver-56656f9798-s75fl" event={"ID":"88850e0f-df47-4650-be7e-d14a83ce05f4","Type":"ContainerStarted","Data":"c09246dccc67fec052ab57b264c0fec707499b48b1a35fc7ce319d0df0df09f9"} Dec 13 06:51:35 crc kubenswrapper[4971]: I1213 06:51:35.470182 4971 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-authentication/oauth-openshift-558db77b4-nzzbm" event={"ID":"23e16ced-5b54-4a51-a1bc-4d11c907c90a","Type":"ContainerStarted","Data":"eae50d9db612b30a2aa3d28acec408f9df99cb8757dc54079b0da3f4449bb56b"} Dec 13 06:51:35 crc kubenswrapper[4971]: I1213 06:51:35.477924 4971 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-apiserver/apiserver-76f77b778f-2n4cq" Dec 13 06:51:35 crc kubenswrapper[4971]: I1213 06:51:35.484074 4971 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"node-bootstrapper-token" Dec 13 06:51:35 crc kubenswrapper[4971]: I1213 06:51:35.489661 4971 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-zqt2t" Dec 13 06:51:35 crc kubenswrapper[4971]: I1213 06:51:35.506020 4971 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ingress-canary"/"canary-serving-cert" Dec 13 06:51:35 crc kubenswrapper[4971]: I1213 06:51:35.524166 4971 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ingress-canary"/"default-dockercfg-2llfx" Dec 13 06:51:35 crc kubenswrapper[4971]: I1213 06:51:35.536431 4971 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-controller-manager/controller-manager-879f6c89f-xtxr8" Dec 13 06:51:35 crc kubenswrapper[4971]: I1213 06:51:35.551165 4971 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ingress-canary"/"openshift-service-ca.crt" Dec 13 06:51:35 crc kubenswrapper[4971]: I1213 06:51:35.563587 4971 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ingress-canary"/"kube-root-ca.crt" Dec 13 06:51:35 crc kubenswrapper[4971]: I1213 06:51:35.616824 4971 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-machine-api/machine-api-operator-5694c8668f-t24dm"] Dec 13 06:51:35 crc kubenswrapper[4971]: I1213 06:51:35.624087 4971 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-96z5g\" (UniqueName: \"kubernetes.io/projected/bb0ead7b-eba0-4d16-9363-df54caee0a32-kube-api-access-96z5g\") pod \"openshift-controller-manager-operator-756b6f6bc6-gtrwg\" (UID: \"bb0ead7b-eba0-4d16-9363-df54caee0a32\") " pod="openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-gtrwg" Dec 13 06:51:35 crc kubenswrapper[4971]: I1213 06:51:35.627965 4971 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-vbcwz\" (UniqueName: \"kubernetes.io/projected/dcdf01b4-83e9-4d28-821a-e4321ce6b92c-kube-api-access-vbcwz\") pod \"service-ca-9c57cc56f-q99rh\" (UID: \"dcdf01b4-83e9-4d28-821a-e4321ce6b92c\") " pod="openshift-service-ca/service-ca-9c57cc56f-q99rh" Dec 13 06:51:35 crc kubenswrapper[4971]: I1213 06:51:35.634365 4971 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-gtrwg" Dec 13 06:51:35 crc kubenswrapper[4971]: I1213 06:51:35.645304 4971 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-pbjnf\" (UniqueName: \"kubernetes.io/projected/56c62b80-0a33-4156-b52b-cb54c98a4343-kube-api-access-pbjnf\") pod \"etcd-operator-b45778765-6xfwq\" (UID: \"56c62b80-0a33-4156-b52b-cb54c98a4343\") " pod="openshift-etcd-operator/etcd-operator-b45778765-6xfwq" Dec 13 06:51:35 crc kubenswrapper[4971]: I1213 06:51:35.646579 4971 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-oauth-apiserver/apiserver-7bbb656c7d-mb2q7"] Dec 13 06:51:35 crc kubenswrapper[4971]: W1213 06:51:35.648826 4971 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podc45a314c_77da_43fc_b2ae_9de350540c2b.slice/crio-9c4e4064236c42265993bd0592a5564ea72a0b1010749d1dab5988fc7f547753 WatchSource:0}: Error finding container 9c4e4064236c42265993bd0592a5564ea72a0b1010749d1dab5988fc7f547753: Status 404 returned error can't find the container with id 9c4e4064236c42265993bd0592a5564ea72a0b1010749d1dab5988fc7f547753 Dec 13 06:51:35 crc kubenswrapper[4971]: I1213 06:51:35.665623 4971 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-etcd-operator/etcd-operator-b45778765-6xfwq" Dec 13 06:51:35 crc kubenswrapper[4971]: I1213 06:51:35.670261 4971 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-7ddbb"] Dec 13 06:51:35 crc kubenswrapper[4971]: I1213 06:51:35.671442 4971 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-tllvh\" (UniqueName: \"kubernetes.io/projected/f14c8a94-8589-4059-bddf-329452355fab-kube-api-access-tllvh\") pod \"marketplace-operator-79b997595-j9zqf\" (UID: \"f14c8a94-8589-4059-bddf-329452355fab\") " pod="openshift-marketplace/marketplace-operator-79b997595-j9zqf" Dec 13 06:51:35 crc kubenswrapper[4971]: I1213 06:51:35.677905 4971 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-2sjfn\" (UniqueName: \"kubernetes.io/projected/32c8b95e-38c8-43fb-b1e4-86ac9a27c2f0-kube-api-access-2sjfn\") pod \"multus-admission-controller-857f4d67dd-z68ks\" (UID: \"32c8b95e-38c8-43fb-b1e4-86ac9a27c2f0\") " pod="openshift-multus/multus-admission-controller-857f4d67dd-z68ks" Dec 13 06:51:35 crc kubenswrapper[4971]: I1213 06:51:35.695094 4971 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/multus-admission-controller-857f4d67dd-z68ks" Dec 13 06:51:35 crc kubenswrapper[4971]: I1213 06:51:35.702430 4971 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-fs9ph\" (UniqueName: \"kubernetes.io/projected/2f87c67a-fcf1-46e3-9bbd-9bf03ee55da3-kube-api-access-fs9ph\") pod \"machine-config-controller-84d6567774-kf7sp\" (UID: \"2f87c67a-fcf1-46e3-9bbd-9bf03ee55da3\") " pod="openshift-machine-config-operator/machine-config-controller-84d6567774-kf7sp" Dec 13 06:51:35 crc kubenswrapper[4971]: I1213 06:51:35.704965 4971 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/marketplace-operator-79b997595-j9zqf" Dec 13 06:51:35 crc kubenswrapper[4971]: I1213 06:51:35.721224 4971 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-mlbwt"] Dec 13 06:51:35 crc kubenswrapper[4971]: I1213 06:51:35.721809 4971 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-service-ca/service-ca-9c57cc56f-q99rh" Dec 13 06:51:35 crc kubenswrapper[4971]: I1213 06:51:35.807888 4971 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-fs22c\" (UniqueName: \"kubernetes.io/projected/cba893c6-ff99-466a-9db0-a62e7a87a334-kube-api-access-fs22c\") pod \"image-registry-697d97f7c8-mdd8q\" (UID: \"cba893c6-ff99-466a-9db0-a62e7a87a334\") " pod="openshift-image-registry/image-registry-697d97f7c8-mdd8q" Dec 13 06:51:35 crc kubenswrapper[4971]: I1213 06:51:35.808208 4971 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/22f3a4f0-1686-4087-bcec-e48f2015ffe2-bound-sa-token\") pod \"ingress-operator-5b745b69d9-pkmb5\" (UID: \"22f3a4f0-1686-4087-bcec-e48f2015ffe2\") " pod="openshift-ingress-operator/ingress-operator-5b745b69d9-pkmb5" Dec 13 06:51:35 crc kubenswrapper[4971]: I1213 06:51:35.808231 4971 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/375613f8-1e50-42fe-b593-2a6a07afbe64-kube-api-access\") pod \"kube-apiserver-operator-766d6c64bb-xfvfg\" (UID: \"375613f8-1e50-42fe-b593-2a6a07afbe64\") " pod="openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-xfvfg" Dec 13 06:51:35 crc kubenswrapper[4971]: I1213 06:51:35.808291 4971 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/874a112c-101d-4a74-a674-d9afe62b6dc8-proxy-tls\") pod \"machine-config-operator-74547568cd-4cn2z\" (UID: \"874a112c-101d-4a74-a674-d9afe62b6dc8\") " pod="openshift-machine-config-operator/machine-config-operator-74547568cd-4cn2z" Dec 13 06:51:35 crc kubenswrapper[4971]: I1213 06:51:35.808311 4971 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/874a112c-101d-4a74-a674-d9afe62b6dc8-auth-proxy-config\") pod \"machine-config-operator-74547568cd-4cn2z\" (UID: \"874a112c-101d-4a74-a674-d9afe62b6dc8\") " pod="openshift-machine-config-operator/machine-config-operator-74547568cd-4cn2z" Dec 13 06:51:35 crc kubenswrapper[4971]: I1213 06:51:35.808330 4971 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"default-certificate\" (UniqueName: \"kubernetes.io/secret/d89e3583-3f66-4259-b6f9-bffe8ddbde12-default-certificate\") pod \"router-default-5444994796-mzhp6\" (UID: \"d89e3583-3f66-4259-b6f9-bffe8ddbde12\") " pod="openshift-ingress/router-default-5444994796-mzhp6" Dec 13 06:51:35 crc kubenswrapper[4971]: I1213 06:51:35.808373 4971 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/2cf6e363-8511-4c79-9181-79dba726e5af-client-ca\") pod \"route-controller-manager-6576b87f9c-9d86n\" (UID: \"2cf6e363-8511-4c79-9181-79dba726e5af\") " pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-9d86n" Dec 13 06:51:35 crc kubenswrapper[4971]: I1213 06:51:35.808413 4971 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-lltlz\" (UniqueName: \"kubernetes.io/projected/874a112c-101d-4a74-a674-d9afe62b6dc8-kube-api-access-lltlz\") pod \"machine-config-operator-74547568cd-4cn2z\" (UID: \"874a112c-101d-4a74-a674-d9afe62b6dc8\") " pod="openshift-machine-config-operator/machine-config-operator-74547568cd-4cn2z" Dec 13 06:51:35 crc kubenswrapper[4971]: I1213 06:51:35.808456 4971 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-fzfk6\" (UniqueName: \"kubernetes.io/projected/d89e3583-3f66-4259-b6f9-bffe8ddbde12-kube-api-access-fzfk6\") pod \"router-default-5444994796-mzhp6\" (UID: \"d89e3583-3f66-4259-b6f9-bffe8ddbde12\") " pod="openshift-ingress/router-default-5444994796-mzhp6" Dec 13 06:51:35 crc kubenswrapper[4971]: I1213 06:51:35.808481 4971 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-mdd8q\" (UID: \"cba893c6-ff99-466a-9db0-a62e7a87a334\") " pod="openshift-image-registry/image-registry-697d97f7c8-mdd8q" Dec 13 06:51:35 crc kubenswrapper[4971]: I1213 06:51:35.808502 4971 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/375613f8-1e50-42fe-b593-2a6a07afbe64-config\") pod \"kube-apiserver-operator-766d6c64bb-xfvfg\" (UID: \"375613f8-1e50-42fe-b593-2a6a07afbe64\") " pod="openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-xfvfg" Dec 13 06:51:35 crc kubenswrapper[4971]: I1213 06:51:35.808558 4971 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-t2s9j\" (UniqueName: \"kubernetes.io/projected/ad158968-e110-4b4c-9f42-da9c08274dab-kube-api-access-t2s9j\") pod \"dns-operator-744455d44c-b999k\" (UID: \"ad158968-e110-4b4c-9f42-da9c08274dab\") " pod="openshift-dns-operator/dns-operator-744455d44c-b999k" Dec 13 06:51:35 crc kubenswrapper[4971]: I1213 06:51:35.808577 4971 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"registry-tls\" (UniqueName: \"kubernetes.io/projected/cba893c6-ff99-466a-9db0-a62e7a87a334-registry-tls\") pod \"image-registry-697d97f7c8-mdd8q\" (UID: \"cba893c6-ff99-466a-9db0-a62e7a87a334\") " pod="openshift-image-registry/image-registry-697d97f7c8-mdd8q" Dec 13 06:51:35 crc kubenswrapper[4971]: I1213 06:51:35.808596 4971 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/2cf6e363-8511-4c79-9181-79dba726e5af-config\") pod \"route-controller-manager-6576b87f9c-9d86n\" (UID: \"2cf6e363-8511-4c79-9181-79dba726e5af\") " pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-9d86n" Dec 13 06:51:35 crc kubenswrapper[4971]: I1213 06:51:35.808630 4971 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/375613f8-1e50-42fe-b593-2a6a07afbe64-serving-cert\") pod \"kube-apiserver-operator-766d6c64bb-xfvfg\" (UID: \"375613f8-1e50-42fe-b593-2a6a07afbe64\") " pod="openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-xfvfg" Dec 13 06:51:35 crc kubenswrapper[4971]: I1213 06:51:35.808674 4971 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/cba893c6-ff99-466a-9db0-a62e7a87a334-trusted-ca\") pod \"image-registry-697d97f7c8-mdd8q\" (UID: \"cba893c6-ff99-466a-9db0-a62e7a87a334\") " pod="openshift-image-registry/image-registry-697d97f7c8-mdd8q" Dec 13 06:51:35 crc kubenswrapper[4971]: I1213 06:51:35.808693 4971 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-jk6gm\" (UniqueName: \"kubernetes.io/projected/ee2e94f4-cadb-4b17-8853-a1292cb02404-kube-api-access-jk6gm\") pod \"downloads-7954f5f757-dqk9z\" (UID: \"ee2e94f4-cadb-4b17-8853-a1292cb02404\") " pod="openshift-console/downloads-7954f5f757-dqk9z" Dec 13 06:51:35 crc kubenswrapper[4971]: I1213 06:51:35.808734 4971 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-2gk4l\" (UniqueName: \"kubernetes.io/projected/94469d04-47fd-4bd8-8258-9e874767f250-kube-api-access-2gk4l\") pod \"console-operator-58897d9998-kdfcd\" (UID: \"94469d04-47fd-4bd8-8258-9e874767f250\") " pod="openshift-console-operator/console-operator-58897d9998-kdfcd" Dec 13 06:51:35 crc kubenswrapper[4971]: I1213 06:51:35.808754 4971 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"images\" (UniqueName: \"kubernetes.io/configmap/874a112c-101d-4a74-a674-d9afe62b6dc8-images\") pod \"machine-config-operator-74547568cd-4cn2z\" (UID: \"874a112c-101d-4a74-a674-d9afe62b6dc8\") " pod="openshift-machine-config-operator/machine-config-operator-74547568cd-4cn2z" Dec 13 06:51:35 crc kubenswrapper[4971]: I1213 06:51:35.808771 4971 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/ad158968-e110-4b4c-9f42-da9c08274dab-metrics-tls\") pod \"dns-operator-744455d44c-b999k\" (UID: \"ad158968-e110-4b4c-9f42-da9c08274dab\") " pod="openshift-dns-operator/dns-operator-744455d44c-b999k" Dec 13 06:51:35 crc kubenswrapper[4971]: I1213 06:51:35.808801 4971 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"stats-auth\" (UniqueName: \"kubernetes.io/secret/d89e3583-3f66-4259-b6f9-bffe8ddbde12-stats-auth\") pod \"router-default-5444994796-mzhp6\" (UID: \"d89e3583-3f66-4259-b6f9-bffe8ddbde12\") " pod="openshift-ingress/router-default-5444994796-mzhp6" Dec 13 06:51:35 crc kubenswrapper[4971]: I1213 06:51:35.808822 4971 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/94469d04-47fd-4bd8-8258-9e874767f250-serving-cert\") pod \"console-operator-58897d9998-kdfcd\" (UID: \"94469d04-47fd-4bd8-8258-9e874767f250\") " pod="openshift-console-operator/console-operator-58897d9998-kdfcd" Dec 13 06:51:35 crc kubenswrapper[4971]: I1213 06:51:35.808841 4971 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/22f3a4f0-1686-4087-bcec-e48f2015ffe2-trusted-ca\") pod \"ingress-operator-5b745b69d9-pkmb5\" (UID: \"22f3a4f0-1686-4087-bcec-e48f2015ffe2\") " pod="openshift-ingress-operator/ingress-operator-5b745b69d9-pkmb5" Dec 13 06:51:35 crc kubenswrapper[4971]: I1213 06:51:35.808880 4971 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"package-server-manager-serving-cert\" (UniqueName: \"kubernetes.io/secret/479b25e6-6618-4123-a913-ca50d6c4ba28-package-server-manager-serving-cert\") pod \"package-server-manager-789f6589d5-g8vfp\" (UID: \"479b25e6-6618-4123-a913-ca50d6c4ba28\") " pod="openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-g8vfp" Dec 13 06:51:35 crc kubenswrapper[4971]: I1213 06:51:35.808955 4971 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/cba893c6-ff99-466a-9db0-a62e7a87a334-bound-sa-token\") pod \"image-registry-697d97f7c8-mdd8q\" (UID: \"cba893c6-ff99-466a-9db0-a62e7a87a334\") " pod="openshift-image-registry/image-registry-697d97f7c8-mdd8q" Dec 13 06:51:35 crc kubenswrapper[4971]: I1213 06:51:35.809008 4971 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/94469d04-47fd-4bd8-8258-9e874767f250-config\") pod \"console-operator-58897d9998-kdfcd\" (UID: \"94469d04-47fd-4bd8-8258-9e874767f250\") " pod="openshift-console-operator/console-operator-58897d9998-kdfcd" Dec 13 06:51:35 crc kubenswrapper[4971]: I1213 06:51:35.809029 4971 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-mkgwg\" (UniqueName: \"kubernetes.io/projected/2cf6e363-8511-4c79-9181-79dba726e5af-kube-api-access-mkgwg\") pod \"route-controller-manager-6576b87f9c-9d86n\" (UID: \"2cf6e363-8511-4c79-9181-79dba726e5af\") " pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-9d86n" Dec 13 06:51:35 crc kubenswrapper[4971]: I1213 06:51:35.809064 4971 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-v95qb\" (UniqueName: \"kubernetes.io/projected/479b25e6-6618-4123-a913-ca50d6c4ba28-kube-api-access-v95qb\") pod \"package-server-manager-789f6589d5-g8vfp\" (UID: \"479b25e6-6618-4123-a913-ca50d6c4ba28\") " pod="openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-g8vfp" Dec 13 06:51:35 crc kubenswrapper[4971]: I1213 06:51:35.809112 4971 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/22f3a4f0-1686-4087-bcec-e48f2015ffe2-metrics-tls\") pod \"ingress-operator-5b745b69d9-pkmb5\" (UID: \"22f3a4f0-1686-4087-bcec-e48f2015ffe2\") " pod="openshift-ingress-operator/ingress-operator-5b745b69d9-pkmb5" Dec 13 06:51:35 crc kubenswrapper[4971]: I1213 06:51:35.809134 4971 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-64btv\" (UniqueName: \"kubernetes.io/projected/930fe38c-5ff7-45a8-a292-cd69d07e86f7-kube-api-access-64btv\") pod \"control-plane-machine-set-operator-78cbb6b69f-bgdbk\" (UID: \"930fe38c-5ff7-45a8-a292-cd69d07e86f7\") " pod="openshift-machine-api/control-plane-machine-set-operator-78cbb6b69f-bgdbk" Dec 13 06:51:35 crc kubenswrapper[4971]: I1213 06:51:35.809170 4971 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"registry-certificates\" (UniqueName: \"kubernetes.io/configmap/cba893c6-ff99-466a-9db0-a62e7a87a334-registry-certificates\") pod \"image-registry-697d97f7c8-mdd8q\" (UID: \"cba893c6-ff99-466a-9db0-a62e7a87a334\") " pod="openshift-image-registry/image-registry-697d97f7c8-mdd8q" Dec 13 06:51:35 crc kubenswrapper[4971]: I1213 06:51:35.809204 4971 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ca-trust-extracted\" (UniqueName: \"kubernetes.io/empty-dir/cba893c6-ff99-466a-9db0-a62e7a87a334-ca-trust-extracted\") pod \"image-registry-697d97f7c8-mdd8q\" (UID: \"cba893c6-ff99-466a-9db0-a62e7a87a334\") " pod="openshift-image-registry/image-registry-697d97f7c8-mdd8q" Dec 13 06:51:35 crc kubenswrapper[4971]: I1213 06:51:35.809227 4971 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"control-plane-machine-set-operator-tls\" (UniqueName: \"kubernetes.io/secret/930fe38c-5ff7-45a8-a292-cd69d07e86f7-control-plane-machine-set-operator-tls\") pod \"control-plane-machine-set-operator-78cbb6b69f-bgdbk\" (UID: \"930fe38c-5ff7-45a8-a292-cd69d07e86f7\") " pod="openshift-machine-api/control-plane-machine-set-operator-78cbb6b69f-bgdbk" Dec 13 06:51:35 crc kubenswrapper[4971]: I1213 06:51:35.809258 4971 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/94469d04-47fd-4bd8-8258-9e874767f250-trusted-ca\") pod \"console-operator-58897d9998-kdfcd\" (UID: \"94469d04-47fd-4bd8-8258-9e874767f250\") " pod="openshift-console-operator/console-operator-58897d9998-kdfcd" Dec 13 06:51:35 crc kubenswrapper[4971]: I1213 06:51:35.809280 4971 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"service-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/d89e3583-3f66-4259-b6f9-bffe8ddbde12-service-ca-bundle\") pod \"router-default-5444994796-mzhp6\" (UID: \"d89e3583-3f66-4259-b6f9-bffe8ddbde12\") " pod="openshift-ingress/router-default-5444994796-mzhp6" Dec 13 06:51:35 crc kubenswrapper[4971]: I1213 06:51:35.809299 4971 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"installation-pull-secrets\" (UniqueName: \"kubernetes.io/secret/cba893c6-ff99-466a-9db0-a62e7a87a334-installation-pull-secrets\") pod \"image-registry-697d97f7c8-mdd8q\" (UID: \"cba893c6-ff99-466a-9db0-a62e7a87a334\") " pod="openshift-image-registry/image-registry-697d97f7c8-mdd8q" Dec 13 06:51:35 crc kubenswrapper[4971]: I1213 06:51:35.809320 4971 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/2cf6e363-8511-4c79-9181-79dba726e5af-serving-cert\") pod \"route-controller-manager-6576b87f9c-9d86n\" (UID: \"2cf6e363-8511-4c79-9181-79dba726e5af\") " pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-9d86n" Dec 13 06:51:35 crc kubenswrapper[4971]: I1213 06:51:35.809395 4971 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/d89e3583-3f66-4259-b6f9-bffe8ddbde12-metrics-certs\") pod \"router-default-5444994796-mzhp6\" (UID: \"d89e3583-3f66-4259-b6f9-bffe8ddbde12\") " pod="openshift-ingress/router-default-5444994796-mzhp6" Dec 13 06:51:35 crc kubenswrapper[4971]: I1213 06:51:35.809444 4971 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-vxdq2\" (UniqueName: \"kubernetes.io/projected/22f3a4f0-1686-4087-bcec-e48f2015ffe2-kube-api-access-vxdq2\") pod \"ingress-operator-5b745b69d9-pkmb5\" (UID: \"22f3a4f0-1686-4087-bcec-e48f2015ffe2\") " pod="openshift-ingress-operator/ingress-operator-5b745b69d9-pkmb5" Dec 13 06:51:35 crc kubenswrapper[4971]: E1213 06:51:35.812807 4971 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-12-13 06:51:36.312792864 +0000 UTC m=+152.917202312 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-mdd8q" (UID: "cba893c6-ff99-466a-9db0-a62e7a87a334") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 13 06:51:35 crc kubenswrapper[4971]: I1213 06:51:35.911055 4971 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 13 06:51:35 crc kubenswrapper[4971]: I1213 06:51:35.911273 4971 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/94469d04-47fd-4bd8-8258-9e874767f250-serving-cert\") pod \"console-operator-58897d9998-kdfcd\" (UID: \"94469d04-47fd-4bd8-8258-9e874767f250\") " pod="openshift-console-operator/console-operator-58897d9998-kdfcd" Dec 13 06:51:35 crc kubenswrapper[4971]: I1213 06:51:35.911302 4971 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/22f3a4f0-1686-4087-bcec-e48f2015ffe2-trusted-ca\") pod \"ingress-operator-5b745b69d9-pkmb5\" (UID: \"22f3a4f0-1686-4087-bcec-e48f2015ffe2\") " pod="openshift-ingress-operator/ingress-operator-5b745b69d9-pkmb5" Dec 13 06:51:35 crc kubenswrapper[4971]: I1213 06:51:35.911331 4971 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-njt75\" (UniqueName: \"kubernetes.io/projected/5764a8c4-18e5-4025-8d36-15a7b01e9169-kube-api-access-njt75\") pod \"migrator-59844c95c7-z2p8z\" (UID: \"5764a8c4-18e5-4025-8d36-15a7b01e9169\") " pod="openshift-kube-storage-version-migrator/migrator-59844c95c7-z2p8z" Dec 13 06:51:35 crc kubenswrapper[4971]: I1213 06:51:35.911383 4971 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-qdm7j\" (UniqueName: \"kubernetes.io/projected/c1575ff9-1c53-4732-bc48-252f9f270597-kube-api-access-qdm7j\") pod \"collect-profiles-29426805-fxxrz\" (UID: \"c1575ff9-1c53-4732-bc48-252f9f270597\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29426805-fxxrz" Dec 13 06:51:35 crc kubenswrapper[4971]: I1213 06:51:35.911406 4971 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"package-server-manager-serving-cert\" (UniqueName: \"kubernetes.io/secret/479b25e6-6618-4123-a913-ca50d6c4ba28-package-server-manager-serving-cert\") pod \"package-server-manager-789f6589d5-g8vfp\" (UID: \"479b25e6-6618-4123-a913-ca50d6c4ba28\") " pod="openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-g8vfp" Dec 13 06:51:35 crc kubenswrapper[4971]: I1213 06:51:35.911429 4971 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-tlp42\" (UniqueName: \"kubernetes.io/projected/49f3dc41-5142-4cdb-9bb3-6d7248228535-kube-api-access-tlp42\") pod \"service-ca-operator-777779d784-drq5g\" (UID: \"49f3dc41-5142-4cdb-9bb3-6d7248228535\") " pod="openshift-service-ca-operator/service-ca-operator-777779d784-drq5g" Dec 13 06:51:35 crc kubenswrapper[4971]: I1213 06:51:35.911451 4971 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-8zjpk\" (UniqueName: \"kubernetes.io/projected/ea24f72e-df9d-418f-aaef-dfc0573eface-kube-api-access-8zjpk\") pod \"olm-operator-6b444d44fb-pxjqw\" (UID: \"ea24f72e-df9d-418f-aaef-dfc0573eface\") " pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-pxjqw" Dec 13 06:51:35 crc kubenswrapper[4971]: I1213 06:51:35.911572 4971 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/cba893c6-ff99-466a-9db0-a62e7a87a334-bound-sa-token\") pod \"image-registry-697d97f7c8-mdd8q\" (UID: \"cba893c6-ff99-466a-9db0-a62e7a87a334\") " pod="openshift-image-registry/image-registry-697d97f7c8-mdd8q" Dec 13 06:51:35 crc kubenswrapper[4971]: I1213 06:51:35.911639 4971 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"service-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/c0ca0ce3-929c-48d6-a3f7-6d334e3c3709-service-ca-bundle\") pod \"authentication-operator-69f744f599-wjdd5\" (UID: \"c0ca0ce3-929c-48d6-a3f7-6d334e3c3709\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-wjdd5" Dec 13 06:51:35 crc kubenswrapper[4971]: I1213 06:51:35.911689 4971 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/c0ca0ce3-929c-48d6-a3f7-6d334e3c3709-trusted-ca-bundle\") pod \"authentication-operator-69f744f599-wjdd5\" (UID: \"c0ca0ce3-929c-48d6-a3f7-6d334e3c3709\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-wjdd5" Dec 13 06:51:35 crc kubenswrapper[4971]: I1213 06:51:35.911712 4971 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-mkgwg\" (UniqueName: \"kubernetes.io/projected/2cf6e363-8511-4c79-9181-79dba726e5af-kube-api-access-mkgwg\") pod \"route-controller-manager-6576b87f9c-9d86n\" (UID: \"2cf6e363-8511-4c79-9181-79dba726e5af\") " pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-9d86n" Dec 13 06:51:35 crc kubenswrapper[4971]: I1213 06:51:35.911767 4971 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/94469d04-47fd-4bd8-8258-9e874767f250-config\") pod \"console-operator-58897d9998-kdfcd\" (UID: \"94469d04-47fd-4bd8-8258-9e874767f250\") " pod="openshift-console-operator/console-operator-58897d9998-kdfcd" Dec 13 06:51:35 crc kubenswrapper[4971]: I1213 06:51:35.911790 4971 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-v95qb\" (UniqueName: \"kubernetes.io/projected/479b25e6-6618-4123-a913-ca50d6c4ba28-kube-api-access-v95qb\") pod \"package-server-manager-789f6589d5-g8vfp\" (UID: \"479b25e6-6618-4123-a913-ca50d6c4ba28\") " pod="openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-g8vfp" Dec 13 06:51:35 crc kubenswrapper[4971]: I1213 06:51:35.911812 4971 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-vsp8z\" (UniqueName: \"kubernetes.io/projected/4d847949-530c-4cbc-b11d-c1c8d80d3db7-kube-api-access-vsp8z\") pod \"ingress-canary-ffhlq\" (UID: \"4d847949-530c-4cbc-b11d-c1c8d80d3db7\") " pod="openshift-ingress-canary/ingress-canary-ffhlq" Dec 13 06:51:35 crc kubenswrapper[4971]: I1213 06:51:35.911847 4971 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"profile-collector-cert\" (UniqueName: \"kubernetes.io/secret/ea24f72e-df9d-418f-aaef-dfc0573eface-profile-collector-cert\") pod \"olm-operator-6b444d44fb-pxjqw\" (UID: \"ea24f72e-df9d-418f-aaef-dfc0573eface\") " pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-pxjqw" Dec 13 06:51:35 crc kubenswrapper[4971]: I1213 06:51:35.911869 4971 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/49f3dc41-5142-4cdb-9bb3-6d7248228535-serving-cert\") pod \"service-ca-operator-777779d784-drq5g\" (UID: \"49f3dc41-5142-4cdb-9bb3-6d7248228535\") " pod="openshift-service-ca-operator/service-ca-operator-777779d784-drq5g" Dec 13 06:51:35 crc kubenswrapper[4971]: I1213 06:51:35.911888 4971 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/87f8bc63-d8c1-415f-975f-b8418034197b-webhook-cert\") pod \"packageserver-d55dfcdfc-lnrkm\" (UID: \"87f8bc63-d8c1-415f-975f-b8418034197b\") " pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-lnrkm" Dec 13 06:51:35 crc kubenswrapper[4971]: I1213 06:51:35.911907 4971 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/637e887c-db01-4b47-9f17-2d61b731a118-config-volume\") pod \"dns-default-dntmz\" (UID: \"637e887c-db01-4b47-9f17-2d61b731a118\") " pod="openshift-dns/dns-default-dntmz" Dec 13 06:51:35 crc kubenswrapper[4971]: I1213 06:51:35.911931 4971 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/22f3a4f0-1686-4087-bcec-e48f2015ffe2-metrics-tls\") pod \"ingress-operator-5b745b69d9-pkmb5\" (UID: \"22f3a4f0-1686-4087-bcec-e48f2015ffe2\") " pod="openshift-ingress-operator/ingress-operator-5b745b69d9-pkmb5" Dec 13 06:51:35 crc kubenswrapper[4971]: I1213 06:51:35.911964 4971 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-64btv\" (UniqueName: \"kubernetes.io/projected/930fe38c-5ff7-45a8-a292-cd69d07e86f7-kube-api-access-64btv\") pod \"control-plane-machine-set-operator-78cbb6b69f-bgdbk\" (UID: \"930fe38c-5ff7-45a8-a292-cd69d07e86f7\") " pod="openshift-machine-api/control-plane-machine-set-operator-78cbb6b69f-bgdbk" Dec 13 06:51:35 crc kubenswrapper[4971]: I1213 06:51:35.911997 4971 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"registry-certificates\" (UniqueName: \"kubernetes.io/configmap/cba893c6-ff99-466a-9db0-a62e7a87a334-registry-certificates\") pod \"image-registry-697d97f7c8-mdd8q\" (UID: \"cba893c6-ff99-466a-9db0-a62e7a87a334\") " pod="openshift-image-registry/image-registry-697d97f7c8-mdd8q" Dec 13 06:51:35 crc kubenswrapper[4971]: I1213 06:51:35.912027 4971 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ca-trust-extracted\" (UniqueName: \"kubernetes.io/empty-dir/cba893c6-ff99-466a-9db0-a62e7a87a334-ca-trust-extracted\") pod \"image-registry-697d97f7c8-mdd8q\" (UID: \"cba893c6-ff99-466a-9db0-a62e7a87a334\") " pod="openshift-image-registry/image-registry-697d97f7c8-mdd8q" Dec 13 06:51:35 crc kubenswrapper[4971]: I1213 06:51:35.912050 4971 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"control-plane-machine-set-operator-tls\" (UniqueName: \"kubernetes.io/secret/930fe38c-5ff7-45a8-a292-cd69d07e86f7-control-plane-machine-set-operator-tls\") pod \"control-plane-machine-set-operator-78cbb6b69f-bgdbk\" (UID: \"930fe38c-5ff7-45a8-a292-cd69d07e86f7\") " pod="openshift-machine-api/control-plane-machine-set-operator-78cbb6b69f-bgdbk" Dec 13 06:51:35 crc kubenswrapper[4971]: I1213 06:51:35.912071 4971 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/c0ca0ce3-929c-48d6-a3f7-6d334e3c3709-serving-cert\") pod \"authentication-operator-69f744f599-wjdd5\" (UID: \"c0ca0ce3-929c-48d6-a3f7-6d334e3c3709\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-wjdd5" Dec 13 06:51:35 crc kubenswrapper[4971]: I1213 06:51:35.912092 4971 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/94469d04-47fd-4bd8-8258-9e874767f250-trusted-ca\") pod \"console-operator-58897d9998-kdfcd\" (UID: \"94469d04-47fd-4bd8-8258-9e874767f250\") " pod="openshift-console-operator/console-operator-58897d9998-kdfcd" Dec 13 06:51:35 crc kubenswrapper[4971]: I1213 06:51:35.912124 4971 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"service-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/d89e3583-3f66-4259-b6f9-bffe8ddbde12-service-ca-bundle\") pod \"router-default-5444994796-mzhp6\" (UID: \"d89e3583-3f66-4259-b6f9-bffe8ddbde12\") " pod="openshift-ingress/router-default-5444994796-mzhp6" Dec 13 06:51:35 crc kubenswrapper[4971]: I1213 06:51:35.912144 4971 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"installation-pull-secrets\" (UniqueName: \"kubernetes.io/secret/cba893c6-ff99-466a-9db0-a62e7a87a334-installation-pull-secrets\") pod \"image-registry-697d97f7c8-mdd8q\" (UID: \"cba893c6-ff99-466a-9db0-a62e7a87a334\") " pod="openshift-image-registry/image-registry-697d97f7c8-mdd8q" Dec 13 06:51:35 crc kubenswrapper[4971]: I1213 06:51:35.912164 4971 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/2cf6e363-8511-4c79-9181-79dba726e5af-serving-cert\") pod \"route-controller-manager-6576b87f9c-9d86n\" (UID: \"2cf6e363-8511-4c79-9181-79dba726e5af\") " pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-9d86n" Dec 13 06:51:35 crc kubenswrapper[4971]: I1213 06:51:35.912209 4971 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-fwvzw\" (UniqueName: \"kubernetes.io/projected/87f8bc63-d8c1-415f-975f-b8418034197b-kube-api-access-fwvzw\") pod \"packageserver-d55dfcdfc-lnrkm\" (UID: \"87f8bc63-d8c1-415f-975f-b8418034197b\") " pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-lnrkm" Dec 13 06:51:35 crc kubenswrapper[4971]: I1213 06:51:35.912243 4971 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/d89e3583-3f66-4259-b6f9-bffe8ddbde12-metrics-certs\") pod \"router-default-5444994796-mzhp6\" (UID: \"d89e3583-3f66-4259-b6f9-bffe8ddbde12\") " pod="openshift-ingress/router-default-5444994796-mzhp6" Dec 13 06:51:35 crc kubenswrapper[4971]: I1213 06:51:35.912265 4971 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-4ccc9\" (UniqueName: \"kubernetes.io/projected/3052c2e3-f9bf-442b-b153-d60eb86f9c3a-kube-api-access-4ccc9\") pod \"catalog-operator-68c6474976-l4rzp\" (UID: \"3052c2e3-f9bf-442b-b153-d60eb86f9c3a\") " pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-l4rzp" Dec 13 06:51:35 crc kubenswrapper[4971]: I1213 06:51:35.912285 4971 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-jdnll\" (UniqueName: \"kubernetes.io/projected/1c4c5572-81b1-4270-b519-b077f9682f49-kube-api-access-jdnll\") pod \"machine-config-server-xdxjz\" (UID: \"1c4c5572-81b1-4270-b519-b077f9682f49\") " pod="openshift-machine-config-operator/machine-config-server-xdxjz" Dec 13 06:51:35 crc kubenswrapper[4971]: I1213 06:51:35.912305 4971 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/6e6307eb-6448-4cbb-b2a3-93e340f5b3a4-kube-api-access\") pod \"kube-controller-manager-operator-78b949d7b-cwwf9\" (UID: \"6e6307eb-6448-4cbb-b2a3-93e340f5b3a4\") " pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-cwwf9" Dec 13 06:51:35 crc kubenswrapper[4971]: I1213 06:51:35.912340 4971 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-vxdq2\" (UniqueName: \"kubernetes.io/projected/22f3a4f0-1686-4087-bcec-e48f2015ffe2-kube-api-access-vxdq2\") pod \"ingress-operator-5b745b69d9-pkmb5\" (UID: \"22f3a4f0-1686-4087-bcec-e48f2015ffe2\") " pod="openshift-ingress-operator/ingress-operator-5b745b69d9-pkmb5" Dec 13 06:51:35 crc kubenswrapper[4971]: I1213 06:51:35.912417 4971 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/c0ca0ce3-929c-48d6-a3f7-6d334e3c3709-config\") pod \"authentication-operator-69f744f599-wjdd5\" (UID: \"c0ca0ce3-929c-48d6-a3f7-6d334e3c3709\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-wjdd5" Dec 13 06:51:35 crc kubenswrapper[4971]: I1213 06:51:35.912440 4971 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"node-bootstrap-token\" (UniqueName: \"kubernetes.io/secret/1c4c5572-81b1-4270-b519-b077f9682f49-node-bootstrap-token\") pod \"machine-config-server-xdxjz\" (UID: \"1c4c5572-81b1-4270-b519-b077f9682f49\") " pod="openshift-machine-config-operator/machine-config-server-xdxjz" Dec 13 06:51:35 crc kubenswrapper[4971]: I1213 06:51:35.912473 4971 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/4b58bd40-cb76-46ba-9391-a350d830d729-serving-cert\") pod \"openshift-kube-scheduler-operator-5fdd9b5758-fz6wk\" (UID: \"4b58bd40-cb76-46ba-9391-a350d830d729\") " pod="openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-fz6wk" Dec 13 06:51:35 crc kubenswrapper[4971]: I1213 06:51:35.912492 4971 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/c1575ff9-1c53-4732-bc48-252f9f270597-secret-volume\") pod \"collect-profiles-29426805-fxxrz\" (UID: \"c1575ff9-1c53-4732-bc48-252f9f270597\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29426805-fxxrz" Dec 13 06:51:35 crc kubenswrapper[4971]: I1213 06:51:35.912563 4971 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-c5lc5\" (UniqueName: \"kubernetes.io/projected/6a3649bc-2d2e-4897-8977-99c08aa77e01-kube-api-access-c5lc5\") pod \"kube-storage-version-migrator-operator-b67b599dd-6blxl\" (UID: \"6a3649bc-2d2e-4897-8977-99c08aa77e01\") " pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-6blxl" Dec 13 06:51:35 crc kubenswrapper[4971]: I1213 06:51:35.912618 4971 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"socket-dir\" (UniqueName: \"kubernetes.io/host-path/5fd61507-84bf-4ef5-8bd4-731075ef58ef-socket-dir\") pod \"csi-hostpathplugin-7nhhk\" (UID: \"5fd61507-84bf-4ef5-8bd4-731075ef58ef\") " pod="hostpath-provisioner/csi-hostpathplugin-7nhhk" Dec 13 06:51:35 crc kubenswrapper[4971]: I1213 06:51:35.912640 4971 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-wnzs2\" (UniqueName: \"kubernetes.io/projected/5fd61507-84bf-4ef5-8bd4-731075ef58ef-kube-api-access-wnzs2\") pod \"csi-hostpathplugin-7nhhk\" (UID: \"5fd61507-84bf-4ef5-8bd4-731075ef58ef\") " pod="hostpath-provisioner/csi-hostpathplugin-7nhhk" Dec 13 06:51:35 crc kubenswrapper[4971]: I1213 06:51:35.912700 4971 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-fs22c\" (UniqueName: \"kubernetes.io/projected/cba893c6-ff99-466a-9db0-a62e7a87a334-kube-api-access-fs22c\") pod \"image-registry-697d97f7c8-mdd8q\" (UID: \"cba893c6-ff99-466a-9db0-a62e7a87a334\") " pod="openshift-image-registry/image-registry-697d97f7c8-mdd8q" Dec 13 06:51:35 crc kubenswrapper[4971]: I1213 06:51:35.912722 4971 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/6e6307eb-6448-4cbb-b2a3-93e340f5b3a4-config\") pod \"kube-controller-manager-operator-78b949d7b-cwwf9\" (UID: \"6e6307eb-6448-4cbb-b2a3-93e340f5b3a4\") " pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-cwwf9" Dec 13 06:51:35 crc kubenswrapper[4971]: I1213 06:51:35.912758 4971 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/22f3a4f0-1686-4087-bcec-e48f2015ffe2-bound-sa-token\") pod \"ingress-operator-5b745b69d9-pkmb5\" (UID: \"22f3a4f0-1686-4087-bcec-e48f2015ffe2\") " pod="openshift-ingress-operator/ingress-operator-5b745b69d9-pkmb5" Dec 13 06:51:35 crc kubenswrapper[4971]: I1213 06:51:35.912780 4971 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/375613f8-1e50-42fe-b593-2a6a07afbe64-kube-api-access\") pod \"kube-apiserver-operator-766d6c64bb-xfvfg\" (UID: \"375613f8-1e50-42fe-b593-2a6a07afbe64\") " pod="openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-xfvfg" Dec 13 06:51:35 crc kubenswrapper[4971]: I1213 06:51:35.912801 4971 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/4b58bd40-cb76-46ba-9391-a350d830d729-config\") pod \"openshift-kube-scheduler-operator-5fdd9b5758-fz6wk\" (UID: \"4b58bd40-cb76-46ba-9391-a350d830d729\") " pod="openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-fz6wk" Dec 13 06:51:35 crc kubenswrapper[4971]: I1213 06:51:35.912927 4971 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/6a3649bc-2d2e-4897-8977-99c08aa77e01-serving-cert\") pod \"kube-storage-version-migrator-operator-b67b599dd-6blxl\" (UID: \"6a3649bc-2d2e-4897-8977-99c08aa77e01\") " pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-6blxl" Dec 13 06:51:35 crc kubenswrapper[4971]: I1213 06:51:35.912953 4971 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"srv-cert\" (UniqueName: \"kubernetes.io/secret/3052c2e3-f9bf-442b-b153-d60eb86f9c3a-srv-cert\") pod \"catalog-operator-68c6474976-l4rzp\" (UID: \"3052c2e3-f9bf-442b-b153-d60eb86f9c3a\") " pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-l4rzp" Dec 13 06:51:35 crc kubenswrapper[4971]: I1213 06:51:35.912971 4971 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/49f3dc41-5142-4cdb-9bb3-6d7248228535-config\") pod \"service-ca-operator-777779d784-drq5g\" (UID: \"49f3dc41-5142-4cdb-9bb3-6d7248228535\") " pod="openshift-service-ca-operator/service-ca-operator-777779d784-drq5g" Dec 13 06:51:35 crc kubenswrapper[4971]: I1213 06:51:35.912990 4971 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/c1575ff9-1c53-4732-bc48-252f9f270597-config-volume\") pod \"collect-profiles-29426805-fxxrz\" (UID: \"c1575ff9-1c53-4732-bc48-252f9f270597\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29426805-fxxrz" Dec 13 06:51:35 crc kubenswrapper[4971]: I1213 06:51:35.913012 4971 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-fdpnd\" (UniqueName: \"kubernetes.io/projected/c0ca0ce3-929c-48d6-a3f7-6d334e3c3709-kube-api-access-fdpnd\") pod \"authentication-operator-69f744f599-wjdd5\" (UID: \"c0ca0ce3-929c-48d6-a3f7-6d334e3c3709\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-wjdd5" Dec 13 06:51:35 crc kubenswrapper[4971]: I1213 06:51:35.913036 4971 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/874a112c-101d-4a74-a674-d9afe62b6dc8-proxy-tls\") pod \"machine-config-operator-74547568cd-4cn2z\" (UID: \"874a112c-101d-4a74-a674-d9afe62b6dc8\") " pod="openshift-machine-config-operator/machine-config-operator-74547568cd-4cn2z" Dec 13 06:51:35 crc kubenswrapper[4971]: I1213 06:51:35.913056 4971 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/874a112c-101d-4a74-a674-d9afe62b6dc8-auth-proxy-config\") pod \"machine-config-operator-74547568cd-4cn2z\" (UID: \"874a112c-101d-4a74-a674-d9afe62b6dc8\") " pod="openshift-machine-config-operator/machine-config-operator-74547568cd-4cn2z" Dec 13 06:51:35 crc kubenswrapper[4971]: I1213 06:51:35.913078 4971 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/6e6307eb-6448-4cbb-b2a3-93e340f5b3a4-serving-cert\") pod \"kube-controller-manager-operator-78b949d7b-cwwf9\" (UID: \"6e6307eb-6448-4cbb-b2a3-93e340f5b3a4\") " pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-cwwf9" Dec 13 06:51:35 crc kubenswrapper[4971]: I1213 06:51:35.913110 4971 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"default-certificate\" (UniqueName: \"kubernetes.io/secret/d89e3583-3f66-4259-b6f9-bffe8ddbde12-default-certificate\") pod \"router-default-5444994796-mzhp6\" (UID: \"d89e3583-3f66-4259-b6f9-bffe8ddbde12\") " pod="openshift-ingress/router-default-5444994796-mzhp6" Dec 13 06:51:35 crc kubenswrapper[4971]: I1213 06:51:35.913144 4971 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"tmpfs\" (UniqueName: \"kubernetes.io/empty-dir/87f8bc63-d8c1-415f-975f-b8418034197b-tmpfs\") pod \"packageserver-d55dfcdfc-lnrkm\" (UID: \"87f8bc63-d8c1-415f-975f-b8418034197b\") " pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-lnrkm" Dec 13 06:51:35 crc kubenswrapper[4971]: I1213 06:51:35.913211 4971 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/2cf6e363-8511-4c79-9181-79dba726e5af-client-ca\") pod \"route-controller-manager-6576b87f9c-9d86n\" (UID: \"2cf6e363-8511-4c79-9181-79dba726e5af\") " pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-9d86n" Dec 13 06:51:35 crc kubenswrapper[4971]: I1213 06:51:35.913234 4971 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/4b58bd40-cb76-46ba-9391-a350d830d729-kube-api-access\") pod \"openshift-kube-scheduler-operator-5fdd9b5758-fz6wk\" (UID: \"4b58bd40-cb76-46ba-9391-a350d830d729\") " pod="openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-fz6wk" Dec 13 06:51:35 crc kubenswrapper[4971]: I1213 06:51:35.913274 4971 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-lltlz\" (UniqueName: \"kubernetes.io/projected/874a112c-101d-4a74-a674-d9afe62b6dc8-kube-api-access-lltlz\") pod \"machine-config-operator-74547568cd-4cn2z\" (UID: \"874a112c-101d-4a74-a674-d9afe62b6dc8\") " pod="openshift-machine-config-operator/machine-config-operator-74547568cd-4cn2z" Dec 13 06:51:35 crc kubenswrapper[4971]: I1213 06:51:35.913294 4971 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"srv-cert\" (UniqueName: \"kubernetes.io/secret/ea24f72e-df9d-418f-aaef-dfc0573eface-srv-cert\") pod \"olm-operator-6b444d44fb-pxjqw\" (UID: \"ea24f72e-df9d-418f-aaef-dfc0573eface\") " pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-pxjqw" Dec 13 06:51:35 crc kubenswrapper[4971]: I1213 06:51:35.913315 4971 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/87f8bc63-d8c1-415f-975f-b8418034197b-apiservice-cert\") pod \"packageserver-d55dfcdfc-lnrkm\" (UID: \"87f8bc63-d8c1-415f-975f-b8418034197b\") " pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-lnrkm" Dec 13 06:51:35 crc kubenswrapper[4971]: I1213 06:51:35.913334 4971 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/6a3649bc-2d2e-4897-8977-99c08aa77e01-config\") pod \"kube-storage-version-migrator-operator-b67b599dd-6blxl\" (UID: \"6a3649bc-2d2e-4897-8977-99c08aa77e01\") " pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-6blxl" Dec 13 06:51:35 crc kubenswrapper[4971]: I1213 06:51:35.913358 4971 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-fzfk6\" (UniqueName: \"kubernetes.io/projected/d89e3583-3f66-4259-b6f9-bffe8ddbde12-kube-api-access-fzfk6\") pod \"router-default-5444994796-mzhp6\" (UID: \"d89e3583-3f66-4259-b6f9-bffe8ddbde12\") " pod="openshift-ingress/router-default-5444994796-mzhp6" Dec 13 06:51:35 crc kubenswrapper[4971]: I1213 06:51:35.913390 4971 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/375613f8-1e50-42fe-b593-2a6a07afbe64-config\") pod \"kube-apiserver-operator-766d6c64bb-xfvfg\" (UID: \"375613f8-1e50-42fe-b593-2a6a07afbe64\") " pod="openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-xfvfg" Dec 13 06:51:35 crc kubenswrapper[4971]: I1213 06:51:35.913411 4971 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"mountpoint-dir\" (UniqueName: \"kubernetes.io/host-path/5fd61507-84bf-4ef5-8bd4-731075ef58ef-mountpoint-dir\") pod \"csi-hostpathplugin-7nhhk\" (UID: \"5fd61507-84bf-4ef5-8bd4-731075ef58ef\") " pod="hostpath-provisioner/csi-hostpathplugin-7nhhk" Dec 13 06:51:35 crc kubenswrapper[4971]: I1213 06:51:35.913430 4971 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"certs\" (UniqueName: \"kubernetes.io/secret/1c4c5572-81b1-4270-b519-b077f9682f49-certs\") pod \"machine-config-server-xdxjz\" (UID: \"1c4c5572-81b1-4270-b519-b077f9682f49\") " pod="openshift-machine-config-operator/machine-config-server-xdxjz" Dec 13 06:51:35 crc kubenswrapper[4971]: I1213 06:51:35.913456 4971 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-t2s9j\" (UniqueName: \"kubernetes.io/projected/ad158968-e110-4b4c-9f42-da9c08274dab-kube-api-access-t2s9j\") pod \"dns-operator-744455d44c-b999k\" (UID: \"ad158968-e110-4b4c-9f42-da9c08274dab\") " pod="openshift-dns-operator/dns-operator-744455d44c-b999k" Dec 13 06:51:35 crc kubenswrapper[4971]: I1213 06:51:35.913477 4971 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/4d847949-530c-4cbc-b11d-c1c8d80d3db7-cert\") pod \"ingress-canary-ffhlq\" (UID: \"4d847949-530c-4cbc-b11d-c1c8d80d3db7\") " pod="openshift-ingress-canary/ingress-canary-ffhlq" Dec 13 06:51:35 crc kubenswrapper[4971]: I1213 06:51:35.913546 4971 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"registry-tls\" (UniqueName: \"kubernetes.io/projected/cba893c6-ff99-466a-9db0-a62e7a87a334-registry-tls\") pod \"image-registry-697d97f7c8-mdd8q\" (UID: \"cba893c6-ff99-466a-9db0-a62e7a87a334\") " pod="openshift-image-registry/image-registry-697d97f7c8-mdd8q" Dec 13 06:51:35 crc kubenswrapper[4971]: I1213 06:51:35.913568 4971 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-2m57w\" (UniqueName: \"kubernetes.io/projected/637e887c-db01-4b47-9f17-2d61b731a118-kube-api-access-2m57w\") pod \"dns-default-dntmz\" (UID: \"637e887c-db01-4b47-9f17-2d61b731a118\") " pod="openshift-dns/dns-default-dntmz" Dec 13 06:51:35 crc kubenswrapper[4971]: I1213 06:51:35.913593 4971 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/2cf6e363-8511-4c79-9181-79dba726e5af-config\") pod \"route-controller-manager-6576b87f9c-9d86n\" (UID: \"2cf6e363-8511-4c79-9181-79dba726e5af\") " pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-9d86n" Dec 13 06:51:35 crc kubenswrapper[4971]: I1213 06:51:35.913617 4971 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"registration-dir\" (UniqueName: \"kubernetes.io/host-path/5fd61507-84bf-4ef5-8bd4-731075ef58ef-registration-dir\") pod \"csi-hostpathplugin-7nhhk\" (UID: \"5fd61507-84bf-4ef5-8bd4-731075ef58ef\") " pod="hostpath-provisioner/csi-hostpathplugin-7nhhk" Dec 13 06:51:35 crc kubenswrapper[4971]: I1213 06:51:35.913653 4971 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/375613f8-1e50-42fe-b593-2a6a07afbe64-serving-cert\") pod \"kube-apiserver-operator-766d6c64bb-xfvfg\" (UID: \"375613f8-1e50-42fe-b593-2a6a07afbe64\") " pod="openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-xfvfg" Dec 13 06:51:35 crc kubenswrapper[4971]: I1213 06:51:35.913674 4971 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"plugins-dir\" (UniqueName: \"kubernetes.io/host-path/5fd61507-84bf-4ef5-8bd4-731075ef58ef-plugins-dir\") pod \"csi-hostpathplugin-7nhhk\" (UID: \"5fd61507-84bf-4ef5-8bd4-731075ef58ef\") " pod="hostpath-provisioner/csi-hostpathplugin-7nhhk" Dec 13 06:51:35 crc kubenswrapper[4971]: I1213 06:51:35.913695 4971 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"profile-collector-cert\" (UniqueName: \"kubernetes.io/secret/3052c2e3-f9bf-442b-b153-d60eb86f9c3a-profile-collector-cert\") pod \"catalog-operator-68c6474976-l4rzp\" (UID: \"3052c2e3-f9bf-442b-b153-d60eb86f9c3a\") " pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-l4rzp" Dec 13 06:51:35 crc kubenswrapper[4971]: I1213 06:51:35.913715 4971 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"csi-data-dir\" (UniqueName: \"kubernetes.io/host-path/5fd61507-84bf-4ef5-8bd4-731075ef58ef-csi-data-dir\") pod \"csi-hostpathplugin-7nhhk\" (UID: \"5fd61507-84bf-4ef5-8bd4-731075ef58ef\") " pod="hostpath-provisioner/csi-hostpathplugin-7nhhk" Dec 13 06:51:35 crc kubenswrapper[4971]: I1213 06:51:35.913739 4971 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/cba893c6-ff99-466a-9db0-a62e7a87a334-trusted-ca\") pod \"image-registry-697d97f7c8-mdd8q\" (UID: \"cba893c6-ff99-466a-9db0-a62e7a87a334\") " pod="openshift-image-registry/image-registry-697d97f7c8-mdd8q" Dec 13 06:51:35 crc kubenswrapper[4971]: I1213 06:51:35.913760 4971 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-jk6gm\" (UniqueName: \"kubernetes.io/projected/ee2e94f4-cadb-4b17-8853-a1292cb02404-kube-api-access-jk6gm\") pod \"downloads-7954f5f757-dqk9z\" (UID: \"ee2e94f4-cadb-4b17-8853-a1292cb02404\") " pod="openshift-console/downloads-7954f5f757-dqk9z" Dec 13 06:51:35 crc kubenswrapper[4971]: I1213 06:51:35.913807 4971 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-2gk4l\" (UniqueName: \"kubernetes.io/projected/94469d04-47fd-4bd8-8258-9e874767f250-kube-api-access-2gk4l\") pod \"console-operator-58897d9998-kdfcd\" (UID: \"94469d04-47fd-4bd8-8258-9e874767f250\") " pod="openshift-console-operator/console-operator-58897d9998-kdfcd" Dec 13 06:51:35 crc kubenswrapper[4971]: I1213 06:51:35.913830 4971 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/637e887c-db01-4b47-9f17-2d61b731a118-metrics-tls\") pod \"dns-default-dntmz\" (UID: \"637e887c-db01-4b47-9f17-2d61b731a118\") " pod="openshift-dns/dns-default-dntmz" Dec 13 06:51:35 crc kubenswrapper[4971]: I1213 06:51:35.913851 4971 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"images\" (UniqueName: \"kubernetes.io/configmap/874a112c-101d-4a74-a674-d9afe62b6dc8-images\") pod \"machine-config-operator-74547568cd-4cn2z\" (UID: \"874a112c-101d-4a74-a674-d9afe62b6dc8\") " pod="openshift-machine-config-operator/machine-config-operator-74547568cd-4cn2z" Dec 13 06:51:35 crc kubenswrapper[4971]: I1213 06:51:35.913870 4971 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/ad158968-e110-4b4c-9f42-da9c08274dab-metrics-tls\") pod \"dns-operator-744455d44c-b999k\" (UID: \"ad158968-e110-4b4c-9f42-da9c08274dab\") " pod="openshift-dns-operator/dns-operator-744455d44c-b999k" Dec 13 06:51:35 crc kubenswrapper[4971]: I1213 06:51:35.913896 4971 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"stats-auth\" (UniqueName: \"kubernetes.io/secret/d89e3583-3f66-4259-b6f9-bffe8ddbde12-stats-auth\") pod \"router-default-5444994796-mzhp6\" (UID: \"d89e3583-3f66-4259-b6f9-bffe8ddbde12\") " pod="openshift-ingress/router-default-5444994796-mzhp6" Dec 13 06:51:35 crc kubenswrapper[4971]: E1213 06:51:35.914865 4971 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-13 06:51:36.414844457 +0000 UTC m=+153.019253905 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 13 06:51:35 crc kubenswrapper[4971]: I1213 06:51:35.922178 4971 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/22f3a4f0-1686-4087-bcec-e48f2015ffe2-trusted-ca\") pod \"ingress-operator-5b745b69d9-pkmb5\" (UID: \"22f3a4f0-1686-4087-bcec-e48f2015ffe2\") " pod="openshift-ingress-operator/ingress-operator-5b745b69d9-pkmb5" Dec 13 06:51:35 crc kubenswrapper[4971]: I1213 06:51:35.923349 4971 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"stats-auth\" (UniqueName: \"kubernetes.io/secret/d89e3583-3f66-4259-b6f9-bffe8ddbde12-stats-auth\") pod \"router-default-5444994796-mzhp6\" (UID: \"d89e3583-3f66-4259-b6f9-bffe8ddbde12\") " pod="openshift-ingress/router-default-5444994796-mzhp6" Dec 13 06:51:35 crc kubenswrapper[4971]: I1213 06:51:35.924197 4971 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/874a112c-101d-4a74-a674-d9afe62b6dc8-auth-proxy-config\") pod \"machine-config-operator-74547568cd-4cn2z\" (UID: \"874a112c-101d-4a74-a674-d9afe62b6dc8\") " pod="openshift-machine-config-operator/machine-config-operator-74547568cd-4cn2z" Dec 13 06:51:35 crc kubenswrapper[4971]: I1213 06:51:35.924509 4971 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/94469d04-47fd-4bd8-8258-9e874767f250-serving-cert\") pod \"console-operator-58897d9998-kdfcd\" (UID: \"94469d04-47fd-4bd8-8258-9e874767f250\") " pod="openshift-console-operator/console-operator-58897d9998-kdfcd" Dec 13 06:51:35 crc kubenswrapper[4971]: I1213 06:51:35.925511 4971 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/94469d04-47fd-4bd8-8258-9e874767f250-config\") pod \"console-operator-58897d9998-kdfcd\" (UID: \"94469d04-47fd-4bd8-8258-9e874767f250\") " pod="openshift-console-operator/console-operator-58897d9998-kdfcd" Dec 13 06:51:35 crc kubenswrapper[4971]: I1213 06:51:35.927371 4971 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/2cf6e363-8511-4c79-9181-79dba726e5af-config\") pod \"route-controller-manager-6576b87f9c-9d86n\" (UID: \"2cf6e363-8511-4c79-9181-79dba726e5af\") " pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-9d86n" Dec 13 06:51:35 crc kubenswrapper[4971]: I1213 06:51:35.927817 4971 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/375613f8-1e50-42fe-b593-2a6a07afbe64-config\") pod \"kube-apiserver-operator-766d6c64bb-xfvfg\" (UID: \"375613f8-1e50-42fe-b593-2a6a07afbe64\") " pod="openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-xfvfg" Dec 13 06:51:35 crc kubenswrapper[4971]: I1213 06:51:35.928620 4971 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ca-trust-extracted\" (UniqueName: \"kubernetes.io/empty-dir/cba893c6-ff99-466a-9db0-a62e7a87a334-ca-trust-extracted\") pod \"image-registry-697d97f7c8-mdd8q\" (UID: \"cba893c6-ff99-466a-9db0-a62e7a87a334\") " pod="openshift-image-registry/image-registry-697d97f7c8-mdd8q" Dec 13 06:51:35 crc kubenswrapper[4971]: I1213 06:51:35.930194 4971 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"registry-certificates\" (UniqueName: \"kubernetes.io/configmap/cba893c6-ff99-466a-9db0-a62e7a87a334-registry-certificates\") pod \"image-registry-697d97f7c8-mdd8q\" (UID: \"cba893c6-ff99-466a-9db0-a62e7a87a334\") " pod="openshift-image-registry/image-registry-697d97f7c8-mdd8q" Dec 13 06:51:35 crc kubenswrapper[4971]: I1213 06:51:35.931141 4971 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/22f3a4f0-1686-4087-bcec-e48f2015ffe2-metrics-tls\") pod \"ingress-operator-5b745b69d9-pkmb5\" (UID: \"22f3a4f0-1686-4087-bcec-e48f2015ffe2\") " pod="openshift-ingress-operator/ingress-operator-5b745b69d9-pkmb5" Dec 13 06:51:35 crc kubenswrapper[4971]: I1213 06:51:35.931585 4971 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"service-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/d89e3583-3f66-4259-b6f9-bffe8ddbde12-service-ca-bundle\") pod \"router-default-5444994796-mzhp6\" (UID: \"d89e3583-3f66-4259-b6f9-bffe8ddbde12\") " pod="openshift-ingress/router-default-5444994796-mzhp6" Dec 13 06:51:35 crc kubenswrapper[4971]: I1213 06:51:35.935702 4971 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"images\" (UniqueName: \"kubernetes.io/configmap/874a112c-101d-4a74-a674-d9afe62b6dc8-images\") pod \"machine-config-operator-74547568cd-4cn2z\" (UID: \"874a112c-101d-4a74-a674-d9afe62b6dc8\") " pod="openshift-machine-config-operator/machine-config-operator-74547568cd-4cn2z" Dec 13 06:51:35 crc kubenswrapper[4971]: I1213 06:51:35.935968 4971 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/cba893c6-ff99-466a-9db0-a62e7a87a334-trusted-ca\") pod \"image-registry-697d97f7c8-mdd8q\" (UID: \"cba893c6-ff99-466a-9db0-a62e7a87a334\") " pod="openshift-image-registry/image-registry-697d97f7c8-mdd8q" Dec 13 06:51:35 crc kubenswrapper[4971]: I1213 06:51:35.936895 4971 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/94469d04-47fd-4bd8-8258-9e874767f250-trusted-ca\") pod \"console-operator-58897d9998-kdfcd\" (UID: \"94469d04-47fd-4bd8-8258-9e874767f250\") " pod="openshift-console-operator/console-operator-58897d9998-kdfcd" Dec 13 06:51:35 crc kubenswrapper[4971]: I1213 06:51:35.937027 4971 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"default-certificate\" (UniqueName: \"kubernetes.io/secret/d89e3583-3f66-4259-b6f9-bffe8ddbde12-default-certificate\") pod \"router-default-5444994796-mzhp6\" (UID: \"d89e3583-3f66-4259-b6f9-bffe8ddbde12\") " pod="openshift-ingress/router-default-5444994796-mzhp6" Dec 13 06:51:35 crc kubenswrapper[4971]: I1213 06:51:35.937801 4971 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"registry-tls\" (UniqueName: \"kubernetes.io/projected/cba893c6-ff99-466a-9db0-a62e7a87a334-registry-tls\") pod \"image-registry-697d97f7c8-mdd8q\" (UID: \"cba893c6-ff99-466a-9db0-a62e7a87a334\") " pod="openshift-image-registry/image-registry-697d97f7c8-mdd8q" Dec 13 06:51:35 crc kubenswrapper[4971]: I1213 06:51:35.939646 4971 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/874a112c-101d-4a74-a674-d9afe62b6dc8-proxy-tls\") pod \"machine-config-operator-74547568cd-4cn2z\" (UID: \"874a112c-101d-4a74-a674-d9afe62b6dc8\") " pod="openshift-machine-config-operator/machine-config-operator-74547568cd-4cn2z" Dec 13 06:51:35 crc kubenswrapper[4971]: I1213 06:51:35.940231 4971 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/2cf6e363-8511-4c79-9181-79dba726e5af-serving-cert\") pod \"route-controller-manager-6576b87f9c-9d86n\" (UID: \"2cf6e363-8511-4c79-9181-79dba726e5af\") " pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-9d86n" Dec 13 06:51:35 crc kubenswrapper[4971]: I1213 06:51:35.940368 4971 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"installation-pull-secrets\" (UniqueName: \"kubernetes.io/secret/cba893c6-ff99-466a-9db0-a62e7a87a334-installation-pull-secrets\") pod \"image-registry-697d97f7c8-mdd8q\" (UID: \"cba893c6-ff99-466a-9db0-a62e7a87a334\") " pod="openshift-image-registry/image-registry-697d97f7c8-mdd8q" Dec 13 06:51:35 crc kubenswrapper[4971]: I1213 06:51:35.940648 4971 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"package-server-manager-serving-cert\" (UniqueName: \"kubernetes.io/secret/479b25e6-6618-4123-a913-ca50d6c4ba28-package-server-manager-serving-cert\") pod \"package-server-manager-789f6589d5-g8vfp\" (UID: \"479b25e6-6618-4123-a913-ca50d6c4ba28\") " pod="openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-g8vfp" Dec 13 06:51:35 crc kubenswrapper[4971]: I1213 06:51:35.946380 4971 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/2cf6e363-8511-4c79-9181-79dba726e5af-client-ca\") pod \"route-controller-manager-6576b87f9c-9d86n\" (UID: \"2cf6e363-8511-4c79-9181-79dba726e5af\") " pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-9d86n" Dec 13 06:51:35 crc kubenswrapper[4971]: I1213 06:51:35.946655 4971 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/375613f8-1e50-42fe-b593-2a6a07afbe64-serving-cert\") pod \"kube-apiserver-operator-766d6c64bb-xfvfg\" (UID: \"375613f8-1e50-42fe-b593-2a6a07afbe64\") " pod="openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-xfvfg" Dec 13 06:51:35 crc kubenswrapper[4971]: I1213 06:51:35.955404 4971 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/d89e3583-3f66-4259-b6f9-bffe8ddbde12-metrics-certs\") pod \"router-default-5444994796-mzhp6\" (UID: \"d89e3583-3f66-4259-b6f9-bffe8ddbde12\") " pod="openshift-ingress/router-default-5444994796-mzhp6" Dec 13 06:51:35 crc kubenswrapper[4971]: I1213 06:51:35.958470 4971 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/ad158968-e110-4b4c-9f42-da9c08274dab-metrics-tls\") pod \"dns-operator-744455d44c-b999k\" (UID: \"ad158968-e110-4b4c-9f42-da9c08274dab\") " pod="openshift-dns-operator/dns-operator-744455d44c-b999k" Dec 13 06:51:35 crc kubenswrapper[4971]: I1213 06:51:35.959278 4971 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"control-plane-machine-set-operator-tls\" (UniqueName: \"kubernetes.io/secret/930fe38c-5ff7-45a8-a292-cd69d07e86f7-control-plane-machine-set-operator-tls\") pod \"control-plane-machine-set-operator-78cbb6b69f-bgdbk\" (UID: \"930fe38c-5ff7-45a8-a292-cd69d07e86f7\") " pod="openshift-machine-api/control-plane-machine-set-operator-78cbb6b69f-bgdbk" Dec 13 06:51:35 crc kubenswrapper[4971]: I1213 06:51:35.960104 4971 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-fs22c\" (UniqueName: \"kubernetes.io/projected/cba893c6-ff99-466a-9db0-a62e7a87a334-kube-api-access-fs22c\") pod \"image-registry-697d97f7c8-mdd8q\" (UID: \"cba893c6-ff99-466a-9db0-a62e7a87a334\") " pod="openshift-image-registry/image-registry-697d97f7c8-mdd8q" Dec 13 06:51:35 crc kubenswrapper[4971]: I1213 06:51:35.963100 4971 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/22f3a4f0-1686-4087-bcec-e48f2015ffe2-bound-sa-token\") pod \"ingress-operator-5b745b69d9-pkmb5\" (UID: \"22f3a4f0-1686-4087-bcec-e48f2015ffe2\") " pod="openshift-ingress-operator/ingress-operator-5b745b69d9-pkmb5" Dec 13 06:51:35 crc kubenswrapper[4971]: I1213 06:51:35.971872 4971 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-machine-config-operator/machine-config-controller-84d6567774-kf7sp" Dec 13 06:51:35 crc kubenswrapper[4971]: I1213 06:51:35.984642 4971 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/375613f8-1e50-42fe-b593-2a6a07afbe64-kube-api-access\") pod \"kube-apiserver-operator-766d6c64bb-xfvfg\" (UID: \"375613f8-1e50-42fe-b593-2a6a07afbe64\") " pod="openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-xfvfg" Dec 13 06:51:35 crc kubenswrapper[4971]: I1213 06:51:35.986362 4971 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-apiserver/apiserver-76f77b778f-2n4cq"] Dec 13 06:51:36 crc kubenswrapper[4971]: I1213 06:51:36.015181 4971 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/c0ca0ce3-929c-48d6-a3f7-6d334e3c3709-trusted-ca-bundle\") pod \"authentication-operator-69f744f599-wjdd5\" (UID: \"c0ca0ce3-929c-48d6-a3f7-6d334e3c3709\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-wjdd5" Dec 13 06:51:36 crc kubenswrapper[4971]: I1213 06:51:36.015235 4971 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-vsp8z\" (UniqueName: \"kubernetes.io/projected/4d847949-530c-4cbc-b11d-c1c8d80d3db7-kube-api-access-vsp8z\") pod \"ingress-canary-ffhlq\" (UID: \"4d847949-530c-4cbc-b11d-c1c8d80d3db7\") " pod="openshift-ingress-canary/ingress-canary-ffhlq" Dec 13 06:51:36 crc kubenswrapper[4971]: I1213 06:51:36.015261 4971 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"profile-collector-cert\" (UniqueName: \"kubernetes.io/secret/ea24f72e-df9d-418f-aaef-dfc0573eface-profile-collector-cert\") pod \"olm-operator-6b444d44fb-pxjqw\" (UID: \"ea24f72e-df9d-418f-aaef-dfc0573eface\") " pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-pxjqw" Dec 13 06:51:36 crc kubenswrapper[4971]: I1213 06:51:36.015285 4971 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/49f3dc41-5142-4cdb-9bb3-6d7248228535-serving-cert\") pod \"service-ca-operator-777779d784-drq5g\" (UID: \"49f3dc41-5142-4cdb-9bb3-6d7248228535\") " pod="openshift-service-ca-operator/service-ca-operator-777779d784-drq5g" Dec 13 06:51:36 crc kubenswrapper[4971]: I1213 06:51:36.015306 4971 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/87f8bc63-d8c1-415f-975f-b8418034197b-webhook-cert\") pod \"packageserver-d55dfcdfc-lnrkm\" (UID: \"87f8bc63-d8c1-415f-975f-b8418034197b\") " pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-lnrkm" Dec 13 06:51:36 crc kubenswrapper[4971]: I1213 06:51:36.015331 4971 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/637e887c-db01-4b47-9f17-2d61b731a118-config-volume\") pod \"dns-default-dntmz\" (UID: \"637e887c-db01-4b47-9f17-2d61b731a118\") " pod="openshift-dns/dns-default-dntmz" Dec 13 06:51:36 crc kubenswrapper[4971]: I1213 06:51:36.015371 4971 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/c0ca0ce3-929c-48d6-a3f7-6d334e3c3709-serving-cert\") pod \"authentication-operator-69f744f599-wjdd5\" (UID: \"c0ca0ce3-929c-48d6-a3f7-6d334e3c3709\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-wjdd5" Dec 13 06:51:36 crc kubenswrapper[4971]: I1213 06:51:36.015400 4971 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-fwvzw\" (UniqueName: \"kubernetes.io/projected/87f8bc63-d8c1-415f-975f-b8418034197b-kube-api-access-fwvzw\") pod \"packageserver-d55dfcdfc-lnrkm\" (UID: \"87f8bc63-d8c1-415f-975f-b8418034197b\") " pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-lnrkm" Dec 13 06:51:36 crc kubenswrapper[4971]: I1213 06:51:36.015436 4971 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-4ccc9\" (UniqueName: \"kubernetes.io/projected/3052c2e3-f9bf-442b-b153-d60eb86f9c3a-kube-api-access-4ccc9\") pod \"catalog-operator-68c6474976-l4rzp\" (UID: \"3052c2e3-f9bf-442b-b153-d60eb86f9c3a\") " pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-l4rzp" Dec 13 06:51:36 crc kubenswrapper[4971]: I1213 06:51:36.015461 4971 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-jdnll\" (UniqueName: \"kubernetes.io/projected/1c4c5572-81b1-4270-b519-b077f9682f49-kube-api-access-jdnll\") pod \"machine-config-server-xdxjz\" (UID: \"1c4c5572-81b1-4270-b519-b077f9682f49\") " pod="openshift-machine-config-operator/machine-config-server-xdxjz" Dec 13 06:51:36 crc kubenswrapper[4971]: I1213 06:51:36.015484 4971 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/6e6307eb-6448-4cbb-b2a3-93e340f5b3a4-kube-api-access\") pod \"kube-controller-manager-operator-78b949d7b-cwwf9\" (UID: \"6e6307eb-6448-4cbb-b2a3-93e340f5b3a4\") " pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-cwwf9" Dec 13 06:51:36 crc kubenswrapper[4971]: I1213 06:51:36.015535 4971 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/c0ca0ce3-929c-48d6-a3f7-6d334e3c3709-config\") pod \"authentication-operator-69f744f599-wjdd5\" (UID: \"c0ca0ce3-929c-48d6-a3f7-6d334e3c3709\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-wjdd5" Dec 13 06:51:36 crc kubenswrapper[4971]: I1213 06:51:36.015560 4971 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/4b58bd40-cb76-46ba-9391-a350d830d729-serving-cert\") pod \"openshift-kube-scheduler-operator-5fdd9b5758-fz6wk\" (UID: \"4b58bd40-cb76-46ba-9391-a350d830d729\") " pod="openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-fz6wk" Dec 13 06:51:36 crc kubenswrapper[4971]: I1213 06:51:36.015581 4971 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/c1575ff9-1c53-4732-bc48-252f9f270597-secret-volume\") pod \"collect-profiles-29426805-fxxrz\" (UID: \"c1575ff9-1c53-4732-bc48-252f9f270597\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29426805-fxxrz" Dec 13 06:51:36 crc kubenswrapper[4971]: I1213 06:51:36.015600 4971 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"node-bootstrap-token\" (UniqueName: \"kubernetes.io/secret/1c4c5572-81b1-4270-b519-b077f9682f49-node-bootstrap-token\") pod \"machine-config-server-xdxjz\" (UID: \"1c4c5572-81b1-4270-b519-b077f9682f49\") " pod="openshift-machine-config-operator/machine-config-server-xdxjz" Dec 13 06:51:36 crc kubenswrapper[4971]: I1213 06:51:36.015625 4971 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-c5lc5\" (UniqueName: \"kubernetes.io/projected/6a3649bc-2d2e-4897-8977-99c08aa77e01-kube-api-access-c5lc5\") pod \"kube-storage-version-migrator-operator-b67b599dd-6blxl\" (UID: \"6a3649bc-2d2e-4897-8977-99c08aa77e01\") " pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-6blxl" Dec 13 06:51:36 crc kubenswrapper[4971]: I1213 06:51:36.015652 4971 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"socket-dir\" (UniqueName: \"kubernetes.io/host-path/5fd61507-84bf-4ef5-8bd4-731075ef58ef-socket-dir\") pod \"csi-hostpathplugin-7nhhk\" (UID: \"5fd61507-84bf-4ef5-8bd4-731075ef58ef\") " pod="hostpath-provisioner/csi-hostpathplugin-7nhhk" Dec 13 06:51:36 crc kubenswrapper[4971]: I1213 06:51:36.015674 4971 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-wnzs2\" (UniqueName: \"kubernetes.io/projected/5fd61507-84bf-4ef5-8bd4-731075ef58ef-kube-api-access-wnzs2\") pod \"csi-hostpathplugin-7nhhk\" (UID: \"5fd61507-84bf-4ef5-8bd4-731075ef58ef\") " pod="hostpath-provisioner/csi-hostpathplugin-7nhhk" Dec 13 06:51:36 crc kubenswrapper[4971]: I1213 06:51:36.015698 4971 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/4b58bd40-cb76-46ba-9391-a350d830d729-config\") pod \"openshift-kube-scheduler-operator-5fdd9b5758-fz6wk\" (UID: \"4b58bd40-cb76-46ba-9391-a350d830d729\") " pod="openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-fz6wk" Dec 13 06:51:36 crc kubenswrapper[4971]: I1213 06:51:36.015718 4971 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/6e6307eb-6448-4cbb-b2a3-93e340f5b3a4-config\") pod \"kube-controller-manager-operator-78b949d7b-cwwf9\" (UID: \"6e6307eb-6448-4cbb-b2a3-93e340f5b3a4\") " pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-cwwf9" Dec 13 06:51:36 crc kubenswrapper[4971]: I1213 06:51:36.015739 4971 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"srv-cert\" (UniqueName: \"kubernetes.io/secret/3052c2e3-f9bf-442b-b153-d60eb86f9c3a-srv-cert\") pod \"catalog-operator-68c6474976-l4rzp\" (UID: \"3052c2e3-f9bf-442b-b153-d60eb86f9c3a\") " pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-l4rzp" Dec 13 06:51:36 crc kubenswrapper[4971]: I1213 06:51:36.015760 4971 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/49f3dc41-5142-4cdb-9bb3-6d7248228535-config\") pod \"service-ca-operator-777779d784-drq5g\" (UID: \"49f3dc41-5142-4cdb-9bb3-6d7248228535\") " pod="openshift-service-ca-operator/service-ca-operator-777779d784-drq5g" Dec 13 06:51:36 crc kubenswrapper[4971]: I1213 06:51:36.015779 4971 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/6a3649bc-2d2e-4897-8977-99c08aa77e01-serving-cert\") pod \"kube-storage-version-migrator-operator-b67b599dd-6blxl\" (UID: \"6a3649bc-2d2e-4897-8977-99c08aa77e01\") " pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-6blxl" Dec 13 06:51:36 crc kubenswrapper[4971]: I1213 06:51:36.015801 4971 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/c1575ff9-1c53-4732-bc48-252f9f270597-config-volume\") pod \"collect-profiles-29426805-fxxrz\" (UID: \"c1575ff9-1c53-4732-bc48-252f9f270597\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29426805-fxxrz" Dec 13 06:51:36 crc kubenswrapper[4971]: I1213 06:51:36.015824 4971 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-fdpnd\" (UniqueName: \"kubernetes.io/projected/c0ca0ce3-929c-48d6-a3f7-6d334e3c3709-kube-api-access-fdpnd\") pod \"authentication-operator-69f744f599-wjdd5\" (UID: \"c0ca0ce3-929c-48d6-a3f7-6d334e3c3709\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-wjdd5" Dec 13 06:51:36 crc kubenswrapper[4971]: I1213 06:51:36.015849 4971 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/6e6307eb-6448-4cbb-b2a3-93e340f5b3a4-serving-cert\") pod \"kube-controller-manager-operator-78b949d7b-cwwf9\" (UID: \"6e6307eb-6448-4cbb-b2a3-93e340f5b3a4\") " pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-cwwf9" Dec 13 06:51:36 crc kubenswrapper[4971]: I1213 06:51:36.015878 4971 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"tmpfs\" (UniqueName: \"kubernetes.io/empty-dir/87f8bc63-d8c1-415f-975f-b8418034197b-tmpfs\") pod \"packageserver-d55dfcdfc-lnrkm\" (UID: \"87f8bc63-d8c1-415f-975f-b8418034197b\") " pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-lnrkm" Dec 13 06:51:36 crc kubenswrapper[4971]: I1213 06:51:36.015899 4971 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/4b58bd40-cb76-46ba-9391-a350d830d729-kube-api-access\") pod \"openshift-kube-scheduler-operator-5fdd9b5758-fz6wk\" (UID: \"4b58bd40-cb76-46ba-9391-a350d830d729\") " pod="openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-fz6wk" Dec 13 06:51:36 crc kubenswrapper[4971]: I1213 06:51:36.015925 4971 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"srv-cert\" (UniqueName: \"kubernetes.io/secret/ea24f72e-df9d-418f-aaef-dfc0573eface-srv-cert\") pod \"olm-operator-6b444d44fb-pxjqw\" (UID: \"ea24f72e-df9d-418f-aaef-dfc0573eface\") " pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-pxjqw" Dec 13 06:51:36 crc kubenswrapper[4971]: I1213 06:51:36.015944 4971 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/6a3649bc-2d2e-4897-8977-99c08aa77e01-config\") pod \"kube-storage-version-migrator-operator-b67b599dd-6blxl\" (UID: \"6a3649bc-2d2e-4897-8977-99c08aa77e01\") " pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-6blxl" Dec 13 06:51:36 crc kubenswrapper[4971]: I1213 06:51:36.015964 4971 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/87f8bc63-d8c1-415f-975f-b8418034197b-apiservice-cert\") pod \"packageserver-d55dfcdfc-lnrkm\" (UID: \"87f8bc63-d8c1-415f-975f-b8418034197b\") " pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-lnrkm" Dec 13 06:51:36 crc kubenswrapper[4971]: I1213 06:51:36.015997 4971 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-mdd8q\" (UID: \"cba893c6-ff99-466a-9db0-a62e7a87a334\") " pod="openshift-image-registry/image-registry-697d97f7c8-mdd8q" Dec 13 06:51:36 crc kubenswrapper[4971]: I1213 06:51:36.016020 4971 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"mountpoint-dir\" (UniqueName: \"kubernetes.io/host-path/5fd61507-84bf-4ef5-8bd4-731075ef58ef-mountpoint-dir\") pod \"csi-hostpathplugin-7nhhk\" (UID: \"5fd61507-84bf-4ef5-8bd4-731075ef58ef\") " pod="hostpath-provisioner/csi-hostpathplugin-7nhhk" Dec 13 06:51:36 crc kubenswrapper[4971]: I1213 06:51:36.016041 4971 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"certs\" (UniqueName: \"kubernetes.io/secret/1c4c5572-81b1-4270-b519-b077f9682f49-certs\") pod \"machine-config-server-xdxjz\" (UID: \"1c4c5572-81b1-4270-b519-b077f9682f49\") " pod="openshift-machine-config-operator/machine-config-server-xdxjz" Dec 13 06:51:36 crc kubenswrapper[4971]: I1213 06:51:36.016069 4971 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/4d847949-530c-4cbc-b11d-c1c8d80d3db7-cert\") pod \"ingress-canary-ffhlq\" (UID: \"4d847949-530c-4cbc-b11d-c1c8d80d3db7\") " pod="openshift-ingress-canary/ingress-canary-ffhlq" Dec 13 06:51:36 crc kubenswrapper[4971]: I1213 06:51:36.016092 4971 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-2m57w\" (UniqueName: \"kubernetes.io/projected/637e887c-db01-4b47-9f17-2d61b731a118-kube-api-access-2m57w\") pod \"dns-default-dntmz\" (UID: \"637e887c-db01-4b47-9f17-2d61b731a118\") " pod="openshift-dns/dns-default-dntmz" Dec 13 06:51:36 crc kubenswrapper[4971]: I1213 06:51:36.016116 4971 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"registration-dir\" (UniqueName: \"kubernetes.io/host-path/5fd61507-84bf-4ef5-8bd4-731075ef58ef-registration-dir\") pod \"csi-hostpathplugin-7nhhk\" (UID: \"5fd61507-84bf-4ef5-8bd4-731075ef58ef\") " pod="hostpath-provisioner/csi-hostpathplugin-7nhhk" Dec 13 06:51:36 crc kubenswrapper[4971]: I1213 06:51:36.016139 4971 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"plugins-dir\" (UniqueName: \"kubernetes.io/host-path/5fd61507-84bf-4ef5-8bd4-731075ef58ef-plugins-dir\") pod \"csi-hostpathplugin-7nhhk\" (UID: \"5fd61507-84bf-4ef5-8bd4-731075ef58ef\") " pod="hostpath-provisioner/csi-hostpathplugin-7nhhk" Dec 13 06:51:36 crc kubenswrapper[4971]: I1213 06:51:36.016172 4971 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"profile-collector-cert\" (UniqueName: \"kubernetes.io/secret/3052c2e3-f9bf-442b-b153-d60eb86f9c3a-profile-collector-cert\") pod \"catalog-operator-68c6474976-l4rzp\" (UID: \"3052c2e3-f9bf-442b-b153-d60eb86f9c3a\") " pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-l4rzp" Dec 13 06:51:36 crc kubenswrapper[4971]: I1213 06:51:36.016193 4971 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"csi-data-dir\" (UniqueName: \"kubernetes.io/host-path/5fd61507-84bf-4ef5-8bd4-731075ef58ef-csi-data-dir\") pod \"csi-hostpathplugin-7nhhk\" (UID: \"5fd61507-84bf-4ef5-8bd4-731075ef58ef\") " pod="hostpath-provisioner/csi-hostpathplugin-7nhhk" Dec 13 06:51:36 crc kubenswrapper[4971]: I1213 06:51:36.016245 4971 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/637e887c-db01-4b47-9f17-2d61b731a118-metrics-tls\") pod \"dns-default-dntmz\" (UID: \"637e887c-db01-4b47-9f17-2d61b731a118\") " pod="openshift-dns/dns-default-dntmz" Dec 13 06:51:36 crc kubenswrapper[4971]: I1213 06:51:36.016276 4971 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-njt75\" (UniqueName: \"kubernetes.io/projected/5764a8c4-18e5-4025-8d36-15a7b01e9169-kube-api-access-njt75\") pod \"migrator-59844c95c7-z2p8z\" (UID: \"5764a8c4-18e5-4025-8d36-15a7b01e9169\") " pod="openshift-kube-storage-version-migrator/migrator-59844c95c7-z2p8z" Dec 13 06:51:36 crc kubenswrapper[4971]: I1213 06:51:36.016301 4971 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-tlp42\" (UniqueName: \"kubernetes.io/projected/49f3dc41-5142-4cdb-9bb3-6d7248228535-kube-api-access-tlp42\") pod \"service-ca-operator-777779d784-drq5g\" (UID: \"49f3dc41-5142-4cdb-9bb3-6d7248228535\") " pod="openshift-service-ca-operator/service-ca-operator-777779d784-drq5g" Dec 13 06:51:36 crc kubenswrapper[4971]: I1213 06:51:36.016325 4971 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-8zjpk\" (UniqueName: \"kubernetes.io/projected/ea24f72e-df9d-418f-aaef-dfc0573eface-kube-api-access-8zjpk\") pod \"olm-operator-6b444d44fb-pxjqw\" (UID: \"ea24f72e-df9d-418f-aaef-dfc0573eface\") " pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-pxjqw" Dec 13 06:51:36 crc kubenswrapper[4971]: I1213 06:51:36.016346 4971 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-qdm7j\" (UniqueName: \"kubernetes.io/projected/c1575ff9-1c53-4732-bc48-252f9f270597-kube-api-access-qdm7j\") pod \"collect-profiles-29426805-fxxrz\" (UID: \"c1575ff9-1c53-4732-bc48-252f9f270597\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29426805-fxxrz" Dec 13 06:51:36 crc kubenswrapper[4971]: I1213 06:51:36.016374 4971 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"service-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/c0ca0ce3-929c-48d6-a3f7-6d334e3c3709-service-ca-bundle\") pod \"authentication-operator-69f744f599-wjdd5\" (UID: \"c0ca0ce3-929c-48d6-a3f7-6d334e3c3709\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-wjdd5" Dec 13 06:51:36 crc kubenswrapper[4971]: I1213 06:51:36.017116 4971 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"service-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/c0ca0ce3-929c-48d6-a3f7-6d334e3c3709-service-ca-bundle\") pod \"authentication-operator-69f744f599-wjdd5\" (UID: \"c0ca0ce3-929c-48d6-a3f7-6d334e3c3709\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-wjdd5" Dec 13 06:51:36 crc kubenswrapper[4971]: I1213 06:51:36.019059 4971 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/c0ca0ce3-929c-48d6-a3f7-6d334e3c3709-trusted-ca-bundle\") pod \"authentication-operator-69f744f599-wjdd5\" (UID: \"c0ca0ce3-929c-48d6-a3f7-6d334e3c3709\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-wjdd5" Dec 13 06:51:36 crc kubenswrapper[4971]: I1213 06:51:36.019574 4971 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"csi-data-dir\" (UniqueName: \"kubernetes.io/host-path/5fd61507-84bf-4ef5-8bd4-731075ef58ef-csi-data-dir\") pod \"csi-hostpathplugin-7nhhk\" (UID: \"5fd61507-84bf-4ef5-8bd4-731075ef58ef\") " pod="hostpath-provisioner/csi-hostpathplugin-7nhhk" Dec 13 06:51:36 crc kubenswrapper[4971]: E1213 06:51:36.020125 4971 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-12-13 06:51:36.52011134 +0000 UTC m=+153.124520788 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-mdd8q" (UID: "cba893c6-ff99-466a-9db0-a62e7a87a334") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 13 06:51:36 crc kubenswrapper[4971]: I1213 06:51:36.020695 4971 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/c1575ff9-1c53-4732-bc48-252f9f270597-config-volume\") pod \"collect-profiles-29426805-fxxrz\" (UID: \"c1575ff9-1c53-4732-bc48-252f9f270597\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29426805-fxxrz" Dec 13 06:51:36 crc kubenswrapper[4971]: I1213 06:51:36.021003 4971 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"tmpfs\" (UniqueName: \"kubernetes.io/empty-dir/87f8bc63-d8c1-415f-975f-b8418034197b-tmpfs\") pod \"packageserver-d55dfcdfc-lnrkm\" (UID: \"87f8bc63-d8c1-415f-975f-b8418034197b\") " pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-lnrkm" Dec 13 06:51:36 crc kubenswrapper[4971]: I1213 06:51:36.021156 4971 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"mountpoint-dir\" (UniqueName: \"kubernetes.io/host-path/5fd61507-84bf-4ef5-8bd4-731075ef58ef-mountpoint-dir\") pod \"csi-hostpathplugin-7nhhk\" (UID: \"5fd61507-84bf-4ef5-8bd4-731075ef58ef\") " pod="hostpath-provisioner/csi-hostpathplugin-7nhhk" Dec 13 06:51:36 crc kubenswrapper[4971]: I1213 06:51:36.021602 4971 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"socket-dir\" (UniqueName: \"kubernetes.io/host-path/5fd61507-84bf-4ef5-8bd4-731075ef58ef-socket-dir\") pod \"csi-hostpathplugin-7nhhk\" (UID: \"5fd61507-84bf-4ef5-8bd4-731075ef58ef\") " pod="hostpath-provisioner/csi-hostpathplugin-7nhhk" Dec 13 06:51:36 crc kubenswrapper[4971]: I1213 06:51:36.021614 4971 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/c0ca0ce3-929c-48d6-a3f7-6d334e3c3709-config\") pod \"authentication-operator-69f744f599-wjdd5\" (UID: \"c0ca0ce3-929c-48d6-a3f7-6d334e3c3709\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-wjdd5" Dec 13 06:51:36 crc kubenswrapper[4971]: I1213 06:51:36.021674 4971 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/6a3649bc-2d2e-4897-8977-99c08aa77e01-config\") pod \"kube-storage-version-migrator-operator-b67b599dd-6blxl\" (UID: \"6a3649bc-2d2e-4897-8977-99c08aa77e01\") " pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-6blxl" Dec 13 06:51:36 crc kubenswrapper[4971]: I1213 06:51:36.021874 4971 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"registration-dir\" (UniqueName: \"kubernetes.io/host-path/5fd61507-84bf-4ef5-8bd4-731075ef58ef-registration-dir\") pod \"csi-hostpathplugin-7nhhk\" (UID: \"5fd61507-84bf-4ef5-8bd4-731075ef58ef\") " pod="hostpath-provisioner/csi-hostpathplugin-7nhhk" Dec 13 06:51:36 crc kubenswrapper[4971]: I1213 06:51:36.022010 4971 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"plugins-dir\" (UniqueName: \"kubernetes.io/host-path/5fd61507-84bf-4ef5-8bd4-731075ef58ef-plugins-dir\") pod \"csi-hostpathplugin-7nhhk\" (UID: \"5fd61507-84bf-4ef5-8bd4-731075ef58ef\") " pod="hostpath-provisioner/csi-hostpathplugin-7nhhk" Dec 13 06:51:36 crc kubenswrapper[4971]: I1213 06:51:36.022685 4971 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/6e6307eb-6448-4cbb-b2a3-93e340f5b3a4-config\") pod \"kube-controller-manager-operator-78b949d7b-cwwf9\" (UID: \"6e6307eb-6448-4cbb-b2a3-93e340f5b3a4\") " pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-cwwf9" Dec 13 06:51:36 crc kubenswrapper[4971]: I1213 06:51:36.024231 4971 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/4b58bd40-cb76-46ba-9391-a350d830d729-config\") pod \"openshift-kube-scheduler-operator-5fdd9b5758-fz6wk\" (UID: \"4b58bd40-cb76-46ba-9391-a350d830d729\") " pod="openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-fz6wk" Dec 13 06:51:36 crc kubenswrapper[4971]: I1213 06:51:36.024782 4971 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/637e887c-db01-4b47-9f17-2d61b731a118-config-volume\") pod \"dns-default-dntmz\" (UID: \"637e887c-db01-4b47-9f17-2d61b731a118\") " pod="openshift-dns/dns-default-dntmz" Dec 13 06:51:36 crc kubenswrapper[4971]: I1213 06:51:36.025295 4971 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/6e6307eb-6448-4cbb-b2a3-93e340f5b3a4-serving-cert\") pod \"kube-controller-manager-operator-78b949d7b-cwwf9\" (UID: \"6e6307eb-6448-4cbb-b2a3-93e340f5b3a4\") " pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-cwwf9" Dec 13 06:51:36 crc kubenswrapper[4971]: I1213 06:51:36.027681 4971 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/49f3dc41-5142-4cdb-9bb3-6d7248228535-config\") pod \"service-ca-operator-777779d784-drq5g\" (UID: \"49f3dc41-5142-4cdb-9bb3-6d7248228535\") " pod="openshift-service-ca-operator/service-ca-operator-777779d784-drq5g" Dec 13 06:51:36 crc kubenswrapper[4971]: I1213 06:51:36.028009 4971 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/637e887c-db01-4b47-9f17-2d61b731a118-metrics-tls\") pod \"dns-default-dntmz\" (UID: \"637e887c-db01-4b47-9f17-2d61b731a118\") " pod="openshift-dns/dns-default-dntmz" Dec 13 06:51:36 crc kubenswrapper[4971]: I1213 06:51:36.029058 4971 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"profile-collector-cert\" (UniqueName: \"kubernetes.io/secret/3052c2e3-f9bf-442b-b153-d60eb86f9c3a-profile-collector-cert\") pod \"catalog-operator-68c6474976-l4rzp\" (UID: \"3052c2e3-f9bf-442b-b153-d60eb86f9c3a\") " pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-l4rzp" Dec 13 06:51:36 crc kubenswrapper[4971]: I1213 06:51:36.029655 4971 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"node-bootstrap-token\" (UniqueName: \"kubernetes.io/secret/1c4c5572-81b1-4270-b519-b077f9682f49-node-bootstrap-token\") pod \"machine-config-server-xdxjz\" (UID: \"1c4c5572-81b1-4270-b519-b077f9682f49\") " pod="openshift-machine-config-operator/machine-config-server-xdxjz" Dec 13 06:51:36 crc kubenswrapper[4971]: I1213 06:51:36.030827 4971 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/6a3649bc-2d2e-4897-8977-99c08aa77e01-serving-cert\") pod \"kube-storage-version-migrator-operator-b67b599dd-6blxl\" (UID: \"6a3649bc-2d2e-4897-8977-99c08aa77e01\") " pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-6blxl" Dec 13 06:51:36 crc kubenswrapper[4971]: I1213 06:51:36.031173 4971 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cert\" (UniqueName: \"kubernetes.io/secret/4d847949-530c-4cbc-b11d-c1c8d80d3db7-cert\") pod \"ingress-canary-ffhlq\" (UID: \"4d847949-530c-4cbc-b11d-c1c8d80d3db7\") " pod="openshift-ingress-canary/ingress-canary-ffhlq" Dec 13 06:51:36 crc kubenswrapper[4971]: I1213 06:51:36.031207 4971 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/4b58bd40-cb76-46ba-9391-a350d830d729-serving-cert\") pod \"openshift-kube-scheduler-operator-5fdd9b5758-fz6wk\" (UID: \"4b58bd40-cb76-46ba-9391-a350d830d729\") " pod="openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-fz6wk" Dec 13 06:51:36 crc kubenswrapper[4971]: I1213 06:51:36.031604 4971 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"profile-collector-cert\" (UniqueName: \"kubernetes.io/secret/ea24f72e-df9d-418f-aaef-dfc0573eface-profile-collector-cert\") pod \"olm-operator-6b444d44fb-pxjqw\" (UID: \"ea24f72e-df9d-418f-aaef-dfc0573eface\") " pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-pxjqw" Dec 13 06:51:36 crc kubenswrapper[4971]: I1213 06:51:36.031912 4971 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"srv-cert\" (UniqueName: \"kubernetes.io/secret/ea24f72e-df9d-418f-aaef-dfc0573eface-srv-cert\") pod \"olm-operator-6b444d44fb-pxjqw\" (UID: \"ea24f72e-df9d-418f-aaef-dfc0573eface\") " pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-pxjqw" Dec 13 06:51:36 crc kubenswrapper[4971]: I1213 06:51:36.032018 4971 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-mkgwg\" (UniqueName: \"kubernetes.io/projected/2cf6e363-8511-4c79-9181-79dba726e5af-kube-api-access-mkgwg\") pod \"route-controller-manager-6576b87f9c-9d86n\" (UID: \"2cf6e363-8511-4c79-9181-79dba726e5af\") " pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-9d86n" Dec 13 06:51:36 crc kubenswrapper[4971]: I1213 06:51:36.032396 4971 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"certs\" (UniqueName: \"kubernetes.io/secret/1c4c5572-81b1-4270-b519-b077f9682f49-certs\") pod \"machine-config-server-xdxjz\" (UID: \"1c4c5572-81b1-4270-b519-b077f9682f49\") " pod="openshift-machine-config-operator/machine-config-server-xdxjz" Dec 13 06:51:36 crc kubenswrapper[4971]: I1213 06:51:36.032718 4971 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/49f3dc41-5142-4cdb-9bb3-6d7248228535-serving-cert\") pod \"service-ca-operator-777779d784-drq5g\" (UID: \"49f3dc41-5142-4cdb-9bb3-6d7248228535\") " pod="openshift-service-ca-operator/service-ca-operator-777779d784-drq5g" Dec 13 06:51:36 crc kubenswrapper[4971]: I1213 06:51:36.034140 4971 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"srv-cert\" (UniqueName: \"kubernetes.io/secret/3052c2e3-f9bf-442b-b153-d60eb86f9c3a-srv-cert\") pod \"catalog-operator-68c6474976-l4rzp\" (UID: \"3052c2e3-f9bf-442b-b153-d60eb86f9c3a\") " pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-l4rzp" Dec 13 06:51:36 crc kubenswrapper[4971]: I1213 06:51:36.037224 4971 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/87f8bc63-d8c1-415f-975f-b8418034197b-apiservice-cert\") pod \"packageserver-d55dfcdfc-lnrkm\" (UID: \"87f8bc63-d8c1-415f-975f-b8418034197b\") " pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-lnrkm" Dec 13 06:51:36 crc kubenswrapper[4971]: I1213 06:51:36.038253 4971 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/87f8bc63-d8c1-415f-975f-b8418034197b-webhook-cert\") pod \"packageserver-d55dfcdfc-lnrkm\" (UID: \"87f8bc63-d8c1-415f-975f-b8418034197b\") " pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-lnrkm" Dec 13 06:51:36 crc kubenswrapper[4971]: I1213 06:51:36.038780 4971 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/c1575ff9-1c53-4732-bc48-252f9f270597-secret-volume\") pod \"collect-profiles-29426805-fxxrz\" (UID: \"c1575ff9-1c53-4732-bc48-252f9f270597\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29426805-fxxrz" Dec 13 06:51:36 crc kubenswrapper[4971]: I1213 06:51:36.039886 4971 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/c0ca0ce3-929c-48d6-a3f7-6d334e3c3709-serving-cert\") pod \"authentication-operator-69f744f599-wjdd5\" (UID: \"c0ca0ce3-929c-48d6-a3f7-6d334e3c3709\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-wjdd5" Dec 13 06:51:36 crc kubenswrapper[4971]: I1213 06:51:36.044656 4971 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-fzfk6\" (UniqueName: \"kubernetes.io/projected/d89e3583-3f66-4259-b6f9-bffe8ddbde12-kube-api-access-fzfk6\") pod \"router-default-5444994796-mzhp6\" (UID: \"d89e3583-3f66-4259-b6f9-bffe8ddbde12\") " pod="openshift-ingress/router-default-5444994796-mzhp6" Dec 13 06:51:36 crc kubenswrapper[4971]: I1213 06:51:36.047681 4971 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-xfvfg" Dec 13 06:51:36 crc kubenswrapper[4971]: I1213 06:51:36.056503 4971 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-ingress/router-default-5444994796-mzhp6" Dec 13 06:51:36 crc kubenswrapper[4971]: I1213 06:51:36.073882 4971 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-zqt2t"] Dec 13 06:51:36 crc kubenswrapper[4971]: I1213 06:51:36.079983 4971 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-controller-manager/controller-manager-879f6c89f-xtxr8"] Dec 13 06:51:36 crc kubenswrapper[4971]: I1213 06:51:36.080835 4971 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-multus/multus-admission-controller-857f4d67dd-z68ks"] Dec 13 06:51:36 crc kubenswrapper[4971]: I1213 06:51:36.081618 4971 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-v95qb\" (UniqueName: \"kubernetes.io/projected/479b25e6-6618-4123-a913-ca50d6c4ba28-kube-api-access-v95qb\") pod \"package-server-manager-789f6589d5-g8vfp\" (UID: \"479b25e6-6618-4123-a913-ca50d6c4ba28\") " pod="openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-g8vfp" Dec 13 06:51:36 crc kubenswrapper[4971]: I1213 06:51:36.102295 4971 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-vxdq2\" (UniqueName: \"kubernetes.io/projected/22f3a4f0-1686-4087-bcec-e48f2015ffe2-kube-api-access-vxdq2\") pod \"ingress-operator-5b745b69d9-pkmb5\" (UID: \"22f3a4f0-1686-4087-bcec-e48f2015ffe2\") " pod="openshift-ingress-operator/ingress-operator-5b745b69d9-pkmb5" Dec 13 06:51:36 crc kubenswrapper[4971]: I1213 06:51:36.112731 4971 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/cba893c6-ff99-466a-9db0-a62e7a87a334-bound-sa-token\") pod \"image-registry-697d97f7c8-mdd8q\" (UID: \"cba893c6-ff99-466a-9db0-a62e7a87a334\") " pod="openshift-image-registry/image-registry-697d97f7c8-mdd8q" Dec 13 06:51:36 crc kubenswrapper[4971]: I1213 06:51:36.118057 4971 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 13 06:51:36 crc kubenswrapper[4971]: E1213 06:51:36.118709 4971 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-13 06:51:36.618689975 +0000 UTC m=+153.223099423 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 13 06:51:36 crc kubenswrapper[4971]: I1213 06:51:36.154711 4971 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-64btv\" (UniqueName: \"kubernetes.io/projected/930fe38c-5ff7-45a8-a292-cd69d07e86f7-kube-api-access-64btv\") pod \"control-plane-machine-set-operator-78cbb6b69f-bgdbk\" (UID: \"930fe38c-5ff7-45a8-a292-cd69d07e86f7\") " pod="openshift-machine-api/control-plane-machine-set-operator-78cbb6b69f-bgdbk" Dec 13 06:51:36 crc kubenswrapper[4971]: I1213 06:51:36.173749 4971 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/marketplace-operator-79b997595-j9zqf"] Dec 13 06:51:36 crc kubenswrapper[4971]: I1213 06:51:36.174510 4971 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-2gk4l\" (UniqueName: \"kubernetes.io/projected/94469d04-47fd-4bd8-8258-9e874767f250-kube-api-access-2gk4l\") pod \"console-operator-58897d9998-kdfcd\" (UID: \"94469d04-47fd-4bd8-8258-9e874767f250\") " pod="openshift-console-operator/console-operator-58897d9998-kdfcd" Dec 13 06:51:36 crc kubenswrapper[4971]: I1213 06:51:36.175134 4971 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-9d86n" Dec 13 06:51:36 crc kubenswrapper[4971]: I1213 06:51:36.190832 4971 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-t2s9j\" (UniqueName: \"kubernetes.io/projected/ad158968-e110-4b4c-9f42-da9c08274dab-kube-api-access-t2s9j\") pod \"dns-operator-744455d44c-b999k\" (UID: \"ad158968-e110-4b4c-9f42-da9c08274dab\") " pod="openshift-dns-operator/dns-operator-744455d44c-b999k" Dec 13 06:51:36 crc kubenswrapper[4971]: W1213 06:51:36.194967 4971 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podd89e3583_3f66_4259_b6f9_bffe8ddbde12.slice/crio-46f0e746181c8f41e7f9c597d795e818dad8cb15ef34b2c6740076ce085432eb WatchSource:0}: Error finding container 46f0e746181c8f41e7f9c597d795e818dad8cb15ef34b2c6740076ce085432eb: Status 404 returned error can't find the container with id 46f0e746181c8f41e7f9c597d795e818dad8cb15ef34b2c6740076ce085432eb Dec 13 06:51:36 crc kubenswrapper[4971]: I1213 06:51:36.199694 4971 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-service-ca/service-ca-9c57cc56f-q99rh"] Dec 13 06:51:36 crc kubenswrapper[4971]: I1213 06:51:36.200975 4971 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-lltlz\" (UniqueName: \"kubernetes.io/projected/874a112c-101d-4a74-a674-d9afe62b6dc8-kube-api-access-lltlz\") pod \"machine-config-operator-74547568cd-4cn2z\" (UID: \"874a112c-101d-4a74-a674-d9afe62b6dc8\") " pod="openshift-machine-config-operator/machine-config-operator-74547568cd-4cn2z" Dec 13 06:51:36 crc kubenswrapper[4971]: I1213 06:51:36.213794 4971 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-jk6gm\" (UniqueName: \"kubernetes.io/projected/ee2e94f4-cadb-4b17-8853-a1292cb02404-kube-api-access-jk6gm\") pod \"downloads-7954f5f757-dqk9z\" (UID: \"ee2e94f4-cadb-4b17-8853-a1292cb02404\") " pod="openshift-console/downloads-7954f5f757-dqk9z" Dec 13 06:51:36 crc kubenswrapper[4971]: I1213 06:51:36.222893 4971 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-mdd8q\" (UID: \"cba893c6-ff99-466a-9db0-a62e7a87a334\") " pod="openshift-image-registry/image-registry-697d97f7c8-mdd8q" Dec 13 06:51:36 crc kubenswrapper[4971]: I1213 06:51:36.222963 4971 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-console/downloads-7954f5f757-dqk9z" Dec 13 06:51:36 crc kubenswrapper[4971]: E1213 06:51:36.223449 4971 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-12-13 06:51:36.723431614 +0000 UTC m=+153.327841072 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-mdd8q" (UID: "cba893c6-ff99-466a-9db0-a62e7a87a334") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 13 06:51:36 crc kubenswrapper[4971]: I1213 06:51:36.231855 4971 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-console-operator/console-operator-58897d9998-kdfcd" Dec 13 06:51:36 crc kubenswrapper[4971]: I1213 06:51:36.246737 4971 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/4b58bd40-cb76-46ba-9391-a350d830d729-kube-api-access\") pod \"openshift-kube-scheduler-operator-5fdd9b5758-fz6wk\" (UID: \"4b58bd40-cb76-46ba-9391-a350d830d729\") " pod="openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-fz6wk" Dec 13 06:51:36 crc kubenswrapper[4971]: I1213 06:51:36.249330 4971 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-ingress-operator/ingress-operator-5b745b69d9-pkmb5" Dec 13 06:51:36 crc kubenswrapper[4971]: I1213 06:51:36.251408 4971 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-gtrwg"] Dec 13 06:51:36 crc kubenswrapper[4971]: I1213 06:51:36.255732 4971 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-etcd-operator/etcd-operator-b45778765-6xfwq"] Dec 13 06:51:36 crc kubenswrapper[4971]: I1213 06:51:36.255953 4971 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-dns-operator/dns-operator-744455d44c-b999k" Dec 13 06:51:36 crc kubenswrapper[4971]: I1213 06:51:36.272203 4971 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-tlp42\" (UniqueName: \"kubernetes.io/projected/49f3dc41-5142-4cdb-9bb3-6d7248228535-kube-api-access-tlp42\") pod \"service-ca-operator-777779d784-drq5g\" (UID: \"49f3dc41-5142-4cdb-9bb3-6d7248228535\") " pod="openshift-service-ca-operator/service-ca-operator-777779d784-drq5g" Dec 13 06:51:36 crc kubenswrapper[4971]: I1213 06:51:36.279796 4971 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-machine-api/control-plane-machine-set-operator-78cbb6b69f-bgdbk" Dec 13 06:51:36 crc kubenswrapper[4971]: I1213 06:51:36.290292 4971 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-njt75\" (UniqueName: \"kubernetes.io/projected/5764a8c4-18e5-4025-8d36-15a7b01e9169-kube-api-access-njt75\") pod \"migrator-59844c95c7-z2p8z\" (UID: \"5764a8c4-18e5-4025-8d36-15a7b01e9169\") " pod="openshift-kube-storage-version-migrator/migrator-59844c95c7-z2p8z" Dec 13 06:51:36 crc kubenswrapper[4971]: I1213 06:51:36.298369 4971 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-8zjpk\" (UniqueName: \"kubernetes.io/projected/ea24f72e-df9d-418f-aaef-dfc0573eface-kube-api-access-8zjpk\") pod \"olm-operator-6b444d44fb-pxjqw\" (UID: \"ea24f72e-df9d-418f-aaef-dfc0573eface\") " pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-pxjqw" Dec 13 06:51:36 crc kubenswrapper[4971]: W1213 06:51:36.301967 4971 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod56c62b80_0a33_4156_b52b_cb54c98a4343.slice/crio-7c8cacdc2911ff16c349a62962ad34eb9b6eedf77eb20dbcf6a6cd3da46ae058 WatchSource:0}: Error finding container 7c8cacdc2911ff16c349a62962ad34eb9b6eedf77eb20dbcf6a6cd3da46ae058: Status 404 returned error can't find the container with id 7c8cacdc2911ff16c349a62962ad34eb9b6eedf77eb20dbcf6a6cd3da46ae058 Dec 13 06:51:36 crc kubenswrapper[4971]: I1213 06:51:36.314936 4971 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-g8vfp" Dec 13 06:51:36 crc kubenswrapper[4971]: I1213 06:51:36.324593 4971 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 13 06:51:36 crc kubenswrapper[4971]: E1213 06:51:36.325222 4971 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-13 06:51:36.82520012 +0000 UTC m=+153.429609578 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 13 06:51:36 crc kubenswrapper[4971]: I1213 06:51:36.326243 4971 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-qdm7j\" (UniqueName: \"kubernetes.io/projected/c1575ff9-1c53-4732-bc48-252f9f270597-kube-api-access-qdm7j\") pod \"collect-profiles-29426805-fxxrz\" (UID: \"c1575ff9-1c53-4732-bc48-252f9f270597\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29426805-fxxrz" Dec 13 06:51:36 crc kubenswrapper[4971]: I1213 06:51:36.327172 4971 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-mdd8q\" (UID: \"cba893c6-ff99-466a-9db0-a62e7a87a334\") " pod="openshift-image-registry/image-registry-697d97f7c8-mdd8q" Dec 13 06:51:36 crc kubenswrapper[4971]: E1213 06:51:36.327816 4971 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-12-13 06:51:36.827803706 +0000 UTC m=+153.432213154 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-mdd8q" (UID: "cba893c6-ff99-466a-9db0-a62e7a87a334") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 13 06:51:36 crc kubenswrapper[4971]: I1213 06:51:36.331326 4971 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-machine-config-operator/machine-config-operator-74547568cd-4cn2z" Dec 13 06:51:36 crc kubenswrapper[4971]: I1213 06:51:36.340419 4971 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-fdpnd\" (UniqueName: \"kubernetes.io/projected/c0ca0ce3-929c-48d6-a3f7-6d334e3c3709-kube-api-access-fdpnd\") pod \"authentication-operator-69f744f599-wjdd5\" (UID: \"c0ca0ce3-929c-48d6-a3f7-6d334e3c3709\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-wjdd5" Dec 13 06:51:36 crc kubenswrapper[4971]: I1213 06:51:36.358169 4971 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-machine-config-operator/machine-config-controller-84d6567774-kf7sp"] Dec 13 06:51:36 crc kubenswrapper[4971]: I1213 06:51:36.361769 4971 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-fz6wk" Dec 13 06:51:36 crc kubenswrapper[4971]: I1213 06:51:36.367089 4971 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-vsp8z\" (UniqueName: \"kubernetes.io/projected/4d847949-530c-4cbc-b11d-c1c8d80d3db7-kube-api-access-vsp8z\") pod \"ingress-canary-ffhlq\" (UID: \"4d847949-530c-4cbc-b11d-c1c8d80d3db7\") " pod="openshift-ingress-canary/ingress-canary-ffhlq" Dec 13 06:51:36 crc kubenswrapper[4971]: I1213 06:51:36.382274 4971 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-2m57w\" (UniqueName: \"kubernetes.io/projected/637e887c-db01-4b47-9f17-2d61b731a118-kube-api-access-2m57w\") pod \"dns-default-dntmz\" (UID: \"637e887c-db01-4b47-9f17-2d61b731a118\") " pod="openshift-dns/dns-default-dntmz" Dec 13 06:51:36 crc kubenswrapper[4971]: I1213 06:51:36.388792 4971 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-pxjqw" Dec 13 06:51:36 crc kubenswrapper[4971]: I1213 06:51:36.399717 4971 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-wnzs2\" (UniqueName: \"kubernetes.io/projected/5fd61507-84bf-4ef5-8bd4-731075ef58ef-kube-api-access-wnzs2\") pod \"csi-hostpathplugin-7nhhk\" (UID: \"5fd61507-84bf-4ef5-8bd4-731075ef58ef\") " pod="hostpath-provisioner/csi-hostpathplugin-7nhhk" Dec 13 06:51:36 crc kubenswrapper[4971]: I1213 06:51:36.404718 4971 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-service-ca-operator/service-ca-operator-777779d784-drq5g" Dec 13 06:51:36 crc kubenswrapper[4971]: I1213 06:51:36.419655 4971 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29426805-fxxrz" Dec 13 06:51:36 crc kubenswrapper[4971]: I1213 06:51:36.426491 4971 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-fwvzw\" (UniqueName: \"kubernetes.io/projected/87f8bc63-d8c1-415f-975f-b8418034197b-kube-api-access-fwvzw\") pod \"packageserver-d55dfcdfc-lnrkm\" (UID: \"87f8bc63-d8c1-415f-975f-b8418034197b\") " pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-lnrkm" Dec 13 06:51:36 crc kubenswrapper[4971]: I1213 06:51:36.428201 4971 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-storage-version-migrator/migrator-59844c95c7-z2p8z" Dec 13 06:51:36 crc kubenswrapper[4971]: I1213 06:51:36.428878 4971 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 13 06:51:36 crc kubenswrapper[4971]: E1213 06:51:36.429038 4971 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-13 06:51:36.929015426 +0000 UTC m=+153.533424884 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 13 06:51:36 crc kubenswrapper[4971]: I1213 06:51:36.429262 4971 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-mdd8q\" (UID: \"cba893c6-ff99-466a-9db0-a62e7a87a334\") " pod="openshift-image-registry/image-registry-697d97f7c8-mdd8q" Dec 13 06:51:36 crc kubenswrapper[4971]: E1213 06:51:36.429591 4971 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-12-13 06:51:36.929580201 +0000 UTC m=+153.533989649 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-mdd8q" (UID: "cba893c6-ff99-466a-9db0-a62e7a87a334") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 13 06:51:36 crc kubenswrapper[4971]: I1213 06:51:36.436974 4971 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-authentication-operator/authentication-operator-69f744f599-wjdd5" Dec 13 06:51:36 crc kubenswrapper[4971]: I1213 06:51:36.446002 4971 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-lnrkm" Dec 13 06:51:36 crc kubenswrapper[4971]: I1213 06:51:36.453444 4971 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-4ccc9\" (UniqueName: \"kubernetes.io/projected/3052c2e3-f9bf-442b-b153-d60eb86f9c3a-kube-api-access-4ccc9\") pod \"catalog-operator-68c6474976-l4rzp\" (UID: \"3052c2e3-f9bf-442b-b153-d60eb86f9c3a\") " pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-l4rzp" Dec 13 06:51:36 crc kubenswrapper[4971]: I1213 06:51:36.461866 4971 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="hostpath-provisioner/csi-hostpathplugin-7nhhk" Dec 13 06:51:36 crc kubenswrapper[4971]: I1213 06:51:36.464227 4971 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-jdnll\" (UniqueName: \"kubernetes.io/projected/1c4c5572-81b1-4270-b519-b077f9682f49-kube-api-access-jdnll\") pod \"machine-config-server-xdxjz\" (UID: \"1c4c5572-81b1-4270-b519-b077f9682f49\") " pod="openshift-machine-config-operator/machine-config-server-xdxjz" Dec 13 06:51:36 crc kubenswrapper[4971]: I1213 06:51:36.487031 4971 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-dns/dns-default-dntmz" Dec 13 06:51:36 crc kubenswrapper[4971]: I1213 06:51:36.498155 4971 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/6e6307eb-6448-4cbb-b2a3-93e340f5b3a4-kube-api-access\") pod \"kube-controller-manager-operator-78b949d7b-cwwf9\" (UID: \"6e6307eb-6448-4cbb-b2a3-93e340f5b3a4\") " pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-cwwf9" Dec 13 06:51:36 crc kubenswrapper[4971]: I1213 06:51:36.498153 4971 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-authentication/oauth-openshift-558db77b4-nzzbm" event={"ID":"23e16ced-5b54-4a51-a1bc-4d11c907c90a","Type":"ContainerStarted","Data":"3fa8695071968f87224b501328ee01062999751e51021bd23f6e31d535f5c42d"} Dec 13 06:51:36 crc kubenswrapper[4971]: I1213 06:51:36.498215 4971 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-machine-config-operator/machine-config-server-xdxjz" Dec 13 06:51:36 crc kubenswrapper[4971]: I1213 06:51:36.498558 4971 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-authentication/oauth-openshift-558db77b4-nzzbm" Dec 13 06:51:36 crc kubenswrapper[4971]: I1213 06:51:36.499282 4971 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-ingress-canary/ingress-canary-ffhlq" Dec 13 06:51:36 crc kubenswrapper[4971]: I1213 06:51:36.501178 4971 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-xfvfg"] Dec 13 06:51:36 crc kubenswrapper[4971]: I1213 06:51:36.524239 4971 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-c5lc5\" (UniqueName: \"kubernetes.io/projected/6a3649bc-2d2e-4897-8977-99c08aa77e01-kube-api-access-c5lc5\") pod \"kube-storage-version-migrator-operator-b67b599dd-6blxl\" (UID: \"6a3649bc-2d2e-4897-8977-99c08aa77e01\") " pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-6blxl" Dec 13 06:51:36 crc kubenswrapper[4971]: I1213 06:51:36.530382 4971 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 13 06:51:36 crc kubenswrapper[4971]: E1213 06:51:36.530737 4971 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-13 06:51:37.0307225 +0000 UTC m=+153.635131948 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 13 06:51:36 crc kubenswrapper[4971]: I1213 06:51:36.543711 4971 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-service-ca/service-ca-9c57cc56f-q99rh" event={"ID":"dcdf01b4-83e9-4d28-821a-e4321ce6b92c","Type":"ContainerStarted","Data":"86c775fa788328c6558789b48d3e8e7e85a3addc1338a17ffe8c2882a3ecc11f"} Dec 13 06:51:36 crc kubenswrapper[4971]: I1213 06:51:36.547018 4971 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/marketplace-operator-79b997595-j9zqf" event={"ID":"f14c8a94-8589-4059-bddf-329452355fab","Type":"ContainerStarted","Data":"59144fe25fa154135eea30f6ed1f54f23ce9a403973267f4c8d8e235c632965a"} Dec 13 06:51:36 crc kubenswrapper[4971]: I1213 06:51:36.556453 4971 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-route-controller-manager/route-controller-manager-6576b87f9c-9d86n"] Dec 13 06:51:36 crc kubenswrapper[4971]: I1213 06:51:36.559894 4971 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-console/console-f9d7485db-5fd9b" event={"ID":"9f0db656-7a94-415b-b737-30ed342817ee","Type":"ContainerStarted","Data":"8e76dd9bdab28393e8fb3d0e9f5f44af7e8d1a8c9d3e19e32359e99815f53655"} Dec 13 06:51:36 crc kubenswrapper[4971]: I1213 06:51:36.561743 4971 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ingress/router-default-5444994796-mzhp6" event={"ID":"d89e3583-3f66-4259-b6f9-bffe8ddbde12","Type":"ContainerStarted","Data":"46f0e746181c8f41e7f9c597d795e818dad8cb15ef34b2c6740076ce085432eb"} Dec 13 06:51:36 crc kubenswrapper[4971]: I1213 06:51:36.569981 4971 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-apiserver/apiserver-76f77b778f-2n4cq" event={"ID":"bec6e53b-3998-4402-b77b-87ff79f31827","Type":"ContainerStarted","Data":"8aec7a88f79f8bcfef79483b7e348f7df3dea518c2d8eda95826cfb3e4905b21"} Dec 13 06:51:36 crc kubenswrapper[4971]: I1213 06:51:36.574353 4971 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-7ddbb" event={"ID":"97219ee9-4c52-49aa-903d-044a079d4ecb","Type":"ContainerStarted","Data":"7d168b4907756ee048b4518c5d45a587dd4e440251ddbe1d2a6a60994f0bf2bb"} Dec 13 06:51:36 crc kubenswrapper[4971]: I1213 06:51:36.574398 4971 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-7ddbb" event={"ID":"97219ee9-4c52-49aa-903d-044a079d4ecb","Type":"ContainerStarted","Data":"f4575b1add536593887d04c8462f0853926b44b1df45c465424c7b12e6dde417"} Dec 13 06:51:36 crc kubenswrapper[4971]: I1213 06:51:36.588726 4971 generic.go:334] "Generic (PLEG): container finished" podID="874901b0-9650-4029-8d21-74441f40e201" containerID="ed73e9b2cfd37518aa5746ba469ca355d12224b1daf2a71c90f4fa12294a6c84" exitCode=0 Dec 13 06:51:36 crc kubenswrapper[4971]: I1213 06:51:36.588859 4971 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-config-operator/openshift-config-operator-7777fb866f-vmdsp" event={"ID":"874901b0-9650-4029-8d21-74441f40e201","Type":"ContainerDied","Data":"ed73e9b2cfd37518aa5746ba469ca355d12224b1daf2a71c90f4fa12294a6c84"} Dec 13 06:51:36 crc kubenswrapper[4971]: I1213 06:51:36.588895 4971 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-config-operator/openshift-config-operator-7777fb866f-vmdsp" event={"ID":"874901b0-9650-4029-8d21-74441f40e201","Type":"ContainerStarted","Data":"938c74553bc5e3263d5013110fe9acf746789b9824f013405ce7ac7d4d9c182f"} Dec 13 06:51:36 crc kubenswrapper[4971]: I1213 06:51:36.592055 4971 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-admission-controller-857f4d67dd-z68ks" event={"ID":"32c8b95e-38c8-43fb-b1e4-86ac9a27c2f0","Type":"ContainerStarted","Data":"730c698afe4dba5a8c5c685e108fb77724e72cc4f246395e85300dcc1040f6c0"} Dec 13 06:51:36 crc kubenswrapper[4971]: I1213 06:51:36.593323 4971 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-zqt2t" event={"ID":"8294e250-74ea-497c-b6ea-08b957184eec","Type":"ContainerStarted","Data":"118a88bd555c19c77e9525e0adf9987a58432cac052d21aa7c543c8863a5ab3b"} Dec 13 06:51:36 crc kubenswrapper[4971]: I1213 06:51:36.595119 4971 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-mlbwt" event={"ID":"817cbb9d-ed5b-4955-8b6a-4b33155b744c","Type":"ContainerStarted","Data":"66a1ab9a1319cc1f651874f3b6cb74871b3a031252445f8a7eb8a06d0e7bb467"} Dec 13 06:51:36 crc kubenswrapper[4971]: I1213 06:51:36.595146 4971 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-mlbwt" event={"ID":"817cbb9d-ed5b-4955-8b6a-4b33155b744c","Type":"ContainerStarted","Data":"b792d528fca59a88e9d463b79fe3aa0ca0d7ea6b8996cf9ca2d744ae7fa0fe94"} Dec 13 06:51:36 crc kubenswrapper[4971]: I1213 06:51:36.595159 4971 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-mlbwt" event={"ID":"817cbb9d-ed5b-4955-8b6a-4b33155b744c","Type":"ContainerStarted","Data":"6a194bf19ea2f61260b2b693069e550250ea6f21a2e446a73da0df896d8b0f78"} Dec 13 06:51:36 crc kubenswrapper[4971]: I1213 06:51:36.599027 4971 generic.go:334] "Generic (PLEG): container finished" podID="8e7aefba-ae2e-4069-892e-544c85b9af15" containerID="cee2cae5e3d45265b0bb811279b18d695c88b9fd3db4b3d78c3b5faddc15dc96" exitCode=0 Dec 13 06:51:36 crc kubenswrapper[4971]: I1213 06:51:36.599564 4971 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-mb2q7" event={"ID":"8e7aefba-ae2e-4069-892e-544c85b9af15","Type":"ContainerDied","Data":"cee2cae5e3d45265b0bb811279b18d695c88b9fd3db4b3d78c3b5faddc15dc96"} Dec 13 06:51:36 crc kubenswrapper[4971]: I1213 06:51:36.599686 4971 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-mb2q7" event={"ID":"8e7aefba-ae2e-4069-892e-544c85b9af15","Type":"ContainerStarted","Data":"0c046f50f363868d44561771d6f5685bf6ce0fa79349f04052efdb1c5f0cd086"} Dec 13 06:51:36 crc kubenswrapper[4971]: I1213 06:51:36.608770 4971 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-etcd-operator/etcd-operator-b45778765-6xfwq" event={"ID":"56c62b80-0a33-4156-b52b-cb54c98a4343","Type":"ContainerStarted","Data":"7c8cacdc2911ff16c349a62962ad34eb9b6eedf77eb20dbcf6a6cd3da46ae058"} Dec 13 06:51:36 crc kubenswrapper[4971]: I1213 06:51:36.626042 4971 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-gtrwg" event={"ID":"bb0ead7b-eba0-4d16-9363-df54caee0a32","Type":"ContainerStarted","Data":"0b3d437398e6e1f9032006811a93c33a40fb4c638ce89999d6caac1aaf96e914"} Dec 13 06:51:36 crc kubenswrapper[4971]: I1213 06:51:36.627008 4971 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-ingress-operator/ingress-operator-5b745b69d9-pkmb5"] Dec 13 06:51:36 crc kubenswrapper[4971]: I1213 06:51:36.632216 4971 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-mdd8q\" (UID: \"cba893c6-ff99-466a-9db0-a62e7a87a334\") " pod="openshift-image-registry/image-registry-697d97f7c8-mdd8q" Dec 13 06:51:36 crc kubenswrapper[4971]: E1213 06:51:36.634647 4971 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-12-13 06:51:37.134629009 +0000 UTC m=+153.739038457 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-mdd8q" (UID: "cba893c6-ff99-466a-9db0-a62e7a87a334") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 13 06:51:36 crc kubenswrapper[4971]: I1213 06:51:36.656381 4971 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-console-operator/console-operator-58897d9998-kdfcd"] Dec 13 06:51:36 crc kubenswrapper[4971]: I1213 06:51:36.662331 4971 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-authentication/oauth-openshift-558db77b4-nzzbm" Dec 13 06:51:36 crc kubenswrapper[4971]: I1213 06:51:36.669476 4971 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-api/machine-api-operator-5694c8668f-t24dm" event={"ID":"c45a314c-77da-43fc-b2ae-9de350540c2b","Type":"ContainerStarted","Data":"8ef8c7faef69804d610ded319e0752b48aa63868e528b98087486902343f21d4"} Dec 13 06:51:36 crc kubenswrapper[4971]: I1213 06:51:36.669563 4971 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-api/machine-api-operator-5694c8668f-t24dm" event={"ID":"c45a314c-77da-43fc-b2ae-9de350540c2b","Type":"ContainerStarted","Data":"8b1a8560f5dc294b1557815cf47362b5e9f0d71691a25371bc36dfdd07b9fb47"} Dec 13 06:51:36 crc kubenswrapper[4971]: I1213 06:51:36.669576 4971 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-api/machine-api-operator-5694c8668f-t24dm" event={"ID":"c45a314c-77da-43fc-b2ae-9de350540c2b","Type":"ContainerStarted","Data":"9c4e4064236c42265993bd0592a5564ea72a0b1010749d1dab5988fc7f547753"} Dec 13 06:51:36 crc kubenswrapper[4971]: I1213 06:51:36.671842 4971 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-6blxl" Dec 13 06:51:36 crc kubenswrapper[4971]: I1213 06:51:36.674886 4971 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-controller-manager/controller-manager-879f6c89f-xtxr8" event={"ID":"7148e573-e2f5-4fc5-b30c-4e777516509e","Type":"ContainerStarted","Data":"45ca28d8f97bfc12cd2dcba8da0fabcb8a352f2d2188c0d53d940710d41bbb65"} Dec 13 06:51:36 crc kubenswrapper[4971]: I1213 06:51:36.678442 4971 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-cwwf9" Dec 13 06:51:36 crc kubenswrapper[4971]: I1213 06:51:36.681963 4971 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-cluster-machine-approver/machine-approver-56656f9798-s75fl" event={"ID":"88850e0f-df47-4650-be7e-d14a83ce05f4","Type":"ContainerStarted","Data":"f56ceba0a541d876dad769d61972305df19e623e2de2afc81807deda1b67e807"} Dec 13 06:51:36 crc kubenswrapper[4971]: W1213 06:51:36.688957 4971 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod2cf6e363_8511_4c79_9181_79dba726e5af.slice/crio-5fa0eec368450ce9c968373d6851400be728d4c59120d3368fa30d48866623e7 WatchSource:0}: Error finding container 5fa0eec368450ce9c968373d6851400be728d4c59120d3368fa30d48866623e7: Status 404 returned error can't find the container with id 5fa0eec368450ce9c968373d6851400be728d4c59120d3368fa30d48866623e7 Dec 13 06:51:36 crc kubenswrapper[4971]: I1213 06:51:36.695058 4971 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-l4rzp" Dec 13 06:51:36 crc kubenswrapper[4971]: I1213 06:51:36.706757 4971 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-console/downloads-7954f5f757-dqk9z"] Dec 13 06:51:36 crc kubenswrapper[4971]: I1213 06:51:36.733984 4971 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 13 06:51:36 crc kubenswrapper[4971]: E1213 06:51:36.734743 4971 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-13 06:51:37.234717772 +0000 UTC m=+153.839127230 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 13 06:51:36 crc kubenswrapper[4971]: I1213 06:51:36.734936 4971 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-mdd8q\" (UID: \"cba893c6-ff99-466a-9db0-a62e7a87a334\") " pod="openshift-image-registry/image-registry-697d97f7c8-mdd8q" Dec 13 06:51:36 crc kubenswrapper[4971]: E1213 06:51:36.735288 4971 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-12-13 06:51:37.235278506 +0000 UTC m=+153.839688024 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-mdd8q" (UID: "cba893c6-ff99-466a-9db0-a62e7a87a334") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 13 06:51:36 crc kubenswrapper[4971]: I1213 06:51:36.793734 4971 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-dns-operator/dns-operator-744455d44c-b999k"] Dec 13 06:51:36 crc kubenswrapper[4971]: I1213 06:51:36.827999 4971 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-machine-api/control-plane-machine-set-operator-78cbb6b69f-bgdbk"] Dec 13 06:51:36 crc kubenswrapper[4971]: I1213 06:51:36.835669 4971 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 13 06:51:36 crc kubenswrapper[4971]: E1213 06:51:36.836742 4971 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-13 06:51:37.336721302 +0000 UTC m=+153.941130750 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 13 06:51:36 crc kubenswrapper[4971]: I1213 06:51:36.938625 4971 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-mdd8q\" (UID: \"cba893c6-ff99-466a-9db0-a62e7a87a334\") " pod="openshift-image-registry/image-registry-697d97f7c8-mdd8q" Dec 13 06:51:36 crc kubenswrapper[4971]: E1213 06:51:36.938997 4971 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-12-13 06:51:37.438971581 +0000 UTC m=+154.043381029 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-mdd8q" (UID: "cba893c6-ff99-466a-9db0-a62e7a87a334") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 13 06:51:37 crc kubenswrapper[4971]: I1213 06:51:37.011129 4971 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-g8vfp"] Dec 13 06:51:37 crc kubenswrapper[4971]: I1213 06:51:37.039666 4971 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 13 06:51:37 crc kubenswrapper[4971]: E1213 06:51:37.040164 4971 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-13 06:51:37.54014628 +0000 UTC m=+154.144555728 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 13 06:51:37 crc kubenswrapper[4971]: I1213 06:51:37.108391 4971 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-machine-config-operator/machine-config-operator-74547568cd-4cn2z"] Dec 13 06:51:37 crc kubenswrapper[4971]: I1213 06:51:37.110384 4971 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29426805-fxxrz"] Dec 13 06:51:37 crc kubenswrapper[4971]: I1213 06:51:37.139668 4971 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-pxjqw"] Dec 13 06:51:37 crc kubenswrapper[4971]: I1213 06:51:37.141894 4971 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-mdd8q\" (UID: \"cba893c6-ff99-466a-9db0-a62e7a87a334\") " pod="openshift-image-registry/image-registry-697d97f7c8-mdd8q" Dec 13 06:51:37 crc kubenswrapper[4971]: E1213 06:51:37.142214 4971 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-12-13 06:51:37.642200953 +0000 UTC m=+154.246610411 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-mdd8q" (UID: "cba893c6-ff99-466a-9db0-a62e7a87a334") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 13 06:51:37 crc kubenswrapper[4971]: I1213 06:51:37.244717 4971 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 13 06:51:37 crc kubenswrapper[4971]: E1213 06:51:37.245162 4971 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-13 06:51:37.745140437 +0000 UTC m=+154.349549885 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 13 06:51:37 crc kubenswrapper[4971]: I1213 06:51:37.322660 4971 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-fz6wk"] Dec 13 06:51:37 crc kubenswrapper[4971]: I1213 06:51:37.342623 4971 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["hostpath-provisioner/csi-hostpathplugin-7nhhk"] Dec 13 06:51:37 crc kubenswrapper[4971]: I1213 06:51:37.347353 4971 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-mdd8q\" (UID: \"cba893c6-ff99-466a-9db0-a62e7a87a334\") " pod="openshift-image-registry/image-registry-697d97f7c8-mdd8q" Dec 13 06:51:37 crc kubenswrapper[4971]: E1213 06:51:37.347892 4971 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-12-13 06:51:37.847877037 +0000 UTC m=+154.452286485 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-mdd8q" (UID: "cba893c6-ff99-466a-9db0-a62e7a87a334") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 13 06:51:37 crc kubenswrapper[4971]: I1213 06:51:37.382127 4971 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-storage-version-migrator/migrator-59844c95c7-z2p8z"] Dec 13 06:51:37 crc kubenswrapper[4971]: I1213 06:51:37.449414 4971 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 13 06:51:37 crc kubenswrapper[4971]: E1213 06:51:37.450176 4971 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-13 06:51:37.950157225 +0000 UTC m=+154.554566673 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 13 06:51:37 crc kubenswrapper[4971]: I1213 06:51:37.551841 4971 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-mdd8q\" (UID: \"cba893c6-ff99-466a-9db0-a62e7a87a334\") " pod="openshift-image-registry/image-registry-697d97f7c8-mdd8q" Dec 13 06:51:37 crc kubenswrapper[4971]: E1213 06:51:37.552163 4971 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-12-13 06:51:38.052150695 +0000 UTC m=+154.656560143 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-mdd8q" (UID: "cba893c6-ff99-466a-9db0-a62e7a87a334") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 13 06:51:37 crc kubenswrapper[4971]: I1213 06:51:37.599443 4971 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-cwwf9"] Dec 13 06:51:37 crc kubenswrapper[4971]: I1213 06:51:37.617155 4971 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-service-ca-operator/service-ca-operator-777779d784-drq5g"] Dec 13 06:51:37 crc kubenswrapper[4971]: I1213 06:51:37.638763 4971 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-authentication-operator/authentication-operator-69f744f599-wjdd5"] Dec 13 06:51:37 crc kubenswrapper[4971]: I1213 06:51:37.659986 4971 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 13 06:51:37 crc kubenswrapper[4971]: E1213 06:51:37.660411 4971 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-13 06:51:38.160390344 +0000 UTC m=+154.764799792 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 13 06:51:37 crc kubenswrapper[4971]: I1213 06:51:37.667230 4971 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-lnrkm"] Dec 13 06:51:37 crc kubenswrapper[4971]: I1213 06:51:37.766066 4971 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-mdd8q\" (UID: \"cba893c6-ff99-466a-9db0-a62e7a87a334\") " pod="openshift-image-registry/image-registry-697d97f7c8-mdd8q" Dec 13 06:51:37 crc kubenswrapper[4971]: E1213 06:51:37.766370 4971 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-12-13 06:51:38.266359985 +0000 UTC m=+154.870769423 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-mdd8q" (UID: "cba893c6-ff99-466a-9db0-a62e7a87a334") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 13 06:51:37 crc kubenswrapper[4971]: I1213 06:51:37.821704 4971 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-server-xdxjz" event={"ID":"1c4c5572-81b1-4270-b519-b077f9682f49","Type":"ContainerStarted","Data":"0ac3a1312aea7be14bc6c7927e2e0261d47c6528e33fe31f7d7a6962d5a3fee3"} Dec 13 06:51:37 crc kubenswrapper[4971]: I1213 06:51:37.823103 4971 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-console-operator/console-operator-58897d9998-kdfcd" event={"ID":"94469d04-47fd-4bd8-8258-9e874767f250","Type":"ContainerStarted","Data":"c9923fd85c9c3aa5471d9dbc821cc5d0f5d49ab01fac0e027600342980b3bed6"} Dec 13 06:51:37 crc kubenswrapper[4971]: I1213 06:51:37.831035 4971 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-g8vfp" event={"ID":"479b25e6-6618-4123-a913-ca50d6c4ba28","Type":"ContainerStarted","Data":"b2b90444fd549fa283d43505c1bdb15b4ff64ec6752d610917cc28cba6672fff"} Dec 13 06:51:37 crc kubenswrapper[4971]: W1213 06:51:37.832500 4971 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod87f8bc63_d8c1_415f_975f_b8418034197b.slice/crio-41aca581416eec7c9faeb5722769713273f09449d8841c874c68969c01dc4f28 WatchSource:0}: Error finding container 41aca581416eec7c9faeb5722769713273f09449d8841c874c68969c01dc4f28: Status 404 returned error can't find the container with id 41aca581416eec7c9faeb5722769713273f09449d8841c874c68969c01dc4f28 Dec 13 06:51:37 crc kubenswrapper[4971]: I1213 06:51:37.856722 4971 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-operator-74547568cd-4cn2z" event={"ID":"874a112c-101d-4a74-a674-d9afe62b6dc8","Type":"ContainerStarted","Data":"39b1e7bd916d9215ff8583d53e3b8cd00d109e2ee64c3db0782fc761245da077"} Dec 13 06:51:37 crc kubenswrapper[4971]: I1213 06:51:37.862489 4971 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-console/downloads-7954f5f757-dqk9z" event={"ID":"ee2e94f4-cadb-4b17-8853-a1292cb02404","Type":"ContainerStarted","Data":"e1d13cb59eb966db3e288778dc399d8a54d527b995d8e046e529e247a1804b90"} Dec 13 06:51:37 crc kubenswrapper[4971]: I1213 06:51:37.862539 4971 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-console/downloads-7954f5f757-dqk9z" event={"ID":"ee2e94f4-cadb-4b17-8853-a1292cb02404","Type":"ContainerStarted","Data":"928871d09dfb91909527e7e8a12c5a2cf316641c1df7d167bb0f5fd16269cd87"} Dec 13 06:51:37 crc kubenswrapper[4971]: I1213 06:51:37.863259 4971 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-console/downloads-7954f5f757-dqk9z" Dec 13 06:51:37 crc kubenswrapper[4971]: I1213 06:51:37.867804 4971 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-service-ca/service-ca-9c57cc56f-q99rh" event={"ID":"dcdf01b4-83e9-4d28-821a-e4321ce6b92c","Type":"ContainerStarted","Data":"9795fa73c69fff8da3192bf7c3703c0a399fdc37ebc97e35f054a64d9090be57"} Dec 13 06:51:37 crc kubenswrapper[4971]: I1213 06:51:37.870439 4971 patch_prober.go:28] interesting pod/downloads-7954f5f757-dqk9z container/download-server namespace/openshift-console: Readiness probe status=failure output="Get \"http://10.217.0.20:8080/\": dial tcp 10.217.0.20:8080: connect: connection refused" start-of-body= Dec 13 06:51:37 crc kubenswrapper[4971]: I1213 06:51:37.870850 4971 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-console/downloads-7954f5f757-dqk9z" podUID="ee2e94f4-cadb-4b17-8853-a1292cb02404" containerName="download-server" probeResult="failure" output="Get \"http://10.217.0.20:8080/\": dial tcp 10.217.0.20:8080: connect: connection refused" Dec 13 06:51:37 crc kubenswrapper[4971]: I1213 06:51:37.871816 4971 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 13 06:51:37 crc kubenswrapper[4971]: E1213 06:51:37.872196 4971 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-13 06:51:38.372176193 +0000 UTC m=+154.976585641 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 13 06:51:37 crc kubenswrapper[4971]: I1213 06:51:37.872329 4971 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-fz6wk" event={"ID":"4b58bd40-cb76-46ba-9391-a350d830d729","Type":"ContainerStarted","Data":"edec19d3f0a26073559ae9611185b426846377168810d526e4fec8636dbb8c6c"} Dec 13 06:51:37 crc kubenswrapper[4971]: I1213 06:51:37.876612 4971 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-cluster-machine-approver/machine-approver-56656f9798-s75fl" podStartSLOduration=132.876592124 podStartE2EDuration="2m12.876592124s" podCreationTimestamp="2025-12-13 06:49:25 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-13 06:51:37.875703162 +0000 UTC m=+154.480112611" watchObservedRunningTime="2025-12-13 06:51:37.876592124 +0000 UTC m=+154.481001572" Dec 13 06:51:37 crc kubenswrapper[4971]: I1213 06:51:37.894370 4971 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-gtrwg" event={"ID":"bb0ead7b-eba0-4d16-9363-df54caee0a32","Type":"ContainerStarted","Data":"885dc91301557e19bbf37afeaa19febd6d7e13d7b12f3e655d9f0ef8669eea59"} Dec 13 06:51:37 crc kubenswrapper[4971]: I1213 06:51:37.903625 4971 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-console/console-f9d7485db-5fd9b" podStartSLOduration=132.903502155 podStartE2EDuration="2m12.903502155s" podCreationTimestamp="2025-12-13 06:49:25 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-13 06:51:37.903165677 +0000 UTC m=+154.507575125" watchObservedRunningTime="2025-12-13 06:51:37.903502155 +0000 UTC m=+154.507911613" Dec 13 06:51:37 crc kubenswrapper[4971]: I1213 06:51:37.906598 4971 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-controller-manager/controller-manager-879f6c89f-xtxr8" event={"ID":"7148e573-e2f5-4fc5-b30c-4e777516509e","Type":"ContainerStarted","Data":"07720aa0b0d810dbae552adbd6b261970ba16c7af295ef016ea351b90dade542"} Dec 13 06:51:37 crc kubenswrapper[4971]: I1213 06:51:37.913767 4971 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-controller-manager/controller-manager-879f6c89f-xtxr8" Dec 13 06:51:37 crc kubenswrapper[4971]: I1213 06:51:37.938400 4971 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ingress-operator/ingress-operator-5b745b69d9-pkmb5" event={"ID":"22f3a4f0-1686-4087-bcec-e48f2015ffe2","Type":"ContainerStarted","Data":"fa8184c38ddb6b4c38a1d0f6f6c4f5e964b0610158e0d6937943e7907dcf8c57"} Dec 13 06:51:37 crc kubenswrapper[4971]: I1213 06:51:37.938448 4971 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ingress-operator/ingress-operator-5b745b69d9-pkmb5" event={"ID":"22f3a4f0-1686-4087-bcec-e48f2015ffe2","Type":"ContainerStarted","Data":"61cb0b42531a9bd09334decfcb3592face4ce97e3bff7fd54c969a20a9087392"} Dec 13 06:51:37 crc kubenswrapper[4971]: I1213 06:51:37.941164 4971 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-6blxl"] Dec 13 06:51:37 crc kubenswrapper[4971]: I1213 06:51:37.946586 4971 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-authentication/oauth-openshift-558db77b4-nzzbm" podStartSLOduration=132.946567325 podStartE2EDuration="2m12.946567325s" podCreationTimestamp="2025-12-13 06:49:25 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-13 06:51:37.94476116 +0000 UTC m=+154.549170618" watchObservedRunningTime="2025-12-13 06:51:37.946567325 +0000 UTC m=+154.550976783" Dec 13 06:51:37 crc kubenswrapper[4971]: I1213 06:51:37.946673 4971 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-admission-controller-857f4d67dd-z68ks" event={"ID":"32c8b95e-38c8-43fb-b1e4-86ac9a27c2f0","Type":"ContainerStarted","Data":"6e5c5da9ffd7934aeef0625bc9cac0dcb401bca1fe8dd2dd38605788adf8332b"} Dec 13 06:51:37 crc kubenswrapper[4971]: I1213 06:51:37.953960 4971 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-controller-84d6567774-kf7sp" event={"ID":"2f87c67a-fcf1-46e3-9bbd-9bf03ee55da3","Type":"ContainerStarted","Data":"6ddb7367a3b39a1cc186931579616a204b5eceb503bf3a17ee7b626768c134ac"} Dec 13 06:51:37 crc kubenswrapper[4971]: I1213 06:51:37.954009 4971 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-controller-84d6567774-kf7sp" event={"ID":"2f87c67a-fcf1-46e3-9bbd-9bf03ee55da3","Type":"ContainerStarted","Data":"50f9d903b6e0e890213ead29d8bf868a1e0230c3818c77c3243b602af2acb948"} Dec 13 06:51:37 crc kubenswrapper[4971]: I1213 06:51:37.970054 4971 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ingress/router-default-5444994796-mzhp6" event={"ID":"d89e3583-3f66-4259-b6f9-bffe8ddbde12","Type":"ContainerStarted","Data":"83bc98325ca70abda5ef46ceb83413c82df6cab39a63830e3dc15e71c5884ac1"} Dec 13 06:51:37 crc kubenswrapper[4971]: I1213 06:51:37.973424 4971 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-mdd8q\" (UID: \"cba893c6-ff99-466a-9db0-a62e7a87a334\") " pod="openshift-image-registry/image-registry-697d97f7c8-mdd8q" Dec 13 06:51:37 crc kubenswrapper[4971]: E1213 06:51:37.975893 4971 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-12-13 06:51:38.475872577 +0000 UTC m=+155.080282035 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-mdd8q" (UID: "cba893c6-ff99-466a-9db0-a62e7a87a334") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 13 06:51:37 crc kubenswrapper[4971]: I1213 06:51:37.994227 4971 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-controller-manager/controller-manager-879f6c89f-xtxr8" Dec 13 06:51:38 crc kubenswrapper[4971]: I1213 06:51:37.996922 4971 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-zqt2t" event={"ID":"8294e250-74ea-497c-b6ea-08b957184eec","Type":"ContainerStarted","Data":"fe00af26f1ea70fb5b39e765e92bcb9e467ff25d3b2782c23ff01a277b751645"} Dec 13 06:51:38 crc kubenswrapper[4971]: I1213 06:51:38.021953 4971 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-machine-api/machine-api-operator-5694c8668f-t24dm" podStartSLOduration=133.021936332 podStartE2EDuration="2m13.021936332s" podCreationTimestamp="2025-12-13 06:49:25 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-13 06:51:38.021476081 +0000 UTC m=+154.625885539" watchObservedRunningTime="2025-12-13 06:51:38.021936332 +0000 UTC m=+154.626345780" Dec 13 06:51:38 crc kubenswrapper[4971]: I1213 06:51:38.024290 4971 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29426805-fxxrz" event={"ID":"c1575ff9-1c53-4732-bc48-252f9f270597","Type":"ContainerStarted","Data":"f938bdc1991e28af8004493352b892f7c2166554190ee959d89d1ec0f31229f0"} Dec 13 06:51:38 crc kubenswrapper[4971]: I1213 06:51:38.036294 4971 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-xfvfg" event={"ID":"375613f8-1e50-42fe-b593-2a6a07afbe64","Type":"ContainerStarted","Data":"0cee4bef63b66a0377d3a761cf2e3ffd4b475a352ec2d32fe9abad8a5ef93c1e"} Dec 13 06:51:38 crc kubenswrapper[4971]: I1213 06:51:38.036348 4971 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-xfvfg" event={"ID":"375613f8-1e50-42fe-b593-2a6a07afbe64","Type":"ContainerStarted","Data":"3aede2d9c7f5c78426eae8471e0d4131af249d5b69aa0e0db3d3c04610fe9485"} Dec 13 06:51:38 crc kubenswrapper[4971]: I1213 06:51:38.056424 4971 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-ingress-canary/ingress-canary-ffhlq"] Dec 13 06:51:38 crc kubenswrapper[4971]: I1213 06:51:38.059981 4971 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-ingress/router-default-5444994796-mzhp6" Dec 13 06:51:38 crc kubenswrapper[4971]: I1213 06:51:38.068675 4971 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-pxjqw" event={"ID":"ea24f72e-df9d-418f-aaef-dfc0573eface","Type":"ContainerStarted","Data":"34b15ccb7b1a5a97f783472510e7e47db9236bf93017bf83c32bcdc9443a46b8"} Dec 13 06:51:38 crc kubenswrapper[4971]: I1213 06:51:38.068887 4971 patch_prober.go:28] interesting pod/router-default-5444994796-mzhp6 container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Dec 13 06:51:38 crc kubenswrapper[4971]: [-]has-synced failed: reason withheld Dec 13 06:51:38 crc kubenswrapper[4971]: [+]process-running ok Dec 13 06:51:38 crc kubenswrapper[4971]: healthz check failed Dec 13 06:51:38 crc kubenswrapper[4971]: I1213 06:51:38.068972 4971 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-mzhp6" podUID="d89e3583-3f66-4259-b6f9-bffe8ddbde12" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Dec 13 06:51:38 crc kubenswrapper[4971]: I1213 06:51:38.074578 4971 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 13 06:51:38 crc kubenswrapper[4971]: E1213 06:51:38.076906 4971 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-13 06:51:38.576883172 +0000 UTC m=+155.181292620 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 13 06:51:38 crc kubenswrapper[4971]: I1213 06:51:38.081650 4971 generic.go:334] "Generic (PLEG): container finished" podID="bec6e53b-3998-4402-b77b-87ff79f31827" containerID="f2d15207e09e4f539136ff5670cdf6d5a7f929e93ccde67628bddf442cba0a52" exitCode=0 Dec 13 06:51:38 crc kubenswrapper[4971]: I1213 06:51:38.081869 4971 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-apiserver/apiserver-76f77b778f-2n4cq" event={"ID":"bec6e53b-3998-4402-b77b-87ff79f31827","Type":"ContainerDied","Data":"f2d15207e09e4f539136ff5670cdf6d5a7f929e93ccde67628bddf442cba0a52"} Dec 13 06:51:38 crc kubenswrapper[4971]: I1213 06:51:38.100894 4971 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-dns-operator/dns-operator-744455d44c-b999k" event={"ID":"ad158968-e110-4b4c-9f42-da9c08274dab","Type":"ContainerStarted","Data":"bd1ef807ac2974b3206add5893120dc1f3c48750d6f2ec28ce1d705a9cd8773e"} Dec 13 06:51:38 crc kubenswrapper[4971]: I1213 06:51:38.126470 4971 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-7ddbb" podStartSLOduration=133.126423586 podStartE2EDuration="2m13.126423586s" podCreationTimestamp="2025-12-13 06:49:25 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-13 06:51:38.110761369 +0000 UTC m=+154.715170827" watchObservedRunningTime="2025-12-13 06:51:38.126423586 +0000 UTC m=+154.730833044" Dec 13 06:51:38 crc kubenswrapper[4971]: I1213 06:51:38.148038 4971 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-mlbwt" podStartSLOduration=133.148017632 podStartE2EDuration="2m13.148017632s" podCreationTimestamp="2025-12-13 06:49:25 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-13 06:51:38.142148724 +0000 UTC m=+154.746558172" watchObservedRunningTime="2025-12-13 06:51:38.148017632 +0000 UTC m=+154.752427080" Dec 13 06:51:38 crc kubenswrapper[4971]: I1213 06:51:38.156747 4971 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-9d86n" event={"ID":"2cf6e363-8511-4c79-9181-79dba726e5af","Type":"ContainerStarted","Data":"5fa0eec368450ce9c968373d6851400be728d4c59120d3368fa30d48866623e7"} Dec 13 06:51:38 crc kubenswrapper[4971]: I1213 06:51:38.157681 4971 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-9d86n" Dec 13 06:51:38 crc kubenswrapper[4971]: I1213 06:51:38.177075 4971 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-cwwf9" event={"ID":"6e6307eb-6448-4cbb-b2a3-93e340f5b3a4","Type":"ContainerStarted","Data":"eebbfd197b34b99d43ee8f80cdd56fd766f54c80883e5f9c96de5bab18d0d933"} Dec 13 06:51:38 crc kubenswrapper[4971]: I1213 06:51:38.178179 4971 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-mdd8q\" (UID: \"cba893c6-ff99-466a-9db0-a62e7a87a334\") " pod="openshift-image-registry/image-registry-697d97f7c8-mdd8q" Dec 13 06:51:38 crc kubenswrapper[4971]: E1213 06:51:38.181030 4971 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-12-13 06:51:38.681011898 +0000 UTC m=+155.285421346 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-mdd8q" (UID: "cba893c6-ff99-466a-9db0-a62e7a87a334") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 13 06:51:38 crc kubenswrapper[4971]: I1213 06:51:38.188974 4971 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-console/downloads-7954f5f757-dqk9z" podStartSLOduration=133.188951248 podStartE2EDuration="2m13.188951248s" podCreationTimestamp="2025-12-13 06:49:25 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-13 06:51:38.18626629 +0000 UTC m=+154.790675748" watchObservedRunningTime="2025-12-13 06:51:38.188951248 +0000 UTC m=+154.793360696" Dec 13 06:51:38 crc kubenswrapper[4971]: I1213 06:51:38.193009 4971 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-dns/dns-default-dntmz"] Dec 13 06:51:38 crc kubenswrapper[4971]: I1213 06:51:38.201659 4971 patch_prober.go:28] interesting pod/route-controller-manager-6576b87f9c-9d86n container/route-controller-manager namespace/openshift-route-controller-manager: Readiness probe status=failure output="Get \"https://10.217.0.15:8443/healthz\": dial tcp 10.217.0.15:8443: connect: connection refused" start-of-body= Dec 13 06:51:38 crc kubenswrapper[4971]: I1213 06:51:38.201721 4971 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-9d86n" podUID="2cf6e363-8511-4c79-9181-79dba726e5af" containerName="route-controller-manager" probeResult="failure" output="Get \"https://10.217.0.15:8443/healthz\": dial tcp 10.217.0.15:8443: connect: connection refused" Dec 13 06:51:38 crc kubenswrapper[4971]: I1213 06:51:38.203000 4971 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-cluster-machine-approver/machine-approver-56656f9798-s75fl" event={"ID":"88850e0f-df47-4650-be7e-d14a83ce05f4","Type":"ContainerStarted","Data":"00bd391df6d8d6ca1144259fa97953ad3b5f0c377f58af0065a0599307ccde66"} Dec 13 06:51:38 crc kubenswrapper[4971]: I1213 06:51:38.210261 4971 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/catalog-operator-68c6474976-l4rzp"] Dec 13 06:51:38 crc kubenswrapper[4971]: I1213 06:51:38.212381 4971 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/marketplace-operator-79b997595-j9zqf" event={"ID":"f14c8a94-8589-4059-bddf-329452355fab","Type":"ContainerStarted","Data":"524506c76ba56186a2daa60747573baef0eb42cebe8bb321e82be093150f35d7"} Dec 13 06:51:38 crc kubenswrapper[4971]: I1213 06:51:38.213380 4971 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/marketplace-operator-79b997595-j9zqf" Dec 13 06:51:38 crc kubenswrapper[4971]: I1213 06:51:38.218604 4971 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-service-ca/service-ca-9c57cc56f-q99rh" podStartSLOduration=132.218589808 podStartE2EDuration="2m12.218589808s" podCreationTimestamp="2025-12-13 06:49:26 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-13 06:51:38.216740001 +0000 UTC m=+154.821149459" watchObservedRunningTime="2025-12-13 06:51:38.218589808 +0000 UTC m=+154.822999266" Dec 13 06:51:38 crc kubenswrapper[4971]: I1213 06:51:38.220485 4971 patch_prober.go:28] interesting pod/marketplace-operator-79b997595-j9zqf container/marketplace-operator namespace/openshift-marketplace: Readiness probe status=failure output="Get \"http://10.217.0.17:8080/healthz\": dial tcp 10.217.0.17:8080: connect: connection refused" start-of-body= Dec 13 06:51:38 crc kubenswrapper[4971]: I1213 06:51:38.220772 4971 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-marketplace/marketplace-operator-79b997595-j9zqf" podUID="f14c8a94-8589-4059-bddf-329452355fab" containerName="marketplace-operator" probeResult="failure" output="Get \"http://10.217.0.17:8080/healthz\": dial tcp 10.217.0.17:8080: connect: connection refused" Dec 13 06:51:38 crc kubenswrapper[4971]: W1213 06:51:38.246577 4971 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod6a3649bc_2d2e_4897_8977_99c08aa77e01.slice/crio-f91351dfad5e336053c8b82f4524daa30eeac3d89e5c03b3c4f4833129a3a298 WatchSource:0}: Error finding container f91351dfad5e336053c8b82f4524daa30eeac3d89e5c03b3c4f4833129a3a298: Status 404 returned error can't find the container with id f91351dfad5e336053c8b82f4524daa30eeac3d89e5c03b3c4f4833129a3a298 Dec 13 06:51:38 crc kubenswrapper[4971]: I1213 06:51:38.246721 4971 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="hostpath-provisioner/csi-hostpathplugin-7nhhk" event={"ID":"5fd61507-84bf-4ef5-8bd4-731075ef58ef","Type":"ContainerStarted","Data":"263aca208fb57bfb312b38a87a45b61dea397f6e9f9510a2e1a44fc08f6ba27b"} Dec 13 06:51:38 crc kubenswrapper[4971]: I1213 06:51:38.271466 4971 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-storage-version-migrator/migrator-59844c95c7-z2p8z" event={"ID":"5764a8c4-18e5-4025-8d36-15a7b01e9169","Type":"ContainerStarted","Data":"832eb08b7dce0d8868f603b415f6c6747ba4e3e1adc62d1d2f30aa0450ea24f8"} Dec 13 06:51:38 crc kubenswrapper[4971]: I1213 06:51:38.282746 4971 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 13 06:51:38 crc kubenswrapper[4971]: E1213 06:51:38.284079 4971 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-13 06:51:38.784057864 +0000 UTC m=+155.388467312 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 13 06:51:38 crc kubenswrapper[4971]: W1213 06:51:38.305896 4971 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod4d847949_530c_4cbc_b11d_c1c8d80d3db7.slice/crio-4884cd7b3616a12b3dacb046dc8baa93087857ec0fedc3a041a52eef783ced64 WatchSource:0}: Error finding container 4884cd7b3616a12b3dacb046dc8baa93087857ec0fedc3a041a52eef783ced64: Status 404 returned error can't find the container with id 4884cd7b3616a12b3dacb046dc8baa93087857ec0fedc3a041a52eef783ced64 Dec 13 06:51:38 crc kubenswrapper[4971]: I1213 06:51:38.310349 4971 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-api/control-plane-machine-set-operator-78cbb6b69f-bgdbk" event={"ID":"930fe38c-5ff7-45a8-a292-cd69d07e86f7","Type":"ContainerStarted","Data":"41abeacc7da0b141c34eeeafb458db4b6b65b943c812ef8d24767d25ac5040d0"} Dec 13 06:51:38 crc kubenswrapper[4971]: I1213 06:51:38.384582 4971 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-mdd8q\" (UID: \"cba893c6-ff99-466a-9db0-a62e7a87a334\") " pod="openshift-image-registry/image-registry-697d97f7c8-mdd8q" Dec 13 06:51:38 crc kubenswrapper[4971]: E1213 06:51:38.392261 4971 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-12-13 06:51:38.892244561 +0000 UTC m=+155.496654009 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-mdd8q" (UID: "cba893c6-ff99-466a-9db0-a62e7a87a334") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 13 06:51:38 crc kubenswrapper[4971]: I1213 06:51:38.394393 4971 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-zqt2t" podStartSLOduration=133.394347515 podStartE2EDuration="2m13.394347515s" podCreationTimestamp="2025-12-13 06:49:25 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-13 06:51:38.337869367 +0000 UTC m=+154.942278815" watchObservedRunningTime="2025-12-13 06:51:38.394347515 +0000 UTC m=+154.998756963" Dec 13 06:51:38 crc kubenswrapper[4971]: W1213 06:51:38.425187 4971 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod637e887c_db01_4b47_9f17_2d61b731a118.slice/crio-e07ca3ca1803b4f976ff1b8e72d60c1287ffa56576b40f065178b4408595e350 WatchSource:0}: Error finding container e07ca3ca1803b4f976ff1b8e72d60c1287ffa56576b40f065178b4408595e350: Status 404 returned error can't find the container with id e07ca3ca1803b4f976ff1b8e72d60c1287ffa56576b40f065178b4408595e350 Dec 13 06:51:38 crc kubenswrapper[4971]: I1213 06:51:38.504563 4971 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 13 06:51:38 crc kubenswrapper[4971]: E1213 06:51:38.505730 4971 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-13 06:51:39.005711774 +0000 UTC m=+155.610121222 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 13 06:51:38 crc kubenswrapper[4971]: I1213 06:51:38.522428 4971 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-ingress/router-default-5444994796-mzhp6" podStartSLOduration=133.522405705 podStartE2EDuration="2m13.522405705s" podCreationTimestamp="2025-12-13 06:49:25 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-13 06:51:38.409681123 +0000 UTC m=+155.014090591" watchObservedRunningTime="2025-12-13 06:51:38.522405705 +0000 UTC m=+155.126815153" Dec 13 06:51:38 crc kubenswrapper[4971]: I1213 06:51:38.533988 4971 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-controller-manager/controller-manager-879f6c89f-xtxr8" podStartSLOduration=133.533961078 podStartE2EDuration="2m13.533961078s" podCreationTimestamp="2025-12-13 06:49:25 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-13 06:51:38.490148709 +0000 UTC m=+155.094558167" watchObservedRunningTime="2025-12-13 06:51:38.533961078 +0000 UTC m=+155.138370546" Dec 13 06:51:38 crc kubenswrapper[4971]: I1213 06:51:38.565993 4971 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-gtrwg" podStartSLOduration=133.565968018 podStartE2EDuration="2m13.565968018s" podCreationTimestamp="2025-12-13 06:49:25 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-13 06:51:38.543568301 +0000 UTC m=+155.147977749" watchObservedRunningTime="2025-12-13 06:51:38.565968018 +0000 UTC m=+155.170377466" Dec 13 06:51:38 crc kubenswrapper[4971]: I1213 06:51:38.609268 4971 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-mdd8q\" (UID: \"cba893c6-ff99-466a-9db0-a62e7a87a334\") " pod="openshift-image-registry/image-registry-697d97f7c8-mdd8q" Dec 13 06:51:38 crc kubenswrapper[4971]: E1213 06:51:38.609616 4971 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-12-13 06:51:39.109603622 +0000 UTC m=+155.714013070 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-mdd8q" (UID: "cba893c6-ff99-466a-9db0-a62e7a87a334") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 13 06:51:38 crc kubenswrapper[4971]: I1213 06:51:38.661668 4971 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-xfvfg" podStartSLOduration=133.661649439 podStartE2EDuration="2m13.661649439s" podCreationTimestamp="2025-12-13 06:49:25 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-13 06:51:38.632484901 +0000 UTC m=+155.236894369" watchObservedRunningTime="2025-12-13 06:51:38.661649439 +0000 UTC m=+155.266058887" Dec 13 06:51:38 crc kubenswrapper[4971]: I1213 06:51:38.698019 4971 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/marketplace-operator-79b997595-j9zqf" podStartSLOduration=133.697911726 podStartE2EDuration="2m13.697911726s" podCreationTimestamp="2025-12-13 06:49:25 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-13 06:51:38.663603688 +0000 UTC m=+155.268013156" watchObservedRunningTime="2025-12-13 06:51:38.697911726 +0000 UTC m=+155.302321174" Dec 13 06:51:38 crc kubenswrapper[4971]: I1213 06:51:38.735446 4971 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-9d86n" podStartSLOduration=132.735401535 podStartE2EDuration="2m12.735401535s" podCreationTimestamp="2025-12-13 06:49:26 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-13 06:51:38.729067355 +0000 UTC m=+155.333476803" watchObservedRunningTime="2025-12-13 06:51:38.735401535 +0000 UTC m=+155.339810993" Dec 13 06:51:38 crc kubenswrapper[4971]: I1213 06:51:38.736854 4971 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-machine-api/control-plane-machine-set-operator-78cbb6b69f-bgdbk" podStartSLOduration=133.736827281 podStartE2EDuration="2m13.736827281s" podCreationTimestamp="2025-12-13 06:49:25 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-13 06:51:38.691371551 +0000 UTC m=+155.295781009" watchObservedRunningTime="2025-12-13 06:51:38.736827281 +0000 UTC m=+155.341236729" Dec 13 06:51:38 crc kubenswrapper[4971]: I1213 06:51:38.740695 4971 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 13 06:51:38 crc kubenswrapper[4971]: E1213 06:51:38.740783 4971 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-13 06:51:39.24076343 +0000 UTC m=+155.845172878 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 13 06:51:38 crc kubenswrapper[4971]: I1213 06:51:38.743590 4971 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-mdd8q\" (UID: \"cba893c6-ff99-466a-9db0-a62e7a87a334\") " pod="openshift-image-registry/image-registry-697d97f7c8-mdd8q" Dec 13 06:51:38 crc kubenswrapper[4971]: E1213 06:51:38.750951 4971 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-12-13 06:51:39.250922648 +0000 UTC m=+155.855332096 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-mdd8q" (UID: "cba893c6-ff99-466a-9db0-a62e7a87a334") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 13 06:51:38 crc kubenswrapper[4971]: I1213 06:51:38.846191 4971 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 13 06:51:38 crc kubenswrapper[4971]: E1213 06:51:38.846567 4971 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-13 06:51:39.346548467 +0000 UTC m=+155.950957915 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 13 06:51:38 crc kubenswrapper[4971]: I1213 06:51:38.947513 4971 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-mdd8q\" (UID: \"cba893c6-ff99-466a-9db0-a62e7a87a334\") " pod="openshift-image-registry/image-registry-697d97f7c8-mdd8q" Dec 13 06:51:38 crc kubenswrapper[4971]: E1213 06:51:38.948125 4971 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-12-13 06:51:39.448112936 +0000 UTC m=+156.052522384 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-mdd8q" (UID: "cba893c6-ff99-466a-9db0-a62e7a87a334") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 13 06:51:39 crc kubenswrapper[4971]: I1213 06:51:39.049108 4971 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 13 06:51:39 crc kubenswrapper[4971]: E1213 06:51:39.049858 4971 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-13 06:51:39.549836231 +0000 UTC m=+156.154245689 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 13 06:51:39 crc kubenswrapper[4971]: I1213 06:51:39.070931 4971 patch_prober.go:28] interesting pod/router-default-5444994796-mzhp6 container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Dec 13 06:51:39 crc kubenswrapper[4971]: [-]has-synced failed: reason withheld Dec 13 06:51:39 crc kubenswrapper[4971]: [+]process-running ok Dec 13 06:51:39 crc kubenswrapper[4971]: healthz check failed Dec 13 06:51:39 crc kubenswrapper[4971]: I1213 06:51:39.071003 4971 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-mzhp6" podUID="d89e3583-3f66-4259-b6f9-bffe8ddbde12" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Dec 13 06:51:39 crc kubenswrapper[4971]: I1213 06:51:39.150230 4971 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-mdd8q\" (UID: \"cba893c6-ff99-466a-9db0-a62e7a87a334\") " pod="openshift-image-registry/image-registry-697d97f7c8-mdd8q" Dec 13 06:51:39 crc kubenswrapper[4971]: E1213 06:51:39.150599 4971 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-12-13 06:51:39.65058701 +0000 UTC m=+156.254996458 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-mdd8q" (UID: "cba893c6-ff99-466a-9db0-a62e7a87a334") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 13 06:51:39 crc kubenswrapper[4971]: I1213 06:51:39.258025 4971 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 13 06:51:39 crc kubenswrapper[4971]: E1213 06:51:39.258183 4971 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-13 06:51:39.758155872 +0000 UTC m=+156.362565320 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 13 06:51:39 crc kubenswrapper[4971]: I1213 06:51:39.258686 4971 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-mdd8q\" (UID: \"cba893c6-ff99-466a-9db0-a62e7a87a334\") " pod="openshift-image-registry/image-registry-697d97f7c8-mdd8q" Dec 13 06:51:39 crc kubenswrapper[4971]: E1213 06:51:39.259042 4971 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-12-13 06:51:39.759027064 +0000 UTC m=+156.363436512 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-mdd8q" (UID: "cba893c6-ff99-466a-9db0-a62e7a87a334") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 13 06:51:39 crc kubenswrapper[4971]: I1213 06:51:39.337905 4971 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29426805-fxxrz" event={"ID":"c1575ff9-1c53-4732-bc48-252f9f270597","Type":"ContainerStarted","Data":"7279bfe22b885129c7c9a8d98bc482013cb0cdf21dfdc5a026eda208800095de"} Dec 13 06:51:39 crc kubenswrapper[4971]: I1213 06:51:39.363043 4971 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 13 06:51:39 crc kubenswrapper[4971]: E1213 06:51:39.363146 4971 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-13 06:51:39.863124508 +0000 UTC m=+156.467533956 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 13 06:51:39 crc kubenswrapper[4971]: I1213 06:51:39.363261 4971 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-mdd8q\" (UID: \"cba893c6-ff99-466a-9db0-a62e7a87a334\") " pod="openshift-image-registry/image-registry-697d97f7c8-mdd8q" Dec 13 06:51:39 crc kubenswrapper[4971]: E1213 06:51:39.364400 4971 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-12-13 06:51:39.86439245 +0000 UTC m=+156.468801898 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-mdd8q" (UID: "cba893c6-ff99-466a-9db0-a62e7a87a334") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 13 06:51:39 crc kubenswrapper[4971]: I1213 06:51:39.370063 4971 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ingress-canary/ingress-canary-ffhlq" event={"ID":"4d847949-530c-4cbc-b11d-c1c8d80d3db7","Type":"ContainerStarted","Data":"4884cd7b3616a12b3dacb046dc8baa93087857ec0fedc3a041a52eef783ced64"} Dec 13 06:51:39 crc kubenswrapper[4971]: I1213 06:51:39.382797 4971 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-server-xdxjz" event={"ID":"1c4c5572-81b1-4270-b519-b077f9682f49","Type":"ContainerStarted","Data":"d4d255bf855e87d93e0d6121637847a3d88d2f4b05f87d5a6817879da161aea5"} Dec 13 06:51:39 crc kubenswrapper[4971]: I1213 06:51:39.404790 4971 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-authentication-operator/authentication-operator-69f744f599-wjdd5" event={"ID":"c0ca0ce3-929c-48d6-a3f7-6d334e3c3709","Type":"ContainerStarted","Data":"256518f2103ad634202ad20df2404927d75add2283125b14cdd66e12ec9d117d"} Dec 13 06:51:39 crc kubenswrapper[4971]: I1213 06:51:39.404838 4971 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-authentication-operator/authentication-operator-69f744f599-wjdd5" event={"ID":"c0ca0ce3-929c-48d6-a3f7-6d334e3c3709","Type":"ContainerStarted","Data":"6437d756c9a4cba751bba9bf75eab97a3f9c092dbf646f8f5fb381b0f591fb14"} Dec 13 06:51:39 crc kubenswrapper[4971]: I1213 06:51:39.406476 4971 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-pxjqw" event={"ID":"ea24f72e-df9d-418f-aaef-dfc0573eface","Type":"ContainerStarted","Data":"e3a14beb20513b57a8c72dd7d0b5393646970d2a497fade19375e13d63d95fdb"} Dec 13 06:51:39 crc kubenswrapper[4971]: I1213 06:51:39.407160 4971 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-pxjqw" Dec 13 06:51:39 crc kubenswrapper[4971]: I1213 06:51:39.411390 4971 patch_prober.go:28] interesting pod/olm-operator-6b444d44fb-pxjqw container/olm-operator namespace/openshift-operator-lifecycle-manager: Readiness probe status=failure output="Get \"https://10.217.0.30:8443/healthz\": dial tcp 10.217.0.30:8443: connect: connection refused" start-of-body= Dec 13 06:51:39 crc kubenswrapper[4971]: I1213 06:51:39.411456 4971 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-pxjqw" podUID="ea24f72e-df9d-418f-aaef-dfc0573eface" containerName="olm-operator" probeResult="failure" output="Get \"https://10.217.0.30:8443/healthz\": dial tcp 10.217.0.30:8443: connect: connection refused" Dec 13 06:51:39 crc kubenswrapper[4971]: I1213 06:51:39.449626 4971 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-config-operator/openshift-config-operator-7777fb866f-vmdsp" event={"ID":"874901b0-9650-4029-8d21-74441f40e201","Type":"ContainerStarted","Data":"051786f84a1d2106da36c5d04c89a1834299b88129f334ad7a8668e28c1b8d33"} Dec 13 06:51:39 crc kubenswrapper[4971]: I1213 06:51:39.450326 4971 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-config-operator/openshift-config-operator-7777fb866f-vmdsp" Dec 13 06:51:39 crc kubenswrapper[4971]: I1213 06:51:39.456404 4971 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-6blxl" event={"ID":"6a3649bc-2d2e-4897-8977-99c08aa77e01","Type":"ContainerStarted","Data":"6793de8835c3f4faca9114e12d48909aa65599c2ccbdfe7ba8dd65f908f4811d"} Dec 13 06:51:39 crc kubenswrapper[4971]: I1213 06:51:39.456445 4971 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-6blxl" event={"ID":"6a3649bc-2d2e-4897-8977-99c08aa77e01","Type":"ContainerStarted","Data":"f91351dfad5e336053c8b82f4524daa30eeac3d89e5c03b3c4f4833129a3a298"} Dec 13 06:51:39 crc kubenswrapper[4971]: I1213 06:51:39.474138 4971 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 13 06:51:39 crc kubenswrapper[4971]: E1213 06:51:39.475413 4971 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-13 06:51:39.974614319 +0000 UTC m=+156.579023777 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 13 06:51:39 crc kubenswrapper[4971]: I1213 06:51:39.475708 4971 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-mdd8q\" (UID: \"cba893c6-ff99-466a-9db0-a62e7a87a334\") " pod="openshift-image-registry/image-registry-697d97f7c8-mdd8q" Dec 13 06:51:39 crc kubenswrapper[4971]: E1213 06:51:39.478053 4971 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-12-13 06:51:39.978039336 +0000 UTC m=+156.582448784 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-mdd8q" (UID: "cba893c6-ff99-466a-9db0-a62e7a87a334") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 13 06:51:39 crc kubenswrapper[4971]: I1213 06:51:39.484668 4971 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-etcd-operator/etcd-operator-b45778765-6xfwq" event={"ID":"56c62b80-0a33-4156-b52b-cb54c98a4343","Type":"ContainerStarted","Data":"3ba1abf791c3d7266ed9ffba5beacd1612d97fa86c76c6662d9f954f5b154ddb"} Dec 13 06:51:39 crc kubenswrapper[4971]: I1213 06:51:39.508023 4971 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-operator-lifecycle-manager/collect-profiles-29426805-fxxrz" podStartSLOduration=134.508007103 podStartE2EDuration="2m14.508007103s" podCreationTimestamp="2025-12-13 06:49:25 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-13 06:51:39.446907148 +0000 UTC m=+156.051316596" watchObservedRunningTime="2025-12-13 06:51:39.508007103 +0000 UTC m=+156.112416551" Dec 13 06:51:39 crc kubenswrapper[4971]: I1213 06:51:39.508708 4971 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-l4rzp" event={"ID":"3052c2e3-f9bf-442b-b153-d60eb86f9c3a","Type":"ContainerStarted","Data":"bad4177ef667281b9fc37fe11538c5d3dbadd0db956ccb5365b2cd29e0764a81"} Dec 13 06:51:39 crc kubenswrapper[4971]: I1213 06:51:39.521561 4971 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-9d86n" event={"ID":"2cf6e363-8511-4c79-9181-79dba726e5af","Type":"ContainerStarted","Data":"6ccbca076f1fdb45dc7a3f3efdf6a363f8806ecdef5e1047dc12136783d5c199"} Dec 13 06:51:39 crc kubenswrapper[4971]: I1213 06:51:39.531800 4971 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-9d86n" Dec 13 06:51:39 crc kubenswrapper[4971]: I1213 06:51:39.580218 4971 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 13 06:51:39 crc kubenswrapper[4971]: E1213 06:51:39.581201 4971 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-13 06:51:40.081184436 +0000 UTC m=+156.685593884 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 13 06:51:39 crc kubenswrapper[4971]: I1213 06:51:39.593727 4971 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-service-ca-operator/service-ca-operator-777779d784-drq5g" event={"ID":"49f3dc41-5142-4cdb-9bb3-6d7248228535","Type":"ContainerStarted","Data":"21a517c93a2e7c0f1cab1e5cbb2df071ffe9a1e17bb01d93389f0050a5f933e2"} Dec 13 06:51:39 crc kubenswrapper[4971]: I1213 06:51:39.593767 4971 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-service-ca-operator/service-ca-operator-777779d784-drq5g" event={"ID":"49f3dc41-5142-4cdb-9bb3-6d7248228535","Type":"ContainerStarted","Data":"830bda5db05b1e5a6df6688570fc467a7902d9477bd73a95e505db7eb33f6f77"} Dec 13 06:51:39 crc kubenswrapper[4971]: I1213 06:51:39.609947 4971 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-mb2q7" event={"ID":"8e7aefba-ae2e-4069-892e-544c85b9af15","Type":"ContainerStarted","Data":"d9a36381d45913e9a366015a3560b4949b545324d9af53031df1aa53f10a9687"} Dec 13 06:51:39 crc kubenswrapper[4971]: I1213 06:51:39.616882 4971 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-storage-version-migrator/migrator-59844c95c7-z2p8z" event={"ID":"5764a8c4-18e5-4025-8d36-15a7b01e9169","Type":"ContainerStarted","Data":"b31536e449a22439e8ff7269ef02982dd58f045048efe1fdd1c7a3ff1e424cca"} Dec 13 06:51:39 crc kubenswrapper[4971]: I1213 06:51:39.630099 4971 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-dns-operator/dns-operator-744455d44c-b999k" event={"ID":"ad158968-e110-4b4c-9f42-da9c08274dab","Type":"ContainerStarted","Data":"13e690bf71b421be9a360908de8897f0524d1895a8d790fcdc85d5abacaa28a4"} Dec 13 06:51:39 crc kubenswrapper[4971]: I1213 06:51:39.636465 4971 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-machine-config-operator/machine-config-server-xdxjz" podStartSLOduration=6.636446454 podStartE2EDuration="6.636446454s" podCreationTimestamp="2025-12-13 06:51:33 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-13 06:51:39.553850344 +0000 UTC m=+156.158259792" watchObservedRunningTime="2025-12-13 06:51:39.636446454 +0000 UTC m=+156.240855902" Dec 13 06:51:39 crc kubenswrapper[4971]: I1213 06:51:39.652329 4971 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-operator-74547568cd-4cn2z" event={"ID":"874a112c-101d-4a74-a674-d9afe62b6dc8","Type":"ContainerStarted","Data":"dd8b78c5064743fb896ac677dc66013ebce268ebafd9f87955da33a1c7507302"} Dec 13 06:51:39 crc kubenswrapper[4971]: I1213 06:51:39.681989 4971 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-mdd8q\" (UID: \"cba893c6-ff99-466a-9db0-a62e7a87a334\") " pod="openshift-image-registry/image-registry-697d97f7c8-mdd8q" Dec 13 06:51:39 crc kubenswrapper[4971]: I1213 06:51:39.682602 4971 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-controller-84d6567774-kf7sp" event={"ID":"2f87c67a-fcf1-46e3-9bbd-9bf03ee55da3","Type":"ContainerStarted","Data":"724d4ee3e3446dfd9e021f58fa7cde97cccbd52efb48e38cf898249362388300"} Dec 13 06:51:39 crc kubenswrapper[4971]: E1213 06:51:39.683444 4971 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-12-13 06:51:40.183429493 +0000 UTC m=+156.787838941 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-mdd8q" (UID: "cba893c6-ff99-466a-9db0-a62e7a87a334") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 13 06:51:39 crc kubenswrapper[4971]: I1213 06:51:39.702804 4971 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-pxjqw" podStartSLOduration=134.702790572 podStartE2EDuration="2m14.702790572s" podCreationTimestamp="2025-12-13 06:49:25 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-13 06:51:39.645887463 +0000 UTC m=+156.250296921" watchObservedRunningTime="2025-12-13 06:51:39.702790572 +0000 UTC m=+156.307200020" Dec 13 06:51:39 crc kubenswrapper[4971]: I1213 06:51:39.709833 4971 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-lnrkm" event={"ID":"87f8bc63-d8c1-415f-975f-b8418034197b","Type":"ContainerStarted","Data":"41aca581416eec7c9faeb5722769713273f09449d8841c874c68969c01dc4f28"} Dec 13 06:51:39 crc kubenswrapper[4971]: I1213 06:51:39.710958 4971 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-lnrkm" Dec 13 06:51:39 crc kubenswrapper[4971]: I1213 06:51:39.713032 4971 patch_prober.go:28] interesting pod/packageserver-d55dfcdfc-lnrkm container/packageserver namespace/openshift-operator-lifecycle-manager: Readiness probe status=failure output="Get \"https://10.217.0.29:5443/healthz\": dial tcp 10.217.0.29:5443: connect: connection refused" start-of-body= Dec 13 06:51:39 crc kubenswrapper[4971]: I1213 06:51:39.713070 4971 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-lnrkm" podUID="87f8bc63-d8c1-415f-975f-b8418034197b" containerName="packageserver" probeResult="failure" output="Get \"https://10.217.0.29:5443/healthz\": dial tcp 10.217.0.29:5443: connect: connection refused" Dec 13 06:51:39 crc kubenswrapper[4971]: I1213 06:51:39.745080 4971 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-api/control-plane-machine-set-operator-78cbb6b69f-bgdbk" event={"ID":"930fe38c-5ff7-45a8-a292-cd69d07e86f7","Type":"ContainerStarted","Data":"ca97f6df84f2bb9c2c527ad84f854eee8e1ab1ddc029a01f818b246add6d886a"} Dec 13 06:51:39 crc kubenswrapper[4971]: I1213 06:51:39.764166 4971 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-authentication-operator/authentication-operator-69f744f599-wjdd5" podStartSLOduration=134.764153055 podStartE2EDuration="2m14.764153055s" podCreationTimestamp="2025-12-13 06:49:25 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-13 06:51:39.703257294 +0000 UTC m=+156.307666742" watchObservedRunningTime="2025-12-13 06:51:39.764153055 +0000 UTC m=+156.368562503" Dec 13 06:51:39 crc kubenswrapper[4971]: I1213 06:51:39.790608 4971 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 13 06:51:39 crc kubenswrapper[4971]: E1213 06:51:39.791861 4971 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-13 06:51:40.291827525 +0000 UTC m=+156.896236973 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 13 06:51:39 crc kubenswrapper[4971]: I1213 06:51:39.825370 4971 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-g8vfp" event={"ID":"479b25e6-6618-4123-a913-ca50d6c4ba28","Type":"ContainerStarted","Data":"401d6b22daa0a3c63d4233898e103b84d9bd5f3de15a32f1d9556240d34c592b"} Dec 13 06:51:39 crc kubenswrapper[4971]: I1213 06:51:39.825808 4971 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-g8vfp" Dec 13 06:51:39 crc kubenswrapper[4971]: I1213 06:51:39.857183 4971 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-service-ca-operator/service-ca-operator-777779d784-drq5g" podStartSLOduration=133.857161219 podStartE2EDuration="2m13.857161219s" podCreationTimestamp="2025-12-13 06:49:26 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-13 06:51:39.763806776 +0000 UTC m=+156.368216234" watchObservedRunningTime="2025-12-13 06:51:39.857161219 +0000 UTC m=+156.461570667" Dec 13 06:51:39 crc kubenswrapper[4971]: I1213 06:51:39.859195 4971 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-console-operator/console-operator-58897d9998-kdfcd" event={"ID":"94469d04-47fd-4bd8-8258-9e874767f250","Type":"ContainerStarted","Data":"4d137a816e6017817cc140b360887550ddfde858e153bee6cf222c8ad8f2f9e6"} Dec 13 06:51:39 crc kubenswrapper[4971]: I1213 06:51:39.870072 4971 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-console-operator/console-operator-58897d9998-kdfcd" Dec 13 06:51:39 crc kubenswrapper[4971]: I1213 06:51:39.881484 4971 patch_prober.go:28] interesting pod/console-operator-58897d9998-kdfcd container/console-operator namespace/openshift-console-operator: Readiness probe status=failure output="Get \"https://10.217.0.35:8443/readyz\": dial tcp 10.217.0.35:8443: connect: connection refused" start-of-body= Dec 13 06:51:39 crc kubenswrapper[4971]: I1213 06:51:39.881546 4971 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-console-operator/console-operator-58897d9998-kdfcd" podUID="94469d04-47fd-4bd8-8258-9e874767f250" containerName="console-operator" probeResult="failure" output="Get \"https://10.217.0.35:8443/readyz\": dial tcp 10.217.0.35:8443: connect: connection refused" Dec 13 06:51:39 crc kubenswrapper[4971]: I1213 06:51:39.895062 4971 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-mdd8q\" (UID: \"cba893c6-ff99-466a-9db0-a62e7a87a334\") " pod="openshift-image-registry/image-registry-697d97f7c8-mdd8q" Dec 13 06:51:39 crc kubenswrapper[4971]: E1213 06:51:39.896675 4971 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-12-13 06:51:40.396655768 +0000 UTC m=+157.001065216 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-mdd8q" (UID: "cba893c6-ff99-466a-9db0-a62e7a87a334") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 13 06:51:39 crc kubenswrapper[4971]: I1213 06:51:39.907595 4971 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-mb2q7" podStartSLOduration=134.907574734 podStartE2EDuration="2m14.907574734s" podCreationTimestamp="2025-12-13 06:49:25 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-13 06:51:39.859800146 +0000 UTC m=+156.464209594" watchObservedRunningTime="2025-12-13 06:51:39.907574734 +0000 UTC m=+156.511984192" Dec 13 06:51:39 crc kubenswrapper[4971]: I1213 06:51:39.913795 4971 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ingress-operator/ingress-operator-5b745b69d9-pkmb5" event={"ID":"22f3a4f0-1686-4087-bcec-e48f2015ffe2","Type":"ContainerStarted","Data":"ba0164803328dbf682a2b7cb921567f2182629060e391934f3cf14834c276d4c"} Dec 13 06:51:39 crc kubenswrapper[4971]: I1213 06:51:39.949643 4971 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-etcd-operator/etcd-operator-b45778765-6xfwq" podStartSLOduration=134.949625528 podStartE2EDuration="2m14.949625528s" podCreationTimestamp="2025-12-13 06:49:25 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-13 06:51:39.903697866 +0000 UTC m=+156.508107314" watchObservedRunningTime="2025-12-13 06:51:39.949625528 +0000 UTC m=+156.554034976" Dec 13 06:51:39 crc kubenswrapper[4971]: I1213 06:51:39.993707 4971 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-dns/dns-default-dntmz" event={"ID":"637e887c-db01-4b47-9f17-2d61b731a118","Type":"ContainerStarted","Data":"e07ca3ca1803b4f976ff1b8e72d60c1287ffa56576b40f065178b4408595e350"} Dec 13 06:51:39 crc kubenswrapper[4971]: I1213 06:51:39.995866 4971 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 13 06:51:39 crc kubenswrapper[4971]: E1213 06:51:39.996838 4971 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-13 06:51:40.496824692 +0000 UTC m=+157.101234140 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 13 06:51:39 crc kubenswrapper[4971]: I1213 06:51:39.997287 4971 patch_prober.go:28] interesting pod/marketplace-operator-79b997595-j9zqf container/marketplace-operator namespace/openshift-marketplace: Readiness probe status=failure output="Get \"http://10.217.0.17:8080/healthz\": dial tcp 10.217.0.17:8080: connect: connection refused" start-of-body= Dec 13 06:51:39 crc kubenswrapper[4971]: I1213 06:51:39.997428 4971 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-marketplace/marketplace-operator-79b997595-j9zqf" podUID="f14c8a94-8589-4059-bddf-329452355fab" containerName="marketplace-operator" probeResult="failure" output="Get \"http://10.217.0.17:8080/healthz\": dial tcp 10.217.0.17:8080: connect: connection refused" Dec 13 06:51:39 crc kubenswrapper[4971]: I1213 06:51:39.997564 4971 patch_prober.go:28] interesting pod/downloads-7954f5f757-dqk9z container/download-server namespace/openshift-console: Readiness probe status=failure output="Get \"http://10.217.0.20:8080/\": dial tcp 10.217.0.20:8080: connect: connection refused" start-of-body= Dec 13 06:51:39 crc kubenswrapper[4971]: I1213 06:51:39.997671 4971 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-console/downloads-7954f5f757-dqk9z" podUID="ee2e94f4-cadb-4b17-8853-a1292cb02404" containerName="download-server" probeResult="failure" output="Get \"http://10.217.0.20:8080/\": dial tcp 10.217.0.20:8080: connect: connection refused" Dec 13 06:51:40 crc kubenswrapper[4971]: I1213 06:51:40.011562 4971 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-machine-config-operator/machine-config-operator-74547568cd-4cn2z" podStartSLOduration=135.011548765 podStartE2EDuration="2m15.011548765s" podCreationTimestamp="2025-12-13 06:49:25 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-13 06:51:39.951664369 +0000 UTC m=+156.556073817" watchObservedRunningTime="2025-12-13 06:51:40.011548765 +0000 UTC m=+156.615958213" Dec 13 06:51:40 crc kubenswrapper[4971]: I1213 06:51:40.073700 4971 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-config-operator/openshift-config-operator-7777fb866f-vmdsp" podStartSLOduration=135.073680767 podStartE2EDuration="2m15.073680767s" podCreationTimestamp="2025-12-13 06:49:25 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-13 06:51:40.073474432 +0000 UTC m=+156.677883880" watchObservedRunningTime="2025-12-13 06:51:40.073680767 +0000 UTC m=+156.678090215" Dec 13 06:51:40 crc kubenswrapper[4971]: I1213 06:51:40.083557 4971 patch_prober.go:28] interesting pod/router-default-5444994796-mzhp6 container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Dec 13 06:51:40 crc kubenswrapper[4971]: [-]has-synced failed: reason withheld Dec 13 06:51:40 crc kubenswrapper[4971]: [+]process-running ok Dec 13 06:51:40 crc kubenswrapper[4971]: healthz check failed Dec 13 06:51:40 crc kubenswrapper[4971]: I1213 06:51:40.083600 4971 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-mzhp6" podUID="d89e3583-3f66-4259-b6f9-bffe8ddbde12" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Dec 13 06:51:40 crc kubenswrapper[4971]: I1213 06:51:40.101001 4971 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-mdd8q\" (UID: \"cba893c6-ff99-466a-9db0-a62e7a87a334\") " pod="openshift-image-registry/image-registry-697d97f7c8-mdd8q" Dec 13 06:51:40 crc kubenswrapper[4971]: E1213 06:51:40.103246 4971 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-12-13 06:51:40.603232934 +0000 UTC m=+157.207642382 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-mdd8q" (UID: "cba893c6-ff99-466a-9db0-a62e7a87a334") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 13 06:51:40 crc kubenswrapper[4971]: I1213 06:51:40.116601 4971 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-machine-config-operator/machine-config-controller-84d6567774-kf7sp" podStartSLOduration=135.116587263 podStartE2EDuration="2m15.116587263s" podCreationTimestamp="2025-12-13 06:49:25 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-13 06:51:40.11647443 +0000 UTC m=+156.720883878" watchObservedRunningTime="2025-12-13 06:51:40.116587263 +0000 UTC m=+156.720996711" Dec 13 06:51:40 crc kubenswrapper[4971]: I1213 06:51:40.192071 4971 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-6blxl" podStartSLOduration=135.192056152 podStartE2EDuration="2m15.192056152s" podCreationTimestamp="2025-12-13 06:49:25 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-13 06:51:40.160435172 +0000 UTC m=+156.764844620" watchObservedRunningTime="2025-12-13 06:51:40.192056152 +0000 UTC m=+156.796465600" Dec 13 06:51:40 crc kubenswrapper[4971]: I1213 06:51:40.193097 4971 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-multus/multus-admission-controller-857f4d67dd-z68ks" podStartSLOduration=135.193093839 podStartE2EDuration="2m15.193093839s" podCreationTimestamp="2025-12-13 06:49:25 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-13 06:51:40.19118183 +0000 UTC m=+156.795591278" watchObservedRunningTime="2025-12-13 06:51:40.193093839 +0000 UTC m=+156.797503287" Dec 13 06:51:40 crc kubenswrapper[4971]: I1213 06:51:40.207132 4971 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 13 06:51:40 crc kubenswrapper[4971]: E1213 06:51:40.207462 4971 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-13 06:51:40.707443092 +0000 UTC m=+157.311852550 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 13 06:51:40 crc kubenswrapper[4971]: I1213 06:51:40.286724 4971 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-lnrkm" podStartSLOduration=135.286705048 podStartE2EDuration="2m15.286705048s" podCreationTimestamp="2025-12-13 06:49:25 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-13 06:51:40.248953982 +0000 UTC m=+156.853363430" watchObservedRunningTime="2025-12-13 06:51:40.286705048 +0000 UTC m=+156.891114496" Dec 13 06:51:40 crc kubenswrapper[4971]: I1213 06:51:40.309114 4971 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-mdd8q\" (UID: \"cba893c6-ff99-466a-9db0-a62e7a87a334\") " pod="openshift-image-registry/image-registry-697d97f7c8-mdd8q" Dec 13 06:51:40 crc kubenswrapper[4971]: E1213 06:51:40.309457 4971 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-12-13 06:51:40.809443893 +0000 UTC m=+157.413853341 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-mdd8q" (UID: "cba893c6-ff99-466a-9db0-a62e7a87a334") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 13 06:51:40 crc kubenswrapper[4971]: I1213 06:51:40.331415 4971 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-ingress-operator/ingress-operator-5b745b69d9-pkmb5" podStartSLOduration=135.331395548 podStartE2EDuration="2m15.331395548s" podCreationTimestamp="2025-12-13 06:49:25 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-13 06:51:40.285176648 +0000 UTC m=+156.889586106" watchObservedRunningTime="2025-12-13 06:51:40.331395548 +0000 UTC m=+156.935804996" Dec 13 06:51:40 crc kubenswrapper[4971]: I1213 06:51:40.332961 4971 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-console-operator/console-operator-58897d9998-kdfcd" podStartSLOduration=135.332951028 podStartE2EDuration="2m15.332951028s" podCreationTimestamp="2025-12-13 06:49:25 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-13 06:51:40.329799697 +0000 UTC m=+156.934209145" watchObservedRunningTime="2025-12-13 06:51:40.332951028 +0000 UTC m=+156.937360476" Dec 13 06:51:40 crc kubenswrapper[4971]: I1213 06:51:40.382580 4971 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-mb2q7" Dec 13 06:51:40 crc kubenswrapper[4971]: I1213 06:51:40.383121 4971 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-mb2q7" Dec 13 06:51:40 crc kubenswrapper[4971]: I1213 06:51:40.396153 4971 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-g8vfp" podStartSLOduration=135.396135836 podStartE2EDuration="2m15.396135836s" podCreationTimestamp="2025-12-13 06:49:25 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-13 06:51:40.392086414 +0000 UTC m=+156.996495862" watchObservedRunningTime="2025-12-13 06:51:40.396135836 +0000 UTC m=+157.000545284" Dec 13 06:51:40 crc kubenswrapper[4971]: I1213 06:51:40.399267 4971 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-mb2q7" Dec 13 06:51:40 crc kubenswrapper[4971]: I1213 06:51:40.409948 4971 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 13 06:51:40 crc kubenswrapper[4971]: E1213 06:51:40.410231 4971 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-13 06:51:40.910212232 +0000 UTC m=+157.514621680 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 13 06:51:40 crc kubenswrapper[4971]: I1213 06:51:40.515367 4971 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-mdd8q\" (UID: \"cba893c6-ff99-466a-9db0-a62e7a87a334\") " pod="openshift-image-registry/image-registry-697d97f7c8-mdd8q" Dec 13 06:51:40 crc kubenswrapper[4971]: E1213 06:51:40.515856 4971 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-12-13 06:51:41.015841215 +0000 UTC m=+157.620250663 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-mdd8q" (UID: "cba893c6-ff99-466a-9db0-a62e7a87a334") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 13 06:51:40 crc kubenswrapper[4971]: I1213 06:51:40.617131 4971 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 13 06:51:40 crc kubenswrapper[4971]: E1213 06:51:40.617397 4971 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-13 06:51:41.117362794 +0000 UTC m=+157.721772252 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 13 06:51:40 crc kubenswrapper[4971]: I1213 06:51:40.719291 4971 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-mdd8q\" (UID: \"cba893c6-ff99-466a-9db0-a62e7a87a334\") " pod="openshift-image-registry/image-registry-697d97f7c8-mdd8q" Dec 13 06:51:40 crc kubenswrapper[4971]: E1213 06:51:40.719777 4971 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-12-13 06:51:41.219756445 +0000 UTC m=+157.824165893 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-mdd8q" (UID: "cba893c6-ff99-466a-9db0-a62e7a87a334") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 13 06:51:40 crc kubenswrapper[4971]: I1213 06:51:40.820071 4971 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 13 06:51:40 crc kubenswrapper[4971]: E1213 06:51:40.820306 4971 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-13 06:51:41.320273728 +0000 UTC m=+157.924683186 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 13 06:51:40 crc kubenswrapper[4971]: I1213 06:51:40.820510 4971 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-mdd8q\" (UID: \"cba893c6-ff99-466a-9db0-a62e7a87a334\") " pod="openshift-image-registry/image-registry-697d97f7c8-mdd8q" Dec 13 06:51:40 crc kubenswrapper[4971]: E1213 06:51:40.820855 4971 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-12-13 06:51:41.320840162 +0000 UTC m=+157.925249610 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-mdd8q" (UID: "cba893c6-ff99-466a-9db0-a62e7a87a334") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 13 06:51:40 crc kubenswrapper[4971]: I1213 06:51:40.921438 4971 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 13 06:51:40 crc kubenswrapper[4971]: E1213 06:51:40.921666 4971 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-13 06:51:41.421635783 +0000 UTC m=+158.026045231 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 13 06:51:40 crc kubenswrapper[4971]: I1213 06:51:40.921747 4971 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-mdd8q\" (UID: \"cba893c6-ff99-466a-9db0-a62e7a87a334\") " pod="openshift-image-registry/image-registry-697d97f7c8-mdd8q" Dec 13 06:51:40 crc kubenswrapper[4971]: E1213 06:51:40.922073 4971 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-12-13 06:51:41.422062833 +0000 UTC m=+158.026472281 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-mdd8q" (UID: "cba893c6-ff99-466a-9db0-a62e7a87a334") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 13 06:51:41 crc kubenswrapper[4971]: I1213 06:51:41.015222 4971 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-dns-operator/dns-operator-744455d44c-b999k" event={"ID":"ad158968-e110-4b4c-9f42-da9c08274dab","Type":"ContainerStarted","Data":"41fbe28f760ea2fe5ee7ae7468dfd301300b78d220cf90a63cc443bef8d46c20"} Dec 13 06:51:41 crc kubenswrapper[4971]: I1213 06:51:41.022886 4971 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 13 06:51:41 crc kubenswrapper[4971]: E1213 06:51:41.023028 4971 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-13 06:51:41.523006147 +0000 UTC m=+158.127415605 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 13 06:51:41 crc kubenswrapper[4971]: I1213 06:51:41.023077 4971 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-mdd8q\" (UID: \"cba893c6-ff99-466a-9db0-a62e7a87a334\") " pod="openshift-image-registry/image-registry-697d97f7c8-mdd8q" Dec 13 06:51:41 crc kubenswrapper[4971]: E1213 06:51:41.023405 4971 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-12-13 06:51:41.523395318 +0000 UTC m=+158.127804766 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-mdd8q" (UID: "cba893c6-ff99-466a-9db0-a62e7a87a334") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 13 06:51:41 crc kubenswrapper[4971]: I1213 06:51:41.028223 4971 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-g8vfp" event={"ID":"479b25e6-6618-4123-a913-ca50d6c4ba28","Type":"ContainerStarted","Data":"9fe7f104167a764e4c112e8d2345c621c3367fcc1745f1069739598e1b9c0393"} Dec 13 06:51:41 crc kubenswrapper[4971]: I1213 06:51:41.036773 4971 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-cwwf9" event={"ID":"6e6307eb-6448-4cbb-b2a3-93e340f5b3a4","Type":"ContainerStarted","Data":"5ab2dd0842d845734d3234713c4c4e2999ccee6a32f50cd4f4f85542a6d9f0ba"} Dec 13 06:51:41 crc kubenswrapper[4971]: I1213 06:51:41.043024 4971 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-dns-operator/dns-operator-744455d44c-b999k" podStartSLOduration=136.042986153 podStartE2EDuration="2m16.042986153s" podCreationTimestamp="2025-12-13 06:49:25 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-13 06:51:41.042647864 +0000 UTC m=+157.647057322" watchObservedRunningTime="2025-12-13 06:51:41.042986153 +0000 UTC m=+157.647395621" Dec 13 06:51:41 crc kubenswrapper[4971]: I1213 06:51:41.051978 4971 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-operator-74547568cd-4cn2z" event={"ID":"874a112c-101d-4a74-a674-d9afe62b6dc8","Type":"ContainerStarted","Data":"178e01802250c4d77b91c91e8bfe066065712e5466c2eb72d56fc4d9d6d83607"} Dec 13 06:51:41 crc kubenswrapper[4971]: I1213 06:51:41.066154 4971 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-dns/dns-default-dntmz" event={"ID":"637e887c-db01-4b47-9f17-2d61b731a118","Type":"ContainerStarted","Data":"ffe52ba93d3c11974b5974a4e37326acb85258f267f5ff5114e45a38f5be0bd8"} Dec 13 06:51:41 crc kubenswrapper[4971]: I1213 06:51:41.066201 4971 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-dns/dns-default-dntmz" event={"ID":"637e887c-db01-4b47-9f17-2d61b731a118","Type":"ContainerStarted","Data":"c50ea9c471de0faed46b9e33a489535fec2f307d7fc8381ea6b6fa006c9f878e"} Dec 13 06:51:41 crc kubenswrapper[4971]: I1213 06:51:41.066321 4971 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-dns/dns-default-dntmz" Dec 13 06:51:41 crc kubenswrapper[4971]: I1213 06:51:41.067546 4971 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="hostpath-provisioner/csi-hostpathplugin-7nhhk" event={"ID":"5fd61507-84bf-4ef5-8bd4-731075ef58ef","Type":"ContainerStarted","Data":"475732f6c1147cedb1049a0402b4fd44abfbc6d7153f7c0c755f4e17b24a0af6"} Dec 13 06:51:41 crc kubenswrapper[4971]: I1213 06:51:41.068643 4971 patch_prober.go:28] interesting pod/router-default-5444994796-mzhp6 container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Dec 13 06:51:41 crc kubenswrapper[4971]: [-]has-synced failed: reason withheld Dec 13 06:51:41 crc kubenswrapper[4971]: [+]process-running ok Dec 13 06:51:41 crc kubenswrapper[4971]: healthz check failed Dec 13 06:51:41 crc kubenswrapper[4971]: I1213 06:51:41.068741 4971 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-mzhp6" podUID="d89e3583-3f66-4259-b6f9-bffe8ddbde12" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Dec 13 06:51:41 crc kubenswrapper[4971]: I1213 06:51:41.073178 4971 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-lnrkm" event={"ID":"87f8bc63-d8c1-415f-975f-b8418034197b","Type":"ContainerStarted","Data":"2e3cc346ffde3b0ad24e43fa21a0b97390d6d3010b2019a54bad5c1afaec5b7f"} Dec 13 06:51:41 crc kubenswrapper[4971]: I1213 06:51:41.094284 4971 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-cwwf9" podStartSLOduration=136.09426774 podStartE2EDuration="2m16.09426774s" podCreationTimestamp="2025-12-13 06:49:25 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-13 06:51:41.091114371 +0000 UTC m=+157.695523839" watchObservedRunningTime="2025-12-13 06:51:41.09426774 +0000 UTC m=+157.698677188" Dec 13 06:51:41 crc kubenswrapper[4971]: I1213 06:51:41.106813 4971 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ingress-canary/ingress-canary-ffhlq" event={"ID":"4d847949-530c-4cbc-b11d-c1c8d80d3db7","Type":"ContainerStarted","Data":"e091d8264b24bed5f3ba7826591ca5b5c5c0a201897251e28725e1b3301f6bef"} Dec 13 06:51:41 crc kubenswrapper[4971]: I1213 06:51:41.124096 4971 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 13 06:51:41 crc kubenswrapper[4971]: E1213 06:51:41.125401 4971 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-13 06:51:41.625381908 +0000 UTC m=+158.229791376 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 13 06:51:41 crc kubenswrapper[4971]: I1213 06:51:41.143727 4971 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-admission-controller-857f4d67dd-z68ks" event={"ID":"32c8b95e-38c8-43fb-b1e4-86ac9a27c2f0","Type":"ContainerStarted","Data":"1895b0e06543d075d90e241cec56f07d815d4e5f127ccce3f7684ff7dedb88de"} Dec 13 06:51:41 crc kubenswrapper[4971]: I1213 06:51:41.150936 4971 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-dns/dns-default-dntmz" podStartSLOduration=8.150884874 podStartE2EDuration="8.150884874s" podCreationTimestamp="2025-12-13 06:51:33 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-13 06:51:41.139247538 +0000 UTC m=+157.743656996" watchObservedRunningTime="2025-12-13 06:51:41.150884874 +0000 UTC m=+157.755294342" Dec 13 06:51:41 crc kubenswrapper[4971]: I1213 06:51:41.177003 4971 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-ingress-canary/ingress-canary-ffhlq" podStartSLOduration=8.176979433 podStartE2EDuration="8.176979433s" podCreationTimestamp="2025-12-13 06:51:33 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-13 06:51:41.175402673 +0000 UTC m=+157.779812121" watchObservedRunningTime="2025-12-13 06:51:41.176979433 +0000 UTC m=+157.781388881" Dec 13 06:51:41 crc kubenswrapper[4971]: I1213 06:51:41.180498 4971 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-apiserver/apiserver-76f77b778f-2n4cq" event={"ID":"bec6e53b-3998-4402-b77b-87ff79f31827","Type":"ContainerStarted","Data":"e296213f418445426c172140851afee883cf1a7547e79d8760701ec4aadaef66"} Dec 13 06:51:41 crc kubenswrapper[4971]: I1213 06:51:41.180582 4971 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-apiserver/apiserver-76f77b778f-2n4cq" event={"ID":"bec6e53b-3998-4402-b77b-87ff79f31827","Type":"ContainerStarted","Data":"907696efb698457f991e4e2785b85f0054e079d58f626c73ec1bf29883a4f04f"} Dec 13 06:51:41 crc kubenswrapper[4971]: I1213 06:51:41.205244 4971 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-fz6wk" event={"ID":"4b58bd40-cb76-46ba-9391-a350d830d729","Type":"ContainerStarted","Data":"f16a0b7f8ad0e1ec2f18552c8518249ded3c2fcb3f26eff42a00b1feabd1dd9d"} Dec 13 06:51:41 crc kubenswrapper[4971]: I1213 06:51:41.231535 4971 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-mdd8q\" (UID: \"cba893c6-ff99-466a-9db0-a62e7a87a334\") " pod="openshift-image-registry/image-registry-697d97f7c8-mdd8q" Dec 13 06:51:41 crc kubenswrapper[4971]: E1213 06:51:41.235636 4971 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-12-13 06:51:41.735614717 +0000 UTC m=+158.340024175 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-mdd8q" (UID: "cba893c6-ff99-466a-9db0-a62e7a87a334") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 13 06:51:41 crc kubenswrapper[4971]: I1213 06:51:41.247249 4971 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-storage-version-migrator/migrator-59844c95c7-z2p8z" event={"ID":"5764a8c4-18e5-4025-8d36-15a7b01e9169","Type":"ContainerStarted","Data":"2b868365c9a1e06f0eacccb4a60d4aecdd7aa7a090768ca8579c80a82d293f91"} Dec 13 06:51:41 crc kubenswrapper[4971]: I1213 06:51:41.260559 4971 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-fz6wk" podStartSLOduration=136.260541698 podStartE2EDuration="2m16.260541698s" podCreationTimestamp="2025-12-13 06:49:25 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-13 06:51:41.259690776 +0000 UTC m=+157.864100224" watchObservedRunningTime="2025-12-13 06:51:41.260541698 +0000 UTC m=+157.864951156" Dec 13 06:51:41 crc kubenswrapper[4971]: I1213 06:51:41.261568 4971 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-apiserver/apiserver-76f77b778f-2n4cq" podStartSLOduration=136.261561784 podStartE2EDuration="2m16.261561784s" podCreationTimestamp="2025-12-13 06:49:25 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-13 06:51:41.230168619 +0000 UTC m=+157.834578067" watchObservedRunningTime="2025-12-13 06:51:41.261561784 +0000 UTC m=+157.865971232" Dec 13 06:51:41 crc kubenswrapper[4971]: I1213 06:51:41.288889 4971 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-config-operator/openshift-config-operator-7777fb866f-vmdsp" Dec 13 06:51:41 crc kubenswrapper[4971]: I1213 06:51:41.290953 4971 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-l4rzp" event={"ID":"3052c2e3-f9bf-442b-b153-d60eb86f9c3a","Type":"ContainerStarted","Data":"e6e5d277f9169c0a5814239c58e0bfcbd90a9cadf9fada4ed5504e0e74637464"} Dec 13 06:51:41 crc kubenswrapper[4971]: I1213 06:51:41.291041 4971 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-l4rzp" Dec 13 06:51:41 crc kubenswrapper[4971]: I1213 06:51:41.292331 4971 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-kube-storage-version-migrator/migrator-59844c95c7-z2p8z" podStartSLOduration=136.292318682 podStartE2EDuration="2m16.292318682s" podCreationTimestamp="2025-12-13 06:49:25 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-13 06:51:41.291296466 +0000 UTC m=+157.895705914" watchObservedRunningTime="2025-12-13 06:51:41.292318682 +0000 UTC m=+157.896728140" Dec 13 06:51:41 crc kubenswrapper[4971]: I1213 06:51:41.292896 4971 patch_prober.go:28] interesting pod/downloads-7954f5f757-dqk9z container/download-server namespace/openshift-console: Readiness probe status=failure output="Get \"http://10.217.0.20:8080/\": dial tcp 10.217.0.20:8080: connect: connection refused" start-of-body= Dec 13 06:51:41 crc kubenswrapper[4971]: I1213 06:51:41.293017 4971 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-console/downloads-7954f5f757-dqk9z" podUID="ee2e94f4-cadb-4b17-8853-a1292cb02404" containerName="download-server" probeResult="failure" output="Get \"http://10.217.0.20:8080/\": dial tcp 10.217.0.20:8080: connect: connection refused" Dec 13 06:51:41 crc kubenswrapper[4971]: I1213 06:51:41.311693 4971 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/marketplace-operator-79b997595-j9zqf" Dec 13 06:51:41 crc kubenswrapper[4971]: I1213 06:51:41.316203 4971 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-pxjqw" Dec 13 06:51:41 crc kubenswrapper[4971]: I1213 06:51:41.325890 4971 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-mb2q7" Dec 13 06:51:41 crc kubenswrapper[4971]: I1213 06:51:41.336818 4971 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 13 06:51:41 crc kubenswrapper[4971]: E1213 06:51:41.338696 4971 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-13 06:51:41.838681655 +0000 UTC m=+158.443091103 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 13 06:51:41 crc kubenswrapper[4971]: I1213 06:51:41.340962 4971 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-l4rzp" Dec 13 06:51:41 crc kubenswrapper[4971]: I1213 06:51:41.353740 4971 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-l4rzp" podStartSLOduration=136.353722146 podStartE2EDuration="2m16.353722146s" podCreationTimestamp="2025-12-13 06:49:25 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-13 06:51:41.353433068 +0000 UTC m=+157.957842516" watchObservedRunningTime="2025-12-13 06:51:41.353722146 +0000 UTC m=+157.958131594" Dec 13 06:51:41 crc kubenswrapper[4971]: I1213 06:51:41.371670 4971 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-console-operator/console-operator-58897d9998-kdfcd" Dec 13 06:51:41 crc kubenswrapper[4971]: I1213 06:51:41.441332 4971 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-mdd8q\" (UID: \"cba893c6-ff99-466a-9db0-a62e7a87a334\") " pod="openshift-image-registry/image-registry-697d97f7c8-mdd8q" Dec 13 06:51:41 crc kubenswrapper[4971]: E1213 06:51:41.441737 4971 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-12-13 06:51:41.941722752 +0000 UTC m=+158.546132200 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-mdd8q" (UID: "cba893c6-ff99-466a-9db0-a62e7a87a334") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 13 06:51:41 crc kubenswrapper[4971]: I1213 06:51:41.542599 4971 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 13 06:51:41 crc kubenswrapper[4971]: E1213 06:51:41.543262 4971 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-13 06:51:42.043242781 +0000 UTC m=+158.647652229 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 13 06:51:41 crc kubenswrapper[4971]: I1213 06:51:41.646247 4971 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-mdd8q\" (UID: \"cba893c6-ff99-466a-9db0-a62e7a87a334\") " pod="openshift-image-registry/image-registry-697d97f7c8-mdd8q" Dec 13 06:51:41 crc kubenswrapper[4971]: E1213 06:51:41.646680 4971 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-12-13 06:51:42.146667038 +0000 UTC m=+158.751076486 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-mdd8q" (UID: "cba893c6-ff99-466a-9db0-a62e7a87a334") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 13 06:51:41 crc kubenswrapper[4971]: I1213 06:51:41.748115 4971 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 13 06:51:41 crc kubenswrapper[4971]: E1213 06:51:41.748551 4971 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-13 06:51:42.248530916 +0000 UTC m=+158.852940364 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 13 06:51:41 crc kubenswrapper[4971]: I1213 06:51:41.853302 4971 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-mdd8q\" (UID: \"cba893c6-ff99-466a-9db0-a62e7a87a334\") " pod="openshift-image-registry/image-registry-697d97f7c8-mdd8q" Dec 13 06:51:41 crc kubenswrapper[4971]: E1213 06:51:41.853788 4971 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-12-13 06:51:42.353766868 +0000 UTC m=+158.958176326 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-mdd8q" (UID: "cba893c6-ff99-466a-9db0-a62e7a87a334") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 13 06:51:41 crc kubenswrapper[4971]: I1213 06:51:41.954070 4971 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 13 06:51:41 crc kubenswrapper[4971]: E1213 06:51:41.954260 4971 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-13 06:51:42.45423331 +0000 UTC m=+159.058642758 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 13 06:51:41 crc kubenswrapper[4971]: I1213 06:51:41.954414 4971 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-mdd8q\" (UID: \"cba893c6-ff99-466a-9db0-a62e7a87a334\") " pod="openshift-image-registry/image-registry-697d97f7c8-mdd8q" Dec 13 06:51:41 crc kubenswrapper[4971]: E1213 06:51:41.954764 4971 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-12-13 06:51:42.454752124 +0000 UTC m=+159.059161562 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-mdd8q" (UID: "cba893c6-ff99-466a-9db0-a62e7a87a334") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 13 06:51:42 crc kubenswrapper[4971]: I1213 06:51:42.055204 4971 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 13 06:51:42 crc kubenswrapper[4971]: E1213 06:51:42.055364 4971 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-13 06:51:42.555336179 +0000 UTC m=+159.159745627 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 13 06:51:42 crc kubenswrapper[4971]: I1213 06:51:42.055559 4971 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-mdd8q\" (UID: \"cba893c6-ff99-466a-9db0-a62e7a87a334\") " pod="openshift-image-registry/image-registry-697d97f7c8-mdd8q" Dec 13 06:51:42 crc kubenswrapper[4971]: E1213 06:51:42.055929 4971 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-12-13 06:51:42.555919313 +0000 UTC m=+159.160328811 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-mdd8q" (UID: "cba893c6-ff99-466a-9db0-a62e7a87a334") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 13 06:51:42 crc kubenswrapper[4971]: I1213 06:51:42.070105 4971 patch_prober.go:28] interesting pod/router-default-5444994796-mzhp6 container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Dec 13 06:51:42 crc kubenswrapper[4971]: [-]has-synced failed: reason withheld Dec 13 06:51:42 crc kubenswrapper[4971]: [+]process-running ok Dec 13 06:51:42 crc kubenswrapper[4971]: healthz check failed Dec 13 06:51:42 crc kubenswrapper[4971]: I1213 06:51:42.070160 4971 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-mzhp6" podUID="d89e3583-3f66-4259-b6f9-bffe8ddbde12" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Dec 13 06:51:42 crc kubenswrapper[4971]: I1213 06:51:42.074347 4971 patch_prober.go:28] interesting pod/packageserver-d55dfcdfc-lnrkm container/packageserver namespace/openshift-operator-lifecycle-manager: Readiness probe status=failure output="Get \"https://10.217.0.29:5443/healthz\": net/http: request canceled while waiting for connection (Client.Timeout exceeded while awaiting headers)" start-of-body= Dec 13 06:51:42 crc kubenswrapper[4971]: I1213 06:51:42.074415 4971 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-lnrkm" podUID="87f8bc63-d8c1-415f-975f-b8418034197b" containerName="packageserver" probeResult="failure" output="Get \"https://10.217.0.29:5443/healthz\": net/http: request canceled while waiting for connection (Client.Timeout exceeded while awaiting headers)" Dec 13 06:51:42 crc kubenswrapper[4971]: I1213 06:51:42.157251 4971 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 13 06:51:42 crc kubenswrapper[4971]: E1213 06:51:42.157464 4971 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-13 06:51:42.657444072 +0000 UTC m=+159.261853600 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 13 06:51:42 crc kubenswrapper[4971]: I1213 06:51:42.157824 4971 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-mdd8q\" (UID: \"cba893c6-ff99-466a-9db0-a62e7a87a334\") " pod="openshift-image-registry/image-registry-697d97f7c8-mdd8q" Dec 13 06:51:42 crc kubenswrapper[4971]: E1213 06:51:42.158137 4971 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-12-13 06:51:42.658128529 +0000 UTC m=+159.262537977 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-mdd8q" (UID: "cba893c6-ff99-466a-9db0-a62e7a87a334") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 13 06:51:42 crc kubenswrapper[4971]: I1213 06:51:42.259219 4971 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 13 06:51:42 crc kubenswrapper[4971]: E1213 06:51:42.259393 4971 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-13 06:51:42.759367881 +0000 UTC m=+159.363777319 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 13 06:51:42 crc kubenswrapper[4971]: I1213 06:51:42.259474 4971 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-mdd8q\" (UID: \"cba893c6-ff99-466a-9db0-a62e7a87a334\") " pod="openshift-image-registry/image-registry-697d97f7c8-mdd8q" Dec 13 06:51:42 crc kubenswrapper[4971]: E1213 06:51:42.259787 4971 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-12-13 06:51:42.759778632 +0000 UTC m=+159.364188080 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-mdd8q" (UID: "cba893c6-ff99-466a-9db0-a62e7a87a334") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 13 06:51:42 crc kubenswrapper[4971]: I1213 06:51:42.296250 4971 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="hostpath-provisioner/csi-hostpathplugin-7nhhk" event={"ID":"5fd61507-84bf-4ef5-8bd4-731075ef58ef","Type":"ContainerStarted","Data":"48ee158d5b9070a8f7fffbce0b03d94ff3d39fd1ae5a21650b2d8bebaf5ba938"} Dec 13 06:51:42 crc kubenswrapper[4971]: I1213 06:51:42.296302 4971 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="hostpath-provisioner/csi-hostpathplugin-7nhhk" event={"ID":"5fd61507-84bf-4ef5-8bd4-731075ef58ef","Type":"ContainerStarted","Data":"b7565f6ca82b357a6ceb2a490fdfc08ef77836edf714d857f59a95c5bf665e5e"} Dec 13 06:51:42 crc kubenswrapper[4971]: I1213 06:51:42.360197 4971 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 13 06:51:42 crc kubenswrapper[4971]: E1213 06:51:42.361671 4971 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-13 06:51:42.861655219 +0000 UTC m=+159.466064667 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 13 06:51:42 crc kubenswrapper[4971]: I1213 06:51:42.462413 4971 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-mdd8q\" (UID: \"cba893c6-ff99-466a-9db0-a62e7a87a334\") " pod="openshift-image-registry/image-registry-697d97f7c8-mdd8q" Dec 13 06:51:42 crc kubenswrapper[4971]: E1213 06:51:42.462873 4971 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-12-13 06:51:42.96285352 +0000 UTC m=+159.567262958 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-mdd8q" (UID: "cba893c6-ff99-466a-9db0-a62e7a87a334") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 13 06:51:42 crc kubenswrapper[4971]: I1213 06:51:42.491921 4971 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-lnrkm" Dec 13 06:51:42 crc kubenswrapper[4971]: I1213 06:51:42.563572 4971 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 13 06:51:42 crc kubenswrapper[4971]: E1213 06:51:42.564043 4971 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-13 06:51:43.06402172 +0000 UTC m=+159.668431168 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 13 06:51:42 crc kubenswrapper[4971]: I1213 06:51:42.665075 4971 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-mdd8q\" (UID: \"cba893c6-ff99-466a-9db0-a62e7a87a334\") " pod="openshift-image-registry/image-registry-697d97f7c8-mdd8q" Dec 13 06:51:42 crc kubenswrapper[4971]: E1213 06:51:42.665430 4971 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-12-13 06:51:43.165418155 +0000 UTC m=+159.769827603 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-mdd8q" (UID: "cba893c6-ff99-466a-9db0-a62e7a87a334") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 13 06:51:42 crc kubenswrapper[4971]: I1213 06:51:42.766118 4971 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 13 06:51:42 crc kubenswrapper[4971]: E1213 06:51:42.766483 4971 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-13 06:51:43.266465702 +0000 UTC m=+159.870875150 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 13 06:51:42 crc kubenswrapper[4971]: I1213 06:51:42.867587 4971 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-mdd8q\" (UID: \"cba893c6-ff99-466a-9db0-a62e7a87a334\") " pod="openshift-image-registry/image-registry-697d97f7c8-mdd8q" Dec 13 06:51:42 crc kubenswrapper[4971]: E1213 06:51:42.868014 4971 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-12-13 06:51:43.367998871 +0000 UTC m=+159.972408329 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-mdd8q" (UID: "cba893c6-ff99-466a-9db0-a62e7a87a334") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 13 06:51:42 crc kubenswrapper[4971]: I1213 06:51:42.968880 4971 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 13 06:51:42 crc kubenswrapper[4971]: E1213 06:51:42.969049 4971 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-13 06:51:43.469019507 +0000 UTC m=+160.073428955 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 13 06:51:42 crc kubenswrapper[4971]: I1213 06:51:42.969152 4971 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-mdd8q\" (UID: \"cba893c6-ff99-466a-9db0-a62e7a87a334\") " pod="openshift-image-registry/image-registry-697d97f7c8-mdd8q" Dec 13 06:51:42 crc kubenswrapper[4971]: E1213 06:51:42.969508 4971 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-12-13 06:51:43.46949283 +0000 UTC m=+160.073902278 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-mdd8q" (UID: "cba893c6-ff99-466a-9db0-a62e7a87a334") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 13 06:51:43 crc kubenswrapper[4971]: I1213 06:51:43.031505 4971 plugin_watcher.go:194] "Adding socket path or updating timestamp to desired state cache" path="/var/lib/kubelet/plugins_registry/kubevirt.io.hostpath-provisioner-reg.sock" Dec 13 06:51:43 crc kubenswrapper[4971]: I1213 06:51:43.067623 4971 patch_prober.go:28] interesting pod/router-default-5444994796-mzhp6 container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Dec 13 06:51:43 crc kubenswrapper[4971]: [-]has-synced failed: reason withheld Dec 13 06:51:43 crc kubenswrapper[4971]: [+]process-running ok Dec 13 06:51:43 crc kubenswrapper[4971]: healthz check failed Dec 13 06:51:43 crc kubenswrapper[4971]: I1213 06:51:43.067746 4971 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-mzhp6" podUID="d89e3583-3f66-4259-b6f9-bffe8ddbde12" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Dec 13 06:51:43 crc kubenswrapper[4971]: I1213 06:51:43.070499 4971 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 13 06:51:43 crc kubenswrapper[4971]: E1213 06:51:43.070759 4971 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-13 06:51:43.570730931 +0000 UTC m=+160.175140379 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 13 06:51:43 crc kubenswrapper[4971]: I1213 06:51:43.070822 4971 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-mdd8q\" (UID: \"cba893c6-ff99-466a-9db0-a62e7a87a334\") " pod="openshift-image-registry/image-registry-697d97f7c8-mdd8q" Dec 13 06:51:43 crc kubenswrapper[4971]: E1213 06:51:43.071169 4971 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-12-13 06:51:43.571161932 +0000 UTC m=+160.175571380 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-mdd8q" (UID: "cba893c6-ff99-466a-9db0-a62e7a87a334") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 13 06:51:43 crc kubenswrapper[4971]: I1213 06:51:43.102496 4971 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/community-operators-997kc"] Dec 13 06:51:43 crc kubenswrapper[4971]: I1213 06:51:43.103704 4971 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-997kc" Dec 13 06:51:43 crc kubenswrapper[4971]: I1213 06:51:43.106546 4971 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"community-operators-dockercfg-dmngl" Dec 13 06:51:43 crc kubenswrapper[4971]: I1213 06:51:43.126461 4971 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-997kc"] Dec 13 06:51:43 crc kubenswrapper[4971]: I1213 06:51:43.228821 4971 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 13 06:51:43 crc kubenswrapper[4971]: E1213 06:51:43.228958 4971 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-13 06:51:43.728932174 +0000 UTC m=+160.333341622 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 13 06:51:43 crc kubenswrapper[4971]: I1213 06:51:43.229430 4971 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/cf417851-ae43-462f-a678-125593fb653b-utilities\") pod \"community-operators-997kc\" (UID: \"cf417851-ae43-462f-a678-125593fb653b\") " pod="openshift-marketplace/community-operators-997kc" Dec 13 06:51:43 crc kubenswrapper[4971]: I1213 06:51:43.229509 4971 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/cf417851-ae43-462f-a678-125593fb653b-catalog-content\") pod \"community-operators-997kc\" (UID: \"cf417851-ae43-462f-a678-125593fb653b\") " pod="openshift-marketplace/community-operators-997kc" Dec 13 06:51:43 crc kubenswrapper[4971]: I1213 06:51:43.229692 4971 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-t7lfq\" (UniqueName: \"kubernetes.io/projected/cf417851-ae43-462f-a678-125593fb653b-kube-api-access-t7lfq\") pod \"community-operators-997kc\" (UID: \"cf417851-ae43-462f-a678-125593fb653b\") " pod="openshift-marketplace/community-operators-997kc" Dec 13 06:51:43 crc kubenswrapper[4971]: I1213 06:51:43.229771 4971 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-mdd8q\" (UID: \"cba893c6-ff99-466a-9db0-a62e7a87a334\") " pod="openshift-image-registry/image-registry-697d97f7c8-mdd8q" Dec 13 06:51:43 crc kubenswrapper[4971]: E1213 06:51:43.230130 4971 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-12-13 06:51:43.730110484 +0000 UTC m=+160.334519932 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-mdd8q" (UID: "cba893c6-ff99-466a-9db0-a62e7a87a334") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 13 06:51:43 crc kubenswrapper[4971]: I1213 06:51:43.286163 4971 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/certified-operators-rkd9c"] Dec 13 06:51:43 crc kubenswrapper[4971]: I1213 06:51:43.288748 4971 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-rkd9c" Dec 13 06:51:43 crc kubenswrapper[4971]: I1213 06:51:43.290967 4971 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"certified-operators-dockercfg-4rs5g" Dec 13 06:51:43 crc kubenswrapper[4971]: I1213 06:51:43.307947 4971 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="hostpath-provisioner/csi-hostpathplugin-7nhhk" event={"ID":"5fd61507-84bf-4ef5-8bd4-731075ef58ef","Type":"ContainerStarted","Data":"1c62f62b227cfeb22a5113e10a169cd07db3c5f0f825c0173a9bed2c0f92be77"} Dec 13 06:51:43 crc kubenswrapper[4971]: I1213 06:51:43.311463 4971 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-rkd9c"] Dec 13 06:51:43 crc kubenswrapper[4971]: I1213 06:51:43.330706 4971 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 13 06:51:43 crc kubenswrapper[4971]: E1213 06:51:43.330930 4971 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-13 06:51:43.830898054 +0000 UTC m=+160.435307512 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 13 06:51:43 crc kubenswrapper[4971]: I1213 06:51:43.330986 4971 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/cf417851-ae43-462f-a678-125593fb653b-utilities\") pod \"community-operators-997kc\" (UID: \"cf417851-ae43-462f-a678-125593fb653b\") " pod="openshift-marketplace/community-operators-997kc" Dec 13 06:51:43 crc kubenswrapper[4971]: I1213 06:51:43.331490 4971 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/cf417851-ae43-462f-a678-125593fb653b-utilities\") pod \"community-operators-997kc\" (UID: \"cf417851-ae43-462f-a678-125593fb653b\") " pod="openshift-marketplace/community-operators-997kc" Dec 13 06:51:43 crc kubenswrapper[4971]: I1213 06:51:43.332027 4971 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/cf417851-ae43-462f-a678-125593fb653b-catalog-content\") pod \"community-operators-997kc\" (UID: \"cf417851-ae43-462f-a678-125593fb653b\") " pod="openshift-marketplace/community-operators-997kc" Dec 13 06:51:43 crc kubenswrapper[4971]: I1213 06:51:43.332433 4971 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-t7lfq\" (UniqueName: \"kubernetes.io/projected/cf417851-ae43-462f-a678-125593fb653b-kube-api-access-t7lfq\") pod \"community-operators-997kc\" (UID: \"cf417851-ae43-462f-a678-125593fb653b\") " pod="openshift-marketplace/community-operators-997kc" Dec 13 06:51:43 crc kubenswrapper[4971]: I1213 06:51:43.332548 4971 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-mdd8q\" (UID: \"cba893c6-ff99-466a-9db0-a62e7a87a334\") " pod="openshift-image-registry/image-registry-697d97f7c8-mdd8q" Dec 13 06:51:43 crc kubenswrapper[4971]: I1213 06:51:43.332985 4971 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/cf417851-ae43-462f-a678-125593fb653b-catalog-content\") pod \"community-operators-997kc\" (UID: \"cf417851-ae43-462f-a678-125593fb653b\") " pod="openshift-marketplace/community-operators-997kc" Dec 13 06:51:43 crc kubenswrapper[4971]: E1213 06:51:43.335118 4971 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-12-13 06:51:43.83508976 +0000 UTC m=+160.439499208 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-mdd8q" (UID: "cba893c6-ff99-466a-9db0-a62e7a87a334") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 13 06:51:43 crc kubenswrapper[4971]: I1213 06:51:43.359611 4971 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-t7lfq\" (UniqueName: \"kubernetes.io/projected/cf417851-ae43-462f-a678-125593fb653b-kube-api-access-t7lfq\") pod \"community-operators-997kc\" (UID: \"cf417851-ae43-462f-a678-125593fb653b\") " pod="openshift-marketplace/community-operators-997kc" Dec 13 06:51:43 crc kubenswrapper[4971]: I1213 06:51:43.422651 4971 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-997kc" Dec 13 06:51:43 crc kubenswrapper[4971]: I1213 06:51:43.433334 4971 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 13 06:51:43 crc kubenswrapper[4971]: E1213 06:51:43.433489 4971 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-13 06:51:43.933455949 +0000 UTC m=+160.537865407 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 13 06:51:43 crc kubenswrapper[4971]: I1213 06:51:43.433667 4971 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/8f50f5ba-3ec3-4fa2-9f5d-51e1c43f6352-catalog-content\") pod \"certified-operators-rkd9c\" (UID: \"8f50f5ba-3ec3-4fa2-9f5d-51e1c43f6352\") " pod="openshift-marketplace/certified-operators-rkd9c" Dec 13 06:51:43 crc kubenswrapper[4971]: I1213 06:51:43.433829 4971 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-mdd8q\" (UID: \"cba893c6-ff99-466a-9db0-a62e7a87a334\") " pod="openshift-image-registry/image-registry-697d97f7c8-mdd8q" Dec 13 06:51:43 crc kubenswrapper[4971]: I1213 06:51:43.433899 4971 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/8f50f5ba-3ec3-4fa2-9f5d-51e1c43f6352-utilities\") pod \"certified-operators-rkd9c\" (UID: \"8f50f5ba-3ec3-4fa2-9f5d-51e1c43f6352\") " pod="openshift-marketplace/certified-operators-rkd9c" Dec 13 06:51:43 crc kubenswrapper[4971]: I1213 06:51:43.433927 4971 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-stj24\" (UniqueName: \"kubernetes.io/projected/8f50f5ba-3ec3-4fa2-9f5d-51e1c43f6352-kube-api-access-stj24\") pod \"certified-operators-rkd9c\" (UID: \"8f50f5ba-3ec3-4fa2-9f5d-51e1c43f6352\") " pod="openshift-marketplace/certified-operators-rkd9c" Dec 13 06:51:43 crc kubenswrapper[4971]: E1213 06:51:43.434265 4971 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-12-13 06:51:43.934250249 +0000 UTC m=+160.538659697 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-mdd8q" (UID: "cba893c6-ff99-466a-9db0-a62e7a87a334") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 13 06:51:43 crc kubenswrapper[4971]: I1213 06:51:43.481638 4971 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="hostpath-provisioner/csi-hostpathplugin-7nhhk" podStartSLOduration=10.481618407 podStartE2EDuration="10.481618407s" podCreationTimestamp="2025-12-13 06:51:33 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-13 06:51:43.358746079 +0000 UTC m=+159.963155527" watchObservedRunningTime="2025-12-13 06:51:43.481618407 +0000 UTC m=+160.086027845" Dec 13 06:51:43 crc kubenswrapper[4971]: I1213 06:51:43.482972 4971 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/community-operators-7qgvq"] Dec 13 06:51:43 crc kubenswrapper[4971]: I1213 06:51:43.483967 4971 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-7qgvq" Dec 13 06:51:43 crc kubenswrapper[4971]: I1213 06:51:43.492836 4971 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-7qgvq"] Dec 13 06:51:43 crc kubenswrapper[4971]: I1213 06:51:43.534762 4971 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 13 06:51:43 crc kubenswrapper[4971]: E1213 06:51:43.534877 4971 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-13 06:51:44.034852374 +0000 UTC m=+160.639261822 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 13 06:51:43 crc kubenswrapper[4971]: I1213 06:51:43.535250 4971 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/8f50f5ba-3ec3-4fa2-9f5d-51e1c43f6352-catalog-content\") pod \"certified-operators-rkd9c\" (UID: \"8f50f5ba-3ec3-4fa2-9f5d-51e1c43f6352\") " pod="openshift-marketplace/certified-operators-rkd9c" Dec 13 06:51:43 crc kubenswrapper[4971]: I1213 06:51:43.535299 4971 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/89f6de2b-6afc-4782-97ca-5601253c688f-utilities\") pod \"community-operators-7qgvq\" (UID: \"89f6de2b-6afc-4782-97ca-5601253c688f\") " pod="openshift-marketplace/community-operators-7qgvq" Dec 13 06:51:43 crc kubenswrapper[4971]: I1213 06:51:43.535317 4971 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-j5glc\" (UniqueName: \"kubernetes.io/projected/89f6de2b-6afc-4782-97ca-5601253c688f-kube-api-access-j5glc\") pod \"community-operators-7qgvq\" (UID: \"89f6de2b-6afc-4782-97ca-5601253c688f\") " pod="openshift-marketplace/community-operators-7qgvq" Dec 13 06:51:43 crc kubenswrapper[4971]: I1213 06:51:43.535341 4971 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/89f6de2b-6afc-4782-97ca-5601253c688f-catalog-content\") pod \"community-operators-7qgvq\" (UID: \"89f6de2b-6afc-4782-97ca-5601253c688f\") " pod="openshift-marketplace/community-operators-7qgvq" Dec 13 06:51:43 crc kubenswrapper[4971]: I1213 06:51:43.535369 4971 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-mdd8q\" (UID: \"cba893c6-ff99-466a-9db0-a62e7a87a334\") " pod="openshift-image-registry/image-registry-697d97f7c8-mdd8q" Dec 13 06:51:43 crc kubenswrapper[4971]: I1213 06:51:43.535403 4971 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/8f50f5ba-3ec3-4fa2-9f5d-51e1c43f6352-utilities\") pod \"certified-operators-rkd9c\" (UID: \"8f50f5ba-3ec3-4fa2-9f5d-51e1c43f6352\") " pod="openshift-marketplace/certified-operators-rkd9c" Dec 13 06:51:43 crc kubenswrapper[4971]: I1213 06:51:43.535421 4971 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-stj24\" (UniqueName: \"kubernetes.io/projected/8f50f5ba-3ec3-4fa2-9f5d-51e1c43f6352-kube-api-access-stj24\") pod \"certified-operators-rkd9c\" (UID: \"8f50f5ba-3ec3-4fa2-9f5d-51e1c43f6352\") " pod="openshift-marketplace/certified-operators-rkd9c" Dec 13 06:51:43 crc kubenswrapper[4971]: E1213 06:51:43.535896 4971 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-12-13 06:51:44.03588236 +0000 UTC m=+160.640291808 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-mdd8q" (UID: "cba893c6-ff99-466a-9db0-a62e7a87a334") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 13 06:51:43 crc kubenswrapper[4971]: I1213 06:51:43.536277 4971 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/8f50f5ba-3ec3-4fa2-9f5d-51e1c43f6352-catalog-content\") pod \"certified-operators-rkd9c\" (UID: \"8f50f5ba-3ec3-4fa2-9f5d-51e1c43f6352\") " pod="openshift-marketplace/certified-operators-rkd9c" Dec 13 06:51:43 crc kubenswrapper[4971]: I1213 06:51:43.536471 4971 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/8f50f5ba-3ec3-4fa2-9f5d-51e1c43f6352-utilities\") pod \"certified-operators-rkd9c\" (UID: \"8f50f5ba-3ec3-4fa2-9f5d-51e1c43f6352\") " pod="openshift-marketplace/certified-operators-rkd9c" Dec 13 06:51:43 crc kubenswrapper[4971]: I1213 06:51:43.565906 4971 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-stj24\" (UniqueName: \"kubernetes.io/projected/8f50f5ba-3ec3-4fa2-9f5d-51e1c43f6352-kube-api-access-stj24\") pod \"certified-operators-rkd9c\" (UID: \"8f50f5ba-3ec3-4fa2-9f5d-51e1c43f6352\") " pod="openshift-marketplace/certified-operators-rkd9c" Dec 13 06:51:43 crc kubenswrapper[4971]: I1213 06:51:43.605198 4971 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-rkd9c" Dec 13 06:51:43 crc kubenswrapper[4971]: I1213 06:51:43.618831 4971 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-997kc"] Dec 13 06:51:43 crc kubenswrapper[4971]: I1213 06:51:43.638147 4971 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 13 06:51:43 crc kubenswrapper[4971]: I1213 06:51:43.638508 4971 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/89f6de2b-6afc-4782-97ca-5601253c688f-utilities\") pod \"community-operators-7qgvq\" (UID: \"89f6de2b-6afc-4782-97ca-5601253c688f\") " pod="openshift-marketplace/community-operators-7qgvq" Dec 13 06:51:43 crc kubenswrapper[4971]: I1213 06:51:43.638564 4971 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-j5glc\" (UniqueName: \"kubernetes.io/projected/89f6de2b-6afc-4782-97ca-5601253c688f-kube-api-access-j5glc\") pod \"community-operators-7qgvq\" (UID: \"89f6de2b-6afc-4782-97ca-5601253c688f\") " pod="openshift-marketplace/community-operators-7qgvq" Dec 13 06:51:43 crc kubenswrapper[4971]: I1213 06:51:43.638609 4971 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/89f6de2b-6afc-4782-97ca-5601253c688f-catalog-content\") pod \"community-operators-7qgvq\" (UID: \"89f6de2b-6afc-4782-97ca-5601253c688f\") " pod="openshift-marketplace/community-operators-7qgvq" Dec 13 06:51:43 crc kubenswrapper[4971]: I1213 06:51:43.639331 4971 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/89f6de2b-6afc-4782-97ca-5601253c688f-catalog-content\") pod \"community-operators-7qgvq\" (UID: \"89f6de2b-6afc-4782-97ca-5601253c688f\") " pod="openshift-marketplace/community-operators-7qgvq" Dec 13 06:51:43 crc kubenswrapper[4971]: E1213 06:51:43.639449 4971 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-13 06:51:44.13942443 +0000 UTC m=+160.743833878 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 13 06:51:43 crc kubenswrapper[4971]: I1213 06:51:43.639750 4971 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/89f6de2b-6afc-4782-97ca-5601253c688f-utilities\") pod \"community-operators-7qgvq\" (UID: \"89f6de2b-6afc-4782-97ca-5601253c688f\") " pod="openshift-marketplace/community-operators-7qgvq" Dec 13 06:51:43 crc kubenswrapper[4971]: I1213 06:51:43.658289 4971 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-j5glc\" (UniqueName: \"kubernetes.io/projected/89f6de2b-6afc-4782-97ca-5601253c688f-kube-api-access-j5glc\") pod \"community-operators-7qgvq\" (UID: \"89f6de2b-6afc-4782-97ca-5601253c688f\") " pod="openshift-marketplace/community-operators-7qgvq" Dec 13 06:51:43 crc kubenswrapper[4971]: I1213 06:51:43.683024 4971 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/certified-operators-82f6m"] Dec 13 06:51:43 crc kubenswrapper[4971]: I1213 06:51:43.684276 4971 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-82f6m" Dec 13 06:51:43 crc kubenswrapper[4971]: I1213 06:51:43.692733 4971 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-82f6m"] Dec 13 06:51:43 crc kubenswrapper[4971]: I1213 06:51:43.740086 4971 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-mdd8q\" (UID: \"cba893c6-ff99-466a-9db0-a62e7a87a334\") " pod="openshift-image-registry/image-registry-697d97f7c8-mdd8q" Dec 13 06:51:43 crc kubenswrapper[4971]: I1213 06:51:43.740471 4971 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/9404e8eb-0f18-4a49-a716-f076273e7179-utilities\") pod \"certified-operators-82f6m\" (UID: \"9404e8eb-0f18-4a49-a716-f076273e7179\") " pod="openshift-marketplace/certified-operators-82f6m" Dec 13 06:51:43 crc kubenswrapper[4971]: I1213 06:51:43.740752 4971 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/9404e8eb-0f18-4a49-a716-f076273e7179-catalog-content\") pod \"certified-operators-82f6m\" (UID: \"9404e8eb-0f18-4a49-a716-f076273e7179\") " pod="openshift-marketplace/certified-operators-82f6m" Dec 13 06:51:43 crc kubenswrapper[4971]: I1213 06:51:43.740833 4971 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-wjf2d\" (UniqueName: \"kubernetes.io/projected/9404e8eb-0f18-4a49-a716-f076273e7179-kube-api-access-wjf2d\") pod \"certified-operators-82f6m\" (UID: \"9404e8eb-0f18-4a49-a716-f076273e7179\") " pod="openshift-marketplace/certified-operators-82f6m" Dec 13 06:51:43 crc kubenswrapper[4971]: E1213 06:51:43.741029 4971 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-12-13 06:51:44.241009231 +0000 UTC m=+160.845418679 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-mdd8q" (UID: "cba893c6-ff99-466a-9db0-a62e7a87a334") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 13 06:51:43 crc kubenswrapper[4971]: I1213 06:51:43.800658 4971 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-7qgvq" Dec 13 06:51:43 crc kubenswrapper[4971]: I1213 06:51:43.841642 4971 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 13 06:51:43 crc kubenswrapper[4971]: E1213 06:51:43.841821 4971 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-13 06:51:44.341794701 +0000 UTC m=+160.946204149 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 13 06:51:43 crc kubenswrapper[4971]: I1213 06:51:43.841947 4971 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/9404e8eb-0f18-4a49-a716-f076273e7179-utilities\") pod \"certified-operators-82f6m\" (UID: \"9404e8eb-0f18-4a49-a716-f076273e7179\") " pod="openshift-marketplace/certified-operators-82f6m" Dec 13 06:51:43 crc kubenswrapper[4971]: I1213 06:51:43.842004 4971 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/9404e8eb-0f18-4a49-a716-f076273e7179-catalog-content\") pod \"certified-operators-82f6m\" (UID: \"9404e8eb-0f18-4a49-a716-f076273e7179\") " pod="openshift-marketplace/certified-operators-82f6m" Dec 13 06:51:43 crc kubenswrapper[4971]: I1213 06:51:43.842072 4971 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-wjf2d\" (UniqueName: \"kubernetes.io/projected/9404e8eb-0f18-4a49-a716-f076273e7179-kube-api-access-wjf2d\") pod \"certified-operators-82f6m\" (UID: \"9404e8eb-0f18-4a49-a716-f076273e7179\") " pod="openshift-marketplace/certified-operators-82f6m" Dec 13 06:51:43 crc kubenswrapper[4971]: I1213 06:51:43.842108 4971 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-mdd8q\" (UID: \"cba893c6-ff99-466a-9db0-a62e7a87a334\") " pod="openshift-image-registry/image-registry-697d97f7c8-mdd8q" Dec 13 06:51:43 crc kubenswrapper[4971]: I1213 06:51:43.842437 4971 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/9404e8eb-0f18-4a49-a716-f076273e7179-utilities\") pod \"certified-operators-82f6m\" (UID: \"9404e8eb-0f18-4a49-a716-f076273e7179\") " pod="openshift-marketplace/certified-operators-82f6m" Dec 13 06:51:43 crc kubenswrapper[4971]: E1213 06:51:43.842456 4971 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-12-13 06:51:44.342447487 +0000 UTC m=+160.946856935 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-mdd8q" (UID: "cba893c6-ff99-466a-9db0-a62e7a87a334") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 13 06:51:43 crc kubenswrapper[4971]: I1213 06:51:43.842963 4971 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/9404e8eb-0f18-4a49-a716-f076273e7179-catalog-content\") pod \"certified-operators-82f6m\" (UID: \"9404e8eb-0f18-4a49-a716-f076273e7179\") " pod="openshift-marketplace/certified-operators-82f6m" Dec 13 06:51:43 crc kubenswrapper[4971]: I1213 06:51:43.860235 4971 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-wjf2d\" (UniqueName: \"kubernetes.io/projected/9404e8eb-0f18-4a49-a716-f076273e7179-kube-api-access-wjf2d\") pod \"certified-operators-82f6m\" (UID: \"9404e8eb-0f18-4a49-a716-f076273e7179\") " pod="openshift-marketplace/certified-operators-82f6m" Dec 13 06:51:43 crc kubenswrapper[4971]: I1213 06:51:43.942625 4971 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 13 06:51:43 crc kubenswrapper[4971]: E1213 06:51:43.942814 4971 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-13 06:51:44.442785656 +0000 UTC m=+161.047195104 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 13 06:51:43 crc kubenswrapper[4971]: I1213 06:51:43.942917 4971 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-mdd8q\" (UID: \"cba893c6-ff99-466a-9db0-a62e7a87a334\") " pod="openshift-image-registry/image-registry-697d97f7c8-mdd8q" Dec 13 06:51:43 crc kubenswrapper[4971]: E1213 06:51:43.943267 4971 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-12-13 06:51:44.443252518 +0000 UTC m=+161.047661966 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-mdd8q" (UID: "cba893c6-ff99-466a-9db0-a62e7a87a334") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 13 06:51:43 crc kubenswrapper[4971]: I1213 06:51:43.949885 4971 reconciler.go:161] "OperationExecutor.RegisterPlugin started" plugin={"SocketPath":"/var/lib/kubelet/plugins_registry/kubevirt.io.hostpath-provisioner-reg.sock","Timestamp":"2025-12-13T06:51:43.0315785Z","Handler":null,"Name":""} Dec 13 06:51:43 crc kubenswrapper[4971]: I1213 06:51:43.952265 4971 csi_plugin.go:100] kubernetes.io/csi: Trying to validate a new CSI Driver with name: kubevirt.io.hostpath-provisioner endpoint: /var/lib/kubelet/plugins/csi-hostpath/csi.sock versions: 1.0.0 Dec 13 06:51:43 crc kubenswrapper[4971]: I1213 06:51:43.952299 4971 csi_plugin.go:113] kubernetes.io/csi: Register new plugin with name: kubevirt.io.hostpath-provisioner at endpoint: /var/lib/kubelet/plugins/csi-hostpath/csi.sock Dec 13 06:51:44 crc kubenswrapper[4971]: I1213 06:51:44.012726 4971 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-82f6m" Dec 13 06:51:44 crc kubenswrapper[4971]: I1213 06:51:44.043661 4971 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 13 06:51:44 crc kubenswrapper[4971]: I1213 06:51:44.047072 4971 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (OuterVolumeSpecName: "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b"). InnerVolumeSpecName "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8". PluginName "kubernetes.io/csi", VolumeGidValue "" Dec 13 06:51:44 crc kubenswrapper[4971]: I1213 06:51:44.060501 4971 patch_prober.go:28] interesting pod/router-default-5444994796-mzhp6 container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Dec 13 06:51:44 crc kubenswrapper[4971]: [-]has-synced failed: reason withheld Dec 13 06:51:44 crc kubenswrapper[4971]: [+]process-running ok Dec 13 06:51:44 crc kubenswrapper[4971]: healthz check failed Dec 13 06:51:44 crc kubenswrapper[4971]: I1213 06:51:44.060577 4971 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-mzhp6" podUID="d89e3583-3f66-4259-b6f9-bffe8ddbde12" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Dec 13 06:51:44 crc kubenswrapper[4971]: I1213 06:51:44.144832 4971 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-mdd8q\" (UID: \"cba893c6-ff99-466a-9db0-a62e7a87a334\") " pod="openshift-image-registry/image-registry-697d97f7c8-mdd8q" Dec 13 06:51:44 crc kubenswrapper[4971]: I1213 06:51:44.234769 4971 csi_attacher.go:380] kubernetes.io/csi: attacher.MountDevice STAGE_UNSTAGE_VOLUME capability not set. Skipping MountDevice... Dec 13 06:51:44 crc kubenswrapper[4971]: I1213 06:51:44.234824 4971 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-mdd8q\" (UID: \"cba893c6-ff99-466a-9db0-a62e7a87a334\") device mount path \"/var/lib/kubelet/plugins/kubernetes.io/csi/kubevirt.io.hostpath-provisioner/1f4776af88835e41c12b831b4c9fed40233456d14189815a54dbe7f892fc1983/globalmount\"" pod="openshift-image-registry/image-registry-697d97f7c8-mdd8q" Dec 13 06:51:44 crc kubenswrapper[4971]: I1213 06:51:44.276468 4971 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-mdd8q\" (UID: \"cba893c6-ff99-466a-9db0-a62e7a87a334\") " pod="openshift-image-registry/image-registry-697d97f7c8-mdd8q" Dec 13 06:51:44 crc kubenswrapper[4971]: I1213 06:51:44.342810 4971 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-997kc" event={"ID":"cf417851-ae43-462f-a678-125593fb653b","Type":"ContainerStarted","Data":"258f4a913f8ba4a34d86e83ac34969fb30afa3dfb66b8f141e0c6bf3b63eaab0"} Dec 13 06:51:44 crc kubenswrapper[4971]: I1213 06:51:44.343542 4971 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-image-registry/image-registry-697d97f7c8-mdd8q" Dec 13 06:51:44 crc kubenswrapper[4971]: I1213 06:51:44.350904 4971 generic.go:334] "Generic (PLEG): container finished" podID="c1575ff9-1c53-4732-bc48-252f9f270597" containerID="7279bfe22b885129c7c9a8d98bc482013cb0cdf21dfdc5a026eda208800095de" exitCode=0 Dec 13 06:51:44 crc kubenswrapper[4971]: I1213 06:51:44.351279 4971 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29426805-fxxrz" event={"ID":"c1575ff9-1c53-4732-bc48-252f9f270597","Type":"ContainerDied","Data":"7279bfe22b885129c7c9a8d98bc482013cb0cdf21dfdc5a026eda208800095de"} Dec 13 06:51:44 crc kubenswrapper[4971]: I1213 06:51:44.701315 4971 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-82f6m"] Dec 13 06:51:44 crc kubenswrapper[4971]: I1213 06:51:44.705177 4971 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-7qgvq"] Dec 13 06:51:44 crc kubenswrapper[4971]: W1213 06:51:44.746859 4971 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod89f6de2b_6afc_4782_97ca_5601253c688f.slice/crio-bec0257f0dfd4cf28c11dae346efe7a50a920c0b7d88591f6b542f2f4b6bcdc4 WatchSource:0}: Error finding container bec0257f0dfd4cf28c11dae346efe7a50a920c0b7d88591f6b542f2f4b6bcdc4: Status 404 returned error can't find the container with id bec0257f0dfd4cf28c11dae346efe7a50a920c0b7d88591f6b542f2f4b6bcdc4 Dec 13 06:51:44 crc kubenswrapper[4971]: I1213 06:51:44.748835 4971 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-rkd9c"] Dec 13 06:51:44 crc kubenswrapper[4971]: W1213 06:51:44.756735 4971 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod8f50f5ba_3ec3_4fa2_9f5d_51e1c43f6352.slice/crio-d46dca0b4e85de9494ecaf1e3a49c4a9d7eb1b1e29a3f6dd19588adef058cc3e WatchSource:0}: Error finding container d46dca0b4e85de9494ecaf1e3a49c4a9d7eb1b1e29a3f6dd19588adef058cc3e: Status 404 returned error can't find the container with id d46dca0b4e85de9494ecaf1e3a49c4a9d7eb1b1e29a3f6dd19588adef058cc3e Dec 13 06:51:44 crc kubenswrapper[4971]: I1213 06:51:44.826164 4971 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-image-registry/image-registry-697d97f7c8-mdd8q"] Dec 13 06:51:44 crc kubenswrapper[4971]: W1213 06:51:44.860158 4971 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podcba893c6_ff99_466a_9db0_a62e7a87a334.slice/crio-7ba0c024c728c897f54acbb50586782f4f871548f0d1f5406f676b8e1beeefeb WatchSource:0}: Error finding container 7ba0c024c728c897f54acbb50586782f4f871548f0d1f5406f676b8e1beeefeb: Status 404 returned error can't find the container with id 7ba0c024c728c897f54acbb50586782f4f871548f0d1f5406f676b8e1beeefeb Dec 13 06:51:45 crc kubenswrapper[4971]: I1213 06:51:45.001541 4971 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-ovn-kubernetes/ovnkube-node-zlcg5" Dec 13 06:51:45 crc kubenswrapper[4971]: I1213 06:51:45.061909 4971 patch_prober.go:28] interesting pod/router-default-5444994796-mzhp6 container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Dec 13 06:51:45 crc kubenswrapper[4971]: [-]has-synced failed: reason withheld Dec 13 06:51:45 crc kubenswrapper[4971]: [+]process-running ok Dec 13 06:51:45 crc kubenswrapper[4971]: healthz check failed Dec 13 06:51:45 crc kubenswrapper[4971]: I1213 06:51:45.061980 4971 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-mzhp6" podUID="d89e3583-3f66-4259-b6f9-bffe8ddbde12" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Dec 13 06:51:45 crc kubenswrapper[4971]: I1213 06:51:45.080810 4971 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-marketplace-xwxlf"] Dec 13 06:51:45 crc kubenswrapper[4971]: I1213 06:51:45.082217 4971 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-xwxlf" Dec 13 06:51:45 crc kubenswrapper[4971]: I1213 06:51:45.084266 4971 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"redhat-marketplace-dockercfg-x2ctb" Dec 13 06:51:45 crc kubenswrapper[4971]: I1213 06:51:45.090014 4971 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-xwxlf"] Dec 13 06:51:45 crc kubenswrapper[4971]: I1213 06:51:45.186692 4971 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-25p8j\" (UniqueName: \"kubernetes.io/projected/22037bdc-fe6d-4411-be2b-aec38524a8dd-kube-api-access-25p8j\") pod \"redhat-marketplace-xwxlf\" (UID: \"22037bdc-fe6d-4411-be2b-aec38524a8dd\") " pod="openshift-marketplace/redhat-marketplace-xwxlf" Dec 13 06:51:45 crc kubenswrapper[4971]: I1213 06:51:45.186906 4971 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/22037bdc-fe6d-4411-be2b-aec38524a8dd-utilities\") pod \"redhat-marketplace-xwxlf\" (UID: \"22037bdc-fe6d-4411-be2b-aec38524a8dd\") " pod="openshift-marketplace/redhat-marketplace-xwxlf" Dec 13 06:51:45 crc kubenswrapper[4971]: I1213 06:51:45.187025 4971 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/22037bdc-fe6d-4411-be2b-aec38524a8dd-catalog-content\") pod \"redhat-marketplace-xwxlf\" (UID: \"22037bdc-fe6d-4411-be2b-aec38524a8dd\") " pod="openshift-marketplace/redhat-marketplace-xwxlf" Dec 13 06:51:45 crc kubenswrapper[4971]: I1213 06:51:45.225959 4971 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-console/console-f9d7485db-5fd9b" Dec 13 06:51:45 crc kubenswrapper[4971]: I1213 06:51:45.226029 4971 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-console/console-f9d7485db-5fd9b" Dec 13 06:51:45 crc kubenswrapper[4971]: I1213 06:51:45.227279 4971 patch_prober.go:28] interesting pod/console-f9d7485db-5fd9b container/console namespace/openshift-console: Startup probe status=failure output="Get \"https://10.217.0.10:8443/health\": dial tcp 10.217.0.10:8443: connect: connection refused" start-of-body= Dec 13 06:51:45 crc kubenswrapper[4971]: I1213 06:51:45.227316 4971 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-console/console-f9d7485db-5fd9b" podUID="9f0db656-7a94-415b-b737-30ed342817ee" containerName="console" probeResult="failure" output="Get \"https://10.217.0.10:8443/health\": dial tcp 10.217.0.10:8443: connect: connection refused" Dec 13 06:51:45 crc kubenswrapper[4971]: I1213 06:51:45.288685 4971 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-25p8j\" (UniqueName: \"kubernetes.io/projected/22037bdc-fe6d-4411-be2b-aec38524a8dd-kube-api-access-25p8j\") pod \"redhat-marketplace-xwxlf\" (UID: \"22037bdc-fe6d-4411-be2b-aec38524a8dd\") " pod="openshift-marketplace/redhat-marketplace-xwxlf" Dec 13 06:51:45 crc kubenswrapper[4971]: I1213 06:51:45.288921 4971 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/22037bdc-fe6d-4411-be2b-aec38524a8dd-utilities\") pod \"redhat-marketplace-xwxlf\" (UID: \"22037bdc-fe6d-4411-be2b-aec38524a8dd\") " pod="openshift-marketplace/redhat-marketplace-xwxlf" Dec 13 06:51:45 crc kubenswrapper[4971]: I1213 06:51:45.288954 4971 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/22037bdc-fe6d-4411-be2b-aec38524a8dd-catalog-content\") pod \"redhat-marketplace-xwxlf\" (UID: \"22037bdc-fe6d-4411-be2b-aec38524a8dd\") " pod="openshift-marketplace/redhat-marketplace-xwxlf" Dec 13 06:51:45 crc kubenswrapper[4971]: I1213 06:51:45.289772 4971 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/22037bdc-fe6d-4411-be2b-aec38524a8dd-catalog-content\") pod \"redhat-marketplace-xwxlf\" (UID: \"22037bdc-fe6d-4411-be2b-aec38524a8dd\") " pod="openshift-marketplace/redhat-marketplace-xwxlf" Dec 13 06:51:45 crc kubenswrapper[4971]: I1213 06:51:45.289883 4971 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/22037bdc-fe6d-4411-be2b-aec38524a8dd-utilities\") pod \"redhat-marketplace-xwxlf\" (UID: \"22037bdc-fe6d-4411-be2b-aec38524a8dd\") " pod="openshift-marketplace/redhat-marketplace-xwxlf" Dec 13 06:51:45 crc kubenswrapper[4971]: I1213 06:51:45.314766 4971 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-25p8j\" (UniqueName: \"kubernetes.io/projected/22037bdc-fe6d-4411-be2b-aec38524a8dd-kube-api-access-25p8j\") pod \"redhat-marketplace-xwxlf\" (UID: \"22037bdc-fe6d-4411-be2b-aec38524a8dd\") " pod="openshift-marketplace/redhat-marketplace-xwxlf" Dec 13 06:51:45 crc kubenswrapper[4971]: I1213 06:51:45.359226 4971 generic.go:334] "Generic (PLEG): container finished" podID="9404e8eb-0f18-4a49-a716-f076273e7179" containerID="8d4628a40eb05a6df7bd6b0fece84c9e24e6b6112b6289309830e6027e1bb986" exitCode=0 Dec 13 06:51:45 crc kubenswrapper[4971]: I1213 06:51:45.359361 4971 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-82f6m" event={"ID":"9404e8eb-0f18-4a49-a716-f076273e7179","Type":"ContainerDied","Data":"8d4628a40eb05a6df7bd6b0fece84c9e24e6b6112b6289309830e6027e1bb986"} Dec 13 06:51:45 crc kubenswrapper[4971]: I1213 06:51:45.359686 4971 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-82f6m" event={"ID":"9404e8eb-0f18-4a49-a716-f076273e7179","Type":"ContainerStarted","Data":"873ed9c2e934008905cffeebd260086c4af4aeb66a1946f5fa1eed62c4d72268"} Dec 13 06:51:45 crc kubenswrapper[4971]: I1213 06:51:45.361210 4971 provider.go:102] Refreshing cache for provider: *credentialprovider.defaultDockerConfigProvider Dec 13 06:51:45 crc kubenswrapper[4971]: I1213 06:51:45.362410 4971 generic.go:334] "Generic (PLEG): container finished" podID="cf417851-ae43-462f-a678-125593fb653b" containerID="ef186a9e3ccd3c7dd0a3bc655e584757165c6f621358f6432a5768b19e98cbc8" exitCode=0 Dec 13 06:51:45 crc kubenswrapper[4971]: I1213 06:51:45.362456 4971 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-997kc" event={"ID":"cf417851-ae43-462f-a678-125593fb653b","Type":"ContainerDied","Data":"ef186a9e3ccd3c7dd0a3bc655e584757165c6f621358f6432a5768b19e98cbc8"} Dec 13 06:51:45 crc kubenswrapper[4971]: I1213 06:51:45.365122 4971 generic.go:334] "Generic (PLEG): container finished" podID="89f6de2b-6afc-4782-97ca-5601253c688f" containerID="89235ff3b064024ef637ca519f2c1ffc06586998c87c15d021996b7b76a7f9e3" exitCode=0 Dec 13 06:51:45 crc kubenswrapper[4971]: I1213 06:51:45.365211 4971 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-7qgvq" event={"ID":"89f6de2b-6afc-4782-97ca-5601253c688f","Type":"ContainerDied","Data":"89235ff3b064024ef637ca519f2c1ffc06586998c87c15d021996b7b76a7f9e3"} Dec 13 06:51:45 crc kubenswrapper[4971]: I1213 06:51:45.365251 4971 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-7qgvq" event={"ID":"89f6de2b-6afc-4782-97ca-5601253c688f","Type":"ContainerStarted","Data":"bec0257f0dfd4cf28c11dae346efe7a50a920c0b7d88591f6b542f2f4b6bcdc4"} Dec 13 06:51:45 crc kubenswrapper[4971]: I1213 06:51:45.367006 4971 generic.go:334] "Generic (PLEG): container finished" podID="8f50f5ba-3ec3-4fa2-9f5d-51e1c43f6352" containerID="f1a912a2a413b3b52e51cb9027a159d5bb254455287625669896a9d6362ddc4f" exitCode=0 Dec 13 06:51:45 crc kubenswrapper[4971]: I1213 06:51:45.367059 4971 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-rkd9c" event={"ID":"8f50f5ba-3ec3-4fa2-9f5d-51e1c43f6352","Type":"ContainerDied","Data":"f1a912a2a413b3b52e51cb9027a159d5bb254455287625669896a9d6362ddc4f"} Dec 13 06:51:45 crc kubenswrapper[4971]: I1213 06:51:45.367198 4971 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-rkd9c" event={"ID":"8f50f5ba-3ec3-4fa2-9f5d-51e1c43f6352","Type":"ContainerStarted","Data":"d46dca0b4e85de9494ecaf1e3a49c4a9d7eb1b1e29a3f6dd19588adef058cc3e"} Dec 13 06:51:45 crc kubenswrapper[4971]: I1213 06:51:45.371796 4971 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-image-registry/image-registry-697d97f7c8-mdd8q" event={"ID":"cba893c6-ff99-466a-9db0-a62e7a87a334","Type":"ContainerStarted","Data":"db855d10cdb4b3b9b54e0ece89243b84cac1c530b376b33acceeff82e7f769b3"} Dec 13 06:51:45 crc kubenswrapper[4971]: I1213 06:51:45.371839 4971 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-image-registry/image-registry-697d97f7c8-mdd8q" event={"ID":"cba893c6-ff99-466a-9db0-a62e7a87a334","Type":"ContainerStarted","Data":"7ba0c024c728c897f54acbb50586782f4f871548f0d1f5406f676b8e1beeefeb"} Dec 13 06:51:45 crc kubenswrapper[4971]: I1213 06:51:45.429806 4971 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-xwxlf" Dec 13 06:51:45 crc kubenswrapper[4971]: I1213 06:51:45.459984 4971 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-image-registry/image-registry-697d97f7c8-mdd8q" podStartSLOduration=140.459962654 podStartE2EDuration="2m20.459962654s" podCreationTimestamp="2025-12-13 06:49:25 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-13 06:51:45.455625165 +0000 UTC m=+162.060034613" watchObservedRunningTime="2025-12-13 06:51:45.459962654 +0000 UTC m=+162.064372102" Dec 13 06:51:45 crc kubenswrapper[4971]: I1213 06:51:45.478909 4971 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-apiserver/apiserver-76f77b778f-2n4cq" Dec 13 06:51:45 crc kubenswrapper[4971]: I1213 06:51:45.478979 4971 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-apiserver/apiserver-76f77b778f-2n4cq" Dec 13 06:51:45 crc kubenswrapper[4971]: I1213 06:51:45.489437 4971 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-apiserver/apiserver-76f77b778f-2n4cq" Dec 13 06:51:45 crc kubenswrapper[4971]: I1213 06:51:45.501299 4971 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-marketplace-mkqps"] Dec 13 06:51:45 crc kubenswrapper[4971]: I1213 06:51:45.505770 4971 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-mkqps" Dec 13 06:51:45 crc kubenswrapper[4971]: I1213 06:51:45.509024 4971 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-mkqps"] Dec 13 06:51:45 crc kubenswrapper[4971]: I1213 06:51:45.595061 4971 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/9d4f1ea5-8472-4320-820b-8090ff6f5ca6-catalog-content\") pod \"redhat-marketplace-mkqps\" (UID: \"9d4f1ea5-8472-4320-820b-8090ff6f5ca6\") " pod="openshift-marketplace/redhat-marketplace-mkqps" Dec 13 06:51:45 crc kubenswrapper[4971]: I1213 06:51:45.595148 4971 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/9d4f1ea5-8472-4320-820b-8090ff6f5ca6-utilities\") pod \"redhat-marketplace-mkqps\" (UID: \"9d4f1ea5-8472-4320-820b-8090ff6f5ca6\") " pod="openshift-marketplace/redhat-marketplace-mkqps" Dec 13 06:51:45 crc kubenswrapper[4971]: I1213 06:51:45.595217 4971 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-npf8d\" (UniqueName: \"kubernetes.io/projected/9d4f1ea5-8472-4320-820b-8090ff6f5ca6-kube-api-access-npf8d\") pod \"redhat-marketplace-mkqps\" (UID: \"9d4f1ea5-8472-4320-820b-8090ff6f5ca6\") " pod="openshift-marketplace/redhat-marketplace-mkqps" Dec 13 06:51:45 crc kubenswrapper[4971]: I1213 06:51:45.635051 4971 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29426805-fxxrz" Dec 13 06:51:45 crc kubenswrapper[4971]: I1213 06:51:45.691320 4971 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-xwxlf"] Dec 13 06:51:45 crc kubenswrapper[4971]: I1213 06:51:45.696339 4971 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/c1575ff9-1c53-4732-bc48-252f9f270597-config-volume\") pod \"c1575ff9-1c53-4732-bc48-252f9f270597\" (UID: \"c1575ff9-1c53-4732-bc48-252f9f270597\") " Dec 13 06:51:45 crc kubenswrapper[4971]: I1213 06:51:45.696735 4971 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/c1575ff9-1c53-4732-bc48-252f9f270597-secret-volume\") pod \"c1575ff9-1c53-4732-bc48-252f9f270597\" (UID: \"c1575ff9-1c53-4732-bc48-252f9f270597\") " Dec 13 06:51:45 crc kubenswrapper[4971]: I1213 06:51:45.697025 4971 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-qdm7j\" (UniqueName: \"kubernetes.io/projected/c1575ff9-1c53-4732-bc48-252f9f270597-kube-api-access-qdm7j\") pod \"c1575ff9-1c53-4732-bc48-252f9f270597\" (UID: \"c1575ff9-1c53-4732-bc48-252f9f270597\") " Dec 13 06:51:45 crc kubenswrapper[4971]: I1213 06:51:45.697367 4971 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/9d4f1ea5-8472-4320-820b-8090ff6f5ca6-utilities\") pod \"redhat-marketplace-mkqps\" (UID: \"9d4f1ea5-8472-4320-820b-8090ff6f5ca6\") " pod="openshift-marketplace/redhat-marketplace-mkqps" Dec 13 06:51:45 crc kubenswrapper[4971]: I1213 06:51:45.697508 4971 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-npf8d\" (UniqueName: \"kubernetes.io/projected/9d4f1ea5-8472-4320-820b-8090ff6f5ca6-kube-api-access-npf8d\") pod \"redhat-marketplace-mkqps\" (UID: \"9d4f1ea5-8472-4320-820b-8090ff6f5ca6\") " pod="openshift-marketplace/redhat-marketplace-mkqps" Dec 13 06:51:45 crc kubenswrapper[4971]: I1213 06:51:45.697664 4971 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/9d4f1ea5-8472-4320-820b-8090ff6f5ca6-catalog-content\") pod \"redhat-marketplace-mkqps\" (UID: \"9d4f1ea5-8472-4320-820b-8090ff6f5ca6\") " pod="openshift-marketplace/redhat-marketplace-mkqps" Dec 13 06:51:45 crc kubenswrapper[4971]: I1213 06:51:45.698213 4971 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/9d4f1ea5-8472-4320-820b-8090ff6f5ca6-catalog-content\") pod \"redhat-marketplace-mkqps\" (UID: \"9d4f1ea5-8472-4320-820b-8090ff6f5ca6\") " pod="openshift-marketplace/redhat-marketplace-mkqps" Dec 13 06:51:45 crc kubenswrapper[4971]: I1213 06:51:45.698532 4971 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/9d4f1ea5-8472-4320-820b-8090ff6f5ca6-utilities\") pod \"redhat-marketplace-mkqps\" (UID: \"9d4f1ea5-8472-4320-820b-8090ff6f5ca6\") " pod="openshift-marketplace/redhat-marketplace-mkqps" Dec 13 06:51:45 crc kubenswrapper[4971]: I1213 06:51:45.698657 4971 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/c1575ff9-1c53-4732-bc48-252f9f270597-config-volume" (OuterVolumeSpecName: "config-volume") pod "c1575ff9-1c53-4732-bc48-252f9f270597" (UID: "c1575ff9-1c53-4732-bc48-252f9f270597"). InnerVolumeSpecName "config-volume". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 13 06:51:45 crc kubenswrapper[4971]: W1213 06:51:45.702660 4971 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod22037bdc_fe6d_4411_be2b_aec38524a8dd.slice/crio-5c5265d52117bbfcb392ee1a7a528d5ed5ff1be0ce3712238a7ea4e6263303a3 WatchSource:0}: Error finding container 5c5265d52117bbfcb392ee1a7a528d5ed5ff1be0ce3712238a7ea4e6263303a3: Status 404 returned error can't find the container with id 5c5265d52117bbfcb392ee1a7a528d5ed5ff1be0ce3712238a7ea4e6263303a3 Dec 13 06:51:45 crc kubenswrapper[4971]: I1213 06:51:45.702675 4971 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/c1575ff9-1c53-4732-bc48-252f9f270597-secret-volume" (OuterVolumeSpecName: "secret-volume") pod "c1575ff9-1c53-4732-bc48-252f9f270597" (UID: "c1575ff9-1c53-4732-bc48-252f9f270597"). InnerVolumeSpecName "secret-volume". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 13 06:51:45 crc kubenswrapper[4971]: I1213 06:51:45.703219 4971 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/c1575ff9-1c53-4732-bc48-252f9f270597-kube-api-access-qdm7j" (OuterVolumeSpecName: "kube-api-access-qdm7j") pod "c1575ff9-1c53-4732-bc48-252f9f270597" (UID: "c1575ff9-1c53-4732-bc48-252f9f270597"). InnerVolumeSpecName "kube-api-access-qdm7j". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 13 06:51:45 crc kubenswrapper[4971]: I1213 06:51:45.717971 4971 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-npf8d\" (UniqueName: \"kubernetes.io/projected/9d4f1ea5-8472-4320-820b-8090ff6f5ca6-kube-api-access-npf8d\") pod \"redhat-marketplace-mkqps\" (UID: \"9d4f1ea5-8472-4320-820b-8090ff6f5ca6\") " pod="openshift-marketplace/redhat-marketplace-mkqps" Dec 13 06:51:45 crc kubenswrapper[4971]: I1213 06:51:45.780093 4971 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="8f668bae-612b-4b75-9490-919e737c6a3b" path="/var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes" Dec 13 06:51:45 crc kubenswrapper[4971]: I1213 06:51:45.799633 4971 reconciler_common.go:293] "Volume detached for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/c1575ff9-1c53-4732-bc48-252f9f270597-secret-volume\") on node \"crc\" DevicePath \"\"" Dec 13 06:51:45 crc kubenswrapper[4971]: I1213 06:51:45.799680 4971 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-qdm7j\" (UniqueName: \"kubernetes.io/projected/c1575ff9-1c53-4732-bc48-252f9f270597-kube-api-access-qdm7j\") on node \"crc\" DevicePath \"\"" Dec 13 06:51:45 crc kubenswrapper[4971]: I1213 06:51:45.799692 4971 reconciler_common.go:293] "Volume detached for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/c1575ff9-1c53-4732-bc48-252f9f270597-config-volume\") on node \"crc\" DevicePath \"\"" Dec 13 06:51:45 crc kubenswrapper[4971]: I1213 06:51:45.853830 4971 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-mkqps" Dec 13 06:51:45 crc kubenswrapper[4971]: I1213 06:51:45.980135 4971 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-kube-controller-manager/revision-pruner-9-crc"] Dec 13 06:51:45 crc kubenswrapper[4971]: E1213 06:51:45.982566 4971 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="c1575ff9-1c53-4732-bc48-252f9f270597" containerName="collect-profiles" Dec 13 06:51:45 crc kubenswrapper[4971]: I1213 06:51:45.982596 4971 state_mem.go:107] "Deleted CPUSet assignment" podUID="c1575ff9-1c53-4732-bc48-252f9f270597" containerName="collect-profiles" Dec 13 06:51:45 crc kubenswrapper[4971]: I1213 06:51:45.982858 4971 memory_manager.go:354] "RemoveStaleState removing state" podUID="c1575ff9-1c53-4732-bc48-252f9f270597" containerName="collect-profiles" Dec 13 06:51:45 crc kubenswrapper[4971]: I1213 06:51:45.983697 4971 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-controller-manager/revision-pruner-9-crc" Dec 13 06:51:45 crc kubenswrapper[4971]: I1213 06:51:45.987968 4971 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-controller-manager/revision-pruner-9-crc"] Dec 13 06:51:45 crc kubenswrapper[4971]: I1213 06:51:45.988426 4971 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-controller-manager"/"kube-root-ca.crt" Dec 13 06:51:45 crc kubenswrapper[4971]: I1213 06:51:45.988862 4971 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-controller-manager"/"installer-sa-dockercfg-kjl2n" Dec 13 06:51:46 crc kubenswrapper[4971]: I1213 06:51:46.005752 4971 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/e2829ff0-23f7-48af-8076-263273592801-kubelet-dir\") pod \"revision-pruner-9-crc\" (UID: \"e2829ff0-23f7-48af-8076-263273592801\") " pod="openshift-kube-controller-manager/revision-pruner-9-crc" Dec 13 06:51:46 crc kubenswrapper[4971]: I1213 06:51:46.005873 4971 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/e2829ff0-23f7-48af-8076-263273592801-kube-api-access\") pod \"revision-pruner-9-crc\" (UID: \"e2829ff0-23f7-48af-8076-263273592801\") " pod="openshift-kube-controller-manager/revision-pruner-9-crc" Dec 13 06:51:46 crc kubenswrapper[4971]: I1213 06:51:46.059108 4971 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-ingress/router-default-5444994796-mzhp6" Dec 13 06:51:46 crc kubenswrapper[4971]: I1213 06:51:46.063326 4971 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-mkqps"] Dec 13 06:51:46 crc kubenswrapper[4971]: I1213 06:51:46.063665 4971 patch_prober.go:28] interesting pod/router-default-5444994796-mzhp6 container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Dec 13 06:51:46 crc kubenswrapper[4971]: [-]has-synced failed: reason withheld Dec 13 06:51:46 crc kubenswrapper[4971]: [+]process-running ok Dec 13 06:51:46 crc kubenswrapper[4971]: healthz check failed Dec 13 06:51:46 crc kubenswrapper[4971]: I1213 06:51:46.063751 4971 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-mzhp6" podUID="d89e3583-3f66-4259-b6f9-bffe8ddbde12" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Dec 13 06:51:46 crc kubenswrapper[4971]: W1213 06:51:46.075915 4971 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod9d4f1ea5_8472_4320_820b_8090ff6f5ca6.slice/crio-26bef07a986f75c2981599189dc717ed62ed1b5ab6f7f1b51c940aa244522c9e WatchSource:0}: Error finding container 26bef07a986f75c2981599189dc717ed62ed1b5ab6f7f1b51c940aa244522c9e: Status 404 returned error can't find the container with id 26bef07a986f75c2981599189dc717ed62ed1b5ab6f7f1b51c940aa244522c9e Dec 13 06:51:46 crc kubenswrapper[4971]: I1213 06:51:46.107275 4971 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/e2829ff0-23f7-48af-8076-263273592801-kube-api-access\") pod \"revision-pruner-9-crc\" (UID: \"e2829ff0-23f7-48af-8076-263273592801\") " pod="openshift-kube-controller-manager/revision-pruner-9-crc" Dec 13 06:51:46 crc kubenswrapper[4971]: I1213 06:51:46.107378 4971 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/e2829ff0-23f7-48af-8076-263273592801-kubelet-dir\") pod \"revision-pruner-9-crc\" (UID: \"e2829ff0-23f7-48af-8076-263273592801\") " pod="openshift-kube-controller-manager/revision-pruner-9-crc" Dec 13 06:51:46 crc kubenswrapper[4971]: I1213 06:51:46.107568 4971 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/e2829ff0-23f7-48af-8076-263273592801-kubelet-dir\") pod \"revision-pruner-9-crc\" (UID: \"e2829ff0-23f7-48af-8076-263273592801\") " pod="openshift-kube-controller-manager/revision-pruner-9-crc" Dec 13 06:51:46 crc kubenswrapper[4971]: I1213 06:51:46.131777 4971 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/e2829ff0-23f7-48af-8076-263273592801-kube-api-access\") pod \"revision-pruner-9-crc\" (UID: \"e2829ff0-23f7-48af-8076-263273592801\") " pod="openshift-kube-controller-manager/revision-pruner-9-crc" Dec 13 06:51:46 crc kubenswrapper[4971]: I1213 06:51:46.155578 4971 patch_prober.go:28] interesting pod/machine-config-daemon-82xjz container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 13 06:51:46 crc kubenswrapper[4971]: I1213 06:51:46.155677 4971 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-82xjz" podUID="e0c4af12-edda-41d7-9b44-f87396174ef2" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 13 06:51:46 crc kubenswrapper[4971]: I1213 06:51:46.223985 4971 patch_prober.go:28] interesting pod/downloads-7954f5f757-dqk9z container/download-server namespace/openshift-console: Liveness probe status=failure output="Get \"http://10.217.0.20:8080/\": dial tcp 10.217.0.20:8080: connect: connection refused" start-of-body= Dec 13 06:51:46 crc kubenswrapper[4971]: I1213 06:51:46.224096 4971 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-console/downloads-7954f5f757-dqk9z" podUID="ee2e94f4-cadb-4b17-8853-a1292cb02404" containerName="download-server" probeResult="failure" output="Get \"http://10.217.0.20:8080/\": dial tcp 10.217.0.20:8080: connect: connection refused" Dec 13 06:51:46 crc kubenswrapper[4971]: I1213 06:51:46.224446 4971 patch_prober.go:28] interesting pod/downloads-7954f5f757-dqk9z container/download-server namespace/openshift-console: Readiness probe status=failure output="Get \"http://10.217.0.20:8080/\": dial tcp 10.217.0.20:8080: connect: connection refused" start-of-body= Dec 13 06:51:46 crc kubenswrapper[4971]: I1213 06:51:46.224504 4971 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-console/downloads-7954f5f757-dqk9z" podUID="ee2e94f4-cadb-4b17-8853-a1292cb02404" containerName="download-server" probeResult="failure" output="Get \"http://10.217.0.20:8080/\": dial tcp 10.217.0.20:8080: connect: connection refused" Dec 13 06:51:46 crc kubenswrapper[4971]: I1213 06:51:46.279879 4971 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-operators-d8n8p"] Dec 13 06:51:46 crc kubenswrapper[4971]: I1213 06:51:46.281631 4971 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-d8n8p" Dec 13 06:51:46 crc kubenswrapper[4971]: I1213 06:51:46.283460 4971 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"redhat-operators-dockercfg-ct8rh" Dec 13 06:51:46 crc kubenswrapper[4971]: I1213 06:51:46.294485 4971 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-d8n8p"] Dec 13 06:51:46 crc kubenswrapper[4971]: I1213 06:51:46.310732 4971 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/b9a0bc5d-3a39-4732-a849-22d351a4cc18-utilities\") pod \"redhat-operators-d8n8p\" (UID: \"b9a0bc5d-3a39-4732-a849-22d351a4cc18\") " pod="openshift-marketplace/redhat-operators-d8n8p" Dec 13 06:51:46 crc kubenswrapper[4971]: I1213 06:51:46.310917 4971 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/b9a0bc5d-3a39-4732-a849-22d351a4cc18-catalog-content\") pod \"redhat-operators-d8n8p\" (UID: \"b9a0bc5d-3a39-4732-a849-22d351a4cc18\") " pod="openshift-marketplace/redhat-operators-d8n8p" Dec 13 06:51:46 crc kubenswrapper[4971]: I1213 06:51:46.311126 4971 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-55kcz\" (UniqueName: \"kubernetes.io/projected/b9a0bc5d-3a39-4732-a849-22d351a4cc18-kube-api-access-55kcz\") pod \"redhat-operators-d8n8p\" (UID: \"b9a0bc5d-3a39-4732-a849-22d351a4cc18\") " pod="openshift-marketplace/redhat-operators-d8n8p" Dec 13 06:51:46 crc kubenswrapper[4971]: I1213 06:51:46.378768 4971 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29426805-fxxrz" event={"ID":"c1575ff9-1c53-4732-bc48-252f9f270597","Type":"ContainerDied","Data":"f938bdc1991e28af8004493352b892f7c2166554190ee959d89d1ec0f31229f0"} Dec 13 06:51:46 crc kubenswrapper[4971]: I1213 06:51:46.378808 4971 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="f938bdc1991e28af8004493352b892f7c2166554190ee959d89d1ec0f31229f0" Dec 13 06:51:46 crc kubenswrapper[4971]: I1213 06:51:46.378859 4971 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29426805-fxxrz" Dec 13 06:51:46 crc kubenswrapper[4971]: I1213 06:51:46.381256 4971 generic.go:334] "Generic (PLEG): container finished" podID="9d4f1ea5-8472-4320-820b-8090ff6f5ca6" containerID="4c8765305a845952aa8d2a0c9bb6e24513864afdd01cf9e5bbed8dd7ea15cf36" exitCode=0 Dec 13 06:51:46 crc kubenswrapper[4971]: I1213 06:51:46.381325 4971 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-mkqps" event={"ID":"9d4f1ea5-8472-4320-820b-8090ff6f5ca6","Type":"ContainerDied","Data":"4c8765305a845952aa8d2a0c9bb6e24513864afdd01cf9e5bbed8dd7ea15cf36"} Dec 13 06:51:46 crc kubenswrapper[4971]: I1213 06:51:46.381345 4971 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-mkqps" event={"ID":"9d4f1ea5-8472-4320-820b-8090ff6f5ca6","Type":"ContainerStarted","Data":"26bef07a986f75c2981599189dc717ed62ed1b5ab6f7f1b51c940aa244522c9e"} Dec 13 06:51:46 crc kubenswrapper[4971]: I1213 06:51:46.384130 4971 generic.go:334] "Generic (PLEG): container finished" podID="22037bdc-fe6d-4411-be2b-aec38524a8dd" containerID="a8cf7f8ba01f1f3be0caea6b89cddcbc28e056346def1520e8197eaa54f64016" exitCode=0 Dec 13 06:51:46 crc kubenswrapper[4971]: I1213 06:51:46.384203 4971 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-xwxlf" event={"ID":"22037bdc-fe6d-4411-be2b-aec38524a8dd","Type":"ContainerDied","Data":"a8cf7f8ba01f1f3be0caea6b89cddcbc28e056346def1520e8197eaa54f64016"} Dec 13 06:51:46 crc kubenswrapper[4971]: I1213 06:51:46.384228 4971 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-xwxlf" event={"ID":"22037bdc-fe6d-4411-be2b-aec38524a8dd","Type":"ContainerStarted","Data":"5c5265d52117bbfcb392ee1a7a528d5ed5ff1be0ce3712238a7ea4e6263303a3"} Dec 13 06:51:46 crc kubenswrapper[4971]: I1213 06:51:46.385172 4971 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-image-registry/image-registry-697d97f7c8-mdd8q" Dec 13 06:51:46 crc kubenswrapper[4971]: I1213 06:51:46.386186 4971 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-controller-manager/revision-pruner-9-crc" Dec 13 06:51:46 crc kubenswrapper[4971]: I1213 06:51:46.389337 4971 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-apiserver/apiserver-76f77b778f-2n4cq" Dec 13 06:51:46 crc kubenswrapper[4971]: I1213 06:51:46.415543 4971 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-55kcz\" (UniqueName: \"kubernetes.io/projected/b9a0bc5d-3a39-4732-a849-22d351a4cc18-kube-api-access-55kcz\") pod \"redhat-operators-d8n8p\" (UID: \"b9a0bc5d-3a39-4732-a849-22d351a4cc18\") " pod="openshift-marketplace/redhat-operators-d8n8p" Dec 13 06:51:46 crc kubenswrapper[4971]: I1213 06:51:46.415972 4971 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/b9a0bc5d-3a39-4732-a849-22d351a4cc18-utilities\") pod \"redhat-operators-d8n8p\" (UID: \"b9a0bc5d-3a39-4732-a849-22d351a4cc18\") " pod="openshift-marketplace/redhat-operators-d8n8p" Dec 13 06:51:46 crc kubenswrapper[4971]: I1213 06:51:46.416015 4971 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/b9a0bc5d-3a39-4732-a849-22d351a4cc18-catalog-content\") pod \"redhat-operators-d8n8p\" (UID: \"b9a0bc5d-3a39-4732-a849-22d351a4cc18\") " pod="openshift-marketplace/redhat-operators-d8n8p" Dec 13 06:51:46 crc kubenswrapper[4971]: I1213 06:51:46.417920 4971 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/b9a0bc5d-3a39-4732-a849-22d351a4cc18-utilities\") pod \"redhat-operators-d8n8p\" (UID: \"b9a0bc5d-3a39-4732-a849-22d351a4cc18\") " pod="openshift-marketplace/redhat-operators-d8n8p" Dec 13 06:51:46 crc kubenswrapper[4971]: I1213 06:51:46.418219 4971 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/b9a0bc5d-3a39-4732-a849-22d351a4cc18-catalog-content\") pod \"redhat-operators-d8n8p\" (UID: \"b9a0bc5d-3a39-4732-a849-22d351a4cc18\") " pod="openshift-marketplace/redhat-operators-d8n8p" Dec 13 06:51:46 crc kubenswrapper[4971]: I1213 06:51:46.440947 4971 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-55kcz\" (UniqueName: \"kubernetes.io/projected/b9a0bc5d-3a39-4732-a849-22d351a4cc18-kube-api-access-55kcz\") pod \"redhat-operators-d8n8p\" (UID: \"b9a0bc5d-3a39-4732-a849-22d351a4cc18\") " pod="openshift-marketplace/redhat-operators-d8n8p" Dec 13 06:51:46 crc kubenswrapper[4971]: I1213 06:51:46.602922 4971 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-d8n8p" Dec 13 06:51:46 crc kubenswrapper[4971]: I1213 06:51:46.619853 4971 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-kube-apiserver/revision-pruner-8-crc"] Dec 13 06:51:46 crc kubenswrapper[4971]: I1213 06:51:46.620540 4971 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/revision-pruner-8-crc" Dec 13 06:51:46 crc kubenswrapper[4971]: I1213 06:51:46.630118 4971 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-apiserver"/"kube-root-ca.crt" Dec 13 06:51:46 crc kubenswrapper[4971]: I1213 06:51:46.630415 4971 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-apiserver"/"installer-sa-dockercfg-5pr6n" Dec 13 06:51:46 crc kubenswrapper[4971]: I1213 06:51:46.631534 4971 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-apiserver/revision-pruner-8-crc"] Dec 13 06:51:46 crc kubenswrapper[4971]: I1213 06:51:46.696203 4971 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-operators-wpxfk"] Dec 13 06:51:46 crc kubenswrapper[4971]: I1213 06:51:46.698928 4971 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-wpxfk" Dec 13 06:51:46 crc kubenswrapper[4971]: I1213 06:51:46.702317 4971 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-wpxfk"] Dec 13 06:51:46 crc kubenswrapper[4971]: I1213 06:51:46.730139 4971 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/4c6bfe92-35ba-4a4f-a0b6-fb81683ca493-kube-api-access\") pod \"revision-pruner-8-crc\" (UID: \"4c6bfe92-35ba-4a4f-a0b6-fb81683ca493\") " pod="openshift-kube-apiserver/revision-pruner-8-crc" Dec 13 06:51:46 crc kubenswrapper[4971]: I1213 06:51:46.730210 4971 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/4c6bfe92-35ba-4a4f-a0b6-fb81683ca493-kubelet-dir\") pod \"revision-pruner-8-crc\" (UID: \"4c6bfe92-35ba-4a4f-a0b6-fb81683ca493\") " pod="openshift-kube-apiserver/revision-pruner-8-crc" Dec 13 06:51:46 crc kubenswrapper[4971]: I1213 06:51:46.730255 4971 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-d9kbv\" (UniqueName: \"kubernetes.io/projected/02f6e9f2-6bad-441c-8e2e-9acfc71f3ab4-kube-api-access-d9kbv\") pod \"redhat-operators-wpxfk\" (UID: \"02f6e9f2-6bad-441c-8e2e-9acfc71f3ab4\") " pod="openshift-marketplace/redhat-operators-wpxfk" Dec 13 06:51:46 crc kubenswrapper[4971]: I1213 06:51:46.730281 4971 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/02f6e9f2-6bad-441c-8e2e-9acfc71f3ab4-catalog-content\") pod \"redhat-operators-wpxfk\" (UID: \"02f6e9f2-6bad-441c-8e2e-9acfc71f3ab4\") " pod="openshift-marketplace/redhat-operators-wpxfk" Dec 13 06:51:46 crc kubenswrapper[4971]: I1213 06:51:46.730383 4971 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/02f6e9f2-6bad-441c-8e2e-9acfc71f3ab4-utilities\") pod \"redhat-operators-wpxfk\" (UID: \"02f6e9f2-6bad-441c-8e2e-9acfc71f3ab4\") " pod="openshift-marketplace/redhat-operators-wpxfk" Dec 13 06:51:46 crc kubenswrapper[4971]: I1213 06:51:46.831611 4971 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/02f6e9f2-6bad-441c-8e2e-9acfc71f3ab4-catalog-content\") pod \"redhat-operators-wpxfk\" (UID: \"02f6e9f2-6bad-441c-8e2e-9acfc71f3ab4\") " pod="openshift-marketplace/redhat-operators-wpxfk" Dec 13 06:51:46 crc kubenswrapper[4971]: I1213 06:51:46.831687 4971 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/02f6e9f2-6bad-441c-8e2e-9acfc71f3ab4-utilities\") pod \"redhat-operators-wpxfk\" (UID: \"02f6e9f2-6bad-441c-8e2e-9acfc71f3ab4\") " pod="openshift-marketplace/redhat-operators-wpxfk" Dec 13 06:51:46 crc kubenswrapper[4971]: I1213 06:51:46.831732 4971 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/4c6bfe92-35ba-4a4f-a0b6-fb81683ca493-kube-api-access\") pod \"revision-pruner-8-crc\" (UID: \"4c6bfe92-35ba-4a4f-a0b6-fb81683ca493\") " pod="openshift-kube-apiserver/revision-pruner-8-crc" Dec 13 06:51:46 crc kubenswrapper[4971]: I1213 06:51:46.831787 4971 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/4c6bfe92-35ba-4a4f-a0b6-fb81683ca493-kubelet-dir\") pod \"revision-pruner-8-crc\" (UID: \"4c6bfe92-35ba-4a4f-a0b6-fb81683ca493\") " pod="openshift-kube-apiserver/revision-pruner-8-crc" Dec 13 06:51:46 crc kubenswrapper[4971]: I1213 06:51:46.831818 4971 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-d9kbv\" (UniqueName: \"kubernetes.io/projected/02f6e9f2-6bad-441c-8e2e-9acfc71f3ab4-kube-api-access-d9kbv\") pod \"redhat-operators-wpxfk\" (UID: \"02f6e9f2-6bad-441c-8e2e-9acfc71f3ab4\") " pod="openshift-marketplace/redhat-operators-wpxfk" Dec 13 06:51:46 crc kubenswrapper[4971]: I1213 06:51:46.832657 4971 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/02f6e9f2-6bad-441c-8e2e-9acfc71f3ab4-catalog-content\") pod \"redhat-operators-wpxfk\" (UID: \"02f6e9f2-6bad-441c-8e2e-9acfc71f3ab4\") " pod="openshift-marketplace/redhat-operators-wpxfk" Dec 13 06:51:46 crc kubenswrapper[4971]: I1213 06:51:46.832922 4971 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/02f6e9f2-6bad-441c-8e2e-9acfc71f3ab4-utilities\") pod \"redhat-operators-wpxfk\" (UID: \"02f6e9f2-6bad-441c-8e2e-9acfc71f3ab4\") " pod="openshift-marketplace/redhat-operators-wpxfk" Dec 13 06:51:46 crc kubenswrapper[4971]: I1213 06:51:46.833130 4971 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/4c6bfe92-35ba-4a4f-a0b6-fb81683ca493-kubelet-dir\") pod \"revision-pruner-8-crc\" (UID: \"4c6bfe92-35ba-4a4f-a0b6-fb81683ca493\") " pod="openshift-kube-apiserver/revision-pruner-8-crc" Dec 13 06:51:46 crc kubenswrapper[4971]: I1213 06:51:46.852050 4971 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/4c6bfe92-35ba-4a4f-a0b6-fb81683ca493-kube-api-access\") pod \"revision-pruner-8-crc\" (UID: \"4c6bfe92-35ba-4a4f-a0b6-fb81683ca493\") " pod="openshift-kube-apiserver/revision-pruner-8-crc" Dec 13 06:51:46 crc kubenswrapper[4971]: I1213 06:51:46.859368 4971 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-d9kbv\" (UniqueName: \"kubernetes.io/projected/02f6e9f2-6bad-441c-8e2e-9acfc71f3ab4-kube-api-access-d9kbv\") pod \"redhat-operators-wpxfk\" (UID: \"02f6e9f2-6bad-441c-8e2e-9acfc71f3ab4\") " pod="openshift-marketplace/redhat-operators-wpxfk" Dec 13 06:51:46 crc kubenswrapper[4971]: I1213 06:51:46.943486 4971 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-controller-manager/revision-pruner-9-crc"] Dec 13 06:51:46 crc kubenswrapper[4971]: I1213 06:51:46.986672 4971 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-d8n8p"] Dec 13 06:51:46 crc kubenswrapper[4971]: I1213 06:51:46.988245 4971 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/revision-pruner-8-crc" Dec 13 06:51:47 crc kubenswrapper[4971]: I1213 06:51:47.011083 4971 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-wpxfk" Dec 13 06:51:47 crc kubenswrapper[4971]: W1213 06:51:47.053671 4971 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podb9a0bc5d_3a39_4732_a849_22d351a4cc18.slice/crio-a4a3ad87df8940b34556eacf79e1d8f7f5e82ac7fba0af8aaea503b398caa4cb WatchSource:0}: Error finding container a4a3ad87df8940b34556eacf79e1d8f7f5e82ac7fba0af8aaea503b398caa4cb: Status 404 returned error can't find the container with id a4a3ad87df8940b34556eacf79e1d8f7f5e82ac7fba0af8aaea503b398caa4cb Dec 13 06:51:47 crc kubenswrapper[4971]: I1213 06:51:47.064864 4971 patch_prober.go:28] interesting pod/router-default-5444994796-mzhp6 container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Dec 13 06:51:47 crc kubenswrapper[4971]: [-]has-synced failed: reason withheld Dec 13 06:51:47 crc kubenswrapper[4971]: [+]process-running ok Dec 13 06:51:47 crc kubenswrapper[4971]: healthz check failed Dec 13 06:51:47 crc kubenswrapper[4971]: I1213 06:51:47.064919 4971 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-mzhp6" podUID="d89e3583-3f66-4259-b6f9-bffe8ddbde12" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Dec 13 06:51:47 crc kubenswrapper[4971]: I1213 06:51:47.433416 4971 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-controller-manager/revision-pruner-9-crc" event={"ID":"e2829ff0-23f7-48af-8076-263273592801","Type":"ContainerStarted","Data":"1ccc0fc9e685df94740d400e18b12d0881b1bae5be0a26f14253a658c14e218b"} Dec 13 06:51:47 crc kubenswrapper[4971]: I1213 06:51:47.436994 4971 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-d8n8p" event={"ID":"b9a0bc5d-3a39-4732-a849-22d351a4cc18","Type":"ContainerStarted","Data":"a4a3ad87df8940b34556eacf79e1d8f7f5e82ac7fba0af8aaea503b398caa4cb"} Dec 13 06:51:47 crc kubenswrapper[4971]: I1213 06:51:47.449915 4971 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/c7185dec-3240-42e8-9d90-6182a5e8b65c-metrics-certs\") pod \"network-metrics-daemon-x98k7\" (UID: \"c7185dec-3240-42e8-9d90-6182a5e8b65c\") " pod="openshift-multus/network-metrics-daemon-x98k7" Dec 13 06:51:47 crc kubenswrapper[4971]: I1213 06:51:47.456354 4971 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/c7185dec-3240-42e8-9d90-6182a5e8b65c-metrics-certs\") pod \"network-metrics-daemon-x98k7\" (UID: \"c7185dec-3240-42e8-9d90-6182a5e8b65c\") " pod="openshift-multus/network-metrics-daemon-x98k7" Dec 13 06:51:47 crc kubenswrapper[4971]: I1213 06:51:47.477764 4971 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-wpxfk"] Dec 13 06:51:47 crc kubenswrapper[4971]: W1213 06:51:47.508783 4971 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod02f6e9f2_6bad_441c_8e2e_9acfc71f3ab4.slice/crio-63c622776bc727571794ee6aabc7d9588866c6f1e1a30002c0e83865eaaa3bac WatchSource:0}: Error finding container 63c622776bc727571794ee6aabc7d9588866c6f1e1a30002c0e83865eaaa3bac: Status 404 returned error can't find the container with id 63c622776bc727571794ee6aabc7d9588866c6f1e1a30002c0e83865eaaa3bac Dec 13 06:51:47 crc kubenswrapper[4971]: I1213 06:51:47.588905 4971 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-apiserver/revision-pruner-8-crc"] Dec 13 06:51:47 crc kubenswrapper[4971]: W1213 06:51:47.617755 4971 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-pod4c6bfe92_35ba_4a4f_a0b6_fb81683ca493.slice/crio-f1fed253cd0500dab2d408394ce07c796d9eab875d6dc4ff0551cfe41954c904 WatchSource:0}: Error finding container f1fed253cd0500dab2d408394ce07c796d9eab875d6dc4ff0551cfe41954c904: Status 404 returned error can't find the container with id f1fed253cd0500dab2d408394ce07c796d9eab875d6dc4ff0551cfe41954c904 Dec 13 06:51:47 crc kubenswrapper[4971]: I1213 06:51:47.694268 4971 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-x98k7" Dec 13 06:51:48 crc kubenswrapper[4971]: I1213 06:51:48.060128 4971 patch_prober.go:28] interesting pod/router-default-5444994796-mzhp6 container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Dec 13 06:51:48 crc kubenswrapper[4971]: [-]has-synced failed: reason withheld Dec 13 06:51:48 crc kubenswrapper[4971]: [+]process-running ok Dec 13 06:51:48 crc kubenswrapper[4971]: healthz check failed Dec 13 06:51:48 crc kubenswrapper[4971]: I1213 06:51:48.060372 4971 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-mzhp6" podUID="d89e3583-3f66-4259-b6f9-bffe8ddbde12" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Dec 13 06:51:48 crc kubenswrapper[4971]: I1213 06:51:48.119604 4971 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-multus/network-metrics-daemon-x98k7"] Dec 13 06:51:48 crc kubenswrapper[4971]: W1213 06:51:48.264306 4971 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podc7185dec_3240_42e8_9d90_6182a5e8b65c.slice/crio-46226de139f8f6bcc91eb8671f612526959fe9d19c3cbf1bb73dbd73581bf5f7 WatchSource:0}: Error finding container 46226de139f8f6bcc91eb8671f612526959fe9d19c3cbf1bb73dbd73581bf5f7: Status 404 returned error can't find the container with id 46226de139f8f6bcc91eb8671f612526959fe9d19c3cbf1bb73dbd73581bf5f7 Dec 13 06:51:48 crc kubenswrapper[4971]: I1213 06:51:48.469507 4971 generic.go:334] "Generic (PLEG): container finished" podID="b9a0bc5d-3a39-4732-a849-22d351a4cc18" containerID="1d8c1c7524f867e2c7ec3ccc601f78308c890a37019036f24c1eb839af11281b" exitCode=0 Dec 13 06:51:48 crc kubenswrapper[4971]: I1213 06:51:48.469916 4971 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-d8n8p" event={"ID":"b9a0bc5d-3a39-4732-a849-22d351a4cc18","Type":"ContainerDied","Data":"1d8c1c7524f867e2c7ec3ccc601f78308c890a37019036f24c1eb839af11281b"} Dec 13 06:51:48 crc kubenswrapper[4971]: I1213 06:51:48.486008 4971 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/revision-pruner-8-crc" event={"ID":"4c6bfe92-35ba-4a4f-a0b6-fb81683ca493","Type":"ContainerStarted","Data":"f1fed253cd0500dab2d408394ce07c796d9eab875d6dc4ff0551cfe41954c904"} Dec 13 06:51:48 crc kubenswrapper[4971]: I1213 06:51:48.497303 4971 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-controller-manager/revision-pruner-9-crc" event={"ID":"e2829ff0-23f7-48af-8076-263273592801","Type":"ContainerStarted","Data":"836a15cbba35f8a401a33be13f762496065df415ad5ce3396968733b6bcc1b74"} Dec 13 06:51:48 crc kubenswrapper[4971]: I1213 06:51:48.499082 4971 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/network-metrics-daemon-x98k7" event={"ID":"c7185dec-3240-42e8-9d90-6182a5e8b65c","Type":"ContainerStarted","Data":"46226de139f8f6bcc91eb8671f612526959fe9d19c3cbf1bb73dbd73581bf5f7"} Dec 13 06:51:48 crc kubenswrapper[4971]: I1213 06:51:48.504270 4971 generic.go:334] "Generic (PLEG): container finished" podID="02f6e9f2-6bad-441c-8e2e-9acfc71f3ab4" containerID="970df14f0e90612c6aee36fde7557ac705f1ce238a087fd366a6b85f623dba26" exitCode=0 Dec 13 06:51:48 crc kubenswrapper[4971]: I1213 06:51:48.504319 4971 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-wpxfk" event={"ID":"02f6e9f2-6bad-441c-8e2e-9acfc71f3ab4","Type":"ContainerDied","Data":"970df14f0e90612c6aee36fde7557ac705f1ce238a087fd366a6b85f623dba26"} Dec 13 06:51:48 crc kubenswrapper[4971]: I1213 06:51:48.504346 4971 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-wpxfk" event={"ID":"02f6e9f2-6bad-441c-8e2e-9acfc71f3ab4","Type":"ContainerStarted","Data":"63c622776bc727571794ee6aabc7d9588866c6f1e1a30002c0e83865eaaa3bac"} Dec 13 06:51:48 crc kubenswrapper[4971]: I1213 06:51:48.516539 4971 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-kube-controller-manager/revision-pruner-9-crc" podStartSLOduration=3.516455132 podStartE2EDuration="3.516455132s" podCreationTimestamp="2025-12-13 06:51:45 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-13 06:51:48.514306988 +0000 UTC m=+165.118716446" watchObservedRunningTime="2025-12-13 06:51:48.516455132 +0000 UTC m=+165.120864580" Dec 13 06:51:49 crc kubenswrapper[4971]: I1213 06:51:49.061204 4971 patch_prober.go:28] interesting pod/router-default-5444994796-mzhp6 container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Dec 13 06:51:49 crc kubenswrapper[4971]: [-]has-synced failed: reason withheld Dec 13 06:51:49 crc kubenswrapper[4971]: [+]process-running ok Dec 13 06:51:49 crc kubenswrapper[4971]: healthz check failed Dec 13 06:51:49 crc kubenswrapper[4971]: I1213 06:51:49.061268 4971 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-mzhp6" podUID="d89e3583-3f66-4259-b6f9-bffe8ddbde12" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Dec 13 06:51:49 crc kubenswrapper[4971]: I1213 06:51:49.515593 4971 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/revision-pruner-8-crc" event={"ID":"4c6bfe92-35ba-4a4f-a0b6-fb81683ca493","Type":"ContainerStarted","Data":"9e825f10a21f625f184e4d35bbfee6dfdaabd8018ba052abb3a959c5b5c4e936"} Dec 13 06:51:49 crc kubenswrapper[4971]: I1213 06:51:49.520535 4971 generic.go:334] "Generic (PLEG): container finished" podID="e2829ff0-23f7-48af-8076-263273592801" containerID="836a15cbba35f8a401a33be13f762496065df415ad5ce3396968733b6bcc1b74" exitCode=0 Dec 13 06:51:49 crc kubenswrapper[4971]: I1213 06:51:49.520635 4971 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-controller-manager/revision-pruner-9-crc" event={"ID":"e2829ff0-23f7-48af-8076-263273592801","Type":"ContainerDied","Data":"836a15cbba35f8a401a33be13f762496065df415ad5ce3396968733b6bcc1b74"} Dec 13 06:51:49 crc kubenswrapper[4971]: I1213 06:51:49.529752 4971 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-kube-apiserver/revision-pruner-8-crc" podStartSLOduration=3.529729751 podStartE2EDuration="3.529729751s" podCreationTimestamp="2025-12-13 06:51:46 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-13 06:51:49.529410333 +0000 UTC m=+166.133819781" watchObservedRunningTime="2025-12-13 06:51:49.529729751 +0000 UTC m=+166.134139209" Dec 13 06:51:50 crc kubenswrapper[4971]: I1213 06:51:50.060424 4971 patch_prober.go:28] interesting pod/router-default-5444994796-mzhp6 container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Dec 13 06:51:50 crc kubenswrapper[4971]: [-]has-synced failed: reason withheld Dec 13 06:51:50 crc kubenswrapper[4971]: [+]process-running ok Dec 13 06:51:50 crc kubenswrapper[4971]: healthz check failed Dec 13 06:51:50 crc kubenswrapper[4971]: I1213 06:51:50.060859 4971 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-mzhp6" podUID="d89e3583-3f66-4259-b6f9-bffe8ddbde12" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Dec 13 06:51:50 crc kubenswrapper[4971]: I1213 06:51:50.639782 4971 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/network-metrics-daemon-x98k7" event={"ID":"c7185dec-3240-42e8-9d90-6182a5e8b65c","Type":"ContainerStarted","Data":"efec61c5ae1f4bfd1318104d35622463e7e6748897830007ff0898e8ee98e0f2"} Dec 13 06:51:50 crc kubenswrapper[4971]: I1213 06:51:50.655347 4971 generic.go:334] "Generic (PLEG): container finished" podID="4c6bfe92-35ba-4a4f-a0b6-fb81683ca493" containerID="9e825f10a21f625f184e4d35bbfee6dfdaabd8018ba052abb3a959c5b5c4e936" exitCode=0 Dec 13 06:51:50 crc kubenswrapper[4971]: I1213 06:51:50.655889 4971 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/revision-pruner-8-crc" event={"ID":"4c6bfe92-35ba-4a4f-a0b6-fb81683ca493","Type":"ContainerDied","Data":"9e825f10a21f625f184e4d35bbfee6dfdaabd8018ba052abb3a959c5b5c4e936"} Dec 13 06:51:51 crc kubenswrapper[4971]: I1213 06:51:51.068214 4971 patch_prober.go:28] interesting pod/router-default-5444994796-mzhp6 container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Dec 13 06:51:51 crc kubenswrapper[4971]: [-]has-synced failed: reason withheld Dec 13 06:51:51 crc kubenswrapper[4971]: [+]process-running ok Dec 13 06:51:51 crc kubenswrapper[4971]: healthz check failed Dec 13 06:51:51 crc kubenswrapper[4971]: I1213 06:51:51.068453 4971 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-mzhp6" podUID="d89e3583-3f66-4259-b6f9-bffe8ddbde12" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Dec 13 06:51:51 crc kubenswrapper[4971]: I1213 06:51:51.214583 4971 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-kube-controller-manager/revision-pruner-9-crc" Dec 13 06:51:51 crc kubenswrapper[4971]: I1213 06:51:51.307960 4971 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/e2829ff0-23f7-48af-8076-263273592801-kube-api-access\") pod \"e2829ff0-23f7-48af-8076-263273592801\" (UID: \"e2829ff0-23f7-48af-8076-263273592801\") " Dec 13 06:51:51 crc kubenswrapper[4971]: I1213 06:51:51.308044 4971 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/e2829ff0-23f7-48af-8076-263273592801-kubelet-dir\") pod \"e2829ff0-23f7-48af-8076-263273592801\" (UID: \"e2829ff0-23f7-48af-8076-263273592801\") " Dec 13 06:51:51 crc kubenswrapper[4971]: I1213 06:51:51.308278 4971 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/e2829ff0-23f7-48af-8076-263273592801-kubelet-dir" (OuterVolumeSpecName: "kubelet-dir") pod "e2829ff0-23f7-48af-8076-263273592801" (UID: "e2829ff0-23f7-48af-8076-263273592801"). InnerVolumeSpecName "kubelet-dir". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 13 06:51:51 crc kubenswrapper[4971]: I1213 06:51:51.316781 4971 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/e2829ff0-23f7-48af-8076-263273592801-kube-api-access" (OuterVolumeSpecName: "kube-api-access") pod "e2829ff0-23f7-48af-8076-263273592801" (UID: "e2829ff0-23f7-48af-8076-263273592801"). InnerVolumeSpecName "kube-api-access". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 13 06:51:51 crc kubenswrapper[4971]: I1213 06:51:51.409638 4971 reconciler_common.go:293] "Volume detached for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/e2829ff0-23f7-48af-8076-263273592801-kubelet-dir\") on node \"crc\" DevicePath \"\"" Dec 13 06:51:51 crc kubenswrapper[4971]: I1213 06:51:51.409933 4971 reconciler_common.go:293] "Volume detached for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/e2829ff0-23f7-48af-8076-263273592801-kube-api-access\") on node \"crc\" DevicePath \"\"" Dec 13 06:51:51 crc kubenswrapper[4971]: I1213 06:51:51.508828 4971 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-dns/dns-default-dntmz" Dec 13 06:51:51 crc kubenswrapper[4971]: I1213 06:51:51.670539 4971 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-kube-controller-manager/revision-pruner-9-crc" Dec 13 06:51:51 crc kubenswrapper[4971]: I1213 06:51:51.670571 4971 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-controller-manager/revision-pruner-9-crc" event={"ID":"e2829ff0-23f7-48af-8076-263273592801","Type":"ContainerDied","Data":"1ccc0fc9e685df94740d400e18b12d0881b1bae5be0a26f14253a658c14e218b"} Dec 13 06:51:51 crc kubenswrapper[4971]: I1213 06:51:51.670617 4971 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="1ccc0fc9e685df94740d400e18b12d0881b1bae5be0a26f14253a658c14e218b" Dec 13 06:51:51 crc kubenswrapper[4971]: I1213 06:51:51.674886 4971 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/network-metrics-daemon-x98k7" event={"ID":"c7185dec-3240-42e8-9d90-6182a5e8b65c","Type":"ContainerStarted","Data":"b114fa9edd5d2a79a52d15e8f3c3fe30688b98f00c4abd5f4c94d3cb8b85192e"} Dec 13 06:51:52 crc kubenswrapper[4971]: I1213 06:51:52.053438 4971 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/revision-pruner-8-crc" Dec 13 06:51:52 crc kubenswrapper[4971]: I1213 06:51:52.061468 4971 patch_prober.go:28] interesting pod/router-default-5444994796-mzhp6 container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Dec 13 06:51:52 crc kubenswrapper[4971]: [+]has-synced ok Dec 13 06:51:52 crc kubenswrapper[4971]: [+]process-running ok Dec 13 06:51:52 crc kubenswrapper[4971]: healthz check failed Dec 13 06:51:52 crc kubenswrapper[4971]: I1213 06:51:52.061544 4971 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-mzhp6" podUID="d89e3583-3f66-4259-b6f9-bffe8ddbde12" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Dec 13 06:51:52 crc kubenswrapper[4971]: I1213 06:51:52.228663 4971 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/4c6bfe92-35ba-4a4f-a0b6-fb81683ca493-kubelet-dir\") pod \"4c6bfe92-35ba-4a4f-a0b6-fb81683ca493\" (UID: \"4c6bfe92-35ba-4a4f-a0b6-fb81683ca493\") " Dec 13 06:51:52 crc kubenswrapper[4971]: I1213 06:51:52.228740 4971 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/4c6bfe92-35ba-4a4f-a0b6-fb81683ca493-kubelet-dir" (OuterVolumeSpecName: "kubelet-dir") pod "4c6bfe92-35ba-4a4f-a0b6-fb81683ca493" (UID: "4c6bfe92-35ba-4a4f-a0b6-fb81683ca493"). InnerVolumeSpecName "kubelet-dir". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 13 06:51:52 crc kubenswrapper[4971]: I1213 06:51:52.229825 4971 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/4c6bfe92-35ba-4a4f-a0b6-fb81683ca493-kube-api-access\") pod \"4c6bfe92-35ba-4a4f-a0b6-fb81683ca493\" (UID: \"4c6bfe92-35ba-4a4f-a0b6-fb81683ca493\") " Dec 13 06:51:52 crc kubenswrapper[4971]: I1213 06:51:52.230158 4971 reconciler_common.go:293] "Volume detached for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/4c6bfe92-35ba-4a4f-a0b6-fb81683ca493-kubelet-dir\") on node \"crc\" DevicePath \"\"" Dec 13 06:51:52 crc kubenswrapper[4971]: I1213 06:51:52.237124 4971 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/4c6bfe92-35ba-4a4f-a0b6-fb81683ca493-kube-api-access" (OuterVolumeSpecName: "kube-api-access") pod "4c6bfe92-35ba-4a4f-a0b6-fb81683ca493" (UID: "4c6bfe92-35ba-4a4f-a0b6-fb81683ca493"). InnerVolumeSpecName "kube-api-access". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 13 06:51:52 crc kubenswrapper[4971]: I1213 06:51:52.331990 4971 reconciler_common.go:293] "Volume detached for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/4c6bfe92-35ba-4a4f-a0b6-fb81683ca493-kube-api-access\") on node \"crc\" DevicePath \"\"" Dec 13 06:51:52 crc kubenswrapper[4971]: I1213 06:51:52.683435 4971 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/revision-pruner-8-crc" event={"ID":"4c6bfe92-35ba-4a4f-a0b6-fb81683ca493","Type":"ContainerDied","Data":"f1fed253cd0500dab2d408394ce07c796d9eab875d6dc4ff0551cfe41954c904"} Dec 13 06:51:52 crc kubenswrapper[4971]: I1213 06:51:52.683478 4971 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="f1fed253cd0500dab2d408394ce07c796d9eab875d6dc4ff0551cfe41954c904" Dec 13 06:51:52 crc kubenswrapper[4971]: I1213 06:51:52.683456 4971 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/revision-pruner-8-crc" Dec 13 06:51:52 crc kubenswrapper[4971]: I1213 06:51:52.703894 4971 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-multus/network-metrics-daemon-x98k7" podStartSLOduration=147.703856955 podStartE2EDuration="2m27.703856955s" podCreationTimestamp="2025-12-13 06:49:25 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-13 06:51:52.700285424 +0000 UTC m=+169.304694882" watchObservedRunningTime="2025-12-13 06:51:52.703856955 +0000 UTC m=+169.308266403" Dec 13 06:51:53 crc kubenswrapper[4971]: I1213 06:51:53.059602 4971 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-ingress/router-default-5444994796-mzhp6" Dec 13 06:51:53 crc kubenswrapper[4971]: I1213 06:51:53.061877 4971 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-ingress/router-default-5444994796-mzhp6" Dec 13 06:51:55 crc kubenswrapper[4971]: I1213 06:51:55.228556 4971 patch_prober.go:28] interesting pod/console-f9d7485db-5fd9b container/console namespace/openshift-console: Startup probe status=failure output="Get \"https://10.217.0.10:8443/health\": dial tcp 10.217.0.10:8443: connect: connection refused" start-of-body= Dec 13 06:51:55 crc kubenswrapper[4971]: I1213 06:51:55.228955 4971 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-console/console-f9d7485db-5fd9b" podUID="9f0db656-7a94-415b-b737-30ed342817ee" containerName="console" probeResult="failure" output="Get \"https://10.217.0.10:8443/health\": dial tcp 10.217.0.10:8443: connect: connection refused" Dec 13 06:51:56 crc kubenswrapper[4971]: I1213 06:51:56.238046 4971 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-console/downloads-7954f5f757-dqk9z" Dec 13 06:52:04 crc kubenswrapper[4971]: I1213 06:52:04.350733 4971 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-image-registry/image-registry-697d97f7c8-mdd8q" Dec 13 06:52:05 crc kubenswrapper[4971]: I1213 06:52:05.230920 4971 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-console/console-f9d7485db-5fd9b" Dec 13 06:52:05 crc kubenswrapper[4971]: I1213 06:52:05.238259 4971 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-console/console-f9d7485db-5fd9b" Dec 13 06:52:11 crc kubenswrapper[4971]: I1213 06:52:11.896820 4971 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 13 06:52:16 crc kubenswrapper[4971]: I1213 06:52:16.154125 4971 patch_prober.go:28] interesting pod/machine-config-daemon-82xjz container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 13 06:52:16 crc kubenswrapper[4971]: I1213 06:52:16.154593 4971 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-82xjz" podUID="e0c4af12-edda-41d7-9b44-f87396174ef2" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 13 06:52:16 crc kubenswrapper[4971]: I1213 06:52:16.334552 4971 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-g8vfp" Dec 13 06:52:23 crc kubenswrapper[4971]: E1213 06:52:23.079436 4971 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled" image="registry.redhat.io/redhat/certified-operator-index:v4.18" Dec 13 06:52:23 crc kubenswrapper[4971]: E1213 06:52:23.080097 4971 kuberuntime_manager.go:1274] "Unhandled Error" err="init container &Container{Name:extract-content,Image:registry.redhat.io/redhat/certified-operator-index:v4.18,Command:[/utilities/copy-content],Args:[--catalog.from=/configs --catalog.to=/extracted-catalog/catalog --cache.from=/tmp/cache --cache.to=/extracted-catalog/cache],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:utilities,ReadOnly:false,MountPath:/utilities,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:catalog-content,ReadOnly:false,MountPath:/extracted-catalog,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-stj24,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:Always,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*1000170000,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:FallbackToLogsOnError,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod certified-operators-rkd9c_openshift-marketplace(8f50f5ba-3ec3-4fa2-9f5d-51e1c43f6352): ErrImagePull: rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled" logger="UnhandledError" Dec 13 06:52:23 crc kubenswrapper[4971]: E1213 06:52:23.081267 4971 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ErrImagePull: \"rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled\"" pod="openshift-marketplace/certified-operators-rkd9c" podUID="8f50f5ba-3ec3-4fa2-9f5d-51e1c43f6352" Dec 13 06:52:23 crc kubenswrapper[4971]: E1213 06:52:23.114034 4971 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled" image="registry.redhat.io/redhat/certified-operator-index:v4.18" Dec 13 06:52:23 crc kubenswrapper[4971]: E1213 06:52:23.114233 4971 kuberuntime_manager.go:1274] "Unhandled Error" err="init container &Container{Name:extract-content,Image:registry.redhat.io/redhat/certified-operator-index:v4.18,Command:[/utilities/copy-content],Args:[--catalog.from=/configs --catalog.to=/extracted-catalog/catalog --cache.from=/tmp/cache --cache.to=/extracted-catalog/cache],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:utilities,ReadOnly:false,MountPath:/utilities,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:catalog-content,ReadOnly:false,MountPath:/extracted-catalog,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-wjf2d,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:Always,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*1000170000,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:FallbackToLogsOnError,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod certified-operators-82f6m_openshift-marketplace(9404e8eb-0f18-4a49-a716-f076273e7179): ErrImagePull: rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled" logger="UnhandledError" Dec 13 06:52:23 crc kubenswrapper[4971]: E1213 06:52:23.115424 4971 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ErrImagePull: \"rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled\"" pod="openshift-marketplace/certified-operators-82f6m" podUID="9404e8eb-0f18-4a49-a716-f076273e7179" Dec 13 06:52:23 crc kubenswrapper[4971]: I1213 06:52:23.764199 4971 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-kube-apiserver/revision-pruner-9-crc"] Dec 13 06:52:23 crc kubenswrapper[4971]: E1213 06:52:23.764506 4971 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="e2829ff0-23f7-48af-8076-263273592801" containerName="pruner" Dec 13 06:52:23 crc kubenswrapper[4971]: I1213 06:52:23.764537 4971 state_mem.go:107] "Deleted CPUSet assignment" podUID="e2829ff0-23f7-48af-8076-263273592801" containerName="pruner" Dec 13 06:52:23 crc kubenswrapper[4971]: E1213 06:52:23.764551 4971 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="4c6bfe92-35ba-4a4f-a0b6-fb81683ca493" containerName="pruner" Dec 13 06:52:23 crc kubenswrapper[4971]: I1213 06:52:23.764557 4971 state_mem.go:107] "Deleted CPUSet assignment" podUID="4c6bfe92-35ba-4a4f-a0b6-fb81683ca493" containerName="pruner" Dec 13 06:52:23 crc kubenswrapper[4971]: I1213 06:52:23.764664 4971 memory_manager.go:354] "RemoveStaleState removing state" podUID="e2829ff0-23f7-48af-8076-263273592801" containerName="pruner" Dec 13 06:52:23 crc kubenswrapper[4971]: I1213 06:52:23.764675 4971 memory_manager.go:354] "RemoveStaleState removing state" podUID="4c6bfe92-35ba-4a4f-a0b6-fb81683ca493" containerName="pruner" Dec 13 06:52:23 crc kubenswrapper[4971]: I1213 06:52:23.765091 4971 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/revision-pruner-9-crc" Dec 13 06:52:23 crc kubenswrapper[4971]: I1213 06:52:23.767485 4971 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-apiserver"/"installer-sa-dockercfg-5pr6n" Dec 13 06:52:23 crc kubenswrapper[4971]: I1213 06:52:23.768321 4971 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-apiserver"/"kube-root-ca.crt" Dec 13 06:52:23 crc kubenswrapper[4971]: I1213 06:52:23.789927 4971 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-apiserver/revision-pruner-9-crc"] Dec 13 06:52:23 crc kubenswrapper[4971]: I1213 06:52:23.962933 4971 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/fb460f16-f1e2-4ad7-a4bb-a2f5ef0cfe7a-kubelet-dir\") pod \"revision-pruner-9-crc\" (UID: \"fb460f16-f1e2-4ad7-a4bb-a2f5ef0cfe7a\") " pod="openshift-kube-apiserver/revision-pruner-9-crc" Dec 13 06:52:23 crc kubenswrapper[4971]: I1213 06:52:23.963010 4971 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/fb460f16-f1e2-4ad7-a4bb-a2f5ef0cfe7a-kube-api-access\") pod \"revision-pruner-9-crc\" (UID: \"fb460f16-f1e2-4ad7-a4bb-a2f5ef0cfe7a\") " pod="openshift-kube-apiserver/revision-pruner-9-crc" Dec 13 06:52:24 crc kubenswrapper[4971]: I1213 06:52:24.065401 4971 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/fb460f16-f1e2-4ad7-a4bb-a2f5ef0cfe7a-kubelet-dir\") pod \"revision-pruner-9-crc\" (UID: \"fb460f16-f1e2-4ad7-a4bb-a2f5ef0cfe7a\") " pod="openshift-kube-apiserver/revision-pruner-9-crc" Dec 13 06:52:24 crc kubenswrapper[4971]: I1213 06:52:24.065496 4971 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/fb460f16-f1e2-4ad7-a4bb-a2f5ef0cfe7a-kube-api-access\") pod \"revision-pruner-9-crc\" (UID: \"fb460f16-f1e2-4ad7-a4bb-a2f5ef0cfe7a\") " pod="openshift-kube-apiserver/revision-pruner-9-crc" Dec 13 06:52:24 crc kubenswrapper[4971]: I1213 06:52:24.065534 4971 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/fb460f16-f1e2-4ad7-a4bb-a2f5ef0cfe7a-kubelet-dir\") pod \"revision-pruner-9-crc\" (UID: \"fb460f16-f1e2-4ad7-a4bb-a2f5ef0cfe7a\") " pod="openshift-kube-apiserver/revision-pruner-9-crc" Dec 13 06:52:24 crc kubenswrapper[4971]: I1213 06:52:24.089878 4971 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/fb460f16-f1e2-4ad7-a4bb-a2f5ef0cfe7a-kube-api-access\") pod \"revision-pruner-9-crc\" (UID: \"fb460f16-f1e2-4ad7-a4bb-a2f5ef0cfe7a\") " pod="openshift-kube-apiserver/revision-pruner-9-crc" Dec 13 06:52:24 crc kubenswrapper[4971]: I1213 06:52:24.093713 4971 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/revision-pruner-9-crc" Dec 13 06:52:24 crc kubenswrapper[4971]: E1213 06:52:24.801254 4971 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ImagePullBackOff: \"Back-off pulling image \\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"\"" pod="openshift-marketplace/certified-operators-rkd9c" podUID="8f50f5ba-3ec3-4fa2-9f5d-51e1c43f6352" Dec 13 06:52:24 crc kubenswrapper[4971]: E1213 06:52:24.801299 4971 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ImagePullBackOff: \"Back-off pulling image \\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"\"" pod="openshift-marketplace/certified-operators-82f6m" podUID="9404e8eb-0f18-4a49-a716-f076273e7179" Dec 13 06:52:24 crc kubenswrapper[4971]: E1213 06:52:24.873039 4971 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled" image="registry.redhat.io/redhat/community-operator-index:v4.18" Dec 13 06:52:24 crc kubenswrapper[4971]: E1213 06:52:24.873306 4971 kuberuntime_manager.go:1274] "Unhandled Error" err="init container &Container{Name:extract-content,Image:registry.redhat.io/redhat/community-operator-index:v4.18,Command:[/utilities/copy-content],Args:[--catalog.from=/configs --catalog.to=/extracted-catalog/catalog --cache.from=/tmp/cache --cache.to=/extracted-catalog/cache],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:utilities,ReadOnly:false,MountPath:/utilities,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:catalog-content,ReadOnly:false,MountPath:/extracted-catalog,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-j5glc,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:Always,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*1000170000,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:FallbackToLogsOnError,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod community-operators-7qgvq_openshift-marketplace(89f6de2b-6afc-4782-97ca-5601253c688f): ErrImagePull: rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled" logger="UnhandledError" Dec 13 06:52:24 crc kubenswrapper[4971]: E1213 06:52:24.874660 4971 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ErrImagePull: \"rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled\"" pod="openshift-marketplace/community-operators-7qgvq" podUID="89f6de2b-6afc-4782-97ca-5601253c688f" Dec 13 06:52:24 crc kubenswrapper[4971]: E1213 06:52:24.879419 4971 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled" image="registry.redhat.io/redhat/community-operator-index:v4.18" Dec 13 06:52:24 crc kubenswrapper[4971]: E1213 06:52:24.879584 4971 kuberuntime_manager.go:1274] "Unhandled Error" err="init container &Container{Name:extract-content,Image:registry.redhat.io/redhat/community-operator-index:v4.18,Command:[/utilities/copy-content],Args:[--catalog.from=/configs --catalog.to=/extracted-catalog/catalog --cache.from=/tmp/cache --cache.to=/extracted-catalog/cache],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:utilities,ReadOnly:false,MountPath:/utilities,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:catalog-content,ReadOnly:false,MountPath:/extracted-catalog,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-t7lfq,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:Always,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*1000170000,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:FallbackToLogsOnError,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod community-operators-997kc_openshift-marketplace(cf417851-ae43-462f-a678-125593fb653b): ErrImagePull: rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled" logger="UnhandledError" Dec 13 06:52:24 crc kubenswrapper[4971]: E1213 06:52:24.880722 4971 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ErrImagePull: \"rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled\"" pod="openshift-marketplace/community-operators-997kc" podUID="cf417851-ae43-462f-a678-125593fb653b" Dec 13 06:52:25 crc kubenswrapper[4971]: E1213 06:52:25.849188 4971 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ImagePullBackOff: \"Back-off pulling image \\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"\"" pod="openshift-marketplace/community-operators-997kc" podUID="cf417851-ae43-462f-a678-125593fb653b" Dec 13 06:52:25 crc kubenswrapper[4971]: E1213 06:52:25.849171 4971 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ImagePullBackOff: \"Back-off pulling image \\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"\"" pod="openshift-marketplace/community-operators-7qgvq" podUID="89f6de2b-6afc-4782-97ca-5601253c688f" Dec 13 06:52:25 crc kubenswrapper[4971]: E1213 06:52:25.904555 4971 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled" image="registry.redhat.io/redhat/redhat-marketplace-index:v4.18" Dec 13 06:52:25 crc kubenswrapper[4971]: E1213 06:52:25.904965 4971 kuberuntime_manager.go:1274] "Unhandled Error" err="init container &Container{Name:extract-content,Image:registry.redhat.io/redhat/redhat-marketplace-index:v4.18,Command:[/utilities/copy-content],Args:[--catalog.from=/configs --catalog.to=/extracted-catalog/catalog --cache.from=/tmp/cache --cache.to=/extracted-catalog/cache],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:utilities,ReadOnly:false,MountPath:/utilities,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:catalog-content,ReadOnly:false,MountPath:/extracted-catalog,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-25p8j,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:Always,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*1000170000,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:FallbackToLogsOnError,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod redhat-marketplace-xwxlf_openshift-marketplace(22037bdc-fe6d-4411-be2b-aec38524a8dd): ErrImagePull: rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled" logger="UnhandledError" Dec 13 06:52:25 crc kubenswrapper[4971]: E1213 06:52:25.906166 4971 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ErrImagePull: \"rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled\"" pod="openshift-marketplace/redhat-marketplace-xwxlf" podUID="22037bdc-fe6d-4411-be2b-aec38524a8dd" Dec 13 06:52:28 crc kubenswrapper[4971]: I1213 06:52:28.159019 4971 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-kube-apiserver/installer-9-crc"] Dec 13 06:52:28 crc kubenswrapper[4971]: I1213 06:52:28.161804 4971 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/installer-9-crc" Dec 13 06:52:28 crc kubenswrapper[4971]: I1213 06:52:28.167984 4971 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-apiserver/installer-9-crc"] Dec 13 06:52:28 crc kubenswrapper[4971]: I1213 06:52:28.338077 4971 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/53fdcbc2-900b-48eb-8be3-cd72940096c3-kube-api-access\") pod \"installer-9-crc\" (UID: \"53fdcbc2-900b-48eb-8be3-cd72940096c3\") " pod="openshift-kube-apiserver/installer-9-crc" Dec 13 06:52:28 crc kubenswrapper[4971]: I1213 06:52:28.338201 4971 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/53fdcbc2-900b-48eb-8be3-cd72940096c3-kubelet-dir\") pod \"installer-9-crc\" (UID: \"53fdcbc2-900b-48eb-8be3-cd72940096c3\") " pod="openshift-kube-apiserver/installer-9-crc" Dec 13 06:52:28 crc kubenswrapper[4971]: I1213 06:52:28.338236 4971 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-lock\" (UniqueName: \"kubernetes.io/host-path/53fdcbc2-900b-48eb-8be3-cd72940096c3-var-lock\") pod \"installer-9-crc\" (UID: \"53fdcbc2-900b-48eb-8be3-cd72940096c3\") " pod="openshift-kube-apiserver/installer-9-crc" Dec 13 06:52:28 crc kubenswrapper[4971]: I1213 06:52:28.439845 4971 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/53fdcbc2-900b-48eb-8be3-cd72940096c3-kubelet-dir\") pod \"installer-9-crc\" (UID: \"53fdcbc2-900b-48eb-8be3-cd72940096c3\") " pod="openshift-kube-apiserver/installer-9-crc" Dec 13 06:52:28 crc kubenswrapper[4971]: I1213 06:52:28.439901 4971 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-lock\" (UniqueName: \"kubernetes.io/host-path/53fdcbc2-900b-48eb-8be3-cd72940096c3-var-lock\") pod \"installer-9-crc\" (UID: \"53fdcbc2-900b-48eb-8be3-cd72940096c3\") " pod="openshift-kube-apiserver/installer-9-crc" Dec 13 06:52:28 crc kubenswrapper[4971]: I1213 06:52:28.439943 4971 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/53fdcbc2-900b-48eb-8be3-cd72940096c3-kube-api-access\") pod \"installer-9-crc\" (UID: \"53fdcbc2-900b-48eb-8be3-cd72940096c3\") " pod="openshift-kube-apiserver/installer-9-crc" Dec 13 06:52:28 crc kubenswrapper[4971]: I1213 06:52:28.439988 4971 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/53fdcbc2-900b-48eb-8be3-cd72940096c3-kubelet-dir\") pod \"installer-9-crc\" (UID: \"53fdcbc2-900b-48eb-8be3-cd72940096c3\") " pod="openshift-kube-apiserver/installer-9-crc" Dec 13 06:52:28 crc kubenswrapper[4971]: I1213 06:52:28.440094 4971 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-lock\" (UniqueName: \"kubernetes.io/host-path/53fdcbc2-900b-48eb-8be3-cd72940096c3-var-lock\") pod \"installer-9-crc\" (UID: \"53fdcbc2-900b-48eb-8be3-cd72940096c3\") " pod="openshift-kube-apiserver/installer-9-crc" Dec 13 06:52:28 crc kubenswrapper[4971]: I1213 06:52:28.459732 4971 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/53fdcbc2-900b-48eb-8be3-cd72940096c3-kube-api-access\") pod \"installer-9-crc\" (UID: \"53fdcbc2-900b-48eb-8be3-cd72940096c3\") " pod="openshift-kube-apiserver/installer-9-crc" Dec 13 06:52:28 crc kubenswrapper[4971]: I1213 06:52:28.486686 4971 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/installer-9-crc" Dec 13 06:52:29 crc kubenswrapper[4971]: E1213 06:52:29.105220 4971 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ImagePullBackOff: \"Back-off pulling image \\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"\"" pod="openshift-marketplace/redhat-marketplace-xwxlf" podUID="22037bdc-fe6d-4411-be2b-aec38524a8dd" Dec 13 06:52:29 crc kubenswrapper[4971]: E1213 06:52:29.146790 4971 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled" image="registry.redhat.io/redhat/redhat-operator-index:v4.18" Dec 13 06:52:29 crc kubenswrapper[4971]: E1213 06:52:29.146925 4971 kuberuntime_manager.go:1274] "Unhandled Error" err="init container &Container{Name:extract-content,Image:registry.redhat.io/redhat/redhat-operator-index:v4.18,Command:[/utilities/copy-content],Args:[--catalog.from=/configs --catalog.to=/extracted-catalog/catalog --cache.from=/tmp/cache --cache.to=/extracted-catalog/cache],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:utilities,ReadOnly:false,MountPath:/utilities,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:catalog-content,ReadOnly:false,MountPath:/extracted-catalog,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-55kcz,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:Always,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*1000170000,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:FallbackToLogsOnError,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod redhat-operators-d8n8p_openshift-marketplace(b9a0bc5d-3a39-4732-a849-22d351a4cc18): ErrImagePull: rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled" logger="UnhandledError" Dec 13 06:52:29 crc kubenswrapper[4971]: E1213 06:52:29.149213 4971 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ErrImagePull: \"rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled\"" pod="openshift-marketplace/redhat-operators-d8n8p" podUID="b9a0bc5d-3a39-4732-a849-22d351a4cc18" Dec 13 06:52:29 crc kubenswrapper[4971]: E1213 06:52:29.178103 4971 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled" image="registry.redhat.io/redhat/redhat-operator-index:v4.18" Dec 13 06:52:29 crc kubenswrapper[4971]: E1213 06:52:29.178301 4971 kuberuntime_manager.go:1274] "Unhandled Error" err="init container &Container{Name:extract-content,Image:registry.redhat.io/redhat/redhat-operator-index:v4.18,Command:[/utilities/copy-content],Args:[--catalog.from=/configs --catalog.to=/extracted-catalog/catalog --cache.from=/tmp/cache --cache.to=/extracted-catalog/cache],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:utilities,ReadOnly:false,MountPath:/utilities,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:catalog-content,ReadOnly:false,MountPath:/extracted-catalog,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-d9kbv,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:Always,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*1000170000,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:FallbackToLogsOnError,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod redhat-operators-wpxfk_openshift-marketplace(02f6e9f2-6bad-441c-8e2e-9acfc71f3ab4): ErrImagePull: rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled" logger="UnhandledError" Dec 13 06:52:29 crc kubenswrapper[4971]: E1213 06:52:29.179468 4971 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ErrImagePull: \"rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled\"" pod="openshift-marketplace/redhat-operators-wpxfk" podUID="02f6e9f2-6bad-441c-8e2e-9acfc71f3ab4" Dec 13 06:52:29 crc kubenswrapper[4971]: E1213 06:52:29.202605 4971 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled" image="registry.redhat.io/redhat/redhat-marketplace-index:v4.18" Dec 13 06:52:29 crc kubenswrapper[4971]: E1213 06:52:29.202992 4971 kuberuntime_manager.go:1274] "Unhandled Error" err="init container &Container{Name:extract-content,Image:registry.redhat.io/redhat/redhat-marketplace-index:v4.18,Command:[/utilities/copy-content],Args:[--catalog.from=/configs --catalog.to=/extracted-catalog/catalog --cache.from=/tmp/cache --cache.to=/extracted-catalog/cache],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:utilities,ReadOnly:false,MountPath:/utilities,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:catalog-content,ReadOnly:false,MountPath:/extracted-catalog,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-npf8d,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:Always,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*1000170000,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:FallbackToLogsOnError,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod redhat-marketplace-mkqps_openshift-marketplace(9d4f1ea5-8472-4320-820b-8090ff6f5ca6): ErrImagePull: rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled" logger="UnhandledError" Dec 13 06:52:29 crc kubenswrapper[4971]: E1213 06:52:29.204136 4971 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ErrImagePull: \"rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled\"" pod="openshift-marketplace/redhat-marketplace-mkqps" podUID="9d4f1ea5-8472-4320-820b-8090ff6f5ca6" Dec 13 06:52:29 crc kubenswrapper[4971]: I1213 06:52:29.348757 4971 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-apiserver/installer-9-crc"] Dec 13 06:52:29 crc kubenswrapper[4971]: I1213 06:52:29.504226 4971 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-apiserver/revision-pruner-9-crc"] Dec 13 06:52:30 crc kubenswrapper[4971]: I1213 06:52:30.050347 4971 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/revision-pruner-9-crc" event={"ID":"fb460f16-f1e2-4ad7-a4bb-a2f5ef0cfe7a","Type":"ContainerStarted","Data":"3bb383a216929e591595f8dfcc3a8a0907bc5f04cb964bfcc3a98ae92d3d0e19"} Dec 13 06:52:30 crc kubenswrapper[4971]: I1213 06:52:30.050748 4971 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/revision-pruner-9-crc" event={"ID":"fb460f16-f1e2-4ad7-a4bb-a2f5ef0cfe7a","Type":"ContainerStarted","Data":"430652c014b62db4e4af07c75ed2c0aa5f9993d8068b5479422c83f96ce5f2b9"} Dec 13 06:52:30 crc kubenswrapper[4971]: I1213 06:52:30.053676 4971 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/installer-9-crc" event={"ID":"53fdcbc2-900b-48eb-8be3-cd72940096c3","Type":"ContainerStarted","Data":"0e7d033dee64f00f95af0d1a290dc8fafbfbfd51a34dabf8fca951acfe190bae"} Dec 13 06:52:30 crc kubenswrapper[4971]: I1213 06:52:30.053717 4971 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/installer-9-crc" event={"ID":"53fdcbc2-900b-48eb-8be3-cd72940096c3","Type":"ContainerStarted","Data":"1fd88163f64ef84917208a4b013342315222b8bfaf5a456f69c32021ec433c8f"} Dec 13 06:52:30 crc kubenswrapper[4971]: E1213 06:52:30.055453 4971 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ImagePullBackOff: \"Back-off pulling image \\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"\"" pod="openshift-marketplace/redhat-operators-d8n8p" podUID="b9a0bc5d-3a39-4732-a849-22d351a4cc18" Dec 13 06:52:30 crc kubenswrapper[4971]: E1213 06:52:30.056260 4971 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ImagePullBackOff: \"Back-off pulling image \\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"\"" pod="openshift-marketplace/redhat-operators-wpxfk" podUID="02f6e9f2-6bad-441c-8e2e-9acfc71f3ab4" Dec 13 06:52:30 crc kubenswrapper[4971]: E1213 06:52:30.056841 4971 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ImagePullBackOff: \"Back-off pulling image \\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"\"" pod="openshift-marketplace/redhat-marketplace-mkqps" podUID="9d4f1ea5-8472-4320-820b-8090ff6f5ca6" Dec 13 06:52:30 crc kubenswrapper[4971]: I1213 06:52:30.069738 4971 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-kube-apiserver/revision-pruner-9-crc" podStartSLOduration=7.069717995 podStartE2EDuration="7.069717995s" podCreationTimestamp="2025-12-13 06:52:23 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-13 06:52:30.068492076 +0000 UTC m=+206.672901524" watchObservedRunningTime="2025-12-13 06:52:30.069717995 +0000 UTC m=+206.674127443" Dec 13 06:52:31 crc kubenswrapper[4971]: I1213 06:52:31.060067 4971 generic.go:334] "Generic (PLEG): container finished" podID="fb460f16-f1e2-4ad7-a4bb-a2f5ef0cfe7a" containerID="3bb383a216929e591595f8dfcc3a8a0907bc5f04cb964bfcc3a98ae92d3d0e19" exitCode=0 Dec 13 06:52:31 crc kubenswrapper[4971]: I1213 06:52:31.060168 4971 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/revision-pruner-9-crc" event={"ID":"fb460f16-f1e2-4ad7-a4bb-a2f5ef0cfe7a","Type":"ContainerDied","Data":"3bb383a216929e591595f8dfcc3a8a0907bc5f04cb964bfcc3a98ae92d3d0e19"} Dec 13 06:52:31 crc kubenswrapper[4971]: I1213 06:52:31.076340 4971 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-kube-apiserver/installer-9-crc" podStartSLOduration=3.076323435 podStartE2EDuration="3.076323435s" podCreationTimestamp="2025-12-13 06:52:28 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-13 06:52:30.142467872 +0000 UTC m=+206.746877320" watchObservedRunningTime="2025-12-13 06:52:31.076323435 +0000 UTC m=+207.680732883" Dec 13 06:52:32 crc kubenswrapper[4971]: I1213 06:52:32.263033 4971 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/revision-pruner-9-crc" Dec 13 06:52:32 crc kubenswrapper[4971]: I1213 06:52:32.291655 4971 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/fb460f16-f1e2-4ad7-a4bb-a2f5ef0cfe7a-kubelet-dir\") pod \"fb460f16-f1e2-4ad7-a4bb-a2f5ef0cfe7a\" (UID: \"fb460f16-f1e2-4ad7-a4bb-a2f5ef0cfe7a\") " Dec 13 06:52:32 crc kubenswrapper[4971]: I1213 06:52:32.291711 4971 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/fb460f16-f1e2-4ad7-a4bb-a2f5ef0cfe7a-kube-api-access\") pod \"fb460f16-f1e2-4ad7-a4bb-a2f5ef0cfe7a\" (UID: \"fb460f16-f1e2-4ad7-a4bb-a2f5ef0cfe7a\") " Dec 13 06:52:32 crc kubenswrapper[4971]: I1213 06:52:32.293015 4971 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/fb460f16-f1e2-4ad7-a4bb-a2f5ef0cfe7a-kubelet-dir" (OuterVolumeSpecName: "kubelet-dir") pod "fb460f16-f1e2-4ad7-a4bb-a2f5ef0cfe7a" (UID: "fb460f16-f1e2-4ad7-a4bb-a2f5ef0cfe7a"). InnerVolumeSpecName "kubelet-dir". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 13 06:52:32 crc kubenswrapper[4971]: I1213 06:52:32.307408 4971 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/fb460f16-f1e2-4ad7-a4bb-a2f5ef0cfe7a-kube-api-access" (OuterVolumeSpecName: "kube-api-access") pod "fb460f16-f1e2-4ad7-a4bb-a2f5ef0cfe7a" (UID: "fb460f16-f1e2-4ad7-a4bb-a2f5ef0cfe7a"). InnerVolumeSpecName "kube-api-access". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 13 06:52:32 crc kubenswrapper[4971]: I1213 06:52:32.392370 4971 reconciler_common.go:293] "Volume detached for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/fb460f16-f1e2-4ad7-a4bb-a2f5ef0cfe7a-kubelet-dir\") on node \"crc\" DevicePath \"\"" Dec 13 06:52:32 crc kubenswrapper[4971]: I1213 06:52:32.392424 4971 reconciler_common.go:293] "Volume detached for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/fb460f16-f1e2-4ad7-a4bb-a2f5ef0cfe7a-kube-api-access\") on node \"crc\" DevicePath \"\"" Dec 13 06:52:33 crc kubenswrapper[4971]: I1213 06:52:33.071051 4971 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/revision-pruner-9-crc" event={"ID":"fb460f16-f1e2-4ad7-a4bb-a2f5ef0cfe7a","Type":"ContainerDied","Data":"430652c014b62db4e4af07c75ed2c0aa5f9993d8068b5479422c83f96ce5f2b9"} Dec 13 06:52:33 crc kubenswrapper[4971]: I1213 06:52:33.071095 4971 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="430652c014b62db4e4af07c75ed2c0aa5f9993d8068b5479422c83f96ce5f2b9" Dec 13 06:52:33 crc kubenswrapper[4971]: I1213 06:52:33.071156 4971 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/revision-pruner-9-crc" Dec 13 06:52:37 crc kubenswrapper[4971]: I1213 06:52:37.091325 4971 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-rkd9c" event={"ID":"8f50f5ba-3ec3-4fa2-9f5d-51e1c43f6352","Type":"ContainerStarted","Data":"13e3b312f5f1c8e5fb3236ed71e96a35521fd5172708bc5ba5e184501233d15a"} Dec 13 06:52:38 crc kubenswrapper[4971]: I1213 06:52:38.103829 4971 generic.go:334] "Generic (PLEG): container finished" podID="8f50f5ba-3ec3-4fa2-9f5d-51e1c43f6352" containerID="13e3b312f5f1c8e5fb3236ed71e96a35521fd5172708bc5ba5e184501233d15a" exitCode=0 Dec 13 06:52:38 crc kubenswrapper[4971]: I1213 06:52:38.103930 4971 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-rkd9c" event={"ID":"8f50f5ba-3ec3-4fa2-9f5d-51e1c43f6352","Type":"ContainerDied","Data":"13e3b312f5f1c8e5fb3236ed71e96a35521fd5172708bc5ba5e184501233d15a"} Dec 13 06:52:39 crc kubenswrapper[4971]: I1213 06:52:39.113184 4971 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-rkd9c" event={"ID":"8f50f5ba-3ec3-4fa2-9f5d-51e1c43f6352","Type":"ContainerStarted","Data":"cf192b8ffdcb439c2bd3f71dc37ac0dbd8e41220c791304f4d26a5436e9cc7e5"} Dec 13 06:52:39 crc kubenswrapper[4971]: I1213 06:52:39.115221 4971 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-82f6m" event={"ID":"9404e8eb-0f18-4a49-a716-f076273e7179","Type":"ContainerStarted","Data":"21babded3d425f16dab11c5599b2bab2aa17940f56592b3c6953d021bcc968e7"} Dec 13 06:52:39 crc kubenswrapper[4971]: I1213 06:52:39.133956 4971 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/certified-operators-rkd9c" podStartSLOduration=2.9302366490000002 podStartE2EDuration="56.133939283s" podCreationTimestamp="2025-12-13 06:51:43 +0000 UTC" firstStartedPulling="2025-12-13 06:51:45.368642024 +0000 UTC m=+161.973051472" lastFinishedPulling="2025-12-13 06:52:38.572344658 +0000 UTC m=+215.176754106" observedRunningTime="2025-12-13 06:52:39.133861601 +0000 UTC m=+215.738271069" watchObservedRunningTime="2025-12-13 06:52:39.133939283 +0000 UTC m=+215.738348731" Dec 13 06:52:40 crc kubenswrapper[4971]: I1213 06:52:40.122213 4971 generic.go:334] "Generic (PLEG): container finished" podID="cf417851-ae43-462f-a678-125593fb653b" containerID="69a96a08af13f360b70743d81bdc43ccbf3371412d42bbd6979545432c6afe67" exitCode=0 Dec 13 06:52:40 crc kubenswrapper[4971]: I1213 06:52:40.122271 4971 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-997kc" event={"ID":"cf417851-ae43-462f-a678-125593fb653b","Type":"ContainerDied","Data":"69a96a08af13f360b70743d81bdc43ccbf3371412d42bbd6979545432c6afe67"} Dec 13 06:52:40 crc kubenswrapper[4971]: I1213 06:52:40.125470 4971 generic.go:334] "Generic (PLEG): container finished" podID="89f6de2b-6afc-4782-97ca-5601253c688f" containerID="e1c1c10115925655314e1901c42c06170fa624b6bcac9bda258eda6d0be9fcfd" exitCode=0 Dec 13 06:52:40 crc kubenswrapper[4971]: I1213 06:52:40.125492 4971 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-7qgvq" event={"ID":"89f6de2b-6afc-4782-97ca-5601253c688f","Type":"ContainerDied","Data":"e1c1c10115925655314e1901c42c06170fa624b6bcac9bda258eda6d0be9fcfd"} Dec 13 06:52:40 crc kubenswrapper[4971]: I1213 06:52:40.128773 4971 generic.go:334] "Generic (PLEG): container finished" podID="9404e8eb-0f18-4a49-a716-f076273e7179" containerID="21babded3d425f16dab11c5599b2bab2aa17940f56592b3c6953d021bcc968e7" exitCode=0 Dec 13 06:52:40 crc kubenswrapper[4971]: I1213 06:52:40.128811 4971 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-82f6m" event={"ID":"9404e8eb-0f18-4a49-a716-f076273e7179","Type":"ContainerDied","Data":"21babded3d425f16dab11c5599b2bab2aa17940f56592b3c6953d021bcc968e7"} Dec 13 06:52:41 crc kubenswrapper[4971]: I1213 06:52:41.136383 4971 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-7qgvq" event={"ID":"89f6de2b-6afc-4782-97ca-5601253c688f","Type":"ContainerStarted","Data":"464d89ce642a124ec73b1eeb422bb495b868a3bd58953b273e654c0f41fea569"} Dec 13 06:52:41 crc kubenswrapper[4971]: I1213 06:52:41.139013 4971 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-82f6m" event={"ID":"9404e8eb-0f18-4a49-a716-f076273e7179","Type":"ContainerStarted","Data":"2729945958d5c3b79f733cfe2d61a0aa1865d2605fdc84f4aab0367c7ebf86e5"} Dec 13 06:52:41 crc kubenswrapper[4971]: I1213 06:52:41.140740 4971 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-997kc" event={"ID":"cf417851-ae43-462f-a678-125593fb653b","Type":"ContainerStarted","Data":"6e9990367538c8d28196c610c227d5744eafd1a07907c88fcc963201ce0f06cc"} Dec 13 06:52:41 crc kubenswrapper[4971]: I1213 06:52:41.155810 4971 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/community-operators-7qgvq" podStartSLOduration=2.906153202 podStartE2EDuration="58.155794551s" podCreationTimestamp="2025-12-13 06:51:43 +0000 UTC" firstStartedPulling="2025-12-13 06:51:45.367142346 +0000 UTC m=+161.971551794" lastFinishedPulling="2025-12-13 06:52:40.616783695 +0000 UTC m=+217.221193143" observedRunningTime="2025-12-13 06:52:41.152758507 +0000 UTC m=+217.757167965" watchObservedRunningTime="2025-12-13 06:52:41.155794551 +0000 UTC m=+217.760203999" Dec 13 06:52:41 crc kubenswrapper[4971]: I1213 06:52:41.177187 4971 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/community-operators-997kc" podStartSLOduration=3.037604571 podStartE2EDuration="58.17716882s" podCreationTimestamp="2025-12-13 06:51:43 +0000 UTC" firstStartedPulling="2025-12-13 06:51:45.364484609 +0000 UTC m=+161.968894067" lastFinishedPulling="2025-12-13 06:52:40.504048868 +0000 UTC m=+217.108458316" observedRunningTime="2025-12-13 06:52:41.17262699 +0000 UTC m=+217.777036458" watchObservedRunningTime="2025-12-13 06:52:41.17716882 +0000 UTC m=+217.781578268" Dec 13 06:52:41 crc kubenswrapper[4971]: I1213 06:52:41.192412 4971 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/certified-operators-82f6m" podStartSLOduration=3.006226657 podStartE2EDuration="58.19239221s" podCreationTimestamp="2025-12-13 06:51:43 +0000 UTC" firstStartedPulling="2025-12-13 06:51:45.36093528 +0000 UTC m=+161.965344718" lastFinishedPulling="2025-12-13 06:52:40.547100823 +0000 UTC m=+217.151510271" observedRunningTime="2025-12-13 06:52:41.190312869 +0000 UTC m=+217.794722337" watchObservedRunningTime="2025-12-13 06:52:41.19239221 +0000 UTC m=+217.796801658" Dec 13 06:52:43 crc kubenswrapper[4971]: I1213 06:52:43.424213 4971 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/community-operators-997kc" Dec 13 06:52:43 crc kubenswrapper[4971]: I1213 06:52:43.424585 4971 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/community-operators-997kc" Dec 13 06:52:43 crc kubenswrapper[4971]: I1213 06:52:43.606352 4971 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/certified-operators-rkd9c" Dec 13 06:52:43 crc kubenswrapper[4971]: I1213 06:52:43.606759 4971 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/certified-operators-rkd9c" Dec 13 06:52:43 crc kubenswrapper[4971]: I1213 06:52:43.689199 4971 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/certified-operators-rkd9c" Dec 13 06:52:43 crc kubenswrapper[4971]: I1213 06:52:43.690178 4971 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/community-operators-997kc" Dec 13 06:52:43 crc kubenswrapper[4971]: I1213 06:52:43.801417 4971 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/community-operators-7qgvq" Dec 13 06:52:43 crc kubenswrapper[4971]: I1213 06:52:43.801491 4971 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/community-operators-7qgvq" Dec 13 06:52:43 crc kubenswrapper[4971]: I1213 06:52:43.838015 4971 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/community-operators-7qgvq" Dec 13 06:52:44 crc kubenswrapper[4971]: I1213 06:52:44.014443 4971 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/certified-operators-82f6m" Dec 13 06:52:44 crc kubenswrapper[4971]: I1213 06:52:44.014488 4971 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/certified-operators-82f6m" Dec 13 06:52:44 crc kubenswrapper[4971]: I1213 06:52:44.061233 4971 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/certified-operators-82f6m" Dec 13 06:52:44 crc kubenswrapper[4971]: I1213 06:52:44.204059 4971 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/certified-operators-rkd9c" Dec 13 06:52:46 crc kubenswrapper[4971]: I1213 06:52:46.153146 4971 patch_prober.go:28] interesting pod/machine-config-daemon-82xjz container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 13 06:52:46 crc kubenswrapper[4971]: I1213 06:52:46.153204 4971 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-82xjz" podUID="e0c4af12-edda-41d7-9b44-f87396174ef2" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 13 06:52:46 crc kubenswrapper[4971]: I1213 06:52:46.153280 4971 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-82xjz" Dec 13 06:52:46 crc kubenswrapper[4971]: I1213 06:52:46.153877 4971 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"7e0d2020e62f1a277cff7047e1907dc747ed393a1f6c983e1ee258bd156c16c7"} pod="openshift-machine-config-operator/machine-config-daemon-82xjz" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Dec 13 06:52:46 crc kubenswrapper[4971]: I1213 06:52:46.153985 4971 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-82xjz" podUID="e0c4af12-edda-41d7-9b44-f87396174ef2" containerName="machine-config-daemon" containerID="cri-o://7e0d2020e62f1a277cff7047e1907dc747ed393a1f6c983e1ee258bd156c16c7" gracePeriod=600 Dec 13 06:52:46 crc kubenswrapper[4971]: I1213 06:52:46.171144 4971 generic.go:334] "Generic (PLEG): container finished" podID="22037bdc-fe6d-4411-be2b-aec38524a8dd" containerID="a45464afad8c3643d3a5c8eb822b09c0d2807f574631a94b542744363a306d6c" exitCode=0 Dec 13 06:52:46 crc kubenswrapper[4971]: I1213 06:52:46.171199 4971 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-xwxlf" event={"ID":"22037bdc-fe6d-4411-be2b-aec38524a8dd","Type":"ContainerDied","Data":"a45464afad8c3643d3a5c8eb822b09c0d2807f574631a94b542744363a306d6c"} Dec 13 06:52:48 crc kubenswrapper[4971]: I1213 06:52:48.186221 4971 generic.go:334] "Generic (PLEG): container finished" podID="e0c4af12-edda-41d7-9b44-f87396174ef2" containerID="7e0d2020e62f1a277cff7047e1907dc747ed393a1f6c983e1ee258bd156c16c7" exitCode=0 Dec 13 06:52:48 crc kubenswrapper[4971]: I1213 06:52:48.186310 4971 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-82xjz" event={"ID":"e0c4af12-edda-41d7-9b44-f87396174ef2","Type":"ContainerDied","Data":"7e0d2020e62f1a277cff7047e1907dc747ed393a1f6c983e1ee258bd156c16c7"} Dec 13 06:52:52 crc kubenswrapper[4971]: I1213 06:52:52.212228 4971 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-82xjz" event={"ID":"e0c4af12-edda-41d7-9b44-f87396174ef2","Type":"ContainerStarted","Data":"b43d8045f85a66d814d3d526124e6e6bee306cba177b3bfd2a41d456c09be2eb"} Dec 13 06:52:53 crc kubenswrapper[4971]: I1213 06:52:53.467094 4971 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/community-operators-997kc" Dec 13 06:52:53 crc kubenswrapper[4971]: I1213 06:52:53.864457 4971 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/community-operators-7qgvq" Dec 13 06:52:53 crc kubenswrapper[4971]: I1213 06:52:53.907113 4971 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-7qgvq"] Dec 13 06:52:54 crc kubenswrapper[4971]: I1213 06:52:54.221140 4971 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/certified-operators-82f6m" Dec 13 06:52:54 crc kubenswrapper[4971]: I1213 06:52:54.222848 4971 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-d8n8p" event={"ID":"b9a0bc5d-3a39-4732-a849-22d351a4cc18","Type":"ContainerStarted","Data":"8ead1d45e4814ca9d385e346db4d8556a4ac91e83105c52ee55c9b712da52a36"} Dec 13 06:52:54 crc kubenswrapper[4971]: I1213 06:52:54.224821 4971 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-mkqps" event={"ID":"9d4f1ea5-8472-4320-820b-8090ff6f5ca6","Type":"ContainerStarted","Data":"8b1751665ebf5281772672dfa3219f016e81b5f54d3c9a85761125ad6f32cf0b"} Dec 13 06:52:54 crc kubenswrapper[4971]: I1213 06:52:54.229902 4971 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-xwxlf" event={"ID":"22037bdc-fe6d-4411-be2b-aec38524a8dd","Type":"ContainerStarted","Data":"5e118dbb3b89959ae750afb59cd8dca72e876ecbeb215d5fb1c410b36e5e10c0"} Dec 13 06:52:54 crc kubenswrapper[4971]: I1213 06:52:54.231484 4971 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/community-operators-7qgvq" podUID="89f6de2b-6afc-4782-97ca-5601253c688f" containerName="registry-server" containerID="cri-o://464d89ce642a124ec73b1eeb422bb495b868a3bd58953b273e654c0f41fea569" gracePeriod=2 Dec 13 06:52:54 crc kubenswrapper[4971]: I1213 06:52:54.231794 4971 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-wpxfk" event={"ID":"02f6e9f2-6bad-441c-8e2e-9acfc71f3ab4","Type":"ContainerStarted","Data":"ed8f0295bdfe44b9c612064aba54bb4629a51eab4eebe12267dfaf5b4c03532b"} Dec 13 06:52:54 crc kubenswrapper[4971]: I1213 06:52:54.266758 4971 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-marketplace-xwxlf" podStartSLOduration=1.895981478 podStartE2EDuration="1m9.266736264s" podCreationTimestamp="2025-12-13 06:51:45 +0000 UTC" firstStartedPulling="2025-12-13 06:51:46.38573093 +0000 UTC m=+162.990140378" lastFinishedPulling="2025-12-13 06:52:53.756485716 +0000 UTC m=+230.360895164" observedRunningTime="2025-12-13 06:52:54.264618253 +0000 UTC m=+230.869027711" watchObservedRunningTime="2025-12-13 06:52:54.266736264 +0000 UTC m=+230.871145702" Dec 13 06:52:55 crc kubenswrapper[4971]: I1213 06:52:55.239554 4971 generic.go:334] "Generic (PLEG): container finished" podID="89f6de2b-6afc-4782-97ca-5601253c688f" containerID="464d89ce642a124ec73b1eeb422bb495b868a3bd58953b273e654c0f41fea569" exitCode=0 Dec 13 06:52:55 crc kubenswrapper[4971]: I1213 06:52:55.239632 4971 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-7qgvq" event={"ID":"89f6de2b-6afc-4782-97ca-5601253c688f","Type":"ContainerDied","Data":"464d89ce642a124ec73b1eeb422bb495b868a3bd58953b273e654c0f41fea569"} Dec 13 06:52:55 crc kubenswrapper[4971]: I1213 06:52:55.242061 4971 generic.go:334] "Generic (PLEG): container finished" podID="9d4f1ea5-8472-4320-820b-8090ff6f5ca6" containerID="8b1751665ebf5281772672dfa3219f016e81b5f54d3c9a85761125ad6f32cf0b" exitCode=0 Dec 13 06:52:55 crc kubenswrapper[4971]: I1213 06:52:55.242100 4971 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-mkqps" event={"ID":"9d4f1ea5-8472-4320-820b-8090ff6f5ca6","Type":"ContainerDied","Data":"8b1751665ebf5281772672dfa3219f016e81b5f54d3c9a85761125ad6f32cf0b"} Dec 13 06:52:55 crc kubenswrapper[4971]: I1213 06:52:55.308912 4971 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-7qgvq" Dec 13 06:52:55 crc kubenswrapper[4971]: I1213 06:52:55.377528 4971 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/89f6de2b-6afc-4782-97ca-5601253c688f-utilities\") pod \"89f6de2b-6afc-4782-97ca-5601253c688f\" (UID: \"89f6de2b-6afc-4782-97ca-5601253c688f\") " Dec 13 06:52:55 crc kubenswrapper[4971]: I1213 06:52:55.377593 4971 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-j5glc\" (UniqueName: \"kubernetes.io/projected/89f6de2b-6afc-4782-97ca-5601253c688f-kube-api-access-j5glc\") pod \"89f6de2b-6afc-4782-97ca-5601253c688f\" (UID: \"89f6de2b-6afc-4782-97ca-5601253c688f\") " Dec 13 06:52:55 crc kubenswrapper[4971]: I1213 06:52:55.377669 4971 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/89f6de2b-6afc-4782-97ca-5601253c688f-catalog-content\") pod \"89f6de2b-6afc-4782-97ca-5601253c688f\" (UID: \"89f6de2b-6afc-4782-97ca-5601253c688f\") " Dec 13 06:52:55 crc kubenswrapper[4971]: I1213 06:52:55.378497 4971 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/89f6de2b-6afc-4782-97ca-5601253c688f-utilities" (OuterVolumeSpecName: "utilities") pod "89f6de2b-6afc-4782-97ca-5601253c688f" (UID: "89f6de2b-6afc-4782-97ca-5601253c688f"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 13 06:52:55 crc kubenswrapper[4971]: I1213 06:52:55.430949 4971 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/89f6de2b-6afc-4782-97ca-5601253c688f-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "89f6de2b-6afc-4782-97ca-5601253c688f" (UID: "89f6de2b-6afc-4782-97ca-5601253c688f"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 13 06:52:55 crc kubenswrapper[4971]: I1213 06:52:55.431090 4971 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-marketplace-xwxlf" Dec 13 06:52:55 crc kubenswrapper[4971]: I1213 06:52:55.431131 4971 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-marketplace-xwxlf" Dec 13 06:52:55 crc kubenswrapper[4971]: I1213 06:52:55.437777 4971 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/89f6de2b-6afc-4782-97ca-5601253c688f-kube-api-access-j5glc" (OuterVolumeSpecName: "kube-api-access-j5glc") pod "89f6de2b-6afc-4782-97ca-5601253c688f" (UID: "89f6de2b-6afc-4782-97ca-5601253c688f"). InnerVolumeSpecName "kube-api-access-j5glc". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 13 06:52:55 crc kubenswrapper[4971]: I1213 06:52:55.479997 4971 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/89f6de2b-6afc-4782-97ca-5601253c688f-utilities\") on node \"crc\" DevicePath \"\"" Dec 13 06:52:55 crc kubenswrapper[4971]: I1213 06:52:55.480040 4971 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-j5glc\" (UniqueName: \"kubernetes.io/projected/89f6de2b-6afc-4782-97ca-5601253c688f-kube-api-access-j5glc\") on node \"crc\" DevicePath \"\"" Dec 13 06:52:55 crc kubenswrapper[4971]: I1213 06:52:55.480051 4971 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/89f6de2b-6afc-4782-97ca-5601253c688f-catalog-content\") on node \"crc\" DevicePath \"\"" Dec 13 06:52:56 crc kubenswrapper[4971]: I1213 06:52:56.255430 4971 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-7qgvq" event={"ID":"89f6de2b-6afc-4782-97ca-5601253c688f","Type":"ContainerDied","Data":"bec0257f0dfd4cf28c11dae346efe7a50a920c0b7d88591f6b542f2f4b6bcdc4"} Dec 13 06:52:56 crc kubenswrapper[4971]: I1213 06:52:56.255814 4971 scope.go:117] "RemoveContainer" containerID="464d89ce642a124ec73b1eeb422bb495b868a3bd58953b273e654c0f41fea569" Dec 13 06:52:56 crc kubenswrapper[4971]: I1213 06:52:56.255734 4971 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-7qgvq" Dec 13 06:52:56 crc kubenswrapper[4971]: I1213 06:52:56.363441 4971 scope.go:117] "RemoveContainer" containerID="e1c1c10115925655314e1901c42c06170fa624b6bcac9bda258eda6d0be9fcfd" Dec 13 06:52:56 crc kubenswrapper[4971]: I1213 06:52:56.379847 4971 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-7qgvq"] Dec 13 06:52:56 crc kubenswrapper[4971]: I1213 06:52:56.383455 4971 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/community-operators-7qgvq"] Dec 13 06:52:56 crc kubenswrapper[4971]: I1213 06:52:56.395890 4971 scope.go:117] "RemoveContainer" containerID="89235ff3b064024ef637ca519f2c1ffc06586998c87c15d021996b7b76a7f9e3" Dec 13 06:52:56 crc kubenswrapper[4971]: I1213 06:52:56.486386 4971 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-marketplace/redhat-marketplace-xwxlf" podUID="22037bdc-fe6d-4411-be2b-aec38524a8dd" containerName="registry-server" probeResult="failure" output=< Dec 13 06:52:56 crc kubenswrapper[4971]: timeout: failed to connect service ":50051" within 1s Dec 13 06:52:56 crc kubenswrapper[4971]: > Dec 13 06:52:56 crc kubenswrapper[4971]: I1213 06:52:56.503277 4971 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-82f6m"] Dec 13 06:52:56 crc kubenswrapper[4971]: I1213 06:52:56.503817 4971 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/certified-operators-82f6m" podUID="9404e8eb-0f18-4a49-a716-f076273e7179" containerName="registry-server" containerID="cri-o://2729945958d5c3b79f733cfe2d61a0aa1865d2605fdc84f4aab0367c7ebf86e5" gracePeriod=2 Dec 13 06:52:56 crc kubenswrapper[4971]: I1213 06:52:56.860416 4971 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-82f6m" Dec 13 06:52:56 crc kubenswrapper[4971]: I1213 06:52:56.958208 4971 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/9404e8eb-0f18-4a49-a716-f076273e7179-catalog-content\") pod \"9404e8eb-0f18-4a49-a716-f076273e7179\" (UID: \"9404e8eb-0f18-4a49-a716-f076273e7179\") " Dec 13 06:52:56 crc kubenswrapper[4971]: I1213 06:52:56.958302 4971 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-wjf2d\" (UniqueName: \"kubernetes.io/projected/9404e8eb-0f18-4a49-a716-f076273e7179-kube-api-access-wjf2d\") pod \"9404e8eb-0f18-4a49-a716-f076273e7179\" (UID: \"9404e8eb-0f18-4a49-a716-f076273e7179\") " Dec 13 06:52:56 crc kubenswrapper[4971]: I1213 06:52:56.958357 4971 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/9404e8eb-0f18-4a49-a716-f076273e7179-utilities\") pod \"9404e8eb-0f18-4a49-a716-f076273e7179\" (UID: \"9404e8eb-0f18-4a49-a716-f076273e7179\") " Dec 13 06:52:56 crc kubenswrapper[4971]: I1213 06:52:56.959382 4971 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/9404e8eb-0f18-4a49-a716-f076273e7179-utilities" (OuterVolumeSpecName: "utilities") pod "9404e8eb-0f18-4a49-a716-f076273e7179" (UID: "9404e8eb-0f18-4a49-a716-f076273e7179"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 13 06:52:56 crc kubenswrapper[4971]: I1213 06:52:56.964851 4971 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/9404e8eb-0f18-4a49-a716-f076273e7179-kube-api-access-wjf2d" (OuterVolumeSpecName: "kube-api-access-wjf2d") pod "9404e8eb-0f18-4a49-a716-f076273e7179" (UID: "9404e8eb-0f18-4a49-a716-f076273e7179"). InnerVolumeSpecName "kube-api-access-wjf2d". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 13 06:52:57 crc kubenswrapper[4971]: I1213 06:52:57.009039 4971 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/9404e8eb-0f18-4a49-a716-f076273e7179-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "9404e8eb-0f18-4a49-a716-f076273e7179" (UID: "9404e8eb-0f18-4a49-a716-f076273e7179"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 13 06:52:57 crc kubenswrapper[4971]: I1213 06:52:57.060052 4971 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/9404e8eb-0f18-4a49-a716-f076273e7179-catalog-content\") on node \"crc\" DevicePath \"\"" Dec 13 06:52:57 crc kubenswrapper[4971]: I1213 06:52:57.060099 4971 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-wjf2d\" (UniqueName: \"kubernetes.io/projected/9404e8eb-0f18-4a49-a716-f076273e7179-kube-api-access-wjf2d\") on node \"crc\" DevicePath \"\"" Dec 13 06:52:57 crc kubenswrapper[4971]: I1213 06:52:57.060113 4971 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/9404e8eb-0f18-4a49-a716-f076273e7179-utilities\") on node \"crc\" DevicePath \"\"" Dec 13 06:52:57 crc kubenswrapper[4971]: I1213 06:52:57.262018 4971 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-mkqps" event={"ID":"9d4f1ea5-8472-4320-820b-8090ff6f5ca6","Type":"ContainerStarted","Data":"b98959ce12c11f9c6b25aaca858920c32124fb7a92b0ddee184e3e434e02168a"} Dec 13 06:52:57 crc kubenswrapper[4971]: I1213 06:52:57.264408 4971 generic.go:334] "Generic (PLEG): container finished" podID="9404e8eb-0f18-4a49-a716-f076273e7179" containerID="2729945958d5c3b79f733cfe2d61a0aa1865d2605fdc84f4aab0367c7ebf86e5" exitCode=0 Dec 13 06:52:57 crc kubenswrapper[4971]: I1213 06:52:57.264470 4971 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-82f6m" Dec 13 06:52:57 crc kubenswrapper[4971]: I1213 06:52:57.264494 4971 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-82f6m" event={"ID":"9404e8eb-0f18-4a49-a716-f076273e7179","Type":"ContainerDied","Data":"2729945958d5c3b79f733cfe2d61a0aa1865d2605fdc84f4aab0367c7ebf86e5"} Dec 13 06:52:57 crc kubenswrapper[4971]: I1213 06:52:57.264553 4971 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-82f6m" event={"ID":"9404e8eb-0f18-4a49-a716-f076273e7179","Type":"ContainerDied","Data":"873ed9c2e934008905cffeebd260086c4af4aeb66a1946f5fa1eed62c4d72268"} Dec 13 06:52:57 crc kubenswrapper[4971]: I1213 06:52:57.264576 4971 scope.go:117] "RemoveContainer" containerID="2729945958d5c3b79f733cfe2d61a0aa1865d2605fdc84f4aab0367c7ebf86e5" Dec 13 06:52:57 crc kubenswrapper[4971]: I1213 06:52:57.266610 4971 generic.go:334] "Generic (PLEG): container finished" podID="02f6e9f2-6bad-441c-8e2e-9acfc71f3ab4" containerID="ed8f0295bdfe44b9c612064aba54bb4629a51eab4eebe12267dfaf5b4c03532b" exitCode=0 Dec 13 06:52:57 crc kubenswrapper[4971]: I1213 06:52:57.266648 4971 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-wpxfk" event={"ID":"02f6e9f2-6bad-441c-8e2e-9acfc71f3ab4","Type":"ContainerDied","Data":"ed8f0295bdfe44b9c612064aba54bb4629a51eab4eebe12267dfaf5b4c03532b"} Dec 13 06:52:57 crc kubenswrapper[4971]: I1213 06:52:57.269341 4971 generic.go:334] "Generic (PLEG): container finished" podID="b9a0bc5d-3a39-4732-a849-22d351a4cc18" containerID="8ead1d45e4814ca9d385e346db4d8556a4ac91e83105c52ee55c9b712da52a36" exitCode=0 Dec 13 06:52:57 crc kubenswrapper[4971]: I1213 06:52:57.269428 4971 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-d8n8p" event={"ID":"b9a0bc5d-3a39-4732-a849-22d351a4cc18","Type":"ContainerDied","Data":"8ead1d45e4814ca9d385e346db4d8556a4ac91e83105c52ee55c9b712da52a36"} Dec 13 06:52:57 crc kubenswrapper[4971]: I1213 06:52:57.282591 4971 scope.go:117] "RemoveContainer" containerID="21babded3d425f16dab11c5599b2bab2aa17940f56592b3c6953d021bcc968e7" Dec 13 06:52:57 crc kubenswrapper[4971]: I1213 06:52:57.288359 4971 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-marketplace-mkqps" podStartSLOduration=2.844149786 podStartE2EDuration="1m12.288334094s" podCreationTimestamp="2025-12-13 06:51:45 +0000 UTC" firstStartedPulling="2025-12-13 06:51:46.383841671 +0000 UTC m=+162.988251119" lastFinishedPulling="2025-12-13 06:52:55.828025979 +0000 UTC m=+232.432435427" observedRunningTime="2025-12-13 06:52:57.286430388 +0000 UTC m=+233.890839846" watchObservedRunningTime="2025-12-13 06:52:57.288334094 +0000 UTC m=+233.892743542" Dec 13 06:52:57 crc kubenswrapper[4971]: I1213 06:52:57.319205 4971 scope.go:117] "RemoveContainer" containerID="8d4628a40eb05a6df7bd6b0fece84c9e24e6b6112b6289309830e6027e1bb986" Dec 13 06:52:57 crc kubenswrapper[4971]: I1213 06:52:57.338912 4971 scope.go:117] "RemoveContainer" containerID="2729945958d5c3b79f733cfe2d61a0aa1865d2605fdc84f4aab0367c7ebf86e5" Dec 13 06:52:57 crc kubenswrapper[4971]: E1213 06:52:57.339720 4971 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"2729945958d5c3b79f733cfe2d61a0aa1865d2605fdc84f4aab0367c7ebf86e5\": container with ID starting with 2729945958d5c3b79f733cfe2d61a0aa1865d2605fdc84f4aab0367c7ebf86e5 not found: ID does not exist" containerID="2729945958d5c3b79f733cfe2d61a0aa1865d2605fdc84f4aab0367c7ebf86e5" Dec 13 06:52:57 crc kubenswrapper[4971]: I1213 06:52:57.339772 4971 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"2729945958d5c3b79f733cfe2d61a0aa1865d2605fdc84f4aab0367c7ebf86e5"} err="failed to get container status \"2729945958d5c3b79f733cfe2d61a0aa1865d2605fdc84f4aab0367c7ebf86e5\": rpc error: code = NotFound desc = could not find container \"2729945958d5c3b79f733cfe2d61a0aa1865d2605fdc84f4aab0367c7ebf86e5\": container with ID starting with 2729945958d5c3b79f733cfe2d61a0aa1865d2605fdc84f4aab0367c7ebf86e5 not found: ID does not exist" Dec 13 06:52:57 crc kubenswrapper[4971]: I1213 06:52:57.339804 4971 scope.go:117] "RemoveContainer" containerID="21babded3d425f16dab11c5599b2bab2aa17940f56592b3c6953d021bcc968e7" Dec 13 06:52:57 crc kubenswrapper[4971]: E1213 06:52:57.340977 4971 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"21babded3d425f16dab11c5599b2bab2aa17940f56592b3c6953d021bcc968e7\": container with ID starting with 21babded3d425f16dab11c5599b2bab2aa17940f56592b3c6953d021bcc968e7 not found: ID does not exist" containerID="21babded3d425f16dab11c5599b2bab2aa17940f56592b3c6953d021bcc968e7" Dec 13 06:52:57 crc kubenswrapper[4971]: I1213 06:52:57.341017 4971 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"21babded3d425f16dab11c5599b2bab2aa17940f56592b3c6953d021bcc968e7"} err="failed to get container status \"21babded3d425f16dab11c5599b2bab2aa17940f56592b3c6953d021bcc968e7\": rpc error: code = NotFound desc = could not find container \"21babded3d425f16dab11c5599b2bab2aa17940f56592b3c6953d021bcc968e7\": container with ID starting with 21babded3d425f16dab11c5599b2bab2aa17940f56592b3c6953d021bcc968e7 not found: ID does not exist" Dec 13 06:52:57 crc kubenswrapper[4971]: I1213 06:52:57.341031 4971 scope.go:117] "RemoveContainer" containerID="8d4628a40eb05a6df7bd6b0fece84c9e24e6b6112b6289309830e6027e1bb986" Dec 13 06:52:57 crc kubenswrapper[4971]: E1213 06:52:57.343309 4971 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"8d4628a40eb05a6df7bd6b0fece84c9e24e6b6112b6289309830e6027e1bb986\": container with ID starting with 8d4628a40eb05a6df7bd6b0fece84c9e24e6b6112b6289309830e6027e1bb986 not found: ID does not exist" containerID="8d4628a40eb05a6df7bd6b0fece84c9e24e6b6112b6289309830e6027e1bb986" Dec 13 06:52:57 crc kubenswrapper[4971]: I1213 06:52:57.343334 4971 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"8d4628a40eb05a6df7bd6b0fece84c9e24e6b6112b6289309830e6027e1bb986"} err="failed to get container status \"8d4628a40eb05a6df7bd6b0fece84c9e24e6b6112b6289309830e6027e1bb986\": rpc error: code = NotFound desc = could not find container \"8d4628a40eb05a6df7bd6b0fece84c9e24e6b6112b6289309830e6027e1bb986\": container with ID starting with 8d4628a40eb05a6df7bd6b0fece84c9e24e6b6112b6289309830e6027e1bb986 not found: ID does not exist" Dec 13 06:52:57 crc kubenswrapper[4971]: I1213 06:52:57.348293 4971 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-82f6m"] Dec 13 06:52:57 crc kubenswrapper[4971]: I1213 06:52:57.356773 4971 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/certified-operators-82f6m"] Dec 13 06:52:57 crc kubenswrapper[4971]: I1213 06:52:57.776756 4971 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="89f6de2b-6afc-4782-97ca-5601253c688f" path="/var/lib/kubelet/pods/89f6de2b-6afc-4782-97ca-5601253c688f/volumes" Dec 13 06:52:57 crc kubenswrapper[4971]: I1213 06:52:57.777830 4971 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="9404e8eb-0f18-4a49-a716-f076273e7179" path="/var/lib/kubelet/pods/9404e8eb-0f18-4a49-a716-f076273e7179/volumes" Dec 13 06:52:58 crc kubenswrapper[4971]: I1213 06:52:58.295268 4971 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-wpxfk" event={"ID":"02f6e9f2-6bad-441c-8e2e-9acfc71f3ab4","Type":"ContainerStarted","Data":"cf5a955add1c7bf17976d195f2fba78c93d0fd447a0684c6e8377e595e9546ad"} Dec 13 06:52:58 crc kubenswrapper[4971]: I1213 06:52:58.298800 4971 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-d8n8p" event={"ID":"b9a0bc5d-3a39-4732-a849-22d351a4cc18","Type":"ContainerStarted","Data":"14fb548fd8178b7fa4ade14c8907b148dd031b77bdcd95bcd98d8b54fc688a83"} Dec 13 06:52:58 crc kubenswrapper[4971]: I1213 06:52:58.339715 4971 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-operators-d8n8p" podStartSLOduration=3.082656604 podStartE2EDuration="1m12.339697319s" podCreationTimestamp="2025-12-13 06:51:46 +0000 UTC" firstStartedPulling="2025-12-13 06:51:48.472928601 +0000 UTC m=+165.077338049" lastFinishedPulling="2025-12-13 06:52:57.729969316 +0000 UTC m=+234.334378764" observedRunningTime="2025-12-13 06:52:58.338271705 +0000 UTC m=+234.942681163" watchObservedRunningTime="2025-12-13 06:52:58.339697319 +0000 UTC m=+234.944106767" Dec 13 06:52:58 crc kubenswrapper[4971]: I1213 06:52:58.340452 4971 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-operators-wpxfk" podStartSLOduration=2.840858459 podStartE2EDuration="1m12.340443958s" podCreationTimestamp="2025-12-13 06:51:46 +0000 UTC" firstStartedPulling="2025-12-13 06:51:48.506510211 +0000 UTC m=+165.110919659" lastFinishedPulling="2025-12-13 06:52:58.00609571 +0000 UTC m=+234.610505158" observedRunningTime="2025-12-13 06:52:58.323849935 +0000 UTC m=+234.928259383" watchObservedRunningTime="2025-12-13 06:52:58.340443958 +0000 UTC m=+234.944853406" Dec 13 06:53:04 crc kubenswrapper[4971]: I1213 06:53:04.738952 4971 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-authentication/oauth-openshift-558db77b4-nzzbm"] Dec 13 06:53:05 crc kubenswrapper[4971]: I1213 06:53:05.473360 4971 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-marketplace-xwxlf" Dec 13 06:53:05 crc kubenswrapper[4971]: I1213 06:53:05.517968 4971 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-marketplace-xwxlf" Dec 13 06:53:05 crc kubenswrapper[4971]: I1213 06:53:05.854692 4971 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-marketplace-mkqps" Dec 13 06:53:05 crc kubenswrapper[4971]: I1213 06:53:05.855064 4971 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-marketplace-mkqps" Dec 13 06:53:05 crc kubenswrapper[4971]: I1213 06:53:05.898518 4971 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-marketplace-mkqps" Dec 13 06:53:06 crc kubenswrapper[4971]: I1213 06:53:06.384323 4971 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-marketplace-mkqps" Dec 13 06:53:06 crc kubenswrapper[4971]: I1213 06:53:06.514963 4971 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-rkd9c"] Dec 13 06:53:06 crc kubenswrapper[4971]: I1213 06:53:06.515241 4971 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/certified-operators-rkd9c" podUID="8f50f5ba-3ec3-4fa2-9f5d-51e1c43f6352" containerName="registry-server" containerID="cri-o://cf192b8ffdcb439c2bd3f71dc37ac0dbd8e41220c791304f4d26a5436e9cc7e5" gracePeriod=30 Dec 13 06:53:06 crc kubenswrapper[4971]: I1213 06:53:06.521812 4971 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-997kc"] Dec 13 06:53:06 crc kubenswrapper[4971]: I1213 06:53:06.523898 4971 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/community-operators-997kc" podUID="cf417851-ae43-462f-a678-125593fb653b" containerName="registry-server" containerID="cri-o://6e9990367538c8d28196c610c227d5744eafd1a07907c88fcc963201ce0f06cc" gracePeriod=30 Dec 13 06:53:06 crc kubenswrapper[4971]: I1213 06:53:06.528168 4971 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/marketplace-operator-79b997595-j9zqf"] Dec 13 06:53:06 crc kubenswrapper[4971]: I1213 06:53:06.528373 4971 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/marketplace-operator-79b997595-j9zqf" podUID="f14c8a94-8589-4059-bddf-329452355fab" containerName="marketplace-operator" containerID="cri-o://524506c76ba56186a2daa60747573baef0eb42cebe8bb321e82be093150f35d7" gracePeriod=30 Dec 13 06:53:06 crc kubenswrapper[4971]: I1213 06:53:06.542195 4971 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/marketplace-operator-79b997595-6rvc9"] Dec 13 06:53:06 crc kubenswrapper[4971]: E1213 06:53:06.542401 4971 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="89f6de2b-6afc-4782-97ca-5601253c688f" containerName="registry-server" Dec 13 06:53:06 crc kubenswrapper[4971]: I1213 06:53:06.542412 4971 state_mem.go:107] "Deleted CPUSet assignment" podUID="89f6de2b-6afc-4782-97ca-5601253c688f" containerName="registry-server" Dec 13 06:53:06 crc kubenswrapper[4971]: E1213 06:53:06.542419 4971 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="fb460f16-f1e2-4ad7-a4bb-a2f5ef0cfe7a" containerName="pruner" Dec 13 06:53:06 crc kubenswrapper[4971]: I1213 06:53:06.542425 4971 state_mem.go:107] "Deleted CPUSet assignment" podUID="fb460f16-f1e2-4ad7-a4bb-a2f5ef0cfe7a" containerName="pruner" Dec 13 06:53:06 crc kubenswrapper[4971]: E1213 06:53:06.542441 4971 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="89f6de2b-6afc-4782-97ca-5601253c688f" containerName="extract-content" Dec 13 06:53:06 crc kubenswrapper[4971]: I1213 06:53:06.542447 4971 state_mem.go:107] "Deleted CPUSet assignment" podUID="89f6de2b-6afc-4782-97ca-5601253c688f" containerName="extract-content" Dec 13 06:53:06 crc kubenswrapper[4971]: E1213 06:53:06.542455 4971 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="9404e8eb-0f18-4a49-a716-f076273e7179" containerName="extract-utilities" Dec 13 06:53:06 crc kubenswrapper[4971]: I1213 06:53:06.542461 4971 state_mem.go:107] "Deleted CPUSet assignment" podUID="9404e8eb-0f18-4a49-a716-f076273e7179" containerName="extract-utilities" Dec 13 06:53:06 crc kubenswrapper[4971]: E1213 06:53:06.542472 4971 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="89f6de2b-6afc-4782-97ca-5601253c688f" containerName="extract-utilities" Dec 13 06:53:06 crc kubenswrapper[4971]: I1213 06:53:06.542477 4971 state_mem.go:107] "Deleted CPUSet assignment" podUID="89f6de2b-6afc-4782-97ca-5601253c688f" containerName="extract-utilities" Dec 13 06:53:06 crc kubenswrapper[4971]: E1213 06:53:06.542485 4971 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="9404e8eb-0f18-4a49-a716-f076273e7179" containerName="registry-server" Dec 13 06:53:06 crc kubenswrapper[4971]: I1213 06:53:06.542491 4971 state_mem.go:107] "Deleted CPUSet assignment" podUID="9404e8eb-0f18-4a49-a716-f076273e7179" containerName="registry-server" Dec 13 06:53:06 crc kubenswrapper[4971]: E1213 06:53:06.542499 4971 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="9404e8eb-0f18-4a49-a716-f076273e7179" containerName="extract-content" Dec 13 06:53:06 crc kubenswrapper[4971]: I1213 06:53:06.542504 4971 state_mem.go:107] "Deleted CPUSet assignment" podUID="9404e8eb-0f18-4a49-a716-f076273e7179" containerName="extract-content" Dec 13 06:53:06 crc kubenswrapper[4971]: I1213 06:53:06.542635 4971 memory_manager.go:354] "RemoveStaleState removing state" podUID="9404e8eb-0f18-4a49-a716-f076273e7179" containerName="registry-server" Dec 13 06:53:06 crc kubenswrapper[4971]: I1213 06:53:06.542650 4971 memory_manager.go:354] "RemoveStaleState removing state" podUID="89f6de2b-6afc-4782-97ca-5601253c688f" containerName="registry-server" Dec 13 06:53:06 crc kubenswrapper[4971]: I1213 06:53:06.542658 4971 memory_manager.go:354] "RemoveStaleState removing state" podUID="fb460f16-f1e2-4ad7-a4bb-a2f5ef0cfe7a" containerName="pruner" Dec 13 06:53:06 crc kubenswrapper[4971]: I1213 06:53:06.543024 4971 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/marketplace-operator-79b997595-6rvc9" Dec 13 06:53:06 crc kubenswrapper[4971]: I1213 06:53:06.551145 4971 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-mkqps"] Dec 13 06:53:06 crc kubenswrapper[4971]: I1213 06:53:06.555228 4971 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/marketplace-operator-79b997595-6rvc9"] Dec 13 06:53:06 crc kubenswrapper[4971]: I1213 06:53:06.576811 4971 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-xwxlf"] Dec 13 06:53:06 crc kubenswrapper[4971]: I1213 06:53:06.579387 4971 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-d8n8p"] Dec 13 06:53:06 crc kubenswrapper[4971]: I1213 06:53:06.579707 4971 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-operators-d8n8p" podUID="b9a0bc5d-3a39-4732-a849-22d351a4cc18" containerName="registry-server" containerID="cri-o://14fb548fd8178b7fa4ade14c8907b148dd031b77bdcd95bcd98d8b54fc688a83" gracePeriod=30 Dec 13 06:53:06 crc kubenswrapper[4971]: I1213 06:53:06.582315 4971 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-wpxfk"] Dec 13 06:53:06 crc kubenswrapper[4971]: I1213 06:53:06.582772 4971 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-operators-wpxfk" podUID="02f6e9f2-6bad-441c-8e2e-9acfc71f3ab4" containerName="registry-server" containerID="cri-o://cf5a955add1c7bf17976d195f2fba78c93d0fd447a0684c6e8377e595e9546ad" gracePeriod=30 Dec 13 06:53:06 crc kubenswrapper[4971]: I1213 06:53:06.603817 4971 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-operators-d8n8p" Dec 13 06:53:06 crc kubenswrapper[4971]: I1213 06:53:06.678735 4971 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"marketplace-trusted-ca\" (UniqueName: \"kubernetes.io/configmap/5c88a864-89a9-4f5e-b713-385170b9bf17-marketplace-trusted-ca\") pod \"marketplace-operator-79b997595-6rvc9\" (UID: \"5c88a864-89a9-4f5e-b713-385170b9bf17\") " pod="openshift-marketplace/marketplace-operator-79b997595-6rvc9" Dec 13 06:53:06 crc kubenswrapper[4971]: I1213 06:53:06.678810 4971 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-7zzwm\" (UniqueName: \"kubernetes.io/projected/5c88a864-89a9-4f5e-b713-385170b9bf17-kube-api-access-7zzwm\") pod \"marketplace-operator-79b997595-6rvc9\" (UID: \"5c88a864-89a9-4f5e-b713-385170b9bf17\") " pod="openshift-marketplace/marketplace-operator-79b997595-6rvc9" Dec 13 06:53:06 crc kubenswrapper[4971]: I1213 06:53:06.678851 4971 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"marketplace-operator-metrics\" (UniqueName: \"kubernetes.io/secret/5c88a864-89a9-4f5e-b713-385170b9bf17-marketplace-operator-metrics\") pod \"marketplace-operator-79b997595-6rvc9\" (UID: \"5c88a864-89a9-4f5e-b713-385170b9bf17\") " pod="openshift-marketplace/marketplace-operator-79b997595-6rvc9" Dec 13 06:53:06 crc kubenswrapper[4971]: I1213 06:53:06.780276 4971 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-7zzwm\" (UniqueName: \"kubernetes.io/projected/5c88a864-89a9-4f5e-b713-385170b9bf17-kube-api-access-7zzwm\") pod \"marketplace-operator-79b997595-6rvc9\" (UID: \"5c88a864-89a9-4f5e-b713-385170b9bf17\") " pod="openshift-marketplace/marketplace-operator-79b997595-6rvc9" Dec 13 06:53:06 crc kubenswrapper[4971]: I1213 06:53:06.780338 4971 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"marketplace-operator-metrics\" (UniqueName: \"kubernetes.io/secret/5c88a864-89a9-4f5e-b713-385170b9bf17-marketplace-operator-metrics\") pod \"marketplace-operator-79b997595-6rvc9\" (UID: \"5c88a864-89a9-4f5e-b713-385170b9bf17\") " pod="openshift-marketplace/marketplace-operator-79b997595-6rvc9" Dec 13 06:53:06 crc kubenswrapper[4971]: I1213 06:53:06.780417 4971 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"marketplace-trusted-ca\" (UniqueName: \"kubernetes.io/configmap/5c88a864-89a9-4f5e-b713-385170b9bf17-marketplace-trusted-ca\") pod \"marketplace-operator-79b997595-6rvc9\" (UID: \"5c88a864-89a9-4f5e-b713-385170b9bf17\") " pod="openshift-marketplace/marketplace-operator-79b997595-6rvc9" Dec 13 06:53:06 crc kubenswrapper[4971]: I1213 06:53:06.781468 4971 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"marketplace-trusted-ca\" (UniqueName: \"kubernetes.io/configmap/5c88a864-89a9-4f5e-b713-385170b9bf17-marketplace-trusted-ca\") pod \"marketplace-operator-79b997595-6rvc9\" (UID: \"5c88a864-89a9-4f5e-b713-385170b9bf17\") " pod="openshift-marketplace/marketplace-operator-79b997595-6rvc9" Dec 13 06:53:06 crc kubenswrapper[4971]: I1213 06:53:06.787234 4971 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"marketplace-operator-metrics\" (UniqueName: \"kubernetes.io/secret/5c88a864-89a9-4f5e-b713-385170b9bf17-marketplace-operator-metrics\") pod \"marketplace-operator-79b997595-6rvc9\" (UID: \"5c88a864-89a9-4f5e-b713-385170b9bf17\") " pod="openshift-marketplace/marketplace-operator-79b997595-6rvc9" Dec 13 06:53:06 crc kubenswrapper[4971]: I1213 06:53:06.796473 4971 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-7zzwm\" (UniqueName: \"kubernetes.io/projected/5c88a864-89a9-4f5e-b713-385170b9bf17-kube-api-access-7zzwm\") pod \"marketplace-operator-79b997595-6rvc9\" (UID: \"5c88a864-89a9-4f5e-b713-385170b9bf17\") " pod="openshift-marketplace/marketplace-operator-79b997595-6rvc9" Dec 13 06:53:06 crc kubenswrapper[4971]: I1213 06:53:06.867939 4971 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/marketplace-operator-79b997595-6rvc9" Dec 13 06:53:07 crc kubenswrapper[4971]: I1213 06:53:07.011689 4971 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-operators-wpxfk" Dec 13 06:53:07 crc kubenswrapper[4971]: I1213 06:53:07.279725 4971 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/marketplace-operator-79b997595-6rvc9"] Dec 13 06:53:07 crc kubenswrapper[4971]: W1213 06:53:07.283804 4971 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod5c88a864_89a9_4f5e_b713_385170b9bf17.slice/crio-f93dd5acab7a80e5d527432e821725f3324ff27fa1176c1a26c332a22354f3be WatchSource:0}: Error finding container f93dd5acab7a80e5d527432e821725f3324ff27fa1176c1a26c332a22354f3be: Status 404 returned error can't find the container with id f93dd5acab7a80e5d527432e821725f3324ff27fa1176c1a26c332a22354f3be Dec 13 06:53:07 crc kubenswrapper[4971]: I1213 06:53:07.338027 4971 kubelet.go:2421] "SyncLoop ADD" source="file" pods=["openshift-kube-apiserver/kube-apiserver-startup-monitor-crc"] Dec 13 06:53:07 crc kubenswrapper[4971]: I1213 06:53:07.339021 4971 kubelet.go:2431] "SyncLoop REMOVE" source="file" pods=["openshift-kube-apiserver/kube-apiserver-crc"] Dec 13 06:53:07 crc kubenswrapper[4971]: I1213 06:53:07.339288 4971 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver" containerID="cri-o://3716afdf2bc693be80a09fc9f7727d334f435caf1d4c8ec4811b3a1f1652cdfe" gracePeriod=15 Dec 13 06:53:07 crc kubenswrapper[4971]: I1213 06:53:07.339505 4971 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" Dec 13 06:53:07 crc kubenswrapper[4971]: I1213 06:53:07.339793 4971 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-check-endpoints" containerID="cri-o://770b73dd0c2b6f38744e2c1cd6aa186a69bb87d1cd015a603ad2e33be06a8f6b" gracePeriod=15 Dec 13 06:53:07 crc kubenswrapper[4971]: I1213 06:53:07.339890 4971 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-cert-syncer" containerID="cri-o://ccdf65f6d2b1eb3647a03ba37c305c32e34addfdf79ff937c8dddaa3e5d5b5c6" gracePeriod=15 Dec 13 06:53:07 crc kubenswrapper[4971]: I1213 06:53:07.340273 4971 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-cert-regeneration-controller" containerID="cri-o://30df3abf8c79db8af52c6a918f73e1ce410ddafcf341636424c267e1eeb195a7" gracePeriod=15 Dec 13 06:53:07 crc kubenswrapper[4971]: I1213 06:53:07.340800 4971 kubelet.go:2421] "SyncLoop ADD" source="file" pods=["openshift-kube-apiserver/kube-apiserver-crc"] Dec 13 06:53:07 crc kubenswrapper[4971]: I1213 06:53:07.340814 4971 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-insecure-readyz" containerID="cri-o://9acd07066deed483bd7fa61ee45fda09539d110717a134a56827f871186f92fc" gracePeriod=15 Dec 13 06:53:07 crc kubenswrapper[4971]: E1213 06:53:07.341038 4971 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-cert-syncer" Dec 13 06:53:07 crc kubenswrapper[4971]: I1213 06:53:07.341049 4971 state_mem.go:107] "Deleted CPUSet assignment" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-cert-syncer" Dec 13 06:53:07 crc kubenswrapper[4971]: E1213 06:53:07.341058 4971 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-cert-regeneration-controller" Dec 13 06:53:07 crc kubenswrapper[4971]: I1213 06:53:07.341080 4971 state_mem.go:107] "Deleted CPUSet assignment" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-cert-regeneration-controller" Dec 13 06:53:07 crc kubenswrapper[4971]: E1213 06:53:07.341088 4971 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-check-endpoints" Dec 13 06:53:07 crc kubenswrapper[4971]: I1213 06:53:07.341093 4971 state_mem.go:107] "Deleted CPUSet assignment" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-check-endpoints" Dec 13 06:53:07 crc kubenswrapper[4971]: E1213 06:53:07.341101 4971 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-check-endpoints" Dec 13 06:53:07 crc kubenswrapper[4971]: I1213 06:53:07.341108 4971 state_mem.go:107] "Deleted CPUSet assignment" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-check-endpoints" Dec 13 06:53:07 crc kubenswrapper[4971]: E1213 06:53:07.341125 4971 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-insecure-readyz" Dec 13 06:53:07 crc kubenswrapper[4971]: I1213 06:53:07.341131 4971 state_mem.go:107] "Deleted CPUSet assignment" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-insecure-readyz" Dec 13 06:53:07 crc kubenswrapper[4971]: E1213 06:53:07.341142 4971 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="setup" Dec 13 06:53:07 crc kubenswrapper[4971]: I1213 06:53:07.341163 4971 state_mem.go:107] "Deleted CPUSet assignment" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="setup" Dec 13 06:53:07 crc kubenswrapper[4971]: E1213 06:53:07.341169 4971 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver" Dec 13 06:53:07 crc kubenswrapper[4971]: I1213 06:53:07.341175 4971 state_mem.go:107] "Deleted CPUSet assignment" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver" Dec 13 06:53:07 crc kubenswrapper[4971]: I1213 06:53:07.341280 4971 memory_manager.go:354] "RemoveStaleState removing state" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-check-endpoints" Dec 13 06:53:07 crc kubenswrapper[4971]: I1213 06:53:07.341292 4971 memory_manager.go:354] "RemoveStaleState removing state" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-insecure-readyz" Dec 13 06:53:07 crc kubenswrapper[4971]: I1213 06:53:07.341319 4971 memory_manager.go:354] "RemoveStaleState removing state" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-cert-syncer" Dec 13 06:53:07 crc kubenswrapper[4971]: I1213 06:53:07.341328 4971 memory_manager.go:354] "RemoveStaleState removing state" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-cert-regeneration-controller" Dec 13 06:53:07 crc kubenswrapper[4971]: I1213 06:53:07.341335 4971 memory_manager.go:354] "RemoveStaleState removing state" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver" Dec 13 06:53:07 crc kubenswrapper[4971]: I1213 06:53:07.341567 4971 memory_manager.go:354] "RemoveStaleState removing state" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-check-endpoints" Dec 13 06:53:07 crc kubenswrapper[4971]: I1213 06:53:07.349266 4971 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/marketplace-operator-79b997595-6rvc9" event={"ID":"5c88a864-89a9-4f5e-b713-385170b9bf17","Type":"ContainerStarted","Data":"f93dd5acab7a80e5d527432e821725f3324ff27fa1176c1a26c332a22354f3be"} Dec 13 06:53:07 crc kubenswrapper[4971]: I1213 06:53:07.349392 4971 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-marketplace-xwxlf" podUID="22037bdc-fe6d-4411-be2b-aec38524a8dd" containerName="registry-server" containerID="cri-o://5e118dbb3b89959ae750afb59cd8dca72e876ecbeb215d5fb1c410b36e5e10c0" gracePeriod=30 Dec 13 06:53:07 crc kubenswrapper[4971]: I1213 06:53:07.364565 4971 status_manager.go:861] "Pod was deleted and then recreated, skipping status update" pod="openshift-kube-apiserver/kube-apiserver-crc" oldPodUID="f4b27818a5e8e43d0dc095d08835c792" podUID="71bb4a3aecc4ba5b26c4b7318770ce13" Dec 13 06:53:07 crc kubenswrapper[4971]: I1213 06:53:07.385089 4971 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-kube-apiserver/kube-apiserver-startup-monitor-crc"] Dec 13 06:53:07 crc kubenswrapper[4971]: I1213 06:53:07.489429 4971 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-resource-dir\") pod \"kube-apiserver-startup-monitor-crc\" (UID: \"f85e55b1a89d02b0cb034b1ea31ed45a\") " pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" Dec 13 06:53:07 crc kubenswrapper[4971]: I1213 06:53:07.489798 4971 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/71bb4a3aecc4ba5b26c4b7318770ce13-resource-dir\") pod \"kube-apiserver-crc\" (UID: \"71bb4a3aecc4ba5b26c4b7318770ce13\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Dec 13 06:53:07 crc kubenswrapper[4971]: I1213 06:53:07.489817 4971 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-lock\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-var-lock\") pod \"kube-apiserver-startup-monitor-crc\" (UID: \"f85e55b1a89d02b0cb034b1ea31ed45a\") " pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" Dec 13 06:53:07 crc kubenswrapper[4971]: I1213 06:53:07.489833 4971 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"pod-resource-dir\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-pod-resource-dir\") pod \"kube-apiserver-startup-monitor-crc\" (UID: \"f85e55b1a89d02b0cb034b1ea31ed45a\") " pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" Dec 13 06:53:07 crc kubenswrapper[4971]: I1213 06:53:07.490035 4971 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/71bb4a3aecc4ba5b26c4b7318770ce13-cert-dir\") pod \"kube-apiserver-crc\" (UID: \"71bb4a3aecc4ba5b26c4b7318770ce13\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Dec 13 06:53:07 crc kubenswrapper[4971]: I1213 06:53:07.490141 4971 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-log\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-var-log\") pod \"kube-apiserver-startup-monitor-crc\" (UID: \"f85e55b1a89d02b0cb034b1ea31ed45a\") " pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" Dec 13 06:53:07 crc kubenswrapper[4971]: I1213 06:53:07.490182 4971 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/71bb4a3aecc4ba5b26c4b7318770ce13-audit-dir\") pod \"kube-apiserver-crc\" (UID: \"71bb4a3aecc4ba5b26c4b7318770ce13\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Dec 13 06:53:07 crc kubenswrapper[4971]: I1213 06:53:07.490244 4971 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"manifests\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-manifests\") pod \"kube-apiserver-startup-monitor-crc\" (UID: \"f85e55b1a89d02b0cb034b1ea31ed45a\") " pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" Dec 13 06:53:07 crc kubenswrapper[4971]: I1213 06:53:07.591795 4971 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-resource-dir\") pod \"kube-apiserver-startup-monitor-crc\" (UID: \"f85e55b1a89d02b0cb034b1ea31ed45a\") " pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" Dec 13 06:53:07 crc kubenswrapper[4971]: I1213 06:53:07.591853 4971 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/71bb4a3aecc4ba5b26c4b7318770ce13-resource-dir\") pod \"kube-apiserver-crc\" (UID: \"71bb4a3aecc4ba5b26c4b7318770ce13\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Dec 13 06:53:07 crc kubenswrapper[4971]: I1213 06:53:07.591877 4971 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-lock\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-var-lock\") pod \"kube-apiserver-startup-monitor-crc\" (UID: \"f85e55b1a89d02b0cb034b1ea31ed45a\") " pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" Dec 13 06:53:07 crc kubenswrapper[4971]: I1213 06:53:07.591899 4971 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pod-resource-dir\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-pod-resource-dir\") pod \"kube-apiserver-startup-monitor-crc\" (UID: \"f85e55b1a89d02b0cb034b1ea31ed45a\") " pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" Dec 13 06:53:07 crc kubenswrapper[4971]: I1213 06:53:07.591937 4971 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/71bb4a3aecc4ba5b26c4b7318770ce13-cert-dir\") pod \"kube-apiserver-crc\" (UID: \"71bb4a3aecc4ba5b26c4b7318770ce13\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Dec 13 06:53:07 crc kubenswrapper[4971]: I1213 06:53:07.591937 4971 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/71bb4a3aecc4ba5b26c4b7318770ce13-resource-dir\") pod \"kube-apiserver-crc\" (UID: \"71bb4a3aecc4ba5b26c4b7318770ce13\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Dec 13 06:53:07 crc kubenswrapper[4971]: I1213 06:53:07.591945 4971 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-lock\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-var-lock\") pod \"kube-apiserver-startup-monitor-crc\" (UID: \"f85e55b1a89d02b0cb034b1ea31ed45a\") " pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" Dec 13 06:53:07 crc kubenswrapper[4971]: I1213 06:53:07.591946 4971 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-resource-dir\") pod \"kube-apiserver-startup-monitor-crc\" (UID: \"f85e55b1a89d02b0cb034b1ea31ed45a\") " pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" Dec 13 06:53:07 crc kubenswrapper[4971]: I1213 06:53:07.591995 4971 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/71bb4a3aecc4ba5b26c4b7318770ce13-cert-dir\") pod \"kube-apiserver-crc\" (UID: \"71bb4a3aecc4ba5b26c4b7318770ce13\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Dec 13 06:53:07 crc kubenswrapper[4971]: I1213 06:53:07.591996 4971 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-log\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-var-log\") pod \"kube-apiserver-startup-monitor-crc\" (UID: \"f85e55b1a89d02b0cb034b1ea31ed45a\") " pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" Dec 13 06:53:07 crc kubenswrapper[4971]: I1213 06:53:07.591969 4971 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-log\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-var-log\") pod \"kube-apiserver-startup-monitor-crc\" (UID: \"f85e55b1a89d02b0cb034b1ea31ed45a\") " pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" Dec 13 06:53:07 crc kubenswrapper[4971]: I1213 06:53:07.591996 4971 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"pod-resource-dir\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-pod-resource-dir\") pod \"kube-apiserver-startup-monitor-crc\" (UID: \"f85e55b1a89d02b0cb034b1ea31ed45a\") " pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" Dec 13 06:53:07 crc kubenswrapper[4971]: I1213 06:53:07.592063 4971 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/71bb4a3aecc4ba5b26c4b7318770ce13-audit-dir\") pod \"kube-apiserver-crc\" (UID: \"71bb4a3aecc4ba5b26c4b7318770ce13\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Dec 13 06:53:07 crc kubenswrapper[4971]: I1213 06:53:07.592106 4971 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"manifests\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-manifests\") pod \"kube-apiserver-startup-monitor-crc\" (UID: \"f85e55b1a89d02b0cb034b1ea31ed45a\") " pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" Dec 13 06:53:07 crc kubenswrapper[4971]: I1213 06:53:07.592170 4971 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/71bb4a3aecc4ba5b26c4b7318770ce13-audit-dir\") pod \"kube-apiserver-crc\" (UID: \"71bb4a3aecc4ba5b26c4b7318770ce13\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Dec 13 06:53:07 crc kubenswrapper[4971]: I1213 06:53:07.592237 4971 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"manifests\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-manifests\") pod \"kube-apiserver-startup-monitor-crc\" (UID: \"f85e55b1a89d02b0cb034b1ea31ed45a\") " pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" Dec 13 06:53:07 crc kubenswrapper[4971]: I1213 06:53:07.676130 4971 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" Dec 13 06:53:07 crc kubenswrapper[4971]: W1213 06:53:07.698331 4971 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podf85e55b1a89d02b0cb034b1ea31ed45a.slice/crio-295b500301b45d67ae0a74a2f97f744e423c169d2294d00479c6a454866d548e WatchSource:0}: Error finding container 295b500301b45d67ae0a74a2f97f744e423c169d2294d00479c6a454866d548e: Status 404 returned error can't find the container with id 295b500301b45d67ae0a74a2f97f744e423c169d2294d00479c6a454866d548e Dec 13 06:53:08 crc kubenswrapper[4971]: I1213 06:53:08.355132 4971 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" event={"ID":"f85e55b1a89d02b0cb034b1ea31ed45a","Type":"ContainerStarted","Data":"295b500301b45d67ae0a74a2f97f744e423c169d2294d00479c6a454866d548e"} Dec 13 06:53:08 crc kubenswrapper[4971]: I1213 06:53:08.355235 4971 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-marketplace-mkqps" podUID="9d4f1ea5-8472-4320-820b-8090ff6f5ca6" containerName="registry-server" containerID="cri-o://b98959ce12c11f9c6b25aaca858920c32124fb7a92b0ddee184e3e434e02168a" gracePeriod=30 Dec 13 06:53:10 crc kubenswrapper[4971]: E1213 06:53:10.263579 4971 event.go:368] "Unable to write event (may retry after sleeping)" err="Post \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/events\": dial tcp 38.102.83.195:6443: connect: connection refused" event="&Event{ObjectMeta:{kube-apiserver-startup-monitor-crc.1880b3d713eba826 openshift-kube-apiserver 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-kube-apiserver,Name:kube-apiserver-startup-monitor-crc,UID:f85e55b1a89d02b0cb034b1ea31ed45a,APIVersion:v1,ResourceVersion:,FieldPath:spec.containers{startup-monitor},},Reason:Created,Message:Created container startup-monitor,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2025-12-13 06:53:10.262786086 +0000 UTC m=+246.867195554,LastTimestamp:2025-12-13 06:53:10.262786086 +0000 UTC m=+246.867195554,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Dec 13 06:53:10 crc kubenswrapper[4971]: I1213 06:53:10.366754 4971 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/marketplace-operator-79b997595-6rvc9" event={"ID":"5c88a864-89a9-4f5e-b713-385170b9bf17","Type":"ContainerStarted","Data":"7a53386540871a14f654bd05dae1a7988d7a3396a65ccf45bb2e98846899cc87"} Dec 13 06:53:10 crc kubenswrapper[4971]: I1213 06:53:10.367225 4971 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/marketplace-operator-79b997595-6rvc9" Dec 13 06:53:10 crc kubenswrapper[4971]: I1213 06:53:10.367639 4971 status_manager.go:851] "Failed to get status for pod" podUID="5c88a864-89a9-4f5e-b713-385170b9bf17" pod="openshift-marketplace/marketplace-operator-79b997595-6rvc9" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/marketplace-operator-79b997595-6rvc9\": dial tcp 38.102.83.195:6443: connect: connection refused" Dec 13 06:53:10 crc kubenswrapper[4971]: I1213 06:53:10.368608 4971 patch_prober.go:28] interesting pod/marketplace-operator-79b997595-6rvc9 container/marketplace-operator namespace/openshift-marketplace: Readiness probe status=failure output="Get \"http://10.217.0.56:8080/healthz\": dial tcp 10.217.0.56:8080: connect: connection refused" start-of-body= Dec 13 06:53:10 crc kubenswrapper[4971]: I1213 06:53:10.368662 4971 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-marketplace/marketplace-operator-79b997595-6rvc9" podUID="5c88a864-89a9-4f5e-b713-385170b9bf17" containerName="marketplace-operator" probeResult="failure" output="Get \"http://10.217.0.56:8080/healthz\": dial tcp 10.217.0.56:8080: connect: connection refused" Dec 13 06:53:10 crc kubenswrapper[4971]: I1213 06:53:10.371124 4971 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-kube-apiserver_kube-apiserver-crc_f4b27818a5e8e43d0dc095d08835c792/kube-apiserver-check-endpoints/0.log" Dec 13 06:53:10 crc kubenswrapper[4971]: I1213 06:53:10.372468 4971 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-kube-apiserver_kube-apiserver-crc_f4b27818a5e8e43d0dc095d08835c792/kube-apiserver-cert-syncer/0.log" Dec 13 06:53:10 crc kubenswrapper[4971]: I1213 06:53:10.373118 4971 generic.go:334] "Generic (PLEG): container finished" podID="f4b27818a5e8e43d0dc095d08835c792" containerID="770b73dd0c2b6f38744e2c1cd6aa186a69bb87d1cd015a603ad2e33be06a8f6b" exitCode=0 Dec 13 06:53:10 crc kubenswrapper[4971]: I1213 06:53:10.373146 4971 generic.go:334] "Generic (PLEG): container finished" podID="f4b27818a5e8e43d0dc095d08835c792" containerID="9acd07066deed483bd7fa61ee45fda09539d110717a134a56827f871186f92fc" exitCode=0 Dec 13 06:53:10 crc kubenswrapper[4971]: I1213 06:53:10.373156 4971 generic.go:334] "Generic (PLEG): container finished" podID="f4b27818a5e8e43d0dc095d08835c792" containerID="30df3abf8c79db8af52c6a918f73e1ce410ddafcf341636424c267e1eeb195a7" exitCode=0 Dec 13 06:53:10 crc kubenswrapper[4971]: I1213 06:53:10.373166 4971 generic.go:334] "Generic (PLEG): container finished" podID="f4b27818a5e8e43d0dc095d08835c792" containerID="ccdf65f6d2b1eb3647a03ba37c305c32e34addfdf79ff937c8dddaa3e5d5b5c6" exitCode=2 Dec 13 06:53:10 crc kubenswrapper[4971]: I1213 06:53:10.373244 4971 scope.go:117] "RemoveContainer" containerID="b3c1bab3d53f975b2e24582ce2771bc26d600e0dd7c30339dcfd18a071162ad4" Dec 13 06:53:10 crc kubenswrapper[4971]: I1213 06:53:10.374851 4971 generic.go:334] "Generic (PLEG): container finished" podID="53fdcbc2-900b-48eb-8be3-cd72940096c3" containerID="0e7d033dee64f00f95af0d1a290dc8fafbfbfd51a34dabf8fca951acfe190bae" exitCode=0 Dec 13 06:53:10 crc kubenswrapper[4971]: I1213 06:53:10.374889 4971 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/installer-9-crc" event={"ID":"53fdcbc2-900b-48eb-8be3-cd72940096c3","Type":"ContainerDied","Data":"0e7d033dee64f00f95af0d1a290dc8fafbfbfd51a34dabf8fca951acfe190bae"} Dec 13 06:53:10 crc kubenswrapper[4971]: I1213 06:53:10.375446 4971 status_manager.go:851] "Failed to get status for pod" podUID="53fdcbc2-900b-48eb-8be3-cd72940096c3" pod="openshift-kube-apiserver/installer-9-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/installer-9-crc\": dial tcp 38.102.83.195:6443: connect: connection refused" Dec 13 06:53:10 crc kubenswrapper[4971]: I1213 06:53:10.375997 4971 status_manager.go:851] "Failed to get status for pod" podUID="5c88a864-89a9-4f5e-b713-385170b9bf17" pod="openshift-marketplace/marketplace-operator-79b997595-6rvc9" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/marketplace-operator-79b997595-6rvc9\": dial tcp 38.102.83.195:6443: connect: connection refused" Dec 13 06:53:10 crc kubenswrapper[4971]: I1213 06:53:10.377816 4971 generic.go:334] "Generic (PLEG): container finished" podID="8f50f5ba-3ec3-4fa2-9f5d-51e1c43f6352" containerID="cf192b8ffdcb439c2bd3f71dc37ac0dbd8e41220c791304f4d26a5436e9cc7e5" exitCode=0 Dec 13 06:53:10 crc kubenswrapper[4971]: I1213 06:53:10.377870 4971 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-rkd9c" event={"ID":"8f50f5ba-3ec3-4fa2-9f5d-51e1c43f6352","Type":"ContainerDied","Data":"cf192b8ffdcb439c2bd3f71dc37ac0dbd8e41220c791304f4d26a5436e9cc7e5"} Dec 13 06:53:10 crc kubenswrapper[4971]: I1213 06:53:10.379403 4971 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" event={"ID":"f85e55b1a89d02b0cb034b1ea31ed45a","Type":"ContainerStarted","Data":"a60d4a360ca22d5ab008969e58dd981a7858882ea44246f6b56a407dba706501"} Dec 13 06:53:10 crc kubenswrapper[4971]: I1213 06:53:10.380244 4971 status_manager.go:851] "Failed to get status for pod" podUID="53fdcbc2-900b-48eb-8be3-cd72940096c3" pod="openshift-kube-apiserver/installer-9-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/installer-9-crc\": dial tcp 38.102.83.195:6443: connect: connection refused" Dec 13 06:53:10 crc kubenswrapper[4971]: I1213 06:53:10.380432 4971 status_manager.go:851] "Failed to get status for pod" podUID="f85e55b1a89d02b0cb034b1ea31ed45a" pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-startup-monitor-crc\": dial tcp 38.102.83.195:6443: connect: connection refused" Dec 13 06:53:10 crc kubenswrapper[4971]: I1213 06:53:10.380887 4971 status_manager.go:851] "Failed to get status for pod" podUID="5c88a864-89a9-4f5e-b713-385170b9bf17" pod="openshift-marketplace/marketplace-operator-79b997595-6rvc9" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/marketplace-operator-79b997595-6rvc9\": dial tcp 38.102.83.195:6443: connect: connection refused" Dec 13 06:53:10 crc kubenswrapper[4971]: I1213 06:53:10.381841 4971 generic.go:334] "Generic (PLEG): container finished" podID="22037bdc-fe6d-4411-be2b-aec38524a8dd" containerID="5e118dbb3b89959ae750afb59cd8dca72e876ecbeb215d5fb1c410b36e5e10c0" exitCode=0 Dec 13 06:53:10 crc kubenswrapper[4971]: I1213 06:53:10.381877 4971 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-xwxlf" event={"ID":"22037bdc-fe6d-4411-be2b-aec38524a8dd","Type":"ContainerDied","Data":"5e118dbb3b89959ae750afb59cd8dca72e876ecbeb215d5fb1c410b36e5e10c0"} Dec 13 06:53:10 crc kubenswrapper[4971]: I1213 06:53:10.384085 4971 generic.go:334] "Generic (PLEG): container finished" podID="cf417851-ae43-462f-a678-125593fb653b" containerID="6e9990367538c8d28196c610c227d5744eafd1a07907c88fcc963201ce0f06cc" exitCode=0 Dec 13 06:53:10 crc kubenswrapper[4971]: I1213 06:53:10.384152 4971 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-997kc" event={"ID":"cf417851-ae43-462f-a678-125593fb653b","Type":"ContainerDied","Data":"6e9990367538c8d28196c610c227d5744eafd1a07907c88fcc963201ce0f06cc"} Dec 13 06:53:10 crc kubenswrapper[4971]: I1213 06:53:10.386624 4971 generic.go:334] "Generic (PLEG): container finished" podID="9d4f1ea5-8472-4320-820b-8090ff6f5ca6" containerID="b98959ce12c11f9c6b25aaca858920c32124fb7a92b0ddee184e3e434e02168a" exitCode=0 Dec 13 06:53:10 crc kubenswrapper[4971]: I1213 06:53:10.386693 4971 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-mkqps" event={"ID":"9d4f1ea5-8472-4320-820b-8090ff6f5ca6","Type":"ContainerDied","Data":"b98959ce12c11f9c6b25aaca858920c32124fb7a92b0ddee184e3e434e02168a"} Dec 13 06:53:10 crc kubenswrapper[4971]: I1213 06:53:10.388338 4971 generic.go:334] "Generic (PLEG): container finished" podID="02f6e9f2-6bad-441c-8e2e-9acfc71f3ab4" containerID="cf5a955add1c7bf17976d195f2fba78c93d0fd447a0684c6e8377e595e9546ad" exitCode=0 Dec 13 06:53:10 crc kubenswrapper[4971]: I1213 06:53:10.388412 4971 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-wpxfk" event={"ID":"02f6e9f2-6bad-441c-8e2e-9acfc71f3ab4","Type":"ContainerDied","Data":"cf5a955add1c7bf17976d195f2fba78c93d0fd447a0684c6e8377e595e9546ad"} Dec 13 06:53:10 crc kubenswrapper[4971]: I1213 06:53:10.389896 4971 generic.go:334] "Generic (PLEG): container finished" podID="b9a0bc5d-3a39-4732-a849-22d351a4cc18" containerID="14fb548fd8178b7fa4ade14c8907b148dd031b77bdcd95bcd98d8b54fc688a83" exitCode=0 Dec 13 06:53:10 crc kubenswrapper[4971]: I1213 06:53:10.389970 4971 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-d8n8p" event={"ID":"b9a0bc5d-3a39-4732-a849-22d351a4cc18","Type":"ContainerDied","Data":"14fb548fd8178b7fa4ade14c8907b148dd031b77bdcd95bcd98d8b54fc688a83"} Dec 13 06:53:10 crc kubenswrapper[4971]: I1213 06:53:10.390819 4971 generic.go:334] "Generic (PLEG): container finished" podID="f14c8a94-8589-4059-bddf-329452355fab" containerID="524506c76ba56186a2daa60747573baef0eb42cebe8bb321e82be093150f35d7" exitCode=0 Dec 13 06:53:10 crc kubenswrapper[4971]: I1213 06:53:10.390850 4971 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/marketplace-operator-79b997595-j9zqf" event={"ID":"f14c8a94-8589-4059-bddf-329452355fab","Type":"ContainerDied","Data":"524506c76ba56186a2daa60747573baef0eb42cebe8bb321e82be093150f35d7"} Dec 13 06:53:10 crc kubenswrapper[4971]: I1213 06:53:10.826984 4971 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-rkd9c" Dec 13 06:53:10 crc kubenswrapper[4971]: I1213 06:53:10.828334 4971 status_manager.go:851] "Failed to get status for pod" podUID="8f50f5ba-3ec3-4fa2-9f5d-51e1c43f6352" pod="openshift-marketplace/certified-operators-rkd9c" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/certified-operators-rkd9c\": dial tcp 38.102.83.195:6443: connect: connection refused" Dec 13 06:53:10 crc kubenswrapper[4971]: I1213 06:53:10.828996 4971 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/marketplace-operator-79b997595-j9zqf" Dec 13 06:53:10 crc kubenswrapper[4971]: I1213 06:53:10.829327 4971 status_manager.go:851] "Failed to get status for pod" podUID="5c88a864-89a9-4f5e-b713-385170b9bf17" pod="openshift-marketplace/marketplace-operator-79b997595-6rvc9" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/marketplace-operator-79b997595-6rvc9\": dial tcp 38.102.83.195:6443: connect: connection refused" Dec 13 06:53:10 crc kubenswrapper[4971]: I1213 06:53:10.829573 4971 status_manager.go:851] "Failed to get status for pod" podUID="53fdcbc2-900b-48eb-8be3-cd72940096c3" pod="openshift-kube-apiserver/installer-9-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/installer-9-crc\": dial tcp 38.102.83.195:6443: connect: connection refused" Dec 13 06:53:10 crc kubenswrapper[4971]: I1213 06:53:10.829806 4971 status_manager.go:851] "Failed to get status for pod" podUID="f85e55b1a89d02b0cb034b1ea31ed45a" pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-startup-monitor-crc\": dial tcp 38.102.83.195:6443: connect: connection refused" Dec 13 06:53:10 crc kubenswrapper[4971]: I1213 06:53:10.830532 4971 status_manager.go:851] "Failed to get status for pod" podUID="8f50f5ba-3ec3-4fa2-9f5d-51e1c43f6352" pod="openshift-marketplace/certified-operators-rkd9c" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/certified-operators-rkd9c\": dial tcp 38.102.83.195:6443: connect: connection refused" Dec 13 06:53:10 crc kubenswrapper[4971]: I1213 06:53:10.830931 4971 status_manager.go:851] "Failed to get status for pod" podUID="5c88a864-89a9-4f5e-b713-385170b9bf17" pod="openshift-marketplace/marketplace-operator-79b997595-6rvc9" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/marketplace-operator-79b997595-6rvc9\": dial tcp 38.102.83.195:6443: connect: connection refused" Dec 13 06:53:10 crc kubenswrapper[4971]: I1213 06:53:10.831159 4971 status_manager.go:851] "Failed to get status for pod" podUID="53fdcbc2-900b-48eb-8be3-cd72940096c3" pod="openshift-kube-apiserver/installer-9-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/installer-9-crc\": dial tcp 38.102.83.195:6443: connect: connection refused" Dec 13 06:53:10 crc kubenswrapper[4971]: I1213 06:53:10.831418 4971 status_manager.go:851] "Failed to get status for pod" podUID="f14c8a94-8589-4059-bddf-329452355fab" pod="openshift-marketplace/marketplace-operator-79b997595-j9zqf" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/marketplace-operator-79b997595-j9zqf\": dial tcp 38.102.83.195:6443: connect: connection refused" Dec 13 06:53:10 crc kubenswrapper[4971]: I1213 06:53:10.831917 4971 status_manager.go:851] "Failed to get status for pod" podUID="f85e55b1a89d02b0cb034b1ea31ed45a" pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-startup-monitor-crc\": dial tcp 38.102.83.195:6443: connect: connection refused" Dec 13 06:53:10 crc kubenswrapper[4971]: I1213 06:53:10.833425 4971 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-wpxfk" Dec 13 06:53:10 crc kubenswrapper[4971]: I1213 06:53:10.833659 4971 status_manager.go:851] "Failed to get status for pod" podUID="5c88a864-89a9-4f5e-b713-385170b9bf17" pod="openshift-marketplace/marketplace-operator-79b997595-6rvc9" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/marketplace-operator-79b997595-6rvc9\": dial tcp 38.102.83.195:6443: connect: connection refused" Dec 13 06:53:10 crc kubenswrapper[4971]: I1213 06:53:10.833809 4971 status_manager.go:851] "Failed to get status for pod" podUID="8f50f5ba-3ec3-4fa2-9f5d-51e1c43f6352" pod="openshift-marketplace/certified-operators-rkd9c" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/certified-operators-rkd9c\": dial tcp 38.102.83.195:6443: connect: connection refused" Dec 13 06:53:10 crc kubenswrapper[4971]: I1213 06:53:10.834039 4971 status_manager.go:851] "Failed to get status for pod" podUID="02f6e9f2-6bad-441c-8e2e-9acfc71f3ab4" pod="openshift-marketplace/redhat-operators-wpxfk" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/redhat-operators-wpxfk\": dial tcp 38.102.83.195:6443: connect: connection refused" Dec 13 06:53:10 crc kubenswrapper[4971]: I1213 06:53:10.834400 4971 status_manager.go:851] "Failed to get status for pod" podUID="53fdcbc2-900b-48eb-8be3-cd72940096c3" pod="openshift-kube-apiserver/installer-9-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/installer-9-crc\": dial tcp 38.102.83.195:6443: connect: connection refused" Dec 13 06:53:10 crc kubenswrapper[4971]: I1213 06:53:10.834624 4971 status_manager.go:851] "Failed to get status for pod" podUID="f85e55b1a89d02b0cb034b1ea31ed45a" pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-startup-monitor-crc\": dial tcp 38.102.83.195:6443: connect: connection refused" Dec 13 06:53:10 crc kubenswrapper[4971]: I1213 06:53:10.834966 4971 status_manager.go:851] "Failed to get status for pod" podUID="f14c8a94-8589-4059-bddf-329452355fab" pod="openshift-marketplace/marketplace-operator-79b997595-j9zqf" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/marketplace-operator-79b997595-j9zqf\": dial tcp 38.102.83.195:6443: connect: connection refused" Dec 13 06:53:10 crc kubenswrapper[4971]: I1213 06:53:10.838165 4971 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-d8n8p" Dec 13 06:53:10 crc kubenswrapper[4971]: I1213 06:53:10.838569 4971 status_manager.go:851] "Failed to get status for pod" podUID="8f50f5ba-3ec3-4fa2-9f5d-51e1c43f6352" pod="openshift-marketplace/certified-operators-rkd9c" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/certified-operators-rkd9c\": dial tcp 38.102.83.195:6443: connect: connection refused" Dec 13 06:53:10 crc kubenswrapper[4971]: I1213 06:53:10.838991 4971 status_manager.go:851] "Failed to get status for pod" podUID="02f6e9f2-6bad-441c-8e2e-9acfc71f3ab4" pod="openshift-marketplace/redhat-operators-wpxfk" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/redhat-operators-wpxfk\": dial tcp 38.102.83.195:6443: connect: connection refused" Dec 13 06:53:10 crc kubenswrapper[4971]: I1213 06:53:10.839298 4971 status_manager.go:851] "Failed to get status for pod" podUID="5c88a864-89a9-4f5e-b713-385170b9bf17" pod="openshift-marketplace/marketplace-operator-79b997595-6rvc9" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/marketplace-operator-79b997595-6rvc9\": dial tcp 38.102.83.195:6443: connect: connection refused" Dec 13 06:53:10 crc kubenswrapper[4971]: I1213 06:53:10.840145 4971 status_manager.go:851] "Failed to get status for pod" podUID="53fdcbc2-900b-48eb-8be3-cd72940096c3" pod="openshift-kube-apiserver/installer-9-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/installer-9-crc\": dial tcp 38.102.83.195:6443: connect: connection refused" Dec 13 06:53:10 crc kubenswrapper[4971]: I1213 06:53:10.840370 4971 status_manager.go:851] "Failed to get status for pod" podUID="f14c8a94-8589-4059-bddf-329452355fab" pod="openshift-marketplace/marketplace-operator-79b997595-j9zqf" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/marketplace-operator-79b997595-j9zqf\": dial tcp 38.102.83.195:6443: connect: connection refused" Dec 13 06:53:10 crc kubenswrapper[4971]: I1213 06:53:10.840580 4971 status_manager.go:851] "Failed to get status for pod" podUID="b9a0bc5d-3a39-4732-a849-22d351a4cc18" pod="openshift-marketplace/redhat-operators-d8n8p" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/redhat-operators-d8n8p\": dial tcp 38.102.83.195:6443: connect: connection refused" Dec 13 06:53:10 crc kubenswrapper[4971]: I1213 06:53:10.840792 4971 status_manager.go:851] "Failed to get status for pod" podUID="f85e55b1a89d02b0cb034b1ea31ed45a" pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-startup-monitor-crc\": dial tcp 38.102.83.195:6443: connect: connection refused" Dec 13 06:53:10 crc kubenswrapper[4971]: I1213 06:53:10.842631 4971 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-997kc" Dec 13 06:53:10 crc kubenswrapper[4971]: I1213 06:53:10.842944 4971 status_manager.go:851] "Failed to get status for pod" podUID="cf417851-ae43-462f-a678-125593fb653b" pod="openshift-marketplace/community-operators-997kc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/community-operators-997kc\": dial tcp 38.102.83.195:6443: connect: connection refused" Dec 13 06:53:10 crc kubenswrapper[4971]: I1213 06:53:10.843169 4971 status_manager.go:851] "Failed to get status for pod" podUID="8f50f5ba-3ec3-4fa2-9f5d-51e1c43f6352" pod="openshift-marketplace/certified-operators-rkd9c" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/certified-operators-rkd9c\": dial tcp 38.102.83.195:6443: connect: connection refused" Dec 13 06:53:10 crc kubenswrapper[4971]: I1213 06:53:10.843329 4971 status_manager.go:851] "Failed to get status for pod" podUID="02f6e9f2-6bad-441c-8e2e-9acfc71f3ab4" pod="openshift-marketplace/redhat-operators-wpxfk" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/redhat-operators-wpxfk\": dial tcp 38.102.83.195:6443: connect: connection refused" Dec 13 06:53:10 crc kubenswrapper[4971]: I1213 06:53:10.843493 4971 status_manager.go:851] "Failed to get status for pod" podUID="5c88a864-89a9-4f5e-b713-385170b9bf17" pod="openshift-marketplace/marketplace-operator-79b997595-6rvc9" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/marketplace-operator-79b997595-6rvc9\": dial tcp 38.102.83.195:6443: connect: connection refused" Dec 13 06:53:10 crc kubenswrapper[4971]: I1213 06:53:10.843845 4971 status_manager.go:851] "Failed to get status for pod" podUID="53fdcbc2-900b-48eb-8be3-cd72940096c3" pod="openshift-kube-apiserver/installer-9-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/installer-9-crc\": dial tcp 38.102.83.195:6443: connect: connection refused" Dec 13 06:53:10 crc kubenswrapper[4971]: I1213 06:53:10.844028 4971 status_manager.go:851] "Failed to get status for pod" podUID="f14c8a94-8589-4059-bddf-329452355fab" pod="openshift-marketplace/marketplace-operator-79b997595-j9zqf" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/marketplace-operator-79b997595-j9zqf\": dial tcp 38.102.83.195:6443: connect: connection refused" Dec 13 06:53:10 crc kubenswrapper[4971]: I1213 06:53:10.844207 4971 status_manager.go:851] "Failed to get status for pod" podUID="b9a0bc5d-3a39-4732-a849-22d351a4cc18" pod="openshift-marketplace/redhat-operators-d8n8p" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/redhat-operators-d8n8p\": dial tcp 38.102.83.195:6443: connect: connection refused" Dec 13 06:53:10 crc kubenswrapper[4971]: I1213 06:53:10.844409 4971 status_manager.go:851] "Failed to get status for pod" podUID="f85e55b1a89d02b0cb034b1ea31ed45a" pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-startup-monitor-crc\": dial tcp 38.102.83.195:6443: connect: connection refused" Dec 13 06:53:10 crc kubenswrapper[4971]: I1213 06:53:10.916016 4971 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-xwxlf" Dec 13 06:53:10 crc kubenswrapper[4971]: I1213 06:53:10.916475 4971 status_manager.go:851] "Failed to get status for pod" podUID="b9a0bc5d-3a39-4732-a849-22d351a4cc18" pod="openshift-marketplace/redhat-operators-d8n8p" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/redhat-operators-d8n8p\": dial tcp 38.102.83.195:6443: connect: connection refused" Dec 13 06:53:10 crc kubenswrapper[4971]: I1213 06:53:10.916654 4971 status_manager.go:851] "Failed to get status for pod" podUID="f85e55b1a89d02b0cb034b1ea31ed45a" pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-startup-monitor-crc\": dial tcp 38.102.83.195:6443: connect: connection refused" Dec 13 06:53:10 crc kubenswrapper[4971]: I1213 06:53:10.917006 4971 status_manager.go:851] "Failed to get status for pod" podUID="f14c8a94-8589-4059-bddf-329452355fab" pod="openshift-marketplace/marketplace-operator-79b997595-j9zqf" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/marketplace-operator-79b997595-j9zqf\": dial tcp 38.102.83.195:6443: connect: connection refused" Dec 13 06:53:10 crc kubenswrapper[4971]: I1213 06:53:10.917469 4971 status_manager.go:851] "Failed to get status for pod" podUID="cf417851-ae43-462f-a678-125593fb653b" pod="openshift-marketplace/community-operators-997kc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/community-operators-997kc\": dial tcp 38.102.83.195:6443: connect: connection refused" Dec 13 06:53:10 crc kubenswrapper[4971]: I1213 06:53:10.917688 4971 status_manager.go:851] "Failed to get status for pod" podUID="5c88a864-89a9-4f5e-b713-385170b9bf17" pod="openshift-marketplace/marketplace-operator-79b997595-6rvc9" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/marketplace-operator-79b997595-6rvc9\": dial tcp 38.102.83.195:6443: connect: connection refused" Dec 13 06:53:10 crc kubenswrapper[4971]: I1213 06:53:10.917885 4971 status_manager.go:851] "Failed to get status for pod" podUID="8f50f5ba-3ec3-4fa2-9f5d-51e1c43f6352" pod="openshift-marketplace/certified-operators-rkd9c" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/certified-operators-rkd9c\": dial tcp 38.102.83.195:6443: connect: connection refused" Dec 13 06:53:10 crc kubenswrapper[4971]: I1213 06:53:10.918060 4971 status_manager.go:851] "Failed to get status for pod" podUID="22037bdc-fe6d-4411-be2b-aec38524a8dd" pod="openshift-marketplace/redhat-marketplace-xwxlf" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/redhat-marketplace-xwxlf\": dial tcp 38.102.83.195:6443: connect: connection refused" Dec 13 06:53:10 crc kubenswrapper[4971]: I1213 06:53:10.918205 4971 status_manager.go:851] "Failed to get status for pod" podUID="02f6e9f2-6bad-441c-8e2e-9acfc71f3ab4" pod="openshift-marketplace/redhat-operators-wpxfk" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/redhat-operators-wpxfk\": dial tcp 38.102.83.195:6443: connect: connection refused" Dec 13 06:53:10 crc kubenswrapper[4971]: I1213 06:53:10.918369 4971 status_manager.go:851] "Failed to get status for pod" podUID="53fdcbc2-900b-48eb-8be3-cd72940096c3" pod="openshift-kube-apiserver/installer-9-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/installer-9-crc\": dial tcp 38.102.83.195:6443: connect: connection refused" Dec 13 06:53:10 crc kubenswrapper[4971]: I1213 06:53:10.921664 4971 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-mkqps" Dec 13 06:53:10 crc kubenswrapper[4971]: I1213 06:53:10.921971 4971 status_manager.go:851] "Failed to get status for pod" podUID="5c88a864-89a9-4f5e-b713-385170b9bf17" pod="openshift-marketplace/marketplace-operator-79b997595-6rvc9" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/marketplace-operator-79b997595-6rvc9\": dial tcp 38.102.83.195:6443: connect: connection refused" Dec 13 06:53:10 crc kubenswrapper[4971]: I1213 06:53:10.922165 4971 status_manager.go:851] "Failed to get status for pod" podUID="8f50f5ba-3ec3-4fa2-9f5d-51e1c43f6352" pod="openshift-marketplace/certified-operators-rkd9c" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/certified-operators-rkd9c\": dial tcp 38.102.83.195:6443: connect: connection refused" Dec 13 06:53:10 crc kubenswrapper[4971]: I1213 06:53:10.922390 4971 status_manager.go:851] "Failed to get status for pod" podUID="22037bdc-fe6d-4411-be2b-aec38524a8dd" pod="openshift-marketplace/redhat-marketplace-xwxlf" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/redhat-marketplace-xwxlf\": dial tcp 38.102.83.195:6443: connect: connection refused" Dec 13 06:53:10 crc kubenswrapper[4971]: I1213 06:53:10.922611 4971 status_manager.go:851] "Failed to get status for pod" podUID="02f6e9f2-6bad-441c-8e2e-9acfc71f3ab4" pod="openshift-marketplace/redhat-operators-wpxfk" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/redhat-operators-wpxfk\": dial tcp 38.102.83.195:6443: connect: connection refused" Dec 13 06:53:10 crc kubenswrapper[4971]: I1213 06:53:10.922795 4971 status_manager.go:851] "Failed to get status for pod" podUID="9d4f1ea5-8472-4320-820b-8090ff6f5ca6" pod="openshift-marketplace/redhat-marketplace-mkqps" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/redhat-marketplace-mkqps\": dial tcp 38.102.83.195:6443: connect: connection refused" Dec 13 06:53:10 crc kubenswrapper[4971]: I1213 06:53:10.922992 4971 status_manager.go:851] "Failed to get status for pod" podUID="53fdcbc2-900b-48eb-8be3-cd72940096c3" pod="openshift-kube-apiserver/installer-9-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/installer-9-crc\": dial tcp 38.102.83.195:6443: connect: connection refused" Dec 13 06:53:10 crc kubenswrapper[4971]: I1213 06:53:10.923173 4971 status_manager.go:851] "Failed to get status for pod" podUID="b9a0bc5d-3a39-4732-a849-22d351a4cc18" pod="openshift-marketplace/redhat-operators-d8n8p" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/redhat-operators-d8n8p\": dial tcp 38.102.83.195:6443: connect: connection refused" Dec 13 06:53:10 crc kubenswrapper[4971]: I1213 06:53:10.923395 4971 status_manager.go:851] "Failed to get status for pod" podUID="f85e55b1a89d02b0cb034b1ea31ed45a" pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-startup-monitor-crc\": dial tcp 38.102.83.195:6443: connect: connection refused" Dec 13 06:53:10 crc kubenswrapper[4971]: I1213 06:53:10.923616 4971 status_manager.go:851] "Failed to get status for pod" podUID="f14c8a94-8589-4059-bddf-329452355fab" pod="openshift-marketplace/marketplace-operator-79b997595-j9zqf" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/marketplace-operator-79b997595-j9zqf\": dial tcp 38.102.83.195:6443: connect: connection refused" Dec 13 06:53:10 crc kubenswrapper[4971]: I1213 06:53:10.923798 4971 status_manager.go:851] "Failed to get status for pod" podUID="cf417851-ae43-462f-a678-125593fb653b" pod="openshift-marketplace/community-operators-997kc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/community-operators-997kc\": dial tcp 38.102.83.195:6443: connect: connection refused" Dec 13 06:53:10 crc kubenswrapper[4971]: I1213 06:53:10.941331 4971 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/b9a0bc5d-3a39-4732-a849-22d351a4cc18-catalog-content\") pod \"b9a0bc5d-3a39-4732-a849-22d351a4cc18\" (UID: \"b9a0bc5d-3a39-4732-a849-22d351a4cc18\") " Dec 13 06:53:10 crc kubenswrapper[4971]: I1213 06:53:10.941399 4971 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/02f6e9f2-6bad-441c-8e2e-9acfc71f3ab4-utilities\") pod \"02f6e9f2-6bad-441c-8e2e-9acfc71f3ab4\" (UID: \"02f6e9f2-6bad-441c-8e2e-9acfc71f3ab4\") " Dec 13 06:53:10 crc kubenswrapper[4971]: I1213 06:53:10.941431 4971 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/cf417851-ae43-462f-a678-125593fb653b-utilities\") pod \"cf417851-ae43-462f-a678-125593fb653b\" (UID: \"cf417851-ae43-462f-a678-125593fb653b\") " Dec 13 06:53:10 crc kubenswrapper[4971]: I1213 06:53:10.941480 4971 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/b9a0bc5d-3a39-4732-a849-22d351a4cc18-utilities\") pod \"b9a0bc5d-3a39-4732-a849-22d351a4cc18\" (UID: \"b9a0bc5d-3a39-4732-a849-22d351a4cc18\") " Dec 13 06:53:10 crc kubenswrapper[4971]: I1213 06:53:10.941497 4971 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"marketplace-trusted-ca\" (UniqueName: \"kubernetes.io/configmap/f14c8a94-8589-4059-bddf-329452355fab-marketplace-trusted-ca\") pod \"f14c8a94-8589-4059-bddf-329452355fab\" (UID: \"f14c8a94-8589-4059-bddf-329452355fab\") " Dec 13 06:53:10 crc kubenswrapper[4971]: I1213 06:53:10.941529 4971 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-55kcz\" (UniqueName: \"kubernetes.io/projected/b9a0bc5d-3a39-4732-a849-22d351a4cc18-kube-api-access-55kcz\") pod \"b9a0bc5d-3a39-4732-a849-22d351a4cc18\" (UID: \"b9a0bc5d-3a39-4732-a849-22d351a4cc18\") " Dec 13 06:53:10 crc kubenswrapper[4971]: I1213 06:53:10.941546 4971 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-t7lfq\" (UniqueName: \"kubernetes.io/projected/cf417851-ae43-462f-a678-125593fb653b-kube-api-access-t7lfq\") pod \"cf417851-ae43-462f-a678-125593fb653b\" (UID: \"cf417851-ae43-462f-a678-125593fb653b\") " Dec 13 06:53:10 crc kubenswrapper[4971]: I1213 06:53:10.941570 4971 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-d9kbv\" (UniqueName: \"kubernetes.io/projected/02f6e9f2-6bad-441c-8e2e-9acfc71f3ab4-kube-api-access-d9kbv\") pod \"02f6e9f2-6bad-441c-8e2e-9acfc71f3ab4\" (UID: \"02f6e9f2-6bad-441c-8e2e-9acfc71f3ab4\") " Dec 13 06:53:10 crc kubenswrapper[4971]: I1213 06:53:10.941612 4971 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-tllvh\" (UniqueName: \"kubernetes.io/projected/f14c8a94-8589-4059-bddf-329452355fab-kube-api-access-tllvh\") pod \"f14c8a94-8589-4059-bddf-329452355fab\" (UID: \"f14c8a94-8589-4059-bddf-329452355fab\") " Dec 13 06:53:10 crc kubenswrapper[4971]: I1213 06:53:10.941644 4971 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"marketplace-operator-metrics\" (UniqueName: \"kubernetes.io/secret/f14c8a94-8589-4059-bddf-329452355fab-marketplace-operator-metrics\") pod \"f14c8a94-8589-4059-bddf-329452355fab\" (UID: \"f14c8a94-8589-4059-bddf-329452355fab\") " Dec 13 06:53:10 crc kubenswrapper[4971]: I1213 06:53:10.941677 4971 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/cf417851-ae43-462f-a678-125593fb653b-catalog-content\") pod \"cf417851-ae43-462f-a678-125593fb653b\" (UID: \"cf417851-ae43-462f-a678-125593fb653b\") " Dec 13 06:53:10 crc kubenswrapper[4971]: I1213 06:53:10.941703 4971 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/8f50f5ba-3ec3-4fa2-9f5d-51e1c43f6352-utilities\") pod \"8f50f5ba-3ec3-4fa2-9f5d-51e1c43f6352\" (UID: \"8f50f5ba-3ec3-4fa2-9f5d-51e1c43f6352\") " Dec 13 06:53:10 crc kubenswrapper[4971]: I1213 06:53:10.941723 4971 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-stj24\" (UniqueName: \"kubernetes.io/projected/8f50f5ba-3ec3-4fa2-9f5d-51e1c43f6352-kube-api-access-stj24\") pod \"8f50f5ba-3ec3-4fa2-9f5d-51e1c43f6352\" (UID: \"8f50f5ba-3ec3-4fa2-9f5d-51e1c43f6352\") " Dec 13 06:53:10 crc kubenswrapper[4971]: I1213 06:53:10.941751 4971 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/8f50f5ba-3ec3-4fa2-9f5d-51e1c43f6352-catalog-content\") pod \"8f50f5ba-3ec3-4fa2-9f5d-51e1c43f6352\" (UID: \"8f50f5ba-3ec3-4fa2-9f5d-51e1c43f6352\") " Dec 13 06:53:10 crc kubenswrapper[4971]: I1213 06:53:10.941768 4971 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/02f6e9f2-6bad-441c-8e2e-9acfc71f3ab4-catalog-content\") pod \"02f6e9f2-6bad-441c-8e2e-9acfc71f3ab4\" (UID: \"02f6e9f2-6bad-441c-8e2e-9acfc71f3ab4\") " Dec 13 06:53:10 crc kubenswrapper[4971]: I1213 06:53:10.942290 4971 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/02f6e9f2-6bad-441c-8e2e-9acfc71f3ab4-utilities" (OuterVolumeSpecName: "utilities") pod "02f6e9f2-6bad-441c-8e2e-9acfc71f3ab4" (UID: "02f6e9f2-6bad-441c-8e2e-9acfc71f3ab4"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 13 06:53:10 crc kubenswrapper[4971]: I1213 06:53:10.942296 4971 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/f14c8a94-8589-4059-bddf-329452355fab-marketplace-trusted-ca" (OuterVolumeSpecName: "marketplace-trusted-ca") pod "f14c8a94-8589-4059-bddf-329452355fab" (UID: "f14c8a94-8589-4059-bddf-329452355fab"). InnerVolumeSpecName "marketplace-trusted-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 13 06:53:10 crc kubenswrapper[4971]: I1213 06:53:10.942639 4971 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/8f50f5ba-3ec3-4fa2-9f5d-51e1c43f6352-utilities" (OuterVolumeSpecName: "utilities") pod "8f50f5ba-3ec3-4fa2-9f5d-51e1c43f6352" (UID: "8f50f5ba-3ec3-4fa2-9f5d-51e1c43f6352"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 13 06:53:10 crc kubenswrapper[4971]: I1213 06:53:10.942864 4971 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/b9a0bc5d-3a39-4732-a849-22d351a4cc18-utilities" (OuterVolumeSpecName: "utilities") pod "b9a0bc5d-3a39-4732-a849-22d351a4cc18" (UID: "b9a0bc5d-3a39-4732-a849-22d351a4cc18"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 13 06:53:10 crc kubenswrapper[4971]: I1213 06:53:10.942925 4971 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/cf417851-ae43-462f-a678-125593fb653b-utilities" (OuterVolumeSpecName: "utilities") pod "cf417851-ae43-462f-a678-125593fb653b" (UID: "cf417851-ae43-462f-a678-125593fb653b"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 13 06:53:10 crc kubenswrapper[4971]: I1213 06:53:10.946871 4971 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/8f50f5ba-3ec3-4fa2-9f5d-51e1c43f6352-kube-api-access-stj24" (OuterVolumeSpecName: "kube-api-access-stj24") pod "8f50f5ba-3ec3-4fa2-9f5d-51e1c43f6352" (UID: "8f50f5ba-3ec3-4fa2-9f5d-51e1c43f6352"). InnerVolumeSpecName "kube-api-access-stj24". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 13 06:53:10 crc kubenswrapper[4971]: I1213 06:53:10.946913 4971 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/cf417851-ae43-462f-a678-125593fb653b-kube-api-access-t7lfq" (OuterVolumeSpecName: "kube-api-access-t7lfq") pod "cf417851-ae43-462f-a678-125593fb653b" (UID: "cf417851-ae43-462f-a678-125593fb653b"). InnerVolumeSpecName "kube-api-access-t7lfq". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 13 06:53:10 crc kubenswrapper[4971]: I1213 06:53:10.947075 4971 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/b9a0bc5d-3a39-4732-a849-22d351a4cc18-kube-api-access-55kcz" (OuterVolumeSpecName: "kube-api-access-55kcz") pod "b9a0bc5d-3a39-4732-a849-22d351a4cc18" (UID: "b9a0bc5d-3a39-4732-a849-22d351a4cc18"). InnerVolumeSpecName "kube-api-access-55kcz". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 13 06:53:10 crc kubenswrapper[4971]: I1213 06:53:10.947249 4971 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/02f6e9f2-6bad-441c-8e2e-9acfc71f3ab4-kube-api-access-d9kbv" (OuterVolumeSpecName: "kube-api-access-d9kbv") pod "02f6e9f2-6bad-441c-8e2e-9acfc71f3ab4" (UID: "02f6e9f2-6bad-441c-8e2e-9acfc71f3ab4"). InnerVolumeSpecName "kube-api-access-d9kbv". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 13 06:53:10 crc kubenswrapper[4971]: I1213 06:53:10.948043 4971 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/f14c8a94-8589-4059-bddf-329452355fab-marketplace-operator-metrics" (OuterVolumeSpecName: "marketplace-operator-metrics") pod "f14c8a94-8589-4059-bddf-329452355fab" (UID: "f14c8a94-8589-4059-bddf-329452355fab"). InnerVolumeSpecName "marketplace-operator-metrics". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 13 06:53:10 crc kubenswrapper[4971]: I1213 06:53:10.956168 4971 reconciler_common.go:293] "Volume detached for volume \"marketplace-operator-metrics\" (UniqueName: \"kubernetes.io/secret/f14c8a94-8589-4059-bddf-329452355fab-marketplace-operator-metrics\") on node \"crc\" DevicePath \"\"" Dec 13 06:53:10 crc kubenswrapper[4971]: I1213 06:53:10.956346 4971 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-stj24\" (UniqueName: \"kubernetes.io/projected/8f50f5ba-3ec3-4fa2-9f5d-51e1c43f6352-kube-api-access-stj24\") on node \"crc\" DevicePath \"\"" Dec 13 06:53:10 crc kubenswrapper[4971]: I1213 06:53:10.956390 4971 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/8f50f5ba-3ec3-4fa2-9f5d-51e1c43f6352-utilities\") on node \"crc\" DevicePath \"\"" Dec 13 06:53:10 crc kubenswrapper[4971]: I1213 06:53:10.956418 4971 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/02f6e9f2-6bad-441c-8e2e-9acfc71f3ab4-utilities\") on node \"crc\" DevicePath \"\"" Dec 13 06:53:10 crc kubenswrapper[4971]: I1213 06:53:10.956428 4971 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/cf417851-ae43-462f-a678-125593fb653b-utilities\") on node \"crc\" DevicePath \"\"" Dec 13 06:53:10 crc kubenswrapper[4971]: I1213 06:53:10.956437 4971 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/b9a0bc5d-3a39-4732-a849-22d351a4cc18-utilities\") on node \"crc\" DevicePath \"\"" Dec 13 06:53:10 crc kubenswrapper[4971]: I1213 06:53:10.956458 4971 reconciler_common.go:293] "Volume detached for volume \"marketplace-trusted-ca\" (UniqueName: \"kubernetes.io/configmap/f14c8a94-8589-4059-bddf-329452355fab-marketplace-trusted-ca\") on node \"crc\" DevicePath \"\"" Dec 13 06:53:10 crc kubenswrapper[4971]: I1213 06:53:10.956475 4971 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-55kcz\" (UniqueName: \"kubernetes.io/projected/b9a0bc5d-3a39-4732-a849-22d351a4cc18-kube-api-access-55kcz\") on node \"crc\" DevicePath \"\"" Dec 13 06:53:10 crc kubenswrapper[4971]: I1213 06:53:10.956490 4971 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-t7lfq\" (UniqueName: \"kubernetes.io/projected/cf417851-ae43-462f-a678-125593fb653b-kube-api-access-t7lfq\") on node \"crc\" DevicePath \"\"" Dec 13 06:53:10 crc kubenswrapper[4971]: I1213 06:53:10.956502 4971 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-d9kbv\" (UniqueName: \"kubernetes.io/projected/02f6e9f2-6bad-441c-8e2e-9acfc71f3ab4-kube-api-access-d9kbv\") on node \"crc\" DevicePath \"\"" Dec 13 06:53:10 crc kubenswrapper[4971]: I1213 06:53:10.961673 4971 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/f14c8a94-8589-4059-bddf-329452355fab-kube-api-access-tllvh" (OuterVolumeSpecName: "kube-api-access-tllvh") pod "f14c8a94-8589-4059-bddf-329452355fab" (UID: "f14c8a94-8589-4059-bddf-329452355fab"). InnerVolumeSpecName "kube-api-access-tllvh". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 13 06:53:10 crc kubenswrapper[4971]: I1213 06:53:10.990665 4971 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/cf417851-ae43-462f-a678-125593fb653b-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "cf417851-ae43-462f-a678-125593fb653b" (UID: "cf417851-ae43-462f-a678-125593fb653b"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 13 06:53:11 crc kubenswrapper[4971]: I1213 06:53:11.005956 4971 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/8f50f5ba-3ec3-4fa2-9f5d-51e1c43f6352-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "8f50f5ba-3ec3-4fa2-9f5d-51e1c43f6352" (UID: "8f50f5ba-3ec3-4fa2-9f5d-51e1c43f6352"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 13 06:53:11 crc kubenswrapper[4971]: I1213 06:53:11.049211 4971 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/02f6e9f2-6bad-441c-8e2e-9acfc71f3ab4-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "02f6e9f2-6bad-441c-8e2e-9acfc71f3ab4" (UID: "02f6e9f2-6bad-441c-8e2e-9acfc71f3ab4"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 13 06:53:11 crc kubenswrapper[4971]: I1213 06:53:11.057046 4971 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-25p8j\" (UniqueName: \"kubernetes.io/projected/22037bdc-fe6d-4411-be2b-aec38524a8dd-kube-api-access-25p8j\") pod \"22037bdc-fe6d-4411-be2b-aec38524a8dd\" (UID: \"22037bdc-fe6d-4411-be2b-aec38524a8dd\") " Dec 13 06:53:11 crc kubenswrapper[4971]: I1213 06:53:11.057104 4971 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/22037bdc-fe6d-4411-be2b-aec38524a8dd-catalog-content\") pod \"22037bdc-fe6d-4411-be2b-aec38524a8dd\" (UID: \"22037bdc-fe6d-4411-be2b-aec38524a8dd\") " Dec 13 06:53:11 crc kubenswrapper[4971]: I1213 06:53:11.057168 4971 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/22037bdc-fe6d-4411-be2b-aec38524a8dd-utilities\") pod \"22037bdc-fe6d-4411-be2b-aec38524a8dd\" (UID: \"22037bdc-fe6d-4411-be2b-aec38524a8dd\") " Dec 13 06:53:11 crc kubenswrapper[4971]: I1213 06:53:11.057189 4971 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/9d4f1ea5-8472-4320-820b-8090ff6f5ca6-utilities\") pod \"9d4f1ea5-8472-4320-820b-8090ff6f5ca6\" (UID: \"9d4f1ea5-8472-4320-820b-8090ff6f5ca6\") " Dec 13 06:53:11 crc kubenswrapper[4971]: I1213 06:53:11.057226 4971 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-npf8d\" (UniqueName: \"kubernetes.io/projected/9d4f1ea5-8472-4320-820b-8090ff6f5ca6-kube-api-access-npf8d\") pod \"9d4f1ea5-8472-4320-820b-8090ff6f5ca6\" (UID: \"9d4f1ea5-8472-4320-820b-8090ff6f5ca6\") " Dec 13 06:53:11 crc kubenswrapper[4971]: I1213 06:53:11.057265 4971 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/9d4f1ea5-8472-4320-820b-8090ff6f5ca6-catalog-content\") pod \"9d4f1ea5-8472-4320-820b-8090ff6f5ca6\" (UID: \"9d4f1ea5-8472-4320-820b-8090ff6f5ca6\") " Dec 13 06:53:11 crc kubenswrapper[4971]: I1213 06:53:11.057481 4971 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/cf417851-ae43-462f-a678-125593fb653b-catalog-content\") on node \"crc\" DevicePath \"\"" Dec 13 06:53:11 crc kubenswrapper[4971]: I1213 06:53:11.057492 4971 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/8f50f5ba-3ec3-4fa2-9f5d-51e1c43f6352-catalog-content\") on node \"crc\" DevicePath \"\"" Dec 13 06:53:11 crc kubenswrapper[4971]: I1213 06:53:11.057503 4971 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/02f6e9f2-6bad-441c-8e2e-9acfc71f3ab4-catalog-content\") on node \"crc\" DevicePath \"\"" Dec 13 06:53:11 crc kubenswrapper[4971]: I1213 06:53:11.057513 4971 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-tllvh\" (UniqueName: \"kubernetes.io/projected/f14c8a94-8589-4059-bddf-329452355fab-kube-api-access-tllvh\") on node \"crc\" DevicePath \"\"" Dec 13 06:53:11 crc kubenswrapper[4971]: I1213 06:53:11.059575 4971 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/9d4f1ea5-8472-4320-820b-8090ff6f5ca6-utilities" (OuterVolumeSpecName: "utilities") pod "9d4f1ea5-8472-4320-820b-8090ff6f5ca6" (UID: "9d4f1ea5-8472-4320-820b-8090ff6f5ca6"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 13 06:53:11 crc kubenswrapper[4971]: I1213 06:53:11.060135 4971 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/22037bdc-fe6d-4411-be2b-aec38524a8dd-utilities" (OuterVolumeSpecName: "utilities") pod "22037bdc-fe6d-4411-be2b-aec38524a8dd" (UID: "22037bdc-fe6d-4411-be2b-aec38524a8dd"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 13 06:53:11 crc kubenswrapper[4971]: I1213 06:53:11.061639 4971 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/9d4f1ea5-8472-4320-820b-8090ff6f5ca6-kube-api-access-npf8d" (OuterVolumeSpecName: "kube-api-access-npf8d") pod "9d4f1ea5-8472-4320-820b-8090ff6f5ca6" (UID: "9d4f1ea5-8472-4320-820b-8090ff6f5ca6"). InnerVolumeSpecName "kube-api-access-npf8d". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 13 06:53:11 crc kubenswrapper[4971]: I1213 06:53:11.061696 4971 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/22037bdc-fe6d-4411-be2b-aec38524a8dd-kube-api-access-25p8j" (OuterVolumeSpecName: "kube-api-access-25p8j") pod "22037bdc-fe6d-4411-be2b-aec38524a8dd" (UID: "22037bdc-fe6d-4411-be2b-aec38524a8dd"). InnerVolumeSpecName "kube-api-access-25p8j". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 13 06:53:11 crc kubenswrapper[4971]: I1213 06:53:11.078462 4971 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/9d4f1ea5-8472-4320-820b-8090ff6f5ca6-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "9d4f1ea5-8472-4320-820b-8090ff6f5ca6" (UID: "9d4f1ea5-8472-4320-820b-8090ff6f5ca6"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 13 06:53:11 crc kubenswrapper[4971]: I1213 06:53:11.084008 4971 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/b9a0bc5d-3a39-4732-a849-22d351a4cc18-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "b9a0bc5d-3a39-4732-a849-22d351a4cc18" (UID: "b9a0bc5d-3a39-4732-a849-22d351a4cc18"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 13 06:53:11 crc kubenswrapper[4971]: I1213 06:53:11.085517 4971 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/22037bdc-fe6d-4411-be2b-aec38524a8dd-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "22037bdc-fe6d-4411-be2b-aec38524a8dd" (UID: "22037bdc-fe6d-4411-be2b-aec38524a8dd"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 13 06:53:11 crc kubenswrapper[4971]: I1213 06:53:11.158330 4971 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/22037bdc-fe6d-4411-be2b-aec38524a8dd-utilities\") on node \"crc\" DevicePath \"\"" Dec 13 06:53:11 crc kubenswrapper[4971]: I1213 06:53:11.158370 4971 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/9d4f1ea5-8472-4320-820b-8090ff6f5ca6-utilities\") on node \"crc\" DevicePath \"\"" Dec 13 06:53:11 crc kubenswrapper[4971]: I1213 06:53:11.158384 4971 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-npf8d\" (UniqueName: \"kubernetes.io/projected/9d4f1ea5-8472-4320-820b-8090ff6f5ca6-kube-api-access-npf8d\") on node \"crc\" DevicePath \"\"" Dec 13 06:53:11 crc kubenswrapper[4971]: I1213 06:53:11.158398 4971 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/9d4f1ea5-8472-4320-820b-8090ff6f5ca6-catalog-content\") on node \"crc\" DevicePath \"\"" Dec 13 06:53:11 crc kubenswrapper[4971]: I1213 06:53:11.158410 4971 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/b9a0bc5d-3a39-4732-a849-22d351a4cc18-catalog-content\") on node \"crc\" DevicePath \"\"" Dec 13 06:53:11 crc kubenswrapper[4971]: I1213 06:53:11.158422 4971 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-25p8j\" (UniqueName: \"kubernetes.io/projected/22037bdc-fe6d-4411-be2b-aec38524a8dd-kube-api-access-25p8j\") on node \"crc\" DevicePath \"\"" Dec 13 06:53:11 crc kubenswrapper[4971]: I1213 06:53:11.158434 4971 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/22037bdc-fe6d-4411-be2b-aec38524a8dd-catalog-content\") on node \"crc\" DevicePath \"\"" Dec 13 06:53:11 crc kubenswrapper[4971]: I1213 06:53:11.399464 4971 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-mkqps" event={"ID":"9d4f1ea5-8472-4320-820b-8090ff6f5ca6","Type":"ContainerDied","Data":"26bef07a986f75c2981599189dc717ed62ed1b5ab6f7f1b51c940aa244522c9e"} Dec 13 06:53:11 crc kubenswrapper[4971]: I1213 06:53:11.399547 4971 scope.go:117] "RemoveContainer" containerID="b98959ce12c11f9c6b25aaca858920c32124fb7a92b0ddee184e3e434e02168a" Dec 13 06:53:11 crc kubenswrapper[4971]: I1213 06:53:11.399672 4971 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-mkqps" Dec 13 06:53:11 crc kubenswrapper[4971]: I1213 06:53:11.401679 4971 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-997kc" event={"ID":"cf417851-ae43-462f-a678-125593fb653b","Type":"ContainerDied","Data":"258f4a913f8ba4a34d86e83ac34969fb30afa3dfb66b8f141e0c6bf3b63eaab0"} Dec 13 06:53:11 crc kubenswrapper[4971]: I1213 06:53:11.401718 4971 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-997kc" Dec 13 06:53:11 crc kubenswrapper[4971]: I1213 06:53:11.402107 4971 status_manager.go:851] "Failed to get status for pod" podUID="f14c8a94-8589-4059-bddf-329452355fab" pod="openshift-marketplace/marketplace-operator-79b997595-j9zqf" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/marketplace-operator-79b997595-j9zqf\": dial tcp 38.102.83.195:6443: connect: connection refused" Dec 13 06:53:11 crc kubenswrapper[4971]: I1213 06:53:11.402715 4971 status_manager.go:851] "Failed to get status for pod" podUID="b9a0bc5d-3a39-4732-a849-22d351a4cc18" pod="openshift-marketplace/redhat-operators-d8n8p" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/redhat-operators-d8n8p\": dial tcp 38.102.83.195:6443: connect: connection refused" Dec 13 06:53:11 crc kubenswrapper[4971]: I1213 06:53:11.403011 4971 status_manager.go:851] "Failed to get status for pod" podUID="f85e55b1a89d02b0cb034b1ea31ed45a" pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-startup-monitor-crc\": dial tcp 38.102.83.195:6443: connect: connection refused" Dec 13 06:53:11 crc kubenswrapper[4971]: I1213 06:53:11.403300 4971 status_manager.go:851] "Failed to get status for pod" podUID="cf417851-ae43-462f-a678-125593fb653b" pod="openshift-marketplace/community-operators-997kc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/community-operators-997kc\": dial tcp 38.102.83.195:6443: connect: connection refused" Dec 13 06:53:11 crc kubenswrapper[4971]: I1213 06:53:11.403675 4971 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_marketplace-operator-79b997595-6rvc9_5c88a864-89a9-4f5e-b713-385170b9bf17/marketplace-operator/0.log" Dec 13 06:53:11 crc kubenswrapper[4971]: I1213 06:53:11.403721 4971 generic.go:334] "Generic (PLEG): container finished" podID="5c88a864-89a9-4f5e-b713-385170b9bf17" containerID="7a53386540871a14f654bd05dae1a7988d7a3396a65ccf45bb2e98846899cc87" exitCode=1 Dec 13 06:53:11 crc kubenswrapper[4971]: I1213 06:53:11.403805 4971 status_manager.go:851] "Failed to get status for pod" podUID="8f50f5ba-3ec3-4fa2-9f5d-51e1c43f6352" pod="openshift-marketplace/certified-operators-rkd9c" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/certified-operators-rkd9c\": dial tcp 38.102.83.195:6443: connect: connection refused" Dec 13 06:53:11 crc kubenswrapper[4971]: I1213 06:53:11.404093 4971 status_manager.go:851] "Failed to get status for pod" podUID="22037bdc-fe6d-4411-be2b-aec38524a8dd" pod="openshift-marketplace/redhat-marketplace-xwxlf" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/redhat-marketplace-xwxlf\": dial tcp 38.102.83.195:6443: connect: connection refused" Dec 13 06:53:11 crc kubenswrapper[4971]: I1213 06:53:11.404182 4971 scope.go:117] "RemoveContainer" containerID="7a53386540871a14f654bd05dae1a7988d7a3396a65ccf45bb2e98846899cc87" Dec 13 06:53:11 crc kubenswrapper[4971]: I1213 06:53:11.404602 4971 status_manager.go:851] "Failed to get status for pod" podUID="02f6e9f2-6bad-441c-8e2e-9acfc71f3ab4" pod="openshift-marketplace/redhat-operators-wpxfk" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/redhat-operators-wpxfk\": dial tcp 38.102.83.195:6443: connect: connection refused" Dec 13 06:53:11 crc kubenswrapper[4971]: I1213 06:53:11.404192 4971 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/marketplace-operator-79b997595-6rvc9" event={"ID":"5c88a864-89a9-4f5e-b713-385170b9bf17","Type":"ContainerDied","Data":"7a53386540871a14f654bd05dae1a7988d7a3396a65ccf45bb2e98846899cc87"} Dec 13 06:53:11 crc kubenswrapper[4971]: I1213 06:53:11.405695 4971 status_manager.go:851] "Failed to get status for pod" podUID="5c88a864-89a9-4f5e-b713-385170b9bf17" pod="openshift-marketplace/marketplace-operator-79b997595-6rvc9" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/marketplace-operator-79b997595-6rvc9\": dial tcp 38.102.83.195:6443: connect: connection refused" Dec 13 06:53:11 crc kubenswrapper[4971]: I1213 06:53:11.406193 4971 status_manager.go:851] "Failed to get status for pod" podUID="9d4f1ea5-8472-4320-820b-8090ff6f5ca6" pod="openshift-marketplace/redhat-marketplace-mkqps" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/redhat-marketplace-mkqps\": dial tcp 38.102.83.195:6443: connect: connection refused" Dec 13 06:53:11 crc kubenswrapper[4971]: I1213 06:53:11.406311 4971 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/marketplace-operator-79b997595-j9zqf" Dec 13 06:53:11 crc kubenswrapper[4971]: I1213 06:53:11.406215 4971 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/marketplace-operator-79b997595-j9zqf" event={"ID":"f14c8a94-8589-4059-bddf-329452355fab","Type":"ContainerDied","Data":"59144fe25fa154135eea30f6ed1f54f23ce9a403973267f4c8d8e235c632965a"} Dec 13 06:53:11 crc kubenswrapper[4971]: I1213 06:53:11.406635 4971 status_manager.go:851] "Failed to get status for pod" podUID="53fdcbc2-900b-48eb-8be3-cd72940096c3" pod="openshift-kube-apiserver/installer-9-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/installer-9-crc\": dial tcp 38.102.83.195:6443: connect: connection refused" Dec 13 06:53:11 crc kubenswrapper[4971]: I1213 06:53:11.408048 4971 status_manager.go:851] "Failed to get status for pod" podUID="cf417851-ae43-462f-a678-125593fb653b" pod="openshift-marketplace/community-operators-997kc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/community-operators-997kc\": dial tcp 38.102.83.195:6443: connect: connection refused" Dec 13 06:53:11 crc kubenswrapper[4971]: I1213 06:53:11.408443 4971 status_manager.go:851] "Failed to get status for pod" podUID="02f6e9f2-6bad-441c-8e2e-9acfc71f3ab4" pod="openshift-marketplace/redhat-operators-wpxfk" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/redhat-operators-wpxfk\": dial tcp 38.102.83.195:6443: connect: connection refused" Dec 13 06:53:11 crc kubenswrapper[4971]: I1213 06:53:11.408732 4971 status_manager.go:851] "Failed to get status for pod" podUID="5c88a864-89a9-4f5e-b713-385170b9bf17" pod="openshift-marketplace/marketplace-operator-79b997595-6rvc9" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/marketplace-operator-79b997595-6rvc9\": dial tcp 38.102.83.195:6443: connect: connection refused" Dec 13 06:53:11 crc kubenswrapper[4971]: I1213 06:53:11.408951 4971 status_manager.go:851] "Failed to get status for pod" podUID="8f50f5ba-3ec3-4fa2-9f5d-51e1c43f6352" pod="openshift-marketplace/certified-operators-rkd9c" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/certified-operators-rkd9c\": dial tcp 38.102.83.195:6443: connect: connection refused" Dec 13 06:53:11 crc kubenswrapper[4971]: I1213 06:53:11.409125 4971 status_manager.go:851] "Failed to get status for pod" podUID="22037bdc-fe6d-4411-be2b-aec38524a8dd" pod="openshift-marketplace/redhat-marketplace-xwxlf" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/redhat-marketplace-xwxlf\": dial tcp 38.102.83.195:6443: connect: connection refused" Dec 13 06:53:11 crc kubenswrapper[4971]: I1213 06:53:11.409894 4971 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-rkd9c" event={"ID":"8f50f5ba-3ec3-4fa2-9f5d-51e1c43f6352","Type":"ContainerDied","Data":"d46dca0b4e85de9494ecaf1e3a49c4a9d7eb1b1e29a3f6dd19588adef058cc3e"} Dec 13 06:53:11 crc kubenswrapper[4971]: I1213 06:53:11.409912 4971 status_manager.go:851] "Failed to get status for pod" podUID="9d4f1ea5-8472-4320-820b-8090ff6f5ca6" pod="openshift-marketplace/redhat-marketplace-mkqps" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/redhat-marketplace-mkqps\": dial tcp 38.102.83.195:6443: connect: connection refused" Dec 13 06:53:11 crc kubenswrapper[4971]: I1213 06:53:11.409951 4971 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-rkd9c" Dec 13 06:53:11 crc kubenswrapper[4971]: I1213 06:53:11.410670 4971 status_manager.go:851] "Failed to get status for pod" podUID="53fdcbc2-900b-48eb-8be3-cd72940096c3" pod="openshift-kube-apiserver/installer-9-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/installer-9-crc\": dial tcp 38.102.83.195:6443: connect: connection refused" Dec 13 06:53:11 crc kubenswrapper[4971]: I1213 06:53:11.410867 4971 status_manager.go:851] "Failed to get status for pod" podUID="b9a0bc5d-3a39-4732-a849-22d351a4cc18" pod="openshift-marketplace/redhat-operators-d8n8p" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/redhat-operators-d8n8p\": dial tcp 38.102.83.195:6443: connect: connection refused" Dec 13 06:53:11 crc kubenswrapper[4971]: I1213 06:53:11.411047 4971 status_manager.go:851] "Failed to get status for pod" podUID="f85e55b1a89d02b0cb034b1ea31ed45a" pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-startup-monitor-crc\": dial tcp 38.102.83.195:6443: connect: connection refused" Dec 13 06:53:11 crc kubenswrapper[4971]: I1213 06:53:11.411368 4971 status_manager.go:851] "Failed to get status for pod" podUID="f14c8a94-8589-4059-bddf-329452355fab" pod="openshift-marketplace/marketplace-operator-79b997595-j9zqf" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/marketplace-operator-79b997595-j9zqf\": dial tcp 38.102.83.195:6443: connect: connection refused" Dec 13 06:53:11 crc kubenswrapper[4971]: I1213 06:53:11.411831 4971 status_manager.go:851] "Failed to get status for pod" podUID="8f50f5ba-3ec3-4fa2-9f5d-51e1c43f6352" pod="openshift-marketplace/certified-operators-rkd9c" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/certified-operators-rkd9c\": dial tcp 38.102.83.195:6443: connect: connection refused" Dec 13 06:53:11 crc kubenswrapper[4971]: I1213 06:53:11.412234 4971 status_manager.go:851] "Failed to get status for pod" podUID="22037bdc-fe6d-4411-be2b-aec38524a8dd" pod="openshift-marketplace/redhat-marketplace-xwxlf" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/redhat-marketplace-xwxlf\": dial tcp 38.102.83.195:6443: connect: connection refused" Dec 13 06:53:11 crc kubenswrapper[4971]: I1213 06:53:11.412807 4971 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-xwxlf" event={"ID":"22037bdc-fe6d-4411-be2b-aec38524a8dd","Type":"ContainerDied","Data":"5c5265d52117bbfcb392ee1a7a528d5ed5ff1be0ce3712238a7ea4e6263303a3"} Dec 13 06:53:11 crc kubenswrapper[4971]: I1213 06:53:11.412843 4971 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-xwxlf" Dec 13 06:53:11 crc kubenswrapper[4971]: I1213 06:53:11.412840 4971 status_manager.go:851] "Failed to get status for pod" podUID="02f6e9f2-6bad-441c-8e2e-9acfc71f3ab4" pod="openshift-marketplace/redhat-operators-wpxfk" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/redhat-operators-wpxfk\": dial tcp 38.102.83.195:6443: connect: connection refused" Dec 13 06:53:11 crc kubenswrapper[4971]: I1213 06:53:11.413697 4971 status_manager.go:851] "Failed to get status for pod" podUID="5c88a864-89a9-4f5e-b713-385170b9bf17" pod="openshift-marketplace/marketplace-operator-79b997595-6rvc9" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/marketplace-operator-79b997595-6rvc9\": dial tcp 38.102.83.195:6443: connect: connection refused" Dec 13 06:53:11 crc kubenswrapper[4971]: I1213 06:53:11.414010 4971 status_manager.go:851] "Failed to get status for pod" podUID="9d4f1ea5-8472-4320-820b-8090ff6f5ca6" pod="openshift-marketplace/redhat-marketplace-mkqps" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/redhat-marketplace-mkqps\": dial tcp 38.102.83.195:6443: connect: connection refused" Dec 13 06:53:11 crc kubenswrapper[4971]: I1213 06:53:11.414485 4971 status_manager.go:851] "Failed to get status for pod" podUID="53fdcbc2-900b-48eb-8be3-cd72940096c3" pod="openshift-kube-apiserver/installer-9-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/installer-9-crc\": dial tcp 38.102.83.195:6443: connect: connection refused" Dec 13 06:53:11 crc kubenswrapper[4971]: I1213 06:53:11.414506 4971 scope.go:117] "RemoveContainer" containerID="8b1751665ebf5281772672dfa3219f016e81b5f54d3c9a85761125ad6f32cf0b" Dec 13 06:53:11 crc kubenswrapper[4971]: I1213 06:53:11.415197 4971 status_manager.go:851] "Failed to get status for pod" podUID="f14c8a94-8589-4059-bddf-329452355fab" pod="openshift-marketplace/marketplace-operator-79b997595-j9zqf" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/marketplace-operator-79b997595-j9zqf\": dial tcp 38.102.83.195:6443: connect: connection refused" Dec 13 06:53:11 crc kubenswrapper[4971]: I1213 06:53:11.415674 4971 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-wpxfk" event={"ID":"02f6e9f2-6bad-441c-8e2e-9acfc71f3ab4","Type":"ContainerDied","Data":"63c622776bc727571794ee6aabc7d9588866c6f1e1a30002c0e83865eaaa3bac"} Dec 13 06:53:11 crc kubenswrapper[4971]: I1213 06:53:11.415748 4971 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-wpxfk" Dec 13 06:53:11 crc kubenswrapper[4971]: I1213 06:53:11.415947 4971 status_manager.go:851] "Failed to get status for pod" podUID="b9a0bc5d-3a39-4732-a849-22d351a4cc18" pod="openshift-marketplace/redhat-operators-d8n8p" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/redhat-operators-d8n8p\": dial tcp 38.102.83.195:6443: connect: connection refused" Dec 13 06:53:11 crc kubenswrapper[4971]: I1213 06:53:11.416309 4971 status_manager.go:851] "Failed to get status for pod" podUID="f85e55b1a89d02b0cb034b1ea31ed45a" pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-startup-monitor-crc\": dial tcp 38.102.83.195:6443: connect: connection refused" Dec 13 06:53:11 crc kubenswrapper[4971]: I1213 06:53:11.416654 4971 status_manager.go:851] "Failed to get status for pod" podUID="cf417851-ae43-462f-a678-125593fb653b" pod="openshift-marketplace/community-operators-997kc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/community-operators-997kc\": dial tcp 38.102.83.195:6443: connect: connection refused" Dec 13 06:53:11 crc kubenswrapper[4971]: I1213 06:53:11.417058 4971 status_manager.go:851] "Failed to get status for pod" podUID="8f50f5ba-3ec3-4fa2-9f5d-51e1c43f6352" pod="openshift-marketplace/certified-operators-rkd9c" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/certified-operators-rkd9c\": dial tcp 38.102.83.195:6443: connect: connection refused" Dec 13 06:53:11 crc kubenswrapper[4971]: I1213 06:53:11.417365 4971 status_manager.go:851] "Failed to get status for pod" podUID="22037bdc-fe6d-4411-be2b-aec38524a8dd" pod="openshift-marketplace/redhat-marketplace-xwxlf" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/redhat-marketplace-xwxlf\": dial tcp 38.102.83.195:6443: connect: connection refused" Dec 13 06:53:11 crc kubenswrapper[4971]: I1213 06:53:11.417650 4971 status_manager.go:851] "Failed to get status for pod" podUID="02f6e9f2-6bad-441c-8e2e-9acfc71f3ab4" pod="openshift-marketplace/redhat-operators-wpxfk" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/redhat-operators-wpxfk\": dial tcp 38.102.83.195:6443: connect: connection refused" Dec 13 06:53:11 crc kubenswrapper[4971]: I1213 06:53:11.417986 4971 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-d8n8p" event={"ID":"b9a0bc5d-3a39-4732-a849-22d351a4cc18","Type":"ContainerDied","Data":"a4a3ad87df8940b34556eacf79e1d8f7f5e82ac7fba0af8aaea503b398caa4cb"} Dec 13 06:53:11 crc kubenswrapper[4971]: I1213 06:53:11.418010 4971 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-d8n8p" Dec 13 06:53:11 crc kubenswrapper[4971]: I1213 06:53:11.418086 4971 status_manager.go:851] "Failed to get status for pod" podUID="5c88a864-89a9-4f5e-b713-385170b9bf17" pod="openshift-marketplace/marketplace-operator-79b997595-6rvc9" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/marketplace-operator-79b997595-6rvc9\": dial tcp 38.102.83.195:6443: connect: connection refused" Dec 13 06:53:11 crc kubenswrapper[4971]: I1213 06:53:11.418508 4971 status_manager.go:851] "Failed to get status for pod" podUID="9d4f1ea5-8472-4320-820b-8090ff6f5ca6" pod="openshift-marketplace/redhat-marketplace-mkqps" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/redhat-marketplace-mkqps\": dial tcp 38.102.83.195:6443: connect: connection refused" Dec 13 06:53:11 crc kubenswrapper[4971]: I1213 06:53:11.418834 4971 status_manager.go:851] "Failed to get status for pod" podUID="53fdcbc2-900b-48eb-8be3-cd72940096c3" pod="openshift-kube-apiserver/installer-9-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/installer-9-crc\": dial tcp 38.102.83.195:6443: connect: connection refused" Dec 13 06:53:11 crc kubenswrapper[4971]: I1213 06:53:11.419098 4971 status_manager.go:851] "Failed to get status for pod" podUID="f14c8a94-8589-4059-bddf-329452355fab" pod="openshift-marketplace/marketplace-operator-79b997595-j9zqf" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/marketplace-operator-79b997595-j9zqf\": dial tcp 38.102.83.195:6443: connect: connection refused" Dec 13 06:53:11 crc kubenswrapper[4971]: I1213 06:53:11.419347 4971 status_manager.go:851] "Failed to get status for pod" podUID="b9a0bc5d-3a39-4732-a849-22d351a4cc18" pod="openshift-marketplace/redhat-operators-d8n8p" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/redhat-operators-d8n8p\": dial tcp 38.102.83.195:6443: connect: connection refused" Dec 13 06:53:11 crc kubenswrapper[4971]: I1213 06:53:11.419608 4971 status_manager.go:851] "Failed to get status for pod" podUID="f85e55b1a89d02b0cb034b1ea31ed45a" pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-startup-monitor-crc\": dial tcp 38.102.83.195:6443: connect: connection refused" Dec 13 06:53:11 crc kubenswrapper[4971]: I1213 06:53:11.419903 4971 status_manager.go:851] "Failed to get status for pod" podUID="cf417851-ae43-462f-a678-125593fb653b" pod="openshift-marketplace/community-operators-997kc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/community-operators-997kc\": dial tcp 38.102.83.195:6443: connect: connection refused" Dec 13 06:53:11 crc kubenswrapper[4971]: I1213 06:53:11.420117 4971 status_manager.go:851] "Failed to get status for pod" podUID="9d4f1ea5-8472-4320-820b-8090ff6f5ca6" pod="openshift-marketplace/redhat-marketplace-mkqps" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/redhat-marketplace-mkqps\": dial tcp 38.102.83.195:6443: connect: connection refused" Dec 13 06:53:11 crc kubenswrapper[4971]: I1213 06:53:11.420312 4971 status_manager.go:851] "Failed to get status for pod" podUID="53fdcbc2-900b-48eb-8be3-cd72940096c3" pod="openshift-kube-apiserver/installer-9-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/installer-9-crc\": dial tcp 38.102.83.195:6443: connect: connection refused" Dec 13 06:53:11 crc kubenswrapper[4971]: I1213 06:53:11.420558 4971 status_manager.go:851] "Failed to get status for pod" podUID="b9a0bc5d-3a39-4732-a849-22d351a4cc18" pod="openshift-marketplace/redhat-operators-d8n8p" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/redhat-operators-d8n8p\": dial tcp 38.102.83.195:6443: connect: connection refused" Dec 13 06:53:11 crc kubenswrapper[4971]: I1213 06:53:11.420792 4971 status_manager.go:851] "Failed to get status for pod" podUID="f85e55b1a89d02b0cb034b1ea31ed45a" pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-startup-monitor-crc\": dial tcp 38.102.83.195:6443: connect: connection refused" Dec 13 06:53:11 crc kubenswrapper[4971]: I1213 06:53:11.421007 4971 status_manager.go:851] "Failed to get status for pod" podUID="f14c8a94-8589-4059-bddf-329452355fab" pod="openshift-marketplace/marketplace-operator-79b997595-j9zqf" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/marketplace-operator-79b997595-j9zqf\": dial tcp 38.102.83.195:6443: connect: connection refused" Dec 13 06:53:11 crc kubenswrapper[4971]: I1213 06:53:11.421015 4971 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-kube-apiserver_kube-apiserver-crc_f4b27818a5e8e43d0dc095d08835c792/kube-apiserver-cert-syncer/0.log" Dec 13 06:53:11 crc kubenswrapper[4971]: I1213 06:53:11.421239 4971 status_manager.go:851] "Failed to get status for pod" podUID="cf417851-ae43-462f-a678-125593fb653b" pod="openshift-marketplace/community-operators-997kc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/community-operators-997kc\": dial tcp 38.102.83.195:6443: connect: connection refused" Dec 13 06:53:11 crc kubenswrapper[4971]: I1213 06:53:11.421488 4971 status_manager.go:851] "Failed to get status for pod" podUID="5c88a864-89a9-4f5e-b713-385170b9bf17" pod="openshift-marketplace/marketplace-operator-79b997595-6rvc9" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/marketplace-operator-79b997595-6rvc9\": dial tcp 38.102.83.195:6443: connect: connection refused" Dec 13 06:53:11 crc kubenswrapper[4971]: I1213 06:53:11.421741 4971 status_manager.go:851] "Failed to get status for pod" podUID="8f50f5ba-3ec3-4fa2-9f5d-51e1c43f6352" pod="openshift-marketplace/certified-operators-rkd9c" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/certified-operators-rkd9c\": dial tcp 38.102.83.195:6443: connect: connection refused" Dec 13 06:53:11 crc kubenswrapper[4971]: I1213 06:53:11.421949 4971 status_manager.go:851] "Failed to get status for pod" podUID="22037bdc-fe6d-4411-be2b-aec38524a8dd" pod="openshift-marketplace/redhat-marketplace-xwxlf" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/redhat-marketplace-xwxlf\": dial tcp 38.102.83.195:6443: connect: connection refused" Dec 13 06:53:11 crc kubenswrapper[4971]: I1213 06:53:11.422292 4971 status_manager.go:851] "Failed to get status for pod" podUID="02f6e9f2-6bad-441c-8e2e-9acfc71f3ab4" pod="openshift-marketplace/redhat-operators-wpxfk" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/redhat-operators-wpxfk\": dial tcp 38.102.83.195:6443: connect: connection refused" Dec 13 06:53:11 crc kubenswrapper[4971]: I1213 06:53:11.435356 4971 scope.go:117] "RemoveContainer" containerID="4c8765305a845952aa8d2a0c9bb6e24513864afdd01cf9e5bbed8dd7ea15cf36" Dec 13 06:53:11 crc kubenswrapper[4971]: I1213 06:53:11.463874 4971 status_manager.go:851] "Failed to get status for pod" podUID="cf417851-ae43-462f-a678-125593fb653b" pod="openshift-marketplace/community-operators-997kc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/community-operators-997kc\": dial tcp 38.102.83.195:6443: connect: connection refused" Dec 13 06:53:11 crc kubenswrapper[4971]: I1213 06:53:11.464621 4971 status_manager.go:851] "Failed to get status for pod" podUID="8f50f5ba-3ec3-4fa2-9f5d-51e1c43f6352" pod="openshift-marketplace/certified-operators-rkd9c" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/certified-operators-rkd9c\": dial tcp 38.102.83.195:6443: connect: connection refused" Dec 13 06:53:11 crc kubenswrapper[4971]: I1213 06:53:11.464836 4971 status_manager.go:851] "Failed to get status for pod" podUID="22037bdc-fe6d-4411-be2b-aec38524a8dd" pod="openshift-marketplace/redhat-marketplace-xwxlf" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/redhat-marketplace-xwxlf\": dial tcp 38.102.83.195:6443: connect: connection refused" Dec 13 06:53:11 crc kubenswrapper[4971]: I1213 06:53:11.465019 4971 status_manager.go:851] "Failed to get status for pod" podUID="02f6e9f2-6bad-441c-8e2e-9acfc71f3ab4" pod="openshift-marketplace/redhat-operators-wpxfk" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/redhat-operators-wpxfk\": dial tcp 38.102.83.195:6443: connect: connection refused" Dec 13 06:53:11 crc kubenswrapper[4971]: I1213 06:53:11.465216 4971 status_manager.go:851] "Failed to get status for pod" podUID="5c88a864-89a9-4f5e-b713-385170b9bf17" pod="openshift-marketplace/marketplace-operator-79b997595-6rvc9" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/marketplace-operator-79b997595-6rvc9\": dial tcp 38.102.83.195:6443: connect: connection refused" Dec 13 06:53:11 crc kubenswrapper[4971]: I1213 06:53:11.465446 4971 status_manager.go:851] "Failed to get status for pod" podUID="9d4f1ea5-8472-4320-820b-8090ff6f5ca6" pod="openshift-marketplace/redhat-marketplace-mkqps" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/redhat-marketplace-mkqps\": dial tcp 38.102.83.195:6443: connect: connection refused" Dec 13 06:53:11 crc kubenswrapper[4971]: I1213 06:53:11.465766 4971 status_manager.go:851] "Failed to get status for pod" podUID="53fdcbc2-900b-48eb-8be3-cd72940096c3" pod="openshift-kube-apiserver/installer-9-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/installer-9-crc\": dial tcp 38.102.83.195:6443: connect: connection refused" Dec 13 06:53:11 crc kubenswrapper[4971]: I1213 06:53:11.465979 4971 status_manager.go:851] "Failed to get status for pod" podUID="b9a0bc5d-3a39-4732-a849-22d351a4cc18" pod="openshift-marketplace/redhat-operators-d8n8p" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/redhat-operators-d8n8p\": dial tcp 38.102.83.195:6443: connect: connection refused" Dec 13 06:53:11 crc kubenswrapper[4971]: I1213 06:53:11.466332 4971 status_manager.go:851] "Failed to get status for pod" podUID="f85e55b1a89d02b0cb034b1ea31ed45a" pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-startup-monitor-crc\": dial tcp 38.102.83.195:6443: connect: connection refused" Dec 13 06:53:11 crc kubenswrapper[4971]: I1213 06:53:11.466672 4971 status_manager.go:851] "Failed to get status for pod" podUID="f14c8a94-8589-4059-bddf-329452355fab" pod="openshift-marketplace/marketplace-operator-79b997595-j9zqf" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/marketplace-operator-79b997595-j9zqf\": dial tcp 38.102.83.195:6443: connect: connection refused" Dec 13 06:53:11 crc kubenswrapper[4971]: I1213 06:53:11.470547 4971 scope.go:117] "RemoveContainer" containerID="6e9990367538c8d28196c610c227d5744eafd1a07907c88fcc963201ce0f06cc" Dec 13 06:53:11 crc kubenswrapper[4971]: I1213 06:53:11.529792 4971 scope.go:117] "RemoveContainer" containerID="69a96a08af13f360b70743d81bdc43ccbf3371412d42bbd6979545432c6afe67" Dec 13 06:53:11 crc kubenswrapper[4971]: I1213 06:53:11.557209 4971 status_manager.go:851] "Failed to get status for pod" podUID="b9a0bc5d-3a39-4732-a849-22d351a4cc18" pod="openshift-marketplace/redhat-operators-d8n8p" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/redhat-operators-d8n8p\": dial tcp 38.102.83.195:6443: connect: connection refused" Dec 13 06:53:11 crc kubenswrapper[4971]: I1213 06:53:11.557648 4971 status_manager.go:851] "Failed to get status for pod" podUID="f85e55b1a89d02b0cb034b1ea31ed45a" pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-startup-monitor-crc\": dial tcp 38.102.83.195:6443: connect: connection refused" Dec 13 06:53:11 crc kubenswrapper[4971]: I1213 06:53:11.558152 4971 status_manager.go:851] "Failed to get status for pod" podUID="f14c8a94-8589-4059-bddf-329452355fab" pod="openshift-marketplace/marketplace-operator-79b997595-j9zqf" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/marketplace-operator-79b997595-j9zqf\": dial tcp 38.102.83.195:6443: connect: connection refused" Dec 13 06:53:11 crc kubenswrapper[4971]: I1213 06:53:11.558350 4971 status_manager.go:851] "Failed to get status for pod" podUID="cf417851-ae43-462f-a678-125593fb653b" pod="openshift-marketplace/community-operators-997kc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/community-operators-997kc\": dial tcp 38.102.83.195:6443: connect: connection refused" Dec 13 06:53:11 crc kubenswrapper[4971]: I1213 06:53:11.558568 4971 status_manager.go:851] "Failed to get status for pod" podUID="5c88a864-89a9-4f5e-b713-385170b9bf17" pod="openshift-marketplace/marketplace-operator-79b997595-6rvc9" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/marketplace-operator-79b997595-6rvc9\": dial tcp 38.102.83.195:6443: connect: connection refused" Dec 13 06:53:11 crc kubenswrapper[4971]: I1213 06:53:11.558769 4971 status_manager.go:851] "Failed to get status for pod" podUID="8f50f5ba-3ec3-4fa2-9f5d-51e1c43f6352" pod="openshift-marketplace/certified-operators-rkd9c" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/certified-operators-rkd9c\": dial tcp 38.102.83.195:6443: connect: connection refused" Dec 13 06:53:11 crc kubenswrapper[4971]: I1213 06:53:11.558976 4971 status_manager.go:851] "Failed to get status for pod" podUID="22037bdc-fe6d-4411-be2b-aec38524a8dd" pod="openshift-marketplace/redhat-marketplace-xwxlf" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/redhat-marketplace-xwxlf\": dial tcp 38.102.83.195:6443: connect: connection refused" Dec 13 06:53:11 crc kubenswrapper[4971]: I1213 06:53:11.559189 4971 status_manager.go:851] "Failed to get status for pod" podUID="02f6e9f2-6bad-441c-8e2e-9acfc71f3ab4" pod="openshift-marketplace/redhat-operators-wpxfk" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/redhat-operators-wpxfk\": dial tcp 38.102.83.195:6443: connect: connection refused" Dec 13 06:53:11 crc kubenswrapper[4971]: I1213 06:53:11.559397 4971 status_manager.go:851] "Failed to get status for pod" podUID="9d4f1ea5-8472-4320-820b-8090ff6f5ca6" pod="openshift-marketplace/redhat-marketplace-mkqps" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/redhat-marketplace-mkqps\": dial tcp 38.102.83.195:6443: connect: connection refused" Dec 13 06:53:11 crc kubenswrapper[4971]: I1213 06:53:11.559623 4971 status_manager.go:851] "Failed to get status for pod" podUID="53fdcbc2-900b-48eb-8be3-cd72940096c3" pod="openshift-kube-apiserver/installer-9-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/installer-9-crc\": dial tcp 38.102.83.195:6443: connect: connection refused" Dec 13 06:53:11 crc kubenswrapper[4971]: I1213 06:53:11.559863 4971 status_manager.go:851] "Failed to get status for pod" podUID="9d4f1ea5-8472-4320-820b-8090ff6f5ca6" pod="openshift-marketplace/redhat-marketplace-mkqps" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/redhat-marketplace-mkqps\": dial tcp 38.102.83.195:6443: connect: connection refused" Dec 13 06:53:11 crc kubenswrapper[4971]: I1213 06:53:11.560059 4971 status_manager.go:851] "Failed to get status for pod" podUID="53fdcbc2-900b-48eb-8be3-cd72940096c3" pod="openshift-kube-apiserver/installer-9-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/installer-9-crc\": dial tcp 38.102.83.195:6443: connect: connection refused" Dec 13 06:53:11 crc kubenswrapper[4971]: I1213 06:53:11.560251 4971 status_manager.go:851] "Failed to get status for pod" podUID="f14c8a94-8589-4059-bddf-329452355fab" pod="openshift-marketplace/marketplace-operator-79b997595-j9zqf" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/marketplace-operator-79b997595-j9zqf\": dial tcp 38.102.83.195:6443: connect: connection refused" Dec 13 06:53:11 crc kubenswrapper[4971]: I1213 06:53:11.560419 4971 status_manager.go:851] "Failed to get status for pod" podUID="b9a0bc5d-3a39-4732-a849-22d351a4cc18" pod="openshift-marketplace/redhat-operators-d8n8p" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/redhat-operators-d8n8p\": dial tcp 38.102.83.195:6443: connect: connection refused" Dec 13 06:53:11 crc kubenswrapper[4971]: I1213 06:53:11.560930 4971 status_manager.go:851] "Failed to get status for pod" podUID="f85e55b1a89d02b0cb034b1ea31ed45a" pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-startup-monitor-crc\": dial tcp 38.102.83.195:6443: connect: connection refused" Dec 13 06:53:11 crc kubenswrapper[4971]: I1213 06:53:11.561168 4971 status_manager.go:851] "Failed to get status for pod" podUID="cf417851-ae43-462f-a678-125593fb653b" pod="openshift-marketplace/community-operators-997kc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/community-operators-997kc\": dial tcp 38.102.83.195:6443: connect: connection refused" Dec 13 06:53:11 crc kubenswrapper[4971]: I1213 06:53:11.561441 4971 status_manager.go:851] "Failed to get status for pod" podUID="8f50f5ba-3ec3-4fa2-9f5d-51e1c43f6352" pod="openshift-marketplace/certified-operators-rkd9c" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/certified-operators-rkd9c\": dial tcp 38.102.83.195:6443: connect: connection refused" Dec 13 06:53:11 crc kubenswrapper[4971]: I1213 06:53:11.561872 4971 status_manager.go:851] "Failed to get status for pod" podUID="22037bdc-fe6d-4411-be2b-aec38524a8dd" pod="openshift-marketplace/redhat-marketplace-xwxlf" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/redhat-marketplace-xwxlf\": dial tcp 38.102.83.195:6443: connect: connection refused" Dec 13 06:53:11 crc kubenswrapper[4971]: I1213 06:53:11.562062 4971 status_manager.go:851] "Failed to get status for pod" podUID="02f6e9f2-6bad-441c-8e2e-9acfc71f3ab4" pod="openshift-marketplace/redhat-operators-wpxfk" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/redhat-operators-wpxfk\": dial tcp 38.102.83.195:6443: connect: connection refused" Dec 13 06:53:11 crc kubenswrapper[4971]: I1213 06:53:11.562288 4971 status_manager.go:851] "Failed to get status for pod" podUID="5c88a864-89a9-4f5e-b713-385170b9bf17" pod="openshift-marketplace/marketplace-operator-79b997595-6rvc9" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/marketplace-operator-79b997595-6rvc9\": dial tcp 38.102.83.195:6443: connect: connection refused" Dec 13 06:53:11 crc kubenswrapper[4971]: I1213 06:53:11.586985 4971 scope.go:117] "RemoveContainer" containerID="ef186a9e3ccd3c7dd0a3bc655e584757165c6f621358f6432a5768b19e98cbc8" Dec 13 06:53:11 crc kubenswrapper[4971]: I1213 06:53:11.603258 4971 scope.go:117] "RemoveContainer" containerID="524506c76ba56186a2daa60747573baef0eb42cebe8bb321e82be093150f35d7" Dec 13 06:53:11 crc kubenswrapper[4971]: I1213 06:53:11.617562 4971 scope.go:117] "RemoveContainer" containerID="cf192b8ffdcb439c2bd3f71dc37ac0dbd8e41220c791304f4d26a5436e9cc7e5" Dec 13 06:53:11 crc kubenswrapper[4971]: I1213 06:53:11.619148 4971 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/installer-9-crc" Dec 13 06:53:11 crc kubenswrapper[4971]: I1213 06:53:11.619746 4971 status_manager.go:851] "Failed to get status for pod" podUID="b9a0bc5d-3a39-4732-a849-22d351a4cc18" pod="openshift-marketplace/redhat-operators-d8n8p" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/redhat-operators-d8n8p\": dial tcp 38.102.83.195:6443: connect: connection refused" Dec 13 06:53:11 crc kubenswrapper[4971]: I1213 06:53:11.619981 4971 status_manager.go:851] "Failed to get status for pod" podUID="f85e55b1a89d02b0cb034b1ea31ed45a" pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-startup-monitor-crc\": dial tcp 38.102.83.195:6443: connect: connection refused" Dec 13 06:53:11 crc kubenswrapper[4971]: I1213 06:53:11.620288 4971 status_manager.go:851] "Failed to get status for pod" podUID="f14c8a94-8589-4059-bddf-329452355fab" pod="openshift-marketplace/marketplace-operator-79b997595-j9zqf" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/marketplace-operator-79b997595-j9zqf\": dial tcp 38.102.83.195:6443: connect: connection refused" Dec 13 06:53:11 crc kubenswrapper[4971]: I1213 06:53:11.620698 4971 status_manager.go:851] "Failed to get status for pod" podUID="cf417851-ae43-462f-a678-125593fb653b" pod="openshift-marketplace/community-operators-997kc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/community-operators-997kc\": dial tcp 38.102.83.195:6443: connect: connection refused" Dec 13 06:53:11 crc kubenswrapper[4971]: I1213 06:53:11.620913 4971 status_manager.go:851] "Failed to get status for pod" podUID="22037bdc-fe6d-4411-be2b-aec38524a8dd" pod="openshift-marketplace/redhat-marketplace-xwxlf" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/redhat-marketplace-xwxlf\": dial tcp 38.102.83.195:6443: connect: connection refused" Dec 13 06:53:11 crc kubenswrapper[4971]: I1213 06:53:11.621164 4971 status_manager.go:851] "Failed to get status for pod" podUID="02f6e9f2-6bad-441c-8e2e-9acfc71f3ab4" pod="openshift-marketplace/redhat-operators-wpxfk" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/redhat-operators-wpxfk\": dial tcp 38.102.83.195:6443: connect: connection refused" Dec 13 06:53:11 crc kubenswrapper[4971]: I1213 06:53:11.621397 4971 status_manager.go:851] "Failed to get status for pod" podUID="5c88a864-89a9-4f5e-b713-385170b9bf17" pod="openshift-marketplace/marketplace-operator-79b997595-6rvc9" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/marketplace-operator-79b997595-6rvc9\": dial tcp 38.102.83.195:6443: connect: connection refused" Dec 13 06:53:11 crc kubenswrapper[4971]: I1213 06:53:11.621682 4971 status_manager.go:851] "Failed to get status for pod" podUID="8f50f5ba-3ec3-4fa2-9f5d-51e1c43f6352" pod="openshift-marketplace/certified-operators-rkd9c" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/certified-operators-rkd9c\": dial tcp 38.102.83.195:6443: connect: connection refused" Dec 13 06:53:11 crc kubenswrapper[4971]: I1213 06:53:11.621918 4971 status_manager.go:851] "Failed to get status for pod" podUID="9d4f1ea5-8472-4320-820b-8090ff6f5ca6" pod="openshift-marketplace/redhat-marketplace-mkqps" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/redhat-marketplace-mkqps\": dial tcp 38.102.83.195:6443: connect: connection refused" Dec 13 06:53:11 crc kubenswrapper[4971]: I1213 06:53:11.622113 4971 status_manager.go:851] "Failed to get status for pod" podUID="53fdcbc2-900b-48eb-8be3-cd72940096c3" pod="openshift-kube-apiserver/installer-9-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/installer-9-crc\": dial tcp 38.102.83.195:6443: connect: connection refused" Dec 13 06:53:11 crc kubenswrapper[4971]: I1213 06:53:11.632878 4971 scope.go:117] "RemoveContainer" containerID="13e3b312f5f1c8e5fb3236ed71e96a35521fd5172708bc5ba5e184501233d15a" Dec 13 06:53:11 crc kubenswrapper[4971]: I1213 06:53:11.646746 4971 scope.go:117] "RemoveContainer" containerID="f1a912a2a413b3b52e51cb9027a159d5bb254455287625669896a9d6362ddc4f" Dec 13 06:53:11 crc kubenswrapper[4971]: I1213 06:53:11.659220 4971 scope.go:117] "RemoveContainer" containerID="5e118dbb3b89959ae750afb59cd8dca72e876ecbeb215d5fb1c410b36e5e10c0" Dec 13 06:53:11 crc kubenswrapper[4971]: I1213 06:53:11.684116 4971 scope.go:117] "RemoveContainer" containerID="a45464afad8c3643d3a5c8eb822b09c0d2807f574631a94b542744363a306d6c" Dec 13 06:53:11 crc kubenswrapper[4971]: I1213 06:53:11.702289 4971 scope.go:117] "RemoveContainer" containerID="a8cf7f8ba01f1f3be0caea6b89cddcbc28e056346def1520e8197eaa54f64016" Dec 13 06:53:11 crc kubenswrapper[4971]: I1213 06:53:11.715440 4971 scope.go:117] "RemoveContainer" containerID="cf5a955add1c7bf17976d195f2fba78c93d0fd447a0684c6e8377e595e9546ad" Dec 13 06:53:11 crc kubenswrapper[4971]: I1213 06:53:11.726087 4971 scope.go:117] "RemoveContainer" containerID="ed8f0295bdfe44b9c612064aba54bb4629a51eab4eebe12267dfaf5b4c03532b" Dec 13 06:53:11 crc kubenswrapper[4971]: I1213 06:53:11.751818 4971 scope.go:117] "RemoveContainer" containerID="970df14f0e90612c6aee36fde7557ac705f1ce238a087fd366a6b85f623dba26" Dec 13 06:53:11 crc kubenswrapper[4971]: I1213 06:53:11.766739 4971 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/53fdcbc2-900b-48eb-8be3-cd72940096c3-kube-api-access\") pod \"53fdcbc2-900b-48eb-8be3-cd72940096c3\" (UID: \"53fdcbc2-900b-48eb-8be3-cd72940096c3\") " Dec 13 06:53:11 crc kubenswrapper[4971]: I1213 06:53:11.766790 4971 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/53fdcbc2-900b-48eb-8be3-cd72940096c3-kubelet-dir\") pod \"53fdcbc2-900b-48eb-8be3-cd72940096c3\" (UID: \"53fdcbc2-900b-48eb-8be3-cd72940096c3\") " Dec 13 06:53:11 crc kubenswrapper[4971]: I1213 06:53:11.766881 4971 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"var-lock\" (UniqueName: \"kubernetes.io/host-path/53fdcbc2-900b-48eb-8be3-cd72940096c3-var-lock\") pod \"53fdcbc2-900b-48eb-8be3-cd72940096c3\" (UID: \"53fdcbc2-900b-48eb-8be3-cd72940096c3\") " Dec 13 06:53:11 crc kubenswrapper[4971]: I1213 06:53:11.767010 4971 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/53fdcbc2-900b-48eb-8be3-cd72940096c3-var-lock" (OuterVolumeSpecName: "var-lock") pod "53fdcbc2-900b-48eb-8be3-cd72940096c3" (UID: "53fdcbc2-900b-48eb-8be3-cd72940096c3"). InnerVolumeSpecName "var-lock". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 13 06:53:11 crc kubenswrapper[4971]: I1213 06:53:11.766973 4971 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/53fdcbc2-900b-48eb-8be3-cd72940096c3-kubelet-dir" (OuterVolumeSpecName: "kubelet-dir") pod "53fdcbc2-900b-48eb-8be3-cd72940096c3" (UID: "53fdcbc2-900b-48eb-8be3-cd72940096c3"). InnerVolumeSpecName "kubelet-dir". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 13 06:53:11 crc kubenswrapper[4971]: I1213 06:53:11.767176 4971 reconciler_common.go:293] "Volume detached for volume \"var-lock\" (UniqueName: \"kubernetes.io/host-path/53fdcbc2-900b-48eb-8be3-cd72940096c3-var-lock\") on node \"crc\" DevicePath \"\"" Dec 13 06:53:11 crc kubenswrapper[4971]: I1213 06:53:11.767194 4971 reconciler_common.go:293] "Volume detached for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/53fdcbc2-900b-48eb-8be3-cd72940096c3-kubelet-dir\") on node \"crc\" DevicePath \"\"" Dec 13 06:53:11 crc kubenswrapper[4971]: I1213 06:53:11.771922 4971 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/53fdcbc2-900b-48eb-8be3-cd72940096c3-kube-api-access" (OuterVolumeSpecName: "kube-api-access") pod "53fdcbc2-900b-48eb-8be3-cd72940096c3" (UID: "53fdcbc2-900b-48eb-8be3-cd72940096c3"). InnerVolumeSpecName "kube-api-access". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 13 06:53:11 crc kubenswrapper[4971]: I1213 06:53:11.779730 4971 scope.go:117] "RemoveContainer" containerID="14fb548fd8178b7fa4ade14c8907b148dd031b77bdcd95bcd98d8b54fc688a83" Dec 13 06:53:11 crc kubenswrapper[4971]: I1213 06:53:11.848913 4971 scope.go:117] "RemoveContainer" containerID="8ead1d45e4814ca9d385e346db4d8556a4ac91e83105c52ee55c9b712da52a36" Dec 13 06:53:11 crc kubenswrapper[4971]: I1213 06:53:11.863526 4971 scope.go:117] "RemoveContainer" containerID="1d8c1c7524f867e2c7ec3ccc601f78308c890a37019036f24c1eb839af11281b" Dec 13 06:53:11 crc kubenswrapper[4971]: I1213 06:53:11.868435 4971 reconciler_common.go:293] "Volume detached for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/53fdcbc2-900b-48eb-8be3-cd72940096c3-kube-api-access\") on node \"crc\" DevicePath \"\"" Dec 13 06:53:12 crc kubenswrapper[4971]: I1213 06:53:12.222017 4971 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-kube-apiserver_kube-apiserver-crc_f4b27818a5e8e43d0dc095d08835c792/kube-apiserver-cert-syncer/0.log" Dec 13 06:53:12 crc kubenswrapper[4971]: I1213 06:53:12.223131 4971 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/kube-apiserver-crc" Dec 13 06:53:12 crc kubenswrapper[4971]: I1213 06:53:12.223828 4971 status_manager.go:851] "Failed to get status for pod" podUID="cf417851-ae43-462f-a678-125593fb653b" pod="openshift-marketplace/community-operators-997kc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/community-operators-997kc\": dial tcp 38.102.83.195:6443: connect: connection refused" Dec 13 06:53:12 crc kubenswrapper[4971]: I1213 06:53:12.224221 4971 status_manager.go:851] "Failed to get status for pod" podUID="22037bdc-fe6d-4411-be2b-aec38524a8dd" pod="openshift-marketplace/redhat-marketplace-xwxlf" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/redhat-marketplace-xwxlf\": dial tcp 38.102.83.195:6443: connect: connection refused" Dec 13 06:53:12 crc kubenswrapper[4971]: I1213 06:53:12.224546 4971 status_manager.go:851] "Failed to get status for pod" podUID="02f6e9f2-6bad-441c-8e2e-9acfc71f3ab4" pod="openshift-marketplace/redhat-operators-wpxfk" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/redhat-operators-wpxfk\": dial tcp 38.102.83.195:6443: connect: connection refused" Dec 13 06:53:12 crc kubenswrapper[4971]: I1213 06:53:12.224872 4971 status_manager.go:851] "Failed to get status for pod" podUID="5c88a864-89a9-4f5e-b713-385170b9bf17" pod="openshift-marketplace/marketplace-operator-79b997595-6rvc9" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/marketplace-operator-79b997595-6rvc9\": dial tcp 38.102.83.195:6443: connect: connection refused" Dec 13 06:53:12 crc kubenswrapper[4971]: I1213 06:53:12.225092 4971 status_manager.go:851] "Failed to get status for pod" podUID="8f50f5ba-3ec3-4fa2-9f5d-51e1c43f6352" pod="openshift-marketplace/certified-operators-rkd9c" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/certified-operators-rkd9c\": dial tcp 38.102.83.195:6443: connect: connection refused" Dec 13 06:53:12 crc kubenswrapper[4971]: I1213 06:53:12.225363 4971 status_manager.go:851] "Failed to get status for pod" podUID="f4b27818a5e8e43d0dc095d08835c792" pod="openshift-kube-apiserver/kube-apiserver-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\": dial tcp 38.102.83.195:6443: connect: connection refused" Dec 13 06:53:12 crc kubenswrapper[4971]: I1213 06:53:12.225697 4971 status_manager.go:851] "Failed to get status for pod" podUID="9d4f1ea5-8472-4320-820b-8090ff6f5ca6" pod="openshift-marketplace/redhat-marketplace-mkqps" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/redhat-marketplace-mkqps\": dial tcp 38.102.83.195:6443: connect: connection refused" Dec 13 06:53:12 crc kubenswrapper[4971]: I1213 06:53:12.225934 4971 status_manager.go:851] "Failed to get status for pod" podUID="53fdcbc2-900b-48eb-8be3-cd72940096c3" pod="openshift-kube-apiserver/installer-9-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/installer-9-crc\": dial tcp 38.102.83.195:6443: connect: connection refused" Dec 13 06:53:12 crc kubenswrapper[4971]: I1213 06:53:12.226231 4971 status_manager.go:851] "Failed to get status for pod" podUID="b9a0bc5d-3a39-4732-a849-22d351a4cc18" pod="openshift-marketplace/redhat-operators-d8n8p" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/redhat-operators-d8n8p\": dial tcp 38.102.83.195:6443: connect: connection refused" Dec 13 06:53:12 crc kubenswrapper[4971]: I1213 06:53:12.226555 4971 status_manager.go:851] "Failed to get status for pod" podUID="f85e55b1a89d02b0cb034b1ea31ed45a" pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-startup-monitor-crc\": dial tcp 38.102.83.195:6443: connect: connection refused" Dec 13 06:53:12 crc kubenswrapper[4971]: I1213 06:53:12.226792 4971 status_manager.go:851] "Failed to get status for pod" podUID="f14c8a94-8589-4059-bddf-329452355fab" pod="openshift-marketplace/marketplace-operator-79b997595-j9zqf" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/marketplace-operator-79b997595-j9zqf\": dial tcp 38.102.83.195:6443: connect: connection refused" Dec 13 06:53:12 crc kubenswrapper[4971]: I1213 06:53:12.374039 4971 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-audit-dir\") pod \"f4b27818a5e8e43d0dc095d08835c792\" (UID: \"f4b27818a5e8e43d0dc095d08835c792\") " Dec 13 06:53:12 crc kubenswrapper[4971]: I1213 06:53:12.374330 4971 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-audit-dir" (OuterVolumeSpecName: "audit-dir") pod "f4b27818a5e8e43d0dc095d08835c792" (UID: "f4b27818a5e8e43d0dc095d08835c792"). InnerVolumeSpecName "audit-dir". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 13 06:53:12 crc kubenswrapper[4971]: I1213 06:53:12.374638 4971 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-cert-dir\") pod \"f4b27818a5e8e43d0dc095d08835c792\" (UID: \"f4b27818a5e8e43d0dc095d08835c792\") " Dec 13 06:53:12 crc kubenswrapper[4971]: I1213 06:53:12.374833 4971 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-resource-dir\") pod \"f4b27818a5e8e43d0dc095d08835c792\" (UID: \"f4b27818a5e8e43d0dc095d08835c792\") " Dec 13 06:53:12 crc kubenswrapper[4971]: I1213 06:53:12.375242 4971 reconciler_common.go:293] "Volume detached for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-audit-dir\") on node \"crc\" DevicePath \"\"" Dec 13 06:53:12 crc kubenswrapper[4971]: I1213 06:53:12.375369 4971 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-resource-dir" (OuterVolumeSpecName: "resource-dir") pod "f4b27818a5e8e43d0dc095d08835c792" (UID: "f4b27818a5e8e43d0dc095d08835c792"). InnerVolumeSpecName "resource-dir". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 13 06:53:12 crc kubenswrapper[4971]: I1213 06:53:12.375482 4971 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-cert-dir" (OuterVolumeSpecName: "cert-dir") pod "f4b27818a5e8e43d0dc095d08835c792" (UID: "f4b27818a5e8e43d0dc095d08835c792"). InnerVolumeSpecName "cert-dir". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 13 06:53:12 crc kubenswrapper[4971]: I1213 06:53:12.429294 4971 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_marketplace-operator-79b997595-6rvc9_5c88a864-89a9-4f5e-b713-385170b9bf17/marketplace-operator/1.log" Dec 13 06:53:12 crc kubenswrapper[4971]: I1213 06:53:12.430841 4971 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_marketplace-operator-79b997595-6rvc9_5c88a864-89a9-4f5e-b713-385170b9bf17/marketplace-operator/0.log" Dec 13 06:53:12 crc kubenswrapper[4971]: I1213 06:53:12.430877 4971 generic.go:334] "Generic (PLEG): container finished" podID="5c88a864-89a9-4f5e-b713-385170b9bf17" containerID="eec75e44095b014d79821c2ca91b51f25998f1dd4330e3e1bdc38a006708a436" exitCode=1 Dec 13 06:53:12 crc kubenswrapper[4971]: I1213 06:53:12.431352 4971 scope.go:117] "RemoveContainer" containerID="eec75e44095b014d79821c2ca91b51f25998f1dd4330e3e1bdc38a006708a436" Dec 13 06:53:12 crc kubenswrapper[4971]: I1213 06:53:12.431673 4971 status_manager.go:851] "Failed to get status for pod" podUID="8f50f5ba-3ec3-4fa2-9f5d-51e1c43f6352" pod="openshift-marketplace/certified-operators-rkd9c" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/certified-operators-rkd9c\": dial tcp 38.102.83.195:6443: connect: connection refused" Dec 13 06:53:12 crc kubenswrapper[4971]: I1213 06:53:12.431027 4971 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/marketplace-operator-79b997595-6rvc9" event={"ID":"5c88a864-89a9-4f5e-b713-385170b9bf17","Type":"ContainerDied","Data":"eec75e44095b014d79821c2ca91b51f25998f1dd4330e3e1bdc38a006708a436"} Dec 13 06:53:12 crc kubenswrapper[4971]: I1213 06:53:12.432002 4971 status_manager.go:851] "Failed to get status for pod" podUID="22037bdc-fe6d-4411-be2b-aec38524a8dd" pod="openshift-marketplace/redhat-marketplace-xwxlf" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/redhat-marketplace-xwxlf\": dial tcp 38.102.83.195:6443: connect: connection refused" Dec 13 06:53:12 crc kubenswrapper[4971]: I1213 06:53:12.432319 4971 scope.go:117] "RemoveContainer" containerID="7a53386540871a14f654bd05dae1a7988d7a3396a65ccf45bb2e98846899cc87" Dec 13 06:53:12 crc kubenswrapper[4971]: I1213 06:53:12.432451 4971 status_manager.go:851] "Failed to get status for pod" podUID="02f6e9f2-6bad-441c-8e2e-9acfc71f3ab4" pod="openshift-marketplace/redhat-operators-wpxfk" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/redhat-operators-wpxfk\": dial tcp 38.102.83.195:6443: connect: connection refused" Dec 13 06:53:12 crc kubenswrapper[4971]: E1213 06:53:12.432474 4971 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"marketplace-operator\" with CrashLoopBackOff: \"back-off 10s restarting failed container=marketplace-operator pod=marketplace-operator-79b997595-6rvc9_openshift-marketplace(5c88a864-89a9-4f5e-b713-385170b9bf17)\"" pod="openshift-marketplace/marketplace-operator-79b997595-6rvc9" podUID="5c88a864-89a9-4f5e-b713-385170b9bf17" Dec 13 06:53:12 crc kubenswrapper[4971]: I1213 06:53:12.432887 4971 status_manager.go:851] "Failed to get status for pod" podUID="5c88a864-89a9-4f5e-b713-385170b9bf17" pod="openshift-marketplace/marketplace-operator-79b997595-6rvc9" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/marketplace-operator-79b997595-6rvc9\": dial tcp 38.102.83.195:6443: connect: connection refused" Dec 13 06:53:12 crc kubenswrapper[4971]: I1213 06:53:12.433340 4971 status_manager.go:851] "Failed to get status for pod" podUID="f4b27818a5e8e43d0dc095d08835c792" pod="openshift-kube-apiserver/kube-apiserver-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\": dial tcp 38.102.83.195:6443: connect: connection refused" Dec 13 06:53:12 crc kubenswrapper[4971]: I1213 06:53:12.433582 4971 status_manager.go:851] "Failed to get status for pod" podUID="9d4f1ea5-8472-4320-820b-8090ff6f5ca6" pod="openshift-marketplace/redhat-marketplace-mkqps" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/redhat-marketplace-mkqps\": dial tcp 38.102.83.195:6443: connect: connection refused" Dec 13 06:53:12 crc kubenswrapper[4971]: I1213 06:53:12.433790 4971 status_manager.go:851] "Failed to get status for pod" podUID="53fdcbc2-900b-48eb-8be3-cd72940096c3" pod="openshift-kube-apiserver/installer-9-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/installer-9-crc\": dial tcp 38.102.83.195:6443: connect: connection refused" Dec 13 06:53:12 crc kubenswrapper[4971]: I1213 06:53:12.434055 4971 status_manager.go:851] "Failed to get status for pod" podUID="f14c8a94-8589-4059-bddf-329452355fab" pod="openshift-marketplace/marketplace-operator-79b997595-j9zqf" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/marketplace-operator-79b997595-j9zqf\": dial tcp 38.102.83.195:6443: connect: connection refused" Dec 13 06:53:12 crc kubenswrapper[4971]: I1213 06:53:12.437012 4971 status_manager.go:851] "Failed to get status for pod" podUID="b9a0bc5d-3a39-4732-a849-22d351a4cc18" pod="openshift-marketplace/redhat-operators-d8n8p" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/redhat-operators-d8n8p\": dial tcp 38.102.83.195:6443: connect: connection refused" Dec 13 06:53:12 crc kubenswrapper[4971]: I1213 06:53:12.437690 4971 status_manager.go:851] "Failed to get status for pod" podUID="f85e55b1a89d02b0cb034b1ea31ed45a" pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-startup-monitor-crc\": dial tcp 38.102.83.195:6443: connect: connection refused" Dec 13 06:53:12 crc kubenswrapper[4971]: I1213 06:53:12.437930 4971 status_manager.go:851] "Failed to get status for pod" podUID="cf417851-ae43-462f-a678-125593fb653b" pod="openshift-marketplace/community-operators-997kc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/community-operators-997kc\": dial tcp 38.102.83.195:6443: connect: connection refused" Dec 13 06:53:12 crc kubenswrapper[4971]: I1213 06:53:12.438458 4971 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-kube-apiserver_kube-apiserver-crc_f4b27818a5e8e43d0dc095d08835c792/kube-apiserver-cert-syncer/0.log" Dec 13 06:53:12 crc kubenswrapper[4971]: I1213 06:53:12.439609 4971 generic.go:334] "Generic (PLEG): container finished" podID="f4b27818a5e8e43d0dc095d08835c792" containerID="3716afdf2bc693be80a09fc9f7727d334f435caf1d4c8ec4811b3a1f1652cdfe" exitCode=0 Dec 13 06:53:12 crc kubenswrapper[4971]: I1213 06:53:12.440124 4971 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/kube-apiserver-crc" Dec 13 06:53:12 crc kubenswrapper[4971]: I1213 06:53:12.451874 4971 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/installer-9-crc" Dec 13 06:53:12 crc kubenswrapper[4971]: I1213 06:53:12.452486 4971 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/installer-9-crc" event={"ID":"53fdcbc2-900b-48eb-8be3-cd72940096c3","Type":"ContainerDied","Data":"1fd88163f64ef84917208a4b013342315222b8bfaf5a456f69c32021ec433c8f"} Dec 13 06:53:12 crc kubenswrapper[4971]: I1213 06:53:12.452783 4971 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="1fd88163f64ef84917208a4b013342315222b8bfaf5a456f69c32021ec433c8f" Dec 13 06:53:12 crc kubenswrapper[4971]: I1213 06:53:12.457593 4971 scope.go:117] "RemoveContainer" containerID="770b73dd0c2b6f38744e2c1cd6aa186a69bb87d1cd015a603ad2e33be06a8f6b" Dec 13 06:53:12 crc kubenswrapper[4971]: I1213 06:53:12.458191 4971 status_manager.go:851] "Failed to get status for pod" podUID="22037bdc-fe6d-4411-be2b-aec38524a8dd" pod="openshift-marketplace/redhat-marketplace-xwxlf" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/redhat-marketplace-xwxlf\": dial tcp 38.102.83.195:6443: connect: connection refused" Dec 13 06:53:12 crc kubenswrapper[4971]: I1213 06:53:12.458689 4971 status_manager.go:851] "Failed to get status for pod" podUID="02f6e9f2-6bad-441c-8e2e-9acfc71f3ab4" pod="openshift-marketplace/redhat-operators-wpxfk" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/redhat-operators-wpxfk\": dial tcp 38.102.83.195:6443: connect: connection refused" Dec 13 06:53:12 crc kubenswrapper[4971]: I1213 06:53:12.459310 4971 status_manager.go:851] "Failed to get status for pod" podUID="5c88a864-89a9-4f5e-b713-385170b9bf17" pod="openshift-marketplace/marketplace-operator-79b997595-6rvc9" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/marketplace-operator-79b997595-6rvc9\": dial tcp 38.102.83.195:6443: connect: connection refused" Dec 13 06:53:12 crc kubenswrapper[4971]: I1213 06:53:12.459675 4971 status_manager.go:851] "Failed to get status for pod" podUID="8f50f5ba-3ec3-4fa2-9f5d-51e1c43f6352" pod="openshift-marketplace/certified-operators-rkd9c" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/certified-operators-rkd9c\": dial tcp 38.102.83.195:6443: connect: connection refused" Dec 13 06:53:12 crc kubenswrapper[4971]: I1213 06:53:12.459834 4971 status_manager.go:851] "Failed to get status for pod" podUID="f4b27818a5e8e43d0dc095d08835c792" pod="openshift-kube-apiserver/kube-apiserver-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\": dial tcp 38.102.83.195:6443: connect: connection refused" Dec 13 06:53:12 crc kubenswrapper[4971]: I1213 06:53:12.460005 4971 status_manager.go:851] "Failed to get status for pod" podUID="9d4f1ea5-8472-4320-820b-8090ff6f5ca6" pod="openshift-marketplace/redhat-marketplace-mkqps" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/redhat-marketplace-mkqps\": dial tcp 38.102.83.195:6443: connect: connection refused" Dec 13 06:53:12 crc kubenswrapper[4971]: I1213 06:53:12.460636 4971 status_manager.go:851] "Failed to get status for pod" podUID="53fdcbc2-900b-48eb-8be3-cd72940096c3" pod="openshift-kube-apiserver/installer-9-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/installer-9-crc\": dial tcp 38.102.83.195:6443: connect: connection refused" Dec 13 06:53:12 crc kubenswrapper[4971]: I1213 06:53:12.461918 4971 status_manager.go:851] "Failed to get status for pod" podUID="b9a0bc5d-3a39-4732-a849-22d351a4cc18" pod="openshift-marketplace/redhat-operators-d8n8p" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/redhat-operators-d8n8p\": dial tcp 38.102.83.195:6443: connect: connection refused" Dec 13 06:53:12 crc kubenswrapper[4971]: I1213 06:53:12.463122 4971 status_manager.go:851] "Failed to get status for pod" podUID="f85e55b1a89d02b0cb034b1ea31ed45a" pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-startup-monitor-crc\": dial tcp 38.102.83.195:6443: connect: connection refused" Dec 13 06:53:12 crc kubenswrapper[4971]: I1213 06:53:12.463665 4971 status_manager.go:851] "Failed to get status for pod" podUID="f14c8a94-8589-4059-bddf-329452355fab" pod="openshift-marketplace/marketplace-operator-79b997595-j9zqf" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/marketplace-operator-79b997595-j9zqf\": dial tcp 38.102.83.195:6443: connect: connection refused" Dec 13 06:53:12 crc kubenswrapper[4971]: I1213 06:53:12.464496 4971 status_manager.go:851] "Failed to get status for pod" podUID="cf417851-ae43-462f-a678-125593fb653b" pod="openshift-marketplace/community-operators-997kc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/community-operators-997kc\": dial tcp 38.102.83.195:6443: connect: connection refused" Dec 13 06:53:12 crc kubenswrapper[4971]: I1213 06:53:12.465176 4971 status_manager.go:851] "Failed to get status for pod" podUID="8f50f5ba-3ec3-4fa2-9f5d-51e1c43f6352" pod="openshift-marketplace/certified-operators-rkd9c" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/certified-operators-rkd9c\": dial tcp 38.102.83.195:6443: connect: connection refused" Dec 13 06:53:12 crc kubenswrapper[4971]: I1213 06:53:12.465553 4971 status_manager.go:851] "Failed to get status for pod" podUID="22037bdc-fe6d-4411-be2b-aec38524a8dd" pod="openshift-marketplace/redhat-marketplace-xwxlf" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/redhat-marketplace-xwxlf\": dial tcp 38.102.83.195:6443: connect: connection refused" Dec 13 06:53:12 crc kubenswrapper[4971]: I1213 06:53:12.466166 4971 status_manager.go:851] "Failed to get status for pod" podUID="02f6e9f2-6bad-441c-8e2e-9acfc71f3ab4" pod="openshift-marketplace/redhat-operators-wpxfk" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/redhat-operators-wpxfk\": dial tcp 38.102.83.195:6443: connect: connection refused" Dec 13 06:53:12 crc kubenswrapper[4971]: I1213 06:53:12.466432 4971 status_manager.go:851] "Failed to get status for pod" podUID="5c88a864-89a9-4f5e-b713-385170b9bf17" pod="openshift-marketplace/marketplace-operator-79b997595-6rvc9" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/marketplace-operator-79b997595-6rvc9\": dial tcp 38.102.83.195:6443: connect: connection refused" Dec 13 06:53:12 crc kubenswrapper[4971]: I1213 06:53:12.467111 4971 status_manager.go:851] "Failed to get status for pod" podUID="f4b27818a5e8e43d0dc095d08835c792" pod="openshift-kube-apiserver/kube-apiserver-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\": dial tcp 38.102.83.195:6443: connect: connection refused" Dec 13 06:53:12 crc kubenswrapper[4971]: I1213 06:53:12.467350 4971 status_manager.go:851] "Failed to get status for pod" podUID="9d4f1ea5-8472-4320-820b-8090ff6f5ca6" pod="openshift-marketplace/redhat-marketplace-mkqps" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/redhat-marketplace-mkqps\": dial tcp 38.102.83.195:6443: connect: connection refused" Dec 13 06:53:12 crc kubenswrapper[4971]: I1213 06:53:12.467555 4971 status_manager.go:851] "Failed to get status for pod" podUID="53fdcbc2-900b-48eb-8be3-cd72940096c3" pod="openshift-kube-apiserver/installer-9-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/installer-9-crc\": dial tcp 38.102.83.195:6443: connect: connection refused" Dec 13 06:53:12 crc kubenswrapper[4971]: I1213 06:53:12.467846 4971 status_manager.go:851] "Failed to get status for pod" podUID="f14c8a94-8589-4059-bddf-329452355fab" pod="openshift-marketplace/marketplace-operator-79b997595-j9zqf" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/marketplace-operator-79b997595-j9zqf\": dial tcp 38.102.83.195:6443: connect: connection refused" Dec 13 06:53:12 crc kubenswrapper[4971]: I1213 06:53:12.468325 4971 status_manager.go:851] "Failed to get status for pod" podUID="b9a0bc5d-3a39-4732-a849-22d351a4cc18" pod="openshift-marketplace/redhat-operators-d8n8p" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/redhat-operators-d8n8p\": dial tcp 38.102.83.195:6443: connect: connection refused" Dec 13 06:53:12 crc kubenswrapper[4971]: I1213 06:53:12.468592 4971 status_manager.go:851] "Failed to get status for pod" podUID="f85e55b1a89d02b0cb034b1ea31ed45a" pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-startup-monitor-crc\": dial tcp 38.102.83.195:6443: connect: connection refused" Dec 13 06:53:12 crc kubenswrapper[4971]: I1213 06:53:12.474903 4971 scope.go:117] "RemoveContainer" containerID="9acd07066deed483bd7fa61ee45fda09539d110717a134a56827f871186f92fc" Dec 13 06:53:12 crc kubenswrapper[4971]: I1213 06:53:12.476207 4971 reconciler_common.go:293] "Volume detached for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-cert-dir\") on node \"crc\" DevicePath \"\"" Dec 13 06:53:12 crc kubenswrapper[4971]: I1213 06:53:12.476236 4971 reconciler_common.go:293] "Volume detached for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-resource-dir\") on node \"crc\" DevicePath \"\"" Dec 13 06:53:12 crc kubenswrapper[4971]: I1213 06:53:12.488371 4971 status_manager.go:851] "Failed to get status for pod" podUID="cf417851-ae43-462f-a678-125593fb653b" pod="openshift-marketplace/community-operators-997kc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/community-operators-997kc\": dial tcp 38.102.83.195:6443: connect: connection refused" Dec 13 06:53:12 crc kubenswrapper[4971]: I1213 06:53:12.489159 4971 scope.go:117] "RemoveContainer" containerID="30df3abf8c79db8af52c6a918f73e1ce410ddafcf341636424c267e1eeb195a7" Dec 13 06:53:12 crc kubenswrapper[4971]: I1213 06:53:12.501586 4971 scope.go:117] "RemoveContainer" containerID="ccdf65f6d2b1eb3647a03ba37c305c32e34addfdf79ff937c8dddaa3e5d5b5c6" Dec 13 06:53:12 crc kubenswrapper[4971]: I1213 06:53:12.511859 4971 scope.go:117] "RemoveContainer" containerID="3716afdf2bc693be80a09fc9f7727d334f435caf1d4c8ec4811b3a1f1652cdfe" Dec 13 06:53:12 crc kubenswrapper[4971]: I1213 06:53:12.541155 4971 scope.go:117] "RemoveContainer" containerID="38c91cfbaecfdbd3e3b663422a91ab4294610afb6bbf9ebdb77db24e2b76c682" Dec 13 06:53:12 crc kubenswrapper[4971]: I1213 06:53:12.569872 4971 scope.go:117] "RemoveContainer" containerID="770b73dd0c2b6f38744e2c1cd6aa186a69bb87d1cd015a603ad2e33be06a8f6b" Dec 13 06:53:12 crc kubenswrapper[4971]: E1213 06:53:12.570440 4971 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"770b73dd0c2b6f38744e2c1cd6aa186a69bb87d1cd015a603ad2e33be06a8f6b\": container with ID starting with 770b73dd0c2b6f38744e2c1cd6aa186a69bb87d1cd015a603ad2e33be06a8f6b not found: ID does not exist" containerID="770b73dd0c2b6f38744e2c1cd6aa186a69bb87d1cd015a603ad2e33be06a8f6b" Dec 13 06:53:12 crc kubenswrapper[4971]: I1213 06:53:12.570481 4971 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"770b73dd0c2b6f38744e2c1cd6aa186a69bb87d1cd015a603ad2e33be06a8f6b"} err="failed to get container status \"770b73dd0c2b6f38744e2c1cd6aa186a69bb87d1cd015a603ad2e33be06a8f6b\": rpc error: code = NotFound desc = could not find container \"770b73dd0c2b6f38744e2c1cd6aa186a69bb87d1cd015a603ad2e33be06a8f6b\": container with ID starting with 770b73dd0c2b6f38744e2c1cd6aa186a69bb87d1cd015a603ad2e33be06a8f6b not found: ID does not exist" Dec 13 06:53:12 crc kubenswrapper[4971]: I1213 06:53:12.570508 4971 scope.go:117] "RemoveContainer" containerID="9acd07066deed483bd7fa61ee45fda09539d110717a134a56827f871186f92fc" Dec 13 06:53:12 crc kubenswrapper[4971]: E1213 06:53:12.570899 4971 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"9acd07066deed483bd7fa61ee45fda09539d110717a134a56827f871186f92fc\": container with ID starting with 9acd07066deed483bd7fa61ee45fda09539d110717a134a56827f871186f92fc not found: ID does not exist" containerID="9acd07066deed483bd7fa61ee45fda09539d110717a134a56827f871186f92fc" Dec 13 06:53:12 crc kubenswrapper[4971]: I1213 06:53:12.570944 4971 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"9acd07066deed483bd7fa61ee45fda09539d110717a134a56827f871186f92fc"} err="failed to get container status \"9acd07066deed483bd7fa61ee45fda09539d110717a134a56827f871186f92fc\": rpc error: code = NotFound desc = could not find container \"9acd07066deed483bd7fa61ee45fda09539d110717a134a56827f871186f92fc\": container with ID starting with 9acd07066deed483bd7fa61ee45fda09539d110717a134a56827f871186f92fc not found: ID does not exist" Dec 13 06:53:12 crc kubenswrapper[4971]: I1213 06:53:12.570974 4971 scope.go:117] "RemoveContainer" containerID="30df3abf8c79db8af52c6a918f73e1ce410ddafcf341636424c267e1eeb195a7" Dec 13 06:53:12 crc kubenswrapper[4971]: E1213 06:53:12.571299 4971 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"30df3abf8c79db8af52c6a918f73e1ce410ddafcf341636424c267e1eeb195a7\": container with ID starting with 30df3abf8c79db8af52c6a918f73e1ce410ddafcf341636424c267e1eeb195a7 not found: ID does not exist" containerID="30df3abf8c79db8af52c6a918f73e1ce410ddafcf341636424c267e1eeb195a7" Dec 13 06:53:12 crc kubenswrapper[4971]: I1213 06:53:12.571341 4971 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"30df3abf8c79db8af52c6a918f73e1ce410ddafcf341636424c267e1eeb195a7"} err="failed to get container status \"30df3abf8c79db8af52c6a918f73e1ce410ddafcf341636424c267e1eeb195a7\": rpc error: code = NotFound desc = could not find container \"30df3abf8c79db8af52c6a918f73e1ce410ddafcf341636424c267e1eeb195a7\": container with ID starting with 30df3abf8c79db8af52c6a918f73e1ce410ddafcf341636424c267e1eeb195a7 not found: ID does not exist" Dec 13 06:53:12 crc kubenswrapper[4971]: I1213 06:53:12.571371 4971 scope.go:117] "RemoveContainer" containerID="ccdf65f6d2b1eb3647a03ba37c305c32e34addfdf79ff937c8dddaa3e5d5b5c6" Dec 13 06:53:12 crc kubenswrapper[4971]: E1213 06:53:12.571848 4971 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"ccdf65f6d2b1eb3647a03ba37c305c32e34addfdf79ff937c8dddaa3e5d5b5c6\": container with ID starting with ccdf65f6d2b1eb3647a03ba37c305c32e34addfdf79ff937c8dddaa3e5d5b5c6 not found: ID does not exist" containerID="ccdf65f6d2b1eb3647a03ba37c305c32e34addfdf79ff937c8dddaa3e5d5b5c6" Dec 13 06:53:12 crc kubenswrapper[4971]: I1213 06:53:12.571871 4971 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"ccdf65f6d2b1eb3647a03ba37c305c32e34addfdf79ff937c8dddaa3e5d5b5c6"} err="failed to get container status \"ccdf65f6d2b1eb3647a03ba37c305c32e34addfdf79ff937c8dddaa3e5d5b5c6\": rpc error: code = NotFound desc = could not find container \"ccdf65f6d2b1eb3647a03ba37c305c32e34addfdf79ff937c8dddaa3e5d5b5c6\": container with ID starting with ccdf65f6d2b1eb3647a03ba37c305c32e34addfdf79ff937c8dddaa3e5d5b5c6 not found: ID does not exist" Dec 13 06:53:12 crc kubenswrapper[4971]: I1213 06:53:12.571887 4971 scope.go:117] "RemoveContainer" containerID="3716afdf2bc693be80a09fc9f7727d334f435caf1d4c8ec4811b3a1f1652cdfe" Dec 13 06:53:12 crc kubenswrapper[4971]: E1213 06:53:12.572298 4971 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"3716afdf2bc693be80a09fc9f7727d334f435caf1d4c8ec4811b3a1f1652cdfe\": container with ID starting with 3716afdf2bc693be80a09fc9f7727d334f435caf1d4c8ec4811b3a1f1652cdfe not found: ID does not exist" containerID="3716afdf2bc693be80a09fc9f7727d334f435caf1d4c8ec4811b3a1f1652cdfe" Dec 13 06:53:12 crc kubenswrapper[4971]: I1213 06:53:12.572323 4971 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"3716afdf2bc693be80a09fc9f7727d334f435caf1d4c8ec4811b3a1f1652cdfe"} err="failed to get container status \"3716afdf2bc693be80a09fc9f7727d334f435caf1d4c8ec4811b3a1f1652cdfe\": rpc error: code = NotFound desc = could not find container \"3716afdf2bc693be80a09fc9f7727d334f435caf1d4c8ec4811b3a1f1652cdfe\": container with ID starting with 3716afdf2bc693be80a09fc9f7727d334f435caf1d4c8ec4811b3a1f1652cdfe not found: ID does not exist" Dec 13 06:53:12 crc kubenswrapper[4971]: I1213 06:53:12.572339 4971 scope.go:117] "RemoveContainer" containerID="38c91cfbaecfdbd3e3b663422a91ab4294610afb6bbf9ebdb77db24e2b76c682" Dec 13 06:53:12 crc kubenswrapper[4971]: E1213 06:53:12.572627 4971 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"38c91cfbaecfdbd3e3b663422a91ab4294610afb6bbf9ebdb77db24e2b76c682\": container with ID starting with 38c91cfbaecfdbd3e3b663422a91ab4294610afb6bbf9ebdb77db24e2b76c682 not found: ID does not exist" containerID="38c91cfbaecfdbd3e3b663422a91ab4294610afb6bbf9ebdb77db24e2b76c682" Dec 13 06:53:12 crc kubenswrapper[4971]: I1213 06:53:12.572649 4971 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"38c91cfbaecfdbd3e3b663422a91ab4294610afb6bbf9ebdb77db24e2b76c682"} err="failed to get container status \"38c91cfbaecfdbd3e3b663422a91ab4294610afb6bbf9ebdb77db24e2b76c682\": rpc error: code = NotFound desc = could not find container \"38c91cfbaecfdbd3e3b663422a91ab4294610afb6bbf9ebdb77db24e2b76c682\": container with ID starting with 38c91cfbaecfdbd3e3b663422a91ab4294610afb6bbf9ebdb77db24e2b76c682 not found: ID does not exist" Dec 13 06:53:13 crc kubenswrapper[4971]: I1213 06:53:13.482206 4971 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_marketplace-operator-79b997595-6rvc9_5c88a864-89a9-4f5e-b713-385170b9bf17/marketplace-operator/1.log" Dec 13 06:53:13 crc kubenswrapper[4971]: I1213 06:53:13.482643 4971 scope.go:117] "RemoveContainer" containerID="eec75e44095b014d79821c2ca91b51f25998f1dd4330e3e1bdc38a006708a436" Dec 13 06:53:13 crc kubenswrapper[4971]: E1213 06:53:13.482797 4971 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"marketplace-operator\" with CrashLoopBackOff: \"back-off 10s restarting failed container=marketplace-operator pod=marketplace-operator-79b997595-6rvc9_openshift-marketplace(5c88a864-89a9-4f5e-b713-385170b9bf17)\"" pod="openshift-marketplace/marketplace-operator-79b997595-6rvc9" podUID="5c88a864-89a9-4f5e-b713-385170b9bf17" Dec 13 06:53:13 crc kubenswrapper[4971]: I1213 06:53:13.483087 4971 status_manager.go:851] "Failed to get status for pod" podUID="02f6e9f2-6bad-441c-8e2e-9acfc71f3ab4" pod="openshift-marketplace/redhat-operators-wpxfk" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/redhat-operators-wpxfk\": dial tcp 38.102.83.195:6443: connect: connection refused" Dec 13 06:53:13 crc kubenswrapper[4971]: I1213 06:53:13.483236 4971 status_manager.go:851] "Failed to get status for pod" podUID="5c88a864-89a9-4f5e-b713-385170b9bf17" pod="openshift-marketplace/marketplace-operator-79b997595-6rvc9" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/marketplace-operator-79b997595-6rvc9\": dial tcp 38.102.83.195:6443: connect: connection refused" Dec 13 06:53:13 crc kubenswrapper[4971]: I1213 06:53:13.483421 4971 status_manager.go:851] "Failed to get status for pod" podUID="8f50f5ba-3ec3-4fa2-9f5d-51e1c43f6352" pod="openshift-marketplace/certified-operators-rkd9c" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/certified-operators-rkd9c\": dial tcp 38.102.83.195:6443: connect: connection refused" Dec 13 06:53:13 crc kubenswrapper[4971]: I1213 06:53:13.483606 4971 status_manager.go:851] "Failed to get status for pod" podUID="22037bdc-fe6d-4411-be2b-aec38524a8dd" pod="openshift-marketplace/redhat-marketplace-xwxlf" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/redhat-marketplace-xwxlf\": dial tcp 38.102.83.195:6443: connect: connection refused" Dec 13 06:53:13 crc kubenswrapper[4971]: I1213 06:53:13.483786 4971 status_manager.go:851] "Failed to get status for pod" podUID="f4b27818a5e8e43d0dc095d08835c792" pod="openshift-kube-apiserver/kube-apiserver-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\": dial tcp 38.102.83.195:6443: connect: connection refused" Dec 13 06:53:13 crc kubenswrapper[4971]: I1213 06:53:13.484088 4971 status_manager.go:851] "Failed to get status for pod" podUID="9d4f1ea5-8472-4320-820b-8090ff6f5ca6" pod="openshift-marketplace/redhat-marketplace-mkqps" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/redhat-marketplace-mkqps\": dial tcp 38.102.83.195:6443: connect: connection refused" Dec 13 06:53:13 crc kubenswrapper[4971]: I1213 06:53:13.484388 4971 status_manager.go:851] "Failed to get status for pod" podUID="53fdcbc2-900b-48eb-8be3-cd72940096c3" pod="openshift-kube-apiserver/installer-9-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/installer-9-crc\": dial tcp 38.102.83.195:6443: connect: connection refused" Dec 13 06:53:13 crc kubenswrapper[4971]: I1213 06:53:13.484964 4971 status_manager.go:851] "Failed to get status for pod" podUID="b9a0bc5d-3a39-4732-a849-22d351a4cc18" pod="openshift-marketplace/redhat-operators-d8n8p" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/redhat-operators-d8n8p\": dial tcp 38.102.83.195:6443: connect: connection refused" Dec 13 06:53:13 crc kubenswrapper[4971]: I1213 06:53:13.485127 4971 status_manager.go:851] "Failed to get status for pod" podUID="f85e55b1a89d02b0cb034b1ea31ed45a" pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-startup-monitor-crc\": dial tcp 38.102.83.195:6443: connect: connection refused" Dec 13 06:53:13 crc kubenswrapper[4971]: I1213 06:53:13.485510 4971 status_manager.go:851] "Failed to get status for pod" podUID="f14c8a94-8589-4059-bddf-329452355fab" pod="openshift-marketplace/marketplace-operator-79b997595-j9zqf" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/marketplace-operator-79b997595-j9zqf\": dial tcp 38.102.83.195:6443: connect: connection refused" Dec 13 06:53:13 crc kubenswrapper[4971]: I1213 06:53:13.486400 4971 status_manager.go:851] "Failed to get status for pod" podUID="cf417851-ae43-462f-a678-125593fb653b" pod="openshift-marketplace/community-operators-997kc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/community-operators-997kc\": dial tcp 38.102.83.195:6443: connect: connection refused" Dec 13 06:53:13 crc kubenswrapper[4971]: I1213 06:53:13.774429 4971 status_manager.go:851] "Failed to get status for pod" podUID="f4b27818a5e8e43d0dc095d08835c792" pod="openshift-kube-apiserver/kube-apiserver-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\": dial tcp 38.102.83.195:6443: connect: connection refused" Dec 13 06:53:13 crc kubenswrapper[4971]: I1213 06:53:13.775184 4971 status_manager.go:851] "Failed to get status for pod" podUID="9d4f1ea5-8472-4320-820b-8090ff6f5ca6" pod="openshift-marketplace/redhat-marketplace-mkqps" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/redhat-marketplace-mkqps\": dial tcp 38.102.83.195:6443: connect: connection refused" Dec 13 06:53:13 crc kubenswrapper[4971]: I1213 06:53:13.775892 4971 status_manager.go:851] "Failed to get status for pod" podUID="53fdcbc2-900b-48eb-8be3-cd72940096c3" pod="openshift-kube-apiserver/installer-9-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/installer-9-crc\": dial tcp 38.102.83.195:6443: connect: connection refused" Dec 13 06:53:13 crc kubenswrapper[4971]: I1213 06:53:13.777122 4971 status_manager.go:851] "Failed to get status for pod" podUID="f14c8a94-8589-4059-bddf-329452355fab" pod="openshift-marketplace/marketplace-operator-79b997595-j9zqf" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/marketplace-operator-79b997595-j9zqf\": dial tcp 38.102.83.195:6443: connect: connection refused" Dec 13 06:53:13 crc kubenswrapper[4971]: I1213 06:53:13.777928 4971 status_manager.go:851] "Failed to get status for pod" podUID="b9a0bc5d-3a39-4732-a849-22d351a4cc18" pod="openshift-marketplace/redhat-operators-d8n8p" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/redhat-operators-d8n8p\": dial tcp 38.102.83.195:6443: connect: connection refused" Dec 13 06:53:13 crc kubenswrapper[4971]: I1213 06:53:13.778554 4971 status_manager.go:851] "Failed to get status for pod" podUID="f85e55b1a89d02b0cb034b1ea31ed45a" pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-startup-monitor-crc\": dial tcp 38.102.83.195:6443: connect: connection refused" Dec 13 06:53:13 crc kubenswrapper[4971]: I1213 06:53:13.779009 4971 status_manager.go:851] "Failed to get status for pod" podUID="cf417851-ae43-462f-a678-125593fb653b" pod="openshift-marketplace/community-operators-997kc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/community-operators-997kc\": dial tcp 38.102.83.195:6443: connect: connection refused" Dec 13 06:53:13 crc kubenswrapper[4971]: I1213 06:53:13.781089 4971 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="f4b27818a5e8e43d0dc095d08835c792" path="/var/lib/kubelet/pods/f4b27818a5e8e43d0dc095d08835c792/volumes" Dec 13 06:53:13 crc kubenswrapper[4971]: I1213 06:53:13.781203 4971 status_manager.go:851] "Failed to get status for pod" podUID="8f50f5ba-3ec3-4fa2-9f5d-51e1c43f6352" pod="openshift-marketplace/certified-operators-rkd9c" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/certified-operators-rkd9c\": dial tcp 38.102.83.195:6443: connect: connection refused" Dec 13 06:53:13 crc kubenswrapper[4971]: I1213 06:53:13.781559 4971 status_manager.go:851] "Failed to get status for pod" podUID="22037bdc-fe6d-4411-be2b-aec38524a8dd" pod="openshift-marketplace/redhat-marketplace-xwxlf" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/redhat-marketplace-xwxlf\": dial tcp 38.102.83.195:6443: connect: connection refused" Dec 13 06:53:13 crc kubenswrapper[4971]: I1213 06:53:13.782175 4971 status_manager.go:851] "Failed to get status for pod" podUID="02f6e9f2-6bad-441c-8e2e-9acfc71f3ab4" pod="openshift-marketplace/redhat-operators-wpxfk" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/redhat-operators-wpxfk\": dial tcp 38.102.83.195:6443: connect: connection refused" Dec 13 06:53:13 crc kubenswrapper[4971]: I1213 06:53:13.782349 4971 status_manager.go:851] "Failed to get status for pod" podUID="5c88a864-89a9-4f5e-b713-385170b9bf17" pod="openshift-marketplace/marketplace-operator-79b997595-6rvc9" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/marketplace-operator-79b997595-6rvc9\": dial tcp 38.102.83.195:6443: connect: connection refused" Dec 13 06:53:14 crc kubenswrapper[4971]: E1213 06:53:14.387146 4971 controller.go:195] "Failed to update lease" err="Put \"https://api-int.crc.testing:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/crc?timeout=10s\": dial tcp 38.102.83.195:6443: connect: connection refused" Dec 13 06:53:14 crc kubenswrapper[4971]: E1213 06:53:14.387638 4971 controller.go:195] "Failed to update lease" err="Put \"https://api-int.crc.testing:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/crc?timeout=10s\": dial tcp 38.102.83.195:6443: connect: connection refused" Dec 13 06:53:14 crc kubenswrapper[4971]: E1213 06:53:14.387993 4971 controller.go:195] "Failed to update lease" err="Put \"https://api-int.crc.testing:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/crc?timeout=10s\": dial tcp 38.102.83.195:6443: connect: connection refused" Dec 13 06:53:14 crc kubenswrapper[4971]: E1213 06:53:14.388380 4971 controller.go:195] "Failed to update lease" err="Put \"https://api-int.crc.testing:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/crc?timeout=10s\": dial tcp 38.102.83.195:6443: connect: connection refused" Dec 13 06:53:14 crc kubenswrapper[4971]: E1213 06:53:14.388801 4971 controller.go:195] "Failed to update lease" err="Put \"https://api-int.crc.testing:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/crc?timeout=10s\": dial tcp 38.102.83.195:6443: connect: connection refused" Dec 13 06:53:14 crc kubenswrapper[4971]: I1213 06:53:14.388853 4971 controller.go:115] "failed to update lease using latest lease, fallback to ensure lease" err="failed 5 attempts to update lease" Dec 13 06:53:14 crc kubenswrapper[4971]: E1213 06:53:14.389103 4971 controller.go:145] "Failed to ensure lease exists, will retry" err="Get \"https://api-int.crc.testing:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/crc?timeout=10s\": dial tcp 38.102.83.195:6443: connect: connection refused" interval="200ms" Dec 13 06:53:14 crc kubenswrapper[4971]: E1213 06:53:14.590396 4971 controller.go:145] "Failed to ensure lease exists, will retry" err="Get \"https://api-int.crc.testing:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/crc?timeout=10s\": dial tcp 38.102.83.195:6443: connect: connection refused" interval="400ms" Dec 13 06:53:14 crc kubenswrapper[4971]: E1213 06:53:14.991590 4971 controller.go:145] "Failed to ensure lease exists, will retry" err="Get \"https://api-int.crc.testing:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/crc?timeout=10s\": dial tcp 38.102.83.195:6443: connect: connection refused" interval="800ms" Dec 13 06:53:15 crc kubenswrapper[4971]: E1213 06:53:15.793412 4971 controller.go:145] "Failed to ensure lease exists, will retry" err="Get \"https://api-int.crc.testing:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/crc?timeout=10s\": dial tcp 38.102.83.195:6443: connect: connection refused" interval="1.6s" Dec 13 06:53:16 crc kubenswrapper[4971]: I1213 06:53:16.868573 4971 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-marketplace/marketplace-operator-79b997595-6rvc9" Dec 13 06:53:16 crc kubenswrapper[4971]: I1213 06:53:16.868947 4971 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/marketplace-operator-79b997595-6rvc9" Dec 13 06:53:16 crc kubenswrapper[4971]: I1213 06:53:16.869612 4971 scope.go:117] "RemoveContainer" containerID="eec75e44095b014d79821c2ca91b51f25998f1dd4330e3e1bdc38a006708a436" Dec 13 06:53:16 crc kubenswrapper[4971]: E1213 06:53:16.869840 4971 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"marketplace-operator\" with CrashLoopBackOff: \"back-off 10s restarting failed container=marketplace-operator pod=marketplace-operator-79b997595-6rvc9_openshift-marketplace(5c88a864-89a9-4f5e-b713-385170b9bf17)\"" pod="openshift-marketplace/marketplace-operator-79b997595-6rvc9" podUID="5c88a864-89a9-4f5e-b713-385170b9bf17" Dec 13 06:53:17 crc kubenswrapper[4971]: E1213 06:53:17.395415 4971 controller.go:145] "Failed to ensure lease exists, will retry" err="Get \"https://api-int.crc.testing:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/crc?timeout=10s\": dial tcp 38.102.83.195:6443: connect: connection refused" interval="3.2s" Dec 13 06:53:20 crc kubenswrapper[4971]: E1213 06:53:20.251378 4971 event.go:368] "Unable to write event (may retry after sleeping)" err="Post \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/events\": dial tcp 38.102.83.195:6443: connect: connection refused" event="&Event{ObjectMeta:{kube-apiserver-startup-monitor-crc.1880b3d713eba826 openshift-kube-apiserver 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-kube-apiserver,Name:kube-apiserver-startup-monitor-crc,UID:f85e55b1a89d02b0cb034b1ea31ed45a,APIVersion:v1,ResourceVersion:,FieldPath:spec.containers{startup-monitor},},Reason:Created,Message:Created container startup-monitor,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2025-12-13 06:53:10.262786086 +0000 UTC m=+246.867195554,LastTimestamp:2025-12-13 06:53:10.262786086 +0000 UTC m=+246.867195554,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Dec 13 06:53:20 crc kubenswrapper[4971]: E1213 06:53:20.596721 4971 controller.go:145] "Failed to ensure lease exists, will retry" err="Get \"https://api-int.crc.testing:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/crc?timeout=10s\": dial tcp 38.102.83.195:6443: connect: connection refused" interval="6.4s" Dec 13 06:53:20 crc kubenswrapper[4971]: I1213 06:53:20.767934 4971 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/kube-apiserver-crc" Dec 13 06:53:20 crc kubenswrapper[4971]: I1213 06:53:20.769156 4971 status_manager.go:851] "Failed to get status for pod" podUID="8f50f5ba-3ec3-4fa2-9f5d-51e1c43f6352" pod="openshift-marketplace/certified-operators-rkd9c" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/certified-operators-rkd9c\": dial tcp 38.102.83.195:6443: connect: connection refused" Dec 13 06:53:20 crc kubenswrapper[4971]: I1213 06:53:20.769947 4971 status_manager.go:851] "Failed to get status for pod" podUID="22037bdc-fe6d-4411-be2b-aec38524a8dd" pod="openshift-marketplace/redhat-marketplace-xwxlf" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/redhat-marketplace-xwxlf\": dial tcp 38.102.83.195:6443: connect: connection refused" Dec 13 06:53:20 crc kubenswrapper[4971]: I1213 06:53:20.770416 4971 status_manager.go:851] "Failed to get status for pod" podUID="02f6e9f2-6bad-441c-8e2e-9acfc71f3ab4" pod="openshift-marketplace/redhat-operators-wpxfk" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/redhat-operators-wpxfk\": dial tcp 38.102.83.195:6443: connect: connection refused" Dec 13 06:53:20 crc kubenswrapper[4971]: I1213 06:53:20.770904 4971 status_manager.go:851] "Failed to get status for pod" podUID="5c88a864-89a9-4f5e-b713-385170b9bf17" pod="openshift-marketplace/marketplace-operator-79b997595-6rvc9" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/marketplace-operator-79b997595-6rvc9\": dial tcp 38.102.83.195:6443: connect: connection refused" Dec 13 06:53:20 crc kubenswrapper[4971]: I1213 06:53:20.771424 4971 status_manager.go:851] "Failed to get status for pod" podUID="9d4f1ea5-8472-4320-820b-8090ff6f5ca6" pod="openshift-marketplace/redhat-marketplace-mkqps" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/redhat-marketplace-mkqps\": dial tcp 38.102.83.195:6443: connect: connection refused" Dec 13 06:53:20 crc kubenswrapper[4971]: I1213 06:53:20.771964 4971 status_manager.go:851] "Failed to get status for pod" podUID="53fdcbc2-900b-48eb-8be3-cd72940096c3" pod="openshift-kube-apiserver/installer-9-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/installer-9-crc\": dial tcp 38.102.83.195:6443: connect: connection refused" Dec 13 06:53:20 crc kubenswrapper[4971]: I1213 06:53:20.772495 4971 status_manager.go:851] "Failed to get status for pod" podUID="f14c8a94-8589-4059-bddf-329452355fab" pod="openshift-marketplace/marketplace-operator-79b997595-j9zqf" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/marketplace-operator-79b997595-j9zqf\": dial tcp 38.102.83.195:6443: connect: connection refused" Dec 13 06:53:20 crc kubenswrapper[4971]: I1213 06:53:20.773172 4971 status_manager.go:851] "Failed to get status for pod" podUID="b9a0bc5d-3a39-4732-a849-22d351a4cc18" pod="openshift-marketplace/redhat-operators-d8n8p" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/redhat-operators-d8n8p\": dial tcp 38.102.83.195:6443: connect: connection refused" Dec 13 06:53:20 crc kubenswrapper[4971]: I1213 06:53:20.773399 4971 status_manager.go:851] "Failed to get status for pod" podUID="f85e55b1a89d02b0cb034b1ea31ed45a" pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-startup-monitor-crc\": dial tcp 38.102.83.195:6443: connect: connection refused" Dec 13 06:53:20 crc kubenswrapper[4971]: I1213 06:53:20.773685 4971 status_manager.go:851] "Failed to get status for pod" podUID="cf417851-ae43-462f-a678-125593fb653b" pod="openshift-marketplace/community-operators-997kc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/community-operators-997kc\": dial tcp 38.102.83.195:6443: connect: connection refused" Dec 13 06:53:20 crc kubenswrapper[4971]: I1213 06:53:20.784782 4971 kubelet.go:1909] "Trying to delete pod" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="3ad3268a-a982-428c-9c06-4aaeddd372d9" Dec 13 06:53:20 crc kubenswrapper[4971]: I1213 06:53:20.784826 4971 mirror_client.go:130] "Deleting a mirror pod" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="3ad3268a-a982-428c-9c06-4aaeddd372d9" Dec 13 06:53:20 crc kubenswrapper[4971]: E1213 06:53:20.785391 4971 mirror_client.go:138] "Failed deleting a mirror pod" err="Delete \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\": dial tcp 38.102.83.195:6443: connect: connection refused" pod="openshift-kube-apiserver/kube-apiserver-crc" Dec 13 06:53:20 crc kubenswrapper[4971]: I1213 06:53:20.785963 4971 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/kube-apiserver-crc" Dec 13 06:53:20 crc kubenswrapper[4971]: W1213 06:53:20.811785 4971 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod71bb4a3aecc4ba5b26c4b7318770ce13.slice/crio-e7fbc9915d0d06e9da832313900d43c43a82fd6e0184671809fac6a35f9b7060 WatchSource:0}: Error finding container e7fbc9915d0d06e9da832313900d43c43a82fd6e0184671809fac6a35f9b7060: Status 404 returned error can't find the container with id e7fbc9915d0d06e9da832313900d43c43a82fd6e0184671809fac6a35f9b7060 Dec 13 06:53:21 crc kubenswrapper[4971]: I1213 06:53:21.526257 4971 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"71bb4a3aecc4ba5b26c4b7318770ce13","Type":"ContainerStarted","Data":"e7fbc9915d0d06e9da832313900d43c43a82fd6e0184671809fac6a35f9b7060"} Dec 13 06:53:22 crc kubenswrapper[4971]: I1213 06:53:22.289972 4971 patch_prober.go:28] interesting pod/kube-controller-manager-crc container/kube-controller-manager namespace/openshift-kube-controller-manager: Liveness probe status=failure output="Get \"https://192.168.126.11:10257/healthz\": dial tcp 192.168.126.11:10257: connect: connection refused" start-of-body= Dec 13 06:53:22 crc kubenswrapper[4971]: I1213 06:53:22.290034 4971 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-kube-controller-manager/kube-controller-manager-crc" podUID="f614b9022728cf315e60c057852e563e" containerName="kube-controller-manager" probeResult="failure" output="Get \"https://192.168.126.11:10257/healthz\": dial tcp 192.168.126.11:10257: connect: connection refused" Dec 13 06:53:22 crc kubenswrapper[4971]: I1213 06:53:22.532743 4971 generic.go:334] "Generic (PLEG): container finished" podID="71bb4a3aecc4ba5b26c4b7318770ce13" containerID="d333ac093a8d2516fc90ecd3b2a3079d849abec5f23fd67df88347e6c4f2eb62" exitCode=0 Dec 13 06:53:22 crc kubenswrapper[4971]: I1213 06:53:22.532831 4971 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"71bb4a3aecc4ba5b26c4b7318770ce13","Type":"ContainerDied","Data":"d333ac093a8d2516fc90ecd3b2a3079d849abec5f23fd67df88347e6c4f2eb62"} Dec 13 06:53:22 crc kubenswrapper[4971]: I1213 06:53:22.533016 4971 kubelet.go:1909] "Trying to delete pod" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="3ad3268a-a982-428c-9c06-4aaeddd372d9" Dec 13 06:53:22 crc kubenswrapper[4971]: I1213 06:53:22.533033 4971 mirror_client.go:130] "Deleting a mirror pod" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="3ad3268a-a982-428c-9c06-4aaeddd372d9" Dec 13 06:53:22 crc kubenswrapper[4971]: I1213 06:53:22.533390 4971 status_manager.go:851] "Failed to get status for pod" podUID="cf417851-ae43-462f-a678-125593fb653b" pod="openshift-marketplace/community-operators-997kc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/community-operators-997kc\": dial tcp 38.102.83.195:6443: connect: connection refused" Dec 13 06:53:22 crc kubenswrapper[4971]: E1213 06:53:22.533750 4971 mirror_client.go:138] "Failed deleting a mirror pod" err="Delete \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\": dial tcp 38.102.83.195:6443: connect: connection refused" pod="openshift-kube-apiserver/kube-apiserver-crc" Dec 13 06:53:22 crc kubenswrapper[4971]: I1213 06:53:22.533786 4971 status_manager.go:851] "Failed to get status for pod" podUID="22037bdc-fe6d-4411-be2b-aec38524a8dd" pod="openshift-marketplace/redhat-marketplace-xwxlf" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/redhat-marketplace-xwxlf\": dial tcp 38.102.83.195:6443: connect: connection refused" Dec 13 06:53:22 crc kubenswrapper[4971]: I1213 06:53:22.534192 4971 status_manager.go:851] "Failed to get status for pod" podUID="02f6e9f2-6bad-441c-8e2e-9acfc71f3ab4" pod="openshift-marketplace/redhat-operators-wpxfk" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/redhat-operators-wpxfk\": dial tcp 38.102.83.195:6443: connect: connection refused" Dec 13 06:53:22 crc kubenswrapper[4971]: I1213 06:53:22.534417 4971 status_manager.go:851] "Failed to get status for pod" podUID="5c88a864-89a9-4f5e-b713-385170b9bf17" pod="openshift-marketplace/marketplace-operator-79b997595-6rvc9" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/marketplace-operator-79b997595-6rvc9\": dial tcp 38.102.83.195:6443: connect: connection refused" Dec 13 06:53:22 crc kubenswrapper[4971]: I1213 06:53:22.534691 4971 status_manager.go:851] "Failed to get status for pod" podUID="8f50f5ba-3ec3-4fa2-9f5d-51e1c43f6352" pod="openshift-marketplace/certified-operators-rkd9c" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/certified-operators-rkd9c\": dial tcp 38.102.83.195:6443: connect: connection refused" Dec 13 06:53:22 crc kubenswrapper[4971]: I1213 06:53:22.534982 4971 status_manager.go:851] "Failed to get status for pod" podUID="9d4f1ea5-8472-4320-820b-8090ff6f5ca6" pod="openshift-marketplace/redhat-marketplace-mkqps" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/redhat-marketplace-mkqps\": dial tcp 38.102.83.195:6443: connect: connection refused" Dec 13 06:53:22 crc kubenswrapper[4971]: I1213 06:53:22.535440 4971 status_manager.go:851] "Failed to get status for pod" podUID="53fdcbc2-900b-48eb-8be3-cd72940096c3" pod="openshift-kube-apiserver/installer-9-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/installer-9-crc\": dial tcp 38.102.83.195:6443: connect: connection refused" Dec 13 06:53:22 crc kubenswrapper[4971]: I1213 06:53:22.535807 4971 status_manager.go:851] "Failed to get status for pod" podUID="b9a0bc5d-3a39-4732-a849-22d351a4cc18" pod="openshift-marketplace/redhat-operators-d8n8p" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/redhat-operators-d8n8p\": dial tcp 38.102.83.195:6443: connect: connection refused" Dec 13 06:53:22 crc kubenswrapper[4971]: I1213 06:53:22.536143 4971 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-kube-controller-manager_kube-controller-manager-crc_f614b9022728cf315e60c057852e563e/kube-controller-manager/0.log" Dec 13 06:53:22 crc kubenswrapper[4971]: I1213 06:53:22.536188 4971 generic.go:334] "Generic (PLEG): container finished" podID="f614b9022728cf315e60c057852e563e" containerID="c788de86754f0bf37ee184672fc1dcb388ccebc246d4f15be4b5da661f725f8f" exitCode=1 Dec 13 06:53:22 crc kubenswrapper[4971]: I1213 06:53:22.536205 4971 status_manager.go:851] "Failed to get status for pod" podUID="f85e55b1a89d02b0cb034b1ea31ed45a" pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-startup-monitor-crc\": dial tcp 38.102.83.195:6443: connect: connection refused" Dec 13 06:53:22 crc kubenswrapper[4971]: I1213 06:53:22.536247 4971 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" event={"ID":"f614b9022728cf315e60c057852e563e","Type":"ContainerDied","Data":"c788de86754f0bf37ee184672fc1dcb388ccebc246d4f15be4b5da661f725f8f"} Dec 13 06:53:22 crc kubenswrapper[4971]: I1213 06:53:22.537068 4971 status_manager.go:851] "Failed to get status for pod" podUID="f14c8a94-8589-4059-bddf-329452355fab" pod="openshift-marketplace/marketplace-operator-79b997595-j9zqf" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/marketplace-operator-79b997595-j9zqf\": dial tcp 38.102.83.195:6443: connect: connection refused" Dec 13 06:53:22 crc kubenswrapper[4971]: I1213 06:53:22.537373 4971 scope.go:117] "RemoveContainer" containerID="c788de86754f0bf37ee184672fc1dcb388ccebc246d4f15be4b5da661f725f8f" Dec 13 06:53:22 crc kubenswrapper[4971]: I1213 06:53:22.537469 4971 status_manager.go:851] "Failed to get status for pod" podUID="02f6e9f2-6bad-441c-8e2e-9acfc71f3ab4" pod="openshift-marketplace/redhat-operators-wpxfk" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/redhat-operators-wpxfk\": dial tcp 38.102.83.195:6443: connect: connection refused" Dec 13 06:53:22 crc kubenswrapper[4971]: I1213 06:53:22.537788 4971 status_manager.go:851] "Failed to get status for pod" podUID="5c88a864-89a9-4f5e-b713-385170b9bf17" pod="openshift-marketplace/marketplace-operator-79b997595-6rvc9" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/marketplace-operator-79b997595-6rvc9\": dial tcp 38.102.83.195:6443: connect: connection refused" Dec 13 06:53:22 crc kubenswrapper[4971]: I1213 06:53:22.538318 4971 status_manager.go:851] "Failed to get status for pod" podUID="8f50f5ba-3ec3-4fa2-9f5d-51e1c43f6352" pod="openshift-marketplace/certified-operators-rkd9c" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/certified-operators-rkd9c\": dial tcp 38.102.83.195:6443: connect: connection refused" Dec 13 06:53:22 crc kubenswrapper[4971]: I1213 06:53:22.539017 4971 status_manager.go:851] "Failed to get status for pod" podUID="22037bdc-fe6d-4411-be2b-aec38524a8dd" pod="openshift-marketplace/redhat-marketplace-xwxlf" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/redhat-marketplace-xwxlf\": dial tcp 38.102.83.195:6443: connect: connection refused" Dec 13 06:53:22 crc kubenswrapper[4971]: I1213 06:53:22.540088 4971 status_manager.go:851] "Failed to get status for pod" podUID="9d4f1ea5-8472-4320-820b-8090ff6f5ca6" pod="openshift-marketplace/redhat-marketplace-mkqps" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/redhat-marketplace-mkqps\": dial tcp 38.102.83.195:6443: connect: connection refused" Dec 13 06:53:22 crc kubenswrapper[4971]: I1213 06:53:22.540857 4971 status_manager.go:851] "Failed to get status for pod" podUID="53fdcbc2-900b-48eb-8be3-cd72940096c3" pod="openshift-kube-apiserver/installer-9-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/installer-9-crc\": dial tcp 38.102.83.195:6443: connect: connection refused" Dec 13 06:53:22 crc kubenswrapper[4971]: I1213 06:53:22.541329 4971 status_manager.go:851] "Failed to get status for pod" podUID="f614b9022728cf315e60c057852e563e" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-controller-manager/pods/kube-controller-manager-crc\": dial tcp 38.102.83.195:6443: connect: connection refused" Dec 13 06:53:22 crc kubenswrapper[4971]: I1213 06:53:22.541673 4971 status_manager.go:851] "Failed to get status for pod" podUID="b9a0bc5d-3a39-4732-a849-22d351a4cc18" pod="openshift-marketplace/redhat-operators-d8n8p" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/redhat-operators-d8n8p\": dial tcp 38.102.83.195:6443: connect: connection refused" Dec 13 06:53:22 crc kubenswrapper[4971]: I1213 06:53:22.542127 4971 status_manager.go:851] "Failed to get status for pod" podUID="f85e55b1a89d02b0cb034b1ea31ed45a" pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-startup-monitor-crc\": dial tcp 38.102.83.195:6443: connect: connection refused" Dec 13 06:53:22 crc kubenswrapper[4971]: I1213 06:53:22.542441 4971 status_manager.go:851] "Failed to get status for pod" podUID="f14c8a94-8589-4059-bddf-329452355fab" pod="openshift-marketplace/marketplace-operator-79b997595-j9zqf" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/marketplace-operator-79b997595-j9zqf\": dial tcp 38.102.83.195:6443: connect: connection refused" Dec 13 06:53:22 crc kubenswrapper[4971]: I1213 06:53:22.542737 4971 status_manager.go:851] "Failed to get status for pod" podUID="cf417851-ae43-462f-a678-125593fb653b" pod="openshift-marketplace/community-operators-997kc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/community-operators-997kc\": dial tcp 38.102.83.195:6443: connect: connection refused" Dec 13 06:53:23 crc kubenswrapper[4971]: I1213 06:53:23.558592 4971 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"71bb4a3aecc4ba5b26c4b7318770ce13","Type":"ContainerStarted","Data":"1991f4205ad4b510ae0bcc1106d1ae3d9f8c1207c718b2b46ebedcab98b46317"} Dec 13 06:53:23 crc kubenswrapper[4971]: I1213 06:53:23.558945 4971 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"71bb4a3aecc4ba5b26c4b7318770ce13","Type":"ContainerStarted","Data":"8c7c35ea94b8845bd25cf131ad41a81e1fb4a426112ecd6d06b81a27de26a215"} Dec 13 06:53:23 crc kubenswrapper[4971]: I1213 06:53:23.558960 4971 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"71bb4a3aecc4ba5b26c4b7318770ce13","Type":"ContainerStarted","Data":"ce6a3fdc13813e2d58b745bdbffdeab41df53ecdc9dd2b2bad89263dd1660cd0"} Dec 13 06:53:23 crc kubenswrapper[4971]: I1213 06:53:23.558976 4971 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-kube-apiserver/kube-apiserver-crc" Dec 13 06:53:23 crc kubenswrapper[4971]: I1213 06:53:23.558985 4971 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"71bb4a3aecc4ba5b26c4b7318770ce13","Type":"ContainerStarted","Data":"6d8844875d01c5771cbbab260af3e35bb73bbfaba187f6eef6534e4a94f4231b"} Dec 13 06:53:23 crc kubenswrapper[4971]: I1213 06:53:23.558995 4971 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"71bb4a3aecc4ba5b26c4b7318770ce13","Type":"ContainerStarted","Data":"87ace4b5fac9a59ff221f89a90c064f6ff00bdf554d6e0890c8ddbe5062312c3"} Dec 13 06:53:23 crc kubenswrapper[4971]: I1213 06:53:23.558876 4971 kubelet.go:1909] "Trying to delete pod" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="3ad3268a-a982-428c-9c06-4aaeddd372d9" Dec 13 06:53:23 crc kubenswrapper[4971]: I1213 06:53:23.559017 4971 mirror_client.go:130] "Deleting a mirror pod" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="3ad3268a-a982-428c-9c06-4aaeddd372d9" Dec 13 06:53:23 crc kubenswrapper[4971]: I1213 06:53:23.562384 4971 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-kube-controller-manager_kube-controller-manager-crc_f614b9022728cf315e60c057852e563e/kube-controller-manager/0.log" Dec 13 06:53:23 crc kubenswrapper[4971]: I1213 06:53:23.562425 4971 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" event={"ID":"f614b9022728cf315e60c057852e563e","Type":"ContainerStarted","Data":"65bdca95a7c288de60f6ad348f489507c045b5ffcbae188538f47763c1f891c2"} Dec 13 06:53:25 crc kubenswrapper[4971]: I1213 06:53:25.780548 4971 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-kube-controller-manager/kube-controller-manager-crc" Dec 13 06:53:25 crc kubenswrapper[4971]: I1213 06:53:25.780782 4971 patch_prober.go:28] interesting pod/kube-controller-manager-crc container/kube-controller-manager namespace/openshift-kube-controller-manager: Startup probe status=failure output="Get \"https://192.168.126.11:10257/healthz\": dial tcp 192.168.126.11:10257: connect: connection refused" start-of-body= Dec 13 06:53:25 crc kubenswrapper[4971]: I1213 06:53:25.780969 4971 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-kube-controller-manager/kube-controller-manager-crc" podUID="f614b9022728cf315e60c057852e563e" containerName="kube-controller-manager" probeResult="failure" output="Get \"https://192.168.126.11:10257/healthz\": dial tcp 192.168.126.11:10257: connect: connection refused" Dec 13 06:53:25 crc kubenswrapper[4971]: I1213 06:53:25.787139 4971 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-kube-apiserver/kube-apiserver-crc" Dec 13 06:53:25 crc kubenswrapper[4971]: I1213 06:53:25.787206 4971 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-kube-apiserver/kube-apiserver-crc" Dec 13 06:53:25 crc kubenswrapper[4971]: I1213 06:53:25.792082 4971 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-kube-apiserver/kube-apiserver-crc" Dec 13 06:53:27 crc kubenswrapper[4971]: I1213 06:53:27.704482 4971 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-kube-controller-manager/kube-controller-manager-crc" Dec 13 06:53:28 crc kubenswrapper[4971]: I1213 06:53:28.769283 4971 scope.go:117] "RemoveContainer" containerID="eec75e44095b014d79821c2ca91b51f25998f1dd4330e3e1bdc38a006708a436" Dec 13 06:53:28 crc kubenswrapper[4971]: I1213 06:53:28.974038 4971 kubelet.go:1914] "Deleted mirror pod because it is outdated" pod="openshift-kube-apiserver/kube-apiserver-crc" Dec 13 06:53:29 crc kubenswrapper[4971]: I1213 06:53:29.253028 4971 status_manager.go:861] "Pod was deleted and then recreated, skipping status update" pod="openshift-kube-apiserver/kube-apiserver-crc" oldPodUID="71bb4a3aecc4ba5b26c4b7318770ce13" podUID="20070b14-54ae-4c85-88d6-c07b7ef3e355" Dec 13 06:53:29 crc kubenswrapper[4971]: I1213 06:53:29.591590 4971 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_marketplace-operator-79b997595-6rvc9_5c88a864-89a9-4f5e-b713-385170b9bf17/marketplace-operator/2.log" Dec 13 06:53:29 crc kubenswrapper[4971]: I1213 06:53:29.592001 4971 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_marketplace-operator-79b997595-6rvc9_5c88a864-89a9-4f5e-b713-385170b9bf17/marketplace-operator/1.log" Dec 13 06:53:29 crc kubenswrapper[4971]: I1213 06:53:29.592041 4971 generic.go:334] "Generic (PLEG): container finished" podID="5c88a864-89a9-4f5e-b713-385170b9bf17" containerID="286c449dc5a9a113f79f93efbaf4070fc4e73a43894b9a9329a6c56a40aef05b" exitCode=1 Dec 13 06:53:29 crc kubenswrapper[4971]: I1213 06:53:29.592307 4971 kubelet.go:1909] "Trying to delete pod" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="3ad3268a-a982-428c-9c06-4aaeddd372d9" Dec 13 06:53:29 crc kubenswrapper[4971]: I1213 06:53:29.592327 4971 mirror_client.go:130] "Deleting a mirror pod" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="3ad3268a-a982-428c-9c06-4aaeddd372d9" Dec 13 06:53:29 crc kubenswrapper[4971]: I1213 06:53:29.592511 4971 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/marketplace-operator-79b997595-6rvc9" event={"ID":"5c88a864-89a9-4f5e-b713-385170b9bf17","Type":"ContainerDied","Data":"286c449dc5a9a113f79f93efbaf4070fc4e73a43894b9a9329a6c56a40aef05b"} Dec 13 06:53:29 crc kubenswrapper[4971]: I1213 06:53:29.592567 4971 scope.go:117] "RemoveContainer" containerID="eec75e44095b014d79821c2ca91b51f25998f1dd4330e3e1bdc38a006708a436" Dec 13 06:53:29 crc kubenswrapper[4971]: I1213 06:53:29.593343 4971 scope.go:117] "RemoveContainer" containerID="286c449dc5a9a113f79f93efbaf4070fc4e73a43894b9a9329a6c56a40aef05b" Dec 13 06:53:29 crc kubenswrapper[4971]: E1213 06:53:29.593680 4971 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"marketplace-operator\" with CrashLoopBackOff: \"back-off 20s restarting failed container=marketplace-operator pod=marketplace-operator-79b997595-6rvc9_openshift-marketplace(5c88a864-89a9-4f5e-b713-385170b9bf17)\"" pod="openshift-marketplace/marketplace-operator-79b997595-6rvc9" podUID="5c88a864-89a9-4f5e-b713-385170b9bf17" Dec 13 06:53:29 crc kubenswrapper[4971]: I1213 06:53:29.596194 4971 status_manager.go:861] "Pod was deleted and then recreated, skipping status update" pod="openshift-kube-apiserver/kube-apiserver-crc" oldPodUID="71bb4a3aecc4ba5b26c4b7318770ce13" podUID="20070b14-54ae-4c85-88d6-c07b7ef3e355" Dec 13 06:53:29 crc kubenswrapper[4971]: I1213 06:53:29.606083 4971 status_manager.go:308] "Container readiness changed before pod has synced" pod="openshift-kube-apiserver/kube-apiserver-crc" containerID="cri-o://87ace4b5fac9a59ff221f89a90c064f6ff00bdf554d6e0890c8ddbe5062312c3" Dec 13 06:53:29 crc kubenswrapper[4971]: I1213 06:53:29.606120 4971 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-kube-apiserver/kube-apiserver-crc" Dec 13 06:53:29 crc kubenswrapper[4971]: I1213 06:53:29.765639 4971 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-authentication/oauth-openshift-558db77b4-nzzbm" podUID="23e16ced-5b54-4a51-a1bc-4d11c907c90a" containerName="oauth-openshift" containerID="cri-o://3fa8695071968f87224b501328ee01062999751e51021bd23f6e31d535f5c42d" gracePeriod=15 Dec 13 06:53:30 crc kubenswrapper[4971]: I1213 06:53:30.075154 4971 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-authentication/oauth-openshift-558db77b4-nzzbm" Dec 13 06:53:30 crc kubenswrapper[4971]: I1213 06:53:30.191983 4971 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/23e16ced-5b54-4a51-a1bc-4d11c907c90a-audit-dir\") pod \"23e16ced-5b54-4a51-a1bc-4d11c907c90a\" (UID: \"23e16ced-5b54-4a51-a1bc-4d11c907c90a\") " Dec 13 06:53:30 crc kubenswrapper[4971]: I1213 06:53:30.192409 4971 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-user-template-error\" (UniqueName: \"kubernetes.io/secret/23e16ced-5b54-4a51-a1bc-4d11c907c90a-v4-0-config-user-template-error\") pod \"23e16ced-5b54-4a51-a1bc-4d11c907c90a\" (UID: \"23e16ced-5b54-4a51-a1bc-4d11c907c90a\") " Dec 13 06:53:30 crc kubenswrapper[4971]: I1213 06:53:30.192440 4971 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-system-cliconfig\" (UniqueName: \"kubernetes.io/configmap/23e16ced-5b54-4a51-a1bc-4d11c907c90a-v4-0-config-system-cliconfig\") pod \"23e16ced-5b54-4a51-a1bc-4d11c907c90a\" (UID: \"23e16ced-5b54-4a51-a1bc-4d11c907c90a\") " Dec 13 06:53:30 crc kubenswrapper[4971]: I1213 06:53:30.192477 4971 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-user-template-provider-selection\" (UniqueName: \"kubernetes.io/secret/23e16ced-5b54-4a51-a1bc-4d11c907c90a-v4-0-config-user-template-provider-selection\") pod \"23e16ced-5b54-4a51-a1bc-4d11c907c90a\" (UID: \"23e16ced-5b54-4a51-a1bc-4d11c907c90a\") " Dec 13 06:53:30 crc kubenswrapper[4971]: I1213 06:53:30.192500 4971 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-system-serving-cert\" (UniqueName: \"kubernetes.io/secret/23e16ced-5b54-4a51-a1bc-4d11c907c90a-v4-0-config-system-serving-cert\") pod \"23e16ced-5b54-4a51-a1bc-4d11c907c90a\" (UID: \"23e16ced-5b54-4a51-a1bc-4d11c907c90a\") " Dec 13 06:53:30 crc kubenswrapper[4971]: I1213 06:53:30.192542 4971 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-system-session\" (UniqueName: \"kubernetes.io/secret/23e16ced-5b54-4a51-a1bc-4d11c907c90a-v4-0-config-system-session\") pod \"23e16ced-5b54-4a51-a1bc-4d11c907c90a\" (UID: \"23e16ced-5b54-4a51-a1bc-4d11c907c90a\") " Dec 13 06:53:30 crc kubenswrapper[4971]: I1213 06:53:30.192738 4971 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-system-ocp-branding-template\" (UniqueName: \"kubernetes.io/secret/23e16ced-5b54-4a51-a1bc-4d11c907c90a-v4-0-config-system-ocp-branding-template\") pod \"23e16ced-5b54-4a51-a1bc-4d11c907c90a\" (UID: \"23e16ced-5b54-4a51-a1bc-4d11c907c90a\") " Dec 13 06:53:30 crc kubenswrapper[4971]: I1213 06:53:30.192740 4971 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/23e16ced-5b54-4a51-a1bc-4d11c907c90a-audit-dir" (OuterVolumeSpecName: "audit-dir") pod "23e16ced-5b54-4a51-a1bc-4d11c907c90a" (UID: "23e16ced-5b54-4a51-a1bc-4d11c907c90a"). InnerVolumeSpecName "audit-dir". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 13 06:53:30 crc kubenswrapper[4971]: I1213 06:53:30.192785 4971 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-system-trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/23e16ced-5b54-4a51-a1bc-4d11c907c90a-v4-0-config-system-trusted-ca-bundle\") pod \"23e16ced-5b54-4a51-a1bc-4d11c907c90a\" (UID: \"23e16ced-5b54-4a51-a1bc-4d11c907c90a\") " Dec 13 06:53:30 crc kubenswrapper[4971]: I1213 06:53:30.192812 4971 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-user-template-login\" (UniqueName: \"kubernetes.io/secret/23e16ced-5b54-4a51-a1bc-4d11c907c90a-v4-0-config-user-template-login\") pod \"23e16ced-5b54-4a51-a1bc-4d11c907c90a\" (UID: \"23e16ced-5b54-4a51-a1bc-4d11c907c90a\") " Dec 13 06:53:30 crc kubenswrapper[4971]: I1213 06:53:30.192853 4971 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-krxlt\" (UniqueName: \"kubernetes.io/projected/23e16ced-5b54-4a51-a1bc-4d11c907c90a-kube-api-access-krxlt\") pod \"23e16ced-5b54-4a51-a1bc-4d11c907c90a\" (UID: \"23e16ced-5b54-4a51-a1bc-4d11c907c90a\") " Dec 13 06:53:30 crc kubenswrapper[4971]: I1213 06:53:30.192891 4971 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-user-idp-0-file-data\" (UniqueName: \"kubernetes.io/secret/23e16ced-5b54-4a51-a1bc-4d11c907c90a-v4-0-config-user-idp-0-file-data\") pod \"23e16ced-5b54-4a51-a1bc-4d11c907c90a\" (UID: \"23e16ced-5b54-4a51-a1bc-4d11c907c90a\") " Dec 13 06:53:30 crc kubenswrapper[4971]: I1213 06:53:30.192925 4971 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/23e16ced-5b54-4a51-a1bc-4d11c907c90a-audit-policies\") pod \"23e16ced-5b54-4a51-a1bc-4d11c907c90a\" (UID: \"23e16ced-5b54-4a51-a1bc-4d11c907c90a\") " Dec 13 06:53:30 crc kubenswrapper[4971]: I1213 06:53:30.192956 4971 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-system-service-ca\" (UniqueName: \"kubernetes.io/configmap/23e16ced-5b54-4a51-a1bc-4d11c907c90a-v4-0-config-system-service-ca\") pod \"23e16ced-5b54-4a51-a1bc-4d11c907c90a\" (UID: \"23e16ced-5b54-4a51-a1bc-4d11c907c90a\") " Dec 13 06:53:30 crc kubenswrapper[4971]: I1213 06:53:30.192991 4971 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-system-router-certs\" (UniqueName: \"kubernetes.io/secret/23e16ced-5b54-4a51-a1bc-4d11c907c90a-v4-0-config-system-router-certs\") pod \"23e16ced-5b54-4a51-a1bc-4d11c907c90a\" (UID: \"23e16ced-5b54-4a51-a1bc-4d11c907c90a\") " Dec 13 06:53:30 crc kubenswrapper[4971]: I1213 06:53:30.193245 4971 reconciler_common.go:293] "Volume detached for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/23e16ced-5b54-4a51-a1bc-4d11c907c90a-audit-dir\") on node \"crc\" DevicePath \"\"" Dec 13 06:53:30 crc kubenswrapper[4971]: I1213 06:53:30.193732 4971 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/23e16ced-5b54-4a51-a1bc-4d11c907c90a-v4-0-config-system-trusted-ca-bundle" (OuterVolumeSpecName: "v4-0-config-system-trusted-ca-bundle") pod "23e16ced-5b54-4a51-a1bc-4d11c907c90a" (UID: "23e16ced-5b54-4a51-a1bc-4d11c907c90a"). InnerVolumeSpecName "v4-0-config-system-trusted-ca-bundle". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 13 06:53:30 crc kubenswrapper[4971]: I1213 06:53:30.193935 4971 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/23e16ced-5b54-4a51-a1bc-4d11c907c90a-v4-0-config-system-cliconfig" (OuterVolumeSpecName: "v4-0-config-system-cliconfig") pod "23e16ced-5b54-4a51-a1bc-4d11c907c90a" (UID: "23e16ced-5b54-4a51-a1bc-4d11c907c90a"). InnerVolumeSpecName "v4-0-config-system-cliconfig". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 13 06:53:30 crc kubenswrapper[4971]: I1213 06:53:30.195535 4971 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/23e16ced-5b54-4a51-a1bc-4d11c907c90a-v4-0-config-system-service-ca" (OuterVolumeSpecName: "v4-0-config-system-service-ca") pod "23e16ced-5b54-4a51-a1bc-4d11c907c90a" (UID: "23e16ced-5b54-4a51-a1bc-4d11c907c90a"). InnerVolumeSpecName "v4-0-config-system-service-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 13 06:53:30 crc kubenswrapper[4971]: I1213 06:53:30.195752 4971 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/23e16ced-5b54-4a51-a1bc-4d11c907c90a-audit-policies" (OuterVolumeSpecName: "audit-policies") pod "23e16ced-5b54-4a51-a1bc-4d11c907c90a" (UID: "23e16ced-5b54-4a51-a1bc-4d11c907c90a"). InnerVolumeSpecName "audit-policies". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 13 06:53:30 crc kubenswrapper[4971]: I1213 06:53:30.199920 4971 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/23e16ced-5b54-4a51-a1bc-4d11c907c90a-kube-api-access-krxlt" (OuterVolumeSpecName: "kube-api-access-krxlt") pod "23e16ced-5b54-4a51-a1bc-4d11c907c90a" (UID: "23e16ced-5b54-4a51-a1bc-4d11c907c90a"). InnerVolumeSpecName "kube-api-access-krxlt". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 13 06:53:30 crc kubenswrapper[4971]: I1213 06:53:30.199949 4971 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/23e16ced-5b54-4a51-a1bc-4d11c907c90a-v4-0-config-system-ocp-branding-template" (OuterVolumeSpecName: "v4-0-config-system-ocp-branding-template") pod "23e16ced-5b54-4a51-a1bc-4d11c907c90a" (UID: "23e16ced-5b54-4a51-a1bc-4d11c907c90a"). InnerVolumeSpecName "v4-0-config-system-ocp-branding-template". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 13 06:53:30 crc kubenswrapper[4971]: I1213 06:53:30.200028 4971 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/23e16ced-5b54-4a51-a1bc-4d11c907c90a-v4-0-config-user-idp-0-file-data" (OuterVolumeSpecName: "v4-0-config-user-idp-0-file-data") pod "23e16ced-5b54-4a51-a1bc-4d11c907c90a" (UID: "23e16ced-5b54-4a51-a1bc-4d11c907c90a"). InnerVolumeSpecName "v4-0-config-user-idp-0-file-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 13 06:53:30 crc kubenswrapper[4971]: I1213 06:53:30.200658 4971 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/23e16ced-5b54-4a51-a1bc-4d11c907c90a-v4-0-config-system-serving-cert" (OuterVolumeSpecName: "v4-0-config-system-serving-cert") pod "23e16ced-5b54-4a51-a1bc-4d11c907c90a" (UID: "23e16ced-5b54-4a51-a1bc-4d11c907c90a"). InnerVolumeSpecName "v4-0-config-system-serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 13 06:53:30 crc kubenswrapper[4971]: I1213 06:53:30.200730 4971 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/23e16ced-5b54-4a51-a1bc-4d11c907c90a-v4-0-config-system-session" (OuterVolumeSpecName: "v4-0-config-system-session") pod "23e16ced-5b54-4a51-a1bc-4d11c907c90a" (UID: "23e16ced-5b54-4a51-a1bc-4d11c907c90a"). InnerVolumeSpecName "v4-0-config-system-session". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 13 06:53:30 crc kubenswrapper[4971]: I1213 06:53:30.200994 4971 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/23e16ced-5b54-4a51-a1bc-4d11c907c90a-v4-0-config-user-template-provider-selection" (OuterVolumeSpecName: "v4-0-config-user-template-provider-selection") pod "23e16ced-5b54-4a51-a1bc-4d11c907c90a" (UID: "23e16ced-5b54-4a51-a1bc-4d11c907c90a"). InnerVolumeSpecName "v4-0-config-user-template-provider-selection". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 13 06:53:30 crc kubenswrapper[4971]: I1213 06:53:30.201141 4971 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/23e16ced-5b54-4a51-a1bc-4d11c907c90a-v4-0-config-system-router-certs" (OuterVolumeSpecName: "v4-0-config-system-router-certs") pod "23e16ced-5b54-4a51-a1bc-4d11c907c90a" (UID: "23e16ced-5b54-4a51-a1bc-4d11c907c90a"). InnerVolumeSpecName "v4-0-config-system-router-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 13 06:53:30 crc kubenswrapper[4971]: I1213 06:53:30.201302 4971 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/23e16ced-5b54-4a51-a1bc-4d11c907c90a-v4-0-config-user-template-login" (OuterVolumeSpecName: "v4-0-config-user-template-login") pod "23e16ced-5b54-4a51-a1bc-4d11c907c90a" (UID: "23e16ced-5b54-4a51-a1bc-4d11c907c90a"). InnerVolumeSpecName "v4-0-config-user-template-login". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 13 06:53:30 crc kubenswrapper[4971]: I1213 06:53:30.201349 4971 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/23e16ced-5b54-4a51-a1bc-4d11c907c90a-v4-0-config-user-template-error" (OuterVolumeSpecName: "v4-0-config-user-template-error") pod "23e16ced-5b54-4a51-a1bc-4d11c907c90a" (UID: "23e16ced-5b54-4a51-a1bc-4d11c907c90a"). InnerVolumeSpecName "v4-0-config-user-template-error". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 13 06:53:30 crc kubenswrapper[4971]: I1213 06:53:30.294853 4971 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-user-template-error\" (UniqueName: \"kubernetes.io/secret/23e16ced-5b54-4a51-a1bc-4d11c907c90a-v4-0-config-user-template-error\") on node \"crc\" DevicePath \"\"" Dec 13 06:53:30 crc kubenswrapper[4971]: I1213 06:53:30.294894 4971 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-system-cliconfig\" (UniqueName: \"kubernetes.io/configmap/23e16ced-5b54-4a51-a1bc-4d11c907c90a-v4-0-config-system-cliconfig\") on node \"crc\" DevicePath \"\"" Dec 13 06:53:30 crc kubenswrapper[4971]: I1213 06:53:30.294908 4971 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-user-template-provider-selection\" (UniqueName: \"kubernetes.io/secret/23e16ced-5b54-4a51-a1bc-4d11c907c90a-v4-0-config-user-template-provider-selection\") on node \"crc\" DevicePath \"\"" Dec 13 06:53:30 crc kubenswrapper[4971]: I1213 06:53:30.294922 4971 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-system-serving-cert\" (UniqueName: \"kubernetes.io/secret/23e16ced-5b54-4a51-a1bc-4d11c907c90a-v4-0-config-system-serving-cert\") on node \"crc\" DevicePath \"\"" Dec 13 06:53:30 crc kubenswrapper[4971]: I1213 06:53:30.294934 4971 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-system-session\" (UniqueName: \"kubernetes.io/secret/23e16ced-5b54-4a51-a1bc-4d11c907c90a-v4-0-config-system-session\") on node \"crc\" DevicePath \"\"" Dec 13 06:53:30 crc kubenswrapper[4971]: I1213 06:53:30.294947 4971 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-system-ocp-branding-template\" (UniqueName: \"kubernetes.io/secret/23e16ced-5b54-4a51-a1bc-4d11c907c90a-v4-0-config-system-ocp-branding-template\") on node \"crc\" DevicePath \"\"" Dec 13 06:53:30 crc kubenswrapper[4971]: I1213 06:53:30.294960 4971 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-system-trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/23e16ced-5b54-4a51-a1bc-4d11c907c90a-v4-0-config-system-trusted-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 13 06:53:30 crc kubenswrapper[4971]: I1213 06:53:30.294971 4971 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-user-template-login\" (UniqueName: \"kubernetes.io/secret/23e16ced-5b54-4a51-a1bc-4d11c907c90a-v4-0-config-user-template-login\") on node \"crc\" DevicePath \"\"" Dec 13 06:53:30 crc kubenswrapper[4971]: I1213 06:53:30.294983 4971 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-krxlt\" (UniqueName: \"kubernetes.io/projected/23e16ced-5b54-4a51-a1bc-4d11c907c90a-kube-api-access-krxlt\") on node \"crc\" DevicePath \"\"" Dec 13 06:53:30 crc kubenswrapper[4971]: I1213 06:53:30.294994 4971 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-user-idp-0-file-data\" (UniqueName: \"kubernetes.io/secret/23e16ced-5b54-4a51-a1bc-4d11c907c90a-v4-0-config-user-idp-0-file-data\") on node \"crc\" DevicePath \"\"" Dec 13 06:53:30 crc kubenswrapper[4971]: I1213 06:53:30.295007 4971 reconciler_common.go:293] "Volume detached for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/23e16ced-5b54-4a51-a1bc-4d11c907c90a-audit-policies\") on node \"crc\" DevicePath \"\"" Dec 13 06:53:30 crc kubenswrapper[4971]: I1213 06:53:30.295020 4971 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-system-service-ca\" (UniqueName: \"kubernetes.io/configmap/23e16ced-5b54-4a51-a1bc-4d11c907c90a-v4-0-config-system-service-ca\") on node \"crc\" DevicePath \"\"" Dec 13 06:53:30 crc kubenswrapper[4971]: I1213 06:53:30.295035 4971 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-system-router-certs\" (UniqueName: \"kubernetes.io/secret/23e16ced-5b54-4a51-a1bc-4d11c907c90a-v4-0-config-system-router-certs\") on node \"crc\" DevicePath \"\"" Dec 13 06:53:30 crc kubenswrapper[4971]: I1213 06:53:30.606198 4971 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_marketplace-operator-79b997595-6rvc9_5c88a864-89a9-4f5e-b713-385170b9bf17/marketplace-operator/2.log" Dec 13 06:53:30 crc kubenswrapper[4971]: I1213 06:53:30.608804 4971 generic.go:334] "Generic (PLEG): container finished" podID="23e16ced-5b54-4a51-a1bc-4d11c907c90a" containerID="3fa8695071968f87224b501328ee01062999751e51021bd23f6e31d535f5c42d" exitCode=0 Dec 13 06:53:30 crc kubenswrapper[4971]: I1213 06:53:30.608900 4971 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-authentication/oauth-openshift-558db77b4-nzzbm" Dec 13 06:53:30 crc kubenswrapper[4971]: I1213 06:53:30.608983 4971 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-authentication/oauth-openshift-558db77b4-nzzbm" event={"ID":"23e16ced-5b54-4a51-a1bc-4d11c907c90a","Type":"ContainerDied","Data":"3fa8695071968f87224b501328ee01062999751e51021bd23f6e31d535f5c42d"} Dec 13 06:53:30 crc kubenswrapper[4971]: I1213 06:53:30.609031 4971 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-authentication/oauth-openshift-558db77b4-nzzbm" event={"ID":"23e16ced-5b54-4a51-a1bc-4d11c907c90a","Type":"ContainerDied","Data":"eae50d9db612b30a2aa3d28acec408f9df99cb8757dc54079b0da3f4449bb56b"} Dec 13 06:53:30 crc kubenswrapper[4971]: I1213 06:53:30.609050 4971 scope.go:117] "RemoveContainer" containerID="3fa8695071968f87224b501328ee01062999751e51021bd23f6e31d535f5c42d" Dec 13 06:53:30 crc kubenswrapper[4971]: I1213 06:53:30.609460 4971 kubelet.go:1909] "Trying to delete pod" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="3ad3268a-a982-428c-9c06-4aaeddd372d9" Dec 13 06:53:30 crc kubenswrapper[4971]: I1213 06:53:30.609483 4971 mirror_client.go:130] "Deleting a mirror pod" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="3ad3268a-a982-428c-9c06-4aaeddd372d9" Dec 13 06:53:30 crc kubenswrapper[4971]: I1213 06:53:30.615332 4971 status_manager.go:861] "Pod was deleted and then recreated, skipping status update" pod="openshift-kube-apiserver/kube-apiserver-crc" oldPodUID="71bb4a3aecc4ba5b26c4b7318770ce13" podUID="20070b14-54ae-4c85-88d6-c07b7ef3e355" Dec 13 06:53:30 crc kubenswrapper[4971]: I1213 06:53:30.630965 4971 scope.go:117] "RemoveContainer" containerID="3fa8695071968f87224b501328ee01062999751e51021bd23f6e31d535f5c42d" Dec 13 06:53:30 crc kubenswrapper[4971]: E1213 06:53:30.631373 4971 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"3fa8695071968f87224b501328ee01062999751e51021bd23f6e31d535f5c42d\": container with ID starting with 3fa8695071968f87224b501328ee01062999751e51021bd23f6e31d535f5c42d not found: ID does not exist" containerID="3fa8695071968f87224b501328ee01062999751e51021bd23f6e31d535f5c42d" Dec 13 06:53:30 crc kubenswrapper[4971]: I1213 06:53:30.631398 4971 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"3fa8695071968f87224b501328ee01062999751e51021bd23f6e31d535f5c42d"} err="failed to get container status \"3fa8695071968f87224b501328ee01062999751e51021bd23f6e31d535f5c42d\": rpc error: code = NotFound desc = could not find container \"3fa8695071968f87224b501328ee01062999751e51021bd23f6e31d535f5c42d\": container with ID starting with 3fa8695071968f87224b501328ee01062999751e51021bd23f6e31d535f5c42d not found: ID does not exist" Dec 13 06:53:35 crc kubenswrapper[4971]: I1213 06:53:35.780360 4971 patch_prober.go:28] interesting pod/kube-controller-manager-crc container/kube-controller-manager namespace/openshift-kube-controller-manager: Startup probe status=failure output="Get \"https://192.168.126.11:10257/healthz\": dial tcp 192.168.126.11:10257: connect: connection refused" start-of-body= Dec 13 06:53:35 crc kubenswrapper[4971]: I1213 06:53:35.781130 4971 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-kube-controller-manager/kube-controller-manager-crc" podUID="f614b9022728cf315e60c057852e563e" containerName="kube-controller-manager" probeResult="failure" output="Get \"https://192.168.126.11:10257/healthz\": dial tcp 192.168.126.11:10257: connect: connection refused" Dec 13 06:53:36 crc kubenswrapper[4971]: I1213 06:53:36.868508 4971 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/marketplace-operator-79b997595-6rvc9" Dec 13 06:53:36 crc kubenswrapper[4971]: I1213 06:53:36.868612 4971 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-marketplace/marketplace-operator-79b997595-6rvc9" Dec 13 06:53:36 crc kubenswrapper[4971]: I1213 06:53:36.869240 4971 scope.go:117] "RemoveContainer" containerID="286c449dc5a9a113f79f93efbaf4070fc4e73a43894b9a9329a6c56a40aef05b" Dec 13 06:53:36 crc kubenswrapper[4971]: E1213 06:53:36.869579 4971 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"marketplace-operator\" with CrashLoopBackOff: \"back-off 20s restarting failed container=marketplace-operator pod=marketplace-operator-79b997595-6rvc9_openshift-marketplace(5c88a864-89a9-4f5e-b713-385170b9bf17)\"" pod="openshift-marketplace/marketplace-operator-79b997595-6rvc9" podUID="5c88a864-89a9-4f5e-b713-385170b9bf17" Dec 13 06:53:39 crc kubenswrapper[4971]: I1213 06:53:39.797877 4971 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-diagnostics"/"kube-root-ca.crt" Dec 13 06:53:39 crc kubenswrapper[4971]: I1213 06:53:39.993844 4971 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-node-identity"/"ovnkube-identity-cm" Dec 13 06:53:40 crc kubenswrapper[4971]: I1213 06:53:40.725129 4971 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-cluster-version"/"openshift-service-ca.crt" Dec 13 06:53:40 crc kubenswrapper[4971]: I1213 06:53:40.979065 4971 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-storage-version-migrator-operator"/"config" Dec 13 06:53:41 crc kubenswrapper[4971]: I1213 06:53:41.023219 4971 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager-operator"/"openshift-controller-manager-operator-config" Dec 13 06:53:41 crc kubenswrapper[4971]: I1213 06:53:41.060377 4971 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-api"/"control-plane-machine-set-operator-dockercfg-k9rxt" Dec 13 06:53:41 crc kubenswrapper[4971]: I1213 06:53:41.061729 4971 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-console-operator"/"serving-cert" Dec 13 06:53:41 crc kubenswrapper[4971]: I1213 06:53:41.116981 4971 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-machine-api"/"kube-root-ca.crt" Dec 13 06:53:41 crc kubenswrapper[4971]: I1213 06:53:41.127680 4971 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"marketplace-operator-dockercfg-5nsgg" Dec 13 06:53:41 crc kubenswrapper[4971]: I1213 06:53:41.134941 4971 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console"/"service-ca" Dec 13 06:53:41 crc kubenswrapper[4971]: I1213 06:53:41.196315 4971 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-diagnostics"/"openshift-service-ca.crt" Dec 13 06:53:41 crc kubenswrapper[4971]: I1213 06:53:41.375219 4971 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication-operator"/"serving-cert" Dec 13 06:53:41 crc kubenswrapper[4971]: I1213 06:53:41.529690 4971 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-cluster-version"/"cluster-version-operator-serving-cert" Dec 13 06:53:41 crc kubenswrapper[4971]: I1213 06:53:41.553428 4971 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-machine-config-operator"/"kube-rbac-proxy" Dec 13 06:53:41 crc kubenswrapper[4971]: I1213 06:53:41.786436 4971 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console"/"kube-root-ca.crt" Dec 13 06:53:41 crc kubenswrapper[4971]: I1213 06:53:41.927967 4971 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ovn-kubernetes"/"ovn-kubernetes-node-dockercfg-pwtwl" Dec 13 06:53:42 crc kubenswrapper[4971]: I1213 06:53:42.065987 4971 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ingress-operator"/"trusted-ca" Dec 13 06:53:42 crc kubenswrapper[4971]: I1213 06:53:42.094257 4971 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-service-ca"/"service-ca-dockercfg-pn86c" Dec 13 06:53:42 crc kubenswrapper[4971]: I1213 06:53:42.154596 4971 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-machine-config-operator"/"kube-root-ca.crt" Dec 13 06:53:42 crc kubenswrapper[4971]: I1213 06:53:42.300415 4971 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-oauth-apiserver"/"openshift-service-ca.crt" Dec 13 06:53:42 crc kubenswrapper[4971]: I1213 06:53:42.364283 4971 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"kube-root-ca.crt" Dec 13 06:53:42 crc kubenswrapper[4971]: I1213 06:53:42.821858 4971 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-console-operator"/"console-operator-dockercfg-4xjcr" Dec 13 06:53:42 crc kubenswrapper[4971]: I1213 06:53:42.934189 4971 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-service-ca-operator"/"openshift-service-ca.crt" Dec 13 06:53:43 crc kubenswrapper[4971]: I1213 06:53:43.035487 4971 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-config-operator"/"openshift-config-operator-dockercfg-7pc5z" Dec 13 06:53:43 crc kubenswrapper[4971]: I1213 06:53:43.083959 4971 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-dns"/"kube-root-ca.crt" Dec 13 06:53:43 crc kubenswrapper[4971]: I1213 06:53:43.203152 4971 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"olm-operator-serving-cert" Dec 13 06:53:43 crc kubenswrapper[4971]: I1213 06:53:43.394807 4971 reflector.go:368] Caches populated for *v1.RuntimeClass from k8s.io/client-go/informers/factory.go:160 Dec 13 06:53:43 crc kubenswrapper[4971]: I1213 06:53:43.463202 4971 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver-operator"/"kube-root-ca.crt" Dec 13 06:53:43 crc kubenswrapper[4971]: I1213 06:53:43.486454 4971 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication-operator"/"authentication-operator-config" Dec 13 06:53:43 crc kubenswrapper[4971]: I1213 06:53:43.492051 4971 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ovn-kubernetes"/"openshift-service-ca.crt" Dec 13 06:53:43 crc kubenswrapper[4971]: I1213 06:53:43.560936 4971 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ingress-canary"/"kube-root-ca.crt" Dec 13 06:53:43 crc kubenswrapper[4971]: I1213 06:53:43.623661 4971 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver"/"trusted-ca-bundle" Dec 13 06:53:43 crc kubenswrapper[4971]: I1213 06:53:43.744612 4971 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-cluster-version"/"default-dockercfg-gxtc4" Dec 13 06:53:43 crc kubenswrapper[4971]: I1213 06:53:43.879057 4971 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-image-registry"/"kube-root-ca.crt" Dec 13 06:53:43 crc kubenswrapper[4971]: I1213 06:53:43.901318 4971 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-config-operator"/"config-operator-serving-cert" Dec 13 06:53:44 crc kubenswrapper[4971]: I1213 06:53:44.125280 4971 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-console"/"console-dockercfg-f62pw" Dec 13 06:53:44 crc kubenswrapper[4971]: I1213 06:53:44.130668 4971 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-multus"/"metrics-daemon-secret" Dec 13 06:53:44 crc kubenswrapper[4971]: I1213 06:53:44.171448 4971 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-network-console"/"networking-console-plugin-cert" Dec 13 06:53:44 crc kubenswrapper[4971]: I1213 06:53:44.173715 4971 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-console"/"default-dockercfg-chnjx" Dec 13 06:53:44 crc kubenswrapper[4971]: I1213 06:53:44.203558 4971 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-service-ca-operator"/"serving-cert" Dec 13 06:53:44 crc kubenswrapper[4971]: I1213 06:53:44.224551 4971 reflector.go:368] Caches populated for *v1.Pod from pkg/kubelet/config/apiserver.go:66 Dec 13 06:53:44 crc kubenswrapper[4971]: I1213 06:53:44.226030 4971 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" podStartSLOduration=37.226011966 podStartE2EDuration="37.226011966s" podCreationTimestamp="2025-12-13 06:53:07 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-13 06:53:29.066560947 +0000 UTC m=+265.670970395" watchObservedRunningTime="2025-12-13 06:53:44.226011966 +0000 UTC m=+280.830421424" Dec 13 06:53:44 crc kubenswrapper[4971]: I1213 06:53:44.229489 4971 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"openshift-global-ca" Dec 13 06:53:44 crc kubenswrapper[4971]: I1213 06:53:44.230290 4971 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-authentication/oauth-openshift-558db77b4-nzzbm","openshift-kube-apiserver/kube-apiserver-crc","openshift-marketplace/redhat-operators-wpxfk","openshift-marketplace/community-operators-997kc","openshift-marketplace/marketplace-operator-79b997595-j9zqf","openshift-marketplace/certified-operators-rkd9c","openshift-marketplace/redhat-operators-d8n8p","openshift-marketplace/redhat-marketplace-xwxlf","openshift-marketplace/redhat-marketplace-mkqps"] Dec 13 06:53:44 crc kubenswrapper[4971]: I1213 06:53:44.230412 4971 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-kube-apiserver/kube-apiserver-crc"] Dec 13 06:53:44 crc kubenswrapper[4971]: I1213 06:53:44.236762 4971 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-kube-apiserver/kube-apiserver-crc" Dec 13 06:53:44 crc kubenswrapper[4971]: I1213 06:53:44.251764 4971 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-kube-apiserver/kube-apiserver-crc" podStartSLOduration=16.251732601 podStartE2EDuration="16.251732601s" podCreationTimestamp="2025-12-13 06:53:28 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-13 06:53:44.248102622 +0000 UTC m=+280.852512070" watchObservedRunningTime="2025-12-13 06:53:44.251732601 +0000 UTC m=+280.856142049" Dec 13 06:53:44 crc kubenswrapper[4971]: I1213 06:53:44.320288 4971 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-storage-version-migrator-operator"/"kube-root-ca.crt" Dec 13 06:53:44 crc kubenswrapper[4971]: I1213 06:53:44.375835 4971 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-etcd-operator"/"etcd-operator-serving-cert" Dec 13 06:53:44 crc kubenswrapper[4971]: I1213 06:53:44.518045 4971 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ingress"/"service-ca-bundle" Dec 13 06:53:44 crc kubenswrapper[4971]: I1213 06:53:44.574841 4971 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ovn-kubernetes"/"ovnkube-config" Dec 13 06:53:44 crc kubenswrapper[4971]: I1213 06:53:44.693896 4971 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver"/"image-import-ca" Dec 13 06:53:44 crc kubenswrapper[4971]: I1213 06:53:44.738971 4971 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver-operator"/"openshift-service-ca.crt" Dec 13 06:53:44 crc kubenswrapper[4971]: I1213 06:53:44.816498 4971 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ingress-operator"/"metrics-tls" Dec 13 06:53:44 crc kubenswrapper[4971]: I1213 06:53:44.821812 4971 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-operator"/"kube-root-ca.crt" Dec 13 06:53:44 crc kubenswrapper[4971]: I1213 06:53:44.857417 4971 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-route-controller-manager"/"config" Dec 13 06:53:44 crc kubenswrapper[4971]: I1213 06:53:44.917611 4971 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-oauth-apiserver"/"etcd-serving-ca" Dec 13 06:53:44 crc kubenswrapper[4971]: I1213 06:53:44.985713 4971 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-etcd-operator"/"etcd-client" Dec 13 06:53:44 crc kubenswrapper[4971]: I1213 06:53:44.998488 4971 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"olm-operator-serviceaccount-dockercfg-rq7zk" Dec 13 06:53:45 crc kubenswrapper[4971]: I1213 06:53:45.028132 4971 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver"/"openshift-service-ca.crt" Dec 13 06:53:45 crc kubenswrapper[4971]: I1213 06:53:45.335990 4971 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"machine-config-operator-dockercfg-98p87" Dec 13 06:53:45 crc kubenswrapper[4971]: I1213 06:53:45.348602 4971 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-apiserver-operator"/"kube-root-ca.crt" Dec 13 06:53:45 crc kubenswrapper[4971]: I1213 06:53:45.417612 4971 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-cluster-version"/"kube-root-ca.crt" Dec 13 06:53:45 crc kubenswrapper[4971]: I1213 06:53:45.611064 4971 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-controller-manager-operator"/"openshift-controller-manager-operator-serving-cert" Dec 13 06:53:45 crc kubenswrapper[4971]: I1213 06:53:45.665692 4971 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ovn-kubernetes"/"ovn-kubernetes-control-plane-dockercfg-gs7dd" Dec 13 06:53:45 crc kubenswrapper[4971]: I1213 06:53:45.777119 4971 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="02f6e9f2-6bad-441c-8e2e-9acfc71f3ab4" path="/var/lib/kubelet/pods/02f6e9f2-6bad-441c-8e2e-9acfc71f3ab4/volumes" Dec 13 06:53:45 crc kubenswrapper[4971]: I1213 06:53:45.778367 4971 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="22037bdc-fe6d-4411-be2b-aec38524a8dd" path="/var/lib/kubelet/pods/22037bdc-fe6d-4411-be2b-aec38524a8dd/volumes" Dec 13 06:53:45 crc kubenswrapper[4971]: I1213 06:53:45.779298 4971 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="23e16ced-5b54-4a51-a1bc-4d11c907c90a" path="/var/lib/kubelet/pods/23e16ced-5b54-4a51-a1bc-4d11c907c90a/volumes" Dec 13 06:53:45 crc kubenswrapper[4971]: I1213 06:53:45.780428 4971 patch_prober.go:28] interesting pod/kube-controller-manager-crc container/kube-controller-manager namespace/openshift-kube-controller-manager: Startup probe status=failure output="Get \"https://192.168.126.11:10257/healthz\": dial tcp 192.168.126.11:10257: connect: connection refused" start-of-body= Dec 13 06:53:45 crc kubenswrapper[4971]: I1213 06:53:45.780493 4971 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-kube-controller-manager/kube-controller-manager-crc" podUID="f614b9022728cf315e60c057852e563e" containerName="kube-controller-manager" probeResult="failure" output="Get \"https://192.168.126.11:10257/healthz\": dial tcp 192.168.126.11:10257: connect: connection refused" Dec 13 06:53:45 crc kubenswrapper[4971]: I1213 06:53:45.780804 4971 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="8f50f5ba-3ec3-4fa2-9f5d-51e1c43f6352" path="/var/lib/kubelet/pods/8f50f5ba-3ec3-4fa2-9f5d-51e1c43f6352/volumes" Dec 13 06:53:45 crc kubenswrapper[4971]: I1213 06:53:45.781816 4971 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="9d4f1ea5-8472-4320-820b-8090ff6f5ca6" path="/var/lib/kubelet/pods/9d4f1ea5-8472-4320-820b-8090ff6f5ca6/volumes" Dec 13 06:53:45 crc kubenswrapper[4971]: I1213 06:53:45.783392 4971 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="b9a0bc5d-3a39-4732-a849-22d351a4cc18" path="/var/lib/kubelet/pods/b9a0bc5d-3a39-4732-a849-22d351a4cc18/volumes" Dec 13 06:53:45 crc kubenswrapper[4971]: I1213 06:53:45.784385 4971 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="cf417851-ae43-462f-a678-125593fb653b" path="/var/lib/kubelet/pods/cf417851-ae43-462f-a678-125593fb653b/volumes" Dec 13 06:53:45 crc kubenswrapper[4971]: I1213 06:53:45.785402 4971 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="f14c8a94-8589-4059-bddf-329452355fab" path="/var/lib/kubelet/pods/f14c8a94-8589-4059-bddf-329452355fab/volumes" Dec 13 06:53:45 crc kubenswrapper[4971]: I1213 06:53:45.787318 4971 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-kube-controller-manager/kube-controller-manager-crc" Dec 13 06:53:45 crc kubenswrapper[4971]: I1213 06:53:45.787995 4971 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="kube-controller-manager" containerStatusID={"Type":"cri-o","ID":"65bdca95a7c288de60f6ad348f489507c045b5ffcbae188538f47763c1f891c2"} pod="openshift-kube-controller-manager/kube-controller-manager-crc" containerMessage="Container kube-controller-manager failed startup probe, will be restarted" Dec 13 06:53:45 crc kubenswrapper[4971]: I1213 06:53:45.788163 4971 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-kube-controller-manager/kube-controller-manager-crc" podUID="f614b9022728cf315e60c057852e563e" containerName="kube-controller-manager" containerID="cri-o://65bdca95a7c288de60f6ad348f489507c045b5ffcbae188538f47763c1f891c2" gracePeriod=30 Dec 13 06:53:45 crc kubenswrapper[4971]: I1213 06:53:45.889649 4971 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication-operator"/"service-ca-bundle" Dec 13 06:53:46 crc kubenswrapper[4971]: I1213 06:53:45.917427 4971 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication-operator"/"trusted-ca-bundle" Dec 13 06:53:46 crc kubenswrapper[4971]: I1213 06:53:46.009153 4971 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-oauth-apiserver"/"kube-root-ca.crt" Dec 13 06:53:46 crc kubenswrapper[4971]: I1213 06:53:46.133107 4971 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-service-ca-operator"/"service-ca-operator-config" Dec 13 06:53:46 crc kubenswrapper[4971]: I1213 06:53:46.142984 4971 reflector.go:368] Caches populated for *v1.Secret from object-"hostpath-provisioner"/"csi-hostpath-provisioner-sa-dockercfg-qd74k" Dec 13 06:53:46 crc kubenswrapper[4971]: I1213 06:53:46.144687 4971 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-config-operator"/"kube-root-ca.crt" Dec 13 06:53:46 crc kubenswrapper[4971]: I1213 06:53:46.182141 4971 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-machine-config-operator"/"machine-config-operator-images" Dec 13 06:53:46 crc kubenswrapper[4971]: I1213 06:53:46.220610 4971 reflector.go:368] Caches populated for *v1.CSIDriver from k8s.io/client-go/informers/factory.go:160 Dec 13 06:53:46 crc kubenswrapper[4971]: I1213 06:53:46.263595 4971 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-storage-version-migrator"/"kube-root-ca.crt" Dec 13 06:53:46 crc kubenswrapper[4971]: I1213 06:53:46.274367 4971 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-multus"/"openshift-service-ca.crt" Dec 13 06:53:46 crc kubenswrapper[4971]: I1213 06:53:46.315859 4971 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-multus"/"multus-ancillary-tools-dockercfg-vnmsz" Dec 13 06:53:46 crc kubenswrapper[4971]: I1213 06:53:46.351695 4971 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-multus"/"default-cni-sysctl-allowlist" Dec 13 06:53:46 crc kubenswrapper[4971]: I1213 06:53:46.415755 4971 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-marketplace"/"marketplace-trusted-ca" Dec 13 06:53:46 crc kubenswrapper[4971]: I1213 06:53:46.518825 4971 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-service-ca"/"signing-cabundle" Dec 13 06:53:46 crc kubenswrapper[4971]: I1213 06:53:46.547973 4971 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication-operator"/"openshift-service-ca.crt" Dec 13 06:53:46 crc kubenswrapper[4971]: I1213 06:53:46.550453 4971 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-controller-manager-operator"/"kube-root-ca.crt" Dec 13 06:53:46 crc kubenswrapper[4971]: I1213 06:53:46.560653 4971 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-multus"/"multus-admission-controller-secret" Dec 13 06:53:46 crc kubenswrapper[4971]: I1213 06:53:46.644945 4971 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-service-ca"/"signing-key" Dec 13 06:53:46 crc kubenswrapper[4971]: I1213 06:53:46.716861 4971 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-route-controller-manager"/"route-controller-manager-sa-dockercfg-h2zr2" Dec 13 06:53:46 crc kubenswrapper[4971]: I1213 06:53:46.820251 4971 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ingress-canary"/"openshift-service-ca.crt" Dec 13 06:53:46 crc kubenswrapper[4971]: I1213 06:53:46.836598 4971 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"catalog-operator-serving-cert" Dec 13 06:53:46 crc kubenswrapper[4971]: I1213 06:53:46.898214 4971 reflector.go:368] Caches populated for *v1.ConfigMap from object-"hostpath-provisioner"/"kube-root-ca.crt" Dec 13 06:53:46 crc kubenswrapper[4971]: I1213 06:53:46.981264 4971 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ingress"/"openshift-service-ca.crt" Dec 13 06:53:47 crc kubenswrapper[4971]: I1213 06:53:47.021239 4971 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-operator-lifecycle-manager"/"kube-root-ca.crt" Dec 13 06:53:47 crc kubenswrapper[4971]: I1213 06:53:47.091206 4971 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-image-registry"/"image-registry-tls" Dec 13 06:53:47 crc kubenswrapper[4971]: I1213 06:53:47.203576 4971 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ingress"/"router-metrics-certs-default" Dec 13 06:53:47 crc kubenswrapper[4971]: I1213 06:53:47.231559 4971 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-service-ca-operator"/"kube-root-ca.crt" Dec 13 06:53:47 crc kubenswrapper[4971]: I1213 06:53:47.237955 4971 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-apiserver"/"openshift-apiserver-sa-dockercfg-djjff" Dec 13 06:53:47 crc kubenswrapper[4971]: I1213 06:53:47.312752 4971 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-image-registry"/"trusted-ca" Dec 13 06:53:47 crc kubenswrapper[4971]: I1213 06:53:47.331166 4971 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-apiserver-operator"/"openshift-apiserver-operator-serving-cert" Dec 13 06:53:47 crc kubenswrapper[4971]: I1213 06:53:47.347332 4971 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ovn-kubernetes"/"kube-root-ca.crt" Dec 13 06:53:47 crc kubenswrapper[4971]: I1213 06:53:47.430109 4971 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"package-server-manager-serving-cert" Dec 13 06:53:47 crc kubenswrapper[4971]: I1213 06:53:47.433829 4971 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-apiserver-operator"/"kube-apiserver-operator-config" Dec 13 06:53:47 crc kubenswrapper[4971]: I1213 06:53:47.477949 4971 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-route-controller-manager"/"kube-root-ca.crt" Dec 13 06:53:47 crc kubenswrapper[4971]: I1213 06:53:47.486818 4971 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console"/"trusted-ca-bundle" Dec 13 06:53:47 crc kubenswrapper[4971]: I1213 06:53:47.523163 4971 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-node-identity"/"env-overrides" Dec 13 06:53:47 crc kubenswrapper[4971]: I1213 06:53:47.523880 4971 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-etcd-operator"/"etcd-service-ca-bundle" Dec 13 06:53:47 crc kubenswrapper[4971]: I1213 06:53:47.597996 4971 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-config-operator"/"openshift-service-ca.crt" Dec 13 06:53:47 crc kubenswrapper[4971]: I1213 06:53:47.676104 4971 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-image-registry"/"cluster-image-registry-operator-dockercfg-m4qtx" Dec 13 06:53:47 crc kubenswrapper[4971]: I1213 06:53:47.726404 4971 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-image-registry"/"openshift-service-ca.crt" Dec 13 06:53:47 crc kubenswrapper[4971]: I1213 06:53:47.789964 4971 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-machine-api"/"kube-rbac-proxy" Dec 13 06:53:47 crc kubenswrapper[4971]: I1213 06:53:47.807842 4971 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-etcd-operator"/"etcd-ca-bundle" Dec 13 06:53:47 crc kubenswrapper[4971]: I1213 06:53:47.832002 4971 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"mcc-proxy-tls" Dec 13 06:53:47 crc kubenswrapper[4971]: I1213 06:53:47.887320 4971 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"mco-proxy-tls" Dec 13 06:53:47 crc kubenswrapper[4971]: I1213 06:53:47.937146 4971 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"openshift-service-ca.crt" Dec 13 06:53:48 crc kubenswrapper[4971]: I1213 06:53:48.095548 4971 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-etcd-operator"/"etcd-operator-config" Dec 13 06:53:48 crc kubenswrapper[4971]: I1213 06:53:48.110370 4971 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-console"/"networking-console-plugin" Dec 13 06:53:48 crc kubenswrapper[4971]: I1213 06:53:48.125200 4971 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-controller-manager-operator"/"kube-controller-manager-operator-serving-cert" Dec 13 06:53:48 crc kubenswrapper[4971]: I1213 06:53:48.169680 4971 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-console"/"console-serving-cert" Dec 13 06:53:48 crc kubenswrapper[4971]: I1213 06:53:48.215437 4971 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-apiserver"/"serving-cert" Dec 13 06:53:48 crc kubenswrapper[4971]: I1213 06:53:48.340084 4971 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-scheduler-operator"/"openshift-kube-scheduler-operator-dockercfg-qt55r" Dec 13 06:53:48 crc kubenswrapper[4971]: I1213 06:53:48.413973 4971 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-scheduler-operator"/"kube-root-ca.crt" Dec 13 06:53:48 crc kubenswrapper[4971]: I1213 06:53:48.423723 4971 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-dns-operator"/"openshift-service-ca.crt" Dec 13 06:53:48 crc kubenswrapper[4971]: I1213 06:53:48.438624 4971 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-machine-api"/"machine-api-operator-images" Dec 13 06:53:48 crc kubenswrapper[4971]: I1213 06:53:48.440664 4971 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-dns"/"dns-dockercfg-jwfmh" Dec 13 06:53:48 crc kubenswrapper[4971]: I1213 06:53:48.516552 4971 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-api"/"machine-api-operator-dockercfg-mfbb7" Dec 13 06:53:48 crc kubenswrapper[4971]: I1213 06:53:48.524556 4971 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-controller-manager-operator"/"openshift-controller-manager-operator-dockercfg-vw8fw" Dec 13 06:53:48 crc kubenswrapper[4971]: I1213 06:53:48.525205 4971 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver"/"etcd-serving-ca" Dec 13 06:53:48 crc kubenswrapper[4971]: I1213 06:53:48.527924 4971 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-apiserver-operator"/"kube-apiserver-operator-serving-cert" Dec 13 06:53:48 crc kubenswrapper[4971]: I1213 06:53:48.549059 4971 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-storage-version-migrator"/"openshift-service-ca.crt" Dec 13 06:53:48 crc kubenswrapper[4971]: I1213 06:53:48.582118 4971 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console"/"console-config" Dec 13 06:53:48 crc kubenswrapper[4971]: I1213 06:53:48.601600 4971 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-dns"/"dns-default" Dec 13 06:53:48 crc kubenswrapper[4971]: I1213 06:53:48.612689 4971 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ingress-operator"/"ingress-operator-dockercfg-7lnqk" Dec 13 06:53:48 crc kubenswrapper[4971]: I1213 06:53:48.616295 4971 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-oauth-apiserver"/"audit-1" Dec 13 06:53:48 crc kubenswrapper[4971]: I1213 06:53:48.642988 4971 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-marketplace"/"openshift-service-ca.crt" Dec 13 06:53:48 crc kubenswrapper[4971]: I1213 06:53:48.833359 4971 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-network-operator"/"metrics-tls" Dec 13 06:53:48 crc kubenswrapper[4971]: I1213 06:53:48.847410 4971 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ovn-kubernetes"/"ovnkube-script-lib" Dec 13 06:53:48 crc kubenswrapper[4971]: I1213 06:53:48.936573 4971 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console-operator"/"trusted-ca" Dec 13 06:53:49 crc kubenswrapper[4971]: I1213 06:53:49.018106 4971 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"marketplace-operator-metrics" Dec 13 06:53:49 crc kubenswrapper[4971]: I1213 06:53:49.021969 4971 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-controller-manager"/"openshift-controller-manager-sa-dockercfg-msq4c" Dec 13 06:53:49 crc kubenswrapper[4971]: I1213 06:53:49.025964 4971 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-image-registry"/"installation-pull-secrets" Dec 13 06:53:49 crc kubenswrapper[4971]: I1213 06:53:49.037754 4971 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-dns-operator"/"dns-operator-dockercfg-9mqw5" Dec 13 06:53:49 crc kubenswrapper[4971]: I1213 06:53:49.061302 4971 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-multus"/"multus-ac-dockercfg-9lkdf" Dec 13 06:53:49 crc kubenswrapper[4971]: I1213 06:53:49.068313 4971 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-oauth-apiserver"/"encryption-config-1" Dec 13 06:53:49 crc kubenswrapper[4971]: I1213 06:53:49.204921 4971 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"machine-config-server-dockercfg-qx5rd" Dec 13 06:53:49 crc kubenswrapper[4971]: I1213 06:53:49.249960 4971 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-node-identity"/"kube-root-ca.crt" Dec 13 06:53:49 crc kubenswrapper[4971]: I1213 06:53:49.334300 4971 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-cluster-machine-approver"/"machine-approver-tls" Dec 13 06:53:49 crc kubenswrapper[4971]: I1213 06:53:49.347552 4971 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-dns"/"dns-default-metrics-tls" Dec 13 06:53:49 crc kubenswrapper[4971]: I1213 06:53:49.408689 4971 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ingress-operator"/"kube-root-ca.crt" Dec 13 06:53:49 crc kubenswrapper[4971]: I1213 06:53:49.458069 4971 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-scheduler-operator"/"openshift-kube-scheduler-operator-config" Dec 13 06:53:49 crc kubenswrapper[4971]: I1213 06:53:49.487399 4971 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-etcd-operator"/"etcd-operator-dockercfg-r9srn" Dec 13 06:53:49 crc kubenswrapper[4971]: I1213 06:53:49.572821 4971 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-apiserver-operator"/"kube-apiserver-operator-dockercfg-x57mr" Dec 13 06:53:49 crc kubenswrapper[4971]: I1213 06:53:49.587550 4971 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console-operator"/"console-operator-config" Dec 13 06:53:49 crc kubenswrapper[4971]: I1213 06:53:49.587601 4971 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-cluster-machine-approver"/"kube-rbac-proxy" Dec 13 06:53:49 crc kubenswrapper[4971]: I1213 06:53:49.721417 4971 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-service-ca"/"openshift-service-ca.crt" Dec 13 06:53:49 crc kubenswrapper[4971]: I1213 06:53:49.761322 4971 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-cluster-samples-operator"/"openshift-service-ca.crt" Dec 13 06:53:49 crc kubenswrapper[4971]: I1213 06:53:49.786376 4971 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager-operator"/"kube-root-ca.crt" Dec 13 06:53:49 crc kubenswrapper[4971]: I1213 06:53:49.789283 4971 reflector.go:368] Caches populated for *v1.Node from k8s.io/client-go/informers/factory.go:160 Dec 13 06:53:49 crc kubenswrapper[4971]: I1213 06:53:49.940991 4971 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-storage-version-migrator"/"kube-storage-version-migrator-sa-dockercfg-5xfcg" Dec 13 06:53:49 crc kubenswrapper[4971]: I1213 06:53:49.962822 4971 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-oauth-apiserver"/"etcd-client" Dec 13 06:53:50 crc kubenswrapper[4971]: I1213 06:53:50.002597 4971 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-service-ca"/"kube-root-ca.crt" Dec 13 06:53:50 crc kubenswrapper[4971]: I1213 06:53:50.012253 4971 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-apiserver-operator"/"openshift-apiserver-operator-dockercfg-xtcjv" Dec 13 06:53:50 crc kubenswrapper[4971]: I1213 06:53:50.041262 4971 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-storage-version-migrator-operator"/"openshift-service-ca.crt" Dec 13 06:53:50 crc kubenswrapper[4971]: I1213 06:53:50.235783 4971 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-route-controller-manager"/"openshift-service-ca.crt" Dec 13 06:53:50 crc kubenswrapper[4971]: I1213 06:53:50.437048 4971 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"config" Dec 13 06:53:50 crc kubenswrapper[4971]: I1213 06:53:50.482059 4971 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-multus"/"multus-daemon-config" Dec 13 06:53:50 crc kubenswrapper[4971]: I1213 06:53:50.513765 4971 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-console"/"console-oauth-config" Dec 13 06:53:50 crc kubenswrapper[4971]: I1213 06:53:50.585394 4971 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ingress"/"router-stats-default" Dec 13 06:53:50 crc kubenswrapper[4971]: I1213 06:53:50.718965 4971 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-apiserver"/"etcd-client" Dec 13 06:53:50 crc kubenswrapper[4971]: I1213 06:53:50.721452 4971 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-dns-operator"/"metrics-tls" Dec 13 06:53:50 crc kubenswrapper[4971]: I1213 06:53:50.723387 4971 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-multus"/"kube-root-ca.crt" Dec 13 06:53:50 crc kubenswrapper[4971]: I1213 06:53:50.776066 4971 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-etcd-operator"/"openshift-service-ca.crt" Dec 13 06:53:50 crc kubenswrapper[4971]: I1213 06:53:50.793476 4971 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-cluster-machine-approver"/"openshift-service-ca.crt" Dec 13 06:53:50 crc kubenswrapper[4971]: I1213 06:53:50.993717 4971 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-oauth-apiserver"/"trusted-ca-bundle" Dec 13 06:53:51 crc kubenswrapper[4971]: I1213 06:53:51.108577 4971 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-scheduler-operator"/"kube-scheduler-operator-serving-cert" Dec 13 06:53:51 crc kubenswrapper[4971]: I1213 06:53:51.121802 4971 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ingress"/"kube-root-ca.crt" Dec 13 06:53:51 crc kubenswrapper[4971]: I1213 06:53:51.127223 4971 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-machine-config-operator"/"openshift-service-ca.crt" Dec 13 06:53:51 crc kubenswrapper[4971]: I1213 06:53:51.135062 4971 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console"/"oauth-serving-cert" Dec 13 06:53:51 crc kubenswrapper[4971]: I1213 06:53:51.156991 4971 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-etcd-operator"/"kube-root-ca.crt" Dec 13 06:53:51 crc kubenswrapper[4971]: I1213 06:53:51.268998 4971 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-cluster-samples-operator"/"kube-root-ca.crt" Dec 13 06:53:51 crc kubenswrapper[4971]: I1213 06:53:51.346671 4971 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ovn-kubernetes"/"ovn-control-plane-metrics-cert" Dec 13 06:53:51 crc kubenswrapper[4971]: I1213 06:53:51.373470 4971 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-image-registry"/"image-registry-operator-tls" Dec 13 06:53:51 crc kubenswrapper[4971]: I1213 06:53:51.410764 4971 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-marketplace"/"kube-root-ca.crt" Dec 13 06:53:51 crc kubenswrapper[4971]: I1213 06:53:51.495546 4971 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-cluster-samples-operator"/"samples-operator-tls" Dec 13 06:53:51 crc kubenswrapper[4971]: I1213 06:53:51.502861 4971 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-apiserver"/"encryption-config-1" Dec 13 06:53:51 crc kubenswrapper[4971]: I1213 06:53:51.671248 4971 kubelet.go:2431] "SyncLoop REMOVE" source="file" pods=["openshift-kube-apiserver/kube-apiserver-startup-monitor-crc"] Dec 13 06:53:51 crc kubenswrapper[4971]: I1213 06:53:51.671493 4971 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" podUID="f85e55b1a89d02b0cb034b1ea31ed45a" containerName="startup-monitor" containerID="cri-o://a60d4a360ca22d5ab008969e58dd981a7858882ea44246f6b56a407dba706501" gracePeriod=5 Dec 13 06:53:51 crc kubenswrapper[4971]: I1213 06:53:51.674158 4971 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-image-registry"/"node-ca-dockercfg-4777p" Dec 13 06:53:51 crc kubenswrapper[4971]: I1213 06:53:51.690915 4971 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver"/"config" Dec 13 06:53:51 crc kubenswrapper[4971]: I1213 06:53:51.768965 4971 scope.go:117] "RemoveContainer" containerID="286c449dc5a9a113f79f93efbaf4070fc4e73a43894b9a9329a6c56a40aef05b" Dec 13 06:53:51 crc kubenswrapper[4971]: I1213 06:53:51.795061 4971 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-service-ca-operator"/"service-ca-operator-dockercfg-rg9jl" Dec 13 06:53:51 crc kubenswrapper[4971]: I1213 06:53:51.812862 4971 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver"/"audit-1" Dec 13 06:53:51 crc kubenswrapper[4971]: I1213 06:53:51.843686 4971 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-api"/"control-plane-machine-set-operator-tls" Dec 13 06:53:52 crc kubenswrapper[4971]: I1213 06:53:52.006315 4971 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-dns-operator"/"kube-root-ca.crt" Dec 13 06:53:52 crc kubenswrapper[4971]: I1213 06:53:52.111571 4971 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ovn-kubernetes"/"env-overrides" Dec 13 06:53:52 crc kubenswrapper[4971]: I1213 06:53:52.133538 4971 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-dns"/"openshift-service-ca.crt" Dec 13 06:53:52 crc kubenswrapper[4971]: I1213 06:53:52.179829 4971 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-route-controller-manager"/"serving-cert" Dec 13 06:53:52 crc kubenswrapper[4971]: I1213 06:53:52.189139 4971 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-image-registry"/"image-registry-certificates" Dec 13 06:53:52 crc kubenswrapper[4971]: I1213 06:53:52.193078 4971 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-route-controller-manager"/"client-ca" Dec 13 06:53:52 crc kubenswrapper[4971]: I1213 06:53:52.235434 4971 reflector.go:368] Caches populated for *v1.Service from k8s.io/client-go/informers/factory.go:160 Dec 13 06:53:52 crc kubenswrapper[4971]: I1213 06:53:52.244985 4971 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"node-bootstrapper-token" Dec 13 06:53:52 crc kubenswrapper[4971]: I1213 06:53:52.299734 4971 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-cluster-samples-operator"/"cluster-samples-operator-dockercfg-xpp9w" Dec 13 06:53:52 crc kubenswrapper[4971]: I1213 06:53:52.425121 4971 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-oauth-apiserver"/"oauth-apiserver-sa-dockercfg-6r2bq" Dec 13 06:53:52 crc kubenswrapper[4971]: I1213 06:53:52.454607 4971 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-multus"/"default-dockercfg-2q5b6" Dec 13 06:53:52 crc kubenswrapper[4971]: I1213 06:53:52.515566 4971 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication-operator"/"authentication-operator-dockercfg-mz9bj" Dec 13 06:53:52 crc kubenswrapper[4971]: I1213 06:53:52.520683 4971 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver"/"kube-root-ca.crt" Dec 13 06:53:52 crc kubenswrapper[4971]: I1213 06:53:52.716174 4971 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-operator-lifecycle-manager"/"openshift-service-ca.crt" Dec 13 06:53:52 crc kubenswrapper[4971]: I1213 06:53:52.724581 4971 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_marketplace-operator-79b997595-6rvc9_5c88a864-89a9-4f5e-b713-385170b9bf17/marketplace-operator/2.log" Dec 13 06:53:52 crc kubenswrapper[4971]: I1213 06:53:52.724829 4971 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/marketplace-operator-79b997595-6rvc9" event={"ID":"5c88a864-89a9-4f5e-b713-385170b9bf17","Type":"ContainerStarted","Data":"c61b97c1755702dffd406eec5db5b5f8bb5f4fa4b6d081c8392bce7eb961c949"} Dec 13 06:53:52 crc kubenswrapper[4971]: I1213 06:53:52.725273 4971 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/marketplace-operator-79b997595-6rvc9" Dec 13 06:53:52 crc kubenswrapper[4971]: I1213 06:53:52.736067 4971 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/marketplace-operator-79b997595-6rvc9" Dec 13 06:53:52 crc kubenswrapper[4971]: I1213 06:53:52.741874 4971 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/marketplace-operator-79b997595-6rvc9" podStartSLOduration=46.741856678 podStartE2EDuration="46.741856678s" podCreationTimestamp="2025-12-13 06:53:06 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-13 06:53:52.738748083 +0000 UTC m=+289.343157531" watchObservedRunningTime="2025-12-13 06:53:52.741856678 +0000 UTC m=+289.346266126" Dec 13 06:53:52 crc kubenswrapper[4971]: I1213 06:53:52.870272 4971 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication-operator"/"kube-root-ca.crt" Dec 13 06:53:53 crc kubenswrapper[4971]: I1213 06:53:53.257097 4971 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager-operator"/"openshift-service-ca.crt" Dec 13 06:53:53 crc kubenswrapper[4971]: I1213 06:53:53.422395 4971 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"machine-config-daemon-dockercfg-r5tcq" Dec 13 06:53:53 crc kubenswrapper[4971]: I1213 06:53:53.470641 4971 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ingress-operator"/"openshift-service-ca.crt" Dec 13 06:53:53 crc kubenswrapper[4971]: I1213 06:53:53.476976 4971 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-cluster-machine-approver"/"kube-root-ca.crt" Dec 13 06:53:53 crc kubenswrapper[4971]: I1213 06:53:53.611779 4971 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console-operator"/"openshift-service-ca.crt" Dec 13 06:53:53 crc kubenswrapper[4971]: I1213 06:53:53.645579 4971 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver-operator"/"openshift-apiserver-operator-config" Dec 13 06:53:53 crc kubenswrapper[4971]: I1213 06:53:53.715230 4971 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-node-identity"/"openshift-service-ca.crt" Dec 13 06:53:53 crc kubenswrapper[4971]: I1213 06:53:53.715233 4971 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-multus"/"cni-copy-resources" Dec 13 06:53:53 crc kubenswrapper[4971]: I1213 06:53:53.753100 4971 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-machine-api"/"openshift-service-ca.crt" Dec 13 06:53:53 crc kubenswrapper[4971]: I1213 06:53:53.765957 4971 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-storage-version-migrator-operator"/"serving-cert" Dec 13 06:53:53 crc kubenswrapper[4971]: I1213 06:53:53.824721 4971 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-api"/"machine-api-operator-tls" Dec 13 06:53:53 crc kubenswrapper[4971]: I1213 06:53:53.925025 4971 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"machine-config-server-tls" Dec 13 06:53:53 crc kubenswrapper[4971]: I1213 06:53:53.971628 4971 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"client-ca" Dec 13 06:53:54 crc kubenswrapper[4971]: I1213 06:53:54.006614 4971 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ingress"/"router-dockercfg-zdk86" Dec 13 06:53:54 crc kubenswrapper[4971]: I1213 06:53:54.048700 4971 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ovn-kubernetes"/"ovn-node-metrics-cert" Dec 13 06:53:54 crc kubenswrapper[4971]: I1213 06:53:54.121547 4971 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"machine-config-controller-dockercfg-c2lfx" Dec 13 06:53:54 crc kubenswrapper[4971]: I1213 06:53:54.163943 4971 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console"/"openshift-service-ca.crt" Dec 13 06:53:54 crc kubenswrapper[4971]: I1213 06:53:54.392384 4971 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-controller-manager-operator"/"kube-controller-manager-operator-config" Dec 13 06:53:54 crc kubenswrapper[4971]: I1213 06:53:54.436421 4971 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-storage-version-migrator-operator"/"kube-storage-version-migrator-operator-dockercfg-2bh8d" Dec 13 06:53:54 crc kubenswrapper[4971]: I1213 06:53:54.748658 4971 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"pprof-cert" Dec 13 06:53:54 crc kubenswrapper[4971]: I1213 06:53:54.958192 4971 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-oauth-apiserver"/"serving-cert" Dec 13 06:53:54 crc kubenswrapper[4971]: I1213 06:53:54.984863 4971 reflector.go:368] Caches populated for *v1.ConfigMap from object-"hostpath-provisioner"/"openshift-service-ca.crt" Dec 13 06:53:55 crc kubenswrapper[4971]: I1213 06:53:55.107316 4971 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-network-node-identity"/"network-node-identity-cert" Dec 13 06:53:55 crc kubenswrapper[4971]: I1213 06:53:55.132109 4971 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ingress-canary"/"default-dockercfg-2llfx" Dec 13 06:53:55 crc kubenswrapper[4971]: I1213 06:53:55.186421 4971 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-multus"/"metrics-daemon-sa-dockercfg-d427c" Dec 13 06:53:55 crc kubenswrapper[4971]: I1213 06:53:55.191442 4971 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ingress-canary"/"canary-serving-cert" Dec 13 06:53:55 crc kubenswrapper[4971]: I1213 06:53:55.383474 4971 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-cluster-machine-approver"/"machine-approver-config" Dec 13 06:53:55 crc kubenswrapper[4971]: I1213 06:53:55.522453 4971 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-authentication/oauth-openshift-684d58cf9b-xzljc"] Dec 13 06:53:55 crc kubenswrapper[4971]: E1213 06:53:55.522678 4971 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f14c8a94-8589-4059-bddf-329452355fab" containerName="marketplace-operator" Dec 13 06:53:55 crc kubenswrapper[4971]: I1213 06:53:55.522689 4971 state_mem.go:107] "Deleted CPUSet assignment" podUID="f14c8a94-8589-4059-bddf-329452355fab" containerName="marketplace-operator" Dec 13 06:53:55 crc kubenswrapper[4971]: E1213 06:53:55.522699 4971 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="22037bdc-fe6d-4411-be2b-aec38524a8dd" containerName="extract-utilities" Dec 13 06:53:55 crc kubenswrapper[4971]: I1213 06:53:55.522705 4971 state_mem.go:107] "Deleted CPUSet assignment" podUID="22037bdc-fe6d-4411-be2b-aec38524a8dd" containerName="extract-utilities" Dec 13 06:53:55 crc kubenswrapper[4971]: E1213 06:53:55.522713 4971 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="cf417851-ae43-462f-a678-125593fb653b" containerName="extract-utilities" Dec 13 06:53:55 crc kubenswrapper[4971]: I1213 06:53:55.522719 4971 state_mem.go:107] "Deleted CPUSet assignment" podUID="cf417851-ae43-462f-a678-125593fb653b" containerName="extract-utilities" Dec 13 06:53:55 crc kubenswrapper[4971]: E1213 06:53:55.522731 4971 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="22037bdc-fe6d-4411-be2b-aec38524a8dd" containerName="extract-content" Dec 13 06:53:55 crc kubenswrapper[4971]: I1213 06:53:55.522737 4971 state_mem.go:107] "Deleted CPUSet assignment" podUID="22037bdc-fe6d-4411-be2b-aec38524a8dd" containerName="extract-content" Dec 13 06:53:55 crc kubenswrapper[4971]: E1213 06:53:55.522745 4971 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="02f6e9f2-6bad-441c-8e2e-9acfc71f3ab4" containerName="registry-server" Dec 13 06:53:55 crc kubenswrapper[4971]: I1213 06:53:55.522750 4971 state_mem.go:107] "Deleted CPUSet assignment" podUID="02f6e9f2-6bad-441c-8e2e-9acfc71f3ab4" containerName="registry-server" Dec 13 06:53:55 crc kubenswrapper[4971]: E1213 06:53:55.522758 4971 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="b9a0bc5d-3a39-4732-a849-22d351a4cc18" containerName="extract-utilities" Dec 13 06:53:55 crc kubenswrapper[4971]: I1213 06:53:55.522764 4971 state_mem.go:107] "Deleted CPUSet assignment" podUID="b9a0bc5d-3a39-4732-a849-22d351a4cc18" containerName="extract-utilities" Dec 13 06:53:55 crc kubenswrapper[4971]: E1213 06:53:55.522773 4971 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="cf417851-ae43-462f-a678-125593fb653b" containerName="extract-content" Dec 13 06:53:55 crc kubenswrapper[4971]: I1213 06:53:55.522778 4971 state_mem.go:107] "Deleted CPUSet assignment" podUID="cf417851-ae43-462f-a678-125593fb653b" containerName="extract-content" Dec 13 06:53:55 crc kubenswrapper[4971]: E1213 06:53:55.522787 4971 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="9d4f1ea5-8472-4320-820b-8090ff6f5ca6" containerName="extract-content" Dec 13 06:53:55 crc kubenswrapper[4971]: I1213 06:53:55.522792 4971 state_mem.go:107] "Deleted CPUSet assignment" podUID="9d4f1ea5-8472-4320-820b-8090ff6f5ca6" containerName="extract-content" Dec 13 06:53:55 crc kubenswrapper[4971]: E1213 06:53:55.522799 4971 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="b9a0bc5d-3a39-4732-a849-22d351a4cc18" containerName="extract-content" Dec 13 06:53:55 crc kubenswrapper[4971]: I1213 06:53:55.522804 4971 state_mem.go:107] "Deleted CPUSet assignment" podUID="b9a0bc5d-3a39-4732-a849-22d351a4cc18" containerName="extract-content" Dec 13 06:53:55 crc kubenswrapper[4971]: E1213 06:53:55.522812 4971 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="9d4f1ea5-8472-4320-820b-8090ff6f5ca6" containerName="extract-utilities" Dec 13 06:53:55 crc kubenswrapper[4971]: I1213 06:53:55.522817 4971 state_mem.go:107] "Deleted CPUSet assignment" podUID="9d4f1ea5-8472-4320-820b-8090ff6f5ca6" containerName="extract-utilities" Dec 13 06:53:55 crc kubenswrapper[4971]: E1213 06:53:55.522824 4971 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="8f50f5ba-3ec3-4fa2-9f5d-51e1c43f6352" containerName="extract-content" Dec 13 06:53:55 crc kubenswrapper[4971]: I1213 06:53:55.522830 4971 state_mem.go:107] "Deleted CPUSet assignment" podUID="8f50f5ba-3ec3-4fa2-9f5d-51e1c43f6352" containerName="extract-content" Dec 13 06:53:55 crc kubenswrapper[4971]: E1213 06:53:55.522837 4971 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="22037bdc-fe6d-4411-be2b-aec38524a8dd" containerName="registry-server" Dec 13 06:53:55 crc kubenswrapper[4971]: I1213 06:53:55.522842 4971 state_mem.go:107] "Deleted CPUSet assignment" podUID="22037bdc-fe6d-4411-be2b-aec38524a8dd" containerName="registry-server" Dec 13 06:53:55 crc kubenswrapper[4971]: E1213 06:53:55.522850 4971 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="53fdcbc2-900b-48eb-8be3-cd72940096c3" containerName="installer" Dec 13 06:53:55 crc kubenswrapper[4971]: I1213 06:53:55.522856 4971 state_mem.go:107] "Deleted CPUSet assignment" podUID="53fdcbc2-900b-48eb-8be3-cd72940096c3" containerName="installer" Dec 13 06:53:55 crc kubenswrapper[4971]: E1213 06:53:55.522865 4971 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="cf417851-ae43-462f-a678-125593fb653b" containerName="registry-server" Dec 13 06:53:55 crc kubenswrapper[4971]: I1213 06:53:55.522870 4971 state_mem.go:107] "Deleted CPUSet assignment" podUID="cf417851-ae43-462f-a678-125593fb653b" containerName="registry-server" Dec 13 06:53:55 crc kubenswrapper[4971]: E1213 06:53:55.522878 4971 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="b9a0bc5d-3a39-4732-a849-22d351a4cc18" containerName="registry-server" Dec 13 06:53:55 crc kubenswrapper[4971]: I1213 06:53:55.522885 4971 state_mem.go:107] "Deleted CPUSet assignment" podUID="b9a0bc5d-3a39-4732-a849-22d351a4cc18" containerName="registry-server" Dec 13 06:53:55 crc kubenswrapper[4971]: E1213 06:53:55.522893 4971 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="02f6e9f2-6bad-441c-8e2e-9acfc71f3ab4" containerName="extract-utilities" Dec 13 06:53:55 crc kubenswrapper[4971]: I1213 06:53:55.522900 4971 state_mem.go:107] "Deleted CPUSet assignment" podUID="02f6e9f2-6bad-441c-8e2e-9acfc71f3ab4" containerName="extract-utilities" Dec 13 06:53:55 crc kubenswrapper[4971]: E1213 06:53:55.522909 4971 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="9d4f1ea5-8472-4320-820b-8090ff6f5ca6" containerName="registry-server" Dec 13 06:53:55 crc kubenswrapper[4971]: I1213 06:53:55.522914 4971 state_mem.go:107] "Deleted CPUSet assignment" podUID="9d4f1ea5-8472-4320-820b-8090ff6f5ca6" containerName="registry-server" Dec 13 06:53:55 crc kubenswrapper[4971]: E1213 06:53:55.522922 4971 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="8f50f5ba-3ec3-4fa2-9f5d-51e1c43f6352" containerName="extract-utilities" Dec 13 06:53:55 crc kubenswrapper[4971]: I1213 06:53:55.522928 4971 state_mem.go:107] "Deleted CPUSet assignment" podUID="8f50f5ba-3ec3-4fa2-9f5d-51e1c43f6352" containerName="extract-utilities" Dec 13 06:53:55 crc kubenswrapper[4971]: E1213 06:53:55.522935 4971 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="8f50f5ba-3ec3-4fa2-9f5d-51e1c43f6352" containerName="registry-server" Dec 13 06:53:55 crc kubenswrapper[4971]: I1213 06:53:55.522941 4971 state_mem.go:107] "Deleted CPUSet assignment" podUID="8f50f5ba-3ec3-4fa2-9f5d-51e1c43f6352" containerName="registry-server" Dec 13 06:53:55 crc kubenswrapper[4971]: E1213 06:53:55.522950 4971 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f85e55b1a89d02b0cb034b1ea31ed45a" containerName="startup-monitor" Dec 13 06:53:55 crc kubenswrapper[4971]: I1213 06:53:55.522955 4971 state_mem.go:107] "Deleted CPUSet assignment" podUID="f85e55b1a89d02b0cb034b1ea31ed45a" containerName="startup-monitor" Dec 13 06:53:55 crc kubenswrapper[4971]: E1213 06:53:55.522963 4971 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="23e16ced-5b54-4a51-a1bc-4d11c907c90a" containerName="oauth-openshift" Dec 13 06:53:55 crc kubenswrapper[4971]: I1213 06:53:55.522969 4971 state_mem.go:107] "Deleted CPUSet assignment" podUID="23e16ced-5b54-4a51-a1bc-4d11c907c90a" containerName="oauth-openshift" Dec 13 06:53:55 crc kubenswrapper[4971]: E1213 06:53:55.522979 4971 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="02f6e9f2-6bad-441c-8e2e-9acfc71f3ab4" containerName="extract-content" Dec 13 06:53:55 crc kubenswrapper[4971]: I1213 06:53:55.522984 4971 state_mem.go:107] "Deleted CPUSet assignment" podUID="02f6e9f2-6bad-441c-8e2e-9acfc71f3ab4" containerName="extract-content" Dec 13 06:53:55 crc kubenswrapper[4971]: I1213 06:53:55.523060 4971 memory_manager.go:354] "RemoveStaleState removing state" podUID="9d4f1ea5-8472-4320-820b-8090ff6f5ca6" containerName="registry-server" Dec 13 06:53:55 crc kubenswrapper[4971]: I1213 06:53:55.523071 4971 memory_manager.go:354] "RemoveStaleState removing state" podUID="f14c8a94-8589-4059-bddf-329452355fab" containerName="marketplace-operator" Dec 13 06:53:55 crc kubenswrapper[4971]: I1213 06:53:55.523077 4971 memory_manager.go:354] "RemoveStaleState removing state" podUID="f85e55b1a89d02b0cb034b1ea31ed45a" containerName="startup-monitor" Dec 13 06:53:55 crc kubenswrapper[4971]: I1213 06:53:55.523086 4971 memory_manager.go:354] "RemoveStaleState removing state" podUID="53fdcbc2-900b-48eb-8be3-cd72940096c3" containerName="installer" Dec 13 06:53:55 crc kubenswrapper[4971]: I1213 06:53:55.523095 4971 memory_manager.go:354] "RemoveStaleState removing state" podUID="02f6e9f2-6bad-441c-8e2e-9acfc71f3ab4" containerName="registry-server" Dec 13 06:53:55 crc kubenswrapper[4971]: I1213 06:53:55.523104 4971 memory_manager.go:354] "RemoveStaleState removing state" podUID="22037bdc-fe6d-4411-be2b-aec38524a8dd" containerName="registry-server" Dec 13 06:53:55 crc kubenswrapper[4971]: I1213 06:53:55.523112 4971 memory_manager.go:354] "RemoveStaleState removing state" podUID="23e16ced-5b54-4a51-a1bc-4d11c907c90a" containerName="oauth-openshift" Dec 13 06:53:55 crc kubenswrapper[4971]: I1213 06:53:55.523121 4971 memory_manager.go:354] "RemoveStaleState removing state" podUID="b9a0bc5d-3a39-4732-a849-22d351a4cc18" containerName="registry-server" Dec 13 06:53:55 crc kubenswrapper[4971]: I1213 06:53:55.523130 4971 memory_manager.go:354] "RemoveStaleState removing state" podUID="8f50f5ba-3ec3-4fa2-9f5d-51e1c43f6352" containerName="registry-server" Dec 13 06:53:55 crc kubenswrapper[4971]: I1213 06:53:55.523138 4971 memory_manager.go:354] "RemoveStaleState removing state" podUID="cf417851-ae43-462f-a678-125593fb653b" containerName="registry-server" Dec 13 06:53:55 crc kubenswrapper[4971]: I1213 06:53:55.526181 4971 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-authentication/oauth-openshift-684d58cf9b-xzljc" Dec 13 06:53:55 crc kubenswrapper[4971]: I1213 06:53:55.529830 4971 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-user-template-error" Dec 13 06:53:55 crc kubenswrapper[4971]: I1213 06:53:55.530082 4971 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication"/"v4-0-config-system-cliconfig" Dec 13 06:53:55 crc kubenswrapper[4971]: I1213 06:53:55.530259 4971 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-system-serving-cert" Dec 13 06:53:55 crc kubenswrapper[4971]: I1213 06:53:55.530334 4971 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-user-idp-0-file-data" Dec 13 06:53:55 crc kubenswrapper[4971]: I1213 06:53:55.530400 4971 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication"/"audit" Dec 13 06:53:55 crc kubenswrapper[4971]: I1213 06:53:55.530631 4971 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication"/"v4-0-config-system-service-ca" Dec 13 06:53:55 crc kubenswrapper[4971]: I1213 06:53:55.530644 4971 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication"/"kube-root-ca.crt" Dec 13 06:53:55 crc kubenswrapper[4971]: I1213 06:53:55.530800 4971 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"oauth-openshift-dockercfg-znhcc" Dec 13 06:53:55 crc kubenswrapper[4971]: I1213 06:53:55.531342 4971 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-system-router-certs" Dec 13 06:53:55 crc kubenswrapper[4971]: I1213 06:53:55.531584 4971 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-user-template-provider-selection" Dec 13 06:53:55 crc kubenswrapper[4971]: I1213 06:53:55.531682 4971 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication"/"openshift-service-ca.crt" Dec 13 06:53:55 crc kubenswrapper[4971]: I1213 06:53:55.533360 4971 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-system-session" Dec 13 06:53:55 crc kubenswrapper[4971]: I1213 06:53:55.535844 4971 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"proxy-tls" Dec 13 06:53:55 crc kubenswrapper[4971]: I1213 06:53:55.539427 4971 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-user-template-login" Dec 13 06:53:55 crc kubenswrapper[4971]: I1213 06:53:55.542278 4971 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-authentication/oauth-openshift-684d58cf9b-xzljc"] Dec 13 06:53:55 crc kubenswrapper[4971]: I1213 06:53:55.547876 4971 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication"/"v4-0-config-system-trusted-ca-bundle" Dec 13 06:53:55 crc kubenswrapper[4971]: I1213 06:53:55.552348 4971 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-system-ocp-branding-template" Dec 13 06:53:55 crc kubenswrapper[4971]: I1213 06:53:55.611069 4971 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-user-template-provider-selection\" (UniqueName: \"kubernetes.io/secret/7c44c8d7-f038-4b8b-aadc-3ac7434f3b53-v4-0-config-user-template-provider-selection\") pod \"oauth-openshift-684d58cf9b-xzljc\" (UID: \"7c44c8d7-f038-4b8b-aadc-3ac7434f3b53\") " pod="openshift-authentication/oauth-openshift-684d58cf9b-xzljc" Dec 13 06:53:55 crc kubenswrapper[4971]: I1213 06:53:55.611129 4971 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-system-router-certs\" (UniqueName: \"kubernetes.io/secret/7c44c8d7-f038-4b8b-aadc-3ac7434f3b53-v4-0-config-system-router-certs\") pod \"oauth-openshift-684d58cf9b-xzljc\" (UID: \"7c44c8d7-f038-4b8b-aadc-3ac7434f3b53\") " pod="openshift-authentication/oauth-openshift-684d58cf9b-xzljc" Dec 13 06:53:55 crc kubenswrapper[4971]: I1213 06:53:55.611175 4971 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-system-session\" (UniqueName: \"kubernetes.io/secret/7c44c8d7-f038-4b8b-aadc-3ac7434f3b53-v4-0-config-system-session\") pod \"oauth-openshift-684d58cf9b-xzljc\" (UID: \"7c44c8d7-f038-4b8b-aadc-3ac7434f3b53\") " pod="openshift-authentication/oauth-openshift-684d58cf9b-xzljc" Dec 13 06:53:55 crc kubenswrapper[4971]: I1213 06:53:55.611203 4971 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-system-service-ca\" (UniqueName: \"kubernetes.io/configmap/7c44c8d7-f038-4b8b-aadc-3ac7434f3b53-v4-0-config-system-service-ca\") pod \"oauth-openshift-684d58cf9b-xzljc\" (UID: \"7c44c8d7-f038-4b8b-aadc-3ac7434f3b53\") " pod="openshift-authentication/oauth-openshift-684d58cf9b-xzljc" Dec 13 06:53:55 crc kubenswrapper[4971]: I1213 06:53:55.611233 4971 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-system-cliconfig\" (UniqueName: \"kubernetes.io/configmap/7c44c8d7-f038-4b8b-aadc-3ac7434f3b53-v4-0-config-system-cliconfig\") pod \"oauth-openshift-684d58cf9b-xzljc\" (UID: \"7c44c8d7-f038-4b8b-aadc-3ac7434f3b53\") " pod="openshift-authentication/oauth-openshift-684d58cf9b-xzljc" Dec 13 06:53:55 crc kubenswrapper[4971]: I1213 06:53:55.611261 4971 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-system-serving-cert\" (UniqueName: \"kubernetes.io/secret/7c44c8d7-f038-4b8b-aadc-3ac7434f3b53-v4-0-config-system-serving-cert\") pod \"oauth-openshift-684d58cf9b-xzljc\" (UID: \"7c44c8d7-f038-4b8b-aadc-3ac7434f3b53\") " pod="openshift-authentication/oauth-openshift-684d58cf9b-xzljc" Dec 13 06:53:55 crc kubenswrapper[4971]: I1213 06:53:55.611284 4971 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-system-ocp-branding-template\" (UniqueName: \"kubernetes.io/secret/7c44c8d7-f038-4b8b-aadc-3ac7434f3b53-v4-0-config-system-ocp-branding-template\") pod \"oauth-openshift-684d58cf9b-xzljc\" (UID: \"7c44c8d7-f038-4b8b-aadc-3ac7434f3b53\") " pod="openshift-authentication/oauth-openshift-684d58cf9b-xzljc" Dec 13 06:53:55 crc kubenswrapper[4971]: I1213 06:53:55.611311 4971 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-qt2gv\" (UniqueName: \"kubernetes.io/projected/7c44c8d7-f038-4b8b-aadc-3ac7434f3b53-kube-api-access-qt2gv\") pod \"oauth-openshift-684d58cf9b-xzljc\" (UID: \"7c44c8d7-f038-4b8b-aadc-3ac7434f3b53\") " pod="openshift-authentication/oauth-openshift-684d58cf9b-xzljc" Dec 13 06:53:55 crc kubenswrapper[4971]: I1213 06:53:55.611334 4971 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-system-trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/7c44c8d7-f038-4b8b-aadc-3ac7434f3b53-v4-0-config-system-trusted-ca-bundle\") pod \"oauth-openshift-684d58cf9b-xzljc\" (UID: \"7c44c8d7-f038-4b8b-aadc-3ac7434f3b53\") " pod="openshift-authentication/oauth-openshift-684d58cf9b-xzljc" Dec 13 06:53:55 crc kubenswrapper[4971]: I1213 06:53:55.611358 4971 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-user-idp-0-file-data\" (UniqueName: \"kubernetes.io/secret/7c44c8d7-f038-4b8b-aadc-3ac7434f3b53-v4-0-config-user-idp-0-file-data\") pod \"oauth-openshift-684d58cf9b-xzljc\" (UID: \"7c44c8d7-f038-4b8b-aadc-3ac7434f3b53\") " pod="openshift-authentication/oauth-openshift-684d58cf9b-xzljc" Dec 13 06:53:55 crc kubenswrapper[4971]: I1213 06:53:55.611384 4971 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-user-template-error\" (UniqueName: \"kubernetes.io/secret/7c44c8d7-f038-4b8b-aadc-3ac7434f3b53-v4-0-config-user-template-error\") pod \"oauth-openshift-684d58cf9b-xzljc\" (UID: \"7c44c8d7-f038-4b8b-aadc-3ac7434f3b53\") " pod="openshift-authentication/oauth-openshift-684d58cf9b-xzljc" Dec 13 06:53:55 crc kubenswrapper[4971]: I1213 06:53:55.611411 4971 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/7c44c8d7-f038-4b8b-aadc-3ac7434f3b53-audit-dir\") pod \"oauth-openshift-684d58cf9b-xzljc\" (UID: \"7c44c8d7-f038-4b8b-aadc-3ac7434f3b53\") " pod="openshift-authentication/oauth-openshift-684d58cf9b-xzljc" Dec 13 06:53:55 crc kubenswrapper[4971]: I1213 06:53:55.611437 4971 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-user-template-login\" (UniqueName: \"kubernetes.io/secret/7c44c8d7-f038-4b8b-aadc-3ac7434f3b53-v4-0-config-user-template-login\") pod \"oauth-openshift-684d58cf9b-xzljc\" (UID: \"7c44c8d7-f038-4b8b-aadc-3ac7434f3b53\") " pod="openshift-authentication/oauth-openshift-684d58cf9b-xzljc" Dec 13 06:53:55 crc kubenswrapper[4971]: I1213 06:53:55.611464 4971 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/7c44c8d7-f038-4b8b-aadc-3ac7434f3b53-audit-policies\") pod \"oauth-openshift-684d58cf9b-xzljc\" (UID: \"7c44c8d7-f038-4b8b-aadc-3ac7434f3b53\") " pod="openshift-authentication/oauth-openshift-684d58cf9b-xzljc" Dec 13 06:53:55 crc kubenswrapper[4971]: I1213 06:53:55.713103 4971 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-user-template-provider-selection\" (UniqueName: \"kubernetes.io/secret/7c44c8d7-f038-4b8b-aadc-3ac7434f3b53-v4-0-config-user-template-provider-selection\") pod \"oauth-openshift-684d58cf9b-xzljc\" (UID: \"7c44c8d7-f038-4b8b-aadc-3ac7434f3b53\") " pod="openshift-authentication/oauth-openshift-684d58cf9b-xzljc" Dec 13 06:53:55 crc kubenswrapper[4971]: I1213 06:53:55.713167 4971 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-system-router-certs\" (UniqueName: \"kubernetes.io/secret/7c44c8d7-f038-4b8b-aadc-3ac7434f3b53-v4-0-config-system-router-certs\") pod \"oauth-openshift-684d58cf9b-xzljc\" (UID: \"7c44c8d7-f038-4b8b-aadc-3ac7434f3b53\") " pod="openshift-authentication/oauth-openshift-684d58cf9b-xzljc" Dec 13 06:53:55 crc kubenswrapper[4971]: I1213 06:53:55.713225 4971 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-system-session\" (UniqueName: \"kubernetes.io/secret/7c44c8d7-f038-4b8b-aadc-3ac7434f3b53-v4-0-config-system-session\") pod \"oauth-openshift-684d58cf9b-xzljc\" (UID: \"7c44c8d7-f038-4b8b-aadc-3ac7434f3b53\") " pod="openshift-authentication/oauth-openshift-684d58cf9b-xzljc" Dec 13 06:53:55 crc kubenswrapper[4971]: I1213 06:53:55.713262 4971 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-system-service-ca\" (UniqueName: \"kubernetes.io/configmap/7c44c8d7-f038-4b8b-aadc-3ac7434f3b53-v4-0-config-system-service-ca\") pod \"oauth-openshift-684d58cf9b-xzljc\" (UID: \"7c44c8d7-f038-4b8b-aadc-3ac7434f3b53\") " pod="openshift-authentication/oauth-openshift-684d58cf9b-xzljc" Dec 13 06:53:55 crc kubenswrapper[4971]: I1213 06:53:55.713304 4971 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-system-cliconfig\" (UniqueName: \"kubernetes.io/configmap/7c44c8d7-f038-4b8b-aadc-3ac7434f3b53-v4-0-config-system-cliconfig\") pod \"oauth-openshift-684d58cf9b-xzljc\" (UID: \"7c44c8d7-f038-4b8b-aadc-3ac7434f3b53\") " pod="openshift-authentication/oauth-openshift-684d58cf9b-xzljc" Dec 13 06:53:55 crc kubenswrapper[4971]: I1213 06:53:55.713346 4971 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-system-serving-cert\" (UniqueName: \"kubernetes.io/secret/7c44c8d7-f038-4b8b-aadc-3ac7434f3b53-v4-0-config-system-serving-cert\") pod \"oauth-openshift-684d58cf9b-xzljc\" (UID: \"7c44c8d7-f038-4b8b-aadc-3ac7434f3b53\") " pod="openshift-authentication/oauth-openshift-684d58cf9b-xzljc" Dec 13 06:53:55 crc kubenswrapper[4971]: I1213 06:53:55.713379 4971 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-system-ocp-branding-template\" (UniqueName: \"kubernetes.io/secret/7c44c8d7-f038-4b8b-aadc-3ac7434f3b53-v4-0-config-system-ocp-branding-template\") pod \"oauth-openshift-684d58cf9b-xzljc\" (UID: \"7c44c8d7-f038-4b8b-aadc-3ac7434f3b53\") " pod="openshift-authentication/oauth-openshift-684d58cf9b-xzljc" Dec 13 06:53:55 crc kubenswrapper[4971]: I1213 06:53:55.713421 4971 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-qt2gv\" (UniqueName: \"kubernetes.io/projected/7c44c8d7-f038-4b8b-aadc-3ac7434f3b53-kube-api-access-qt2gv\") pod \"oauth-openshift-684d58cf9b-xzljc\" (UID: \"7c44c8d7-f038-4b8b-aadc-3ac7434f3b53\") " pod="openshift-authentication/oauth-openshift-684d58cf9b-xzljc" Dec 13 06:53:55 crc kubenswrapper[4971]: I1213 06:53:55.713451 4971 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-system-trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/7c44c8d7-f038-4b8b-aadc-3ac7434f3b53-v4-0-config-system-trusted-ca-bundle\") pod \"oauth-openshift-684d58cf9b-xzljc\" (UID: \"7c44c8d7-f038-4b8b-aadc-3ac7434f3b53\") " pod="openshift-authentication/oauth-openshift-684d58cf9b-xzljc" Dec 13 06:53:55 crc kubenswrapper[4971]: I1213 06:53:55.713486 4971 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-user-idp-0-file-data\" (UniqueName: \"kubernetes.io/secret/7c44c8d7-f038-4b8b-aadc-3ac7434f3b53-v4-0-config-user-idp-0-file-data\") pod \"oauth-openshift-684d58cf9b-xzljc\" (UID: \"7c44c8d7-f038-4b8b-aadc-3ac7434f3b53\") " pod="openshift-authentication/oauth-openshift-684d58cf9b-xzljc" Dec 13 06:53:55 crc kubenswrapper[4971]: I1213 06:53:55.713544 4971 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-user-template-error\" (UniqueName: \"kubernetes.io/secret/7c44c8d7-f038-4b8b-aadc-3ac7434f3b53-v4-0-config-user-template-error\") pod \"oauth-openshift-684d58cf9b-xzljc\" (UID: \"7c44c8d7-f038-4b8b-aadc-3ac7434f3b53\") " pod="openshift-authentication/oauth-openshift-684d58cf9b-xzljc" Dec 13 06:53:55 crc kubenswrapper[4971]: I1213 06:53:55.713587 4971 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/7c44c8d7-f038-4b8b-aadc-3ac7434f3b53-audit-dir\") pod \"oauth-openshift-684d58cf9b-xzljc\" (UID: \"7c44c8d7-f038-4b8b-aadc-3ac7434f3b53\") " pod="openshift-authentication/oauth-openshift-684d58cf9b-xzljc" Dec 13 06:53:55 crc kubenswrapper[4971]: I1213 06:53:55.713621 4971 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-user-template-login\" (UniqueName: \"kubernetes.io/secret/7c44c8d7-f038-4b8b-aadc-3ac7434f3b53-v4-0-config-user-template-login\") pod \"oauth-openshift-684d58cf9b-xzljc\" (UID: \"7c44c8d7-f038-4b8b-aadc-3ac7434f3b53\") " pod="openshift-authentication/oauth-openshift-684d58cf9b-xzljc" Dec 13 06:53:55 crc kubenswrapper[4971]: I1213 06:53:55.713660 4971 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/7c44c8d7-f038-4b8b-aadc-3ac7434f3b53-audit-policies\") pod \"oauth-openshift-684d58cf9b-xzljc\" (UID: \"7c44c8d7-f038-4b8b-aadc-3ac7434f3b53\") " pod="openshift-authentication/oauth-openshift-684d58cf9b-xzljc" Dec 13 06:53:55 crc kubenswrapper[4971]: I1213 06:53:55.714756 4971 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/7c44c8d7-f038-4b8b-aadc-3ac7434f3b53-audit-policies\") pod \"oauth-openshift-684d58cf9b-xzljc\" (UID: \"7c44c8d7-f038-4b8b-aadc-3ac7434f3b53\") " pod="openshift-authentication/oauth-openshift-684d58cf9b-xzljc" Dec 13 06:53:55 crc kubenswrapper[4971]: I1213 06:53:55.716404 4971 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/7c44c8d7-f038-4b8b-aadc-3ac7434f3b53-audit-dir\") pod \"oauth-openshift-684d58cf9b-xzljc\" (UID: \"7c44c8d7-f038-4b8b-aadc-3ac7434f3b53\") " pod="openshift-authentication/oauth-openshift-684d58cf9b-xzljc" Dec 13 06:53:55 crc kubenswrapper[4971]: I1213 06:53:55.716451 4971 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-system-service-ca\" (UniqueName: \"kubernetes.io/configmap/7c44c8d7-f038-4b8b-aadc-3ac7434f3b53-v4-0-config-system-service-ca\") pod \"oauth-openshift-684d58cf9b-xzljc\" (UID: \"7c44c8d7-f038-4b8b-aadc-3ac7434f3b53\") " pod="openshift-authentication/oauth-openshift-684d58cf9b-xzljc" Dec 13 06:53:55 crc kubenswrapper[4971]: I1213 06:53:55.716653 4971 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-system-cliconfig\" (UniqueName: \"kubernetes.io/configmap/7c44c8d7-f038-4b8b-aadc-3ac7434f3b53-v4-0-config-system-cliconfig\") pod \"oauth-openshift-684d58cf9b-xzljc\" (UID: \"7c44c8d7-f038-4b8b-aadc-3ac7434f3b53\") " pod="openshift-authentication/oauth-openshift-684d58cf9b-xzljc" Dec 13 06:53:55 crc kubenswrapper[4971]: I1213 06:53:55.717531 4971 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-system-trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/7c44c8d7-f038-4b8b-aadc-3ac7434f3b53-v4-0-config-system-trusted-ca-bundle\") pod \"oauth-openshift-684d58cf9b-xzljc\" (UID: \"7c44c8d7-f038-4b8b-aadc-3ac7434f3b53\") " pod="openshift-authentication/oauth-openshift-684d58cf9b-xzljc" Dec 13 06:53:55 crc kubenswrapper[4971]: I1213 06:53:55.720831 4971 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-user-template-provider-selection\" (UniqueName: \"kubernetes.io/secret/7c44c8d7-f038-4b8b-aadc-3ac7434f3b53-v4-0-config-user-template-provider-selection\") pod \"oauth-openshift-684d58cf9b-xzljc\" (UID: \"7c44c8d7-f038-4b8b-aadc-3ac7434f3b53\") " pod="openshift-authentication/oauth-openshift-684d58cf9b-xzljc" Dec 13 06:53:55 crc kubenswrapper[4971]: I1213 06:53:55.722065 4971 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-system-session\" (UniqueName: \"kubernetes.io/secret/7c44c8d7-f038-4b8b-aadc-3ac7434f3b53-v4-0-config-system-session\") pod \"oauth-openshift-684d58cf9b-xzljc\" (UID: \"7c44c8d7-f038-4b8b-aadc-3ac7434f3b53\") " pod="openshift-authentication/oauth-openshift-684d58cf9b-xzljc" Dec 13 06:53:55 crc kubenswrapper[4971]: I1213 06:53:55.722194 4971 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-system-ocp-branding-template\" (UniqueName: \"kubernetes.io/secret/7c44c8d7-f038-4b8b-aadc-3ac7434f3b53-v4-0-config-system-ocp-branding-template\") pod \"oauth-openshift-684d58cf9b-xzljc\" (UID: \"7c44c8d7-f038-4b8b-aadc-3ac7434f3b53\") " pod="openshift-authentication/oauth-openshift-684d58cf9b-xzljc" Dec 13 06:53:55 crc kubenswrapper[4971]: I1213 06:53:55.723224 4971 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-user-idp-0-file-data\" (UniqueName: \"kubernetes.io/secret/7c44c8d7-f038-4b8b-aadc-3ac7434f3b53-v4-0-config-user-idp-0-file-data\") pod \"oauth-openshift-684d58cf9b-xzljc\" (UID: \"7c44c8d7-f038-4b8b-aadc-3ac7434f3b53\") " pod="openshift-authentication/oauth-openshift-684d58cf9b-xzljc" Dec 13 06:53:55 crc kubenswrapper[4971]: I1213 06:53:55.724153 4971 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-user-template-login\" (UniqueName: \"kubernetes.io/secret/7c44c8d7-f038-4b8b-aadc-3ac7434f3b53-v4-0-config-user-template-login\") pod \"oauth-openshift-684d58cf9b-xzljc\" (UID: \"7c44c8d7-f038-4b8b-aadc-3ac7434f3b53\") " pod="openshift-authentication/oauth-openshift-684d58cf9b-xzljc" Dec 13 06:53:55 crc kubenswrapper[4971]: I1213 06:53:55.727151 4971 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-system-router-certs\" (UniqueName: \"kubernetes.io/secret/7c44c8d7-f038-4b8b-aadc-3ac7434f3b53-v4-0-config-system-router-certs\") pod \"oauth-openshift-684d58cf9b-xzljc\" (UID: \"7c44c8d7-f038-4b8b-aadc-3ac7434f3b53\") " pod="openshift-authentication/oauth-openshift-684d58cf9b-xzljc" Dec 13 06:53:55 crc kubenswrapper[4971]: I1213 06:53:55.728468 4971 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-cluster-machine-approver"/"machine-approver-sa-dockercfg-nl2j4" Dec 13 06:53:55 crc kubenswrapper[4971]: I1213 06:53:55.733660 4971 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-qt2gv\" (UniqueName: \"kubernetes.io/projected/7c44c8d7-f038-4b8b-aadc-3ac7434f3b53-kube-api-access-qt2gv\") pod \"oauth-openshift-684d58cf9b-xzljc\" (UID: \"7c44c8d7-f038-4b8b-aadc-3ac7434f3b53\") " pod="openshift-authentication/oauth-openshift-684d58cf9b-xzljc" Dec 13 06:53:55 crc kubenswrapper[4971]: I1213 06:53:55.734054 4971 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-user-template-error\" (UniqueName: \"kubernetes.io/secret/7c44c8d7-f038-4b8b-aadc-3ac7434f3b53-v4-0-config-user-template-error\") pod \"oauth-openshift-684d58cf9b-xzljc\" (UID: \"7c44c8d7-f038-4b8b-aadc-3ac7434f3b53\") " pod="openshift-authentication/oauth-openshift-684d58cf9b-xzljc" Dec 13 06:53:55 crc kubenswrapper[4971]: I1213 06:53:55.740998 4971 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-system-serving-cert\" (UniqueName: \"kubernetes.io/secret/7c44c8d7-f038-4b8b-aadc-3ac7434f3b53-v4-0-config-system-serving-cert\") pod \"oauth-openshift-684d58cf9b-xzljc\" (UID: \"7c44c8d7-f038-4b8b-aadc-3ac7434f3b53\") " pod="openshift-authentication/oauth-openshift-684d58cf9b-xzljc" Dec 13 06:53:55 crc kubenswrapper[4971]: I1213 06:53:55.808823 4971 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-operator"/"openshift-service-ca.crt" Dec 13 06:53:55 crc kubenswrapper[4971]: I1213 06:53:55.862958 4971 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-authentication/oauth-openshift-684d58cf9b-xzljc" Dec 13 06:53:55 crc kubenswrapper[4971]: I1213 06:53:55.866008 4971 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-operator"/"iptables-alerter-script" Dec 13 06:53:55 crc kubenswrapper[4971]: I1213 06:53:55.923246 4971 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ingress"/"router-certs-default" Dec 13 06:53:55 crc kubenswrapper[4971]: I1213 06:53:55.941771 4971 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-dns"/"node-resolver-dockercfg-kz9s7" Dec 13 06:53:56 crc kubenswrapper[4971]: I1213 06:53:56.082443 4971 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-authentication/oauth-openshift-684d58cf9b-xzljc"] Dec 13 06:53:56 crc kubenswrapper[4971]: W1213 06:53:56.093729 4971 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod7c44c8d7_f038_4b8b_aadc_3ac7434f3b53.slice/crio-cc6cb8d8da165a574f733b9c52ccba02371e3115fe5c1bc0290a3905f299134d WatchSource:0}: Error finding container cc6cb8d8da165a574f733b9c52ccba02371e3115fe5c1bc0290a3905f299134d: Status 404 returned error can't find the container with id cc6cb8d8da165a574f733b9c52ccba02371e3115fe5c1bc0290a3905f299134d Dec 13 06:53:56 crc kubenswrapper[4971]: I1213 06:53:56.754474 4971 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-authentication/oauth-openshift-684d58cf9b-xzljc" event={"ID":"7c44c8d7-f038-4b8b-aadc-3ac7434f3b53","Type":"ContainerStarted","Data":"0d9c63abce7df5b1567a09b9ad4e0d44c495869036fc5aea35352c689f14aa4f"} Dec 13 06:53:56 crc kubenswrapper[4971]: I1213 06:53:56.755147 4971 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-authentication/oauth-openshift-684d58cf9b-xzljc" event={"ID":"7c44c8d7-f038-4b8b-aadc-3ac7434f3b53","Type":"ContainerStarted","Data":"cc6cb8d8da165a574f733b9c52ccba02371e3115fe5c1bc0290a3905f299134d"} Dec 13 06:53:56 crc kubenswrapper[4971]: I1213 06:53:56.755908 4971 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-authentication/oauth-openshift-684d58cf9b-xzljc" Dec 13 06:53:56 crc kubenswrapper[4971]: I1213 06:53:56.758765 4971 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-kube-apiserver_kube-apiserver-startup-monitor-crc_f85e55b1a89d02b0cb034b1ea31ed45a/startup-monitor/0.log" Dec 13 06:53:56 crc kubenswrapper[4971]: I1213 06:53:56.758813 4971 generic.go:334] "Generic (PLEG): container finished" podID="f85e55b1a89d02b0cb034b1ea31ed45a" containerID="a60d4a360ca22d5ab008969e58dd981a7858882ea44246f6b56a407dba706501" exitCode=137 Dec 13 06:53:56 crc kubenswrapper[4971]: I1213 06:53:56.758847 4971 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="295b500301b45d67ae0a74a2f97f744e423c169d2294d00479c6a454866d548e" Dec 13 06:53:56 crc kubenswrapper[4971]: I1213 06:53:56.775589 4971 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-authentication/oauth-openshift-684d58cf9b-xzljc" podStartSLOduration=52.775572381 podStartE2EDuration="52.775572381s" podCreationTimestamp="2025-12-13 06:53:04 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-13 06:53:56.773638604 +0000 UTC m=+293.378048052" watchObservedRunningTime="2025-12-13 06:53:56.775572381 +0000 UTC m=+293.379981829" Dec 13 06:53:56 crc kubenswrapper[4971]: I1213 06:53:56.788694 4971 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-authentication/oauth-openshift-684d58cf9b-xzljc" Dec 13 06:53:56 crc kubenswrapper[4971]: I1213 06:53:56.798693 4971 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-controller-manager-operator"/"kube-controller-manager-operator-dockercfg-gkqpw" Dec 13 06:53:57 crc kubenswrapper[4971]: I1213 06:53:57.126620 4971 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console-operator"/"kube-root-ca.crt" Dec 13 06:53:57 crc kubenswrapper[4971]: I1213 06:53:57.130410 4971 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-image-registry"/"registry-dockercfg-kzzsd" Dec 13 06:53:57 crc kubenswrapper[4971]: I1213 06:53:57.157767 4971 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-controller-manager"/"serving-cert" Dec 13 06:53:57 crc kubenswrapper[4971]: I1213 06:53:57.544341 4971 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-kube-apiserver_kube-apiserver-startup-monitor-crc_f85e55b1a89d02b0cb034b1ea31ed45a/startup-monitor/0.log" Dec 13 06:53:57 crc kubenswrapper[4971]: I1213 06:53:57.544434 4971 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" Dec 13 06:53:57 crc kubenswrapper[4971]: I1213 06:53:57.637248 4971 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"var-log\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-var-log\") pod \"f85e55b1a89d02b0cb034b1ea31ed45a\" (UID: \"f85e55b1a89d02b0cb034b1ea31ed45a\") " Dec 13 06:53:57 crc kubenswrapper[4971]: I1213 06:53:57.637382 4971 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"var-lock\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-var-lock\") pod \"f85e55b1a89d02b0cb034b1ea31ed45a\" (UID: \"f85e55b1a89d02b0cb034b1ea31ed45a\") " Dec 13 06:53:57 crc kubenswrapper[4971]: I1213 06:53:57.637425 4971 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pod-resource-dir\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-pod-resource-dir\") pod \"f85e55b1a89d02b0cb034b1ea31ed45a\" (UID: \"f85e55b1a89d02b0cb034b1ea31ed45a\") " Dec 13 06:53:57 crc kubenswrapper[4971]: I1213 06:53:57.637442 4971 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-resource-dir\") pod \"f85e55b1a89d02b0cb034b1ea31ed45a\" (UID: \"f85e55b1a89d02b0cb034b1ea31ed45a\") " Dec 13 06:53:57 crc kubenswrapper[4971]: I1213 06:53:57.637461 4971 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"manifests\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-manifests\") pod \"f85e55b1a89d02b0cb034b1ea31ed45a\" (UID: \"f85e55b1a89d02b0cb034b1ea31ed45a\") " Dec 13 06:53:57 crc kubenswrapper[4971]: I1213 06:53:57.637771 4971 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-manifests" (OuterVolumeSpecName: "manifests") pod "f85e55b1a89d02b0cb034b1ea31ed45a" (UID: "f85e55b1a89d02b0cb034b1ea31ed45a"). InnerVolumeSpecName "manifests". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 13 06:53:57 crc kubenswrapper[4971]: I1213 06:53:57.637802 4971 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-var-log" (OuterVolumeSpecName: "var-log") pod "f85e55b1a89d02b0cb034b1ea31ed45a" (UID: "f85e55b1a89d02b0cb034b1ea31ed45a"). InnerVolumeSpecName "var-log". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 13 06:53:57 crc kubenswrapper[4971]: I1213 06:53:57.637820 4971 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-var-lock" (OuterVolumeSpecName: "var-lock") pod "f85e55b1a89d02b0cb034b1ea31ed45a" (UID: "f85e55b1a89d02b0cb034b1ea31ed45a"). InnerVolumeSpecName "var-lock". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 13 06:53:57 crc kubenswrapper[4971]: I1213 06:53:57.638490 4971 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-resource-dir" (OuterVolumeSpecName: "resource-dir") pod "f85e55b1a89d02b0cb034b1ea31ed45a" (UID: "f85e55b1a89d02b0cb034b1ea31ed45a"). InnerVolumeSpecName "resource-dir". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 13 06:53:57 crc kubenswrapper[4971]: I1213 06:53:57.644319 4971 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-pod-resource-dir" (OuterVolumeSpecName: "pod-resource-dir") pod "f85e55b1a89d02b0cb034b1ea31ed45a" (UID: "f85e55b1a89d02b0cb034b1ea31ed45a"). InnerVolumeSpecName "pod-resource-dir". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 13 06:53:57 crc kubenswrapper[4971]: I1213 06:53:57.738660 4971 reconciler_common.go:293] "Volume detached for volume \"var-lock\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-var-lock\") on node \"crc\" DevicePath \"\"" Dec 13 06:53:57 crc kubenswrapper[4971]: I1213 06:53:57.738710 4971 reconciler_common.go:293] "Volume detached for volume \"pod-resource-dir\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-pod-resource-dir\") on node \"crc\" DevicePath \"\"" Dec 13 06:53:57 crc kubenswrapper[4971]: I1213 06:53:57.738725 4971 reconciler_common.go:293] "Volume detached for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-resource-dir\") on node \"crc\" DevicePath \"\"" Dec 13 06:53:57 crc kubenswrapper[4971]: I1213 06:53:57.738736 4971 reconciler_common.go:293] "Volume detached for volume \"manifests\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-manifests\") on node \"crc\" DevicePath \"\"" Dec 13 06:53:57 crc kubenswrapper[4971]: I1213 06:53:57.738749 4971 reconciler_common.go:293] "Volume detached for volume \"var-log\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-var-log\") on node \"crc\" DevicePath \"\"" Dec 13 06:53:57 crc kubenswrapper[4971]: I1213 06:53:57.763698 4971 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" Dec 13 06:53:57 crc kubenswrapper[4971]: I1213 06:53:57.774274 4971 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="f85e55b1a89d02b0cb034b1ea31ed45a" path="/var/lib/kubelet/pods/f85e55b1a89d02b0cb034b1ea31ed45a/volumes" Dec 13 06:53:57 crc kubenswrapper[4971]: I1213 06:53:57.774580 4971 mirror_client.go:130] "Deleting a mirror pod" pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" podUID="" Dec 13 06:53:57 crc kubenswrapper[4971]: I1213 06:53:57.785787 4971 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-kube-apiserver/kube-apiserver-startup-monitor-crc"] Dec 13 06:53:57 crc kubenswrapper[4971]: I1213 06:53:57.785824 4971 kubelet.go:2649] "Unable to find pod for mirror pod, skipping" mirrorPod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" mirrorPodUID="d319392a-a3da-4bf4-a4b7-b33567d1bce2" Dec 13 06:53:57 crc kubenswrapper[4971]: I1213 06:53:57.789848 4971 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-kube-apiserver/kube-apiserver-startup-monitor-crc"] Dec 13 06:53:57 crc kubenswrapper[4971]: I1213 06:53:57.789899 4971 kubelet.go:2673] "Unable to find pod for mirror pod, skipping" mirrorPod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" mirrorPodUID="d319392a-a3da-4bf4-a4b7-b33567d1bce2" Dec 13 06:53:58 crc kubenswrapper[4971]: I1213 06:53:58.898812 4971 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"packageserver-service-cert" Dec 13 06:54:16 crc kubenswrapper[4971]: I1213 06:54:16.856347 4971 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-kube-controller-manager_kube-controller-manager-crc_f614b9022728cf315e60c057852e563e/kube-controller-manager/1.log" Dec 13 06:54:16 crc kubenswrapper[4971]: I1213 06:54:16.858336 4971 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-kube-controller-manager_kube-controller-manager-crc_f614b9022728cf315e60c057852e563e/kube-controller-manager/0.log" Dec 13 06:54:16 crc kubenswrapper[4971]: I1213 06:54:16.858395 4971 generic.go:334] "Generic (PLEG): container finished" podID="f614b9022728cf315e60c057852e563e" containerID="65bdca95a7c288de60f6ad348f489507c045b5ffcbae188538f47763c1f891c2" exitCode=137 Dec 13 06:54:16 crc kubenswrapper[4971]: I1213 06:54:16.858436 4971 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" event={"ID":"f614b9022728cf315e60c057852e563e","Type":"ContainerDied","Data":"65bdca95a7c288de60f6ad348f489507c045b5ffcbae188538f47763c1f891c2"} Dec 13 06:54:16 crc kubenswrapper[4971]: I1213 06:54:16.858482 4971 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" event={"ID":"f614b9022728cf315e60c057852e563e","Type":"ContainerStarted","Data":"0b0c570f8f985c93e4e5ab170c4b7c3fd4cf2e1842a54ac64afbfca3ead108bb"} Dec 13 06:54:16 crc kubenswrapper[4971]: I1213 06:54:16.858508 4971 scope.go:117] "RemoveContainer" containerID="c788de86754f0bf37ee184672fc1dcb388ccebc246d4f15be4b5da661f725f8f" Dec 13 06:54:17 crc kubenswrapper[4971]: I1213 06:54:17.705278 4971 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-kube-controller-manager/kube-controller-manager-crc" Dec 13 06:54:17 crc kubenswrapper[4971]: I1213 06:54:17.866494 4971 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-kube-controller-manager_kube-controller-manager-crc_f614b9022728cf315e60c057852e563e/kube-controller-manager/1.log" Dec 13 06:54:25 crc kubenswrapper[4971]: I1213 06:54:25.780029 4971 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-kube-controller-manager/kube-controller-manager-crc" Dec 13 06:54:25 crc kubenswrapper[4971]: I1213 06:54:25.783380 4971 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-kube-controller-manager/kube-controller-manager-crc" Dec 13 06:54:27 crc kubenswrapper[4971]: I1213 06:54:27.718834 4971 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-kube-controller-manager/kube-controller-manager-crc" Dec 13 06:54:34 crc kubenswrapper[4971]: I1213 06:54:34.838436 4971 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-route-controller-manager/route-controller-manager-6576b87f9c-9d86n"] Dec 13 06:54:34 crc kubenswrapper[4971]: I1213 06:54:34.838967 4971 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-9d86n" podUID="2cf6e363-8511-4c79-9181-79dba726e5af" containerName="route-controller-manager" containerID="cri-o://6ccbca076f1fdb45dc7a3f3efdf6a363f8806ecdef5e1047dc12136783d5c199" gracePeriod=30 Dec 13 06:54:34 crc kubenswrapper[4971]: I1213 06:54:34.931469 4971 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-controller-manager/controller-manager-879f6c89f-xtxr8"] Dec 13 06:54:34 crc kubenswrapper[4971]: I1213 06:54:34.932021 4971 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-controller-manager/controller-manager-879f6c89f-xtxr8" podUID="7148e573-e2f5-4fc5-b30c-4e777516509e" containerName="controller-manager" containerID="cri-o://07720aa0b0d810dbae552adbd6b261970ba16c7af295ef016ea351b90dade542" gracePeriod=30 Dec 13 06:54:35 crc kubenswrapper[4971]: I1213 06:54:35.307290 4971 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-controller-manager/controller-manager-879f6c89f-xtxr8" Dec 13 06:54:35 crc kubenswrapper[4971]: I1213 06:54:35.405381 4971 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/7148e573-e2f5-4fc5-b30c-4e777516509e-serving-cert\") pod \"7148e573-e2f5-4fc5-b30c-4e777516509e\" (UID: \"7148e573-e2f5-4fc5-b30c-4e777516509e\") " Dec 13 06:54:35 crc kubenswrapper[4971]: I1213 06:54:35.405501 4971 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/7148e573-e2f5-4fc5-b30c-4e777516509e-client-ca\") pod \"7148e573-e2f5-4fc5-b30c-4e777516509e\" (UID: \"7148e573-e2f5-4fc5-b30c-4e777516509e\") " Dec 13 06:54:35 crc kubenswrapper[4971]: I1213 06:54:35.405542 4971 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/7148e573-e2f5-4fc5-b30c-4e777516509e-config\") pod \"7148e573-e2f5-4fc5-b30c-4e777516509e\" (UID: \"7148e573-e2f5-4fc5-b30c-4e777516509e\") " Dec 13 06:54:35 crc kubenswrapper[4971]: I1213 06:54:35.405570 4971 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-lmhvx\" (UniqueName: \"kubernetes.io/projected/7148e573-e2f5-4fc5-b30c-4e777516509e-kube-api-access-lmhvx\") pod \"7148e573-e2f5-4fc5-b30c-4e777516509e\" (UID: \"7148e573-e2f5-4fc5-b30c-4e777516509e\") " Dec 13 06:54:35 crc kubenswrapper[4971]: I1213 06:54:35.406319 4971 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/7148e573-e2f5-4fc5-b30c-4e777516509e-client-ca" (OuterVolumeSpecName: "client-ca") pod "7148e573-e2f5-4fc5-b30c-4e777516509e" (UID: "7148e573-e2f5-4fc5-b30c-4e777516509e"). InnerVolumeSpecName "client-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 13 06:54:35 crc kubenswrapper[4971]: I1213 06:54:35.406399 4971 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/7148e573-e2f5-4fc5-b30c-4e777516509e-config" (OuterVolumeSpecName: "config") pod "7148e573-e2f5-4fc5-b30c-4e777516509e" (UID: "7148e573-e2f5-4fc5-b30c-4e777516509e"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 13 06:54:35 crc kubenswrapper[4971]: I1213 06:54:35.406459 4971 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/7148e573-e2f5-4fc5-b30c-4e777516509e-proxy-ca-bundles\") pod \"7148e573-e2f5-4fc5-b30c-4e777516509e\" (UID: \"7148e573-e2f5-4fc5-b30c-4e777516509e\") " Dec 13 06:54:35 crc kubenswrapper[4971]: I1213 06:54:35.406969 4971 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/7148e573-e2f5-4fc5-b30c-4e777516509e-proxy-ca-bundles" (OuterVolumeSpecName: "proxy-ca-bundles") pod "7148e573-e2f5-4fc5-b30c-4e777516509e" (UID: "7148e573-e2f5-4fc5-b30c-4e777516509e"). InnerVolumeSpecName "proxy-ca-bundles". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 13 06:54:35 crc kubenswrapper[4971]: I1213 06:54:35.407207 4971 reconciler_common.go:293] "Volume detached for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/7148e573-e2f5-4fc5-b30c-4e777516509e-client-ca\") on node \"crc\" DevicePath \"\"" Dec 13 06:54:35 crc kubenswrapper[4971]: I1213 06:54:35.407221 4971 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/7148e573-e2f5-4fc5-b30c-4e777516509e-config\") on node \"crc\" DevicePath \"\"" Dec 13 06:54:35 crc kubenswrapper[4971]: I1213 06:54:35.407232 4971 reconciler_common.go:293] "Volume detached for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/7148e573-e2f5-4fc5-b30c-4e777516509e-proxy-ca-bundles\") on node \"crc\" DevicePath \"\"" Dec 13 06:54:35 crc kubenswrapper[4971]: I1213 06:54:35.411303 4971 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/7148e573-e2f5-4fc5-b30c-4e777516509e-kube-api-access-lmhvx" (OuterVolumeSpecName: "kube-api-access-lmhvx") pod "7148e573-e2f5-4fc5-b30c-4e777516509e" (UID: "7148e573-e2f5-4fc5-b30c-4e777516509e"). InnerVolumeSpecName "kube-api-access-lmhvx". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 13 06:54:35 crc kubenswrapper[4971]: I1213 06:54:35.411323 4971 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/7148e573-e2f5-4fc5-b30c-4e777516509e-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "7148e573-e2f5-4fc5-b30c-4e777516509e" (UID: "7148e573-e2f5-4fc5-b30c-4e777516509e"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 13 06:54:35 crc kubenswrapper[4971]: I1213 06:54:35.507920 4971 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-lmhvx\" (UniqueName: \"kubernetes.io/projected/7148e573-e2f5-4fc5-b30c-4e777516509e-kube-api-access-lmhvx\") on node \"crc\" DevicePath \"\"" Dec 13 06:54:35 crc kubenswrapper[4971]: I1213 06:54:35.507953 4971 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/7148e573-e2f5-4fc5-b30c-4e777516509e-serving-cert\") on node \"crc\" DevicePath \"\"" Dec 13 06:54:35 crc kubenswrapper[4971]: I1213 06:54:35.626774 4971 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-9d86n" Dec 13 06:54:35 crc kubenswrapper[4971]: I1213 06:54:35.709007 4971 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/2cf6e363-8511-4c79-9181-79dba726e5af-client-ca\") pod \"2cf6e363-8511-4c79-9181-79dba726e5af\" (UID: \"2cf6e363-8511-4c79-9181-79dba726e5af\") " Dec 13 06:54:35 crc kubenswrapper[4971]: I1213 06:54:35.709386 4971 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-mkgwg\" (UniqueName: \"kubernetes.io/projected/2cf6e363-8511-4c79-9181-79dba726e5af-kube-api-access-mkgwg\") pod \"2cf6e363-8511-4c79-9181-79dba726e5af\" (UID: \"2cf6e363-8511-4c79-9181-79dba726e5af\") " Dec 13 06:54:35 crc kubenswrapper[4971]: I1213 06:54:35.709460 4971 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/2cf6e363-8511-4c79-9181-79dba726e5af-config\") pod \"2cf6e363-8511-4c79-9181-79dba726e5af\" (UID: \"2cf6e363-8511-4c79-9181-79dba726e5af\") " Dec 13 06:54:35 crc kubenswrapper[4971]: I1213 06:54:35.709546 4971 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/2cf6e363-8511-4c79-9181-79dba726e5af-serving-cert\") pod \"2cf6e363-8511-4c79-9181-79dba726e5af\" (UID: \"2cf6e363-8511-4c79-9181-79dba726e5af\") " Dec 13 06:54:35 crc kubenswrapper[4971]: I1213 06:54:35.709883 4971 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/2cf6e363-8511-4c79-9181-79dba726e5af-client-ca" (OuterVolumeSpecName: "client-ca") pod "2cf6e363-8511-4c79-9181-79dba726e5af" (UID: "2cf6e363-8511-4c79-9181-79dba726e5af"). InnerVolumeSpecName "client-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 13 06:54:35 crc kubenswrapper[4971]: I1213 06:54:35.712407 4971 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/2cf6e363-8511-4c79-9181-79dba726e5af-config" (OuterVolumeSpecName: "config") pod "2cf6e363-8511-4c79-9181-79dba726e5af" (UID: "2cf6e363-8511-4c79-9181-79dba726e5af"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 13 06:54:35 crc kubenswrapper[4971]: I1213 06:54:35.713565 4971 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/2cf6e363-8511-4c79-9181-79dba726e5af-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "2cf6e363-8511-4c79-9181-79dba726e5af" (UID: "2cf6e363-8511-4c79-9181-79dba726e5af"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 13 06:54:35 crc kubenswrapper[4971]: I1213 06:54:35.715485 4971 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/2cf6e363-8511-4c79-9181-79dba726e5af-kube-api-access-mkgwg" (OuterVolumeSpecName: "kube-api-access-mkgwg") pod "2cf6e363-8511-4c79-9181-79dba726e5af" (UID: "2cf6e363-8511-4c79-9181-79dba726e5af"). InnerVolumeSpecName "kube-api-access-mkgwg". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 13 06:54:35 crc kubenswrapper[4971]: I1213 06:54:35.811015 4971 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/2cf6e363-8511-4c79-9181-79dba726e5af-config\") on node \"crc\" DevicePath \"\"" Dec 13 06:54:35 crc kubenswrapper[4971]: I1213 06:54:35.811056 4971 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/2cf6e363-8511-4c79-9181-79dba726e5af-serving-cert\") on node \"crc\" DevicePath \"\"" Dec 13 06:54:35 crc kubenswrapper[4971]: I1213 06:54:35.811070 4971 reconciler_common.go:293] "Volume detached for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/2cf6e363-8511-4c79-9181-79dba726e5af-client-ca\") on node \"crc\" DevicePath \"\"" Dec 13 06:54:35 crc kubenswrapper[4971]: I1213 06:54:35.811082 4971 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-mkgwg\" (UniqueName: \"kubernetes.io/projected/2cf6e363-8511-4c79-9181-79dba726e5af-kube-api-access-mkgwg\") on node \"crc\" DevicePath \"\"" Dec 13 06:54:35 crc kubenswrapper[4971]: I1213 06:54:35.951494 4971 generic.go:334] "Generic (PLEG): container finished" podID="2cf6e363-8511-4c79-9181-79dba726e5af" containerID="6ccbca076f1fdb45dc7a3f3efdf6a363f8806ecdef5e1047dc12136783d5c199" exitCode=0 Dec 13 06:54:35 crc kubenswrapper[4971]: I1213 06:54:35.951543 4971 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-9d86n" Dec 13 06:54:35 crc kubenswrapper[4971]: I1213 06:54:35.951581 4971 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-9d86n" event={"ID":"2cf6e363-8511-4c79-9181-79dba726e5af","Type":"ContainerDied","Data":"6ccbca076f1fdb45dc7a3f3efdf6a363f8806ecdef5e1047dc12136783d5c199"} Dec 13 06:54:35 crc kubenswrapper[4971]: I1213 06:54:35.951606 4971 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-9d86n" event={"ID":"2cf6e363-8511-4c79-9181-79dba726e5af","Type":"ContainerDied","Data":"5fa0eec368450ce9c968373d6851400be728d4c59120d3368fa30d48866623e7"} Dec 13 06:54:35 crc kubenswrapper[4971]: I1213 06:54:35.951623 4971 scope.go:117] "RemoveContainer" containerID="6ccbca076f1fdb45dc7a3f3efdf6a363f8806ecdef5e1047dc12136783d5c199" Dec 13 06:54:35 crc kubenswrapper[4971]: I1213 06:54:35.953456 4971 generic.go:334] "Generic (PLEG): container finished" podID="7148e573-e2f5-4fc5-b30c-4e777516509e" containerID="07720aa0b0d810dbae552adbd6b261970ba16c7af295ef016ea351b90dade542" exitCode=0 Dec 13 06:54:35 crc kubenswrapper[4971]: I1213 06:54:35.953480 4971 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-controller-manager/controller-manager-879f6c89f-xtxr8" event={"ID":"7148e573-e2f5-4fc5-b30c-4e777516509e","Type":"ContainerDied","Data":"07720aa0b0d810dbae552adbd6b261970ba16c7af295ef016ea351b90dade542"} Dec 13 06:54:35 crc kubenswrapper[4971]: I1213 06:54:35.953498 4971 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-controller-manager/controller-manager-879f6c89f-xtxr8" event={"ID":"7148e573-e2f5-4fc5-b30c-4e777516509e","Type":"ContainerDied","Data":"45ca28d8f97bfc12cd2dcba8da0fabcb8a352f2d2188c0d53d940710d41bbb65"} Dec 13 06:54:35 crc kubenswrapper[4971]: I1213 06:54:35.953562 4971 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-controller-manager/controller-manager-879f6c89f-xtxr8" Dec 13 06:54:35 crc kubenswrapper[4971]: I1213 06:54:35.983368 4971 scope.go:117] "RemoveContainer" containerID="6ccbca076f1fdb45dc7a3f3efdf6a363f8806ecdef5e1047dc12136783d5c199" Dec 13 06:54:35 crc kubenswrapper[4971]: E1213 06:54:35.983763 4971 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"6ccbca076f1fdb45dc7a3f3efdf6a363f8806ecdef5e1047dc12136783d5c199\": container with ID starting with 6ccbca076f1fdb45dc7a3f3efdf6a363f8806ecdef5e1047dc12136783d5c199 not found: ID does not exist" containerID="6ccbca076f1fdb45dc7a3f3efdf6a363f8806ecdef5e1047dc12136783d5c199" Dec 13 06:54:35 crc kubenswrapper[4971]: I1213 06:54:35.983794 4971 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"6ccbca076f1fdb45dc7a3f3efdf6a363f8806ecdef5e1047dc12136783d5c199"} err="failed to get container status \"6ccbca076f1fdb45dc7a3f3efdf6a363f8806ecdef5e1047dc12136783d5c199\": rpc error: code = NotFound desc = could not find container \"6ccbca076f1fdb45dc7a3f3efdf6a363f8806ecdef5e1047dc12136783d5c199\": container with ID starting with 6ccbca076f1fdb45dc7a3f3efdf6a363f8806ecdef5e1047dc12136783d5c199 not found: ID does not exist" Dec 13 06:54:35 crc kubenswrapper[4971]: I1213 06:54:35.983821 4971 scope.go:117] "RemoveContainer" containerID="07720aa0b0d810dbae552adbd6b261970ba16c7af295ef016ea351b90dade542" Dec 13 06:54:35 crc kubenswrapper[4971]: I1213 06:54:35.985166 4971 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-route-controller-manager/route-controller-manager-6576b87f9c-9d86n"] Dec 13 06:54:35 crc kubenswrapper[4971]: I1213 06:54:35.996968 4971 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-route-controller-manager/route-controller-manager-6576b87f9c-9d86n"] Dec 13 06:54:35 crc kubenswrapper[4971]: I1213 06:54:35.998994 4971 scope.go:117] "RemoveContainer" containerID="07720aa0b0d810dbae552adbd6b261970ba16c7af295ef016ea351b90dade542" Dec 13 06:54:36 crc kubenswrapper[4971]: E1213 06:54:36.000021 4971 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"07720aa0b0d810dbae552adbd6b261970ba16c7af295ef016ea351b90dade542\": container with ID starting with 07720aa0b0d810dbae552adbd6b261970ba16c7af295ef016ea351b90dade542 not found: ID does not exist" containerID="07720aa0b0d810dbae552adbd6b261970ba16c7af295ef016ea351b90dade542" Dec 13 06:54:36 crc kubenswrapper[4971]: I1213 06:54:36.000070 4971 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"07720aa0b0d810dbae552adbd6b261970ba16c7af295ef016ea351b90dade542"} err="failed to get container status \"07720aa0b0d810dbae552adbd6b261970ba16c7af295ef016ea351b90dade542\": rpc error: code = NotFound desc = could not find container \"07720aa0b0d810dbae552adbd6b261970ba16c7af295ef016ea351b90dade542\": container with ID starting with 07720aa0b0d810dbae552adbd6b261970ba16c7af295ef016ea351b90dade542 not found: ID does not exist" Dec 13 06:54:36 crc kubenswrapper[4971]: I1213 06:54:36.001908 4971 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-controller-manager/controller-manager-879f6c89f-xtxr8"] Dec 13 06:54:36 crc kubenswrapper[4971]: I1213 06:54:36.005315 4971 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-controller-manager/controller-manager-879f6c89f-xtxr8"] Dec 13 06:54:36 crc kubenswrapper[4971]: I1213 06:54:36.545902 4971 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-route-controller-manager/route-controller-manager-5dcf5cc8c9-j7frp"] Dec 13 06:54:36 crc kubenswrapper[4971]: E1213 06:54:36.546171 4971 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="2cf6e363-8511-4c79-9181-79dba726e5af" containerName="route-controller-manager" Dec 13 06:54:36 crc kubenswrapper[4971]: I1213 06:54:36.546186 4971 state_mem.go:107] "Deleted CPUSet assignment" podUID="2cf6e363-8511-4c79-9181-79dba726e5af" containerName="route-controller-manager" Dec 13 06:54:36 crc kubenswrapper[4971]: E1213 06:54:36.546195 4971 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="7148e573-e2f5-4fc5-b30c-4e777516509e" containerName="controller-manager" Dec 13 06:54:36 crc kubenswrapper[4971]: I1213 06:54:36.546202 4971 state_mem.go:107] "Deleted CPUSet assignment" podUID="7148e573-e2f5-4fc5-b30c-4e777516509e" containerName="controller-manager" Dec 13 06:54:36 crc kubenswrapper[4971]: I1213 06:54:36.546280 4971 memory_manager.go:354] "RemoveStaleState removing state" podUID="7148e573-e2f5-4fc5-b30c-4e777516509e" containerName="controller-manager" Dec 13 06:54:36 crc kubenswrapper[4971]: I1213 06:54:36.546295 4971 memory_manager.go:354] "RemoveStaleState removing state" podUID="2cf6e363-8511-4c79-9181-79dba726e5af" containerName="route-controller-manager" Dec 13 06:54:36 crc kubenswrapper[4971]: I1213 06:54:36.546656 4971 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-route-controller-manager/route-controller-manager-5dcf5cc8c9-j7frp" Dec 13 06:54:36 crc kubenswrapper[4971]: I1213 06:54:36.549265 4971 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-controller-manager/controller-manager-cd9f4966d-ljtrh"] Dec 13 06:54:36 crc kubenswrapper[4971]: I1213 06:54:36.549825 4971 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-controller-manager/controller-manager-cd9f4966d-ljtrh" Dec 13 06:54:36 crc kubenswrapper[4971]: I1213 06:54:36.550960 4971 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-route-controller-manager"/"kube-root-ca.crt" Dec 13 06:54:36 crc kubenswrapper[4971]: I1213 06:54:36.551116 4971 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-route-controller-manager"/"serving-cert" Dec 13 06:54:36 crc kubenswrapper[4971]: I1213 06:54:36.551230 4971 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-route-controller-manager"/"openshift-service-ca.crt" Dec 13 06:54:36 crc kubenswrapper[4971]: I1213 06:54:36.551309 4971 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-route-controller-manager"/"config" Dec 13 06:54:36 crc kubenswrapper[4971]: I1213 06:54:36.551454 4971 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-route-controller-manager"/"route-controller-manager-sa-dockercfg-h2zr2" Dec 13 06:54:36 crc kubenswrapper[4971]: I1213 06:54:36.551810 4971 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-route-controller-manager"/"client-ca" Dec 13 06:54:36 crc kubenswrapper[4971]: I1213 06:54:36.553007 4971 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"config" Dec 13 06:54:36 crc kubenswrapper[4971]: I1213 06:54:36.553142 4971 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-controller-manager"/"openshift-controller-manager-sa-dockercfg-msq4c" Dec 13 06:54:36 crc kubenswrapper[4971]: I1213 06:54:36.553227 4971 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"kube-root-ca.crt" Dec 13 06:54:36 crc kubenswrapper[4971]: I1213 06:54:36.553505 4971 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-controller-manager"/"serving-cert" Dec 13 06:54:36 crc kubenswrapper[4971]: I1213 06:54:36.553598 4971 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"client-ca" Dec 13 06:54:36 crc kubenswrapper[4971]: I1213 06:54:36.555311 4971 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"openshift-service-ca.crt" Dec 13 06:54:36 crc kubenswrapper[4971]: I1213 06:54:36.559986 4971 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-route-controller-manager/route-controller-manager-5dcf5cc8c9-j7frp"] Dec 13 06:54:36 crc kubenswrapper[4971]: I1213 06:54:36.562279 4971 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"openshift-global-ca" Dec 13 06:54:36 crc kubenswrapper[4971]: I1213 06:54:36.564161 4971 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-controller-manager/controller-manager-cd9f4966d-ljtrh"] Dec 13 06:54:36 crc kubenswrapper[4971]: I1213 06:54:36.620671 4971 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-wdpt8\" (UniqueName: \"kubernetes.io/projected/ed96d9f8-2ba9-4f6e-b759-4c7c607801c2-kube-api-access-wdpt8\") pod \"route-controller-manager-5dcf5cc8c9-j7frp\" (UID: \"ed96d9f8-2ba9-4f6e-b759-4c7c607801c2\") " pod="openshift-route-controller-manager/route-controller-manager-5dcf5cc8c9-j7frp" Dec 13 06:54:36 crc kubenswrapper[4971]: I1213 06:54:36.620716 4971 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/5cccf3e0-5335-44fc-a685-df1f5d46554d-proxy-ca-bundles\") pod \"controller-manager-cd9f4966d-ljtrh\" (UID: \"5cccf3e0-5335-44fc-a685-df1f5d46554d\") " pod="openshift-controller-manager/controller-manager-cd9f4966d-ljtrh" Dec 13 06:54:36 crc kubenswrapper[4971]: I1213 06:54:36.620745 4971 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/ed96d9f8-2ba9-4f6e-b759-4c7c607801c2-client-ca\") pod \"route-controller-manager-5dcf5cc8c9-j7frp\" (UID: \"ed96d9f8-2ba9-4f6e-b759-4c7c607801c2\") " pod="openshift-route-controller-manager/route-controller-manager-5dcf5cc8c9-j7frp" Dec 13 06:54:36 crc kubenswrapper[4971]: I1213 06:54:36.620770 4971 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-44spt\" (UniqueName: \"kubernetes.io/projected/5cccf3e0-5335-44fc-a685-df1f5d46554d-kube-api-access-44spt\") pod \"controller-manager-cd9f4966d-ljtrh\" (UID: \"5cccf3e0-5335-44fc-a685-df1f5d46554d\") " pod="openshift-controller-manager/controller-manager-cd9f4966d-ljtrh" Dec 13 06:54:36 crc kubenswrapper[4971]: I1213 06:54:36.620885 4971 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/5cccf3e0-5335-44fc-a685-df1f5d46554d-client-ca\") pod \"controller-manager-cd9f4966d-ljtrh\" (UID: \"5cccf3e0-5335-44fc-a685-df1f5d46554d\") " pod="openshift-controller-manager/controller-manager-cd9f4966d-ljtrh" Dec 13 06:54:36 crc kubenswrapper[4971]: I1213 06:54:36.620916 4971 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/5cccf3e0-5335-44fc-a685-df1f5d46554d-serving-cert\") pod \"controller-manager-cd9f4966d-ljtrh\" (UID: \"5cccf3e0-5335-44fc-a685-df1f5d46554d\") " pod="openshift-controller-manager/controller-manager-cd9f4966d-ljtrh" Dec 13 06:54:36 crc kubenswrapper[4971]: I1213 06:54:36.620947 4971 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/ed96d9f8-2ba9-4f6e-b759-4c7c607801c2-serving-cert\") pod \"route-controller-manager-5dcf5cc8c9-j7frp\" (UID: \"ed96d9f8-2ba9-4f6e-b759-4c7c607801c2\") " pod="openshift-route-controller-manager/route-controller-manager-5dcf5cc8c9-j7frp" Dec 13 06:54:36 crc kubenswrapper[4971]: I1213 06:54:36.620994 4971 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/ed96d9f8-2ba9-4f6e-b759-4c7c607801c2-config\") pod \"route-controller-manager-5dcf5cc8c9-j7frp\" (UID: \"ed96d9f8-2ba9-4f6e-b759-4c7c607801c2\") " pod="openshift-route-controller-manager/route-controller-manager-5dcf5cc8c9-j7frp" Dec 13 06:54:36 crc kubenswrapper[4971]: I1213 06:54:36.621020 4971 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/5cccf3e0-5335-44fc-a685-df1f5d46554d-config\") pod \"controller-manager-cd9f4966d-ljtrh\" (UID: \"5cccf3e0-5335-44fc-a685-df1f5d46554d\") " pod="openshift-controller-manager/controller-manager-cd9f4966d-ljtrh" Dec 13 06:54:36 crc kubenswrapper[4971]: I1213 06:54:36.721763 4971 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/5cccf3e0-5335-44fc-a685-df1f5d46554d-client-ca\") pod \"controller-manager-cd9f4966d-ljtrh\" (UID: \"5cccf3e0-5335-44fc-a685-df1f5d46554d\") " pod="openshift-controller-manager/controller-manager-cd9f4966d-ljtrh" Dec 13 06:54:36 crc kubenswrapper[4971]: I1213 06:54:36.721800 4971 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/5cccf3e0-5335-44fc-a685-df1f5d46554d-serving-cert\") pod \"controller-manager-cd9f4966d-ljtrh\" (UID: \"5cccf3e0-5335-44fc-a685-df1f5d46554d\") " pod="openshift-controller-manager/controller-manager-cd9f4966d-ljtrh" Dec 13 06:54:36 crc kubenswrapper[4971]: I1213 06:54:36.721815 4971 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/ed96d9f8-2ba9-4f6e-b759-4c7c607801c2-serving-cert\") pod \"route-controller-manager-5dcf5cc8c9-j7frp\" (UID: \"ed96d9f8-2ba9-4f6e-b759-4c7c607801c2\") " pod="openshift-route-controller-manager/route-controller-manager-5dcf5cc8c9-j7frp" Dec 13 06:54:36 crc kubenswrapper[4971]: I1213 06:54:36.721847 4971 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/ed96d9f8-2ba9-4f6e-b759-4c7c607801c2-config\") pod \"route-controller-manager-5dcf5cc8c9-j7frp\" (UID: \"ed96d9f8-2ba9-4f6e-b759-4c7c607801c2\") " pod="openshift-route-controller-manager/route-controller-manager-5dcf5cc8c9-j7frp" Dec 13 06:54:36 crc kubenswrapper[4971]: I1213 06:54:36.721865 4971 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/5cccf3e0-5335-44fc-a685-df1f5d46554d-config\") pod \"controller-manager-cd9f4966d-ljtrh\" (UID: \"5cccf3e0-5335-44fc-a685-df1f5d46554d\") " pod="openshift-controller-manager/controller-manager-cd9f4966d-ljtrh" Dec 13 06:54:36 crc kubenswrapper[4971]: I1213 06:54:36.721886 4971 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-wdpt8\" (UniqueName: \"kubernetes.io/projected/ed96d9f8-2ba9-4f6e-b759-4c7c607801c2-kube-api-access-wdpt8\") pod \"route-controller-manager-5dcf5cc8c9-j7frp\" (UID: \"ed96d9f8-2ba9-4f6e-b759-4c7c607801c2\") " pod="openshift-route-controller-manager/route-controller-manager-5dcf5cc8c9-j7frp" Dec 13 06:54:36 crc kubenswrapper[4971]: I1213 06:54:36.721903 4971 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/5cccf3e0-5335-44fc-a685-df1f5d46554d-proxy-ca-bundles\") pod \"controller-manager-cd9f4966d-ljtrh\" (UID: \"5cccf3e0-5335-44fc-a685-df1f5d46554d\") " pod="openshift-controller-manager/controller-manager-cd9f4966d-ljtrh" Dec 13 06:54:36 crc kubenswrapper[4971]: I1213 06:54:36.721927 4971 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/ed96d9f8-2ba9-4f6e-b759-4c7c607801c2-client-ca\") pod \"route-controller-manager-5dcf5cc8c9-j7frp\" (UID: \"ed96d9f8-2ba9-4f6e-b759-4c7c607801c2\") " pod="openshift-route-controller-manager/route-controller-manager-5dcf5cc8c9-j7frp" Dec 13 06:54:36 crc kubenswrapper[4971]: I1213 06:54:36.721950 4971 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-44spt\" (UniqueName: \"kubernetes.io/projected/5cccf3e0-5335-44fc-a685-df1f5d46554d-kube-api-access-44spt\") pod \"controller-manager-cd9f4966d-ljtrh\" (UID: \"5cccf3e0-5335-44fc-a685-df1f5d46554d\") " pod="openshift-controller-manager/controller-manager-cd9f4966d-ljtrh" Dec 13 06:54:36 crc kubenswrapper[4971]: I1213 06:54:36.722756 4971 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/5cccf3e0-5335-44fc-a685-df1f5d46554d-client-ca\") pod \"controller-manager-cd9f4966d-ljtrh\" (UID: \"5cccf3e0-5335-44fc-a685-df1f5d46554d\") " pod="openshift-controller-manager/controller-manager-cd9f4966d-ljtrh" Dec 13 06:54:36 crc kubenswrapper[4971]: I1213 06:54:36.723180 4971 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/5cccf3e0-5335-44fc-a685-df1f5d46554d-config\") pod \"controller-manager-cd9f4966d-ljtrh\" (UID: \"5cccf3e0-5335-44fc-a685-df1f5d46554d\") " pod="openshift-controller-manager/controller-manager-cd9f4966d-ljtrh" Dec 13 06:54:36 crc kubenswrapper[4971]: I1213 06:54:36.723348 4971 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/5cccf3e0-5335-44fc-a685-df1f5d46554d-proxy-ca-bundles\") pod \"controller-manager-cd9f4966d-ljtrh\" (UID: \"5cccf3e0-5335-44fc-a685-df1f5d46554d\") " pod="openshift-controller-manager/controller-manager-cd9f4966d-ljtrh" Dec 13 06:54:36 crc kubenswrapper[4971]: I1213 06:54:36.724095 4971 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/ed96d9f8-2ba9-4f6e-b759-4c7c607801c2-config\") pod \"route-controller-manager-5dcf5cc8c9-j7frp\" (UID: \"ed96d9f8-2ba9-4f6e-b759-4c7c607801c2\") " pod="openshift-route-controller-manager/route-controller-manager-5dcf5cc8c9-j7frp" Dec 13 06:54:36 crc kubenswrapper[4971]: I1213 06:54:36.724441 4971 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/ed96d9f8-2ba9-4f6e-b759-4c7c607801c2-client-ca\") pod \"route-controller-manager-5dcf5cc8c9-j7frp\" (UID: \"ed96d9f8-2ba9-4f6e-b759-4c7c607801c2\") " pod="openshift-route-controller-manager/route-controller-manager-5dcf5cc8c9-j7frp" Dec 13 06:54:36 crc kubenswrapper[4971]: I1213 06:54:36.730160 4971 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/ed96d9f8-2ba9-4f6e-b759-4c7c607801c2-serving-cert\") pod \"route-controller-manager-5dcf5cc8c9-j7frp\" (UID: \"ed96d9f8-2ba9-4f6e-b759-4c7c607801c2\") " pod="openshift-route-controller-manager/route-controller-manager-5dcf5cc8c9-j7frp" Dec 13 06:54:36 crc kubenswrapper[4971]: I1213 06:54:36.738155 4971 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/5cccf3e0-5335-44fc-a685-df1f5d46554d-serving-cert\") pod \"controller-manager-cd9f4966d-ljtrh\" (UID: \"5cccf3e0-5335-44fc-a685-df1f5d46554d\") " pod="openshift-controller-manager/controller-manager-cd9f4966d-ljtrh" Dec 13 06:54:36 crc kubenswrapper[4971]: I1213 06:54:36.738368 4971 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-44spt\" (UniqueName: \"kubernetes.io/projected/5cccf3e0-5335-44fc-a685-df1f5d46554d-kube-api-access-44spt\") pod \"controller-manager-cd9f4966d-ljtrh\" (UID: \"5cccf3e0-5335-44fc-a685-df1f5d46554d\") " pod="openshift-controller-manager/controller-manager-cd9f4966d-ljtrh" Dec 13 06:54:36 crc kubenswrapper[4971]: I1213 06:54:36.739245 4971 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-wdpt8\" (UniqueName: \"kubernetes.io/projected/ed96d9f8-2ba9-4f6e-b759-4c7c607801c2-kube-api-access-wdpt8\") pod \"route-controller-manager-5dcf5cc8c9-j7frp\" (UID: \"ed96d9f8-2ba9-4f6e-b759-4c7c607801c2\") " pod="openshift-route-controller-manager/route-controller-manager-5dcf5cc8c9-j7frp" Dec 13 06:54:36 crc kubenswrapper[4971]: I1213 06:54:36.864922 4971 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-route-controller-manager/route-controller-manager-5dcf5cc8c9-j7frp" Dec 13 06:54:36 crc kubenswrapper[4971]: I1213 06:54:36.882624 4971 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-controller-manager/controller-manager-cd9f4966d-ljtrh" Dec 13 06:54:37 crc kubenswrapper[4971]: I1213 06:54:37.081915 4971 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-route-controller-manager/route-controller-manager-5dcf5cc8c9-j7frp"] Dec 13 06:54:37 crc kubenswrapper[4971]: I1213 06:54:37.152733 4971 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-controller-manager/controller-manager-cd9f4966d-ljtrh"] Dec 13 06:54:37 crc kubenswrapper[4971]: I1213 06:54:37.774091 4971 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="2cf6e363-8511-4c79-9181-79dba726e5af" path="/var/lib/kubelet/pods/2cf6e363-8511-4c79-9181-79dba726e5af/volumes" Dec 13 06:54:37 crc kubenswrapper[4971]: I1213 06:54:37.774827 4971 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="7148e573-e2f5-4fc5-b30c-4e777516509e" path="/var/lib/kubelet/pods/7148e573-e2f5-4fc5-b30c-4e777516509e/volumes" Dec 13 06:54:37 crc kubenswrapper[4971]: I1213 06:54:37.973916 4971 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-controller-manager/controller-manager-cd9f4966d-ljtrh" event={"ID":"5cccf3e0-5335-44fc-a685-df1f5d46554d","Type":"ContainerStarted","Data":"f4413363f71e7d620ad5dba57c9f5ec39bfc596edc644fc513bd46337c7da433"} Dec 13 06:54:37 crc kubenswrapper[4971]: I1213 06:54:37.973976 4971 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-controller-manager/controller-manager-cd9f4966d-ljtrh" event={"ID":"5cccf3e0-5335-44fc-a685-df1f5d46554d","Type":"ContainerStarted","Data":"7869998dfde07ea2d637be942b34afe0de15a896b79ed5efe257b059df782183"} Dec 13 06:54:37 crc kubenswrapper[4971]: I1213 06:54:37.974361 4971 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-controller-manager/controller-manager-cd9f4966d-ljtrh" Dec 13 06:54:37 crc kubenswrapper[4971]: I1213 06:54:37.976344 4971 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-route-controller-manager/route-controller-manager-5dcf5cc8c9-j7frp" event={"ID":"ed96d9f8-2ba9-4f6e-b759-4c7c607801c2","Type":"ContainerStarted","Data":"19df763bdfda675e280685d70df318b8d070eab9b7ca12b1a11c8bbca6d5b54a"} Dec 13 06:54:37 crc kubenswrapper[4971]: I1213 06:54:37.976499 4971 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-route-controller-manager/route-controller-manager-5dcf5cc8c9-j7frp" Dec 13 06:54:37 crc kubenswrapper[4971]: I1213 06:54:37.976624 4971 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-route-controller-manager/route-controller-manager-5dcf5cc8c9-j7frp" event={"ID":"ed96d9f8-2ba9-4f6e-b759-4c7c607801c2","Type":"ContainerStarted","Data":"e9c6bd8e5b24b9c507d354d4557dce71baa78535881e6f910893d51794ef06fb"} Dec 13 06:54:37 crc kubenswrapper[4971]: I1213 06:54:37.978502 4971 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-controller-manager/controller-manager-cd9f4966d-ljtrh" Dec 13 06:54:37 crc kubenswrapper[4971]: I1213 06:54:37.981828 4971 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-route-controller-manager/route-controller-manager-5dcf5cc8c9-j7frp" Dec 13 06:54:37 crc kubenswrapper[4971]: I1213 06:54:37.993979 4971 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-controller-manager/controller-manager-cd9f4966d-ljtrh" podStartSLOduration=3.99395901 podStartE2EDuration="3.99395901s" podCreationTimestamp="2025-12-13 06:54:34 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-13 06:54:37.993588851 +0000 UTC m=+334.597998309" watchObservedRunningTime="2025-12-13 06:54:37.99395901 +0000 UTC m=+334.598368468" Dec 13 06:54:38 crc kubenswrapper[4971]: I1213 06:54:38.044281 4971 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-route-controller-manager/route-controller-manager-5dcf5cc8c9-j7frp" podStartSLOduration=4.044263127 podStartE2EDuration="4.044263127s" podCreationTimestamp="2025-12-13 06:54:34 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-13 06:54:38.042275659 +0000 UTC m=+334.646685117" watchObservedRunningTime="2025-12-13 06:54:38.044263127 +0000 UTC m=+334.648672575" Dec 13 06:55:04 crc kubenswrapper[4971]: I1213 06:55:04.345165 4971 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-controller-manager/controller-manager-cd9f4966d-ljtrh"] Dec 13 06:55:04 crc kubenswrapper[4971]: I1213 06:55:04.345852 4971 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-controller-manager/controller-manager-cd9f4966d-ljtrh" podUID="5cccf3e0-5335-44fc-a685-df1f5d46554d" containerName="controller-manager" containerID="cri-o://f4413363f71e7d620ad5dba57c9f5ec39bfc596edc644fc513bd46337c7da433" gracePeriod=30 Dec 13 06:55:04 crc kubenswrapper[4971]: I1213 06:55:04.835118 4971 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-controller-manager/controller-manager-cd9f4966d-ljtrh" Dec 13 06:55:04 crc kubenswrapper[4971]: I1213 06:55:04.990860 4971 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-44spt\" (UniqueName: \"kubernetes.io/projected/5cccf3e0-5335-44fc-a685-df1f5d46554d-kube-api-access-44spt\") pod \"5cccf3e0-5335-44fc-a685-df1f5d46554d\" (UID: \"5cccf3e0-5335-44fc-a685-df1f5d46554d\") " Dec 13 06:55:04 crc kubenswrapper[4971]: I1213 06:55:04.990945 4971 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/5cccf3e0-5335-44fc-a685-df1f5d46554d-proxy-ca-bundles\") pod \"5cccf3e0-5335-44fc-a685-df1f5d46554d\" (UID: \"5cccf3e0-5335-44fc-a685-df1f5d46554d\") " Dec 13 06:55:04 crc kubenswrapper[4971]: I1213 06:55:04.990991 4971 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/5cccf3e0-5335-44fc-a685-df1f5d46554d-serving-cert\") pod \"5cccf3e0-5335-44fc-a685-df1f5d46554d\" (UID: \"5cccf3e0-5335-44fc-a685-df1f5d46554d\") " Dec 13 06:55:04 crc kubenswrapper[4971]: I1213 06:55:04.991059 4971 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/5cccf3e0-5335-44fc-a685-df1f5d46554d-config\") pod \"5cccf3e0-5335-44fc-a685-df1f5d46554d\" (UID: \"5cccf3e0-5335-44fc-a685-df1f5d46554d\") " Dec 13 06:55:04 crc kubenswrapper[4971]: I1213 06:55:04.991088 4971 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/5cccf3e0-5335-44fc-a685-df1f5d46554d-client-ca\") pod \"5cccf3e0-5335-44fc-a685-df1f5d46554d\" (UID: \"5cccf3e0-5335-44fc-a685-df1f5d46554d\") " Dec 13 06:55:04 crc kubenswrapper[4971]: I1213 06:55:04.991610 4971 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/5cccf3e0-5335-44fc-a685-df1f5d46554d-proxy-ca-bundles" (OuterVolumeSpecName: "proxy-ca-bundles") pod "5cccf3e0-5335-44fc-a685-df1f5d46554d" (UID: "5cccf3e0-5335-44fc-a685-df1f5d46554d"). InnerVolumeSpecName "proxy-ca-bundles". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 13 06:55:04 crc kubenswrapper[4971]: I1213 06:55:04.991695 4971 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/5cccf3e0-5335-44fc-a685-df1f5d46554d-config" (OuterVolumeSpecName: "config") pod "5cccf3e0-5335-44fc-a685-df1f5d46554d" (UID: "5cccf3e0-5335-44fc-a685-df1f5d46554d"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 13 06:55:04 crc kubenswrapper[4971]: I1213 06:55:04.991871 4971 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/5cccf3e0-5335-44fc-a685-df1f5d46554d-client-ca" (OuterVolumeSpecName: "client-ca") pod "5cccf3e0-5335-44fc-a685-df1f5d46554d" (UID: "5cccf3e0-5335-44fc-a685-df1f5d46554d"). InnerVolumeSpecName "client-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 13 06:55:04 crc kubenswrapper[4971]: I1213 06:55:04.997065 4971 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/5cccf3e0-5335-44fc-a685-df1f5d46554d-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "5cccf3e0-5335-44fc-a685-df1f5d46554d" (UID: "5cccf3e0-5335-44fc-a685-df1f5d46554d"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 13 06:55:04 crc kubenswrapper[4971]: I1213 06:55:04.997229 4971 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/5cccf3e0-5335-44fc-a685-df1f5d46554d-kube-api-access-44spt" (OuterVolumeSpecName: "kube-api-access-44spt") pod "5cccf3e0-5335-44fc-a685-df1f5d46554d" (UID: "5cccf3e0-5335-44fc-a685-df1f5d46554d"). InnerVolumeSpecName "kube-api-access-44spt". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 13 06:55:05 crc kubenswrapper[4971]: I1213 06:55:05.092915 4971 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/5cccf3e0-5335-44fc-a685-df1f5d46554d-config\") on node \"crc\" DevicePath \"\"" Dec 13 06:55:05 crc kubenswrapper[4971]: I1213 06:55:05.092969 4971 reconciler_common.go:293] "Volume detached for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/5cccf3e0-5335-44fc-a685-df1f5d46554d-client-ca\") on node \"crc\" DevicePath \"\"" Dec 13 06:55:05 crc kubenswrapper[4971]: I1213 06:55:05.092996 4971 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-44spt\" (UniqueName: \"kubernetes.io/projected/5cccf3e0-5335-44fc-a685-df1f5d46554d-kube-api-access-44spt\") on node \"crc\" DevicePath \"\"" Dec 13 06:55:05 crc kubenswrapper[4971]: I1213 06:55:05.093025 4971 reconciler_common.go:293] "Volume detached for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/5cccf3e0-5335-44fc-a685-df1f5d46554d-proxy-ca-bundles\") on node \"crc\" DevicePath \"\"" Dec 13 06:55:05 crc kubenswrapper[4971]: I1213 06:55:05.093053 4971 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/5cccf3e0-5335-44fc-a685-df1f5d46554d-serving-cert\") on node \"crc\" DevicePath \"\"" Dec 13 06:55:05 crc kubenswrapper[4971]: I1213 06:55:05.107217 4971 generic.go:334] "Generic (PLEG): container finished" podID="5cccf3e0-5335-44fc-a685-df1f5d46554d" containerID="f4413363f71e7d620ad5dba57c9f5ec39bfc596edc644fc513bd46337c7da433" exitCode=0 Dec 13 06:55:05 crc kubenswrapper[4971]: I1213 06:55:05.107295 4971 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-controller-manager/controller-manager-cd9f4966d-ljtrh" event={"ID":"5cccf3e0-5335-44fc-a685-df1f5d46554d","Type":"ContainerDied","Data":"f4413363f71e7d620ad5dba57c9f5ec39bfc596edc644fc513bd46337c7da433"} Dec 13 06:55:05 crc kubenswrapper[4971]: I1213 06:55:05.107346 4971 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-controller-manager/controller-manager-cd9f4966d-ljtrh" event={"ID":"5cccf3e0-5335-44fc-a685-df1f5d46554d","Type":"ContainerDied","Data":"7869998dfde07ea2d637be942b34afe0de15a896b79ed5efe257b059df782183"} Dec 13 06:55:05 crc kubenswrapper[4971]: I1213 06:55:05.107384 4971 scope.go:117] "RemoveContainer" containerID="f4413363f71e7d620ad5dba57c9f5ec39bfc596edc644fc513bd46337c7da433" Dec 13 06:55:05 crc kubenswrapper[4971]: I1213 06:55:05.107616 4971 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-controller-manager/controller-manager-cd9f4966d-ljtrh" Dec 13 06:55:05 crc kubenswrapper[4971]: I1213 06:55:05.127558 4971 scope.go:117] "RemoveContainer" containerID="f4413363f71e7d620ad5dba57c9f5ec39bfc596edc644fc513bd46337c7da433" Dec 13 06:55:05 crc kubenswrapper[4971]: E1213 06:55:05.128310 4971 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"f4413363f71e7d620ad5dba57c9f5ec39bfc596edc644fc513bd46337c7da433\": container with ID starting with f4413363f71e7d620ad5dba57c9f5ec39bfc596edc644fc513bd46337c7da433 not found: ID does not exist" containerID="f4413363f71e7d620ad5dba57c9f5ec39bfc596edc644fc513bd46337c7da433" Dec 13 06:55:05 crc kubenswrapper[4971]: I1213 06:55:05.128419 4971 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"f4413363f71e7d620ad5dba57c9f5ec39bfc596edc644fc513bd46337c7da433"} err="failed to get container status \"f4413363f71e7d620ad5dba57c9f5ec39bfc596edc644fc513bd46337c7da433\": rpc error: code = NotFound desc = could not find container \"f4413363f71e7d620ad5dba57c9f5ec39bfc596edc644fc513bd46337c7da433\": container with ID starting with f4413363f71e7d620ad5dba57c9f5ec39bfc596edc644fc513bd46337c7da433 not found: ID does not exist" Dec 13 06:55:05 crc kubenswrapper[4971]: I1213 06:55:05.143710 4971 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-controller-manager/controller-manager-cd9f4966d-ljtrh"] Dec 13 06:55:05 crc kubenswrapper[4971]: I1213 06:55:05.146905 4971 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-controller-manager/controller-manager-cd9f4966d-ljtrh"] Dec 13 06:55:05 crc kubenswrapper[4971]: I1213 06:55:05.575725 4971 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-controller-manager/controller-manager-5597b99f57-jpz8s"] Dec 13 06:55:05 crc kubenswrapper[4971]: E1213 06:55:05.576109 4971 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="5cccf3e0-5335-44fc-a685-df1f5d46554d" containerName="controller-manager" Dec 13 06:55:05 crc kubenswrapper[4971]: I1213 06:55:05.576126 4971 state_mem.go:107] "Deleted CPUSet assignment" podUID="5cccf3e0-5335-44fc-a685-df1f5d46554d" containerName="controller-manager" Dec 13 06:55:05 crc kubenswrapper[4971]: I1213 06:55:05.576439 4971 memory_manager.go:354] "RemoveStaleState removing state" podUID="5cccf3e0-5335-44fc-a685-df1f5d46554d" containerName="controller-manager" Dec 13 06:55:05 crc kubenswrapper[4971]: I1213 06:55:05.578341 4971 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-controller-manager/controller-manager-5597b99f57-jpz8s" Dec 13 06:55:05 crc kubenswrapper[4971]: I1213 06:55:05.581573 4971 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"config" Dec 13 06:55:05 crc kubenswrapper[4971]: I1213 06:55:05.581588 4971 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-controller-manager"/"openshift-controller-manager-sa-dockercfg-msq4c" Dec 13 06:55:05 crc kubenswrapper[4971]: I1213 06:55:05.581606 4971 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"openshift-service-ca.crt" Dec 13 06:55:05 crc kubenswrapper[4971]: I1213 06:55:05.581795 4971 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-controller-manager"/"serving-cert" Dec 13 06:55:05 crc kubenswrapper[4971]: I1213 06:55:05.582234 4971 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"kube-root-ca.crt" Dec 13 06:55:05 crc kubenswrapper[4971]: I1213 06:55:05.583250 4971 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"client-ca" Dec 13 06:55:05 crc kubenswrapper[4971]: I1213 06:55:05.585597 4971 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-controller-manager/controller-manager-5597b99f57-jpz8s"] Dec 13 06:55:05 crc kubenswrapper[4971]: I1213 06:55:05.589756 4971 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"openshift-global-ca" Dec 13 06:55:05 crc kubenswrapper[4971]: I1213 06:55:05.699572 4971 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/8fd596c1-3076-44e4-affa-7e323492ad88-config\") pod \"controller-manager-5597b99f57-jpz8s\" (UID: \"8fd596c1-3076-44e4-affa-7e323492ad88\") " pod="openshift-controller-manager/controller-manager-5597b99f57-jpz8s" Dec 13 06:55:05 crc kubenswrapper[4971]: I1213 06:55:05.699658 4971 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/8fd596c1-3076-44e4-affa-7e323492ad88-proxy-ca-bundles\") pod \"controller-manager-5597b99f57-jpz8s\" (UID: \"8fd596c1-3076-44e4-affa-7e323492ad88\") " pod="openshift-controller-manager/controller-manager-5597b99f57-jpz8s" Dec 13 06:55:05 crc kubenswrapper[4971]: I1213 06:55:05.699690 4971 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-t4cs2\" (UniqueName: \"kubernetes.io/projected/8fd596c1-3076-44e4-affa-7e323492ad88-kube-api-access-t4cs2\") pod \"controller-manager-5597b99f57-jpz8s\" (UID: \"8fd596c1-3076-44e4-affa-7e323492ad88\") " pod="openshift-controller-manager/controller-manager-5597b99f57-jpz8s" Dec 13 06:55:05 crc kubenswrapper[4971]: I1213 06:55:05.699731 4971 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/8fd596c1-3076-44e4-affa-7e323492ad88-client-ca\") pod \"controller-manager-5597b99f57-jpz8s\" (UID: \"8fd596c1-3076-44e4-affa-7e323492ad88\") " pod="openshift-controller-manager/controller-manager-5597b99f57-jpz8s" Dec 13 06:55:05 crc kubenswrapper[4971]: I1213 06:55:05.699762 4971 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/8fd596c1-3076-44e4-affa-7e323492ad88-serving-cert\") pod \"controller-manager-5597b99f57-jpz8s\" (UID: \"8fd596c1-3076-44e4-affa-7e323492ad88\") " pod="openshift-controller-manager/controller-manager-5597b99f57-jpz8s" Dec 13 06:55:05 crc kubenswrapper[4971]: I1213 06:55:05.775352 4971 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="5cccf3e0-5335-44fc-a685-df1f5d46554d" path="/var/lib/kubelet/pods/5cccf3e0-5335-44fc-a685-df1f5d46554d/volumes" Dec 13 06:55:05 crc kubenswrapper[4971]: I1213 06:55:05.801431 4971 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/8fd596c1-3076-44e4-affa-7e323492ad88-config\") pod \"controller-manager-5597b99f57-jpz8s\" (UID: \"8fd596c1-3076-44e4-affa-7e323492ad88\") " pod="openshift-controller-manager/controller-manager-5597b99f57-jpz8s" Dec 13 06:55:05 crc kubenswrapper[4971]: I1213 06:55:05.801534 4971 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/8fd596c1-3076-44e4-affa-7e323492ad88-proxy-ca-bundles\") pod \"controller-manager-5597b99f57-jpz8s\" (UID: \"8fd596c1-3076-44e4-affa-7e323492ad88\") " pod="openshift-controller-manager/controller-manager-5597b99f57-jpz8s" Dec 13 06:55:05 crc kubenswrapper[4971]: I1213 06:55:05.801567 4971 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-t4cs2\" (UniqueName: \"kubernetes.io/projected/8fd596c1-3076-44e4-affa-7e323492ad88-kube-api-access-t4cs2\") pod \"controller-manager-5597b99f57-jpz8s\" (UID: \"8fd596c1-3076-44e4-affa-7e323492ad88\") " pod="openshift-controller-manager/controller-manager-5597b99f57-jpz8s" Dec 13 06:55:05 crc kubenswrapper[4971]: I1213 06:55:05.801613 4971 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/8fd596c1-3076-44e4-affa-7e323492ad88-client-ca\") pod \"controller-manager-5597b99f57-jpz8s\" (UID: \"8fd596c1-3076-44e4-affa-7e323492ad88\") " pod="openshift-controller-manager/controller-manager-5597b99f57-jpz8s" Dec 13 06:55:05 crc kubenswrapper[4971]: I1213 06:55:05.801643 4971 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/8fd596c1-3076-44e4-affa-7e323492ad88-serving-cert\") pod \"controller-manager-5597b99f57-jpz8s\" (UID: \"8fd596c1-3076-44e4-affa-7e323492ad88\") " pod="openshift-controller-manager/controller-manager-5597b99f57-jpz8s" Dec 13 06:55:05 crc kubenswrapper[4971]: I1213 06:55:05.802799 4971 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/8fd596c1-3076-44e4-affa-7e323492ad88-client-ca\") pod \"controller-manager-5597b99f57-jpz8s\" (UID: \"8fd596c1-3076-44e4-affa-7e323492ad88\") " pod="openshift-controller-manager/controller-manager-5597b99f57-jpz8s" Dec 13 06:55:05 crc kubenswrapper[4971]: I1213 06:55:05.803110 4971 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/8fd596c1-3076-44e4-affa-7e323492ad88-proxy-ca-bundles\") pod \"controller-manager-5597b99f57-jpz8s\" (UID: \"8fd596c1-3076-44e4-affa-7e323492ad88\") " pod="openshift-controller-manager/controller-manager-5597b99f57-jpz8s" Dec 13 06:55:05 crc kubenswrapper[4971]: I1213 06:55:05.803412 4971 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/8fd596c1-3076-44e4-affa-7e323492ad88-config\") pod \"controller-manager-5597b99f57-jpz8s\" (UID: \"8fd596c1-3076-44e4-affa-7e323492ad88\") " pod="openshift-controller-manager/controller-manager-5597b99f57-jpz8s" Dec 13 06:55:05 crc kubenswrapper[4971]: I1213 06:55:05.811168 4971 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/8fd596c1-3076-44e4-affa-7e323492ad88-serving-cert\") pod \"controller-manager-5597b99f57-jpz8s\" (UID: \"8fd596c1-3076-44e4-affa-7e323492ad88\") " pod="openshift-controller-manager/controller-manager-5597b99f57-jpz8s" Dec 13 06:55:05 crc kubenswrapper[4971]: I1213 06:55:05.818061 4971 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-t4cs2\" (UniqueName: \"kubernetes.io/projected/8fd596c1-3076-44e4-affa-7e323492ad88-kube-api-access-t4cs2\") pod \"controller-manager-5597b99f57-jpz8s\" (UID: \"8fd596c1-3076-44e4-affa-7e323492ad88\") " pod="openshift-controller-manager/controller-manager-5597b99f57-jpz8s" Dec 13 06:55:05 crc kubenswrapper[4971]: I1213 06:55:05.903903 4971 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-controller-manager/controller-manager-5597b99f57-jpz8s" Dec 13 06:55:06 crc kubenswrapper[4971]: I1213 06:55:06.301428 4971 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-controller-manager/controller-manager-5597b99f57-jpz8s"] Dec 13 06:55:07 crc kubenswrapper[4971]: I1213 06:55:07.126914 4971 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-controller-manager/controller-manager-5597b99f57-jpz8s" event={"ID":"8fd596c1-3076-44e4-affa-7e323492ad88","Type":"ContainerStarted","Data":"407fcfcff6f16b76345b0bfadc6b78c8040a1ac6773444714191939cf386206b"} Dec 13 06:55:07 crc kubenswrapper[4971]: I1213 06:55:07.126996 4971 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-controller-manager/controller-manager-5597b99f57-jpz8s" event={"ID":"8fd596c1-3076-44e4-affa-7e323492ad88","Type":"ContainerStarted","Data":"cc3149f165eda47ed477526912e3009e9e3fef34d88eb3e0078a50975f0b9662"} Dec 13 06:55:07 crc kubenswrapper[4971]: I1213 06:55:07.127314 4971 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-controller-manager/controller-manager-5597b99f57-jpz8s" Dec 13 06:55:07 crc kubenswrapper[4971]: I1213 06:55:07.133874 4971 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-controller-manager/controller-manager-5597b99f57-jpz8s" Dec 13 06:55:07 crc kubenswrapper[4971]: I1213 06:55:07.145934 4971 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-controller-manager/controller-manager-5597b99f57-jpz8s" podStartSLOduration=3.14588405 podStartE2EDuration="3.14588405s" podCreationTimestamp="2025-12-13 06:55:04 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-13 06:55:07.142693912 +0000 UTC m=+363.747103380" watchObservedRunningTime="2025-12-13 06:55:07.14588405 +0000 UTC m=+363.750293508" Dec 13 06:55:13 crc kubenswrapper[4971]: I1213 06:55:13.931152 4971 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-image-registry/image-registry-66df7c8f76-xjwwz"] Dec 13 06:55:13 crc kubenswrapper[4971]: I1213 06:55:13.932497 4971 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-image-registry/image-registry-66df7c8f76-xjwwz" Dec 13 06:55:13 crc kubenswrapper[4971]: I1213 06:55:13.944916 4971 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-image-registry/image-registry-66df7c8f76-xjwwz"] Dec 13 06:55:14 crc kubenswrapper[4971]: I1213 06:55:14.110091 4971 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/c7796491-08b9-4ba0-ba02-229f6362b638-bound-sa-token\") pod \"image-registry-66df7c8f76-xjwwz\" (UID: \"c7796491-08b9-4ba0-ba02-229f6362b638\") " pod="openshift-image-registry/image-registry-66df7c8f76-xjwwz" Dec 13 06:55:14 crc kubenswrapper[4971]: I1213 06:55:14.110143 4971 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/c7796491-08b9-4ba0-ba02-229f6362b638-trusted-ca\") pod \"image-registry-66df7c8f76-xjwwz\" (UID: \"c7796491-08b9-4ba0-ba02-229f6362b638\") " pod="openshift-image-registry/image-registry-66df7c8f76-xjwwz" Dec 13 06:55:14 crc kubenswrapper[4971]: I1213 06:55:14.110201 4971 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-qgh24\" (UniqueName: \"kubernetes.io/projected/c7796491-08b9-4ba0-ba02-229f6362b638-kube-api-access-qgh24\") pod \"image-registry-66df7c8f76-xjwwz\" (UID: \"c7796491-08b9-4ba0-ba02-229f6362b638\") " pod="openshift-image-registry/image-registry-66df7c8f76-xjwwz" Dec 13 06:55:14 crc kubenswrapper[4971]: I1213 06:55:14.110237 4971 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"installation-pull-secrets\" (UniqueName: \"kubernetes.io/secret/c7796491-08b9-4ba0-ba02-229f6362b638-installation-pull-secrets\") pod \"image-registry-66df7c8f76-xjwwz\" (UID: \"c7796491-08b9-4ba0-ba02-229f6362b638\") " pod="openshift-image-registry/image-registry-66df7c8f76-xjwwz" Dec 13 06:55:14 crc kubenswrapper[4971]: I1213 06:55:14.110325 4971 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"registry-certificates\" (UniqueName: \"kubernetes.io/configmap/c7796491-08b9-4ba0-ba02-229f6362b638-registry-certificates\") pod \"image-registry-66df7c8f76-xjwwz\" (UID: \"c7796491-08b9-4ba0-ba02-229f6362b638\") " pod="openshift-image-registry/image-registry-66df7c8f76-xjwwz" Dec 13 06:55:14 crc kubenswrapper[4971]: I1213 06:55:14.110373 4971 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ca-trust-extracted\" (UniqueName: \"kubernetes.io/empty-dir/c7796491-08b9-4ba0-ba02-229f6362b638-ca-trust-extracted\") pod \"image-registry-66df7c8f76-xjwwz\" (UID: \"c7796491-08b9-4ba0-ba02-229f6362b638\") " pod="openshift-image-registry/image-registry-66df7c8f76-xjwwz" Dec 13 06:55:14 crc kubenswrapper[4971]: I1213 06:55:14.110398 4971 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"registry-tls\" (UniqueName: \"kubernetes.io/projected/c7796491-08b9-4ba0-ba02-229f6362b638-registry-tls\") pod \"image-registry-66df7c8f76-xjwwz\" (UID: \"c7796491-08b9-4ba0-ba02-229f6362b638\") " pod="openshift-image-registry/image-registry-66df7c8f76-xjwwz" Dec 13 06:55:14 crc kubenswrapper[4971]: I1213 06:55:14.110496 4971 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-66df7c8f76-xjwwz\" (UID: \"c7796491-08b9-4ba0-ba02-229f6362b638\") " pod="openshift-image-registry/image-registry-66df7c8f76-xjwwz" Dec 13 06:55:14 crc kubenswrapper[4971]: I1213 06:55:14.138304 4971 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-66df7c8f76-xjwwz\" (UID: \"c7796491-08b9-4ba0-ba02-229f6362b638\") " pod="openshift-image-registry/image-registry-66df7c8f76-xjwwz" Dec 13 06:55:14 crc kubenswrapper[4971]: I1213 06:55:14.214203 4971 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"registry-tls\" (UniqueName: \"kubernetes.io/projected/c7796491-08b9-4ba0-ba02-229f6362b638-registry-tls\") pod \"image-registry-66df7c8f76-xjwwz\" (UID: \"c7796491-08b9-4ba0-ba02-229f6362b638\") " pod="openshift-image-registry/image-registry-66df7c8f76-xjwwz" Dec 13 06:55:14 crc kubenswrapper[4971]: I1213 06:55:14.214299 4971 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/c7796491-08b9-4ba0-ba02-229f6362b638-bound-sa-token\") pod \"image-registry-66df7c8f76-xjwwz\" (UID: \"c7796491-08b9-4ba0-ba02-229f6362b638\") " pod="openshift-image-registry/image-registry-66df7c8f76-xjwwz" Dec 13 06:55:14 crc kubenswrapper[4971]: I1213 06:55:14.214321 4971 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/c7796491-08b9-4ba0-ba02-229f6362b638-trusted-ca\") pod \"image-registry-66df7c8f76-xjwwz\" (UID: \"c7796491-08b9-4ba0-ba02-229f6362b638\") " pod="openshift-image-registry/image-registry-66df7c8f76-xjwwz" Dec 13 06:55:14 crc kubenswrapper[4971]: I1213 06:55:14.214383 4971 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-qgh24\" (UniqueName: \"kubernetes.io/projected/c7796491-08b9-4ba0-ba02-229f6362b638-kube-api-access-qgh24\") pod \"image-registry-66df7c8f76-xjwwz\" (UID: \"c7796491-08b9-4ba0-ba02-229f6362b638\") " pod="openshift-image-registry/image-registry-66df7c8f76-xjwwz" Dec 13 06:55:14 crc kubenswrapper[4971]: I1213 06:55:14.214427 4971 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"installation-pull-secrets\" (UniqueName: \"kubernetes.io/secret/c7796491-08b9-4ba0-ba02-229f6362b638-installation-pull-secrets\") pod \"image-registry-66df7c8f76-xjwwz\" (UID: \"c7796491-08b9-4ba0-ba02-229f6362b638\") " pod="openshift-image-registry/image-registry-66df7c8f76-xjwwz" Dec 13 06:55:14 crc kubenswrapper[4971]: I1213 06:55:14.214462 4971 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"registry-certificates\" (UniqueName: \"kubernetes.io/configmap/c7796491-08b9-4ba0-ba02-229f6362b638-registry-certificates\") pod \"image-registry-66df7c8f76-xjwwz\" (UID: \"c7796491-08b9-4ba0-ba02-229f6362b638\") " pod="openshift-image-registry/image-registry-66df7c8f76-xjwwz" Dec 13 06:55:14 crc kubenswrapper[4971]: I1213 06:55:14.214567 4971 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ca-trust-extracted\" (UniqueName: \"kubernetes.io/empty-dir/c7796491-08b9-4ba0-ba02-229f6362b638-ca-trust-extracted\") pod \"image-registry-66df7c8f76-xjwwz\" (UID: \"c7796491-08b9-4ba0-ba02-229f6362b638\") " pod="openshift-image-registry/image-registry-66df7c8f76-xjwwz" Dec 13 06:55:14 crc kubenswrapper[4971]: I1213 06:55:14.215986 4971 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ca-trust-extracted\" (UniqueName: \"kubernetes.io/empty-dir/c7796491-08b9-4ba0-ba02-229f6362b638-ca-trust-extracted\") pod \"image-registry-66df7c8f76-xjwwz\" (UID: \"c7796491-08b9-4ba0-ba02-229f6362b638\") " pod="openshift-image-registry/image-registry-66df7c8f76-xjwwz" Dec 13 06:55:14 crc kubenswrapper[4971]: I1213 06:55:14.217282 4971 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"registry-certificates\" (UniqueName: \"kubernetes.io/configmap/c7796491-08b9-4ba0-ba02-229f6362b638-registry-certificates\") pod \"image-registry-66df7c8f76-xjwwz\" (UID: \"c7796491-08b9-4ba0-ba02-229f6362b638\") " pod="openshift-image-registry/image-registry-66df7c8f76-xjwwz" Dec 13 06:55:14 crc kubenswrapper[4971]: I1213 06:55:14.221473 4971 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/c7796491-08b9-4ba0-ba02-229f6362b638-trusted-ca\") pod \"image-registry-66df7c8f76-xjwwz\" (UID: \"c7796491-08b9-4ba0-ba02-229f6362b638\") " pod="openshift-image-registry/image-registry-66df7c8f76-xjwwz" Dec 13 06:55:14 crc kubenswrapper[4971]: I1213 06:55:14.222441 4971 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"registry-tls\" (UniqueName: \"kubernetes.io/projected/c7796491-08b9-4ba0-ba02-229f6362b638-registry-tls\") pod \"image-registry-66df7c8f76-xjwwz\" (UID: \"c7796491-08b9-4ba0-ba02-229f6362b638\") " pod="openshift-image-registry/image-registry-66df7c8f76-xjwwz" Dec 13 06:55:14 crc kubenswrapper[4971]: I1213 06:55:14.223016 4971 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"installation-pull-secrets\" (UniqueName: \"kubernetes.io/secret/c7796491-08b9-4ba0-ba02-229f6362b638-installation-pull-secrets\") pod \"image-registry-66df7c8f76-xjwwz\" (UID: \"c7796491-08b9-4ba0-ba02-229f6362b638\") " pod="openshift-image-registry/image-registry-66df7c8f76-xjwwz" Dec 13 06:55:14 crc kubenswrapper[4971]: I1213 06:55:14.230741 4971 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/c7796491-08b9-4ba0-ba02-229f6362b638-bound-sa-token\") pod \"image-registry-66df7c8f76-xjwwz\" (UID: \"c7796491-08b9-4ba0-ba02-229f6362b638\") " pod="openshift-image-registry/image-registry-66df7c8f76-xjwwz" Dec 13 06:55:14 crc kubenswrapper[4971]: I1213 06:55:14.232216 4971 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-qgh24\" (UniqueName: \"kubernetes.io/projected/c7796491-08b9-4ba0-ba02-229f6362b638-kube-api-access-qgh24\") pod \"image-registry-66df7c8f76-xjwwz\" (UID: \"c7796491-08b9-4ba0-ba02-229f6362b638\") " pod="openshift-image-registry/image-registry-66df7c8f76-xjwwz" Dec 13 06:55:14 crc kubenswrapper[4971]: I1213 06:55:14.249315 4971 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-image-registry/image-registry-66df7c8f76-xjwwz" Dec 13 06:55:14 crc kubenswrapper[4971]: I1213 06:55:14.373349 4971 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-operators-5qtd2"] Dec 13 06:55:14 crc kubenswrapper[4971]: I1213 06:55:14.375055 4971 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-5qtd2" Dec 13 06:55:14 crc kubenswrapper[4971]: I1213 06:55:14.381135 4971 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"redhat-operators-dockercfg-ct8rh" Dec 13 06:55:14 crc kubenswrapper[4971]: I1213 06:55:14.385420 4971 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-5qtd2"] Dec 13 06:55:14 crc kubenswrapper[4971]: I1213 06:55:14.517855 4971 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-zsstt\" (UniqueName: \"kubernetes.io/projected/9fd587ab-4839-4479-80f8-f97d78d833b9-kube-api-access-zsstt\") pod \"redhat-operators-5qtd2\" (UID: \"9fd587ab-4839-4479-80f8-f97d78d833b9\") " pod="openshift-marketplace/redhat-operators-5qtd2" Dec 13 06:55:14 crc kubenswrapper[4971]: I1213 06:55:14.517917 4971 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/9fd587ab-4839-4479-80f8-f97d78d833b9-utilities\") pod \"redhat-operators-5qtd2\" (UID: \"9fd587ab-4839-4479-80f8-f97d78d833b9\") " pod="openshift-marketplace/redhat-operators-5qtd2" Dec 13 06:55:14 crc kubenswrapper[4971]: I1213 06:55:14.518051 4971 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/9fd587ab-4839-4479-80f8-f97d78d833b9-catalog-content\") pod \"redhat-operators-5qtd2\" (UID: \"9fd587ab-4839-4479-80f8-f97d78d833b9\") " pod="openshift-marketplace/redhat-operators-5qtd2" Dec 13 06:55:14 crc kubenswrapper[4971]: I1213 06:55:14.619144 4971 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/9fd587ab-4839-4479-80f8-f97d78d833b9-catalog-content\") pod \"redhat-operators-5qtd2\" (UID: \"9fd587ab-4839-4479-80f8-f97d78d833b9\") " pod="openshift-marketplace/redhat-operators-5qtd2" Dec 13 06:55:14 crc kubenswrapper[4971]: I1213 06:55:14.619215 4971 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-zsstt\" (UniqueName: \"kubernetes.io/projected/9fd587ab-4839-4479-80f8-f97d78d833b9-kube-api-access-zsstt\") pod \"redhat-operators-5qtd2\" (UID: \"9fd587ab-4839-4479-80f8-f97d78d833b9\") " pod="openshift-marketplace/redhat-operators-5qtd2" Dec 13 06:55:14 crc kubenswrapper[4971]: I1213 06:55:14.619254 4971 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/9fd587ab-4839-4479-80f8-f97d78d833b9-utilities\") pod \"redhat-operators-5qtd2\" (UID: \"9fd587ab-4839-4479-80f8-f97d78d833b9\") " pod="openshift-marketplace/redhat-operators-5qtd2" Dec 13 06:55:14 crc kubenswrapper[4971]: I1213 06:55:14.620005 4971 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/9fd587ab-4839-4479-80f8-f97d78d833b9-utilities\") pod \"redhat-operators-5qtd2\" (UID: \"9fd587ab-4839-4479-80f8-f97d78d833b9\") " pod="openshift-marketplace/redhat-operators-5qtd2" Dec 13 06:55:14 crc kubenswrapper[4971]: I1213 06:55:14.620112 4971 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/9fd587ab-4839-4479-80f8-f97d78d833b9-catalog-content\") pod \"redhat-operators-5qtd2\" (UID: \"9fd587ab-4839-4479-80f8-f97d78d833b9\") " pod="openshift-marketplace/redhat-operators-5qtd2" Dec 13 06:55:14 crc kubenswrapper[4971]: I1213 06:55:14.643303 4971 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-zsstt\" (UniqueName: \"kubernetes.io/projected/9fd587ab-4839-4479-80f8-f97d78d833b9-kube-api-access-zsstt\") pod \"redhat-operators-5qtd2\" (UID: \"9fd587ab-4839-4479-80f8-f97d78d833b9\") " pod="openshift-marketplace/redhat-operators-5qtd2" Dec 13 06:55:14 crc kubenswrapper[4971]: I1213 06:55:14.662016 4971 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-image-registry/image-registry-66df7c8f76-xjwwz"] Dec 13 06:55:14 crc kubenswrapper[4971]: I1213 06:55:14.697467 4971 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-5qtd2" Dec 13 06:55:15 crc kubenswrapper[4971]: I1213 06:55:15.128611 4971 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-5qtd2"] Dec 13 06:55:15 crc kubenswrapper[4971]: W1213 06:55:15.134022 4971 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod9fd587ab_4839_4479_80f8_f97d78d833b9.slice/crio-36a034a89d0f03625ad83f07e5763c9c99810c4d95c944e9d25c4733e3d8d87e WatchSource:0}: Error finding container 36a034a89d0f03625ad83f07e5763c9c99810c4d95c944e9d25c4733e3d8d87e: Status 404 returned error can't find the container with id 36a034a89d0f03625ad83f07e5763c9c99810c4d95c944e9d25c4733e3d8d87e Dec 13 06:55:15 crc kubenswrapper[4971]: I1213 06:55:15.169460 4971 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-5qtd2" event={"ID":"9fd587ab-4839-4479-80f8-f97d78d833b9","Type":"ContainerStarted","Data":"36a034a89d0f03625ad83f07e5763c9c99810c4d95c944e9d25c4733e3d8d87e"} Dec 13 06:55:15 crc kubenswrapper[4971]: I1213 06:55:15.170996 4971 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-image-registry/image-registry-66df7c8f76-xjwwz" event={"ID":"c7796491-08b9-4ba0-ba02-229f6362b638","Type":"ContainerStarted","Data":"2811cdbc1b10c63138a878fb95cafbcdfe21ccf69bff0dd2345be3862a729d84"} Dec 13 06:55:15 crc kubenswrapper[4971]: I1213 06:55:15.171028 4971 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-image-registry/image-registry-66df7c8f76-xjwwz" event={"ID":"c7796491-08b9-4ba0-ba02-229f6362b638","Type":"ContainerStarted","Data":"5223d890ef730381f32e07b6b0bf599c1b5251932235ff8dbb19e5b99d58c740"} Dec 13 06:55:15 crc kubenswrapper[4971]: I1213 06:55:15.171196 4971 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-image-registry/image-registry-66df7c8f76-xjwwz" Dec 13 06:55:15 crc kubenswrapper[4971]: I1213 06:55:15.193254 4971 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-image-registry/image-registry-66df7c8f76-xjwwz" podStartSLOduration=2.193236424 podStartE2EDuration="2.193236424s" podCreationTimestamp="2025-12-13 06:55:13 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-13 06:55:15.191980484 +0000 UTC m=+371.796389932" watchObservedRunningTime="2025-12-13 06:55:15.193236424 +0000 UTC m=+371.797645872" Dec 13 06:55:15 crc kubenswrapper[4971]: I1213 06:55:15.362810 4971 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-marketplace-68kxv"] Dec 13 06:55:15 crc kubenswrapper[4971]: I1213 06:55:15.364039 4971 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-68kxv" Dec 13 06:55:15 crc kubenswrapper[4971]: I1213 06:55:15.368316 4971 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"redhat-marketplace-dockercfg-x2ctb" Dec 13 06:55:15 crc kubenswrapper[4971]: I1213 06:55:15.373051 4971 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-68kxv"] Dec 13 06:55:15 crc kubenswrapper[4971]: I1213 06:55:15.530590 4971 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-frhc2\" (UniqueName: \"kubernetes.io/projected/5eea3e58-f9fb-4529-b645-0f5193d5a8a9-kube-api-access-frhc2\") pod \"redhat-marketplace-68kxv\" (UID: \"5eea3e58-f9fb-4529-b645-0f5193d5a8a9\") " pod="openshift-marketplace/redhat-marketplace-68kxv" Dec 13 06:55:15 crc kubenswrapper[4971]: I1213 06:55:15.531064 4971 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/5eea3e58-f9fb-4529-b645-0f5193d5a8a9-catalog-content\") pod \"redhat-marketplace-68kxv\" (UID: \"5eea3e58-f9fb-4529-b645-0f5193d5a8a9\") " pod="openshift-marketplace/redhat-marketplace-68kxv" Dec 13 06:55:15 crc kubenswrapper[4971]: I1213 06:55:15.531108 4971 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/5eea3e58-f9fb-4529-b645-0f5193d5a8a9-utilities\") pod \"redhat-marketplace-68kxv\" (UID: \"5eea3e58-f9fb-4529-b645-0f5193d5a8a9\") " pod="openshift-marketplace/redhat-marketplace-68kxv" Dec 13 06:55:15 crc kubenswrapper[4971]: I1213 06:55:15.632341 4971 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/5eea3e58-f9fb-4529-b645-0f5193d5a8a9-catalog-content\") pod \"redhat-marketplace-68kxv\" (UID: \"5eea3e58-f9fb-4529-b645-0f5193d5a8a9\") " pod="openshift-marketplace/redhat-marketplace-68kxv" Dec 13 06:55:15 crc kubenswrapper[4971]: I1213 06:55:15.632406 4971 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/5eea3e58-f9fb-4529-b645-0f5193d5a8a9-utilities\") pod \"redhat-marketplace-68kxv\" (UID: \"5eea3e58-f9fb-4529-b645-0f5193d5a8a9\") " pod="openshift-marketplace/redhat-marketplace-68kxv" Dec 13 06:55:15 crc kubenswrapper[4971]: I1213 06:55:15.632534 4971 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-frhc2\" (UniqueName: \"kubernetes.io/projected/5eea3e58-f9fb-4529-b645-0f5193d5a8a9-kube-api-access-frhc2\") pod \"redhat-marketplace-68kxv\" (UID: \"5eea3e58-f9fb-4529-b645-0f5193d5a8a9\") " pod="openshift-marketplace/redhat-marketplace-68kxv" Dec 13 06:55:15 crc kubenswrapper[4971]: I1213 06:55:15.632887 4971 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/5eea3e58-f9fb-4529-b645-0f5193d5a8a9-catalog-content\") pod \"redhat-marketplace-68kxv\" (UID: \"5eea3e58-f9fb-4529-b645-0f5193d5a8a9\") " pod="openshift-marketplace/redhat-marketplace-68kxv" Dec 13 06:55:15 crc kubenswrapper[4971]: I1213 06:55:15.632951 4971 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/5eea3e58-f9fb-4529-b645-0f5193d5a8a9-utilities\") pod \"redhat-marketplace-68kxv\" (UID: \"5eea3e58-f9fb-4529-b645-0f5193d5a8a9\") " pod="openshift-marketplace/redhat-marketplace-68kxv" Dec 13 06:55:15 crc kubenswrapper[4971]: I1213 06:55:15.658890 4971 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-frhc2\" (UniqueName: \"kubernetes.io/projected/5eea3e58-f9fb-4529-b645-0f5193d5a8a9-kube-api-access-frhc2\") pod \"redhat-marketplace-68kxv\" (UID: \"5eea3e58-f9fb-4529-b645-0f5193d5a8a9\") " pod="openshift-marketplace/redhat-marketplace-68kxv" Dec 13 06:55:15 crc kubenswrapper[4971]: I1213 06:55:15.677957 4971 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-68kxv" Dec 13 06:55:16 crc kubenswrapper[4971]: I1213 06:55:16.067086 4971 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-68kxv"] Dec 13 06:55:16 crc kubenswrapper[4971]: W1213 06:55:16.077862 4971 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod5eea3e58_f9fb_4529_b645_0f5193d5a8a9.slice/crio-b37e9a06348fd99e96df1f61c0823b165c6ca5197051b3514fe96e71ec77cc6d WatchSource:0}: Error finding container b37e9a06348fd99e96df1f61c0823b165c6ca5197051b3514fe96e71ec77cc6d: Status 404 returned error can't find the container with id b37e9a06348fd99e96df1f61c0823b165c6ca5197051b3514fe96e71ec77cc6d Dec 13 06:55:16 crc kubenswrapper[4971]: I1213 06:55:16.153590 4971 patch_prober.go:28] interesting pod/machine-config-daemon-82xjz container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 13 06:55:16 crc kubenswrapper[4971]: I1213 06:55:16.153636 4971 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-82xjz" podUID="e0c4af12-edda-41d7-9b44-f87396174ef2" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 13 06:55:16 crc kubenswrapper[4971]: I1213 06:55:16.176635 4971 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-68kxv" event={"ID":"5eea3e58-f9fb-4529-b645-0f5193d5a8a9","Type":"ContainerStarted","Data":"b37e9a06348fd99e96df1f61c0823b165c6ca5197051b3514fe96e71ec77cc6d"} Dec 13 06:55:16 crc kubenswrapper[4971]: I1213 06:55:16.178202 4971 generic.go:334] "Generic (PLEG): container finished" podID="9fd587ab-4839-4479-80f8-f97d78d833b9" containerID="33ef1eb9944bba090fda89c5cbe65281414e52f467a649d0395eae60a5916a4f" exitCode=0 Dec 13 06:55:16 crc kubenswrapper[4971]: I1213 06:55:16.178280 4971 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-5qtd2" event={"ID":"9fd587ab-4839-4479-80f8-f97d78d833b9","Type":"ContainerDied","Data":"33ef1eb9944bba090fda89c5cbe65281414e52f467a649d0395eae60a5916a4f"} Dec 13 06:55:16 crc kubenswrapper[4971]: I1213 06:55:16.966307 4971 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/community-operators-kbz6d"] Dec 13 06:55:16 crc kubenswrapper[4971]: I1213 06:55:16.967875 4971 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-kbz6d" Dec 13 06:55:16 crc kubenswrapper[4971]: I1213 06:55:16.977083 4971 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"community-operators-dockercfg-dmngl" Dec 13 06:55:16 crc kubenswrapper[4971]: I1213 06:55:16.980660 4971 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-kbz6d"] Dec 13 06:55:17 crc kubenswrapper[4971]: I1213 06:55:17.053352 4971 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-tl949\" (UniqueName: \"kubernetes.io/projected/d44cf79f-dbdf-4e9b-ba48-7351d398c38c-kube-api-access-tl949\") pod \"community-operators-kbz6d\" (UID: \"d44cf79f-dbdf-4e9b-ba48-7351d398c38c\") " pod="openshift-marketplace/community-operators-kbz6d" Dec 13 06:55:17 crc kubenswrapper[4971]: I1213 06:55:17.053418 4971 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/d44cf79f-dbdf-4e9b-ba48-7351d398c38c-catalog-content\") pod \"community-operators-kbz6d\" (UID: \"d44cf79f-dbdf-4e9b-ba48-7351d398c38c\") " pod="openshift-marketplace/community-operators-kbz6d" Dec 13 06:55:17 crc kubenswrapper[4971]: I1213 06:55:17.053446 4971 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/d44cf79f-dbdf-4e9b-ba48-7351d398c38c-utilities\") pod \"community-operators-kbz6d\" (UID: \"d44cf79f-dbdf-4e9b-ba48-7351d398c38c\") " pod="openshift-marketplace/community-operators-kbz6d" Dec 13 06:55:17 crc kubenswrapper[4971]: I1213 06:55:17.155353 4971 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/d44cf79f-dbdf-4e9b-ba48-7351d398c38c-catalog-content\") pod \"community-operators-kbz6d\" (UID: \"d44cf79f-dbdf-4e9b-ba48-7351d398c38c\") " pod="openshift-marketplace/community-operators-kbz6d" Dec 13 06:55:17 crc kubenswrapper[4971]: I1213 06:55:17.155788 4971 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/d44cf79f-dbdf-4e9b-ba48-7351d398c38c-utilities\") pod \"community-operators-kbz6d\" (UID: \"d44cf79f-dbdf-4e9b-ba48-7351d398c38c\") " pod="openshift-marketplace/community-operators-kbz6d" Dec 13 06:55:17 crc kubenswrapper[4971]: I1213 06:55:17.155992 4971 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-tl949\" (UniqueName: \"kubernetes.io/projected/d44cf79f-dbdf-4e9b-ba48-7351d398c38c-kube-api-access-tl949\") pod \"community-operators-kbz6d\" (UID: \"d44cf79f-dbdf-4e9b-ba48-7351d398c38c\") " pod="openshift-marketplace/community-operators-kbz6d" Dec 13 06:55:17 crc kubenswrapper[4971]: I1213 06:55:17.156941 4971 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/d44cf79f-dbdf-4e9b-ba48-7351d398c38c-catalog-content\") pod \"community-operators-kbz6d\" (UID: \"d44cf79f-dbdf-4e9b-ba48-7351d398c38c\") " pod="openshift-marketplace/community-operators-kbz6d" Dec 13 06:55:17 crc kubenswrapper[4971]: I1213 06:55:17.158393 4971 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/d44cf79f-dbdf-4e9b-ba48-7351d398c38c-utilities\") pod \"community-operators-kbz6d\" (UID: \"d44cf79f-dbdf-4e9b-ba48-7351d398c38c\") " pod="openshift-marketplace/community-operators-kbz6d" Dec 13 06:55:17 crc kubenswrapper[4971]: I1213 06:55:17.178952 4971 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-tl949\" (UniqueName: \"kubernetes.io/projected/d44cf79f-dbdf-4e9b-ba48-7351d398c38c-kube-api-access-tl949\") pod \"community-operators-kbz6d\" (UID: \"d44cf79f-dbdf-4e9b-ba48-7351d398c38c\") " pod="openshift-marketplace/community-operators-kbz6d" Dec 13 06:55:17 crc kubenswrapper[4971]: I1213 06:55:17.188081 4971 generic.go:334] "Generic (PLEG): container finished" podID="5eea3e58-f9fb-4529-b645-0f5193d5a8a9" containerID="6566c640713f28f685b0401d2835ce790afd33ffaabf644d1b149c8f84e21298" exitCode=0 Dec 13 06:55:17 crc kubenswrapper[4971]: I1213 06:55:17.188165 4971 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-68kxv" event={"ID":"5eea3e58-f9fb-4529-b645-0f5193d5a8a9","Type":"ContainerDied","Data":"6566c640713f28f685b0401d2835ce790afd33ffaabf644d1b149c8f84e21298"} Dec 13 06:55:17 crc kubenswrapper[4971]: I1213 06:55:17.190746 4971 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-5qtd2" event={"ID":"9fd587ab-4839-4479-80f8-f97d78d833b9","Type":"ContainerStarted","Data":"eb6d0a86233100a22e7b2f97ac48557afe6abc4b331a3f5a0d445a5986cab038"} Dec 13 06:55:17 crc kubenswrapper[4971]: I1213 06:55:17.292044 4971 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-kbz6d" Dec 13 06:55:17 crc kubenswrapper[4971]: I1213 06:55:17.676281 4971 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-kbz6d"] Dec 13 06:55:17 crc kubenswrapper[4971]: W1213 06:55:17.682297 4971 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podd44cf79f_dbdf_4e9b_ba48_7351d398c38c.slice/crio-764b5c5c96560b3780d70e038b8c532f9b2a02615adefa94de4d3a0d72234cd0 WatchSource:0}: Error finding container 764b5c5c96560b3780d70e038b8c532f9b2a02615adefa94de4d3a0d72234cd0: Status 404 returned error can't find the container with id 764b5c5c96560b3780d70e038b8c532f9b2a02615adefa94de4d3a0d72234cd0 Dec 13 06:55:17 crc kubenswrapper[4971]: I1213 06:55:17.967372 4971 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/certified-operators-4pzkf"] Dec 13 06:55:17 crc kubenswrapper[4971]: I1213 06:55:17.968942 4971 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-4pzkf" Dec 13 06:55:17 crc kubenswrapper[4971]: I1213 06:55:17.973627 4971 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"certified-operators-dockercfg-4rs5g" Dec 13 06:55:17 crc kubenswrapper[4971]: I1213 06:55:17.978117 4971 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-4pzkf"] Dec 13 06:55:18 crc kubenswrapper[4971]: I1213 06:55:18.068710 4971 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/97cf4a02-e236-49c1-904f-21b25e87eba4-catalog-content\") pod \"certified-operators-4pzkf\" (UID: \"97cf4a02-e236-49c1-904f-21b25e87eba4\") " pod="openshift-marketplace/certified-operators-4pzkf" Dec 13 06:55:18 crc kubenswrapper[4971]: I1213 06:55:18.068937 4971 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-wnvmg\" (UniqueName: \"kubernetes.io/projected/97cf4a02-e236-49c1-904f-21b25e87eba4-kube-api-access-wnvmg\") pod \"certified-operators-4pzkf\" (UID: \"97cf4a02-e236-49c1-904f-21b25e87eba4\") " pod="openshift-marketplace/certified-operators-4pzkf" Dec 13 06:55:18 crc kubenswrapper[4971]: I1213 06:55:18.069081 4971 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/97cf4a02-e236-49c1-904f-21b25e87eba4-utilities\") pod \"certified-operators-4pzkf\" (UID: \"97cf4a02-e236-49c1-904f-21b25e87eba4\") " pod="openshift-marketplace/certified-operators-4pzkf" Dec 13 06:55:18 crc kubenswrapper[4971]: I1213 06:55:18.170798 4971 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-wnvmg\" (UniqueName: \"kubernetes.io/projected/97cf4a02-e236-49c1-904f-21b25e87eba4-kube-api-access-wnvmg\") pod \"certified-operators-4pzkf\" (UID: \"97cf4a02-e236-49c1-904f-21b25e87eba4\") " pod="openshift-marketplace/certified-operators-4pzkf" Dec 13 06:55:18 crc kubenswrapper[4971]: I1213 06:55:18.170901 4971 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/97cf4a02-e236-49c1-904f-21b25e87eba4-utilities\") pod \"certified-operators-4pzkf\" (UID: \"97cf4a02-e236-49c1-904f-21b25e87eba4\") " pod="openshift-marketplace/certified-operators-4pzkf" Dec 13 06:55:18 crc kubenswrapper[4971]: I1213 06:55:18.170962 4971 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/97cf4a02-e236-49c1-904f-21b25e87eba4-catalog-content\") pod \"certified-operators-4pzkf\" (UID: \"97cf4a02-e236-49c1-904f-21b25e87eba4\") " pod="openshift-marketplace/certified-operators-4pzkf" Dec 13 06:55:18 crc kubenswrapper[4971]: I1213 06:55:18.171600 4971 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/97cf4a02-e236-49c1-904f-21b25e87eba4-catalog-content\") pod \"certified-operators-4pzkf\" (UID: \"97cf4a02-e236-49c1-904f-21b25e87eba4\") " pod="openshift-marketplace/certified-operators-4pzkf" Dec 13 06:55:18 crc kubenswrapper[4971]: I1213 06:55:18.171708 4971 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/97cf4a02-e236-49c1-904f-21b25e87eba4-utilities\") pod \"certified-operators-4pzkf\" (UID: \"97cf4a02-e236-49c1-904f-21b25e87eba4\") " pod="openshift-marketplace/certified-operators-4pzkf" Dec 13 06:55:18 crc kubenswrapper[4971]: I1213 06:55:18.194856 4971 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-wnvmg\" (UniqueName: \"kubernetes.io/projected/97cf4a02-e236-49c1-904f-21b25e87eba4-kube-api-access-wnvmg\") pod \"certified-operators-4pzkf\" (UID: \"97cf4a02-e236-49c1-904f-21b25e87eba4\") " pod="openshift-marketplace/certified-operators-4pzkf" Dec 13 06:55:18 crc kubenswrapper[4971]: I1213 06:55:18.223533 4971 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-5qtd2" event={"ID":"9fd587ab-4839-4479-80f8-f97d78d833b9","Type":"ContainerDied","Data":"eb6d0a86233100a22e7b2f97ac48557afe6abc4b331a3f5a0d445a5986cab038"} Dec 13 06:55:18 crc kubenswrapper[4971]: I1213 06:55:18.223730 4971 generic.go:334] "Generic (PLEG): container finished" podID="9fd587ab-4839-4479-80f8-f97d78d833b9" containerID="eb6d0a86233100a22e7b2f97ac48557afe6abc4b331a3f5a0d445a5986cab038" exitCode=0 Dec 13 06:55:18 crc kubenswrapper[4971]: I1213 06:55:18.226822 4971 generic.go:334] "Generic (PLEG): container finished" podID="d44cf79f-dbdf-4e9b-ba48-7351d398c38c" containerID="f1d0237bf148bee90f27653f9d643336dd61d6e7de7d05f7bd4324aeaca9dffb" exitCode=0 Dec 13 06:55:18 crc kubenswrapper[4971]: I1213 06:55:18.226855 4971 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-kbz6d" event={"ID":"d44cf79f-dbdf-4e9b-ba48-7351d398c38c","Type":"ContainerDied","Data":"f1d0237bf148bee90f27653f9d643336dd61d6e7de7d05f7bd4324aeaca9dffb"} Dec 13 06:55:18 crc kubenswrapper[4971]: I1213 06:55:18.226872 4971 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-kbz6d" event={"ID":"d44cf79f-dbdf-4e9b-ba48-7351d398c38c","Type":"ContainerStarted","Data":"764b5c5c96560b3780d70e038b8c532f9b2a02615adefa94de4d3a0d72234cd0"} Dec 13 06:55:18 crc kubenswrapper[4971]: I1213 06:55:18.290199 4971 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-4pzkf" Dec 13 06:55:18 crc kubenswrapper[4971]: I1213 06:55:18.706467 4971 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-4pzkf"] Dec 13 06:55:19 crc kubenswrapper[4971]: I1213 06:55:19.234041 4971 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-5qtd2" event={"ID":"9fd587ab-4839-4479-80f8-f97d78d833b9","Type":"ContainerStarted","Data":"91507b889fb1db4fed8d82e2f27769c51087db01b1cd1c9b2b06d317b8ef3ff0"} Dec 13 06:55:19 crc kubenswrapper[4971]: I1213 06:55:19.236190 4971 generic.go:334] "Generic (PLEG): container finished" podID="97cf4a02-e236-49c1-904f-21b25e87eba4" containerID="5eeb9f49b31482f869b7d83a6c04554577d9b80da86141ce746ec2edfdce19c5" exitCode=0 Dec 13 06:55:19 crc kubenswrapper[4971]: I1213 06:55:19.236241 4971 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-4pzkf" event={"ID":"97cf4a02-e236-49c1-904f-21b25e87eba4","Type":"ContainerDied","Data":"5eeb9f49b31482f869b7d83a6c04554577d9b80da86141ce746ec2edfdce19c5"} Dec 13 06:55:19 crc kubenswrapper[4971]: I1213 06:55:19.236259 4971 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-4pzkf" event={"ID":"97cf4a02-e236-49c1-904f-21b25e87eba4","Type":"ContainerStarted","Data":"8b5a62fad31d2bf2ba6c9d85a43034e6ad383943670a489067b36c2011e02513"} Dec 13 06:55:19 crc kubenswrapper[4971]: I1213 06:55:19.241693 4971 generic.go:334] "Generic (PLEG): container finished" podID="5eea3e58-f9fb-4529-b645-0f5193d5a8a9" containerID="48d12ce796ac32c962947d15f042cb190a9dacbd93a6c1aa9892bf9301e57461" exitCode=0 Dec 13 06:55:19 crc kubenswrapper[4971]: I1213 06:55:19.241835 4971 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-68kxv" event={"ID":"5eea3e58-f9fb-4529-b645-0f5193d5a8a9","Type":"ContainerDied","Data":"48d12ce796ac32c962947d15f042cb190a9dacbd93a6c1aa9892bf9301e57461"} Dec 13 06:55:19 crc kubenswrapper[4971]: I1213 06:55:19.247275 4971 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-kbz6d" event={"ID":"d44cf79f-dbdf-4e9b-ba48-7351d398c38c","Type":"ContainerStarted","Data":"9c03c8d63a2c6e9b660e112dcfa1f170515d6921c706e6db2bfc1bb20df69234"} Dec 13 06:55:19 crc kubenswrapper[4971]: I1213 06:55:19.257422 4971 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-operators-5qtd2" podStartSLOduration=2.546648336 podStartE2EDuration="5.257401148s" podCreationTimestamp="2025-12-13 06:55:14 +0000 UTC" firstStartedPulling="2025-12-13 06:55:16.179630014 +0000 UTC m=+372.784039462" lastFinishedPulling="2025-12-13 06:55:18.890382826 +0000 UTC m=+375.494792274" observedRunningTime="2025-12-13 06:55:19.253636007 +0000 UTC m=+375.858045465" watchObservedRunningTime="2025-12-13 06:55:19.257401148 +0000 UTC m=+375.861810596" Dec 13 06:55:20 crc kubenswrapper[4971]: I1213 06:55:20.253481 4971 generic.go:334] "Generic (PLEG): container finished" podID="d44cf79f-dbdf-4e9b-ba48-7351d398c38c" containerID="9c03c8d63a2c6e9b660e112dcfa1f170515d6921c706e6db2bfc1bb20df69234" exitCode=0 Dec 13 06:55:20 crc kubenswrapper[4971]: I1213 06:55:20.253578 4971 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-kbz6d" event={"ID":"d44cf79f-dbdf-4e9b-ba48-7351d398c38c","Type":"ContainerDied","Data":"9c03c8d63a2c6e9b660e112dcfa1f170515d6921c706e6db2bfc1bb20df69234"} Dec 13 06:55:20 crc kubenswrapper[4971]: I1213 06:55:20.256015 4971 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-4pzkf" event={"ID":"97cf4a02-e236-49c1-904f-21b25e87eba4","Type":"ContainerStarted","Data":"39bded6183ae1297006f892ec4367021f144a8659108a22755ab06e082f8574f"} Dec 13 06:55:20 crc kubenswrapper[4971]: I1213 06:55:20.258636 4971 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-68kxv" event={"ID":"5eea3e58-f9fb-4529-b645-0f5193d5a8a9","Type":"ContainerStarted","Data":"937af01c300b74d0c52a530d08bf8ad033b99588471fb7be33345a2e3fb07a39"} Dec 13 06:55:20 crc kubenswrapper[4971]: I1213 06:55:20.287780 4971 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-marketplace-68kxv" podStartSLOduration=2.607447372 podStartE2EDuration="5.287758791s" podCreationTimestamp="2025-12-13 06:55:15 +0000 UTC" firstStartedPulling="2025-12-13 06:55:17.190173904 +0000 UTC m=+373.794583392" lastFinishedPulling="2025-12-13 06:55:19.870485323 +0000 UTC m=+376.474894811" observedRunningTime="2025-12-13 06:55:20.28688597 +0000 UTC m=+376.891295418" watchObservedRunningTime="2025-12-13 06:55:20.287758791 +0000 UTC m=+376.892168239" Dec 13 06:55:21 crc kubenswrapper[4971]: I1213 06:55:21.266962 4971 generic.go:334] "Generic (PLEG): container finished" podID="97cf4a02-e236-49c1-904f-21b25e87eba4" containerID="39bded6183ae1297006f892ec4367021f144a8659108a22755ab06e082f8574f" exitCode=0 Dec 13 06:55:21 crc kubenswrapper[4971]: I1213 06:55:21.267209 4971 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-4pzkf" event={"ID":"97cf4a02-e236-49c1-904f-21b25e87eba4","Type":"ContainerDied","Data":"39bded6183ae1297006f892ec4367021f144a8659108a22755ab06e082f8574f"} Dec 13 06:55:21 crc kubenswrapper[4971]: I1213 06:55:21.270536 4971 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-kbz6d" event={"ID":"d44cf79f-dbdf-4e9b-ba48-7351d398c38c","Type":"ContainerStarted","Data":"0cbd2685b75d6e865d414a1ccb83ec0b3fa61e649b012dabdb47bd9f09c8c2c2"} Dec 13 06:55:21 crc kubenswrapper[4971]: I1213 06:55:21.306570 4971 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/community-operators-kbz6d" podStartSLOduration=2.669261482 podStartE2EDuration="5.306549792s" podCreationTimestamp="2025-12-13 06:55:16 +0000 UTC" firstStartedPulling="2025-12-13 06:55:18.230846628 +0000 UTC m=+374.835256086" lastFinishedPulling="2025-12-13 06:55:20.868134948 +0000 UTC m=+377.472544396" observedRunningTime="2025-12-13 06:55:21.302630176 +0000 UTC m=+377.907039644" watchObservedRunningTime="2025-12-13 06:55:21.306549792 +0000 UTC m=+377.910959240" Dec 13 06:55:24 crc kubenswrapper[4971]: I1213 06:55:24.287032 4971 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-4pzkf" event={"ID":"97cf4a02-e236-49c1-904f-21b25e87eba4","Type":"ContainerStarted","Data":"47325854a36fbd957987f114286b98caf8038283e15ce75adc5c479b0787910c"} Dec 13 06:55:24 crc kubenswrapper[4971]: I1213 06:55:24.311185 4971 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/certified-operators-4pzkf" podStartSLOduration=3.795891477 podStartE2EDuration="7.311167162s" podCreationTimestamp="2025-12-13 06:55:17 +0000 UTC" firstStartedPulling="2025-12-13 06:55:19.237942214 +0000 UTC m=+375.842351662" lastFinishedPulling="2025-12-13 06:55:22.753217899 +0000 UTC m=+379.357627347" observedRunningTime="2025-12-13 06:55:24.306602121 +0000 UTC m=+380.911011589" watchObservedRunningTime="2025-12-13 06:55:24.311167162 +0000 UTC m=+380.915576610" Dec 13 06:55:24 crc kubenswrapper[4971]: I1213 06:55:24.697896 4971 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-operators-5qtd2" Dec 13 06:55:24 crc kubenswrapper[4971]: I1213 06:55:24.697948 4971 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-operators-5qtd2" Dec 13 06:55:24 crc kubenswrapper[4971]: I1213 06:55:24.735981 4971 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-operators-5qtd2" Dec 13 06:55:25 crc kubenswrapper[4971]: I1213 06:55:25.330602 4971 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-operators-5qtd2" Dec 13 06:55:25 crc kubenswrapper[4971]: I1213 06:55:25.678408 4971 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-marketplace-68kxv" Dec 13 06:55:25 crc kubenswrapper[4971]: I1213 06:55:25.678810 4971 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-marketplace-68kxv" Dec 13 06:55:25 crc kubenswrapper[4971]: I1213 06:55:25.718006 4971 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-marketplace-68kxv" Dec 13 06:55:26 crc kubenswrapper[4971]: I1213 06:55:26.335584 4971 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-marketplace-68kxv" Dec 13 06:55:27 crc kubenswrapper[4971]: I1213 06:55:27.292543 4971 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/community-operators-kbz6d" Dec 13 06:55:27 crc kubenswrapper[4971]: I1213 06:55:27.292787 4971 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/community-operators-kbz6d" Dec 13 06:55:27 crc kubenswrapper[4971]: I1213 06:55:27.329019 4971 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/community-operators-kbz6d" Dec 13 06:55:28 crc kubenswrapper[4971]: I1213 06:55:28.290489 4971 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/certified-operators-4pzkf" Dec 13 06:55:28 crc kubenswrapper[4971]: I1213 06:55:28.290563 4971 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/certified-operators-4pzkf" Dec 13 06:55:28 crc kubenswrapper[4971]: I1213 06:55:28.333816 4971 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/certified-operators-4pzkf" Dec 13 06:55:28 crc kubenswrapper[4971]: I1213 06:55:28.349718 4971 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/community-operators-kbz6d" Dec 13 06:55:28 crc kubenswrapper[4971]: I1213 06:55:28.380456 4971 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/certified-operators-4pzkf" Dec 13 06:55:34 crc kubenswrapper[4971]: I1213 06:55:34.257038 4971 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-image-registry/image-registry-66df7c8f76-xjwwz" Dec 13 06:55:34 crc kubenswrapper[4971]: I1213 06:55:34.301690 4971 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-image-registry/image-registry-697d97f7c8-mdd8q"] Dec 13 06:55:46 crc kubenswrapper[4971]: I1213 06:55:46.153558 4971 patch_prober.go:28] interesting pod/machine-config-daemon-82xjz container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 13 06:55:46 crc kubenswrapper[4971]: I1213 06:55:46.153924 4971 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-82xjz" podUID="e0c4af12-edda-41d7-9b44-f87396174ef2" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 13 06:55:59 crc kubenswrapper[4971]: I1213 06:55:59.339122 4971 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-image-registry/image-registry-697d97f7c8-mdd8q" podUID="cba893c6-ff99-466a-9db0-a62e7a87a334" containerName="registry" containerID="cri-o://db855d10cdb4b3b9b54e0ece89243b84cac1c530b376b33acceeff82e7f769b3" gracePeriod=30 Dec 13 06:55:59 crc kubenswrapper[4971]: I1213 06:55:59.697744 4971 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-image-registry/image-registry-697d97f7c8-mdd8q" Dec 13 06:55:59 crc kubenswrapper[4971]: I1213 06:55:59.838787 4971 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-fs22c\" (UniqueName: \"kubernetes.io/projected/cba893c6-ff99-466a-9db0-a62e7a87a334-kube-api-access-fs22c\") pod \"cba893c6-ff99-466a-9db0-a62e7a87a334\" (UID: \"cba893c6-ff99-466a-9db0-a62e7a87a334\") " Dec 13 06:55:59 crc kubenswrapper[4971]: I1213 06:55:59.838863 4971 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"installation-pull-secrets\" (UniqueName: \"kubernetes.io/secret/cba893c6-ff99-466a-9db0-a62e7a87a334-installation-pull-secrets\") pod \"cba893c6-ff99-466a-9db0-a62e7a87a334\" (UID: \"cba893c6-ff99-466a-9db0-a62e7a87a334\") " Dec 13 06:55:59 crc kubenswrapper[4971]: I1213 06:55:59.838912 4971 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"registry-certificates\" (UniqueName: \"kubernetes.io/configmap/cba893c6-ff99-466a-9db0-a62e7a87a334-registry-certificates\") pod \"cba893c6-ff99-466a-9db0-a62e7a87a334\" (UID: \"cba893c6-ff99-466a-9db0-a62e7a87a334\") " Dec 13 06:55:59 crc kubenswrapper[4971]: I1213 06:55:59.838949 4971 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/cba893c6-ff99-466a-9db0-a62e7a87a334-trusted-ca\") pod \"cba893c6-ff99-466a-9db0-a62e7a87a334\" (UID: \"cba893c6-ff99-466a-9db0-a62e7a87a334\") " Dec 13 06:55:59 crc kubenswrapper[4971]: I1213 06:55:59.839001 4971 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ca-trust-extracted\" (UniqueName: \"kubernetes.io/empty-dir/cba893c6-ff99-466a-9db0-a62e7a87a334-ca-trust-extracted\") pod \"cba893c6-ff99-466a-9db0-a62e7a87a334\" (UID: \"cba893c6-ff99-466a-9db0-a62e7a87a334\") " Dec 13 06:55:59 crc kubenswrapper[4971]: I1213 06:55:59.839031 4971 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"registry-tls\" (UniqueName: \"kubernetes.io/projected/cba893c6-ff99-466a-9db0-a62e7a87a334-registry-tls\") pod \"cba893c6-ff99-466a-9db0-a62e7a87a334\" (UID: \"cba893c6-ff99-466a-9db0-a62e7a87a334\") " Dec 13 06:55:59 crc kubenswrapper[4971]: I1213 06:55:59.839058 4971 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/cba893c6-ff99-466a-9db0-a62e7a87a334-bound-sa-token\") pod \"cba893c6-ff99-466a-9db0-a62e7a87a334\" (UID: \"cba893c6-ff99-466a-9db0-a62e7a87a334\") " Dec 13 06:55:59 crc kubenswrapper[4971]: I1213 06:55:59.839232 4971 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"registry-storage\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"cba893c6-ff99-466a-9db0-a62e7a87a334\" (UID: \"cba893c6-ff99-466a-9db0-a62e7a87a334\") " Dec 13 06:55:59 crc kubenswrapper[4971]: I1213 06:55:59.840097 4971 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/cba893c6-ff99-466a-9db0-a62e7a87a334-trusted-ca" (OuterVolumeSpecName: "trusted-ca") pod "cba893c6-ff99-466a-9db0-a62e7a87a334" (UID: "cba893c6-ff99-466a-9db0-a62e7a87a334"). InnerVolumeSpecName "trusted-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 13 06:55:59 crc kubenswrapper[4971]: I1213 06:55:59.840230 4971 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/cba893c6-ff99-466a-9db0-a62e7a87a334-registry-certificates" (OuterVolumeSpecName: "registry-certificates") pod "cba893c6-ff99-466a-9db0-a62e7a87a334" (UID: "cba893c6-ff99-466a-9db0-a62e7a87a334"). InnerVolumeSpecName "registry-certificates". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 13 06:55:59 crc kubenswrapper[4971]: I1213 06:55:59.844737 4971 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/cba893c6-ff99-466a-9db0-a62e7a87a334-registry-tls" (OuterVolumeSpecName: "registry-tls") pod "cba893c6-ff99-466a-9db0-a62e7a87a334" (UID: "cba893c6-ff99-466a-9db0-a62e7a87a334"). InnerVolumeSpecName "registry-tls". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 13 06:55:59 crc kubenswrapper[4971]: I1213 06:55:59.844976 4971 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/cba893c6-ff99-466a-9db0-a62e7a87a334-kube-api-access-fs22c" (OuterVolumeSpecName: "kube-api-access-fs22c") pod "cba893c6-ff99-466a-9db0-a62e7a87a334" (UID: "cba893c6-ff99-466a-9db0-a62e7a87a334"). InnerVolumeSpecName "kube-api-access-fs22c". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 13 06:55:59 crc kubenswrapper[4971]: I1213 06:55:59.846570 4971 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/cba893c6-ff99-466a-9db0-a62e7a87a334-installation-pull-secrets" (OuterVolumeSpecName: "installation-pull-secrets") pod "cba893c6-ff99-466a-9db0-a62e7a87a334" (UID: "cba893c6-ff99-466a-9db0-a62e7a87a334"). InnerVolumeSpecName "installation-pull-secrets". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 13 06:55:59 crc kubenswrapper[4971]: I1213 06:55:59.847659 4971 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (OuterVolumeSpecName: "registry-storage") pod "cba893c6-ff99-466a-9db0-a62e7a87a334" (UID: "cba893c6-ff99-466a-9db0-a62e7a87a334"). InnerVolumeSpecName "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8". PluginName "kubernetes.io/csi", VolumeGidValue "" Dec 13 06:55:59 crc kubenswrapper[4971]: I1213 06:55:59.856167 4971 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/cba893c6-ff99-466a-9db0-a62e7a87a334-bound-sa-token" (OuterVolumeSpecName: "bound-sa-token") pod "cba893c6-ff99-466a-9db0-a62e7a87a334" (UID: "cba893c6-ff99-466a-9db0-a62e7a87a334"). InnerVolumeSpecName "bound-sa-token". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 13 06:55:59 crc kubenswrapper[4971]: I1213 06:55:59.856541 4971 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/cba893c6-ff99-466a-9db0-a62e7a87a334-ca-trust-extracted" (OuterVolumeSpecName: "ca-trust-extracted") pod "cba893c6-ff99-466a-9db0-a62e7a87a334" (UID: "cba893c6-ff99-466a-9db0-a62e7a87a334"). InnerVolumeSpecName "ca-trust-extracted". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 13 06:55:59 crc kubenswrapper[4971]: I1213 06:55:59.941911 4971 reconciler_common.go:293] "Volume detached for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/cba893c6-ff99-466a-9db0-a62e7a87a334-trusted-ca\") on node \"crc\" DevicePath \"\"" Dec 13 06:55:59 crc kubenswrapper[4971]: I1213 06:55:59.942214 4971 reconciler_common.go:293] "Volume detached for volume \"ca-trust-extracted\" (UniqueName: \"kubernetes.io/empty-dir/cba893c6-ff99-466a-9db0-a62e7a87a334-ca-trust-extracted\") on node \"crc\" DevicePath \"\"" Dec 13 06:55:59 crc kubenswrapper[4971]: I1213 06:55:59.942232 4971 reconciler_common.go:293] "Volume detached for volume \"registry-tls\" (UniqueName: \"kubernetes.io/projected/cba893c6-ff99-466a-9db0-a62e7a87a334-registry-tls\") on node \"crc\" DevicePath \"\"" Dec 13 06:55:59 crc kubenswrapper[4971]: I1213 06:55:59.942241 4971 reconciler_common.go:293] "Volume detached for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/cba893c6-ff99-466a-9db0-a62e7a87a334-bound-sa-token\") on node \"crc\" DevicePath \"\"" Dec 13 06:55:59 crc kubenswrapper[4971]: I1213 06:55:59.942250 4971 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-fs22c\" (UniqueName: \"kubernetes.io/projected/cba893c6-ff99-466a-9db0-a62e7a87a334-kube-api-access-fs22c\") on node \"crc\" DevicePath \"\"" Dec 13 06:55:59 crc kubenswrapper[4971]: I1213 06:55:59.942258 4971 reconciler_common.go:293] "Volume detached for volume \"installation-pull-secrets\" (UniqueName: \"kubernetes.io/secret/cba893c6-ff99-466a-9db0-a62e7a87a334-installation-pull-secrets\") on node \"crc\" DevicePath \"\"" Dec 13 06:55:59 crc kubenswrapper[4971]: I1213 06:55:59.942267 4971 reconciler_common.go:293] "Volume detached for volume \"registry-certificates\" (UniqueName: \"kubernetes.io/configmap/cba893c6-ff99-466a-9db0-a62e7a87a334-registry-certificates\") on node \"crc\" DevicePath \"\"" Dec 13 06:56:00 crc kubenswrapper[4971]: I1213 06:56:00.101300 4971 generic.go:334] "Generic (PLEG): container finished" podID="cba893c6-ff99-466a-9db0-a62e7a87a334" containerID="db855d10cdb4b3b9b54e0ece89243b84cac1c530b376b33acceeff82e7f769b3" exitCode=0 Dec 13 06:56:00 crc kubenswrapper[4971]: I1213 06:56:00.101339 4971 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-image-registry/image-registry-697d97f7c8-mdd8q" event={"ID":"cba893c6-ff99-466a-9db0-a62e7a87a334","Type":"ContainerDied","Data":"db855d10cdb4b3b9b54e0ece89243b84cac1c530b376b33acceeff82e7f769b3"} Dec 13 06:56:00 crc kubenswrapper[4971]: I1213 06:56:00.101356 4971 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-image-registry/image-registry-697d97f7c8-mdd8q" Dec 13 06:56:00 crc kubenswrapper[4971]: I1213 06:56:00.101375 4971 scope.go:117] "RemoveContainer" containerID="db855d10cdb4b3b9b54e0ece89243b84cac1c530b376b33acceeff82e7f769b3" Dec 13 06:56:00 crc kubenswrapper[4971]: I1213 06:56:00.101363 4971 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-image-registry/image-registry-697d97f7c8-mdd8q" event={"ID":"cba893c6-ff99-466a-9db0-a62e7a87a334","Type":"ContainerDied","Data":"7ba0c024c728c897f54acbb50586782f4f871548f0d1f5406f676b8e1beeefeb"} Dec 13 06:56:00 crc kubenswrapper[4971]: I1213 06:56:00.118127 4971 scope.go:117] "RemoveContainer" containerID="db855d10cdb4b3b9b54e0ece89243b84cac1c530b376b33acceeff82e7f769b3" Dec 13 06:56:00 crc kubenswrapper[4971]: E1213 06:56:00.118475 4971 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"db855d10cdb4b3b9b54e0ece89243b84cac1c530b376b33acceeff82e7f769b3\": container with ID starting with db855d10cdb4b3b9b54e0ece89243b84cac1c530b376b33acceeff82e7f769b3 not found: ID does not exist" containerID="db855d10cdb4b3b9b54e0ece89243b84cac1c530b376b33acceeff82e7f769b3" Dec 13 06:56:00 crc kubenswrapper[4971]: I1213 06:56:00.118532 4971 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"db855d10cdb4b3b9b54e0ece89243b84cac1c530b376b33acceeff82e7f769b3"} err="failed to get container status \"db855d10cdb4b3b9b54e0ece89243b84cac1c530b376b33acceeff82e7f769b3\": rpc error: code = NotFound desc = could not find container \"db855d10cdb4b3b9b54e0ece89243b84cac1c530b376b33acceeff82e7f769b3\": container with ID starting with db855d10cdb4b3b9b54e0ece89243b84cac1c530b376b33acceeff82e7f769b3 not found: ID does not exist" Dec 13 06:56:00 crc kubenswrapper[4971]: I1213 06:56:00.134893 4971 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-image-registry/image-registry-697d97f7c8-mdd8q"] Dec 13 06:56:00 crc kubenswrapper[4971]: I1213 06:56:00.139114 4971 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-image-registry/image-registry-697d97f7c8-mdd8q"] Dec 13 06:56:01 crc kubenswrapper[4971]: I1213 06:56:01.777698 4971 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="cba893c6-ff99-466a-9db0-a62e7a87a334" path="/var/lib/kubelet/pods/cba893c6-ff99-466a-9db0-a62e7a87a334/volumes" Dec 13 06:56:16 crc kubenswrapper[4971]: I1213 06:56:16.153150 4971 patch_prober.go:28] interesting pod/machine-config-daemon-82xjz container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 13 06:56:16 crc kubenswrapper[4971]: I1213 06:56:16.153769 4971 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-82xjz" podUID="e0c4af12-edda-41d7-9b44-f87396174ef2" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 13 06:56:16 crc kubenswrapper[4971]: I1213 06:56:16.153822 4971 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-82xjz" Dec 13 06:56:16 crc kubenswrapper[4971]: I1213 06:56:16.154440 4971 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"b43d8045f85a66d814d3d526124e6e6bee306cba177b3bfd2a41d456c09be2eb"} pod="openshift-machine-config-operator/machine-config-daemon-82xjz" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Dec 13 06:56:16 crc kubenswrapper[4971]: I1213 06:56:16.154502 4971 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-82xjz" podUID="e0c4af12-edda-41d7-9b44-f87396174ef2" containerName="machine-config-daemon" containerID="cri-o://b43d8045f85a66d814d3d526124e6e6bee306cba177b3bfd2a41d456c09be2eb" gracePeriod=600 Dec 13 06:56:17 crc kubenswrapper[4971]: I1213 06:56:17.186383 4971 generic.go:334] "Generic (PLEG): container finished" podID="e0c4af12-edda-41d7-9b44-f87396174ef2" containerID="b43d8045f85a66d814d3d526124e6e6bee306cba177b3bfd2a41d456c09be2eb" exitCode=0 Dec 13 06:56:17 crc kubenswrapper[4971]: I1213 06:56:17.186468 4971 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-82xjz" event={"ID":"e0c4af12-edda-41d7-9b44-f87396174ef2","Type":"ContainerDied","Data":"b43d8045f85a66d814d3d526124e6e6bee306cba177b3bfd2a41d456c09be2eb"} Dec 13 06:56:17 crc kubenswrapper[4971]: I1213 06:56:17.186795 4971 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-82xjz" event={"ID":"e0c4af12-edda-41d7-9b44-f87396174ef2","Type":"ContainerStarted","Data":"85537ceec5f79277ceb7773d56141774b3a454a245f9047cd2720b76888daff3"} Dec 13 06:56:17 crc kubenswrapper[4971]: I1213 06:56:17.186829 4971 scope.go:117] "RemoveContainer" containerID="7e0d2020e62f1a277cff7047e1907dc747ed393a1f6c983e1ee258bd156c16c7" Dec 13 06:58:16 crc kubenswrapper[4971]: I1213 06:58:16.154198 4971 patch_prober.go:28] interesting pod/machine-config-daemon-82xjz container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 13 06:58:16 crc kubenswrapper[4971]: I1213 06:58:16.155401 4971 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-82xjz" podUID="e0c4af12-edda-41d7-9b44-f87396174ef2" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 13 06:58:46 crc kubenswrapper[4971]: I1213 06:58:46.154114 4971 patch_prober.go:28] interesting pod/machine-config-daemon-82xjz container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 13 06:58:46 crc kubenswrapper[4971]: I1213 06:58:46.154640 4971 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-82xjz" podUID="e0c4af12-edda-41d7-9b44-f87396174ef2" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 13 06:59:16 crc kubenswrapper[4971]: I1213 06:59:16.154269 4971 patch_prober.go:28] interesting pod/machine-config-daemon-82xjz container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 13 06:59:16 crc kubenswrapper[4971]: I1213 06:59:16.154993 4971 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-82xjz" podUID="e0c4af12-edda-41d7-9b44-f87396174ef2" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 13 06:59:16 crc kubenswrapper[4971]: I1213 06:59:16.155069 4971 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-82xjz" Dec 13 06:59:16 crc kubenswrapper[4971]: I1213 06:59:16.155880 4971 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"85537ceec5f79277ceb7773d56141774b3a454a245f9047cd2720b76888daff3"} pod="openshift-machine-config-operator/machine-config-daemon-82xjz" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Dec 13 06:59:16 crc kubenswrapper[4971]: I1213 06:59:16.155972 4971 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-82xjz" podUID="e0c4af12-edda-41d7-9b44-f87396174ef2" containerName="machine-config-daemon" containerID="cri-o://85537ceec5f79277ceb7773d56141774b3a454a245f9047cd2720b76888daff3" gracePeriod=600 Dec 13 06:59:17 crc kubenswrapper[4971]: I1213 06:59:17.123757 4971 generic.go:334] "Generic (PLEG): container finished" podID="e0c4af12-edda-41d7-9b44-f87396174ef2" containerID="85537ceec5f79277ceb7773d56141774b3a454a245f9047cd2720b76888daff3" exitCode=0 Dec 13 06:59:17 crc kubenswrapper[4971]: I1213 06:59:17.123814 4971 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-82xjz" event={"ID":"e0c4af12-edda-41d7-9b44-f87396174ef2","Type":"ContainerDied","Data":"85537ceec5f79277ceb7773d56141774b3a454a245f9047cd2720b76888daff3"} Dec 13 06:59:17 crc kubenswrapper[4971]: I1213 06:59:17.124033 4971 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-82xjz" event={"ID":"e0c4af12-edda-41d7-9b44-f87396174ef2","Type":"ContainerStarted","Data":"358e8ac17d5a2f5ef3b4deef1d163694a45d4c70827405c28f4c81a7c84b880e"} Dec 13 06:59:17 crc kubenswrapper[4971]: I1213 06:59:17.124056 4971 scope.go:117] "RemoveContainer" containerID="b43d8045f85a66d814d3d526124e6e6bee306cba177b3bfd2a41d456c09be2eb" Dec 13 07:00:00 crc kubenswrapper[4971]: I1213 07:00:00.165096 4971 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29426820-pbkhw"] Dec 13 07:00:00 crc kubenswrapper[4971]: E1213 07:00:00.165977 4971 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="cba893c6-ff99-466a-9db0-a62e7a87a334" containerName="registry" Dec 13 07:00:00 crc kubenswrapper[4971]: I1213 07:00:00.165994 4971 state_mem.go:107] "Deleted CPUSet assignment" podUID="cba893c6-ff99-466a-9db0-a62e7a87a334" containerName="registry" Dec 13 07:00:00 crc kubenswrapper[4971]: I1213 07:00:00.166111 4971 memory_manager.go:354] "RemoveStaleState removing state" podUID="cba893c6-ff99-466a-9db0-a62e7a87a334" containerName="registry" Dec 13 07:00:00 crc kubenswrapper[4971]: I1213 07:00:00.166581 4971 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29426820-pbkhw" Dec 13 07:00:00 crc kubenswrapper[4971]: I1213 07:00:00.169060 4971 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-operator-lifecycle-manager"/"collect-profiles-config" Dec 13 07:00:00 crc kubenswrapper[4971]: I1213 07:00:00.169309 4971 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"collect-profiles-dockercfg-kzf4t" Dec 13 07:00:00 crc kubenswrapper[4971]: I1213 07:00:00.181729 4971 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29426820-pbkhw"] Dec 13 07:00:00 crc kubenswrapper[4971]: I1213 07:00:00.206604 4971 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/48585415-a489-4ea0-9d68-981bbebd10cb-secret-volume\") pod \"collect-profiles-29426820-pbkhw\" (UID: \"48585415-a489-4ea0-9d68-981bbebd10cb\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29426820-pbkhw" Dec 13 07:00:00 crc kubenswrapper[4971]: I1213 07:00:00.206790 4971 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-4lgnd\" (UniqueName: \"kubernetes.io/projected/48585415-a489-4ea0-9d68-981bbebd10cb-kube-api-access-4lgnd\") pod \"collect-profiles-29426820-pbkhw\" (UID: \"48585415-a489-4ea0-9d68-981bbebd10cb\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29426820-pbkhw" Dec 13 07:00:00 crc kubenswrapper[4971]: I1213 07:00:00.206830 4971 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/48585415-a489-4ea0-9d68-981bbebd10cb-config-volume\") pod \"collect-profiles-29426820-pbkhw\" (UID: \"48585415-a489-4ea0-9d68-981bbebd10cb\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29426820-pbkhw" Dec 13 07:00:00 crc kubenswrapper[4971]: I1213 07:00:00.307759 4971 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/48585415-a489-4ea0-9d68-981bbebd10cb-secret-volume\") pod \"collect-profiles-29426820-pbkhw\" (UID: \"48585415-a489-4ea0-9d68-981bbebd10cb\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29426820-pbkhw" Dec 13 07:00:00 crc kubenswrapper[4971]: I1213 07:00:00.307807 4971 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-4lgnd\" (UniqueName: \"kubernetes.io/projected/48585415-a489-4ea0-9d68-981bbebd10cb-kube-api-access-4lgnd\") pod \"collect-profiles-29426820-pbkhw\" (UID: \"48585415-a489-4ea0-9d68-981bbebd10cb\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29426820-pbkhw" Dec 13 07:00:00 crc kubenswrapper[4971]: I1213 07:00:00.307827 4971 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/48585415-a489-4ea0-9d68-981bbebd10cb-config-volume\") pod \"collect-profiles-29426820-pbkhw\" (UID: \"48585415-a489-4ea0-9d68-981bbebd10cb\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29426820-pbkhw" Dec 13 07:00:00 crc kubenswrapper[4971]: I1213 07:00:00.309238 4971 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/48585415-a489-4ea0-9d68-981bbebd10cb-config-volume\") pod \"collect-profiles-29426820-pbkhw\" (UID: \"48585415-a489-4ea0-9d68-981bbebd10cb\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29426820-pbkhw" Dec 13 07:00:00 crc kubenswrapper[4971]: I1213 07:00:00.316005 4971 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/48585415-a489-4ea0-9d68-981bbebd10cb-secret-volume\") pod \"collect-profiles-29426820-pbkhw\" (UID: \"48585415-a489-4ea0-9d68-981bbebd10cb\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29426820-pbkhw" Dec 13 07:00:00 crc kubenswrapper[4971]: I1213 07:00:00.325124 4971 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-4lgnd\" (UniqueName: \"kubernetes.io/projected/48585415-a489-4ea0-9d68-981bbebd10cb-kube-api-access-4lgnd\") pod \"collect-profiles-29426820-pbkhw\" (UID: \"48585415-a489-4ea0-9d68-981bbebd10cb\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29426820-pbkhw" Dec 13 07:00:00 crc kubenswrapper[4971]: I1213 07:00:00.495408 4971 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29426820-pbkhw" Dec 13 07:00:00 crc kubenswrapper[4971]: I1213 07:00:00.691970 4971 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29426820-pbkhw"] Dec 13 07:00:01 crc kubenswrapper[4971]: I1213 07:00:01.365159 4971 generic.go:334] "Generic (PLEG): container finished" podID="48585415-a489-4ea0-9d68-981bbebd10cb" containerID="2c7b6bd72e36013d046f5737a2251f46017870ee010e28c0aa3441cf21916bf0" exitCode=0 Dec 13 07:00:01 crc kubenswrapper[4971]: I1213 07:00:01.365210 4971 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29426820-pbkhw" event={"ID":"48585415-a489-4ea0-9d68-981bbebd10cb","Type":"ContainerDied","Data":"2c7b6bd72e36013d046f5737a2251f46017870ee010e28c0aa3441cf21916bf0"} Dec 13 07:00:01 crc kubenswrapper[4971]: I1213 07:00:01.365458 4971 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29426820-pbkhw" event={"ID":"48585415-a489-4ea0-9d68-981bbebd10cb","Type":"ContainerStarted","Data":"90b5951e48a7f65d2022bf0507d04efe01e71e8abf242315097f332ab330d641"} Dec 13 07:00:02 crc kubenswrapper[4971]: I1213 07:00:02.552491 4971 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29426820-pbkhw" Dec 13 07:00:02 crc kubenswrapper[4971]: I1213 07:00:02.633035 4971 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/48585415-a489-4ea0-9d68-981bbebd10cb-secret-volume\") pod \"48585415-a489-4ea0-9d68-981bbebd10cb\" (UID: \"48585415-a489-4ea0-9d68-981bbebd10cb\") " Dec 13 07:00:02 crc kubenswrapper[4971]: I1213 07:00:02.633133 4971 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-4lgnd\" (UniqueName: \"kubernetes.io/projected/48585415-a489-4ea0-9d68-981bbebd10cb-kube-api-access-4lgnd\") pod \"48585415-a489-4ea0-9d68-981bbebd10cb\" (UID: \"48585415-a489-4ea0-9d68-981bbebd10cb\") " Dec 13 07:00:02 crc kubenswrapper[4971]: I1213 07:00:02.633213 4971 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/48585415-a489-4ea0-9d68-981bbebd10cb-config-volume\") pod \"48585415-a489-4ea0-9d68-981bbebd10cb\" (UID: \"48585415-a489-4ea0-9d68-981bbebd10cb\") " Dec 13 07:00:02 crc kubenswrapper[4971]: I1213 07:00:02.633953 4971 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/48585415-a489-4ea0-9d68-981bbebd10cb-config-volume" (OuterVolumeSpecName: "config-volume") pod "48585415-a489-4ea0-9d68-981bbebd10cb" (UID: "48585415-a489-4ea0-9d68-981bbebd10cb"). InnerVolumeSpecName "config-volume". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 13 07:00:02 crc kubenswrapper[4971]: I1213 07:00:02.638634 4971 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/48585415-a489-4ea0-9d68-981bbebd10cb-kube-api-access-4lgnd" (OuterVolumeSpecName: "kube-api-access-4lgnd") pod "48585415-a489-4ea0-9d68-981bbebd10cb" (UID: "48585415-a489-4ea0-9d68-981bbebd10cb"). InnerVolumeSpecName "kube-api-access-4lgnd". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 13 07:00:02 crc kubenswrapper[4971]: I1213 07:00:02.638651 4971 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/48585415-a489-4ea0-9d68-981bbebd10cb-secret-volume" (OuterVolumeSpecName: "secret-volume") pod "48585415-a489-4ea0-9d68-981bbebd10cb" (UID: "48585415-a489-4ea0-9d68-981bbebd10cb"). InnerVolumeSpecName "secret-volume". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 13 07:00:02 crc kubenswrapper[4971]: I1213 07:00:02.734682 4971 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-4lgnd\" (UniqueName: \"kubernetes.io/projected/48585415-a489-4ea0-9d68-981bbebd10cb-kube-api-access-4lgnd\") on node \"crc\" DevicePath \"\"" Dec 13 07:00:02 crc kubenswrapper[4971]: I1213 07:00:02.734722 4971 reconciler_common.go:293] "Volume detached for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/48585415-a489-4ea0-9d68-981bbebd10cb-config-volume\") on node \"crc\" DevicePath \"\"" Dec 13 07:00:02 crc kubenswrapper[4971]: I1213 07:00:02.734734 4971 reconciler_common.go:293] "Volume detached for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/48585415-a489-4ea0-9d68-981bbebd10cb-secret-volume\") on node \"crc\" DevicePath \"\"" Dec 13 07:00:03 crc kubenswrapper[4971]: I1213 07:00:03.376410 4971 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29426820-pbkhw" event={"ID":"48585415-a489-4ea0-9d68-981bbebd10cb","Type":"ContainerDied","Data":"90b5951e48a7f65d2022bf0507d04efe01e71e8abf242315097f332ab330d641"} Dec 13 07:00:03 crc kubenswrapper[4971]: I1213 07:00:03.376455 4971 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="90b5951e48a7f65d2022bf0507d04efe01e71e8abf242315097f332ab330d641" Dec 13 07:00:03 crc kubenswrapper[4971]: I1213 07:00:03.376485 4971 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29426820-pbkhw" Dec 13 07:00:03 crc kubenswrapper[4971]: I1213 07:00:03.898184 4971 scope.go:117] "RemoveContainer" containerID="a60d4a360ca22d5ab008969e58dd981a7858882ea44246f6b56a407dba706501" Dec 13 07:01:16 crc kubenswrapper[4971]: I1213 07:01:16.153606 4971 patch_prober.go:28] interesting pod/machine-config-daemon-82xjz container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 13 07:01:16 crc kubenswrapper[4971]: I1213 07:01:16.154172 4971 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-82xjz" podUID="e0c4af12-edda-41d7-9b44-f87396174ef2" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 13 07:01:46 crc kubenswrapper[4971]: I1213 07:01:46.154224 4971 patch_prober.go:28] interesting pod/machine-config-daemon-82xjz container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 13 07:01:46 crc kubenswrapper[4971]: I1213 07:01:46.154721 4971 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-82xjz" podUID="e0c4af12-edda-41d7-9b44-f87396174ef2" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 13 07:01:47 crc kubenswrapper[4971]: I1213 07:01:47.831786 4971 dynamic_cafile_content.go:123] "Loaded a new CA Bundle and Verifier" name="client-ca-bundle::/etc/kubernetes/kubelet-ca.crt" Dec 13 07:02:16 crc kubenswrapper[4971]: I1213 07:02:16.153175 4971 patch_prober.go:28] interesting pod/machine-config-daemon-82xjz container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 13 07:02:16 crc kubenswrapper[4971]: I1213 07:02:16.153665 4971 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-82xjz" podUID="e0c4af12-edda-41d7-9b44-f87396174ef2" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 13 07:02:16 crc kubenswrapper[4971]: I1213 07:02:16.153708 4971 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-82xjz" Dec 13 07:02:16 crc kubenswrapper[4971]: I1213 07:02:16.154260 4971 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"358e8ac17d5a2f5ef3b4deef1d163694a45d4c70827405c28f4c81a7c84b880e"} pod="openshift-machine-config-operator/machine-config-daemon-82xjz" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Dec 13 07:02:16 crc kubenswrapper[4971]: I1213 07:02:16.154316 4971 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-82xjz" podUID="e0c4af12-edda-41d7-9b44-f87396174ef2" containerName="machine-config-daemon" containerID="cri-o://358e8ac17d5a2f5ef3b4deef1d163694a45d4c70827405c28f4c81a7c84b880e" gracePeriod=600 Dec 13 07:02:17 crc kubenswrapper[4971]: I1213 07:02:17.094979 4971 generic.go:334] "Generic (PLEG): container finished" podID="e0c4af12-edda-41d7-9b44-f87396174ef2" containerID="358e8ac17d5a2f5ef3b4deef1d163694a45d4c70827405c28f4c81a7c84b880e" exitCode=0 Dec 13 07:02:17 crc kubenswrapper[4971]: I1213 07:02:17.095054 4971 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-82xjz" event={"ID":"e0c4af12-edda-41d7-9b44-f87396174ef2","Type":"ContainerDied","Data":"358e8ac17d5a2f5ef3b4deef1d163694a45d4c70827405c28f4c81a7c84b880e"} Dec 13 07:02:17 crc kubenswrapper[4971]: I1213 07:02:17.095324 4971 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-82xjz" event={"ID":"e0c4af12-edda-41d7-9b44-f87396174ef2","Type":"ContainerStarted","Data":"c41da72089cbd159f1c07bf9d29b0bd93e1593b7236249fdd03c408a03b5a8ca"} Dec 13 07:02:17 crc kubenswrapper[4971]: I1213 07:02:17.095342 4971 scope.go:117] "RemoveContainer" containerID="85537ceec5f79277ceb7773d56141774b3a454a245f9047cd2720b76888daff3" Dec 13 07:03:02 crc kubenswrapper[4971]: I1213 07:03:02.019782 4971 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["cert-manager/cert-manager-5b446d88c5-bmjtf"] Dec 13 07:03:02 crc kubenswrapper[4971]: E1213 07:03:02.020616 4971 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="48585415-a489-4ea0-9d68-981bbebd10cb" containerName="collect-profiles" Dec 13 07:03:02 crc kubenswrapper[4971]: I1213 07:03:02.020633 4971 state_mem.go:107] "Deleted CPUSet assignment" podUID="48585415-a489-4ea0-9d68-981bbebd10cb" containerName="collect-profiles" Dec 13 07:03:02 crc kubenswrapper[4971]: I1213 07:03:02.020763 4971 memory_manager.go:354] "RemoveStaleState removing state" podUID="48585415-a489-4ea0-9d68-981bbebd10cb" containerName="collect-profiles" Dec 13 07:03:02 crc kubenswrapper[4971]: I1213 07:03:02.021289 4971 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="cert-manager/cert-manager-5b446d88c5-bmjtf" Dec 13 07:03:02 crc kubenswrapper[4971]: I1213 07:03:02.023846 4971 reflector.go:368] Caches populated for *v1.ConfigMap from object-"cert-manager"/"kube-root-ca.crt" Dec 13 07:03:02 crc kubenswrapper[4971]: I1213 07:03:02.024848 4971 reflector.go:368] Caches populated for *v1.ConfigMap from object-"cert-manager"/"openshift-service-ca.crt" Dec 13 07:03:02 crc kubenswrapper[4971]: I1213 07:03:02.026297 4971 reflector.go:368] Caches populated for *v1.Secret from object-"cert-manager"/"cert-manager-dockercfg-g2fb7" Dec 13 07:03:02 crc kubenswrapper[4971]: I1213 07:03:02.027612 4971 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["cert-manager/cert-manager-cainjector-7f985d654d-dp572"] Dec 13 07:03:02 crc kubenswrapper[4971]: I1213 07:03:02.028341 4971 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="cert-manager/cert-manager-cainjector-7f985d654d-dp572" Dec 13 07:03:02 crc kubenswrapper[4971]: I1213 07:03:02.029550 4971 reflector.go:368] Caches populated for *v1.Secret from object-"cert-manager"/"cert-manager-cainjector-dockercfg-vvfbn" Dec 13 07:03:02 crc kubenswrapper[4971]: I1213 07:03:02.033713 4971 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["cert-manager/cert-manager-cainjector-7f985d654d-dp572"] Dec 13 07:03:02 crc kubenswrapper[4971]: I1213 07:03:02.039914 4971 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["cert-manager/cert-manager-5b446d88c5-bmjtf"] Dec 13 07:03:02 crc kubenswrapper[4971]: I1213 07:03:02.061414 4971 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["cert-manager/cert-manager-webhook-5655c58dd6-qbbjr"] Dec 13 07:03:02 crc kubenswrapper[4971]: I1213 07:03:02.062263 4971 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="cert-manager/cert-manager-webhook-5655c58dd6-qbbjr" Dec 13 07:03:02 crc kubenswrapper[4971]: I1213 07:03:02.064220 4971 reflector.go:368] Caches populated for *v1.Secret from object-"cert-manager"/"cert-manager-webhook-dockercfg-2pbdj" Dec 13 07:03:02 crc kubenswrapper[4971]: I1213 07:03:02.078495 4971 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["cert-manager/cert-manager-webhook-5655c58dd6-qbbjr"] Dec 13 07:03:02 crc kubenswrapper[4971]: I1213 07:03:02.178621 4971 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-5lg8k\" (UniqueName: \"kubernetes.io/projected/65bdb0e4-18e8-4db6-af21-341a748772e9-kube-api-access-5lg8k\") pod \"cert-manager-5b446d88c5-bmjtf\" (UID: \"65bdb0e4-18e8-4db6-af21-341a748772e9\") " pod="cert-manager/cert-manager-5b446d88c5-bmjtf" Dec 13 07:03:02 crc kubenswrapper[4971]: I1213 07:03:02.178712 4971 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-2d8tq\" (UniqueName: \"kubernetes.io/projected/fb8b0fc5-85d4-4df8-9b43-8a567a888b20-kube-api-access-2d8tq\") pod \"cert-manager-cainjector-7f985d654d-dp572\" (UID: \"fb8b0fc5-85d4-4df8-9b43-8a567a888b20\") " pod="cert-manager/cert-manager-cainjector-7f985d654d-dp572" Dec 13 07:03:02 crc kubenswrapper[4971]: I1213 07:03:02.178750 4971 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-78vjs\" (UniqueName: \"kubernetes.io/projected/974beed7-18ac-4d6b-8192-d844b50b7ad3-kube-api-access-78vjs\") pod \"cert-manager-webhook-5655c58dd6-qbbjr\" (UID: \"974beed7-18ac-4d6b-8192-d844b50b7ad3\") " pod="cert-manager/cert-manager-webhook-5655c58dd6-qbbjr" Dec 13 07:03:02 crc kubenswrapper[4971]: I1213 07:03:02.280012 4971 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-5lg8k\" (UniqueName: \"kubernetes.io/projected/65bdb0e4-18e8-4db6-af21-341a748772e9-kube-api-access-5lg8k\") pod \"cert-manager-5b446d88c5-bmjtf\" (UID: \"65bdb0e4-18e8-4db6-af21-341a748772e9\") " pod="cert-manager/cert-manager-5b446d88c5-bmjtf" Dec 13 07:03:02 crc kubenswrapper[4971]: I1213 07:03:02.280076 4971 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-2d8tq\" (UniqueName: \"kubernetes.io/projected/fb8b0fc5-85d4-4df8-9b43-8a567a888b20-kube-api-access-2d8tq\") pod \"cert-manager-cainjector-7f985d654d-dp572\" (UID: \"fb8b0fc5-85d4-4df8-9b43-8a567a888b20\") " pod="cert-manager/cert-manager-cainjector-7f985d654d-dp572" Dec 13 07:03:02 crc kubenswrapper[4971]: I1213 07:03:02.280112 4971 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-78vjs\" (UniqueName: \"kubernetes.io/projected/974beed7-18ac-4d6b-8192-d844b50b7ad3-kube-api-access-78vjs\") pod \"cert-manager-webhook-5655c58dd6-qbbjr\" (UID: \"974beed7-18ac-4d6b-8192-d844b50b7ad3\") " pod="cert-manager/cert-manager-webhook-5655c58dd6-qbbjr" Dec 13 07:03:02 crc kubenswrapper[4971]: I1213 07:03:02.302731 4971 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-2d8tq\" (UniqueName: \"kubernetes.io/projected/fb8b0fc5-85d4-4df8-9b43-8a567a888b20-kube-api-access-2d8tq\") pod \"cert-manager-cainjector-7f985d654d-dp572\" (UID: \"fb8b0fc5-85d4-4df8-9b43-8a567a888b20\") " pod="cert-manager/cert-manager-cainjector-7f985d654d-dp572" Dec 13 07:03:02 crc kubenswrapper[4971]: I1213 07:03:02.307650 4971 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-5lg8k\" (UniqueName: \"kubernetes.io/projected/65bdb0e4-18e8-4db6-af21-341a748772e9-kube-api-access-5lg8k\") pod \"cert-manager-5b446d88c5-bmjtf\" (UID: \"65bdb0e4-18e8-4db6-af21-341a748772e9\") " pod="cert-manager/cert-manager-5b446d88c5-bmjtf" Dec 13 07:03:02 crc kubenswrapper[4971]: I1213 07:03:02.310842 4971 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-78vjs\" (UniqueName: \"kubernetes.io/projected/974beed7-18ac-4d6b-8192-d844b50b7ad3-kube-api-access-78vjs\") pod \"cert-manager-webhook-5655c58dd6-qbbjr\" (UID: \"974beed7-18ac-4d6b-8192-d844b50b7ad3\") " pod="cert-manager/cert-manager-webhook-5655c58dd6-qbbjr" Dec 13 07:03:02 crc kubenswrapper[4971]: I1213 07:03:02.344091 4971 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="cert-manager/cert-manager-5b446d88c5-bmjtf" Dec 13 07:03:02 crc kubenswrapper[4971]: I1213 07:03:02.358201 4971 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="cert-manager/cert-manager-cainjector-7f985d654d-dp572" Dec 13 07:03:02 crc kubenswrapper[4971]: I1213 07:03:02.379088 4971 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="cert-manager/cert-manager-webhook-5655c58dd6-qbbjr" Dec 13 07:03:02 crc kubenswrapper[4971]: I1213 07:03:02.592546 4971 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["cert-manager/cert-manager-cainjector-7f985d654d-dp572"] Dec 13 07:03:02 crc kubenswrapper[4971]: W1213 07:03:02.596324 4971 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podfb8b0fc5_85d4_4df8_9b43_8a567a888b20.slice/crio-2c3997a300339794da9807d5b03508a7a16df42c73a36060621512aa78cf45bd WatchSource:0}: Error finding container 2c3997a300339794da9807d5b03508a7a16df42c73a36060621512aa78cf45bd: Status 404 returned error can't find the container with id 2c3997a300339794da9807d5b03508a7a16df42c73a36060621512aa78cf45bd Dec 13 07:03:02 crc kubenswrapper[4971]: I1213 07:03:02.599501 4971 provider.go:102] Refreshing cache for provider: *credentialprovider.defaultDockerConfigProvider Dec 13 07:03:02 crc kubenswrapper[4971]: I1213 07:03:02.735678 4971 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["cert-manager/cert-manager-5b446d88c5-bmjtf"] Dec 13 07:03:02 crc kubenswrapper[4971]: I1213 07:03:02.739945 4971 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="cert-manager/cert-manager-cainjector-7f985d654d-dp572" event={"ID":"fb8b0fc5-85d4-4df8-9b43-8a567a888b20","Type":"ContainerStarted","Data":"2c3997a300339794da9807d5b03508a7a16df42c73a36060621512aa78cf45bd"} Dec 13 07:03:02 crc kubenswrapper[4971]: I1213 07:03:02.825256 4971 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["cert-manager/cert-manager-webhook-5655c58dd6-qbbjr"] Dec 13 07:03:02 crc kubenswrapper[4971]: W1213 07:03:02.827032 4971 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod974beed7_18ac_4d6b_8192_d844b50b7ad3.slice/crio-05a5dc9b5b92add1c13f0cfa5e04f6a66604a401c7794c122e4bc934e5b46eef WatchSource:0}: Error finding container 05a5dc9b5b92add1c13f0cfa5e04f6a66604a401c7794c122e4bc934e5b46eef: Status 404 returned error can't find the container with id 05a5dc9b5b92add1c13f0cfa5e04f6a66604a401c7794c122e4bc934e5b46eef Dec 13 07:03:03 crc kubenswrapper[4971]: I1213 07:03:03.747368 4971 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="cert-manager/cert-manager-5b446d88c5-bmjtf" event={"ID":"65bdb0e4-18e8-4db6-af21-341a748772e9","Type":"ContainerStarted","Data":"220ec096f8523bc3e998d53ada38d8243fdb7f894dfcb252d75d522e4f8eaec7"} Dec 13 07:03:03 crc kubenswrapper[4971]: I1213 07:03:03.749105 4971 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="cert-manager/cert-manager-webhook-5655c58dd6-qbbjr" event={"ID":"974beed7-18ac-4d6b-8192-d844b50b7ad3","Type":"ContainerStarted","Data":"05a5dc9b5b92add1c13f0cfa5e04f6a66604a401c7794c122e4bc934e5b46eef"} Dec 13 07:03:08 crc kubenswrapper[4971]: I1213 07:03:08.783872 4971 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="cert-manager/cert-manager-webhook-5655c58dd6-qbbjr" event={"ID":"974beed7-18ac-4d6b-8192-d844b50b7ad3","Type":"ContainerStarted","Data":"8d7440ccb7a9d21963cbe8dfdddcdc97126051b51618995b60f517e172cf8abb"} Dec 13 07:03:08 crc kubenswrapper[4971]: I1213 07:03:08.786099 4971 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="cert-manager/cert-manager-cainjector-7f985d654d-dp572" event={"ID":"fb8b0fc5-85d4-4df8-9b43-8a567a888b20","Type":"ContainerStarted","Data":"fed1852f76f50813d62413efc7f0731eccb3e7d2090fdec29a636099f9cd4131"} Dec 13 07:03:08 crc kubenswrapper[4971]: I1213 07:03:08.790602 4971 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="cert-manager/cert-manager-5b446d88c5-bmjtf" event={"ID":"65bdb0e4-18e8-4db6-af21-341a748772e9","Type":"ContainerStarted","Data":"f331bf5cc1e9d1846edbe9a824f88e95a21e3fe0439e893ddc3e1b5af273adc6"} Dec 13 07:03:09 crc kubenswrapper[4971]: I1213 07:03:09.816759 4971 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="cert-manager/cert-manager-5b446d88c5-bmjtf" podStartSLOduration=1.9773531800000002 podStartE2EDuration="7.81673276s" podCreationTimestamp="2025-12-13 07:03:02 +0000 UTC" firstStartedPulling="2025-12-13 07:03:02.743385869 +0000 UTC m=+839.347795317" lastFinishedPulling="2025-12-13 07:03:08.582765449 +0000 UTC m=+845.187174897" observedRunningTime="2025-12-13 07:03:08.807500842 +0000 UTC m=+845.411910290" watchObservedRunningTime="2025-12-13 07:03:09.81673276 +0000 UTC m=+846.421142248" Dec 13 07:03:09 crc kubenswrapper[4971]: I1213 07:03:09.823180 4971 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="cert-manager/cert-manager-cainjector-7f985d654d-dp572" podStartSLOduration=1.783644378 podStartE2EDuration="7.82315616s" podCreationTimestamp="2025-12-13 07:03:02 +0000 UTC" firstStartedPulling="2025-12-13 07:03:02.599266583 +0000 UTC m=+839.203676031" lastFinishedPulling="2025-12-13 07:03:08.638778365 +0000 UTC m=+845.243187813" observedRunningTime="2025-12-13 07:03:09.810937587 +0000 UTC m=+846.415347045" watchObservedRunningTime="2025-12-13 07:03:09.82315616 +0000 UTC m=+846.427565608" Dec 13 07:03:09 crc kubenswrapper[4971]: I1213 07:03:09.823476 4971 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="cert-manager/cert-manager-webhook-5655c58dd6-qbbjr" podStartSLOduration=2.069604204 podStartE2EDuration="7.823469137s" podCreationTimestamp="2025-12-13 07:03:02 +0000 UTC" firstStartedPulling="2025-12-13 07:03:02.828942527 +0000 UTC m=+839.433351975" lastFinishedPulling="2025-12-13 07:03:08.58280746 +0000 UTC m=+845.187216908" observedRunningTime="2025-12-13 07:03:09.822248967 +0000 UTC m=+846.426658415" watchObservedRunningTime="2025-12-13 07:03:09.823469137 +0000 UTC m=+846.427878605" Dec 13 07:03:12 crc kubenswrapper[4971]: I1213 07:03:12.149953 4971 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-ovn-kubernetes/ovnkube-node-zlcg5"] Dec 13 07:03:12 crc kubenswrapper[4971]: I1213 07:03:12.150651 4971 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-ovn-kubernetes/ovnkube-node-zlcg5" podUID="736c6b1a-93dd-4f71-81b6-1c43fcf5f556" containerName="ovn-controller" containerID="cri-o://f137f4b441c2402093ab430a9689a46221c4fb4bc70d99fb992f62b0a989c465" gracePeriod=30 Dec 13 07:03:12 crc kubenswrapper[4971]: I1213 07:03:12.150698 4971 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-ovn-kubernetes/ovnkube-node-zlcg5" podUID="736c6b1a-93dd-4f71-81b6-1c43fcf5f556" containerName="nbdb" containerID="cri-o://f6073140d9f86929b6d4bd2960ede0617a5c9620b269bbab043bae37e0d389f5" gracePeriod=30 Dec 13 07:03:12 crc kubenswrapper[4971]: I1213 07:03:12.150764 4971 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-ovn-kubernetes/ovnkube-node-zlcg5" podUID="736c6b1a-93dd-4f71-81b6-1c43fcf5f556" containerName="northd" containerID="cri-o://693a2ab434c250cb140a866e8068b156df9191414fed1d63bee919a2528ea6e5" gracePeriod=30 Dec 13 07:03:12 crc kubenswrapper[4971]: I1213 07:03:12.150808 4971 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-ovn-kubernetes/ovnkube-node-zlcg5" podUID="736c6b1a-93dd-4f71-81b6-1c43fcf5f556" containerName="kube-rbac-proxy-ovn-metrics" containerID="cri-o://f22a4ec756119c4f791fb5105cf1f31df3e548b2b07b1f909a62802e9b6035ae" gracePeriod=30 Dec 13 07:03:12 crc kubenswrapper[4971]: I1213 07:03:12.150842 4971 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-ovn-kubernetes/ovnkube-node-zlcg5" podUID="736c6b1a-93dd-4f71-81b6-1c43fcf5f556" containerName="kube-rbac-proxy-node" containerID="cri-o://325223543b3c7db417f769533559a1ca2b23a723494e1885b0fd0480b5a6399f" gracePeriod=30 Dec 13 07:03:12 crc kubenswrapper[4971]: I1213 07:03:12.150844 4971 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-ovn-kubernetes/ovnkube-node-zlcg5" podUID="736c6b1a-93dd-4f71-81b6-1c43fcf5f556" containerName="sbdb" containerID="cri-o://868049c6a3a4ebe4036d1d22e02ba6b76bced2ee7e269bb228accc4a0636678a" gracePeriod=30 Dec 13 07:03:12 crc kubenswrapper[4971]: I1213 07:03:12.150903 4971 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-ovn-kubernetes/ovnkube-node-zlcg5" podUID="736c6b1a-93dd-4f71-81b6-1c43fcf5f556" containerName="ovn-acl-logging" containerID="cri-o://219620f51474e60b3e1f880c0972eb612dffc4bad24462bbba2c093ae817d852" gracePeriod=30 Dec 13 07:03:12 crc kubenswrapper[4971]: I1213 07:03:12.191166 4971 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-ovn-kubernetes/ovnkube-node-zlcg5" podUID="736c6b1a-93dd-4f71-81b6-1c43fcf5f556" containerName="ovnkube-controller" containerID="cri-o://6afaff46af01dcc99e08b76967fe4181e615cd044389b9b8cea1581505eef247" gracePeriod=30 Dec 13 07:03:12 crc kubenswrapper[4971]: I1213 07:03:12.380268 4971 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="cert-manager/cert-manager-webhook-5655c58dd6-qbbjr" Dec 13 07:03:12 crc kubenswrapper[4971]: I1213 07:03:12.449158 4971 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-zlcg5_736c6b1a-93dd-4f71-81b6-1c43fcf5f556/ovnkube-controller/3.log" Dec 13 07:03:12 crc kubenswrapper[4971]: I1213 07:03:12.451324 4971 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-zlcg5_736c6b1a-93dd-4f71-81b6-1c43fcf5f556/ovn-acl-logging/0.log" Dec 13 07:03:12 crc kubenswrapper[4971]: I1213 07:03:12.451738 4971 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-zlcg5_736c6b1a-93dd-4f71-81b6-1c43fcf5f556/ovn-controller/0.log" Dec 13 07:03:12 crc kubenswrapper[4971]: I1213 07:03:12.452220 4971 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-ovn-kubernetes/ovnkube-node-zlcg5" Dec 13 07:03:12 crc kubenswrapper[4971]: I1213 07:03:12.504864 4971 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-ovn-kubernetes/ovnkube-node-fq8jh"] Dec 13 07:03:12 crc kubenswrapper[4971]: E1213 07:03:12.505107 4971 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="736c6b1a-93dd-4f71-81b6-1c43fcf5f556" containerName="sbdb" Dec 13 07:03:12 crc kubenswrapper[4971]: I1213 07:03:12.505120 4971 state_mem.go:107] "Deleted CPUSet assignment" podUID="736c6b1a-93dd-4f71-81b6-1c43fcf5f556" containerName="sbdb" Dec 13 07:03:12 crc kubenswrapper[4971]: E1213 07:03:12.505133 4971 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="736c6b1a-93dd-4f71-81b6-1c43fcf5f556" containerName="ovnkube-controller" Dec 13 07:03:12 crc kubenswrapper[4971]: I1213 07:03:12.505139 4971 state_mem.go:107] "Deleted CPUSet assignment" podUID="736c6b1a-93dd-4f71-81b6-1c43fcf5f556" containerName="ovnkube-controller" Dec 13 07:03:12 crc kubenswrapper[4971]: E1213 07:03:12.505147 4971 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="736c6b1a-93dd-4f71-81b6-1c43fcf5f556" containerName="ovnkube-controller" Dec 13 07:03:12 crc kubenswrapper[4971]: I1213 07:03:12.505153 4971 state_mem.go:107] "Deleted CPUSet assignment" podUID="736c6b1a-93dd-4f71-81b6-1c43fcf5f556" containerName="ovnkube-controller" Dec 13 07:03:12 crc kubenswrapper[4971]: E1213 07:03:12.505161 4971 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="736c6b1a-93dd-4f71-81b6-1c43fcf5f556" containerName="kubecfg-setup" Dec 13 07:03:12 crc kubenswrapper[4971]: I1213 07:03:12.505167 4971 state_mem.go:107] "Deleted CPUSet assignment" podUID="736c6b1a-93dd-4f71-81b6-1c43fcf5f556" containerName="kubecfg-setup" Dec 13 07:03:12 crc kubenswrapper[4971]: E1213 07:03:12.505174 4971 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="736c6b1a-93dd-4f71-81b6-1c43fcf5f556" containerName="northd" Dec 13 07:03:12 crc kubenswrapper[4971]: I1213 07:03:12.505180 4971 state_mem.go:107] "Deleted CPUSet assignment" podUID="736c6b1a-93dd-4f71-81b6-1c43fcf5f556" containerName="northd" Dec 13 07:03:12 crc kubenswrapper[4971]: E1213 07:03:12.505188 4971 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="736c6b1a-93dd-4f71-81b6-1c43fcf5f556" containerName="kube-rbac-proxy-ovn-metrics" Dec 13 07:03:12 crc kubenswrapper[4971]: I1213 07:03:12.505194 4971 state_mem.go:107] "Deleted CPUSet assignment" podUID="736c6b1a-93dd-4f71-81b6-1c43fcf5f556" containerName="kube-rbac-proxy-ovn-metrics" Dec 13 07:03:12 crc kubenswrapper[4971]: E1213 07:03:12.505202 4971 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="736c6b1a-93dd-4f71-81b6-1c43fcf5f556" containerName="ovnkube-controller" Dec 13 07:03:12 crc kubenswrapper[4971]: I1213 07:03:12.505207 4971 state_mem.go:107] "Deleted CPUSet assignment" podUID="736c6b1a-93dd-4f71-81b6-1c43fcf5f556" containerName="ovnkube-controller" Dec 13 07:03:12 crc kubenswrapper[4971]: E1213 07:03:12.505219 4971 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="736c6b1a-93dd-4f71-81b6-1c43fcf5f556" containerName="kube-rbac-proxy-node" Dec 13 07:03:12 crc kubenswrapper[4971]: I1213 07:03:12.505225 4971 state_mem.go:107] "Deleted CPUSet assignment" podUID="736c6b1a-93dd-4f71-81b6-1c43fcf5f556" containerName="kube-rbac-proxy-node" Dec 13 07:03:12 crc kubenswrapper[4971]: E1213 07:03:12.505231 4971 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="736c6b1a-93dd-4f71-81b6-1c43fcf5f556" containerName="nbdb" Dec 13 07:03:12 crc kubenswrapper[4971]: I1213 07:03:12.505236 4971 state_mem.go:107] "Deleted CPUSet assignment" podUID="736c6b1a-93dd-4f71-81b6-1c43fcf5f556" containerName="nbdb" Dec 13 07:03:12 crc kubenswrapper[4971]: E1213 07:03:12.505244 4971 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="736c6b1a-93dd-4f71-81b6-1c43fcf5f556" containerName="ovn-acl-logging" Dec 13 07:03:12 crc kubenswrapper[4971]: I1213 07:03:12.505249 4971 state_mem.go:107] "Deleted CPUSet assignment" podUID="736c6b1a-93dd-4f71-81b6-1c43fcf5f556" containerName="ovn-acl-logging" Dec 13 07:03:12 crc kubenswrapper[4971]: E1213 07:03:12.505259 4971 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="736c6b1a-93dd-4f71-81b6-1c43fcf5f556" containerName="ovn-controller" Dec 13 07:03:12 crc kubenswrapper[4971]: I1213 07:03:12.505265 4971 state_mem.go:107] "Deleted CPUSet assignment" podUID="736c6b1a-93dd-4f71-81b6-1c43fcf5f556" containerName="ovn-controller" Dec 13 07:03:12 crc kubenswrapper[4971]: I1213 07:03:12.505367 4971 memory_manager.go:354] "RemoveStaleState removing state" podUID="736c6b1a-93dd-4f71-81b6-1c43fcf5f556" containerName="ovnkube-controller" Dec 13 07:03:12 crc kubenswrapper[4971]: I1213 07:03:12.505375 4971 memory_manager.go:354] "RemoveStaleState removing state" podUID="736c6b1a-93dd-4f71-81b6-1c43fcf5f556" containerName="kube-rbac-proxy-ovn-metrics" Dec 13 07:03:12 crc kubenswrapper[4971]: I1213 07:03:12.505383 4971 memory_manager.go:354] "RemoveStaleState removing state" podUID="736c6b1a-93dd-4f71-81b6-1c43fcf5f556" containerName="nbdb" Dec 13 07:03:12 crc kubenswrapper[4971]: I1213 07:03:12.505390 4971 memory_manager.go:354] "RemoveStaleState removing state" podUID="736c6b1a-93dd-4f71-81b6-1c43fcf5f556" containerName="ovn-acl-logging" Dec 13 07:03:12 crc kubenswrapper[4971]: I1213 07:03:12.505400 4971 memory_manager.go:354] "RemoveStaleState removing state" podUID="736c6b1a-93dd-4f71-81b6-1c43fcf5f556" containerName="ovn-controller" Dec 13 07:03:12 crc kubenswrapper[4971]: I1213 07:03:12.505408 4971 memory_manager.go:354] "RemoveStaleState removing state" podUID="736c6b1a-93dd-4f71-81b6-1c43fcf5f556" containerName="ovnkube-controller" Dec 13 07:03:12 crc kubenswrapper[4971]: I1213 07:03:12.505414 4971 memory_manager.go:354] "RemoveStaleState removing state" podUID="736c6b1a-93dd-4f71-81b6-1c43fcf5f556" containerName="ovnkube-controller" Dec 13 07:03:12 crc kubenswrapper[4971]: I1213 07:03:12.505422 4971 memory_manager.go:354] "RemoveStaleState removing state" podUID="736c6b1a-93dd-4f71-81b6-1c43fcf5f556" containerName="kube-rbac-proxy-node" Dec 13 07:03:12 crc kubenswrapper[4971]: I1213 07:03:12.505428 4971 memory_manager.go:354] "RemoveStaleState removing state" podUID="736c6b1a-93dd-4f71-81b6-1c43fcf5f556" containerName="sbdb" Dec 13 07:03:12 crc kubenswrapper[4971]: I1213 07:03:12.505435 4971 memory_manager.go:354] "RemoveStaleState removing state" podUID="736c6b1a-93dd-4f71-81b6-1c43fcf5f556" containerName="ovnkube-controller" Dec 13 07:03:12 crc kubenswrapper[4971]: I1213 07:03:12.505442 4971 memory_manager.go:354] "RemoveStaleState removing state" podUID="736c6b1a-93dd-4f71-81b6-1c43fcf5f556" containerName="northd" Dec 13 07:03:12 crc kubenswrapper[4971]: I1213 07:03:12.505451 4971 memory_manager.go:354] "RemoveStaleState removing state" podUID="736c6b1a-93dd-4f71-81b6-1c43fcf5f556" containerName="ovnkube-controller" Dec 13 07:03:12 crc kubenswrapper[4971]: E1213 07:03:12.505550 4971 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="736c6b1a-93dd-4f71-81b6-1c43fcf5f556" containerName="ovnkube-controller" Dec 13 07:03:12 crc kubenswrapper[4971]: I1213 07:03:12.505558 4971 state_mem.go:107] "Deleted CPUSet assignment" podUID="736c6b1a-93dd-4f71-81b6-1c43fcf5f556" containerName="ovnkube-controller" Dec 13 07:03:12 crc kubenswrapper[4971]: E1213 07:03:12.505715 4971 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="736c6b1a-93dd-4f71-81b6-1c43fcf5f556" containerName="ovnkube-controller" Dec 13 07:03:12 crc kubenswrapper[4971]: I1213 07:03:12.505723 4971 state_mem.go:107] "Deleted CPUSet assignment" podUID="736c6b1a-93dd-4f71-81b6-1c43fcf5f556" containerName="ovnkube-controller" Dec 13 07:03:12 crc kubenswrapper[4971]: I1213 07:03:12.507081 4971 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-ovn-kubernetes/ovnkube-node-fq8jh" Dec 13 07:03:12 crc kubenswrapper[4971]: I1213 07:03:12.602779 4971 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/736c6b1a-93dd-4f71-81b6-1c43fcf5f556-ovnkube-config\") pod \"736c6b1a-93dd-4f71-81b6-1c43fcf5f556\" (UID: \"736c6b1a-93dd-4f71-81b6-1c43fcf5f556\") " Dec 13 07:03:12 crc kubenswrapper[4971]: I1213 07:03:12.603106 4971 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etc-openvswitch\" (UniqueName: \"kubernetes.io/host-path/736c6b1a-93dd-4f71-81b6-1c43fcf5f556-etc-openvswitch\") pod \"736c6b1a-93dd-4f71-81b6-1c43fcf5f556\" (UID: \"736c6b1a-93dd-4f71-81b6-1c43fcf5f556\") " Dec 13 07:03:12 crc kubenswrapper[4971]: I1213 07:03:12.603253 4971 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovnkube-script-lib\" (UniqueName: \"kubernetes.io/configmap/736c6b1a-93dd-4f71-81b6-1c43fcf5f556-ovnkube-script-lib\") pod \"736c6b1a-93dd-4f71-81b6-1c43fcf5f556\" (UID: \"736c6b1a-93dd-4f71-81b6-1c43fcf5f556\") " Dec 13 07:03:12 crc kubenswrapper[4971]: I1213 07:03:12.603394 4971 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"node-log\" (UniqueName: \"kubernetes.io/host-path/736c6b1a-93dd-4f71-81b6-1c43fcf5f556-node-log\") pod \"736c6b1a-93dd-4f71-81b6-1c43fcf5f556\" (UID: \"736c6b1a-93dd-4f71-81b6-1c43fcf5f556\") " Dec 13 07:03:12 crc kubenswrapper[4971]: I1213 07:03:12.603190 4971 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/736c6b1a-93dd-4f71-81b6-1c43fcf5f556-etc-openvswitch" (OuterVolumeSpecName: "etc-openvswitch") pod "736c6b1a-93dd-4f71-81b6-1c43fcf5f556" (UID: "736c6b1a-93dd-4f71-81b6-1c43fcf5f556"). InnerVolumeSpecName "etc-openvswitch". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 13 07:03:12 crc kubenswrapper[4971]: I1213 07:03:12.603323 4971 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/736c6b1a-93dd-4f71-81b6-1c43fcf5f556-ovnkube-config" (OuterVolumeSpecName: "ovnkube-config") pod "736c6b1a-93dd-4f71-81b6-1c43fcf5f556" (UID: "736c6b1a-93dd-4f71-81b6-1c43fcf5f556"). InnerVolumeSpecName "ovnkube-config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 13 07:03:12 crc kubenswrapper[4971]: I1213 07:03:12.603485 4971 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/736c6b1a-93dd-4f71-81b6-1c43fcf5f556-node-log" (OuterVolumeSpecName: "node-log") pod "736c6b1a-93dd-4f71-81b6-1c43fcf5f556" (UID: "736c6b1a-93dd-4f71-81b6-1c43fcf5f556"). InnerVolumeSpecName "node-log". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 13 07:03:12 crc kubenswrapper[4971]: I1213 07:03:12.603832 4971 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"host-kubelet\" (UniqueName: \"kubernetes.io/host-path/736c6b1a-93dd-4f71-81b6-1c43fcf5f556-host-kubelet\") pod \"736c6b1a-93dd-4f71-81b6-1c43fcf5f556\" (UID: \"736c6b1a-93dd-4f71-81b6-1c43fcf5f556\") " Dec 13 07:03:12 crc kubenswrapper[4971]: I1213 07:03:12.603982 4971 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovn-node-metrics-cert\" (UniqueName: \"kubernetes.io/secret/736c6b1a-93dd-4f71-81b6-1c43fcf5f556-ovn-node-metrics-cert\") pod \"736c6b1a-93dd-4f71-81b6-1c43fcf5f556\" (UID: \"736c6b1a-93dd-4f71-81b6-1c43fcf5f556\") " Dec 13 07:03:12 crc kubenswrapper[4971]: I1213 07:03:12.604157 4971 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"host-run-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/736c6b1a-93dd-4f71-81b6-1c43fcf5f556-host-run-ovn-kubernetes\") pod \"736c6b1a-93dd-4f71-81b6-1c43fcf5f556\" (UID: \"736c6b1a-93dd-4f71-81b6-1c43fcf5f556\") " Dec 13 07:03:12 crc kubenswrapper[4971]: I1213 07:03:12.604548 4971 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"run-openvswitch\" (UniqueName: \"kubernetes.io/host-path/736c6b1a-93dd-4f71-81b6-1c43fcf5f556-run-openvswitch\") pod \"736c6b1a-93dd-4f71-81b6-1c43fcf5f556\" (UID: \"736c6b1a-93dd-4f71-81b6-1c43fcf5f556\") " Dec 13 07:03:12 crc kubenswrapper[4971]: I1213 07:03:12.604711 4971 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"host-var-lib-cni-networks-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/736c6b1a-93dd-4f71-81b6-1c43fcf5f556-host-var-lib-cni-networks-ovn-kubernetes\") pod \"736c6b1a-93dd-4f71-81b6-1c43fcf5f556\" (UID: \"736c6b1a-93dd-4f71-81b6-1c43fcf5f556\") " Dec 13 07:03:12 crc kubenswrapper[4971]: I1213 07:03:12.604846 4971 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"host-run-netns\" (UniqueName: \"kubernetes.io/host-path/736c6b1a-93dd-4f71-81b6-1c43fcf5f556-host-run-netns\") pod \"736c6b1a-93dd-4f71-81b6-1c43fcf5f556\" (UID: \"736c6b1a-93dd-4f71-81b6-1c43fcf5f556\") " Dec 13 07:03:12 crc kubenswrapper[4971]: I1213 07:03:12.604963 4971 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"systemd-units\" (UniqueName: \"kubernetes.io/host-path/736c6b1a-93dd-4f71-81b6-1c43fcf5f556-systemd-units\") pod \"736c6b1a-93dd-4f71-81b6-1c43fcf5f556\" (UID: \"736c6b1a-93dd-4f71-81b6-1c43fcf5f556\") " Dec 13 07:03:12 crc kubenswrapper[4971]: I1213 07:03:12.605100 4971 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/736c6b1a-93dd-4f71-81b6-1c43fcf5f556-env-overrides\") pod \"736c6b1a-93dd-4f71-81b6-1c43fcf5f556\" (UID: \"736c6b1a-93dd-4f71-81b6-1c43fcf5f556\") " Dec 13 07:03:12 crc kubenswrapper[4971]: I1213 07:03:12.605240 4971 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"log-socket\" (UniqueName: \"kubernetes.io/host-path/736c6b1a-93dd-4f71-81b6-1c43fcf5f556-log-socket\") pod \"736c6b1a-93dd-4f71-81b6-1c43fcf5f556\" (UID: \"736c6b1a-93dd-4f71-81b6-1c43fcf5f556\") " Dec 13 07:03:12 crc kubenswrapper[4971]: I1213 07:03:12.605367 4971 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"host-cni-netd\" (UniqueName: \"kubernetes.io/host-path/736c6b1a-93dd-4f71-81b6-1c43fcf5f556-host-cni-netd\") pod \"736c6b1a-93dd-4f71-81b6-1c43fcf5f556\" (UID: \"736c6b1a-93dd-4f71-81b6-1c43fcf5f556\") " Dec 13 07:03:12 crc kubenswrapper[4971]: I1213 07:03:12.605500 4971 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-vzmsr\" (UniqueName: \"kubernetes.io/projected/736c6b1a-93dd-4f71-81b6-1c43fcf5f556-kube-api-access-vzmsr\") pod \"736c6b1a-93dd-4f71-81b6-1c43fcf5f556\" (UID: \"736c6b1a-93dd-4f71-81b6-1c43fcf5f556\") " Dec 13 07:03:12 crc kubenswrapper[4971]: I1213 07:03:12.605631 4971 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"run-systemd\" (UniqueName: \"kubernetes.io/host-path/736c6b1a-93dd-4f71-81b6-1c43fcf5f556-run-systemd\") pod \"736c6b1a-93dd-4f71-81b6-1c43fcf5f556\" (UID: \"736c6b1a-93dd-4f71-81b6-1c43fcf5f556\") " Dec 13 07:03:12 crc kubenswrapper[4971]: I1213 07:03:12.605752 4971 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"run-ovn\" (UniqueName: \"kubernetes.io/host-path/736c6b1a-93dd-4f71-81b6-1c43fcf5f556-run-ovn\") pod \"736c6b1a-93dd-4f71-81b6-1c43fcf5f556\" (UID: \"736c6b1a-93dd-4f71-81b6-1c43fcf5f556\") " Dec 13 07:03:12 crc kubenswrapper[4971]: I1213 07:03:12.605867 4971 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"var-lib-openvswitch\" (UniqueName: \"kubernetes.io/host-path/736c6b1a-93dd-4f71-81b6-1c43fcf5f556-var-lib-openvswitch\") pod \"736c6b1a-93dd-4f71-81b6-1c43fcf5f556\" (UID: \"736c6b1a-93dd-4f71-81b6-1c43fcf5f556\") " Dec 13 07:03:12 crc kubenswrapper[4971]: I1213 07:03:12.605977 4971 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"host-cni-bin\" (UniqueName: \"kubernetes.io/host-path/736c6b1a-93dd-4f71-81b6-1c43fcf5f556-host-cni-bin\") pod \"736c6b1a-93dd-4f71-81b6-1c43fcf5f556\" (UID: \"736c6b1a-93dd-4f71-81b6-1c43fcf5f556\") " Dec 13 07:03:12 crc kubenswrapper[4971]: I1213 07:03:12.603868 4971 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/736c6b1a-93dd-4f71-81b6-1c43fcf5f556-ovnkube-script-lib" (OuterVolumeSpecName: "ovnkube-script-lib") pod "736c6b1a-93dd-4f71-81b6-1c43fcf5f556" (UID: "736c6b1a-93dd-4f71-81b6-1c43fcf5f556"). InnerVolumeSpecName "ovnkube-script-lib". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 13 07:03:12 crc kubenswrapper[4971]: I1213 07:03:12.603907 4971 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/736c6b1a-93dd-4f71-81b6-1c43fcf5f556-host-kubelet" (OuterVolumeSpecName: "host-kubelet") pod "736c6b1a-93dd-4f71-81b6-1c43fcf5f556" (UID: "736c6b1a-93dd-4f71-81b6-1c43fcf5f556"). InnerVolumeSpecName "host-kubelet". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 13 07:03:12 crc kubenswrapper[4971]: I1213 07:03:12.606073 4971 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/736c6b1a-93dd-4f71-81b6-1c43fcf5f556-host-cni-bin" (OuterVolumeSpecName: "host-cni-bin") pod "736c6b1a-93dd-4f71-81b6-1c43fcf5f556" (UID: "736c6b1a-93dd-4f71-81b6-1c43fcf5f556"). InnerVolumeSpecName "host-cni-bin". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 13 07:03:12 crc kubenswrapper[4971]: I1213 07:03:12.604207 4971 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/736c6b1a-93dd-4f71-81b6-1c43fcf5f556-host-run-ovn-kubernetes" (OuterVolumeSpecName: "host-run-ovn-kubernetes") pod "736c6b1a-93dd-4f71-81b6-1c43fcf5f556" (UID: "736c6b1a-93dd-4f71-81b6-1c43fcf5f556"). InnerVolumeSpecName "host-run-ovn-kubernetes". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 13 07:03:12 crc kubenswrapper[4971]: I1213 07:03:12.604598 4971 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/736c6b1a-93dd-4f71-81b6-1c43fcf5f556-run-openvswitch" (OuterVolumeSpecName: "run-openvswitch") pod "736c6b1a-93dd-4f71-81b6-1c43fcf5f556" (UID: "736c6b1a-93dd-4f71-81b6-1c43fcf5f556"). InnerVolumeSpecName "run-openvswitch". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 13 07:03:12 crc kubenswrapper[4971]: I1213 07:03:12.604808 4971 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/736c6b1a-93dd-4f71-81b6-1c43fcf5f556-host-var-lib-cni-networks-ovn-kubernetes" (OuterVolumeSpecName: "host-var-lib-cni-networks-ovn-kubernetes") pod "736c6b1a-93dd-4f71-81b6-1c43fcf5f556" (UID: "736c6b1a-93dd-4f71-81b6-1c43fcf5f556"). InnerVolumeSpecName "host-var-lib-cni-networks-ovn-kubernetes". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 13 07:03:12 crc kubenswrapper[4971]: I1213 07:03:12.604920 4971 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/736c6b1a-93dd-4f71-81b6-1c43fcf5f556-host-run-netns" (OuterVolumeSpecName: "host-run-netns") pod "736c6b1a-93dd-4f71-81b6-1c43fcf5f556" (UID: "736c6b1a-93dd-4f71-81b6-1c43fcf5f556"). InnerVolumeSpecName "host-run-netns". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 13 07:03:12 crc kubenswrapper[4971]: I1213 07:03:12.605033 4971 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/736c6b1a-93dd-4f71-81b6-1c43fcf5f556-systemd-units" (OuterVolumeSpecName: "systemd-units") pod "736c6b1a-93dd-4f71-81b6-1c43fcf5f556" (UID: "736c6b1a-93dd-4f71-81b6-1c43fcf5f556"). InnerVolumeSpecName "systemd-units". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 13 07:03:12 crc kubenswrapper[4971]: I1213 07:03:12.605332 4971 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/736c6b1a-93dd-4f71-81b6-1c43fcf5f556-log-socket" (OuterVolumeSpecName: "log-socket") pod "736c6b1a-93dd-4f71-81b6-1c43fcf5f556" (UID: "736c6b1a-93dd-4f71-81b6-1c43fcf5f556"). InnerVolumeSpecName "log-socket". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 13 07:03:12 crc kubenswrapper[4971]: I1213 07:03:12.605443 4971 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/736c6b1a-93dd-4f71-81b6-1c43fcf5f556-env-overrides" (OuterVolumeSpecName: "env-overrides") pod "736c6b1a-93dd-4f71-81b6-1c43fcf5f556" (UID: "736c6b1a-93dd-4f71-81b6-1c43fcf5f556"). InnerVolumeSpecName "env-overrides". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 13 07:03:12 crc kubenswrapper[4971]: I1213 07:03:12.605452 4971 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/736c6b1a-93dd-4f71-81b6-1c43fcf5f556-host-cni-netd" (OuterVolumeSpecName: "host-cni-netd") pod "736c6b1a-93dd-4f71-81b6-1c43fcf5f556" (UID: "736c6b1a-93dd-4f71-81b6-1c43fcf5f556"). InnerVolumeSpecName "host-cni-netd". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 13 07:03:12 crc kubenswrapper[4971]: I1213 07:03:12.605841 4971 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/736c6b1a-93dd-4f71-81b6-1c43fcf5f556-run-ovn" (OuterVolumeSpecName: "run-ovn") pod "736c6b1a-93dd-4f71-81b6-1c43fcf5f556" (UID: "736c6b1a-93dd-4f71-81b6-1c43fcf5f556"). InnerVolumeSpecName "run-ovn". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 13 07:03:12 crc kubenswrapper[4971]: I1213 07:03:12.605911 4971 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/736c6b1a-93dd-4f71-81b6-1c43fcf5f556-var-lib-openvswitch" (OuterVolumeSpecName: "var-lib-openvswitch") pod "736c6b1a-93dd-4f71-81b6-1c43fcf5f556" (UID: "736c6b1a-93dd-4f71-81b6-1c43fcf5f556"). InnerVolumeSpecName "var-lib-openvswitch". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 13 07:03:12 crc kubenswrapper[4971]: I1213 07:03:12.606180 4971 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/736c6b1a-93dd-4f71-81b6-1c43fcf5f556-host-slash" (OuterVolumeSpecName: "host-slash") pod "736c6b1a-93dd-4f71-81b6-1c43fcf5f556" (UID: "736c6b1a-93dd-4f71-81b6-1c43fcf5f556"). InnerVolumeSpecName "host-slash". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 13 07:03:12 crc kubenswrapper[4971]: I1213 07:03:12.606644 4971 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"host-slash\" (UniqueName: \"kubernetes.io/host-path/736c6b1a-93dd-4f71-81b6-1c43fcf5f556-host-slash\") pod \"736c6b1a-93dd-4f71-81b6-1c43fcf5f556\" (UID: \"736c6b1a-93dd-4f71-81b6-1c43fcf5f556\") " Dec 13 07:03:12 crc kubenswrapper[4971]: I1213 07:03:12.606888 4971 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-cni-bin\" (UniqueName: \"kubernetes.io/host-path/45f87d92-a73d-47ed-83da-400dc4d72255-host-cni-bin\") pod \"ovnkube-node-fq8jh\" (UID: \"45f87d92-a73d-47ed-83da-400dc4d72255\") " pod="openshift-ovn-kubernetes/ovnkube-node-fq8jh" Dec 13 07:03:12 crc kubenswrapper[4971]: I1213 07:03:12.607031 4971 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run-openvswitch\" (UniqueName: \"kubernetes.io/host-path/45f87d92-a73d-47ed-83da-400dc4d72255-run-openvswitch\") pod \"ovnkube-node-fq8jh\" (UID: \"45f87d92-a73d-47ed-83da-400dc4d72255\") " pod="openshift-ovn-kubernetes/ovnkube-node-fq8jh" Dec 13 07:03:12 crc kubenswrapper[4971]: I1213 07:03:12.607139 4971 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/45f87d92-a73d-47ed-83da-400dc4d72255-env-overrides\") pod \"ovnkube-node-fq8jh\" (UID: \"45f87d92-a73d-47ed-83da-400dc4d72255\") " pod="openshift-ovn-kubernetes/ovnkube-node-fq8jh" Dec 13 07:03:12 crc kubenswrapper[4971]: I1213 07:03:12.607249 4971 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovnkube-script-lib\" (UniqueName: \"kubernetes.io/configmap/45f87d92-a73d-47ed-83da-400dc4d72255-ovnkube-script-lib\") pod \"ovnkube-node-fq8jh\" (UID: \"45f87d92-a73d-47ed-83da-400dc4d72255\") " pod="openshift-ovn-kubernetes/ovnkube-node-fq8jh" Dec 13 07:03:12 crc kubenswrapper[4971]: I1213 07:03:12.607388 4971 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run-systemd\" (UniqueName: \"kubernetes.io/host-path/45f87d92-a73d-47ed-83da-400dc4d72255-run-systemd\") pod \"ovnkube-node-fq8jh\" (UID: \"45f87d92-a73d-47ed-83da-400dc4d72255\") " pod="openshift-ovn-kubernetes/ovnkube-node-fq8jh" Dec 13 07:03:12 crc kubenswrapper[4971]: I1213 07:03:12.607564 4971 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-var-lib-cni-networks-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/45f87d92-a73d-47ed-83da-400dc4d72255-host-var-lib-cni-networks-ovn-kubernetes\") pod \"ovnkube-node-fq8jh\" (UID: \"45f87d92-a73d-47ed-83da-400dc4d72255\") " pod="openshift-ovn-kubernetes/ovnkube-node-fq8jh" Dec 13 07:03:12 crc kubenswrapper[4971]: I1213 07:03:12.607689 4971 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-kubelet\" (UniqueName: \"kubernetes.io/host-path/45f87d92-a73d-47ed-83da-400dc4d72255-host-kubelet\") pod \"ovnkube-node-fq8jh\" (UID: \"45f87d92-a73d-47ed-83da-400dc4d72255\") " pod="openshift-ovn-kubernetes/ovnkube-node-fq8jh" Dec 13 07:03:12 crc kubenswrapper[4971]: I1213 07:03:12.607795 4971 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-run-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/45f87d92-a73d-47ed-83da-400dc4d72255-host-run-ovn-kubernetes\") pod \"ovnkube-node-fq8jh\" (UID: \"45f87d92-a73d-47ed-83da-400dc4d72255\") " pod="openshift-ovn-kubernetes/ovnkube-node-fq8jh" Dec 13 07:03:12 crc kubenswrapper[4971]: I1213 07:03:12.607907 4971 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-slash\" (UniqueName: \"kubernetes.io/host-path/45f87d92-a73d-47ed-83da-400dc4d72255-host-slash\") pod \"ovnkube-node-fq8jh\" (UID: \"45f87d92-a73d-47ed-83da-400dc4d72255\") " pod="openshift-ovn-kubernetes/ovnkube-node-fq8jh" Dec 13 07:03:12 crc kubenswrapper[4971]: I1213 07:03:12.608005 4971 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"systemd-units\" (UniqueName: \"kubernetes.io/host-path/45f87d92-a73d-47ed-83da-400dc4d72255-systemd-units\") pod \"ovnkube-node-fq8jh\" (UID: \"45f87d92-a73d-47ed-83da-400dc4d72255\") " pod="openshift-ovn-kubernetes/ovnkube-node-fq8jh" Dec 13 07:03:12 crc kubenswrapper[4971]: I1213 07:03:12.608110 4971 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"log-socket\" (UniqueName: \"kubernetes.io/host-path/45f87d92-a73d-47ed-83da-400dc4d72255-log-socket\") pod \"ovnkube-node-fq8jh\" (UID: \"45f87d92-a73d-47ed-83da-400dc4d72255\") " pod="openshift-ovn-kubernetes/ovnkube-node-fq8jh" Dec 13 07:03:12 crc kubenswrapper[4971]: I1213 07:03:12.608207 4971 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-run-netns\" (UniqueName: \"kubernetes.io/host-path/45f87d92-a73d-47ed-83da-400dc4d72255-host-run-netns\") pod \"ovnkube-node-fq8jh\" (UID: \"45f87d92-a73d-47ed-83da-400dc4d72255\") " pod="openshift-ovn-kubernetes/ovnkube-node-fq8jh" Dec 13 07:03:12 crc kubenswrapper[4971]: I1213 07:03:12.608297 4971 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-2fbhj\" (UniqueName: \"kubernetes.io/projected/45f87d92-a73d-47ed-83da-400dc4d72255-kube-api-access-2fbhj\") pod \"ovnkube-node-fq8jh\" (UID: \"45f87d92-a73d-47ed-83da-400dc4d72255\") " pod="openshift-ovn-kubernetes/ovnkube-node-fq8jh" Dec 13 07:03:12 crc kubenswrapper[4971]: I1213 07:03:12.608389 4971 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"node-log\" (UniqueName: \"kubernetes.io/host-path/45f87d92-a73d-47ed-83da-400dc4d72255-node-log\") pod \"ovnkube-node-fq8jh\" (UID: \"45f87d92-a73d-47ed-83da-400dc4d72255\") " pod="openshift-ovn-kubernetes/ovnkube-node-fq8jh" Dec 13 07:03:12 crc kubenswrapper[4971]: I1213 07:03:12.608543 4971 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run-ovn\" (UniqueName: \"kubernetes.io/host-path/45f87d92-a73d-47ed-83da-400dc4d72255-run-ovn\") pod \"ovnkube-node-fq8jh\" (UID: \"45f87d92-a73d-47ed-83da-400dc4d72255\") " pod="openshift-ovn-kubernetes/ovnkube-node-fq8jh" Dec 13 07:03:12 crc kubenswrapper[4971]: I1213 07:03:12.608591 4971 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-lib-openvswitch\" (UniqueName: \"kubernetes.io/host-path/45f87d92-a73d-47ed-83da-400dc4d72255-var-lib-openvswitch\") pod \"ovnkube-node-fq8jh\" (UID: \"45f87d92-a73d-47ed-83da-400dc4d72255\") " pod="openshift-ovn-kubernetes/ovnkube-node-fq8jh" Dec 13 07:03:12 crc kubenswrapper[4971]: I1213 07:03:12.608618 4971 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/45f87d92-a73d-47ed-83da-400dc4d72255-ovnkube-config\") pod \"ovnkube-node-fq8jh\" (UID: \"45f87d92-a73d-47ed-83da-400dc4d72255\") " pod="openshift-ovn-kubernetes/ovnkube-node-fq8jh" Dec 13 07:03:12 crc kubenswrapper[4971]: I1213 07:03:12.608670 4971 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-openvswitch\" (UniqueName: \"kubernetes.io/host-path/45f87d92-a73d-47ed-83da-400dc4d72255-etc-openvswitch\") pod \"ovnkube-node-fq8jh\" (UID: \"45f87d92-a73d-47ed-83da-400dc4d72255\") " pod="openshift-ovn-kubernetes/ovnkube-node-fq8jh" Dec 13 07:03:12 crc kubenswrapper[4971]: I1213 07:03:12.608721 4971 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-cni-netd\" (UniqueName: \"kubernetes.io/host-path/45f87d92-a73d-47ed-83da-400dc4d72255-host-cni-netd\") pod \"ovnkube-node-fq8jh\" (UID: \"45f87d92-a73d-47ed-83da-400dc4d72255\") " pod="openshift-ovn-kubernetes/ovnkube-node-fq8jh" Dec 13 07:03:12 crc kubenswrapper[4971]: I1213 07:03:12.608835 4971 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovn-node-metrics-cert\" (UniqueName: \"kubernetes.io/secret/45f87d92-a73d-47ed-83da-400dc4d72255-ovn-node-metrics-cert\") pod \"ovnkube-node-fq8jh\" (UID: \"45f87d92-a73d-47ed-83da-400dc4d72255\") " pod="openshift-ovn-kubernetes/ovnkube-node-fq8jh" Dec 13 07:03:12 crc kubenswrapper[4971]: I1213 07:03:12.608964 4971 reconciler_common.go:293] "Volume detached for volume \"host-kubelet\" (UniqueName: \"kubernetes.io/host-path/736c6b1a-93dd-4f71-81b6-1c43fcf5f556-host-kubelet\") on node \"crc\" DevicePath \"\"" Dec 13 07:03:12 crc kubenswrapper[4971]: I1213 07:03:12.608978 4971 reconciler_common.go:293] "Volume detached for volume \"host-run-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/736c6b1a-93dd-4f71-81b6-1c43fcf5f556-host-run-ovn-kubernetes\") on node \"crc\" DevicePath \"\"" Dec 13 07:03:12 crc kubenswrapper[4971]: I1213 07:03:12.608990 4971 reconciler_common.go:293] "Volume detached for volume \"run-openvswitch\" (UniqueName: \"kubernetes.io/host-path/736c6b1a-93dd-4f71-81b6-1c43fcf5f556-run-openvswitch\") on node \"crc\" DevicePath \"\"" Dec 13 07:03:12 crc kubenswrapper[4971]: I1213 07:03:12.609001 4971 reconciler_common.go:293] "Volume detached for volume \"host-var-lib-cni-networks-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/736c6b1a-93dd-4f71-81b6-1c43fcf5f556-host-var-lib-cni-networks-ovn-kubernetes\") on node \"crc\" DevicePath \"\"" Dec 13 07:03:12 crc kubenswrapper[4971]: I1213 07:03:12.609014 4971 reconciler_common.go:293] "Volume detached for volume \"host-run-netns\" (UniqueName: \"kubernetes.io/host-path/736c6b1a-93dd-4f71-81b6-1c43fcf5f556-host-run-netns\") on node \"crc\" DevicePath \"\"" Dec 13 07:03:12 crc kubenswrapper[4971]: I1213 07:03:12.609040 4971 reconciler_common.go:293] "Volume detached for volume \"systemd-units\" (UniqueName: \"kubernetes.io/host-path/736c6b1a-93dd-4f71-81b6-1c43fcf5f556-systemd-units\") on node \"crc\" DevicePath \"\"" Dec 13 07:03:12 crc kubenswrapper[4971]: I1213 07:03:12.609052 4971 reconciler_common.go:293] "Volume detached for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/736c6b1a-93dd-4f71-81b6-1c43fcf5f556-env-overrides\") on node \"crc\" DevicePath \"\"" Dec 13 07:03:12 crc kubenswrapper[4971]: I1213 07:03:12.609063 4971 reconciler_common.go:293] "Volume detached for volume \"log-socket\" (UniqueName: \"kubernetes.io/host-path/736c6b1a-93dd-4f71-81b6-1c43fcf5f556-log-socket\") on node \"crc\" DevicePath \"\"" Dec 13 07:03:12 crc kubenswrapper[4971]: I1213 07:03:12.609076 4971 reconciler_common.go:293] "Volume detached for volume \"host-cni-netd\" (UniqueName: \"kubernetes.io/host-path/736c6b1a-93dd-4f71-81b6-1c43fcf5f556-host-cni-netd\") on node \"crc\" DevicePath \"\"" Dec 13 07:03:12 crc kubenswrapper[4971]: I1213 07:03:12.609093 4971 reconciler_common.go:293] "Volume detached for volume \"run-ovn\" (UniqueName: \"kubernetes.io/host-path/736c6b1a-93dd-4f71-81b6-1c43fcf5f556-run-ovn\") on node \"crc\" DevicePath \"\"" Dec 13 07:03:12 crc kubenswrapper[4971]: I1213 07:03:12.609104 4971 reconciler_common.go:293] "Volume detached for volume \"var-lib-openvswitch\" (UniqueName: \"kubernetes.io/host-path/736c6b1a-93dd-4f71-81b6-1c43fcf5f556-var-lib-openvswitch\") on node \"crc\" DevicePath \"\"" Dec 13 07:03:12 crc kubenswrapper[4971]: I1213 07:03:12.609115 4971 reconciler_common.go:293] "Volume detached for volume \"host-cni-bin\" (UniqueName: \"kubernetes.io/host-path/736c6b1a-93dd-4f71-81b6-1c43fcf5f556-host-cni-bin\") on node \"crc\" DevicePath \"\"" Dec 13 07:03:12 crc kubenswrapper[4971]: I1213 07:03:12.609125 4971 reconciler_common.go:293] "Volume detached for volume \"host-slash\" (UniqueName: \"kubernetes.io/host-path/736c6b1a-93dd-4f71-81b6-1c43fcf5f556-host-slash\") on node \"crc\" DevicePath \"\"" Dec 13 07:03:12 crc kubenswrapper[4971]: I1213 07:03:12.609135 4971 reconciler_common.go:293] "Volume detached for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/736c6b1a-93dd-4f71-81b6-1c43fcf5f556-ovnkube-config\") on node \"crc\" DevicePath \"\"" Dec 13 07:03:12 crc kubenswrapper[4971]: I1213 07:03:12.609146 4971 reconciler_common.go:293] "Volume detached for volume \"etc-openvswitch\" (UniqueName: \"kubernetes.io/host-path/736c6b1a-93dd-4f71-81b6-1c43fcf5f556-etc-openvswitch\") on node \"crc\" DevicePath \"\"" Dec 13 07:03:12 crc kubenswrapper[4971]: I1213 07:03:12.609155 4971 reconciler_common.go:293] "Volume detached for volume \"ovnkube-script-lib\" (UniqueName: \"kubernetes.io/configmap/736c6b1a-93dd-4f71-81b6-1c43fcf5f556-ovnkube-script-lib\") on node \"crc\" DevicePath \"\"" Dec 13 07:03:12 crc kubenswrapper[4971]: I1213 07:03:12.609164 4971 reconciler_common.go:293] "Volume detached for volume \"node-log\" (UniqueName: \"kubernetes.io/host-path/736c6b1a-93dd-4f71-81b6-1c43fcf5f556-node-log\") on node \"crc\" DevicePath \"\"" Dec 13 07:03:12 crc kubenswrapper[4971]: I1213 07:03:12.609721 4971 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/736c6b1a-93dd-4f71-81b6-1c43fcf5f556-kube-api-access-vzmsr" (OuterVolumeSpecName: "kube-api-access-vzmsr") pod "736c6b1a-93dd-4f71-81b6-1c43fcf5f556" (UID: "736c6b1a-93dd-4f71-81b6-1c43fcf5f556"). InnerVolumeSpecName "kube-api-access-vzmsr". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 13 07:03:12 crc kubenswrapper[4971]: I1213 07:03:12.609812 4971 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/736c6b1a-93dd-4f71-81b6-1c43fcf5f556-ovn-node-metrics-cert" (OuterVolumeSpecName: "ovn-node-metrics-cert") pod "736c6b1a-93dd-4f71-81b6-1c43fcf5f556" (UID: "736c6b1a-93dd-4f71-81b6-1c43fcf5f556"). InnerVolumeSpecName "ovn-node-metrics-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 13 07:03:12 crc kubenswrapper[4971]: I1213 07:03:12.621807 4971 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/736c6b1a-93dd-4f71-81b6-1c43fcf5f556-run-systemd" (OuterVolumeSpecName: "run-systemd") pod "736c6b1a-93dd-4f71-81b6-1c43fcf5f556" (UID: "736c6b1a-93dd-4f71-81b6-1c43fcf5f556"). InnerVolumeSpecName "run-systemd". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 13 07:03:12 crc kubenswrapper[4971]: I1213 07:03:12.709896 4971 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-var-lib-cni-networks-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/45f87d92-a73d-47ed-83da-400dc4d72255-host-var-lib-cni-networks-ovn-kubernetes\") pod \"ovnkube-node-fq8jh\" (UID: \"45f87d92-a73d-47ed-83da-400dc4d72255\") " pod="openshift-ovn-kubernetes/ovnkube-node-fq8jh" Dec 13 07:03:12 crc kubenswrapper[4971]: I1213 07:03:12.709948 4971 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-kubelet\" (UniqueName: \"kubernetes.io/host-path/45f87d92-a73d-47ed-83da-400dc4d72255-host-kubelet\") pod \"ovnkube-node-fq8jh\" (UID: \"45f87d92-a73d-47ed-83da-400dc4d72255\") " pod="openshift-ovn-kubernetes/ovnkube-node-fq8jh" Dec 13 07:03:12 crc kubenswrapper[4971]: I1213 07:03:12.709968 4971 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-run-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/45f87d92-a73d-47ed-83da-400dc4d72255-host-run-ovn-kubernetes\") pod \"ovnkube-node-fq8jh\" (UID: \"45f87d92-a73d-47ed-83da-400dc4d72255\") " pod="openshift-ovn-kubernetes/ovnkube-node-fq8jh" Dec 13 07:03:12 crc kubenswrapper[4971]: I1213 07:03:12.710023 4971 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-slash\" (UniqueName: \"kubernetes.io/host-path/45f87d92-a73d-47ed-83da-400dc4d72255-host-slash\") pod \"ovnkube-node-fq8jh\" (UID: \"45f87d92-a73d-47ed-83da-400dc4d72255\") " pod="openshift-ovn-kubernetes/ovnkube-node-fq8jh" Dec 13 07:03:12 crc kubenswrapper[4971]: I1213 07:03:12.710056 4971 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"systemd-units\" (UniqueName: \"kubernetes.io/host-path/45f87d92-a73d-47ed-83da-400dc4d72255-systemd-units\") pod \"ovnkube-node-fq8jh\" (UID: \"45f87d92-a73d-47ed-83da-400dc4d72255\") " pod="openshift-ovn-kubernetes/ovnkube-node-fq8jh" Dec 13 07:03:12 crc kubenswrapper[4971]: I1213 07:03:12.710079 4971 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"log-socket\" (UniqueName: \"kubernetes.io/host-path/45f87d92-a73d-47ed-83da-400dc4d72255-log-socket\") pod \"ovnkube-node-fq8jh\" (UID: \"45f87d92-a73d-47ed-83da-400dc4d72255\") " pod="openshift-ovn-kubernetes/ovnkube-node-fq8jh" Dec 13 07:03:12 crc kubenswrapper[4971]: I1213 07:03:12.710105 4971 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-run-netns\" (UniqueName: \"kubernetes.io/host-path/45f87d92-a73d-47ed-83da-400dc4d72255-host-run-netns\") pod \"ovnkube-node-fq8jh\" (UID: \"45f87d92-a73d-47ed-83da-400dc4d72255\") " pod="openshift-ovn-kubernetes/ovnkube-node-fq8jh" Dec 13 07:03:12 crc kubenswrapper[4971]: I1213 07:03:12.710109 4971 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-kubelet\" (UniqueName: \"kubernetes.io/host-path/45f87d92-a73d-47ed-83da-400dc4d72255-host-kubelet\") pod \"ovnkube-node-fq8jh\" (UID: \"45f87d92-a73d-47ed-83da-400dc4d72255\") " pod="openshift-ovn-kubernetes/ovnkube-node-fq8jh" Dec 13 07:03:12 crc kubenswrapper[4971]: I1213 07:03:12.710059 4971 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-run-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/45f87d92-a73d-47ed-83da-400dc4d72255-host-run-ovn-kubernetes\") pod \"ovnkube-node-fq8jh\" (UID: \"45f87d92-a73d-47ed-83da-400dc4d72255\") " pod="openshift-ovn-kubernetes/ovnkube-node-fq8jh" Dec 13 07:03:12 crc kubenswrapper[4971]: I1213 07:03:12.710157 4971 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"systemd-units\" (UniqueName: \"kubernetes.io/host-path/45f87d92-a73d-47ed-83da-400dc4d72255-systemd-units\") pod \"ovnkube-node-fq8jh\" (UID: \"45f87d92-a73d-47ed-83da-400dc4d72255\") " pod="openshift-ovn-kubernetes/ovnkube-node-fq8jh" Dec 13 07:03:12 crc kubenswrapper[4971]: I1213 07:03:12.710124 4971 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-2fbhj\" (UniqueName: \"kubernetes.io/projected/45f87d92-a73d-47ed-83da-400dc4d72255-kube-api-access-2fbhj\") pod \"ovnkube-node-fq8jh\" (UID: \"45f87d92-a73d-47ed-83da-400dc4d72255\") " pod="openshift-ovn-kubernetes/ovnkube-node-fq8jh" Dec 13 07:03:12 crc kubenswrapper[4971]: I1213 07:03:12.710151 4971 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"log-socket\" (UniqueName: \"kubernetes.io/host-path/45f87d92-a73d-47ed-83da-400dc4d72255-log-socket\") pod \"ovnkube-node-fq8jh\" (UID: \"45f87d92-a73d-47ed-83da-400dc4d72255\") " pod="openshift-ovn-kubernetes/ovnkube-node-fq8jh" Dec 13 07:03:12 crc kubenswrapper[4971]: I1213 07:03:12.710285 4971 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"node-log\" (UniqueName: \"kubernetes.io/host-path/45f87d92-a73d-47ed-83da-400dc4d72255-node-log\") pod \"ovnkube-node-fq8jh\" (UID: \"45f87d92-a73d-47ed-83da-400dc4d72255\") " pod="openshift-ovn-kubernetes/ovnkube-node-fq8jh" Dec 13 07:03:12 crc kubenswrapper[4971]: I1213 07:03:12.710085 4971 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-slash\" (UniqueName: \"kubernetes.io/host-path/45f87d92-a73d-47ed-83da-400dc4d72255-host-slash\") pod \"ovnkube-node-fq8jh\" (UID: \"45f87d92-a73d-47ed-83da-400dc4d72255\") " pod="openshift-ovn-kubernetes/ovnkube-node-fq8jh" Dec 13 07:03:12 crc kubenswrapper[4971]: I1213 07:03:12.710016 4971 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-var-lib-cni-networks-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/45f87d92-a73d-47ed-83da-400dc4d72255-host-var-lib-cni-networks-ovn-kubernetes\") pod \"ovnkube-node-fq8jh\" (UID: \"45f87d92-a73d-47ed-83da-400dc4d72255\") " pod="openshift-ovn-kubernetes/ovnkube-node-fq8jh" Dec 13 07:03:12 crc kubenswrapper[4971]: I1213 07:03:12.710258 4971 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"node-log\" (UniqueName: \"kubernetes.io/host-path/45f87d92-a73d-47ed-83da-400dc4d72255-node-log\") pod \"ovnkube-node-fq8jh\" (UID: \"45f87d92-a73d-47ed-83da-400dc4d72255\") " pod="openshift-ovn-kubernetes/ovnkube-node-fq8jh" Dec 13 07:03:12 crc kubenswrapper[4971]: I1213 07:03:12.710362 4971 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run-ovn\" (UniqueName: \"kubernetes.io/host-path/45f87d92-a73d-47ed-83da-400dc4d72255-run-ovn\") pod \"ovnkube-node-fq8jh\" (UID: \"45f87d92-a73d-47ed-83da-400dc4d72255\") " pod="openshift-ovn-kubernetes/ovnkube-node-fq8jh" Dec 13 07:03:12 crc kubenswrapper[4971]: I1213 07:03:12.710175 4971 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-run-netns\" (UniqueName: \"kubernetes.io/host-path/45f87d92-a73d-47ed-83da-400dc4d72255-host-run-netns\") pod \"ovnkube-node-fq8jh\" (UID: \"45f87d92-a73d-47ed-83da-400dc4d72255\") " pod="openshift-ovn-kubernetes/ovnkube-node-fq8jh" Dec 13 07:03:12 crc kubenswrapper[4971]: I1213 07:03:12.710397 4971 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-lib-openvswitch\" (UniqueName: \"kubernetes.io/host-path/45f87d92-a73d-47ed-83da-400dc4d72255-var-lib-openvswitch\") pod \"ovnkube-node-fq8jh\" (UID: \"45f87d92-a73d-47ed-83da-400dc4d72255\") " pod="openshift-ovn-kubernetes/ovnkube-node-fq8jh" Dec 13 07:03:12 crc kubenswrapper[4971]: I1213 07:03:12.710379 4971 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-lib-openvswitch\" (UniqueName: \"kubernetes.io/host-path/45f87d92-a73d-47ed-83da-400dc4d72255-var-lib-openvswitch\") pod \"ovnkube-node-fq8jh\" (UID: \"45f87d92-a73d-47ed-83da-400dc4d72255\") " pod="openshift-ovn-kubernetes/ovnkube-node-fq8jh" Dec 13 07:03:12 crc kubenswrapper[4971]: I1213 07:03:12.710442 4971 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run-ovn\" (UniqueName: \"kubernetes.io/host-path/45f87d92-a73d-47ed-83da-400dc4d72255-run-ovn\") pod \"ovnkube-node-fq8jh\" (UID: \"45f87d92-a73d-47ed-83da-400dc4d72255\") " pod="openshift-ovn-kubernetes/ovnkube-node-fq8jh" Dec 13 07:03:12 crc kubenswrapper[4971]: I1213 07:03:12.710459 4971 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/45f87d92-a73d-47ed-83da-400dc4d72255-ovnkube-config\") pod \"ovnkube-node-fq8jh\" (UID: \"45f87d92-a73d-47ed-83da-400dc4d72255\") " pod="openshift-ovn-kubernetes/ovnkube-node-fq8jh" Dec 13 07:03:12 crc kubenswrapper[4971]: I1213 07:03:12.710499 4971 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-openvswitch\" (UniqueName: \"kubernetes.io/host-path/45f87d92-a73d-47ed-83da-400dc4d72255-etc-openvswitch\") pod \"ovnkube-node-fq8jh\" (UID: \"45f87d92-a73d-47ed-83da-400dc4d72255\") " pod="openshift-ovn-kubernetes/ovnkube-node-fq8jh" Dec 13 07:03:12 crc kubenswrapper[4971]: I1213 07:03:12.710563 4971 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-cni-netd\" (UniqueName: \"kubernetes.io/host-path/45f87d92-a73d-47ed-83da-400dc4d72255-host-cni-netd\") pod \"ovnkube-node-fq8jh\" (UID: \"45f87d92-a73d-47ed-83da-400dc4d72255\") " pod="openshift-ovn-kubernetes/ovnkube-node-fq8jh" Dec 13 07:03:12 crc kubenswrapper[4971]: I1213 07:03:12.710583 4971 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-openvswitch\" (UniqueName: \"kubernetes.io/host-path/45f87d92-a73d-47ed-83da-400dc4d72255-etc-openvswitch\") pod \"ovnkube-node-fq8jh\" (UID: \"45f87d92-a73d-47ed-83da-400dc4d72255\") " pod="openshift-ovn-kubernetes/ovnkube-node-fq8jh" Dec 13 07:03:12 crc kubenswrapper[4971]: I1213 07:03:12.710610 4971 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-cni-netd\" (UniqueName: \"kubernetes.io/host-path/45f87d92-a73d-47ed-83da-400dc4d72255-host-cni-netd\") pod \"ovnkube-node-fq8jh\" (UID: \"45f87d92-a73d-47ed-83da-400dc4d72255\") " pod="openshift-ovn-kubernetes/ovnkube-node-fq8jh" Dec 13 07:03:12 crc kubenswrapper[4971]: I1213 07:03:12.710629 4971 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovn-node-metrics-cert\" (UniqueName: \"kubernetes.io/secret/45f87d92-a73d-47ed-83da-400dc4d72255-ovn-node-metrics-cert\") pod \"ovnkube-node-fq8jh\" (UID: \"45f87d92-a73d-47ed-83da-400dc4d72255\") " pod="openshift-ovn-kubernetes/ovnkube-node-fq8jh" Dec 13 07:03:12 crc kubenswrapper[4971]: I1213 07:03:12.710669 4971 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-cni-bin\" (UniqueName: \"kubernetes.io/host-path/45f87d92-a73d-47ed-83da-400dc4d72255-host-cni-bin\") pod \"ovnkube-node-fq8jh\" (UID: \"45f87d92-a73d-47ed-83da-400dc4d72255\") " pod="openshift-ovn-kubernetes/ovnkube-node-fq8jh" Dec 13 07:03:12 crc kubenswrapper[4971]: I1213 07:03:12.710712 4971 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run-openvswitch\" (UniqueName: \"kubernetes.io/host-path/45f87d92-a73d-47ed-83da-400dc4d72255-run-openvswitch\") pod \"ovnkube-node-fq8jh\" (UID: \"45f87d92-a73d-47ed-83da-400dc4d72255\") " pod="openshift-ovn-kubernetes/ovnkube-node-fq8jh" Dec 13 07:03:12 crc kubenswrapper[4971]: I1213 07:03:12.710744 4971 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/45f87d92-a73d-47ed-83da-400dc4d72255-env-overrides\") pod \"ovnkube-node-fq8jh\" (UID: \"45f87d92-a73d-47ed-83da-400dc4d72255\") " pod="openshift-ovn-kubernetes/ovnkube-node-fq8jh" Dec 13 07:03:12 crc kubenswrapper[4971]: I1213 07:03:12.710775 4971 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovnkube-script-lib\" (UniqueName: \"kubernetes.io/configmap/45f87d92-a73d-47ed-83da-400dc4d72255-ovnkube-script-lib\") pod \"ovnkube-node-fq8jh\" (UID: \"45f87d92-a73d-47ed-83da-400dc4d72255\") " pod="openshift-ovn-kubernetes/ovnkube-node-fq8jh" Dec 13 07:03:12 crc kubenswrapper[4971]: I1213 07:03:12.710789 4971 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-cni-bin\" (UniqueName: \"kubernetes.io/host-path/45f87d92-a73d-47ed-83da-400dc4d72255-host-cni-bin\") pod \"ovnkube-node-fq8jh\" (UID: \"45f87d92-a73d-47ed-83da-400dc4d72255\") " pod="openshift-ovn-kubernetes/ovnkube-node-fq8jh" Dec 13 07:03:12 crc kubenswrapper[4971]: I1213 07:03:12.710827 4971 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run-openvswitch\" (UniqueName: \"kubernetes.io/host-path/45f87d92-a73d-47ed-83da-400dc4d72255-run-openvswitch\") pod \"ovnkube-node-fq8jh\" (UID: \"45f87d92-a73d-47ed-83da-400dc4d72255\") " pod="openshift-ovn-kubernetes/ovnkube-node-fq8jh" Dec 13 07:03:12 crc kubenswrapper[4971]: I1213 07:03:12.710851 4971 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run-systemd\" (UniqueName: \"kubernetes.io/host-path/45f87d92-a73d-47ed-83da-400dc4d72255-run-systemd\") pod \"ovnkube-node-fq8jh\" (UID: \"45f87d92-a73d-47ed-83da-400dc4d72255\") " pod="openshift-ovn-kubernetes/ovnkube-node-fq8jh" Dec 13 07:03:12 crc kubenswrapper[4971]: I1213 07:03:12.710958 4971 reconciler_common.go:293] "Volume detached for volume \"ovn-node-metrics-cert\" (UniqueName: \"kubernetes.io/secret/736c6b1a-93dd-4f71-81b6-1c43fcf5f556-ovn-node-metrics-cert\") on node \"crc\" DevicePath \"\"" Dec 13 07:03:12 crc kubenswrapper[4971]: I1213 07:03:12.710975 4971 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-vzmsr\" (UniqueName: \"kubernetes.io/projected/736c6b1a-93dd-4f71-81b6-1c43fcf5f556-kube-api-access-vzmsr\") on node \"crc\" DevicePath \"\"" Dec 13 07:03:12 crc kubenswrapper[4971]: I1213 07:03:12.710989 4971 reconciler_common.go:293] "Volume detached for volume \"run-systemd\" (UniqueName: \"kubernetes.io/host-path/736c6b1a-93dd-4f71-81b6-1c43fcf5f556-run-systemd\") on node \"crc\" DevicePath \"\"" Dec 13 07:03:12 crc kubenswrapper[4971]: I1213 07:03:12.711006 4971 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run-systemd\" (UniqueName: \"kubernetes.io/host-path/45f87d92-a73d-47ed-83da-400dc4d72255-run-systemd\") pod \"ovnkube-node-fq8jh\" (UID: \"45f87d92-a73d-47ed-83da-400dc4d72255\") " pod="openshift-ovn-kubernetes/ovnkube-node-fq8jh" Dec 13 07:03:12 crc kubenswrapper[4971]: I1213 07:03:12.711263 4971 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/45f87d92-a73d-47ed-83da-400dc4d72255-env-overrides\") pod \"ovnkube-node-fq8jh\" (UID: \"45f87d92-a73d-47ed-83da-400dc4d72255\") " pod="openshift-ovn-kubernetes/ovnkube-node-fq8jh" Dec 13 07:03:12 crc kubenswrapper[4971]: I1213 07:03:12.711314 4971 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/45f87d92-a73d-47ed-83da-400dc4d72255-ovnkube-config\") pod \"ovnkube-node-fq8jh\" (UID: \"45f87d92-a73d-47ed-83da-400dc4d72255\") " pod="openshift-ovn-kubernetes/ovnkube-node-fq8jh" Dec 13 07:03:12 crc kubenswrapper[4971]: I1213 07:03:12.711482 4971 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovnkube-script-lib\" (UniqueName: \"kubernetes.io/configmap/45f87d92-a73d-47ed-83da-400dc4d72255-ovnkube-script-lib\") pod \"ovnkube-node-fq8jh\" (UID: \"45f87d92-a73d-47ed-83da-400dc4d72255\") " pod="openshift-ovn-kubernetes/ovnkube-node-fq8jh" Dec 13 07:03:12 crc kubenswrapper[4971]: I1213 07:03:12.715746 4971 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovn-node-metrics-cert\" (UniqueName: \"kubernetes.io/secret/45f87d92-a73d-47ed-83da-400dc4d72255-ovn-node-metrics-cert\") pod \"ovnkube-node-fq8jh\" (UID: \"45f87d92-a73d-47ed-83da-400dc4d72255\") " pod="openshift-ovn-kubernetes/ovnkube-node-fq8jh" Dec 13 07:03:12 crc kubenswrapper[4971]: I1213 07:03:12.726272 4971 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-2fbhj\" (UniqueName: \"kubernetes.io/projected/45f87d92-a73d-47ed-83da-400dc4d72255-kube-api-access-2fbhj\") pod \"ovnkube-node-fq8jh\" (UID: \"45f87d92-a73d-47ed-83da-400dc4d72255\") " pod="openshift-ovn-kubernetes/ovnkube-node-fq8jh" Dec 13 07:03:12 crc kubenswrapper[4971]: I1213 07:03:12.814725 4971 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-multus_multus-wwvpl_dddf744b-2122-49ad-a980-105c09636e0f/kube-multus/2.log" Dec 13 07:03:12 crc kubenswrapper[4971]: I1213 07:03:12.815617 4971 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-multus_multus-wwvpl_dddf744b-2122-49ad-a980-105c09636e0f/kube-multus/1.log" Dec 13 07:03:12 crc kubenswrapper[4971]: I1213 07:03:12.815690 4971 generic.go:334] "Generic (PLEG): container finished" podID="dddf744b-2122-49ad-a980-105c09636e0f" containerID="b02826c7d90e9a7a2ed0fa1dd785f2690d152ec880d5eee474ddd4d2d89c8f75" exitCode=2 Dec 13 07:03:12 crc kubenswrapper[4971]: I1213 07:03:12.815796 4971 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-wwvpl" event={"ID":"dddf744b-2122-49ad-a980-105c09636e0f","Type":"ContainerDied","Data":"b02826c7d90e9a7a2ed0fa1dd785f2690d152ec880d5eee474ddd4d2d89c8f75"} Dec 13 07:03:12 crc kubenswrapper[4971]: I1213 07:03:12.815869 4971 scope.go:117] "RemoveContainer" containerID="c8680aaa9e05095ac7e467c896d9e927bccff908925768ce5ed469ab89a72475" Dec 13 07:03:12 crc kubenswrapper[4971]: I1213 07:03:12.816630 4971 scope.go:117] "RemoveContainer" containerID="b02826c7d90e9a7a2ed0fa1dd785f2690d152ec880d5eee474ddd4d2d89c8f75" Dec 13 07:03:12 crc kubenswrapper[4971]: I1213 07:03:12.818913 4971 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-zlcg5_736c6b1a-93dd-4f71-81b6-1c43fcf5f556/ovnkube-controller/3.log" Dec 13 07:03:12 crc kubenswrapper[4971]: I1213 07:03:12.824184 4971 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-ovn-kubernetes/ovnkube-node-fq8jh" Dec 13 07:03:12 crc kubenswrapper[4971]: I1213 07:03:12.825681 4971 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-zlcg5_736c6b1a-93dd-4f71-81b6-1c43fcf5f556/ovn-acl-logging/0.log" Dec 13 07:03:12 crc kubenswrapper[4971]: I1213 07:03:12.826928 4971 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-zlcg5_736c6b1a-93dd-4f71-81b6-1c43fcf5f556/ovn-controller/0.log" Dec 13 07:03:12 crc kubenswrapper[4971]: I1213 07:03:12.827300 4971 generic.go:334] "Generic (PLEG): container finished" podID="736c6b1a-93dd-4f71-81b6-1c43fcf5f556" containerID="6afaff46af01dcc99e08b76967fe4181e615cd044389b9b8cea1581505eef247" exitCode=0 Dec 13 07:03:12 crc kubenswrapper[4971]: I1213 07:03:12.827341 4971 generic.go:334] "Generic (PLEG): container finished" podID="736c6b1a-93dd-4f71-81b6-1c43fcf5f556" containerID="868049c6a3a4ebe4036d1d22e02ba6b76bced2ee7e269bb228accc4a0636678a" exitCode=0 Dec 13 07:03:12 crc kubenswrapper[4971]: I1213 07:03:12.827350 4971 generic.go:334] "Generic (PLEG): container finished" podID="736c6b1a-93dd-4f71-81b6-1c43fcf5f556" containerID="f6073140d9f86929b6d4bd2960ede0617a5c9620b269bbab043bae37e0d389f5" exitCode=0 Dec 13 07:03:12 crc kubenswrapper[4971]: I1213 07:03:12.827358 4971 generic.go:334] "Generic (PLEG): container finished" podID="736c6b1a-93dd-4f71-81b6-1c43fcf5f556" containerID="693a2ab434c250cb140a866e8068b156df9191414fed1d63bee919a2528ea6e5" exitCode=0 Dec 13 07:03:12 crc kubenswrapper[4971]: I1213 07:03:12.827364 4971 generic.go:334] "Generic (PLEG): container finished" podID="736c6b1a-93dd-4f71-81b6-1c43fcf5f556" containerID="f22a4ec756119c4f791fb5105cf1f31df3e548b2b07b1f909a62802e9b6035ae" exitCode=0 Dec 13 07:03:12 crc kubenswrapper[4971]: I1213 07:03:12.827371 4971 generic.go:334] "Generic (PLEG): container finished" podID="736c6b1a-93dd-4f71-81b6-1c43fcf5f556" containerID="325223543b3c7db417f769533559a1ca2b23a723494e1885b0fd0480b5a6399f" exitCode=0 Dec 13 07:03:12 crc kubenswrapper[4971]: I1213 07:03:12.827379 4971 generic.go:334] "Generic (PLEG): container finished" podID="736c6b1a-93dd-4f71-81b6-1c43fcf5f556" containerID="219620f51474e60b3e1f880c0972eb612dffc4bad24462bbba2c093ae817d852" exitCode=143 Dec 13 07:03:12 crc kubenswrapper[4971]: I1213 07:03:12.827390 4971 generic.go:334] "Generic (PLEG): container finished" podID="736c6b1a-93dd-4f71-81b6-1c43fcf5f556" containerID="f137f4b441c2402093ab430a9689a46221c4fb4bc70d99fb992f62b0a989c465" exitCode=143 Dec 13 07:03:12 crc kubenswrapper[4971]: I1213 07:03:12.827422 4971 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-zlcg5" event={"ID":"736c6b1a-93dd-4f71-81b6-1c43fcf5f556","Type":"ContainerDied","Data":"6afaff46af01dcc99e08b76967fe4181e615cd044389b9b8cea1581505eef247"} Dec 13 07:03:12 crc kubenswrapper[4971]: I1213 07:03:12.827450 4971 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-zlcg5" event={"ID":"736c6b1a-93dd-4f71-81b6-1c43fcf5f556","Type":"ContainerDied","Data":"868049c6a3a4ebe4036d1d22e02ba6b76bced2ee7e269bb228accc4a0636678a"} Dec 13 07:03:12 crc kubenswrapper[4971]: I1213 07:03:12.827463 4971 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-zlcg5" event={"ID":"736c6b1a-93dd-4f71-81b6-1c43fcf5f556","Type":"ContainerDied","Data":"f6073140d9f86929b6d4bd2960ede0617a5c9620b269bbab043bae37e0d389f5"} Dec 13 07:03:12 crc kubenswrapper[4971]: I1213 07:03:12.827482 4971 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-zlcg5" event={"ID":"736c6b1a-93dd-4f71-81b6-1c43fcf5f556","Type":"ContainerDied","Data":"693a2ab434c250cb140a866e8068b156df9191414fed1d63bee919a2528ea6e5"} Dec 13 07:03:12 crc kubenswrapper[4971]: I1213 07:03:12.827495 4971 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-zlcg5" event={"ID":"736c6b1a-93dd-4f71-81b6-1c43fcf5f556","Type":"ContainerDied","Data":"f22a4ec756119c4f791fb5105cf1f31df3e548b2b07b1f909a62802e9b6035ae"} Dec 13 07:03:12 crc kubenswrapper[4971]: I1213 07:03:12.827505 4971 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-zlcg5" event={"ID":"736c6b1a-93dd-4f71-81b6-1c43fcf5f556","Type":"ContainerDied","Data":"325223543b3c7db417f769533559a1ca2b23a723494e1885b0fd0480b5a6399f"} Dec 13 07:03:12 crc kubenswrapper[4971]: I1213 07:03:12.827579 4971 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"6afaff46af01dcc99e08b76967fe4181e615cd044389b9b8cea1581505eef247"} Dec 13 07:03:12 crc kubenswrapper[4971]: I1213 07:03:12.827591 4971 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"7dd9913a65acba73ae1f32f4398207c5e276a1e979d86f9e720e4fd761bae4a3"} Dec 13 07:03:12 crc kubenswrapper[4971]: I1213 07:03:12.827597 4971 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"868049c6a3a4ebe4036d1d22e02ba6b76bced2ee7e269bb228accc4a0636678a"} Dec 13 07:03:12 crc kubenswrapper[4971]: I1213 07:03:12.827603 4971 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"f6073140d9f86929b6d4bd2960ede0617a5c9620b269bbab043bae37e0d389f5"} Dec 13 07:03:12 crc kubenswrapper[4971]: I1213 07:03:12.827609 4971 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"693a2ab434c250cb140a866e8068b156df9191414fed1d63bee919a2528ea6e5"} Dec 13 07:03:12 crc kubenswrapper[4971]: I1213 07:03:12.827614 4971 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"f22a4ec756119c4f791fb5105cf1f31df3e548b2b07b1f909a62802e9b6035ae"} Dec 13 07:03:12 crc kubenswrapper[4971]: I1213 07:03:12.827620 4971 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"325223543b3c7db417f769533559a1ca2b23a723494e1885b0fd0480b5a6399f"} Dec 13 07:03:12 crc kubenswrapper[4971]: I1213 07:03:12.827625 4971 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"219620f51474e60b3e1f880c0972eb612dffc4bad24462bbba2c093ae817d852"} Dec 13 07:03:12 crc kubenswrapper[4971]: I1213 07:03:12.827630 4971 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"f137f4b441c2402093ab430a9689a46221c4fb4bc70d99fb992f62b0a989c465"} Dec 13 07:03:12 crc kubenswrapper[4971]: I1213 07:03:12.827636 4971 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"451b22df6973ddc8fa2209a65d64d9e41b33eb231bb4adfab88cf0e5943bb246"} Dec 13 07:03:12 crc kubenswrapper[4971]: I1213 07:03:12.827644 4971 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-zlcg5" event={"ID":"736c6b1a-93dd-4f71-81b6-1c43fcf5f556","Type":"ContainerDied","Data":"219620f51474e60b3e1f880c0972eb612dffc4bad24462bbba2c093ae817d852"} Dec 13 07:03:12 crc kubenswrapper[4971]: I1213 07:03:12.827653 4971 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"6afaff46af01dcc99e08b76967fe4181e615cd044389b9b8cea1581505eef247"} Dec 13 07:03:12 crc kubenswrapper[4971]: I1213 07:03:12.827662 4971 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"7dd9913a65acba73ae1f32f4398207c5e276a1e979d86f9e720e4fd761bae4a3"} Dec 13 07:03:12 crc kubenswrapper[4971]: I1213 07:03:12.827671 4971 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"868049c6a3a4ebe4036d1d22e02ba6b76bced2ee7e269bb228accc4a0636678a"} Dec 13 07:03:12 crc kubenswrapper[4971]: I1213 07:03:12.827676 4971 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"f6073140d9f86929b6d4bd2960ede0617a5c9620b269bbab043bae37e0d389f5"} Dec 13 07:03:12 crc kubenswrapper[4971]: I1213 07:03:12.827682 4971 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"693a2ab434c250cb140a866e8068b156df9191414fed1d63bee919a2528ea6e5"} Dec 13 07:03:12 crc kubenswrapper[4971]: I1213 07:03:12.827687 4971 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"f22a4ec756119c4f791fb5105cf1f31df3e548b2b07b1f909a62802e9b6035ae"} Dec 13 07:03:12 crc kubenswrapper[4971]: I1213 07:03:12.827692 4971 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"325223543b3c7db417f769533559a1ca2b23a723494e1885b0fd0480b5a6399f"} Dec 13 07:03:12 crc kubenswrapper[4971]: I1213 07:03:12.827697 4971 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"219620f51474e60b3e1f880c0972eb612dffc4bad24462bbba2c093ae817d852"} Dec 13 07:03:12 crc kubenswrapper[4971]: I1213 07:03:12.827703 4971 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"f137f4b441c2402093ab430a9689a46221c4fb4bc70d99fb992f62b0a989c465"} Dec 13 07:03:12 crc kubenswrapper[4971]: I1213 07:03:12.827708 4971 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"451b22df6973ddc8fa2209a65d64d9e41b33eb231bb4adfab88cf0e5943bb246"} Dec 13 07:03:12 crc kubenswrapper[4971]: I1213 07:03:12.827716 4971 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-zlcg5" event={"ID":"736c6b1a-93dd-4f71-81b6-1c43fcf5f556","Type":"ContainerDied","Data":"f137f4b441c2402093ab430a9689a46221c4fb4bc70d99fb992f62b0a989c465"} Dec 13 07:03:12 crc kubenswrapper[4971]: I1213 07:03:12.827725 4971 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"6afaff46af01dcc99e08b76967fe4181e615cd044389b9b8cea1581505eef247"} Dec 13 07:03:12 crc kubenswrapper[4971]: I1213 07:03:12.827732 4971 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"7dd9913a65acba73ae1f32f4398207c5e276a1e979d86f9e720e4fd761bae4a3"} Dec 13 07:03:12 crc kubenswrapper[4971]: I1213 07:03:12.827738 4971 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"868049c6a3a4ebe4036d1d22e02ba6b76bced2ee7e269bb228accc4a0636678a"} Dec 13 07:03:12 crc kubenswrapper[4971]: I1213 07:03:12.827744 4971 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"f6073140d9f86929b6d4bd2960ede0617a5c9620b269bbab043bae37e0d389f5"} Dec 13 07:03:12 crc kubenswrapper[4971]: I1213 07:03:12.827750 4971 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"693a2ab434c250cb140a866e8068b156df9191414fed1d63bee919a2528ea6e5"} Dec 13 07:03:12 crc kubenswrapper[4971]: I1213 07:03:12.827758 4971 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"f22a4ec756119c4f791fb5105cf1f31df3e548b2b07b1f909a62802e9b6035ae"} Dec 13 07:03:12 crc kubenswrapper[4971]: I1213 07:03:12.827763 4971 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"325223543b3c7db417f769533559a1ca2b23a723494e1885b0fd0480b5a6399f"} Dec 13 07:03:12 crc kubenswrapper[4971]: I1213 07:03:12.827769 4971 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"219620f51474e60b3e1f880c0972eb612dffc4bad24462bbba2c093ae817d852"} Dec 13 07:03:12 crc kubenswrapper[4971]: I1213 07:03:12.827774 4971 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"f137f4b441c2402093ab430a9689a46221c4fb4bc70d99fb992f62b0a989c465"} Dec 13 07:03:12 crc kubenswrapper[4971]: I1213 07:03:12.827779 4971 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"451b22df6973ddc8fa2209a65d64d9e41b33eb231bb4adfab88cf0e5943bb246"} Dec 13 07:03:12 crc kubenswrapper[4971]: I1213 07:03:12.827788 4971 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-zlcg5" event={"ID":"736c6b1a-93dd-4f71-81b6-1c43fcf5f556","Type":"ContainerDied","Data":"527d12e90bd8f5cf8ef2baf875cc761d6d71ff609b96d3fc24245a28d60416ab"} Dec 13 07:03:12 crc kubenswrapper[4971]: I1213 07:03:12.827797 4971 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"6afaff46af01dcc99e08b76967fe4181e615cd044389b9b8cea1581505eef247"} Dec 13 07:03:12 crc kubenswrapper[4971]: I1213 07:03:12.827804 4971 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"7dd9913a65acba73ae1f32f4398207c5e276a1e979d86f9e720e4fd761bae4a3"} Dec 13 07:03:12 crc kubenswrapper[4971]: I1213 07:03:12.827810 4971 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"868049c6a3a4ebe4036d1d22e02ba6b76bced2ee7e269bb228accc4a0636678a"} Dec 13 07:03:12 crc kubenswrapper[4971]: I1213 07:03:12.827815 4971 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"f6073140d9f86929b6d4bd2960ede0617a5c9620b269bbab043bae37e0d389f5"} Dec 13 07:03:12 crc kubenswrapper[4971]: I1213 07:03:12.827821 4971 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"693a2ab434c250cb140a866e8068b156df9191414fed1d63bee919a2528ea6e5"} Dec 13 07:03:12 crc kubenswrapper[4971]: I1213 07:03:12.827826 4971 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"f22a4ec756119c4f791fb5105cf1f31df3e548b2b07b1f909a62802e9b6035ae"} Dec 13 07:03:12 crc kubenswrapper[4971]: I1213 07:03:12.827831 4971 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"325223543b3c7db417f769533559a1ca2b23a723494e1885b0fd0480b5a6399f"} Dec 13 07:03:12 crc kubenswrapper[4971]: I1213 07:03:12.827836 4971 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"219620f51474e60b3e1f880c0972eb612dffc4bad24462bbba2c093ae817d852"} Dec 13 07:03:12 crc kubenswrapper[4971]: I1213 07:03:12.827842 4971 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"f137f4b441c2402093ab430a9689a46221c4fb4bc70d99fb992f62b0a989c465"} Dec 13 07:03:12 crc kubenswrapper[4971]: I1213 07:03:12.827851 4971 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"451b22df6973ddc8fa2209a65d64d9e41b33eb231bb4adfab88cf0e5943bb246"} Dec 13 07:03:12 crc kubenswrapper[4971]: I1213 07:03:12.828064 4971 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-ovn-kubernetes/ovnkube-node-zlcg5" Dec 13 07:03:12 crc kubenswrapper[4971]: I1213 07:03:12.855619 4971 scope.go:117] "RemoveContainer" containerID="6afaff46af01dcc99e08b76967fe4181e615cd044389b9b8cea1581505eef247" Dec 13 07:03:12 crc kubenswrapper[4971]: I1213 07:03:12.866221 4971 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-ovn-kubernetes/ovnkube-node-zlcg5"] Dec 13 07:03:12 crc kubenswrapper[4971]: I1213 07:03:12.872436 4971 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-ovn-kubernetes/ovnkube-node-zlcg5"] Dec 13 07:03:12 crc kubenswrapper[4971]: I1213 07:03:12.883890 4971 scope.go:117] "RemoveContainer" containerID="7dd9913a65acba73ae1f32f4398207c5e276a1e979d86f9e720e4fd761bae4a3" Dec 13 07:03:12 crc kubenswrapper[4971]: I1213 07:03:12.901989 4971 scope.go:117] "RemoveContainer" containerID="868049c6a3a4ebe4036d1d22e02ba6b76bced2ee7e269bb228accc4a0636678a" Dec 13 07:03:12 crc kubenswrapper[4971]: I1213 07:03:12.917090 4971 scope.go:117] "RemoveContainer" containerID="f6073140d9f86929b6d4bd2960ede0617a5c9620b269bbab043bae37e0d389f5" Dec 13 07:03:12 crc kubenswrapper[4971]: I1213 07:03:12.932810 4971 scope.go:117] "RemoveContainer" containerID="693a2ab434c250cb140a866e8068b156df9191414fed1d63bee919a2528ea6e5" Dec 13 07:03:12 crc kubenswrapper[4971]: I1213 07:03:12.946267 4971 scope.go:117] "RemoveContainer" containerID="f22a4ec756119c4f791fb5105cf1f31df3e548b2b07b1f909a62802e9b6035ae" Dec 13 07:03:12 crc kubenswrapper[4971]: I1213 07:03:12.965323 4971 scope.go:117] "RemoveContainer" containerID="325223543b3c7db417f769533559a1ca2b23a723494e1885b0fd0480b5a6399f" Dec 13 07:03:13 crc kubenswrapper[4971]: I1213 07:03:13.023028 4971 scope.go:117] "RemoveContainer" containerID="219620f51474e60b3e1f880c0972eb612dffc4bad24462bbba2c093ae817d852" Dec 13 07:03:13 crc kubenswrapper[4971]: I1213 07:03:13.035669 4971 scope.go:117] "RemoveContainer" containerID="f137f4b441c2402093ab430a9689a46221c4fb4bc70d99fb992f62b0a989c465" Dec 13 07:03:13 crc kubenswrapper[4971]: I1213 07:03:13.051656 4971 scope.go:117] "RemoveContainer" containerID="451b22df6973ddc8fa2209a65d64d9e41b33eb231bb4adfab88cf0e5943bb246" Dec 13 07:03:13 crc kubenswrapper[4971]: I1213 07:03:13.063324 4971 scope.go:117] "RemoveContainer" containerID="6afaff46af01dcc99e08b76967fe4181e615cd044389b9b8cea1581505eef247" Dec 13 07:03:13 crc kubenswrapper[4971]: E1213 07:03:13.063772 4971 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"6afaff46af01dcc99e08b76967fe4181e615cd044389b9b8cea1581505eef247\": container with ID starting with 6afaff46af01dcc99e08b76967fe4181e615cd044389b9b8cea1581505eef247 not found: ID does not exist" containerID="6afaff46af01dcc99e08b76967fe4181e615cd044389b9b8cea1581505eef247" Dec 13 07:03:13 crc kubenswrapper[4971]: I1213 07:03:13.063816 4971 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"6afaff46af01dcc99e08b76967fe4181e615cd044389b9b8cea1581505eef247"} err="failed to get container status \"6afaff46af01dcc99e08b76967fe4181e615cd044389b9b8cea1581505eef247\": rpc error: code = NotFound desc = could not find container \"6afaff46af01dcc99e08b76967fe4181e615cd044389b9b8cea1581505eef247\": container with ID starting with 6afaff46af01dcc99e08b76967fe4181e615cd044389b9b8cea1581505eef247 not found: ID does not exist" Dec 13 07:03:13 crc kubenswrapper[4971]: I1213 07:03:13.063848 4971 scope.go:117] "RemoveContainer" containerID="7dd9913a65acba73ae1f32f4398207c5e276a1e979d86f9e720e4fd761bae4a3" Dec 13 07:03:13 crc kubenswrapper[4971]: E1213 07:03:13.064179 4971 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"7dd9913a65acba73ae1f32f4398207c5e276a1e979d86f9e720e4fd761bae4a3\": container with ID starting with 7dd9913a65acba73ae1f32f4398207c5e276a1e979d86f9e720e4fd761bae4a3 not found: ID does not exist" containerID="7dd9913a65acba73ae1f32f4398207c5e276a1e979d86f9e720e4fd761bae4a3" Dec 13 07:03:13 crc kubenswrapper[4971]: I1213 07:03:13.064209 4971 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"7dd9913a65acba73ae1f32f4398207c5e276a1e979d86f9e720e4fd761bae4a3"} err="failed to get container status \"7dd9913a65acba73ae1f32f4398207c5e276a1e979d86f9e720e4fd761bae4a3\": rpc error: code = NotFound desc = could not find container \"7dd9913a65acba73ae1f32f4398207c5e276a1e979d86f9e720e4fd761bae4a3\": container with ID starting with 7dd9913a65acba73ae1f32f4398207c5e276a1e979d86f9e720e4fd761bae4a3 not found: ID does not exist" Dec 13 07:03:13 crc kubenswrapper[4971]: I1213 07:03:13.064230 4971 scope.go:117] "RemoveContainer" containerID="868049c6a3a4ebe4036d1d22e02ba6b76bced2ee7e269bb228accc4a0636678a" Dec 13 07:03:13 crc kubenswrapper[4971]: E1213 07:03:13.064620 4971 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"868049c6a3a4ebe4036d1d22e02ba6b76bced2ee7e269bb228accc4a0636678a\": container with ID starting with 868049c6a3a4ebe4036d1d22e02ba6b76bced2ee7e269bb228accc4a0636678a not found: ID does not exist" containerID="868049c6a3a4ebe4036d1d22e02ba6b76bced2ee7e269bb228accc4a0636678a" Dec 13 07:03:13 crc kubenswrapper[4971]: I1213 07:03:13.064676 4971 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"868049c6a3a4ebe4036d1d22e02ba6b76bced2ee7e269bb228accc4a0636678a"} err="failed to get container status \"868049c6a3a4ebe4036d1d22e02ba6b76bced2ee7e269bb228accc4a0636678a\": rpc error: code = NotFound desc = could not find container \"868049c6a3a4ebe4036d1d22e02ba6b76bced2ee7e269bb228accc4a0636678a\": container with ID starting with 868049c6a3a4ebe4036d1d22e02ba6b76bced2ee7e269bb228accc4a0636678a not found: ID does not exist" Dec 13 07:03:13 crc kubenswrapper[4971]: I1213 07:03:13.064709 4971 scope.go:117] "RemoveContainer" containerID="f6073140d9f86929b6d4bd2960ede0617a5c9620b269bbab043bae37e0d389f5" Dec 13 07:03:13 crc kubenswrapper[4971]: E1213 07:03:13.065008 4971 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"f6073140d9f86929b6d4bd2960ede0617a5c9620b269bbab043bae37e0d389f5\": container with ID starting with f6073140d9f86929b6d4bd2960ede0617a5c9620b269bbab043bae37e0d389f5 not found: ID does not exist" containerID="f6073140d9f86929b6d4bd2960ede0617a5c9620b269bbab043bae37e0d389f5" Dec 13 07:03:13 crc kubenswrapper[4971]: I1213 07:03:13.065040 4971 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"f6073140d9f86929b6d4bd2960ede0617a5c9620b269bbab043bae37e0d389f5"} err="failed to get container status \"f6073140d9f86929b6d4bd2960ede0617a5c9620b269bbab043bae37e0d389f5\": rpc error: code = NotFound desc = could not find container \"f6073140d9f86929b6d4bd2960ede0617a5c9620b269bbab043bae37e0d389f5\": container with ID starting with f6073140d9f86929b6d4bd2960ede0617a5c9620b269bbab043bae37e0d389f5 not found: ID does not exist" Dec 13 07:03:13 crc kubenswrapper[4971]: I1213 07:03:13.065060 4971 scope.go:117] "RemoveContainer" containerID="693a2ab434c250cb140a866e8068b156df9191414fed1d63bee919a2528ea6e5" Dec 13 07:03:13 crc kubenswrapper[4971]: E1213 07:03:13.065307 4971 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"693a2ab434c250cb140a866e8068b156df9191414fed1d63bee919a2528ea6e5\": container with ID starting with 693a2ab434c250cb140a866e8068b156df9191414fed1d63bee919a2528ea6e5 not found: ID does not exist" containerID="693a2ab434c250cb140a866e8068b156df9191414fed1d63bee919a2528ea6e5" Dec 13 07:03:13 crc kubenswrapper[4971]: I1213 07:03:13.065337 4971 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"693a2ab434c250cb140a866e8068b156df9191414fed1d63bee919a2528ea6e5"} err="failed to get container status \"693a2ab434c250cb140a866e8068b156df9191414fed1d63bee919a2528ea6e5\": rpc error: code = NotFound desc = could not find container \"693a2ab434c250cb140a866e8068b156df9191414fed1d63bee919a2528ea6e5\": container with ID starting with 693a2ab434c250cb140a866e8068b156df9191414fed1d63bee919a2528ea6e5 not found: ID does not exist" Dec 13 07:03:13 crc kubenswrapper[4971]: I1213 07:03:13.065365 4971 scope.go:117] "RemoveContainer" containerID="f22a4ec756119c4f791fb5105cf1f31df3e548b2b07b1f909a62802e9b6035ae" Dec 13 07:03:13 crc kubenswrapper[4971]: E1213 07:03:13.065649 4971 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"f22a4ec756119c4f791fb5105cf1f31df3e548b2b07b1f909a62802e9b6035ae\": container with ID starting with f22a4ec756119c4f791fb5105cf1f31df3e548b2b07b1f909a62802e9b6035ae not found: ID does not exist" containerID="f22a4ec756119c4f791fb5105cf1f31df3e548b2b07b1f909a62802e9b6035ae" Dec 13 07:03:13 crc kubenswrapper[4971]: I1213 07:03:13.065675 4971 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"f22a4ec756119c4f791fb5105cf1f31df3e548b2b07b1f909a62802e9b6035ae"} err="failed to get container status \"f22a4ec756119c4f791fb5105cf1f31df3e548b2b07b1f909a62802e9b6035ae\": rpc error: code = NotFound desc = could not find container \"f22a4ec756119c4f791fb5105cf1f31df3e548b2b07b1f909a62802e9b6035ae\": container with ID starting with f22a4ec756119c4f791fb5105cf1f31df3e548b2b07b1f909a62802e9b6035ae not found: ID does not exist" Dec 13 07:03:13 crc kubenswrapper[4971]: I1213 07:03:13.065692 4971 scope.go:117] "RemoveContainer" containerID="325223543b3c7db417f769533559a1ca2b23a723494e1885b0fd0480b5a6399f" Dec 13 07:03:13 crc kubenswrapper[4971]: E1213 07:03:13.065924 4971 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"325223543b3c7db417f769533559a1ca2b23a723494e1885b0fd0480b5a6399f\": container with ID starting with 325223543b3c7db417f769533559a1ca2b23a723494e1885b0fd0480b5a6399f not found: ID does not exist" containerID="325223543b3c7db417f769533559a1ca2b23a723494e1885b0fd0480b5a6399f" Dec 13 07:03:13 crc kubenswrapper[4971]: I1213 07:03:13.065952 4971 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"325223543b3c7db417f769533559a1ca2b23a723494e1885b0fd0480b5a6399f"} err="failed to get container status \"325223543b3c7db417f769533559a1ca2b23a723494e1885b0fd0480b5a6399f\": rpc error: code = NotFound desc = could not find container \"325223543b3c7db417f769533559a1ca2b23a723494e1885b0fd0480b5a6399f\": container with ID starting with 325223543b3c7db417f769533559a1ca2b23a723494e1885b0fd0480b5a6399f not found: ID does not exist" Dec 13 07:03:13 crc kubenswrapper[4971]: I1213 07:03:13.065970 4971 scope.go:117] "RemoveContainer" containerID="219620f51474e60b3e1f880c0972eb612dffc4bad24462bbba2c093ae817d852" Dec 13 07:03:13 crc kubenswrapper[4971]: E1213 07:03:13.066250 4971 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"219620f51474e60b3e1f880c0972eb612dffc4bad24462bbba2c093ae817d852\": container with ID starting with 219620f51474e60b3e1f880c0972eb612dffc4bad24462bbba2c093ae817d852 not found: ID does not exist" containerID="219620f51474e60b3e1f880c0972eb612dffc4bad24462bbba2c093ae817d852" Dec 13 07:03:13 crc kubenswrapper[4971]: I1213 07:03:13.066272 4971 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"219620f51474e60b3e1f880c0972eb612dffc4bad24462bbba2c093ae817d852"} err="failed to get container status \"219620f51474e60b3e1f880c0972eb612dffc4bad24462bbba2c093ae817d852\": rpc error: code = NotFound desc = could not find container \"219620f51474e60b3e1f880c0972eb612dffc4bad24462bbba2c093ae817d852\": container with ID starting with 219620f51474e60b3e1f880c0972eb612dffc4bad24462bbba2c093ae817d852 not found: ID does not exist" Dec 13 07:03:13 crc kubenswrapper[4971]: I1213 07:03:13.066286 4971 scope.go:117] "RemoveContainer" containerID="f137f4b441c2402093ab430a9689a46221c4fb4bc70d99fb992f62b0a989c465" Dec 13 07:03:13 crc kubenswrapper[4971]: E1213 07:03:13.066488 4971 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"f137f4b441c2402093ab430a9689a46221c4fb4bc70d99fb992f62b0a989c465\": container with ID starting with f137f4b441c2402093ab430a9689a46221c4fb4bc70d99fb992f62b0a989c465 not found: ID does not exist" containerID="f137f4b441c2402093ab430a9689a46221c4fb4bc70d99fb992f62b0a989c465" Dec 13 07:03:13 crc kubenswrapper[4971]: I1213 07:03:13.066507 4971 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"f137f4b441c2402093ab430a9689a46221c4fb4bc70d99fb992f62b0a989c465"} err="failed to get container status \"f137f4b441c2402093ab430a9689a46221c4fb4bc70d99fb992f62b0a989c465\": rpc error: code = NotFound desc = could not find container \"f137f4b441c2402093ab430a9689a46221c4fb4bc70d99fb992f62b0a989c465\": container with ID starting with f137f4b441c2402093ab430a9689a46221c4fb4bc70d99fb992f62b0a989c465 not found: ID does not exist" Dec 13 07:03:13 crc kubenswrapper[4971]: I1213 07:03:13.066550 4971 scope.go:117] "RemoveContainer" containerID="451b22df6973ddc8fa2209a65d64d9e41b33eb231bb4adfab88cf0e5943bb246" Dec 13 07:03:13 crc kubenswrapper[4971]: E1213 07:03:13.066787 4971 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"451b22df6973ddc8fa2209a65d64d9e41b33eb231bb4adfab88cf0e5943bb246\": container with ID starting with 451b22df6973ddc8fa2209a65d64d9e41b33eb231bb4adfab88cf0e5943bb246 not found: ID does not exist" containerID="451b22df6973ddc8fa2209a65d64d9e41b33eb231bb4adfab88cf0e5943bb246" Dec 13 07:03:13 crc kubenswrapper[4971]: I1213 07:03:13.066805 4971 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"451b22df6973ddc8fa2209a65d64d9e41b33eb231bb4adfab88cf0e5943bb246"} err="failed to get container status \"451b22df6973ddc8fa2209a65d64d9e41b33eb231bb4adfab88cf0e5943bb246\": rpc error: code = NotFound desc = could not find container \"451b22df6973ddc8fa2209a65d64d9e41b33eb231bb4adfab88cf0e5943bb246\": container with ID starting with 451b22df6973ddc8fa2209a65d64d9e41b33eb231bb4adfab88cf0e5943bb246 not found: ID does not exist" Dec 13 07:03:13 crc kubenswrapper[4971]: I1213 07:03:13.066819 4971 scope.go:117] "RemoveContainer" containerID="6afaff46af01dcc99e08b76967fe4181e615cd044389b9b8cea1581505eef247" Dec 13 07:03:13 crc kubenswrapper[4971]: I1213 07:03:13.067014 4971 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"6afaff46af01dcc99e08b76967fe4181e615cd044389b9b8cea1581505eef247"} err="failed to get container status \"6afaff46af01dcc99e08b76967fe4181e615cd044389b9b8cea1581505eef247\": rpc error: code = NotFound desc = could not find container \"6afaff46af01dcc99e08b76967fe4181e615cd044389b9b8cea1581505eef247\": container with ID starting with 6afaff46af01dcc99e08b76967fe4181e615cd044389b9b8cea1581505eef247 not found: ID does not exist" Dec 13 07:03:13 crc kubenswrapper[4971]: I1213 07:03:13.067039 4971 scope.go:117] "RemoveContainer" containerID="7dd9913a65acba73ae1f32f4398207c5e276a1e979d86f9e720e4fd761bae4a3" Dec 13 07:03:13 crc kubenswrapper[4971]: I1213 07:03:13.067250 4971 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"7dd9913a65acba73ae1f32f4398207c5e276a1e979d86f9e720e4fd761bae4a3"} err="failed to get container status \"7dd9913a65acba73ae1f32f4398207c5e276a1e979d86f9e720e4fd761bae4a3\": rpc error: code = NotFound desc = could not find container \"7dd9913a65acba73ae1f32f4398207c5e276a1e979d86f9e720e4fd761bae4a3\": container with ID starting with 7dd9913a65acba73ae1f32f4398207c5e276a1e979d86f9e720e4fd761bae4a3 not found: ID does not exist" Dec 13 07:03:13 crc kubenswrapper[4971]: I1213 07:03:13.067270 4971 scope.go:117] "RemoveContainer" containerID="868049c6a3a4ebe4036d1d22e02ba6b76bced2ee7e269bb228accc4a0636678a" Dec 13 07:03:13 crc kubenswrapper[4971]: I1213 07:03:13.067431 4971 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"868049c6a3a4ebe4036d1d22e02ba6b76bced2ee7e269bb228accc4a0636678a"} err="failed to get container status \"868049c6a3a4ebe4036d1d22e02ba6b76bced2ee7e269bb228accc4a0636678a\": rpc error: code = NotFound desc = could not find container \"868049c6a3a4ebe4036d1d22e02ba6b76bced2ee7e269bb228accc4a0636678a\": container with ID starting with 868049c6a3a4ebe4036d1d22e02ba6b76bced2ee7e269bb228accc4a0636678a not found: ID does not exist" Dec 13 07:03:13 crc kubenswrapper[4971]: I1213 07:03:13.067449 4971 scope.go:117] "RemoveContainer" containerID="f6073140d9f86929b6d4bd2960ede0617a5c9620b269bbab043bae37e0d389f5" Dec 13 07:03:13 crc kubenswrapper[4971]: I1213 07:03:13.067666 4971 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"f6073140d9f86929b6d4bd2960ede0617a5c9620b269bbab043bae37e0d389f5"} err="failed to get container status \"f6073140d9f86929b6d4bd2960ede0617a5c9620b269bbab043bae37e0d389f5\": rpc error: code = NotFound desc = could not find container \"f6073140d9f86929b6d4bd2960ede0617a5c9620b269bbab043bae37e0d389f5\": container with ID starting with f6073140d9f86929b6d4bd2960ede0617a5c9620b269bbab043bae37e0d389f5 not found: ID does not exist" Dec 13 07:03:13 crc kubenswrapper[4971]: I1213 07:03:13.067684 4971 scope.go:117] "RemoveContainer" containerID="693a2ab434c250cb140a866e8068b156df9191414fed1d63bee919a2528ea6e5" Dec 13 07:03:13 crc kubenswrapper[4971]: I1213 07:03:13.067851 4971 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"693a2ab434c250cb140a866e8068b156df9191414fed1d63bee919a2528ea6e5"} err="failed to get container status \"693a2ab434c250cb140a866e8068b156df9191414fed1d63bee919a2528ea6e5\": rpc error: code = NotFound desc = could not find container \"693a2ab434c250cb140a866e8068b156df9191414fed1d63bee919a2528ea6e5\": container with ID starting with 693a2ab434c250cb140a866e8068b156df9191414fed1d63bee919a2528ea6e5 not found: ID does not exist" Dec 13 07:03:13 crc kubenswrapper[4971]: I1213 07:03:13.067870 4971 scope.go:117] "RemoveContainer" containerID="f22a4ec756119c4f791fb5105cf1f31df3e548b2b07b1f909a62802e9b6035ae" Dec 13 07:03:13 crc kubenswrapper[4971]: I1213 07:03:13.068008 4971 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"f22a4ec756119c4f791fb5105cf1f31df3e548b2b07b1f909a62802e9b6035ae"} err="failed to get container status \"f22a4ec756119c4f791fb5105cf1f31df3e548b2b07b1f909a62802e9b6035ae\": rpc error: code = NotFound desc = could not find container \"f22a4ec756119c4f791fb5105cf1f31df3e548b2b07b1f909a62802e9b6035ae\": container with ID starting with f22a4ec756119c4f791fb5105cf1f31df3e548b2b07b1f909a62802e9b6035ae not found: ID does not exist" Dec 13 07:03:13 crc kubenswrapper[4971]: I1213 07:03:13.068030 4971 scope.go:117] "RemoveContainer" containerID="325223543b3c7db417f769533559a1ca2b23a723494e1885b0fd0480b5a6399f" Dec 13 07:03:13 crc kubenswrapper[4971]: I1213 07:03:13.068342 4971 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"325223543b3c7db417f769533559a1ca2b23a723494e1885b0fd0480b5a6399f"} err="failed to get container status \"325223543b3c7db417f769533559a1ca2b23a723494e1885b0fd0480b5a6399f\": rpc error: code = NotFound desc = could not find container \"325223543b3c7db417f769533559a1ca2b23a723494e1885b0fd0480b5a6399f\": container with ID starting with 325223543b3c7db417f769533559a1ca2b23a723494e1885b0fd0480b5a6399f not found: ID does not exist" Dec 13 07:03:13 crc kubenswrapper[4971]: I1213 07:03:13.068361 4971 scope.go:117] "RemoveContainer" containerID="219620f51474e60b3e1f880c0972eb612dffc4bad24462bbba2c093ae817d852" Dec 13 07:03:13 crc kubenswrapper[4971]: I1213 07:03:13.068547 4971 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"219620f51474e60b3e1f880c0972eb612dffc4bad24462bbba2c093ae817d852"} err="failed to get container status \"219620f51474e60b3e1f880c0972eb612dffc4bad24462bbba2c093ae817d852\": rpc error: code = NotFound desc = could not find container \"219620f51474e60b3e1f880c0972eb612dffc4bad24462bbba2c093ae817d852\": container with ID starting with 219620f51474e60b3e1f880c0972eb612dffc4bad24462bbba2c093ae817d852 not found: ID does not exist" Dec 13 07:03:13 crc kubenswrapper[4971]: I1213 07:03:13.068567 4971 scope.go:117] "RemoveContainer" containerID="f137f4b441c2402093ab430a9689a46221c4fb4bc70d99fb992f62b0a989c465" Dec 13 07:03:13 crc kubenswrapper[4971]: I1213 07:03:13.068835 4971 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"f137f4b441c2402093ab430a9689a46221c4fb4bc70d99fb992f62b0a989c465"} err="failed to get container status \"f137f4b441c2402093ab430a9689a46221c4fb4bc70d99fb992f62b0a989c465\": rpc error: code = NotFound desc = could not find container \"f137f4b441c2402093ab430a9689a46221c4fb4bc70d99fb992f62b0a989c465\": container with ID starting with f137f4b441c2402093ab430a9689a46221c4fb4bc70d99fb992f62b0a989c465 not found: ID does not exist" Dec 13 07:03:13 crc kubenswrapper[4971]: I1213 07:03:13.068854 4971 scope.go:117] "RemoveContainer" containerID="451b22df6973ddc8fa2209a65d64d9e41b33eb231bb4adfab88cf0e5943bb246" Dec 13 07:03:13 crc kubenswrapper[4971]: I1213 07:03:13.069040 4971 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"451b22df6973ddc8fa2209a65d64d9e41b33eb231bb4adfab88cf0e5943bb246"} err="failed to get container status \"451b22df6973ddc8fa2209a65d64d9e41b33eb231bb4adfab88cf0e5943bb246\": rpc error: code = NotFound desc = could not find container \"451b22df6973ddc8fa2209a65d64d9e41b33eb231bb4adfab88cf0e5943bb246\": container with ID starting with 451b22df6973ddc8fa2209a65d64d9e41b33eb231bb4adfab88cf0e5943bb246 not found: ID does not exist" Dec 13 07:03:13 crc kubenswrapper[4971]: I1213 07:03:13.069057 4971 scope.go:117] "RemoveContainer" containerID="6afaff46af01dcc99e08b76967fe4181e615cd044389b9b8cea1581505eef247" Dec 13 07:03:13 crc kubenswrapper[4971]: I1213 07:03:13.069349 4971 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"6afaff46af01dcc99e08b76967fe4181e615cd044389b9b8cea1581505eef247"} err="failed to get container status \"6afaff46af01dcc99e08b76967fe4181e615cd044389b9b8cea1581505eef247\": rpc error: code = NotFound desc = could not find container \"6afaff46af01dcc99e08b76967fe4181e615cd044389b9b8cea1581505eef247\": container with ID starting with 6afaff46af01dcc99e08b76967fe4181e615cd044389b9b8cea1581505eef247 not found: ID does not exist" Dec 13 07:03:13 crc kubenswrapper[4971]: I1213 07:03:13.069369 4971 scope.go:117] "RemoveContainer" containerID="7dd9913a65acba73ae1f32f4398207c5e276a1e979d86f9e720e4fd761bae4a3" Dec 13 07:03:13 crc kubenswrapper[4971]: I1213 07:03:13.069607 4971 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"7dd9913a65acba73ae1f32f4398207c5e276a1e979d86f9e720e4fd761bae4a3"} err="failed to get container status \"7dd9913a65acba73ae1f32f4398207c5e276a1e979d86f9e720e4fd761bae4a3\": rpc error: code = NotFound desc = could not find container \"7dd9913a65acba73ae1f32f4398207c5e276a1e979d86f9e720e4fd761bae4a3\": container with ID starting with 7dd9913a65acba73ae1f32f4398207c5e276a1e979d86f9e720e4fd761bae4a3 not found: ID does not exist" Dec 13 07:03:13 crc kubenswrapper[4971]: I1213 07:03:13.069626 4971 scope.go:117] "RemoveContainer" containerID="868049c6a3a4ebe4036d1d22e02ba6b76bced2ee7e269bb228accc4a0636678a" Dec 13 07:03:13 crc kubenswrapper[4971]: I1213 07:03:13.069859 4971 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"868049c6a3a4ebe4036d1d22e02ba6b76bced2ee7e269bb228accc4a0636678a"} err="failed to get container status \"868049c6a3a4ebe4036d1d22e02ba6b76bced2ee7e269bb228accc4a0636678a\": rpc error: code = NotFound desc = could not find container \"868049c6a3a4ebe4036d1d22e02ba6b76bced2ee7e269bb228accc4a0636678a\": container with ID starting with 868049c6a3a4ebe4036d1d22e02ba6b76bced2ee7e269bb228accc4a0636678a not found: ID does not exist" Dec 13 07:03:13 crc kubenswrapper[4971]: I1213 07:03:13.069876 4971 scope.go:117] "RemoveContainer" containerID="f6073140d9f86929b6d4bd2960ede0617a5c9620b269bbab043bae37e0d389f5" Dec 13 07:03:13 crc kubenswrapper[4971]: I1213 07:03:13.070068 4971 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"f6073140d9f86929b6d4bd2960ede0617a5c9620b269bbab043bae37e0d389f5"} err="failed to get container status \"f6073140d9f86929b6d4bd2960ede0617a5c9620b269bbab043bae37e0d389f5\": rpc error: code = NotFound desc = could not find container \"f6073140d9f86929b6d4bd2960ede0617a5c9620b269bbab043bae37e0d389f5\": container with ID starting with f6073140d9f86929b6d4bd2960ede0617a5c9620b269bbab043bae37e0d389f5 not found: ID does not exist" Dec 13 07:03:13 crc kubenswrapper[4971]: I1213 07:03:13.070084 4971 scope.go:117] "RemoveContainer" containerID="693a2ab434c250cb140a866e8068b156df9191414fed1d63bee919a2528ea6e5" Dec 13 07:03:13 crc kubenswrapper[4971]: I1213 07:03:13.070264 4971 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"693a2ab434c250cb140a866e8068b156df9191414fed1d63bee919a2528ea6e5"} err="failed to get container status \"693a2ab434c250cb140a866e8068b156df9191414fed1d63bee919a2528ea6e5\": rpc error: code = NotFound desc = could not find container \"693a2ab434c250cb140a866e8068b156df9191414fed1d63bee919a2528ea6e5\": container with ID starting with 693a2ab434c250cb140a866e8068b156df9191414fed1d63bee919a2528ea6e5 not found: ID does not exist" Dec 13 07:03:13 crc kubenswrapper[4971]: I1213 07:03:13.070280 4971 scope.go:117] "RemoveContainer" containerID="f22a4ec756119c4f791fb5105cf1f31df3e548b2b07b1f909a62802e9b6035ae" Dec 13 07:03:13 crc kubenswrapper[4971]: I1213 07:03:13.070439 4971 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"f22a4ec756119c4f791fb5105cf1f31df3e548b2b07b1f909a62802e9b6035ae"} err="failed to get container status \"f22a4ec756119c4f791fb5105cf1f31df3e548b2b07b1f909a62802e9b6035ae\": rpc error: code = NotFound desc = could not find container \"f22a4ec756119c4f791fb5105cf1f31df3e548b2b07b1f909a62802e9b6035ae\": container with ID starting with f22a4ec756119c4f791fb5105cf1f31df3e548b2b07b1f909a62802e9b6035ae not found: ID does not exist" Dec 13 07:03:13 crc kubenswrapper[4971]: I1213 07:03:13.070455 4971 scope.go:117] "RemoveContainer" containerID="325223543b3c7db417f769533559a1ca2b23a723494e1885b0fd0480b5a6399f" Dec 13 07:03:13 crc kubenswrapper[4971]: I1213 07:03:13.070730 4971 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"325223543b3c7db417f769533559a1ca2b23a723494e1885b0fd0480b5a6399f"} err="failed to get container status \"325223543b3c7db417f769533559a1ca2b23a723494e1885b0fd0480b5a6399f\": rpc error: code = NotFound desc = could not find container \"325223543b3c7db417f769533559a1ca2b23a723494e1885b0fd0480b5a6399f\": container with ID starting with 325223543b3c7db417f769533559a1ca2b23a723494e1885b0fd0480b5a6399f not found: ID does not exist" Dec 13 07:03:13 crc kubenswrapper[4971]: I1213 07:03:13.070747 4971 scope.go:117] "RemoveContainer" containerID="219620f51474e60b3e1f880c0972eb612dffc4bad24462bbba2c093ae817d852" Dec 13 07:03:13 crc kubenswrapper[4971]: I1213 07:03:13.070877 4971 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"219620f51474e60b3e1f880c0972eb612dffc4bad24462bbba2c093ae817d852"} err="failed to get container status \"219620f51474e60b3e1f880c0972eb612dffc4bad24462bbba2c093ae817d852\": rpc error: code = NotFound desc = could not find container \"219620f51474e60b3e1f880c0972eb612dffc4bad24462bbba2c093ae817d852\": container with ID starting with 219620f51474e60b3e1f880c0972eb612dffc4bad24462bbba2c093ae817d852 not found: ID does not exist" Dec 13 07:03:13 crc kubenswrapper[4971]: I1213 07:03:13.070895 4971 scope.go:117] "RemoveContainer" containerID="f137f4b441c2402093ab430a9689a46221c4fb4bc70d99fb992f62b0a989c465" Dec 13 07:03:13 crc kubenswrapper[4971]: I1213 07:03:13.071091 4971 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"f137f4b441c2402093ab430a9689a46221c4fb4bc70d99fb992f62b0a989c465"} err="failed to get container status \"f137f4b441c2402093ab430a9689a46221c4fb4bc70d99fb992f62b0a989c465\": rpc error: code = NotFound desc = could not find container \"f137f4b441c2402093ab430a9689a46221c4fb4bc70d99fb992f62b0a989c465\": container with ID starting with f137f4b441c2402093ab430a9689a46221c4fb4bc70d99fb992f62b0a989c465 not found: ID does not exist" Dec 13 07:03:13 crc kubenswrapper[4971]: I1213 07:03:13.071112 4971 scope.go:117] "RemoveContainer" containerID="451b22df6973ddc8fa2209a65d64d9e41b33eb231bb4adfab88cf0e5943bb246" Dec 13 07:03:13 crc kubenswrapper[4971]: I1213 07:03:13.071337 4971 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"451b22df6973ddc8fa2209a65d64d9e41b33eb231bb4adfab88cf0e5943bb246"} err="failed to get container status \"451b22df6973ddc8fa2209a65d64d9e41b33eb231bb4adfab88cf0e5943bb246\": rpc error: code = NotFound desc = could not find container \"451b22df6973ddc8fa2209a65d64d9e41b33eb231bb4adfab88cf0e5943bb246\": container with ID starting with 451b22df6973ddc8fa2209a65d64d9e41b33eb231bb4adfab88cf0e5943bb246 not found: ID does not exist" Dec 13 07:03:13 crc kubenswrapper[4971]: I1213 07:03:13.071357 4971 scope.go:117] "RemoveContainer" containerID="6afaff46af01dcc99e08b76967fe4181e615cd044389b9b8cea1581505eef247" Dec 13 07:03:13 crc kubenswrapper[4971]: I1213 07:03:13.071581 4971 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"6afaff46af01dcc99e08b76967fe4181e615cd044389b9b8cea1581505eef247"} err="failed to get container status \"6afaff46af01dcc99e08b76967fe4181e615cd044389b9b8cea1581505eef247\": rpc error: code = NotFound desc = could not find container \"6afaff46af01dcc99e08b76967fe4181e615cd044389b9b8cea1581505eef247\": container with ID starting with 6afaff46af01dcc99e08b76967fe4181e615cd044389b9b8cea1581505eef247 not found: ID does not exist" Dec 13 07:03:13 crc kubenswrapper[4971]: I1213 07:03:13.071607 4971 scope.go:117] "RemoveContainer" containerID="7dd9913a65acba73ae1f32f4398207c5e276a1e979d86f9e720e4fd761bae4a3" Dec 13 07:03:13 crc kubenswrapper[4971]: I1213 07:03:13.071947 4971 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"7dd9913a65acba73ae1f32f4398207c5e276a1e979d86f9e720e4fd761bae4a3"} err="failed to get container status \"7dd9913a65acba73ae1f32f4398207c5e276a1e979d86f9e720e4fd761bae4a3\": rpc error: code = NotFound desc = could not find container \"7dd9913a65acba73ae1f32f4398207c5e276a1e979d86f9e720e4fd761bae4a3\": container with ID starting with 7dd9913a65acba73ae1f32f4398207c5e276a1e979d86f9e720e4fd761bae4a3 not found: ID does not exist" Dec 13 07:03:13 crc kubenswrapper[4971]: I1213 07:03:13.071973 4971 scope.go:117] "RemoveContainer" containerID="868049c6a3a4ebe4036d1d22e02ba6b76bced2ee7e269bb228accc4a0636678a" Dec 13 07:03:13 crc kubenswrapper[4971]: I1213 07:03:13.072163 4971 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"868049c6a3a4ebe4036d1d22e02ba6b76bced2ee7e269bb228accc4a0636678a"} err="failed to get container status \"868049c6a3a4ebe4036d1d22e02ba6b76bced2ee7e269bb228accc4a0636678a\": rpc error: code = NotFound desc = could not find container \"868049c6a3a4ebe4036d1d22e02ba6b76bced2ee7e269bb228accc4a0636678a\": container with ID starting with 868049c6a3a4ebe4036d1d22e02ba6b76bced2ee7e269bb228accc4a0636678a not found: ID does not exist" Dec 13 07:03:13 crc kubenswrapper[4971]: I1213 07:03:13.072186 4971 scope.go:117] "RemoveContainer" containerID="f6073140d9f86929b6d4bd2960ede0617a5c9620b269bbab043bae37e0d389f5" Dec 13 07:03:13 crc kubenswrapper[4971]: I1213 07:03:13.072388 4971 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"f6073140d9f86929b6d4bd2960ede0617a5c9620b269bbab043bae37e0d389f5"} err="failed to get container status \"f6073140d9f86929b6d4bd2960ede0617a5c9620b269bbab043bae37e0d389f5\": rpc error: code = NotFound desc = could not find container \"f6073140d9f86929b6d4bd2960ede0617a5c9620b269bbab043bae37e0d389f5\": container with ID starting with f6073140d9f86929b6d4bd2960ede0617a5c9620b269bbab043bae37e0d389f5 not found: ID does not exist" Dec 13 07:03:13 crc kubenswrapper[4971]: I1213 07:03:13.072407 4971 scope.go:117] "RemoveContainer" containerID="693a2ab434c250cb140a866e8068b156df9191414fed1d63bee919a2528ea6e5" Dec 13 07:03:13 crc kubenswrapper[4971]: I1213 07:03:13.072598 4971 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"693a2ab434c250cb140a866e8068b156df9191414fed1d63bee919a2528ea6e5"} err="failed to get container status \"693a2ab434c250cb140a866e8068b156df9191414fed1d63bee919a2528ea6e5\": rpc error: code = NotFound desc = could not find container \"693a2ab434c250cb140a866e8068b156df9191414fed1d63bee919a2528ea6e5\": container with ID starting with 693a2ab434c250cb140a866e8068b156df9191414fed1d63bee919a2528ea6e5 not found: ID does not exist" Dec 13 07:03:13 crc kubenswrapper[4971]: I1213 07:03:13.072625 4971 scope.go:117] "RemoveContainer" containerID="f22a4ec756119c4f791fb5105cf1f31df3e548b2b07b1f909a62802e9b6035ae" Dec 13 07:03:13 crc kubenswrapper[4971]: I1213 07:03:13.072827 4971 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"f22a4ec756119c4f791fb5105cf1f31df3e548b2b07b1f909a62802e9b6035ae"} err="failed to get container status \"f22a4ec756119c4f791fb5105cf1f31df3e548b2b07b1f909a62802e9b6035ae\": rpc error: code = NotFound desc = could not find container \"f22a4ec756119c4f791fb5105cf1f31df3e548b2b07b1f909a62802e9b6035ae\": container with ID starting with f22a4ec756119c4f791fb5105cf1f31df3e548b2b07b1f909a62802e9b6035ae not found: ID does not exist" Dec 13 07:03:13 crc kubenswrapper[4971]: I1213 07:03:13.072847 4971 scope.go:117] "RemoveContainer" containerID="325223543b3c7db417f769533559a1ca2b23a723494e1885b0fd0480b5a6399f" Dec 13 07:03:13 crc kubenswrapper[4971]: I1213 07:03:13.073028 4971 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"325223543b3c7db417f769533559a1ca2b23a723494e1885b0fd0480b5a6399f"} err="failed to get container status \"325223543b3c7db417f769533559a1ca2b23a723494e1885b0fd0480b5a6399f\": rpc error: code = NotFound desc = could not find container \"325223543b3c7db417f769533559a1ca2b23a723494e1885b0fd0480b5a6399f\": container with ID starting with 325223543b3c7db417f769533559a1ca2b23a723494e1885b0fd0480b5a6399f not found: ID does not exist" Dec 13 07:03:13 crc kubenswrapper[4971]: I1213 07:03:13.073045 4971 scope.go:117] "RemoveContainer" containerID="219620f51474e60b3e1f880c0972eb612dffc4bad24462bbba2c093ae817d852" Dec 13 07:03:13 crc kubenswrapper[4971]: I1213 07:03:13.073259 4971 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"219620f51474e60b3e1f880c0972eb612dffc4bad24462bbba2c093ae817d852"} err="failed to get container status \"219620f51474e60b3e1f880c0972eb612dffc4bad24462bbba2c093ae817d852\": rpc error: code = NotFound desc = could not find container \"219620f51474e60b3e1f880c0972eb612dffc4bad24462bbba2c093ae817d852\": container with ID starting with 219620f51474e60b3e1f880c0972eb612dffc4bad24462bbba2c093ae817d852 not found: ID does not exist" Dec 13 07:03:13 crc kubenswrapper[4971]: I1213 07:03:13.073277 4971 scope.go:117] "RemoveContainer" containerID="f137f4b441c2402093ab430a9689a46221c4fb4bc70d99fb992f62b0a989c465" Dec 13 07:03:13 crc kubenswrapper[4971]: I1213 07:03:13.073460 4971 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"f137f4b441c2402093ab430a9689a46221c4fb4bc70d99fb992f62b0a989c465"} err="failed to get container status \"f137f4b441c2402093ab430a9689a46221c4fb4bc70d99fb992f62b0a989c465\": rpc error: code = NotFound desc = could not find container \"f137f4b441c2402093ab430a9689a46221c4fb4bc70d99fb992f62b0a989c465\": container with ID starting with f137f4b441c2402093ab430a9689a46221c4fb4bc70d99fb992f62b0a989c465 not found: ID does not exist" Dec 13 07:03:13 crc kubenswrapper[4971]: I1213 07:03:13.073477 4971 scope.go:117] "RemoveContainer" containerID="451b22df6973ddc8fa2209a65d64d9e41b33eb231bb4adfab88cf0e5943bb246" Dec 13 07:03:13 crc kubenswrapper[4971]: I1213 07:03:13.073667 4971 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"451b22df6973ddc8fa2209a65d64d9e41b33eb231bb4adfab88cf0e5943bb246"} err="failed to get container status \"451b22df6973ddc8fa2209a65d64d9e41b33eb231bb4adfab88cf0e5943bb246\": rpc error: code = NotFound desc = could not find container \"451b22df6973ddc8fa2209a65d64d9e41b33eb231bb4adfab88cf0e5943bb246\": container with ID starting with 451b22df6973ddc8fa2209a65d64d9e41b33eb231bb4adfab88cf0e5943bb246 not found: ID does not exist" Dec 13 07:03:13 crc kubenswrapper[4971]: I1213 07:03:13.777604 4971 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="736c6b1a-93dd-4f71-81b6-1c43fcf5f556" path="/var/lib/kubelet/pods/736c6b1a-93dd-4f71-81b6-1c43fcf5f556/volumes" Dec 13 07:03:13 crc kubenswrapper[4971]: I1213 07:03:13.835474 4971 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-multus_multus-wwvpl_dddf744b-2122-49ad-a980-105c09636e0f/kube-multus/2.log" Dec 13 07:03:13 crc kubenswrapper[4971]: I1213 07:03:13.835643 4971 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-wwvpl" event={"ID":"dddf744b-2122-49ad-a980-105c09636e0f","Type":"ContainerStarted","Data":"9ea5338b17a10a9995a8d36b80c7df595983904ce4411e16c160e9532de7f31a"} Dec 13 07:03:13 crc kubenswrapper[4971]: I1213 07:03:13.839318 4971 generic.go:334] "Generic (PLEG): container finished" podID="45f87d92-a73d-47ed-83da-400dc4d72255" containerID="6ed6771dd4ed71530c3d5f6777ef14b76b88e5812133ad8ce2646b5e127eb8fc" exitCode=0 Dec 13 07:03:13 crc kubenswrapper[4971]: I1213 07:03:13.839370 4971 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-fq8jh" event={"ID":"45f87d92-a73d-47ed-83da-400dc4d72255","Type":"ContainerDied","Data":"6ed6771dd4ed71530c3d5f6777ef14b76b88e5812133ad8ce2646b5e127eb8fc"} Dec 13 07:03:13 crc kubenswrapper[4971]: I1213 07:03:13.839402 4971 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-fq8jh" event={"ID":"45f87d92-a73d-47ed-83da-400dc4d72255","Type":"ContainerStarted","Data":"0b59bfdcec167c4529b63d292dce92d0db3c815ab8abc9e50ae49b2673e3db03"} Dec 13 07:03:14 crc kubenswrapper[4971]: I1213 07:03:14.852803 4971 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-fq8jh" event={"ID":"45f87d92-a73d-47ed-83da-400dc4d72255","Type":"ContainerStarted","Data":"9894e8f21f8889684f454b8df0e0d08bcb18e8641c6772f3c709bfb13b33aed6"} Dec 13 07:03:14 crc kubenswrapper[4971]: I1213 07:03:14.852854 4971 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-fq8jh" event={"ID":"45f87d92-a73d-47ed-83da-400dc4d72255","Type":"ContainerStarted","Data":"33fbfd3799fce7a3055a36c3dbfe50934a39d3e89014b599314e5417d35fb7a7"} Dec 13 07:03:14 crc kubenswrapper[4971]: I1213 07:03:14.852867 4971 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-fq8jh" event={"ID":"45f87d92-a73d-47ed-83da-400dc4d72255","Type":"ContainerStarted","Data":"52c1811ec1c3ea675ebd74da9bd19378035429b998c0b415b155829bd60fdc49"} Dec 13 07:03:14 crc kubenswrapper[4971]: I1213 07:03:14.852879 4971 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-fq8jh" event={"ID":"45f87d92-a73d-47ed-83da-400dc4d72255","Type":"ContainerStarted","Data":"9f60feb5efe8866336eaf2e236db57659cf60323935370f4aacb00c090c9b014"} Dec 13 07:03:14 crc kubenswrapper[4971]: I1213 07:03:14.852891 4971 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-fq8jh" event={"ID":"45f87d92-a73d-47ed-83da-400dc4d72255","Type":"ContainerStarted","Data":"4a00580eda2581c305e410847ae3fe0e82b31175b85780b04ac88e24e003968c"} Dec 13 07:03:14 crc kubenswrapper[4971]: I1213 07:03:14.852901 4971 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-fq8jh" event={"ID":"45f87d92-a73d-47ed-83da-400dc4d72255","Type":"ContainerStarted","Data":"5a9243a838aefe86f06f7c6dbde2963c9e44493c5ac30e372dc3a398f3897e88"} Dec 13 07:03:16 crc kubenswrapper[4971]: I1213 07:03:16.870062 4971 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-fq8jh" event={"ID":"45f87d92-a73d-47ed-83da-400dc4d72255","Type":"ContainerStarted","Data":"e0b229534eb428da170ef4c53327491e8bf82d45d4e9ba3004ce1b99cf3857c5"} Dec 13 07:03:17 crc kubenswrapper[4971]: I1213 07:03:17.383809 4971 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="cert-manager/cert-manager-webhook-5655c58dd6-qbbjr" Dec 13 07:03:19 crc kubenswrapper[4971]: I1213 07:03:19.888086 4971 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-fq8jh" event={"ID":"45f87d92-a73d-47ed-83da-400dc4d72255","Type":"ContainerStarted","Data":"ff17f12814566e7fe3dc73f7080926296a07494974fa909b604628f936986021"} Dec 13 07:03:19 crc kubenswrapper[4971]: I1213 07:03:19.888754 4971 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-ovn-kubernetes/ovnkube-node-fq8jh" Dec 13 07:03:19 crc kubenswrapper[4971]: I1213 07:03:19.888770 4971 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-ovn-kubernetes/ovnkube-node-fq8jh" Dec 13 07:03:19 crc kubenswrapper[4971]: I1213 07:03:19.888781 4971 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-ovn-kubernetes/ovnkube-node-fq8jh" Dec 13 07:03:19 crc kubenswrapper[4971]: I1213 07:03:19.918076 4971 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-ovn-kubernetes/ovnkube-node-fq8jh" podStartSLOduration=7.918049807 podStartE2EDuration="7.918049807s" podCreationTimestamp="2025-12-13 07:03:12 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-13 07:03:19.914944981 +0000 UTC m=+856.519354479" watchObservedRunningTime="2025-12-13 07:03:19.918049807 +0000 UTC m=+856.522459295" Dec 13 07:03:19 crc kubenswrapper[4971]: I1213 07:03:19.923263 4971 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-ovn-kubernetes/ovnkube-node-fq8jh" Dec 13 07:03:19 crc kubenswrapper[4971]: I1213 07:03:19.924596 4971 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-ovn-kubernetes/ovnkube-node-fq8jh" Dec 13 07:03:42 crc kubenswrapper[4971]: I1213 07:03:42.863178 4971 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-ovn-kubernetes/ovnkube-node-fq8jh" Dec 13 07:04:08 crc kubenswrapper[4971]: I1213 07:04:08.153129 4971 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/community-operators-5kgpk"] Dec 13 07:04:08 crc kubenswrapper[4971]: I1213 07:04:08.154933 4971 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-5kgpk" Dec 13 07:04:08 crc kubenswrapper[4971]: I1213 07:04:08.161325 4971 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-5kgpk"] Dec 13 07:04:08 crc kubenswrapper[4971]: I1213 07:04:08.332465 4971 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-7fx8q\" (UniqueName: \"kubernetes.io/projected/7e9ec79e-730d-4d91-b8f9-851c2b1cb769-kube-api-access-7fx8q\") pod \"community-operators-5kgpk\" (UID: \"7e9ec79e-730d-4d91-b8f9-851c2b1cb769\") " pod="openshift-marketplace/community-operators-5kgpk" Dec 13 07:04:08 crc kubenswrapper[4971]: I1213 07:04:08.332567 4971 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/7e9ec79e-730d-4d91-b8f9-851c2b1cb769-catalog-content\") pod \"community-operators-5kgpk\" (UID: \"7e9ec79e-730d-4d91-b8f9-851c2b1cb769\") " pod="openshift-marketplace/community-operators-5kgpk" Dec 13 07:04:08 crc kubenswrapper[4971]: I1213 07:04:08.332692 4971 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/7e9ec79e-730d-4d91-b8f9-851c2b1cb769-utilities\") pod \"community-operators-5kgpk\" (UID: \"7e9ec79e-730d-4d91-b8f9-851c2b1cb769\") " pod="openshift-marketplace/community-operators-5kgpk" Dec 13 07:04:08 crc kubenswrapper[4971]: I1213 07:04:08.433917 4971 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/7e9ec79e-730d-4d91-b8f9-851c2b1cb769-catalog-content\") pod \"community-operators-5kgpk\" (UID: \"7e9ec79e-730d-4d91-b8f9-851c2b1cb769\") " pod="openshift-marketplace/community-operators-5kgpk" Dec 13 07:04:08 crc kubenswrapper[4971]: I1213 07:04:08.433992 4971 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/7e9ec79e-730d-4d91-b8f9-851c2b1cb769-utilities\") pod \"community-operators-5kgpk\" (UID: \"7e9ec79e-730d-4d91-b8f9-851c2b1cb769\") " pod="openshift-marketplace/community-operators-5kgpk" Dec 13 07:04:08 crc kubenswrapper[4971]: I1213 07:04:08.434036 4971 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-7fx8q\" (UniqueName: \"kubernetes.io/projected/7e9ec79e-730d-4d91-b8f9-851c2b1cb769-kube-api-access-7fx8q\") pod \"community-operators-5kgpk\" (UID: \"7e9ec79e-730d-4d91-b8f9-851c2b1cb769\") " pod="openshift-marketplace/community-operators-5kgpk" Dec 13 07:04:08 crc kubenswrapper[4971]: I1213 07:04:08.435116 4971 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/7e9ec79e-730d-4d91-b8f9-851c2b1cb769-utilities\") pod \"community-operators-5kgpk\" (UID: \"7e9ec79e-730d-4d91-b8f9-851c2b1cb769\") " pod="openshift-marketplace/community-operators-5kgpk" Dec 13 07:04:08 crc kubenswrapper[4971]: I1213 07:04:08.435278 4971 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/7e9ec79e-730d-4d91-b8f9-851c2b1cb769-catalog-content\") pod \"community-operators-5kgpk\" (UID: \"7e9ec79e-730d-4d91-b8f9-851c2b1cb769\") " pod="openshift-marketplace/community-operators-5kgpk" Dec 13 07:04:08 crc kubenswrapper[4971]: I1213 07:04:08.459913 4971 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-7fx8q\" (UniqueName: \"kubernetes.io/projected/7e9ec79e-730d-4d91-b8f9-851c2b1cb769-kube-api-access-7fx8q\") pod \"community-operators-5kgpk\" (UID: \"7e9ec79e-730d-4d91-b8f9-851c2b1cb769\") " pod="openshift-marketplace/community-operators-5kgpk" Dec 13 07:04:08 crc kubenswrapper[4971]: I1213 07:04:08.473686 4971 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-5kgpk" Dec 13 07:04:08 crc kubenswrapper[4971]: I1213 07:04:08.700680 4971 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-5kgpk"] Dec 13 07:04:09 crc kubenswrapper[4971]: I1213 07:04:09.153432 4971 generic.go:334] "Generic (PLEG): container finished" podID="7e9ec79e-730d-4d91-b8f9-851c2b1cb769" containerID="2123533c6c514a612c02c3519057f2c3521d7a1afa3187b6cd26d425858d5230" exitCode=0 Dec 13 07:04:09 crc kubenswrapper[4971]: I1213 07:04:09.153481 4971 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-5kgpk" event={"ID":"7e9ec79e-730d-4d91-b8f9-851c2b1cb769","Type":"ContainerDied","Data":"2123533c6c514a612c02c3519057f2c3521d7a1afa3187b6cd26d425858d5230"} Dec 13 07:04:09 crc kubenswrapper[4971]: I1213 07:04:09.153537 4971 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-5kgpk" event={"ID":"7e9ec79e-730d-4d91-b8f9-851c2b1cb769","Type":"ContainerStarted","Data":"637cbbc8b0da971b274574b1075e1e413f27fa5283a753b58f382f1c4980eecf"} Dec 13 07:04:10 crc kubenswrapper[4971]: I1213 07:04:10.160603 4971 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-5kgpk" event={"ID":"7e9ec79e-730d-4d91-b8f9-851c2b1cb769","Type":"ContainerStarted","Data":"a5b55297439bc82bb4cc7c6eb000526c88488ac57a9f9efad54f0b90a45993f6"} Dec 13 07:04:11 crc kubenswrapper[4971]: I1213 07:04:11.167607 4971 generic.go:334] "Generic (PLEG): container finished" podID="7e9ec79e-730d-4d91-b8f9-851c2b1cb769" containerID="a5b55297439bc82bb4cc7c6eb000526c88488ac57a9f9efad54f0b90a45993f6" exitCode=0 Dec 13 07:04:11 crc kubenswrapper[4971]: I1213 07:04:11.167662 4971 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-5kgpk" event={"ID":"7e9ec79e-730d-4d91-b8f9-851c2b1cb769","Type":"ContainerDied","Data":"a5b55297439bc82bb4cc7c6eb000526c88488ac57a9f9efad54f0b90a45993f6"} Dec 13 07:04:11 crc kubenswrapper[4971]: I1213 07:04:11.220197 4971 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-4pzkf"] Dec 13 07:04:11 crc kubenswrapper[4971]: I1213 07:04:11.220493 4971 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/certified-operators-4pzkf" podUID="97cf4a02-e236-49c1-904f-21b25e87eba4" containerName="registry-server" containerID="cri-o://47325854a36fbd957987f114286b98caf8038283e15ce75adc5c479b0787910c" gracePeriod=30 Dec 13 07:04:11 crc kubenswrapper[4971]: I1213 07:04:11.233667 4971 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-5kgpk"] Dec 13 07:04:11 crc kubenswrapper[4971]: I1213 07:04:11.242590 4971 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-kbz6d"] Dec 13 07:04:11 crc kubenswrapper[4971]: I1213 07:04:11.242927 4971 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/community-operators-kbz6d" podUID="d44cf79f-dbdf-4e9b-ba48-7351d398c38c" containerName="registry-server" containerID="cri-o://0cbd2685b75d6e865d414a1ccb83ec0b3fa61e649b012dabdb47bd9f09c8c2c2" gracePeriod=30 Dec 13 07:04:11 crc kubenswrapper[4971]: I1213 07:04:11.244662 4971 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/marketplace-operator-79b997595-6rvc9"] Dec 13 07:04:11 crc kubenswrapper[4971]: I1213 07:04:11.244927 4971 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/marketplace-operator-79b997595-6rvc9" podUID="5c88a864-89a9-4f5e-b713-385170b9bf17" containerName="marketplace-operator" containerID="cri-o://c61b97c1755702dffd406eec5db5b5f8bb5f4fa4b6d081c8392bce7eb961c949" gracePeriod=30 Dec 13 07:04:11 crc kubenswrapper[4971]: I1213 07:04:11.248094 4971 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-68kxv"] Dec 13 07:04:11 crc kubenswrapper[4971]: I1213 07:04:11.248369 4971 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-marketplace-68kxv" podUID="5eea3e58-f9fb-4529-b645-0f5193d5a8a9" containerName="registry-server" containerID="cri-o://937af01c300b74d0c52a530d08bf8ad033b99588471fb7be33345a2e3fb07a39" gracePeriod=30 Dec 13 07:04:11 crc kubenswrapper[4971]: I1213 07:04:11.262149 4971 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/marketplace-operator-79b997595-w4r9w"] Dec 13 07:04:11 crc kubenswrapper[4971]: I1213 07:04:11.263041 4971 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/marketplace-operator-79b997595-w4r9w" Dec 13 07:04:11 crc kubenswrapper[4971]: I1213 07:04:11.266798 4971 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-5qtd2"] Dec 13 07:04:11 crc kubenswrapper[4971]: I1213 07:04:11.266974 4971 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-operators-5qtd2" podUID="9fd587ab-4839-4479-80f8-f97d78d833b9" containerName="registry-server" containerID="cri-o://91507b889fb1db4fed8d82e2f27769c51087db01b1cd1c9b2b06d317b8ef3ff0" gracePeriod=30 Dec 13 07:04:11 crc kubenswrapper[4971]: I1213 07:04:11.267430 4971 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"marketplace-trusted-ca\" (UniqueName: \"kubernetes.io/configmap/f8802833-bb84-4479-be26-f0fae563f1b0-marketplace-trusted-ca\") pod \"marketplace-operator-79b997595-w4r9w\" (UID: \"f8802833-bb84-4479-be26-f0fae563f1b0\") " pod="openshift-marketplace/marketplace-operator-79b997595-w4r9w" Dec 13 07:04:11 crc kubenswrapper[4971]: I1213 07:04:11.267451 4971 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-kvkhn\" (UniqueName: \"kubernetes.io/projected/f8802833-bb84-4479-be26-f0fae563f1b0-kube-api-access-kvkhn\") pod \"marketplace-operator-79b997595-w4r9w\" (UID: \"f8802833-bb84-4479-be26-f0fae563f1b0\") " pod="openshift-marketplace/marketplace-operator-79b997595-w4r9w" Dec 13 07:04:11 crc kubenswrapper[4971]: I1213 07:04:11.267530 4971 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"marketplace-operator-metrics\" (UniqueName: \"kubernetes.io/secret/f8802833-bb84-4479-be26-f0fae563f1b0-marketplace-operator-metrics\") pod \"marketplace-operator-79b997595-w4r9w\" (UID: \"f8802833-bb84-4479-be26-f0fae563f1b0\") " pod="openshift-marketplace/marketplace-operator-79b997595-w4r9w" Dec 13 07:04:11 crc kubenswrapper[4971]: I1213 07:04:11.282363 4971 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/marketplace-operator-79b997595-w4r9w"] Dec 13 07:04:11 crc kubenswrapper[4971]: I1213 07:04:11.368474 4971 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"marketplace-operator-metrics\" (UniqueName: \"kubernetes.io/secret/f8802833-bb84-4479-be26-f0fae563f1b0-marketplace-operator-metrics\") pod \"marketplace-operator-79b997595-w4r9w\" (UID: \"f8802833-bb84-4479-be26-f0fae563f1b0\") " pod="openshift-marketplace/marketplace-operator-79b997595-w4r9w" Dec 13 07:04:11 crc kubenswrapper[4971]: I1213 07:04:11.368544 4971 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"marketplace-trusted-ca\" (UniqueName: \"kubernetes.io/configmap/f8802833-bb84-4479-be26-f0fae563f1b0-marketplace-trusted-ca\") pod \"marketplace-operator-79b997595-w4r9w\" (UID: \"f8802833-bb84-4479-be26-f0fae563f1b0\") " pod="openshift-marketplace/marketplace-operator-79b997595-w4r9w" Dec 13 07:04:11 crc kubenswrapper[4971]: I1213 07:04:11.368562 4971 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-kvkhn\" (UniqueName: \"kubernetes.io/projected/f8802833-bb84-4479-be26-f0fae563f1b0-kube-api-access-kvkhn\") pod \"marketplace-operator-79b997595-w4r9w\" (UID: \"f8802833-bb84-4479-be26-f0fae563f1b0\") " pod="openshift-marketplace/marketplace-operator-79b997595-w4r9w" Dec 13 07:04:11 crc kubenswrapper[4971]: I1213 07:04:11.370878 4971 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"marketplace-trusted-ca\" (UniqueName: \"kubernetes.io/configmap/f8802833-bb84-4479-be26-f0fae563f1b0-marketplace-trusted-ca\") pod \"marketplace-operator-79b997595-w4r9w\" (UID: \"f8802833-bb84-4479-be26-f0fae563f1b0\") " pod="openshift-marketplace/marketplace-operator-79b997595-w4r9w" Dec 13 07:04:11 crc kubenswrapper[4971]: I1213 07:04:11.376318 4971 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"marketplace-operator-metrics\" (UniqueName: \"kubernetes.io/secret/f8802833-bb84-4479-be26-f0fae563f1b0-marketplace-operator-metrics\") pod \"marketplace-operator-79b997595-w4r9w\" (UID: \"f8802833-bb84-4479-be26-f0fae563f1b0\") " pod="openshift-marketplace/marketplace-operator-79b997595-w4r9w" Dec 13 07:04:11 crc kubenswrapper[4971]: I1213 07:04:11.388833 4971 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-kvkhn\" (UniqueName: \"kubernetes.io/projected/f8802833-bb84-4479-be26-f0fae563f1b0-kube-api-access-kvkhn\") pod \"marketplace-operator-79b997595-w4r9w\" (UID: \"f8802833-bb84-4479-be26-f0fae563f1b0\") " pod="openshift-marketplace/marketplace-operator-79b997595-w4r9w" Dec 13 07:04:11 crc kubenswrapper[4971]: I1213 07:04:11.670713 4971 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/marketplace-operator-79b997595-w4r9w" Dec 13 07:04:11 crc kubenswrapper[4971]: I1213 07:04:11.676217 4971 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-4pzkf" Dec 13 07:04:11 crc kubenswrapper[4971]: I1213 07:04:11.685814 4971 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-68kxv" Dec 13 07:04:11 crc kubenswrapper[4971]: I1213 07:04:11.688443 4971 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_marketplace-operator-79b997595-6rvc9_5c88a864-89a9-4f5e-b713-385170b9bf17/marketplace-operator/2.log" Dec 13 07:04:11 crc kubenswrapper[4971]: I1213 07:04:11.688534 4971 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/marketplace-operator-79b997595-6rvc9" Dec 13 07:04:11 crc kubenswrapper[4971]: I1213 07:04:11.706795 4971 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-kbz6d" Dec 13 07:04:11 crc kubenswrapper[4971]: I1213 07:04:11.708796 4971 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-5qtd2" Dec 13 07:04:11 crc kubenswrapper[4971]: I1213 07:04:11.772501 4971 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"marketplace-trusted-ca\" (UniqueName: \"kubernetes.io/configmap/5c88a864-89a9-4f5e-b713-385170b9bf17-marketplace-trusted-ca\") pod \"5c88a864-89a9-4f5e-b713-385170b9bf17\" (UID: \"5c88a864-89a9-4f5e-b713-385170b9bf17\") " Dec 13 07:04:11 crc kubenswrapper[4971]: I1213 07:04:11.772565 4971 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-7zzwm\" (UniqueName: \"kubernetes.io/projected/5c88a864-89a9-4f5e-b713-385170b9bf17-kube-api-access-7zzwm\") pod \"5c88a864-89a9-4f5e-b713-385170b9bf17\" (UID: \"5c88a864-89a9-4f5e-b713-385170b9bf17\") " Dec 13 07:04:11 crc kubenswrapper[4971]: I1213 07:04:11.772621 4971 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/9fd587ab-4839-4479-80f8-f97d78d833b9-catalog-content\") pod \"9fd587ab-4839-4479-80f8-f97d78d833b9\" (UID: \"9fd587ab-4839-4479-80f8-f97d78d833b9\") " Dec 13 07:04:11 crc kubenswrapper[4971]: I1213 07:04:11.772643 4971 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/97cf4a02-e236-49c1-904f-21b25e87eba4-catalog-content\") pod \"97cf4a02-e236-49c1-904f-21b25e87eba4\" (UID: \"97cf4a02-e236-49c1-904f-21b25e87eba4\") " Dec 13 07:04:11 crc kubenswrapper[4971]: I1213 07:04:11.774182 4971 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/5c88a864-89a9-4f5e-b713-385170b9bf17-marketplace-trusted-ca" (OuterVolumeSpecName: "marketplace-trusted-ca") pod "5c88a864-89a9-4f5e-b713-385170b9bf17" (UID: "5c88a864-89a9-4f5e-b713-385170b9bf17"). InnerVolumeSpecName "marketplace-trusted-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 13 07:04:11 crc kubenswrapper[4971]: I1213 07:04:11.788454 4971 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/5c88a864-89a9-4f5e-b713-385170b9bf17-kube-api-access-7zzwm" (OuterVolumeSpecName: "kube-api-access-7zzwm") pod "5c88a864-89a9-4f5e-b713-385170b9bf17" (UID: "5c88a864-89a9-4f5e-b713-385170b9bf17"). InnerVolumeSpecName "kube-api-access-7zzwm". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 13 07:04:11 crc kubenswrapper[4971]: I1213 07:04:11.830261 4971 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/97cf4a02-e236-49c1-904f-21b25e87eba4-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "97cf4a02-e236-49c1-904f-21b25e87eba4" (UID: "97cf4a02-e236-49c1-904f-21b25e87eba4"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 13 07:04:11 crc kubenswrapper[4971]: I1213 07:04:11.873192 4971 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/5eea3e58-f9fb-4529-b645-0f5193d5a8a9-catalog-content\") pod \"5eea3e58-f9fb-4529-b645-0f5193d5a8a9\" (UID: \"5eea3e58-f9fb-4529-b645-0f5193d5a8a9\") " Dec 13 07:04:11 crc kubenswrapper[4971]: I1213 07:04:11.873583 4971 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/d44cf79f-dbdf-4e9b-ba48-7351d398c38c-catalog-content\") pod \"d44cf79f-dbdf-4e9b-ba48-7351d398c38c\" (UID: \"d44cf79f-dbdf-4e9b-ba48-7351d398c38c\") " Dec 13 07:04:11 crc kubenswrapper[4971]: I1213 07:04:11.873631 4971 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/5eea3e58-f9fb-4529-b645-0f5193d5a8a9-utilities\") pod \"5eea3e58-f9fb-4529-b645-0f5193d5a8a9\" (UID: \"5eea3e58-f9fb-4529-b645-0f5193d5a8a9\") " Dec 13 07:04:11 crc kubenswrapper[4971]: I1213 07:04:11.873651 4971 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-wnvmg\" (UniqueName: \"kubernetes.io/projected/97cf4a02-e236-49c1-904f-21b25e87eba4-kube-api-access-wnvmg\") pod \"97cf4a02-e236-49c1-904f-21b25e87eba4\" (UID: \"97cf4a02-e236-49c1-904f-21b25e87eba4\") " Dec 13 07:04:11 crc kubenswrapper[4971]: I1213 07:04:11.873682 4971 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"marketplace-operator-metrics\" (UniqueName: \"kubernetes.io/secret/5c88a864-89a9-4f5e-b713-385170b9bf17-marketplace-operator-metrics\") pod \"5c88a864-89a9-4f5e-b713-385170b9bf17\" (UID: \"5c88a864-89a9-4f5e-b713-385170b9bf17\") " Dec 13 07:04:11 crc kubenswrapper[4971]: I1213 07:04:11.873730 4971 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-zsstt\" (UniqueName: \"kubernetes.io/projected/9fd587ab-4839-4479-80f8-f97d78d833b9-kube-api-access-zsstt\") pod \"9fd587ab-4839-4479-80f8-f97d78d833b9\" (UID: \"9fd587ab-4839-4479-80f8-f97d78d833b9\") " Dec 13 07:04:11 crc kubenswrapper[4971]: I1213 07:04:11.873748 4971 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/9fd587ab-4839-4479-80f8-f97d78d833b9-utilities\") pod \"9fd587ab-4839-4479-80f8-f97d78d833b9\" (UID: \"9fd587ab-4839-4479-80f8-f97d78d833b9\") " Dec 13 07:04:11 crc kubenswrapper[4971]: I1213 07:04:11.873764 4971 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-frhc2\" (UniqueName: \"kubernetes.io/projected/5eea3e58-f9fb-4529-b645-0f5193d5a8a9-kube-api-access-frhc2\") pod \"5eea3e58-f9fb-4529-b645-0f5193d5a8a9\" (UID: \"5eea3e58-f9fb-4529-b645-0f5193d5a8a9\") " Dec 13 07:04:11 crc kubenswrapper[4971]: I1213 07:04:11.873781 4971 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/d44cf79f-dbdf-4e9b-ba48-7351d398c38c-utilities\") pod \"d44cf79f-dbdf-4e9b-ba48-7351d398c38c\" (UID: \"d44cf79f-dbdf-4e9b-ba48-7351d398c38c\") " Dec 13 07:04:11 crc kubenswrapper[4971]: I1213 07:04:11.873802 4971 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/97cf4a02-e236-49c1-904f-21b25e87eba4-utilities\") pod \"97cf4a02-e236-49c1-904f-21b25e87eba4\" (UID: \"97cf4a02-e236-49c1-904f-21b25e87eba4\") " Dec 13 07:04:11 crc kubenswrapper[4971]: I1213 07:04:11.873826 4971 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-tl949\" (UniqueName: \"kubernetes.io/projected/d44cf79f-dbdf-4e9b-ba48-7351d398c38c-kube-api-access-tl949\") pod \"d44cf79f-dbdf-4e9b-ba48-7351d398c38c\" (UID: \"d44cf79f-dbdf-4e9b-ba48-7351d398c38c\") " Dec 13 07:04:11 crc kubenswrapper[4971]: I1213 07:04:11.877651 4971 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/9fd587ab-4839-4479-80f8-f97d78d833b9-utilities" (OuterVolumeSpecName: "utilities") pod "9fd587ab-4839-4479-80f8-f97d78d833b9" (UID: "9fd587ab-4839-4479-80f8-f97d78d833b9"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 13 07:04:11 crc kubenswrapper[4971]: I1213 07:04:11.878587 4971 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/97cf4a02-e236-49c1-904f-21b25e87eba4-utilities" (OuterVolumeSpecName: "utilities") pod "97cf4a02-e236-49c1-904f-21b25e87eba4" (UID: "97cf4a02-e236-49c1-904f-21b25e87eba4"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 13 07:04:11 crc kubenswrapper[4971]: I1213 07:04:11.879174 4971 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/5c88a864-89a9-4f5e-b713-385170b9bf17-marketplace-operator-metrics" (OuterVolumeSpecName: "marketplace-operator-metrics") pod "5c88a864-89a9-4f5e-b713-385170b9bf17" (UID: "5c88a864-89a9-4f5e-b713-385170b9bf17"). InnerVolumeSpecName "marketplace-operator-metrics". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 13 07:04:11 crc kubenswrapper[4971]: I1213 07:04:11.880923 4971 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/5eea3e58-f9fb-4529-b645-0f5193d5a8a9-utilities" (OuterVolumeSpecName: "utilities") pod "5eea3e58-f9fb-4529-b645-0f5193d5a8a9" (UID: "5eea3e58-f9fb-4529-b645-0f5193d5a8a9"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 13 07:04:11 crc kubenswrapper[4971]: I1213 07:04:11.880975 4971 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/9fd587ab-4839-4479-80f8-f97d78d833b9-kube-api-access-zsstt" (OuterVolumeSpecName: "kube-api-access-zsstt") pod "9fd587ab-4839-4479-80f8-f97d78d833b9" (UID: "9fd587ab-4839-4479-80f8-f97d78d833b9"). InnerVolumeSpecName "kube-api-access-zsstt". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 13 07:04:11 crc kubenswrapper[4971]: I1213 07:04:11.880990 4971 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/97cf4a02-e236-49c1-904f-21b25e87eba4-catalog-content\") on node \"crc\" DevicePath \"\"" Dec 13 07:04:11 crc kubenswrapper[4971]: I1213 07:04:11.881057 4971 reconciler_common.go:293] "Volume detached for volume \"marketplace-trusted-ca\" (UniqueName: \"kubernetes.io/configmap/5c88a864-89a9-4f5e-b713-385170b9bf17-marketplace-trusted-ca\") on node \"crc\" DevicePath \"\"" Dec 13 07:04:11 crc kubenswrapper[4971]: I1213 07:04:11.881072 4971 reconciler_common.go:293] "Volume detached for volume \"marketplace-operator-metrics\" (UniqueName: \"kubernetes.io/secret/5c88a864-89a9-4f5e-b713-385170b9bf17-marketplace-operator-metrics\") on node \"crc\" DevicePath \"\"" Dec 13 07:04:11 crc kubenswrapper[4971]: I1213 07:04:11.881083 4971 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-7zzwm\" (UniqueName: \"kubernetes.io/projected/5c88a864-89a9-4f5e-b713-385170b9bf17-kube-api-access-7zzwm\") on node \"crc\" DevicePath \"\"" Dec 13 07:04:11 crc kubenswrapper[4971]: I1213 07:04:11.881092 4971 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/9fd587ab-4839-4479-80f8-f97d78d833b9-utilities\") on node \"crc\" DevicePath \"\"" Dec 13 07:04:11 crc kubenswrapper[4971]: I1213 07:04:11.881101 4971 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/97cf4a02-e236-49c1-904f-21b25e87eba4-utilities\") on node \"crc\" DevicePath \"\"" Dec 13 07:04:11 crc kubenswrapper[4971]: I1213 07:04:11.881624 4971 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/97cf4a02-e236-49c1-904f-21b25e87eba4-kube-api-access-wnvmg" (OuterVolumeSpecName: "kube-api-access-wnvmg") pod "97cf4a02-e236-49c1-904f-21b25e87eba4" (UID: "97cf4a02-e236-49c1-904f-21b25e87eba4"). InnerVolumeSpecName "kube-api-access-wnvmg". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 13 07:04:11 crc kubenswrapper[4971]: I1213 07:04:11.886776 4971 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/5eea3e58-f9fb-4529-b645-0f5193d5a8a9-kube-api-access-frhc2" (OuterVolumeSpecName: "kube-api-access-frhc2") pod "5eea3e58-f9fb-4529-b645-0f5193d5a8a9" (UID: "5eea3e58-f9fb-4529-b645-0f5193d5a8a9"). InnerVolumeSpecName "kube-api-access-frhc2". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 13 07:04:11 crc kubenswrapper[4971]: I1213 07:04:11.895363 4971 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/marketplace-operator-79b997595-w4r9w"] Dec 13 07:04:11 crc kubenswrapper[4971]: I1213 07:04:11.899210 4971 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/d44cf79f-dbdf-4e9b-ba48-7351d398c38c-utilities" (OuterVolumeSpecName: "utilities") pod "d44cf79f-dbdf-4e9b-ba48-7351d398c38c" (UID: "d44cf79f-dbdf-4e9b-ba48-7351d398c38c"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 13 07:04:11 crc kubenswrapper[4971]: W1213 07:04:11.901381 4971 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podf8802833_bb84_4479_be26_f0fae563f1b0.slice/crio-0ccf13188d1a85cee591be593931c07dd049f094afc07e12c368e5f17c0c71db WatchSource:0}: Error finding container 0ccf13188d1a85cee591be593931c07dd049f094afc07e12c368e5f17c0c71db: Status 404 returned error can't find the container with id 0ccf13188d1a85cee591be593931c07dd049f094afc07e12c368e5f17c0c71db Dec 13 07:04:11 crc kubenswrapper[4971]: I1213 07:04:11.901636 4971 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/d44cf79f-dbdf-4e9b-ba48-7351d398c38c-kube-api-access-tl949" (OuterVolumeSpecName: "kube-api-access-tl949") pod "d44cf79f-dbdf-4e9b-ba48-7351d398c38c" (UID: "d44cf79f-dbdf-4e9b-ba48-7351d398c38c"). InnerVolumeSpecName "kube-api-access-tl949". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 13 07:04:11 crc kubenswrapper[4971]: I1213 07:04:11.906891 4971 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/9fd587ab-4839-4479-80f8-f97d78d833b9-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "9fd587ab-4839-4479-80f8-f97d78d833b9" (UID: "9fd587ab-4839-4479-80f8-f97d78d833b9"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 13 07:04:11 crc kubenswrapper[4971]: I1213 07:04:11.907048 4971 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/5eea3e58-f9fb-4529-b645-0f5193d5a8a9-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "5eea3e58-f9fb-4529-b645-0f5193d5a8a9" (UID: "5eea3e58-f9fb-4529-b645-0f5193d5a8a9"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 13 07:04:11 crc kubenswrapper[4971]: I1213 07:04:11.950289 4971 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/d44cf79f-dbdf-4e9b-ba48-7351d398c38c-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "d44cf79f-dbdf-4e9b-ba48-7351d398c38c" (UID: "d44cf79f-dbdf-4e9b-ba48-7351d398c38c"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 13 07:04:11 crc kubenswrapper[4971]: I1213 07:04:11.981825 4971 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/9fd587ab-4839-4479-80f8-f97d78d833b9-catalog-content\") on node \"crc\" DevicePath \"\"" Dec 13 07:04:11 crc kubenswrapper[4971]: I1213 07:04:11.982051 4971 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/d44cf79f-dbdf-4e9b-ba48-7351d398c38c-catalog-content\") on node \"crc\" DevicePath \"\"" Dec 13 07:04:11 crc kubenswrapper[4971]: I1213 07:04:11.982130 4971 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/5eea3e58-f9fb-4529-b645-0f5193d5a8a9-utilities\") on node \"crc\" DevicePath \"\"" Dec 13 07:04:11 crc kubenswrapper[4971]: I1213 07:04:11.982236 4971 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-wnvmg\" (UniqueName: \"kubernetes.io/projected/97cf4a02-e236-49c1-904f-21b25e87eba4-kube-api-access-wnvmg\") on node \"crc\" DevicePath \"\"" Dec 13 07:04:11 crc kubenswrapper[4971]: I1213 07:04:11.982316 4971 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-zsstt\" (UniqueName: \"kubernetes.io/projected/9fd587ab-4839-4479-80f8-f97d78d833b9-kube-api-access-zsstt\") on node \"crc\" DevicePath \"\"" Dec 13 07:04:11 crc kubenswrapper[4971]: I1213 07:04:11.982388 4971 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-frhc2\" (UniqueName: \"kubernetes.io/projected/5eea3e58-f9fb-4529-b645-0f5193d5a8a9-kube-api-access-frhc2\") on node \"crc\" DevicePath \"\"" Dec 13 07:04:11 crc kubenswrapper[4971]: I1213 07:04:11.982449 4971 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/d44cf79f-dbdf-4e9b-ba48-7351d398c38c-utilities\") on node \"crc\" DevicePath \"\"" Dec 13 07:04:11 crc kubenswrapper[4971]: I1213 07:04:11.982499 4971 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-tl949\" (UniqueName: \"kubernetes.io/projected/d44cf79f-dbdf-4e9b-ba48-7351d398c38c-kube-api-access-tl949\") on node \"crc\" DevicePath \"\"" Dec 13 07:04:11 crc kubenswrapper[4971]: I1213 07:04:11.982583 4971 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/5eea3e58-f9fb-4529-b645-0f5193d5a8a9-catalog-content\") on node \"crc\" DevicePath \"\"" Dec 13 07:04:12 crc kubenswrapper[4971]: I1213 07:04:12.173932 4971 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_marketplace-operator-79b997595-6rvc9_5c88a864-89a9-4f5e-b713-385170b9bf17/marketplace-operator/2.log" Dec 13 07:04:12 crc kubenswrapper[4971]: I1213 07:04:12.174986 4971 generic.go:334] "Generic (PLEG): container finished" podID="5c88a864-89a9-4f5e-b713-385170b9bf17" containerID="c61b97c1755702dffd406eec5db5b5f8bb5f4fa4b6d081c8392bce7eb961c949" exitCode=0 Dec 13 07:04:12 crc kubenswrapper[4971]: I1213 07:04:12.175055 4971 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/marketplace-operator-79b997595-6rvc9" Dec 13 07:04:12 crc kubenswrapper[4971]: I1213 07:04:12.175033 4971 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/marketplace-operator-79b997595-6rvc9" event={"ID":"5c88a864-89a9-4f5e-b713-385170b9bf17","Type":"ContainerDied","Data":"c61b97c1755702dffd406eec5db5b5f8bb5f4fa4b6d081c8392bce7eb961c949"} Dec 13 07:04:12 crc kubenswrapper[4971]: I1213 07:04:12.175185 4971 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/marketplace-operator-79b997595-6rvc9" event={"ID":"5c88a864-89a9-4f5e-b713-385170b9bf17","Type":"ContainerDied","Data":"f93dd5acab7a80e5d527432e821725f3324ff27fa1176c1a26c332a22354f3be"} Dec 13 07:04:12 crc kubenswrapper[4971]: I1213 07:04:12.175209 4971 scope.go:117] "RemoveContainer" containerID="c61b97c1755702dffd406eec5db5b5f8bb5f4fa4b6d081c8392bce7eb961c949" Dec 13 07:04:12 crc kubenswrapper[4971]: I1213 07:04:12.177624 4971 generic.go:334] "Generic (PLEG): container finished" podID="9fd587ab-4839-4479-80f8-f97d78d833b9" containerID="91507b889fb1db4fed8d82e2f27769c51087db01b1cd1c9b2b06d317b8ef3ff0" exitCode=0 Dec 13 07:04:12 crc kubenswrapper[4971]: I1213 07:04:12.177706 4971 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-5qtd2" event={"ID":"9fd587ab-4839-4479-80f8-f97d78d833b9","Type":"ContainerDied","Data":"91507b889fb1db4fed8d82e2f27769c51087db01b1cd1c9b2b06d317b8ef3ff0"} Dec 13 07:04:12 crc kubenswrapper[4971]: I1213 07:04:12.177754 4971 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-5qtd2" event={"ID":"9fd587ab-4839-4479-80f8-f97d78d833b9","Type":"ContainerDied","Data":"36a034a89d0f03625ad83f07e5763c9c99810c4d95c944e9d25c4733e3d8d87e"} Dec 13 07:04:12 crc kubenswrapper[4971]: I1213 07:04:12.177861 4971 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-5qtd2" Dec 13 07:04:12 crc kubenswrapper[4971]: I1213 07:04:12.180466 4971 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/marketplace-operator-79b997595-w4r9w" event={"ID":"f8802833-bb84-4479-be26-f0fae563f1b0","Type":"ContainerStarted","Data":"7406fc00fd7af02b9ec434f36e82181aadda1504f2bc950ca1d18b6a70747b5d"} Dec 13 07:04:12 crc kubenswrapper[4971]: I1213 07:04:12.180492 4971 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/marketplace-operator-79b997595-w4r9w" event={"ID":"f8802833-bb84-4479-be26-f0fae563f1b0","Type":"ContainerStarted","Data":"0ccf13188d1a85cee591be593931c07dd049f094afc07e12c368e5f17c0c71db"} Dec 13 07:04:12 crc kubenswrapper[4971]: I1213 07:04:12.186279 4971 generic.go:334] "Generic (PLEG): container finished" podID="97cf4a02-e236-49c1-904f-21b25e87eba4" containerID="47325854a36fbd957987f114286b98caf8038283e15ce75adc5c479b0787910c" exitCode=0 Dec 13 07:04:12 crc kubenswrapper[4971]: I1213 07:04:12.186337 4971 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-4pzkf" event={"ID":"97cf4a02-e236-49c1-904f-21b25e87eba4","Type":"ContainerDied","Data":"47325854a36fbd957987f114286b98caf8038283e15ce75adc5c479b0787910c"} Dec 13 07:04:12 crc kubenswrapper[4971]: I1213 07:04:12.186359 4971 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-4pzkf" event={"ID":"97cf4a02-e236-49c1-904f-21b25e87eba4","Type":"ContainerDied","Data":"8b5a62fad31d2bf2ba6c9d85a43034e6ad383943670a489067b36c2011e02513"} Dec 13 07:04:12 crc kubenswrapper[4971]: I1213 07:04:12.186424 4971 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-4pzkf" Dec 13 07:04:12 crc kubenswrapper[4971]: I1213 07:04:12.192761 4971 generic.go:334] "Generic (PLEG): container finished" podID="5eea3e58-f9fb-4529-b645-0f5193d5a8a9" containerID="937af01c300b74d0c52a530d08bf8ad033b99588471fb7be33345a2e3fb07a39" exitCode=0 Dec 13 07:04:12 crc kubenswrapper[4971]: I1213 07:04:12.192851 4971 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-68kxv" event={"ID":"5eea3e58-f9fb-4529-b645-0f5193d5a8a9","Type":"ContainerDied","Data":"937af01c300b74d0c52a530d08bf8ad033b99588471fb7be33345a2e3fb07a39"} Dec 13 07:04:12 crc kubenswrapper[4971]: I1213 07:04:12.192892 4971 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-68kxv" event={"ID":"5eea3e58-f9fb-4529-b645-0f5193d5a8a9","Type":"ContainerDied","Data":"b37e9a06348fd99e96df1f61c0823b165c6ca5197051b3514fe96e71ec77cc6d"} Dec 13 07:04:12 crc kubenswrapper[4971]: I1213 07:04:12.193017 4971 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-68kxv" Dec 13 07:04:12 crc kubenswrapper[4971]: I1213 07:04:12.194349 4971 scope.go:117] "RemoveContainer" containerID="286c449dc5a9a113f79f93efbaf4070fc4e73a43894b9a9329a6c56a40aef05b" Dec 13 07:04:12 crc kubenswrapper[4971]: I1213 07:04:12.196445 4971 generic.go:334] "Generic (PLEG): container finished" podID="d44cf79f-dbdf-4e9b-ba48-7351d398c38c" containerID="0cbd2685b75d6e865d414a1ccb83ec0b3fa61e649b012dabdb47bd9f09c8c2c2" exitCode=0 Dec 13 07:04:12 crc kubenswrapper[4971]: I1213 07:04:12.196531 4971 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-kbz6d" Dec 13 07:04:12 crc kubenswrapper[4971]: I1213 07:04:12.196529 4971 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-kbz6d" event={"ID":"d44cf79f-dbdf-4e9b-ba48-7351d398c38c","Type":"ContainerDied","Data":"0cbd2685b75d6e865d414a1ccb83ec0b3fa61e649b012dabdb47bd9f09c8c2c2"} Dec 13 07:04:12 crc kubenswrapper[4971]: I1213 07:04:12.196809 4971 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-kbz6d" event={"ID":"d44cf79f-dbdf-4e9b-ba48-7351d398c38c","Type":"ContainerDied","Data":"764b5c5c96560b3780d70e038b8c532f9b2a02615adefa94de4d3a0d72234cd0"} Dec 13 07:04:12 crc kubenswrapper[4971]: I1213 07:04:12.203878 4971 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/marketplace-operator-79b997595-w4r9w" podStartSLOduration=1.203857777 podStartE2EDuration="1.203857777s" podCreationTimestamp="2025-12-13 07:04:11 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-13 07:04:12.201776236 +0000 UTC m=+908.806185684" watchObservedRunningTime="2025-12-13 07:04:12.203857777 +0000 UTC m=+908.808267225" Dec 13 07:04:12 crc kubenswrapper[4971]: I1213 07:04:12.209109 4971 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-5kgpk" event={"ID":"7e9ec79e-730d-4d91-b8f9-851c2b1cb769","Type":"ContainerStarted","Data":"8e90f48a6de8e7d6802430a295023cc6f712227c008c2a36e5a543cbc2c5958f"} Dec 13 07:04:12 crc kubenswrapper[4971]: I1213 07:04:12.209455 4971 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/community-operators-5kgpk" podUID="7e9ec79e-730d-4d91-b8f9-851c2b1cb769" containerName="registry-server" containerID="cri-o://8e90f48a6de8e7d6802430a295023cc6f712227c008c2a36e5a543cbc2c5958f" gracePeriod=30 Dec 13 07:04:12 crc kubenswrapper[4971]: I1213 07:04:12.224995 4971 scope.go:117] "RemoveContainer" containerID="c61b97c1755702dffd406eec5db5b5f8bb5f4fa4b6d081c8392bce7eb961c949" Dec 13 07:04:12 crc kubenswrapper[4971]: E1213 07:04:12.225641 4971 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"c61b97c1755702dffd406eec5db5b5f8bb5f4fa4b6d081c8392bce7eb961c949\": container with ID starting with c61b97c1755702dffd406eec5db5b5f8bb5f4fa4b6d081c8392bce7eb961c949 not found: ID does not exist" containerID="c61b97c1755702dffd406eec5db5b5f8bb5f4fa4b6d081c8392bce7eb961c949" Dec 13 07:04:12 crc kubenswrapper[4971]: I1213 07:04:12.225692 4971 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"c61b97c1755702dffd406eec5db5b5f8bb5f4fa4b6d081c8392bce7eb961c949"} err="failed to get container status \"c61b97c1755702dffd406eec5db5b5f8bb5f4fa4b6d081c8392bce7eb961c949\": rpc error: code = NotFound desc = could not find container \"c61b97c1755702dffd406eec5db5b5f8bb5f4fa4b6d081c8392bce7eb961c949\": container with ID starting with c61b97c1755702dffd406eec5db5b5f8bb5f4fa4b6d081c8392bce7eb961c949 not found: ID does not exist" Dec 13 07:04:12 crc kubenswrapper[4971]: I1213 07:04:12.225721 4971 scope.go:117] "RemoveContainer" containerID="286c449dc5a9a113f79f93efbaf4070fc4e73a43894b9a9329a6c56a40aef05b" Dec 13 07:04:12 crc kubenswrapper[4971]: E1213 07:04:12.226053 4971 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"286c449dc5a9a113f79f93efbaf4070fc4e73a43894b9a9329a6c56a40aef05b\": container with ID starting with 286c449dc5a9a113f79f93efbaf4070fc4e73a43894b9a9329a6c56a40aef05b not found: ID does not exist" containerID="286c449dc5a9a113f79f93efbaf4070fc4e73a43894b9a9329a6c56a40aef05b" Dec 13 07:04:12 crc kubenswrapper[4971]: I1213 07:04:12.226069 4971 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"286c449dc5a9a113f79f93efbaf4070fc4e73a43894b9a9329a6c56a40aef05b"} err="failed to get container status \"286c449dc5a9a113f79f93efbaf4070fc4e73a43894b9a9329a6c56a40aef05b\": rpc error: code = NotFound desc = could not find container \"286c449dc5a9a113f79f93efbaf4070fc4e73a43894b9a9329a6c56a40aef05b\": container with ID starting with 286c449dc5a9a113f79f93efbaf4070fc4e73a43894b9a9329a6c56a40aef05b not found: ID does not exist" Dec 13 07:04:12 crc kubenswrapper[4971]: I1213 07:04:12.226080 4971 scope.go:117] "RemoveContainer" containerID="91507b889fb1db4fed8d82e2f27769c51087db01b1cd1c9b2b06d317b8ef3ff0" Dec 13 07:04:12 crc kubenswrapper[4971]: I1213 07:04:12.226986 4971 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/marketplace-operator-79b997595-6rvc9"] Dec 13 07:04:12 crc kubenswrapper[4971]: I1213 07:04:12.230693 4971 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/marketplace-operator-79b997595-6rvc9"] Dec 13 07:04:12 crc kubenswrapper[4971]: I1213 07:04:12.238702 4971 scope.go:117] "RemoveContainer" containerID="eb6d0a86233100a22e7b2f97ac48557afe6abc4b331a3f5a0d445a5986cab038" Dec 13 07:04:12 crc kubenswrapper[4971]: I1213 07:04:12.253297 4971 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-5qtd2"] Dec 13 07:04:12 crc kubenswrapper[4971]: I1213 07:04:12.260899 4971 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-operators-5qtd2"] Dec 13 07:04:12 crc kubenswrapper[4971]: I1213 07:04:12.265011 4971 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-68kxv"] Dec 13 07:04:12 crc kubenswrapper[4971]: I1213 07:04:12.271682 4971 scope.go:117] "RemoveContainer" containerID="33ef1eb9944bba090fda89c5cbe65281414e52f467a649d0395eae60a5916a4f" Dec 13 07:04:12 crc kubenswrapper[4971]: I1213 07:04:12.274762 4971 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-marketplace-68kxv"] Dec 13 07:04:12 crc kubenswrapper[4971]: I1213 07:04:12.286917 4971 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/community-operators-5kgpk" podStartSLOduration=1.556086032 podStartE2EDuration="4.286897552s" podCreationTimestamp="2025-12-13 07:04:08 +0000 UTC" firstStartedPulling="2025-12-13 07:04:09.156407211 +0000 UTC m=+905.760816659" lastFinishedPulling="2025-12-13 07:04:11.887218731 +0000 UTC m=+908.491628179" observedRunningTime="2025-12-13 07:04:12.282041943 +0000 UTC m=+908.886451391" watchObservedRunningTime="2025-12-13 07:04:12.286897552 +0000 UTC m=+908.891307010" Dec 13 07:04:12 crc kubenswrapper[4971]: I1213 07:04:12.300546 4971 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-4pzkf"] Dec 13 07:04:12 crc kubenswrapper[4971]: I1213 07:04:12.306319 4971 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/certified-operators-4pzkf"] Dec 13 07:04:12 crc kubenswrapper[4971]: I1213 07:04:12.311475 4971 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-kbz6d"] Dec 13 07:04:12 crc kubenswrapper[4971]: I1213 07:04:12.316067 4971 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/community-operators-kbz6d"] Dec 13 07:04:12 crc kubenswrapper[4971]: I1213 07:04:12.316886 4971 scope.go:117] "RemoveContainer" containerID="91507b889fb1db4fed8d82e2f27769c51087db01b1cd1c9b2b06d317b8ef3ff0" Dec 13 07:04:12 crc kubenswrapper[4971]: E1213 07:04:12.317360 4971 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"91507b889fb1db4fed8d82e2f27769c51087db01b1cd1c9b2b06d317b8ef3ff0\": container with ID starting with 91507b889fb1db4fed8d82e2f27769c51087db01b1cd1c9b2b06d317b8ef3ff0 not found: ID does not exist" containerID="91507b889fb1db4fed8d82e2f27769c51087db01b1cd1c9b2b06d317b8ef3ff0" Dec 13 07:04:12 crc kubenswrapper[4971]: I1213 07:04:12.317401 4971 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"91507b889fb1db4fed8d82e2f27769c51087db01b1cd1c9b2b06d317b8ef3ff0"} err="failed to get container status \"91507b889fb1db4fed8d82e2f27769c51087db01b1cd1c9b2b06d317b8ef3ff0\": rpc error: code = NotFound desc = could not find container \"91507b889fb1db4fed8d82e2f27769c51087db01b1cd1c9b2b06d317b8ef3ff0\": container with ID starting with 91507b889fb1db4fed8d82e2f27769c51087db01b1cd1c9b2b06d317b8ef3ff0 not found: ID does not exist" Dec 13 07:04:12 crc kubenswrapper[4971]: I1213 07:04:12.317429 4971 scope.go:117] "RemoveContainer" containerID="eb6d0a86233100a22e7b2f97ac48557afe6abc4b331a3f5a0d445a5986cab038" Dec 13 07:04:12 crc kubenswrapper[4971]: E1213 07:04:12.318072 4971 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"eb6d0a86233100a22e7b2f97ac48557afe6abc4b331a3f5a0d445a5986cab038\": container with ID starting with eb6d0a86233100a22e7b2f97ac48557afe6abc4b331a3f5a0d445a5986cab038 not found: ID does not exist" containerID="eb6d0a86233100a22e7b2f97ac48557afe6abc4b331a3f5a0d445a5986cab038" Dec 13 07:04:12 crc kubenswrapper[4971]: I1213 07:04:12.318103 4971 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"eb6d0a86233100a22e7b2f97ac48557afe6abc4b331a3f5a0d445a5986cab038"} err="failed to get container status \"eb6d0a86233100a22e7b2f97ac48557afe6abc4b331a3f5a0d445a5986cab038\": rpc error: code = NotFound desc = could not find container \"eb6d0a86233100a22e7b2f97ac48557afe6abc4b331a3f5a0d445a5986cab038\": container with ID starting with eb6d0a86233100a22e7b2f97ac48557afe6abc4b331a3f5a0d445a5986cab038 not found: ID does not exist" Dec 13 07:04:12 crc kubenswrapper[4971]: I1213 07:04:12.318121 4971 scope.go:117] "RemoveContainer" containerID="33ef1eb9944bba090fda89c5cbe65281414e52f467a649d0395eae60a5916a4f" Dec 13 07:04:12 crc kubenswrapper[4971]: E1213 07:04:12.318378 4971 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"33ef1eb9944bba090fda89c5cbe65281414e52f467a649d0395eae60a5916a4f\": container with ID starting with 33ef1eb9944bba090fda89c5cbe65281414e52f467a649d0395eae60a5916a4f not found: ID does not exist" containerID="33ef1eb9944bba090fda89c5cbe65281414e52f467a649d0395eae60a5916a4f" Dec 13 07:04:12 crc kubenswrapper[4971]: I1213 07:04:12.318422 4971 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"33ef1eb9944bba090fda89c5cbe65281414e52f467a649d0395eae60a5916a4f"} err="failed to get container status \"33ef1eb9944bba090fda89c5cbe65281414e52f467a649d0395eae60a5916a4f\": rpc error: code = NotFound desc = could not find container \"33ef1eb9944bba090fda89c5cbe65281414e52f467a649d0395eae60a5916a4f\": container with ID starting with 33ef1eb9944bba090fda89c5cbe65281414e52f467a649d0395eae60a5916a4f not found: ID does not exist" Dec 13 07:04:12 crc kubenswrapper[4971]: I1213 07:04:12.318450 4971 scope.go:117] "RemoveContainer" containerID="47325854a36fbd957987f114286b98caf8038283e15ce75adc5c479b0787910c" Dec 13 07:04:12 crc kubenswrapper[4971]: I1213 07:04:12.332729 4971 scope.go:117] "RemoveContainer" containerID="39bded6183ae1297006f892ec4367021f144a8659108a22755ab06e082f8574f" Dec 13 07:04:12 crc kubenswrapper[4971]: I1213 07:04:12.349962 4971 scope.go:117] "RemoveContainer" containerID="5eeb9f49b31482f869b7d83a6c04554577d9b80da86141ce746ec2edfdce19c5" Dec 13 07:04:12 crc kubenswrapper[4971]: I1213 07:04:12.393069 4971 scope.go:117] "RemoveContainer" containerID="47325854a36fbd957987f114286b98caf8038283e15ce75adc5c479b0787910c" Dec 13 07:04:12 crc kubenswrapper[4971]: E1213 07:04:12.393384 4971 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"47325854a36fbd957987f114286b98caf8038283e15ce75adc5c479b0787910c\": container with ID starting with 47325854a36fbd957987f114286b98caf8038283e15ce75adc5c479b0787910c not found: ID does not exist" containerID="47325854a36fbd957987f114286b98caf8038283e15ce75adc5c479b0787910c" Dec 13 07:04:12 crc kubenswrapper[4971]: I1213 07:04:12.393420 4971 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"47325854a36fbd957987f114286b98caf8038283e15ce75adc5c479b0787910c"} err="failed to get container status \"47325854a36fbd957987f114286b98caf8038283e15ce75adc5c479b0787910c\": rpc error: code = NotFound desc = could not find container \"47325854a36fbd957987f114286b98caf8038283e15ce75adc5c479b0787910c\": container with ID starting with 47325854a36fbd957987f114286b98caf8038283e15ce75adc5c479b0787910c not found: ID does not exist" Dec 13 07:04:12 crc kubenswrapper[4971]: I1213 07:04:12.393443 4971 scope.go:117] "RemoveContainer" containerID="39bded6183ae1297006f892ec4367021f144a8659108a22755ab06e082f8574f" Dec 13 07:04:12 crc kubenswrapper[4971]: E1213 07:04:12.393663 4971 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"39bded6183ae1297006f892ec4367021f144a8659108a22755ab06e082f8574f\": container with ID starting with 39bded6183ae1297006f892ec4367021f144a8659108a22755ab06e082f8574f not found: ID does not exist" containerID="39bded6183ae1297006f892ec4367021f144a8659108a22755ab06e082f8574f" Dec 13 07:04:12 crc kubenswrapper[4971]: I1213 07:04:12.393697 4971 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"39bded6183ae1297006f892ec4367021f144a8659108a22755ab06e082f8574f"} err="failed to get container status \"39bded6183ae1297006f892ec4367021f144a8659108a22755ab06e082f8574f\": rpc error: code = NotFound desc = could not find container \"39bded6183ae1297006f892ec4367021f144a8659108a22755ab06e082f8574f\": container with ID starting with 39bded6183ae1297006f892ec4367021f144a8659108a22755ab06e082f8574f not found: ID does not exist" Dec 13 07:04:12 crc kubenswrapper[4971]: I1213 07:04:12.393711 4971 scope.go:117] "RemoveContainer" containerID="5eeb9f49b31482f869b7d83a6c04554577d9b80da86141ce746ec2edfdce19c5" Dec 13 07:04:12 crc kubenswrapper[4971]: E1213 07:04:12.393946 4971 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"5eeb9f49b31482f869b7d83a6c04554577d9b80da86141ce746ec2edfdce19c5\": container with ID starting with 5eeb9f49b31482f869b7d83a6c04554577d9b80da86141ce746ec2edfdce19c5 not found: ID does not exist" containerID="5eeb9f49b31482f869b7d83a6c04554577d9b80da86141ce746ec2edfdce19c5" Dec 13 07:04:12 crc kubenswrapper[4971]: I1213 07:04:12.393969 4971 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"5eeb9f49b31482f869b7d83a6c04554577d9b80da86141ce746ec2edfdce19c5"} err="failed to get container status \"5eeb9f49b31482f869b7d83a6c04554577d9b80da86141ce746ec2edfdce19c5\": rpc error: code = NotFound desc = could not find container \"5eeb9f49b31482f869b7d83a6c04554577d9b80da86141ce746ec2edfdce19c5\": container with ID starting with 5eeb9f49b31482f869b7d83a6c04554577d9b80da86141ce746ec2edfdce19c5 not found: ID does not exist" Dec 13 07:04:12 crc kubenswrapper[4971]: I1213 07:04:12.393982 4971 scope.go:117] "RemoveContainer" containerID="937af01c300b74d0c52a530d08bf8ad033b99588471fb7be33345a2e3fb07a39" Dec 13 07:04:12 crc kubenswrapper[4971]: I1213 07:04:12.412755 4971 scope.go:117] "RemoveContainer" containerID="48d12ce796ac32c962947d15f042cb190a9dacbd93a6c1aa9892bf9301e57461" Dec 13 07:04:12 crc kubenswrapper[4971]: I1213 07:04:12.425742 4971 scope.go:117] "RemoveContainer" containerID="6566c640713f28f685b0401d2835ce790afd33ffaabf644d1b149c8f84e21298" Dec 13 07:04:12 crc kubenswrapper[4971]: I1213 07:04:12.466753 4971 scope.go:117] "RemoveContainer" containerID="937af01c300b74d0c52a530d08bf8ad033b99588471fb7be33345a2e3fb07a39" Dec 13 07:04:12 crc kubenswrapper[4971]: E1213 07:04:12.467204 4971 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"937af01c300b74d0c52a530d08bf8ad033b99588471fb7be33345a2e3fb07a39\": container with ID starting with 937af01c300b74d0c52a530d08bf8ad033b99588471fb7be33345a2e3fb07a39 not found: ID does not exist" containerID="937af01c300b74d0c52a530d08bf8ad033b99588471fb7be33345a2e3fb07a39" Dec 13 07:04:12 crc kubenswrapper[4971]: I1213 07:04:12.467258 4971 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"937af01c300b74d0c52a530d08bf8ad033b99588471fb7be33345a2e3fb07a39"} err="failed to get container status \"937af01c300b74d0c52a530d08bf8ad033b99588471fb7be33345a2e3fb07a39\": rpc error: code = NotFound desc = could not find container \"937af01c300b74d0c52a530d08bf8ad033b99588471fb7be33345a2e3fb07a39\": container with ID starting with 937af01c300b74d0c52a530d08bf8ad033b99588471fb7be33345a2e3fb07a39 not found: ID does not exist" Dec 13 07:04:12 crc kubenswrapper[4971]: I1213 07:04:12.467294 4971 scope.go:117] "RemoveContainer" containerID="48d12ce796ac32c962947d15f042cb190a9dacbd93a6c1aa9892bf9301e57461" Dec 13 07:04:12 crc kubenswrapper[4971]: E1213 07:04:12.467870 4971 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"48d12ce796ac32c962947d15f042cb190a9dacbd93a6c1aa9892bf9301e57461\": container with ID starting with 48d12ce796ac32c962947d15f042cb190a9dacbd93a6c1aa9892bf9301e57461 not found: ID does not exist" containerID="48d12ce796ac32c962947d15f042cb190a9dacbd93a6c1aa9892bf9301e57461" Dec 13 07:04:12 crc kubenswrapper[4971]: I1213 07:04:12.467903 4971 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"48d12ce796ac32c962947d15f042cb190a9dacbd93a6c1aa9892bf9301e57461"} err="failed to get container status \"48d12ce796ac32c962947d15f042cb190a9dacbd93a6c1aa9892bf9301e57461\": rpc error: code = NotFound desc = could not find container \"48d12ce796ac32c962947d15f042cb190a9dacbd93a6c1aa9892bf9301e57461\": container with ID starting with 48d12ce796ac32c962947d15f042cb190a9dacbd93a6c1aa9892bf9301e57461 not found: ID does not exist" Dec 13 07:04:12 crc kubenswrapper[4971]: I1213 07:04:12.467924 4971 scope.go:117] "RemoveContainer" containerID="6566c640713f28f685b0401d2835ce790afd33ffaabf644d1b149c8f84e21298" Dec 13 07:04:12 crc kubenswrapper[4971]: E1213 07:04:12.468207 4971 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"6566c640713f28f685b0401d2835ce790afd33ffaabf644d1b149c8f84e21298\": container with ID starting with 6566c640713f28f685b0401d2835ce790afd33ffaabf644d1b149c8f84e21298 not found: ID does not exist" containerID="6566c640713f28f685b0401d2835ce790afd33ffaabf644d1b149c8f84e21298" Dec 13 07:04:12 crc kubenswrapper[4971]: I1213 07:04:12.468243 4971 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"6566c640713f28f685b0401d2835ce790afd33ffaabf644d1b149c8f84e21298"} err="failed to get container status \"6566c640713f28f685b0401d2835ce790afd33ffaabf644d1b149c8f84e21298\": rpc error: code = NotFound desc = could not find container \"6566c640713f28f685b0401d2835ce790afd33ffaabf644d1b149c8f84e21298\": container with ID starting with 6566c640713f28f685b0401d2835ce790afd33ffaabf644d1b149c8f84e21298 not found: ID does not exist" Dec 13 07:04:12 crc kubenswrapper[4971]: I1213 07:04:12.468263 4971 scope.go:117] "RemoveContainer" containerID="0cbd2685b75d6e865d414a1ccb83ec0b3fa61e649b012dabdb47bd9f09c8c2c2" Dec 13 07:04:12 crc kubenswrapper[4971]: I1213 07:04:12.482288 4971 scope.go:117] "RemoveContainer" containerID="9c03c8d63a2c6e9b660e112dcfa1f170515d6921c706e6db2bfc1bb20df69234" Dec 13 07:04:12 crc kubenswrapper[4971]: I1213 07:04:12.497315 4971 scope.go:117] "RemoveContainer" containerID="f1d0237bf148bee90f27653f9d643336dd61d6e7de7d05f7bd4324aeaca9dffb" Dec 13 07:04:12 crc kubenswrapper[4971]: I1213 07:04:12.512905 4971 scope.go:117] "RemoveContainer" containerID="0cbd2685b75d6e865d414a1ccb83ec0b3fa61e649b012dabdb47bd9f09c8c2c2" Dec 13 07:04:12 crc kubenswrapper[4971]: E1213 07:04:12.513251 4971 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"0cbd2685b75d6e865d414a1ccb83ec0b3fa61e649b012dabdb47bd9f09c8c2c2\": container with ID starting with 0cbd2685b75d6e865d414a1ccb83ec0b3fa61e649b012dabdb47bd9f09c8c2c2 not found: ID does not exist" containerID="0cbd2685b75d6e865d414a1ccb83ec0b3fa61e649b012dabdb47bd9f09c8c2c2" Dec 13 07:04:12 crc kubenswrapper[4971]: I1213 07:04:12.513306 4971 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"0cbd2685b75d6e865d414a1ccb83ec0b3fa61e649b012dabdb47bd9f09c8c2c2"} err="failed to get container status \"0cbd2685b75d6e865d414a1ccb83ec0b3fa61e649b012dabdb47bd9f09c8c2c2\": rpc error: code = NotFound desc = could not find container \"0cbd2685b75d6e865d414a1ccb83ec0b3fa61e649b012dabdb47bd9f09c8c2c2\": container with ID starting with 0cbd2685b75d6e865d414a1ccb83ec0b3fa61e649b012dabdb47bd9f09c8c2c2 not found: ID does not exist" Dec 13 07:04:12 crc kubenswrapper[4971]: I1213 07:04:12.513340 4971 scope.go:117] "RemoveContainer" containerID="9c03c8d63a2c6e9b660e112dcfa1f170515d6921c706e6db2bfc1bb20df69234" Dec 13 07:04:12 crc kubenswrapper[4971]: E1213 07:04:12.513726 4971 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"9c03c8d63a2c6e9b660e112dcfa1f170515d6921c706e6db2bfc1bb20df69234\": container with ID starting with 9c03c8d63a2c6e9b660e112dcfa1f170515d6921c706e6db2bfc1bb20df69234 not found: ID does not exist" containerID="9c03c8d63a2c6e9b660e112dcfa1f170515d6921c706e6db2bfc1bb20df69234" Dec 13 07:04:12 crc kubenswrapper[4971]: I1213 07:04:12.513757 4971 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"9c03c8d63a2c6e9b660e112dcfa1f170515d6921c706e6db2bfc1bb20df69234"} err="failed to get container status \"9c03c8d63a2c6e9b660e112dcfa1f170515d6921c706e6db2bfc1bb20df69234\": rpc error: code = NotFound desc = could not find container \"9c03c8d63a2c6e9b660e112dcfa1f170515d6921c706e6db2bfc1bb20df69234\": container with ID starting with 9c03c8d63a2c6e9b660e112dcfa1f170515d6921c706e6db2bfc1bb20df69234 not found: ID does not exist" Dec 13 07:04:12 crc kubenswrapper[4971]: I1213 07:04:12.513776 4971 scope.go:117] "RemoveContainer" containerID="f1d0237bf148bee90f27653f9d643336dd61d6e7de7d05f7bd4324aeaca9dffb" Dec 13 07:04:12 crc kubenswrapper[4971]: E1213 07:04:12.513967 4971 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"f1d0237bf148bee90f27653f9d643336dd61d6e7de7d05f7bd4324aeaca9dffb\": container with ID starting with f1d0237bf148bee90f27653f9d643336dd61d6e7de7d05f7bd4324aeaca9dffb not found: ID does not exist" containerID="f1d0237bf148bee90f27653f9d643336dd61d6e7de7d05f7bd4324aeaca9dffb" Dec 13 07:04:12 crc kubenswrapper[4971]: I1213 07:04:12.513994 4971 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"f1d0237bf148bee90f27653f9d643336dd61d6e7de7d05f7bd4324aeaca9dffb"} err="failed to get container status \"f1d0237bf148bee90f27653f9d643336dd61d6e7de7d05f7bd4324aeaca9dffb\": rpc error: code = NotFound desc = could not find container \"f1d0237bf148bee90f27653f9d643336dd61d6e7de7d05f7bd4324aeaca9dffb\": container with ID starting with f1d0237bf148bee90f27653f9d643336dd61d6e7de7d05f7bd4324aeaca9dffb not found: ID does not exist" Dec 13 07:04:12 crc kubenswrapper[4971]: I1213 07:04:12.521386 4971 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_community-operators-5kgpk_7e9ec79e-730d-4d91-b8f9-851c2b1cb769/registry-server/0.log" Dec 13 07:04:12 crc kubenswrapper[4971]: I1213 07:04:12.522278 4971 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-5kgpk" Dec 13 07:04:12 crc kubenswrapper[4971]: I1213 07:04:12.690387 4971 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/7e9ec79e-730d-4d91-b8f9-851c2b1cb769-catalog-content\") pod \"7e9ec79e-730d-4d91-b8f9-851c2b1cb769\" (UID: \"7e9ec79e-730d-4d91-b8f9-851c2b1cb769\") " Dec 13 07:04:12 crc kubenswrapper[4971]: I1213 07:04:12.690733 4971 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/7e9ec79e-730d-4d91-b8f9-851c2b1cb769-utilities\") pod \"7e9ec79e-730d-4d91-b8f9-851c2b1cb769\" (UID: \"7e9ec79e-730d-4d91-b8f9-851c2b1cb769\") " Dec 13 07:04:12 crc kubenswrapper[4971]: I1213 07:04:12.690781 4971 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-7fx8q\" (UniqueName: \"kubernetes.io/projected/7e9ec79e-730d-4d91-b8f9-851c2b1cb769-kube-api-access-7fx8q\") pod \"7e9ec79e-730d-4d91-b8f9-851c2b1cb769\" (UID: \"7e9ec79e-730d-4d91-b8f9-851c2b1cb769\") " Dec 13 07:04:12 crc kubenswrapper[4971]: I1213 07:04:12.691620 4971 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/7e9ec79e-730d-4d91-b8f9-851c2b1cb769-utilities" (OuterVolumeSpecName: "utilities") pod "7e9ec79e-730d-4d91-b8f9-851c2b1cb769" (UID: "7e9ec79e-730d-4d91-b8f9-851c2b1cb769"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 13 07:04:12 crc kubenswrapper[4971]: I1213 07:04:12.695947 4971 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/7e9ec79e-730d-4d91-b8f9-851c2b1cb769-kube-api-access-7fx8q" (OuterVolumeSpecName: "kube-api-access-7fx8q") pod "7e9ec79e-730d-4d91-b8f9-851c2b1cb769" (UID: "7e9ec79e-730d-4d91-b8f9-851c2b1cb769"). InnerVolumeSpecName "kube-api-access-7fx8q". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 13 07:04:12 crc kubenswrapper[4971]: I1213 07:04:12.741960 4971 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/7e9ec79e-730d-4d91-b8f9-851c2b1cb769-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "7e9ec79e-730d-4d91-b8f9-851c2b1cb769" (UID: "7e9ec79e-730d-4d91-b8f9-851c2b1cb769"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 13 07:04:12 crc kubenswrapper[4971]: I1213 07:04:12.792485 4971 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/7e9ec79e-730d-4d91-b8f9-851c2b1cb769-utilities\") on node \"crc\" DevicePath \"\"" Dec 13 07:04:12 crc kubenswrapper[4971]: I1213 07:04:12.792536 4971 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-7fx8q\" (UniqueName: \"kubernetes.io/projected/7e9ec79e-730d-4d91-b8f9-851c2b1cb769-kube-api-access-7fx8q\") on node \"crc\" DevicePath \"\"" Dec 13 07:04:12 crc kubenswrapper[4971]: I1213 07:04:12.792547 4971 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/7e9ec79e-730d-4d91-b8f9-851c2b1cb769-catalog-content\") on node \"crc\" DevicePath \"\"" Dec 13 07:04:13 crc kubenswrapper[4971]: I1213 07:04:13.217007 4971 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_community-operators-5kgpk_7e9ec79e-730d-4d91-b8f9-851c2b1cb769/registry-server/0.log" Dec 13 07:04:13 crc kubenswrapper[4971]: I1213 07:04:13.217643 4971 generic.go:334] "Generic (PLEG): container finished" podID="7e9ec79e-730d-4d91-b8f9-851c2b1cb769" containerID="8e90f48a6de8e7d6802430a295023cc6f712227c008c2a36e5a543cbc2c5958f" exitCode=1 Dec 13 07:04:13 crc kubenswrapper[4971]: I1213 07:04:13.217685 4971 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-5kgpk" event={"ID":"7e9ec79e-730d-4d91-b8f9-851c2b1cb769","Type":"ContainerDied","Data":"8e90f48a6de8e7d6802430a295023cc6f712227c008c2a36e5a543cbc2c5958f"} Dec 13 07:04:13 crc kubenswrapper[4971]: I1213 07:04:13.217709 4971 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-5kgpk" event={"ID":"7e9ec79e-730d-4d91-b8f9-851c2b1cb769","Type":"ContainerDied","Data":"637cbbc8b0da971b274574b1075e1e413f27fa5283a753b58f382f1c4980eecf"} Dec 13 07:04:13 crc kubenswrapper[4971]: I1213 07:04:13.217725 4971 scope.go:117] "RemoveContainer" containerID="8e90f48a6de8e7d6802430a295023cc6f712227c008c2a36e5a543cbc2c5958f" Dec 13 07:04:13 crc kubenswrapper[4971]: I1213 07:04:13.217823 4971 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-5kgpk" Dec 13 07:04:13 crc kubenswrapper[4971]: I1213 07:04:13.224065 4971 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/marketplace-operator-79b997595-w4r9w" Dec 13 07:04:13 crc kubenswrapper[4971]: I1213 07:04:13.229883 4971 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/marketplace-operator-79b997595-w4r9w" Dec 13 07:04:13 crc kubenswrapper[4971]: I1213 07:04:13.234632 4971 scope.go:117] "RemoveContainer" containerID="a5b55297439bc82bb4cc7c6eb000526c88488ac57a9f9efad54f0b90a45993f6" Dec 13 07:04:13 crc kubenswrapper[4971]: I1213 07:04:13.268712 4971 scope.go:117] "RemoveContainer" containerID="2123533c6c514a612c02c3519057f2c3521d7a1afa3187b6cd26d425858d5230" Dec 13 07:04:13 crc kubenswrapper[4971]: I1213 07:04:13.273841 4971 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-5kgpk"] Dec 13 07:04:13 crc kubenswrapper[4971]: I1213 07:04:13.278622 4971 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/community-operators-5kgpk"] Dec 13 07:04:13 crc kubenswrapper[4971]: I1213 07:04:13.281883 4971 scope.go:117] "RemoveContainer" containerID="8e90f48a6de8e7d6802430a295023cc6f712227c008c2a36e5a543cbc2c5958f" Dec 13 07:04:13 crc kubenswrapper[4971]: E1213 07:04:13.282414 4971 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"8e90f48a6de8e7d6802430a295023cc6f712227c008c2a36e5a543cbc2c5958f\": container with ID starting with 8e90f48a6de8e7d6802430a295023cc6f712227c008c2a36e5a543cbc2c5958f not found: ID does not exist" containerID="8e90f48a6de8e7d6802430a295023cc6f712227c008c2a36e5a543cbc2c5958f" Dec 13 07:04:13 crc kubenswrapper[4971]: I1213 07:04:13.282444 4971 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"8e90f48a6de8e7d6802430a295023cc6f712227c008c2a36e5a543cbc2c5958f"} err="failed to get container status \"8e90f48a6de8e7d6802430a295023cc6f712227c008c2a36e5a543cbc2c5958f\": rpc error: code = NotFound desc = could not find container \"8e90f48a6de8e7d6802430a295023cc6f712227c008c2a36e5a543cbc2c5958f\": container with ID starting with 8e90f48a6de8e7d6802430a295023cc6f712227c008c2a36e5a543cbc2c5958f not found: ID does not exist" Dec 13 07:04:13 crc kubenswrapper[4971]: I1213 07:04:13.282471 4971 scope.go:117] "RemoveContainer" containerID="a5b55297439bc82bb4cc7c6eb000526c88488ac57a9f9efad54f0b90a45993f6" Dec 13 07:04:13 crc kubenswrapper[4971]: E1213 07:04:13.283076 4971 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"a5b55297439bc82bb4cc7c6eb000526c88488ac57a9f9efad54f0b90a45993f6\": container with ID starting with a5b55297439bc82bb4cc7c6eb000526c88488ac57a9f9efad54f0b90a45993f6 not found: ID does not exist" containerID="a5b55297439bc82bb4cc7c6eb000526c88488ac57a9f9efad54f0b90a45993f6" Dec 13 07:04:13 crc kubenswrapper[4971]: I1213 07:04:13.283102 4971 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"a5b55297439bc82bb4cc7c6eb000526c88488ac57a9f9efad54f0b90a45993f6"} err="failed to get container status \"a5b55297439bc82bb4cc7c6eb000526c88488ac57a9f9efad54f0b90a45993f6\": rpc error: code = NotFound desc = could not find container \"a5b55297439bc82bb4cc7c6eb000526c88488ac57a9f9efad54f0b90a45993f6\": container with ID starting with a5b55297439bc82bb4cc7c6eb000526c88488ac57a9f9efad54f0b90a45993f6 not found: ID does not exist" Dec 13 07:04:13 crc kubenswrapper[4971]: I1213 07:04:13.283119 4971 scope.go:117] "RemoveContainer" containerID="2123533c6c514a612c02c3519057f2c3521d7a1afa3187b6cd26d425858d5230" Dec 13 07:04:13 crc kubenswrapper[4971]: E1213 07:04:13.283372 4971 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"2123533c6c514a612c02c3519057f2c3521d7a1afa3187b6cd26d425858d5230\": container with ID starting with 2123533c6c514a612c02c3519057f2c3521d7a1afa3187b6cd26d425858d5230 not found: ID does not exist" containerID="2123533c6c514a612c02c3519057f2c3521d7a1afa3187b6cd26d425858d5230" Dec 13 07:04:13 crc kubenswrapper[4971]: I1213 07:04:13.283401 4971 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"2123533c6c514a612c02c3519057f2c3521d7a1afa3187b6cd26d425858d5230"} err="failed to get container status \"2123533c6c514a612c02c3519057f2c3521d7a1afa3187b6cd26d425858d5230\": rpc error: code = NotFound desc = could not find container \"2123533c6c514a612c02c3519057f2c3521d7a1afa3187b6cd26d425858d5230\": container with ID starting with 2123533c6c514a612c02c3519057f2c3521d7a1afa3187b6cd26d425858d5230 not found: ID does not exist" Dec 13 07:04:13 crc kubenswrapper[4971]: I1213 07:04:13.535097 4971 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/certified-operators-szqz6"] Dec 13 07:04:13 crc kubenswrapper[4971]: E1213 07:04:13.535283 4971 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="7e9ec79e-730d-4d91-b8f9-851c2b1cb769" containerName="registry-server" Dec 13 07:04:13 crc kubenswrapper[4971]: I1213 07:04:13.535294 4971 state_mem.go:107] "Deleted CPUSet assignment" podUID="7e9ec79e-730d-4d91-b8f9-851c2b1cb769" containerName="registry-server" Dec 13 07:04:13 crc kubenswrapper[4971]: E1213 07:04:13.535305 4971 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="5eea3e58-f9fb-4529-b645-0f5193d5a8a9" containerName="extract-content" Dec 13 07:04:13 crc kubenswrapper[4971]: I1213 07:04:13.535311 4971 state_mem.go:107] "Deleted CPUSet assignment" podUID="5eea3e58-f9fb-4529-b645-0f5193d5a8a9" containerName="extract-content" Dec 13 07:04:13 crc kubenswrapper[4971]: E1213 07:04:13.535321 4971 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="5eea3e58-f9fb-4529-b645-0f5193d5a8a9" containerName="extract-utilities" Dec 13 07:04:13 crc kubenswrapper[4971]: I1213 07:04:13.535327 4971 state_mem.go:107] "Deleted CPUSet assignment" podUID="5eea3e58-f9fb-4529-b645-0f5193d5a8a9" containerName="extract-utilities" Dec 13 07:04:13 crc kubenswrapper[4971]: E1213 07:04:13.535335 4971 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="5eea3e58-f9fb-4529-b645-0f5193d5a8a9" containerName="registry-server" Dec 13 07:04:13 crc kubenswrapper[4971]: I1213 07:04:13.535340 4971 state_mem.go:107] "Deleted CPUSet assignment" podUID="5eea3e58-f9fb-4529-b645-0f5193d5a8a9" containerName="registry-server" Dec 13 07:04:13 crc kubenswrapper[4971]: E1213 07:04:13.535351 4971 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="5c88a864-89a9-4f5e-b713-385170b9bf17" containerName="marketplace-operator" Dec 13 07:04:13 crc kubenswrapper[4971]: I1213 07:04:13.535356 4971 state_mem.go:107] "Deleted CPUSet assignment" podUID="5c88a864-89a9-4f5e-b713-385170b9bf17" containerName="marketplace-operator" Dec 13 07:04:13 crc kubenswrapper[4971]: E1213 07:04:13.535363 4971 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="5c88a864-89a9-4f5e-b713-385170b9bf17" containerName="marketplace-operator" Dec 13 07:04:13 crc kubenswrapper[4971]: I1213 07:04:13.535368 4971 state_mem.go:107] "Deleted CPUSet assignment" podUID="5c88a864-89a9-4f5e-b713-385170b9bf17" containerName="marketplace-operator" Dec 13 07:04:13 crc kubenswrapper[4971]: E1213 07:04:13.535375 4971 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="97cf4a02-e236-49c1-904f-21b25e87eba4" containerName="extract-utilities" Dec 13 07:04:13 crc kubenswrapper[4971]: I1213 07:04:13.535381 4971 state_mem.go:107] "Deleted CPUSet assignment" podUID="97cf4a02-e236-49c1-904f-21b25e87eba4" containerName="extract-utilities" Dec 13 07:04:13 crc kubenswrapper[4971]: E1213 07:04:13.535388 4971 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="97cf4a02-e236-49c1-904f-21b25e87eba4" containerName="registry-server" Dec 13 07:04:13 crc kubenswrapper[4971]: I1213 07:04:13.535394 4971 state_mem.go:107] "Deleted CPUSet assignment" podUID="97cf4a02-e236-49c1-904f-21b25e87eba4" containerName="registry-server" Dec 13 07:04:13 crc kubenswrapper[4971]: E1213 07:04:13.535402 4971 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="7e9ec79e-730d-4d91-b8f9-851c2b1cb769" containerName="extract-content" Dec 13 07:04:13 crc kubenswrapper[4971]: I1213 07:04:13.535408 4971 state_mem.go:107] "Deleted CPUSet assignment" podUID="7e9ec79e-730d-4d91-b8f9-851c2b1cb769" containerName="extract-content" Dec 13 07:04:13 crc kubenswrapper[4971]: E1213 07:04:13.535416 4971 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="97cf4a02-e236-49c1-904f-21b25e87eba4" containerName="extract-content" Dec 13 07:04:13 crc kubenswrapper[4971]: I1213 07:04:13.535422 4971 state_mem.go:107] "Deleted CPUSet assignment" podUID="97cf4a02-e236-49c1-904f-21b25e87eba4" containerName="extract-content" Dec 13 07:04:13 crc kubenswrapper[4971]: E1213 07:04:13.535430 4971 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="9fd587ab-4839-4479-80f8-f97d78d833b9" containerName="extract-utilities" Dec 13 07:04:13 crc kubenswrapper[4971]: I1213 07:04:13.535435 4971 state_mem.go:107] "Deleted CPUSet assignment" podUID="9fd587ab-4839-4479-80f8-f97d78d833b9" containerName="extract-utilities" Dec 13 07:04:13 crc kubenswrapper[4971]: E1213 07:04:13.535442 4971 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="d44cf79f-dbdf-4e9b-ba48-7351d398c38c" containerName="extract-content" Dec 13 07:04:13 crc kubenswrapper[4971]: I1213 07:04:13.535448 4971 state_mem.go:107] "Deleted CPUSet assignment" podUID="d44cf79f-dbdf-4e9b-ba48-7351d398c38c" containerName="extract-content" Dec 13 07:04:13 crc kubenswrapper[4971]: E1213 07:04:13.535460 4971 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="d44cf79f-dbdf-4e9b-ba48-7351d398c38c" containerName="extract-utilities" Dec 13 07:04:13 crc kubenswrapper[4971]: I1213 07:04:13.535466 4971 state_mem.go:107] "Deleted CPUSet assignment" podUID="d44cf79f-dbdf-4e9b-ba48-7351d398c38c" containerName="extract-utilities" Dec 13 07:04:13 crc kubenswrapper[4971]: E1213 07:04:13.535481 4971 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="7e9ec79e-730d-4d91-b8f9-851c2b1cb769" containerName="extract-utilities" Dec 13 07:04:13 crc kubenswrapper[4971]: I1213 07:04:13.535492 4971 state_mem.go:107] "Deleted CPUSet assignment" podUID="7e9ec79e-730d-4d91-b8f9-851c2b1cb769" containerName="extract-utilities" Dec 13 07:04:13 crc kubenswrapper[4971]: E1213 07:04:13.535503 4971 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="5c88a864-89a9-4f5e-b713-385170b9bf17" containerName="marketplace-operator" Dec 13 07:04:13 crc kubenswrapper[4971]: I1213 07:04:13.535510 4971 state_mem.go:107] "Deleted CPUSet assignment" podUID="5c88a864-89a9-4f5e-b713-385170b9bf17" containerName="marketplace-operator" Dec 13 07:04:13 crc kubenswrapper[4971]: E1213 07:04:13.535547 4971 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="9fd587ab-4839-4479-80f8-f97d78d833b9" containerName="registry-server" Dec 13 07:04:13 crc kubenswrapper[4971]: I1213 07:04:13.535562 4971 state_mem.go:107] "Deleted CPUSet assignment" podUID="9fd587ab-4839-4479-80f8-f97d78d833b9" containerName="registry-server" Dec 13 07:04:13 crc kubenswrapper[4971]: E1213 07:04:13.535575 4971 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="9fd587ab-4839-4479-80f8-f97d78d833b9" containerName="extract-content" Dec 13 07:04:13 crc kubenswrapper[4971]: I1213 07:04:13.535773 4971 state_mem.go:107] "Deleted CPUSet assignment" podUID="9fd587ab-4839-4479-80f8-f97d78d833b9" containerName="extract-content" Dec 13 07:04:13 crc kubenswrapper[4971]: E1213 07:04:13.535781 4971 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="d44cf79f-dbdf-4e9b-ba48-7351d398c38c" containerName="registry-server" Dec 13 07:04:13 crc kubenswrapper[4971]: I1213 07:04:13.535787 4971 state_mem.go:107] "Deleted CPUSet assignment" podUID="d44cf79f-dbdf-4e9b-ba48-7351d398c38c" containerName="registry-server" Dec 13 07:04:13 crc kubenswrapper[4971]: I1213 07:04:13.535957 4971 memory_manager.go:354] "RemoveStaleState removing state" podUID="5c88a864-89a9-4f5e-b713-385170b9bf17" containerName="marketplace-operator" Dec 13 07:04:13 crc kubenswrapper[4971]: I1213 07:04:13.535976 4971 memory_manager.go:354] "RemoveStaleState removing state" podUID="5c88a864-89a9-4f5e-b713-385170b9bf17" containerName="marketplace-operator" Dec 13 07:04:13 crc kubenswrapper[4971]: I1213 07:04:13.535988 4971 memory_manager.go:354] "RemoveStaleState removing state" podUID="97cf4a02-e236-49c1-904f-21b25e87eba4" containerName="registry-server" Dec 13 07:04:13 crc kubenswrapper[4971]: I1213 07:04:13.535999 4971 memory_manager.go:354] "RemoveStaleState removing state" podUID="5eea3e58-f9fb-4529-b645-0f5193d5a8a9" containerName="registry-server" Dec 13 07:04:13 crc kubenswrapper[4971]: I1213 07:04:13.536007 4971 memory_manager.go:354] "RemoveStaleState removing state" podUID="d44cf79f-dbdf-4e9b-ba48-7351d398c38c" containerName="registry-server" Dec 13 07:04:13 crc kubenswrapper[4971]: I1213 07:04:13.536017 4971 memory_manager.go:354] "RemoveStaleState removing state" podUID="7e9ec79e-730d-4d91-b8f9-851c2b1cb769" containerName="registry-server" Dec 13 07:04:13 crc kubenswrapper[4971]: I1213 07:04:13.536026 4971 memory_manager.go:354] "RemoveStaleState removing state" podUID="9fd587ab-4839-4479-80f8-f97d78d833b9" containerName="registry-server" Dec 13 07:04:13 crc kubenswrapper[4971]: E1213 07:04:13.536134 4971 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="5c88a864-89a9-4f5e-b713-385170b9bf17" containerName="marketplace-operator" Dec 13 07:04:13 crc kubenswrapper[4971]: I1213 07:04:13.536148 4971 state_mem.go:107] "Deleted CPUSet assignment" podUID="5c88a864-89a9-4f5e-b713-385170b9bf17" containerName="marketplace-operator" Dec 13 07:04:13 crc kubenswrapper[4971]: I1213 07:04:13.536245 4971 memory_manager.go:354] "RemoveStaleState removing state" podUID="5c88a864-89a9-4f5e-b713-385170b9bf17" containerName="marketplace-operator" Dec 13 07:04:13 crc kubenswrapper[4971]: I1213 07:04:13.536257 4971 memory_manager.go:354] "RemoveStaleState removing state" podUID="5c88a864-89a9-4f5e-b713-385170b9bf17" containerName="marketplace-operator" Dec 13 07:04:13 crc kubenswrapper[4971]: I1213 07:04:13.536940 4971 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-szqz6" Dec 13 07:04:13 crc kubenswrapper[4971]: I1213 07:04:13.540562 4971 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"certified-operators-dockercfg-4rs5g" Dec 13 07:04:13 crc kubenswrapper[4971]: I1213 07:04:13.543954 4971 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-szqz6"] Dec 13 07:04:13 crc kubenswrapper[4971]: I1213 07:04:13.701816 4971 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/33212ad2-538d-4059-bc1f-e98a51c93b62-utilities\") pod \"certified-operators-szqz6\" (UID: \"33212ad2-538d-4059-bc1f-e98a51c93b62\") " pod="openshift-marketplace/certified-operators-szqz6" Dec 13 07:04:13 crc kubenswrapper[4971]: I1213 07:04:13.701934 4971 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/33212ad2-538d-4059-bc1f-e98a51c93b62-catalog-content\") pod \"certified-operators-szqz6\" (UID: \"33212ad2-538d-4059-bc1f-e98a51c93b62\") " pod="openshift-marketplace/certified-operators-szqz6" Dec 13 07:04:13 crc kubenswrapper[4971]: I1213 07:04:13.701987 4971 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-fwfsb\" (UniqueName: \"kubernetes.io/projected/33212ad2-538d-4059-bc1f-e98a51c93b62-kube-api-access-fwfsb\") pod \"certified-operators-szqz6\" (UID: \"33212ad2-538d-4059-bc1f-e98a51c93b62\") " pod="openshift-marketplace/certified-operators-szqz6" Dec 13 07:04:13 crc kubenswrapper[4971]: I1213 07:04:13.732384 4971 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-marketplace-hvzhk"] Dec 13 07:04:13 crc kubenswrapper[4971]: I1213 07:04:13.733506 4971 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-hvzhk" Dec 13 07:04:13 crc kubenswrapper[4971]: I1213 07:04:13.735286 4971 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"redhat-marketplace-dockercfg-x2ctb" Dec 13 07:04:13 crc kubenswrapper[4971]: I1213 07:04:13.743124 4971 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-hvzhk"] Dec 13 07:04:13 crc kubenswrapper[4971]: I1213 07:04:13.778581 4971 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="5c88a864-89a9-4f5e-b713-385170b9bf17" path="/var/lib/kubelet/pods/5c88a864-89a9-4f5e-b713-385170b9bf17/volumes" Dec 13 07:04:13 crc kubenswrapper[4971]: I1213 07:04:13.779908 4971 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="5eea3e58-f9fb-4529-b645-0f5193d5a8a9" path="/var/lib/kubelet/pods/5eea3e58-f9fb-4529-b645-0f5193d5a8a9/volumes" Dec 13 07:04:13 crc kubenswrapper[4971]: I1213 07:04:13.780911 4971 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="7e9ec79e-730d-4d91-b8f9-851c2b1cb769" path="/var/lib/kubelet/pods/7e9ec79e-730d-4d91-b8f9-851c2b1cb769/volumes" Dec 13 07:04:13 crc kubenswrapper[4971]: I1213 07:04:13.783674 4971 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="97cf4a02-e236-49c1-904f-21b25e87eba4" path="/var/lib/kubelet/pods/97cf4a02-e236-49c1-904f-21b25e87eba4/volumes" Dec 13 07:04:13 crc kubenswrapper[4971]: I1213 07:04:13.784795 4971 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="9fd587ab-4839-4479-80f8-f97d78d833b9" path="/var/lib/kubelet/pods/9fd587ab-4839-4479-80f8-f97d78d833b9/volumes" Dec 13 07:04:13 crc kubenswrapper[4971]: I1213 07:04:13.786022 4971 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="d44cf79f-dbdf-4e9b-ba48-7351d398c38c" path="/var/lib/kubelet/pods/d44cf79f-dbdf-4e9b-ba48-7351d398c38c/volumes" Dec 13 07:04:13 crc kubenswrapper[4971]: I1213 07:04:13.803968 4971 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-fwfsb\" (UniqueName: \"kubernetes.io/projected/33212ad2-538d-4059-bc1f-e98a51c93b62-kube-api-access-fwfsb\") pod \"certified-operators-szqz6\" (UID: \"33212ad2-538d-4059-bc1f-e98a51c93b62\") " pod="openshift-marketplace/certified-operators-szqz6" Dec 13 07:04:13 crc kubenswrapper[4971]: I1213 07:04:13.804313 4971 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/33212ad2-538d-4059-bc1f-e98a51c93b62-utilities\") pod \"certified-operators-szqz6\" (UID: \"33212ad2-538d-4059-bc1f-e98a51c93b62\") " pod="openshift-marketplace/certified-operators-szqz6" Dec 13 07:04:13 crc kubenswrapper[4971]: I1213 07:04:13.804469 4971 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/33212ad2-538d-4059-bc1f-e98a51c93b62-catalog-content\") pod \"certified-operators-szqz6\" (UID: \"33212ad2-538d-4059-bc1f-e98a51c93b62\") " pod="openshift-marketplace/certified-operators-szqz6" Dec 13 07:04:13 crc kubenswrapper[4971]: I1213 07:04:13.804895 4971 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/33212ad2-538d-4059-bc1f-e98a51c93b62-utilities\") pod \"certified-operators-szqz6\" (UID: \"33212ad2-538d-4059-bc1f-e98a51c93b62\") " pod="openshift-marketplace/certified-operators-szqz6" Dec 13 07:04:13 crc kubenswrapper[4971]: I1213 07:04:13.805062 4971 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/33212ad2-538d-4059-bc1f-e98a51c93b62-catalog-content\") pod \"certified-operators-szqz6\" (UID: \"33212ad2-538d-4059-bc1f-e98a51c93b62\") " pod="openshift-marketplace/certified-operators-szqz6" Dec 13 07:04:13 crc kubenswrapper[4971]: I1213 07:04:13.823250 4971 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-fwfsb\" (UniqueName: \"kubernetes.io/projected/33212ad2-538d-4059-bc1f-e98a51c93b62-kube-api-access-fwfsb\") pod \"certified-operators-szqz6\" (UID: \"33212ad2-538d-4059-bc1f-e98a51c93b62\") " pod="openshift-marketplace/certified-operators-szqz6" Dec 13 07:04:13 crc kubenswrapper[4971]: I1213 07:04:13.871420 4971 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-szqz6" Dec 13 07:04:13 crc kubenswrapper[4971]: I1213 07:04:13.906936 4971 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-mxcsf\" (UniqueName: \"kubernetes.io/projected/4b52f930-cf4c-4a97-b838-9afc9e3d2b31-kube-api-access-mxcsf\") pod \"redhat-marketplace-hvzhk\" (UID: \"4b52f930-cf4c-4a97-b838-9afc9e3d2b31\") " pod="openshift-marketplace/redhat-marketplace-hvzhk" Dec 13 07:04:13 crc kubenswrapper[4971]: I1213 07:04:13.907031 4971 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/4b52f930-cf4c-4a97-b838-9afc9e3d2b31-catalog-content\") pod \"redhat-marketplace-hvzhk\" (UID: \"4b52f930-cf4c-4a97-b838-9afc9e3d2b31\") " pod="openshift-marketplace/redhat-marketplace-hvzhk" Dec 13 07:04:13 crc kubenswrapper[4971]: I1213 07:04:13.907067 4971 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/4b52f930-cf4c-4a97-b838-9afc9e3d2b31-utilities\") pod \"redhat-marketplace-hvzhk\" (UID: \"4b52f930-cf4c-4a97-b838-9afc9e3d2b31\") " pod="openshift-marketplace/redhat-marketplace-hvzhk" Dec 13 07:04:13 crc kubenswrapper[4971]: I1213 07:04:13.934499 4971 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/certified-operators-swsx4"] Dec 13 07:04:13 crc kubenswrapper[4971]: I1213 07:04:13.935473 4971 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-swsx4" Dec 13 07:04:13 crc kubenswrapper[4971]: I1213 07:04:13.952007 4971 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-swsx4"] Dec 13 07:04:14 crc kubenswrapper[4971]: I1213 07:04:14.008216 4971 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/4b52f930-cf4c-4a97-b838-9afc9e3d2b31-utilities\") pod \"redhat-marketplace-hvzhk\" (UID: \"4b52f930-cf4c-4a97-b838-9afc9e3d2b31\") " pod="openshift-marketplace/redhat-marketplace-hvzhk" Dec 13 07:04:14 crc kubenswrapper[4971]: I1213 07:04:14.008306 4971 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-mxcsf\" (UniqueName: \"kubernetes.io/projected/4b52f930-cf4c-4a97-b838-9afc9e3d2b31-kube-api-access-mxcsf\") pod \"redhat-marketplace-hvzhk\" (UID: \"4b52f930-cf4c-4a97-b838-9afc9e3d2b31\") " pod="openshift-marketplace/redhat-marketplace-hvzhk" Dec 13 07:04:14 crc kubenswrapper[4971]: I1213 07:04:14.008329 4971 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/4b52f930-cf4c-4a97-b838-9afc9e3d2b31-catalog-content\") pod \"redhat-marketplace-hvzhk\" (UID: \"4b52f930-cf4c-4a97-b838-9afc9e3d2b31\") " pod="openshift-marketplace/redhat-marketplace-hvzhk" Dec 13 07:04:14 crc kubenswrapper[4971]: I1213 07:04:14.008973 4971 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/4b52f930-cf4c-4a97-b838-9afc9e3d2b31-catalog-content\") pod \"redhat-marketplace-hvzhk\" (UID: \"4b52f930-cf4c-4a97-b838-9afc9e3d2b31\") " pod="openshift-marketplace/redhat-marketplace-hvzhk" Dec 13 07:04:14 crc kubenswrapper[4971]: I1213 07:04:14.009077 4971 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/4b52f930-cf4c-4a97-b838-9afc9e3d2b31-utilities\") pod \"redhat-marketplace-hvzhk\" (UID: \"4b52f930-cf4c-4a97-b838-9afc9e3d2b31\") " pod="openshift-marketplace/redhat-marketplace-hvzhk" Dec 13 07:04:14 crc kubenswrapper[4971]: I1213 07:04:14.028537 4971 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-mxcsf\" (UniqueName: \"kubernetes.io/projected/4b52f930-cf4c-4a97-b838-9afc9e3d2b31-kube-api-access-mxcsf\") pod \"redhat-marketplace-hvzhk\" (UID: \"4b52f930-cf4c-4a97-b838-9afc9e3d2b31\") " pod="openshift-marketplace/redhat-marketplace-hvzhk" Dec 13 07:04:14 crc kubenswrapper[4971]: I1213 07:04:14.055908 4971 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-hvzhk" Dec 13 07:04:14 crc kubenswrapper[4971]: I1213 07:04:14.075411 4971 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-szqz6"] Dec 13 07:04:14 crc kubenswrapper[4971]: W1213 07:04:14.082088 4971 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod33212ad2_538d_4059_bc1f_e98a51c93b62.slice/crio-632edef68ad34e342ecb4aaba5b89a59dc5717da0e0d4c257b789f767c0c9653 WatchSource:0}: Error finding container 632edef68ad34e342ecb4aaba5b89a59dc5717da0e0d4c257b789f767c0c9653: Status 404 returned error can't find the container with id 632edef68ad34e342ecb4aaba5b89a59dc5717da0e0d4c257b789f767c0c9653 Dec 13 07:04:14 crc kubenswrapper[4971]: I1213 07:04:14.110631 4971 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/a1c43378-fa47-4eb5-8c02-f5b229197543-utilities\") pod \"certified-operators-swsx4\" (UID: \"a1c43378-fa47-4eb5-8c02-f5b229197543\") " pod="openshift-marketplace/certified-operators-swsx4" Dec 13 07:04:14 crc kubenswrapper[4971]: I1213 07:04:14.110710 4971 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-dpr4c\" (UniqueName: \"kubernetes.io/projected/a1c43378-fa47-4eb5-8c02-f5b229197543-kube-api-access-dpr4c\") pod \"certified-operators-swsx4\" (UID: \"a1c43378-fa47-4eb5-8c02-f5b229197543\") " pod="openshift-marketplace/certified-operators-swsx4" Dec 13 07:04:14 crc kubenswrapper[4971]: I1213 07:04:14.110804 4971 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/a1c43378-fa47-4eb5-8c02-f5b229197543-catalog-content\") pod \"certified-operators-swsx4\" (UID: \"a1c43378-fa47-4eb5-8c02-f5b229197543\") " pod="openshift-marketplace/certified-operators-swsx4" Dec 13 07:04:14 crc kubenswrapper[4971]: I1213 07:04:14.146294 4971 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-marketplace-shhnv"] Dec 13 07:04:14 crc kubenswrapper[4971]: I1213 07:04:14.148042 4971 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-shhnv" Dec 13 07:04:14 crc kubenswrapper[4971]: I1213 07:04:14.155404 4971 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-shhnv"] Dec 13 07:04:14 crc kubenswrapper[4971]: I1213 07:04:14.211511 4971 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/a1c43378-fa47-4eb5-8c02-f5b229197543-catalog-content\") pod \"certified-operators-swsx4\" (UID: \"a1c43378-fa47-4eb5-8c02-f5b229197543\") " pod="openshift-marketplace/certified-operators-swsx4" Dec 13 07:04:14 crc kubenswrapper[4971]: I1213 07:04:14.211688 4971 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/a1c43378-fa47-4eb5-8c02-f5b229197543-utilities\") pod \"certified-operators-swsx4\" (UID: \"a1c43378-fa47-4eb5-8c02-f5b229197543\") " pod="openshift-marketplace/certified-operators-swsx4" Dec 13 07:04:14 crc kubenswrapper[4971]: I1213 07:04:14.211726 4971 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-dpr4c\" (UniqueName: \"kubernetes.io/projected/a1c43378-fa47-4eb5-8c02-f5b229197543-kube-api-access-dpr4c\") pod \"certified-operators-swsx4\" (UID: \"a1c43378-fa47-4eb5-8c02-f5b229197543\") " pod="openshift-marketplace/certified-operators-swsx4" Dec 13 07:04:14 crc kubenswrapper[4971]: I1213 07:04:14.212133 4971 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/a1c43378-fa47-4eb5-8c02-f5b229197543-utilities\") pod \"certified-operators-swsx4\" (UID: \"a1c43378-fa47-4eb5-8c02-f5b229197543\") " pod="openshift-marketplace/certified-operators-swsx4" Dec 13 07:04:14 crc kubenswrapper[4971]: I1213 07:04:14.212588 4971 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/a1c43378-fa47-4eb5-8c02-f5b229197543-catalog-content\") pod \"certified-operators-swsx4\" (UID: \"a1c43378-fa47-4eb5-8c02-f5b229197543\") " pod="openshift-marketplace/certified-operators-swsx4" Dec 13 07:04:14 crc kubenswrapper[4971]: I1213 07:04:14.228128 4971 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-szqz6" event={"ID":"33212ad2-538d-4059-bc1f-e98a51c93b62","Type":"ContainerStarted","Data":"632edef68ad34e342ecb4aaba5b89a59dc5717da0e0d4c257b789f767c0c9653"} Dec 13 07:04:14 crc kubenswrapper[4971]: I1213 07:04:14.232279 4971 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-dpr4c\" (UniqueName: \"kubernetes.io/projected/a1c43378-fa47-4eb5-8c02-f5b229197543-kube-api-access-dpr4c\") pod \"certified-operators-swsx4\" (UID: \"a1c43378-fa47-4eb5-8c02-f5b229197543\") " pod="openshift-marketplace/certified-operators-swsx4" Dec 13 07:04:14 crc kubenswrapper[4971]: I1213 07:04:14.257254 4971 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-swsx4" Dec 13 07:04:14 crc kubenswrapper[4971]: I1213 07:04:14.316729 4971 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/e80b7d34-d09e-4efe-9879-eb54ea91d724-utilities\") pod \"redhat-marketplace-shhnv\" (UID: \"e80b7d34-d09e-4efe-9879-eb54ea91d724\") " pod="openshift-marketplace/redhat-marketplace-shhnv" Dec 13 07:04:14 crc kubenswrapper[4971]: I1213 07:04:14.316794 4971 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/e80b7d34-d09e-4efe-9879-eb54ea91d724-catalog-content\") pod \"redhat-marketplace-shhnv\" (UID: \"e80b7d34-d09e-4efe-9879-eb54ea91d724\") " pod="openshift-marketplace/redhat-marketplace-shhnv" Dec 13 07:04:14 crc kubenswrapper[4971]: I1213 07:04:14.316842 4971 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-rtlcd\" (UniqueName: \"kubernetes.io/projected/e80b7d34-d09e-4efe-9879-eb54ea91d724-kube-api-access-rtlcd\") pod \"redhat-marketplace-shhnv\" (UID: \"e80b7d34-d09e-4efe-9879-eb54ea91d724\") " pod="openshift-marketplace/redhat-marketplace-shhnv" Dec 13 07:04:14 crc kubenswrapper[4971]: I1213 07:04:14.419637 4971 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/e80b7d34-d09e-4efe-9879-eb54ea91d724-catalog-content\") pod \"redhat-marketplace-shhnv\" (UID: \"e80b7d34-d09e-4efe-9879-eb54ea91d724\") " pod="openshift-marketplace/redhat-marketplace-shhnv" Dec 13 07:04:14 crc kubenswrapper[4971]: I1213 07:04:14.419702 4971 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-rtlcd\" (UniqueName: \"kubernetes.io/projected/e80b7d34-d09e-4efe-9879-eb54ea91d724-kube-api-access-rtlcd\") pod \"redhat-marketplace-shhnv\" (UID: \"e80b7d34-d09e-4efe-9879-eb54ea91d724\") " pod="openshift-marketplace/redhat-marketplace-shhnv" Dec 13 07:04:14 crc kubenswrapper[4971]: I1213 07:04:14.419843 4971 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/e80b7d34-d09e-4efe-9879-eb54ea91d724-utilities\") pod \"redhat-marketplace-shhnv\" (UID: \"e80b7d34-d09e-4efe-9879-eb54ea91d724\") " pod="openshift-marketplace/redhat-marketplace-shhnv" Dec 13 07:04:14 crc kubenswrapper[4971]: I1213 07:04:14.420359 4971 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/e80b7d34-d09e-4efe-9879-eb54ea91d724-utilities\") pod \"redhat-marketplace-shhnv\" (UID: \"e80b7d34-d09e-4efe-9879-eb54ea91d724\") " pod="openshift-marketplace/redhat-marketplace-shhnv" Dec 13 07:04:14 crc kubenswrapper[4971]: I1213 07:04:14.421158 4971 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/e80b7d34-d09e-4efe-9879-eb54ea91d724-catalog-content\") pod \"redhat-marketplace-shhnv\" (UID: \"e80b7d34-d09e-4efe-9879-eb54ea91d724\") " pod="openshift-marketplace/redhat-marketplace-shhnv" Dec 13 07:04:14 crc kubenswrapper[4971]: I1213 07:04:14.438871 4971 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-swsx4"] Dec 13 07:04:14 crc kubenswrapper[4971]: I1213 07:04:14.448296 4971 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-rtlcd\" (UniqueName: \"kubernetes.io/projected/e80b7d34-d09e-4efe-9879-eb54ea91d724-kube-api-access-rtlcd\") pod \"redhat-marketplace-shhnv\" (UID: \"e80b7d34-d09e-4efe-9879-eb54ea91d724\") " pod="openshift-marketplace/redhat-marketplace-shhnv" Dec 13 07:04:14 crc kubenswrapper[4971]: I1213 07:04:14.478778 4971 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-shhnv" Dec 13 07:04:14 crc kubenswrapper[4971]: I1213 07:04:14.514112 4971 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-hvzhk"] Dec 13 07:04:14 crc kubenswrapper[4971]: W1213 07:04:14.554752 4971 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod4b52f930_cf4c_4a97_b838_9afc9e3d2b31.slice/crio-f19fcb49d8f663333fb4803e8fb585adcf014ea8fa0b93958ad37d09b979c700 WatchSource:0}: Error finding container f19fcb49d8f663333fb4803e8fb585adcf014ea8fa0b93958ad37d09b979c700: Status 404 returned error can't find the container with id f19fcb49d8f663333fb4803e8fb585adcf014ea8fa0b93958ad37d09b979c700 Dec 13 07:04:14 crc kubenswrapper[4971]: W1213 07:04:14.555432 4971 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-poda1c43378_fa47_4eb5_8c02_f5b229197543.slice/crio-6db3142585544dc7898dfe4faf03f4f84d8a8689319acc85d3f4860b94fd4b6c WatchSource:0}: Error finding container 6db3142585544dc7898dfe4faf03f4f84d8a8689319acc85d3f4860b94fd4b6c: Status 404 returned error can't find the container with id 6db3142585544dc7898dfe4faf03f4f84d8a8689319acc85d3f4860b94fd4b6c Dec 13 07:04:14 crc kubenswrapper[4971]: I1213 07:04:14.701190 4971 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-shhnv"] Dec 13 07:04:14 crc kubenswrapper[4971]: W1213 07:04:14.706095 4971 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pode80b7d34_d09e_4efe_9879_eb54ea91d724.slice/crio-8dcf1bb6b6cc62194648dd926d4570e9e85c06a944021a584a810c808efecbb3 WatchSource:0}: Error finding container 8dcf1bb6b6cc62194648dd926d4570e9e85c06a944021a584a810c808efecbb3: Status 404 returned error can't find the container with id 8dcf1bb6b6cc62194648dd926d4570e9e85c06a944021a584a810c808efecbb3 Dec 13 07:04:15 crc kubenswrapper[4971]: I1213 07:04:15.236712 4971 generic.go:334] "Generic (PLEG): container finished" podID="33212ad2-538d-4059-bc1f-e98a51c93b62" containerID="fa2289c0d69684e167df7ce6fcdeee2d419f74b749889a05d75074add8e9e739" exitCode=0 Dec 13 07:04:15 crc kubenswrapper[4971]: I1213 07:04:15.236800 4971 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-szqz6" event={"ID":"33212ad2-538d-4059-bc1f-e98a51c93b62","Type":"ContainerDied","Data":"fa2289c0d69684e167df7ce6fcdeee2d419f74b749889a05d75074add8e9e739"} Dec 13 07:04:15 crc kubenswrapper[4971]: I1213 07:04:15.240860 4971 generic.go:334] "Generic (PLEG): container finished" podID="4b52f930-cf4c-4a97-b838-9afc9e3d2b31" containerID="2a896397bcc5c51c1c548637cb1b2324ee3412264b83462f6b07d636c6d32728" exitCode=0 Dec 13 07:04:15 crc kubenswrapper[4971]: I1213 07:04:15.241596 4971 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-hvzhk" event={"ID":"4b52f930-cf4c-4a97-b838-9afc9e3d2b31","Type":"ContainerDied","Data":"2a896397bcc5c51c1c548637cb1b2324ee3412264b83462f6b07d636c6d32728"} Dec 13 07:04:15 crc kubenswrapper[4971]: I1213 07:04:15.241659 4971 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-hvzhk" event={"ID":"4b52f930-cf4c-4a97-b838-9afc9e3d2b31","Type":"ContainerStarted","Data":"f19fcb49d8f663333fb4803e8fb585adcf014ea8fa0b93958ad37d09b979c700"} Dec 13 07:04:15 crc kubenswrapper[4971]: I1213 07:04:15.245803 4971 generic.go:334] "Generic (PLEG): container finished" podID="e80b7d34-d09e-4efe-9879-eb54ea91d724" containerID="d2462b9bb30042d8c344b2f90b21fe0d461a42be2f9be3b1f77ed7d336a8cf49" exitCode=0 Dec 13 07:04:15 crc kubenswrapper[4971]: I1213 07:04:15.245862 4971 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-shhnv" event={"ID":"e80b7d34-d09e-4efe-9879-eb54ea91d724","Type":"ContainerDied","Data":"d2462b9bb30042d8c344b2f90b21fe0d461a42be2f9be3b1f77ed7d336a8cf49"} Dec 13 07:04:15 crc kubenswrapper[4971]: I1213 07:04:15.245889 4971 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-shhnv" event={"ID":"e80b7d34-d09e-4efe-9879-eb54ea91d724","Type":"ContainerStarted","Data":"8dcf1bb6b6cc62194648dd926d4570e9e85c06a944021a584a810c808efecbb3"} Dec 13 07:04:15 crc kubenswrapper[4971]: I1213 07:04:15.249796 4971 generic.go:334] "Generic (PLEG): container finished" podID="a1c43378-fa47-4eb5-8c02-f5b229197543" containerID="37401099d2099e897ac4109f6bbcb2a1b68df9c72d37005a6b6e4c46a86b6956" exitCode=0 Dec 13 07:04:15 crc kubenswrapper[4971]: I1213 07:04:15.250835 4971 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-swsx4" event={"ID":"a1c43378-fa47-4eb5-8c02-f5b229197543","Type":"ContainerDied","Data":"37401099d2099e897ac4109f6bbcb2a1b68df9c72d37005a6b6e4c46a86b6956"} Dec 13 07:04:15 crc kubenswrapper[4971]: I1213 07:04:15.250897 4971 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-swsx4" event={"ID":"a1c43378-fa47-4eb5-8c02-f5b229197543","Type":"ContainerStarted","Data":"6db3142585544dc7898dfe4faf03f4f84d8a8689319acc85d3f4860b94fd4b6c"} Dec 13 07:04:16 crc kubenswrapper[4971]: I1213 07:04:16.153962 4971 patch_prober.go:28] interesting pod/machine-config-daemon-82xjz container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 13 07:04:16 crc kubenswrapper[4971]: I1213 07:04:16.154035 4971 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-82xjz" podUID="e0c4af12-edda-41d7-9b44-f87396174ef2" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 13 07:04:16 crc kubenswrapper[4971]: I1213 07:04:16.256624 4971 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-shhnv" event={"ID":"e80b7d34-d09e-4efe-9879-eb54ea91d724","Type":"ContainerStarted","Data":"56524017b2b0b269c175ed0412e13d140a8e040fe25b81a7284fd32f4fb24d12"} Dec 13 07:04:16 crc kubenswrapper[4971]: I1213 07:04:16.258574 4971 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-swsx4" event={"ID":"a1c43378-fa47-4eb5-8c02-f5b229197543","Type":"ContainerStarted","Data":"d1e668b1d78683413c01dcb122882b84b945a36974492f19eca6e50614791f08"} Dec 13 07:04:16 crc kubenswrapper[4971]: I1213 07:04:16.260929 4971 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-szqz6" event={"ID":"33212ad2-538d-4059-bc1f-e98a51c93b62","Type":"ContainerStarted","Data":"cb66e2d84573ac01f8dc44d393b253a17d8d1620e54ac86445c4f487fb4059c4"} Dec 13 07:04:16 crc kubenswrapper[4971]: I1213 07:04:16.262393 4971 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-hvzhk" event={"ID":"4b52f930-cf4c-4a97-b838-9afc9e3d2b31","Type":"ContainerStarted","Data":"c0e7032640b36f8464b2ae62ca4bb31766253aba44d79e4580007d332ecd111c"} Dec 13 07:04:16 crc kubenswrapper[4971]: I1213 07:04:16.330770 4971 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-operators-cvcsq"] Dec 13 07:04:16 crc kubenswrapper[4971]: I1213 07:04:16.331825 4971 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-cvcsq" Dec 13 07:04:16 crc kubenswrapper[4971]: I1213 07:04:16.333703 4971 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"redhat-operators-dockercfg-ct8rh" Dec 13 07:04:16 crc kubenswrapper[4971]: I1213 07:04:16.345220 4971 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-cvcsq"] Dec 13 07:04:16 crc kubenswrapper[4971]: I1213 07:04:16.447289 4971 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/6fa06763-60e2-4bc5-91f7-58ded0345c6a-utilities\") pod \"redhat-operators-cvcsq\" (UID: \"6fa06763-60e2-4bc5-91f7-58ded0345c6a\") " pod="openshift-marketplace/redhat-operators-cvcsq" Dec 13 07:04:16 crc kubenswrapper[4971]: I1213 07:04:16.447378 4971 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-46c46\" (UniqueName: \"kubernetes.io/projected/6fa06763-60e2-4bc5-91f7-58ded0345c6a-kube-api-access-46c46\") pod \"redhat-operators-cvcsq\" (UID: \"6fa06763-60e2-4bc5-91f7-58ded0345c6a\") " pod="openshift-marketplace/redhat-operators-cvcsq" Dec 13 07:04:16 crc kubenswrapper[4971]: I1213 07:04:16.447428 4971 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/6fa06763-60e2-4bc5-91f7-58ded0345c6a-catalog-content\") pod \"redhat-operators-cvcsq\" (UID: \"6fa06763-60e2-4bc5-91f7-58ded0345c6a\") " pod="openshift-marketplace/redhat-operators-cvcsq" Dec 13 07:04:16 crc kubenswrapper[4971]: I1213 07:04:16.533589 4971 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/community-operators-z62gm"] Dec 13 07:04:16 crc kubenswrapper[4971]: I1213 07:04:16.534675 4971 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-z62gm" Dec 13 07:04:16 crc kubenswrapper[4971]: I1213 07:04:16.537234 4971 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"community-operators-dockercfg-dmngl" Dec 13 07:04:16 crc kubenswrapper[4971]: I1213 07:04:16.546191 4971 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-z62gm"] Dec 13 07:04:16 crc kubenswrapper[4971]: I1213 07:04:16.548267 4971 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/6fa06763-60e2-4bc5-91f7-58ded0345c6a-utilities\") pod \"redhat-operators-cvcsq\" (UID: \"6fa06763-60e2-4bc5-91f7-58ded0345c6a\") " pod="openshift-marketplace/redhat-operators-cvcsq" Dec 13 07:04:16 crc kubenswrapper[4971]: I1213 07:04:16.548336 4971 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-46c46\" (UniqueName: \"kubernetes.io/projected/6fa06763-60e2-4bc5-91f7-58ded0345c6a-kube-api-access-46c46\") pod \"redhat-operators-cvcsq\" (UID: \"6fa06763-60e2-4bc5-91f7-58ded0345c6a\") " pod="openshift-marketplace/redhat-operators-cvcsq" Dec 13 07:04:16 crc kubenswrapper[4971]: I1213 07:04:16.548383 4971 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/6fa06763-60e2-4bc5-91f7-58ded0345c6a-catalog-content\") pod \"redhat-operators-cvcsq\" (UID: \"6fa06763-60e2-4bc5-91f7-58ded0345c6a\") " pod="openshift-marketplace/redhat-operators-cvcsq" Dec 13 07:04:16 crc kubenswrapper[4971]: I1213 07:04:16.548951 4971 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/6fa06763-60e2-4bc5-91f7-58ded0345c6a-catalog-content\") pod \"redhat-operators-cvcsq\" (UID: \"6fa06763-60e2-4bc5-91f7-58ded0345c6a\") " pod="openshift-marketplace/redhat-operators-cvcsq" Dec 13 07:04:16 crc kubenswrapper[4971]: I1213 07:04:16.549066 4971 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/6fa06763-60e2-4bc5-91f7-58ded0345c6a-utilities\") pod \"redhat-operators-cvcsq\" (UID: \"6fa06763-60e2-4bc5-91f7-58ded0345c6a\") " pod="openshift-marketplace/redhat-operators-cvcsq" Dec 13 07:04:16 crc kubenswrapper[4971]: I1213 07:04:16.568907 4971 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-46c46\" (UniqueName: \"kubernetes.io/projected/6fa06763-60e2-4bc5-91f7-58ded0345c6a-kube-api-access-46c46\") pod \"redhat-operators-cvcsq\" (UID: \"6fa06763-60e2-4bc5-91f7-58ded0345c6a\") " pod="openshift-marketplace/redhat-operators-cvcsq" Dec 13 07:04:16 crc kubenswrapper[4971]: I1213 07:04:16.650325 4971 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/af866491-3f91-4f43-809b-47b473e2e53d-catalog-content\") pod \"community-operators-z62gm\" (UID: \"af866491-3f91-4f43-809b-47b473e2e53d\") " pod="openshift-marketplace/community-operators-z62gm" Dec 13 07:04:16 crc kubenswrapper[4971]: I1213 07:04:16.650447 4971 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-dz96m\" (UniqueName: \"kubernetes.io/projected/af866491-3f91-4f43-809b-47b473e2e53d-kube-api-access-dz96m\") pod \"community-operators-z62gm\" (UID: \"af866491-3f91-4f43-809b-47b473e2e53d\") " pod="openshift-marketplace/community-operators-z62gm" Dec 13 07:04:16 crc kubenswrapper[4971]: I1213 07:04:16.650813 4971 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/af866491-3f91-4f43-809b-47b473e2e53d-utilities\") pod \"community-operators-z62gm\" (UID: \"af866491-3f91-4f43-809b-47b473e2e53d\") " pod="openshift-marketplace/community-operators-z62gm" Dec 13 07:04:16 crc kubenswrapper[4971]: I1213 07:04:16.735666 4971 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-operators-nkmq2"] Dec 13 07:04:16 crc kubenswrapper[4971]: I1213 07:04:16.739038 4971 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-nkmq2" Dec 13 07:04:16 crc kubenswrapper[4971]: I1213 07:04:16.753433 4971 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-nkmq2"] Dec 13 07:04:16 crc kubenswrapper[4971]: I1213 07:04:16.753715 4971 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-dz96m\" (UniqueName: \"kubernetes.io/projected/af866491-3f91-4f43-809b-47b473e2e53d-kube-api-access-dz96m\") pod \"community-operators-z62gm\" (UID: \"af866491-3f91-4f43-809b-47b473e2e53d\") " pod="openshift-marketplace/community-operators-z62gm" Dec 13 07:04:16 crc kubenswrapper[4971]: I1213 07:04:16.753837 4971 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/af866491-3f91-4f43-809b-47b473e2e53d-utilities\") pod \"community-operators-z62gm\" (UID: \"af866491-3f91-4f43-809b-47b473e2e53d\") " pod="openshift-marketplace/community-operators-z62gm" Dec 13 07:04:16 crc kubenswrapper[4971]: I1213 07:04:16.753951 4971 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/af866491-3f91-4f43-809b-47b473e2e53d-catalog-content\") pod \"community-operators-z62gm\" (UID: \"af866491-3f91-4f43-809b-47b473e2e53d\") " pod="openshift-marketplace/community-operators-z62gm" Dec 13 07:04:16 crc kubenswrapper[4971]: I1213 07:04:16.754508 4971 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/af866491-3f91-4f43-809b-47b473e2e53d-catalog-content\") pod \"community-operators-z62gm\" (UID: \"af866491-3f91-4f43-809b-47b473e2e53d\") " pod="openshift-marketplace/community-operators-z62gm" Dec 13 07:04:16 crc kubenswrapper[4971]: I1213 07:04:16.754528 4971 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/af866491-3f91-4f43-809b-47b473e2e53d-utilities\") pod \"community-operators-z62gm\" (UID: \"af866491-3f91-4f43-809b-47b473e2e53d\") " pod="openshift-marketplace/community-operators-z62gm" Dec 13 07:04:16 crc kubenswrapper[4971]: I1213 07:04:16.754751 4971 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-cvcsq" Dec 13 07:04:16 crc kubenswrapper[4971]: I1213 07:04:16.773584 4971 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-dz96m\" (UniqueName: \"kubernetes.io/projected/af866491-3f91-4f43-809b-47b473e2e53d-kube-api-access-dz96m\") pod \"community-operators-z62gm\" (UID: \"af866491-3f91-4f43-809b-47b473e2e53d\") " pod="openshift-marketplace/community-operators-z62gm" Dec 13 07:04:16 crc kubenswrapper[4971]: I1213 07:04:16.854851 4971 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/ad20cd8d-6bce-4bb4-b070-79c421d8538b-catalog-content\") pod \"redhat-operators-nkmq2\" (UID: \"ad20cd8d-6bce-4bb4-b070-79c421d8538b\") " pod="openshift-marketplace/redhat-operators-nkmq2" Dec 13 07:04:16 crc kubenswrapper[4971]: I1213 07:04:16.854906 4971 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-zmpxz\" (UniqueName: \"kubernetes.io/projected/ad20cd8d-6bce-4bb4-b070-79c421d8538b-kube-api-access-zmpxz\") pod \"redhat-operators-nkmq2\" (UID: \"ad20cd8d-6bce-4bb4-b070-79c421d8538b\") " pod="openshift-marketplace/redhat-operators-nkmq2" Dec 13 07:04:16 crc kubenswrapper[4971]: I1213 07:04:16.855029 4971 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/ad20cd8d-6bce-4bb4-b070-79c421d8538b-utilities\") pod \"redhat-operators-nkmq2\" (UID: \"ad20cd8d-6bce-4bb4-b070-79c421d8538b\") " pod="openshift-marketplace/redhat-operators-nkmq2" Dec 13 07:04:16 crc kubenswrapper[4971]: I1213 07:04:16.899720 4971 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-z62gm" Dec 13 07:04:16 crc kubenswrapper[4971]: I1213 07:04:16.957013 4971 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/ad20cd8d-6bce-4bb4-b070-79c421d8538b-utilities\") pod \"redhat-operators-nkmq2\" (UID: \"ad20cd8d-6bce-4bb4-b070-79c421d8538b\") " pod="openshift-marketplace/redhat-operators-nkmq2" Dec 13 07:04:16 crc kubenswrapper[4971]: I1213 07:04:16.957108 4971 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/ad20cd8d-6bce-4bb4-b070-79c421d8538b-catalog-content\") pod \"redhat-operators-nkmq2\" (UID: \"ad20cd8d-6bce-4bb4-b070-79c421d8538b\") " pod="openshift-marketplace/redhat-operators-nkmq2" Dec 13 07:04:16 crc kubenswrapper[4971]: I1213 07:04:16.957145 4971 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-zmpxz\" (UniqueName: \"kubernetes.io/projected/ad20cd8d-6bce-4bb4-b070-79c421d8538b-kube-api-access-zmpxz\") pod \"redhat-operators-nkmq2\" (UID: \"ad20cd8d-6bce-4bb4-b070-79c421d8538b\") " pod="openshift-marketplace/redhat-operators-nkmq2" Dec 13 07:04:16 crc kubenswrapper[4971]: I1213 07:04:16.957811 4971 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/ad20cd8d-6bce-4bb4-b070-79c421d8538b-catalog-content\") pod \"redhat-operators-nkmq2\" (UID: \"ad20cd8d-6bce-4bb4-b070-79c421d8538b\") " pod="openshift-marketplace/redhat-operators-nkmq2" Dec 13 07:04:16 crc kubenswrapper[4971]: I1213 07:04:16.957811 4971 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/ad20cd8d-6bce-4bb4-b070-79c421d8538b-utilities\") pod \"redhat-operators-nkmq2\" (UID: \"ad20cd8d-6bce-4bb4-b070-79c421d8538b\") " pod="openshift-marketplace/redhat-operators-nkmq2" Dec 13 07:04:16 crc kubenswrapper[4971]: I1213 07:04:16.982600 4971 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-zmpxz\" (UniqueName: \"kubernetes.io/projected/ad20cd8d-6bce-4bb4-b070-79c421d8538b-kube-api-access-zmpxz\") pod \"redhat-operators-nkmq2\" (UID: \"ad20cd8d-6bce-4bb4-b070-79c421d8538b\") " pod="openshift-marketplace/redhat-operators-nkmq2" Dec 13 07:04:17 crc kubenswrapper[4971]: I1213 07:04:17.065801 4971 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-nkmq2" Dec 13 07:04:17 crc kubenswrapper[4971]: I1213 07:04:17.078879 4971 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-z62gm"] Dec 13 07:04:17 crc kubenswrapper[4971]: W1213 07:04:17.082302 4971 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podaf866491_3f91_4f43_809b_47b473e2e53d.slice/crio-f445f08f77bd2f2654a517a3a35f5a0a57a0a0afb83ad082a26a7401778a6f38 WatchSource:0}: Error finding container f445f08f77bd2f2654a517a3a35f5a0a57a0a0afb83ad082a26a7401778a6f38: Status 404 returned error can't find the container with id f445f08f77bd2f2654a517a3a35f5a0a57a0a0afb83ad082a26a7401778a6f38 Dec 13 07:04:17 crc kubenswrapper[4971]: I1213 07:04:17.142362 4971 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-cvcsq"] Dec 13 07:04:17 crc kubenswrapper[4971]: I1213 07:04:17.274274 4971 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-nkmq2"] Dec 13 07:04:17 crc kubenswrapper[4971]: I1213 07:04:17.274378 4971 generic.go:334] "Generic (PLEG): container finished" podID="e80b7d34-d09e-4efe-9879-eb54ea91d724" containerID="56524017b2b0b269c175ed0412e13d140a8e040fe25b81a7284fd32f4fb24d12" exitCode=0 Dec 13 07:04:17 crc kubenswrapper[4971]: I1213 07:04:17.274471 4971 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-shhnv" event={"ID":"e80b7d34-d09e-4efe-9879-eb54ea91d724","Type":"ContainerDied","Data":"56524017b2b0b269c175ed0412e13d140a8e040fe25b81a7284fd32f4fb24d12"} Dec 13 07:04:17 crc kubenswrapper[4971]: I1213 07:04:17.276082 4971 generic.go:334] "Generic (PLEG): container finished" podID="af866491-3f91-4f43-809b-47b473e2e53d" containerID="9dadbcb68ebd5832cdbbf8fba0f228defdeccac4a5b7715e71465d39c164d555" exitCode=0 Dec 13 07:04:17 crc kubenswrapper[4971]: I1213 07:04:17.276144 4971 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-z62gm" event={"ID":"af866491-3f91-4f43-809b-47b473e2e53d","Type":"ContainerDied","Data":"9dadbcb68ebd5832cdbbf8fba0f228defdeccac4a5b7715e71465d39c164d555"} Dec 13 07:04:17 crc kubenswrapper[4971]: I1213 07:04:17.276170 4971 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-z62gm" event={"ID":"af866491-3f91-4f43-809b-47b473e2e53d","Type":"ContainerStarted","Data":"f445f08f77bd2f2654a517a3a35f5a0a57a0a0afb83ad082a26a7401778a6f38"} Dec 13 07:04:17 crc kubenswrapper[4971]: I1213 07:04:17.278477 4971 generic.go:334] "Generic (PLEG): container finished" podID="a1c43378-fa47-4eb5-8c02-f5b229197543" containerID="d1e668b1d78683413c01dcb122882b84b945a36974492f19eca6e50614791f08" exitCode=0 Dec 13 07:04:17 crc kubenswrapper[4971]: I1213 07:04:17.278562 4971 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-swsx4" event={"ID":"a1c43378-fa47-4eb5-8c02-f5b229197543","Type":"ContainerDied","Data":"d1e668b1d78683413c01dcb122882b84b945a36974492f19eca6e50614791f08"} Dec 13 07:04:17 crc kubenswrapper[4971]: I1213 07:04:17.281679 4971 generic.go:334] "Generic (PLEG): container finished" podID="33212ad2-538d-4059-bc1f-e98a51c93b62" containerID="cb66e2d84573ac01f8dc44d393b253a17d8d1620e54ac86445c4f487fb4059c4" exitCode=0 Dec 13 07:04:17 crc kubenswrapper[4971]: I1213 07:04:17.281753 4971 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-szqz6" event={"ID":"33212ad2-538d-4059-bc1f-e98a51c93b62","Type":"ContainerDied","Data":"cb66e2d84573ac01f8dc44d393b253a17d8d1620e54ac86445c4f487fb4059c4"} Dec 13 07:04:17 crc kubenswrapper[4971]: I1213 07:04:17.286090 4971 generic.go:334] "Generic (PLEG): container finished" podID="4b52f930-cf4c-4a97-b838-9afc9e3d2b31" containerID="c0e7032640b36f8464b2ae62ca4bb31766253aba44d79e4580007d332ecd111c" exitCode=0 Dec 13 07:04:17 crc kubenswrapper[4971]: I1213 07:04:17.286701 4971 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-hvzhk" event={"ID":"4b52f930-cf4c-4a97-b838-9afc9e3d2b31","Type":"ContainerDied","Data":"c0e7032640b36f8464b2ae62ca4bb31766253aba44d79e4580007d332ecd111c"} Dec 13 07:04:17 crc kubenswrapper[4971]: I1213 07:04:17.288555 4971 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-cvcsq" event={"ID":"6fa06763-60e2-4bc5-91f7-58ded0345c6a","Type":"ContainerStarted","Data":"ccb5d3f165efe4941814ef3f0cd330ba5e20ca6aa5dbd724805f3bc56c692b3a"} Dec 13 07:04:18 crc kubenswrapper[4971]: I1213 07:04:18.305323 4971 generic.go:334] "Generic (PLEG): container finished" podID="ad20cd8d-6bce-4bb4-b070-79c421d8538b" containerID="e02224574d97754dca872708b7564e8fd9e0528d19fbf7fe118b6bc93481dbbe" exitCode=0 Dec 13 07:04:18 crc kubenswrapper[4971]: I1213 07:04:18.305367 4971 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-nkmq2" event={"ID":"ad20cd8d-6bce-4bb4-b070-79c421d8538b","Type":"ContainerDied","Data":"e02224574d97754dca872708b7564e8fd9e0528d19fbf7fe118b6bc93481dbbe"} Dec 13 07:04:18 crc kubenswrapper[4971]: I1213 07:04:18.305773 4971 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-nkmq2" event={"ID":"ad20cd8d-6bce-4bb4-b070-79c421d8538b","Type":"ContainerStarted","Data":"87b65ae489e534508ed1ac5ff7f225e95b6846e1c11eba52eb3068d56e274d2e"} Dec 13 07:04:18 crc kubenswrapper[4971]: I1213 07:04:18.311071 4971 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-shhnv" event={"ID":"e80b7d34-d09e-4efe-9879-eb54ea91d724","Type":"ContainerStarted","Data":"0ef38416f602cf2e66532ce278d058da4845515335b64b9ecea801fcb9eb4fe9"} Dec 13 07:04:18 crc kubenswrapper[4971]: I1213 07:04:18.318309 4971 generic.go:334] "Generic (PLEG): container finished" podID="af866491-3f91-4f43-809b-47b473e2e53d" containerID="6e0143cb7ec4ee22e7f2a12722efb6a8345395ac858586f4608d91ec233f7f63" exitCode=0 Dec 13 07:04:18 crc kubenswrapper[4971]: I1213 07:04:18.318389 4971 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-z62gm" event={"ID":"af866491-3f91-4f43-809b-47b473e2e53d","Type":"ContainerDied","Data":"6e0143cb7ec4ee22e7f2a12722efb6a8345395ac858586f4608d91ec233f7f63"} Dec 13 07:04:18 crc kubenswrapper[4971]: I1213 07:04:18.322868 4971 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-swsx4" event={"ID":"a1c43378-fa47-4eb5-8c02-f5b229197543","Type":"ContainerStarted","Data":"8531e9be7ab8f8bb2a1b05961d7e19d670aa22f2b227ba2d0489e24acde6eb47"} Dec 13 07:04:18 crc kubenswrapper[4971]: I1213 07:04:18.331766 4971 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-szqz6" event={"ID":"33212ad2-538d-4059-bc1f-e98a51c93b62","Type":"ContainerStarted","Data":"c70ced632e40817b9f509e434aa8c97d6b0820d939ae0d413cc67401fd45f091"} Dec 13 07:04:18 crc kubenswrapper[4971]: I1213 07:04:18.335645 4971 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-hvzhk" event={"ID":"4b52f930-cf4c-4a97-b838-9afc9e3d2b31","Type":"ContainerStarted","Data":"07bb383dbdc0e76c1c71a9943c4a1dc5dad80b3881e8742b25c371c2a101378f"} Dec 13 07:04:18 crc kubenswrapper[4971]: I1213 07:04:18.337103 4971 generic.go:334] "Generic (PLEG): container finished" podID="6fa06763-60e2-4bc5-91f7-58ded0345c6a" containerID="8b613c55cfed3e81d7708653c1dc0521b726632ae9ae0684c1e3d5d6391065bd" exitCode=0 Dec 13 07:04:18 crc kubenswrapper[4971]: I1213 07:04:18.337142 4971 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-cvcsq" event={"ID":"6fa06763-60e2-4bc5-91f7-58ded0345c6a","Type":"ContainerDied","Data":"8b613c55cfed3e81d7708653c1dc0521b726632ae9ae0684c1e3d5d6391065bd"} Dec 13 07:04:18 crc kubenswrapper[4971]: I1213 07:04:18.348773 4971 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/certified-operators-swsx4" podStartSLOduration=2.636970856 podStartE2EDuration="5.348751157s" podCreationTimestamp="2025-12-13 07:04:13 +0000 UTC" firstStartedPulling="2025-12-13 07:04:15.252809627 +0000 UTC m=+911.857219075" lastFinishedPulling="2025-12-13 07:04:17.964589928 +0000 UTC m=+914.568999376" observedRunningTime="2025-12-13 07:04:18.34846289 +0000 UTC m=+914.952872338" watchObservedRunningTime="2025-12-13 07:04:18.348751157 +0000 UTC m=+914.953160635" Dec 13 07:04:18 crc kubenswrapper[4971]: I1213 07:04:18.371842 4971 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-marketplace-shhnv" podStartSLOduration=1.8753618890000001 podStartE2EDuration="4.371823775s" podCreationTimestamp="2025-12-13 07:04:14 +0000 UTC" firstStartedPulling="2025-12-13 07:04:15.247545027 +0000 UTC m=+911.851954475" lastFinishedPulling="2025-12-13 07:04:17.744006913 +0000 UTC m=+914.348416361" observedRunningTime="2025-12-13 07:04:18.369901538 +0000 UTC m=+914.974310986" watchObservedRunningTime="2025-12-13 07:04:18.371823775 +0000 UTC m=+914.976233223" Dec 13 07:04:18 crc kubenswrapper[4971]: I1213 07:04:18.411416 4971 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/certified-operators-szqz6" podStartSLOduration=2.957915145 podStartE2EDuration="5.411394661s" podCreationTimestamp="2025-12-13 07:04:13 +0000 UTC" firstStartedPulling="2025-12-13 07:04:15.23913864 +0000 UTC m=+911.843548088" lastFinishedPulling="2025-12-13 07:04:17.692618146 +0000 UTC m=+914.297027604" observedRunningTime="2025-12-13 07:04:18.409735119 +0000 UTC m=+915.014144577" watchObservedRunningTime="2025-12-13 07:04:18.411394661 +0000 UTC m=+915.015804119" Dec 13 07:04:18 crc kubenswrapper[4971]: I1213 07:04:18.437405 4971 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-marketplace-hvzhk" podStartSLOduration=2.876105778 podStartE2EDuration="5.437385651s" podCreationTimestamp="2025-12-13 07:04:13 +0000 UTC" firstStartedPulling="2025-12-13 07:04:15.244209825 +0000 UTC m=+911.848619273" lastFinishedPulling="2025-12-13 07:04:17.805489698 +0000 UTC m=+914.409899146" observedRunningTime="2025-12-13 07:04:18.434751656 +0000 UTC m=+915.039161114" watchObservedRunningTime="2025-12-13 07:04:18.437385651 +0000 UTC m=+915.041795099" Dec 13 07:04:19 crc kubenswrapper[4971]: I1213 07:04:19.345098 4971 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-nkmq2" event={"ID":"ad20cd8d-6bce-4bb4-b070-79c421d8538b","Type":"ContainerStarted","Data":"f98c411d11f40083a73ebdbe3de0949233eac394d1da31dfd9dc3f89bac488ec"} Dec 13 07:04:19 crc kubenswrapper[4971]: I1213 07:04:19.349117 4971 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-z62gm" event={"ID":"af866491-3f91-4f43-809b-47b473e2e53d","Type":"ContainerStarted","Data":"16ad47ec9ccf05452ca25d481805aa6c176618de22b4b31f0c2a40625bead6dc"} Dec 13 07:04:19 crc kubenswrapper[4971]: I1213 07:04:19.396183 4971 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/community-operators-z62gm" podStartSLOduration=1.885012368 podStartE2EDuration="3.39616642s" podCreationTimestamp="2025-12-13 07:04:16 +0000 UTC" firstStartedPulling="2025-12-13 07:04:17.280575501 +0000 UTC m=+913.884984949" lastFinishedPulling="2025-12-13 07:04:18.791729543 +0000 UTC m=+915.396139001" observedRunningTime="2025-12-13 07:04:19.393659308 +0000 UTC m=+915.998068756" watchObservedRunningTime="2025-12-13 07:04:19.39616642 +0000 UTC m=+916.000575868" Dec 13 07:04:20 crc kubenswrapper[4971]: I1213 07:04:20.357094 4971 generic.go:334] "Generic (PLEG): container finished" podID="6fa06763-60e2-4bc5-91f7-58ded0345c6a" containerID="3ecdd9f6115c2c199ebc892a032ded0ae0bfd8393e32ca3205bfcc948cef699e" exitCode=0 Dec 13 07:04:20 crc kubenswrapper[4971]: I1213 07:04:20.357145 4971 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-cvcsq" event={"ID":"6fa06763-60e2-4bc5-91f7-58ded0345c6a","Type":"ContainerDied","Data":"3ecdd9f6115c2c199ebc892a032ded0ae0bfd8393e32ca3205bfcc948cef699e"} Dec 13 07:04:20 crc kubenswrapper[4971]: I1213 07:04:20.359342 4971 generic.go:334] "Generic (PLEG): container finished" podID="ad20cd8d-6bce-4bb4-b070-79c421d8538b" containerID="f98c411d11f40083a73ebdbe3de0949233eac394d1da31dfd9dc3f89bac488ec" exitCode=0 Dec 13 07:04:20 crc kubenswrapper[4971]: I1213 07:04:20.359397 4971 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-nkmq2" event={"ID":"ad20cd8d-6bce-4bb4-b070-79c421d8538b","Type":"ContainerDied","Data":"f98c411d11f40083a73ebdbe3de0949233eac394d1da31dfd9dc3f89bac488ec"} Dec 13 07:04:23 crc kubenswrapper[4971]: I1213 07:04:23.872084 4971 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/certified-operators-szqz6" Dec 13 07:04:23 crc kubenswrapper[4971]: I1213 07:04:23.872146 4971 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/certified-operators-szqz6" Dec 13 07:04:23 crc kubenswrapper[4971]: I1213 07:04:23.911890 4971 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/certified-operators-szqz6" Dec 13 07:04:24 crc kubenswrapper[4971]: I1213 07:04:24.056361 4971 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-marketplace-hvzhk" Dec 13 07:04:24 crc kubenswrapper[4971]: I1213 07:04:24.056409 4971 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-marketplace-hvzhk" Dec 13 07:04:24 crc kubenswrapper[4971]: I1213 07:04:24.096744 4971 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-marketplace-hvzhk" Dec 13 07:04:24 crc kubenswrapper[4971]: I1213 07:04:24.258354 4971 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/certified-operators-swsx4" Dec 13 07:04:24 crc kubenswrapper[4971]: I1213 07:04:24.258398 4971 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/certified-operators-swsx4" Dec 13 07:04:24 crc kubenswrapper[4971]: I1213 07:04:24.295168 4971 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/certified-operators-swsx4" Dec 13 07:04:24 crc kubenswrapper[4971]: I1213 07:04:24.418953 4971 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/certified-operators-szqz6" Dec 13 07:04:24 crc kubenswrapper[4971]: I1213 07:04:24.422132 4971 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-marketplace-hvzhk" Dec 13 07:04:24 crc kubenswrapper[4971]: I1213 07:04:24.423897 4971 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/certified-operators-swsx4" Dec 13 07:04:24 crc kubenswrapper[4971]: I1213 07:04:24.479200 4971 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-marketplace-shhnv" Dec 13 07:04:24 crc kubenswrapper[4971]: I1213 07:04:24.479251 4971 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-marketplace-shhnv" Dec 13 07:04:24 crc kubenswrapper[4971]: I1213 07:04:24.523363 4971 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-marketplace-shhnv" Dec 13 07:04:25 crc kubenswrapper[4971]: I1213 07:04:25.426782 4971 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-marketplace-shhnv" Dec 13 07:04:26 crc kubenswrapper[4971]: I1213 07:04:26.396685 4971 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-cvcsq" event={"ID":"6fa06763-60e2-4bc5-91f7-58ded0345c6a","Type":"ContainerStarted","Data":"50a648ed33360842170d924504c03ce609bd5aaf4bc47c3f2e55e53ce73bc0ad"} Dec 13 07:04:26 crc kubenswrapper[4971]: I1213 07:04:26.399611 4971 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-nkmq2" event={"ID":"ad20cd8d-6bce-4bb4-b070-79c421d8538b","Type":"ContainerStarted","Data":"50948e2f3a10dd98e0c3e9107e86400bf58a6a65128383ab0e80f0bf415b5ef8"} Dec 13 07:04:26 crc kubenswrapper[4971]: I1213 07:04:26.439812 4971 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-operators-cvcsq" podStartSLOduration=3.2943912920000002 podStartE2EDuration="10.43978941s" podCreationTimestamp="2025-12-13 07:04:16 +0000 UTC" firstStartedPulling="2025-12-13 07:04:18.338435992 +0000 UTC m=+914.942845440" lastFinishedPulling="2025-12-13 07:04:25.48383411 +0000 UTC m=+922.088243558" observedRunningTime="2025-12-13 07:04:26.417162872 +0000 UTC m=+923.021572320" watchObservedRunningTime="2025-12-13 07:04:26.43978941 +0000 UTC m=+923.044198858" Dec 13 07:04:26 crc kubenswrapper[4971]: I1213 07:04:26.756474 4971 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-operators-cvcsq" Dec 13 07:04:26 crc kubenswrapper[4971]: I1213 07:04:26.758251 4971 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-operators-cvcsq" Dec 13 07:04:26 crc kubenswrapper[4971]: I1213 07:04:26.900415 4971 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/community-operators-z62gm" Dec 13 07:04:26 crc kubenswrapper[4971]: I1213 07:04:26.900476 4971 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/community-operators-z62gm" Dec 13 07:04:26 crc kubenswrapper[4971]: I1213 07:04:26.924168 4971 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-operators-nkmq2" podStartSLOduration=3.7074535920000002 podStartE2EDuration="10.924145807s" podCreationTimestamp="2025-12-13 07:04:16 +0000 UTC" firstStartedPulling="2025-12-13 07:04:18.307695915 +0000 UTC m=+914.912105363" lastFinishedPulling="2025-12-13 07:04:25.52438813 +0000 UTC m=+922.128797578" observedRunningTime="2025-12-13 07:04:26.441949393 +0000 UTC m=+923.046358841" watchObservedRunningTime="2025-12-13 07:04:26.924145807 +0000 UTC m=+923.528555255" Dec 13 07:04:26 crc kubenswrapper[4971]: I1213 07:04:26.927055 4971 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-swsx4"] Dec 13 07:04:26 crc kubenswrapper[4971]: I1213 07:04:26.927310 4971 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/certified-operators-swsx4" podUID="a1c43378-fa47-4eb5-8c02-f5b229197543" containerName="registry-server" containerID="cri-o://8531e9be7ab8f8bb2a1b05961d7e19d670aa22f2b227ba2d0489e24acde6eb47" gracePeriod=2 Dec 13 07:04:26 crc kubenswrapper[4971]: I1213 07:04:26.946185 4971 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/community-operators-z62gm" Dec 13 07:04:27 crc kubenswrapper[4971]: I1213 07:04:27.066933 4971 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-operators-nkmq2" Dec 13 07:04:27 crc kubenswrapper[4971]: I1213 07:04:27.066980 4971 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-operators-nkmq2" Dec 13 07:04:27 crc kubenswrapper[4971]: I1213 07:04:27.124429 4971 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-shhnv"] Dec 13 07:04:27 crc kubenswrapper[4971]: I1213 07:04:27.404050 4971 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-marketplace-shhnv" podUID="e80b7d34-d09e-4efe-9879-eb54ea91d724" containerName="registry-server" containerID="cri-o://0ef38416f602cf2e66532ce278d058da4845515335b64b9ecea801fcb9eb4fe9" gracePeriod=2 Dec 13 07:04:27 crc kubenswrapper[4971]: I1213 07:04:27.459961 4971 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/community-operators-z62gm" Dec 13 07:04:27 crc kubenswrapper[4971]: I1213 07:04:27.797676 4971 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-marketplace/redhat-operators-cvcsq" podUID="6fa06763-60e2-4bc5-91f7-58ded0345c6a" containerName="registry-server" probeResult="failure" output=< Dec 13 07:04:27 crc kubenswrapper[4971]: timeout: failed to connect service ":50051" within 1s Dec 13 07:04:27 crc kubenswrapper[4971]: > Dec 13 07:04:28 crc kubenswrapper[4971]: I1213 07:04:28.106123 4971 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-marketplace/redhat-operators-nkmq2" podUID="ad20cd8d-6bce-4bb4-b070-79c421d8538b" containerName="registry-server" probeResult="failure" output=< Dec 13 07:04:28 crc kubenswrapper[4971]: timeout: failed to connect service ":50051" within 1s Dec 13 07:04:28 crc kubenswrapper[4971]: > Dec 13 07:04:34 crc kubenswrapper[4971]: E1213 07:04:34.258603 4971 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = NotFound desc = container is not created or running: checking if PID of 8531e9be7ab8f8bb2a1b05961d7e19d670aa22f2b227ba2d0489e24acde6eb47 is running failed: container process not found" containerID="8531e9be7ab8f8bb2a1b05961d7e19d670aa22f2b227ba2d0489e24acde6eb47" cmd=["grpc_health_probe","-addr=:50051"] Dec 13 07:04:34 crc kubenswrapper[4971]: E1213 07:04:34.259397 4971 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = NotFound desc = container is not created or running: checking if PID of 8531e9be7ab8f8bb2a1b05961d7e19d670aa22f2b227ba2d0489e24acde6eb47 is running failed: container process not found" containerID="8531e9be7ab8f8bb2a1b05961d7e19d670aa22f2b227ba2d0489e24acde6eb47" cmd=["grpc_health_probe","-addr=:50051"] Dec 13 07:04:34 crc kubenswrapper[4971]: E1213 07:04:34.259658 4971 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = NotFound desc = container is not created or running: checking if PID of 8531e9be7ab8f8bb2a1b05961d7e19d670aa22f2b227ba2d0489e24acde6eb47 is running failed: container process not found" containerID="8531e9be7ab8f8bb2a1b05961d7e19d670aa22f2b227ba2d0489e24acde6eb47" cmd=["grpc_health_probe","-addr=:50051"] Dec 13 07:04:34 crc kubenswrapper[4971]: E1213 07:04:34.259693 4971 prober.go:104] "Probe errored" err="rpc error: code = NotFound desc = container is not created or running: checking if PID of 8531e9be7ab8f8bb2a1b05961d7e19d670aa22f2b227ba2d0489e24acde6eb47 is running failed: container process not found" probeType="Readiness" pod="openshift-marketplace/certified-operators-swsx4" podUID="a1c43378-fa47-4eb5-8c02-f5b229197543" containerName="registry-server" Dec 13 07:04:34 crc kubenswrapper[4971]: E1213 07:04:34.480477 4971 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = NotFound desc = container is not created or running: checking if PID of 0ef38416f602cf2e66532ce278d058da4845515335b64b9ecea801fcb9eb4fe9 is running failed: container process not found" containerID="0ef38416f602cf2e66532ce278d058da4845515335b64b9ecea801fcb9eb4fe9" cmd=["grpc_health_probe","-addr=:50051"] Dec 13 07:04:34 crc kubenswrapper[4971]: E1213 07:04:34.481048 4971 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = NotFound desc = container is not created or running: checking if PID of 0ef38416f602cf2e66532ce278d058da4845515335b64b9ecea801fcb9eb4fe9 is running failed: container process not found" containerID="0ef38416f602cf2e66532ce278d058da4845515335b64b9ecea801fcb9eb4fe9" cmd=["grpc_health_probe","-addr=:50051"] Dec 13 07:04:34 crc kubenswrapper[4971]: E1213 07:04:34.481448 4971 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = NotFound desc = container is not created or running: checking if PID of 0ef38416f602cf2e66532ce278d058da4845515335b64b9ecea801fcb9eb4fe9 is running failed: container process not found" containerID="0ef38416f602cf2e66532ce278d058da4845515335b64b9ecea801fcb9eb4fe9" cmd=["grpc_health_probe","-addr=:50051"] Dec 13 07:04:34 crc kubenswrapper[4971]: E1213 07:04:34.481485 4971 prober.go:104] "Probe errored" err="rpc error: code = NotFound desc = container is not created or running: checking if PID of 0ef38416f602cf2e66532ce278d058da4845515335b64b9ecea801fcb9eb4fe9 is running failed: container process not found" probeType="Readiness" pod="openshift-marketplace/redhat-marketplace-shhnv" podUID="e80b7d34-d09e-4efe-9879-eb54ea91d724" containerName="registry-server" Dec 13 07:04:35 crc kubenswrapper[4971]: I1213 07:04:35.159285 4971 generic.go:334] "Generic (PLEG): container finished" podID="a1c43378-fa47-4eb5-8c02-f5b229197543" containerID="8531e9be7ab8f8bb2a1b05961d7e19d670aa22f2b227ba2d0489e24acde6eb47" exitCode=0 Dec 13 07:04:35 crc kubenswrapper[4971]: I1213 07:04:35.159360 4971 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-swsx4" event={"ID":"a1c43378-fa47-4eb5-8c02-f5b229197543","Type":"ContainerDied","Data":"8531e9be7ab8f8bb2a1b05961d7e19d670aa22f2b227ba2d0489e24acde6eb47"} Dec 13 07:04:35 crc kubenswrapper[4971]: I1213 07:04:35.614126 4971 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-shhnv" Dec 13 07:04:35 crc kubenswrapper[4971]: I1213 07:04:35.646424 4971 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-swsx4" Dec 13 07:04:35 crc kubenswrapper[4971]: I1213 07:04:35.674125 4971 generic.go:334] "Generic (PLEG): container finished" podID="e80b7d34-d09e-4efe-9879-eb54ea91d724" containerID="0ef38416f602cf2e66532ce278d058da4845515335b64b9ecea801fcb9eb4fe9" exitCode=0 Dec 13 07:04:35 crc kubenswrapper[4971]: I1213 07:04:35.674200 4971 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-shhnv" event={"ID":"e80b7d34-d09e-4efe-9879-eb54ea91d724","Type":"ContainerDied","Data":"0ef38416f602cf2e66532ce278d058da4845515335b64b9ecea801fcb9eb4fe9"} Dec 13 07:04:35 crc kubenswrapper[4971]: I1213 07:04:35.674243 4971 scope.go:117] "RemoveContainer" containerID="0ef38416f602cf2e66532ce278d058da4845515335b64b9ecea801fcb9eb4fe9" Dec 13 07:04:35 crc kubenswrapper[4971]: I1213 07:04:35.701220 4971 scope.go:117] "RemoveContainer" containerID="56524017b2b0b269c175ed0412e13d140a8e040fe25b81a7284fd32f4fb24d12" Dec 13 07:04:35 crc kubenswrapper[4971]: I1213 07:04:35.715982 4971 scope.go:117] "RemoveContainer" containerID="d2462b9bb30042d8c344b2f90b21fe0d461a42be2f9be3b1f77ed7d336a8cf49" Dec 13 07:04:35 crc kubenswrapper[4971]: I1213 07:04:35.796418 4971 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/a1c43378-fa47-4eb5-8c02-f5b229197543-catalog-content\") pod \"a1c43378-fa47-4eb5-8c02-f5b229197543\" (UID: \"a1c43378-fa47-4eb5-8c02-f5b229197543\") " Dec 13 07:04:35 crc kubenswrapper[4971]: I1213 07:04:35.796532 4971 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-rtlcd\" (UniqueName: \"kubernetes.io/projected/e80b7d34-d09e-4efe-9879-eb54ea91d724-kube-api-access-rtlcd\") pod \"e80b7d34-d09e-4efe-9879-eb54ea91d724\" (UID: \"e80b7d34-d09e-4efe-9879-eb54ea91d724\") " Dec 13 07:04:35 crc kubenswrapper[4971]: I1213 07:04:35.796569 4971 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/a1c43378-fa47-4eb5-8c02-f5b229197543-utilities\") pod \"a1c43378-fa47-4eb5-8c02-f5b229197543\" (UID: \"a1c43378-fa47-4eb5-8c02-f5b229197543\") " Dec 13 07:04:35 crc kubenswrapper[4971]: I1213 07:04:35.796594 4971 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-dpr4c\" (UniqueName: \"kubernetes.io/projected/a1c43378-fa47-4eb5-8c02-f5b229197543-kube-api-access-dpr4c\") pod \"a1c43378-fa47-4eb5-8c02-f5b229197543\" (UID: \"a1c43378-fa47-4eb5-8c02-f5b229197543\") " Dec 13 07:04:35 crc kubenswrapper[4971]: I1213 07:04:35.796631 4971 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/e80b7d34-d09e-4efe-9879-eb54ea91d724-utilities\") pod \"e80b7d34-d09e-4efe-9879-eb54ea91d724\" (UID: \"e80b7d34-d09e-4efe-9879-eb54ea91d724\") " Dec 13 07:04:35 crc kubenswrapper[4971]: I1213 07:04:35.796721 4971 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/e80b7d34-d09e-4efe-9879-eb54ea91d724-catalog-content\") pod \"e80b7d34-d09e-4efe-9879-eb54ea91d724\" (UID: \"e80b7d34-d09e-4efe-9879-eb54ea91d724\") " Dec 13 07:04:35 crc kubenswrapper[4971]: I1213 07:04:35.797627 4971 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/a1c43378-fa47-4eb5-8c02-f5b229197543-utilities" (OuterVolumeSpecName: "utilities") pod "a1c43378-fa47-4eb5-8c02-f5b229197543" (UID: "a1c43378-fa47-4eb5-8c02-f5b229197543"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 13 07:04:35 crc kubenswrapper[4971]: I1213 07:04:35.798810 4971 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/e80b7d34-d09e-4efe-9879-eb54ea91d724-utilities" (OuterVolumeSpecName: "utilities") pod "e80b7d34-d09e-4efe-9879-eb54ea91d724" (UID: "e80b7d34-d09e-4efe-9879-eb54ea91d724"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 13 07:04:35 crc kubenswrapper[4971]: I1213 07:04:35.806406 4971 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/e80b7d34-d09e-4efe-9879-eb54ea91d724-kube-api-access-rtlcd" (OuterVolumeSpecName: "kube-api-access-rtlcd") pod "e80b7d34-d09e-4efe-9879-eb54ea91d724" (UID: "e80b7d34-d09e-4efe-9879-eb54ea91d724"). InnerVolumeSpecName "kube-api-access-rtlcd". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 13 07:04:35 crc kubenswrapper[4971]: I1213 07:04:35.806541 4971 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/a1c43378-fa47-4eb5-8c02-f5b229197543-kube-api-access-dpr4c" (OuterVolumeSpecName: "kube-api-access-dpr4c") pod "a1c43378-fa47-4eb5-8c02-f5b229197543" (UID: "a1c43378-fa47-4eb5-8c02-f5b229197543"). InnerVolumeSpecName "kube-api-access-dpr4c". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 13 07:04:35 crc kubenswrapper[4971]: I1213 07:04:35.828024 4971 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/e80b7d34-d09e-4efe-9879-eb54ea91d724-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "e80b7d34-d09e-4efe-9879-eb54ea91d724" (UID: "e80b7d34-d09e-4efe-9879-eb54ea91d724"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 13 07:04:35 crc kubenswrapper[4971]: I1213 07:04:35.858777 4971 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/a1c43378-fa47-4eb5-8c02-f5b229197543-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "a1c43378-fa47-4eb5-8c02-f5b229197543" (UID: "a1c43378-fa47-4eb5-8c02-f5b229197543"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 13 07:04:35 crc kubenswrapper[4971]: I1213 07:04:35.897844 4971 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/e80b7d34-d09e-4efe-9879-eb54ea91d724-catalog-content\") on node \"crc\" DevicePath \"\"" Dec 13 07:04:35 crc kubenswrapper[4971]: I1213 07:04:35.897875 4971 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/a1c43378-fa47-4eb5-8c02-f5b229197543-catalog-content\") on node \"crc\" DevicePath \"\"" Dec 13 07:04:35 crc kubenswrapper[4971]: I1213 07:04:35.897884 4971 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-rtlcd\" (UniqueName: \"kubernetes.io/projected/e80b7d34-d09e-4efe-9879-eb54ea91d724-kube-api-access-rtlcd\") on node \"crc\" DevicePath \"\"" Dec 13 07:04:35 crc kubenswrapper[4971]: I1213 07:04:35.897894 4971 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/a1c43378-fa47-4eb5-8c02-f5b229197543-utilities\") on node \"crc\" DevicePath \"\"" Dec 13 07:04:35 crc kubenswrapper[4971]: I1213 07:04:35.897903 4971 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-dpr4c\" (UniqueName: \"kubernetes.io/projected/a1c43378-fa47-4eb5-8c02-f5b229197543-kube-api-access-dpr4c\") on node \"crc\" DevicePath \"\"" Dec 13 07:04:35 crc kubenswrapper[4971]: I1213 07:04:35.897912 4971 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/e80b7d34-d09e-4efe-9879-eb54ea91d724-utilities\") on node \"crc\" DevicePath \"\"" Dec 13 07:04:36 crc kubenswrapper[4971]: I1213 07:04:36.682156 4971 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-swsx4" event={"ID":"a1c43378-fa47-4eb5-8c02-f5b229197543","Type":"ContainerDied","Data":"6db3142585544dc7898dfe4faf03f4f84d8a8689319acc85d3f4860b94fd4b6c"} Dec 13 07:04:36 crc kubenswrapper[4971]: I1213 07:04:36.682169 4971 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-swsx4" Dec 13 07:04:36 crc kubenswrapper[4971]: I1213 07:04:36.682482 4971 scope.go:117] "RemoveContainer" containerID="8531e9be7ab8f8bb2a1b05961d7e19d670aa22f2b227ba2d0489e24acde6eb47" Dec 13 07:04:36 crc kubenswrapper[4971]: I1213 07:04:36.683502 4971 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-shhnv" event={"ID":"e80b7d34-d09e-4efe-9879-eb54ea91d724","Type":"ContainerDied","Data":"8dcf1bb6b6cc62194648dd926d4570e9e85c06a944021a584a810c808efecbb3"} Dec 13 07:04:36 crc kubenswrapper[4971]: I1213 07:04:36.683595 4971 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-shhnv" Dec 13 07:04:36 crc kubenswrapper[4971]: I1213 07:04:36.716485 4971 scope.go:117] "RemoveContainer" containerID="d1e668b1d78683413c01dcb122882b84b945a36974492f19eca6e50614791f08" Dec 13 07:04:36 crc kubenswrapper[4971]: I1213 07:04:36.721966 4971 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-shhnv"] Dec 13 07:04:36 crc kubenswrapper[4971]: I1213 07:04:36.727654 4971 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-marketplace-shhnv"] Dec 13 07:04:36 crc kubenswrapper[4971]: I1213 07:04:36.735637 4971 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-swsx4"] Dec 13 07:04:36 crc kubenswrapper[4971]: I1213 07:04:36.742095 4971 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/certified-operators-swsx4"] Dec 13 07:04:36 crc kubenswrapper[4971]: I1213 07:04:36.749409 4971 scope.go:117] "RemoveContainer" containerID="37401099d2099e897ac4109f6bbcb2a1b68df9c72d37005a6b6e4c46a86b6956" Dec 13 07:04:36 crc kubenswrapper[4971]: I1213 07:04:36.795094 4971 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-operators-cvcsq" Dec 13 07:04:36 crc kubenswrapper[4971]: I1213 07:04:36.834534 4971 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-operators-cvcsq" Dec 13 07:04:37 crc kubenswrapper[4971]: I1213 07:04:37.101461 4971 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-operators-nkmq2" Dec 13 07:04:37 crc kubenswrapper[4971]: I1213 07:04:37.137313 4971 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-operators-nkmq2" Dec 13 07:04:37 crc kubenswrapper[4971]: I1213 07:04:37.774772 4971 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="a1c43378-fa47-4eb5-8c02-f5b229197543" path="/var/lib/kubelet/pods/a1c43378-fa47-4eb5-8c02-f5b229197543/volumes" Dec 13 07:04:37 crc kubenswrapper[4971]: I1213 07:04:37.775498 4971 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="e80b7d34-d09e-4efe-9879-eb54ea91d724" path="/var/lib/kubelet/pods/e80b7d34-d09e-4efe-9879-eb54ea91d724/volumes" Dec 13 07:04:39 crc kubenswrapper[4971]: I1213 07:04:39.927121 4971 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-nkmq2"] Dec 13 07:04:39 crc kubenswrapper[4971]: I1213 07:04:39.927364 4971 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-operators-nkmq2" podUID="ad20cd8d-6bce-4bb4-b070-79c421d8538b" containerName="registry-server" containerID="cri-o://50948e2f3a10dd98e0c3e9107e86400bf58a6a65128383ab0e80f0bf415b5ef8" gracePeriod=2 Dec 13 07:04:40 crc kubenswrapper[4971]: I1213 07:04:40.323813 4971 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-nkmq2" Dec 13 07:04:40 crc kubenswrapper[4971]: I1213 07:04:40.457657 4971 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/ad20cd8d-6bce-4bb4-b070-79c421d8538b-utilities\") pod \"ad20cd8d-6bce-4bb4-b070-79c421d8538b\" (UID: \"ad20cd8d-6bce-4bb4-b070-79c421d8538b\") " Dec 13 07:04:40 crc kubenswrapper[4971]: I1213 07:04:40.457728 4971 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-zmpxz\" (UniqueName: \"kubernetes.io/projected/ad20cd8d-6bce-4bb4-b070-79c421d8538b-kube-api-access-zmpxz\") pod \"ad20cd8d-6bce-4bb4-b070-79c421d8538b\" (UID: \"ad20cd8d-6bce-4bb4-b070-79c421d8538b\") " Dec 13 07:04:40 crc kubenswrapper[4971]: I1213 07:04:40.457811 4971 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/ad20cd8d-6bce-4bb4-b070-79c421d8538b-catalog-content\") pod \"ad20cd8d-6bce-4bb4-b070-79c421d8538b\" (UID: \"ad20cd8d-6bce-4bb4-b070-79c421d8538b\") " Dec 13 07:04:40 crc kubenswrapper[4971]: I1213 07:04:40.458917 4971 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/ad20cd8d-6bce-4bb4-b070-79c421d8538b-utilities" (OuterVolumeSpecName: "utilities") pod "ad20cd8d-6bce-4bb4-b070-79c421d8538b" (UID: "ad20cd8d-6bce-4bb4-b070-79c421d8538b"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 13 07:04:40 crc kubenswrapper[4971]: I1213 07:04:40.463896 4971 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/ad20cd8d-6bce-4bb4-b070-79c421d8538b-kube-api-access-zmpxz" (OuterVolumeSpecName: "kube-api-access-zmpxz") pod "ad20cd8d-6bce-4bb4-b070-79c421d8538b" (UID: "ad20cd8d-6bce-4bb4-b070-79c421d8538b"). InnerVolumeSpecName "kube-api-access-zmpxz". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 13 07:04:40 crc kubenswrapper[4971]: I1213 07:04:40.559586 4971 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-zmpxz\" (UniqueName: \"kubernetes.io/projected/ad20cd8d-6bce-4bb4-b070-79c421d8538b-kube-api-access-zmpxz\") on node \"crc\" DevicePath \"\"" Dec 13 07:04:40 crc kubenswrapper[4971]: I1213 07:04:40.559632 4971 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/ad20cd8d-6bce-4bb4-b070-79c421d8538b-utilities\") on node \"crc\" DevicePath \"\"" Dec 13 07:04:40 crc kubenswrapper[4971]: I1213 07:04:40.580359 4971 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/ad20cd8d-6bce-4bb4-b070-79c421d8538b-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "ad20cd8d-6bce-4bb4-b070-79c421d8538b" (UID: "ad20cd8d-6bce-4bb4-b070-79c421d8538b"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 13 07:04:40 crc kubenswrapper[4971]: I1213 07:04:40.660684 4971 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/ad20cd8d-6bce-4bb4-b070-79c421d8538b-catalog-content\") on node \"crc\" DevicePath \"\"" Dec 13 07:04:40 crc kubenswrapper[4971]: I1213 07:04:40.709147 4971 generic.go:334] "Generic (PLEG): container finished" podID="ad20cd8d-6bce-4bb4-b070-79c421d8538b" containerID="50948e2f3a10dd98e0c3e9107e86400bf58a6a65128383ab0e80f0bf415b5ef8" exitCode=0 Dec 13 07:04:40 crc kubenswrapper[4971]: I1213 07:04:40.709196 4971 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-nkmq2" event={"ID":"ad20cd8d-6bce-4bb4-b070-79c421d8538b","Type":"ContainerDied","Data":"50948e2f3a10dd98e0c3e9107e86400bf58a6a65128383ab0e80f0bf415b5ef8"} Dec 13 07:04:40 crc kubenswrapper[4971]: I1213 07:04:40.709228 4971 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-nkmq2" event={"ID":"ad20cd8d-6bce-4bb4-b070-79c421d8538b","Type":"ContainerDied","Data":"87b65ae489e534508ed1ac5ff7f225e95b6846e1c11eba52eb3068d56e274d2e"} Dec 13 07:04:40 crc kubenswrapper[4971]: I1213 07:04:40.709248 4971 scope.go:117] "RemoveContainer" containerID="50948e2f3a10dd98e0c3e9107e86400bf58a6a65128383ab0e80f0bf415b5ef8" Dec 13 07:04:40 crc kubenswrapper[4971]: I1213 07:04:40.709397 4971 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-nkmq2" Dec 13 07:04:40 crc kubenswrapper[4971]: I1213 07:04:40.728557 4971 scope.go:117] "RemoveContainer" containerID="f98c411d11f40083a73ebdbe3de0949233eac394d1da31dfd9dc3f89bac488ec" Dec 13 07:04:40 crc kubenswrapper[4971]: I1213 07:04:40.736544 4971 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-nkmq2"] Dec 13 07:04:40 crc kubenswrapper[4971]: I1213 07:04:40.745346 4971 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-operators-nkmq2"] Dec 13 07:04:40 crc kubenswrapper[4971]: I1213 07:04:40.758005 4971 scope.go:117] "RemoveContainer" containerID="e02224574d97754dca872708b7564e8fd9e0528d19fbf7fe118b6bc93481dbbe" Dec 13 07:04:40 crc kubenswrapper[4971]: I1213 07:04:40.770612 4971 scope.go:117] "RemoveContainer" containerID="50948e2f3a10dd98e0c3e9107e86400bf58a6a65128383ab0e80f0bf415b5ef8" Dec 13 07:04:40 crc kubenswrapper[4971]: E1213 07:04:40.771128 4971 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"50948e2f3a10dd98e0c3e9107e86400bf58a6a65128383ab0e80f0bf415b5ef8\": container with ID starting with 50948e2f3a10dd98e0c3e9107e86400bf58a6a65128383ab0e80f0bf415b5ef8 not found: ID does not exist" containerID="50948e2f3a10dd98e0c3e9107e86400bf58a6a65128383ab0e80f0bf415b5ef8" Dec 13 07:04:40 crc kubenswrapper[4971]: I1213 07:04:40.771174 4971 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"50948e2f3a10dd98e0c3e9107e86400bf58a6a65128383ab0e80f0bf415b5ef8"} err="failed to get container status \"50948e2f3a10dd98e0c3e9107e86400bf58a6a65128383ab0e80f0bf415b5ef8\": rpc error: code = NotFound desc = could not find container \"50948e2f3a10dd98e0c3e9107e86400bf58a6a65128383ab0e80f0bf415b5ef8\": container with ID starting with 50948e2f3a10dd98e0c3e9107e86400bf58a6a65128383ab0e80f0bf415b5ef8 not found: ID does not exist" Dec 13 07:04:40 crc kubenswrapper[4971]: I1213 07:04:40.771201 4971 scope.go:117] "RemoveContainer" containerID="f98c411d11f40083a73ebdbe3de0949233eac394d1da31dfd9dc3f89bac488ec" Dec 13 07:04:40 crc kubenswrapper[4971]: E1213 07:04:40.771581 4971 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"f98c411d11f40083a73ebdbe3de0949233eac394d1da31dfd9dc3f89bac488ec\": container with ID starting with f98c411d11f40083a73ebdbe3de0949233eac394d1da31dfd9dc3f89bac488ec not found: ID does not exist" containerID="f98c411d11f40083a73ebdbe3de0949233eac394d1da31dfd9dc3f89bac488ec" Dec 13 07:04:40 crc kubenswrapper[4971]: I1213 07:04:40.771663 4971 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"f98c411d11f40083a73ebdbe3de0949233eac394d1da31dfd9dc3f89bac488ec"} err="failed to get container status \"f98c411d11f40083a73ebdbe3de0949233eac394d1da31dfd9dc3f89bac488ec\": rpc error: code = NotFound desc = could not find container \"f98c411d11f40083a73ebdbe3de0949233eac394d1da31dfd9dc3f89bac488ec\": container with ID starting with f98c411d11f40083a73ebdbe3de0949233eac394d1da31dfd9dc3f89bac488ec not found: ID does not exist" Dec 13 07:04:40 crc kubenswrapper[4971]: I1213 07:04:40.771693 4971 scope.go:117] "RemoveContainer" containerID="e02224574d97754dca872708b7564e8fd9e0528d19fbf7fe118b6bc93481dbbe" Dec 13 07:04:40 crc kubenswrapper[4971]: E1213 07:04:40.771987 4971 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"e02224574d97754dca872708b7564e8fd9e0528d19fbf7fe118b6bc93481dbbe\": container with ID starting with e02224574d97754dca872708b7564e8fd9e0528d19fbf7fe118b6bc93481dbbe not found: ID does not exist" containerID="e02224574d97754dca872708b7564e8fd9e0528d19fbf7fe118b6bc93481dbbe" Dec 13 07:04:40 crc kubenswrapper[4971]: I1213 07:04:40.772045 4971 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"e02224574d97754dca872708b7564e8fd9e0528d19fbf7fe118b6bc93481dbbe"} err="failed to get container status \"e02224574d97754dca872708b7564e8fd9e0528d19fbf7fe118b6bc93481dbbe\": rpc error: code = NotFound desc = could not find container \"e02224574d97754dca872708b7564e8fd9e0528d19fbf7fe118b6bc93481dbbe\": container with ID starting with e02224574d97754dca872708b7564e8fd9e0528d19fbf7fe118b6bc93481dbbe not found: ID does not exist" Dec 13 07:04:41 crc kubenswrapper[4971]: I1213 07:04:41.775222 4971 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="ad20cd8d-6bce-4bb4-b070-79c421d8538b" path="/var/lib/kubelet/pods/ad20cd8d-6bce-4bb4-b070-79c421d8538b/volumes" Dec 13 07:04:46 crc kubenswrapper[4971]: I1213 07:04:46.153334 4971 patch_prober.go:28] interesting pod/machine-config-daemon-82xjz container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 13 07:04:46 crc kubenswrapper[4971]: I1213 07:04:46.153710 4971 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-82xjz" podUID="e0c4af12-edda-41d7-9b44-f87396174ef2" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 13 07:04:53 crc kubenswrapper[4971]: I1213 07:04:53.980099 4971 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/98085b0df3808ebec39f9f9529f737144fe2dbcdaa4f334014817c0fa89s2m9"] Dec 13 07:04:53 crc kubenswrapper[4971]: E1213 07:04:53.980874 4971 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="ad20cd8d-6bce-4bb4-b070-79c421d8538b" containerName="registry-server" Dec 13 07:04:53 crc kubenswrapper[4971]: I1213 07:04:53.980888 4971 state_mem.go:107] "Deleted CPUSet assignment" podUID="ad20cd8d-6bce-4bb4-b070-79c421d8538b" containerName="registry-server" Dec 13 07:04:53 crc kubenswrapper[4971]: E1213 07:04:53.980900 4971 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="ad20cd8d-6bce-4bb4-b070-79c421d8538b" containerName="extract-utilities" Dec 13 07:04:53 crc kubenswrapper[4971]: I1213 07:04:53.980905 4971 state_mem.go:107] "Deleted CPUSet assignment" podUID="ad20cd8d-6bce-4bb4-b070-79c421d8538b" containerName="extract-utilities" Dec 13 07:04:53 crc kubenswrapper[4971]: E1213 07:04:53.980914 4971 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="a1c43378-fa47-4eb5-8c02-f5b229197543" containerName="registry-server" Dec 13 07:04:53 crc kubenswrapper[4971]: I1213 07:04:53.980921 4971 state_mem.go:107] "Deleted CPUSet assignment" podUID="a1c43378-fa47-4eb5-8c02-f5b229197543" containerName="registry-server" Dec 13 07:04:53 crc kubenswrapper[4971]: E1213 07:04:53.980930 4971 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="a1c43378-fa47-4eb5-8c02-f5b229197543" containerName="extract-content" Dec 13 07:04:53 crc kubenswrapper[4971]: I1213 07:04:53.980935 4971 state_mem.go:107] "Deleted CPUSet assignment" podUID="a1c43378-fa47-4eb5-8c02-f5b229197543" containerName="extract-content" Dec 13 07:04:53 crc kubenswrapper[4971]: E1213 07:04:53.980946 4971 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="a1c43378-fa47-4eb5-8c02-f5b229197543" containerName="extract-utilities" Dec 13 07:04:53 crc kubenswrapper[4971]: I1213 07:04:53.980952 4971 state_mem.go:107] "Deleted CPUSet assignment" podUID="a1c43378-fa47-4eb5-8c02-f5b229197543" containerName="extract-utilities" Dec 13 07:04:53 crc kubenswrapper[4971]: E1213 07:04:53.980962 4971 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="e80b7d34-d09e-4efe-9879-eb54ea91d724" containerName="extract-content" Dec 13 07:04:53 crc kubenswrapper[4971]: I1213 07:04:53.980968 4971 state_mem.go:107] "Deleted CPUSet assignment" podUID="e80b7d34-d09e-4efe-9879-eb54ea91d724" containerName="extract-content" Dec 13 07:04:53 crc kubenswrapper[4971]: E1213 07:04:53.980980 4971 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="e80b7d34-d09e-4efe-9879-eb54ea91d724" containerName="extract-utilities" Dec 13 07:04:53 crc kubenswrapper[4971]: I1213 07:04:53.980987 4971 state_mem.go:107] "Deleted CPUSet assignment" podUID="e80b7d34-d09e-4efe-9879-eb54ea91d724" containerName="extract-utilities" Dec 13 07:04:53 crc kubenswrapper[4971]: E1213 07:04:53.980995 4971 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="ad20cd8d-6bce-4bb4-b070-79c421d8538b" containerName="extract-content" Dec 13 07:04:53 crc kubenswrapper[4971]: I1213 07:04:53.981001 4971 state_mem.go:107] "Deleted CPUSet assignment" podUID="ad20cd8d-6bce-4bb4-b070-79c421d8538b" containerName="extract-content" Dec 13 07:04:53 crc kubenswrapper[4971]: E1213 07:04:53.981012 4971 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="e80b7d34-d09e-4efe-9879-eb54ea91d724" containerName="registry-server" Dec 13 07:04:53 crc kubenswrapper[4971]: I1213 07:04:53.981017 4971 state_mem.go:107] "Deleted CPUSet assignment" podUID="e80b7d34-d09e-4efe-9879-eb54ea91d724" containerName="registry-server" Dec 13 07:04:53 crc kubenswrapper[4971]: I1213 07:04:53.981103 4971 memory_manager.go:354] "RemoveStaleState removing state" podUID="e80b7d34-d09e-4efe-9879-eb54ea91d724" containerName="registry-server" Dec 13 07:04:53 crc kubenswrapper[4971]: I1213 07:04:53.981114 4971 memory_manager.go:354] "RemoveStaleState removing state" podUID="a1c43378-fa47-4eb5-8c02-f5b229197543" containerName="registry-server" Dec 13 07:04:53 crc kubenswrapper[4971]: I1213 07:04:53.981123 4971 memory_manager.go:354] "RemoveStaleState removing state" podUID="ad20cd8d-6bce-4bb4-b070-79c421d8538b" containerName="registry-server" Dec 13 07:04:53 crc kubenswrapper[4971]: I1213 07:04:53.981809 4971 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/98085b0df3808ebec39f9f9529f737144fe2dbcdaa4f334014817c0fa89s2m9" Dec 13 07:04:53 crc kubenswrapper[4971]: I1213 07:04:53.984278 4971 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"default-dockercfg-vmwhc" Dec 13 07:04:53 crc kubenswrapper[4971]: I1213 07:04:53.989181 4971 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/98085b0df3808ebec39f9f9529f737144fe2dbcdaa4f334014817c0fa89s2m9"] Dec 13 07:04:54 crc kubenswrapper[4971]: I1213 07:04:54.127063 4971 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/c94817e4-313a-4e9a-a0d9-becb8044d4a8-util\") pod \"98085b0df3808ebec39f9f9529f737144fe2dbcdaa4f334014817c0fa89s2m9\" (UID: \"c94817e4-313a-4e9a-a0d9-becb8044d4a8\") " pod="openshift-marketplace/98085b0df3808ebec39f9f9529f737144fe2dbcdaa4f334014817c0fa89s2m9" Dec 13 07:04:54 crc kubenswrapper[4971]: I1213 07:04:54.127123 4971 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/c94817e4-313a-4e9a-a0d9-becb8044d4a8-bundle\") pod \"98085b0df3808ebec39f9f9529f737144fe2dbcdaa4f334014817c0fa89s2m9\" (UID: \"c94817e4-313a-4e9a-a0d9-becb8044d4a8\") " pod="openshift-marketplace/98085b0df3808ebec39f9f9529f737144fe2dbcdaa4f334014817c0fa89s2m9" Dec 13 07:04:54 crc kubenswrapper[4971]: I1213 07:04:54.127380 4971 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-frrc2\" (UniqueName: \"kubernetes.io/projected/c94817e4-313a-4e9a-a0d9-becb8044d4a8-kube-api-access-frrc2\") pod \"98085b0df3808ebec39f9f9529f737144fe2dbcdaa4f334014817c0fa89s2m9\" (UID: \"c94817e4-313a-4e9a-a0d9-becb8044d4a8\") " pod="openshift-marketplace/98085b0df3808ebec39f9f9529f737144fe2dbcdaa4f334014817c0fa89s2m9" Dec 13 07:04:54 crc kubenswrapper[4971]: I1213 07:04:54.228688 4971 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-frrc2\" (UniqueName: \"kubernetes.io/projected/c94817e4-313a-4e9a-a0d9-becb8044d4a8-kube-api-access-frrc2\") pod \"98085b0df3808ebec39f9f9529f737144fe2dbcdaa4f334014817c0fa89s2m9\" (UID: \"c94817e4-313a-4e9a-a0d9-becb8044d4a8\") " pod="openshift-marketplace/98085b0df3808ebec39f9f9529f737144fe2dbcdaa4f334014817c0fa89s2m9" Dec 13 07:04:54 crc kubenswrapper[4971]: I1213 07:04:54.229132 4971 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/c94817e4-313a-4e9a-a0d9-becb8044d4a8-util\") pod \"98085b0df3808ebec39f9f9529f737144fe2dbcdaa4f334014817c0fa89s2m9\" (UID: \"c94817e4-313a-4e9a-a0d9-becb8044d4a8\") " pod="openshift-marketplace/98085b0df3808ebec39f9f9529f737144fe2dbcdaa4f334014817c0fa89s2m9" Dec 13 07:04:54 crc kubenswrapper[4971]: I1213 07:04:54.229166 4971 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/c94817e4-313a-4e9a-a0d9-becb8044d4a8-bundle\") pod \"98085b0df3808ebec39f9f9529f737144fe2dbcdaa4f334014817c0fa89s2m9\" (UID: \"c94817e4-313a-4e9a-a0d9-becb8044d4a8\") " pod="openshift-marketplace/98085b0df3808ebec39f9f9529f737144fe2dbcdaa4f334014817c0fa89s2m9" Dec 13 07:04:54 crc kubenswrapper[4971]: I1213 07:04:54.229727 4971 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/c94817e4-313a-4e9a-a0d9-becb8044d4a8-util\") pod \"98085b0df3808ebec39f9f9529f737144fe2dbcdaa4f334014817c0fa89s2m9\" (UID: \"c94817e4-313a-4e9a-a0d9-becb8044d4a8\") " pod="openshift-marketplace/98085b0df3808ebec39f9f9529f737144fe2dbcdaa4f334014817c0fa89s2m9" Dec 13 07:04:54 crc kubenswrapper[4971]: I1213 07:04:54.230263 4971 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/c94817e4-313a-4e9a-a0d9-becb8044d4a8-bundle\") pod \"98085b0df3808ebec39f9f9529f737144fe2dbcdaa4f334014817c0fa89s2m9\" (UID: \"c94817e4-313a-4e9a-a0d9-becb8044d4a8\") " pod="openshift-marketplace/98085b0df3808ebec39f9f9529f737144fe2dbcdaa4f334014817c0fa89s2m9" Dec 13 07:04:54 crc kubenswrapper[4971]: I1213 07:04:54.253035 4971 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-frrc2\" (UniqueName: \"kubernetes.io/projected/c94817e4-313a-4e9a-a0d9-becb8044d4a8-kube-api-access-frrc2\") pod \"98085b0df3808ebec39f9f9529f737144fe2dbcdaa4f334014817c0fa89s2m9\" (UID: \"c94817e4-313a-4e9a-a0d9-becb8044d4a8\") " pod="openshift-marketplace/98085b0df3808ebec39f9f9529f737144fe2dbcdaa4f334014817c0fa89s2m9" Dec 13 07:04:54 crc kubenswrapper[4971]: I1213 07:04:54.302881 4971 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/98085b0df3808ebec39f9f9529f737144fe2dbcdaa4f334014817c0fa89s2m9" Dec 13 07:04:54 crc kubenswrapper[4971]: I1213 07:04:54.711912 4971 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/98085b0df3808ebec39f9f9529f737144fe2dbcdaa4f334014817c0fa89s2m9"] Dec 13 07:04:54 crc kubenswrapper[4971]: I1213 07:04:54.786948 4971 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/98085b0df3808ebec39f9f9529f737144fe2dbcdaa4f334014817c0fa89s2m9" event={"ID":"c94817e4-313a-4e9a-a0d9-becb8044d4a8","Type":"ContainerStarted","Data":"82df9f27f078d26fc4768246d19243d7961d5c29eba81ff8fd0b78a43ba544f9"} Dec 13 07:04:55 crc kubenswrapper[4971]: I1213 07:04:55.795506 4971 generic.go:334] "Generic (PLEG): container finished" podID="c94817e4-313a-4e9a-a0d9-becb8044d4a8" containerID="65d0308d92796fef07803c85e3847dbd54c9a58382f91904f2acecc37e8f83b3" exitCode=0 Dec 13 07:04:55 crc kubenswrapper[4971]: I1213 07:04:55.795739 4971 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/98085b0df3808ebec39f9f9529f737144fe2dbcdaa4f334014817c0fa89s2m9" event={"ID":"c94817e4-313a-4e9a-a0d9-becb8044d4a8","Type":"ContainerDied","Data":"65d0308d92796fef07803c85e3847dbd54c9a58382f91904f2acecc37e8f83b3"} Dec 13 07:04:57 crc kubenswrapper[4971]: I1213 07:04:57.807666 4971 generic.go:334] "Generic (PLEG): container finished" podID="c94817e4-313a-4e9a-a0d9-becb8044d4a8" containerID="94b0ba2a4249ea5d6568856249d828be608239a2cff9c951fd1047c75b67d228" exitCode=0 Dec 13 07:04:57 crc kubenswrapper[4971]: I1213 07:04:57.807727 4971 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/98085b0df3808ebec39f9f9529f737144fe2dbcdaa4f334014817c0fa89s2m9" event={"ID":"c94817e4-313a-4e9a-a0d9-becb8044d4a8","Type":"ContainerDied","Data":"94b0ba2a4249ea5d6568856249d828be608239a2cff9c951fd1047c75b67d228"} Dec 13 07:04:58 crc kubenswrapper[4971]: I1213 07:04:58.815445 4971 generic.go:334] "Generic (PLEG): container finished" podID="c94817e4-313a-4e9a-a0d9-becb8044d4a8" containerID="a68a153ad016406bb55554bf042cff34e9d6aa1e7bb163da644d890a5c60c775" exitCode=0 Dec 13 07:04:58 crc kubenswrapper[4971]: I1213 07:04:58.815490 4971 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/98085b0df3808ebec39f9f9529f737144fe2dbcdaa4f334014817c0fa89s2m9" event={"ID":"c94817e4-313a-4e9a-a0d9-becb8044d4a8","Type":"ContainerDied","Data":"a68a153ad016406bb55554bf042cff34e9d6aa1e7bb163da644d890a5c60c775"} Dec 13 07:05:00 crc kubenswrapper[4971]: I1213 07:05:00.042007 4971 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/98085b0df3808ebec39f9f9529f737144fe2dbcdaa4f334014817c0fa89s2m9" Dec 13 07:05:00 crc kubenswrapper[4971]: I1213 07:05:00.204012 4971 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-frrc2\" (UniqueName: \"kubernetes.io/projected/c94817e4-313a-4e9a-a0d9-becb8044d4a8-kube-api-access-frrc2\") pod \"c94817e4-313a-4e9a-a0d9-becb8044d4a8\" (UID: \"c94817e4-313a-4e9a-a0d9-becb8044d4a8\") " Dec 13 07:05:00 crc kubenswrapper[4971]: I1213 07:05:00.204185 4971 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/c94817e4-313a-4e9a-a0d9-becb8044d4a8-util\") pod \"c94817e4-313a-4e9a-a0d9-becb8044d4a8\" (UID: \"c94817e4-313a-4e9a-a0d9-becb8044d4a8\") " Dec 13 07:05:00 crc kubenswrapper[4971]: I1213 07:05:00.204222 4971 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/c94817e4-313a-4e9a-a0d9-becb8044d4a8-bundle\") pod \"c94817e4-313a-4e9a-a0d9-becb8044d4a8\" (UID: \"c94817e4-313a-4e9a-a0d9-becb8044d4a8\") " Dec 13 07:05:00 crc kubenswrapper[4971]: I1213 07:05:00.204983 4971 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/c94817e4-313a-4e9a-a0d9-becb8044d4a8-bundle" (OuterVolumeSpecName: "bundle") pod "c94817e4-313a-4e9a-a0d9-becb8044d4a8" (UID: "c94817e4-313a-4e9a-a0d9-becb8044d4a8"). InnerVolumeSpecName "bundle". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 13 07:05:00 crc kubenswrapper[4971]: I1213 07:05:00.217361 4971 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/c94817e4-313a-4e9a-a0d9-becb8044d4a8-kube-api-access-frrc2" (OuterVolumeSpecName: "kube-api-access-frrc2") pod "c94817e4-313a-4e9a-a0d9-becb8044d4a8" (UID: "c94817e4-313a-4e9a-a0d9-becb8044d4a8"). InnerVolumeSpecName "kube-api-access-frrc2". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 13 07:05:00 crc kubenswrapper[4971]: I1213 07:05:00.220923 4971 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/c94817e4-313a-4e9a-a0d9-becb8044d4a8-util" (OuterVolumeSpecName: "util") pod "c94817e4-313a-4e9a-a0d9-becb8044d4a8" (UID: "c94817e4-313a-4e9a-a0d9-becb8044d4a8"). InnerVolumeSpecName "util". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 13 07:05:00 crc kubenswrapper[4971]: I1213 07:05:00.306104 4971 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-frrc2\" (UniqueName: \"kubernetes.io/projected/c94817e4-313a-4e9a-a0d9-becb8044d4a8-kube-api-access-frrc2\") on node \"crc\" DevicePath \"\"" Dec 13 07:05:00 crc kubenswrapper[4971]: I1213 07:05:00.306141 4971 reconciler_common.go:293] "Volume detached for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/c94817e4-313a-4e9a-a0d9-becb8044d4a8-util\") on node \"crc\" DevicePath \"\"" Dec 13 07:05:00 crc kubenswrapper[4971]: I1213 07:05:00.306151 4971 reconciler_common.go:293] "Volume detached for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/c94817e4-313a-4e9a-a0d9-becb8044d4a8-bundle\") on node \"crc\" DevicePath \"\"" Dec 13 07:05:00 crc kubenswrapper[4971]: I1213 07:05:00.828381 4971 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/98085b0df3808ebec39f9f9529f737144fe2dbcdaa4f334014817c0fa89s2m9" event={"ID":"c94817e4-313a-4e9a-a0d9-becb8044d4a8","Type":"ContainerDied","Data":"82df9f27f078d26fc4768246d19243d7961d5c29eba81ff8fd0b78a43ba544f9"} Dec 13 07:05:00 crc kubenswrapper[4971]: I1213 07:05:00.828433 4971 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="82df9f27f078d26fc4768246d19243d7961d5c29eba81ff8fd0b78a43ba544f9" Dec 13 07:05:00 crc kubenswrapper[4971]: I1213 07:05:00.828513 4971 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/98085b0df3808ebec39f9f9529f737144fe2dbcdaa4f334014817c0fa89s2m9" Dec 13 07:05:01 crc kubenswrapper[4971]: I1213 07:05:01.843455 4971 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-nmstate/nmstate-operator-6769fb99d-n9v79"] Dec 13 07:05:01 crc kubenswrapper[4971]: E1213 07:05:01.843681 4971 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="c94817e4-313a-4e9a-a0d9-becb8044d4a8" containerName="util" Dec 13 07:05:01 crc kubenswrapper[4971]: I1213 07:05:01.843692 4971 state_mem.go:107] "Deleted CPUSet assignment" podUID="c94817e4-313a-4e9a-a0d9-becb8044d4a8" containerName="util" Dec 13 07:05:01 crc kubenswrapper[4971]: E1213 07:05:01.843714 4971 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="c94817e4-313a-4e9a-a0d9-becb8044d4a8" containerName="extract" Dec 13 07:05:01 crc kubenswrapper[4971]: I1213 07:05:01.843720 4971 state_mem.go:107] "Deleted CPUSet assignment" podUID="c94817e4-313a-4e9a-a0d9-becb8044d4a8" containerName="extract" Dec 13 07:05:01 crc kubenswrapper[4971]: E1213 07:05:01.843727 4971 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="c94817e4-313a-4e9a-a0d9-becb8044d4a8" containerName="pull" Dec 13 07:05:01 crc kubenswrapper[4971]: I1213 07:05:01.843734 4971 state_mem.go:107] "Deleted CPUSet assignment" podUID="c94817e4-313a-4e9a-a0d9-becb8044d4a8" containerName="pull" Dec 13 07:05:01 crc kubenswrapper[4971]: I1213 07:05:01.843848 4971 memory_manager.go:354] "RemoveStaleState removing state" podUID="c94817e4-313a-4e9a-a0d9-becb8044d4a8" containerName="extract" Dec 13 07:05:01 crc kubenswrapper[4971]: I1213 07:05:01.844248 4971 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-nmstate/nmstate-operator-6769fb99d-n9v79" Dec 13 07:05:01 crc kubenswrapper[4971]: I1213 07:05:01.846136 4971 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-nmstate"/"kube-root-ca.crt" Dec 13 07:05:01 crc kubenswrapper[4971]: I1213 07:05:01.846323 4971 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-nmstate"/"nmstate-operator-dockercfg-z4fxq" Dec 13 07:05:01 crc kubenswrapper[4971]: I1213 07:05:01.852330 4971 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-nmstate"/"openshift-service-ca.crt" Dec 13 07:05:01 crc kubenswrapper[4971]: I1213 07:05:01.858588 4971 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-nmstate/nmstate-operator-6769fb99d-n9v79"] Dec 13 07:05:01 crc kubenswrapper[4971]: I1213 07:05:01.924266 4971 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-vgzzc\" (UniqueName: \"kubernetes.io/projected/658a55da-1b10-45a1-b449-d8d4b8916242-kube-api-access-vgzzc\") pod \"nmstate-operator-6769fb99d-n9v79\" (UID: \"658a55da-1b10-45a1-b449-d8d4b8916242\") " pod="openshift-nmstate/nmstate-operator-6769fb99d-n9v79" Dec 13 07:05:02 crc kubenswrapper[4971]: I1213 07:05:02.025467 4971 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-vgzzc\" (UniqueName: \"kubernetes.io/projected/658a55da-1b10-45a1-b449-d8d4b8916242-kube-api-access-vgzzc\") pod \"nmstate-operator-6769fb99d-n9v79\" (UID: \"658a55da-1b10-45a1-b449-d8d4b8916242\") " pod="openshift-nmstate/nmstate-operator-6769fb99d-n9v79" Dec 13 07:05:02 crc kubenswrapper[4971]: I1213 07:05:02.043909 4971 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-vgzzc\" (UniqueName: \"kubernetes.io/projected/658a55da-1b10-45a1-b449-d8d4b8916242-kube-api-access-vgzzc\") pod \"nmstate-operator-6769fb99d-n9v79\" (UID: \"658a55da-1b10-45a1-b449-d8d4b8916242\") " pod="openshift-nmstate/nmstate-operator-6769fb99d-n9v79" Dec 13 07:05:02 crc kubenswrapper[4971]: I1213 07:05:02.161925 4971 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-nmstate/nmstate-operator-6769fb99d-n9v79" Dec 13 07:05:02 crc kubenswrapper[4971]: I1213 07:05:02.363599 4971 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-nmstate/nmstate-operator-6769fb99d-n9v79"] Dec 13 07:05:02 crc kubenswrapper[4971]: I1213 07:05:02.841305 4971 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-nmstate/nmstate-operator-6769fb99d-n9v79" event={"ID":"658a55da-1b10-45a1-b449-d8d4b8916242","Type":"ContainerStarted","Data":"5f10af10156e1e619225074a94fbeacf1281830e5509a8e379ad8ebe167be7cc"} Dec 13 07:05:05 crc kubenswrapper[4971]: I1213 07:05:05.867470 4971 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-nmstate/nmstate-operator-6769fb99d-n9v79" event={"ID":"658a55da-1b10-45a1-b449-d8d4b8916242","Type":"ContainerStarted","Data":"2ed9dd488507b3131579511f782793662b539690d71692a13f3de60d5baa0170"} Dec 13 07:05:05 crc kubenswrapper[4971]: I1213 07:05:05.893270 4971 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-nmstate/nmstate-operator-6769fb99d-n9v79" podStartSLOduration=1.5862425629999999 podStartE2EDuration="4.893245865s" podCreationTimestamp="2025-12-13 07:05:01 +0000 UTC" firstStartedPulling="2025-12-13 07:05:02.363305819 +0000 UTC m=+958.967715267" lastFinishedPulling="2025-12-13 07:05:05.670309121 +0000 UTC m=+962.274718569" observedRunningTime="2025-12-13 07:05:05.888927168 +0000 UTC m=+962.493336626" watchObservedRunningTime="2025-12-13 07:05:05.893245865 +0000 UTC m=+962.497655313" Dec 13 07:05:07 crc kubenswrapper[4971]: I1213 07:05:07.563297 4971 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-nmstate/nmstate-metrics-7f7f7578db-xsb6k"] Dec 13 07:05:07 crc kubenswrapper[4971]: I1213 07:05:07.564834 4971 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-nmstate/nmstate-metrics-7f7f7578db-xsb6k" Dec 13 07:05:07 crc kubenswrapper[4971]: I1213 07:05:07.570608 4971 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-nmstate"/"nmstate-handler-dockercfg-sr7nl" Dec 13 07:05:07 crc kubenswrapper[4971]: I1213 07:05:07.579247 4971 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-nmstate/nmstate-metrics-7f7f7578db-xsb6k"] Dec 13 07:05:07 crc kubenswrapper[4971]: I1213 07:05:07.583445 4971 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-nmstate/nmstate-webhook-f8fb84555-pwxvn"] Dec 13 07:05:07 crc kubenswrapper[4971]: I1213 07:05:07.585378 4971 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-nmstate/nmstate-webhook-f8fb84555-pwxvn" Dec 13 07:05:07 crc kubenswrapper[4971]: I1213 07:05:07.588029 4971 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-nmstate"/"openshift-nmstate-webhook" Dec 13 07:05:07 crc kubenswrapper[4971]: I1213 07:05:07.621639 4971 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-nmstate/nmstate-webhook-f8fb84555-pwxvn"] Dec 13 07:05:07 crc kubenswrapper[4971]: I1213 07:05:07.635505 4971 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-nmstate/nmstate-handler-5b4bx"] Dec 13 07:05:07 crc kubenswrapper[4971]: I1213 07:05:07.636328 4971 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-nmstate/nmstate-handler-5b4bx" Dec 13 07:05:07 crc kubenswrapper[4971]: I1213 07:05:07.700239 4971 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"tls-key-pair\" (UniqueName: \"kubernetes.io/secret/9e887c99-7f51-4b00-9571-f4d29c191054-tls-key-pair\") pod \"nmstate-webhook-f8fb84555-pwxvn\" (UID: \"9e887c99-7f51-4b00-9571-f4d29c191054\") " pod="openshift-nmstate/nmstate-webhook-f8fb84555-pwxvn" Dec 13 07:05:07 crc kubenswrapper[4971]: I1213 07:05:07.700292 4971 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-xdg8b\" (UniqueName: \"kubernetes.io/projected/9e887c99-7f51-4b00-9571-f4d29c191054-kube-api-access-xdg8b\") pod \"nmstate-webhook-f8fb84555-pwxvn\" (UID: \"9e887c99-7f51-4b00-9571-f4d29c191054\") " pod="openshift-nmstate/nmstate-webhook-f8fb84555-pwxvn" Dec 13 07:05:07 crc kubenswrapper[4971]: I1213 07:05:07.700455 4971 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-g4m8t\" (UniqueName: \"kubernetes.io/projected/f2968080-99c8-4320-9b39-4bcab4dad759-kube-api-access-g4m8t\") pod \"nmstate-metrics-7f7f7578db-xsb6k\" (UID: \"f2968080-99c8-4320-9b39-4bcab4dad759\") " pod="openshift-nmstate/nmstate-metrics-7f7f7578db-xsb6k" Dec 13 07:05:07 crc kubenswrapper[4971]: I1213 07:05:07.723572 4971 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-nmstate/nmstate-console-plugin-6ff7998486-hx58s"] Dec 13 07:05:07 crc kubenswrapper[4971]: I1213 07:05:07.724405 4971 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-nmstate/nmstate-console-plugin-6ff7998486-hx58s" Dec 13 07:05:07 crc kubenswrapper[4971]: I1213 07:05:07.726232 4971 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-nmstate"/"plugin-serving-cert" Dec 13 07:05:07 crc kubenswrapper[4971]: I1213 07:05:07.726369 4971 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-nmstate"/"default-dockercfg-vtsn4" Dec 13 07:05:07 crc kubenswrapper[4971]: I1213 07:05:07.726431 4971 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-nmstate"/"nginx-conf" Dec 13 07:05:07 crc kubenswrapper[4971]: I1213 07:05:07.776618 4971 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-nmstate/nmstate-console-plugin-6ff7998486-hx58s"] Dec 13 07:05:07 crc kubenswrapper[4971]: I1213 07:05:07.801293 4971 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-xdg8b\" (UniqueName: \"kubernetes.io/projected/9e887c99-7f51-4b00-9571-f4d29c191054-kube-api-access-xdg8b\") pod \"nmstate-webhook-f8fb84555-pwxvn\" (UID: \"9e887c99-7f51-4b00-9571-f4d29c191054\") " pod="openshift-nmstate/nmstate-webhook-f8fb84555-pwxvn" Dec 13 07:05:07 crc kubenswrapper[4971]: I1213 07:05:07.801354 4971 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-g4m8t\" (UniqueName: \"kubernetes.io/projected/f2968080-99c8-4320-9b39-4bcab4dad759-kube-api-access-g4m8t\") pod \"nmstate-metrics-7f7f7578db-xsb6k\" (UID: \"f2968080-99c8-4320-9b39-4bcab4dad759\") " pod="openshift-nmstate/nmstate-metrics-7f7f7578db-xsb6k" Dec 13 07:05:07 crc kubenswrapper[4971]: I1213 07:05:07.801397 4971 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dbus-socket\" (UniqueName: \"kubernetes.io/host-path/d8139542-2862-4e1b-8e84-62a77e5bb7e5-dbus-socket\") pod \"nmstate-handler-5b4bx\" (UID: \"d8139542-2862-4e1b-8e84-62a77e5bb7e5\") " pod="openshift-nmstate/nmstate-handler-5b4bx" Dec 13 07:05:07 crc kubenswrapper[4971]: I1213 07:05:07.801448 4971 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovs-socket\" (UniqueName: \"kubernetes.io/host-path/d8139542-2862-4e1b-8e84-62a77e5bb7e5-ovs-socket\") pod \"nmstate-handler-5b4bx\" (UID: \"d8139542-2862-4e1b-8e84-62a77e5bb7e5\") " pod="openshift-nmstate/nmstate-handler-5b4bx" Dec 13 07:05:07 crc kubenswrapper[4971]: I1213 07:05:07.801487 4971 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-2q5fp\" (UniqueName: \"kubernetes.io/projected/d8139542-2862-4e1b-8e84-62a77e5bb7e5-kube-api-access-2q5fp\") pod \"nmstate-handler-5b4bx\" (UID: \"d8139542-2862-4e1b-8e84-62a77e5bb7e5\") " pod="openshift-nmstate/nmstate-handler-5b4bx" Dec 13 07:05:07 crc kubenswrapper[4971]: I1213 07:05:07.801548 4971 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"nmstate-lock\" (UniqueName: \"kubernetes.io/host-path/d8139542-2862-4e1b-8e84-62a77e5bb7e5-nmstate-lock\") pod \"nmstate-handler-5b4bx\" (UID: \"d8139542-2862-4e1b-8e84-62a77e5bb7e5\") " pod="openshift-nmstate/nmstate-handler-5b4bx" Dec 13 07:05:07 crc kubenswrapper[4971]: I1213 07:05:07.801597 4971 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"tls-key-pair\" (UniqueName: \"kubernetes.io/secret/9e887c99-7f51-4b00-9571-f4d29c191054-tls-key-pair\") pod \"nmstate-webhook-f8fb84555-pwxvn\" (UID: \"9e887c99-7f51-4b00-9571-f4d29c191054\") " pod="openshift-nmstate/nmstate-webhook-f8fb84555-pwxvn" Dec 13 07:05:07 crc kubenswrapper[4971]: E1213 07:05:07.801732 4971 secret.go:188] Couldn't get secret openshift-nmstate/openshift-nmstate-webhook: secret "openshift-nmstate-webhook" not found Dec 13 07:05:07 crc kubenswrapper[4971]: E1213 07:05:07.801798 4971 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/9e887c99-7f51-4b00-9571-f4d29c191054-tls-key-pair podName:9e887c99-7f51-4b00-9571-f4d29c191054 nodeName:}" failed. No retries permitted until 2025-12-13 07:05:08.301775611 +0000 UTC m=+964.906185059 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "tls-key-pair" (UniqueName: "kubernetes.io/secret/9e887c99-7f51-4b00-9571-f4d29c191054-tls-key-pair") pod "nmstate-webhook-f8fb84555-pwxvn" (UID: "9e887c99-7f51-4b00-9571-f4d29c191054") : secret "openshift-nmstate-webhook" not found Dec 13 07:05:07 crc kubenswrapper[4971]: I1213 07:05:07.823186 4971 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-g4m8t\" (UniqueName: \"kubernetes.io/projected/f2968080-99c8-4320-9b39-4bcab4dad759-kube-api-access-g4m8t\") pod \"nmstate-metrics-7f7f7578db-xsb6k\" (UID: \"f2968080-99c8-4320-9b39-4bcab4dad759\") " pod="openshift-nmstate/nmstate-metrics-7f7f7578db-xsb6k" Dec 13 07:05:07 crc kubenswrapper[4971]: I1213 07:05:07.823198 4971 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-xdg8b\" (UniqueName: \"kubernetes.io/projected/9e887c99-7f51-4b00-9571-f4d29c191054-kube-api-access-xdg8b\") pod \"nmstate-webhook-f8fb84555-pwxvn\" (UID: \"9e887c99-7f51-4b00-9571-f4d29c191054\") " pod="openshift-nmstate/nmstate-webhook-f8fb84555-pwxvn" Dec 13 07:05:07 crc kubenswrapper[4971]: I1213 07:05:07.885261 4971 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-nmstate/nmstate-metrics-7f7f7578db-xsb6k" Dec 13 07:05:07 crc kubenswrapper[4971]: I1213 07:05:07.902802 4971 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dbus-socket\" (UniqueName: \"kubernetes.io/host-path/d8139542-2862-4e1b-8e84-62a77e5bb7e5-dbus-socket\") pod \"nmstate-handler-5b4bx\" (UID: \"d8139542-2862-4e1b-8e84-62a77e5bb7e5\") " pod="openshift-nmstate/nmstate-handler-5b4bx" Dec 13 07:05:07 crc kubenswrapper[4971]: I1213 07:05:07.902877 4971 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"plugin-serving-cert\" (UniqueName: \"kubernetes.io/secret/dfa2c127-31ee-40a0-814a-5af53e135ae8-plugin-serving-cert\") pod \"nmstate-console-plugin-6ff7998486-hx58s\" (UID: \"dfa2c127-31ee-40a0-814a-5af53e135ae8\") " pod="openshift-nmstate/nmstate-console-plugin-6ff7998486-hx58s" Dec 13 07:05:07 crc kubenswrapper[4971]: I1213 07:05:07.902924 4971 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovs-socket\" (UniqueName: \"kubernetes.io/host-path/d8139542-2862-4e1b-8e84-62a77e5bb7e5-ovs-socket\") pod \"nmstate-handler-5b4bx\" (UID: \"d8139542-2862-4e1b-8e84-62a77e5bb7e5\") " pod="openshift-nmstate/nmstate-handler-5b4bx" Dec 13 07:05:07 crc kubenswrapper[4971]: I1213 07:05:07.902973 4971 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-2q5fp\" (UniqueName: \"kubernetes.io/projected/d8139542-2862-4e1b-8e84-62a77e5bb7e5-kube-api-access-2q5fp\") pod \"nmstate-handler-5b4bx\" (UID: \"d8139542-2862-4e1b-8e84-62a77e5bb7e5\") " pod="openshift-nmstate/nmstate-handler-5b4bx" Dec 13 07:05:07 crc kubenswrapper[4971]: I1213 07:05:07.903020 4971 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-cflx4\" (UniqueName: \"kubernetes.io/projected/dfa2c127-31ee-40a0-814a-5af53e135ae8-kube-api-access-cflx4\") pod \"nmstate-console-plugin-6ff7998486-hx58s\" (UID: \"dfa2c127-31ee-40a0-814a-5af53e135ae8\") " pod="openshift-nmstate/nmstate-console-plugin-6ff7998486-hx58s" Dec 13 07:05:07 crc kubenswrapper[4971]: I1213 07:05:07.903049 4971 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"nginx-conf\" (UniqueName: \"kubernetes.io/configmap/dfa2c127-31ee-40a0-814a-5af53e135ae8-nginx-conf\") pod \"nmstate-console-plugin-6ff7998486-hx58s\" (UID: \"dfa2c127-31ee-40a0-814a-5af53e135ae8\") " pod="openshift-nmstate/nmstate-console-plugin-6ff7998486-hx58s" Dec 13 07:05:07 crc kubenswrapper[4971]: I1213 07:05:07.903078 4971 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nmstate-lock\" (UniqueName: \"kubernetes.io/host-path/d8139542-2862-4e1b-8e84-62a77e5bb7e5-nmstate-lock\") pod \"nmstate-handler-5b4bx\" (UID: \"d8139542-2862-4e1b-8e84-62a77e5bb7e5\") " pod="openshift-nmstate/nmstate-handler-5b4bx" Dec 13 07:05:07 crc kubenswrapper[4971]: I1213 07:05:07.903124 4971 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dbus-socket\" (UniqueName: \"kubernetes.io/host-path/d8139542-2862-4e1b-8e84-62a77e5bb7e5-dbus-socket\") pod \"nmstate-handler-5b4bx\" (UID: \"d8139542-2862-4e1b-8e84-62a77e5bb7e5\") " pod="openshift-nmstate/nmstate-handler-5b4bx" Dec 13 07:05:07 crc kubenswrapper[4971]: I1213 07:05:07.903168 4971 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"nmstate-lock\" (UniqueName: \"kubernetes.io/host-path/d8139542-2862-4e1b-8e84-62a77e5bb7e5-nmstate-lock\") pod \"nmstate-handler-5b4bx\" (UID: \"d8139542-2862-4e1b-8e84-62a77e5bb7e5\") " pod="openshift-nmstate/nmstate-handler-5b4bx" Dec 13 07:05:07 crc kubenswrapper[4971]: I1213 07:05:07.903398 4971 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovs-socket\" (UniqueName: \"kubernetes.io/host-path/d8139542-2862-4e1b-8e84-62a77e5bb7e5-ovs-socket\") pod \"nmstate-handler-5b4bx\" (UID: \"d8139542-2862-4e1b-8e84-62a77e5bb7e5\") " pod="openshift-nmstate/nmstate-handler-5b4bx" Dec 13 07:05:07 crc kubenswrapper[4971]: I1213 07:05:07.926772 4971 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-2q5fp\" (UniqueName: \"kubernetes.io/projected/d8139542-2862-4e1b-8e84-62a77e5bb7e5-kube-api-access-2q5fp\") pod \"nmstate-handler-5b4bx\" (UID: \"d8139542-2862-4e1b-8e84-62a77e5bb7e5\") " pod="openshift-nmstate/nmstate-handler-5b4bx" Dec 13 07:05:07 crc kubenswrapper[4971]: I1213 07:05:07.953725 4971 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-nmstate/nmstate-handler-5b4bx" Dec 13 07:05:07 crc kubenswrapper[4971]: I1213 07:05:07.953784 4971 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-console/console-79476f9bb4-9ngf9"] Dec 13 07:05:07 crc kubenswrapper[4971]: I1213 07:05:07.954665 4971 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-console/console-79476f9bb4-9ngf9" Dec 13 07:05:07 crc kubenswrapper[4971]: I1213 07:05:07.972314 4971 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-console/console-79476f9bb4-9ngf9"] Dec 13 07:05:08 crc kubenswrapper[4971]: I1213 07:05:08.004557 4971 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cflx4\" (UniqueName: \"kubernetes.io/projected/dfa2c127-31ee-40a0-814a-5af53e135ae8-kube-api-access-cflx4\") pod \"nmstate-console-plugin-6ff7998486-hx58s\" (UID: \"dfa2c127-31ee-40a0-814a-5af53e135ae8\") " pod="openshift-nmstate/nmstate-console-plugin-6ff7998486-hx58s" Dec 13 07:05:08 crc kubenswrapper[4971]: I1213 07:05:08.004617 4971 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nginx-conf\" (UniqueName: \"kubernetes.io/configmap/dfa2c127-31ee-40a0-814a-5af53e135ae8-nginx-conf\") pod \"nmstate-console-plugin-6ff7998486-hx58s\" (UID: \"dfa2c127-31ee-40a0-814a-5af53e135ae8\") " pod="openshift-nmstate/nmstate-console-plugin-6ff7998486-hx58s" Dec 13 07:05:08 crc kubenswrapper[4971]: I1213 07:05:08.004708 4971 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"plugin-serving-cert\" (UniqueName: \"kubernetes.io/secret/dfa2c127-31ee-40a0-814a-5af53e135ae8-plugin-serving-cert\") pod \"nmstate-console-plugin-6ff7998486-hx58s\" (UID: \"dfa2c127-31ee-40a0-814a-5af53e135ae8\") " pod="openshift-nmstate/nmstate-console-plugin-6ff7998486-hx58s" Dec 13 07:05:08 crc kubenswrapper[4971]: I1213 07:05:08.007194 4971 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"nginx-conf\" (UniqueName: \"kubernetes.io/configmap/dfa2c127-31ee-40a0-814a-5af53e135ae8-nginx-conf\") pod \"nmstate-console-plugin-6ff7998486-hx58s\" (UID: \"dfa2c127-31ee-40a0-814a-5af53e135ae8\") " pod="openshift-nmstate/nmstate-console-plugin-6ff7998486-hx58s" Dec 13 07:05:08 crc kubenswrapper[4971]: W1213 07:05:08.008074 4971 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podd8139542_2862_4e1b_8e84_62a77e5bb7e5.slice/crio-2422e50f748df877a26bf5ae5dd7a9815afb617b6462b687892179254e0a2d92 WatchSource:0}: Error finding container 2422e50f748df877a26bf5ae5dd7a9815afb617b6462b687892179254e0a2d92: Status 404 returned error can't find the container with id 2422e50f748df877a26bf5ae5dd7a9815afb617b6462b687892179254e0a2d92 Dec 13 07:05:08 crc kubenswrapper[4971]: I1213 07:05:08.010060 4971 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"plugin-serving-cert\" (UniqueName: \"kubernetes.io/secret/dfa2c127-31ee-40a0-814a-5af53e135ae8-plugin-serving-cert\") pod \"nmstate-console-plugin-6ff7998486-hx58s\" (UID: \"dfa2c127-31ee-40a0-814a-5af53e135ae8\") " pod="openshift-nmstate/nmstate-console-plugin-6ff7998486-hx58s" Dec 13 07:05:08 crc kubenswrapper[4971]: I1213 07:05:08.027607 4971 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-cflx4\" (UniqueName: \"kubernetes.io/projected/dfa2c127-31ee-40a0-814a-5af53e135ae8-kube-api-access-cflx4\") pod \"nmstate-console-plugin-6ff7998486-hx58s\" (UID: \"dfa2c127-31ee-40a0-814a-5af53e135ae8\") " pod="openshift-nmstate/nmstate-console-plugin-6ff7998486-hx58s" Dec 13 07:05:08 crc kubenswrapper[4971]: I1213 07:05:08.041476 4971 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-nmstate/nmstate-console-plugin-6ff7998486-hx58s" Dec 13 07:05:08 crc kubenswrapper[4971]: I1213 07:05:08.106141 4971 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/ea4824e7-f6a5-4db3-9e99-680986683b2f-service-ca\") pod \"console-79476f9bb4-9ngf9\" (UID: \"ea4824e7-f6a5-4db3-9e99-680986683b2f\") " pod="openshift-console/console-79476f9bb4-9ngf9" Dec 13 07:05:08 crc kubenswrapper[4971]: I1213 07:05:08.106324 4971 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/ea4824e7-f6a5-4db3-9e99-680986683b2f-trusted-ca-bundle\") pod \"console-79476f9bb4-9ngf9\" (UID: \"ea4824e7-f6a5-4db3-9e99-680986683b2f\") " pod="openshift-console/console-79476f9bb4-9ngf9" Dec 13 07:05:08 crc kubenswrapper[4971]: I1213 07:05:08.106391 4971 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"console-config\" (UniqueName: \"kubernetes.io/configmap/ea4824e7-f6a5-4db3-9e99-680986683b2f-console-config\") pod \"console-79476f9bb4-9ngf9\" (UID: \"ea4824e7-f6a5-4db3-9e99-680986683b2f\") " pod="openshift-console/console-79476f9bb4-9ngf9" Dec 13 07:05:08 crc kubenswrapper[4971]: I1213 07:05:08.106418 4971 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"oauth-serving-cert\" (UniqueName: \"kubernetes.io/configmap/ea4824e7-f6a5-4db3-9e99-680986683b2f-oauth-serving-cert\") pod \"console-79476f9bb4-9ngf9\" (UID: \"ea4824e7-f6a5-4db3-9e99-680986683b2f\") " pod="openshift-console/console-79476f9bb4-9ngf9" Dec 13 07:05:08 crc kubenswrapper[4971]: I1213 07:05:08.106495 4971 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-97hzv\" (UniqueName: \"kubernetes.io/projected/ea4824e7-f6a5-4db3-9e99-680986683b2f-kube-api-access-97hzv\") pod \"console-79476f9bb4-9ngf9\" (UID: \"ea4824e7-f6a5-4db3-9e99-680986683b2f\") " pod="openshift-console/console-79476f9bb4-9ngf9" Dec 13 07:05:08 crc kubenswrapper[4971]: I1213 07:05:08.106631 4971 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"console-oauth-config\" (UniqueName: \"kubernetes.io/secret/ea4824e7-f6a5-4db3-9e99-680986683b2f-console-oauth-config\") pod \"console-79476f9bb4-9ngf9\" (UID: \"ea4824e7-f6a5-4db3-9e99-680986683b2f\") " pod="openshift-console/console-79476f9bb4-9ngf9" Dec 13 07:05:08 crc kubenswrapper[4971]: I1213 07:05:08.106898 4971 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"console-serving-cert\" (UniqueName: \"kubernetes.io/secret/ea4824e7-f6a5-4db3-9e99-680986683b2f-console-serving-cert\") pod \"console-79476f9bb4-9ngf9\" (UID: \"ea4824e7-f6a5-4db3-9e99-680986683b2f\") " pod="openshift-console/console-79476f9bb4-9ngf9" Dec 13 07:05:08 crc kubenswrapper[4971]: I1213 07:05:08.197174 4971 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-nmstate/nmstate-metrics-7f7f7578db-xsb6k"] Dec 13 07:05:08 crc kubenswrapper[4971]: W1213 07:05:08.205957 4971 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podf2968080_99c8_4320_9b39_4bcab4dad759.slice/crio-6e1a90f8962ce1ace6bce35ee220dfc2f04186814c8f31d2f47809284ec7ea50 WatchSource:0}: Error finding container 6e1a90f8962ce1ace6bce35ee220dfc2f04186814c8f31d2f47809284ec7ea50: Status 404 returned error can't find the container with id 6e1a90f8962ce1ace6bce35ee220dfc2f04186814c8f31d2f47809284ec7ea50 Dec 13 07:05:08 crc kubenswrapper[4971]: I1213 07:05:08.207890 4971 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-97hzv\" (UniqueName: \"kubernetes.io/projected/ea4824e7-f6a5-4db3-9e99-680986683b2f-kube-api-access-97hzv\") pod \"console-79476f9bb4-9ngf9\" (UID: \"ea4824e7-f6a5-4db3-9e99-680986683b2f\") " pod="openshift-console/console-79476f9bb4-9ngf9" Dec 13 07:05:08 crc kubenswrapper[4971]: I1213 07:05:08.207936 4971 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"console-oauth-config\" (UniqueName: \"kubernetes.io/secret/ea4824e7-f6a5-4db3-9e99-680986683b2f-console-oauth-config\") pod \"console-79476f9bb4-9ngf9\" (UID: \"ea4824e7-f6a5-4db3-9e99-680986683b2f\") " pod="openshift-console/console-79476f9bb4-9ngf9" Dec 13 07:05:08 crc kubenswrapper[4971]: I1213 07:05:08.207966 4971 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"console-serving-cert\" (UniqueName: \"kubernetes.io/secret/ea4824e7-f6a5-4db3-9e99-680986683b2f-console-serving-cert\") pod \"console-79476f9bb4-9ngf9\" (UID: \"ea4824e7-f6a5-4db3-9e99-680986683b2f\") " pod="openshift-console/console-79476f9bb4-9ngf9" Dec 13 07:05:08 crc kubenswrapper[4971]: I1213 07:05:08.208069 4971 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/ea4824e7-f6a5-4db3-9e99-680986683b2f-service-ca\") pod \"console-79476f9bb4-9ngf9\" (UID: \"ea4824e7-f6a5-4db3-9e99-680986683b2f\") " pod="openshift-console/console-79476f9bb4-9ngf9" Dec 13 07:05:08 crc kubenswrapper[4971]: I1213 07:05:08.208094 4971 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/ea4824e7-f6a5-4db3-9e99-680986683b2f-trusted-ca-bundle\") pod \"console-79476f9bb4-9ngf9\" (UID: \"ea4824e7-f6a5-4db3-9e99-680986683b2f\") " pod="openshift-console/console-79476f9bb4-9ngf9" Dec 13 07:05:08 crc kubenswrapper[4971]: I1213 07:05:08.208124 4971 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"console-config\" (UniqueName: \"kubernetes.io/configmap/ea4824e7-f6a5-4db3-9e99-680986683b2f-console-config\") pod \"console-79476f9bb4-9ngf9\" (UID: \"ea4824e7-f6a5-4db3-9e99-680986683b2f\") " pod="openshift-console/console-79476f9bb4-9ngf9" Dec 13 07:05:08 crc kubenswrapper[4971]: I1213 07:05:08.208145 4971 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"oauth-serving-cert\" (UniqueName: \"kubernetes.io/configmap/ea4824e7-f6a5-4db3-9e99-680986683b2f-oauth-serving-cert\") pod \"console-79476f9bb4-9ngf9\" (UID: \"ea4824e7-f6a5-4db3-9e99-680986683b2f\") " pod="openshift-console/console-79476f9bb4-9ngf9" Dec 13 07:05:08 crc kubenswrapper[4971]: I1213 07:05:08.209116 4971 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"oauth-serving-cert\" (UniqueName: \"kubernetes.io/configmap/ea4824e7-f6a5-4db3-9e99-680986683b2f-oauth-serving-cert\") pod \"console-79476f9bb4-9ngf9\" (UID: \"ea4824e7-f6a5-4db3-9e99-680986683b2f\") " pod="openshift-console/console-79476f9bb4-9ngf9" Dec 13 07:05:08 crc kubenswrapper[4971]: I1213 07:05:08.209875 4971 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"console-config\" (UniqueName: \"kubernetes.io/configmap/ea4824e7-f6a5-4db3-9e99-680986683b2f-console-config\") pod \"console-79476f9bb4-9ngf9\" (UID: \"ea4824e7-f6a5-4db3-9e99-680986683b2f\") " pod="openshift-console/console-79476f9bb4-9ngf9" Dec 13 07:05:08 crc kubenswrapper[4971]: I1213 07:05:08.211065 4971 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/ea4824e7-f6a5-4db3-9e99-680986683b2f-service-ca\") pod \"console-79476f9bb4-9ngf9\" (UID: \"ea4824e7-f6a5-4db3-9e99-680986683b2f\") " pod="openshift-console/console-79476f9bb4-9ngf9" Dec 13 07:05:08 crc kubenswrapper[4971]: I1213 07:05:08.211577 4971 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/ea4824e7-f6a5-4db3-9e99-680986683b2f-trusted-ca-bundle\") pod \"console-79476f9bb4-9ngf9\" (UID: \"ea4824e7-f6a5-4db3-9e99-680986683b2f\") " pod="openshift-console/console-79476f9bb4-9ngf9" Dec 13 07:05:08 crc kubenswrapper[4971]: I1213 07:05:08.214382 4971 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"console-oauth-config\" (UniqueName: \"kubernetes.io/secret/ea4824e7-f6a5-4db3-9e99-680986683b2f-console-oauth-config\") pod \"console-79476f9bb4-9ngf9\" (UID: \"ea4824e7-f6a5-4db3-9e99-680986683b2f\") " pod="openshift-console/console-79476f9bb4-9ngf9" Dec 13 07:05:08 crc kubenswrapper[4971]: I1213 07:05:08.214382 4971 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"console-serving-cert\" (UniqueName: \"kubernetes.io/secret/ea4824e7-f6a5-4db3-9e99-680986683b2f-console-serving-cert\") pod \"console-79476f9bb4-9ngf9\" (UID: \"ea4824e7-f6a5-4db3-9e99-680986683b2f\") " pod="openshift-console/console-79476f9bb4-9ngf9" Dec 13 07:05:08 crc kubenswrapper[4971]: I1213 07:05:08.228725 4971 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-97hzv\" (UniqueName: \"kubernetes.io/projected/ea4824e7-f6a5-4db3-9e99-680986683b2f-kube-api-access-97hzv\") pod \"console-79476f9bb4-9ngf9\" (UID: \"ea4824e7-f6a5-4db3-9e99-680986683b2f\") " pod="openshift-console/console-79476f9bb4-9ngf9" Dec 13 07:05:08 crc kubenswrapper[4971]: I1213 07:05:08.275347 4971 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-nmstate/nmstate-console-plugin-6ff7998486-hx58s"] Dec 13 07:05:08 crc kubenswrapper[4971]: I1213 07:05:08.275911 4971 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-console/console-79476f9bb4-9ngf9" Dec 13 07:05:08 crc kubenswrapper[4971]: I1213 07:05:08.309055 4971 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"tls-key-pair\" (UniqueName: \"kubernetes.io/secret/9e887c99-7f51-4b00-9571-f4d29c191054-tls-key-pair\") pod \"nmstate-webhook-f8fb84555-pwxvn\" (UID: \"9e887c99-7f51-4b00-9571-f4d29c191054\") " pod="openshift-nmstate/nmstate-webhook-f8fb84555-pwxvn" Dec 13 07:05:08 crc kubenswrapper[4971]: I1213 07:05:08.313702 4971 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"tls-key-pair\" (UniqueName: \"kubernetes.io/secret/9e887c99-7f51-4b00-9571-f4d29c191054-tls-key-pair\") pod \"nmstate-webhook-f8fb84555-pwxvn\" (UID: \"9e887c99-7f51-4b00-9571-f4d29c191054\") " pod="openshift-nmstate/nmstate-webhook-f8fb84555-pwxvn" Dec 13 07:05:08 crc kubenswrapper[4971]: I1213 07:05:08.499742 4971 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-console/console-79476f9bb4-9ngf9"] Dec 13 07:05:08 crc kubenswrapper[4971]: I1213 07:05:08.500791 4971 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-nmstate/nmstate-webhook-f8fb84555-pwxvn" Dec 13 07:05:08 crc kubenswrapper[4971]: I1213 07:05:08.717947 4971 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-nmstate/nmstate-webhook-f8fb84555-pwxvn"] Dec 13 07:05:08 crc kubenswrapper[4971]: W1213 07:05:08.721663 4971 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod9e887c99_7f51_4b00_9571_f4d29c191054.slice/crio-eab82e533314a2dbe1039a3ce6db3a92da3187652c28df0d86dae85007452bd2 WatchSource:0}: Error finding container eab82e533314a2dbe1039a3ce6db3a92da3187652c28df0d86dae85007452bd2: Status 404 returned error can't find the container with id eab82e533314a2dbe1039a3ce6db3a92da3187652c28df0d86dae85007452bd2 Dec 13 07:05:08 crc kubenswrapper[4971]: I1213 07:05:08.887543 4971 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-console/console-79476f9bb4-9ngf9" event={"ID":"ea4824e7-f6a5-4db3-9e99-680986683b2f","Type":"ContainerStarted","Data":"abdd645e34e1b41d504a0b2f9f6aa9eed443841cbd54ef9e9180a0d1a1e2cfaa"} Dec 13 07:05:08 crc kubenswrapper[4971]: I1213 07:05:08.887606 4971 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-console/console-79476f9bb4-9ngf9" event={"ID":"ea4824e7-f6a5-4db3-9e99-680986683b2f","Type":"ContainerStarted","Data":"c54e7b2c1cb636adcb83cb7169911fc85e31f3fd3bc0bcd3dcb9130c5fd0ce59"} Dec 13 07:05:08 crc kubenswrapper[4971]: I1213 07:05:08.888883 4971 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-nmstate/nmstate-webhook-f8fb84555-pwxvn" event={"ID":"9e887c99-7f51-4b00-9571-f4d29c191054","Type":"ContainerStarted","Data":"eab82e533314a2dbe1039a3ce6db3a92da3187652c28df0d86dae85007452bd2"} Dec 13 07:05:08 crc kubenswrapper[4971]: I1213 07:05:08.889959 4971 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-nmstate/nmstate-handler-5b4bx" event={"ID":"d8139542-2862-4e1b-8e84-62a77e5bb7e5","Type":"ContainerStarted","Data":"2422e50f748df877a26bf5ae5dd7a9815afb617b6462b687892179254e0a2d92"} Dec 13 07:05:08 crc kubenswrapper[4971]: I1213 07:05:08.891213 4971 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-nmstate/nmstate-metrics-7f7f7578db-xsb6k" event={"ID":"f2968080-99c8-4320-9b39-4bcab4dad759","Type":"ContainerStarted","Data":"6e1a90f8962ce1ace6bce35ee220dfc2f04186814c8f31d2f47809284ec7ea50"} Dec 13 07:05:08 crc kubenswrapper[4971]: I1213 07:05:08.894008 4971 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-nmstate/nmstate-console-plugin-6ff7998486-hx58s" event={"ID":"dfa2c127-31ee-40a0-814a-5af53e135ae8","Type":"ContainerStarted","Data":"9ffb95867f2969633360bda49516350e00751c7cb0684746b4171badf8313519"} Dec 13 07:05:09 crc kubenswrapper[4971]: I1213 07:05:09.938373 4971 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-console/console-79476f9bb4-9ngf9" podStartSLOduration=2.938351607 podStartE2EDuration="2.938351607s" podCreationTimestamp="2025-12-13 07:05:07 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-13 07:05:09.922340712 +0000 UTC m=+966.526750270" watchObservedRunningTime="2025-12-13 07:05:09.938351607 +0000 UTC m=+966.542761055" Dec 13 07:05:12 crc kubenswrapper[4971]: I1213 07:05:12.924393 4971 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-nmstate/nmstate-handler-5b4bx" event={"ID":"d8139542-2862-4e1b-8e84-62a77e5bb7e5","Type":"ContainerStarted","Data":"08db14e557ca70b3a5a7d06e6c2099b957dc45cb5605d0dfcef643c31b6dab84"} Dec 13 07:05:12 crc kubenswrapper[4971]: I1213 07:05:12.924948 4971 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-nmstate/nmstate-handler-5b4bx" Dec 13 07:05:12 crc kubenswrapper[4971]: I1213 07:05:12.925585 4971 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-nmstate/nmstate-metrics-7f7f7578db-xsb6k" event={"ID":"f2968080-99c8-4320-9b39-4bcab4dad759","Type":"ContainerStarted","Data":"8002ff8634e907a0314130a5b23b64b2134a83ddec73b6161c1592f19b4a0054"} Dec 13 07:05:12 crc kubenswrapper[4971]: I1213 07:05:12.927624 4971 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-nmstate/nmstate-console-plugin-6ff7998486-hx58s" event={"ID":"dfa2c127-31ee-40a0-814a-5af53e135ae8","Type":"ContainerStarted","Data":"d5e2fe3a941eeb5b18b49004c45c255385fa284d201412c6450663cf7014ddcf"} Dec 13 07:05:12 crc kubenswrapper[4971]: I1213 07:05:12.932372 4971 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-nmstate/nmstate-webhook-f8fb84555-pwxvn" event={"ID":"9e887c99-7f51-4b00-9571-f4d29c191054","Type":"ContainerStarted","Data":"d337b364f8e6c46dd87044ebb941aebd3995de5587e8249b608107a3d27c8f04"} Dec 13 07:05:12 crc kubenswrapper[4971]: I1213 07:05:12.933119 4971 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-nmstate/nmstate-webhook-f8fb84555-pwxvn" Dec 13 07:05:12 crc kubenswrapper[4971]: I1213 07:05:12.949838 4971 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-nmstate/nmstate-handler-5b4bx" podStartSLOduration=2.228053622 podStartE2EDuration="5.949816365s" podCreationTimestamp="2025-12-13 07:05:07 +0000 UTC" firstStartedPulling="2025-12-13 07:05:08.010128416 +0000 UTC m=+964.614537864" lastFinishedPulling="2025-12-13 07:05:11.731891159 +0000 UTC m=+968.336300607" observedRunningTime="2025-12-13 07:05:12.941729065 +0000 UTC m=+969.546138523" watchObservedRunningTime="2025-12-13 07:05:12.949816365 +0000 UTC m=+969.554225813" Dec 13 07:05:12 crc kubenswrapper[4971]: I1213 07:05:12.965665 4971 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-nmstate/nmstate-webhook-f8fb84555-pwxvn" podStartSLOduration=2.956555636 podStartE2EDuration="5.965645175s" podCreationTimestamp="2025-12-13 07:05:07 +0000 UTC" firstStartedPulling="2025-12-13 07:05:08.723996639 +0000 UTC m=+965.328406087" lastFinishedPulling="2025-12-13 07:05:11.733086168 +0000 UTC m=+968.337495626" observedRunningTime="2025-12-13 07:05:12.959009511 +0000 UTC m=+969.563418969" watchObservedRunningTime="2025-12-13 07:05:12.965645175 +0000 UTC m=+969.570054623" Dec 13 07:05:12 crc kubenswrapper[4971]: I1213 07:05:12.974960 4971 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-nmstate/nmstate-console-plugin-6ff7998486-hx58s" podStartSLOduration=2.534794302 podStartE2EDuration="5.974937004s" podCreationTimestamp="2025-12-13 07:05:07 +0000 UTC" firstStartedPulling="2025-12-13 07:05:08.291188312 +0000 UTC m=+964.895597760" lastFinishedPulling="2025-12-13 07:05:11.731331014 +0000 UTC m=+968.335740462" observedRunningTime="2025-12-13 07:05:12.974305608 +0000 UTC m=+969.578715056" watchObservedRunningTime="2025-12-13 07:05:12.974937004 +0000 UTC m=+969.579346452" Dec 13 07:05:16 crc kubenswrapper[4971]: I1213 07:05:16.154590 4971 patch_prober.go:28] interesting pod/machine-config-daemon-82xjz container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 13 07:05:16 crc kubenswrapper[4971]: I1213 07:05:16.155121 4971 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-82xjz" podUID="e0c4af12-edda-41d7-9b44-f87396174ef2" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 13 07:05:16 crc kubenswrapper[4971]: I1213 07:05:16.155182 4971 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-82xjz" Dec 13 07:05:16 crc kubenswrapper[4971]: I1213 07:05:16.155825 4971 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"c41da72089cbd159f1c07bf9d29b0bd93e1593b7236249fdd03c408a03b5a8ca"} pod="openshift-machine-config-operator/machine-config-daemon-82xjz" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Dec 13 07:05:16 crc kubenswrapper[4971]: I1213 07:05:16.155901 4971 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-82xjz" podUID="e0c4af12-edda-41d7-9b44-f87396174ef2" containerName="machine-config-daemon" containerID="cri-o://c41da72089cbd159f1c07bf9d29b0bd93e1593b7236249fdd03c408a03b5a8ca" gracePeriod=600 Dec 13 07:05:16 crc kubenswrapper[4971]: I1213 07:05:16.962605 4971 generic.go:334] "Generic (PLEG): container finished" podID="e0c4af12-edda-41d7-9b44-f87396174ef2" containerID="c41da72089cbd159f1c07bf9d29b0bd93e1593b7236249fdd03c408a03b5a8ca" exitCode=0 Dec 13 07:05:16 crc kubenswrapper[4971]: I1213 07:05:16.962684 4971 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-82xjz" event={"ID":"e0c4af12-edda-41d7-9b44-f87396174ef2","Type":"ContainerDied","Data":"c41da72089cbd159f1c07bf9d29b0bd93e1593b7236249fdd03c408a03b5a8ca"} Dec 13 07:05:16 crc kubenswrapper[4971]: I1213 07:05:16.962988 4971 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-82xjz" event={"ID":"e0c4af12-edda-41d7-9b44-f87396174ef2","Type":"ContainerStarted","Data":"21a30fd79dbe0c677372d180191864bac6880bbd1aaac9bae997b690cf3d4a99"} Dec 13 07:05:16 crc kubenswrapper[4971]: I1213 07:05:16.963046 4971 scope.go:117] "RemoveContainer" containerID="358e8ac17d5a2f5ef3b4deef1d163694a45d4c70827405c28f4c81a7c84b880e" Dec 13 07:05:16 crc kubenswrapper[4971]: I1213 07:05:16.966836 4971 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-nmstate/nmstate-metrics-7f7f7578db-xsb6k" event={"ID":"f2968080-99c8-4320-9b39-4bcab4dad759","Type":"ContainerStarted","Data":"de414b39a8544ee466dbc14950d570150cc91c4177d60187478be16ec30b39c9"} Dec 13 07:05:17 crc kubenswrapper[4971]: I1213 07:05:17.977646 4971 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-nmstate/nmstate-handler-5b4bx" Dec 13 07:05:17 crc kubenswrapper[4971]: I1213 07:05:17.996952 4971 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-nmstate/nmstate-metrics-7f7f7578db-xsb6k" podStartSLOduration=3.147200013 podStartE2EDuration="10.996926721s" podCreationTimestamp="2025-12-13 07:05:07 +0000 UTC" firstStartedPulling="2025-12-13 07:05:08.208313529 +0000 UTC m=+964.812722987" lastFinishedPulling="2025-12-13 07:05:16.058040247 +0000 UTC m=+972.662449695" observedRunningTime="2025-12-13 07:05:17.000933574 +0000 UTC m=+973.605343022" watchObservedRunningTime="2025-12-13 07:05:17.996926721 +0000 UTC m=+974.601336169" Dec 13 07:05:18 crc kubenswrapper[4971]: I1213 07:05:18.276728 4971 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-console/console-79476f9bb4-9ngf9" Dec 13 07:05:18 crc kubenswrapper[4971]: I1213 07:05:18.276791 4971 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-console/console-79476f9bb4-9ngf9" Dec 13 07:05:18 crc kubenswrapper[4971]: I1213 07:05:18.281436 4971 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-console/console-79476f9bb4-9ngf9" Dec 13 07:05:18 crc kubenswrapper[4971]: I1213 07:05:18.983891 4971 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-console/console-79476f9bb4-9ngf9" Dec 13 07:05:19 crc kubenswrapper[4971]: I1213 07:05:19.033209 4971 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-console/console-f9d7485db-5fd9b"] Dec 13 07:05:28 crc kubenswrapper[4971]: I1213 07:05:28.516037 4971 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-nmstate/nmstate-webhook-f8fb84555-pwxvn" Dec 13 07:05:41 crc kubenswrapper[4971]: I1213 07:05:41.320486 4971 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/5b7fccbebf0e22d2dd769066fa7aaa90fd620c5db34f2af6c91e4319d4gv4b7"] Dec 13 07:05:41 crc kubenswrapper[4971]: I1213 07:05:41.322183 4971 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/5b7fccbebf0e22d2dd769066fa7aaa90fd620c5db34f2af6c91e4319d4gv4b7" Dec 13 07:05:41 crc kubenswrapper[4971]: I1213 07:05:41.324446 4971 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"default-dockercfg-vmwhc" Dec 13 07:05:41 crc kubenswrapper[4971]: I1213 07:05:41.333028 4971 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/5b7fccbebf0e22d2dd769066fa7aaa90fd620c5db34f2af6c91e4319d4gv4b7"] Dec 13 07:05:41 crc kubenswrapper[4971]: I1213 07:05:41.474747 4971 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/75227ae7-8658-447d-90c5-b0619fe75d8f-util\") pod \"5b7fccbebf0e22d2dd769066fa7aaa90fd620c5db34f2af6c91e4319d4gv4b7\" (UID: \"75227ae7-8658-447d-90c5-b0619fe75d8f\") " pod="openshift-marketplace/5b7fccbebf0e22d2dd769066fa7aaa90fd620c5db34f2af6c91e4319d4gv4b7" Dec 13 07:05:41 crc kubenswrapper[4971]: I1213 07:05:41.475112 4971 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-48rf9\" (UniqueName: \"kubernetes.io/projected/75227ae7-8658-447d-90c5-b0619fe75d8f-kube-api-access-48rf9\") pod \"5b7fccbebf0e22d2dd769066fa7aaa90fd620c5db34f2af6c91e4319d4gv4b7\" (UID: \"75227ae7-8658-447d-90c5-b0619fe75d8f\") " pod="openshift-marketplace/5b7fccbebf0e22d2dd769066fa7aaa90fd620c5db34f2af6c91e4319d4gv4b7" Dec 13 07:05:41 crc kubenswrapper[4971]: I1213 07:05:41.475240 4971 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/75227ae7-8658-447d-90c5-b0619fe75d8f-bundle\") pod \"5b7fccbebf0e22d2dd769066fa7aaa90fd620c5db34f2af6c91e4319d4gv4b7\" (UID: \"75227ae7-8658-447d-90c5-b0619fe75d8f\") " pod="openshift-marketplace/5b7fccbebf0e22d2dd769066fa7aaa90fd620c5db34f2af6c91e4319d4gv4b7" Dec 13 07:05:41 crc kubenswrapper[4971]: I1213 07:05:41.576910 4971 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/75227ae7-8658-447d-90c5-b0619fe75d8f-bundle\") pod \"5b7fccbebf0e22d2dd769066fa7aaa90fd620c5db34f2af6c91e4319d4gv4b7\" (UID: \"75227ae7-8658-447d-90c5-b0619fe75d8f\") " pod="openshift-marketplace/5b7fccbebf0e22d2dd769066fa7aaa90fd620c5db34f2af6c91e4319d4gv4b7" Dec 13 07:05:41 crc kubenswrapper[4971]: I1213 07:05:41.577022 4971 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/75227ae7-8658-447d-90c5-b0619fe75d8f-util\") pod \"5b7fccbebf0e22d2dd769066fa7aaa90fd620c5db34f2af6c91e4319d4gv4b7\" (UID: \"75227ae7-8658-447d-90c5-b0619fe75d8f\") " pod="openshift-marketplace/5b7fccbebf0e22d2dd769066fa7aaa90fd620c5db34f2af6c91e4319d4gv4b7" Dec 13 07:05:41 crc kubenswrapper[4971]: I1213 07:05:41.577068 4971 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-48rf9\" (UniqueName: \"kubernetes.io/projected/75227ae7-8658-447d-90c5-b0619fe75d8f-kube-api-access-48rf9\") pod \"5b7fccbebf0e22d2dd769066fa7aaa90fd620c5db34f2af6c91e4319d4gv4b7\" (UID: \"75227ae7-8658-447d-90c5-b0619fe75d8f\") " pod="openshift-marketplace/5b7fccbebf0e22d2dd769066fa7aaa90fd620c5db34f2af6c91e4319d4gv4b7" Dec 13 07:05:41 crc kubenswrapper[4971]: I1213 07:05:41.577923 4971 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/75227ae7-8658-447d-90c5-b0619fe75d8f-util\") pod \"5b7fccbebf0e22d2dd769066fa7aaa90fd620c5db34f2af6c91e4319d4gv4b7\" (UID: \"75227ae7-8658-447d-90c5-b0619fe75d8f\") " pod="openshift-marketplace/5b7fccbebf0e22d2dd769066fa7aaa90fd620c5db34f2af6c91e4319d4gv4b7" Dec 13 07:05:41 crc kubenswrapper[4971]: I1213 07:05:41.577923 4971 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/75227ae7-8658-447d-90c5-b0619fe75d8f-bundle\") pod \"5b7fccbebf0e22d2dd769066fa7aaa90fd620c5db34f2af6c91e4319d4gv4b7\" (UID: \"75227ae7-8658-447d-90c5-b0619fe75d8f\") " pod="openshift-marketplace/5b7fccbebf0e22d2dd769066fa7aaa90fd620c5db34f2af6c91e4319d4gv4b7" Dec 13 07:05:41 crc kubenswrapper[4971]: I1213 07:05:41.603175 4971 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-48rf9\" (UniqueName: \"kubernetes.io/projected/75227ae7-8658-447d-90c5-b0619fe75d8f-kube-api-access-48rf9\") pod \"5b7fccbebf0e22d2dd769066fa7aaa90fd620c5db34f2af6c91e4319d4gv4b7\" (UID: \"75227ae7-8658-447d-90c5-b0619fe75d8f\") " pod="openshift-marketplace/5b7fccbebf0e22d2dd769066fa7aaa90fd620c5db34f2af6c91e4319d4gv4b7" Dec 13 07:05:41 crc kubenswrapper[4971]: I1213 07:05:41.642828 4971 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/5b7fccbebf0e22d2dd769066fa7aaa90fd620c5db34f2af6c91e4319d4gv4b7" Dec 13 07:05:42 crc kubenswrapper[4971]: I1213 07:05:42.085570 4971 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/5b7fccbebf0e22d2dd769066fa7aaa90fd620c5db34f2af6c91e4319d4gv4b7"] Dec 13 07:05:42 crc kubenswrapper[4971]: I1213 07:05:42.116199 4971 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/5b7fccbebf0e22d2dd769066fa7aaa90fd620c5db34f2af6c91e4319d4gv4b7" event={"ID":"75227ae7-8658-447d-90c5-b0619fe75d8f","Type":"ContainerStarted","Data":"0bd0b1ca51d8798a0485c62b2191f2a593a713f27eb99ad8952a88eefca17c7b"} Dec 13 07:05:44 crc kubenswrapper[4971]: I1213 07:05:44.077638 4971 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-console/console-f9d7485db-5fd9b" podUID="9f0db656-7a94-415b-b737-30ed342817ee" containerName="console" containerID="cri-o://8e76dd9bdab28393e8fb3d0e9f5f44af7e8d1a8c9d3e19e32359e99815f53655" gracePeriod=15 Dec 13 07:05:44 crc kubenswrapper[4971]: I1213 07:05:44.135073 4971 generic.go:334] "Generic (PLEG): container finished" podID="75227ae7-8658-447d-90c5-b0619fe75d8f" containerID="91d36005989c7c4ee3384cd018d101323ba9e20452874e17022031dc63dc5fe5" exitCode=0 Dec 13 07:05:44 crc kubenswrapper[4971]: I1213 07:05:44.135130 4971 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/5b7fccbebf0e22d2dd769066fa7aaa90fd620c5db34f2af6c91e4319d4gv4b7" event={"ID":"75227ae7-8658-447d-90c5-b0619fe75d8f","Type":"ContainerDied","Data":"91d36005989c7c4ee3384cd018d101323ba9e20452874e17022031dc63dc5fe5"} Dec 13 07:05:44 crc kubenswrapper[4971]: I1213 07:05:44.444955 4971 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-console_console-f9d7485db-5fd9b_9f0db656-7a94-415b-b737-30ed342817ee/console/0.log" Dec 13 07:05:44 crc kubenswrapper[4971]: I1213 07:05:44.445351 4971 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-console/console-f9d7485db-5fd9b" Dec 13 07:05:44 crc kubenswrapper[4971]: I1213 07:05:44.617390 4971 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/9f0db656-7a94-415b-b737-30ed342817ee-trusted-ca-bundle\") pod \"9f0db656-7a94-415b-b737-30ed342817ee\" (UID: \"9f0db656-7a94-415b-b737-30ed342817ee\") " Dec 13 07:05:44 crc kubenswrapper[4971]: I1213 07:05:44.617469 4971 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"oauth-serving-cert\" (UniqueName: \"kubernetes.io/configmap/9f0db656-7a94-415b-b737-30ed342817ee-oauth-serving-cert\") pod \"9f0db656-7a94-415b-b737-30ed342817ee\" (UID: \"9f0db656-7a94-415b-b737-30ed342817ee\") " Dec 13 07:05:44 crc kubenswrapper[4971]: I1213 07:05:44.617554 4971 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/9f0db656-7a94-415b-b737-30ed342817ee-service-ca\") pod \"9f0db656-7a94-415b-b737-30ed342817ee\" (UID: \"9f0db656-7a94-415b-b737-30ed342817ee\") " Dec 13 07:05:44 crc kubenswrapper[4971]: I1213 07:05:44.617611 4971 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-68rgh\" (UniqueName: \"kubernetes.io/projected/9f0db656-7a94-415b-b737-30ed342817ee-kube-api-access-68rgh\") pod \"9f0db656-7a94-415b-b737-30ed342817ee\" (UID: \"9f0db656-7a94-415b-b737-30ed342817ee\") " Dec 13 07:05:44 crc kubenswrapper[4971]: I1213 07:05:44.617638 4971 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"console-serving-cert\" (UniqueName: \"kubernetes.io/secret/9f0db656-7a94-415b-b737-30ed342817ee-console-serving-cert\") pod \"9f0db656-7a94-415b-b737-30ed342817ee\" (UID: \"9f0db656-7a94-415b-b737-30ed342817ee\") " Dec 13 07:05:44 crc kubenswrapper[4971]: I1213 07:05:44.617669 4971 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"console-config\" (UniqueName: \"kubernetes.io/configmap/9f0db656-7a94-415b-b737-30ed342817ee-console-config\") pod \"9f0db656-7a94-415b-b737-30ed342817ee\" (UID: \"9f0db656-7a94-415b-b737-30ed342817ee\") " Dec 13 07:05:44 crc kubenswrapper[4971]: I1213 07:05:44.617697 4971 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"console-oauth-config\" (UniqueName: \"kubernetes.io/secret/9f0db656-7a94-415b-b737-30ed342817ee-console-oauth-config\") pod \"9f0db656-7a94-415b-b737-30ed342817ee\" (UID: \"9f0db656-7a94-415b-b737-30ed342817ee\") " Dec 13 07:05:44 crc kubenswrapper[4971]: I1213 07:05:44.618573 4971 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/9f0db656-7a94-415b-b737-30ed342817ee-oauth-serving-cert" (OuterVolumeSpecName: "oauth-serving-cert") pod "9f0db656-7a94-415b-b737-30ed342817ee" (UID: "9f0db656-7a94-415b-b737-30ed342817ee"). InnerVolumeSpecName "oauth-serving-cert". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 13 07:05:44 crc kubenswrapper[4971]: I1213 07:05:44.618592 4971 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/9f0db656-7a94-415b-b737-30ed342817ee-service-ca" (OuterVolumeSpecName: "service-ca") pod "9f0db656-7a94-415b-b737-30ed342817ee" (UID: "9f0db656-7a94-415b-b737-30ed342817ee"). InnerVolumeSpecName "service-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 13 07:05:44 crc kubenswrapper[4971]: I1213 07:05:44.618617 4971 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/9f0db656-7a94-415b-b737-30ed342817ee-console-config" (OuterVolumeSpecName: "console-config") pod "9f0db656-7a94-415b-b737-30ed342817ee" (UID: "9f0db656-7a94-415b-b737-30ed342817ee"). InnerVolumeSpecName "console-config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 13 07:05:44 crc kubenswrapper[4971]: I1213 07:05:44.618978 4971 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/9f0db656-7a94-415b-b737-30ed342817ee-trusted-ca-bundle" (OuterVolumeSpecName: "trusted-ca-bundle") pod "9f0db656-7a94-415b-b737-30ed342817ee" (UID: "9f0db656-7a94-415b-b737-30ed342817ee"). InnerVolumeSpecName "trusted-ca-bundle". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 13 07:05:44 crc kubenswrapper[4971]: I1213 07:05:44.623645 4971 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/9f0db656-7a94-415b-b737-30ed342817ee-kube-api-access-68rgh" (OuterVolumeSpecName: "kube-api-access-68rgh") pod "9f0db656-7a94-415b-b737-30ed342817ee" (UID: "9f0db656-7a94-415b-b737-30ed342817ee"). InnerVolumeSpecName "kube-api-access-68rgh". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 13 07:05:44 crc kubenswrapper[4971]: I1213 07:05:44.623667 4971 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/9f0db656-7a94-415b-b737-30ed342817ee-console-oauth-config" (OuterVolumeSpecName: "console-oauth-config") pod "9f0db656-7a94-415b-b737-30ed342817ee" (UID: "9f0db656-7a94-415b-b737-30ed342817ee"). InnerVolumeSpecName "console-oauth-config". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 13 07:05:44 crc kubenswrapper[4971]: I1213 07:05:44.624166 4971 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/9f0db656-7a94-415b-b737-30ed342817ee-console-serving-cert" (OuterVolumeSpecName: "console-serving-cert") pod "9f0db656-7a94-415b-b737-30ed342817ee" (UID: "9f0db656-7a94-415b-b737-30ed342817ee"). InnerVolumeSpecName "console-serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 13 07:05:44 crc kubenswrapper[4971]: I1213 07:05:44.719155 4971 reconciler_common.go:293] "Volume detached for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/9f0db656-7a94-415b-b737-30ed342817ee-service-ca\") on node \"crc\" DevicePath \"\"" Dec 13 07:05:44 crc kubenswrapper[4971]: I1213 07:05:44.719228 4971 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-68rgh\" (UniqueName: \"kubernetes.io/projected/9f0db656-7a94-415b-b737-30ed342817ee-kube-api-access-68rgh\") on node \"crc\" DevicePath \"\"" Dec 13 07:05:44 crc kubenswrapper[4971]: I1213 07:05:44.719250 4971 reconciler_common.go:293] "Volume detached for volume \"console-serving-cert\" (UniqueName: \"kubernetes.io/secret/9f0db656-7a94-415b-b737-30ed342817ee-console-serving-cert\") on node \"crc\" DevicePath \"\"" Dec 13 07:05:44 crc kubenswrapper[4971]: I1213 07:05:44.719268 4971 reconciler_common.go:293] "Volume detached for volume \"console-config\" (UniqueName: \"kubernetes.io/configmap/9f0db656-7a94-415b-b737-30ed342817ee-console-config\") on node \"crc\" DevicePath \"\"" Dec 13 07:05:44 crc kubenswrapper[4971]: I1213 07:05:44.719286 4971 reconciler_common.go:293] "Volume detached for volume \"console-oauth-config\" (UniqueName: \"kubernetes.io/secret/9f0db656-7a94-415b-b737-30ed342817ee-console-oauth-config\") on node \"crc\" DevicePath \"\"" Dec 13 07:05:44 crc kubenswrapper[4971]: I1213 07:05:44.719302 4971 reconciler_common.go:293] "Volume detached for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/9f0db656-7a94-415b-b737-30ed342817ee-trusted-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 13 07:05:44 crc kubenswrapper[4971]: I1213 07:05:44.719318 4971 reconciler_common.go:293] "Volume detached for volume \"oauth-serving-cert\" (UniqueName: \"kubernetes.io/configmap/9f0db656-7a94-415b-b737-30ed342817ee-oauth-serving-cert\") on node \"crc\" DevicePath \"\"" Dec 13 07:05:45 crc kubenswrapper[4971]: I1213 07:05:45.142649 4971 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-console_console-f9d7485db-5fd9b_9f0db656-7a94-415b-b737-30ed342817ee/console/0.log" Dec 13 07:05:45 crc kubenswrapper[4971]: I1213 07:05:45.142696 4971 generic.go:334] "Generic (PLEG): container finished" podID="9f0db656-7a94-415b-b737-30ed342817ee" containerID="8e76dd9bdab28393e8fb3d0e9f5f44af7e8d1a8c9d3e19e32359e99815f53655" exitCode=2 Dec 13 07:05:45 crc kubenswrapper[4971]: I1213 07:05:45.142759 4971 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-console/console-f9d7485db-5fd9b" event={"ID":"9f0db656-7a94-415b-b737-30ed342817ee","Type":"ContainerDied","Data":"8e76dd9bdab28393e8fb3d0e9f5f44af7e8d1a8c9d3e19e32359e99815f53655"} Dec 13 07:05:45 crc kubenswrapper[4971]: I1213 07:05:45.142784 4971 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-console/console-f9d7485db-5fd9b" event={"ID":"9f0db656-7a94-415b-b737-30ed342817ee","Type":"ContainerDied","Data":"13f95dbc7e9f1041c783b48af992f4b000a5028dcacf9bc06266384582807d80"} Dec 13 07:05:45 crc kubenswrapper[4971]: I1213 07:05:45.142801 4971 scope.go:117] "RemoveContainer" containerID="8e76dd9bdab28393e8fb3d0e9f5f44af7e8d1a8c9d3e19e32359e99815f53655" Dec 13 07:05:45 crc kubenswrapper[4971]: I1213 07:05:45.142870 4971 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-console/console-f9d7485db-5fd9b" Dec 13 07:05:45 crc kubenswrapper[4971]: I1213 07:05:45.161423 4971 scope.go:117] "RemoveContainer" containerID="8e76dd9bdab28393e8fb3d0e9f5f44af7e8d1a8c9d3e19e32359e99815f53655" Dec 13 07:05:45 crc kubenswrapper[4971]: E1213 07:05:45.162910 4971 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"8e76dd9bdab28393e8fb3d0e9f5f44af7e8d1a8c9d3e19e32359e99815f53655\": container with ID starting with 8e76dd9bdab28393e8fb3d0e9f5f44af7e8d1a8c9d3e19e32359e99815f53655 not found: ID does not exist" containerID="8e76dd9bdab28393e8fb3d0e9f5f44af7e8d1a8c9d3e19e32359e99815f53655" Dec 13 07:05:45 crc kubenswrapper[4971]: I1213 07:05:45.163046 4971 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"8e76dd9bdab28393e8fb3d0e9f5f44af7e8d1a8c9d3e19e32359e99815f53655"} err="failed to get container status \"8e76dd9bdab28393e8fb3d0e9f5f44af7e8d1a8c9d3e19e32359e99815f53655\": rpc error: code = NotFound desc = could not find container \"8e76dd9bdab28393e8fb3d0e9f5f44af7e8d1a8c9d3e19e32359e99815f53655\": container with ID starting with 8e76dd9bdab28393e8fb3d0e9f5f44af7e8d1a8c9d3e19e32359e99815f53655 not found: ID does not exist" Dec 13 07:05:45 crc kubenswrapper[4971]: I1213 07:05:45.173594 4971 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-console/console-f9d7485db-5fd9b"] Dec 13 07:05:45 crc kubenswrapper[4971]: I1213 07:05:45.177629 4971 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-console/console-f9d7485db-5fd9b"] Dec 13 07:05:45 crc kubenswrapper[4971]: I1213 07:05:45.775851 4971 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="9f0db656-7a94-415b-b737-30ed342817ee" path="/var/lib/kubelet/pods/9f0db656-7a94-415b-b737-30ed342817ee/volumes" Dec 13 07:05:46 crc kubenswrapper[4971]: I1213 07:05:46.149664 4971 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/5b7fccbebf0e22d2dd769066fa7aaa90fd620c5db34f2af6c91e4319d4gv4b7" event={"ID":"75227ae7-8658-447d-90c5-b0619fe75d8f","Type":"ContainerStarted","Data":"20e3355b1b0043cac95a5960b20e2b1ebbd10691a0363ed50da1d526544609f8"} Dec 13 07:05:47 crc kubenswrapper[4971]: I1213 07:05:47.163805 4971 generic.go:334] "Generic (PLEG): container finished" podID="75227ae7-8658-447d-90c5-b0619fe75d8f" containerID="20e3355b1b0043cac95a5960b20e2b1ebbd10691a0363ed50da1d526544609f8" exitCode=0 Dec 13 07:05:47 crc kubenswrapper[4971]: I1213 07:05:47.163944 4971 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/5b7fccbebf0e22d2dd769066fa7aaa90fd620c5db34f2af6c91e4319d4gv4b7" event={"ID":"75227ae7-8658-447d-90c5-b0619fe75d8f","Type":"ContainerDied","Data":"20e3355b1b0043cac95a5960b20e2b1ebbd10691a0363ed50da1d526544609f8"} Dec 13 07:05:48 crc kubenswrapper[4971]: I1213 07:05:48.171624 4971 generic.go:334] "Generic (PLEG): container finished" podID="75227ae7-8658-447d-90c5-b0619fe75d8f" containerID="0c52c68f13fe56afd53764dbd2e7f7905365d5a8b1273f93ca2686ba0530ec0a" exitCode=0 Dec 13 07:05:48 crc kubenswrapper[4971]: I1213 07:05:48.171720 4971 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/5b7fccbebf0e22d2dd769066fa7aaa90fd620c5db34f2af6c91e4319d4gv4b7" event={"ID":"75227ae7-8658-447d-90c5-b0619fe75d8f","Type":"ContainerDied","Data":"0c52c68f13fe56afd53764dbd2e7f7905365d5a8b1273f93ca2686ba0530ec0a"} Dec 13 07:05:49 crc kubenswrapper[4971]: I1213 07:05:49.394695 4971 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/5b7fccbebf0e22d2dd769066fa7aaa90fd620c5db34f2af6c91e4319d4gv4b7" Dec 13 07:05:49 crc kubenswrapper[4971]: I1213 07:05:49.486112 4971 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-48rf9\" (UniqueName: \"kubernetes.io/projected/75227ae7-8658-447d-90c5-b0619fe75d8f-kube-api-access-48rf9\") pod \"75227ae7-8658-447d-90c5-b0619fe75d8f\" (UID: \"75227ae7-8658-447d-90c5-b0619fe75d8f\") " Dec 13 07:05:49 crc kubenswrapper[4971]: I1213 07:05:49.486171 4971 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/75227ae7-8658-447d-90c5-b0619fe75d8f-util\") pod \"75227ae7-8658-447d-90c5-b0619fe75d8f\" (UID: \"75227ae7-8658-447d-90c5-b0619fe75d8f\") " Dec 13 07:05:49 crc kubenswrapper[4971]: I1213 07:05:49.486217 4971 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/75227ae7-8658-447d-90c5-b0619fe75d8f-bundle\") pod \"75227ae7-8658-447d-90c5-b0619fe75d8f\" (UID: \"75227ae7-8658-447d-90c5-b0619fe75d8f\") " Dec 13 07:05:49 crc kubenswrapper[4971]: I1213 07:05:49.487389 4971 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/75227ae7-8658-447d-90c5-b0619fe75d8f-bundle" (OuterVolumeSpecName: "bundle") pod "75227ae7-8658-447d-90c5-b0619fe75d8f" (UID: "75227ae7-8658-447d-90c5-b0619fe75d8f"). InnerVolumeSpecName "bundle". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 13 07:05:49 crc kubenswrapper[4971]: I1213 07:05:49.498038 4971 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/75227ae7-8658-447d-90c5-b0619fe75d8f-util" (OuterVolumeSpecName: "util") pod "75227ae7-8658-447d-90c5-b0619fe75d8f" (UID: "75227ae7-8658-447d-90c5-b0619fe75d8f"). InnerVolumeSpecName "util". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 13 07:05:49 crc kubenswrapper[4971]: I1213 07:05:49.498896 4971 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/75227ae7-8658-447d-90c5-b0619fe75d8f-kube-api-access-48rf9" (OuterVolumeSpecName: "kube-api-access-48rf9") pod "75227ae7-8658-447d-90c5-b0619fe75d8f" (UID: "75227ae7-8658-447d-90c5-b0619fe75d8f"). InnerVolumeSpecName "kube-api-access-48rf9". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 13 07:05:49 crc kubenswrapper[4971]: I1213 07:05:49.587157 4971 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-48rf9\" (UniqueName: \"kubernetes.io/projected/75227ae7-8658-447d-90c5-b0619fe75d8f-kube-api-access-48rf9\") on node \"crc\" DevicePath \"\"" Dec 13 07:05:49 crc kubenswrapper[4971]: I1213 07:05:49.587190 4971 reconciler_common.go:293] "Volume detached for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/75227ae7-8658-447d-90c5-b0619fe75d8f-util\") on node \"crc\" DevicePath \"\"" Dec 13 07:05:49 crc kubenswrapper[4971]: I1213 07:05:49.587199 4971 reconciler_common.go:293] "Volume detached for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/75227ae7-8658-447d-90c5-b0619fe75d8f-bundle\") on node \"crc\" DevicePath \"\"" Dec 13 07:05:50 crc kubenswrapper[4971]: I1213 07:05:50.184063 4971 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/5b7fccbebf0e22d2dd769066fa7aaa90fd620c5db34f2af6c91e4319d4gv4b7" event={"ID":"75227ae7-8658-447d-90c5-b0619fe75d8f","Type":"ContainerDied","Data":"0bd0b1ca51d8798a0485c62b2191f2a593a713f27eb99ad8952a88eefca17c7b"} Dec 13 07:05:50 crc kubenswrapper[4971]: I1213 07:05:50.184104 4971 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="0bd0b1ca51d8798a0485c62b2191f2a593a713f27eb99ad8952a88eefca17c7b" Dec 13 07:05:50 crc kubenswrapper[4971]: I1213 07:05:50.184146 4971 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/5b7fccbebf0e22d2dd769066fa7aaa90fd620c5db34f2af6c91e4319d4gv4b7" Dec 13 07:05:59 crc kubenswrapper[4971]: I1213 07:05:59.871169 4971 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["metallb-system/metallb-operator-controller-manager-9d86ffb46-9kqjk"] Dec 13 07:05:59 crc kubenswrapper[4971]: E1213 07:05:59.871913 4971 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="75227ae7-8658-447d-90c5-b0619fe75d8f" containerName="extract" Dec 13 07:05:59 crc kubenswrapper[4971]: I1213 07:05:59.871927 4971 state_mem.go:107] "Deleted CPUSet assignment" podUID="75227ae7-8658-447d-90c5-b0619fe75d8f" containerName="extract" Dec 13 07:05:59 crc kubenswrapper[4971]: E1213 07:05:59.871942 4971 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="9f0db656-7a94-415b-b737-30ed342817ee" containerName="console" Dec 13 07:05:59 crc kubenswrapper[4971]: I1213 07:05:59.871948 4971 state_mem.go:107] "Deleted CPUSet assignment" podUID="9f0db656-7a94-415b-b737-30ed342817ee" containerName="console" Dec 13 07:05:59 crc kubenswrapper[4971]: E1213 07:05:59.871958 4971 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="75227ae7-8658-447d-90c5-b0619fe75d8f" containerName="util" Dec 13 07:05:59 crc kubenswrapper[4971]: I1213 07:05:59.871964 4971 state_mem.go:107] "Deleted CPUSet assignment" podUID="75227ae7-8658-447d-90c5-b0619fe75d8f" containerName="util" Dec 13 07:05:59 crc kubenswrapper[4971]: E1213 07:05:59.871973 4971 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="75227ae7-8658-447d-90c5-b0619fe75d8f" containerName="pull" Dec 13 07:05:59 crc kubenswrapper[4971]: I1213 07:05:59.871979 4971 state_mem.go:107] "Deleted CPUSet assignment" podUID="75227ae7-8658-447d-90c5-b0619fe75d8f" containerName="pull" Dec 13 07:05:59 crc kubenswrapper[4971]: I1213 07:05:59.872091 4971 memory_manager.go:354] "RemoveStaleState removing state" podUID="9f0db656-7a94-415b-b737-30ed342817ee" containerName="console" Dec 13 07:05:59 crc kubenswrapper[4971]: I1213 07:05:59.872107 4971 memory_manager.go:354] "RemoveStaleState removing state" podUID="75227ae7-8658-447d-90c5-b0619fe75d8f" containerName="extract" Dec 13 07:05:59 crc kubenswrapper[4971]: I1213 07:05:59.872505 4971 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="metallb-system/metallb-operator-controller-manager-9d86ffb46-9kqjk" Dec 13 07:05:59 crc kubenswrapper[4971]: I1213 07:05:59.874857 4971 reflector.go:368] Caches populated for *v1.Secret from object-"metallb-system"/"metallb-operator-controller-manager-service-cert" Dec 13 07:05:59 crc kubenswrapper[4971]: I1213 07:05:59.874862 4971 reflector.go:368] Caches populated for *v1.ConfigMap from object-"metallb-system"/"kube-root-ca.crt" Dec 13 07:05:59 crc kubenswrapper[4971]: I1213 07:05:59.874917 4971 reflector.go:368] Caches populated for *v1.Secret from object-"metallb-system"/"manager-account-dockercfg-8nwqw" Dec 13 07:05:59 crc kubenswrapper[4971]: I1213 07:05:59.875459 4971 reflector.go:368] Caches populated for *v1.ConfigMap from object-"metallb-system"/"openshift-service-ca.crt" Dec 13 07:05:59 crc kubenswrapper[4971]: I1213 07:05:59.876746 4971 reflector.go:368] Caches populated for *v1.Secret from object-"metallb-system"/"metallb-operator-webhook-server-cert" Dec 13 07:05:59 crc kubenswrapper[4971]: I1213 07:05:59.891867 4971 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["metallb-system/metallb-operator-controller-manager-9d86ffb46-9kqjk"] Dec 13 07:05:59 crc kubenswrapper[4971]: I1213 07:05:59.968162 4971 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-2tb48\" (UniqueName: \"kubernetes.io/projected/1a9addf5-4d0c-447d-8a9a-e6c15b9cdb3a-kube-api-access-2tb48\") pod \"metallb-operator-controller-manager-9d86ffb46-9kqjk\" (UID: \"1a9addf5-4d0c-447d-8a9a-e6c15b9cdb3a\") " pod="metallb-system/metallb-operator-controller-manager-9d86ffb46-9kqjk" Dec 13 07:05:59 crc kubenswrapper[4971]: I1213 07:05:59.968213 4971 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/1a9addf5-4d0c-447d-8a9a-e6c15b9cdb3a-webhook-cert\") pod \"metallb-operator-controller-manager-9d86ffb46-9kqjk\" (UID: \"1a9addf5-4d0c-447d-8a9a-e6c15b9cdb3a\") " pod="metallb-system/metallb-operator-controller-manager-9d86ffb46-9kqjk" Dec 13 07:05:59 crc kubenswrapper[4971]: I1213 07:05:59.968257 4971 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/1a9addf5-4d0c-447d-8a9a-e6c15b9cdb3a-apiservice-cert\") pod \"metallb-operator-controller-manager-9d86ffb46-9kqjk\" (UID: \"1a9addf5-4d0c-447d-8a9a-e6c15b9cdb3a\") " pod="metallb-system/metallb-operator-controller-manager-9d86ffb46-9kqjk" Dec 13 07:06:00 crc kubenswrapper[4971]: I1213 07:06:00.068946 4971 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-2tb48\" (UniqueName: \"kubernetes.io/projected/1a9addf5-4d0c-447d-8a9a-e6c15b9cdb3a-kube-api-access-2tb48\") pod \"metallb-operator-controller-manager-9d86ffb46-9kqjk\" (UID: \"1a9addf5-4d0c-447d-8a9a-e6c15b9cdb3a\") " pod="metallb-system/metallb-operator-controller-manager-9d86ffb46-9kqjk" Dec 13 07:06:00 crc kubenswrapper[4971]: I1213 07:06:00.069001 4971 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/1a9addf5-4d0c-447d-8a9a-e6c15b9cdb3a-webhook-cert\") pod \"metallb-operator-controller-manager-9d86ffb46-9kqjk\" (UID: \"1a9addf5-4d0c-447d-8a9a-e6c15b9cdb3a\") " pod="metallb-system/metallb-operator-controller-manager-9d86ffb46-9kqjk" Dec 13 07:06:00 crc kubenswrapper[4971]: I1213 07:06:00.069044 4971 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/1a9addf5-4d0c-447d-8a9a-e6c15b9cdb3a-apiservice-cert\") pod \"metallb-operator-controller-manager-9d86ffb46-9kqjk\" (UID: \"1a9addf5-4d0c-447d-8a9a-e6c15b9cdb3a\") " pod="metallb-system/metallb-operator-controller-manager-9d86ffb46-9kqjk" Dec 13 07:06:00 crc kubenswrapper[4971]: I1213 07:06:00.074805 4971 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/1a9addf5-4d0c-447d-8a9a-e6c15b9cdb3a-webhook-cert\") pod \"metallb-operator-controller-manager-9d86ffb46-9kqjk\" (UID: \"1a9addf5-4d0c-447d-8a9a-e6c15b9cdb3a\") " pod="metallb-system/metallb-operator-controller-manager-9d86ffb46-9kqjk" Dec 13 07:06:00 crc kubenswrapper[4971]: I1213 07:06:00.078084 4971 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/1a9addf5-4d0c-447d-8a9a-e6c15b9cdb3a-apiservice-cert\") pod \"metallb-operator-controller-manager-9d86ffb46-9kqjk\" (UID: \"1a9addf5-4d0c-447d-8a9a-e6c15b9cdb3a\") " pod="metallb-system/metallb-operator-controller-manager-9d86ffb46-9kqjk" Dec 13 07:06:00 crc kubenswrapper[4971]: I1213 07:06:00.084242 4971 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-2tb48\" (UniqueName: \"kubernetes.io/projected/1a9addf5-4d0c-447d-8a9a-e6c15b9cdb3a-kube-api-access-2tb48\") pod \"metallb-operator-controller-manager-9d86ffb46-9kqjk\" (UID: \"1a9addf5-4d0c-447d-8a9a-e6c15b9cdb3a\") " pod="metallb-system/metallb-operator-controller-manager-9d86ffb46-9kqjk" Dec 13 07:06:00 crc kubenswrapper[4971]: I1213 07:06:00.128434 4971 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["metallb-system/metallb-operator-webhook-server-657bbb5cbf-njfvg"] Dec 13 07:06:00 crc kubenswrapper[4971]: I1213 07:06:00.129274 4971 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="metallb-system/metallb-operator-webhook-server-657bbb5cbf-njfvg" Dec 13 07:06:00 crc kubenswrapper[4971]: I1213 07:06:00.131662 4971 reflector.go:368] Caches populated for *v1.Secret from object-"metallb-system"/"metallb-operator-webhook-server-service-cert" Dec 13 07:06:00 crc kubenswrapper[4971]: I1213 07:06:00.131838 4971 reflector.go:368] Caches populated for *v1.Secret from object-"metallb-system"/"metallb-webhook-cert" Dec 13 07:06:00 crc kubenswrapper[4971]: I1213 07:06:00.131976 4971 reflector.go:368] Caches populated for *v1.Secret from object-"metallb-system"/"controller-dockercfg-htdpr" Dec 13 07:06:00 crc kubenswrapper[4971]: I1213 07:06:00.145589 4971 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["metallb-system/metallb-operator-webhook-server-657bbb5cbf-njfvg"] Dec 13 07:06:00 crc kubenswrapper[4971]: I1213 07:06:00.187928 4971 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="metallb-system/metallb-operator-controller-manager-9d86ffb46-9kqjk" Dec 13 07:06:00 crc kubenswrapper[4971]: I1213 07:06:00.272195 4971 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-xt5kb\" (UniqueName: \"kubernetes.io/projected/98e5a1fd-e920-4f5d-82c3-7450cfc0487a-kube-api-access-xt5kb\") pod \"metallb-operator-webhook-server-657bbb5cbf-njfvg\" (UID: \"98e5a1fd-e920-4f5d-82c3-7450cfc0487a\") " pod="metallb-system/metallb-operator-webhook-server-657bbb5cbf-njfvg" Dec 13 07:06:00 crc kubenswrapper[4971]: I1213 07:06:00.272488 4971 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/98e5a1fd-e920-4f5d-82c3-7450cfc0487a-apiservice-cert\") pod \"metallb-operator-webhook-server-657bbb5cbf-njfvg\" (UID: \"98e5a1fd-e920-4f5d-82c3-7450cfc0487a\") " pod="metallb-system/metallb-operator-webhook-server-657bbb5cbf-njfvg" Dec 13 07:06:00 crc kubenswrapper[4971]: I1213 07:06:00.272564 4971 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/98e5a1fd-e920-4f5d-82c3-7450cfc0487a-webhook-cert\") pod \"metallb-operator-webhook-server-657bbb5cbf-njfvg\" (UID: \"98e5a1fd-e920-4f5d-82c3-7450cfc0487a\") " pod="metallb-system/metallb-operator-webhook-server-657bbb5cbf-njfvg" Dec 13 07:06:00 crc kubenswrapper[4971]: I1213 07:06:00.374134 4971 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/98e5a1fd-e920-4f5d-82c3-7450cfc0487a-apiservice-cert\") pod \"metallb-operator-webhook-server-657bbb5cbf-njfvg\" (UID: \"98e5a1fd-e920-4f5d-82c3-7450cfc0487a\") " pod="metallb-system/metallb-operator-webhook-server-657bbb5cbf-njfvg" Dec 13 07:06:00 crc kubenswrapper[4971]: I1213 07:06:00.374195 4971 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/98e5a1fd-e920-4f5d-82c3-7450cfc0487a-webhook-cert\") pod \"metallb-operator-webhook-server-657bbb5cbf-njfvg\" (UID: \"98e5a1fd-e920-4f5d-82c3-7450cfc0487a\") " pod="metallb-system/metallb-operator-webhook-server-657bbb5cbf-njfvg" Dec 13 07:06:00 crc kubenswrapper[4971]: I1213 07:06:00.374246 4971 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-xt5kb\" (UniqueName: \"kubernetes.io/projected/98e5a1fd-e920-4f5d-82c3-7450cfc0487a-kube-api-access-xt5kb\") pod \"metallb-operator-webhook-server-657bbb5cbf-njfvg\" (UID: \"98e5a1fd-e920-4f5d-82c3-7450cfc0487a\") " pod="metallb-system/metallb-operator-webhook-server-657bbb5cbf-njfvg" Dec 13 07:06:00 crc kubenswrapper[4971]: I1213 07:06:00.424674 4971 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/98e5a1fd-e920-4f5d-82c3-7450cfc0487a-apiservice-cert\") pod \"metallb-operator-webhook-server-657bbb5cbf-njfvg\" (UID: \"98e5a1fd-e920-4f5d-82c3-7450cfc0487a\") " pod="metallb-system/metallb-operator-webhook-server-657bbb5cbf-njfvg" Dec 13 07:06:00 crc kubenswrapper[4971]: I1213 07:06:00.425491 4971 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/98e5a1fd-e920-4f5d-82c3-7450cfc0487a-webhook-cert\") pod \"metallb-operator-webhook-server-657bbb5cbf-njfvg\" (UID: \"98e5a1fd-e920-4f5d-82c3-7450cfc0487a\") " pod="metallb-system/metallb-operator-webhook-server-657bbb5cbf-njfvg" Dec 13 07:06:00 crc kubenswrapper[4971]: I1213 07:06:00.454397 4971 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-xt5kb\" (UniqueName: \"kubernetes.io/projected/98e5a1fd-e920-4f5d-82c3-7450cfc0487a-kube-api-access-xt5kb\") pod \"metallb-operator-webhook-server-657bbb5cbf-njfvg\" (UID: \"98e5a1fd-e920-4f5d-82c3-7450cfc0487a\") " pod="metallb-system/metallb-operator-webhook-server-657bbb5cbf-njfvg" Dec 13 07:06:00 crc kubenswrapper[4971]: I1213 07:06:00.471032 4971 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="metallb-system/metallb-operator-webhook-server-657bbb5cbf-njfvg" Dec 13 07:06:00 crc kubenswrapper[4971]: I1213 07:06:00.633064 4971 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["metallb-system/metallb-operator-controller-manager-9d86ffb46-9kqjk"] Dec 13 07:06:00 crc kubenswrapper[4971]: I1213 07:06:00.973582 4971 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["metallb-system/metallb-operator-webhook-server-657bbb5cbf-njfvg"] Dec 13 07:06:00 crc kubenswrapper[4971]: W1213 07:06:00.985673 4971 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod98e5a1fd_e920_4f5d_82c3_7450cfc0487a.slice/crio-827af6a3f7705f698dc10fd6e936d0d34a3e36ee2a515c8117297849457d64e2 WatchSource:0}: Error finding container 827af6a3f7705f698dc10fd6e936d0d34a3e36ee2a515c8117297849457d64e2: Status 404 returned error can't find the container with id 827af6a3f7705f698dc10fd6e936d0d34a3e36ee2a515c8117297849457d64e2 Dec 13 07:06:01 crc kubenswrapper[4971]: I1213 07:06:01.256143 4971 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/metallb-operator-controller-manager-9d86ffb46-9kqjk" event={"ID":"1a9addf5-4d0c-447d-8a9a-e6c15b9cdb3a","Type":"ContainerStarted","Data":"1583e9c6c1c71adcf5d6a738b33853ac3db1d650c52afb090f7147ef8e7e9f3b"} Dec 13 07:06:01 crc kubenswrapper[4971]: I1213 07:06:01.257310 4971 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/metallb-operator-webhook-server-657bbb5cbf-njfvg" event={"ID":"98e5a1fd-e920-4f5d-82c3-7450cfc0487a","Type":"ContainerStarted","Data":"827af6a3f7705f698dc10fd6e936d0d34a3e36ee2a515c8117297849457d64e2"} Dec 13 07:06:04 crc kubenswrapper[4971]: I1213 07:06:04.279021 4971 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/metallb-operator-controller-manager-9d86ffb46-9kqjk" event={"ID":"1a9addf5-4d0c-447d-8a9a-e6c15b9cdb3a","Type":"ContainerStarted","Data":"a7bccee6c67aada4ddddba8909c0164744ea907ddb4bc7fa850c49fba1185765"} Dec 13 07:06:04 crc kubenswrapper[4971]: I1213 07:06:04.279743 4971 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="metallb-system/metallb-operator-controller-manager-9d86ffb46-9kqjk" Dec 13 07:06:07 crc kubenswrapper[4971]: I1213 07:06:07.299955 4971 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/metallb-operator-webhook-server-657bbb5cbf-njfvg" event={"ID":"98e5a1fd-e920-4f5d-82c3-7450cfc0487a","Type":"ContainerStarted","Data":"f4c17c644a50f70a2ca34ed1905936dc2fdbb99f49ad1d674972af297bb832ce"} Dec 13 07:06:07 crc kubenswrapper[4971]: I1213 07:06:07.300688 4971 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="metallb-system/metallb-operator-webhook-server-657bbb5cbf-njfvg" Dec 13 07:06:07 crc kubenswrapper[4971]: I1213 07:06:07.319788 4971 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="metallb-system/metallb-operator-controller-manager-9d86ffb46-9kqjk" podStartSLOduration=5.356310591 podStartE2EDuration="8.319769865s" podCreationTimestamp="2025-12-13 07:05:59 +0000 UTC" firstStartedPulling="2025-12-13 07:06:00.681428009 +0000 UTC m=+1017.285837457" lastFinishedPulling="2025-12-13 07:06:03.644887283 +0000 UTC m=+1020.249296731" observedRunningTime="2025-12-13 07:06:04.301556985 +0000 UTC m=+1020.905966433" watchObservedRunningTime="2025-12-13 07:06:07.319769865 +0000 UTC m=+1023.924179313" Dec 13 07:06:07 crc kubenswrapper[4971]: I1213 07:06:07.321285 4971 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="metallb-system/metallb-operator-webhook-server-657bbb5cbf-njfvg" podStartSLOduration=2.151350986 podStartE2EDuration="7.321274113s" podCreationTimestamp="2025-12-13 07:06:00 +0000 UTC" firstStartedPulling="2025-12-13 07:06:00.989608584 +0000 UTC m=+1017.594018032" lastFinishedPulling="2025-12-13 07:06:06.159531711 +0000 UTC m=+1022.763941159" observedRunningTime="2025-12-13 07:06:07.318509975 +0000 UTC m=+1023.922919433" watchObservedRunningTime="2025-12-13 07:06:07.321274113 +0000 UTC m=+1023.925683561" Dec 13 07:06:20 crc kubenswrapper[4971]: I1213 07:06:20.480584 4971 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="metallb-system/metallb-operator-webhook-server-657bbb5cbf-njfvg" Dec 13 07:06:40 crc kubenswrapper[4971]: I1213 07:06:40.190458 4971 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="metallb-system/metallb-operator-controller-manager-9d86ffb46-9kqjk" Dec 13 07:06:40 crc kubenswrapper[4971]: I1213 07:06:40.908697 4971 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["metallb-system/frr-k8s-webhook-server-7784b6fcf-tkg6z"] Dec 13 07:06:40 crc kubenswrapper[4971]: I1213 07:06:40.909755 4971 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="metallb-system/frr-k8s-webhook-server-7784b6fcf-tkg6z" Dec 13 07:06:40 crc kubenswrapper[4971]: I1213 07:06:40.912434 4971 reflector.go:368] Caches populated for *v1.Secret from object-"metallb-system"/"frr-k8s-webhook-server-cert" Dec 13 07:06:40 crc kubenswrapper[4971]: I1213 07:06:40.914156 4971 reflector.go:368] Caches populated for *v1.Secret from object-"metallb-system"/"frr-k8s-daemon-dockercfg-gxq5h" Dec 13 07:06:40 crc kubenswrapper[4971]: I1213 07:06:40.941750 4971 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["metallb-system/frr-k8s-webhook-server-7784b6fcf-tkg6z"] Dec 13 07:06:40 crc kubenswrapper[4971]: I1213 07:06:40.948973 4971 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["metallb-system/frr-k8s-szzcz"] Dec 13 07:06:40 crc kubenswrapper[4971]: I1213 07:06:40.952238 4971 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="metallb-system/frr-k8s-szzcz" Dec 13 07:06:40 crc kubenswrapper[4971]: I1213 07:06:40.954454 4971 reflector.go:368] Caches populated for *v1.Secret from object-"metallb-system"/"frr-k8s-certs-secret" Dec 13 07:06:40 crc kubenswrapper[4971]: I1213 07:06:40.954837 4971 reflector.go:368] Caches populated for *v1.ConfigMap from object-"metallb-system"/"frr-startup" Dec 13 07:06:41 crc kubenswrapper[4971]: I1213 07:06:41.014759 4971 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["metallb-system/speaker-v229x"] Dec 13 07:06:41 crc kubenswrapper[4971]: I1213 07:06:41.015885 4971 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="metallb-system/speaker-v229x" Dec 13 07:06:41 crc kubenswrapper[4971]: I1213 07:06:41.021451 4971 reflector.go:368] Caches populated for *v1.Secret from object-"metallb-system"/"speaker-dockercfg-9ptjf" Dec 13 07:06:41 crc kubenswrapper[4971]: I1213 07:06:41.021953 4971 reflector.go:368] Caches populated for *v1.Secret from object-"metallb-system"/"speaker-certs-secret" Dec 13 07:06:41 crc kubenswrapper[4971]: I1213 07:06:41.021953 4971 reflector.go:368] Caches populated for *v1.Secret from object-"metallb-system"/"metallb-memberlist" Dec 13 07:06:41 crc kubenswrapper[4971]: I1213 07:06:41.023115 4971 reflector.go:368] Caches populated for *v1.ConfigMap from object-"metallb-system"/"metallb-excludel2" Dec 13 07:06:41 crc kubenswrapper[4971]: I1213 07:06:41.028431 4971 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["metallb-system/controller-5bddd4b946-2kfxb"] Dec 13 07:06:41 crc kubenswrapper[4971]: I1213 07:06:41.029602 4971 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="metallb-system/controller-5bddd4b946-2kfxb" Dec 13 07:06:41 crc kubenswrapper[4971]: I1213 07:06:41.032700 4971 reflector.go:368] Caches populated for *v1.Secret from object-"metallb-system"/"controller-certs-secret" Dec 13 07:06:41 crc kubenswrapper[4971]: I1213 07:06:41.043870 4971 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-45zsl\" (UniqueName: \"kubernetes.io/projected/f093984d-c941-4805-a0c1-0926aa7d7f38-kube-api-access-45zsl\") pod \"frr-k8s-webhook-server-7784b6fcf-tkg6z\" (UID: \"f093984d-c941-4805-a0c1-0926aa7d7f38\") " pod="metallb-system/frr-k8s-webhook-server-7784b6fcf-tkg6z" Dec 13 07:06:41 crc kubenswrapper[4971]: I1213 07:06:41.043946 4971 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/f093984d-c941-4805-a0c1-0926aa7d7f38-cert\") pod \"frr-k8s-webhook-server-7784b6fcf-tkg6z\" (UID: \"f093984d-c941-4805-a0c1-0926aa7d7f38\") " pod="metallb-system/frr-k8s-webhook-server-7784b6fcf-tkg6z" Dec 13 07:06:41 crc kubenswrapper[4971]: I1213 07:06:41.047675 4971 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["metallb-system/controller-5bddd4b946-2kfxb"] Dec 13 07:06:41 crc kubenswrapper[4971]: I1213 07:06:41.145630 4971 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-jzrst\" (UniqueName: \"kubernetes.io/projected/1f567f13-d990-4035-92e4-64835af4aac8-kube-api-access-jzrst\") pod \"frr-k8s-szzcz\" (UID: \"1f567f13-d990-4035-92e4-64835af4aac8\") " pod="metallb-system/frr-k8s-szzcz" Dec 13 07:06:41 crc kubenswrapper[4971]: I1213 07:06:41.145690 4971 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"frr-startup\" (UniqueName: \"kubernetes.io/configmap/1f567f13-d990-4035-92e4-64835af4aac8-frr-startup\") pod \"frr-k8s-szzcz\" (UID: \"1f567f13-d990-4035-92e4-64835af4aac8\") " pod="metallb-system/frr-k8s-szzcz" Dec 13 07:06:41 crc kubenswrapper[4971]: I1213 07:06:41.145723 4971 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/1f567f13-d990-4035-92e4-64835af4aac8-metrics-certs\") pod \"frr-k8s-szzcz\" (UID: \"1f567f13-d990-4035-92e4-64835af4aac8\") " pod="metallb-system/frr-k8s-szzcz" Dec 13 07:06:41 crc kubenswrapper[4971]: I1213 07:06:41.145752 4971 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metallb-excludel2\" (UniqueName: \"kubernetes.io/configmap/c0ee70f0-8158-4fec-bae8-7cd72bdef461-metallb-excludel2\") pod \"speaker-v229x\" (UID: \"c0ee70f0-8158-4fec-bae8-7cd72bdef461\") " pod="metallb-system/speaker-v229x" Dec 13 07:06:41 crc kubenswrapper[4971]: I1213 07:06:41.145778 4971 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-45zsl\" (UniqueName: \"kubernetes.io/projected/f093984d-c941-4805-a0c1-0926aa7d7f38-kube-api-access-45zsl\") pod \"frr-k8s-webhook-server-7784b6fcf-tkg6z\" (UID: \"f093984d-c941-4805-a0c1-0926aa7d7f38\") " pod="metallb-system/frr-k8s-webhook-server-7784b6fcf-tkg6z" Dec 13 07:06:41 crc kubenswrapper[4971]: I1213 07:06:41.145823 4971 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"memberlist\" (UniqueName: \"kubernetes.io/secret/c0ee70f0-8158-4fec-bae8-7cd72bdef461-memberlist\") pod \"speaker-v229x\" (UID: \"c0ee70f0-8158-4fec-bae8-7cd72bdef461\") " pod="metallb-system/speaker-v229x" Dec 13 07:06:41 crc kubenswrapper[4971]: I1213 07:06:41.145853 4971 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/f093984d-c941-4805-a0c1-0926aa7d7f38-cert\") pod \"frr-k8s-webhook-server-7784b6fcf-tkg6z\" (UID: \"f093984d-c941-4805-a0c1-0926aa7d7f38\") " pod="metallb-system/frr-k8s-webhook-server-7784b6fcf-tkg6z" Dec 13 07:06:41 crc kubenswrapper[4971]: I1213 07:06:41.145881 4971 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/7cc4bd2b-7f1a-46d0-bae1-dcd35dc513ad-cert\") pod \"controller-5bddd4b946-2kfxb\" (UID: \"7cc4bd2b-7f1a-46d0-bae1-dcd35dc513ad\") " pod="metallb-system/controller-5bddd4b946-2kfxb" Dec 13 07:06:41 crc kubenswrapper[4971]: I1213 07:06:41.145902 4971 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-z8kml\" (UniqueName: \"kubernetes.io/projected/c0ee70f0-8158-4fec-bae8-7cd72bdef461-kube-api-access-z8kml\") pod \"speaker-v229x\" (UID: \"c0ee70f0-8158-4fec-bae8-7cd72bdef461\") " pod="metallb-system/speaker-v229x" Dec 13 07:06:41 crc kubenswrapper[4971]: I1213 07:06:41.145922 4971 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"frr-conf\" (UniqueName: \"kubernetes.io/empty-dir/1f567f13-d990-4035-92e4-64835af4aac8-frr-conf\") pod \"frr-k8s-szzcz\" (UID: \"1f567f13-d990-4035-92e4-64835af4aac8\") " pod="metallb-system/frr-k8s-szzcz" Dec 13 07:06:41 crc kubenswrapper[4971]: I1213 07:06:41.145952 4971 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/c0ee70f0-8158-4fec-bae8-7cd72bdef461-metrics-certs\") pod \"speaker-v229x\" (UID: \"c0ee70f0-8158-4fec-bae8-7cd72bdef461\") " pod="metallb-system/speaker-v229x" Dec 13 07:06:41 crc kubenswrapper[4971]: I1213 07:06:41.145975 4971 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"frr-sockets\" (UniqueName: \"kubernetes.io/empty-dir/1f567f13-d990-4035-92e4-64835af4aac8-frr-sockets\") pod \"frr-k8s-szzcz\" (UID: \"1f567f13-d990-4035-92e4-64835af4aac8\") " pod="metallb-system/frr-k8s-szzcz" Dec 13 07:06:41 crc kubenswrapper[4971]: I1213 07:06:41.145997 4971 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metrics\" (UniqueName: \"kubernetes.io/empty-dir/1f567f13-d990-4035-92e4-64835af4aac8-metrics\") pod \"frr-k8s-szzcz\" (UID: \"1f567f13-d990-4035-92e4-64835af4aac8\") " pod="metallb-system/frr-k8s-szzcz" Dec 13 07:06:41 crc kubenswrapper[4971]: I1213 07:06:41.146022 4971 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-fvpzv\" (UniqueName: \"kubernetes.io/projected/7cc4bd2b-7f1a-46d0-bae1-dcd35dc513ad-kube-api-access-fvpzv\") pod \"controller-5bddd4b946-2kfxb\" (UID: \"7cc4bd2b-7f1a-46d0-bae1-dcd35dc513ad\") " pod="metallb-system/controller-5bddd4b946-2kfxb" Dec 13 07:06:41 crc kubenswrapper[4971]: I1213 07:06:41.146075 4971 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/7cc4bd2b-7f1a-46d0-bae1-dcd35dc513ad-metrics-certs\") pod \"controller-5bddd4b946-2kfxb\" (UID: \"7cc4bd2b-7f1a-46d0-bae1-dcd35dc513ad\") " pod="metallb-system/controller-5bddd4b946-2kfxb" Dec 13 07:06:41 crc kubenswrapper[4971]: I1213 07:06:41.146136 4971 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"reloader\" (UniqueName: \"kubernetes.io/empty-dir/1f567f13-d990-4035-92e4-64835af4aac8-reloader\") pod \"frr-k8s-szzcz\" (UID: \"1f567f13-d990-4035-92e4-64835af4aac8\") " pod="metallb-system/frr-k8s-szzcz" Dec 13 07:06:41 crc kubenswrapper[4971]: I1213 07:06:41.157060 4971 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cert\" (UniqueName: \"kubernetes.io/secret/f093984d-c941-4805-a0c1-0926aa7d7f38-cert\") pod \"frr-k8s-webhook-server-7784b6fcf-tkg6z\" (UID: \"f093984d-c941-4805-a0c1-0926aa7d7f38\") " pod="metallb-system/frr-k8s-webhook-server-7784b6fcf-tkg6z" Dec 13 07:06:41 crc kubenswrapper[4971]: I1213 07:06:41.164855 4971 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-45zsl\" (UniqueName: \"kubernetes.io/projected/f093984d-c941-4805-a0c1-0926aa7d7f38-kube-api-access-45zsl\") pod \"frr-k8s-webhook-server-7784b6fcf-tkg6z\" (UID: \"f093984d-c941-4805-a0c1-0926aa7d7f38\") " pod="metallb-system/frr-k8s-webhook-server-7784b6fcf-tkg6z" Dec 13 07:06:41 crc kubenswrapper[4971]: I1213 07:06:41.239994 4971 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="metallb-system/frr-k8s-webhook-server-7784b6fcf-tkg6z" Dec 13 07:06:41 crc kubenswrapper[4971]: I1213 07:06:41.246685 4971 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/c0ee70f0-8158-4fec-bae8-7cd72bdef461-metrics-certs\") pod \"speaker-v229x\" (UID: \"c0ee70f0-8158-4fec-bae8-7cd72bdef461\") " pod="metallb-system/speaker-v229x" Dec 13 07:06:41 crc kubenswrapper[4971]: I1213 07:06:41.247239 4971 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"frr-sockets\" (UniqueName: \"kubernetes.io/empty-dir/1f567f13-d990-4035-92e4-64835af4aac8-frr-sockets\") pod \"frr-k8s-szzcz\" (UID: \"1f567f13-d990-4035-92e4-64835af4aac8\") " pod="metallb-system/frr-k8s-szzcz" Dec 13 07:06:41 crc kubenswrapper[4971]: I1213 07:06:41.247383 4971 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics\" (UniqueName: \"kubernetes.io/empty-dir/1f567f13-d990-4035-92e4-64835af4aac8-metrics\") pod \"frr-k8s-szzcz\" (UID: \"1f567f13-d990-4035-92e4-64835af4aac8\") " pod="metallb-system/frr-k8s-szzcz" Dec 13 07:06:41 crc kubenswrapper[4971]: I1213 07:06:41.247498 4971 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-fvpzv\" (UniqueName: \"kubernetes.io/projected/7cc4bd2b-7f1a-46d0-bae1-dcd35dc513ad-kube-api-access-fvpzv\") pod \"controller-5bddd4b946-2kfxb\" (UID: \"7cc4bd2b-7f1a-46d0-bae1-dcd35dc513ad\") " pod="metallb-system/controller-5bddd4b946-2kfxb" Dec 13 07:06:41 crc kubenswrapper[4971]: I1213 07:06:41.247645 4971 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/7cc4bd2b-7f1a-46d0-bae1-dcd35dc513ad-metrics-certs\") pod \"controller-5bddd4b946-2kfxb\" (UID: \"7cc4bd2b-7f1a-46d0-bae1-dcd35dc513ad\") " pod="metallb-system/controller-5bddd4b946-2kfxb" Dec 13 07:06:41 crc kubenswrapper[4971]: I1213 07:06:41.247722 4971 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"frr-sockets\" (UniqueName: \"kubernetes.io/empty-dir/1f567f13-d990-4035-92e4-64835af4aac8-frr-sockets\") pod \"frr-k8s-szzcz\" (UID: \"1f567f13-d990-4035-92e4-64835af4aac8\") " pod="metallb-system/frr-k8s-szzcz" Dec 13 07:06:41 crc kubenswrapper[4971]: E1213 07:06:41.247059 4971 secret.go:188] Couldn't get secret metallb-system/speaker-certs-secret: secret "speaker-certs-secret" not found Dec 13 07:06:41 crc kubenswrapper[4971]: I1213 07:06:41.247815 4971 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metrics\" (UniqueName: \"kubernetes.io/empty-dir/1f567f13-d990-4035-92e4-64835af4aac8-metrics\") pod \"frr-k8s-szzcz\" (UID: \"1f567f13-d990-4035-92e4-64835af4aac8\") " pod="metallb-system/frr-k8s-szzcz" Dec 13 07:06:41 crc kubenswrapper[4971]: I1213 07:06:41.247919 4971 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"reloader\" (UniqueName: \"kubernetes.io/empty-dir/1f567f13-d990-4035-92e4-64835af4aac8-reloader\") pod \"frr-k8s-szzcz\" (UID: \"1f567f13-d990-4035-92e4-64835af4aac8\") " pod="metallb-system/frr-k8s-szzcz" Dec 13 07:06:41 crc kubenswrapper[4971]: I1213 07:06:41.248020 4971 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-jzrst\" (UniqueName: \"kubernetes.io/projected/1f567f13-d990-4035-92e4-64835af4aac8-kube-api-access-jzrst\") pod \"frr-k8s-szzcz\" (UID: \"1f567f13-d990-4035-92e4-64835af4aac8\") " pod="metallb-system/frr-k8s-szzcz" Dec 13 07:06:41 crc kubenswrapper[4971]: I1213 07:06:41.248114 4971 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"frr-startup\" (UniqueName: \"kubernetes.io/configmap/1f567f13-d990-4035-92e4-64835af4aac8-frr-startup\") pod \"frr-k8s-szzcz\" (UID: \"1f567f13-d990-4035-92e4-64835af4aac8\") " pod="metallb-system/frr-k8s-szzcz" Dec 13 07:06:41 crc kubenswrapper[4971]: E1213 07:06:41.248146 4971 secret.go:188] Couldn't get secret metallb-system/controller-certs-secret: secret "controller-certs-secret" not found Dec 13 07:06:41 crc kubenswrapper[4971]: I1213 07:06:41.248610 4971 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"reloader\" (UniqueName: \"kubernetes.io/empty-dir/1f567f13-d990-4035-92e4-64835af4aac8-reloader\") pod \"frr-k8s-szzcz\" (UID: \"1f567f13-d990-4035-92e4-64835af4aac8\") " pod="metallb-system/frr-k8s-szzcz" Dec 13 07:06:41 crc kubenswrapper[4971]: E1213 07:06:41.248366 4971 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/c0ee70f0-8158-4fec-bae8-7cd72bdef461-metrics-certs podName:c0ee70f0-8158-4fec-bae8-7cd72bdef461 nodeName:}" failed. No retries permitted until 2025-12-13 07:06:41.748346612 +0000 UTC m=+1058.352756060 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "metrics-certs" (UniqueName: "kubernetes.io/secret/c0ee70f0-8158-4fec-bae8-7cd72bdef461-metrics-certs") pod "speaker-v229x" (UID: "c0ee70f0-8158-4fec-bae8-7cd72bdef461") : secret "speaker-certs-secret" not found Dec 13 07:06:41 crc kubenswrapper[4971]: I1213 07:06:41.248996 4971 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/1f567f13-d990-4035-92e4-64835af4aac8-metrics-certs\") pod \"frr-k8s-szzcz\" (UID: \"1f567f13-d990-4035-92e4-64835af4aac8\") " pod="metallb-system/frr-k8s-szzcz" Dec 13 07:06:41 crc kubenswrapper[4971]: I1213 07:06:41.249076 4971 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metallb-excludel2\" (UniqueName: \"kubernetes.io/configmap/c0ee70f0-8158-4fec-bae8-7cd72bdef461-metallb-excludel2\") pod \"speaker-v229x\" (UID: \"c0ee70f0-8158-4fec-bae8-7cd72bdef461\") " pod="metallb-system/speaker-v229x" Dec 13 07:06:41 crc kubenswrapper[4971]: I1213 07:06:41.249171 4971 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"memberlist\" (UniqueName: \"kubernetes.io/secret/c0ee70f0-8158-4fec-bae8-7cd72bdef461-memberlist\") pod \"speaker-v229x\" (UID: \"c0ee70f0-8158-4fec-bae8-7cd72bdef461\") " pod="metallb-system/speaker-v229x" Dec 13 07:06:41 crc kubenswrapper[4971]: I1213 07:06:41.249267 4971 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/7cc4bd2b-7f1a-46d0-bae1-dcd35dc513ad-cert\") pod \"controller-5bddd4b946-2kfxb\" (UID: \"7cc4bd2b-7f1a-46d0-bae1-dcd35dc513ad\") " pod="metallb-system/controller-5bddd4b946-2kfxb" Dec 13 07:06:41 crc kubenswrapper[4971]: I1213 07:06:41.249336 4971 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-z8kml\" (UniqueName: \"kubernetes.io/projected/c0ee70f0-8158-4fec-bae8-7cd72bdef461-kube-api-access-z8kml\") pod \"speaker-v229x\" (UID: \"c0ee70f0-8158-4fec-bae8-7cd72bdef461\") " pod="metallb-system/speaker-v229x" Dec 13 07:06:41 crc kubenswrapper[4971]: I1213 07:06:41.249416 4971 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"frr-conf\" (UniqueName: \"kubernetes.io/empty-dir/1f567f13-d990-4035-92e4-64835af4aac8-frr-conf\") pod \"frr-k8s-szzcz\" (UID: \"1f567f13-d990-4035-92e4-64835af4aac8\") " pod="metallb-system/frr-k8s-szzcz" Dec 13 07:06:41 crc kubenswrapper[4971]: I1213 07:06:41.249478 4971 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"frr-startup\" (UniqueName: \"kubernetes.io/configmap/1f567f13-d990-4035-92e4-64835af4aac8-frr-startup\") pod \"frr-k8s-szzcz\" (UID: \"1f567f13-d990-4035-92e4-64835af4aac8\") " pod="metallb-system/frr-k8s-szzcz" Dec 13 07:06:41 crc kubenswrapper[4971]: E1213 07:06:41.249697 4971 secret.go:188] Couldn't get secret metallb-system/metallb-memberlist: secret "metallb-memberlist" not found Dec 13 07:06:41 crc kubenswrapper[4971]: E1213 07:06:41.249787 4971 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/c0ee70f0-8158-4fec-bae8-7cd72bdef461-memberlist podName:c0ee70f0-8158-4fec-bae8-7cd72bdef461 nodeName:}" failed. No retries permitted until 2025-12-13 07:06:41.749770367 +0000 UTC m=+1058.354179815 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "memberlist" (UniqueName: "kubernetes.io/secret/c0ee70f0-8158-4fec-bae8-7cd72bdef461-memberlist") pod "speaker-v229x" (UID: "c0ee70f0-8158-4fec-bae8-7cd72bdef461") : secret "metallb-memberlist" not found Dec 13 07:06:41 crc kubenswrapper[4971]: I1213 07:06:41.250130 4971 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metallb-excludel2\" (UniqueName: \"kubernetes.io/configmap/c0ee70f0-8158-4fec-bae8-7cd72bdef461-metallb-excludel2\") pod \"speaker-v229x\" (UID: \"c0ee70f0-8158-4fec-bae8-7cd72bdef461\") " pod="metallb-system/speaker-v229x" Dec 13 07:06:41 crc kubenswrapper[4971]: I1213 07:06:41.250491 4971 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"frr-conf\" (UniqueName: \"kubernetes.io/empty-dir/1f567f13-d990-4035-92e4-64835af4aac8-frr-conf\") pod \"frr-k8s-szzcz\" (UID: \"1f567f13-d990-4035-92e4-64835af4aac8\") " pod="metallb-system/frr-k8s-szzcz" Dec 13 07:06:41 crc kubenswrapper[4971]: E1213 07:06:41.250753 4971 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/7cc4bd2b-7f1a-46d0-bae1-dcd35dc513ad-metrics-certs podName:7cc4bd2b-7f1a-46d0-bae1-dcd35dc513ad nodeName:}" failed. No retries permitted until 2025-12-13 07:06:41.750714559 +0000 UTC m=+1058.355124007 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "metrics-certs" (UniqueName: "kubernetes.io/secret/7cc4bd2b-7f1a-46d0-bae1-dcd35dc513ad-metrics-certs") pod "controller-5bddd4b946-2kfxb" (UID: "7cc4bd2b-7f1a-46d0-bae1-dcd35dc513ad") : secret "controller-certs-secret" not found Dec 13 07:06:41 crc kubenswrapper[4971]: I1213 07:06:41.253555 4971 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/1f567f13-d990-4035-92e4-64835af4aac8-metrics-certs\") pod \"frr-k8s-szzcz\" (UID: \"1f567f13-d990-4035-92e4-64835af4aac8\") " pod="metallb-system/frr-k8s-szzcz" Dec 13 07:06:41 crc kubenswrapper[4971]: I1213 07:06:41.254834 4971 reflector.go:368] Caches populated for *v1.Secret from object-"metallb-system"/"metallb-webhook-cert" Dec 13 07:06:41 crc kubenswrapper[4971]: I1213 07:06:41.263506 4971 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-jzrst\" (UniqueName: \"kubernetes.io/projected/1f567f13-d990-4035-92e4-64835af4aac8-kube-api-access-jzrst\") pod \"frr-k8s-szzcz\" (UID: \"1f567f13-d990-4035-92e4-64835af4aac8\") " pod="metallb-system/frr-k8s-szzcz" Dec 13 07:06:41 crc kubenswrapper[4971]: I1213 07:06:41.267906 4971 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-fvpzv\" (UniqueName: \"kubernetes.io/projected/7cc4bd2b-7f1a-46d0-bae1-dcd35dc513ad-kube-api-access-fvpzv\") pod \"controller-5bddd4b946-2kfxb\" (UID: \"7cc4bd2b-7f1a-46d0-bae1-dcd35dc513ad\") " pod="metallb-system/controller-5bddd4b946-2kfxb" Dec 13 07:06:41 crc kubenswrapper[4971]: I1213 07:06:41.270367 4971 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cert\" (UniqueName: \"kubernetes.io/secret/7cc4bd2b-7f1a-46d0-bae1-dcd35dc513ad-cert\") pod \"controller-5bddd4b946-2kfxb\" (UID: \"7cc4bd2b-7f1a-46d0-bae1-dcd35dc513ad\") " pod="metallb-system/controller-5bddd4b946-2kfxb" Dec 13 07:06:41 crc kubenswrapper[4971]: I1213 07:06:41.273710 4971 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="metallb-system/frr-k8s-szzcz" Dec 13 07:06:41 crc kubenswrapper[4971]: I1213 07:06:41.274501 4971 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-z8kml\" (UniqueName: \"kubernetes.io/projected/c0ee70f0-8158-4fec-bae8-7cd72bdef461-kube-api-access-z8kml\") pod \"speaker-v229x\" (UID: \"c0ee70f0-8158-4fec-bae8-7cd72bdef461\") " pod="metallb-system/speaker-v229x" Dec 13 07:06:41 crc kubenswrapper[4971]: I1213 07:06:41.490306 4971 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/frr-k8s-szzcz" event={"ID":"1f567f13-d990-4035-92e4-64835af4aac8","Type":"ContainerStarted","Data":"33a191d10164d7bfc1550c966ac17a7f8e384d1387af7493f63cde5e2ac39130"} Dec 13 07:06:41 crc kubenswrapper[4971]: I1213 07:06:41.724709 4971 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["metallb-system/frr-k8s-webhook-server-7784b6fcf-tkg6z"] Dec 13 07:06:41 crc kubenswrapper[4971]: W1213 07:06:41.726902 4971 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podf093984d_c941_4805_a0c1_0926aa7d7f38.slice/crio-f78506b85771be6acdd80f2a5ba8a9c85dfc0a9b316f7ba32123581543a3f263 WatchSource:0}: Error finding container f78506b85771be6acdd80f2a5ba8a9c85dfc0a9b316f7ba32123581543a3f263: Status 404 returned error can't find the container with id f78506b85771be6acdd80f2a5ba8a9c85dfc0a9b316f7ba32123581543a3f263 Dec 13 07:06:41 crc kubenswrapper[4971]: I1213 07:06:41.754422 4971 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"memberlist\" (UniqueName: \"kubernetes.io/secret/c0ee70f0-8158-4fec-bae8-7cd72bdef461-memberlist\") pod \"speaker-v229x\" (UID: \"c0ee70f0-8158-4fec-bae8-7cd72bdef461\") " pod="metallb-system/speaker-v229x" Dec 13 07:06:41 crc kubenswrapper[4971]: I1213 07:06:41.754489 4971 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/c0ee70f0-8158-4fec-bae8-7cd72bdef461-metrics-certs\") pod \"speaker-v229x\" (UID: \"c0ee70f0-8158-4fec-bae8-7cd72bdef461\") " pod="metallb-system/speaker-v229x" Dec 13 07:06:41 crc kubenswrapper[4971]: I1213 07:06:41.754533 4971 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/7cc4bd2b-7f1a-46d0-bae1-dcd35dc513ad-metrics-certs\") pod \"controller-5bddd4b946-2kfxb\" (UID: \"7cc4bd2b-7f1a-46d0-bae1-dcd35dc513ad\") " pod="metallb-system/controller-5bddd4b946-2kfxb" Dec 13 07:06:41 crc kubenswrapper[4971]: E1213 07:06:41.754786 4971 secret.go:188] Couldn't get secret metallb-system/metallb-memberlist: secret "metallb-memberlist" not found Dec 13 07:06:41 crc kubenswrapper[4971]: E1213 07:06:41.754948 4971 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/c0ee70f0-8158-4fec-bae8-7cd72bdef461-memberlist podName:c0ee70f0-8158-4fec-bae8-7cd72bdef461 nodeName:}" failed. No retries permitted until 2025-12-13 07:06:42.754914744 +0000 UTC m=+1059.359324202 (durationBeforeRetry 1s). Error: MountVolume.SetUp failed for volume "memberlist" (UniqueName: "kubernetes.io/secret/c0ee70f0-8158-4fec-bae8-7cd72bdef461-memberlist") pod "speaker-v229x" (UID: "c0ee70f0-8158-4fec-bae8-7cd72bdef461") : secret "metallb-memberlist" not found Dec 13 07:06:41 crc kubenswrapper[4971]: I1213 07:06:41.760148 4971 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/c0ee70f0-8158-4fec-bae8-7cd72bdef461-metrics-certs\") pod \"speaker-v229x\" (UID: \"c0ee70f0-8158-4fec-bae8-7cd72bdef461\") " pod="metallb-system/speaker-v229x" Dec 13 07:06:41 crc kubenswrapper[4971]: I1213 07:06:41.760389 4971 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/7cc4bd2b-7f1a-46d0-bae1-dcd35dc513ad-metrics-certs\") pod \"controller-5bddd4b946-2kfxb\" (UID: \"7cc4bd2b-7f1a-46d0-bae1-dcd35dc513ad\") " pod="metallb-system/controller-5bddd4b946-2kfxb" Dec 13 07:06:41 crc kubenswrapper[4971]: I1213 07:06:41.990040 4971 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="metallb-system/controller-5bddd4b946-2kfxb" Dec 13 07:06:42 crc kubenswrapper[4971]: I1213 07:06:42.421684 4971 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["metallb-system/controller-5bddd4b946-2kfxb"] Dec 13 07:06:42 crc kubenswrapper[4971]: I1213 07:06:42.501808 4971 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/frr-k8s-webhook-server-7784b6fcf-tkg6z" event={"ID":"f093984d-c941-4805-a0c1-0926aa7d7f38","Type":"ContainerStarted","Data":"f78506b85771be6acdd80f2a5ba8a9c85dfc0a9b316f7ba32123581543a3f263"} Dec 13 07:06:42 crc kubenswrapper[4971]: I1213 07:06:42.503252 4971 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/controller-5bddd4b946-2kfxb" event={"ID":"7cc4bd2b-7f1a-46d0-bae1-dcd35dc513ad","Type":"ContainerStarted","Data":"339479f3671ad1632f37de5301aa52f18c3d1cc815cbdf91ca2630ac65cd3557"} Dec 13 07:06:42 crc kubenswrapper[4971]: I1213 07:06:42.785144 4971 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"memberlist\" (UniqueName: \"kubernetes.io/secret/c0ee70f0-8158-4fec-bae8-7cd72bdef461-memberlist\") pod \"speaker-v229x\" (UID: \"c0ee70f0-8158-4fec-bae8-7cd72bdef461\") " pod="metallb-system/speaker-v229x" Dec 13 07:06:42 crc kubenswrapper[4971]: I1213 07:06:42.800680 4971 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"memberlist\" (UniqueName: \"kubernetes.io/secret/c0ee70f0-8158-4fec-bae8-7cd72bdef461-memberlist\") pod \"speaker-v229x\" (UID: \"c0ee70f0-8158-4fec-bae8-7cd72bdef461\") " pod="metallb-system/speaker-v229x" Dec 13 07:06:42 crc kubenswrapper[4971]: I1213 07:06:42.830888 4971 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="metallb-system/speaker-v229x" Dec 13 07:06:43 crc kubenswrapper[4971]: I1213 07:06:43.517893 4971 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/speaker-v229x" event={"ID":"c0ee70f0-8158-4fec-bae8-7cd72bdef461","Type":"ContainerStarted","Data":"660c0402f92d5685a06d271caeb1432631f1ead59fe99ff3ba0fc5571d997eca"} Dec 13 07:06:43 crc kubenswrapper[4971]: I1213 07:06:43.518048 4971 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/speaker-v229x" event={"ID":"c0ee70f0-8158-4fec-bae8-7cd72bdef461","Type":"ContainerStarted","Data":"a69504010a722c605f687297dca4d7e7569d2944f327e1691b5507c5f13a6641"} Dec 13 07:06:43 crc kubenswrapper[4971]: I1213 07:06:43.524726 4971 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/controller-5bddd4b946-2kfxb" event={"ID":"7cc4bd2b-7f1a-46d0-bae1-dcd35dc513ad","Type":"ContainerStarted","Data":"dbbb3981b8f1cfaee9872f141322f93400a2331d88f2bf8dbd914a98d52740c9"} Dec 13 07:06:43 crc kubenswrapper[4971]: I1213 07:06:43.524778 4971 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/controller-5bddd4b946-2kfxb" event={"ID":"7cc4bd2b-7f1a-46d0-bae1-dcd35dc513ad","Type":"ContainerStarted","Data":"a0b9ff8eada383f4a7bd610405b0dbb8b8573481ae4f061e7d8ed3a79ec99b48"} Dec 13 07:06:43 crc kubenswrapper[4971]: I1213 07:06:43.525767 4971 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="metallb-system/controller-5bddd4b946-2kfxb" Dec 13 07:06:43 crc kubenswrapper[4971]: I1213 07:06:43.549594 4971 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="metallb-system/controller-5bddd4b946-2kfxb" podStartSLOduration=2.549568052 podStartE2EDuration="2.549568052s" podCreationTimestamp="2025-12-13 07:06:41 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-13 07:06:43.543272898 +0000 UTC m=+1060.147682346" watchObservedRunningTime="2025-12-13 07:06:43.549568052 +0000 UTC m=+1060.153977520" Dec 13 07:06:44 crc kubenswrapper[4971]: I1213 07:06:44.542799 4971 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/speaker-v229x" event={"ID":"c0ee70f0-8158-4fec-bae8-7cd72bdef461","Type":"ContainerStarted","Data":"295683e6edfad72e5b51f8d5462b1851e5780d8fda6189784563af8a9c17d2aa"} Dec 13 07:06:44 crc kubenswrapper[4971]: I1213 07:06:44.543208 4971 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="metallb-system/speaker-v229x" Dec 13 07:06:44 crc kubenswrapper[4971]: I1213 07:06:44.569182 4971 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="metallb-system/speaker-v229x" podStartSLOduration=4.569157436 podStartE2EDuration="4.569157436s" podCreationTimestamp="2025-12-13 07:06:40 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-13 07:06:44.565772933 +0000 UTC m=+1061.170182391" watchObservedRunningTime="2025-12-13 07:06:44.569157436 +0000 UTC m=+1061.173566894" Dec 13 07:06:50 crc kubenswrapper[4971]: I1213 07:06:50.635555 4971 generic.go:334] "Generic (PLEG): container finished" podID="1f567f13-d990-4035-92e4-64835af4aac8" containerID="871e1b58ab540abf37bb033c0f681f375b524f375362b55ac3b63b284ae208f4" exitCode=0 Dec 13 07:06:50 crc kubenswrapper[4971]: I1213 07:06:50.635832 4971 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/frr-k8s-szzcz" event={"ID":"1f567f13-d990-4035-92e4-64835af4aac8","Type":"ContainerDied","Data":"871e1b58ab540abf37bb033c0f681f375b524f375362b55ac3b63b284ae208f4"} Dec 13 07:06:50 crc kubenswrapper[4971]: I1213 07:06:50.639388 4971 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/frr-k8s-webhook-server-7784b6fcf-tkg6z" event={"ID":"f093984d-c941-4805-a0c1-0926aa7d7f38","Type":"ContainerStarted","Data":"a8d6d0974ccde56ea91a5faa08f28955444fc3b2700124a2d8290ee03bd9b5da"} Dec 13 07:06:50 crc kubenswrapper[4971]: I1213 07:06:50.639645 4971 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="metallb-system/frr-k8s-webhook-server-7784b6fcf-tkg6z" Dec 13 07:06:50 crc kubenswrapper[4971]: I1213 07:06:50.684949 4971 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="metallb-system/frr-k8s-webhook-server-7784b6fcf-tkg6z" podStartSLOduration=2.532616633 podStartE2EDuration="10.684931831s" podCreationTimestamp="2025-12-13 07:06:40 +0000 UTC" firstStartedPulling="2025-12-13 07:06:41.728642529 +0000 UTC m=+1058.333051967" lastFinishedPulling="2025-12-13 07:06:49.880957717 +0000 UTC m=+1066.485367165" observedRunningTime="2025-12-13 07:06:50.683174028 +0000 UTC m=+1067.287583476" watchObservedRunningTime="2025-12-13 07:06:50.684931831 +0000 UTC m=+1067.289341279" Dec 13 07:06:51 crc kubenswrapper[4971]: I1213 07:06:51.647754 4971 generic.go:334] "Generic (PLEG): container finished" podID="1f567f13-d990-4035-92e4-64835af4aac8" containerID="5ab9bba189ba290e0fe89957d053a92aba2f06c6afe3384dfd8b9cabab85ce19" exitCode=0 Dec 13 07:06:51 crc kubenswrapper[4971]: I1213 07:06:51.647832 4971 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/frr-k8s-szzcz" event={"ID":"1f567f13-d990-4035-92e4-64835af4aac8","Type":"ContainerDied","Data":"5ab9bba189ba290e0fe89957d053a92aba2f06c6afe3384dfd8b9cabab85ce19"} Dec 13 07:06:52 crc kubenswrapper[4971]: I1213 07:06:52.656967 4971 generic.go:334] "Generic (PLEG): container finished" podID="1f567f13-d990-4035-92e4-64835af4aac8" containerID="bc04bc9f04a6ebff6c9f4932421a9b22382a7fe2d28048fa927418ddb2ccd955" exitCode=0 Dec 13 07:06:52 crc kubenswrapper[4971]: I1213 07:06:52.657009 4971 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/frr-k8s-szzcz" event={"ID":"1f567f13-d990-4035-92e4-64835af4aac8","Type":"ContainerDied","Data":"bc04bc9f04a6ebff6c9f4932421a9b22382a7fe2d28048fa927418ddb2ccd955"} Dec 13 07:06:53 crc kubenswrapper[4971]: I1213 07:06:53.668692 4971 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/frr-k8s-szzcz" event={"ID":"1f567f13-d990-4035-92e4-64835af4aac8","Type":"ContainerStarted","Data":"5aea7f07b5044265ac5baabae814b4b0a7456b483be1f0d92e33a2e095027a38"} Dec 13 07:06:53 crc kubenswrapper[4971]: I1213 07:06:53.670097 4971 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="metallb-system/frr-k8s-szzcz" Dec 13 07:06:53 crc kubenswrapper[4971]: I1213 07:06:53.670228 4971 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/frr-k8s-szzcz" event={"ID":"1f567f13-d990-4035-92e4-64835af4aac8","Type":"ContainerStarted","Data":"a8f85315ed0b87fb2dac0cb5e543461431113903edcda3e3c8c7ec27726047f3"} Dec 13 07:06:53 crc kubenswrapper[4971]: I1213 07:06:53.670353 4971 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/frr-k8s-szzcz" event={"ID":"1f567f13-d990-4035-92e4-64835af4aac8","Type":"ContainerStarted","Data":"0c77b8850fe0824985e9c019347b00225af79a5f07228e9096e5f0fb0c436e85"} Dec 13 07:06:53 crc kubenswrapper[4971]: I1213 07:06:53.670429 4971 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/frr-k8s-szzcz" event={"ID":"1f567f13-d990-4035-92e4-64835af4aac8","Type":"ContainerStarted","Data":"362776161b0197882f3b6e8ae8d9401babb4f77cb285bb58926b478de4cbca79"} Dec 13 07:06:53 crc kubenswrapper[4971]: I1213 07:06:53.670511 4971 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/frr-k8s-szzcz" event={"ID":"1f567f13-d990-4035-92e4-64835af4aac8","Type":"ContainerStarted","Data":"d5a05a0bf02148187b9be38f709bc8db7187af05e5d5f0ab7f9f08b13db49729"} Dec 13 07:06:53 crc kubenswrapper[4971]: I1213 07:06:53.670617 4971 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/frr-k8s-szzcz" event={"ID":"1f567f13-d990-4035-92e4-64835af4aac8","Type":"ContainerStarted","Data":"0dbf02bb03be06a590d02aec6a21384c194afe3907fd273cc08235b6e85dd05b"} Dec 13 07:06:53 crc kubenswrapper[4971]: I1213 07:06:53.707066 4971 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="metallb-system/frr-k8s-szzcz" podStartSLOduration=5.263569136 podStartE2EDuration="13.707042653s" podCreationTimestamp="2025-12-13 07:06:40 +0000 UTC" firstStartedPulling="2025-12-13 07:06:41.429962506 +0000 UTC m=+1058.034371954" lastFinishedPulling="2025-12-13 07:06:49.873436033 +0000 UTC m=+1066.477845471" observedRunningTime="2025-12-13 07:06:53.701567539 +0000 UTC m=+1070.305977007" watchObservedRunningTime="2025-12-13 07:06:53.707042653 +0000 UTC m=+1070.311452111" Dec 13 07:06:56 crc kubenswrapper[4971]: I1213 07:06:56.274841 4971 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="metallb-system/frr-k8s-szzcz" Dec 13 07:06:56 crc kubenswrapper[4971]: I1213 07:06:56.320130 4971 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="metallb-system/frr-k8s-szzcz" Dec 13 07:07:01 crc kubenswrapper[4971]: I1213 07:07:01.253663 4971 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="metallb-system/frr-k8s-webhook-server-7784b6fcf-tkg6z" Dec 13 07:07:01 crc kubenswrapper[4971]: I1213 07:07:01.994715 4971 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="metallb-system/controller-5bddd4b946-2kfxb" Dec 13 07:07:02 crc kubenswrapper[4971]: I1213 07:07:02.833358 4971 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="metallb-system/speaker-v229x" Dec 13 07:07:05 crc kubenswrapper[4971]: I1213 07:07:05.653703 4971 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/openstack-operator-index-zdlv5"] Dec 13 07:07:05 crc kubenswrapper[4971]: I1213 07:07:05.654967 4971 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/openstack-operator-index-zdlv5" Dec 13 07:07:05 crc kubenswrapper[4971]: I1213 07:07:05.657079 4971 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack-operators"/"kube-root-ca.crt" Dec 13 07:07:05 crc kubenswrapper[4971]: I1213 07:07:05.657646 4971 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack-operators"/"openshift-service-ca.crt" Dec 13 07:07:05 crc kubenswrapper[4971]: I1213 07:07:05.658113 4971 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"openstack-operator-index-dockercfg-84mcp" Dec 13 07:07:05 crc kubenswrapper[4971]: I1213 07:07:05.663627 4971 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/openstack-operator-index-zdlv5"] Dec 13 07:07:05 crc kubenswrapper[4971]: I1213 07:07:05.811367 4971 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-fv76g\" (UniqueName: \"kubernetes.io/projected/d1cfff96-327c-4048-9bba-649a7345108c-kube-api-access-fv76g\") pod \"openstack-operator-index-zdlv5\" (UID: \"d1cfff96-327c-4048-9bba-649a7345108c\") " pod="openstack-operators/openstack-operator-index-zdlv5" Dec 13 07:07:05 crc kubenswrapper[4971]: I1213 07:07:05.912127 4971 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-fv76g\" (UniqueName: \"kubernetes.io/projected/d1cfff96-327c-4048-9bba-649a7345108c-kube-api-access-fv76g\") pod \"openstack-operator-index-zdlv5\" (UID: \"d1cfff96-327c-4048-9bba-649a7345108c\") " pod="openstack-operators/openstack-operator-index-zdlv5" Dec 13 07:07:05 crc kubenswrapper[4971]: I1213 07:07:05.930819 4971 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-fv76g\" (UniqueName: \"kubernetes.io/projected/d1cfff96-327c-4048-9bba-649a7345108c-kube-api-access-fv76g\") pod \"openstack-operator-index-zdlv5\" (UID: \"d1cfff96-327c-4048-9bba-649a7345108c\") " pod="openstack-operators/openstack-operator-index-zdlv5" Dec 13 07:07:05 crc kubenswrapper[4971]: I1213 07:07:05.972831 4971 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/openstack-operator-index-zdlv5" Dec 13 07:07:06 crc kubenswrapper[4971]: I1213 07:07:06.450347 4971 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/openstack-operator-index-zdlv5"] Dec 13 07:07:06 crc kubenswrapper[4971]: I1213 07:07:06.750439 4971 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/openstack-operator-index-zdlv5" event={"ID":"d1cfff96-327c-4048-9bba-649a7345108c","Type":"ContainerStarted","Data":"47487124a04b9621ee8fe280306717b5caabd895d383a9e8f20f2ac841ce7d03"} Dec 13 07:07:08 crc kubenswrapper[4971]: I1213 07:07:08.833326 4971 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack-operators/openstack-operator-index-zdlv5"] Dec 13 07:07:09 crc kubenswrapper[4971]: I1213 07:07:09.440667 4971 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/openstack-operator-index-hbjwh"] Dec 13 07:07:09 crc kubenswrapper[4971]: I1213 07:07:09.441568 4971 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/openstack-operator-index-hbjwh" Dec 13 07:07:09 crc kubenswrapper[4971]: I1213 07:07:09.453039 4971 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/openstack-operator-index-hbjwh"] Dec 13 07:07:09 crc kubenswrapper[4971]: I1213 07:07:09.558895 4971 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-5fqk8\" (UniqueName: \"kubernetes.io/projected/1a3059b7-5a8b-4357-8f6d-caa37f9adef2-kube-api-access-5fqk8\") pod \"openstack-operator-index-hbjwh\" (UID: \"1a3059b7-5a8b-4357-8f6d-caa37f9adef2\") " pod="openstack-operators/openstack-operator-index-hbjwh" Dec 13 07:07:09 crc kubenswrapper[4971]: I1213 07:07:09.660567 4971 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-5fqk8\" (UniqueName: \"kubernetes.io/projected/1a3059b7-5a8b-4357-8f6d-caa37f9adef2-kube-api-access-5fqk8\") pod \"openstack-operator-index-hbjwh\" (UID: \"1a3059b7-5a8b-4357-8f6d-caa37f9adef2\") " pod="openstack-operators/openstack-operator-index-hbjwh" Dec 13 07:07:09 crc kubenswrapper[4971]: I1213 07:07:09.682728 4971 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-5fqk8\" (UniqueName: \"kubernetes.io/projected/1a3059b7-5a8b-4357-8f6d-caa37f9adef2-kube-api-access-5fqk8\") pod \"openstack-operator-index-hbjwh\" (UID: \"1a3059b7-5a8b-4357-8f6d-caa37f9adef2\") " pod="openstack-operators/openstack-operator-index-hbjwh" Dec 13 07:07:09 crc kubenswrapper[4971]: I1213 07:07:09.821611 4971 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/openstack-operator-index-hbjwh" Dec 13 07:07:10 crc kubenswrapper[4971]: I1213 07:07:10.031298 4971 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/openstack-operator-index-hbjwh"] Dec 13 07:07:10 crc kubenswrapper[4971]: I1213 07:07:10.773789 4971 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/openstack-operator-index-hbjwh" event={"ID":"1a3059b7-5a8b-4357-8f6d-caa37f9adef2","Type":"ContainerStarted","Data":"83276c5301f29f617f85f5739367089dc2973d4aa59afde7f52c3d055d2d12b7"} Dec 13 07:07:10 crc kubenswrapper[4971]: I1213 07:07:10.773840 4971 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/openstack-operator-index-hbjwh" event={"ID":"1a3059b7-5a8b-4357-8f6d-caa37f9adef2","Type":"ContainerStarted","Data":"df531ac1157fd277f6ec46249c0f4a028d193a67dd5ab6557e4bc85ad7bdc5e4"} Dec 13 07:07:10 crc kubenswrapper[4971]: I1213 07:07:10.776008 4971 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/openstack-operator-index-zdlv5" event={"ID":"d1cfff96-327c-4048-9bba-649a7345108c","Type":"ContainerStarted","Data":"15b60cd598a528ed939e87047dfb60d907c91fc62a8b3a26bd3f7ebe37b055f8"} Dec 13 07:07:10 crc kubenswrapper[4971]: I1213 07:07:10.776116 4971 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack-operators/openstack-operator-index-zdlv5" podUID="d1cfff96-327c-4048-9bba-649a7345108c" containerName="registry-server" containerID="cri-o://15b60cd598a528ed939e87047dfb60d907c91fc62a8b3a26bd3f7ebe37b055f8" gracePeriod=2 Dec 13 07:07:10 crc kubenswrapper[4971]: I1213 07:07:10.791463 4971 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/openstack-operator-index-hbjwh" podStartSLOduration=1.350181245 podStartE2EDuration="1.791398791s" podCreationTimestamp="2025-12-13 07:07:09 +0000 UTC" firstStartedPulling="2025-12-13 07:07:10.038958953 +0000 UTC m=+1086.643368401" lastFinishedPulling="2025-12-13 07:07:10.480176499 +0000 UTC m=+1087.084585947" observedRunningTime="2025-12-13 07:07:10.787878114 +0000 UTC m=+1087.392287562" watchObservedRunningTime="2025-12-13 07:07:10.791398791 +0000 UTC m=+1087.395808239" Dec 13 07:07:10 crc kubenswrapper[4971]: I1213 07:07:10.804982 4971 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/openstack-operator-index-zdlv5" podStartSLOduration=2.459850581 podStartE2EDuration="5.804958903s" podCreationTimestamp="2025-12-13 07:07:05 +0000 UTC" firstStartedPulling="2025-12-13 07:07:06.450859978 +0000 UTC m=+1083.055269426" lastFinishedPulling="2025-12-13 07:07:09.7959683 +0000 UTC m=+1086.400377748" observedRunningTime="2025-12-13 07:07:10.803936259 +0000 UTC m=+1087.408345727" watchObservedRunningTime="2025-12-13 07:07:10.804958903 +0000 UTC m=+1087.409368371" Dec 13 07:07:11 crc kubenswrapper[4971]: I1213 07:07:11.132735 4971 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack-operators/openstack-operator-index-zdlv5" Dec 13 07:07:11 crc kubenswrapper[4971]: I1213 07:07:11.277265 4971 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="metallb-system/frr-k8s-szzcz" Dec 13 07:07:11 crc kubenswrapper[4971]: I1213 07:07:11.283833 4971 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-fv76g\" (UniqueName: \"kubernetes.io/projected/d1cfff96-327c-4048-9bba-649a7345108c-kube-api-access-fv76g\") pod \"d1cfff96-327c-4048-9bba-649a7345108c\" (UID: \"d1cfff96-327c-4048-9bba-649a7345108c\") " Dec 13 07:07:11 crc kubenswrapper[4971]: I1213 07:07:11.289364 4971 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/d1cfff96-327c-4048-9bba-649a7345108c-kube-api-access-fv76g" (OuterVolumeSpecName: "kube-api-access-fv76g") pod "d1cfff96-327c-4048-9bba-649a7345108c" (UID: "d1cfff96-327c-4048-9bba-649a7345108c"). InnerVolumeSpecName "kube-api-access-fv76g". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 13 07:07:11 crc kubenswrapper[4971]: I1213 07:07:11.385563 4971 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-fv76g\" (UniqueName: \"kubernetes.io/projected/d1cfff96-327c-4048-9bba-649a7345108c-kube-api-access-fv76g\") on node \"crc\" DevicePath \"\"" Dec 13 07:07:11 crc kubenswrapper[4971]: I1213 07:07:11.788998 4971 generic.go:334] "Generic (PLEG): container finished" podID="d1cfff96-327c-4048-9bba-649a7345108c" containerID="15b60cd598a528ed939e87047dfb60d907c91fc62a8b3a26bd3f7ebe37b055f8" exitCode=0 Dec 13 07:07:11 crc kubenswrapper[4971]: I1213 07:07:11.789060 4971 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack-operators/openstack-operator-index-zdlv5" Dec 13 07:07:11 crc kubenswrapper[4971]: I1213 07:07:11.789479 4971 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/openstack-operator-index-zdlv5" event={"ID":"d1cfff96-327c-4048-9bba-649a7345108c","Type":"ContainerDied","Data":"15b60cd598a528ed939e87047dfb60d907c91fc62a8b3a26bd3f7ebe37b055f8"} Dec 13 07:07:11 crc kubenswrapper[4971]: I1213 07:07:11.789582 4971 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/openstack-operator-index-zdlv5" event={"ID":"d1cfff96-327c-4048-9bba-649a7345108c","Type":"ContainerDied","Data":"47487124a04b9621ee8fe280306717b5caabd895d383a9e8f20f2ac841ce7d03"} Dec 13 07:07:11 crc kubenswrapper[4971]: I1213 07:07:11.789604 4971 scope.go:117] "RemoveContainer" containerID="15b60cd598a528ed939e87047dfb60d907c91fc62a8b3a26bd3f7ebe37b055f8" Dec 13 07:07:11 crc kubenswrapper[4971]: I1213 07:07:11.818770 4971 scope.go:117] "RemoveContainer" containerID="15b60cd598a528ed939e87047dfb60d907c91fc62a8b3a26bd3f7ebe37b055f8" Dec 13 07:07:11 crc kubenswrapper[4971]: E1213 07:07:11.819795 4971 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"15b60cd598a528ed939e87047dfb60d907c91fc62a8b3a26bd3f7ebe37b055f8\": container with ID starting with 15b60cd598a528ed939e87047dfb60d907c91fc62a8b3a26bd3f7ebe37b055f8 not found: ID does not exist" containerID="15b60cd598a528ed939e87047dfb60d907c91fc62a8b3a26bd3f7ebe37b055f8" Dec 13 07:07:11 crc kubenswrapper[4971]: I1213 07:07:11.819869 4971 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"15b60cd598a528ed939e87047dfb60d907c91fc62a8b3a26bd3f7ebe37b055f8"} err="failed to get container status \"15b60cd598a528ed939e87047dfb60d907c91fc62a8b3a26bd3f7ebe37b055f8\": rpc error: code = NotFound desc = could not find container \"15b60cd598a528ed939e87047dfb60d907c91fc62a8b3a26bd3f7ebe37b055f8\": container with ID starting with 15b60cd598a528ed939e87047dfb60d907c91fc62a8b3a26bd3f7ebe37b055f8 not found: ID does not exist" Dec 13 07:07:11 crc kubenswrapper[4971]: I1213 07:07:11.824617 4971 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack-operators/openstack-operator-index-zdlv5"] Dec 13 07:07:11 crc kubenswrapper[4971]: I1213 07:07:11.828484 4971 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack-operators/openstack-operator-index-zdlv5"] Dec 13 07:07:13 crc kubenswrapper[4971]: I1213 07:07:13.778231 4971 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="d1cfff96-327c-4048-9bba-649a7345108c" path="/var/lib/kubelet/pods/d1cfff96-327c-4048-9bba-649a7345108c/volumes" Dec 13 07:07:16 crc kubenswrapper[4971]: I1213 07:07:16.154127 4971 patch_prober.go:28] interesting pod/machine-config-daemon-82xjz container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 13 07:07:16 crc kubenswrapper[4971]: I1213 07:07:16.154567 4971 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-82xjz" podUID="e0c4af12-edda-41d7-9b44-f87396174ef2" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 13 07:07:19 crc kubenswrapper[4971]: I1213 07:07:19.822451 4971 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/openstack-operator-index-hbjwh" Dec 13 07:07:19 crc kubenswrapper[4971]: I1213 07:07:19.822937 4971 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack-operators/openstack-operator-index-hbjwh" Dec 13 07:07:19 crc kubenswrapper[4971]: I1213 07:07:19.851165 4971 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack-operators/openstack-operator-index-hbjwh" Dec 13 07:07:19 crc kubenswrapper[4971]: I1213 07:07:19.877631 4971 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/openstack-operator-index-hbjwh" Dec 13 07:07:25 crc kubenswrapper[4971]: I1213 07:07:25.078368 4971 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/ea5d71784a81c95dff2031a02f0a0b3f756f86f14acad8f152d938f56fl9xnm"] Dec 13 07:07:25 crc kubenswrapper[4971]: E1213 07:07:25.078972 4971 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="d1cfff96-327c-4048-9bba-649a7345108c" containerName="registry-server" Dec 13 07:07:25 crc kubenswrapper[4971]: I1213 07:07:25.078990 4971 state_mem.go:107] "Deleted CPUSet assignment" podUID="d1cfff96-327c-4048-9bba-649a7345108c" containerName="registry-server" Dec 13 07:07:25 crc kubenswrapper[4971]: I1213 07:07:25.079185 4971 memory_manager.go:354] "RemoveStaleState removing state" podUID="d1cfff96-327c-4048-9bba-649a7345108c" containerName="registry-server" Dec 13 07:07:25 crc kubenswrapper[4971]: I1213 07:07:25.080283 4971 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/ea5d71784a81c95dff2031a02f0a0b3f756f86f14acad8f152d938f56fl9xnm" Dec 13 07:07:25 crc kubenswrapper[4971]: I1213 07:07:25.083743 4971 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"default-dockercfg-kf9kk" Dec 13 07:07:25 crc kubenswrapper[4971]: I1213 07:07:25.085797 4971 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/ea5d71784a81c95dff2031a02f0a0b3f756f86f14acad8f152d938f56fl9xnm"] Dec 13 07:07:25 crc kubenswrapper[4971]: I1213 07:07:25.167745 4971 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/8fbc1cb9-af8d-4ab7-92e4-63035d0c1d74-bundle\") pod \"ea5d71784a81c95dff2031a02f0a0b3f756f86f14acad8f152d938f56fl9xnm\" (UID: \"8fbc1cb9-af8d-4ab7-92e4-63035d0c1d74\") " pod="openstack-operators/ea5d71784a81c95dff2031a02f0a0b3f756f86f14acad8f152d938f56fl9xnm" Dec 13 07:07:25 crc kubenswrapper[4971]: I1213 07:07:25.167839 4971 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-p7hfg\" (UniqueName: \"kubernetes.io/projected/8fbc1cb9-af8d-4ab7-92e4-63035d0c1d74-kube-api-access-p7hfg\") pod \"ea5d71784a81c95dff2031a02f0a0b3f756f86f14acad8f152d938f56fl9xnm\" (UID: \"8fbc1cb9-af8d-4ab7-92e4-63035d0c1d74\") " pod="openstack-operators/ea5d71784a81c95dff2031a02f0a0b3f756f86f14acad8f152d938f56fl9xnm" Dec 13 07:07:25 crc kubenswrapper[4971]: I1213 07:07:25.167873 4971 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/8fbc1cb9-af8d-4ab7-92e4-63035d0c1d74-util\") pod \"ea5d71784a81c95dff2031a02f0a0b3f756f86f14acad8f152d938f56fl9xnm\" (UID: \"8fbc1cb9-af8d-4ab7-92e4-63035d0c1d74\") " pod="openstack-operators/ea5d71784a81c95dff2031a02f0a0b3f756f86f14acad8f152d938f56fl9xnm" Dec 13 07:07:25 crc kubenswrapper[4971]: I1213 07:07:25.269636 4971 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-p7hfg\" (UniqueName: \"kubernetes.io/projected/8fbc1cb9-af8d-4ab7-92e4-63035d0c1d74-kube-api-access-p7hfg\") pod \"ea5d71784a81c95dff2031a02f0a0b3f756f86f14acad8f152d938f56fl9xnm\" (UID: \"8fbc1cb9-af8d-4ab7-92e4-63035d0c1d74\") " pod="openstack-operators/ea5d71784a81c95dff2031a02f0a0b3f756f86f14acad8f152d938f56fl9xnm" Dec 13 07:07:25 crc kubenswrapper[4971]: I1213 07:07:25.269688 4971 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/8fbc1cb9-af8d-4ab7-92e4-63035d0c1d74-util\") pod \"ea5d71784a81c95dff2031a02f0a0b3f756f86f14acad8f152d938f56fl9xnm\" (UID: \"8fbc1cb9-af8d-4ab7-92e4-63035d0c1d74\") " pod="openstack-operators/ea5d71784a81c95dff2031a02f0a0b3f756f86f14acad8f152d938f56fl9xnm" Dec 13 07:07:25 crc kubenswrapper[4971]: I1213 07:07:25.269760 4971 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/8fbc1cb9-af8d-4ab7-92e4-63035d0c1d74-bundle\") pod \"ea5d71784a81c95dff2031a02f0a0b3f756f86f14acad8f152d938f56fl9xnm\" (UID: \"8fbc1cb9-af8d-4ab7-92e4-63035d0c1d74\") " pod="openstack-operators/ea5d71784a81c95dff2031a02f0a0b3f756f86f14acad8f152d938f56fl9xnm" Dec 13 07:07:25 crc kubenswrapper[4971]: I1213 07:07:25.270331 4971 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/8fbc1cb9-af8d-4ab7-92e4-63035d0c1d74-util\") pod \"ea5d71784a81c95dff2031a02f0a0b3f756f86f14acad8f152d938f56fl9xnm\" (UID: \"8fbc1cb9-af8d-4ab7-92e4-63035d0c1d74\") " pod="openstack-operators/ea5d71784a81c95dff2031a02f0a0b3f756f86f14acad8f152d938f56fl9xnm" Dec 13 07:07:25 crc kubenswrapper[4971]: I1213 07:07:25.270364 4971 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/8fbc1cb9-af8d-4ab7-92e4-63035d0c1d74-bundle\") pod \"ea5d71784a81c95dff2031a02f0a0b3f756f86f14acad8f152d938f56fl9xnm\" (UID: \"8fbc1cb9-af8d-4ab7-92e4-63035d0c1d74\") " pod="openstack-operators/ea5d71784a81c95dff2031a02f0a0b3f756f86f14acad8f152d938f56fl9xnm" Dec 13 07:07:25 crc kubenswrapper[4971]: I1213 07:07:25.291715 4971 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-p7hfg\" (UniqueName: \"kubernetes.io/projected/8fbc1cb9-af8d-4ab7-92e4-63035d0c1d74-kube-api-access-p7hfg\") pod \"ea5d71784a81c95dff2031a02f0a0b3f756f86f14acad8f152d938f56fl9xnm\" (UID: \"8fbc1cb9-af8d-4ab7-92e4-63035d0c1d74\") " pod="openstack-operators/ea5d71784a81c95dff2031a02f0a0b3f756f86f14acad8f152d938f56fl9xnm" Dec 13 07:07:25 crc kubenswrapper[4971]: I1213 07:07:25.398877 4971 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/ea5d71784a81c95dff2031a02f0a0b3f756f86f14acad8f152d938f56fl9xnm" Dec 13 07:07:25 crc kubenswrapper[4971]: I1213 07:07:25.970610 4971 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/ea5d71784a81c95dff2031a02f0a0b3f756f86f14acad8f152d938f56fl9xnm"] Dec 13 07:07:26 crc kubenswrapper[4971]: I1213 07:07:26.884992 4971 generic.go:334] "Generic (PLEG): container finished" podID="8fbc1cb9-af8d-4ab7-92e4-63035d0c1d74" containerID="d242e0f3206602372cd5a56aceb503883ab74a9b340525fe1cb80ed1af4b1673" exitCode=0 Dec 13 07:07:26 crc kubenswrapper[4971]: I1213 07:07:26.885095 4971 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/ea5d71784a81c95dff2031a02f0a0b3f756f86f14acad8f152d938f56fl9xnm" event={"ID":"8fbc1cb9-af8d-4ab7-92e4-63035d0c1d74","Type":"ContainerDied","Data":"d242e0f3206602372cd5a56aceb503883ab74a9b340525fe1cb80ed1af4b1673"} Dec 13 07:07:26 crc kubenswrapper[4971]: I1213 07:07:26.885428 4971 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/ea5d71784a81c95dff2031a02f0a0b3f756f86f14acad8f152d938f56fl9xnm" event={"ID":"8fbc1cb9-af8d-4ab7-92e4-63035d0c1d74","Type":"ContainerStarted","Data":"56d1285884f5d303116c99417ae8b6576cfe513652b8623e6246dd7f409040cd"} Dec 13 07:07:27 crc kubenswrapper[4971]: I1213 07:07:27.892013 4971 generic.go:334] "Generic (PLEG): container finished" podID="8fbc1cb9-af8d-4ab7-92e4-63035d0c1d74" containerID="3f75e518e928d2b5ee9a01aa2b2a3a727b9fb438053bd71ce2716a05d3d93036" exitCode=0 Dec 13 07:07:27 crc kubenswrapper[4971]: I1213 07:07:27.892064 4971 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/ea5d71784a81c95dff2031a02f0a0b3f756f86f14acad8f152d938f56fl9xnm" event={"ID":"8fbc1cb9-af8d-4ab7-92e4-63035d0c1d74","Type":"ContainerDied","Data":"3f75e518e928d2b5ee9a01aa2b2a3a727b9fb438053bd71ce2716a05d3d93036"} Dec 13 07:07:28 crc kubenswrapper[4971]: I1213 07:07:28.902935 4971 generic.go:334] "Generic (PLEG): container finished" podID="8fbc1cb9-af8d-4ab7-92e4-63035d0c1d74" containerID="0c08cb2d640fc48a46caaaba698a35a85e4c784109f18d62a3642d78d4477726" exitCode=0 Dec 13 07:07:28 crc kubenswrapper[4971]: I1213 07:07:28.903075 4971 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/ea5d71784a81c95dff2031a02f0a0b3f756f86f14acad8f152d938f56fl9xnm" event={"ID":"8fbc1cb9-af8d-4ab7-92e4-63035d0c1d74","Type":"ContainerDied","Data":"0c08cb2d640fc48a46caaaba698a35a85e4c784109f18d62a3642d78d4477726"} Dec 13 07:07:30 crc kubenswrapper[4971]: I1213 07:07:30.145241 4971 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack-operators/ea5d71784a81c95dff2031a02f0a0b3f756f86f14acad8f152d938f56fl9xnm" Dec 13 07:07:30 crc kubenswrapper[4971]: I1213 07:07:30.249171 4971 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/8fbc1cb9-af8d-4ab7-92e4-63035d0c1d74-bundle\") pod \"8fbc1cb9-af8d-4ab7-92e4-63035d0c1d74\" (UID: \"8fbc1cb9-af8d-4ab7-92e4-63035d0c1d74\") " Dec 13 07:07:30 crc kubenswrapper[4971]: I1213 07:07:30.249225 4971 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-p7hfg\" (UniqueName: \"kubernetes.io/projected/8fbc1cb9-af8d-4ab7-92e4-63035d0c1d74-kube-api-access-p7hfg\") pod \"8fbc1cb9-af8d-4ab7-92e4-63035d0c1d74\" (UID: \"8fbc1cb9-af8d-4ab7-92e4-63035d0c1d74\") " Dec 13 07:07:30 crc kubenswrapper[4971]: I1213 07:07:30.249268 4971 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/8fbc1cb9-af8d-4ab7-92e4-63035d0c1d74-util\") pod \"8fbc1cb9-af8d-4ab7-92e4-63035d0c1d74\" (UID: \"8fbc1cb9-af8d-4ab7-92e4-63035d0c1d74\") " Dec 13 07:07:30 crc kubenswrapper[4971]: I1213 07:07:30.250005 4971 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/8fbc1cb9-af8d-4ab7-92e4-63035d0c1d74-bundle" (OuterVolumeSpecName: "bundle") pod "8fbc1cb9-af8d-4ab7-92e4-63035d0c1d74" (UID: "8fbc1cb9-af8d-4ab7-92e4-63035d0c1d74"). InnerVolumeSpecName "bundle". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 13 07:07:30 crc kubenswrapper[4971]: I1213 07:07:30.254165 4971 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/8fbc1cb9-af8d-4ab7-92e4-63035d0c1d74-kube-api-access-p7hfg" (OuterVolumeSpecName: "kube-api-access-p7hfg") pod "8fbc1cb9-af8d-4ab7-92e4-63035d0c1d74" (UID: "8fbc1cb9-af8d-4ab7-92e4-63035d0c1d74"). InnerVolumeSpecName "kube-api-access-p7hfg". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 13 07:07:30 crc kubenswrapper[4971]: I1213 07:07:30.264476 4971 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/8fbc1cb9-af8d-4ab7-92e4-63035d0c1d74-util" (OuterVolumeSpecName: "util") pod "8fbc1cb9-af8d-4ab7-92e4-63035d0c1d74" (UID: "8fbc1cb9-af8d-4ab7-92e4-63035d0c1d74"). InnerVolumeSpecName "util". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 13 07:07:30 crc kubenswrapper[4971]: I1213 07:07:30.351301 4971 reconciler_common.go:293] "Volume detached for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/8fbc1cb9-af8d-4ab7-92e4-63035d0c1d74-bundle\") on node \"crc\" DevicePath \"\"" Dec 13 07:07:30 crc kubenswrapper[4971]: I1213 07:07:30.351351 4971 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-p7hfg\" (UniqueName: \"kubernetes.io/projected/8fbc1cb9-af8d-4ab7-92e4-63035d0c1d74-kube-api-access-p7hfg\") on node \"crc\" DevicePath \"\"" Dec 13 07:07:30 crc kubenswrapper[4971]: I1213 07:07:30.351366 4971 reconciler_common.go:293] "Volume detached for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/8fbc1cb9-af8d-4ab7-92e4-63035d0c1d74-util\") on node \"crc\" DevicePath \"\"" Dec 13 07:07:30 crc kubenswrapper[4971]: I1213 07:07:30.918496 4971 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/ea5d71784a81c95dff2031a02f0a0b3f756f86f14acad8f152d938f56fl9xnm" event={"ID":"8fbc1cb9-af8d-4ab7-92e4-63035d0c1d74","Type":"ContainerDied","Data":"56d1285884f5d303116c99417ae8b6576cfe513652b8623e6246dd7f409040cd"} Dec 13 07:07:30 crc kubenswrapper[4971]: I1213 07:07:30.918554 4971 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="56d1285884f5d303116c99417ae8b6576cfe513652b8623e6246dd7f409040cd" Dec 13 07:07:30 crc kubenswrapper[4971]: I1213 07:07:30.918593 4971 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack-operators/ea5d71784a81c95dff2031a02f0a0b3f756f86f14acad8f152d938f56fl9xnm" Dec 13 07:07:32 crc kubenswrapper[4971]: I1213 07:07:32.317958 4971 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/openstack-operator-controller-operator-57bbbf4567-7k465"] Dec 13 07:07:32 crc kubenswrapper[4971]: E1213 07:07:32.318502 4971 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="8fbc1cb9-af8d-4ab7-92e4-63035d0c1d74" containerName="pull" Dec 13 07:07:32 crc kubenswrapper[4971]: I1213 07:07:32.318530 4971 state_mem.go:107] "Deleted CPUSet assignment" podUID="8fbc1cb9-af8d-4ab7-92e4-63035d0c1d74" containerName="pull" Dec 13 07:07:32 crc kubenswrapper[4971]: E1213 07:07:32.318547 4971 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="8fbc1cb9-af8d-4ab7-92e4-63035d0c1d74" containerName="extract" Dec 13 07:07:32 crc kubenswrapper[4971]: I1213 07:07:32.318553 4971 state_mem.go:107] "Deleted CPUSet assignment" podUID="8fbc1cb9-af8d-4ab7-92e4-63035d0c1d74" containerName="extract" Dec 13 07:07:32 crc kubenswrapper[4971]: E1213 07:07:32.318562 4971 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="8fbc1cb9-af8d-4ab7-92e4-63035d0c1d74" containerName="util" Dec 13 07:07:32 crc kubenswrapper[4971]: I1213 07:07:32.318568 4971 state_mem.go:107] "Deleted CPUSet assignment" podUID="8fbc1cb9-af8d-4ab7-92e4-63035d0c1d74" containerName="util" Dec 13 07:07:32 crc kubenswrapper[4971]: I1213 07:07:32.318671 4971 memory_manager.go:354] "RemoveStaleState removing state" podUID="8fbc1cb9-af8d-4ab7-92e4-63035d0c1d74" containerName="extract" Dec 13 07:07:32 crc kubenswrapper[4971]: I1213 07:07:32.319076 4971 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/openstack-operator-controller-operator-57bbbf4567-7k465" Dec 13 07:07:32 crc kubenswrapper[4971]: I1213 07:07:32.321401 4971 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"openstack-operator-controller-operator-dockercfg-8hkr4" Dec 13 07:07:32 crc kubenswrapper[4971]: I1213 07:07:32.366813 4971 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/openstack-operator-controller-operator-57bbbf4567-7k465"] Dec 13 07:07:32 crc kubenswrapper[4971]: I1213 07:07:32.377564 4971 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-v2fx8\" (UniqueName: \"kubernetes.io/projected/8e098434-439a-4045-b1ec-4fd4facfc8b5-kube-api-access-v2fx8\") pod \"openstack-operator-controller-operator-57bbbf4567-7k465\" (UID: \"8e098434-439a-4045-b1ec-4fd4facfc8b5\") " pod="openstack-operators/openstack-operator-controller-operator-57bbbf4567-7k465" Dec 13 07:07:32 crc kubenswrapper[4971]: I1213 07:07:32.478636 4971 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-v2fx8\" (UniqueName: \"kubernetes.io/projected/8e098434-439a-4045-b1ec-4fd4facfc8b5-kube-api-access-v2fx8\") pod \"openstack-operator-controller-operator-57bbbf4567-7k465\" (UID: \"8e098434-439a-4045-b1ec-4fd4facfc8b5\") " pod="openstack-operators/openstack-operator-controller-operator-57bbbf4567-7k465" Dec 13 07:07:32 crc kubenswrapper[4971]: I1213 07:07:32.497651 4971 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-v2fx8\" (UniqueName: \"kubernetes.io/projected/8e098434-439a-4045-b1ec-4fd4facfc8b5-kube-api-access-v2fx8\") pod \"openstack-operator-controller-operator-57bbbf4567-7k465\" (UID: \"8e098434-439a-4045-b1ec-4fd4facfc8b5\") " pod="openstack-operators/openstack-operator-controller-operator-57bbbf4567-7k465" Dec 13 07:07:32 crc kubenswrapper[4971]: I1213 07:07:32.633855 4971 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/openstack-operator-controller-operator-57bbbf4567-7k465" Dec 13 07:07:33 crc kubenswrapper[4971]: I1213 07:07:33.017905 4971 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/openstack-operator-controller-operator-57bbbf4567-7k465"] Dec 13 07:07:33 crc kubenswrapper[4971]: I1213 07:07:33.942016 4971 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/openstack-operator-controller-operator-57bbbf4567-7k465" event={"ID":"8e098434-439a-4045-b1ec-4fd4facfc8b5","Type":"ContainerStarted","Data":"26785ff0eb01b019639095424621c2a90bdb26d995948b6adf9d588225ac9079"} Dec 13 07:07:41 crc kubenswrapper[4971]: I1213 07:07:41.001730 4971 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/openstack-operator-controller-operator-57bbbf4567-7k465" event={"ID":"8e098434-439a-4045-b1ec-4fd4facfc8b5","Type":"ContainerStarted","Data":"a26ade213185196b113d9fe376779358a02eb587caef310deb2018d7638d59e0"} Dec 13 07:07:41 crc kubenswrapper[4971]: I1213 07:07:41.002106 4971 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/openstack-operator-controller-operator-57bbbf4567-7k465" Dec 13 07:07:41 crc kubenswrapper[4971]: I1213 07:07:41.048927 4971 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/openstack-operator-controller-operator-57bbbf4567-7k465" podStartSLOduration=2.000940731 podStartE2EDuration="9.048908009s" podCreationTimestamp="2025-12-13 07:07:32 +0000 UTC" firstStartedPulling="2025-12-13 07:07:33.021083948 +0000 UTC m=+1109.625493396" lastFinishedPulling="2025-12-13 07:07:40.069051226 +0000 UTC m=+1116.673460674" observedRunningTime="2025-12-13 07:07:41.044843548 +0000 UTC m=+1117.649253006" watchObservedRunningTime="2025-12-13 07:07:41.048908009 +0000 UTC m=+1117.653317457" Dec 13 07:07:46 crc kubenswrapper[4971]: I1213 07:07:46.154213 4971 patch_prober.go:28] interesting pod/machine-config-daemon-82xjz container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 13 07:07:46 crc kubenswrapper[4971]: I1213 07:07:46.154633 4971 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-82xjz" podUID="e0c4af12-edda-41d7-9b44-f87396174ef2" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 13 07:07:52 crc kubenswrapper[4971]: I1213 07:07:52.637172 4971 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/openstack-operator-controller-operator-57bbbf4567-7k465" Dec 13 07:08:12 crc kubenswrapper[4971]: I1213 07:08:12.775666 4971 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/barbican-operator-controller-manager-95949466-6hkxl"] Dec 13 07:08:12 crc kubenswrapper[4971]: I1213 07:08:12.777221 4971 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/barbican-operator-controller-manager-95949466-6hkxl" Dec 13 07:08:12 crc kubenswrapper[4971]: I1213 07:08:12.778892 4971 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"barbican-operator-controller-manager-dockercfg-fxv4j" Dec 13 07:08:12 crc kubenswrapper[4971]: I1213 07:08:12.787162 4971 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/cinder-operator-controller-manager-5cf45c46bd-8sg42"] Dec 13 07:08:12 crc kubenswrapper[4971]: I1213 07:08:12.789576 4971 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/cinder-operator-controller-manager-5cf45c46bd-8sg42" Dec 13 07:08:12 crc kubenswrapper[4971]: I1213 07:08:12.794478 4971 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"cinder-operator-controller-manager-dockercfg-z9nqh" Dec 13 07:08:12 crc kubenswrapper[4971]: I1213 07:08:12.795453 4971 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/barbican-operator-controller-manager-95949466-6hkxl"] Dec 13 07:08:12 crc kubenswrapper[4971]: I1213 07:08:12.814138 4971 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/designate-operator-controller-manager-66f8b87655-qxr27"] Dec 13 07:08:12 crc kubenswrapper[4971]: I1213 07:08:12.815946 4971 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/designate-operator-controller-manager-66f8b87655-qxr27" Dec 13 07:08:12 crc kubenswrapper[4971]: I1213 07:08:12.821484 4971 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"designate-operator-controller-manager-dockercfg-q5k7q" Dec 13 07:08:12 crc kubenswrapper[4971]: I1213 07:08:12.822456 4971 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/cinder-operator-controller-manager-5cf45c46bd-8sg42"] Dec 13 07:08:12 crc kubenswrapper[4971]: I1213 07:08:12.835611 4971 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/glance-operator-controller-manager-767f9d7567-tmc6q"] Dec 13 07:08:12 crc kubenswrapper[4971]: I1213 07:08:12.836730 4971 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/glance-operator-controller-manager-767f9d7567-tmc6q" Dec 13 07:08:12 crc kubenswrapper[4971]: I1213 07:08:12.841983 4971 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"glance-operator-controller-manager-dockercfg-s6sxp" Dec 13 07:08:12 crc kubenswrapper[4971]: I1213 07:08:12.852390 4971 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/designate-operator-controller-manager-66f8b87655-qxr27"] Dec 13 07:08:12 crc kubenswrapper[4971]: I1213 07:08:12.916801 4971 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-7dz47\" (UniqueName: \"kubernetes.io/projected/62c93900-e946-4684-9c40-36c1d8ff0468-kube-api-access-7dz47\") pod \"cinder-operator-controller-manager-5cf45c46bd-8sg42\" (UID: \"62c93900-e946-4684-9c40-36c1d8ff0468\") " pod="openstack-operators/cinder-operator-controller-manager-5cf45c46bd-8sg42" Dec 13 07:08:12 crc kubenswrapper[4971]: I1213 07:08:12.916908 4971 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-wlt4j\" (UniqueName: \"kubernetes.io/projected/cfe5eaa1-4beb-49ff-a080-ab230292c8ce-kube-api-access-wlt4j\") pod \"designate-operator-controller-manager-66f8b87655-qxr27\" (UID: \"cfe5eaa1-4beb-49ff-a080-ab230292c8ce\") " pod="openstack-operators/designate-operator-controller-manager-66f8b87655-qxr27" Dec 13 07:08:12 crc kubenswrapper[4971]: I1213 07:08:12.917006 4971 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-5k2mg\" (UniqueName: \"kubernetes.io/projected/3b74a6fd-0613-4f48-b585-64570abdb09f-kube-api-access-5k2mg\") pod \"barbican-operator-controller-manager-95949466-6hkxl\" (UID: \"3b74a6fd-0613-4f48-b585-64570abdb09f\") " pod="openstack-operators/barbican-operator-controller-manager-95949466-6hkxl" Dec 13 07:08:12 crc kubenswrapper[4971]: I1213 07:08:12.922576 4971 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/glance-operator-controller-manager-767f9d7567-tmc6q"] Dec 13 07:08:12 crc kubenswrapper[4971]: I1213 07:08:12.931789 4971 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/heat-operator-controller-manager-59b8dcb766-4xg9v"] Dec 13 07:08:12 crc kubenswrapper[4971]: I1213 07:08:12.946568 4971 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/heat-operator-controller-manager-59b8dcb766-4xg9v"] Dec 13 07:08:12 crc kubenswrapper[4971]: I1213 07:08:12.946689 4971 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/heat-operator-controller-manager-59b8dcb766-4xg9v" Dec 13 07:08:12 crc kubenswrapper[4971]: I1213 07:08:12.948084 4971 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/horizon-operator-controller-manager-6ccf486b9-bb2ft"] Dec 13 07:08:12 crc kubenswrapper[4971]: I1213 07:08:12.949011 4971 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/horizon-operator-controller-manager-6ccf486b9-bb2ft" Dec 13 07:08:12 crc kubenswrapper[4971]: I1213 07:08:12.950945 4971 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"heat-operator-controller-manager-dockercfg-j9wfr" Dec 13 07:08:12 crc kubenswrapper[4971]: I1213 07:08:12.951487 4971 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"horizon-operator-controller-manager-dockercfg-ptpnh" Dec 13 07:08:12 crc kubenswrapper[4971]: I1213 07:08:12.968567 4971 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/horizon-operator-controller-manager-6ccf486b9-bb2ft"] Dec 13 07:08:12 crc kubenswrapper[4971]: I1213 07:08:12.977854 4971 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/infra-operator-controller-manager-58944d7758-59mz9"] Dec 13 07:08:12 crc kubenswrapper[4971]: I1213 07:08:12.978795 4971 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/infra-operator-controller-manager-58944d7758-59mz9" Dec 13 07:08:12 crc kubenswrapper[4971]: I1213 07:08:12.998349 4971 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"infra-operator-controller-manager-dockercfg-wv9bt" Dec 13 07:08:12 crc kubenswrapper[4971]: I1213 07:08:12.998560 4971 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"infra-operator-webhook-server-cert" Dec 13 07:08:12 crc kubenswrapper[4971]: I1213 07:08:12.998678 4971 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/infra-operator-controller-manager-58944d7758-59mz9"] Dec 13 07:08:13 crc kubenswrapper[4971]: I1213 07:08:13.004918 4971 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/keystone-operator-controller-manager-5c7cbf548f-qv7p8"] Dec 13 07:08:13 crc kubenswrapper[4971]: I1213 07:08:13.006035 4971 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/keystone-operator-controller-manager-5c7cbf548f-qv7p8" Dec 13 07:08:13 crc kubenswrapper[4971]: I1213 07:08:13.012966 4971 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"keystone-operator-controller-manager-dockercfg-zvbgp" Dec 13 07:08:13 crc kubenswrapper[4971]: I1213 07:08:13.018861 4971 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-7dz47\" (UniqueName: \"kubernetes.io/projected/62c93900-e946-4684-9c40-36c1d8ff0468-kube-api-access-7dz47\") pod \"cinder-operator-controller-manager-5cf45c46bd-8sg42\" (UID: \"62c93900-e946-4684-9c40-36c1d8ff0468\") " pod="openstack-operators/cinder-operator-controller-manager-5cf45c46bd-8sg42" Dec 13 07:08:13 crc kubenswrapper[4971]: I1213 07:08:13.018915 4971 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-wlt4j\" (UniqueName: \"kubernetes.io/projected/cfe5eaa1-4beb-49ff-a080-ab230292c8ce-kube-api-access-wlt4j\") pod \"designate-operator-controller-manager-66f8b87655-qxr27\" (UID: \"cfe5eaa1-4beb-49ff-a080-ab230292c8ce\") " pod="openstack-operators/designate-operator-controller-manager-66f8b87655-qxr27" Dec 13 07:08:13 crc kubenswrapper[4971]: I1213 07:08:13.021776 4971 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-tj88c\" (UniqueName: \"kubernetes.io/projected/2062b86b-1412-4974-a153-802d9f390944-kube-api-access-tj88c\") pod \"glance-operator-controller-manager-767f9d7567-tmc6q\" (UID: \"2062b86b-1412-4974-a153-802d9f390944\") " pod="openstack-operators/glance-operator-controller-manager-767f9d7567-tmc6q" Dec 13 07:08:13 crc kubenswrapper[4971]: I1213 07:08:13.021842 4971 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-5k2mg\" (UniqueName: \"kubernetes.io/projected/3b74a6fd-0613-4f48-b585-64570abdb09f-kube-api-access-5k2mg\") pod \"barbican-operator-controller-manager-95949466-6hkxl\" (UID: \"3b74a6fd-0613-4f48-b585-64570abdb09f\") " pod="openstack-operators/barbican-operator-controller-manager-95949466-6hkxl" Dec 13 07:08:13 crc kubenswrapper[4971]: I1213 07:08:13.026363 4971 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/ironic-operator-controller-manager-f458558d7-m9grc"] Dec 13 07:08:13 crc kubenswrapper[4971]: I1213 07:08:13.048038 4971 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/ironic-operator-controller-manager-f458558d7-m9grc"] Dec 13 07:08:13 crc kubenswrapper[4971]: I1213 07:08:13.048155 4971 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/ironic-operator-controller-manager-f458558d7-m9grc" Dec 13 07:08:13 crc kubenswrapper[4971]: I1213 07:08:13.051038 4971 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/keystone-operator-controller-manager-5c7cbf548f-qv7p8"] Dec 13 07:08:13 crc kubenswrapper[4971]: I1213 07:08:13.053855 4971 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"ironic-operator-controller-manager-dockercfg-zwvm2" Dec 13 07:08:13 crc kubenswrapper[4971]: I1213 07:08:13.055738 4971 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-7dz47\" (UniqueName: \"kubernetes.io/projected/62c93900-e946-4684-9c40-36c1d8ff0468-kube-api-access-7dz47\") pod \"cinder-operator-controller-manager-5cf45c46bd-8sg42\" (UID: \"62c93900-e946-4684-9c40-36c1d8ff0468\") " pod="openstack-operators/cinder-operator-controller-manager-5cf45c46bd-8sg42" Dec 13 07:08:13 crc kubenswrapper[4971]: I1213 07:08:13.058884 4971 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-5k2mg\" (UniqueName: \"kubernetes.io/projected/3b74a6fd-0613-4f48-b585-64570abdb09f-kube-api-access-5k2mg\") pod \"barbican-operator-controller-manager-95949466-6hkxl\" (UID: \"3b74a6fd-0613-4f48-b585-64570abdb09f\") " pod="openstack-operators/barbican-operator-controller-manager-95949466-6hkxl" Dec 13 07:08:13 crc kubenswrapper[4971]: I1213 07:08:13.059191 4971 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-wlt4j\" (UniqueName: \"kubernetes.io/projected/cfe5eaa1-4beb-49ff-a080-ab230292c8ce-kube-api-access-wlt4j\") pod \"designate-operator-controller-manager-66f8b87655-qxr27\" (UID: \"cfe5eaa1-4beb-49ff-a080-ab230292c8ce\") " pod="openstack-operators/designate-operator-controller-manager-66f8b87655-qxr27" Dec 13 07:08:13 crc kubenswrapper[4971]: I1213 07:08:13.061875 4971 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/manila-operator-controller-manager-5fdd9786f7-jjw6m"] Dec 13 07:08:13 crc kubenswrapper[4971]: I1213 07:08:13.063003 4971 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/manila-operator-controller-manager-5fdd9786f7-jjw6m" Dec 13 07:08:13 crc kubenswrapper[4971]: I1213 07:08:13.069540 4971 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"manila-operator-controller-manager-dockercfg-xpl8h" Dec 13 07:08:13 crc kubenswrapper[4971]: I1213 07:08:13.084815 4971 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/manila-operator-controller-manager-5fdd9786f7-jjw6m"] Dec 13 07:08:13 crc kubenswrapper[4971]: I1213 07:08:13.092936 4971 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/barbican-operator-controller-manager-95949466-6hkxl" Dec 13 07:08:13 crc kubenswrapper[4971]: I1213 07:08:13.103605 4971 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/mariadb-operator-controller-manager-f76f4954c-cst7r"] Dec 13 07:08:13 crc kubenswrapper[4971]: I1213 07:08:13.104382 4971 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/mariadb-operator-controller-manager-f76f4954c-cst7r" Dec 13 07:08:13 crc kubenswrapper[4971]: I1213 07:08:13.108272 4971 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/neutron-operator-controller-manager-7cd87b778f-l6b8q"] Dec 13 07:08:13 crc kubenswrapper[4971]: I1213 07:08:13.109086 4971 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/neutron-operator-controller-manager-7cd87b778f-l6b8q" Dec 13 07:08:13 crc kubenswrapper[4971]: I1213 07:08:13.109585 4971 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/cinder-operator-controller-manager-5cf45c46bd-8sg42" Dec 13 07:08:13 crc kubenswrapper[4971]: I1213 07:08:13.112327 4971 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/mariadb-operator-controller-manager-f76f4954c-cst7r"] Dec 13 07:08:13 crc kubenswrapper[4971]: I1213 07:08:13.122678 4971 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/neutron-operator-controller-manager-7cd87b778f-l6b8q"] Dec 13 07:08:13 crc kubenswrapper[4971]: I1213 07:08:13.123249 4971 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"mariadb-operator-controller-manager-dockercfg-h2vwp" Dec 13 07:08:13 crc kubenswrapper[4971]: I1213 07:08:13.123498 4971 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"neutron-operator-controller-manager-dockercfg-8lbqr" Dec 13 07:08:13 crc kubenswrapper[4971]: I1213 07:08:13.124408 4971 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-c9mpf\" (UniqueName: \"kubernetes.io/projected/8e94aa39-05a2-4a63-8b45-a914021bccd7-kube-api-access-c9mpf\") pod \"horizon-operator-controller-manager-6ccf486b9-bb2ft\" (UID: \"8e94aa39-05a2-4a63-8b45-a914021bccd7\") " pod="openstack-operators/horizon-operator-controller-manager-6ccf486b9-bb2ft" Dec 13 07:08:13 crc kubenswrapper[4971]: I1213 07:08:13.124487 4971 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/846e13b5-58bb-4a12-a77a-0a4a9b53247f-cert\") pod \"infra-operator-controller-manager-58944d7758-59mz9\" (UID: \"846e13b5-58bb-4a12-a77a-0a4a9b53247f\") " pod="openstack-operators/infra-operator-controller-manager-58944d7758-59mz9" Dec 13 07:08:13 crc kubenswrapper[4971]: I1213 07:08:13.124545 4971 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-8bbb8\" (UniqueName: \"kubernetes.io/projected/808957a1-29e0-4991-bc8a-23f5a17ecb74-kube-api-access-8bbb8\") pod \"heat-operator-controller-manager-59b8dcb766-4xg9v\" (UID: \"808957a1-29e0-4991-bc8a-23f5a17ecb74\") " pod="openstack-operators/heat-operator-controller-manager-59b8dcb766-4xg9v" Dec 13 07:08:13 crc kubenswrapper[4971]: I1213 07:08:13.124566 4971 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-tj88c\" (UniqueName: \"kubernetes.io/projected/2062b86b-1412-4974-a153-802d9f390944-kube-api-access-tj88c\") pod \"glance-operator-controller-manager-767f9d7567-tmc6q\" (UID: \"2062b86b-1412-4974-a153-802d9f390944\") " pod="openstack-operators/glance-operator-controller-manager-767f9d7567-tmc6q" Dec 13 07:08:13 crc kubenswrapper[4971]: I1213 07:08:13.124615 4971 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-x4d4c\" (UniqueName: \"kubernetes.io/projected/3e2847dd-dcdc-4740-a4a8-4a2417247dc5-kube-api-access-x4d4c\") pod \"keystone-operator-controller-manager-5c7cbf548f-qv7p8\" (UID: \"3e2847dd-dcdc-4740-a4a8-4a2417247dc5\") " pod="openstack-operators/keystone-operator-controller-manager-5c7cbf548f-qv7p8" Dec 13 07:08:13 crc kubenswrapper[4971]: I1213 07:08:13.124694 4971 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-vvt9j\" (UniqueName: \"kubernetes.io/projected/846e13b5-58bb-4a12-a77a-0a4a9b53247f-kube-api-access-vvt9j\") pod \"infra-operator-controller-manager-58944d7758-59mz9\" (UID: \"846e13b5-58bb-4a12-a77a-0a4a9b53247f\") " pod="openstack-operators/infra-operator-controller-manager-58944d7758-59mz9" Dec 13 07:08:13 crc kubenswrapper[4971]: I1213 07:08:13.128640 4971 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/nova-operator-controller-manager-5fbbf8b6cc-k4vtr"] Dec 13 07:08:13 crc kubenswrapper[4971]: I1213 07:08:13.129382 4971 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/nova-operator-controller-manager-5fbbf8b6cc-k4vtr" Dec 13 07:08:13 crc kubenswrapper[4971]: I1213 07:08:13.130910 4971 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"nova-operator-controller-manager-dockercfg-hs2b2" Dec 13 07:08:13 crc kubenswrapper[4971]: I1213 07:08:13.156830 4971 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/designate-operator-controller-manager-66f8b87655-qxr27" Dec 13 07:08:13 crc kubenswrapper[4971]: I1213 07:08:13.162032 4971 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/octavia-operator-controller-manager-68c649d9d-v9t9p"] Dec 13 07:08:13 crc kubenswrapper[4971]: I1213 07:08:13.165709 4971 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/octavia-operator-controller-manager-68c649d9d-v9t9p" Dec 13 07:08:13 crc kubenswrapper[4971]: I1213 07:08:13.168691 4971 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-tj88c\" (UniqueName: \"kubernetes.io/projected/2062b86b-1412-4974-a153-802d9f390944-kube-api-access-tj88c\") pod \"glance-operator-controller-manager-767f9d7567-tmc6q\" (UID: \"2062b86b-1412-4974-a153-802d9f390944\") " pod="openstack-operators/glance-operator-controller-manager-767f9d7567-tmc6q" Dec 13 07:08:13 crc kubenswrapper[4971]: I1213 07:08:13.171827 4971 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/glance-operator-controller-manager-767f9d7567-tmc6q" Dec 13 07:08:13 crc kubenswrapper[4971]: I1213 07:08:13.171899 4971 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"octavia-operator-controller-manager-dockercfg-52m8q" Dec 13 07:08:13 crc kubenswrapper[4971]: I1213 07:08:13.235630 4971 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-vvt9j\" (UniqueName: \"kubernetes.io/projected/846e13b5-58bb-4a12-a77a-0a4a9b53247f-kube-api-access-vvt9j\") pod \"infra-operator-controller-manager-58944d7758-59mz9\" (UID: \"846e13b5-58bb-4a12-a77a-0a4a9b53247f\") " pod="openstack-operators/infra-operator-controller-manager-58944d7758-59mz9" Dec 13 07:08:13 crc kubenswrapper[4971]: I1213 07:08:13.235802 4971 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-688ml\" (UniqueName: \"kubernetes.io/projected/85d76572-e53b-420f-9ffe-d0534bdb1c96-kube-api-access-688ml\") pod \"ironic-operator-controller-manager-f458558d7-m9grc\" (UID: \"85d76572-e53b-420f-9ffe-d0534bdb1c96\") " pod="openstack-operators/ironic-operator-controller-manager-f458558d7-m9grc" Dec 13 07:08:13 crc kubenswrapper[4971]: I1213 07:08:13.235862 4971 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-x8db2\" (UniqueName: \"kubernetes.io/projected/0c834237-c914-43bc-b312-b208ba303bb3-kube-api-access-x8db2\") pod \"neutron-operator-controller-manager-7cd87b778f-l6b8q\" (UID: \"0c834237-c914-43bc-b312-b208ba303bb3\") " pod="openstack-operators/neutron-operator-controller-manager-7cd87b778f-l6b8q" Dec 13 07:08:13 crc kubenswrapper[4971]: I1213 07:08:13.235889 4971 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-7lqwg\" (UniqueName: \"kubernetes.io/projected/8a160d6b-ce31-4d73-918b-368fbc44eb2b-kube-api-access-7lqwg\") pod \"octavia-operator-controller-manager-68c649d9d-v9t9p\" (UID: \"8a160d6b-ce31-4d73-918b-368fbc44eb2b\") " pod="openstack-operators/octavia-operator-controller-manager-68c649d9d-v9t9p" Dec 13 07:08:13 crc kubenswrapper[4971]: I1213 07:08:13.236035 4971 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-c9mpf\" (UniqueName: \"kubernetes.io/projected/8e94aa39-05a2-4a63-8b45-a914021bccd7-kube-api-access-c9mpf\") pod \"horizon-operator-controller-manager-6ccf486b9-bb2ft\" (UID: \"8e94aa39-05a2-4a63-8b45-a914021bccd7\") " pod="openstack-operators/horizon-operator-controller-manager-6ccf486b9-bb2ft" Dec 13 07:08:13 crc kubenswrapper[4971]: I1213 07:08:13.236084 4971 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-qkq5n\" (UniqueName: \"kubernetes.io/projected/4bd9b122-00a8-4895-a44b-0773aae9f6a9-kube-api-access-qkq5n\") pod \"mariadb-operator-controller-manager-f76f4954c-cst7r\" (UID: \"4bd9b122-00a8-4895-a44b-0773aae9f6a9\") " pod="openstack-operators/mariadb-operator-controller-manager-f76f4954c-cst7r" Dec 13 07:08:13 crc kubenswrapper[4971]: I1213 07:08:13.237030 4971 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-wbkzk\" (UniqueName: \"kubernetes.io/projected/a8909dd9-9995-4694-9b5b-47106910656c-kube-api-access-wbkzk\") pod \"nova-operator-controller-manager-5fbbf8b6cc-k4vtr\" (UID: \"a8909dd9-9995-4694-9b5b-47106910656c\") " pod="openstack-operators/nova-operator-controller-manager-5fbbf8b6cc-k4vtr" Dec 13 07:08:13 crc kubenswrapper[4971]: I1213 07:08:13.237086 4971 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/846e13b5-58bb-4a12-a77a-0a4a9b53247f-cert\") pod \"infra-operator-controller-manager-58944d7758-59mz9\" (UID: \"846e13b5-58bb-4a12-a77a-0a4a9b53247f\") " pod="openstack-operators/infra-operator-controller-manager-58944d7758-59mz9" Dec 13 07:08:13 crc kubenswrapper[4971]: I1213 07:08:13.237122 4971 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-8bbb8\" (UniqueName: \"kubernetes.io/projected/808957a1-29e0-4991-bc8a-23f5a17ecb74-kube-api-access-8bbb8\") pod \"heat-operator-controller-manager-59b8dcb766-4xg9v\" (UID: \"808957a1-29e0-4991-bc8a-23f5a17ecb74\") " pod="openstack-operators/heat-operator-controller-manager-59b8dcb766-4xg9v" Dec 13 07:08:13 crc kubenswrapper[4971]: I1213 07:08:13.237283 4971 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-ptwqp\" (UniqueName: \"kubernetes.io/projected/2761a6c0-8c0d-4430-8976-378c2d6df9f7-kube-api-access-ptwqp\") pod \"manila-operator-controller-manager-5fdd9786f7-jjw6m\" (UID: \"2761a6c0-8c0d-4430-8976-378c2d6df9f7\") " pod="openstack-operators/manila-operator-controller-manager-5fdd9786f7-jjw6m" Dec 13 07:08:13 crc kubenswrapper[4971]: I1213 07:08:13.237335 4971 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-x4d4c\" (UniqueName: \"kubernetes.io/projected/3e2847dd-dcdc-4740-a4a8-4a2417247dc5-kube-api-access-x4d4c\") pod \"keystone-operator-controller-manager-5c7cbf548f-qv7p8\" (UID: \"3e2847dd-dcdc-4740-a4a8-4a2417247dc5\") " pod="openstack-operators/keystone-operator-controller-manager-5c7cbf548f-qv7p8" Dec 13 07:08:13 crc kubenswrapper[4971]: E1213 07:08:13.239704 4971 secret.go:188] Couldn't get secret openstack-operators/infra-operator-webhook-server-cert: secret "infra-operator-webhook-server-cert" not found Dec 13 07:08:13 crc kubenswrapper[4971]: E1213 07:08:13.239778 4971 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/846e13b5-58bb-4a12-a77a-0a4a9b53247f-cert podName:846e13b5-58bb-4a12-a77a-0a4a9b53247f nodeName:}" failed. No retries permitted until 2025-12-13 07:08:13.739742065 +0000 UTC m=+1150.344151513 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "cert" (UniqueName: "kubernetes.io/secret/846e13b5-58bb-4a12-a77a-0a4a9b53247f-cert") pod "infra-operator-controller-manager-58944d7758-59mz9" (UID: "846e13b5-58bb-4a12-a77a-0a4a9b53247f") : secret "infra-operator-webhook-server-cert" not found Dec 13 07:08:13 crc kubenswrapper[4971]: I1213 07:08:13.254895 4971 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/nova-operator-controller-manager-5fbbf8b6cc-k4vtr"] Dec 13 07:08:13 crc kubenswrapper[4971]: I1213 07:08:13.257021 4971 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/octavia-operator-controller-manager-68c649d9d-v9t9p"] Dec 13 07:08:13 crc kubenswrapper[4971]: I1213 07:08:13.263187 4971 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-x4d4c\" (UniqueName: \"kubernetes.io/projected/3e2847dd-dcdc-4740-a4a8-4a2417247dc5-kube-api-access-x4d4c\") pod \"keystone-operator-controller-manager-5c7cbf548f-qv7p8\" (UID: \"3e2847dd-dcdc-4740-a4a8-4a2417247dc5\") " pod="openstack-operators/keystone-operator-controller-manager-5c7cbf548f-qv7p8" Dec 13 07:08:13 crc kubenswrapper[4971]: I1213 07:08:13.267036 4971 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-vvt9j\" (UniqueName: \"kubernetes.io/projected/846e13b5-58bb-4a12-a77a-0a4a9b53247f-kube-api-access-vvt9j\") pod \"infra-operator-controller-manager-58944d7758-59mz9\" (UID: \"846e13b5-58bb-4a12-a77a-0a4a9b53247f\") " pod="openstack-operators/infra-operator-controller-manager-58944d7758-59mz9" Dec 13 07:08:13 crc kubenswrapper[4971]: I1213 07:08:13.291170 4971 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-8bbb8\" (UniqueName: \"kubernetes.io/projected/808957a1-29e0-4991-bc8a-23f5a17ecb74-kube-api-access-8bbb8\") pod \"heat-operator-controller-manager-59b8dcb766-4xg9v\" (UID: \"808957a1-29e0-4991-bc8a-23f5a17ecb74\") " pod="openstack-operators/heat-operator-controller-manager-59b8dcb766-4xg9v" Dec 13 07:08:13 crc kubenswrapper[4971]: I1213 07:08:13.300244 4971 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/openstack-baremetal-operator-controller-manager-689f887b54jnn8l"] Dec 13 07:08:13 crc kubenswrapper[4971]: I1213 07:08:13.301608 4971 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/openstack-baremetal-operator-controller-manager-689f887b54jnn8l" Dec 13 07:08:13 crc kubenswrapper[4971]: I1213 07:08:13.303715 4971 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"openstack-baremetal-operator-webhook-server-cert" Dec 13 07:08:13 crc kubenswrapper[4971]: I1213 07:08:13.303715 4971 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"openstack-baremetal-operator-controller-manager-dockercfg-9929r" Dec 13 07:08:13 crc kubenswrapper[4971]: I1213 07:08:13.312551 4971 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-c9mpf\" (UniqueName: \"kubernetes.io/projected/8e94aa39-05a2-4a63-8b45-a914021bccd7-kube-api-access-c9mpf\") pod \"horizon-operator-controller-manager-6ccf486b9-bb2ft\" (UID: \"8e94aa39-05a2-4a63-8b45-a914021bccd7\") " pod="openstack-operators/horizon-operator-controller-manager-6ccf486b9-bb2ft" Dec 13 07:08:13 crc kubenswrapper[4971]: I1213 07:08:13.318806 4971 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/ovn-operator-controller-manager-bf6d4f946-g2r89"] Dec 13 07:08:13 crc kubenswrapper[4971]: I1213 07:08:13.320221 4971 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/ovn-operator-controller-manager-bf6d4f946-g2r89" Dec 13 07:08:13 crc kubenswrapper[4971]: I1213 07:08:13.322260 4971 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"ovn-operator-controller-manager-dockercfg-2qvlb" Dec 13 07:08:13 crc kubenswrapper[4971]: I1213 07:08:13.339337 4971 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-688ml\" (UniqueName: \"kubernetes.io/projected/85d76572-e53b-420f-9ffe-d0534bdb1c96-kube-api-access-688ml\") pod \"ironic-operator-controller-manager-f458558d7-m9grc\" (UID: \"85d76572-e53b-420f-9ffe-d0534bdb1c96\") " pod="openstack-operators/ironic-operator-controller-manager-f458558d7-m9grc" Dec 13 07:08:13 crc kubenswrapper[4971]: I1213 07:08:13.339426 4971 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-x8db2\" (UniqueName: \"kubernetes.io/projected/0c834237-c914-43bc-b312-b208ba303bb3-kube-api-access-x8db2\") pod \"neutron-operator-controller-manager-7cd87b778f-l6b8q\" (UID: \"0c834237-c914-43bc-b312-b208ba303bb3\") " pod="openstack-operators/neutron-operator-controller-manager-7cd87b778f-l6b8q" Dec 13 07:08:13 crc kubenswrapper[4971]: I1213 07:08:13.339478 4971 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-7lqwg\" (UniqueName: \"kubernetes.io/projected/8a160d6b-ce31-4d73-918b-368fbc44eb2b-kube-api-access-7lqwg\") pod \"octavia-operator-controller-manager-68c649d9d-v9t9p\" (UID: \"8a160d6b-ce31-4d73-918b-368fbc44eb2b\") " pod="openstack-operators/octavia-operator-controller-manager-68c649d9d-v9t9p" Dec 13 07:08:13 crc kubenswrapper[4971]: I1213 07:08:13.339531 4971 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-qkq5n\" (UniqueName: \"kubernetes.io/projected/4bd9b122-00a8-4895-a44b-0773aae9f6a9-kube-api-access-qkq5n\") pod \"mariadb-operator-controller-manager-f76f4954c-cst7r\" (UID: \"4bd9b122-00a8-4895-a44b-0773aae9f6a9\") " pod="openstack-operators/mariadb-operator-controller-manager-f76f4954c-cst7r" Dec 13 07:08:13 crc kubenswrapper[4971]: I1213 07:08:13.339576 4971 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-s66w8\" (UniqueName: \"kubernetes.io/projected/9ed7aa8c-f56d-477d-91f5-ef1944e66d57-kube-api-access-s66w8\") pod \"ovn-operator-controller-manager-bf6d4f946-g2r89\" (UID: \"9ed7aa8c-f56d-477d-91f5-ef1944e66d57\") " pod="openstack-operators/ovn-operator-controller-manager-bf6d4f946-g2r89" Dec 13 07:08:13 crc kubenswrapper[4971]: I1213 07:08:13.339601 4971 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/706451b8-644b-4151-b5b9-9975082e556f-cert\") pod \"openstack-baremetal-operator-controller-manager-689f887b54jnn8l\" (UID: \"706451b8-644b-4151-b5b9-9975082e556f\") " pod="openstack-operators/openstack-baremetal-operator-controller-manager-689f887b54jnn8l" Dec 13 07:08:13 crc kubenswrapper[4971]: I1213 07:08:13.339624 4971 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-wbkzk\" (UniqueName: \"kubernetes.io/projected/a8909dd9-9995-4694-9b5b-47106910656c-kube-api-access-wbkzk\") pod \"nova-operator-controller-manager-5fbbf8b6cc-k4vtr\" (UID: \"a8909dd9-9995-4694-9b5b-47106910656c\") " pod="openstack-operators/nova-operator-controller-manager-5fbbf8b6cc-k4vtr" Dec 13 07:08:13 crc kubenswrapper[4971]: I1213 07:08:13.339690 4971 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-tzj69\" (UniqueName: \"kubernetes.io/projected/706451b8-644b-4151-b5b9-9975082e556f-kube-api-access-tzj69\") pod \"openstack-baremetal-operator-controller-manager-689f887b54jnn8l\" (UID: \"706451b8-644b-4151-b5b9-9975082e556f\") " pod="openstack-operators/openstack-baremetal-operator-controller-manager-689f887b54jnn8l" Dec 13 07:08:13 crc kubenswrapper[4971]: I1213 07:08:13.339725 4971 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-ptwqp\" (UniqueName: \"kubernetes.io/projected/2761a6c0-8c0d-4430-8976-378c2d6df9f7-kube-api-access-ptwqp\") pod \"manila-operator-controller-manager-5fdd9786f7-jjw6m\" (UID: \"2761a6c0-8c0d-4430-8976-378c2d6df9f7\") " pod="openstack-operators/manila-operator-controller-manager-5fdd9786f7-jjw6m" Dec 13 07:08:13 crc kubenswrapper[4971]: I1213 07:08:13.341698 4971 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/placement-operator-controller-manager-8665b56d78-hrflw"] Dec 13 07:08:13 crc kubenswrapper[4971]: I1213 07:08:13.344295 4971 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/placement-operator-controller-manager-8665b56d78-hrflw" Dec 13 07:08:13 crc kubenswrapper[4971]: I1213 07:08:13.347771 4971 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"placement-operator-controller-manager-dockercfg-ngkkt" Dec 13 07:08:13 crc kubenswrapper[4971]: I1213 07:08:13.353645 4971 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/keystone-operator-controller-manager-5c7cbf548f-qv7p8" Dec 13 07:08:13 crc kubenswrapper[4971]: I1213 07:08:13.358251 4971 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/openstack-baremetal-operator-controller-manager-689f887b54jnn8l"] Dec 13 07:08:13 crc kubenswrapper[4971]: I1213 07:08:13.362613 4971 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-688ml\" (UniqueName: \"kubernetes.io/projected/85d76572-e53b-420f-9ffe-d0534bdb1c96-kube-api-access-688ml\") pod \"ironic-operator-controller-manager-f458558d7-m9grc\" (UID: \"85d76572-e53b-420f-9ffe-d0534bdb1c96\") " pod="openstack-operators/ironic-operator-controller-manager-f458558d7-m9grc" Dec 13 07:08:13 crc kubenswrapper[4971]: I1213 07:08:13.363906 4971 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-qkq5n\" (UniqueName: \"kubernetes.io/projected/4bd9b122-00a8-4895-a44b-0773aae9f6a9-kube-api-access-qkq5n\") pod \"mariadb-operator-controller-manager-f76f4954c-cst7r\" (UID: \"4bd9b122-00a8-4895-a44b-0773aae9f6a9\") " pod="openstack-operators/mariadb-operator-controller-manager-f76f4954c-cst7r" Dec 13 07:08:13 crc kubenswrapper[4971]: I1213 07:08:13.368193 4971 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-ptwqp\" (UniqueName: \"kubernetes.io/projected/2761a6c0-8c0d-4430-8976-378c2d6df9f7-kube-api-access-ptwqp\") pod \"manila-operator-controller-manager-5fdd9786f7-jjw6m\" (UID: \"2761a6c0-8c0d-4430-8976-378c2d6df9f7\") " pod="openstack-operators/manila-operator-controller-manager-5fdd9786f7-jjw6m" Dec 13 07:08:13 crc kubenswrapper[4971]: I1213 07:08:13.370628 4971 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-wbkzk\" (UniqueName: \"kubernetes.io/projected/a8909dd9-9995-4694-9b5b-47106910656c-kube-api-access-wbkzk\") pod \"nova-operator-controller-manager-5fbbf8b6cc-k4vtr\" (UID: \"a8909dd9-9995-4694-9b5b-47106910656c\") " pod="openstack-operators/nova-operator-controller-manager-5fbbf8b6cc-k4vtr" Dec 13 07:08:13 crc kubenswrapper[4971]: I1213 07:08:13.383863 4971 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-7lqwg\" (UniqueName: \"kubernetes.io/projected/8a160d6b-ce31-4d73-918b-368fbc44eb2b-kube-api-access-7lqwg\") pod \"octavia-operator-controller-manager-68c649d9d-v9t9p\" (UID: \"8a160d6b-ce31-4d73-918b-368fbc44eb2b\") " pod="openstack-operators/octavia-operator-controller-manager-68c649d9d-v9t9p" Dec 13 07:08:13 crc kubenswrapper[4971]: I1213 07:08:13.385179 4971 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-x8db2\" (UniqueName: \"kubernetes.io/projected/0c834237-c914-43bc-b312-b208ba303bb3-kube-api-access-x8db2\") pod \"neutron-operator-controller-manager-7cd87b778f-l6b8q\" (UID: \"0c834237-c914-43bc-b312-b208ba303bb3\") " pod="openstack-operators/neutron-operator-controller-manager-7cd87b778f-l6b8q" Dec 13 07:08:13 crc kubenswrapper[4971]: I1213 07:08:13.393242 4971 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/swift-operator-controller-manager-5c6df8f9-rt5n5"] Dec 13 07:08:13 crc kubenswrapper[4971]: I1213 07:08:13.394689 4971 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/swift-operator-controller-manager-5c6df8f9-rt5n5" Dec 13 07:08:13 crc kubenswrapper[4971]: I1213 07:08:13.402328 4971 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/placement-operator-controller-manager-8665b56d78-hrflw"] Dec 13 07:08:13 crc kubenswrapper[4971]: I1213 07:08:13.402945 4971 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"swift-operator-controller-manager-dockercfg-9rbv9" Dec 13 07:08:13 crc kubenswrapper[4971]: I1213 07:08:13.406242 4971 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/swift-operator-controller-manager-5c6df8f9-rt5n5"] Dec 13 07:08:13 crc kubenswrapper[4971]: I1213 07:08:13.406531 4971 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/ironic-operator-controller-manager-f458558d7-m9grc" Dec 13 07:08:13 crc kubenswrapper[4971]: I1213 07:08:13.414189 4971 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/ovn-operator-controller-manager-bf6d4f946-g2r89"] Dec 13 07:08:13 crc kubenswrapper[4971]: I1213 07:08:13.419074 4971 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/telemetry-operator-controller-manager-97d456b9-j4fpf"] Dec 13 07:08:13 crc kubenswrapper[4971]: I1213 07:08:13.420119 4971 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/telemetry-operator-controller-manager-97d456b9-j4fpf" Dec 13 07:08:13 crc kubenswrapper[4971]: I1213 07:08:13.423662 4971 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"telemetry-operator-controller-manager-dockercfg-ffxnw" Dec 13 07:08:13 crc kubenswrapper[4971]: I1213 07:08:13.431250 4971 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/telemetry-operator-controller-manager-97d456b9-j4fpf"] Dec 13 07:08:13 crc kubenswrapper[4971]: I1213 07:08:13.448374 4971 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-s66w8\" (UniqueName: \"kubernetes.io/projected/9ed7aa8c-f56d-477d-91f5-ef1944e66d57-kube-api-access-s66w8\") pod \"ovn-operator-controller-manager-bf6d4f946-g2r89\" (UID: \"9ed7aa8c-f56d-477d-91f5-ef1944e66d57\") " pod="openstack-operators/ovn-operator-controller-manager-bf6d4f946-g2r89" Dec 13 07:08:13 crc kubenswrapper[4971]: I1213 07:08:13.448421 4971 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/706451b8-644b-4151-b5b9-9975082e556f-cert\") pod \"openstack-baremetal-operator-controller-manager-689f887b54jnn8l\" (UID: \"706451b8-644b-4151-b5b9-9975082e556f\") " pod="openstack-operators/openstack-baremetal-operator-controller-manager-689f887b54jnn8l" Dec 13 07:08:13 crc kubenswrapper[4971]: I1213 07:08:13.448471 4971 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-j9g4v\" (UniqueName: \"kubernetes.io/projected/bd48845c-5d55-4690-960b-ca319f596beb-kube-api-access-j9g4v\") pod \"swift-operator-controller-manager-5c6df8f9-rt5n5\" (UID: \"bd48845c-5d55-4690-960b-ca319f596beb\") " pod="openstack-operators/swift-operator-controller-manager-5c6df8f9-rt5n5" Dec 13 07:08:13 crc kubenswrapper[4971]: I1213 07:08:13.448491 4971 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-9r8gf\" (UniqueName: \"kubernetes.io/projected/f111629c-eb0b-4294-8f85-d4b0b6500454-kube-api-access-9r8gf\") pod \"telemetry-operator-controller-manager-97d456b9-j4fpf\" (UID: \"f111629c-eb0b-4294-8f85-d4b0b6500454\") " pod="openstack-operators/telemetry-operator-controller-manager-97d456b9-j4fpf" Dec 13 07:08:13 crc kubenswrapper[4971]: I1213 07:08:13.448542 4971 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-tzj69\" (UniqueName: \"kubernetes.io/projected/706451b8-644b-4151-b5b9-9975082e556f-kube-api-access-tzj69\") pod \"openstack-baremetal-operator-controller-manager-689f887b54jnn8l\" (UID: \"706451b8-644b-4151-b5b9-9975082e556f\") " pod="openstack-operators/openstack-baremetal-operator-controller-manager-689f887b54jnn8l" Dec 13 07:08:13 crc kubenswrapper[4971]: I1213 07:08:13.448565 4971 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-hm6sz\" (UniqueName: \"kubernetes.io/projected/6b9f3caa-4280-43f3-8a05-14e72f09c384-kube-api-access-hm6sz\") pod \"placement-operator-controller-manager-8665b56d78-hrflw\" (UID: \"6b9f3caa-4280-43f3-8a05-14e72f09c384\") " pod="openstack-operators/placement-operator-controller-manager-8665b56d78-hrflw" Dec 13 07:08:13 crc kubenswrapper[4971]: E1213 07:08:13.449297 4971 secret.go:188] Couldn't get secret openstack-operators/openstack-baremetal-operator-webhook-server-cert: secret "openstack-baremetal-operator-webhook-server-cert" not found Dec 13 07:08:13 crc kubenswrapper[4971]: E1213 07:08:13.449336 4971 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/706451b8-644b-4151-b5b9-9975082e556f-cert podName:706451b8-644b-4151-b5b9-9975082e556f nodeName:}" failed. No retries permitted until 2025-12-13 07:08:13.949324262 +0000 UTC m=+1150.553733710 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "cert" (UniqueName: "kubernetes.io/secret/706451b8-644b-4151-b5b9-9975082e556f-cert") pod "openstack-baremetal-operator-controller-manager-689f887b54jnn8l" (UID: "706451b8-644b-4151-b5b9-9975082e556f") : secret "openstack-baremetal-operator-webhook-server-cert" not found Dec 13 07:08:13 crc kubenswrapper[4971]: I1213 07:08:13.451405 4971 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/test-operator-controller-manager-756ccf86c7-r8p79"] Dec 13 07:08:13 crc kubenswrapper[4971]: I1213 07:08:13.452218 4971 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/test-operator-controller-manager-756ccf86c7-r8p79" Dec 13 07:08:13 crc kubenswrapper[4971]: I1213 07:08:13.459157 4971 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/test-operator-controller-manager-756ccf86c7-r8p79"] Dec 13 07:08:13 crc kubenswrapper[4971]: I1213 07:08:13.459833 4971 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"test-operator-controller-manager-dockercfg-4j2sg" Dec 13 07:08:13 crc kubenswrapper[4971]: I1213 07:08:13.469541 4971 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/manila-operator-controller-manager-5fdd9786f7-jjw6m" Dec 13 07:08:13 crc kubenswrapper[4971]: I1213 07:08:13.482713 4971 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/watcher-operator-controller-manager-55f78b7c4c-dwmgr"] Dec 13 07:08:13 crc kubenswrapper[4971]: I1213 07:08:13.483647 4971 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/watcher-operator-controller-manager-55f78b7c4c-dwmgr" Dec 13 07:08:13 crc kubenswrapper[4971]: I1213 07:08:13.489336 4971 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/watcher-operator-controller-manager-55f78b7c4c-dwmgr"] Dec 13 07:08:13 crc kubenswrapper[4971]: I1213 07:08:13.541115 4971 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/mariadb-operator-controller-manager-f76f4954c-cst7r" Dec 13 07:08:13 crc kubenswrapper[4971]: I1213 07:08:13.549897 4971 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-9r8gf\" (UniqueName: \"kubernetes.io/projected/f111629c-eb0b-4294-8f85-d4b0b6500454-kube-api-access-9r8gf\") pod \"telemetry-operator-controller-manager-97d456b9-j4fpf\" (UID: \"f111629c-eb0b-4294-8f85-d4b0b6500454\") " pod="openstack-operators/telemetry-operator-controller-manager-97d456b9-j4fpf" Dec 13 07:08:13 crc kubenswrapper[4971]: I1213 07:08:13.549990 4971 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-j9g4v\" (UniqueName: \"kubernetes.io/projected/bd48845c-5d55-4690-960b-ca319f596beb-kube-api-access-j9g4v\") pod \"swift-operator-controller-manager-5c6df8f9-rt5n5\" (UID: \"bd48845c-5d55-4690-960b-ca319f596beb\") " pod="openstack-operators/swift-operator-controller-manager-5c6df8f9-rt5n5" Dec 13 07:08:13 crc kubenswrapper[4971]: I1213 07:08:13.550672 4971 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-cl8ll\" (UniqueName: \"kubernetes.io/projected/ee0625f5-5da4-4dc4-8b2e-f77c9f4459ed-kube-api-access-cl8ll\") pod \"test-operator-controller-manager-756ccf86c7-r8p79\" (UID: \"ee0625f5-5da4-4dc4-8b2e-f77c9f4459ed\") " pod="openstack-operators/test-operator-controller-manager-756ccf86c7-r8p79" Dec 13 07:08:13 crc kubenswrapper[4971]: I1213 07:08:13.550799 4971 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-hm6sz\" (UniqueName: \"kubernetes.io/projected/6b9f3caa-4280-43f3-8a05-14e72f09c384-kube-api-access-hm6sz\") pod \"placement-operator-controller-manager-8665b56d78-hrflw\" (UID: \"6b9f3caa-4280-43f3-8a05-14e72f09c384\") " pod="openstack-operators/placement-operator-controller-manager-8665b56d78-hrflw" Dec 13 07:08:13 crc kubenswrapper[4971]: I1213 07:08:13.550908 4971 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-tlxrq\" (UniqueName: \"kubernetes.io/projected/5bb0c1cd-9a8e-4c03-80c0-d9e66dae91ed-kube-api-access-tlxrq\") pod \"watcher-operator-controller-manager-55f78b7c4c-dwmgr\" (UID: \"5bb0c1cd-9a8e-4c03-80c0-d9e66dae91ed\") " pod="openstack-operators/watcher-operator-controller-manager-55f78b7c4c-dwmgr" Dec 13 07:08:13 crc kubenswrapper[4971]: I1213 07:08:13.563994 4971 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/neutron-operator-controller-manager-7cd87b778f-l6b8q" Dec 13 07:08:13 crc kubenswrapper[4971]: I1213 07:08:13.575055 4971 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/heat-operator-controller-manager-59b8dcb766-4xg9v" Dec 13 07:08:13 crc kubenswrapper[4971]: I1213 07:08:13.603102 4971 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/horizon-operator-controller-manager-6ccf486b9-bb2ft" Dec 13 07:08:13 crc kubenswrapper[4971]: I1213 07:08:13.615070 4971 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/nova-operator-controller-manager-5fbbf8b6cc-k4vtr" Dec 13 07:08:13 crc kubenswrapper[4971]: I1213 07:08:13.636351 4971 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/octavia-operator-controller-manager-68c649d9d-v9t9p" Dec 13 07:08:13 crc kubenswrapper[4971]: I1213 07:08:13.652228 4971 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-tlxrq\" (UniqueName: \"kubernetes.io/projected/5bb0c1cd-9a8e-4c03-80c0-d9e66dae91ed-kube-api-access-tlxrq\") pod \"watcher-operator-controller-manager-55f78b7c4c-dwmgr\" (UID: \"5bb0c1cd-9a8e-4c03-80c0-d9e66dae91ed\") " pod="openstack-operators/watcher-operator-controller-manager-55f78b7c4c-dwmgr" Dec 13 07:08:13 crc kubenswrapper[4971]: I1213 07:08:13.652378 4971 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cl8ll\" (UniqueName: \"kubernetes.io/projected/ee0625f5-5da4-4dc4-8b2e-f77c9f4459ed-kube-api-access-cl8ll\") pod \"test-operator-controller-manager-756ccf86c7-r8p79\" (UID: \"ee0625f5-5da4-4dc4-8b2e-f77c9f4459ed\") " pod="openstack-operators/test-operator-controller-manager-756ccf86c7-r8p79" Dec 13 07:08:13 crc kubenswrapper[4971]: I1213 07:08:13.669653 4971 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"watcher-operator-controller-manager-dockercfg-4fpq9" Dec 13 07:08:13 crc kubenswrapper[4971]: I1213 07:08:13.694294 4971 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-tlxrq\" (UniqueName: \"kubernetes.io/projected/5bb0c1cd-9a8e-4c03-80c0-d9e66dae91ed-kube-api-access-tlxrq\") pod \"watcher-operator-controller-manager-55f78b7c4c-dwmgr\" (UID: \"5bb0c1cd-9a8e-4c03-80c0-d9e66dae91ed\") " pod="openstack-operators/watcher-operator-controller-manager-55f78b7c4c-dwmgr" Dec 13 07:08:13 crc kubenswrapper[4971]: I1213 07:08:13.698457 4971 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-9r8gf\" (UniqueName: \"kubernetes.io/projected/f111629c-eb0b-4294-8f85-d4b0b6500454-kube-api-access-9r8gf\") pod \"telemetry-operator-controller-manager-97d456b9-j4fpf\" (UID: \"f111629c-eb0b-4294-8f85-d4b0b6500454\") " pod="openstack-operators/telemetry-operator-controller-manager-97d456b9-j4fpf" Dec 13 07:08:13 crc kubenswrapper[4971]: I1213 07:08:13.699092 4971 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-s66w8\" (UniqueName: \"kubernetes.io/projected/9ed7aa8c-f56d-477d-91f5-ef1944e66d57-kube-api-access-s66w8\") pod \"ovn-operator-controller-manager-bf6d4f946-g2r89\" (UID: \"9ed7aa8c-f56d-477d-91f5-ef1944e66d57\") " pod="openstack-operators/ovn-operator-controller-manager-bf6d4f946-g2r89" Dec 13 07:08:13 crc kubenswrapper[4971]: I1213 07:08:13.720910 4971 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-j9g4v\" (UniqueName: \"kubernetes.io/projected/bd48845c-5d55-4690-960b-ca319f596beb-kube-api-access-j9g4v\") pod \"swift-operator-controller-manager-5c6df8f9-rt5n5\" (UID: \"bd48845c-5d55-4690-960b-ca319f596beb\") " pod="openstack-operators/swift-operator-controller-manager-5c6df8f9-rt5n5" Dec 13 07:08:13 crc kubenswrapper[4971]: I1213 07:08:13.722476 4971 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-cl8ll\" (UniqueName: \"kubernetes.io/projected/ee0625f5-5da4-4dc4-8b2e-f77c9f4459ed-kube-api-access-cl8ll\") pod \"test-operator-controller-manager-756ccf86c7-r8p79\" (UID: \"ee0625f5-5da4-4dc4-8b2e-f77c9f4459ed\") " pod="openstack-operators/test-operator-controller-manager-756ccf86c7-r8p79" Dec 13 07:08:13 crc kubenswrapper[4971]: I1213 07:08:13.724409 4971 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-tzj69\" (UniqueName: \"kubernetes.io/projected/706451b8-644b-4151-b5b9-9975082e556f-kube-api-access-tzj69\") pod \"openstack-baremetal-operator-controller-manager-689f887b54jnn8l\" (UID: \"706451b8-644b-4151-b5b9-9975082e556f\") " pod="openstack-operators/openstack-baremetal-operator-controller-manager-689f887b54jnn8l" Dec 13 07:08:13 crc kubenswrapper[4971]: I1213 07:08:13.725617 4971 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-hm6sz\" (UniqueName: \"kubernetes.io/projected/6b9f3caa-4280-43f3-8a05-14e72f09c384-kube-api-access-hm6sz\") pod \"placement-operator-controller-manager-8665b56d78-hrflw\" (UID: \"6b9f3caa-4280-43f3-8a05-14e72f09c384\") " pod="openstack-operators/placement-operator-controller-manager-8665b56d78-hrflw" Dec 13 07:08:13 crc kubenswrapper[4971]: I1213 07:08:13.777585 4971 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/846e13b5-58bb-4a12-a77a-0a4a9b53247f-cert\") pod \"infra-operator-controller-manager-58944d7758-59mz9\" (UID: \"846e13b5-58bb-4a12-a77a-0a4a9b53247f\") " pod="openstack-operators/infra-operator-controller-manager-58944d7758-59mz9" Dec 13 07:08:13 crc kubenswrapper[4971]: E1213 07:08:13.778082 4971 secret.go:188] Couldn't get secret openstack-operators/infra-operator-webhook-server-cert: secret "infra-operator-webhook-server-cert" not found Dec 13 07:08:13 crc kubenswrapper[4971]: E1213 07:08:13.778143 4971 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/846e13b5-58bb-4a12-a77a-0a4a9b53247f-cert podName:846e13b5-58bb-4a12-a77a-0a4a9b53247f nodeName:}" failed. No retries permitted until 2025-12-13 07:08:14.778122795 +0000 UTC m=+1151.382532243 (durationBeforeRetry 1s). Error: MountVolume.SetUp failed for volume "cert" (UniqueName: "kubernetes.io/secret/846e13b5-58bb-4a12-a77a-0a4a9b53247f-cert") pod "infra-operator-controller-manager-58944d7758-59mz9" (UID: "846e13b5-58bb-4a12-a77a-0a4a9b53247f") : secret "infra-operator-webhook-server-cert" not found Dec 13 07:08:13 crc kubenswrapper[4971]: I1213 07:08:13.782744 4971 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/swift-operator-controller-manager-5c6df8f9-rt5n5" Dec 13 07:08:13 crc kubenswrapper[4971]: I1213 07:08:13.792035 4971 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/telemetry-operator-controller-manager-97d456b9-j4fpf" Dec 13 07:08:13 crc kubenswrapper[4971]: I1213 07:08:13.802053 4971 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/test-operator-controller-manager-756ccf86c7-r8p79" Dec 13 07:08:13 crc kubenswrapper[4971]: I1213 07:08:13.837451 4971 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/watcher-operator-controller-manager-55f78b7c4c-dwmgr" Dec 13 07:08:13 crc kubenswrapper[4971]: I1213 07:08:13.918703 4971 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/openstack-operator-controller-manager-56f6fbdf6-vzrkl"] Dec 13 07:08:13 crc kubenswrapper[4971]: I1213 07:08:13.919947 4971 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/openstack-operator-controller-manager-56f6fbdf6-vzrkl" Dec 13 07:08:13 crc kubenswrapper[4971]: I1213 07:08:13.923293 4971 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"metrics-server-cert" Dec 13 07:08:13 crc kubenswrapper[4971]: I1213 07:08:13.923486 4971 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"webhook-server-cert" Dec 13 07:08:13 crc kubenswrapper[4971]: I1213 07:08:13.923658 4971 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"openstack-operator-controller-manager-dockercfg-p9xmj" Dec 13 07:08:13 crc kubenswrapper[4971]: I1213 07:08:13.958229 4971 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/ovn-operator-controller-manager-bf6d4f946-g2r89" Dec 13 07:08:14 crc kubenswrapper[4971]: I1213 07:08:14.011045 4971 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/placement-operator-controller-manager-8665b56d78-hrflw" Dec 13 07:08:14 crc kubenswrapper[4971]: I1213 07:08:14.011852 4971 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/openstack-operator-controller-manager-56f6fbdf6-vzrkl"] Dec 13 07:08:14 crc kubenswrapper[4971]: I1213 07:08:14.049149 4971 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/706451b8-644b-4151-b5b9-9975082e556f-cert\") pod \"openstack-baremetal-operator-controller-manager-689f887b54jnn8l\" (UID: \"706451b8-644b-4151-b5b9-9975082e556f\") " pod="openstack-operators/openstack-baremetal-operator-controller-manager-689f887b54jnn8l" Dec 13 07:08:14 crc kubenswrapper[4971]: I1213 07:08:14.049287 4971 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/40f2727e-ef9d-4557-8619-b37174d00b67-metrics-certs\") pod \"openstack-operator-controller-manager-56f6fbdf6-vzrkl\" (UID: \"40f2727e-ef9d-4557-8619-b37174d00b67\") " pod="openstack-operators/openstack-operator-controller-manager-56f6fbdf6-vzrkl" Dec 13 07:08:14 crc kubenswrapper[4971]: I1213 07:08:14.049351 4971 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-5dlnw\" (UniqueName: \"kubernetes.io/projected/40f2727e-ef9d-4557-8619-b37174d00b67-kube-api-access-5dlnw\") pod \"openstack-operator-controller-manager-56f6fbdf6-vzrkl\" (UID: \"40f2727e-ef9d-4557-8619-b37174d00b67\") " pod="openstack-operators/openstack-operator-controller-manager-56f6fbdf6-vzrkl" Dec 13 07:08:14 crc kubenswrapper[4971]: I1213 07:08:14.049379 4971 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"webhook-certs\" (UniqueName: \"kubernetes.io/secret/40f2727e-ef9d-4557-8619-b37174d00b67-webhook-certs\") pod \"openstack-operator-controller-manager-56f6fbdf6-vzrkl\" (UID: \"40f2727e-ef9d-4557-8619-b37174d00b67\") " pod="openstack-operators/openstack-operator-controller-manager-56f6fbdf6-vzrkl" Dec 13 07:08:14 crc kubenswrapper[4971]: E1213 07:08:14.049595 4971 secret.go:188] Couldn't get secret openstack-operators/openstack-baremetal-operator-webhook-server-cert: secret "openstack-baremetal-operator-webhook-server-cert" not found Dec 13 07:08:14 crc kubenswrapper[4971]: E1213 07:08:14.049648 4971 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/706451b8-644b-4151-b5b9-9975082e556f-cert podName:706451b8-644b-4151-b5b9-9975082e556f nodeName:}" failed. No retries permitted until 2025-12-13 07:08:15.049622387 +0000 UTC m=+1151.654031835 (durationBeforeRetry 1s). Error: MountVolume.SetUp failed for volume "cert" (UniqueName: "kubernetes.io/secret/706451b8-644b-4151-b5b9-9975082e556f-cert") pod "openstack-baremetal-operator-controller-manager-689f887b54jnn8l" (UID: "706451b8-644b-4151-b5b9-9975082e556f") : secret "openstack-baremetal-operator-webhook-server-cert" not found Dec 13 07:08:14 crc kubenswrapper[4971]: I1213 07:08:14.068494 4971 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/rabbitmq-cluster-operator-manager-668c99d594-pzcql"] Dec 13 07:08:14 crc kubenswrapper[4971]: I1213 07:08:14.070274 4971 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/rabbitmq-cluster-operator-manager-668c99d594-pzcql" Dec 13 07:08:14 crc kubenswrapper[4971]: I1213 07:08:14.082163 4971 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"rabbitmq-cluster-operator-controller-manager-dockercfg-cjfw6" Dec 13 07:08:14 crc kubenswrapper[4971]: I1213 07:08:14.088910 4971 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/rabbitmq-cluster-operator-manager-668c99d594-pzcql"] Dec 13 07:08:14 crc kubenswrapper[4971]: I1213 07:08:14.140312 4971 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/designate-operator-controller-manager-66f8b87655-qxr27"] Dec 13 07:08:14 crc kubenswrapper[4971]: I1213 07:08:14.155085 4971 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/40f2727e-ef9d-4557-8619-b37174d00b67-metrics-certs\") pod \"openstack-operator-controller-manager-56f6fbdf6-vzrkl\" (UID: \"40f2727e-ef9d-4557-8619-b37174d00b67\") " pod="openstack-operators/openstack-operator-controller-manager-56f6fbdf6-vzrkl" Dec 13 07:08:14 crc kubenswrapper[4971]: I1213 07:08:14.155125 4971 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-5dlnw\" (UniqueName: \"kubernetes.io/projected/40f2727e-ef9d-4557-8619-b37174d00b67-kube-api-access-5dlnw\") pod \"openstack-operator-controller-manager-56f6fbdf6-vzrkl\" (UID: \"40f2727e-ef9d-4557-8619-b37174d00b67\") " pod="openstack-operators/openstack-operator-controller-manager-56f6fbdf6-vzrkl" Dec 13 07:08:14 crc kubenswrapper[4971]: I1213 07:08:14.155164 4971 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"webhook-certs\" (UniqueName: \"kubernetes.io/secret/40f2727e-ef9d-4557-8619-b37174d00b67-webhook-certs\") pod \"openstack-operator-controller-manager-56f6fbdf6-vzrkl\" (UID: \"40f2727e-ef9d-4557-8619-b37174d00b67\") " pod="openstack-operators/openstack-operator-controller-manager-56f6fbdf6-vzrkl" Dec 13 07:08:14 crc kubenswrapper[4971]: I1213 07:08:14.155304 4971 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-tkc2l\" (UniqueName: \"kubernetes.io/projected/14e41caa-c89a-4ca3-98c3-d75114cd2c76-kube-api-access-tkc2l\") pod \"rabbitmq-cluster-operator-manager-668c99d594-pzcql\" (UID: \"14e41caa-c89a-4ca3-98c3-d75114cd2c76\") " pod="openstack-operators/rabbitmq-cluster-operator-manager-668c99d594-pzcql" Dec 13 07:08:14 crc kubenswrapper[4971]: E1213 07:08:14.155468 4971 secret.go:188] Couldn't get secret openstack-operators/metrics-server-cert: secret "metrics-server-cert" not found Dec 13 07:08:14 crc kubenswrapper[4971]: E1213 07:08:14.155539 4971 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/40f2727e-ef9d-4557-8619-b37174d00b67-metrics-certs podName:40f2727e-ef9d-4557-8619-b37174d00b67 nodeName:}" failed. No retries permitted until 2025-12-13 07:08:14.655502715 +0000 UTC m=+1151.259912163 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "metrics-certs" (UniqueName: "kubernetes.io/secret/40f2727e-ef9d-4557-8619-b37174d00b67-metrics-certs") pod "openstack-operator-controller-manager-56f6fbdf6-vzrkl" (UID: "40f2727e-ef9d-4557-8619-b37174d00b67") : secret "metrics-server-cert" not found Dec 13 07:08:14 crc kubenswrapper[4971]: E1213 07:08:14.155874 4971 secret.go:188] Couldn't get secret openstack-operators/webhook-server-cert: secret "webhook-server-cert" not found Dec 13 07:08:14 crc kubenswrapper[4971]: E1213 07:08:14.155908 4971 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/40f2727e-ef9d-4557-8619-b37174d00b67-webhook-certs podName:40f2727e-ef9d-4557-8619-b37174d00b67 nodeName:}" failed. No retries permitted until 2025-12-13 07:08:14.655895054 +0000 UTC m=+1151.260304502 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "webhook-certs" (UniqueName: "kubernetes.io/secret/40f2727e-ef9d-4557-8619-b37174d00b67-webhook-certs") pod "openstack-operator-controller-manager-56f6fbdf6-vzrkl" (UID: "40f2727e-ef9d-4557-8619-b37174d00b67") : secret "webhook-server-cert" not found Dec 13 07:08:14 crc kubenswrapper[4971]: I1213 07:08:14.196537 4971 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-5dlnw\" (UniqueName: \"kubernetes.io/projected/40f2727e-ef9d-4557-8619-b37174d00b67-kube-api-access-5dlnw\") pod \"openstack-operator-controller-manager-56f6fbdf6-vzrkl\" (UID: \"40f2727e-ef9d-4557-8619-b37174d00b67\") " pod="openstack-operators/openstack-operator-controller-manager-56f6fbdf6-vzrkl" Dec 13 07:08:14 crc kubenswrapper[4971]: I1213 07:08:14.210827 4971 provider.go:102] Refreshing cache for provider: *credentialprovider.defaultDockerConfigProvider Dec 13 07:08:14 crc kubenswrapper[4971]: I1213 07:08:14.257051 4971 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-tkc2l\" (UniqueName: \"kubernetes.io/projected/14e41caa-c89a-4ca3-98c3-d75114cd2c76-kube-api-access-tkc2l\") pod \"rabbitmq-cluster-operator-manager-668c99d594-pzcql\" (UID: \"14e41caa-c89a-4ca3-98c3-d75114cd2c76\") " pod="openstack-operators/rabbitmq-cluster-operator-manager-668c99d594-pzcql" Dec 13 07:08:14 crc kubenswrapper[4971]: I1213 07:08:14.292214 4971 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-tkc2l\" (UniqueName: \"kubernetes.io/projected/14e41caa-c89a-4ca3-98c3-d75114cd2c76-kube-api-access-tkc2l\") pod \"rabbitmq-cluster-operator-manager-668c99d594-pzcql\" (UID: \"14e41caa-c89a-4ca3-98c3-d75114cd2c76\") " pod="openstack-operators/rabbitmq-cluster-operator-manager-668c99d594-pzcql" Dec 13 07:08:14 crc kubenswrapper[4971]: I1213 07:08:14.585012 4971 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/rabbitmq-cluster-operator-manager-668c99d594-pzcql" Dec 13 07:08:14 crc kubenswrapper[4971]: I1213 07:08:14.702506 4971 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/40f2727e-ef9d-4557-8619-b37174d00b67-metrics-certs\") pod \"openstack-operator-controller-manager-56f6fbdf6-vzrkl\" (UID: \"40f2727e-ef9d-4557-8619-b37174d00b67\") " pod="openstack-operators/openstack-operator-controller-manager-56f6fbdf6-vzrkl" Dec 13 07:08:14 crc kubenswrapper[4971]: I1213 07:08:14.703464 4971 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"webhook-certs\" (UniqueName: \"kubernetes.io/secret/40f2727e-ef9d-4557-8619-b37174d00b67-webhook-certs\") pod \"openstack-operator-controller-manager-56f6fbdf6-vzrkl\" (UID: \"40f2727e-ef9d-4557-8619-b37174d00b67\") " pod="openstack-operators/openstack-operator-controller-manager-56f6fbdf6-vzrkl" Dec 13 07:08:14 crc kubenswrapper[4971]: E1213 07:08:14.703591 4971 secret.go:188] Couldn't get secret openstack-operators/webhook-server-cert: secret "webhook-server-cert" not found Dec 13 07:08:14 crc kubenswrapper[4971]: E1213 07:08:14.703639 4971 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/40f2727e-ef9d-4557-8619-b37174d00b67-webhook-certs podName:40f2727e-ef9d-4557-8619-b37174d00b67 nodeName:}" failed. No retries permitted until 2025-12-13 07:08:15.703624633 +0000 UTC m=+1152.308034081 (durationBeforeRetry 1s). Error: MountVolume.SetUp failed for volume "webhook-certs" (UniqueName: "kubernetes.io/secret/40f2727e-ef9d-4557-8619-b37174d00b67-webhook-certs") pod "openstack-operator-controller-manager-56f6fbdf6-vzrkl" (UID: "40f2727e-ef9d-4557-8619-b37174d00b67") : secret "webhook-server-cert" not found Dec 13 07:08:14 crc kubenswrapper[4971]: E1213 07:08:14.703652 4971 secret.go:188] Couldn't get secret openstack-operators/metrics-server-cert: secret "metrics-server-cert" not found Dec 13 07:08:14 crc kubenswrapper[4971]: E1213 07:08:14.703732 4971 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/40f2727e-ef9d-4557-8619-b37174d00b67-metrics-certs podName:40f2727e-ef9d-4557-8619-b37174d00b67 nodeName:}" failed. No retries permitted until 2025-12-13 07:08:15.703706625 +0000 UTC m=+1152.308116073 (durationBeforeRetry 1s). Error: MountVolume.SetUp failed for volume "metrics-certs" (UniqueName: "kubernetes.io/secret/40f2727e-ef9d-4557-8619-b37174d00b67-metrics-certs") pod "openstack-operator-controller-manager-56f6fbdf6-vzrkl" (UID: "40f2727e-ef9d-4557-8619-b37174d00b67") : secret "metrics-server-cert" not found Dec 13 07:08:14 crc kubenswrapper[4971]: I1213 07:08:14.745160 4971 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/cinder-operator-controller-manager-5cf45c46bd-8sg42"] Dec 13 07:08:14 crc kubenswrapper[4971]: I1213 07:08:14.758904 4971 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/glance-operator-controller-manager-767f9d7567-tmc6q"] Dec 13 07:08:14 crc kubenswrapper[4971]: I1213 07:08:14.804540 4971 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/846e13b5-58bb-4a12-a77a-0a4a9b53247f-cert\") pod \"infra-operator-controller-manager-58944d7758-59mz9\" (UID: \"846e13b5-58bb-4a12-a77a-0a4a9b53247f\") " pod="openstack-operators/infra-operator-controller-manager-58944d7758-59mz9" Dec 13 07:08:14 crc kubenswrapper[4971]: E1213 07:08:14.805754 4971 secret.go:188] Couldn't get secret openstack-operators/infra-operator-webhook-server-cert: secret "infra-operator-webhook-server-cert" not found Dec 13 07:08:14 crc kubenswrapper[4971]: E1213 07:08:14.805805 4971 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/846e13b5-58bb-4a12-a77a-0a4a9b53247f-cert podName:846e13b5-58bb-4a12-a77a-0a4a9b53247f nodeName:}" failed. No retries permitted until 2025-12-13 07:08:16.805788568 +0000 UTC m=+1153.410198016 (durationBeforeRetry 2s). Error: MountVolume.SetUp failed for volume "cert" (UniqueName: "kubernetes.io/secret/846e13b5-58bb-4a12-a77a-0a4a9b53247f-cert") pod "infra-operator-controller-manager-58944d7758-59mz9" (UID: "846e13b5-58bb-4a12-a77a-0a4a9b53247f") : secret "infra-operator-webhook-server-cert" not found Dec 13 07:08:14 crc kubenswrapper[4971]: I1213 07:08:14.861081 4971 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/barbican-operator-controller-manager-95949466-6hkxl"] Dec 13 07:08:14 crc kubenswrapper[4971]: W1213 07:08:14.929777 4971 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod2062b86b_1412_4974_a153_802d9f390944.slice/crio-e3a0d772d629b6b49ddd9f4da1d3e0af9661328b36fd91fbf582bbcfe744a017 WatchSource:0}: Error finding container e3a0d772d629b6b49ddd9f4da1d3e0af9661328b36fd91fbf582bbcfe744a017: Status 404 returned error can't find the container with id e3a0d772d629b6b49ddd9f4da1d3e0af9661328b36fd91fbf582bbcfe744a017 Dec 13 07:08:15 crc kubenswrapper[4971]: I1213 07:08:15.103567 4971 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/glance-operator-controller-manager-767f9d7567-tmc6q" event={"ID":"2062b86b-1412-4974-a153-802d9f390944","Type":"ContainerStarted","Data":"e3a0d772d629b6b49ddd9f4da1d3e0af9661328b36fd91fbf582bbcfe744a017"} Dec 13 07:08:15 crc kubenswrapper[4971]: I1213 07:08:15.105072 4971 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/barbican-operator-controller-manager-95949466-6hkxl" event={"ID":"3b74a6fd-0613-4f48-b585-64570abdb09f","Type":"ContainerStarted","Data":"49503a3a47cd88aa3dcea55913344ba605a9f2210e0814bf82d5c393bef185bd"} Dec 13 07:08:15 crc kubenswrapper[4971]: I1213 07:08:15.106312 4971 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/designate-operator-controller-manager-66f8b87655-qxr27" event={"ID":"cfe5eaa1-4beb-49ff-a080-ab230292c8ce","Type":"ContainerStarted","Data":"49c3f2230568606d0263a1ad8297b94fc0f234a36b9eb57db82ccce687e9ff7c"} Dec 13 07:08:15 crc kubenswrapper[4971]: I1213 07:08:15.108277 4971 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/706451b8-644b-4151-b5b9-9975082e556f-cert\") pod \"openstack-baremetal-operator-controller-manager-689f887b54jnn8l\" (UID: \"706451b8-644b-4151-b5b9-9975082e556f\") " pod="openstack-operators/openstack-baremetal-operator-controller-manager-689f887b54jnn8l" Dec 13 07:08:15 crc kubenswrapper[4971]: E1213 07:08:15.108472 4971 secret.go:188] Couldn't get secret openstack-operators/openstack-baremetal-operator-webhook-server-cert: secret "openstack-baremetal-operator-webhook-server-cert" not found Dec 13 07:08:15 crc kubenswrapper[4971]: E1213 07:08:15.108573 4971 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/706451b8-644b-4151-b5b9-9975082e556f-cert podName:706451b8-644b-4151-b5b9-9975082e556f nodeName:}" failed. No retries permitted until 2025-12-13 07:08:17.108545508 +0000 UTC m=+1153.712954966 (durationBeforeRetry 2s). Error: MountVolume.SetUp failed for volume "cert" (UniqueName: "kubernetes.io/secret/706451b8-644b-4151-b5b9-9975082e556f-cert") pod "openstack-baremetal-operator-controller-manager-689f887b54jnn8l" (UID: "706451b8-644b-4151-b5b9-9975082e556f") : secret "openstack-baremetal-operator-webhook-server-cert" not found Dec 13 07:08:15 crc kubenswrapper[4971]: I1213 07:08:15.110219 4971 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/cinder-operator-controller-manager-5cf45c46bd-8sg42" event={"ID":"62c93900-e946-4684-9c40-36c1d8ff0468","Type":"ContainerStarted","Data":"083a4d533944c0edc6951c7aa0007266f7ef4a7a2ac1cd1bbc6b3f525ceda41e"} Dec 13 07:08:15 crc kubenswrapper[4971]: I1213 07:08:15.718134 4971 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/40f2727e-ef9d-4557-8619-b37174d00b67-metrics-certs\") pod \"openstack-operator-controller-manager-56f6fbdf6-vzrkl\" (UID: \"40f2727e-ef9d-4557-8619-b37174d00b67\") " pod="openstack-operators/openstack-operator-controller-manager-56f6fbdf6-vzrkl" Dec 13 07:08:15 crc kubenswrapper[4971]: I1213 07:08:15.718201 4971 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"webhook-certs\" (UniqueName: \"kubernetes.io/secret/40f2727e-ef9d-4557-8619-b37174d00b67-webhook-certs\") pod \"openstack-operator-controller-manager-56f6fbdf6-vzrkl\" (UID: \"40f2727e-ef9d-4557-8619-b37174d00b67\") " pod="openstack-operators/openstack-operator-controller-manager-56f6fbdf6-vzrkl" Dec 13 07:08:15 crc kubenswrapper[4971]: E1213 07:08:15.718292 4971 secret.go:188] Couldn't get secret openstack-operators/metrics-server-cert: secret "metrics-server-cert" not found Dec 13 07:08:15 crc kubenswrapper[4971]: E1213 07:08:15.718358 4971 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/40f2727e-ef9d-4557-8619-b37174d00b67-metrics-certs podName:40f2727e-ef9d-4557-8619-b37174d00b67 nodeName:}" failed. No retries permitted until 2025-12-13 07:08:17.718342206 +0000 UTC m=+1154.322751654 (durationBeforeRetry 2s). Error: MountVolume.SetUp failed for volume "metrics-certs" (UniqueName: "kubernetes.io/secret/40f2727e-ef9d-4557-8619-b37174d00b67-metrics-certs") pod "openstack-operator-controller-manager-56f6fbdf6-vzrkl" (UID: "40f2727e-ef9d-4557-8619-b37174d00b67") : secret "metrics-server-cert" not found Dec 13 07:08:15 crc kubenswrapper[4971]: E1213 07:08:15.718388 4971 secret.go:188] Couldn't get secret openstack-operators/webhook-server-cert: secret "webhook-server-cert" not found Dec 13 07:08:15 crc kubenswrapper[4971]: E1213 07:08:15.718458 4971 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/40f2727e-ef9d-4557-8619-b37174d00b67-webhook-certs podName:40f2727e-ef9d-4557-8619-b37174d00b67 nodeName:}" failed. No retries permitted until 2025-12-13 07:08:17.718437618 +0000 UTC m=+1154.322847136 (durationBeforeRetry 2s). Error: MountVolume.SetUp failed for volume "webhook-certs" (UniqueName: "kubernetes.io/secret/40f2727e-ef9d-4557-8619-b37174d00b67-webhook-certs") pod "openstack-operator-controller-manager-56f6fbdf6-vzrkl" (UID: "40f2727e-ef9d-4557-8619-b37174d00b67") : secret "webhook-server-cert" not found Dec 13 07:08:15 crc kubenswrapper[4971]: I1213 07:08:15.724252 4971 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/ironic-operator-controller-manager-f458558d7-m9grc"] Dec 13 07:08:15 crc kubenswrapper[4971]: I1213 07:08:15.731483 4971 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/keystone-operator-controller-manager-5c7cbf548f-qv7p8"] Dec 13 07:08:15 crc kubenswrapper[4971]: I1213 07:08:15.746466 4971 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/horizon-operator-controller-manager-6ccf486b9-bb2ft"] Dec 13 07:08:15 crc kubenswrapper[4971]: W1213 07:08:15.761806 4971 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod3e2847dd_dcdc_4740_a4a8_4a2417247dc5.slice/crio-f7345b36af6ba0d0d17c9ad0247f5e6eb834667615838dfe8a49878cf7e1c5b2 WatchSource:0}: Error finding container f7345b36af6ba0d0d17c9ad0247f5e6eb834667615838dfe8a49878cf7e1c5b2: Status 404 returned error can't find the container with id f7345b36af6ba0d0d17c9ad0247f5e6eb834667615838dfe8a49878cf7e1c5b2 Dec 13 07:08:15 crc kubenswrapper[4971]: I1213 07:08:15.875450 4971 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/manila-operator-controller-manager-5fdd9786f7-jjw6m"] Dec 13 07:08:15 crc kubenswrapper[4971]: I1213 07:08:15.906098 4971 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/neutron-operator-controller-manager-7cd87b778f-l6b8q"] Dec 13 07:08:15 crc kubenswrapper[4971]: I1213 07:08:15.918266 4971 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/ovn-operator-controller-manager-bf6d4f946-g2r89"] Dec 13 07:08:15 crc kubenswrapper[4971]: W1213 07:08:15.929158 4971 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod0c834237_c914_43bc_b312_b208ba303bb3.slice/crio-38d43a768ae3033075296d3f45f53b9627141663a424c93a94b006ca1718a313 WatchSource:0}: Error finding container 38d43a768ae3033075296d3f45f53b9627141663a424c93a94b006ca1718a313: Status 404 returned error can't find the container with id 38d43a768ae3033075296d3f45f53b9627141663a424c93a94b006ca1718a313 Dec 13 07:08:15 crc kubenswrapper[4971]: I1213 07:08:15.942129 4971 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/heat-operator-controller-manager-59b8dcb766-4xg9v"] Dec 13 07:08:15 crc kubenswrapper[4971]: I1213 07:08:15.952264 4971 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/watcher-operator-controller-manager-55f78b7c4c-dwmgr"] Dec 13 07:08:15 crc kubenswrapper[4971]: W1213 07:08:15.957075 4971 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod808957a1_29e0_4991_bc8a_23f5a17ecb74.slice/crio-0c3ef5a5534f81f755e7f248498ab06d48d38de2196e70e9ccc48a71f5e4f7eb WatchSource:0}: Error finding container 0c3ef5a5534f81f755e7f248498ab06d48d38de2196e70e9ccc48a71f5e4f7eb: Status 404 returned error can't find the container with id 0c3ef5a5534f81f755e7f248498ab06d48d38de2196e70e9ccc48a71f5e4f7eb Dec 13 07:08:15 crc kubenswrapper[4971]: W1213 07:08:15.964134 4971 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod9ed7aa8c_f56d_477d_91f5_ef1944e66d57.slice/crio-f7a1e0be50afe2e392a06399b1cd2abb6930c669df263692ac552d19968b2e9d WatchSource:0}: Error finding container f7a1e0be50afe2e392a06399b1cd2abb6930c669df263692ac552d19968b2e9d: Status 404 returned error can't find the container with id f7a1e0be50afe2e392a06399b1cd2abb6930c669df263692ac552d19968b2e9d Dec 13 07:08:16 crc kubenswrapper[4971]: I1213 07:08:16.117600 4971 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/mariadb-operator-controller-manager-f76f4954c-cst7r"] Dec 13 07:08:16 crc kubenswrapper[4971]: I1213 07:08:16.125361 4971 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/rabbitmq-cluster-operator-manager-668c99d594-pzcql"] Dec 13 07:08:16 crc kubenswrapper[4971]: W1213 07:08:16.142111 4971 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod4bd9b122_00a8_4895_a44b_0773aae9f6a9.slice/crio-7680a526bf327cd91246b89a89c17c6156655e83625278fb2130da321f8a83cd WatchSource:0}: Error finding container 7680a526bf327cd91246b89a89c17c6156655e83625278fb2130da321f8a83cd: Status 404 returned error can't find the container with id 7680a526bf327cd91246b89a89c17c6156655e83625278fb2130da321f8a83cd Dec 13 07:08:16 crc kubenswrapper[4971]: I1213 07:08:16.153990 4971 patch_prober.go:28] interesting pod/machine-config-daemon-82xjz container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 13 07:08:16 crc kubenswrapper[4971]: I1213 07:08:16.154048 4971 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-82xjz" podUID="e0c4af12-edda-41d7-9b44-f87396174ef2" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 13 07:08:16 crc kubenswrapper[4971]: I1213 07:08:16.154099 4971 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-82xjz" Dec 13 07:08:16 crc kubenswrapper[4971]: I1213 07:08:16.154150 4971 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/placement-operator-controller-manager-8665b56d78-hrflw"] Dec 13 07:08:16 crc kubenswrapper[4971]: I1213 07:08:16.154705 4971 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"21a30fd79dbe0c677372d180191864bac6880bbd1aaac9bae997b690cf3d4a99"} pod="openshift-machine-config-operator/machine-config-daemon-82xjz" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Dec 13 07:08:16 crc kubenswrapper[4971]: I1213 07:08:16.154765 4971 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-82xjz" podUID="e0c4af12-edda-41d7-9b44-f87396174ef2" containerName="machine-config-daemon" containerID="cri-o://21a30fd79dbe0c677372d180191864bac6880bbd1aaac9bae997b690cf3d4a99" gracePeriod=600 Dec 13 07:08:16 crc kubenswrapper[4971]: I1213 07:08:16.161815 4971 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/heat-operator-controller-manager-59b8dcb766-4xg9v" event={"ID":"808957a1-29e0-4991-bc8a-23f5a17ecb74","Type":"ContainerStarted","Data":"0c3ef5a5534f81f755e7f248498ab06d48d38de2196e70e9ccc48a71f5e4f7eb"} Dec 13 07:08:16 crc kubenswrapper[4971]: I1213 07:08:16.172139 4971 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/manila-operator-controller-manager-5fdd9786f7-jjw6m" event={"ID":"2761a6c0-8c0d-4430-8976-378c2d6df9f7","Type":"ContainerStarted","Data":"16d7da4fdb3fb8171c56f6ed9259cf9e31e8de23541df9c2774a5f39cb99128e"} Dec 13 07:08:16 crc kubenswrapper[4971]: I1213 07:08:16.176129 4971 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/ovn-operator-controller-manager-bf6d4f946-g2r89" event={"ID":"9ed7aa8c-f56d-477d-91f5-ef1944e66d57","Type":"ContainerStarted","Data":"f7a1e0be50afe2e392a06399b1cd2abb6930c669df263692ac552d19968b2e9d"} Dec 13 07:08:16 crc kubenswrapper[4971]: I1213 07:08:16.190170 4971 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/horizon-operator-controller-manager-6ccf486b9-bb2ft" event={"ID":"8e94aa39-05a2-4a63-8b45-a914021bccd7","Type":"ContainerStarted","Data":"90f567a89009050bbc06381be60b42963b3bb22a6fe1ee931461a4bc99180d9f"} Dec 13 07:08:16 crc kubenswrapper[4971]: I1213 07:08:16.201784 4971 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/keystone-operator-controller-manager-5c7cbf548f-qv7p8" event={"ID":"3e2847dd-dcdc-4740-a4a8-4a2417247dc5","Type":"ContainerStarted","Data":"f7345b36af6ba0d0d17c9ad0247f5e6eb834667615838dfe8a49878cf7e1c5b2"} Dec 13 07:08:16 crc kubenswrapper[4971]: I1213 07:08:16.207330 4971 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/watcher-operator-controller-manager-55f78b7c4c-dwmgr" event={"ID":"5bb0c1cd-9a8e-4c03-80c0-d9e66dae91ed","Type":"ContainerStarted","Data":"262fdd6a786acaab3d82af23ad8663deb68aec24f0ed8276007600e54e1e3d96"} Dec 13 07:08:16 crc kubenswrapper[4971]: I1213 07:08:16.210469 4971 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/neutron-operator-controller-manager-7cd87b778f-l6b8q" event={"ID":"0c834237-c914-43bc-b312-b208ba303bb3","Type":"ContainerStarted","Data":"38d43a768ae3033075296d3f45f53b9627141663a424c93a94b006ca1718a313"} Dec 13 07:08:16 crc kubenswrapper[4971]: I1213 07:08:16.214116 4971 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/ironic-operator-controller-manager-f458558d7-m9grc" event={"ID":"85d76572-e53b-420f-9ffe-d0534bdb1c96","Type":"ContainerStarted","Data":"17c7eeb63f5c2edc5fc5432437a519b6fa0d9b71d46d998f569dfa29720fa851"} Dec 13 07:08:16 crc kubenswrapper[4971]: I1213 07:08:16.325194 4971 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/octavia-operator-controller-manager-68c649d9d-v9t9p"] Dec 13 07:08:16 crc kubenswrapper[4971]: I1213 07:08:16.335197 4971 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/swift-operator-controller-manager-5c6df8f9-rt5n5"] Dec 13 07:08:16 crc kubenswrapper[4971]: I1213 07:08:16.360651 4971 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/test-operator-controller-manager-756ccf86c7-r8p79"] Dec 13 07:08:16 crc kubenswrapper[4971]: E1213 07:08:16.370939 4971 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:manager,Image:quay.io/openstack-k8s-operators/swift-operator@sha256:3aa109bb973253ae9dcf339b9b65abbd1176cdb4be672c93e538a5f113816991,Command:[/manager],Args:[--leader-elect --health-probe-bind-address=:8081 --metrics-bind-address=127.0.0.1:8080],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:LEASE_DURATION,Value:30,ValueFrom:nil,},EnvVar{Name:RENEW_DEADLINE,Value:20,ValueFrom:nil,},EnvVar{Name:RETRY_PERIOD,Value:5,ValueFrom:nil,},EnvVar{Name:ENABLE_WEBHOOKS,Value:false,ValueFrom:nil,},EnvVar{Name:METRICS_CERTS,Value:false,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{cpu: {{500 -3} {} 500m DecimalSI},memory: {{536870912 0} {} BinarySI},},Requests:ResourceList{cpu: {{10 -3} {} 10m DecimalSI},memory: {{268435456 0} {} BinarySI},},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:kube-api-access-j9g4v,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/healthz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:15,TimeoutSeconds:1,PeriodSeconds:20,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},ReadinessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/readyz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:5,TimeoutSeconds:1,PeriodSeconds:10,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*1000660000,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod swift-operator-controller-manager-5c6df8f9-rt5n5_openstack-operators(bd48845c-5d55-4690-960b-ca319f596beb): ErrImagePull: pull QPS exceeded" logger="UnhandledError" Dec 13 07:08:16 crc kubenswrapper[4971]: E1213 07:08:16.371905 4971 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:manager,Image:quay.io/openstack-k8s-operators/nova-operator@sha256:779f0cee6024d0fb8f259b036fe790e62aa5a3b0431ea9bf15a6e7d02e2e5670,Command:[/manager],Args:[--leader-elect --health-probe-bind-address=:8081 --metrics-bind-address=127.0.0.1:8080],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:LEASE_DURATION,Value:30,ValueFrom:nil,},EnvVar{Name:RENEW_DEADLINE,Value:20,ValueFrom:nil,},EnvVar{Name:RETRY_PERIOD,Value:5,ValueFrom:nil,},EnvVar{Name:ENABLE_WEBHOOKS,Value:false,ValueFrom:nil,},EnvVar{Name:METRICS_CERTS,Value:false,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{cpu: {{500 -3} {} 500m DecimalSI},memory: {{536870912 0} {} BinarySI},},Requests:ResourceList{cpu: {{10 -3} {} 10m DecimalSI},memory: {{268435456 0} {} BinarySI},},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:kube-api-access-wbkzk,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/healthz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:15,TimeoutSeconds:1,PeriodSeconds:20,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},ReadinessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/readyz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:5,TimeoutSeconds:1,PeriodSeconds:10,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[MKNOD],},Privileged:nil,SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod nova-operator-controller-manager-5fbbf8b6cc-k4vtr_openstack-operators(a8909dd9-9995-4694-9b5b-47106910656c): ErrImagePull: pull QPS exceeded" logger="UnhandledError" Dec 13 07:08:16 crc kubenswrapper[4971]: E1213 07:08:16.372552 4971 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ErrImagePull: \"pull QPS exceeded\"" pod="openstack-operators/swift-operator-controller-manager-5c6df8f9-rt5n5" podUID="bd48845c-5d55-4690-960b-ca319f596beb" Dec 13 07:08:16 crc kubenswrapper[4971]: E1213 07:08:16.374318 4971 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ErrImagePull: \"pull QPS exceeded\"" pod="openstack-operators/nova-operator-controller-manager-5fbbf8b6cc-k4vtr" podUID="a8909dd9-9995-4694-9b5b-47106910656c" Dec 13 07:08:16 crc kubenswrapper[4971]: I1213 07:08:16.379649 4971 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/nova-operator-controller-manager-5fbbf8b6cc-k4vtr"] Dec 13 07:08:16 crc kubenswrapper[4971]: I1213 07:08:16.387138 4971 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/telemetry-operator-controller-manager-97d456b9-j4fpf"] Dec 13 07:08:16 crc kubenswrapper[4971]: E1213 07:08:16.418862 4971 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:manager,Image:quay.io/openstack-k8s-operators/telemetry-operator@sha256:f27e732ec1faee765461bf137d9be81278b2fa39675019a73622755e1e610b6f,Command:[/manager],Args:[--leader-elect --health-probe-bind-address=:8081 --metrics-bind-address=127.0.0.1:8080],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:LEASE_DURATION,Value:30,ValueFrom:nil,},EnvVar{Name:RENEW_DEADLINE,Value:20,ValueFrom:nil,},EnvVar{Name:RETRY_PERIOD,Value:5,ValueFrom:nil,},EnvVar{Name:ENABLE_WEBHOOKS,Value:false,ValueFrom:nil,},EnvVar{Name:METRICS_CERTS,Value:false,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{cpu: {{500 -3} {} 500m DecimalSI},memory: {{536870912 0} {} BinarySI},},Requests:ResourceList{cpu: {{10 -3} {} 10m DecimalSI},memory: {{268435456 0} {} BinarySI},},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:kube-api-access-9r8gf,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/healthz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:15,TimeoutSeconds:1,PeriodSeconds:20,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},ReadinessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/readyz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:5,TimeoutSeconds:1,PeriodSeconds:10,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[MKNOD],},Privileged:nil,SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod telemetry-operator-controller-manager-97d456b9-j4fpf_openstack-operators(f111629c-eb0b-4294-8f85-d4b0b6500454): ErrImagePull: pull QPS exceeded" logger="UnhandledError" Dec 13 07:08:16 crc kubenswrapper[4971]: E1213 07:08:16.420270 4971 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ErrImagePull: \"pull QPS exceeded\"" pod="openstack-operators/telemetry-operator-controller-manager-97d456b9-j4fpf" podUID="f111629c-eb0b-4294-8f85-d4b0b6500454" Dec 13 07:08:16 crc kubenswrapper[4971]: I1213 07:08:16.845605 4971 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/846e13b5-58bb-4a12-a77a-0a4a9b53247f-cert\") pod \"infra-operator-controller-manager-58944d7758-59mz9\" (UID: \"846e13b5-58bb-4a12-a77a-0a4a9b53247f\") " pod="openstack-operators/infra-operator-controller-manager-58944d7758-59mz9" Dec 13 07:08:16 crc kubenswrapper[4971]: E1213 07:08:16.845875 4971 secret.go:188] Couldn't get secret openstack-operators/infra-operator-webhook-server-cert: secret "infra-operator-webhook-server-cert" not found Dec 13 07:08:16 crc kubenswrapper[4971]: E1213 07:08:16.846252 4971 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/846e13b5-58bb-4a12-a77a-0a4a9b53247f-cert podName:846e13b5-58bb-4a12-a77a-0a4a9b53247f nodeName:}" failed. No retries permitted until 2025-12-13 07:08:20.846141418 +0000 UTC m=+1157.450550866 (durationBeforeRetry 4s). Error: MountVolume.SetUp failed for volume "cert" (UniqueName: "kubernetes.io/secret/846e13b5-58bb-4a12-a77a-0a4a9b53247f-cert") pod "infra-operator-controller-manager-58944d7758-59mz9" (UID: "846e13b5-58bb-4a12-a77a-0a4a9b53247f") : secret "infra-operator-webhook-server-cert" not found Dec 13 07:08:17 crc kubenswrapper[4971]: I1213 07:08:17.157177 4971 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/706451b8-644b-4151-b5b9-9975082e556f-cert\") pod \"openstack-baremetal-operator-controller-manager-689f887b54jnn8l\" (UID: \"706451b8-644b-4151-b5b9-9975082e556f\") " pod="openstack-operators/openstack-baremetal-operator-controller-manager-689f887b54jnn8l" Dec 13 07:08:17 crc kubenswrapper[4971]: E1213 07:08:17.157642 4971 secret.go:188] Couldn't get secret openstack-operators/openstack-baremetal-operator-webhook-server-cert: secret "openstack-baremetal-operator-webhook-server-cert" not found Dec 13 07:08:17 crc kubenswrapper[4971]: E1213 07:08:17.157686 4971 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/706451b8-644b-4151-b5b9-9975082e556f-cert podName:706451b8-644b-4151-b5b9-9975082e556f nodeName:}" failed. No retries permitted until 2025-12-13 07:08:21.157672666 +0000 UTC m=+1157.762082114 (durationBeforeRetry 4s). Error: MountVolume.SetUp failed for volume "cert" (UniqueName: "kubernetes.io/secret/706451b8-644b-4151-b5b9-9975082e556f-cert") pod "openstack-baremetal-operator-controller-manager-689f887b54jnn8l" (UID: "706451b8-644b-4151-b5b9-9975082e556f") : secret "openstack-baremetal-operator-webhook-server-cert" not found Dec 13 07:08:17 crc kubenswrapper[4971]: I1213 07:08:17.224860 4971 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/test-operator-controller-manager-756ccf86c7-r8p79" event={"ID":"ee0625f5-5da4-4dc4-8b2e-f77c9f4459ed","Type":"ContainerStarted","Data":"a2a04b4fa316f8603115e52b0c5cbbdc67f569fb0b6fe26e0bc650d4a7867b4c"} Dec 13 07:08:17 crc kubenswrapper[4971]: I1213 07:08:17.227356 4971 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/placement-operator-controller-manager-8665b56d78-hrflw" event={"ID":"6b9f3caa-4280-43f3-8a05-14e72f09c384","Type":"ContainerStarted","Data":"24c4b1fc6c3118996659c5577197b6445e64771d1258a6f6e8a17eb57ebcf52c"} Dec 13 07:08:17 crc kubenswrapper[4971]: I1213 07:08:17.230552 4971 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/nova-operator-controller-manager-5fbbf8b6cc-k4vtr" event={"ID":"a8909dd9-9995-4694-9b5b-47106910656c","Type":"ContainerStarted","Data":"b43aae3962ac4980b08890fdc7c786af422ea1e5e3192a8a51f16a4651f4851c"} Dec 13 07:08:17 crc kubenswrapper[4971]: E1213 07:08:17.242086 4971 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/nova-operator@sha256:779f0cee6024d0fb8f259b036fe790e62aa5a3b0431ea9bf15a6e7d02e2e5670\\\"\"" pod="openstack-operators/nova-operator-controller-manager-5fbbf8b6cc-k4vtr" podUID="a8909dd9-9995-4694-9b5b-47106910656c" Dec 13 07:08:17 crc kubenswrapper[4971]: I1213 07:08:17.243692 4971 generic.go:334] "Generic (PLEG): container finished" podID="e0c4af12-edda-41d7-9b44-f87396174ef2" containerID="21a30fd79dbe0c677372d180191864bac6880bbd1aaac9bae997b690cf3d4a99" exitCode=0 Dec 13 07:08:17 crc kubenswrapper[4971]: I1213 07:08:17.243834 4971 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-82xjz" event={"ID":"e0c4af12-edda-41d7-9b44-f87396174ef2","Type":"ContainerDied","Data":"21a30fd79dbe0c677372d180191864bac6880bbd1aaac9bae997b690cf3d4a99"} Dec 13 07:08:17 crc kubenswrapper[4971]: I1213 07:08:17.243886 4971 scope.go:117] "RemoveContainer" containerID="c41da72089cbd159f1c07bf9d29b0bd93e1593b7236249fdd03c408a03b5a8ca" Dec 13 07:08:17 crc kubenswrapper[4971]: I1213 07:08:17.245648 4971 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/telemetry-operator-controller-manager-97d456b9-j4fpf" event={"ID":"f111629c-eb0b-4294-8f85-d4b0b6500454","Type":"ContainerStarted","Data":"1846337e904ed07b6b5d678a82291a652c07fdd97a0e41efabe8957b6406e889"} Dec 13 07:08:17 crc kubenswrapper[4971]: E1213 07:08:17.248299 4971 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/telemetry-operator@sha256:f27e732ec1faee765461bf137d9be81278b2fa39675019a73622755e1e610b6f\\\"\"" pod="openstack-operators/telemetry-operator-controller-manager-97d456b9-j4fpf" podUID="f111629c-eb0b-4294-8f85-d4b0b6500454" Dec 13 07:08:17 crc kubenswrapper[4971]: I1213 07:08:17.254347 4971 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/swift-operator-controller-manager-5c6df8f9-rt5n5" event={"ID":"bd48845c-5d55-4690-960b-ca319f596beb","Type":"ContainerStarted","Data":"894b4aebff093414e61c15d5576e9f89d26832a8cc4aaa7b6fb904f4730ef368"} Dec 13 07:08:17 crc kubenswrapper[4971]: E1213 07:08:17.255763 4971 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/swift-operator@sha256:3aa109bb973253ae9dcf339b9b65abbd1176cdb4be672c93e538a5f113816991\\\"\"" pod="openstack-operators/swift-operator-controller-manager-5c6df8f9-rt5n5" podUID="bd48845c-5d55-4690-960b-ca319f596beb" Dec 13 07:08:17 crc kubenswrapper[4971]: I1213 07:08:17.258915 4971 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/mariadb-operator-controller-manager-f76f4954c-cst7r" event={"ID":"4bd9b122-00a8-4895-a44b-0773aae9f6a9","Type":"ContainerStarted","Data":"7680a526bf327cd91246b89a89c17c6156655e83625278fb2130da321f8a83cd"} Dec 13 07:08:17 crc kubenswrapper[4971]: I1213 07:08:17.264823 4971 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/octavia-operator-controller-manager-68c649d9d-v9t9p" event={"ID":"8a160d6b-ce31-4d73-918b-368fbc44eb2b","Type":"ContainerStarted","Data":"19082e09ebd01b923aa2365d52d709c1c760085cec2e96e898601c89c590a451"} Dec 13 07:08:17 crc kubenswrapper[4971]: I1213 07:08:17.268628 4971 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/rabbitmq-cluster-operator-manager-668c99d594-pzcql" event={"ID":"14e41caa-c89a-4ca3-98c3-d75114cd2c76","Type":"ContainerStarted","Data":"e9913415273ddd0dd18353c6031e97c5f0e16fa00dc95a15c97c8708cc2bb3dc"} Dec 13 07:08:17 crc kubenswrapper[4971]: I1213 07:08:17.771711 4971 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"webhook-certs\" (UniqueName: \"kubernetes.io/secret/40f2727e-ef9d-4557-8619-b37174d00b67-webhook-certs\") pod \"openstack-operator-controller-manager-56f6fbdf6-vzrkl\" (UID: \"40f2727e-ef9d-4557-8619-b37174d00b67\") " pod="openstack-operators/openstack-operator-controller-manager-56f6fbdf6-vzrkl" Dec 13 07:08:17 crc kubenswrapper[4971]: E1213 07:08:17.776623 4971 secret.go:188] Couldn't get secret openstack-operators/webhook-server-cert: secret "webhook-server-cert" not found Dec 13 07:08:17 crc kubenswrapper[4971]: E1213 07:08:17.776705 4971 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/40f2727e-ef9d-4557-8619-b37174d00b67-webhook-certs podName:40f2727e-ef9d-4557-8619-b37174d00b67 nodeName:}" failed. No retries permitted until 2025-12-13 07:08:21.776688459 +0000 UTC m=+1158.381097907 (durationBeforeRetry 4s). Error: MountVolume.SetUp failed for volume "webhook-certs" (UniqueName: "kubernetes.io/secret/40f2727e-ef9d-4557-8619-b37174d00b67-webhook-certs") pod "openstack-operator-controller-manager-56f6fbdf6-vzrkl" (UID: "40f2727e-ef9d-4557-8619-b37174d00b67") : secret "webhook-server-cert" not found Dec 13 07:08:17 crc kubenswrapper[4971]: I1213 07:08:17.784299 4971 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/40f2727e-ef9d-4557-8619-b37174d00b67-metrics-certs\") pod \"openstack-operator-controller-manager-56f6fbdf6-vzrkl\" (UID: \"40f2727e-ef9d-4557-8619-b37174d00b67\") " pod="openstack-operators/openstack-operator-controller-manager-56f6fbdf6-vzrkl" Dec 13 07:08:17 crc kubenswrapper[4971]: E1213 07:08:17.784420 4971 secret.go:188] Couldn't get secret openstack-operators/metrics-server-cert: secret "metrics-server-cert" not found Dec 13 07:08:17 crc kubenswrapper[4971]: E1213 07:08:17.784626 4971 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/40f2727e-ef9d-4557-8619-b37174d00b67-metrics-certs podName:40f2727e-ef9d-4557-8619-b37174d00b67 nodeName:}" failed. No retries permitted until 2025-12-13 07:08:21.784606015 +0000 UTC m=+1158.389015463 (durationBeforeRetry 4s). Error: MountVolume.SetUp failed for volume "metrics-certs" (UniqueName: "kubernetes.io/secret/40f2727e-ef9d-4557-8619-b37174d00b67-metrics-certs") pod "openstack-operator-controller-manager-56f6fbdf6-vzrkl" (UID: "40f2727e-ef9d-4557-8619-b37174d00b67") : secret "metrics-server-cert" not found Dec 13 07:08:18 crc kubenswrapper[4971]: I1213 07:08:18.558123 4971 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-82xjz" event={"ID":"e0c4af12-edda-41d7-9b44-f87396174ef2","Type":"ContainerStarted","Data":"6543005a61441f4a045d5ad167cc1798a7b51524223dc500db29c8f9518b38f1"} Dec 13 07:08:18 crc kubenswrapper[4971]: E1213 07:08:18.559338 4971 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/nova-operator@sha256:779f0cee6024d0fb8f259b036fe790e62aa5a3b0431ea9bf15a6e7d02e2e5670\\\"\"" pod="openstack-operators/nova-operator-controller-manager-5fbbf8b6cc-k4vtr" podUID="a8909dd9-9995-4694-9b5b-47106910656c" Dec 13 07:08:18 crc kubenswrapper[4971]: E1213 07:08:18.559944 4971 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/telemetry-operator@sha256:f27e732ec1faee765461bf137d9be81278b2fa39675019a73622755e1e610b6f\\\"\"" pod="openstack-operators/telemetry-operator-controller-manager-97d456b9-j4fpf" podUID="f111629c-eb0b-4294-8f85-d4b0b6500454" Dec 13 07:08:18 crc kubenswrapper[4971]: E1213 07:08:18.559958 4971 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/swift-operator@sha256:3aa109bb973253ae9dcf339b9b65abbd1176cdb4be672c93e538a5f113816991\\\"\"" pod="openstack-operators/swift-operator-controller-manager-5c6df8f9-rt5n5" podUID="bd48845c-5d55-4690-960b-ca319f596beb" Dec 13 07:08:20 crc kubenswrapper[4971]: I1213 07:08:20.927182 4971 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/846e13b5-58bb-4a12-a77a-0a4a9b53247f-cert\") pod \"infra-operator-controller-manager-58944d7758-59mz9\" (UID: \"846e13b5-58bb-4a12-a77a-0a4a9b53247f\") " pod="openstack-operators/infra-operator-controller-manager-58944d7758-59mz9" Dec 13 07:08:20 crc kubenswrapper[4971]: E1213 07:08:20.927671 4971 secret.go:188] Couldn't get secret openstack-operators/infra-operator-webhook-server-cert: secret "infra-operator-webhook-server-cert" not found Dec 13 07:08:20 crc kubenswrapper[4971]: E1213 07:08:20.927853 4971 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/846e13b5-58bb-4a12-a77a-0a4a9b53247f-cert podName:846e13b5-58bb-4a12-a77a-0a4a9b53247f nodeName:}" failed. No retries permitted until 2025-12-13 07:08:28.927802707 +0000 UTC m=+1165.532212155 (durationBeforeRetry 8s). Error: MountVolume.SetUp failed for volume "cert" (UniqueName: "kubernetes.io/secret/846e13b5-58bb-4a12-a77a-0a4a9b53247f-cert") pod "infra-operator-controller-manager-58944d7758-59mz9" (UID: "846e13b5-58bb-4a12-a77a-0a4a9b53247f") : secret "infra-operator-webhook-server-cert" not found Dec 13 07:08:21 crc kubenswrapper[4971]: I1213 07:08:21.232130 4971 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/706451b8-644b-4151-b5b9-9975082e556f-cert\") pod \"openstack-baremetal-operator-controller-manager-689f887b54jnn8l\" (UID: \"706451b8-644b-4151-b5b9-9975082e556f\") " pod="openstack-operators/openstack-baremetal-operator-controller-manager-689f887b54jnn8l" Dec 13 07:08:21 crc kubenswrapper[4971]: E1213 07:08:21.232384 4971 secret.go:188] Couldn't get secret openstack-operators/openstack-baremetal-operator-webhook-server-cert: secret "openstack-baremetal-operator-webhook-server-cert" not found Dec 13 07:08:21 crc kubenswrapper[4971]: E1213 07:08:21.232438 4971 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/706451b8-644b-4151-b5b9-9975082e556f-cert podName:706451b8-644b-4151-b5b9-9975082e556f nodeName:}" failed. No retries permitted until 2025-12-13 07:08:29.23242154 +0000 UTC m=+1165.836830988 (durationBeforeRetry 8s). Error: MountVolume.SetUp failed for volume "cert" (UniqueName: "kubernetes.io/secret/706451b8-644b-4151-b5b9-9975082e556f-cert") pod "openstack-baremetal-operator-controller-manager-689f887b54jnn8l" (UID: "706451b8-644b-4151-b5b9-9975082e556f") : secret "openstack-baremetal-operator-webhook-server-cert" not found Dec 13 07:08:21 crc kubenswrapper[4971]: I1213 07:08:21.801347 4971 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/40f2727e-ef9d-4557-8619-b37174d00b67-metrics-certs\") pod \"openstack-operator-controller-manager-56f6fbdf6-vzrkl\" (UID: \"40f2727e-ef9d-4557-8619-b37174d00b67\") " pod="openstack-operators/openstack-operator-controller-manager-56f6fbdf6-vzrkl" Dec 13 07:08:21 crc kubenswrapper[4971]: I1213 07:08:21.801416 4971 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"webhook-certs\" (UniqueName: \"kubernetes.io/secret/40f2727e-ef9d-4557-8619-b37174d00b67-webhook-certs\") pod \"openstack-operator-controller-manager-56f6fbdf6-vzrkl\" (UID: \"40f2727e-ef9d-4557-8619-b37174d00b67\") " pod="openstack-operators/openstack-operator-controller-manager-56f6fbdf6-vzrkl" Dec 13 07:08:21 crc kubenswrapper[4971]: E1213 07:08:21.802274 4971 secret.go:188] Couldn't get secret openstack-operators/metrics-server-cert: secret "metrics-server-cert" not found Dec 13 07:08:21 crc kubenswrapper[4971]: E1213 07:08:21.802336 4971 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/40f2727e-ef9d-4557-8619-b37174d00b67-metrics-certs podName:40f2727e-ef9d-4557-8619-b37174d00b67 nodeName:}" failed. No retries permitted until 2025-12-13 07:08:29.802318345 +0000 UTC m=+1166.406727803 (durationBeforeRetry 8s). Error: MountVolume.SetUp failed for volume "metrics-certs" (UniqueName: "kubernetes.io/secret/40f2727e-ef9d-4557-8619-b37174d00b67-metrics-certs") pod "openstack-operator-controller-manager-56f6fbdf6-vzrkl" (UID: "40f2727e-ef9d-4557-8619-b37174d00b67") : secret "metrics-server-cert" not found Dec 13 07:08:21 crc kubenswrapper[4971]: E1213 07:08:21.802282 4971 secret.go:188] Couldn't get secret openstack-operators/webhook-server-cert: secret "webhook-server-cert" not found Dec 13 07:08:21 crc kubenswrapper[4971]: E1213 07:08:21.802370 4971 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/40f2727e-ef9d-4557-8619-b37174d00b67-webhook-certs podName:40f2727e-ef9d-4557-8619-b37174d00b67 nodeName:}" failed. No retries permitted until 2025-12-13 07:08:29.802362706 +0000 UTC m=+1166.406772164 (durationBeforeRetry 8s). Error: MountVolume.SetUp failed for volume "webhook-certs" (UniqueName: "kubernetes.io/secret/40f2727e-ef9d-4557-8619-b37174d00b67-webhook-certs") pod "openstack-operator-controller-manager-56f6fbdf6-vzrkl" (UID: "40f2727e-ef9d-4557-8619-b37174d00b67") : secret "webhook-server-cert" not found Dec 13 07:08:28 crc kubenswrapper[4971]: I1213 07:08:28.965648 4971 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/846e13b5-58bb-4a12-a77a-0a4a9b53247f-cert\") pod \"infra-operator-controller-manager-58944d7758-59mz9\" (UID: \"846e13b5-58bb-4a12-a77a-0a4a9b53247f\") " pod="openstack-operators/infra-operator-controller-manager-58944d7758-59mz9" Dec 13 07:08:28 crc kubenswrapper[4971]: E1213 07:08:28.965886 4971 secret.go:188] Couldn't get secret openstack-operators/infra-operator-webhook-server-cert: secret "infra-operator-webhook-server-cert" not found Dec 13 07:08:28 crc kubenswrapper[4971]: E1213 07:08:28.966369 4971 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/846e13b5-58bb-4a12-a77a-0a4a9b53247f-cert podName:846e13b5-58bb-4a12-a77a-0a4a9b53247f nodeName:}" failed. No retries permitted until 2025-12-13 07:08:44.9663473 +0000 UTC m=+1181.570756748 (durationBeforeRetry 16s). Error: MountVolume.SetUp failed for volume "cert" (UniqueName: "kubernetes.io/secret/846e13b5-58bb-4a12-a77a-0a4a9b53247f-cert") pod "infra-operator-controller-manager-58944d7758-59mz9" (UID: "846e13b5-58bb-4a12-a77a-0a4a9b53247f") : secret "infra-operator-webhook-server-cert" not found Dec 13 07:08:29 crc kubenswrapper[4971]: I1213 07:08:29.270882 4971 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/706451b8-644b-4151-b5b9-9975082e556f-cert\") pod \"openstack-baremetal-operator-controller-manager-689f887b54jnn8l\" (UID: \"706451b8-644b-4151-b5b9-9975082e556f\") " pod="openstack-operators/openstack-baremetal-operator-controller-manager-689f887b54jnn8l" Dec 13 07:08:29 crc kubenswrapper[4971]: E1213 07:08:29.271177 4971 secret.go:188] Couldn't get secret openstack-operators/openstack-baremetal-operator-webhook-server-cert: secret "openstack-baremetal-operator-webhook-server-cert" not found Dec 13 07:08:29 crc kubenswrapper[4971]: E1213 07:08:29.271274 4971 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/706451b8-644b-4151-b5b9-9975082e556f-cert podName:706451b8-644b-4151-b5b9-9975082e556f nodeName:}" failed. No retries permitted until 2025-12-13 07:08:45.271251279 +0000 UTC m=+1181.875660727 (durationBeforeRetry 16s). Error: MountVolume.SetUp failed for volume "cert" (UniqueName: "kubernetes.io/secret/706451b8-644b-4151-b5b9-9975082e556f-cert") pod "openstack-baremetal-operator-controller-manager-689f887b54jnn8l" (UID: "706451b8-644b-4151-b5b9-9975082e556f") : secret "openstack-baremetal-operator-webhook-server-cert" not found Dec 13 07:08:29 crc kubenswrapper[4971]: I1213 07:08:29.895893 4971 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"webhook-certs\" (UniqueName: \"kubernetes.io/secret/40f2727e-ef9d-4557-8619-b37174d00b67-webhook-certs\") pod \"openstack-operator-controller-manager-56f6fbdf6-vzrkl\" (UID: \"40f2727e-ef9d-4557-8619-b37174d00b67\") " pod="openstack-operators/openstack-operator-controller-manager-56f6fbdf6-vzrkl" Dec 13 07:08:29 crc kubenswrapper[4971]: I1213 07:08:29.896055 4971 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/40f2727e-ef9d-4557-8619-b37174d00b67-metrics-certs\") pod \"openstack-operator-controller-manager-56f6fbdf6-vzrkl\" (UID: \"40f2727e-ef9d-4557-8619-b37174d00b67\") " pod="openstack-operators/openstack-operator-controller-manager-56f6fbdf6-vzrkl" Dec 13 07:08:29 crc kubenswrapper[4971]: E1213 07:08:29.896177 4971 secret.go:188] Couldn't get secret openstack-operators/metrics-server-cert: secret "metrics-server-cert" not found Dec 13 07:08:29 crc kubenswrapper[4971]: E1213 07:08:29.896225 4971 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/40f2727e-ef9d-4557-8619-b37174d00b67-metrics-certs podName:40f2727e-ef9d-4557-8619-b37174d00b67 nodeName:}" failed. No retries permitted until 2025-12-13 07:08:45.896209996 +0000 UTC m=+1182.500619444 (durationBeforeRetry 16s). Error: MountVolume.SetUp failed for volume "metrics-certs" (UniqueName: "kubernetes.io/secret/40f2727e-ef9d-4557-8619-b37174d00b67-metrics-certs") pod "openstack-operator-controller-manager-56f6fbdf6-vzrkl" (UID: "40f2727e-ef9d-4557-8619-b37174d00b67") : secret "metrics-server-cert" not found Dec 13 07:08:29 crc kubenswrapper[4971]: I1213 07:08:29.923784 4971 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"webhook-certs\" (UniqueName: \"kubernetes.io/secret/40f2727e-ef9d-4557-8619-b37174d00b67-webhook-certs\") pod \"openstack-operator-controller-manager-56f6fbdf6-vzrkl\" (UID: \"40f2727e-ef9d-4557-8619-b37174d00b67\") " pod="openstack-operators/openstack-operator-controller-manager-56f6fbdf6-vzrkl" Dec 13 07:08:30 crc kubenswrapper[4971]: E1213 07:08:30.697635 4971 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying config: context canceled" image="quay.io/openstack-k8s-operators/cinder-operator@sha256:981b6a8f95934a86c5f10ef6e198b07265aeba7f11cf84b9ccd13dfaf06f3ca3" Dec 13 07:08:30 crc kubenswrapper[4971]: E1213 07:08:30.698113 4971 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:manager,Image:quay.io/openstack-k8s-operators/cinder-operator@sha256:981b6a8f95934a86c5f10ef6e198b07265aeba7f11cf84b9ccd13dfaf06f3ca3,Command:[/manager],Args:[--leader-elect --health-probe-bind-address=:8081 --metrics-bind-address=127.0.0.1:8080],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:LEASE_DURATION,Value:30,ValueFrom:nil,},EnvVar{Name:RENEW_DEADLINE,Value:20,ValueFrom:nil,},EnvVar{Name:RETRY_PERIOD,Value:5,ValueFrom:nil,},EnvVar{Name:ENABLE_WEBHOOKS,Value:false,ValueFrom:nil,},EnvVar{Name:METRICS_CERTS,Value:false,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{cpu: {{500 -3} {} 500m DecimalSI},memory: {{536870912 0} {} BinarySI},},Requests:ResourceList{cpu: {{10 -3} {} 10m DecimalSI},memory: {{268435456 0} {} BinarySI},},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:kube-api-access-7dz47,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/healthz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:15,TimeoutSeconds:1,PeriodSeconds:20,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},ReadinessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/readyz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:5,TimeoutSeconds:1,PeriodSeconds:10,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[MKNOD],},Privileged:nil,SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod cinder-operator-controller-manager-5cf45c46bd-8sg42_openstack-operators(62c93900-e946-4684-9c40-36c1d8ff0468): ErrImagePull: rpc error: code = Canceled desc = copying config: context canceled" logger="UnhandledError" Dec 13 07:08:30 crc kubenswrapper[4971]: E1213 07:08:30.699397 4971 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ErrImagePull: \"rpc error: code = Canceled desc = copying config: context canceled\"" pod="openstack-operators/cinder-operator-controller-manager-5cf45c46bd-8sg42" podUID="62c93900-e946-4684-9c40-36c1d8ff0468" Dec 13 07:08:31 crc kubenswrapper[4971]: E1213 07:08:31.011571 4971 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/cinder-operator@sha256:981b6a8f95934a86c5f10ef6e198b07265aeba7f11cf84b9ccd13dfaf06f3ca3\\\"\"" pod="openstack-operators/cinder-operator-controller-manager-5cf45c46bd-8sg42" podUID="62c93900-e946-4684-9c40-36c1d8ff0468" Dec 13 07:08:32 crc kubenswrapper[4971]: E1213 07:08:32.011084 4971 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying config: context canceled" image="quay.io/openstack-k8s-operators/barbican-operator@sha256:f6059a0fbf031d34dcf086d14ce8c0546caeaee23c5780e90b5037c5feee9fea" Dec 13 07:08:32 crc kubenswrapper[4971]: E1213 07:08:32.011271 4971 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:manager,Image:quay.io/openstack-k8s-operators/barbican-operator@sha256:f6059a0fbf031d34dcf086d14ce8c0546caeaee23c5780e90b5037c5feee9fea,Command:[/manager],Args:[--leader-elect --health-probe-bind-address=:8081 --metrics-bind-address=127.0.0.1:8080],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:LEASE_DURATION,Value:30,ValueFrom:nil,},EnvVar{Name:RENEW_DEADLINE,Value:20,ValueFrom:nil,},EnvVar{Name:RETRY_PERIOD,Value:5,ValueFrom:nil,},EnvVar{Name:ENABLE_WEBHOOKS,Value:false,ValueFrom:nil,},EnvVar{Name:METRICS_CERTS,Value:false,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{cpu: {{500 -3} {} 500m DecimalSI},memory: {{536870912 0} {} BinarySI},},Requests:ResourceList{cpu: {{10 -3} {} 10m DecimalSI},memory: {{268435456 0} {} BinarySI},},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:kube-api-access-5k2mg,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/healthz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:15,TimeoutSeconds:1,PeriodSeconds:20,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},ReadinessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/readyz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:5,TimeoutSeconds:1,PeriodSeconds:10,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[MKNOD],},Privileged:nil,SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod barbican-operator-controller-manager-95949466-6hkxl_openstack-operators(3b74a6fd-0613-4f48-b585-64570abdb09f): ErrImagePull: rpc error: code = Canceled desc = copying config: context canceled" logger="UnhandledError" Dec 13 07:08:32 crc kubenswrapper[4971]: E1213 07:08:32.013238 4971 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ErrImagePull: \"rpc error: code = Canceled desc = copying config: context canceled\"" pod="openstack-operators/barbican-operator-controller-manager-95949466-6hkxl" podUID="3b74a6fd-0613-4f48-b585-64570abdb09f" Dec 13 07:08:32 crc kubenswrapper[4971]: E1213 07:08:32.950343 4971 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/barbican-operator@sha256:f6059a0fbf031d34dcf086d14ce8c0546caeaee23c5780e90b5037c5feee9fea\\\"\"" pod="openstack-operators/barbican-operator-controller-manager-95949466-6hkxl" podUID="3b74a6fd-0613-4f48-b585-64570abdb09f" Dec 13 07:08:33 crc kubenswrapper[4971]: E1213 07:08:33.328839 4971 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying config: context canceled" image="quay.io/openstack-k8s-operators/heat-operator@sha256:c4abfc148600dfa85915f3dc911d988ea2335f26cb6b8d749fe79bfe53e5e429" Dec 13 07:08:33 crc kubenswrapper[4971]: E1213 07:08:33.329034 4971 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:manager,Image:quay.io/openstack-k8s-operators/heat-operator@sha256:c4abfc148600dfa85915f3dc911d988ea2335f26cb6b8d749fe79bfe53e5e429,Command:[/manager],Args:[--leader-elect --health-probe-bind-address=:8081 --metrics-bind-address=127.0.0.1:8080],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:LEASE_DURATION,Value:30,ValueFrom:nil,},EnvVar{Name:RENEW_DEADLINE,Value:20,ValueFrom:nil,},EnvVar{Name:RETRY_PERIOD,Value:5,ValueFrom:nil,},EnvVar{Name:ENABLE_WEBHOOKS,Value:false,ValueFrom:nil,},EnvVar{Name:METRICS_CERTS,Value:false,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{cpu: {{500 -3} {} 500m DecimalSI},memory: {{536870912 0} {} BinarySI},},Requests:ResourceList{cpu: {{10 -3} {} 10m DecimalSI},memory: {{268435456 0} {} BinarySI},},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:kube-api-access-8bbb8,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/healthz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:15,TimeoutSeconds:1,PeriodSeconds:20,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},ReadinessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/readyz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:5,TimeoutSeconds:1,PeriodSeconds:10,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[MKNOD],},Privileged:nil,SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod heat-operator-controller-manager-59b8dcb766-4xg9v_openstack-operators(808957a1-29e0-4991-bc8a-23f5a17ecb74): ErrImagePull: rpc error: code = Canceled desc = copying config: context canceled" logger="UnhandledError" Dec 13 07:08:33 crc kubenswrapper[4971]: E1213 07:08:33.330207 4971 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ErrImagePull: \"rpc error: code = Canceled desc = copying config: context canceled\"" pod="openstack-operators/heat-operator-controller-manager-59b8dcb766-4xg9v" podUID="808957a1-29e0-4991-bc8a-23f5a17ecb74" Dec 13 07:08:33 crc kubenswrapper[4971]: E1213 07:08:33.952938 4971 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/heat-operator@sha256:c4abfc148600dfa85915f3dc911d988ea2335f26cb6b8d749fe79bfe53e5e429\\\"\"" pod="openstack-operators/heat-operator-controller-manager-59b8dcb766-4xg9v" podUID="808957a1-29e0-4991-bc8a-23f5a17ecb74" Dec 13 07:08:34 crc kubenswrapper[4971]: E1213 07:08:34.050410 4971 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying config: context canceled" image="quay.io/openstack-k8s-operators/test-operator@sha256:101b3e007d8c9f2e183262d7712f986ad51256448099069bc14f1ea5f997ab94" Dec 13 07:08:34 crc kubenswrapper[4971]: E1213 07:08:34.050661 4971 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:manager,Image:quay.io/openstack-k8s-operators/test-operator@sha256:101b3e007d8c9f2e183262d7712f986ad51256448099069bc14f1ea5f997ab94,Command:[/manager],Args:[--leader-elect --health-probe-bind-address=:8081 --metrics-bind-address=127.0.0.1:8080],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:LEASE_DURATION,Value:30,ValueFrom:nil,},EnvVar{Name:RENEW_DEADLINE,Value:20,ValueFrom:nil,},EnvVar{Name:RETRY_PERIOD,Value:5,ValueFrom:nil,},EnvVar{Name:ENABLE_WEBHOOKS,Value:false,ValueFrom:nil,},EnvVar{Name:METRICS_CERTS,Value:false,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{cpu: {{500 -3} {} 500m DecimalSI},memory: {{536870912 0} {} BinarySI},},Requests:ResourceList{cpu: {{10 -3} {} 10m DecimalSI},memory: {{268435456 0} {} BinarySI},},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:kube-api-access-cl8ll,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/healthz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:15,TimeoutSeconds:1,PeriodSeconds:20,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},ReadinessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/readyz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:5,TimeoutSeconds:1,PeriodSeconds:10,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[MKNOD],},Privileged:nil,SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod test-operator-controller-manager-756ccf86c7-r8p79_openstack-operators(ee0625f5-5da4-4dc4-8b2e-f77c9f4459ed): ErrImagePull: rpc error: code = Canceled desc = copying config: context canceled" logger="UnhandledError" Dec 13 07:08:34 crc kubenswrapper[4971]: E1213 07:08:34.052181 4971 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ErrImagePull: \"rpc error: code = Canceled desc = copying config: context canceled\"" pod="openstack-operators/test-operator-controller-manager-756ccf86c7-r8p79" podUID="ee0625f5-5da4-4dc4-8b2e-f77c9f4459ed" Dec 13 07:08:34 crc kubenswrapper[4971]: E1213 07:08:34.958372 4971 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/test-operator@sha256:101b3e007d8c9f2e183262d7712f986ad51256448099069bc14f1ea5f997ab94\\\"\"" pod="openstack-operators/test-operator-controller-manager-756ccf86c7-r8p79" podUID="ee0625f5-5da4-4dc4-8b2e-f77c9f4459ed" Dec 13 07:08:35 crc kubenswrapper[4971]: E1213 07:08:35.558500 4971 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying config: context canceled" image="quay.io/openstack-k8s-operators/horizon-operator@sha256:9e847f4dbdea19ab997f32a02b3680a9bd966f9c705911645c3866a19fda9ea5" Dec 13 07:08:35 crc kubenswrapper[4971]: E1213 07:08:35.558781 4971 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:manager,Image:quay.io/openstack-k8s-operators/horizon-operator@sha256:9e847f4dbdea19ab997f32a02b3680a9bd966f9c705911645c3866a19fda9ea5,Command:[/manager],Args:[--leader-elect --health-probe-bind-address=:8081 --metrics-bind-address=127.0.0.1:8080],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:LEASE_DURATION,Value:30,ValueFrom:nil,},EnvVar{Name:RENEW_DEADLINE,Value:20,ValueFrom:nil,},EnvVar{Name:RETRY_PERIOD,Value:5,ValueFrom:nil,},EnvVar{Name:ENABLE_WEBHOOKS,Value:false,ValueFrom:nil,},EnvVar{Name:METRICS_CERTS,Value:false,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{cpu: {{500 -3} {} 500m DecimalSI},memory: {{536870912 0} {} BinarySI},},Requests:ResourceList{cpu: {{10 -3} {} 10m DecimalSI},memory: {{268435456 0} {} BinarySI},},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:kube-api-access-c9mpf,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/healthz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:15,TimeoutSeconds:1,PeriodSeconds:20,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},ReadinessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/readyz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:5,TimeoutSeconds:1,PeriodSeconds:10,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[MKNOD],},Privileged:nil,SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod horizon-operator-controller-manager-6ccf486b9-bb2ft_openstack-operators(8e94aa39-05a2-4a63-8b45-a914021bccd7): ErrImagePull: rpc error: code = Canceled desc = copying config: context canceled" logger="UnhandledError" Dec 13 07:08:35 crc kubenswrapper[4971]: E1213 07:08:35.559956 4971 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ErrImagePull: \"rpc error: code = Canceled desc = copying config: context canceled\"" pod="openstack-operators/horizon-operator-controller-manager-6ccf486b9-bb2ft" podUID="8e94aa39-05a2-4a63-8b45-a914021bccd7" Dec 13 07:08:35 crc kubenswrapper[4971]: E1213 07:08:35.965332 4971 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/horizon-operator@sha256:9e847f4dbdea19ab997f32a02b3680a9bd966f9c705911645c3866a19fda9ea5\\\"\"" pod="openstack-operators/horizon-operator-controller-manager-6ccf486b9-bb2ft" podUID="8e94aa39-05a2-4a63-8b45-a914021bccd7" Dec 13 07:08:36 crc kubenswrapper[4971]: E1213 07:08:36.635318 4971 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying config: context canceled" image="quay.io/openstack-k8s-operators/placement-operator@sha256:d29650b006da97eb9178fcc58f2eb9fead8c2b414fac18f86a3c3a1507488c4f" Dec 13 07:08:36 crc kubenswrapper[4971]: E1213 07:08:36.635568 4971 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:manager,Image:quay.io/openstack-k8s-operators/placement-operator@sha256:d29650b006da97eb9178fcc58f2eb9fead8c2b414fac18f86a3c3a1507488c4f,Command:[/manager],Args:[--leader-elect --health-probe-bind-address=:8081 --metrics-bind-address=127.0.0.1:8080],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:LEASE_DURATION,Value:30,ValueFrom:nil,},EnvVar{Name:RENEW_DEADLINE,Value:20,ValueFrom:nil,},EnvVar{Name:RETRY_PERIOD,Value:5,ValueFrom:nil,},EnvVar{Name:ENABLE_WEBHOOKS,Value:false,ValueFrom:nil,},EnvVar{Name:METRICS_CERTS,Value:false,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{cpu: {{500 -3} {} 500m DecimalSI},memory: {{536870912 0} {} BinarySI},},Requests:ResourceList{cpu: {{10 -3} {} 10m DecimalSI},memory: {{268435456 0} {} BinarySI},},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:kube-api-access-hm6sz,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/healthz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:15,TimeoutSeconds:1,PeriodSeconds:20,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},ReadinessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/readyz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:5,TimeoutSeconds:1,PeriodSeconds:10,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[MKNOD],},Privileged:nil,SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod placement-operator-controller-manager-8665b56d78-hrflw_openstack-operators(6b9f3caa-4280-43f3-8a05-14e72f09c384): ErrImagePull: rpc error: code = Canceled desc = copying config: context canceled" logger="UnhandledError" Dec 13 07:08:36 crc kubenswrapper[4971]: E1213 07:08:36.636851 4971 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ErrImagePull: \"rpc error: code = Canceled desc = copying config: context canceled\"" pod="openstack-operators/placement-operator-controller-manager-8665b56d78-hrflw" podUID="6b9f3caa-4280-43f3-8a05-14e72f09c384" Dec 13 07:08:36 crc kubenswrapper[4971]: E1213 07:08:36.977702 4971 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/placement-operator@sha256:d29650b006da97eb9178fcc58f2eb9fead8c2b414fac18f86a3c3a1507488c4f\\\"\"" pod="openstack-operators/placement-operator-controller-manager-8665b56d78-hrflw" podUID="6b9f3caa-4280-43f3-8a05-14e72f09c384" Dec 13 07:08:37 crc kubenswrapper[4971]: E1213 07:08:37.695863 4971 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying config: context canceled" image="quay.io/openstack-k8s-operators/octavia-operator@sha256:d9a3694865a7d54ee96397add18c3898886e98d079aa20876a0f4de1fa7a7168" Dec 13 07:08:37 crc kubenswrapper[4971]: E1213 07:08:37.696090 4971 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:manager,Image:quay.io/openstack-k8s-operators/octavia-operator@sha256:d9a3694865a7d54ee96397add18c3898886e98d079aa20876a0f4de1fa7a7168,Command:[/manager],Args:[--leader-elect --health-probe-bind-address=:8081 --metrics-bind-address=127.0.0.1:8080],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:LEASE_DURATION,Value:30,ValueFrom:nil,},EnvVar{Name:RENEW_DEADLINE,Value:20,ValueFrom:nil,},EnvVar{Name:RETRY_PERIOD,Value:5,ValueFrom:nil,},EnvVar{Name:ENABLE_WEBHOOKS,Value:false,ValueFrom:nil,},EnvVar{Name:METRICS_CERTS,Value:false,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{cpu: {{500 -3} {} 500m DecimalSI},memory: {{536870912 0} {} BinarySI},},Requests:ResourceList{cpu: {{10 -3} {} 10m DecimalSI},memory: {{268435456 0} {} BinarySI},},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:kube-api-access-7lqwg,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/healthz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:15,TimeoutSeconds:1,PeriodSeconds:20,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},ReadinessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/readyz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:5,TimeoutSeconds:1,PeriodSeconds:10,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[MKNOD],},Privileged:nil,SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod octavia-operator-controller-manager-68c649d9d-v9t9p_openstack-operators(8a160d6b-ce31-4d73-918b-368fbc44eb2b): ErrImagePull: rpc error: code = Canceled desc = copying config: context canceled" logger="UnhandledError" Dec 13 07:08:37 crc kubenswrapper[4971]: E1213 07:08:37.697474 4971 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ErrImagePull: \"rpc error: code = Canceled desc = copying config: context canceled\"" pod="openstack-operators/octavia-operator-controller-manager-68c649d9d-v9t9p" podUID="8a160d6b-ce31-4d73-918b-368fbc44eb2b" Dec 13 07:08:37 crc kubenswrapper[4971]: E1213 07:08:37.982552 4971 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/octavia-operator@sha256:d9a3694865a7d54ee96397add18c3898886e98d079aa20876a0f4de1fa7a7168\\\"\"" pod="openstack-operators/octavia-operator-controller-manager-68c649d9d-v9t9p" podUID="8a160d6b-ce31-4d73-918b-368fbc44eb2b" Dec 13 07:08:38 crc kubenswrapper[4971]: E1213 07:08:38.997003 4971 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying config: context canceled" image="quay.io/openstack-k8s-operators/neutron-operator@sha256:0b3fb69f35c151895d3dffd514974a9f9fe1c77c3bca69b78b81efb183cf4557" Dec 13 07:08:38 crc kubenswrapper[4971]: E1213 07:08:38.997794 4971 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:manager,Image:quay.io/openstack-k8s-operators/neutron-operator@sha256:0b3fb69f35c151895d3dffd514974a9f9fe1c77c3bca69b78b81efb183cf4557,Command:[/manager],Args:[--leader-elect --health-probe-bind-address=:8081 --metrics-bind-address=127.0.0.1:8080],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:LEASE_DURATION,Value:30,ValueFrom:nil,},EnvVar{Name:RENEW_DEADLINE,Value:20,ValueFrom:nil,},EnvVar{Name:RETRY_PERIOD,Value:5,ValueFrom:nil,},EnvVar{Name:ENABLE_WEBHOOKS,Value:false,ValueFrom:nil,},EnvVar{Name:METRICS_CERTS,Value:false,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{cpu: {{500 -3} {} 500m DecimalSI},memory: {{536870912 0} {} BinarySI},},Requests:ResourceList{cpu: {{10 -3} {} 10m DecimalSI},memory: {{268435456 0} {} BinarySI},},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:kube-api-access-x8db2,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/healthz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:15,TimeoutSeconds:1,PeriodSeconds:20,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},ReadinessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/readyz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:5,TimeoutSeconds:1,PeriodSeconds:10,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[MKNOD],},Privileged:nil,SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod neutron-operator-controller-manager-7cd87b778f-l6b8q_openstack-operators(0c834237-c914-43bc-b312-b208ba303bb3): ErrImagePull: rpc error: code = Canceled desc = copying config: context canceled" logger="UnhandledError" Dec 13 07:08:38 crc kubenswrapper[4971]: E1213 07:08:38.998997 4971 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ErrImagePull: \"rpc error: code = Canceled desc = copying config: context canceled\"" pod="openstack-operators/neutron-operator-controller-manager-7cd87b778f-l6b8q" podUID="0c834237-c914-43bc-b312-b208ba303bb3" Dec 13 07:08:39 crc kubenswrapper[4971]: E1213 07:08:39.994253 4971 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/neutron-operator@sha256:0b3fb69f35c151895d3dffd514974a9f9fe1c77c3bca69b78b81efb183cf4557\\\"\"" pod="openstack-operators/neutron-operator-controller-manager-7cd87b778f-l6b8q" podUID="0c834237-c914-43bc-b312-b208ba303bb3" Dec 13 07:08:45 crc kubenswrapper[4971]: I1213 07:08:45.007840 4971 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/846e13b5-58bb-4a12-a77a-0a4a9b53247f-cert\") pod \"infra-operator-controller-manager-58944d7758-59mz9\" (UID: \"846e13b5-58bb-4a12-a77a-0a4a9b53247f\") " pod="openstack-operators/infra-operator-controller-manager-58944d7758-59mz9" Dec 13 07:08:45 crc kubenswrapper[4971]: I1213 07:08:45.016875 4971 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cert\" (UniqueName: \"kubernetes.io/secret/846e13b5-58bb-4a12-a77a-0a4a9b53247f-cert\") pod \"infra-operator-controller-manager-58944d7758-59mz9\" (UID: \"846e13b5-58bb-4a12-a77a-0a4a9b53247f\") " pod="openstack-operators/infra-operator-controller-manager-58944d7758-59mz9" Dec 13 07:08:45 crc kubenswrapper[4971]: I1213 07:08:45.131888 4971 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/infra-operator-controller-manager-58944d7758-59mz9" Dec 13 07:08:45 crc kubenswrapper[4971]: I1213 07:08:45.312251 4971 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/706451b8-644b-4151-b5b9-9975082e556f-cert\") pod \"openstack-baremetal-operator-controller-manager-689f887b54jnn8l\" (UID: \"706451b8-644b-4151-b5b9-9975082e556f\") " pod="openstack-operators/openstack-baremetal-operator-controller-manager-689f887b54jnn8l" Dec 13 07:08:45 crc kubenswrapper[4971]: I1213 07:08:45.316933 4971 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cert\" (UniqueName: \"kubernetes.io/secret/706451b8-644b-4151-b5b9-9975082e556f-cert\") pod \"openstack-baremetal-operator-controller-manager-689f887b54jnn8l\" (UID: \"706451b8-644b-4151-b5b9-9975082e556f\") " pod="openstack-operators/openstack-baremetal-operator-controller-manager-689f887b54jnn8l" Dec 13 07:08:45 crc kubenswrapper[4971]: I1213 07:08:45.441950 4971 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/openstack-baremetal-operator-controller-manager-689f887b54jnn8l" Dec 13 07:08:45 crc kubenswrapper[4971]: I1213 07:08:45.920410 4971 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/40f2727e-ef9d-4557-8619-b37174d00b67-metrics-certs\") pod \"openstack-operator-controller-manager-56f6fbdf6-vzrkl\" (UID: \"40f2727e-ef9d-4557-8619-b37174d00b67\") " pod="openstack-operators/openstack-operator-controller-manager-56f6fbdf6-vzrkl" Dec 13 07:08:45 crc kubenswrapper[4971]: I1213 07:08:45.936728 4971 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/40f2727e-ef9d-4557-8619-b37174d00b67-metrics-certs\") pod \"openstack-operator-controller-manager-56f6fbdf6-vzrkl\" (UID: \"40f2727e-ef9d-4557-8619-b37174d00b67\") " pod="openstack-operators/openstack-operator-controller-manager-56f6fbdf6-vzrkl" Dec 13 07:08:46 crc kubenswrapper[4971]: I1213 07:08:46.112942 4971 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/openstack-operator-controller-manager-56f6fbdf6-vzrkl" Dec 13 07:08:49 crc kubenswrapper[4971]: E1213 07:08:49.974563 4971 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying config: context canceled" image="quay.io/openstack-k8s-operators/ovn-operator@sha256:635a4aef9d6f0b799e8ec91333dbb312160c001d05b3c63f614c124e0b67cb59" Dec 13 07:08:49 crc kubenswrapper[4971]: E1213 07:08:49.975244 4971 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:manager,Image:quay.io/openstack-k8s-operators/ovn-operator@sha256:635a4aef9d6f0b799e8ec91333dbb312160c001d05b3c63f614c124e0b67cb59,Command:[/manager],Args:[--leader-elect --health-probe-bind-address=:8081 --metrics-bind-address=127.0.0.1:8080],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:LEASE_DURATION,Value:30,ValueFrom:nil,},EnvVar{Name:RENEW_DEADLINE,Value:20,ValueFrom:nil,},EnvVar{Name:RETRY_PERIOD,Value:5,ValueFrom:nil,},EnvVar{Name:ENABLE_WEBHOOKS,Value:false,ValueFrom:nil,},EnvVar{Name:METRICS_CERTS,Value:false,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{cpu: {{500 -3} {} 500m DecimalSI},memory: {{536870912 0} {} BinarySI},},Requests:ResourceList{cpu: {{10 -3} {} 10m DecimalSI},memory: {{268435456 0} {} BinarySI},},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:kube-api-access-s66w8,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/healthz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:15,TimeoutSeconds:1,PeriodSeconds:20,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},ReadinessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/readyz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:5,TimeoutSeconds:1,PeriodSeconds:10,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[MKNOD],},Privileged:nil,SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod ovn-operator-controller-manager-bf6d4f946-g2r89_openstack-operators(9ed7aa8c-f56d-477d-91f5-ef1944e66d57): ErrImagePull: rpc error: code = Canceled desc = copying config: context canceled" logger="UnhandledError" Dec 13 07:08:49 crc kubenswrapper[4971]: E1213 07:08:49.976498 4971 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ErrImagePull: \"rpc error: code = Canceled desc = copying config: context canceled\"" pod="openstack-operators/ovn-operator-controller-manager-bf6d4f946-g2r89" podUID="9ed7aa8c-f56d-477d-91f5-ef1944e66d57" Dec 13 07:08:50 crc kubenswrapper[4971]: E1213 07:08:50.099308 4971 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/ovn-operator@sha256:635a4aef9d6f0b799e8ec91333dbb312160c001d05b3c63f614c124e0b67cb59\\\"\"" pod="openstack-operators/ovn-operator-controller-manager-bf6d4f946-g2r89" podUID="9ed7aa8c-f56d-477d-91f5-ef1944e66d57" Dec 13 07:08:53 crc kubenswrapper[4971]: E1213 07:08:53.371454 4971 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying config: context canceled" image="quay.io/openstack-k8s-operators/manila-operator@sha256:44126f9c6b1d2bf752ddf989e20a4fc4cc1c07723d4fcb78465ccb2f55da6b3a" Dec 13 07:08:53 crc kubenswrapper[4971]: E1213 07:08:53.372010 4971 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:manager,Image:quay.io/openstack-k8s-operators/manila-operator@sha256:44126f9c6b1d2bf752ddf989e20a4fc4cc1c07723d4fcb78465ccb2f55da6b3a,Command:[/manager],Args:[--leader-elect --health-probe-bind-address=:8081 --metrics-bind-address=127.0.0.1:8080],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:LEASE_DURATION,Value:30,ValueFrom:nil,},EnvVar{Name:RENEW_DEADLINE,Value:20,ValueFrom:nil,},EnvVar{Name:RETRY_PERIOD,Value:5,ValueFrom:nil,},EnvVar{Name:ENABLE_WEBHOOKS,Value:false,ValueFrom:nil,},EnvVar{Name:METRICS_CERTS,Value:false,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{cpu: {{500 -3} {} 500m DecimalSI},memory: {{536870912 0} {} BinarySI},},Requests:ResourceList{cpu: {{10 -3} {} 10m DecimalSI},memory: {{268435456 0} {} BinarySI},},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:kube-api-access-ptwqp,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/healthz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:15,TimeoutSeconds:1,PeriodSeconds:20,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},ReadinessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/readyz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:5,TimeoutSeconds:1,PeriodSeconds:10,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[MKNOD],},Privileged:nil,SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod manila-operator-controller-manager-5fdd9786f7-jjw6m_openstack-operators(2761a6c0-8c0d-4430-8976-378c2d6df9f7): ErrImagePull: rpc error: code = Canceled desc = copying config: context canceled" logger="UnhandledError" Dec 13 07:08:53 crc kubenswrapper[4971]: E1213 07:08:53.373353 4971 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ErrImagePull: \"rpc error: code = Canceled desc = copying config: context canceled\"" pod="openstack-operators/manila-operator-controller-manager-5fdd9786f7-jjw6m" podUID="2761a6c0-8c0d-4430-8976-378c2d6df9f7" Dec 13 07:08:54 crc kubenswrapper[4971]: E1213 07:08:54.139233 4971 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/manila-operator@sha256:44126f9c6b1d2bf752ddf989e20a4fc4cc1c07723d4fcb78465ccb2f55da6b3a\\\"\"" pod="openstack-operators/manila-operator-controller-manager-5fdd9786f7-jjw6m" podUID="2761a6c0-8c0d-4430-8976-378c2d6df9f7" Dec 13 07:08:57 crc kubenswrapper[4971]: E1213 07:08:57.497822 4971 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying config: context canceled" image="quay.io/openstack-k8s-operators/mariadb-operator@sha256:424da951f13f1fbe9083215dc9f5088f90676dd813f01fdf3c1a8639b61cbaad" Dec 13 07:08:57 crc kubenswrapper[4971]: E1213 07:08:57.498293 4971 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:manager,Image:quay.io/openstack-k8s-operators/mariadb-operator@sha256:424da951f13f1fbe9083215dc9f5088f90676dd813f01fdf3c1a8639b61cbaad,Command:[/manager],Args:[--leader-elect --health-probe-bind-address=:8081 --metrics-bind-address=127.0.0.1:8080],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:LEASE_DURATION,Value:30,ValueFrom:nil,},EnvVar{Name:RENEW_DEADLINE,Value:20,ValueFrom:nil,},EnvVar{Name:RETRY_PERIOD,Value:5,ValueFrom:nil,},EnvVar{Name:ENABLE_WEBHOOKS,Value:false,ValueFrom:nil,},EnvVar{Name:METRICS_CERTS,Value:false,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{cpu: {{500 -3} {} 500m DecimalSI},memory: {{536870912 0} {} BinarySI},},Requests:ResourceList{cpu: {{10 -3} {} 10m DecimalSI},memory: {{268435456 0} {} BinarySI},},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:kube-api-access-qkq5n,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/healthz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:15,TimeoutSeconds:1,PeriodSeconds:20,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},ReadinessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/readyz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:5,TimeoutSeconds:1,PeriodSeconds:10,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[MKNOD],},Privileged:nil,SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod mariadb-operator-controller-manager-f76f4954c-cst7r_openstack-operators(4bd9b122-00a8-4895-a44b-0773aae9f6a9): ErrImagePull: rpc error: code = Canceled desc = copying config: context canceled" logger="UnhandledError" Dec 13 07:08:57 crc kubenswrapper[4971]: E1213 07:08:57.499401 4971 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ErrImagePull: \"rpc error: code = Canceled desc = copying config: context canceled\"" pod="openstack-operators/mariadb-operator-controller-manager-f76f4954c-cst7r" podUID="4bd9b122-00a8-4895-a44b-0773aae9f6a9" Dec 13 07:08:58 crc kubenswrapper[4971]: E1213 07:08:58.165666 4971 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/mariadb-operator@sha256:424da951f13f1fbe9083215dc9f5088f90676dd813f01fdf3c1a8639b61cbaad\\\"\"" pod="openstack-operators/mariadb-operator-controller-manager-f76f4954c-cst7r" podUID="4bd9b122-00a8-4895-a44b-0773aae9f6a9" Dec 13 07:08:58 crc kubenswrapper[4971]: E1213 07:08:58.398922 4971 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying config: context canceled" image="quay.io/openstack-k8s-operators/ironic-operator@sha256:5bdb3685be3ddc1efd62e16aaf2fa96ead64315e26d52b1b2a7d8ac01baa1e87" Dec 13 07:08:58 crc kubenswrapper[4971]: E1213 07:08:58.399067 4971 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:manager,Image:quay.io/openstack-k8s-operators/ironic-operator@sha256:5bdb3685be3ddc1efd62e16aaf2fa96ead64315e26d52b1b2a7d8ac01baa1e87,Command:[/manager],Args:[--leader-elect --health-probe-bind-address=:8081 --metrics-bind-address=127.0.0.1:8080],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:LEASE_DURATION,Value:30,ValueFrom:nil,},EnvVar{Name:RENEW_DEADLINE,Value:20,ValueFrom:nil,},EnvVar{Name:RETRY_PERIOD,Value:5,ValueFrom:nil,},EnvVar{Name:ENABLE_WEBHOOKS,Value:false,ValueFrom:nil,},EnvVar{Name:METRICS_CERTS,Value:false,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{cpu: {{500 -3} {} 500m DecimalSI},memory: {{536870912 0} {} BinarySI},},Requests:ResourceList{cpu: {{10 -3} {} 10m DecimalSI},memory: {{268435456 0} {} BinarySI},},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:kube-api-access-688ml,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/healthz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:15,TimeoutSeconds:1,PeriodSeconds:20,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},ReadinessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/readyz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:5,TimeoutSeconds:1,PeriodSeconds:10,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[MKNOD],},Privileged:nil,SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod ironic-operator-controller-manager-f458558d7-m9grc_openstack-operators(85d76572-e53b-420f-9ffe-d0534bdb1c96): ErrImagePull: rpc error: code = Canceled desc = copying config: context canceled" logger="UnhandledError" Dec 13 07:08:58 crc kubenswrapper[4971]: E1213 07:08:58.400440 4971 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ErrImagePull: \"rpc error: code = Canceled desc = copying config: context canceled\"" pod="openstack-operators/ironic-operator-controller-manager-f458558d7-m9grc" podUID="85d76572-e53b-420f-9ffe-d0534bdb1c96" Dec 13 07:08:59 crc kubenswrapper[4971]: E1213 07:08:59.174411 4971 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/ironic-operator@sha256:5bdb3685be3ddc1efd62e16aaf2fa96ead64315e26d52b1b2a7d8ac01baa1e87\\\"\"" pod="openstack-operators/ironic-operator-controller-manager-f458558d7-m9grc" podUID="85d76572-e53b-420f-9ffe-d0534bdb1c96" Dec 13 07:08:59 crc kubenswrapper[4971]: E1213 07:08:59.544171 4971 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying config: context canceled" image="quay.io/openstack-k8s-operators/telemetry-operator@sha256:f27e732ec1faee765461bf137d9be81278b2fa39675019a73622755e1e610b6f" Dec 13 07:08:59 crc kubenswrapper[4971]: E1213 07:08:59.544370 4971 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:manager,Image:quay.io/openstack-k8s-operators/telemetry-operator@sha256:f27e732ec1faee765461bf137d9be81278b2fa39675019a73622755e1e610b6f,Command:[/manager],Args:[--leader-elect --health-probe-bind-address=:8081 --metrics-bind-address=127.0.0.1:8080],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:LEASE_DURATION,Value:30,ValueFrom:nil,},EnvVar{Name:RENEW_DEADLINE,Value:20,ValueFrom:nil,},EnvVar{Name:RETRY_PERIOD,Value:5,ValueFrom:nil,},EnvVar{Name:ENABLE_WEBHOOKS,Value:false,ValueFrom:nil,},EnvVar{Name:METRICS_CERTS,Value:false,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{cpu: {{500 -3} {} 500m DecimalSI},memory: {{536870912 0} {} BinarySI},},Requests:ResourceList{cpu: {{10 -3} {} 10m DecimalSI},memory: {{268435456 0} {} BinarySI},},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:kube-api-access-9r8gf,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/healthz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:15,TimeoutSeconds:1,PeriodSeconds:20,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},ReadinessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/readyz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:5,TimeoutSeconds:1,PeriodSeconds:10,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[MKNOD],},Privileged:nil,SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod telemetry-operator-controller-manager-97d456b9-j4fpf_openstack-operators(f111629c-eb0b-4294-8f85-d4b0b6500454): ErrImagePull: rpc error: code = Canceled desc = copying config: context canceled" logger="UnhandledError" Dec 13 07:08:59 crc kubenswrapper[4971]: E1213 07:08:59.545487 4971 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ErrImagePull: \"rpc error: code = Canceled desc = copying config: context canceled\"" pod="openstack-operators/telemetry-operator-controller-manager-97d456b9-j4fpf" podUID="f111629c-eb0b-4294-8f85-d4b0b6500454" Dec 13 07:09:00 crc kubenswrapper[4971]: E1213 07:09:00.785272 4971 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying config: context canceled" image="quay.io/openstack-k8s-operators/swift-operator@sha256:3aa109bb973253ae9dcf339b9b65abbd1176cdb4be672c93e538a5f113816991" Dec 13 07:09:00 crc kubenswrapper[4971]: E1213 07:09:00.785780 4971 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:manager,Image:quay.io/openstack-k8s-operators/swift-operator@sha256:3aa109bb973253ae9dcf339b9b65abbd1176cdb4be672c93e538a5f113816991,Command:[/manager],Args:[--leader-elect --health-probe-bind-address=:8081 --metrics-bind-address=127.0.0.1:8080],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:LEASE_DURATION,Value:30,ValueFrom:nil,},EnvVar{Name:RENEW_DEADLINE,Value:20,ValueFrom:nil,},EnvVar{Name:RETRY_PERIOD,Value:5,ValueFrom:nil,},EnvVar{Name:ENABLE_WEBHOOKS,Value:false,ValueFrom:nil,},EnvVar{Name:METRICS_CERTS,Value:false,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{cpu: {{500 -3} {} 500m DecimalSI},memory: {{536870912 0} {} BinarySI},},Requests:ResourceList{cpu: {{10 -3} {} 10m DecimalSI},memory: {{268435456 0} {} BinarySI},},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:kube-api-access-j9g4v,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/healthz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:15,TimeoutSeconds:1,PeriodSeconds:20,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},ReadinessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/readyz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:5,TimeoutSeconds:1,PeriodSeconds:10,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*1000660000,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod swift-operator-controller-manager-5c6df8f9-rt5n5_openstack-operators(bd48845c-5d55-4690-960b-ca319f596beb): ErrImagePull: rpc error: code = Canceled desc = copying config: context canceled" logger="UnhandledError" Dec 13 07:09:00 crc kubenswrapper[4971]: E1213 07:09:00.786938 4971 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ErrImagePull: \"rpc error: code = Canceled desc = copying config: context canceled\"" pod="openstack-operators/swift-operator-controller-manager-5c6df8f9-rt5n5" podUID="bd48845c-5d55-4690-960b-ca319f596beb" Dec 13 07:09:01 crc kubenswrapper[4971]: E1213 07:09:01.822408 4971 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying config: context canceled" image="quay.io/openstack-k8s-operators/keystone-operator@sha256:72ad6517987f674af0d0ae092cbb874aeae909c8b8b60188099c311762ebc8f7" Dec 13 07:09:01 crc kubenswrapper[4971]: E1213 07:09:01.822753 4971 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:manager,Image:quay.io/openstack-k8s-operators/keystone-operator@sha256:72ad6517987f674af0d0ae092cbb874aeae909c8b8b60188099c311762ebc8f7,Command:[/manager],Args:[--leader-elect --health-probe-bind-address=:8081 --metrics-bind-address=127.0.0.1:8080],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:LEASE_DURATION,Value:30,ValueFrom:nil,},EnvVar{Name:RENEW_DEADLINE,Value:20,ValueFrom:nil,},EnvVar{Name:RETRY_PERIOD,Value:5,ValueFrom:nil,},EnvVar{Name:ENABLE_WEBHOOKS,Value:false,ValueFrom:nil,},EnvVar{Name:METRICS_CERTS,Value:false,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{cpu: {{500 -3} {} 500m DecimalSI},memory: {{536870912 0} {} BinarySI},},Requests:ResourceList{cpu: {{10 -3} {} 10m DecimalSI},memory: {{268435456 0} {} BinarySI},},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:kube-api-access-x4d4c,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/healthz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:15,TimeoutSeconds:1,PeriodSeconds:20,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},ReadinessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/readyz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:5,TimeoutSeconds:1,PeriodSeconds:10,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[MKNOD],},Privileged:nil,SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod keystone-operator-controller-manager-5c7cbf548f-qv7p8_openstack-operators(3e2847dd-dcdc-4740-a4a8-4a2417247dc5): ErrImagePull: rpc error: code = Canceled desc = copying config: context canceled" logger="UnhandledError" Dec 13 07:09:01 crc kubenswrapper[4971]: E1213 07:09:01.823990 4971 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ErrImagePull: \"rpc error: code = Canceled desc = copying config: context canceled\"" pod="openstack-operators/keystone-operator-controller-manager-5c7cbf548f-qv7p8" podUID="3e2847dd-dcdc-4740-a4a8-4a2417247dc5" Dec 13 07:09:02 crc kubenswrapper[4971]: E1213 07:09:02.203034 4971 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/keystone-operator@sha256:72ad6517987f674af0d0ae092cbb874aeae909c8b8b60188099c311762ebc8f7\\\"\"" pod="openstack-operators/keystone-operator-controller-manager-5c7cbf548f-qv7p8" podUID="3e2847dd-dcdc-4740-a4a8-4a2417247dc5" Dec 13 07:09:02 crc kubenswrapper[4971]: E1213 07:09:02.419965 4971 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying config: context canceled" image="quay.io/openstack-k8s-operators/nova-operator@sha256:779f0cee6024d0fb8f259b036fe790e62aa5a3b0431ea9bf15a6e7d02e2e5670" Dec 13 07:09:02 crc kubenswrapper[4971]: E1213 07:09:02.420183 4971 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:manager,Image:quay.io/openstack-k8s-operators/nova-operator@sha256:779f0cee6024d0fb8f259b036fe790e62aa5a3b0431ea9bf15a6e7d02e2e5670,Command:[/manager],Args:[--leader-elect --health-probe-bind-address=:8081 --metrics-bind-address=127.0.0.1:8080],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:LEASE_DURATION,Value:30,ValueFrom:nil,},EnvVar{Name:RENEW_DEADLINE,Value:20,ValueFrom:nil,},EnvVar{Name:RETRY_PERIOD,Value:5,ValueFrom:nil,},EnvVar{Name:ENABLE_WEBHOOKS,Value:false,ValueFrom:nil,},EnvVar{Name:METRICS_CERTS,Value:false,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{cpu: {{500 -3} {} 500m DecimalSI},memory: {{536870912 0} {} BinarySI},},Requests:ResourceList{cpu: {{10 -3} {} 10m DecimalSI},memory: {{268435456 0} {} BinarySI},},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:kube-api-access-wbkzk,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/healthz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:15,TimeoutSeconds:1,PeriodSeconds:20,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},ReadinessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/readyz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:5,TimeoutSeconds:1,PeriodSeconds:10,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[MKNOD],},Privileged:nil,SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod nova-operator-controller-manager-5fbbf8b6cc-k4vtr_openstack-operators(a8909dd9-9995-4694-9b5b-47106910656c): ErrImagePull: rpc error: code = Canceled desc = copying config: context canceled" logger="UnhandledError" Dec 13 07:09:02 crc kubenswrapper[4971]: E1213 07:09:02.421561 4971 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ErrImagePull: \"rpc error: code = Canceled desc = copying config: context canceled\"" pod="openstack-operators/nova-operator-controller-manager-5fbbf8b6cc-k4vtr" podUID="a8909dd9-9995-4694-9b5b-47106910656c" Dec 13 07:09:03 crc kubenswrapper[4971]: E1213 07:09:03.185101 4971 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying config: context canceled" image="quay.io/openstack-k8s-operators/rabbitmq-cluster-operator@sha256:893e66303c1b0bc1d00a299a3f0380bad55c8dc813c8a1c6a4aab379f5aa12a2" Dec 13 07:09:03 crc kubenswrapper[4971]: E1213 07:09:03.185437 4971 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:operator,Image:quay.io/openstack-k8s-operators/rabbitmq-cluster-operator@sha256:893e66303c1b0bc1d00a299a3f0380bad55c8dc813c8a1c6a4aab379f5aa12a2,Command:[/manager],Args:[],WorkingDir:,Ports:[]ContainerPort{ContainerPort{Name:metrics,HostPort:0,ContainerPort:9782,Protocol:TCP,HostIP:,},},Env:[]EnvVar{EnvVar{Name:OPERATOR_NAMESPACE,Value:,ValueFrom:&EnvVarSource{FieldRef:&ObjectFieldSelector{APIVersion:v1,FieldPath:metadata.namespace,},ResourceFieldRef:nil,ConfigMapKeyRef:nil,SecretKeyRef:nil,},},EnvVar{Name:LEASE_DURATION,Value:30,ValueFrom:nil,},EnvVar{Name:RENEW_DEADLINE,Value:20,ValueFrom:nil,},EnvVar{Name:RETRY_PERIOD,Value:5,ValueFrom:nil,},EnvVar{Name:ENABLE_WEBHOOKS,Value:false,ValueFrom:nil,},EnvVar{Name:METRICS_CERTS,Value:false,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{cpu: {{200 -3} {} 200m DecimalSI},memory: {{524288000 0} {} 500Mi BinarySI},},Requests:ResourceList{cpu: {{5 -3} {} 5m DecimalSI},memory: {{67108864 0} {} BinarySI},},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:kube-api-access-tkc2l,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*1000660000,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod rabbitmq-cluster-operator-manager-668c99d594-pzcql_openstack-operators(14e41caa-c89a-4ca3-98c3-d75114cd2c76): ErrImagePull: rpc error: code = Canceled desc = copying config: context canceled" logger="UnhandledError" Dec 13 07:09:03 crc kubenswrapper[4971]: E1213 07:09:03.186841 4971 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"operator\" with ErrImagePull: \"rpc error: code = Canceled desc = copying config: context canceled\"" pod="openstack-operators/rabbitmq-cluster-operator-manager-668c99d594-pzcql" podUID="14e41caa-c89a-4ca3-98c3-d75114cd2c76" Dec 13 07:09:03 crc kubenswrapper[4971]: E1213 07:09:03.348665 4971 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"operator\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/rabbitmq-cluster-operator@sha256:893e66303c1b0bc1d00a299a3f0380bad55c8dc813c8a1c6a4aab379f5aa12a2\\\"\"" pod="openstack-operators/rabbitmq-cluster-operator-manager-668c99d594-pzcql" podUID="14e41caa-c89a-4ca3-98c3-d75114cd2c76" Dec 13 07:09:04 crc kubenswrapper[4971]: I1213 07:09:04.132641 4971 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/openstack-operator-controller-manager-56f6fbdf6-vzrkl"] Dec 13 07:09:04 crc kubenswrapper[4971]: W1213 07:09:04.151739 4971 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod40f2727e_ef9d_4557_8619_b37174d00b67.slice/crio-0c9ac2e739955959ef8f9147e7e29c63b9d6e5e34250ba82b10ed25c0125df46 WatchSource:0}: Error finding container 0c9ac2e739955959ef8f9147e7e29c63b9d6e5e34250ba82b10ed25c0125df46: Status 404 returned error can't find the container with id 0c9ac2e739955959ef8f9147e7e29c63b9d6e5e34250ba82b10ed25c0125df46 Dec 13 07:09:04 crc kubenswrapper[4971]: I1213 07:09:04.229253 4971 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/openstack-operator-controller-manager-56f6fbdf6-vzrkl" event={"ID":"40f2727e-ef9d-4557-8619-b37174d00b67","Type":"ContainerStarted","Data":"0c9ac2e739955959ef8f9147e7e29c63b9d6e5e34250ba82b10ed25c0125df46"} Dec 13 07:09:04 crc kubenswrapper[4971]: I1213 07:09:04.370937 4971 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/infra-operator-controller-manager-58944d7758-59mz9"] Dec 13 07:09:04 crc kubenswrapper[4971]: W1213 07:09:04.403660 4971 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod846e13b5_58bb_4a12_a77a_0a4a9b53247f.slice/crio-430f91cc6c4348f68d17671d9278b68df2eb6b46757cb8d7946a5c0b910e455f WatchSource:0}: Error finding container 430f91cc6c4348f68d17671d9278b68df2eb6b46757cb8d7946a5c0b910e455f: Status 404 returned error can't find the container with id 430f91cc6c4348f68d17671d9278b68df2eb6b46757cb8d7946a5c0b910e455f Dec 13 07:09:04 crc kubenswrapper[4971]: I1213 07:09:04.471818 4971 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/openstack-baremetal-operator-controller-manager-689f887b54jnn8l"] Dec 13 07:09:04 crc kubenswrapper[4971]: W1213 07:09:04.477844 4971 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod706451b8_644b_4151_b5b9_9975082e556f.slice/crio-1eaadc1ca156d67822a831287266bf1e63e775f87739d0a189dbcb47f010e162 WatchSource:0}: Error finding container 1eaadc1ca156d67822a831287266bf1e63e775f87739d0a189dbcb47f010e162: Status 404 returned error can't find the container with id 1eaadc1ca156d67822a831287266bf1e63e775f87739d0a189dbcb47f010e162 Dec 13 07:09:05 crc kubenswrapper[4971]: I1213 07:09:05.273461 4971 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/infra-operator-controller-manager-58944d7758-59mz9" event={"ID":"846e13b5-58bb-4a12-a77a-0a4a9b53247f","Type":"ContainerStarted","Data":"430f91cc6c4348f68d17671d9278b68df2eb6b46757cb8d7946a5c0b910e455f"} Dec 13 07:09:05 crc kubenswrapper[4971]: I1213 07:09:05.276405 4971 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/cinder-operator-controller-manager-5cf45c46bd-8sg42" event={"ID":"62c93900-e946-4684-9c40-36c1d8ff0468","Type":"ContainerStarted","Data":"e34877f3b438d54b575801e22acc946d973237a0dc03169016b09634efe34f6b"} Dec 13 07:09:05 crc kubenswrapper[4971]: I1213 07:09:05.276708 4971 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/cinder-operator-controller-manager-5cf45c46bd-8sg42" Dec 13 07:09:05 crc kubenswrapper[4971]: I1213 07:09:05.278907 4971 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/glance-operator-controller-manager-767f9d7567-tmc6q" event={"ID":"2062b86b-1412-4974-a153-802d9f390944","Type":"ContainerStarted","Data":"ec518ba40318305658f8a67d185421b53d9049a3798a79ba12e2cd02e300b777"} Dec 13 07:09:05 crc kubenswrapper[4971]: I1213 07:09:05.279038 4971 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/glance-operator-controller-manager-767f9d7567-tmc6q" Dec 13 07:09:05 crc kubenswrapper[4971]: I1213 07:09:05.280913 4971 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/barbican-operator-controller-manager-95949466-6hkxl" event={"ID":"3b74a6fd-0613-4f48-b585-64570abdb09f","Type":"ContainerStarted","Data":"9f5bc7e21b240021f90e324184cbc183d069b50a87accdbe89f6a068529f0631"} Dec 13 07:09:05 crc kubenswrapper[4971]: I1213 07:09:05.281145 4971 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/barbican-operator-controller-manager-95949466-6hkxl" Dec 13 07:09:05 crc kubenswrapper[4971]: I1213 07:09:05.282455 4971 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/openstack-baremetal-operator-controller-manager-689f887b54jnn8l" event={"ID":"706451b8-644b-4151-b5b9-9975082e556f","Type":"ContainerStarted","Data":"1eaadc1ca156d67822a831287266bf1e63e775f87739d0a189dbcb47f010e162"} Dec 13 07:09:05 crc kubenswrapper[4971]: I1213 07:09:05.284050 4971 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/horizon-operator-controller-manager-6ccf486b9-bb2ft" event={"ID":"8e94aa39-05a2-4a63-8b45-a914021bccd7","Type":"ContainerStarted","Data":"d2ac4fd866c66948c13096f043d72e5bdd86d70ae977b4c13e02bfb8ea0a388a"} Dec 13 07:09:05 crc kubenswrapper[4971]: I1213 07:09:05.284246 4971 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/horizon-operator-controller-manager-6ccf486b9-bb2ft" Dec 13 07:09:05 crc kubenswrapper[4971]: I1213 07:09:05.285500 4971 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/test-operator-controller-manager-756ccf86c7-r8p79" event={"ID":"ee0625f5-5da4-4dc4-8b2e-f77c9f4459ed","Type":"ContainerStarted","Data":"1da9e8e8e877521032e722c9cfa8f7135044e21263ebfa94f409fb3d06d6e672"} Dec 13 07:09:05 crc kubenswrapper[4971]: I1213 07:09:05.285711 4971 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/test-operator-controller-manager-756ccf86c7-r8p79" Dec 13 07:09:05 crc kubenswrapper[4971]: I1213 07:09:05.287268 4971 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/ovn-operator-controller-manager-bf6d4f946-g2r89" event={"ID":"9ed7aa8c-f56d-477d-91f5-ef1944e66d57","Type":"ContainerStarted","Data":"dfa633cd33c63d0d54e82579aba752227bd54fca95481001636001d1ac09e4e3"} Dec 13 07:09:05 crc kubenswrapper[4971]: I1213 07:09:05.287528 4971 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/ovn-operator-controller-manager-bf6d4f946-g2r89" Dec 13 07:09:05 crc kubenswrapper[4971]: I1213 07:09:05.288643 4971 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/openstack-operator-controller-manager-56f6fbdf6-vzrkl" event={"ID":"40f2727e-ef9d-4557-8619-b37174d00b67","Type":"ContainerStarted","Data":"8e80341c10d94622a78689e764e8c3e83a3e0dd27c630c2ab8f341fc43f7f401"} Dec 13 07:09:05 crc kubenswrapper[4971]: I1213 07:09:05.288749 4971 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/openstack-operator-controller-manager-56f6fbdf6-vzrkl" Dec 13 07:09:05 crc kubenswrapper[4971]: I1213 07:09:05.391355 4971 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/cinder-operator-controller-manager-5cf45c46bd-8sg42" podStartSLOduration=4.045761627 podStartE2EDuration="53.391334805s" podCreationTimestamp="2025-12-13 07:08:12 +0000 UTC" firstStartedPulling="2025-12-13 07:08:14.800286454 +0000 UTC m=+1151.404695902" lastFinishedPulling="2025-12-13 07:09:04.145859632 +0000 UTC m=+1200.750269080" observedRunningTime="2025-12-13 07:09:05.387650843 +0000 UTC m=+1201.992060291" watchObservedRunningTime="2025-12-13 07:09:05.391334805 +0000 UTC m=+1201.995744253" Dec 13 07:09:05 crc kubenswrapper[4971]: I1213 07:09:05.404203 4971 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/placement-operator-controller-manager-8665b56d78-hrflw" event={"ID":"6b9f3caa-4280-43f3-8a05-14e72f09c384","Type":"ContainerStarted","Data":"7ee11a4ec8c73a68f29f2259e3f69d69dd94538e62c273d310740e3a2f5c2267"} Dec 13 07:09:05 crc kubenswrapper[4971]: I1213 07:09:05.404434 4971 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/placement-operator-controller-manager-8665b56d78-hrflw" Dec 13 07:09:05 crc kubenswrapper[4971]: I1213 07:09:05.406867 4971 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/octavia-operator-controller-manager-68c649d9d-v9t9p" event={"ID":"8a160d6b-ce31-4d73-918b-368fbc44eb2b","Type":"ContainerStarted","Data":"ac635af9bd64ebe3054ca526191db50a1916e3db3194858e49a8f34a3a2a6754"} Dec 13 07:09:05 crc kubenswrapper[4971]: I1213 07:09:05.407480 4971 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/octavia-operator-controller-manager-68c649d9d-v9t9p" Dec 13 07:09:05 crc kubenswrapper[4971]: I1213 07:09:05.413271 4971 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/heat-operator-controller-manager-59b8dcb766-4xg9v" event={"ID":"808957a1-29e0-4991-bc8a-23f5a17ecb74","Type":"ContainerStarted","Data":"f211e53cfb44e7df5c1823b0d27ed06a5862c8ceefb3b88d609659c7c052c8ff"} Dec 13 07:09:05 crc kubenswrapper[4971]: I1213 07:09:05.413493 4971 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/heat-operator-controller-manager-59b8dcb766-4xg9v" Dec 13 07:09:05 crc kubenswrapper[4971]: I1213 07:09:05.415557 4971 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/watcher-operator-controller-manager-55f78b7c4c-dwmgr" event={"ID":"5bb0c1cd-9a8e-4c03-80c0-d9e66dae91ed","Type":"ContainerStarted","Data":"68e6743a5394aa5697a56ffd33f0038c8c04214bbfc22aa4d982d5a03341223a"} Dec 13 07:09:05 crc kubenswrapper[4971]: I1213 07:09:05.415931 4971 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/watcher-operator-controller-manager-55f78b7c4c-dwmgr" Dec 13 07:09:05 crc kubenswrapper[4971]: I1213 07:09:05.437093 4971 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/neutron-operator-controller-manager-7cd87b778f-l6b8q" event={"ID":"0c834237-c914-43bc-b312-b208ba303bb3","Type":"ContainerStarted","Data":"1ad20c2cfe1b8de42362014de347b2d0d71afbefad715d1652dfa52a46e3a9ef"} Dec 13 07:09:05 crc kubenswrapper[4971]: I1213 07:09:05.437391 4971 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/neutron-operator-controller-manager-7cd87b778f-l6b8q" Dec 13 07:09:05 crc kubenswrapper[4971]: I1213 07:09:05.510895 4971 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/designate-operator-controller-manager-66f8b87655-qxr27" event={"ID":"cfe5eaa1-4beb-49ff-a080-ab230292c8ce","Type":"ContainerStarted","Data":"dd24e10ea0326850da434911d7a9d799fc80cea3e898d57a73edf7bcfc758409"} Dec 13 07:09:05 crc kubenswrapper[4971]: I1213 07:09:05.511928 4971 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/designate-operator-controller-manager-66f8b87655-qxr27" Dec 13 07:09:05 crc kubenswrapper[4971]: I1213 07:09:05.629633 4971 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/horizon-operator-controller-manager-6ccf486b9-bb2ft" podStartSLOduration=5.234164326 podStartE2EDuration="53.629611304s" podCreationTimestamp="2025-12-13 07:08:12 +0000 UTC" firstStartedPulling="2025-12-13 07:08:15.764599371 +0000 UTC m=+1152.369008819" lastFinishedPulling="2025-12-13 07:09:04.160046349 +0000 UTC m=+1200.764455797" observedRunningTime="2025-12-13 07:09:05.437212101 +0000 UTC m=+1202.041621549" watchObservedRunningTime="2025-12-13 07:09:05.629611304 +0000 UTC m=+1202.234020752" Dec 13 07:09:05 crc kubenswrapper[4971]: I1213 07:09:05.629794 4971 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/openstack-operator-controller-manager-56f6fbdf6-vzrkl" podStartSLOduration=52.629785568 podStartE2EDuration="52.629785568s" podCreationTimestamp="2025-12-13 07:08:13 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-13 07:09:05.625325148 +0000 UTC m=+1202.229734596" watchObservedRunningTime="2025-12-13 07:09:05.629785568 +0000 UTC m=+1202.234195016" Dec 13 07:09:05 crc kubenswrapper[4971]: I1213 07:09:05.723101 4971 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/glance-operator-controller-manager-767f9d7567-tmc6q" podStartSLOduration=18.700615712 podStartE2EDuration="53.723069094s" podCreationTimestamp="2025-12-13 07:08:12 +0000 UTC" firstStartedPulling="2025-12-13 07:08:14.945115913 +0000 UTC m=+1151.549525361" lastFinishedPulling="2025-12-13 07:08:49.967569295 +0000 UTC m=+1186.571978743" observedRunningTime="2025-12-13 07:09:05.691479097 +0000 UTC m=+1202.295888545" watchObservedRunningTime="2025-12-13 07:09:05.723069094 +0000 UTC m=+1202.327478552" Dec 13 07:09:05 crc kubenswrapper[4971]: I1213 07:09:05.731277 4971 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/ovn-operator-controller-manager-bf6d4f946-g2r89" podStartSLOduration=4.534038951 podStartE2EDuration="52.731249057s" podCreationTimestamp="2025-12-13 07:08:13 +0000 UTC" firstStartedPulling="2025-12-13 07:08:15.969670076 +0000 UTC m=+1152.574079524" lastFinishedPulling="2025-12-13 07:09:04.166880182 +0000 UTC m=+1200.771289630" observedRunningTime="2025-12-13 07:09:05.718563043 +0000 UTC m=+1202.322972501" watchObservedRunningTime="2025-12-13 07:09:05.731249057 +0000 UTC m=+1202.335658505" Dec 13 07:09:05 crc kubenswrapper[4971]: I1213 07:09:05.756359 4971 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/barbican-operator-controller-manager-95949466-6hkxl" podStartSLOduration=4.55744863 podStartE2EDuration="53.756342378s" podCreationTimestamp="2025-12-13 07:08:12 +0000 UTC" firstStartedPulling="2025-12-13 07:08:14.96018413 +0000 UTC m=+1151.564593578" lastFinishedPulling="2025-12-13 07:09:04.159077878 +0000 UTC m=+1200.763487326" observedRunningTime="2025-12-13 07:09:05.749950165 +0000 UTC m=+1202.354359623" watchObservedRunningTime="2025-12-13 07:09:05.756342378 +0000 UTC m=+1202.360751826" Dec 13 07:09:05 crc kubenswrapper[4971]: I1213 07:09:05.787102 4971 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/test-operator-controller-manager-756ccf86c7-r8p79" podStartSLOduration=4.968625211 podStartE2EDuration="52.787076015s" podCreationTimestamp="2025-12-13 07:08:13 +0000 UTC" firstStartedPulling="2025-12-13 07:08:16.341458072 +0000 UTC m=+1152.945867520" lastFinishedPulling="2025-12-13 07:09:04.159908876 +0000 UTC m=+1200.764318324" observedRunningTime="2025-12-13 07:09:05.780313104 +0000 UTC m=+1202.384722552" watchObservedRunningTime="2025-12-13 07:09:05.787076015 +0000 UTC m=+1202.391485463" Dec 13 07:09:05 crc kubenswrapper[4971]: I1213 07:09:05.911741 4971 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/watcher-operator-controller-manager-55f78b7c4c-dwmgr" podStartSLOduration=10.469742363 podStartE2EDuration="52.911716612s" podCreationTimestamp="2025-12-13 07:08:13 +0000 UTC" firstStartedPulling="2025-12-13 07:08:15.952262647 +0000 UTC m=+1152.556672095" lastFinishedPulling="2025-12-13 07:08:58.394236906 +0000 UTC m=+1194.998646344" observedRunningTime="2025-12-13 07:09:05.850093424 +0000 UTC m=+1202.454502882" watchObservedRunningTime="2025-12-13 07:09:05.911716612 +0000 UTC m=+1202.516126060" Dec 13 07:09:06 crc kubenswrapper[4971]: I1213 07:09:06.087603 4971 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/designate-operator-controller-manager-66f8b87655-qxr27" podStartSLOduration=30.629979618 podStartE2EDuration="54.087573096s" podCreationTimestamp="2025-12-13 07:08:12 +0000 UTC" firstStartedPulling="2025-12-13 07:08:14.210553476 +0000 UTC m=+1150.814962924" lastFinishedPulling="2025-12-13 07:08:37.668146954 +0000 UTC m=+1174.272556402" observedRunningTime="2025-12-13 07:09:05.915897696 +0000 UTC m=+1202.520307144" watchObservedRunningTime="2025-12-13 07:09:06.087573096 +0000 UTC m=+1202.691982554" Dec 13 07:09:06 crc kubenswrapper[4971]: I1213 07:09:06.089415 4971 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/placement-operator-controller-manager-8665b56d78-hrflw" podStartSLOduration=5.356589306 podStartE2EDuration="53.089407416s" podCreationTimestamp="2025-12-13 07:08:13 +0000 UTC" firstStartedPulling="2025-12-13 07:08:16.176214366 +0000 UTC m=+1152.780623814" lastFinishedPulling="2025-12-13 07:09:03.909032476 +0000 UTC m=+1200.513441924" observedRunningTime="2025-12-13 07:09:05.977965045 +0000 UTC m=+1202.582374503" watchObservedRunningTime="2025-12-13 07:09:06.089407416 +0000 UTC m=+1202.693816874" Dec 13 07:09:06 crc kubenswrapper[4971]: I1213 07:09:06.190564 4971 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/heat-operator-controller-manager-59b8dcb766-4xg9v" podStartSLOduration=6.019421166 podStartE2EDuration="54.190537189s" podCreationTimestamp="2025-12-13 07:08:12 +0000 UTC" firstStartedPulling="2025-12-13 07:08:15.961758689 +0000 UTC m=+1152.566168137" lastFinishedPulling="2025-12-13 07:09:04.132874712 +0000 UTC m=+1200.737284160" observedRunningTime="2025-12-13 07:09:06.184932343 +0000 UTC m=+1202.789341811" watchObservedRunningTime="2025-12-13 07:09:06.190537189 +0000 UTC m=+1202.794946647" Dec 13 07:09:06 crc kubenswrapper[4971]: I1213 07:09:06.216246 4971 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/octavia-operator-controller-manager-68c649d9d-v9t9p" podStartSLOduration=5.438837276 podStartE2EDuration="53.216221583s" podCreationTimestamp="2025-12-13 07:08:13 +0000 UTC" firstStartedPulling="2025-12-13 07:08:16.367540565 +0000 UTC m=+1152.971950013" lastFinishedPulling="2025-12-13 07:09:04.144924872 +0000 UTC m=+1200.749334320" observedRunningTime="2025-12-13 07:09:06.214199117 +0000 UTC m=+1202.818608565" watchObservedRunningTime="2025-12-13 07:09:06.216221583 +0000 UTC m=+1202.820631031" Dec 13 07:09:06 crc kubenswrapper[4971]: I1213 07:09:06.369295 4971 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/neutron-operator-controller-manager-7cd87b778f-l6b8q" podStartSLOduration=6.163817475 podStartE2EDuration="54.367744341s" podCreationTimestamp="2025-12-13 07:08:12 +0000 UTC" firstStartedPulling="2025-12-13 07:08:15.93402558 +0000 UTC m=+1152.538435018" lastFinishedPulling="2025-12-13 07:09:04.137952436 +0000 UTC m=+1200.742361884" observedRunningTime="2025-12-13 07:09:06.353947762 +0000 UTC m=+1202.958357210" watchObservedRunningTime="2025-12-13 07:09:06.367744341 +0000 UTC m=+1202.972153789" Dec 13 07:09:08 crc kubenswrapper[4971]: I1213 07:09:08.610935 4971 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/manila-operator-controller-manager-5fdd9786f7-jjw6m" event={"ID":"2761a6c0-8c0d-4430-8976-378c2d6df9f7","Type":"ContainerStarted","Data":"ee3af1ee8c0ab06e7235742baf65e3334522339c36ebcb4438f7800fcc5606de"} Dec 13 07:09:08 crc kubenswrapper[4971]: I1213 07:09:08.612617 4971 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/manila-operator-controller-manager-5fdd9786f7-jjw6m" Dec 13 07:09:11 crc kubenswrapper[4971]: E1213 07:09:11.770358 4971 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/telemetry-operator@sha256:f27e732ec1faee765461bf137d9be81278b2fa39675019a73622755e1e610b6f\\\"\"" pod="openstack-operators/telemetry-operator-controller-manager-97d456b9-j4fpf" podUID="f111629c-eb0b-4294-8f85-d4b0b6500454" Dec 13 07:09:11 crc kubenswrapper[4971]: I1213 07:09:11.787386 4971 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/manila-operator-controller-manager-5fdd9786f7-jjw6m" podStartSLOduration=8.166360469 podStartE2EDuration="59.787370153s" podCreationTimestamp="2025-12-13 07:08:12 +0000 UTC" firstStartedPulling="2025-12-13 07:08:15.914185705 +0000 UTC m=+1152.518595153" lastFinishedPulling="2025-12-13 07:09:07.535195389 +0000 UTC m=+1204.139604837" observedRunningTime="2025-12-13 07:09:08.643667469 +0000 UTC m=+1205.248076917" watchObservedRunningTime="2025-12-13 07:09:11.787370153 +0000 UTC m=+1208.391779601" Dec 13 07:09:13 crc kubenswrapper[4971]: I1213 07:09:13.096570 4971 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/barbican-operator-controller-manager-95949466-6hkxl" Dec 13 07:09:13 crc kubenswrapper[4971]: I1213 07:09:13.118217 4971 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/cinder-operator-controller-manager-5cf45c46bd-8sg42" Dec 13 07:09:13 crc kubenswrapper[4971]: I1213 07:09:13.159421 4971 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/designate-operator-controller-manager-66f8b87655-qxr27" Dec 13 07:09:13 crc kubenswrapper[4971]: I1213 07:09:13.175015 4971 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/glance-operator-controller-manager-767f9d7567-tmc6q" Dec 13 07:09:13 crc kubenswrapper[4971]: I1213 07:09:13.472226 4971 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/manila-operator-controller-manager-5fdd9786f7-jjw6m" Dec 13 07:09:13 crc kubenswrapper[4971]: I1213 07:09:13.566939 4971 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/neutron-operator-controller-manager-7cd87b778f-l6b8q" Dec 13 07:09:13 crc kubenswrapper[4971]: I1213 07:09:13.577832 4971 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/heat-operator-controller-manager-59b8dcb766-4xg9v" Dec 13 07:09:13 crc kubenswrapper[4971]: I1213 07:09:13.608616 4971 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/horizon-operator-controller-manager-6ccf486b9-bb2ft" Dec 13 07:09:13 crc kubenswrapper[4971]: I1213 07:09:13.640553 4971 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/octavia-operator-controller-manager-68c649d9d-v9t9p" Dec 13 07:09:13 crc kubenswrapper[4971]: I1213 07:09:13.812322 4971 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/test-operator-controller-manager-756ccf86c7-r8p79" Dec 13 07:09:13 crc kubenswrapper[4971]: I1213 07:09:13.842158 4971 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/watcher-operator-controller-manager-55f78b7c4c-dwmgr" Dec 13 07:09:13 crc kubenswrapper[4971]: I1213 07:09:13.961771 4971 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/ovn-operator-controller-manager-bf6d4f946-g2r89" Dec 13 07:09:14 crc kubenswrapper[4971]: I1213 07:09:14.015016 4971 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/placement-operator-controller-manager-8665b56d78-hrflw" Dec 13 07:09:14 crc kubenswrapper[4971]: E1213 07:09:14.771348 4971 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/swift-operator@sha256:3aa109bb973253ae9dcf339b9b65abbd1176cdb4be672c93e538a5f113816991\\\"\"" pod="openstack-operators/swift-operator-controller-manager-5c6df8f9-rt5n5" podUID="bd48845c-5d55-4690-960b-ca319f596beb" Dec 13 07:09:14 crc kubenswrapper[4971]: E1213 07:09:14.773322 4971 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/nova-operator@sha256:779f0cee6024d0fb8f259b036fe790e62aa5a3b0431ea9bf15a6e7d02e2e5670\\\"\"" pod="openstack-operators/nova-operator-controller-manager-5fbbf8b6cc-k4vtr" podUID="a8909dd9-9995-4694-9b5b-47106910656c" Dec 13 07:09:16 crc kubenswrapper[4971]: I1213 07:09:16.123898 4971 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/openstack-operator-controller-manager-56f6fbdf6-vzrkl" Dec 13 07:09:17 crc kubenswrapper[4971]: I1213 07:09:17.828921 4971 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/openstack-baremetal-operator-controller-manager-689f887b54jnn8l" event={"ID":"706451b8-644b-4151-b5b9-9975082e556f","Type":"ContainerStarted","Data":"f381e22563bb3e20aeaafc63639d1586bc3c63a3b40086f43eef882c0afa4d0e"} Dec 13 07:09:17 crc kubenswrapper[4971]: I1213 07:09:17.830320 4971 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/openstack-baremetal-operator-controller-manager-689f887b54jnn8l" Dec 13 07:09:17 crc kubenswrapper[4971]: I1213 07:09:17.832000 4971 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/keystone-operator-controller-manager-5c7cbf548f-qv7p8" event={"ID":"3e2847dd-dcdc-4740-a4a8-4a2417247dc5","Type":"ContainerStarted","Data":"e76bbbf1d2998b79b84f4e18b6cdea797b151760aee7666b64daa465582e7446"} Dec 13 07:09:17 crc kubenswrapper[4971]: I1213 07:09:17.832570 4971 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/keystone-operator-controller-manager-5c7cbf548f-qv7p8" Dec 13 07:09:17 crc kubenswrapper[4971]: I1213 07:09:17.836685 4971 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/ironic-operator-controller-manager-f458558d7-m9grc" event={"ID":"85d76572-e53b-420f-9ffe-d0534bdb1c96","Type":"ContainerStarted","Data":"980535cadee257b4242c1576d645de4d14114079994ebe10395233b47875db20"} Dec 13 07:09:17 crc kubenswrapper[4971]: I1213 07:09:17.837353 4971 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/ironic-operator-controller-manager-f458558d7-m9grc" Dec 13 07:09:17 crc kubenswrapper[4971]: I1213 07:09:17.840833 4971 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/infra-operator-controller-manager-58944d7758-59mz9" event={"ID":"846e13b5-58bb-4a12-a77a-0a4a9b53247f","Type":"ContainerStarted","Data":"5ec20ab0699f6ab14228c952833a4a24735088ffba152d8cede103f36e0237b1"} Dec 13 07:09:17 crc kubenswrapper[4971]: I1213 07:09:17.841403 4971 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/infra-operator-controller-manager-58944d7758-59mz9" Dec 13 07:09:17 crc kubenswrapper[4971]: I1213 07:09:17.843897 4971 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/mariadb-operator-controller-manager-f76f4954c-cst7r" event={"ID":"4bd9b122-00a8-4895-a44b-0773aae9f6a9","Type":"ContainerStarted","Data":"c433d389e14d4babc14724efac0a876010469416a816f787953e245cc8053583"} Dec 13 07:09:17 crc kubenswrapper[4971]: I1213 07:09:17.844440 4971 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/mariadb-operator-controller-manager-f76f4954c-cst7r" Dec 13 07:09:17 crc kubenswrapper[4971]: I1213 07:09:17.878308 4971 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/openstack-baremetal-operator-controller-manager-689f887b54jnn8l" podStartSLOduration=52.606327213 podStartE2EDuration="1m4.878282608s" podCreationTimestamp="2025-12-13 07:08:13 +0000 UTC" firstStartedPulling="2025-12-13 07:09:04.480004355 +0000 UTC m=+1201.084413803" lastFinishedPulling="2025-12-13 07:09:16.75195975 +0000 UTC m=+1213.356369198" observedRunningTime="2025-12-13 07:09:17.878070204 +0000 UTC m=+1214.482479662" watchObservedRunningTime="2025-12-13 07:09:17.878282608 +0000 UTC m=+1214.482692056" Dec 13 07:09:17 crc kubenswrapper[4971]: I1213 07:09:17.903417 4971 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/mariadb-operator-controller-manager-f76f4954c-cst7r" podStartSLOduration=5.295370404 podStartE2EDuration="1m5.90339877s" podCreationTimestamp="2025-12-13 07:08:12 +0000 UTC" firstStartedPulling="2025-12-13 07:08:16.16164585 +0000 UTC m=+1152.766055298" lastFinishedPulling="2025-12-13 07:09:16.769674216 +0000 UTC m=+1213.374083664" observedRunningTime="2025-12-13 07:09:17.902584032 +0000 UTC m=+1214.506993490" watchObservedRunningTime="2025-12-13 07:09:17.90339877 +0000 UTC m=+1214.507808218" Dec 13 07:09:17 crc kubenswrapper[4971]: I1213 07:09:17.919215 4971 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/keystone-operator-controller-manager-5c7cbf548f-qv7p8" podStartSLOduration=4.936012507 podStartE2EDuration="1m5.919195333s" podCreationTimestamp="2025-12-13 07:08:12 +0000 UTC" firstStartedPulling="2025-12-13 07:08:15.782209464 +0000 UTC m=+1152.386618912" lastFinishedPulling="2025-12-13 07:09:16.76539229 +0000 UTC m=+1213.369801738" observedRunningTime="2025-12-13 07:09:17.918621171 +0000 UTC m=+1214.523030649" watchObservedRunningTime="2025-12-13 07:09:17.919195333 +0000 UTC m=+1214.523604781" Dec 13 07:09:17 crc kubenswrapper[4971]: I1213 07:09:17.946986 4971 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/ironic-operator-controller-manager-f458558d7-m9grc" podStartSLOduration=4.932558929 podStartE2EDuration="1m5.946970494s" podCreationTimestamp="2025-12-13 07:08:12 +0000 UTC" firstStartedPulling="2025-12-13 07:08:15.751342624 +0000 UTC m=+1152.355752072" lastFinishedPulling="2025-12-13 07:09:16.765754189 +0000 UTC m=+1213.370163637" observedRunningTime="2025-12-13 07:09:17.943579069 +0000 UTC m=+1214.547988527" watchObservedRunningTime="2025-12-13 07:09:17.946970494 +0000 UTC m=+1214.551379942" Dec 13 07:09:17 crc kubenswrapper[4971]: I1213 07:09:17.983706 4971 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/infra-operator-controller-manager-58944d7758-59mz9" podStartSLOduration=53.637962 podStartE2EDuration="1m5.983671205s" podCreationTimestamp="2025-12-13 07:08:12 +0000 UTC" firstStartedPulling="2025-12-13 07:09:04.406193034 +0000 UTC m=+1201.010602482" lastFinishedPulling="2025-12-13 07:09:16.751902239 +0000 UTC m=+1213.356311687" observedRunningTime="2025-12-13 07:09:17.976006634 +0000 UTC m=+1214.580416092" watchObservedRunningTime="2025-12-13 07:09:17.983671205 +0000 UTC m=+1214.588080663" Dec 13 07:09:20 crc kubenswrapper[4971]: I1213 07:09:20.868936 4971 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/rabbitmq-cluster-operator-manager-668c99d594-pzcql" event={"ID":"14e41caa-c89a-4ca3-98c3-d75114cd2c76","Type":"ContainerStarted","Data":"c568b5d7108ba3168d53d9614edaa2c6af3579750b7781f9a4b088446cb1e924"} Dec 13 07:09:23 crc kubenswrapper[4971]: I1213 07:09:23.356170 4971 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/keystone-operator-controller-manager-5c7cbf548f-qv7p8" Dec 13 07:09:23 crc kubenswrapper[4971]: I1213 07:09:23.375363 4971 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/rabbitmq-cluster-operator-manager-668c99d594-pzcql" podStartSLOduration=6.512868996 podStartE2EDuration="1m10.375342003s" podCreationTimestamp="2025-12-13 07:08:13 +0000 UTC" firstStartedPulling="2025-12-13 07:08:16.174963818 +0000 UTC m=+1152.779373266" lastFinishedPulling="2025-12-13 07:09:20.037436835 +0000 UTC m=+1216.641846273" observedRunningTime="2025-12-13 07:09:20.898597363 +0000 UTC m=+1217.503006831" watchObservedRunningTime="2025-12-13 07:09:23.375342003 +0000 UTC m=+1219.979751451" Dec 13 07:09:23 crc kubenswrapper[4971]: I1213 07:09:23.412334 4971 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/ironic-operator-controller-manager-f458558d7-m9grc" Dec 13 07:09:23 crc kubenswrapper[4971]: I1213 07:09:23.595463 4971 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/mariadb-operator-controller-manager-f76f4954c-cst7r" Dec 13 07:09:25 crc kubenswrapper[4971]: I1213 07:09:25.138863 4971 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/infra-operator-controller-manager-58944d7758-59mz9" Dec 13 07:09:25 crc kubenswrapper[4971]: I1213 07:09:25.446977 4971 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/openstack-baremetal-operator-controller-manager-689f887b54jnn8l" Dec 13 07:09:28 crc kubenswrapper[4971]: I1213 07:09:28.927673 4971 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/nova-operator-controller-manager-5fbbf8b6cc-k4vtr" event={"ID":"a8909dd9-9995-4694-9b5b-47106910656c","Type":"ContainerStarted","Data":"40d24cb0cba11d07ce434071a9b8a0678746b5bf38ac89f07ef7dd4fd72c3f3a"} Dec 13 07:09:28 crc kubenswrapper[4971]: I1213 07:09:28.928434 4971 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/nova-operator-controller-manager-5fbbf8b6cc-k4vtr" Dec 13 07:09:28 crc kubenswrapper[4971]: I1213 07:09:28.929187 4971 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/telemetry-operator-controller-manager-97d456b9-j4fpf" event={"ID":"f111629c-eb0b-4294-8f85-d4b0b6500454","Type":"ContainerStarted","Data":"b277b9cccd50bda49fa96cf561a4faa06f900bde0293660118ed9f77500f0da7"} Dec 13 07:09:28 crc kubenswrapper[4971]: I1213 07:09:28.929448 4971 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/telemetry-operator-controller-manager-97d456b9-j4fpf" Dec 13 07:09:28 crc kubenswrapper[4971]: I1213 07:09:28.971081 4971 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/nova-operator-controller-manager-5fbbf8b6cc-k4vtr" podStartSLOduration=4.129157508 podStartE2EDuration="1m15.971063845s" podCreationTimestamp="2025-12-13 07:08:13 +0000 UTC" firstStartedPulling="2025-12-13 07:08:16.37180717 +0000 UTC m=+1152.976216618" lastFinishedPulling="2025-12-13 07:09:28.213713497 +0000 UTC m=+1224.818122955" observedRunningTime="2025-12-13 07:09:28.967061455 +0000 UTC m=+1225.571470913" watchObservedRunningTime="2025-12-13 07:09:28.971063845 +0000 UTC m=+1225.575473293" Dec 13 07:09:29 crc kubenswrapper[4971]: I1213 07:09:29.951619 4971 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/telemetry-operator-controller-manager-97d456b9-j4fpf" podStartSLOduration=5.156521313 podStartE2EDuration="1m16.951599603s" podCreationTimestamp="2025-12-13 07:08:13 +0000 UTC" firstStartedPulling="2025-12-13 07:08:16.418700939 +0000 UTC m=+1153.023110387" lastFinishedPulling="2025-12-13 07:09:28.213779229 +0000 UTC m=+1224.818188677" observedRunningTime="2025-12-13 07:09:29.002083328 +0000 UTC m=+1225.606492776" watchObservedRunningTime="2025-12-13 07:09:29.951599603 +0000 UTC m=+1226.556009051" Dec 13 07:09:30 crc kubenswrapper[4971]: I1213 07:09:30.966624 4971 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/swift-operator-controller-manager-5c6df8f9-rt5n5" event={"ID":"bd48845c-5d55-4690-960b-ca319f596beb","Type":"ContainerStarted","Data":"09d3b5d46b02cd4b88ac2dc395e7a804bc43d5eb206f3d5fdc2be6941c5b2d3a"} Dec 13 07:09:30 crc kubenswrapper[4971]: I1213 07:09:30.967218 4971 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/swift-operator-controller-manager-5c6df8f9-rt5n5" Dec 13 07:09:30 crc kubenswrapper[4971]: I1213 07:09:30.989693 4971 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/swift-operator-controller-manager-5c6df8f9-rt5n5" podStartSLOduration=3.652935957 podStartE2EDuration="1m17.989671157s" podCreationTimestamp="2025-12-13 07:08:13 +0000 UTC" firstStartedPulling="2025-12-13 07:08:16.370737756 +0000 UTC m=+1152.975147204" lastFinishedPulling="2025-12-13 07:09:30.707472956 +0000 UTC m=+1227.311882404" observedRunningTime="2025-12-13 07:09:30.987664782 +0000 UTC m=+1227.592074230" watchObservedRunningTime="2025-12-13 07:09:30.989671157 +0000 UTC m=+1227.594080605" Dec 13 07:09:33 crc kubenswrapper[4971]: I1213 07:09:33.628457 4971 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/nova-operator-controller-manager-5fbbf8b6cc-k4vtr" Dec 13 07:09:33 crc kubenswrapper[4971]: I1213 07:09:33.795610 4971 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/telemetry-operator-controller-manager-97d456b9-j4fpf" Dec 13 07:09:43 crc kubenswrapper[4971]: I1213 07:09:43.846345 4971 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/swift-operator-controller-manager-5c6df8f9-rt5n5" Dec 13 07:10:01 crc kubenswrapper[4971]: I1213 07:10:01.183548 4971 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-84bb9d8bd9-ljf6c"] Dec 13 07:10:01 crc kubenswrapper[4971]: I1213 07:10:01.185001 4971 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-84bb9d8bd9-ljf6c" Dec 13 07:10:01 crc kubenswrapper[4971]: I1213 07:10:01.187275 4971 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dnsmasq-dns-dockercfg-kp276" Dec 13 07:10:01 crc kubenswrapper[4971]: I1213 07:10:01.187330 4971 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openshift-service-ca.crt" Dec 13 07:10:01 crc kubenswrapper[4971]: I1213 07:10:01.187439 4971 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"kube-root-ca.crt" Dec 13 07:10:01 crc kubenswrapper[4971]: I1213 07:10:01.187715 4971 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"dns" Dec 13 07:10:01 crc kubenswrapper[4971]: I1213 07:10:01.196537 4971 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-84bb9d8bd9-ljf6c"] Dec 13 07:10:01 crc kubenswrapper[4971]: I1213 07:10:01.229475 4971 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-fv74m\" (UniqueName: \"kubernetes.io/projected/ee072669-97ce-4d78-a12f-4169120a00f6-kube-api-access-fv74m\") pod \"dnsmasq-dns-84bb9d8bd9-ljf6c\" (UID: \"ee072669-97ce-4d78-a12f-4169120a00f6\") " pod="openstack/dnsmasq-dns-84bb9d8bd9-ljf6c" Dec 13 07:10:01 crc kubenswrapper[4971]: I1213 07:10:01.229800 4971 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/ee072669-97ce-4d78-a12f-4169120a00f6-config\") pod \"dnsmasq-dns-84bb9d8bd9-ljf6c\" (UID: \"ee072669-97ce-4d78-a12f-4169120a00f6\") " pod="openstack/dnsmasq-dns-84bb9d8bd9-ljf6c" Dec 13 07:10:01 crc kubenswrapper[4971]: I1213 07:10:01.238086 4971 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-5f854695bc-56d9f"] Dec 13 07:10:01 crc kubenswrapper[4971]: I1213 07:10:01.239277 4971 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-5f854695bc-56d9f" Dec 13 07:10:01 crc kubenswrapper[4971]: I1213 07:10:01.243417 4971 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"dns-svc" Dec 13 07:10:01 crc kubenswrapper[4971]: I1213 07:10:01.245542 4971 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-5f854695bc-56d9f"] Dec 13 07:10:01 crc kubenswrapper[4971]: I1213 07:10:01.331551 4971 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/038690e5-17a1-4e19-8d3b-f420dbdf7349-config\") pod \"dnsmasq-dns-5f854695bc-56d9f\" (UID: \"038690e5-17a1-4e19-8d3b-f420dbdf7349\") " pod="openstack/dnsmasq-dns-5f854695bc-56d9f" Dec 13 07:10:01 crc kubenswrapper[4971]: I1213 07:10:01.331629 4971 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-rzb6k\" (UniqueName: \"kubernetes.io/projected/038690e5-17a1-4e19-8d3b-f420dbdf7349-kube-api-access-rzb6k\") pod \"dnsmasq-dns-5f854695bc-56d9f\" (UID: \"038690e5-17a1-4e19-8d3b-f420dbdf7349\") " pod="openstack/dnsmasq-dns-5f854695bc-56d9f" Dec 13 07:10:01 crc kubenswrapper[4971]: I1213 07:10:01.331670 4971 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/038690e5-17a1-4e19-8d3b-f420dbdf7349-dns-svc\") pod \"dnsmasq-dns-5f854695bc-56d9f\" (UID: \"038690e5-17a1-4e19-8d3b-f420dbdf7349\") " pod="openstack/dnsmasq-dns-5f854695bc-56d9f" Dec 13 07:10:01 crc kubenswrapper[4971]: I1213 07:10:01.331710 4971 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-fv74m\" (UniqueName: \"kubernetes.io/projected/ee072669-97ce-4d78-a12f-4169120a00f6-kube-api-access-fv74m\") pod \"dnsmasq-dns-84bb9d8bd9-ljf6c\" (UID: \"ee072669-97ce-4d78-a12f-4169120a00f6\") " pod="openstack/dnsmasq-dns-84bb9d8bd9-ljf6c" Dec 13 07:10:01 crc kubenswrapper[4971]: I1213 07:10:01.331738 4971 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/ee072669-97ce-4d78-a12f-4169120a00f6-config\") pod \"dnsmasq-dns-84bb9d8bd9-ljf6c\" (UID: \"ee072669-97ce-4d78-a12f-4169120a00f6\") " pod="openstack/dnsmasq-dns-84bb9d8bd9-ljf6c" Dec 13 07:10:01 crc kubenswrapper[4971]: I1213 07:10:01.332598 4971 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/ee072669-97ce-4d78-a12f-4169120a00f6-config\") pod \"dnsmasq-dns-84bb9d8bd9-ljf6c\" (UID: \"ee072669-97ce-4d78-a12f-4169120a00f6\") " pod="openstack/dnsmasq-dns-84bb9d8bd9-ljf6c" Dec 13 07:10:01 crc kubenswrapper[4971]: I1213 07:10:01.361502 4971 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-fv74m\" (UniqueName: \"kubernetes.io/projected/ee072669-97ce-4d78-a12f-4169120a00f6-kube-api-access-fv74m\") pod \"dnsmasq-dns-84bb9d8bd9-ljf6c\" (UID: \"ee072669-97ce-4d78-a12f-4169120a00f6\") " pod="openstack/dnsmasq-dns-84bb9d8bd9-ljf6c" Dec 13 07:10:01 crc kubenswrapper[4971]: I1213 07:10:01.433456 4971 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/038690e5-17a1-4e19-8d3b-f420dbdf7349-config\") pod \"dnsmasq-dns-5f854695bc-56d9f\" (UID: \"038690e5-17a1-4e19-8d3b-f420dbdf7349\") " pod="openstack/dnsmasq-dns-5f854695bc-56d9f" Dec 13 07:10:01 crc kubenswrapper[4971]: I1213 07:10:01.433869 4971 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-rzb6k\" (UniqueName: \"kubernetes.io/projected/038690e5-17a1-4e19-8d3b-f420dbdf7349-kube-api-access-rzb6k\") pod \"dnsmasq-dns-5f854695bc-56d9f\" (UID: \"038690e5-17a1-4e19-8d3b-f420dbdf7349\") " pod="openstack/dnsmasq-dns-5f854695bc-56d9f" Dec 13 07:10:01 crc kubenswrapper[4971]: I1213 07:10:01.433899 4971 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/038690e5-17a1-4e19-8d3b-f420dbdf7349-dns-svc\") pod \"dnsmasq-dns-5f854695bc-56d9f\" (UID: \"038690e5-17a1-4e19-8d3b-f420dbdf7349\") " pod="openstack/dnsmasq-dns-5f854695bc-56d9f" Dec 13 07:10:01 crc kubenswrapper[4971]: I1213 07:10:01.435336 4971 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/038690e5-17a1-4e19-8d3b-f420dbdf7349-config\") pod \"dnsmasq-dns-5f854695bc-56d9f\" (UID: \"038690e5-17a1-4e19-8d3b-f420dbdf7349\") " pod="openstack/dnsmasq-dns-5f854695bc-56d9f" Dec 13 07:10:01 crc kubenswrapper[4971]: I1213 07:10:01.435876 4971 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/038690e5-17a1-4e19-8d3b-f420dbdf7349-dns-svc\") pod \"dnsmasq-dns-5f854695bc-56d9f\" (UID: \"038690e5-17a1-4e19-8d3b-f420dbdf7349\") " pod="openstack/dnsmasq-dns-5f854695bc-56d9f" Dec 13 07:10:01 crc kubenswrapper[4971]: I1213 07:10:01.453808 4971 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-rzb6k\" (UniqueName: \"kubernetes.io/projected/038690e5-17a1-4e19-8d3b-f420dbdf7349-kube-api-access-rzb6k\") pod \"dnsmasq-dns-5f854695bc-56d9f\" (UID: \"038690e5-17a1-4e19-8d3b-f420dbdf7349\") " pod="openstack/dnsmasq-dns-5f854695bc-56d9f" Dec 13 07:10:01 crc kubenswrapper[4971]: I1213 07:10:01.503801 4971 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-84bb9d8bd9-ljf6c" Dec 13 07:10:01 crc kubenswrapper[4971]: I1213 07:10:01.565084 4971 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-5f854695bc-56d9f" Dec 13 07:10:01 crc kubenswrapper[4971]: I1213 07:10:01.981069 4971 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-84bb9d8bd9-ljf6c"] Dec 13 07:10:02 crc kubenswrapper[4971]: I1213 07:10:02.072231 4971 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-5f854695bc-56d9f"] Dec 13 07:10:02 crc kubenswrapper[4971]: W1213 07:10:02.075179 4971 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod038690e5_17a1_4e19_8d3b_f420dbdf7349.slice/crio-413ec7db5bd51e525ef83070345f753c3e8cfe41ed771b4f218976331f6f8474 WatchSource:0}: Error finding container 413ec7db5bd51e525ef83070345f753c3e8cfe41ed771b4f218976331f6f8474: Status 404 returned error can't find the container with id 413ec7db5bd51e525ef83070345f753c3e8cfe41ed771b4f218976331f6f8474 Dec 13 07:10:02 crc kubenswrapper[4971]: I1213 07:10:02.373326 4971 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-5f854695bc-56d9f" event={"ID":"038690e5-17a1-4e19-8d3b-f420dbdf7349","Type":"ContainerStarted","Data":"413ec7db5bd51e525ef83070345f753c3e8cfe41ed771b4f218976331f6f8474"} Dec 13 07:10:02 crc kubenswrapper[4971]: I1213 07:10:02.374676 4971 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-84bb9d8bd9-ljf6c" event={"ID":"ee072669-97ce-4d78-a12f-4169120a00f6","Type":"ContainerStarted","Data":"98370c8549a5d34a131f41098b3d60a2ca5bbd6fcbb16443696869a561ab5596"} Dec 13 07:10:04 crc kubenswrapper[4971]: I1213 07:10:04.419004 4971 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-5f854695bc-56d9f"] Dec 13 07:10:04 crc kubenswrapper[4971]: I1213 07:10:04.549600 4971 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-744ffd65bc-zhplm"] Dec 13 07:10:04 crc kubenswrapper[4971]: I1213 07:10:04.551005 4971 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-744ffd65bc-zhplm" Dec 13 07:10:04 crc kubenswrapper[4971]: I1213 07:10:04.566968 4971 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-744ffd65bc-zhplm"] Dec 13 07:10:04 crc kubenswrapper[4971]: I1213 07:10:04.714319 4971 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-5zfr7\" (UniqueName: \"kubernetes.io/projected/b4e999eb-ff02-4c70-b755-f85bd90dd637-kube-api-access-5zfr7\") pod \"dnsmasq-dns-744ffd65bc-zhplm\" (UID: \"b4e999eb-ff02-4c70-b755-f85bd90dd637\") " pod="openstack/dnsmasq-dns-744ffd65bc-zhplm" Dec 13 07:10:04 crc kubenswrapper[4971]: I1213 07:10:04.714555 4971 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/b4e999eb-ff02-4c70-b755-f85bd90dd637-dns-svc\") pod \"dnsmasq-dns-744ffd65bc-zhplm\" (UID: \"b4e999eb-ff02-4c70-b755-f85bd90dd637\") " pod="openstack/dnsmasq-dns-744ffd65bc-zhplm" Dec 13 07:10:04 crc kubenswrapper[4971]: I1213 07:10:04.714671 4971 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/b4e999eb-ff02-4c70-b755-f85bd90dd637-config\") pod \"dnsmasq-dns-744ffd65bc-zhplm\" (UID: \"b4e999eb-ff02-4c70-b755-f85bd90dd637\") " pod="openstack/dnsmasq-dns-744ffd65bc-zhplm" Dec 13 07:10:04 crc kubenswrapper[4971]: I1213 07:10:04.815633 4971 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/b4e999eb-ff02-4c70-b755-f85bd90dd637-dns-svc\") pod \"dnsmasq-dns-744ffd65bc-zhplm\" (UID: \"b4e999eb-ff02-4c70-b755-f85bd90dd637\") " pod="openstack/dnsmasq-dns-744ffd65bc-zhplm" Dec 13 07:10:04 crc kubenswrapper[4971]: I1213 07:10:04.815694 4971 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/b4e999eb-ff02-4c70-b755-f85bd90dd637-config\") pod \"dnsmasq-dns-744ffd65bc-zhplm\" (UID: \"b4e999eb-ff02-4c70-b755-f85bd90dd637\") " pod="openstack/dnsmasq-dns-744ffd65bc-zhplm" Dec 13 07:10:04 crc kubenswrapper[4971]: I1213 07:10:04.815728 4971 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-5zfr7\" (UniqueName: \"kubernetes.io/projected/b4e999eb-ff02-4c70-b755-f85bd90dd637-kube-api-access-5zfr7\") pod \"dnsmasq-dns-744ffd65bc-zhplm\" (UID: \"b4e999eb-ff02-4c70-b755-f85bd90dd637\") " pod="openstack/dnsmasq-dns-744ffd65bc-zhplm" Dec 13 07:10:04 crc kubenswrapper[4971]: I1213 07:10:04.816935 4971 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/b4e999eb-ff02-4c70-b755-f85bd90dd637-dns-svc\") pod \"dnsmasq-dns-744ffd65bc-zhplm\" (UID: \"b4e999eb-ff02-4c70-b755-f85bd90dd637\") " pod="openstack/dnsmasq-dns-744ffd65bc-zhplm" Dec 13 07:10:04 crc kubenswrapper[4971]: I1213 07:10:04.817496 4971 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/b4e999eb-ff02-4c70-b755-f85bd90dd637-config\") pod \"dnsmasq-dns-744ffd65bc-zhplm\" (UID: \"b4e999eb-ff02-4c70-b755-f85bd90dd637\") " pod="openstack/dnsmasq-dns-744ffd65bc-zhplm" Dec 13 07:10:04 crc kubenswrapper[4971]: I1213 07:10:04.817641 4971 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-84bb9d8bd9-ljf6c"] Dec 13 07:10:04 crc kubenswrapper[4971]: I1213 07:10:04.838597 4971 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-5zfr7\" (UniqueName: \"kubernetes.io/projected/b4e999eb-ff02-4c70-b755-f85bd90dd637-kube-api-access-5zfr7\") pod \"dnsmasq-dns-744ffd65bc-zhplm\" (UID: \"b4e999eb-ff02-4c70-b755-f85bd90dd637\") " pod="openstack/dnsmasq-dns-744ffd65bc-zhplm" Dec 13 07:10:04 crc kubenswrapper[4971]: I1213 07:10:04.863831 4971 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-95f5f6995-qv9tw"] Dec 13 07:10:04 crc kubenswrapper[4971]: I1213 07:10:04.865417 4971 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-95f5f6995-qv9tw" Dec 13 07:10:04 crc kubenswrapper[4971]: I1213 07:10:04.877312 4971 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-744ffd65bc-zhplm" Dec 13 07:10:04 crc kubenswrapper[4971]: I1213 07:10:04.878638 4971 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-95f5f6995-qv9tw"] Dec 13 07:10:05 crc kubenswrapper[4971]: I1213 07:10:05.019424 4971 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/9a37323b-7693-41c2-8a8f-661291d1ef5e-dns-svc\") pod \"dnsmasq-dns-95f5f6995-qv9tw\" (UID: \"9a37323b-7693-41c2-8a8f-661291d1ef5e\") " pod="openstack/dnsmasq-dns-95f5f6995-qv9tw" Dec 13 07:10:05 crc kubenswrapper[4971]: I1213 07:10:05.019708 4971 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/9a37323b-7693-41c2-8a8f-661291d1ef5e-config\") pod \"dnsmasq-dns-95f5f6995-qv9tw\" (UID: \"9a37323b-7693-41c2-8a8f-661291d1ef5e\") " pod="openstack/dnsmasq-dns-95f5f6995-qv9tw" Dec 13 07:10:05 crc kubenswrapper[4971]: I1213 07:10:05.022250 4971 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-pszmq\" (UniqueName: \"kubernetes.io/projected/9a37323b-7693-41c2-8a8f-661291d1ef5e-kube-api-access-pszmq\") pod \"dnsmasq-dns-95f5f6995-qv9tw\" (UID: \"9a37323b-7693-41c2-8a8f-661291d1ef5e\") " pod="openstack/dnsmasq-dns-95f5f6995-qv9tw" Dec 13 07:10:05 crc kubenswrapper[4971]: I1213 07:10:05.123416 4971 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-pszmq\" (UniqueName: \"kubernetes.io/projected/9a37323b-7693-41c2-8a8f-661291d1ef5e-kube-api-access-pszmq\") pod \"dnsmasq-dns-95f5f6995-qv9tw\" (UID: \"9a37323b-7693-41c2-8a8f-661291d1ef5e\") " pod="openstack/dnsmasq-dns-95f5f6995-qv9tw" Dec 13 07:10:05 crc kubenswrapper[4971]: I1213 07:10:05.123530 4971 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/9a37323b-7693-41c2-8a8f-661291d1ef5e-dns-svc\") pod \"dnsmasq-dns-95f5f6995-qv9tw\" (UID: \"9a37323b-7693-41c2-8a8f-661291d1ef5e\") " pod="openstack/dnsmasq-dns-95f5f6995-qv9tw" Dec 13 07:10:05 crc kubenswrapper[4971]: I1213 07:10:05.123572 4971 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/9a37323b-7693-41c2-8a8f-661291d1ef5e-config\") pod \"dnsmasq-dns-95f5f6995-qv9tw\" (UID: \"9a37323b-7693-41c2-8a8f-661291d1ef5e\") " pod="openstack/dnsmasq-dns-95f5f6995-qv9tw" Dec 13 07:10:05 crc kubenswrapper[4971]: I1213 07:10:05.124697 4971 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/9a37323b-7693-41c2-8a8f-661291d1ef5e-config\") pod \"dnsmasq-dns-95f5f6995-qv9tw\" (UID: \"9a37323b-7693-41c2-8a8f-661291d1ef5e\") " pod="openstack/dnsmasq-dns-95f5f6995-qv9tw" Dec 13 07:10:05 crc kubenswrapper[4971]: I1213 07:10:05.124820 4971 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/9a37323b-7693-41c2-8a8f-661291d1ef5e-dns-svc\") pod \"dnsmasq-dns-95f5f6995-qv9tw\" (UID: \"9a37323b-7693-41c2-8a8f-661291d1ef5e\") " pod="openstack/dnsmasq-dns-95f5f6995-qv9tw" Dec 13 07:10:05 crc kubenswrapper[4971]: I1213 07:10:05.143222 4971 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-pszmq\" (UniqueName: \"kubernetes.io/projected/9a37323b-7693-41c2-8a8f-661291d1ef5e-kube-api-access-pszmq\") pod \"dnsmasq-dns-95f5f6995-qv9tw\" (UID: \"9a37323b-7693-41c2-8a8f-661291d1ef5e\") " pod="openstack/dnsmasq-dns-95f5f6995-qv9tw" Dec 13 07:10:05 crc kubenswrapper[4971]: I1213 07:10:05.201631 4971 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-95f5f6995-qv9tw" Dec 13 07:10:05 crc kubenswrapper[4971]: I1213 07:10:05.423895 4971 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-744ffd65bc-zhplm"] Dec 13 07:10:05 crc kubenswrapper[4971]: W1213 07:10:05.428359 4971 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podb4e999eb_ff02_4c70_b755_f85bd90dd637.slice/crio-fad71e9c347b23db5981bdc894b841c16cb490dcff033245da131c2ca048068a WatchSource:0}: Error finding container fad71e9c347b23db5981bdc894b841c16cb490dcff033245da131c2ca048068a: Status 404 returned error can't find the container with id fad71e9c347b23db5981bdc894b841c16cb490dcff033245da131c2ca048068a Dec 13 07:10:05 crc kubenswrapper[4971]: I1213 07:10:05.689108 4971 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/rabbitmq-server-0"] Dec 13 07:10:05 crc kubenswrapper[4971]: I1213 07:10:05.690510 4971 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/rabbitmq-server-0" Dec 13 07:10:05 crc kubenswrapper[4971]: I1213 07:10:05.691734 4971 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-rabbitmq-svc" Dec 13 07:10:05 crc kubenswrapper[4971]: I1213 07:10:05.694347 4971 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"rabbitmq-default-user" Dec 13 07:10:05 crc kubenswrapper[4971]: I1213 07:10:05.694462 4971 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"rabbitmq-config-data" Dec 13 07:10:05 crc kubenswrapper[4971]: I1213 07:10:05.694557 4971 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"rabbitmq-server-dockercfg-vzbd9" Dec 13 07:10:05 crc kubenswrapper[4971]: I1213 07:10:05.694622 4971 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"rabbitmq-erlang-cookie" Dec 13 07:10:05 crc kubenswrapper[4971]: I1213 07:10:05.694687 4971 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"rabbitmq-plugins-conf" Dec 13 07:10:05 crc kubenswrapper[4971]: I1213 07:10:05.694763 4971 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"rabbitmq-server-conf" Dec 13 07:10:05 crc kubenswrapper[4971]: I1213 07:10:05.779010 4971 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-95f5f6995-qv9tw"] Dec 13 07:10:05 crc kubenswrapper[4971]: I1213 07:10:05.785237 4971 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/rabbitmq-server-0"] Dec 13 07:10:05 crc kubenswrapper[4971]: W1213 07:10:05.785312 4971 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod9a37323b_7693_41c2_8a8f_661291d1ef5e.slice/crio-85cf34c1f99a8f5cd4f170ba6f7da04e1970efebd7348563a6dc2bfe6bf6b2f9 WatchSource:0}: Error finding container 85cf34c1f99a8f5cd4f170ba6f7da04e1970efebd7348563a6dc2bfe6bf6b2f9: Status 404 returned error can't find the container with id 85cf34c1f99a8f5cd4f170ba6f7da04e1970efebd7348563a6dc2bfe6bf6b2f9 Dec 13 07:10:05 crc kubenswrapper[4971]: I1213 07:10:05.891108 4971 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-wfzdp\" (UniqueName: \"kubernetes.io/projected/a590eb84-c8d5-4cbe-9c03-03a6ccb82073-kube-api-access-wfzdp\") pod \"rabbitmq-server-0\" (UID: \"a590eb84-c8d5-4cbe-9c03-03a6ccb82073\") " pod="openstack/rabbitmq-server-0" Dec 13 07:10:05 crc kubenswrapper[4971]: I1213 07:10:05.891169 4971 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"plugins-conf\" (UniqueName: \"kubernetes.io/configmap/a590eb84-c8d5-4cbe-9c03-03a6ccb82073-plugins-conf\") pod \"rabbitmq-server-0\" (UID: \"a590eb84-c8d5-4cbe-9c03-03a6ccb82073\") " pod="openstack/rabbitmq-server-0" Dec 13 07:10:05 crc kubenswrapper[4971]: I1213 07:10:05.891205 4971 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rabbitmq-tls\" (UniqueName: \"kubernetes.io/projected/a590eb84-c8d5-4cbe-9c03-03a6ccb82073-rabbitmq-tls\") pod \"rabbitmq-server-0\" (UID: \"a590eb84-c8d5-4cbe-9c03-03a6ccb82073\") " pod="openstack/rabbitmq-server-0" Dec 13 07:10:05 crc kubenswrapper[4971]: I1213 07:10:05.891229 4971 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rabbitmq-erlang-cookie\" (UniqueName: \"kubernetes.io/empty-dir/a590eb84-c8d5-4cbe-9c03-03a6ccb82073-rabbitmq-erlang-cookie\") pod \"rabbitmq-server-0\" (UID: \"a590eb84-c8d5-4cbe-9c03-03a6ccb82073\") " pod="openstack/rabbitmq-server-0" Dec 13 07:10:05 crc kubenswrapper[4971]: I1213 07:10:05.891253 4971 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rabbitmq-plugins\" (UniqueName: \"kubernetes.io/empty-dir/a590eb84-c8d5-4cbe-9c03-03a6ccb82073-rabbitmq-plugins\") pod \"rabbitmq-server-0\" (UID: \"a590eb84-c8d5-4cbe-9c03-03a6ccb82073\") " pod="openstack/rabbitmq-server-0" Dec 13 07:10:05 crc kubenswrapper[4971]: I1213 07:10:05.891401 4971 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/a590eb84-c8d5-4cbe-9c03-03a6ccb82073-config-data\") pod \"rabbitmq-server-0\" (UID: \"a590eb84-c8d5-4cbe-9c03-03a6ccb82073\") " pod="openstack/rabbitmq-server-0" Dec 13 07:10:05 crc kubenswrapper[4971]: I1213 07:10:05.891491 4971 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"erlang-cookie-secret\" (UniqueName: \"kubernetes.io/secret/a590eb84-c8d5-4cbe-9c03-03a6ccb82073-erlang-cookie-secret\") pod \"rabbitmq-server-0\" (UID: \"a590eb84-c8d5-4cbe-9c03-03a6ccb82073\") " pod="openstack/rabbitmq-server-0" Dec 13 07:10:05 crc kubenswrapper[4971]: I1213 07:10:05.891590 4971 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"local-storage06-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage06-crc\") pod \"rabbitmq-server-0\" (UID: \"a590eb84-c8d5-4cbe-9c03-03a6ccb82073\") " pod="openstack/rabbitmq-server-0" Dec 13 07:10:05 crc kubenswrapper[4971]: I1213 07:10:05.891634 4971 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rabbitmq-confd\" (UniqueName: \"kubernetes.io/projected/a590eb84-c8d5-4cbe-9c03-03a6ccb82073-rabbitmq-confd\") pod \"rabbitmq-server-0\" (UID: \"a590eb84-c8d5-4cbe-9c03-03a6ccb82073\") " pod="openstack/rabbitmq-server-0" Dec 13 07:10:05 crc kubenswrapper[4971]: I1213 07:10:05.891731 4971 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"pod-info\" (UniqueName: \"kubernetes.io/downward-api/a590eb84-c8d5-4cbe-9c03-03a6ccb82073-pod-info\") pod \"rabbitmq-server-0\" (UID: \"a590eb84-c8d5-4cbe-9c03-03a6ccb82073\") " pod="openstack/rabbitmq-server-0" Dec 13 07:10:05 crc kubenswrapper[4971]: I1213 07:10:05.891826 4971 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"server-conf\" (UniqueName: \"kubernetes.io/configmap/a590eb84-c8d5-4cbe-9c03-03a6ccb82073-server-conf\") pod \"rabbitmq-server-0\" (UID: \"a590eb84-c8d5-4cbe-9c03-03a6ccb82073\") " pod="openstack/rabbitmq-server-0" Dec 13 07:10:05 crc kubenswrapper[4971]: I1213 07:10:05.995472 4971 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-wfzdp\" (UniqueName: \"kubernetes.io/projected/a590eb84-c8d5-4cbe-9c03-03a6ccb82073-kube-api-access-wfzdp\") pod \"rabbitmq-server-0\" (UID: \"a590eb84-c8d5-4cbe-9c03-03a6ccb82073\") " pod="openstack/rabbitmq-server-0" Dec 13 07:10:05 crc kubenswrapper[4971]: I1213 07:10:05.995582 4971 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"plugins-conf\" (UniqueName: \"kubernetes.io/configmap/a590eb84-c8d5-4cbe-9c03-03a6ccb82073-plugins-conf\") pod \"rabbitmq-server-0\" (UID: \"a590eb84-c8d5-4cbe-9c03-03a6ccb82073\") " pod="openstack/rabbitmq-server-0" Dec 13 07:10:05 crc kubenswrapper[4971]: I1213 07:10:05.995644 4971 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rabbitmq-tls\" (UniqueName: \"kubernetes.io/projected/a590eb84-c8d5-4cbe-9c03-03a6ccb82073-rabbitmq-tls\") pod \"rabbitmq-server-0\" (UID: \"a590eb84-c8d5-4cbe-9c03-03a6ccb82073\") " pod="openstack/rabbitmq-server-0" Dec 13 07:10:05 crc kubenswrapper[4971]: I1213 07:10:05.995675 4971 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rabbitmq-erlang-cookie\" (UniqueName: \"kubernetes.io/empty-dir/a590eb84-c8d5-4cbe-9c03-03a6ccb82073-rabbitmq-erlang-cookie\") pod \"rabbitmq-server-0\" (UID: \"a590eb84-c8d5-4cbe-9c03-03a6ccb82073\") " pod="openstack/rabbitmq-server-0" Dec 13 07:10:05 crc kubenswrapper[4971]: I1213 07:10:05.995841 4971 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rabbitmq-plugins\" (UniqueName: \"kubernetes.io/empty-dir/a590eb84-c8d5-4cbe-9c03-03a6ccb82073-rabbitmq-plugins\") pod \"rabbitmq-server-0\" (UID: \"a590eb84-c8d5-4cbe-9c03-03a6ccb82073\") " pod="openstack/rabbitmq-server-0" Dec 13 07:10:05 crc kubenswrapper[4971]: I1213 07:10:05.997677 4971 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rabbitmq-erlang-cookie\" (UniqueName: \"kubernetes.io/empty-dir/a590eb84-c8d5-4cbe-9c03-03a6ccb82073-rabbitmq-erlang-cookie\") pod \"rabbitmq-server-0\" (UID: \"a590eb84-c8d5-4cbe-9c03-03a6ccb82073\") " pod="openstack/rabbitmq-server-0" Dec 13 07:10:05 crc kubenswrapper[4971]: I1213 07:10:05.997833 4971 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"plugins-conf\" (UniqueName: \"kubernetes.io/configmap/a590eb84-c8d5-4cbe-9c03-03a6ccb82073-plugins-conf\") pod \"rabbitmq-server-0\" (UID: \"a590eb84-c8d5-4cbe-9c03-03a6ccb82073\") " pod="openstack/rabbitmq-server-0" Dec 13 07:10:05 crc kubenswrapper[4971]: I1213 07:10:05.999033 4971 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/a590eb84-c8d5-4cbe-9c03-03a6ccb82073-config-data\") pod \"rabbitmq-server-0\" (UID: \"a590eb84-c8d5-4cbe-9c03-03a6ccb82073\") " pod="openstack/rabbitmq-server-0" Dec 13 07:10:05 crc kubenswrapper[4971]: I1213 07:10:05.999773 4971 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/a590eb84-c8d5-4cbe-9c03-03a6ccb82073-config-data\") pod \"rabbitmq-server-0\" (UID: \"a590eb84-c8d5-4cbe-9c03-03a6ccb82073\") " pod="openstack/rabbitmq-server-0" Dec 13 07:10:05 crc kubenswrapper[4971]: I1213 07:10:05.999831 4971 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"erlang-cookie-secret\" (UniqueName: \"kubernetes.io/secret/a590eb84-c8d5-4cbe-9c03-03a6ccb82073-erlang-cookie-secret\") pod \"rabbitmq-server-0\" (UID: \"a590eb84-c8d5-4cbe-9c03-03a6ccb82073\") " pod="openstack/rabbitmq-server-0" Dec 13 07:10:05 crc kubenswrapper[4971]: I1213 07:10:05.999903 4971 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"local-storage06-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage06-crc\") pod \"rabbitmq-server-0\" (UID: \"a590eb84-c8d5-4cbe-9c03-03a6ccb82073\") " pod="openstack/rabbitmq-server-0" Dec 13 07:10:06 crc kubenswrapper[4971]: I1213 07:10:05.999966 4971 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rabbitmq-confd\" (UniqueName: \"kubernetes.io/projected/a590eb84-c8d5-4cbe-9c03-03a6ccb82073-rabbitmq-confd\") pod \"rabbitmq-server-0\" (UID: \"a590eb84-c8d5-4cbe-9c03-03a6ccb82073\") " pod="openstack/rabbitmq-server-0" Dec 13 07:10:06 crc kubenswrapper[4971]: I1213 07:10:06.000002 4971 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pod-info\" (UniqueName: \"kubernetes.io/downward-api/a590eb84-c8d5-4cbe-9c03-03a6ccb82073-pod-info\") pod \"rabbitmq-server-0\" (UID: \"a590eb84-c8d5-4cbe-9c03-03a6ccb82073\") " pod="openstack/rabbitmq-server-0" Dec 13 07:10:06 crc kubenswrapper[4971]: I1213 07:10:06.000245 4971 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"local-storage06-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage06-crc\") pod \"rabbitmq-server-0\" (UID: \"a590eb84-c8d5-4cbe-9c03-03a6ccb82073\") device mount path \"/mnt/openstack/pv06\"" pod="openstack/rabbitmq-server-0" Dec 13 07:10:06 crc kubenswrapper[4971]: I1213 07:10:06.000443 4971 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"server-conf\" (UniqueName: \"kubernetes.io/configmap/a590eb84-c8d5-4cbe-9c03-03a6ccb82073-server-conf\") pod \"rabbitmq-server-0\" (UID: \"a590eb84-c8d5-4cbe-9c03-03a6ccb82073\") " pod="openstack/rabbitmq-server-0" Dec 13 07:10:06 crc kubenswrapper[4971]: I1213 07:10:06.000815 4971 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rabbitmq-plugins\" (UniqueName: \"kubernetes.io/empty-dir/a590eb84-c8d5-4cbe-9c03-03a6ccb82073-rabbitmq-plugins\") pod \"rabbitmq-server-0\" (UID: \"a590eb84-c8d5-4cbe-9c03-03a6ccb82073\") " pod="openstack/rabbitmq-server-0" Dec 13 07:10:06 crc kubenswrapper[4971]: I1213 07:10:06.001927 4971 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"server-conf\" (UniqueName: \"kubernetes.io/configmap/a590eb84-c8d5-4cbe-9c03-03a6ccb82073-server-conf\") pod \"rabbitmq-server-0\" (UID: \"a590eb84-c8d5-4cbe-9c03-03a6ccb82073\") " pod="openstack/rabbitmq-server-0" Dec 13 07:10:06 crc kubenswrapper[4971]: I1213 07:10:06.005592 4971 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"erlang-cookie-secret\" (UniqueName: \"kubernetes.io/secret/a590eb84-c8d5-4cbe-9c03-03a6ccb82073-erlang-cookie-secret\") pod \"rabbitmq-server-0\" (UID: \"a590eb84-c8d5-4cbe-9c03-03a6ccb82073\") " pod="openstack/rabbitmq-server-0" Dec 13 07:10:06 crc kubenswrapper[4971]: I1213 07:10:06.006094 4971 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"pod-info\" (UniqueName: \"kubernetes.io/downward-api/a590eb84-c8d5-4cbe-9c03-03a6ccb82073-pod-info\") pod \"rabbitmq-server-0\" (UID: \"a590eb84-c8d5-4cbe-9c03-03a6ccb82073\") " pod="openstack/rabbitmq-server-0" Dec 13 07:10:06 crc kubenswrapper[4971]: I1213 07:10:06.006395 4971 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rabbitmq-confd\" (UniqueName: \"kubernetes.io/projected/a590eb84-c8d5-4cbe-9c03-03a6ccb82073-rabbitmq-confd\") pod \"rabbitmq-server-0\" (UID: \"a590eb84-c8d5-4cbe-9c03-03a6ccb82073\") " pod="openstack/rabbitmq-server-0" Dec 13 07:10:06 crc kubenswrapper[4971]: I1213 07:10:06.024556 4971 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-wfzdp\" (UniqueName: \"kubernetes.io/projected/a590eb84-c8d5-4cbe-9c03-03a6ccb82073-kube-api-access-wfzdp\") pod \"rabbitmq-server-0\" (UID: \"a590eb84-c8d5-4cbe-9c03-03a6ccb82073\") " pod="openstack/rabbitmq-server-0" Dec 13 07:10:06 crc kubenswrapper[4971]: I1213 07:10:06.031376 4971 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rabbitmq-tls\" (UniqueName: \"kubernetes.io/projected/a590eb84-c8d5-4cbe-9c03-03a6ccb82073-rabbitmq-tls\") pod \"rabbitmq-server-0\" (UID: \"a590eb84-c8d5-4cbe-9c03-03a6ccb82073\") " pod="openstack/rabbitmq-server-0" Dec 13 07:10:06 crc kubenswrapper[4971]: I1213 07:10:06.039162 4971 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"local-storage06-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage06-crc\") pod \"rabbitmq-server-0\" (UID: \"a590eb84-c8d5-4cbe-9c03-03a6ccb82073\") " pod="openstack/rabbitmq-server-0" Dec 13 07:10:06 crc kubenswrapper[4971]: I1213 07:10:06.113123 4971 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/rabbitmq-cell1-server-0"] Dec 13 07:10:06 crc kubenswrapper[4971]: I1213 07:10:06.114730 4971 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/rabbitmq-cell1-server-0" Dec 13 07:10:06 crc kubenswrapper[4971]: I1213 07:10:06.130180 4971 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"rabbitmq-cell1-config-data" Dec 13 07:10:06 crc kubenswrapper[4971]: I1213 07:10:06.130259 4971 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"rabbitmq-cell1-server-conf" Dec 13 07:10:06 crc kubenswrapper[4971]: I1213 07:10:06.130276 4971 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"rabbitmq-cell1-default-user" Dec 13 07:10:06 crc kubenswrapper[4971]: I1213 07:10:06.130484 4971 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"rabbitmq-cell1-server-dockercfg-5725c" Dec 13 07:10:06 crc kubenswrapper[4971]: I1213 07:10:06.135340 4971 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"rabbitmq-cell1-erlang-cookie" Dec 13 07:10:06 crc kubenswrapper[4971]: I1213 07:10:06.135610 4971 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-rabbitmq-cell1-svc" Dec 13 07:10:06 crc kubenswrapper[4971]: I1213 07:10:06.138669 4971 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"rabbitmq-cell1-plugins-conf" Dec 13 07:10:06 crc kubenswrapper[4971]: I1213 07:10:06.141914 4971 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/rabbitmq-cell1-server-0"] Dec 13 07:10:06 crc kubenswrapper[4971]: I1213 07:10:06.311491 4971 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-jpcrh\" (UniqueName: \"kubernetes.io/projected/5eed809d-c915-433c-9915-27e0b01e1ffe-kube-api-access-jpcrh\") pod \"rabbitmq-cell1-server-0\" (UID: \"5eed809d-c915-433c-9915-27e0b01e1ffe\") " pod="openstack/rabbitmq-cell1-server-0" Dec 13 07:10:06 crc kubenswrapper[4971]: I1213 07:10:06.311578 4971 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"local-storage10-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage10-crc\") pod \"rabbitmq-cell1-server-0\" (UID: \"5eed809d-c915-433c-9915-27e0b01e1ffe\") " pod="openstack/rabbitmq-cell1-server-0" Dec 13 07:10:06 crc kubenswrapper[4971]: I1213 07:10:06.311616 4971 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rabbitmq-erlang-cookie\" (UniqueName: \"kubernetes.io/empty-dir/5eed809d-c915-433c-9915-27e0b01e1ffe-rabbitmq-erlang-cookie\") pod \"rabbitmq-cell1-server-0\" (UID: \"5eed809d-c915-433c-9915-27e0b01e1ffe\") " pod="openstack/rabbitmq-cell1-server-0" Dec 13 07:10:06 crc kubenswrapper[4971]: I1213 07:10:06.311632 4971 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"pod-info\" (UniqueName: \"kubernetes.io/downward-api/5eed809d-c915-433c-9915-27e0b01e1ffe-pod-info\") pod \"rabbitmq-cell1-server-0\" (UID: \"5eed809d-c915-433c-9915-27e0b01e1ffe\") " pod="openstack/rabbitmq-cell1-server-0" Dec 13 07:10:06 crc kubenswrapper[4971]: I1213 07:10:06.311684 4971 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"plugins-conf\" (UniqueName: \"kubernetes.io/configmap/5eed809d-c915-433c-9915-27e0b01e1ffe-plugins-conf\") pod \"rabbitmq-cell1-server-0\" (UID: \"5eed809d-c915-433c-9915-27e0b01e1ffe\") " pod="openstack/rabbitmq-cell1-server-0" Dec 13 07:10:06 crc kubenswrapper[4971]: I1213 07:10:06.311707 4971 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rabbitmq-plugins\" (UniqueName: \"kubernetes.io/empty-dir/5eed809d-c915-433c-9915-27e0b01e1ffe-rabbitmq-plugins\") pod \"rabbitmq-cell1-server-0\" (UID: \"5eed809d-c915-433c-9915-27e0b01e1ffe\") " pod="openstack/rabbitmq-cell1-server-0" Dec 13 07:10:06 crc kubenswrapper[4971]: I1213 07:10:06.311737 4971 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"erlang-cookie-secret\" (UniqueName: \"kubernetes.io/secret/5eed809d-c915-433c-9915-27e0b01e1ffe-erlang-cookie-secret\") pod \"rabbitmq-cell1-server-0\" (UID: \"5eed809d-c915-433c-9915-27e0b01e1ffe\") " pod="openstack/rabbitmq-cell1-server-0" Dec 13 07:10:06 crc kubenswrapper[4971]: I1213 07:10:06.311765 4971 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"server-conf\" (UniqueName: \"kubernetes.io/configmap/5eed809d-c915-433c-9915-27e0b01e1ffe-server-conf\") pod \"rabbitmq-cell1-server-0\" (UID: \"5eed809d-c915-433c-9915-27e0b01e1ffe\") " pod="openstack/rabbitmq-cell1-server-0" Dec 13 07:10:06 crc kubenswrapper[4971]: I1213 07:10:06.311782 4971 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rabbitmq-tls\" (UniqueName: \"kubernetes.io/projected/5eed809d-c915-433c-9915-27e0b01e1ffe-rabbitmq-tls\") pod \"rabbitmq-cell1-server-0\" (UID: \"5eed809d-c915-433c-9915-27e0b01e1ffe\") " pod="openstack/rabbitmq-cell1-server-0" Dec 13 07:10:06 crc kubenswrapper[4971]: I1213 07:10:06.311850 4971 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rabbitmq-confd\" (UniqueName: \"kubernetes.io/projected/5eed809d-c915-433c-9915-27e0b01e1ffe-rabbitmq-confd\") pod \"rabbitmq-cell1-server-0\" (UID: \"5eed809d-c915-433c-9915-27e0b01e1ffe\") " pod="openstack/rabbitmq-cell1-server-0" Dec 13 07:10:06 crc kubenswrapper[4971]: I1213 07:10:06.311867 4971 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/5eed809d-c915-433c-9915-27e0b01e1ffe-config-data\") pod \"rabbitmq-cell1-server-0\" (UID: \"5eed809d-c915-433c-9915-27e0b01e1ffe\") " pod="openstack/rabbitmq-cell1-server-0" Dec 13 07:10:06 crc kubenswrapper[4971]: I1213 07:10:06.312675 4971 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/rabbitmq-server-0" Dec 13 07:10:06 crc kubenswrapper[4971]: I1213 07:10:06.413042 4971 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"plugins-conf\" (UniqueName: \"kubernetes.io/configmap/5eed809d-c915-433c-9915-27e0b01e1ffe-plugins-conf\") pod \"rabbitmq-cell1-server-0\" (UID: \"5eed809d-c915-433c-9915-27e0b01e1ffe\") " pod="openstack/rabbitmq-cell1-server-0" Dec 13 07:10:06 crc kubenswrapper[4971]: I1213 07:10:06.413090 4971 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rabbitmq-plugins\" (UniqueName: \"kubernetes.io/empty-dir/5eed809d-c915-433c-9915-27e0b01e1ffe-rabbitmq-plugins\") pod \"rabbitmq-cell1-server-0\" (UID: \"5eed809d-c915-433c-9915-27e0b01e1ffe\") " pod="openstack/rabbitmq-cell1-server-0" Dec 13 07:10:06 crc kubenswrapper[4971]: I1213 07:10:06.413117 4971 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"erlang-cookie-secret\" (UniqueName: \"kubernetes.io/secret/5eed809d-c915-433c-9915-27e0b01e1ffe-erlang-cookie-secret\") pod \"rabbitmq-cell1-server-0\" (UID: \"5eed809d-c915-433c-9915-27e0b01e1ffe\") " pod="openstack/rabbitmq-cell1-server-0" Dec 13 07:10:06 crc kubenswrapper[4971]: I1213 07:10:06.413137 4971 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"server-conf\" (UniqueName: \"kubernetes.io/configmap/5eed809d-c915-433c-9915-27e0b01e1ffe-server-conf\") pod \"rabbitmq-cell1-server-0\" (UID: \"5eed809d-c915-433c-9915-27e0b01e1ffe\") " pod="openstack/rabbitmq-cell1-server-0" Dec 13 07:10:06 crc kubenswrapper[4971]: I1213 07:10:06.413164 4971 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rabbitmq-tls\" (UniqueName: \"kubernetes.io/projected/5eed809d-c915-433c-9915-27e0b01e1ffe-rabbitmq-tls\") pod \"rabbitmq-cell1-server-0\" (UID: \"5eed809d-c915-433c-9915-27e0b01e1ffe\") " pod="openstack/rabbitmq-cell1-server-0" Dec 13 07:10:06 crc kubenswrapper[4971]: I1213 07:10:06.413210 4971 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rabbitmq-confd\" (UniqueName: \"kubernetes.io/projected/5eed809d-c915-433c-9915-27e0b01e1ffe-rabbitmq-confd\") pod \"rabbitmq-cell1-server-0\" (UID: \"5eed809d-c915-433c-9915-27e0b01e1ffe\") " pod="openstack/rabbitmq-cell1-server-0" Dec 13 07:10:06 crc kubenswrapper[4971]: I1213 07:10:06.413228 4971 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/5eed809d-c915-433c-9915-27e0b01e1ffe-config-data\") pod \"rabbitmq-cell1-server-0\" (UID: \"5eed809d-c915-433c-9915-27e0b01e1ffe\") " pod="openstack/rabbitmq-cell1-server-0" Dec 13 07:10:06 crc kubenswrapper[4971]: I1213 07:10:06.413282 4971 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-jpcrh\" (UniqueName: \"kubernetes.io/projected/5eed809d-c915-433c-9915-27e0b01e1ffe-kube-api-access-jpcrh\") pod \"rabbitmq-cell1-server-0\" (UID: \"5eed809d-c915-433c-9915-27e0b01e1ffe\") " pod="openstack/rabbitmq-cell1-server-0" Dec 13 07:10:06 crc kubenswrapper[4971]: I1213 07:10:06.413302 4971 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"local-storage10-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage10-crc\") pod \"rabbitmq-cell1-server-0\" (UID: \"5eed809d-c915-433c-9915-27e0b01e1ffe\") " pod="openstack/rabbitmq-cell1-server-0" Dec 13 07:10:06 crc kubenswrapper[4971]: I1213 07:10:06.413338 4971 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rabbitmq-erlang-cookie\" (UniqueName: \"kubernetes.io/empty-dir/5eed809d-c915-433c-9915-27e0b01e1ffe-rabbitmq-erlang-cookie\") pod \"rabbitmq-cell1-server-0\" (UID: \"5eed809d-c915-433c-9915-27e0b01e1ffe\") " pod="openstack/rabbitmq-cell1-server-0" Dec 13 07:10:06 crc kubenswrapper[4971]: I1213 07:10:06.413352 4971 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pod-info\" (UniqueName: \"kubernetes.io/downward-api/5eed809d-c915-433c-9915-27e0b01e1ffe-pod-info\") pod \"rabbitmq-cell1-server-0\" (UID: \"5eed809d-c915-433c-9915-27e0b01e1ffe\") " pod="openstack/rabbitmq-cell1-server-0" Dec 13 07:10:06 crc kubenswrapper[4971]: I1213 07:10:06.414000 4971 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"local-storage10-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage10-crc\") pod \"rabbitmq-cell1-server-0\" (UID: \"5eed809d-c915-433c-9915-27e0b01e1ffe\") device mount path \"/mnt/openstack/pv10\"" pod="openstack/rabbitmq-cell1-server-0" Dec 13 07:10:06 crc kubenswrapper[4971]: I1213 07:10:06.414998 4971 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rabbitmq-plugins\" (UniqueName: \"kubernetes.io/empty-dir/5eed809d-c915-433c-9915-27e0b01e1ffe-rabbitmq-plugins\") pod \"rabbitmq-cell1-server-0\" (UID: \"5eed809d-c915-433c-9915-27e0b01e1ffe\") " pod="openstack/rabbitmq-cell1-server-0" Dec 13 07:10:06 crc kubenswrapper[4971]: I1213 07:10:06.415377 4971 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rabbitmq-erlang-cookie\" (UniqueName: \"kubernetes.io/empty-dir/5eed809d-c915-433c-9915-27e0b01e1ffe-rabbitmq-erlang-cookie\") pod \"rabbitmq-cell1-server-0\" (UID: \"5eed809d-c915-433c-9915-27e0b01e1ffe\") " pod="openstack/rabbitmq-cell1-server-0" Dec 13 07:10:06 crc kubenswrapper[4971]: I1213 07:10:06.416386 4971 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"plugins-conf\" (UniqueName: \"kubernetes.io/configmap/5eed809d-c915-433c-9915-27e0b01e1ffe-plugins-conf\") pod \"rabbitmq-cell1-server-0\" (UID: \"5eed809d-c915-433c-9915-27e0b01e1ffe\") " pod="openstack/rabbitmq-cell1-server-0" Dec 13 07:10:06 crc kubenswrapper[4971]: I1213 07:10:06.416448 4971 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/5eed809d-c915-433c-9915-27e0b01e1ffe-config-data\") pod \"rabbitmq-cell1-server-0\" (UID: \"5eed809d-c915-433c-9915-27e0b01e1ffe\") " pod="openstack/rabbitmq-cell1-server-0" Dec 13 07:10:06 crc kubenswrapper[4971]: I1213 07:10:06.419150 4971 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"server-conf\" (UniqueName: \"kubernetes.io/configmap/5eed809d-c915-433c-9915-27e0b01e1ffe-server-conf\") pod \"rabbitmq-cell1-server-0\" (UID: \"5eed809d-c915-433c-9915-27e0b01e1ffe\") " pod="openstack/rabbitmq-cell1-server-0" Dec 13 07:10:06 crc kubenswrapper[4971]: I1213 07:10:06.425914 4971 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"pod-info\" (UniqueName: \"kubernetes.io/downward-api/5eed809d-c915-433c-9915-27e0b01e1ffe-pod-info\") pod \"rabbitmq-cell1-server-0\" (UID: \"5eed809d-c915-433c-9915-27e0b01e1ffe\") " pod="openstack/rabbitmq-cell1-server-0" Dec 13 07:10:06 crc kubenswrapper[4971]: I1213 07:10:06.428440 4971 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rabbitmq-tls\" (UniqueName: \"kubernetes.io/projected/5eed809d-c915-433c-9915-27e0b01e1ffe-rabbitmq-tls\") pod \"rabbitmq-cell1-server-0\" (UID: \"5eed809d-c915-433c-9915-27e0b01e1ffe\") " pod="openstack/rabbitmq-cell1-server-0" Dec 13 07:10:06 crc kubenswrapper[4971]: I1213 07:10:06.435872 4971 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rabbitmq-confd\" (UniqueName: \"kubernetes.io/projected/5eed809d-c915-433c-9915-27e0b01e1ffe-rabbitmq-confd\") pod \"rabbitmq-cell1-server-0\" (UID: \"5eed809d-c915-433c-9915-27e0b01e1ffe\") " pod="openstack/rabbitmq-cell1-server-0" Dec 13 07:10:06 crc kubenswrapper[4971]: I1213 07:10:06.446337 4971 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"erlang-cookie-secret\" (UniqueName: \"kubernetes.io/secret/5eed809d-c915-433c-9915-27e0b01e1ffe-erlang-cookie-secret\") pod \"rabbitmq-cell1-server-0\" (UID: \"5eed809d-c915-433c-9915-27e0b01e1ffe\") " pod="openstack/rabbitmq-cell1-server-0" Dec 13 07:10:06 crc kubenswrapper[4971]: I1213 07:10:06.448048 4971 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-jpcrh\" (UniqueName: \"kubernetes.io/projected/5eed809d-c915-433c-9915-27e0b01e1ffe-kube-api-access-jpcrh\") pod \"rabbitmq-cell1-server-0\" (UID: \"5eed809d-c915-433c-9915-27e0b01e1ffe\") " pod="openstack/rabbitmq-cell1-server-0" Dec 13 07:10:06 crc kubenswrapper[4971]: I1213 07:10:06.491071 4971 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-744ffd65bc-zhplm" event={"ID":"b4e999eb-ff02-4c70-b755-f85bd90dd637","Type":"ContainerStarted","Data":"fad71e9c347b23db5981bdc894b841c16cb490dcff033245da131c2ca048068a"} Dec 13 07:10:06 crc kubenswrapper[4971]: I1213 07:10:06.491198 4971 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"local-storage10-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage10-crc\") pod \"rabbitmq-cell1-server-0\" (UID: \"5eed809d-c915-433c-9915-27e0b01e1ffe\") " pod="openstack/rabbitmq-cell1-server-0" Dec 13 07:10:06 crc kubenswrapper[4971]: I1213 07:10:06.493767 4971 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-95f5f6995-qv9tw" event={"ID":"9a37323b-7693-41c2-8a8f-661291d1ef5e","Type":"ContainerStarted","Data":"85cf34c1f99a8f5cd4f170ba6f7da04e1970efebd7348563a6dc2bfe6bf6b2f9"} Dec 13 07:10:06 crc kubenswrapper[4971]: I1213 07:10:06.742350 4971 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/rabbitmq-cell1-server-0" Dec 13 07:10:07 crc kubenswrapper[4971]: I1213 07:10:07.246768 4971 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/rabbitmq-server-0"] Dec 13 07:10:07 crc kubenswrapper[4971]: W1213 07:10:07.311203 4971 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-poda590eb84_c8d5_4cbe_9c03_03a6ccb82073.slice/crio-b3d06f90258a1549aba7fc0a8fdb6a60a2048a29be25f55fa06f83a5ee5cd52c WatchSource:0}: Error finding container b3d06f90258a1549aba7fc0a8fdb6a60a2048a29be25f55fa06f83a5ee5cd52c: Status 404 returned error can't find the container with id b3d06f90258a1549aba7fc0a8fdb6a60a2048a29be25f55fa06f83a5ee5cd52c Dec 13 07:10:07 crc kubenswrapper[4971]: I1213 07:10:07.442800 4971 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/rabbitmq-cell1-server-0"] Dec 13 07:10:07 crc kubenswrapper[4971]: I1213 07:10:07.454681 4971 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/openstack-galera-0"] Dec 13 07:10:07 crc kubenswrapper[4971]: I1213 07:10:07.456312 4971 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/openstack-galera-0" Dec 13 07:10:07 crc kubenswrapper[4971]: I1213 07:10:07.459238 4971 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-config-data" Dec 13 07:10:07 crc kubenswrapper[4971]: I1213 07:10:07.460008 4971 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-scripts" Dec 13 07:10:07 crc kubenswrapper[4971]: I1213 07:10:07.460262 4971 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-galera-openstack-svc" Dec 13 07:10:07 crc kubenswrapper[4971]: I1213 07:10:07.460412 4971 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"galera-openstack-dockercfg-c42qp" Dec 13 07:10:07 crc kubenswrapper[4971]: I1213 07:10:07.468240 4971 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/openstack-galera-0"] Dec 13 07:10:07 crc kubenswrapper[4971]: I1213 07:10:07.472799 4971 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"combined-ca-bundle" Dec 13 07:10:07 crc kubenswrapper[4971]: I1213 07:10:07.535814 4971 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-server-0" event={"ID":"a590eb84-c8d5-4cbe-9c03-03a6ccb82073","Type":"ContainerStarted","Data":"b3d06f90258a1549aba7fc0a8fdb6a60a2048a29be25f55fa06f83a5ee5cd52c"} Dec 13 07:10:07 crc kubenswrapper[4971]: I1213 07:10:07.540594 4971 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-cell1-server-0" event={"ID":"5eed809d-c915-433c-9915-27e0b01e1ffe","Type":"ContainerStarted","Data":"532feca2c32e4b9e2c2fe06c7c12a79ff8d63c0dbe5f42443e2e3c61337e69b7"} Dec 13 07:10:07 crc kubenswrapper[4971]: I1213 07:10:07.551775 4971 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kolla-config\" (UniqueName: \"kubernetes.io/configmap/91831fd5-fc52-46ec-ac49-63859378116b-kolla-config\") pod \"openstack-galera-0\" (UID: \"91831fd5-fc52-46ec-ac49-63859378116b\") " pod="openstack/openstack-galera-0" Dec 13 07:10:07 crc kubenswrapper[4971]: I1213 07:10:07.551837 4971 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-zghhl\" (UniqueName: \"kubernetes.io/projected/91831fd5-fc52-46ec-ac49-63859378116b-kube-api-access-zghhl\") pod \"openstack-galera-0\" (UID: \"91831fd5-fc52-46ec-ac49-63859378116b\") " pod="openstack/openstack-galera-0" Dec 13 07:10:07 crc kubenswrapper[4971]: I1213 07:10:07.551874 4971 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/91831fd5-fc52-46ec-ac49-63859378116b-combined-ca-bundle\") pod \"openstack-galera-0\" (UID: \"91831fd5-fc52-46ec-ac49-63859378116b\") " pod="openstack/openstack-galera-0" Dec 13 07:10:07 crc kubenswrapper[4971]: I1213 07:10:07.551919 4971 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/91831fd5-fc52-46ec-ac49-63859378116b-operator-scripts\") pod \"openstack-galera-0\" (UID: \"91831fd5-fc52-46ec-ac49-63859378116b\") " pod="openstack/openstack-galera-0" Dec 13 07:10:07 crc kubenswrapper[4971]: I1213 07:10:07.551935 4971 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-default\" (UniqueName: \"kubernetes.io/configmap/91831fd5-fc52-46ec-ac49-63859378116b-config-data-default\") pod \"openstack-galera-0\" (UID: \"91831fd5-fc52-46ec-ac49-63859378116b\") " pod="openstack/openstack-galera-0" Dec 13 07:10:07 crc kubenswrapper[4971]: I1213 07:10:07.551953 4971 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-generated\" (UniqueName: \"kubernetes.io/empty-dir/91831fd5-fc52-46ec-ac49-63859378116b-config-data-generated\") pod \"openstack-galera-0\" (UID: \"91831fd5-fc52-46ec-ac49-63859378116b\") " pod="openstack/openstack-galera-0" Dec 13 07:10:07 crc kubenswrapper[4971]: I1213 07:10:07.551971 4971 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"galera-tls-certs\" (UniqueName: \"kubernetes.io/secret/91831fd5-fc52-46ec-ac49-63859378116b-galera-tls-certs\") pod \"openstack-galera-0\" (UID: \"91831fd5-fc52-46ec-ac49-63859378116b\") " pod="openstack/openstack-galera-0" Dec 13 07:10:07 crc kubenswrapper[4971]: I1213 07:10:07.551995 4971 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"local-storage02-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage02-crc\") pod \"openstack-galera-0\" (UID: \"91831fd5-fc52-46ec-ac49-63859378116b\") " pod="openstack/openstack-galera-0" Dec 13 07:10:07 crc kubenswrapper[4971]: I1213 07:10:07.687626 4971 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kolla-config\" (UniqueName: \"kubernetes.io/configmap/91831fd5-fc52-46ec-ac49-63859378116b-kolla-config\") pod \"openstack-galera-0\" (UID: \"91831fd5-fc52-46ec-ac49-63859378116b\") " pod="openstack/openstack-galera-0" Dec 13 07:10:07 crc kubenswrapper[4971]: I1213 07:10:07.687696 4971 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-zghhl\" (UniqueName: \"kubernetes.io/projected/91831fd5-fc52-46ec-ac49-63859378116b-kube-api-access-zghhl\") pod \"openstack-galera-0\" (UID: \"91831fd5-fc52-46ec-ac49-63859378116b\") " pod="openstack/openstack-galera-0" Dec 13 07:10:07 crc kubenswrapper[4971]: I1213 07:10:07.687723 4971 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/91831fd5-fc52-46ec-ac49-63859378116b-combined-ca-bundle\") pod \"openstack-galera-0\" (UID: \"91831fd5-fc52-46ec-ac49-63859378116b\") " pod="openstack/openstack-galera-0" Dec 13 07:10:07 crc kubenswrapper[4971]: I1213 07:10:07.687817 4971 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/91831fd5-fc52-46ec-ac49-63859378116b-operator-scripts\") pod \"openstack-galera-0\" (UID: \"91831fd5-fc52-46ec-ac49-63859378116b\") " pod="openstack/openstack-galera-0" Dec 13 07:10:07 crc kubenswrapper[4971]: I1213 07:10:07.687845 4971 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-default\" (UniqueName: \"kubernetes.io/configmap/91831fd5-fc52-46ec-ac49-63859378116b-config-data-default\") pod \"openstack-galera-0\" (UID: \"91831fd5-fc52-46ec-ac49-63859378116b\") " pod="openstack/openstack-galera-0" Dec 13 07:10:07 crc kubenswrapper[4971]: I1213 07:10:07.687881 4971 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-generated\" (UniqueName: \"kubernetes.io/empty-dir/91831fd5-fc52-46ec-ac49-63859378116b-config-data-generated\") pod \"openstack-galera-0\" (UID: \"91831fd5-fc52-46ec-ac49-63859378116b\") " pod="openstack/openstack-galera-0" Dec 13 07:10:07 crc kubenswrapper[4971]: I1213 07:10:07.687919 4971 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"galera-tls-certs\" (UniqueName: \"kubernetes.io/secret/91831fd5-fc52-46ec-ac49-63859378116b-galera-tls-certs\") pod \"openstack-galera-0\" (UID: \"91831fd5-fc52-46ec-ac49-63859378116b\") " pod="openstack/openstack-galera-0" Dec 13 07:10:07 crc kubenswrapper[4971]: I1213 07:10:07.687951 4971 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"local-storage02-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage02-crc\") pod \"openstack-galera-0\" (UID: \"91831fd5-fc52-46ec-ac49-63859378116b\") " pod="openstack/openstack-galera-0" Dec 13 07:10:07 crc kubenswrapper[4971]: I1213 07:10:07.688432 4971 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"local-storage02-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage02-crc\") pod \"openstack-galera-0\" (UID: \"91831fd5-fc52-46ec-ac49-63859378116b\") device mount path \"/mnt/openstack/pv02\"" pod="openstack/openstack-galera-0" Dec 13 07:10:07 crc kubenswrapper[4971]: I1213 07:10:07.694788 4971 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kolla-config\" (UniqueName: \"kubernetes.io/configmap/91831fd5-fc52-46ec-ac49-63859378116b-kolla-config\") pod \"openstack-galera-0\" (UID: \"91831fd5-fc52-46ec-ac49-63859378116b\") " pod="openstack/openstack-galera-0" Dec 13 07:10:07 crc kubenswrapper[4971]: I1213 07:10:07.695054 4971 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/91831fd5-fc52-46ec-ac49-63859378116b-operator-scripts\") pod \"openstack-galera-0\" (UID: \"91831fd5-fc52-46ec-ac49-63859378116b\") " pod="openstack/openstack-galera-0" Dec 13 07:10:07 crc kubenswrapper[4971]: I1213 07:10:07.696945 4971 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-default\" (UniqueName: \"kubernetes.io/configmap/91831fd5-fc52-46ec-ac49-63859378116b-config-data-default\") pod \"openstack-galera-0\" (UID: \"91831fd5-fc52-46ec-ac49-63859378116b\") " pod="openstack/openstack-galera-0" Dec 13 07:10:07 crc kubenswrapper[4971]: I1213 07:10:07.697495 4971 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-generated\" (UniqueName: \"kubernetes.io/empty-dir/91831fd5-fc52-46ec-ac49-63859378116b-config-data-generated\") pod \"openstack-galera-0\" (UID: \"91831fd5-fc52-46ec-ac49-63859378116b\") " pod="openstack/openstack-galera-0" Dec 13 07:10:07 crc kubenswrapper[4971]: I1213 07:10:07.718152 4971 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"galera-tls-certs\" (UniqueName: \"kubernetes.io/secret/91831fd5-fc52-46ec-ac49-63859378116b-galera-tls-certs\") pod \"openstack-galera-0\" (UID: \"91831fd5-fc52-46ec-ac49-63859378116b\") " pod="openstack/openstack-galera-0" Dec 13 07:10:07 crc kubenswrapper[4971]: I1213 07:10:07.736562 4971 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/91831fd5-fc52-46ec-ac49-63859378116b-combined-ca-bundle\") pod \"openstack-galera-0\" (UID: \"91831fd5-fc52-46ec-ac49-63859378116b\") " pod="openstack/openstack-galera-0" Dec 13 07:10:07 crc kubenswrapper[4971]: I1213 07:10:07.738575 4971 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-zghhl\" (UniqueName: \"kubernetes.io/projected/91831fd5-fc52-46ec-ac49-63859378116b-kube-api-access-zghhl\") pod \"openstack-galera-0\" (UID: \"91831fd5-fc52-46ec-ac49-63859378116b\") " pod="openstack/openstack-galera-0" Dec 13 07:10:07 crc kubenswrapper[4971]: I1213 07:10:07.738901 4971 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"local-storage02-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage02-crc\") pod \"openstack-galera-0\" (UID: \"91831fd5-fc52-46ec-ac49-63859378116b\") " pod="openstack/openstack-galera-0" Dec 13 07:10:07 crc kubenswrapper[4971]: I1213 07:10:07.819356 4971 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/openstack-galera-0" Dec 13 07:10:08 crc kubenswrapper[4971]: I1213 07:10:08.841258 4971 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/openstack-cell1-galera-0"] Dec 13 07:10:08 crc kubenswrapper[4971]: I1213 07:10:08.842885 4971 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/openstack-cell1-galera-0" Dec 13 07:10:08 crc kubenswrapper[4971]: I1213 07:10:08.853395 4971 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-galera-openstack-cell1-svc" Dec 13 07:10:08 crc kubenswrapper[4971]: I1213 07:10:08.853690 4971 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"galera-openstack-cell1-dockercfg-c64jr" Dec 13 07:10:08 crc kubenswrapper[4971]: I1213 07:10:08.854275 4971 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-cell1-scripts" Dec 13 07:10:08 crc kubenswrapper[4971]: I1213 07:10:08.854357 4971 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-cell1-config-data" Dec 13 07:10:08 crc kubenswrapper[4971]: I1213 07:10:08.860297 4971 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/openstack-cell1-galera-0"] Dec 13 07:10:09 crc kubenswrapper[4971]: I1213 07:10:09.062006 4971 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-xxpf9\" (UniqueName: \"kubernetes.io/projected/96571a02-70f1-46e7-bdd3-935b1f43ecc6-kube-api-access-xxpf9\") pod \"openstack-cell1-galera-0\" (UID: \"96571a02-70f1-46e7-bdd3-935b1f43ecc6\") " pod="openstack/openstack-cell1-galera-0" Dec 13 07:10:09 crc kubenswrapper[4971]: I1213 07:10:09.062079 4971 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"local-storage07-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage07-crc\") pod \"openstack-cell1-galera-0\" (UID: \"96571a02-70f1-46e7-bdd3-935b1f43ecc6\") " pod="openstack/openstack-cell1-galera-0" Dec 13 07:10:09 crc kubenswrapper[4971]: I1213 07:10:09.062108 4971 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"galera-tls-certs\" (UniqueName: \"kubernetes.io/secret/96571a02-70f1-46e7-bdd3-935b1f43ecc6-galera-tls-certs\") pod \"openstack-cell1-galera-0\" (UID: \"96571a02-70f1-46e7-bdd3-935b1f43ecc6\") " pod="openstack/openstack-cell1-galera-0" Dec 13 07:10:09 crc kubenswrapper[4971]: I1213 07:10:09.062128 4971 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kolla-config\" (UniqueName: \"kubernetes.io/configmap/96571a02-70f1-46e7-bdd3-935b1f43ecc6-kolla-config\") pod \"openstack-cell1-galera-0\" (UID: \"96571a02-70f1-46e7-bdd3-935b1f43ecc6\") " pod="openstack/openstack-cell1-galera-0" Dec 13 07:10:09 crc kubenswrapper[4971]: I1213 07:10:09.062151 4971 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/96571a02-70f1-46e7-bdd3-935b1f43ecc6-combined-ca-bundle\") pod \"openstack-cell1-galera-0\" (UID: \"96571a02-70f1-46e7-bdd3-935b1f43ecc6\") " pod="openstack/openstack-cell1-galera-0" Dec 13 07:10:09 crc kubenswrapper[4971]: I1213 07:10:09.062171 4971 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-default\" (UniqueName: \"kubernetes.io/configmap/96571a02-70f1-46e7-bdd3-935b1f43ecc6-config-data-default\") pod \"openstack-cell1-galera-0\" (UID: \"96571a02-70f1-46e7-bdd3-935b1f43ecc6\") " pod="openstack/openstack-cell1-galera-0" Dec 13 07:10:09 crc kubenswrapper[4971]: I1213 07:10:09.062196 4971 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-generated\" (UniqueName: \"kubernetes.io/empty-dir/96571a02-70f1-46e7-bdd3-935b1f43ecc6-config-data-generated\") pod \"openstack-cell1-galera-0\" (UID: \"96571a02-70f1-46e7-bdd3-935b1f43ecc6\") " pod="openstack/openstack-cell1-galera-0" Dec 13 07:10:09 crc kubenswrapper[4971]: I1213 07:10:09.062232 4971 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/96571a02-70f1-46e7-bdd3-935b1f43ecc6-operator-scripts\") pod \"openstack-cell1-galera-0\" (UID: \"96571a02-70f1-46e7-bdd3-935b1f43ecc6\") " pod="openstack/openstack-cell1-galera-0" Dec 13 07:10:09 crc kubenswrapper[4971]: I1213 07:10:09.164030 4971 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-default\" (UniqueName: \"kubernetes.io/configmap/96571a02-70f1-46e7-bdd3-935b1f43ecc6-config-data-default\") pod \"openstack-cell1-galera-0\" (UID: \"96571a02-70f1-46e7-bdd3-935b1f43ecc6\") " pod="openstack/openstack-cell1-galera-0" Dec 13 07:10:09 crc kubenswrapper[4971]: I1213 07:10:09.164328 4971 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-generated\" (UniqueName: \"kubernetes.io/empty-dir/96571a02-70f1-46e7-bdd3-935b1f43ecc6-config-data-generated\") pod \"openstack-cell1-galera-0\" (UID: \"96571a02-70f1-46e7-bdd3-935b1f43ecc6\") " pod="openstack/openstack-cell1-galera-0" Dec 13 07:10:09 crc kubenswrapper[4971]: I1213 07:10:09.164369 4971 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/96571a02-70f1-46e7-bdd3-935b1f43ecc6-operator-scripts\") pod \"openstack-cell1-galera-0\" (UID: \"96571a02-70f1-46e7-bdd3-935b1f43ecc6\") " pod="openstack/openstack-cell1-galera-0" Dec 13 07:10:09 crc kubenswrapper[4971]: I1213 07:10:09.164404 4971 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-xxpf9\" (UniqueName: \"kubernetes.io/projected/96571a02-70f1-46e7-bdd3-935b1f43ecc6-kube-api-access-xxpf9\") pod \"openstack-cell1-galera-0\" (UID: \"96571a02-70f1-46e7-bdd3-935b1f43ecc6\") " pod="openstack/openstack-cell1-galera-0" Dec 13 07:10:09 crc kubenswrapper[4971]: I1213 07:10:09.164447 4971 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"local-storage07-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage07-crc\") pod \"openstack-cell1-galera-0\" (UID: \"96571a02-70f1-46e7-bdd3-935b1f43ecc6\") " pod="openstack/openstack-cell1-galera-0" Dec 13 07:10:09 crc kubenswrapper[4971]: I1213 07:10:09.164470 4971 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"galera-tls-certs\" (UniqueName: \"kubernetes.io/secret/96571a02-70f1-46e7-bdd3-935b1f43ecc6-galera-tls-certs\") pod \"openstack-cell1-galera-0\" (UID: \"96571a02-70f1-46e7-bdd3-935b1f43ecc6\") " pod="openstack/openstack-cell1-galera-0" Dec 13 07:10:09 crc kubenswrapper[4971]: I1213 07:10:09.164487 4971 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kolla-config\" (UniqueName: \"kubernetes.io/configmap/96571a02-70f1-46e7-bdd3-935b1f43ecc6-kolla-config\") pod \"openstack-cell1-galera-0\" (UID: \"96571a02-70f1-46e7-bdd3-935b1f43ecc6\") " pod="openstack/openstack-cell1-galera-0" Dec 13 07:10:09 crc kubenswrapper[4971]: I1213 07:10:09.164510 4971 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/96571a02-70f1-46e7-bdd3-935b1f43ecc6-combined-ca-bundle\") pod \"openstack-cell1-galera-0\" (UID: \"96571a02-70f1-46e7-bdd3-935b1f43ecc6\") " pod="openstack/openstack-cell1-galera-0" Dec 13 07:10:09 crc kubenswrapper[4971]: I1213 07:10:09.165477 4971 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/memcached-0"] Dec 13 07:10:09 crc kubenswrapper[4971]: I1213 07:10:09.165996 4971 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-default\" (UniqueName: \"kubernetes.io/configmap/96571a02-70f1-46e7-bdd3-935b1f43ecc6-config-data-default\") pod \"openstack-cell1-galera-0\" (UID: \"96571a02-70f1-46e7-bdd3-935b1f43ecc6\") " pod="openstack/openstack-cell1-galera-0" Dec 13 07:10:09 crc kubenswrapper[4971]: I1213 07:10:09.166219 4971 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-generated\" (UniqueName: \"kubernetes.io/empty-dir/96571a02-70f1-46e7-bdd3-935b1f43ecc6-config-data-generated\") pod \"openstack-cell1-galera-0\" (UID: \"96571a02-70f1-46e7-bdd3-935b1f43ecc6\") " pod="openstack/openstack-cell1-galera-0" Dec 13 07:10:09 crc kubenswrapper[4971]: I1213 07:10:09.170186 4971 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/memcached-0" Dec 13 07:10:09 crc kubenswrapper[4971]: I1213 07:10:09.171922 4971 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"local-storage07-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage07-crc\") pod \"openstack-cell1-galera-0\" (UID: \"96571a02-70f1-46e7-bdd3-935b1f43ecc6\") device mount path \"/mnt/openstack/pv07\"" pod="openstack/openstack-cell1-galera-0" Dec 13 07:10:09 crc kubenswrapper[4971]: I1213 07:10:09.172244 4971 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kolla-config\" (UniqueName: \"kubernetes.io/configmap/96571a02-70f1-46e7-bdd3-935b1f43ecc6-kolla-config\") pod \"openstack-cell1-galera-0\" (UID: \"96571a02-70f1-46e7-bdd3-935b1f43ecc6\") " pod="openstack/openstack-cell1-galera-0" Dec 13 07:10:09 crc kubenswrapper[4971]: I1213 07:10:09.173771 4971 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-memcached-svc" Dec 13 07:10:09 crc kubenswrapper[4971]: I1213 07:10:09.175429 4971 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/96571a02-70f1-46e7-bdd3-935b1f43ecc6-combined-ca-bundle\") pod \"openstack-cell1-galera-0\" (UID: \"96571a02-70f1-46e7-bdd3-935b1f43ecc6\") " pod="openstack/openstack-cell1-galera-0" Dec 13 07:10:09 crc kubenswrapper[4971]: I1213 07:10:09.175503 4971 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"memcached-memcached-dockercfg-g8ddc" Dec 13 07:10:09 crc kubenswrapper[4971]: I1213 07:10:09.175509 4971 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"memcached-config-data" Dec 13 07:10:09 crc kubenswrapper[4971]: I1213 07:10:09.176017 4971 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/96571a02-70f1-46e7-bdd3-935b1f43ecc6-operator-scripts\") pod \"openstack-cell1-galera-0\" (UID: \"96571a02-70f1-46e7-bdd3-935b1f43ecc6\") " pod="openstack/openstack-cell1-galera-0" Dec 13 07:10:09 crc kubenswrapper[4971]: I1213 07:10:09.188040 4971 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"galera-tls-certs\" (UniqueName: \"kubernetes.io/secret/96571a02-70f1-46e7-bdd3-935b1f43ecc6-galera-tls-certs\") pod \"openstack-cell1-galera-0\" (UID: \"96571a02-70f1-46e7-bdd3-935b1f43ecc6\") " pod="openstack/openstack-cell1-galera-0" Dec 13 07:10:09 crc kubenswrapper[4971]: I1213 07:10:09.198276 4971 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/memcached-0"] Dec 13 07:10:09 crc kubenswrapper[4971]: I1213 07:10:09.201074 4971 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-xxpf9\" (UniqueName: \"kubernetes.io/projected/96571a02-70f1-46e7-bdd3-935b1f43ecc6-kube-api-access-xxpf9\") pod \"openstack-cell1-galera-0\" (UID: \"96571a02-70f1-46e7-bdd3-935b1f43ecc6\") " pod="openstack/openstack-cell1-galera-0" Dec 13 07:10:09 crc kubenswrapper[4971]: I1213 07:10:09.224556 4971 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"local-storage07-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage07-crc\") pod \"openstack-cell1-galera-0\" (UID: \"96571a02-70f1-46e7-bdd3-935b1f43ecc6\") " pod="openstack/openstack-cell1-galera-0" Dec 13 07:10:09 crc kubenswrapper[4971]: I1213 07:10:09.299162 4971 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"memcached-tls-certs\" (UniqueName: \"kubernetes.io/secret/e6adf8ca-c137-4b7d-bf8c-74c0766e57f0-memcached-tls-certs\") pod \"memcached-0\" (UID: \"e6adf8ca-c137-4b7d-bf8c-74c0766e57f0\") " pod="openstack/memcached-0" Dec 13 07:10:09 crc kubenswrapper[4971]: I1213 07:10:09.299217 4971 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/e6adf8ca-c137-4b7d-bf8c-74c0766e57f0-config-data\") pod \"memcached-0\" (UID: \"e6adf8ca-c137-4b7d-bf8c-74c0766e57f0\") " pod="openstack/memcached-0" Dec 13 07:10:09 crc kubenswrapper[4971]: I1213 07:10:09.299305 4971 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/e6adf8ca-c137-4b7d-bf8c-74c0766e57f0-combined-ca-bundle\") pod \"memcached-0\" (UID: \"e6adf8ca-c137-4b7d-bf8c-74c0766e57f0\") " pod="openstack/memcached-0" Dec 13 07:10:09 crc kubenswrapper[4971]: I1213 07:10:09.299342 4971 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kolla-config\" (UniqueName: \"kubernetes.io/configmap/e6adf8ca-c137-4b7d-bf8c-74c0766e57f0-kolla-config\") pod \"memcached-0\" (UID: \"e6adf8ca-c137-4b7d-bf8c-74c0766e57f0\") " pod="openstack/memcached-0" Dec 13 07:10:09 crc kubenswrapper[4971]: I1213 07:10:09.299418 4971 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-dmspm\" (UniqueName: \"kubernetes.io/projected/e6adf8ca-c137-4b7d-bf8c-74c0766e57f0-kube-api-access-dmspm\") pod \"memcached-0\" (UID: \"e6adf8ca-c137-4b7d-bf8c-74c0766e57f0\") " pod="openstack/memcached-0" Dec 13 07:10:09 crc kubenswrapper[4971]: I1213 07:10:09.400647 4971 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-dmspm\" (UniqueName: \"kubernetes.io/projected/e6adf8ca-c137-4b7d-bf8c-74c0766e57f0-kube-api-access-dmspm\") pod \"memcached-0\" (UID: \"e6adf8ca-c137-4b7d-bf8c-74c0766e57f0\") " pod="openstack/memcached-0" Dec 13 07:10:09 crc kubenswrapper[4971]: I1213 07:10:09.400743 4971 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"memcached-tls-certs\" (UniqueName: \"kubernetes.io/secret/e6adf8ca-c137-4b7d-bf8c-74c0766e57f0-memcached-tls-certs\") pod \"memcached-0\" (UID: \"e6adf8ca-c137-4b7d-bf8c-74c0766e57f0\") " pod="openstack/memcached-0" Dec 13 07:10:09 crc kubenswrapper[4971]: I1213 07:10:09.400771 4971 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/e6adf8ca-c137-4b7d-bf8c-74c0766e57f0-config-data\") pod \"memcached-0\" (UID: \"e6adf8ca-c137-4b7d-bf8c-74c0766e57f0\") " pod="openstack/memcached-0" Dec 13 07:10:09 crc kubenswrapper[4971]: I1213 07:10:09.400816 4971 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/e6adf8ca-c137-4b7d-bf8c-74c0766e57f0-combined-ca-bundle\") pod \"memcached-0\" (UID: \"e6adf8ca-c137-4b7d-bf8c-74c0766e57f0\") " pod="openstack/memcached-0" Dec 13 07:10:09 crc kubenswrapper[4971]: I1213 07:10:09.400843 4971 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kolla-config\" (UniqueName: \"kubernetes.io/configmap/e6adf8ca-c137-4b7d-bf8c-74c0766e57f0-kolla-config\") pod \"memcached-0\" (UID: \"e6adf8ca-c137-4b7d-bf8c-74c0766e57f0\") " pod="openstack/memcached-0" Dec 13 07:10:09 crc kubenswrapper[4971]: I1213 07:10:09.401795 4971 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kolla-config\" (UniqueName: \"kubernetes.io/configmap/e6adf8ca-c137-4b7d-bf8c-74c0766e57f0-kolla-config\") pod \"memcached-0\" (UID: \"e6adf8ca-c137-4b7d-bf8c-74c0766e57f0\") " pod="openstack/memcached-0" Dec 13 07:10:09 crc kubenswrapper[4971]: I1213 07:10:09.403985 4971 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/e6adf8ca-c137-4b7d-bf8c-74c0766e57f0-config-data\") pod \"memcached-0\" (UID: \"e6adf8ca-c137-4b7d-bf8c-74c0766e57f0\") " pod="openstack/memcached-0" Dec 13 07:10:09 crc kubenswrapper[4971]: I1213 07:10:09.406778 4971 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"memcached-tls-certs\" (UniqueName: \"kubernetes.io/secret/e6adf8ca-c137-4b7d-bf8c-74c0766e57f0-memcached-tls-certs\") pod \"memcached-0\" (UID: \"e6adf8ca-c137-4b7d-bf8c-74c0766e57f0\") " pod="openstack/memcached-0" Dec 13 07:10:09 crc kubenswrapper[4971]: I1213 07:10:09.413724 4971 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/e6adf8ca-c137-4b7d-bf8c-74c0766e57f0-combined-ca-bundle\") pod \"memcached-0\" (UID: \"e6adf8ca-c137-4b7d-bf8c-74c0766e57f0\") " pod="openstack/memcached-0" Dec 13 07:10:09 crc kubenswrapper[4971]: I1213 07:10:09.427179 4971 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-dmspm\" (UniqueName: \"kubernetes.io/projected/e6adf8ca-c137-4b7d-bf8c-74c0766e57f0-kube-api-access-dmspm\") pod \"memcached-0\" (UID: \"e6adf8ca-c137-4b7d-bf8c-74c0766e57f0\") " pod="openstack/memcached-0" Dec 13 07:10:09 crc kubenswrapper[4971]: I1213 07:10:09.433144 4971 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/openstack-galera-0"] Dec 13 07:10:09 crc kubenswrapper[4971]: W1213 07:10:09.451666 4971 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod91831fd5_fc52_46ec_ac49_63859378116b.slice/crio-c292124214e83ba4a37f014cf06287f52f2c3ccb18af30e90dd4e79e88294ba4 WatchSource:0}: Error finding container c292124214e83ba4a37f014cf06287f52f2c3ccb18af30e90dd4e79e88294ba4: Status 404 returned error can't find the container with id c292124214e83ba4a37f014cf06287f52f2c3ccb18af30e90dd4e79e88294ba4 Dec 13 07:10:09 crc kubenswrapper[4971]: I1213 07:10:09.494682 4971 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/openstack-cell1-galera-0" Dec 13 07:10:09 crc kubenswrapper[4971]: I1213 07:10:09.727426 4971 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/openstack-galera-0" event={"ID":"91831fd5-fc52-46ec-ac49-63859378116b","Type":"ContainerStarted","Data":"c292124214e83ba4a37f014cf06287f52f2c3ccb18af30e90dd4e79e88294ba4"} Dec 13 07:10:09 crc kubenswrapper[4971]: I1213 07:10:09.728960 4971 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/memcached-0" Dec 13 07:10:10 crc kubenswrapper[4971]: I1213 07:10:10.397306 4971 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/memcached-0"] Dec 13 07:10:10 crc kubenswrapper[4971]: W1213 07:10:10.410545 4971 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pode6adf8ca_c137_4b7d_bf8c_74c0766e57f0.slice/crio-3467b949c7c327ee6d723d2a3741f528941c8837e23a0778a3c451f1536818e8 WatchSource:0}: Error finding container 3467b949c7c327ee6d723d2a3741f528941c8837e23a0778a3c451f1536818e8: Status 404 returned error can't find the container with id 3467b949c7c327ee6d723d2a3741f528941c8837e23a0778a3c451f1536818e8 Dec 13 07:10:10 crc kubenswrapper[4971]: I1213 07:10:10.752902 4971 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/openstack-cell1-galera-0"] Dec 13 07:10:10 crc kubenswrapper[4971]: W1213 07:10:10.774838 4971 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod96571a02_70f1_46e7_bdd3_935b1f43ecc6.slice/crio-317c0690fd6fcc4d3cdd1b4ba41eced23d5b6564408661d3e455bbb1d485ec5f WatchSource:0}: Error finding container 317c0690fd6fcc4d3cdd1b4ba41eced23d5b6564408661d3e455bbb1d485ec5f: Status 404 returned error can't find the container with id 317c0690fd6fcc4d3cdd1b4ba41eced23d5b6564408661d3e455bbb1d485ec5f Dec 13 07:10:10 crc kubenswrapper[4971]: I1213 07:10:10.783899 4971 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/memcached-0" event={"ID":"e6adf8ca-c137-4b7d-bf8c-74c0766e57f0","Type":"ContainerStarted","Data":"3467b949c7c327ee6d723d2a3741f528941c8837e23a0778a3c451f1536818e8"} Dec 13 07:10:11 crc kubenswrapper[4971]: I1213 07:10:11.611014 4971 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/kube-state-metrics-0"] Dec 13 07:10:11 crc kubenswrapper[4971]: I1213 07:10:11.612956 4971 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/kube-state-metrics-0" Dec 13 07:10:11 crc kubenswrapper[4971]: I1213 07:10:11.615156 4971 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"telemetry-ceilometer-dockercfg-x8ss7" Dec 13 07:10:11 crc kubenswrapper[4971]: I1213 07:10:11.622858 4971 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-c58tt\" (UniqueName: \"kubernetes.io/projected/2bab8c0a-bd76-4265-8355-8df0c6fce6c3-kube-api-access-c58tt\") pod \"kube-state-metrics-0\" (UID: \"2bab8c0a-bd76-4265-8355-8df0c6fce6c3\") " pod="openstack/kube-state-metrics-0" Dec 13 07:10:11 crc kubenswrapper[4971]: I1213 07:10:11.626384 4971 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/kube-state-metrics-0"] Dec 13 07:10:11 crc kubenswrapper[4971]: I1213 07:10:11.724787 4971 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-c58tt\" (UniqueName: \"kubernetes.io/projected/2bab8c0a-bd76-4265-8355-8df0c6fce6c3-kube-api-access-c58tt\") pod \"kube-state-metrics-0\" (UID: \"2bab8c0a-bd76-4265-8355-8df0c6fce6c3\") " pod="openstack/kube-state-metrics-0" Dec 13 07:10:11 crc kubenswrapper[4971]: I1213 07:10:11.753475 4971 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-c58tt\" (UniqueName: \"kubernetes.io/projected/2bab8c0a-bd76-4265-8355-8df0c6fce6c3-kube-api-access-c58tt\") pod \"kube-state-metrics-0\" (UID: \"2bab8c0a-bd76-4265-8355-8df0c6fce6c3\") " pod="openstack/kube-state-metrics-0" Dec 13 07:10:11 crc kubenswrapper[4971]: I1213 07:10:11.899050 4971 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/openstack-cell1-galera-0" event={"ID":"96571a02-70f1-46e7-bdd3-935b1f43ecc6","Type":"ContainerStarted","Data":"317c0690fd6fcc4d3cdd1b4ba41eced23d5b6564408661d3e455bbb1d485ec5f"} Dec 13 07:10:11 crc kubenswrapper[4971]: I1213 07:10:11.931813 4971 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/kube-state-metrics-0" Dec 13 07:10:13 crc kubenswrapper[4971]: I1213 07:10:13.036251 4971 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/kube-state-metrics-0"] Dec 13 07:10:15 crc kubenswrapper[4971]: I1213 07:10:15.014927 4971 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ovn-controller-k5fz9"] Dec 13 07:10:15 crc kubenswrapper[4971]: I1213 07:10:15.017171 4971 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-controller-k5fz9" Dec 13 07:10:15 crc kubenswrapper[4971]: I1213 07:10:15.019265 4971 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-ovncontroller-ovndbs" Dec 13 07:10:15 crc kubenswrapper[4971]: I1213 07:10:15.019553 4971 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ovncontroller-ovncontroller-dockercfg-sgz5v" Dec 13 07:10:15 crc kubenswrapper[4971]: I1213 07:10:15.020913 4971 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"ovncontroller-scripts" Dec 13 07:10:15 crc kubenswrapper[4971]: I1213 07:10:15.031455 4971 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovn-controller-k5fz9"] Dec 13 07:10:15 crc kubenswrapper[4971]: I1213 07:10:15.061287 4971 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-rj7r9\" (UniqueName: \"kubernetes.io/projected/c0bf2685-ca54-4508-a580-531b05473e76-kube-api-access-rj7r9\") pod \"ovn-controller-k5fz9\" (UID: \"c0bf2685-ca54-4508-a580-531b05473e76\") " pod="openstack/ovn-controller-k5fz9" Dec 13 07:10:15 crc kubenswrapper[4971]: I1213 07:10:15.061366 4971 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/c0bf2685-ca54-4508-a580-531b05473e76-scripts\") pod \"ovn-controller-k5fz9\" (UID: \"c0bf2685-ca54-4508-a580-531b05473e76\") " pod="openstack/ovn-controller-k5fz9" Dec 13 07:10:15 crc kubenswrapper[4971]: I1213 07:10:15.061395 4971 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-run-ovn\" (UniqueName: \"kubernetes.io/host-path/c0bf2685-ca54-4508-a580-531b05473e76-var-run-ovn\") pod \"ovn-controller-k5fz9\" (UID: \"c0bf2685-ca54-4508-a580-531b05473e76\") " pod="openstack/ovn-controller-k5fz9" Dec 13 07:10:15 crc kubenswrapper[4971]: I1213 07:10:15.061414 4971 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovn-controller-tls-certs\" (UniqueName: \"kubernetes.io/secret/c0bf2685-ca54-4508-a580-531b05473e76-ovn-controller-tls-certs\") pod \"ovn-controller-k5fz9\" (UID: \"c0bf2685-ca54-4508-a580-531b05473e76\") " pod="openstack/ovn-controller-k5fz9" Dec 13 07:10:15 crc kubenswrapper[4971]: I1213 07:10:15.062967 4971 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/c0bf2685-ca54-4508-a580-531b05473e76-combined-ca-bundle\") pod \"ovn-controller-k5fz9\" (UID: \"c0bf2685-ca54-4508-a580-531b05473e76\") " pod="openstack/ovn-controller-k5fz9" Dec 13 07:10:15 crc kubenswrapper[4971]: I1213 07:10:15.062987 4971 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-log-ovn\" (UniqueName: \"kubernetes.io/host-path/c0bf2685-ca54-4508-a580-531b05473e76-var-log-ovn\") pod \"ovn-controller-k5fz9\" (UID: \"c0bf2685-ca54-4508-a580-531b05473e76\") " pod="openstack/ovn-controller-k5fz9" Dec 13 07:10:15 crc kubenswrapper[4971]: I1213 07:10:15.063008 4971 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-run\" (UniqueName: \"kubernetes.io/host-path/c0bf2685-ca54-4508-a580-531b05473e76-var-run\") pod \"ovn-controller-k5fz9\" (UID: \"c0bf2685-ca54-4508-a580-531b05473e76\") " pod="openstack/ovn-controller-k5fz9" Dec 13 07:10:15 crc kubenswrapper[4971]: I1213 07:10:15.163977 4971 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-run\" (UniqueName: \"kubernetes.io/host-path/c0bf2685-ca54-4508-a580-531b05473e76-var-run\") pod \"ovn-controller-k5fz9\" (UID: \"c0bf2685-ca54-4508-a580-531b05473e76\") " pod="openstack/ovn-controller-k5fz9" Dec 13 07:10:15 crc kubenswrapper[4971]: I1213 07:10:15.164054 4971 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-rj7r9\" (UniqueName: \"kubernetes.io/projected/c0bf2685-ca54-4508-a580-531b05473e76-kube-api-access-rj7r9\") pod \"ovn-controller-k5fz9\" (UID: \"c0bf2685-ca54-4508-a580-531b05473e76\") " pod="openstack/ovn-controller-k5fz9" Dec 13 07:10:15 crc kubenswrapper[4971]: I1213 07:10:15.164143 4971 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/c0bf2685-ca54-4508-a580-531b05473e76-scripts\") pod \"ovn-controller-k5fz9\" (UID: \"c0bf2685-ca54-4508-a580-531b05473e76\") " pod="openstack/ovn-controller-k5fz9" Dec 13 07:10:15 crc kubenswrapper[4971]: I1213 07:10:15.164180 4971 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-run-ovn\" (UniqueName: \"kubernetes.io/host-path/c0bf2685-ca54-4508-a580-531b05473e76-var-run-ovn\") pod \"ovn-controller-k5fz9\" (UID: \"c0bf2685-ca54-4508-a580-531b05473e76\") " pod="openstack/ovn-controller-k5fz9" Dec 13 07:10:15 crc kubenswrapper[4971]: I1213 07:10:15.164208 4971 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovn-controller-tls-certs\" (UniqueName: \"kubernetes.io/secret/c0bf2685-ca54-4508-a580-531b05473e76-ovn-controller-tls-certs\") pod \"ovn-controller-k5fz9\" (UID: \"c0bf2685-ca54-4508-a580-531b05473e76\") " pod="openstack/ovn-controller-k5fz9" Dec 13 07:10:15 crc kubenswrapper[4971]: I1213 07:10:15.164243 4971 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/c0bf2685-ca54-4508-a580-531b05473e76-combined-ca-bundle\") pod \"ovn-controller-k5fz9\" (UID: \"c0bf2685-ca54-4508-a580-531b05473e76\") " pod="openstack/ovn-controller-k5fz9" Dec 13 07:10:15 crc kubenswrapper[4971]: I1213 07:10:15.164264 4971 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-log-ovn\" (UniqueName: \"kubernetes.io/host-path/c0bf2685-ca54-4508-a580-531b05473e76-var-log-ovn\") pod \"ovn-controller-k5fz9\" (UID: \"c0bf2685-ca54-4508-a580-531b05473e76\") " pod="openstack/ovn-controller-k5fz9" Dec 13 07:10:15 crc kubenswrapper[4971]: I1213 07:10:15.215026 4971 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-run\" (UniqueName: \"kubernetes.io/host-path/c0bf2685-ca54-4508-a580-531b05473e76-var-run\") pod \"ovn-controller-k5fz9\" (UID: \"c0bf2685-ca54-4508-a580-531b05473e76\") " pod="openstack/ovn-controller-k5fz9" Dec 13 07:10:15 crc kubenswrapper[4971]: I1213 07:10:15.290765 4971 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-log-ovn\" (UniqueName: \"kubernetes.io/host-path/c0bf2685-ca54-4508-a580-531b05473e76-var-log-ovn\") pod \"ovn-controller-k5fz9\" (UID: \"c0bf2685-ca54-4508-a580-531b05473e76\") " pod="openstack/ovn-controller-k5fz9" Dec 13 07:10:15 crc kubenswrapper[4971]: I1213 07:10:15.296400 4971 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-run-ovn\" (UniqueName: \"kubernetes.io/host-path/c0bf2685-ca54-4508-a580-531b05473e76-var-run-ovn\") pod \"ovn-controller-k5fz9\" (UID: \"c0bf2685-ca54-4508-a580-531b05473e76\") " pod="openstack/ovn-controller-k5fz9" Dec 13 07:10:15 crc kubenswrapper[4971]: I1213 07:10:15.299258 4971 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-rj7r9\" (UniqueName: \"kubernetes.io/projected/c0bf2685-ca54-4508-a580-531b05473e76-kube-api-access-rj7r9\") pod \"ovn-controller-k5fz9\" (UID: \"c0bf2685-ca54-4508-a580-531b05473e76\") " pod="openstack/ovn-controller-k5fz9" Dec 13 07:10:15 crc kubenswrapper[4971]: I1213 07:10:15.302027 4971 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/c0bf2685-ca54-4508-a580-531b05473e76-combined-ca-bundle\") pod \"ovn-controller-k5fz9\" (UID: \"c0bf2685-ca54-4508-a580-531b05473e76\") " pod="openstack/ovn-controller-k5fz9" Dec 13 07:10:15 crc kubenswrapper[4971]: I1213 07:10:15.303782 4971 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/c0bf2685-ca54-4508-a580-531b05473e76-scripts\") pod \"ovn-controller-k5fz9\" (UID: \"c0bf2685-ca54-4508-a580-531b05473e76\") " pod="openstack/ovn-controller-k5fz9" Dec 13 07:10:15 crc kubenswrapper[4971]: I1213 07:10:15.315444 4971 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovn-controller-tls-certs\" (UniqueName: \"kubernetes.io/secret/c0bf2685-ca54-4508-a580-531b05473e76-ovn-controller-tls-certs\") pod \"ovn-controller-k5fz9\" (UID: \"c0bf2685-ca54-4508-a580-531b05473e76\") " pod="openstack/ovn-controller-k5fz9" Dec 13 07:10:15 crc kubenswrapper[4971]: I1213 07:10:15.507177 4971 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-controller-k5fz9" Dec 13 07:10:15 crc kubenswrapper[4971]: I1213 07:10:15.544975 4971 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ovn-controller-ovs-jwh28"] Dec 13 07:10:15 crc kubenswrapper[4971]: I1213 07:10:15.556996 4971 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-controller-ovs-jwh28" Dec 13 07:10:15 crc kubenswrapper[4971]: I1213 07:10:15.618577 4971 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovn-controller-ovs-jwh28"] Dec 13 07:10:15 crc kubenswrapper[4971]: I1213 07:10:15.708644 4971 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ovsdbserver-nb-0"] Dec 13 07:10:15 crc kubenswrapper[4971]: I1213 07:10:15.712903 4971 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovsdbserver-nb-0" Dec 13 07:10:15 crc kubenswrapper[4971]: I1213 07:10:15.715237 4971 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ovncluster-ovndbcluster-nb-dockercfg-q6c75" Dec 13 07:10:15 crc kubenswrapper[4971]: I1213 07:10:15.715633 4971 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-ovn-metrics" Dec 13 07:10:15 crc kubenswrapper[4971]: I1213 07:10:15.715855 4971 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"ovndbcluster-nb-config" Dec 13 07:10:15 crc kubenswrapper[4971]: I1213 07:10:15.716048 4971 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-ovndbcluster-nb-ovndbs" Dec 13 07:10:15 crc kubenswrapper[4971]: I1213 07:10:15.716340 4971 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"ovndbcluster-nb-scripts" Dec 13 07:10:15 crc kubenswrapper[4971]: I1213 07:10:15.720490 4971 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovsdbserver-nb-0"] Dec 13 07:10:15 crc kubenswrapper[4971]: I1213 07:10:15.726220 4971 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-log\" (UniqueName: \"kubernetes.io/host-path/1920db8d-2d52-46fa-a4cc-ad787e9886fc-var-log\") pod \"ovn-controller-ovs-jwh28\" (UID: \"1920db8d-2d52-46fa-a4cc-ad787e9886fc\") " pod="openstack/ovn-controller-ovs-jwh28" Dec 13 07:10:15 crc kubenswrapper[4971]: I1213 07:10:15.726286 4971 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-lib\" (UniqueName: \"kubernetes.io/host-path/1920db8d-2d52-46fa-a4cc-ad787e9886fc-var-lib\") pod \"ovn-controller-ovs-jwh28\" (UID: \"1920db8d-2d52-46fa-a4cc-ad787e9886fc\") " pod="openstack/ovn-controller-ovs-jwh28" Dec 13 07:10:15 crc kubenswrapper[4971]: I1213 07:10:15.726317 4971 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/1920db8d-2d52-46fa-a4cc-ad787e9886fc-scripts\") pod \"ovn-controller-ovs-jwh28\" (UID: \"1920db8d-2d52-46fa-a4cc-ad787e9886fc\") " pod="openstack/ovn-controller-ovs-jwh28" Dec 13 07:10:15 crc kubenswrapper[4971]: I1213 07:10:15.726333 4971 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-run\" (UniqueName: \"kubernetes.io/host-path/1920db8d-2d52-46fa-a4cc-ad787e9886fc-var-run\") pod \"ovn-controller-ovs-jwh28\" (UID: \"1920db8d-2d52-46fa-a4cc-ad787e9886fc\") " pod="openstack/ovn-controller-ovs-jwh28" Dec 13 07:10:15 crc kubenswrapper[4971]: I1213 07:10:15.726358 4971 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-ovs\" (UniqueName: \"kubernetes.io/host-path/1920db8d-2d52-46fa-a4cc-ad787e9886fc-etc-ovs\") pod \"ovn-controller-ovs-jwh28\" (UID: \"1920db8d-2d52-46fa-a4cc-ad787e9886fc\") " pod="openstack/ovn-controller-ovs-jwh28" Dec 13 07:10:15 crc kubenswrapper[4971]: I1213 07:10:15.726380 4971 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-8fpjx\" (UniqueName: \"kubernetes.io/projected/1920db8d-2d52-46fa-a4cc-ad787e9886fc-kube-api-access-8fpjx\") pod \"ovn-controller-ovs-jwh28\" (UID: \"1920db8d-2d52-46fa-a4cc-ad787e9886fc\") " pod="openstack/ovn-controller-ovs-jwh28" Dec 13 07:10:15 crc kubenswrapper[4971]: I1213 07:10:15.827555 4971 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/3952f27c-fb18-4e8a-b3a4-a3fa077378c4-config\") pod \"ovsdbserver-nb-0\" (UID: \"3952f27c-fb18-4e8a-b3a4-a3fa077378c4\") " pod="openstack/ovsdbserver-nb-0" Dec 13 07:10:15 crc kubenswrapper[4971]: I1213 07:10:15.827610 4971 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/3952f27c-fb18-4e8a-b3a4-a3fa077378c4-combined-ca-bundle\") pod \"ovsdbserver-nb-0\" (UID: \"3952f27c-fb18-4e8a-b3a4-a3fa077378c4\") " pod="openstack/ovsdbserver-nb-0" Dec 13 07:10:15 crc kubenswrapper[4971]: I1213 07:10:15.827634 4971 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-nb-tls-certs\" (UniqueName: \"kubernetes.io/secret/3952f27c-fb18-4e8a-b3a4-a3fa077378c4-ovsdbserver-nb-tls-certs\") pod \"ovsdbserver-nb-0\" (UID: \"3952f27c-fb18-4e8a-b3a4-a3fa077378c4\") " pod="openstack/ovsdbserver-nb-0" Dec 13 07:10:15 crc kubenswrapper[4971]: I1213 07:10:15.827652 4971 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-2rhp6\" (UniqueName: \"kubernetes.io/projected/3952f27c-fb18-4e8a-b3a4-a3fa077378c4-kube-api-access-2rhp6\") pod \"ovsdbserver-nb-0\" (UID: \"3952f27c-fb18-4e8a-b3a4-a3fa077378c4\") " pod="openstack/ovsdbserver-nb-0" Dec 13 07:10:15 crc kubenswrapper[4971]: I1213 07:10:15.827697 4971 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-log\" (UniqueName: \"kubernetes.io/host-path/1920db8d-2d52-46fa-a4cc-ad787e9886fc-var-log\") pod \"ovn-controller-ovs-jwh28\" (UID: \"1920db8d-2d52-46fa-a4cc-ad787e9886fc\") " pod="openstack/ovn-controller-ovs-jwh28" Dec 13 07:10:15 crc kubenswrapper[4971]: I1213 07:10:15.827720 4971 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"local-storage08-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage08-crc\") pod \"ovsdbserver-nb-0\" (UID: \"3952f27c-fb18-4e8a-b3a4-a3fa077378c4\") " pod="openstack/ovsdbserver-nb-0" Dec 13 07:10:15 crc kubenswrapper[4971]: I1213 07:10:15.827763 4971 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-lib\" (UniqueName: \"kubernetes.io/host-path/1920db8d-2d52-46fa-a4cc-ad787e9886fc-var-lib\") pod \"ovn-controller-ovs-jwh28\" (UID: \"1920db8d-2d52-46fa-a4cc-ad787e9886fc\") " pod="openstack/ovn-controller-ovs-jwh28" Dec 13 07:10:15 crc kubenswrapper[4971]: I1213 07:10:15.827824 4971 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/3952f27c-fb18-4e8a-b3a4-a3fa077378c4-scripts\") pod \"ovsdbserver-nb-0\" (UID: \"3952f27c-fb18-4e8a-b3a4-a3fa077378c4\") " pod="openstack/ovsdbserver-nb-0" Dec 13 07:10:15 crc kubenswrapper[4971]: I1213 07:10:15.827911 4971 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/1920db8d-2d52-46fa-a4cc-ad787e9886fc-scripts\") pod \"ovn-controller-ovs-jwh28\" (UID: \"1920db8d-2d52-46fa-a4cc-ad787e9886fc\") " pod="openstack/ovn-controller-ovs-jwh28" Dec 13 07:10:15 crc kubenswrapper[4971]: I1213 07:10:15.827951 4971 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-run\" (UniqueName: \"kubernetes.io/host-path/1920db8d-2d52-46fa-a4cc-ad787e9886fc-var-run\") pod \"ovn-controller-ovs-jwh28\" (UID: \"1920db8d-2d52-46fa-a4cc-ad787e9886fc\") " pod="openstack/ovn-controller-ovs-jwh28" Dec 13 07:10:15 crc kubenswrapper[4971]: I1213 07:10:15.828010 4971 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metrics-certs-tls-certs\" (UniqueName: \"kubernetes.io/secret/3952f27c-fb18-4e8a-b3a4-a3fa077378c4-metrics-certs-tls-certs\") pod \"ovsdbserver-nb-0\" (UID: \"3952f27c-fb18-4e8a-b3a4-a3fa077378c4\") " pod="openstack/ovsdbserver-nb-0" Dec 13 07:10:15 crc kubenswrapper[4971]: I1213 07:10:15.828044 4971 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-ovs\" (UniqueName: \"kubernetes.io/host-path/1920db8d-2d52-46fa-a4cc-ad787e9886fc-etc-ovs\") pod \"ovn-controller-ovs-jwh28\" (UID: \"1920db8d-2d52-46fa-a4cc-ad787e9886fc\") " pod="openstack/ovn-controller-ovs-jwh28" Dec 13 07:10:15 crc kubenswrapper[4971]: I1213 07:10:15.828070 4971 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdb-rundir\" (UniqueName: \"kubernetes.io/empty-dir/3952f27c-fb18-4e8a-b3a4-a3fa077378c4-ovsdb-rundir\") pod \"ovsdbserver-nb-0\" (UID: \"3952f27c-fb18-4e8a-b3a4-a3fa077378c4\") " pod="openstack/ovsdbserver-nb-0" Dec 13 07:10:15 crc kubenswrapper[4971]: I1213 07:10:15.828115 4971 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-8fpjx\" (UniqueName: \"kubernetes.io/projected/1920db8d-2d52-46fa-a4cc-ad787e9886fc-kube-api-access-8fpjx\") pod \"ovn-controller-ovs-jwh28\" (UID: \"1920db8d-2d52-46fa-a4cc-ad787e9886fc\") " pod="openstack/ovn-controller-ovs-jwh28" Dec 13 07:10:15 crc kubenswrapper[4971]: I1213 07:10:15.828660 4971 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-lib\" (UniqueName: \"kubernetes.io/host-path/1920db8d-2d52-46fa-a4cc-ad787e9886fc-var-lib\") pod \"ovn-controller-ovs-jwh28\" (UID: \"1920db8d-2d52-46fa-a4cc-ad787e9886fc\") " pod="openstack/ovn-controller-ovs-jwh28" Dec 13 07:10:15 crc kubenswrapper[4971]: I1213 07:10:15.828735 4971 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-log\" (UniqueName: \"kubernetes.io/host-path/1920db8d-2d52-46fa-a4cc-ad787e9886fc-var-log\") pod \"ovn-controller-ovs-jwh28\" (UID: \"1920db8d-2d52-46fa-a4cc-ad787e9886fc\") " pod="openstack/ovn-controller-ovs-jwh28" Dec 13 07:10:15 crc kubenswrapper[4971]: I1213 07:10:15.828950 4971 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-run\" (UniqueName: \"kubernetes.io/host-path/1920db8d-2d52-46fa-a4cc-ad787e9886fc-var-run\") pod \"ovn-controller-ovs-jwh28\" (UID: \"1920db8d-2d52-46fa-a4cc-ad787e9886fc\") " pod="openstack/ovn-controller-ovs-jwh28" Dec 13 07:10:15 crc kubenswrapper[4971]: I1213 07:10:15.829132 4971 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-ovs\" (UniqueName: \"kubernetes.io/host-path/1920db8d-2d52-46fa-a4cc-ad787e9886fc-etc-ovs\") pod \"ovn-controller-ovs-jwh28\" (UID: \"1920db8d-2d52-46fa-a4cc-ad787e9886fc\") " pod="openstack/ovn-controller-ovs-jwh28" Dec 13 07:10:15 crc kubenswrapper[4971]: I1213 07:10:15.833901 4971 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/1920db8d-2d52-46fa-a4cc-ad787e9886fc-scripts\") pod \"ovn-controller-ovs-jwh28\" (UID: \"1920db8d-2d52-46fa-a4cc-ad787e9886fc\") " pod="openstack/ovn-controller-ovs-jwh28" Dec 13 07:10:15 crc kubenswrapper[4971]: I1213 07:10:15.916166 4971 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-8fpjx\" (UniqueName: \"kubernetes.io/projected/1920db8d-2d52-46fa-a4cc-ad787e9886fc-kube-api-access-8fpjx\") pod \"ovn-controller-ovs-jwh28\" (UID: \"1920db8d-2d52-46fa-a4cc-ad787e9886fc\") " pod="openstack/ovn-controller-ovs-jwh28" Dec 13 07:10:15 crc kubenswrapper[4971]: I1213 07:10:15.934009 4971 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-controller-ovs-jwh28" Dec 13 07:10:15 crc kubenswrapper[4971]: I1213 07:10:15.934465 4971 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/3952f27c-fb18-4e8a-b3a4-a3fa077378c4-combined-ca-bundle\") pod \"ovsdbserver-nb-0\" (UID: \"3952f27c-fb18-4e8a-b3a4-a3fa077378c4\") " pod="openstack/ovsdbserver-nb-0" Dec 13 07:10:15 crc kubenswrapper[4971]: I1213 07:10:15.934522 4971 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-nb-tls-certs\" (UniqueName: \"kubernetes.io/secret/3952f27c-fb18-4e8a-b3a4-a3fa077378c4-ovsdbserver-nb-tls-certs\") pod \"ovsdbserver-nb-0\" (UID: \"3952f27c-fb18-4e8a-b3a4-a3fa077378c4\") " pod="openstack/ovsdbserver-nb-0" Dec 13 07:10:15 crc kubenswrapper[4971]: I1213 07:10:15.934575 4971 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-2rhp6\" (UniqueName: \"kubernetes.io/projected/3952f27c-fb18-4e8a-b3a4-a3fa077378c4-kube-api-access-2rhp6\") pod \"ovsdbserver-nb-0\" (UID: \"3952f27c-fb18-4e8a-b3a4-a3fa077378c4\") " pod="openstack/ovsdbserver-nb-0" Dec 13 07:10:15 crc kubenswrapper[4971]: I1213 07:10:15.934648 4971 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"local-storage08-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage08-crc\") pod \"ovsdbserver-nb-0\" (UID: \"3952f27c-fb18-4e8a-b3a4-a3fa077378c4\") " pod="openstack/ovsdbserver-nb-0" Dec 13 07:10:15 crc kubenswrapper[4971]: I1213 07:10:15.934698 4971 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/3952f27c-fb18-4e8a-b3a4-a3fa077378c4-scripts\") pod \"ovsdbserver-nb-0\" (UID: \"3952f27c-fb18-4e8a-b3a4-a3fa077378c4\") " pod="openstack/ovsdbserver-nb-0" Dec 13 07:10:15 crc kubenswrapper[4971]: I1213 07:10:15.934748 4971 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs-tls-certs\" (UniqueName: \"kubernetes.io/secret/3952f27c-fb18-4e8a-b3a4-a3fa077378c4-metrics-certs-tls-certs\") pod \"ovsdbserver-nb-0\" (UID: \"3952f27c-fb18-4e8a-b3a4-a3fa077378c4\") " pod="openstack/ovsdbserver-nb-0" Dec 13 07:10:15 crc kubenswrapper[4971]: I1213 07:10:15.934778 4971 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdb-rundir\" (UniqueName: \"kubernetes.io/empty-dir/3952f27c-fb18-4e8a-b3a4-a3fa077378c4-ovsdb-rundir\") pod \"ovsdbserver-nb-0\" (UID: \"3952f27c-fb18-4e8a-b3a4-a3fa077378c4\") " pod="openstack/ovsdbserver-nb-0" Dec 13 07:10:15 crc kubenswrapper[4971]: I1213 07:10:15.934819 4971 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/3952f27c-fb18-4e8a-b3a4-a3fa077378c4-config\") pod \"ovsdbserver-nb-0\" (UID: \"3952f27c-fb18-4e8a-b3a4-a3fa077378c4\") " pod="openstack/ovsdbserver-nb-0" Dec 13 07:10:15 crc kubenswrapper[4971]: I1213 07:10:15.935318 4971 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"local-storage08-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage08-crc\") pod \"ovsdbserver-nb-0\" (UID: \"3952f27c-fb18-4e8a-b3a4-a3fa077378c4\") device mount path \"/mnt/openstack/pv08\"" pod="openstack/ovsdbserver-nb-0" Dec 13 07:10:15 crc kubenswrapper[4971]: I1213 07:10:15.935837 4971 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/3952f27c-fb18-4e8a-b3a4-a3fa077378c4-config\") pod \"ovsdbserver-nb-0\" (UID: \"3952f27c-fb18-4e8a-b3a4-a3fa077378c4\") " pod="openstack/ovsdbserver-nb-0" Dec 13 07:10:15 crc kubenswrapper[4971]: I1213 07:10:15.937118 4971 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/3952f27c-fb18-4e8a-b3a4-a3fa077378c4-scripts\") pod \"ovsdbserver-nb-0\" (UID: \"3952f27c-fb18-4e8a-b3a4-a3fa077378c4\") " pod="openstack/ovsdbserver-nb-0" Dec 13 07:10:15 crc kubenswrapper[4971]: I1213 07:10:15.939319 4971 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdb-rundir\" (UniqueName: \"kubernetes.io/empty-dir/3952f27c-fb18-4e8a-b3a4-a3fa077378c4-ovsdb-rundir\") pod \"ovsdbserver-nb-0\" (UID: \"3952f27c-fb18-4e8a-b3a4-a3fa077378c4\") " pod="openstack/ovsdbserver-nb-0" Dec 13 07:10:15 crc kubenswrapper[4971]: I1213 07:10:15.947631 4971 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-nb-tls-certs\" (UniqueName: \"kubernetes.io/secret/3952f27c-fb18-4e8a-b3a4-a3fa077378c4-ovsdbserver-nb-tls-certs\") pod \"ovsdbserver-nb-0\" (UID: \"3952f27c-fb18-4e8a-b3a4-a3fa077378c4\") " pod="openstack/ovsdbserver-nb-0" Dec 13 07:10:15 crc kubenswrapper[4971]: I1213 07:10:15.954385 4971 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/3952f27c-fb18-4e8a-b3a4-a3fa077378c4-combined-ca-bundle\") pod \"ovsdbserver-nb-0\" (UID: \"3952f27c-fb18-4e8a-b3a4-a3fa077378c4\") " pod="openstack/ovsdbserver-nb-0" Dec 13 07:10:15 crc kubenswrapper[4971]: I1213 07:10:15.959603 4971 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-2rhp6\" (UniqueName: \"kubernetes.io/projected/3952f27c-fb18-4e8a-b3a4-a3fa077378c4-kube-api-access-2rhp6\") pod \"ovsdbserver-nb-0\" (UID: \"3952f27c-fb18-4e8a-b3a4-a3fa077378c4\") " pod="openstack/ovsdbserver-nb-0" Dec 13 07:10:15 crc kubenswrapper[4971]: I1213 07:10:15.959902 4971 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metrics-certs-tls-certs\" (UniqueName: \"kubernetes.io/secret/3952f27c-fb18-4e8a-b3a4-a3fa077378c4-metrics-certs-tls-certs\") pod \"ovsdbserver-nb-0\" (UID: \"3952f27c-fb18-4e8a-b3a4-a3fa077378c4\") " pod="openstack/ovsdbserver-nb-0" Dec 13 07:10:15 crc kubenswrapper[4971]: I1213 07:10:15.971290 4971 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"local-storage08-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage08-crc\") pod \"ovsdbserver-nb-0\" (UID: \"3952f27c-fb18-4e8a-b3a4-a3fa077378c4\") " pod="openstack/ovsdbserver-nb-0" Dec 13 07:10:16 crc kubenswrapper[4971]: I1213 07:10:16.098191 4971 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovsdbserver-nb-0" Dec 13 07:10:18 crc kubenswrapper[4971]: I1213 07:10:18.799527 4971 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ovsdbserver-sb-0"] Dec 13 07:10:18 crc kubenswrapper[4971]: I1213 07:10:18.802482 4971 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovsdbserver-sb-0" Dec 13 07:10:18 crc kubenswrapper[4971]: I1213 07:10:18.810390 4971 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"ovndbcluster-sb-config" Dec 13 07:10:18 crc kubenswrapper[4971]: I1213 07:10:18.810718 4971 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ovncluster-ovndbcluster-sb-dockercfg-xrk22" Dec 13 07:10:18 crc kubenswrapper[4971]: I1213 07:10:18.810931 4971 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-ovndbcluster-sb-ovndbs" Dec 13 07:10:18 crc kubenswrapper[4971]: I1213 07:10:18.811121 4971 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"ovndbcluster-sb-scripts" Dec 13 07:10:18 crc kubenswrapper[4971]: I1213 07:10:18.884771 4971 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovsdbserver-sb-0"] Dec 13 07:10:18 crc kubenswrapper[4971]: I1213 07:10:18.952084 4971 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/a85733fd-9af5-4fc4-8096-9b368a793fe7-combined-ca-bundle\") pod \"ovsdbserver-sb-0\" (UID: \"a85733fd-9af5-4fc4-8096-9b368a793fe7\") " pod="openstack/ovsdbserver-sb-0" Dec 13 07:10:18 crc kubenswrapper[4971]: I1213 07:10:18.952158 4971 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/a85733fd-9af5-4fc4-8096-9b368a793fe7-config\") pod \"ovsdbserver-sb-0\" (UID: \"a85733fd-9af5-4fc4-8096-9b368a793fe7\") " pod="openstack/ovsdbserver-sb-0" Dec 13 07:10:18 crc kubenswrapper[4971]: I1213 07:10:18.952236 4971 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-sb-tls-certs\" (UniqueName: \"kubernetes.io/secret/a85733fd-9af5-4fc4-8096-9b368a793fe7-ovsdbserver-sb-tls-certs\") pod \"ovsdbserver-sb-0\" (UID: \"a85733fd-9af5-4fc4-8096-9b368a793fe7\") " pod="openstack/ovsdbserver-sb-0" Dec 13 07:10:18 crc kubenswrapper[4971]: I1213 07:10:18.952307 4971 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdb-rundir\" (UniqueName: \"kubernetes.io/empty-dir/a85733fd-9af5-4fc4-8096-9b368a793fe7-ovsdb-rundir\") pod \"ovsdbserver-sb-0\" (UID: \"a85733fd-9af5-4fc4-8096-9b368a793fe7\") " pod="openstack/ovsdbserver-sb-0" Dec 13 07:10:18 crc kubenswrapper[4971]: I1213 07:10:18.952332 4971 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-dz7hc\" (UniqueName: \"kubernetes.io/projected/a85733fd-9af5-4fc4-8096-9b368a793fe7-kube-api-access-dz7hc\") pod \"ovsdbserver-sb-0\" (UID: \"a85733fd-9af5-4fc4-8096-9b368a793fe7\") " pod="openstack/ovsdbserver-sb-0" Dec 13 07:10:18 crc kubenswrapper[4971]: I1213 07:10:18.952355 4971 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metrics-certs-tls-certs\" (UniqueName: \"kubernetes.io/secret/a85733fd-9af5-4fc4-8096-9b368a793fe7-metrics-certs-tls-certs\") pod \"ovsdbserver-sb-0\" (UID: \"a85733fd-9af5-4fc4-8096-9b368a793fe7\") " pod="openstack/ovsdbserver-sb-0" Dec 13 07:10:18 crc kubenswrapper[4971]: I1213 07:10:18.952413 4971 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"local-storage09-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage09-crc\") pod \"ovsdbserver-sb-0\" (UID: \"a85733fd-9af5-4fc4-8096-9b368a793fe7\") " pod="openstack/ovsdbserver-sb-0" Dec 13 07:10:18 crc kubenswrapper[4971]: I1213 07:10:18.953145 4971 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/a85733fd-9af5-4fc4-8096-9b368a793fe7-scripts\") pod \"ovsdbserver-sb-0\" (UID: \"a85733fd-9af5-4fc4-8096-9b368a793fe7\") " pod="openstack/ovsdbserver-sb-0" Dec 13 07:10:19 crc kubenswrapper[4971]: I1213 07:10:19.054823 4971 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/a85733fd-9af5-4fc4-8096-9b368a793fe7-scripts\") pod \"ovsdbserver-sb-0\" (UID: \"a85733fd-9af5-4fc4-8096-9b368a793fe7\") " pod="openstack/ovsdbserver-sb-0" Dec 13 07:10:19 crc kubenswrapper[4971]: I1213 07:10:19.054910 4971 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/a85733fd-9af5-4fc4-8096-9b368a793fe7-combined-ca-bundle\") pod \"ovsdbserver-sb-0\" (UID: \"a85733fd-9af5-4fc4-8096-9b368a793fe7\") " pod="openstack/ovsdbserver-sb-0" Dec 13 07:10:19 crc kubenswrapper[4971]: I1213 07:10:19.054951 4971 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/a85733fd-9af5-4fc4-8096-9b368a793fe7-config\") pod \"ovsdbserver-sb-0\" (UID: \"a85733fd-9af5-4fc4-8096-9b368a793fe7\") " pod="openstack/ovsdbserver-sb-0" Dec 13 07:10:19 crc kubenswrapper[4971]: I1213 07:10:19.054996 4971 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-sb-tls-certs\" (UniqueName: \"kubernetes.io/secret/a85733fd-9af5-4fc4-8096-9b368a793fe7-ovsdbserver-sb-tls-certs\") pod \"ovsdbserver-sb-0\" (UID: \"a85733fd-9af5-4fc4-8096-9b368a793fe7\") " pod="openstack/ovsdbserver-sb-0" Dec 13 07:10:19 crc kubenswrapper[4971]: I1213 07:10:19.055037 4971 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdb-rundir\" (UniqueName: \"kubernetes.io/empty-dir/a85733fd-9af5-4fc4-8096-9b368a793fe7-ovsdb-rundir\") pod \"ovsdbserver-sb-0\" (UID: \"a85733fd-9af5-4fc4-8096-9b368a793fe7\") " pod="openstack/ovsdbserver-sb-0" Dec 13 07:10:19 crc kubenswrapper[4971]: I1213 07:10:19.055053 4971 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-dz7hc\" (UniqueName: \"kubernetes.io/projected/a85733fd-9af5-4fc4-8096-9b368a793fe7-kube-api-access-dz7hc\") pod \"ovsdbserver-sb-0\" (UID: \"a85733fd-9af5-4fc4-8096-9b368a793fe7\") " pod="openstack/ovsdbserver-sb-0" Dec 13 07:10:19 crc kubenswrapper[4971]: I1213 07:10:19.055072 4971 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs-tls-certs\" (UniqueName: \"kubernetes.io/secret/a85733fd-9af5-4fc4-8096-9b368a793fe7-metrics-certs-tls-certs\") pod \"ovsdbserver-sb-0\" (UID: \"a85733fd-9af5-4fc4-8096-9b368a793fe7\") " pod="openstack/ovsdbserver-sb-0" Dec 13 07:10:19 crc kubenswrapper[4971]: I1213 07:10:19.055137 4971 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"local-storage09-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage09-crc\") pod \"ovsdbserver-sb-0\" (UID: \"a85733fd-9af5-4fc4-8096-9b368a793fe7\") " pod="openstack/ovsdbserver-sb-0" Dec 13 07:10:19 crc kubenswrapper[4971]: I1213 07:10:19.055439 4971 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"local-storage09-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage09-crc\") pod \"ovsdbserver-sb-0\" (UID: \"a85733fd-9af5-4fc4-8096-9b368a793fe7\") device mount path \"/mnt/openstack/pv09\"" pod="openstack/ovsdbserver-sb-0" Dec 13 07:10:19 crc kubenswrapper[4971]: I1213 07:10:19.058383 4971 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdb-rundir\" (UniqueName: \"kubernetes.io/empty-dir/a85733fd-9af5-4fc4-8096-9b368a793fe7-ovsdb-rundir\") pod \"ovsdbserver-sb-0\" (UID: \"a85733fd-9af5-4fc4-8096-9b368a793fe7\") " pod="openstack/ovsdbserver-sb-0" Dec 13 07:10:19 crc kubenswrapper[4971]: I1213 07:10:19.116730 4971 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/a85733fd-9af5-4fc4-8096-9b368a793fe7-scripts\") pod \"ovsdbserver-sb-0\" (UID: \"a85733fd-9af5-4fc4-8096-9b368a793fe7\") " pod="openstack/ovsdbserver-sb-0" Dec 13 07:10:19 crc kubenswrapper[4971]: I1213 07:10:19.117683 4971 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-sb-tls-certs\" (UniqueName: \"kubernetes.io/secret/a85733fd-9af5-4fc4-8096-9b368a793fe7-ovsdbserver-sb-tls-certs\") pod \"ovsdbserver-sb-0\" (UID: \"a85733fd-9af5-4fc4-8096-9b368a793fe7\") " pod="openstack/ovsdbserver-sb-0" Dec 13 07:10:19 crc kubenswrapper[4971]: I1213 07:10:19.118032 4971 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/a85733fd-9af5-4fc4-8096-9b368a793fe7-config\") pod \"ovsdbserver-sb-0\" (UID: \"a85733fd-9af5-4fc4-8096-9b368a793fe7\") " pod="openstack/ovsdbserver-sb-0" Dec 13 07:10:19 crc kubenswrapper[4971]: I1213 07:10:19.130213 4971 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metrics-certs-tls-certs\" (UniqueName: \"kubernetes.io/secret/a85733fd-9af5-4fc4-8096-9b368a793fe7-metrics-certs-tls-certs\") pod \"ovsdbserver-sb-0\" (UID: \"a85733fd-9af5-4fc4-8096-9b368a793fe7\") " pod="openstack/ovsdbserver-sb-0" Dec 13 07:10:19 crc kubenswrapper[4971]: I1213 07:10:19.132318 4971 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/a85733fd-9af5-4fc4-8096-9b368a793fe7-combined-ca-bundle\") pod \"ovsdbserver-sb-0\" (UID: \"a85733fd-9af5-4fc4-8096-9b368a793fe7\") " pod="openstack/ovsdbserver-sb-0" Dec 13 07:10:19 crc kubenswrapper[4971]: I1213 07:10:19.132319 4971 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"local-storage09-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage09-crc\") pod \"ovsdbserver-sb-0\" (UID: \"a85733fd-9af5-4fc4-8096-9b368a793fe7\") " pod="openstack/ovsdbserver-sb-0" Dec 13 07:10:19 crc kubenswrapper[4971]: I1213 07:10:19.132366 4971 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-dz7hc\" (UniqueName: \"kubernetes.io/projected/a85733fd-9af5-4fc4-8096-9b368a793fe7-kube-api-access-dz7hc\") pod \"ovsdbserver-sb-0\" (UID: \"a85733fd-9af5-4fc4-8096-9b368a793fe7\") " pod="openstack/ovsdbserver-sb-0" Dec 13 07:10:19 crc kubenswrapper[4971]: I1213 07:10:19.148043 4971 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovsdbserver-sb-0" Dec 13 07:10:26 crc kubenswrapper[4971]: I1213 07:10:26.092211 4971 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/kube-state-metrics-0" event={"ID":"2bab8c0a-bd76-4265-8355-8df0c6fce6c3","Type":"ContainerStarted","Data":"76129f35a3b0ed0f3d63fd9275ae2264a1ef0588120ed9ab9858a169a60b5f88"} Dec 13 07:10:33 crc kubenswrapper[4971]: E1213 07:10:33.983361 4971 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying config: context canceled" image="quay.io/podified-antelope-centos9/openstack-rabbitmq@sha256:e733252aab7f4bc0efbdd712bcd88e44c5498bf1773dba843bc9dcfac324fe3d" Dec 13 07:10:33 crc kubenswrapper[4971]: E1213 07:10:33.984169 4971 kuberuntime_manager.go:1274] "Unhandled Error" err="init container &Container{Name:setup-container,Image:quay.io/podified-antelope-centos9/openstack-rabbitmq@sha256:e733252aab7f4bc0efbdd712bcd88e44c5498bf1773dba843bc9dcfac324fe3d,Command:[sh -c cp /tmp/erlang-cookie-secret/.erlang.cookie /var/lib/rabbitmq/.erlang.cookie && chmod 600 /var/lib/rabbitmq/.erlang.cookie ; cp /tmp/rabbitmq-plugins/enabled_plugins /operator/enabled_plugins ; echo '[default]' > /var/lib/rabbitmq/.rabbitmqadmin.conf && sed -e 's/default_user/username/' -e 's/default_pass/password/' /tmp/default_user.conf >> /var/lib/rabbitmq/.rabbitmqadmin.conf && chmod 600 /var/lib/rabbitmq/.rabbitmqadmin.conf ; sleep 30],Args:[],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{},Resources:ResourceRequirements{Limits:ResourceList{cpu: {{20 -3} {} 20m DecimalSI},memory: {{67108864 0} {} BinarySI},},Requests:ResourceList{cpu: {{20 -3} {} 20m DecimalSI},memory: {{67108864 0} {} BinarySI},},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:plugins-conf,ReadOnly:false,MountPath:/tmp/rabbitmq-plugins/,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:rabbitmq-erlang-cookie,ReadOnly:false,MountPath:/var/lib/rabbitmq/,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:erlang-cookie-secret,ReadOnly:false,MountPath:/tmp/erlang-cookie-secret/,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:rabbitmq-plugins,ReadOnly:false,MountPath:/operator,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:persistence,ReadOnly:false,MountPath:/var/lib/rabbitmq/mnesia/,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:rabbitmq-confd,ReadOnly:false,MountPath:/tmp/default_user.conf,SubPath:default_user.conf,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-wfzdp,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*1000650000,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod rabbitmq-server-0_openstack(a590eb84-c8d5-4cbe-9c03-03a6ccb82073): ErrImagePull: rpc error: code = Canceled desc = copying config: context canceled" logger="UnhandledError" Dec 13 07:10:33 crc kubenswrapper[4971]: E1213 07:10:33.985688 4971 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"setup-container\" with ErrImagePull: \"rpc error: code = Canceled desc = copying config: context canceled\"" pod="openstack/rabbitmq-server-0" podUID="a590eb84-c8d5-4cbe-9c03-03a6ccb82073" Dec 13 07:10:34 crc kubenswrapper[4971]: E1213 07:10:34.006628 4971 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying config: context canceled" image="quay.io/podified-antelope-centos9/openstack-rabbitmq@sha256:e733252aab7f4bc0efbdd712bcd88e44c5498bf1773dba843bc9dcfac324fe3d" Dec 13 07:10:34 crc kubenswrapper[4971]: E1213 07:10:34.007198 4971 kuberuntime_manager.go:1274] "Unhandled Error" err="init container &Container{Name:setup-container,Image:quay.io/podified-antelope-centos9/openstack-rabbitmq@sha256:e733252aab7f4bc0efbdd712bcd88e44c5498bf1773dba843bc9dcfac324fe3d,Command:[sh -c cp /tmp/erlang-cookie-secret/.erlang.cookie /var/lib/rabbitmq/.erlang.cookie && chmod 600 /var/lib/rabbitmq/.erlang.cookie ; cp /tmp/rabbitmq-plugins/enabled_plugins /operator/enabled_plugins ; echo '[default]' > /var/lib/rabbitmq/.rabbitmqadmin.conf && sed -e 's/default_user/username/' -e 's/default_pass/password/' /tmp/default_user.conf >> /var/lib/rabbitmq/.rabbitmqadmin.conf && chmod 600 /var/lib/rabbitmq/.rabbitmqadmin.conf ; sleep 30],Args:[],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{},Resources:ResourceRequirements{Limits:ResourceList{cpu: {{20 -3} {} 20m DecimalSI},memory: {{67108864 0} {} BinarySI},},Requests:ResourceList{cpu: {{20 -3} {} 20m DecimalSI},memory: {{67108864 0} {} BinarySI},},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:plugins-conf,ReadOnly:false,MountPath:/tmp/rabbitmq-plugins/,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:rabbitmq-erlang-cookie,ReadOnly:false,MountPath:/var/lib/rabbitmq/,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:erlang-cookie-secret,ReadOnly:false,MountPath:/tmp/erlang-cookie-secret/,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:rabbitmq-plugins,ReadOnly:false,MountPath:/operator,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:persistence,ReadOnly:false,MountPath:/var/lib/rabbitmq/mnesia/,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:rabbitmq-confd,ReadOnly:false,MountPath:/tmp/default_user.conf,SubPath:default_user.conf,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-jpcrh,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*1000650000,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod rabbitmq-cell1-server-0_openstack(5eed809d-c915-433c-9915-27e0b01e1ffe): ErrImagePull: rpc error: code = Canceled desc = copying config: context canceled" logger="UnhandledError" Dec 13 07:10:34 crc kubenswrapper[4971]: E1213 07:10:34.008420 4971 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"setup-container\" with ErrImagePull: \"rpc error: code = Canceled desc = copying config: context canceled\"" pod="openstack/rabbitmq-cell1-server-0" podUID="5eed809d-c915-433c-9915-27e0b01e1ffe" Dec 13 07:10:34 crc kubenswrapper[4971]: E1213 07:10:34.148019 4971 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"setup-container\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/podified-antelope-centos9/openstack-rabbitmq@sha256:e733252aab7f4bc0efbdd712bcd88e44c5498bf1773dba843bc9dcfac324fe3d\\\"\"" pod="openstack/rabbitmq-cell1-server-0" podUID="5eed809d-c915-433c-9915-27e0b01e1ffe" Dec 13 07:10:34 crc kubenswrapper[4971]: E1213 07:10:34.148200 4971 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"setup-container\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/podified-antelope-centos9/openstack-rabbitmq@sha256:e733252aab7f4bc0efbdd712bcd88e44c5498bf1773dba843bc9dcfac324fe3d\\\"\"" pod="openstack/rabbitmq-server-0" podUID="a590eb84-c8d5-4cbe-9c03-03a6ccb82073" Dec 13 07:10:34 crc kubenswrapper[4971]: E1213 07:10:34.685258 4971 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying config: context canceled" image="quay.io/podified-antelope-centos9/openstack-memcached@sha256:e47191ba776414b781b3e27b856ab45a03b9480c7dc2b1addb939608794882dc" Dec 13 07:10:34 crc kubenswrapper[4971]: E1213 07:10:34.685737 4971 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:memcached,Image:quay.io/podified-antelope-centos9/openstack-memcached@sha256:e47191ba776414b781b3e27b856ab45a03b9480c7dc2b1addb939608794882dc,Command:[/usr/bin/dumb-init -- /usr/local/bin/kolla_start],Args:[],WorkingDir:,Ports:[]ContainerPort{ContainerPort{Name:memcached,HostPort:0,ContainerPort:11211,Protocol:TCP,HostIP:,},ContainerPort{Name:memcached-tls,HostPort:0,ContainerPort:11212,Protocol:TCP,HostIP:,},},Env:[]EnvVar{EnvVar{Name:KOLLA_CONFIG_STRATEGY,Value:COPY_ALWAYS,ValueFrom:nil,},EnvVar{Name:POD_IPS,Value:,ValueFrom:&EnvVarSource{FieldRef:&ObjectFieldSelector{APIVersion:v1,FieldPath:status.podIPs,},ResourceFieldRef:nil,ConfigMapKeyRef:nil,SecretKeyRef:nil,},},EnvVar{Name:CONFIG_HASH,Value:n64hb6h78hd5h5cfhc6h98h76hffh54h4h6bh64ch658h585hd5h5f6h5d4hd4h577h596h58dh6chdh688h68dhdh56fh594h648h665h697q,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:config-data,ReadOnly:true,MountPath:/var/lib/kolla/config_files/src,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kolla-config,ReadOnly:true,MountPath:/var/lib/kolla/config_files,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:memcached-tls-certs,ReadOnly:true,MountPath:/var/lib/config-data/tls/certs/memcached.crt,SubPath:tls.crt,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:memcached-tls-certs,ReadOnly:true,MountPath:/var/lib/config-data/tls/private/memcached.key,SubPath:tls.key,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:combined-ca-bundle,ReadOnly:true,MountPath:/etc/pki/ca-trust/extracted/pem/tls-ca-bundle.pem,SubPath:tls-ca-bundle.pem,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-dmspm,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:nil,TCPSocket:&TCPSocketAction{Port:{0 11211 },Host:,},GRPC:nil,},InitialDelaySeconds:3,TimeoutSeconds:5,PeriodSeconds:3,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},ReadinessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:nil,TCPSocket:&TCPSocketAction{Port:{0 11211 },Host:,},GRPC:nil,},InitialDelaySeconds:5,TimeoutSeconds:5,PeriodSeconds:5,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[MKNOD],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*42457,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:nil,RunAsGroup:*42457,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod memcached-0_openstack(e6adf8ca-c137-4b7d-bf8c-74c0766e57f0): ErrImagePull: rpc error: code = Canceled desc = copying config: context canceled" logger="UnhandledError" Dec 13 07:10:34 crc kubenswrapper[4971]: E1213 07:10:34.687430 4971 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"memcached\" with ErrImagePull: \"rpc error: code = Canceled desc = copying config: context canceled\"" pod="openstack/memcached-0" podUID="e6adf8ca-c137-4b7d-bf8c-74c0766e57f0" Dec 13 07:10:35 crc kubenswrapper[4971]: E1213 07:10:35.154656 4971 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"memcached\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/podified-antelope-centos9/openstack-memcached@sha256:e47191ba776414b781b3e27b856ab45a03b9480c7dc2b1addb939608794882dc\\\"\"" pod="openstack/memcached-0" podUID="e6adf8ca-c137-4b7d-bf8c-74c0766e57f0" Dec 13 07:10:39 crc kubenswrapper[4971]: I1213 07:10:39.542590 4971 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovn-controller-k5fz9"] Dec 13 07:10:39 crc kubenswrapper[4971]: E1213 07:10:39.924095 4971 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying config: context canceled" image="quay.io/podified-antelope-centos9/openstack-neutron-server@sha256:ea0bf67f1aa5d95a9a07b9c8692c293470f1311792c55d3d57f1f92e56689c33" Dec 13 07:10:39 crc kubenswrapper[4971]: E1213 07:10:39.925620 4971 kuberuntime_manager.go:1274] "Unhandled Error" err="init container &Container{Name:init,Image:quay.io/podified-antelope-centos9/openstack-neutron-server@sha256:ea0bf67f1aa5d95a9a07b9c8692c293470f1311792c55d3d57f1f92e56689c33,Command:[/bin/bash],Args:[-c dnsmasq --interface=* --conf-dir=/etc/dnsmasq.d --hostsdir=/etc/dnsmasq.d/hosts --keep-in-foreground --log-debug --bind-interfaces --listen-address=$(POD_IP) --port 5353 --log-facility=- --no-hosts --domain-needed --no-resolv --bogus-priv --log-queries --test],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:CONFIG_HASH,Value:ndfhb5h667h568h584h5f9h58dh565h664h587h597h577h64bh5c4h66fh647hbdh68ch5c5h68dh686h5f7h64hd7hc6h55fh57bh98h57fh87h5fh57fq,ValueFrom:nil,},EnvVar{Name:POD_IP,Value:,ValueFrom:&EnvVarSource{FieldRef:&ObjectFieldSelector{APIVersion:v1,FieldPath:status.podIP,},ResourceFieldRef:nil,ConfigMapKeyRef:nil,SecretKeyRef:nil,},},},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:config,ReadOnly:true,MountPath:/etc/dnsmasq.d/config.cfg,SubPath:dns,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:dns-svc,ReadOnly:true,MountPath:/etc/dnsmasq.d/hosts/dns-svc,SubPath:dns-svc,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-rzb6k,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*1000650000,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:&SeccompProfile{Type:RuntimeDefault,LocalhostProfile:nil,},AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod dnsmasq-dns-5f854695bc-56d9f_openstack(038690e5-17a1-4e19-8d3b-f420dbdf7349): ErrImagePull: rpc error: code = Canceled desc = copying config: context canceled" logger="UnhandledError" Dec 13 07:10:39 crc kubenswrapper[4971]: E1213 07:10:39.927326 4971 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"init\" with ErrImagePull: \"rpc error: code = Canceled desc = copying config: context canceled\"" pod="openstack/dnsmasq-dns-5f854695bc-56d9f" podUID="038690e5-17a1-4e19-8d3b-f420dbdf7349" Dec 13 07:10:39 crc kubenswrapper[4971]: E1213 07:10:39.997904 4971 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying config: context canceled" image="quay.io/podified-antelope-centos9/openstack-neutron-server@sha256:ea0bf67f1aa5d95a9a07b9c8692c293470f1311792c55d3d57f1f92e56689c33" Dec 13 07:10:39 crc kubenswrapper[4971]: E1213 07:10:39.998041 4971 kuberuntime_manager.go:1274] "Unhandled Error" err="init container &Container{Name:init,Image:quay.io/podified-antelope-centos9/openstack-neutron-server@sha256:ea0bf67f1aa5d95a9a07b9c8692c293470f1311792c55d3d57f1f92e56689c33,Command:[/bin/bash],Args:[-c dnsmasq --interface=* --conf-dir=/etc/dnsmasq.d --hostsdir=/etc/dnsmasq.d/hosts --keep-in-foreground --log-debug --bind-interfaces --listen-address=$(POD_IP) --port 5353 --log-facility=- --no-hosts --domain-needed --no-resolv --bogus-priv --log-queries --test],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:CONFIG_HASH,Value:nffh5bdhf4h5f8h79h55h77h58fh56dh7bh6fh578hbch55dh68h56bhd9h65dh57ch658hc9h566h666h688h58h65dh684h5d7h6ch575h5d6h88q,ValueFrom:nil,},EnvVar{Name:POD_IP,Value:,ValueFrom:&EnvVarSource{FieldRef:&ObjectFieldSelector{APIVersion:v1,FieldPath:status.podIP,},ResourceFieldRef:nil,ConfigMapKeyRef:nil,SecretKeyRef:nil,},},},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:config,ReadOnly:true,MountPath:/etc/dnsmasq.d/config.cfg,SubPath:dns,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-fv74m,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*1000650000,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:&SeccompProfile{Type:RuntimeDefault,LocalhostProfile:nil,},AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod dnsmasq-dns-84bb9d8bd9-ljf6c_openstack(ee072669-97ce-4d78-a12f-4169120a00f6): ErrImagePull: rpc error: code = Canceled desc = copying config: context canceled" logger="UnhandledError" Dec 13 07:10:39 crc kubenswrapper[4971]: E1213 07:10:39.999150 4971 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"init\" with ErrImagePull: \"rpc error: code = Canceled desc = copying config: context canceled\"" pod="openstack/dnsmasq-dns-84bb9d8bd9-ljf6c" podUID="ee072669-97ce-4d78-a12f-4169120a00f6" Dec 13 07:10:40 crc kubenswrapper[4971]: E1213 07:10:40.065183 4971 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying config: context canceled" image="quay.io/podified-antelope-centos9/openstack-neutron-server@sha256:ea0bf67f1aa5d95a9a07b9c8692c293470f1311792c55d3d57f1f92e56689c33" Dec 13 07:10:40 crc kubenswrapper[4971]: E1213 07:10:40.065660 4971 kuberuntime_manager.go:1274] "Unhandled Error" err="init container &Container{Name:init,Image:quay.io/podified-antelope-centos9/openstack-neutron-server@sha256:ea0bf67f1aa5d95a9a07b9c8692c293470f1311792c55d3d57f1f92e56689c33,Command:[/bin/bash],Args:[-c dnsmasq --interface=* --conf-dir=/etc/dnsmasq.d --hostsdir=/etc/dnsmasq.d/hosts --keep-in-foreground --log-debug --bind-interfaces --listen-address=$(POD_IP) --port 5353 --log-facility=- --no-hosts --domain-needed --no-resolv --bogus-priv --log-queries --test],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:CONFIG_HASH,Value:n659h4h664hbh658h587h67ch89h587h8fh679hc6hf9h55fh644h5d5h698h68dh5cdh5ffh669h54ch9h689hb8hd4h5bfhd8h5d7h5fh665h574q,ValueFrom:nil,},EnvVar{Name:POD_IP,Value:,ValueFrom:&EnvVarSource{FieldRef:&ObjectFieldSelector{APIVersion:v1,FieldPath:status.podIP,},ResourceFieldRef:nil,ConfigMapKeyRef:nil,SecretKeyRef:nil,},},},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:config,ReadOnly:true,MountPath:/etc/dnsmasq.d/config.cfg,SubPath:dns,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:dns-svc,ReadOnly:true,MountPath:/etc/dnsmasq.d/hosts/dns-svc,SubPath:dns-svc,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-pszmq,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*1000650000,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:&SeccompProfile{Type:RuntimeDefault,LocalhostProfile:nil,},AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod dnsmasq-dns-95f5f6995-qv9tw_openstack(9a37323b-7693-41c2-8a8f-661291d1ef5e): ErrImagePull: rpc error: code = Canceled desc = copying config: context canceled" logger="UnhandledError" Dec 13 07:10:40 crc kubenswrapper[4971]: E1213 07:10:40.067657 4971 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"init\" with ErrImagePull: \"rpc error: code = Canceled desc = copying config: context canceled\"" pod="openstack/dnsmasq-dns-95f5f6995-qv9tw" podUID="9a37323b-7693-41c2-8a8f-661291d1ef5e" Dec 13 07:10:40 crc kubenswrapper[4971]: E1213 07:10:40.179607 4971 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying config: context canceled" image="quay.io/podified-antelope-centos9/openstack-neutron-server@sha256:ea0bf67f1aa5d95a9a07b9c8692c293470f1311792c55d3d57f1f92e56689c33" Dec 13 07:10:40 crc kubenswrapper[4971]: E1213 07:10:40.179790 4971 kuberuntime_manager.go:1274] "Unhandled Error" err="init container &Container{Name:init,Image:quay.io/podified-antelope-centos9/openstack-neutron-server@sha256:ea0bf67f1aa5d95a9a07b9c8692c293470f1311792c55d3d57f1f92e56689c33,Command:[/bin/bash],Args:[-c dnsmasq --interface=* --conf-dir=/etc/dnsmasq.d --hostsdir=/etc/dnsmasq.d/hosts --keep-in-foreground --log-debug --bind-interfaces --listen-address=$(POD_IP) --port 5353 --log-facility=- --no-hosts --domain-needed --no-resolv --bogus-priv --log-queries --test],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:CONFIG_HASH,Value:n68chd6h679hbfh55fhc6h5ffh5d8h94h56ch589hb4hc5h57bh677hcdh655h8dh667h675h654h66ch567h8fh659h5b4h675h566h55bh54h67dh6dq,ValueFrom:nil,},EnvVar{Name:POD_IP,Value:,ValueFrom:&EnvVarSource{FieldRef:&ObjectFieldSelector{APIVersion:v1,FieldPath:status.podIP,},ResourceFieldRef:nil,ConfigMapKeyRef:nil,SecretKeyRef:nil,},},},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:config,ReadOnly:true,MountPath:/etc/dnsmasq.d/config.cfg,SubPath:dns,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:dns-svc,ReadOnly:true,MountPath:/etc/dnsmasq.d/hosts/dns-svc,SubPath:dns-svc,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-5zfr7,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*1000650000,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:&SeccompProfile{Type:RuntimeDefault,LocalhostProfile:nil,},AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod dnsmasq-dns-744ffd65bc-zhplm_openstack(b4e999eb-ff02-4c70-b755-f85bd90dd637): ErrImagePull: rpc error: code = Canceled desc = copying config: context canceled" logger="UnhandledError" Dec 13 07:10:40 crc kubenswrapper[4971]: E1213 07:10:40.182416 4971 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"init\" with ErrImagePull: \"rpc error: code = Canceled desc = copying config: context canceled\"" pod="openstack/dnsmasq-dns-744ffd65bc-zhplm" podUID="b4e999eb-ff02-4c70-b755-f85bd90dd637" Dec 13 07:10:40 crc kubenswrapper[4971]: I1213 07:10:40.211619 4971 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-controller-k5fz9" event={"ID":"c0bf2685-ca54-4508-a580-531b05473e76","Type":"ContainerStarted","Data":"f86f65433d42103d4157937e1e8d9d448e69739771f05ec66ac5ce83bb412c12"} Dec 13 07:10:40 crc kubenswrapper[4971]: E1213 07:10:40.213738 4971 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"init\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/podified-antelope-centos9/openstack-neutron-server@sha256:ea0bf67f1aa5d95a9a07b9c8692c293470f1311792c55d3d57f1f92e56689c33\\\"\"" pod="openstack/dnsmasq-dns-744ffd65bc-zhplm" podUID="b4e999eb-ff02-4c70-b755-f85bd90dd637" Dec 13 07:10:40 crc kubenswrapper[4971]: E1213 07:10:40.213749 4971 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"init\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/podified-antelope-centos9/openstack-neutron-server@sha256:ea0bf67f1aa5d95a9a07b9c8692c293470f1311792c55d3d57f1f92e56689c33\\\"\"" pod="openstack/dnsmasq-dns-95f5f6995-qv9tw" podUID="9a37323b-7693-41c2-8a8f-661291d1ef5e" Dec 13 07:10:40 crc kubenswrapper[4971]: I1213 07:10:40.595377 4971 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovn-controller-ovs-jwh28"] Dec 13 07:10:40 crc kubenswrapper[4971]: W1213 07:10:40.696146 4971 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod1920db8d_2d52_46fa_a4cc_ad787e9886fc.slice/crio-7fb2f961dd61d1a1fae7421b75fedf90bc59323f72d4bb5b8402f8f59f76a427 WatchSource:0}: Error finding container 7fb2f961dd61d1a1fae7421b75fedf90bc59323f72d4bb5b8402f8f59f76a427: Status 404 returned error can't find the container with id 7fb2f961dd61d1a1fae7421b75fedf90bc59323f72d4bb5b8402f8f59f76a427 Dec 13 07:10:40 crc kubenswrapper[4971]: I1213 07:10:40.768835 4971 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-5f854695bc-56d9f" Dec 13 07:10:40 crc kubenswrapper[4971]: I1213 07:10:40.784289 4971 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-84bb9d8bd9-ljf6c" Dec 13 07:10:40 crc kubenswrapper[4971]: I1213 07:10:40.935198 4971 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-fv74m\" (UniqueName: \"kubernetes.io/projected/ee072669-97ce-4d78-a12f-4169120a00f6-kube-api-access-fv74m\") pod \"ee072669-97ce-4d78-a12f-4169120a00f6\" (UID: \"ee072669-97ce-4d78-a12f-4169120a00f6\") " Dec 13 07:10:40 crc kubenswrapper[4971]: I1213 07:10:40.935235 4971 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-rzb6k\" (UniqueName: \"kubernetes.io/projected/038690e5-17a1-4e19-8d3b-f420dbdf7349-kube-api-access-rzb6k\") pod \"038690e5-17a1-4e19-8d3b-f420dbdf7349\" (UID: \"038690e5-17a1-4e19-8d3b-f420dbdf7349\") " Dec 13 07:10:40 crc kubenswrapper[4971]: I1213 07:10:40.935278 4971 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/038690e5-17a1-4e19-8d3b-f420dbdf7349-config\") pod \"038690e5-17a1-4e19-8d3b-f420dbdf7349\" (UID: \"038690e5-17a1-4e19-8d3b-f420dbdf7349\") " Dec 13 07:10:40 crc kubenswrapper[4971]: I1213 07:10:40.935304 4971 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/ee072669-97ce-4d78-a12f-4169120a00f6-config\") pod \"ee072669-97ce-4d78-a12f-4169120a00f6\" (UID: \"ee072669-97ce-4d78-a12f-4169120a00f6\") " Dec 13 07:10:40 crc kubenswrapper[4971]: I1213 07:10:40.935324 4971 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/038690e5-17a1-4e19-8d3b-f420dbdf7349-dns-svc\") pod \"038690e5-17a1-4e19-8d3b-f420dbdf7349\" (UID: \"038690e5-17a1-4e19-8d3b-f420dbdf7349\") " Dec 13 07:10:40 crc kubenswrapper[4971]: I1213 07:10:40.937423 4971 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/ee072669-97ce-4d78-a12f-4169120a00f6-config" (OuterVolumeSpecName: "config") pod "ee072669-97ce-4d78-a12f-4169120a00f6" (UID: "ee072669-97ce-4d78-a12f-4169120a00f6"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 13 07:10:40 crc kubenswrapper[4971]: I1213 07:10:40.937438 4971 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/038690e5-17a1-4e19-8d3b-f420dbdf7349-config" (OuterVolumeSpecName: "config") pod "038690e5-17a1-4e19-8d3b-f420dbdf7349" (UID: "038690e5-17a1-4e19-8d3b-f420dbdf7349"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 13 07:10:40 crc kubenswrapper[4971]: I1213 07:10:40.937950 4971 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/038690e5-17a1-4e19-8d3b-f420dbdf7349-dns-svc" (OuterVolumeSpecName: "dns-svc") pod "038690e5-17a1-4e19-8d3b-f420dbdf7349" (UID: "038690e5-17a1-4e19-8d3b-f420dbdf7349"). InnerVolumeSpecName "dns-svc". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 13 07:10:40 crc kubenswrapper[4971]: I1213 07:10:40.941364 4971 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/038690e5-17a1-4e19-8d3b-f420dbdf7349-kube-api-access-rzb6k" (OuterVolumeSpecName: "kube-api-access-rzb6k") pod "038690e5-17a1-4e19-8d3b-f420dbdf7349" (UID: "038690e5-17a1-4e19-8d3b-f420dbdf7349"). InnerVolumeSpecName "kube-api-access-rzb6k". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 13 07:10:40 crc kubenswrapper[4971]: I1213 07:10:40.941755 4971 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/ee072669-97ce-4d78-a12f-4169120a00f6-kube-api-access-fv74m" (OuterVolumeSpecName: "kube-api-access-fv74m") pod "ee072669-97ce-4d78-a12f-4169120a00f6" (UID: "ee072669-97ce-4d78-a12f-4169120a00f6"). InnerVolumeSpecName "kube-api-access-fv74m". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 13 07:10:41 crc kubenswrapper[4971]: I1213 07:10:41.036800 4971 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-rzb6k\" (UniqueName: \"kubernetes.io/projected/038690e5-17a1-4e19-8d3b-f420dbdf7349-kube-api-access-rzb6k\") on node \"crc\" DevicePath \"\"" Dec 13 07:10:41 crc kubenswrapper[4971]: I1213 07:10:41.036853 4971 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/038690e5-17a1-4e19-8d3b-f420dbdf7349-config\") on node \"crc\" DevicePath \"\"" Dec 13 07:10:41 crc kubenswrapper[4971]: I1213 07:10:41.036863 4971 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/ee072669-97ce-4d78-a12f-4169120a00f6-config\") on node \"crc\" DevicePath \"\"" Dec 13 07:10:41 crc kubenswrapper[4971]: I1213 07:10:41.036872 4971 reconciler_common.go:293] "Volume detached for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/038690e5-17a1-4e19-8d3b-f420dbdf7349-dns-svc\") on node \"crc\" DevicePath \"\"" Dec 13 07:10:41 crc kubenswrapper[4971]: I1213 07:10:41.036881 4971 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-fv74m\" (UniqueName: \"kubernetes.io/projected/ee072669-97ce-4d78-a12f-4169120a00f6-kube-api-access-fv74m\") on node \"crc\" DevicePath \"\"" Dec 13 07:10:41 crc kubenswrapper[4971]: I1213 07:10:41.221859 4971 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-5f854695bc-56d9f" event={"ID":"038690e5-17a1-4e19-8d3b-f420dbdf7349","Type":"ContainerDied","Data":"413ec7db5bd51e525ef83070345f753c3e8cfe41ed771b4f218976331f6f8474"} Dec 13 07:10:41 crc kubenswrapper[4971]: I1213 07:10:41.221882 4971 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-5f854695bc-56d9f" Dec 13 07:10:41 crc kubenswrapper[4971]: I1213 07:10:41.224023 4971 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-84bb9d8bd9-ljf6c" event={"ID":"ee072669-97ce-4d78-a12f-4169120a00f6","Type":"ContainerDied","Data":"98370c8549a5d34a131f41098b3d60a2ca5bbd6fcbb16443696869a561ab5596"} Dec 13 07:10:41 crc kubenswrapper[4971]: I1213 07:10:41.224132 4971 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-84bb9d8bd9-ljf6c" Dec 13 07:10:41 crc kubenswrapper[4971]: I1213 07:10:41.230764 4971 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-controller-ovs-jwh28" event={"ID":"1920db8d-2d52-46fa-a4cc-ad787e9886fc","Type":"ContainerStarted","Data":"7fb2f961dd61d1a1fae7421b75fedf90bc59323f72d4bb5b8402f8f59f76a427"} Dec 13 07:10:41 crc kubenswrapper[4971]: I1213 07:10:41.233486 4971 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/openstack-cell1-galera-0" event={"ID":"96571a02-70f1-46e7-bdd3-935b1f43ecc6","Type":"ContainerStarted","Data":"a5c86247233cf334176c684d1baace4f6267148ecbbe37f773b4c855004df7a6"} Dec 13 07:10:41 crc kubenswrapper[4971]: I1213 07:10:41.236600 4971 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/openstack-galera-0" event={"ID":"91831fd5-fc52-46ec-ac49-63859378116b","Type":"ContainerStarted","Data":"d3f99891e8765e96e30503a2ead01526c3bc2a939b1583fc00c460e636b0105f"} Dec 13 07:10:41 crc kubenswrapper[4971]: I1213 07:10:41.293280 4971 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-84bb9d8bd9-ljf6c"] Dec 13 07:10:41 crc kubenswrapper[4971]: I1213 07:10:41.298504 4971 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-84bb9d8bd9-ljf6c"] Dec 13 07:10:41 crc kubenswrapper[4971]: I1213 07:10:41.326191 4971 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-5f854695bc-56d9f"] Dec 13 07:10:41 crc kubenswrapper[4971]: I1213 07:10:41.337180 4971 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-5f854695bc-56d9f"] Dec 13 07:10:41 crc kubenswrapper[4971]: I1213 07:10:41.472933 4971 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovsdbserver-sb-0"] Dec 13 07:10:41 crc kubenswrapper[4971]: I1213 07:10:41.666894 4971 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovsdbserver-nb-0"] Dec 13 07:10:41 crc kubenswrapper[4971]: I1213 07:10:41.809912 4971 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="038690e5-17a1-4e19-8d3b-f420dbdf7349" path="/var/lib/kubelet/pods/038690e5-17a1-4e19-8d3b-f420dbdf7349/volumes" Dec 13 07:10:41 crc kubenswrapper[4971]: I1213 07:10:41.810358 4971 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="ee072669-97ce-4d78-a12f-4169120a00f6" path="/var/lib/kubelet/pods/ee072669-97ce-4d78-a12f-4169120a00f6/volumes" Dec 13 07:10:41 crc kubenswrapper[4971]: W1213 07:10:41.875732 4971 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-poda85733fd_9af5_4fc4_8096_9b368a793fe7.slice/crio-e7c5eb7ac246b0ecc2e31deb6485e93c292a8feb84c45edbf5878e55dc893cde WatchSource:0}: Error finding container e7c5eb7ac246b0ecc2e31deb6485e93c292a8feb84c45edbf5878e55dc893cde: Status 404 returned error can't find the container with id e7c5eb7ac246b0ecc2e31deb6485e93c292a8feb84c45edbf5878e55dc893cde Dec 13 07:10:41 crc kubenswrapper[4971]: W1213 07:10:41.876118 4971 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod3952f27c_fb18_4e8a_b3a4_a3fa077378c4.slice/crio-f886fad9421de92bee4d5d7c862ea32f6c2ff8b4b711037c29a6bee99562e95d WatchSource:0}: Error finding container f886fad9421de92bee4d5d7c862ea32f6c2ff8b4b711037c29a6bee99562e95d: Status 404 returned error can't find the container with id f886fad9421de92bee4d5d7c862ea32f6c2ff8b4b711037c29a6bee99562e95d Dec 13 07:10:42 crc kubenswrapper[4971]: I1213 07:10:42.245126 4971 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovsdbserver-sb-0" event={"ID":"a85733fd-9af5-4fc4-8096-9b368a793fe7","Type":"ContainerStarted","Data":"e7c5eb7ac246b0ecc2e31deb6485e93c292a8feb84c45edbf5878e55dc893cde"} Dec 13 07:10:42 crc kubenswrapper[4971]: I1213 07:10:42.247841 4971 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/kube-state-metrics-0" event={"ID":"2bab8c0a-bd76-4265-8355-8df0c6fce6c3","Type":"ContainerStarted","Data":"7a965470ccb73744ddf8839144cea274fc2632c1baaa2c97193772f8c5e14440"} Dec 13 07:10:42 crc kubenswrapper[4971]: I1213 07:10:42.247954 4971 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/kube-state-metrics-0" Dec 13 07:10:42 crc kubenswrapper[4971]: I1213 07:10:42.249603 4971 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovsdbserver-nb-0" event={"ID":"3952f27c-fb18-4e8a-b3a4-a3fa077378c4","Type":"ContainerStarted","Data":"f886fad9421de92bee4d5d7c862ea32f6c2ff8b4b711037c29a6bee99562e95d"} Dec 13 07:10:42 crc kubenswrapper[4971]: I1213 07:10:42.268957 4971 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/kube-state-metrics-0" podStartSLOduration=14.999730543 podStartE2EDuration="31.268935167s" podCreationTimestamp="2025-12-13 07:10:11 +0000 UTC" firstStartedPulling="2025-12-13 07:10:25.652467943 +0000 UTC m=+1282.256877401" lastFinishedPulling="2025-12-13 07:10:41.921672567 +0000 UTC m=+1298.526082025" observedRunningTime="2025-12-13 07:10:42.261922994 +0000 UTC m=+1298.866332472" watchObservedRunningTime="2025-12-13 07:10:42.268935167 +0000 UTC m=+1298.873344615" Dec 13 07:10:44 crc kubenswrapper[4971]: I1213 07:10:44.265399 4971 generic.go:334] "Generic (PLEG): container finished" podID="96571a02-70f1-46e7-bdd3-935b1f43ecc6" containerID="a5c86247233cf334176c684d1baace4f6267148ecbbe37f773b4c855004df7a6" exitCode=0 Dec 13 07:10:44 crc kubenswrapper[4971]: I1213 07:10:44.265693 4971 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/openstack-cell1-galera-0" event={"ID":"96571a02-70f1-46e7-bdd3-935b1f43ecc6","Type":"ContainerDied","Data":"a5c86247233cf334176c684d1baace4f6267148ecbbe37f773b4c855004df7a6"} Dec 13 07:10:44 crc kubenswrapper[4971]: I1213 07:10:44.270081 4971 generic.go:334] "Generic (PLEG): container finished" podID="91831fd5-fc52-46ec-ac49-63859378116b" containerID="d3f99891e8765e96e30503a2ead01526c3bc2a939b1583fc00c460e636b0105f" exitCode=0 Dec 13 07:10:44 crc kubenswrapper[4971]: I1213 07:10:44.270178 4971 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/openstack-galera-0" event={"ID":"91831fd5-fc52-46ec-ac49-63859378116b","Type":"ContainerDied","Data":"d3f99891e8765e96e30503a2ead01526c3bc2a939b1583fc00c460e636b0105f"} Dec 13 07:10:44 crc kubenswrapper[4971]: I1213 07:10:44.274009 4971 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-controller-ovs-jwh28" event={"ID":"1920db8d-2d52-46fa-a4cc-ad787e9886fc","Type":"ContainerStarted","Data":"e2598d64391bc83ad14b9e2b00c2cec0a72d692efb8c2d6fbb7cd6a0e5831ca5"} Dec 13 07:10:45 crc kubenswrapper[4971]: I1213 07:10:45.283477 4971 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovsdbserver-sb-0" event={"ID":"a85733fd-9af5-4fc4-8096-9b368a793fe7","Type":"ContainerStarted","Data":"c9e97efec5854b40a7ff94d26e756413e1b2cbc1cda1a26a1cb4fa86fe1d5130"} Dec 13 07:10:45 crc kubenswrapper[4971]: I1213 07:10:45.285075 4971 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-controller-k5fz9" event={"ID":"c0bf2685-ca54-4508-a580-531b05473e76","Type":"ContainerStarted","Data":"9cd9e5d36ac4ce3ea02e50acba7ef18ff948d6c3b77f2c98e3614edbbebf48ca"} Dec 13 07:10:45 crc kubenswrapper[4971]: I1213 07:10:45.285201 4971 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/ovn-controller-k5fz9" Dec 13 07:10:45 crc kubenswrapper[4971]: I1213 07:10:45.287865 4971 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/openstack-cell1-galera-0" event={"ID":"96571a02-70f1-46e7-bdd3-935b1f43ecc6","Type":"ContainerStarted","Data":"f0f0af597a780891be3c816c94adaa9278edb108ead08128f1d224af2f29abeb"} Dec 13 07:10:45 crc kubenswrapper[4971]: I1213 07:10:45.291438 4971 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/openstack-galera-0" event={"ID":"91831fd5-fc52-46ec-ac49-63859378116b","Type":"ContainerStarted","Data":"4355bade7915a5130d577c69d3a29ad82f33251bd809c84c63697d3e9663cb1b"} Dec 13 07:10:45 crc kubenswrapper[4971]: I1213 07:10:45.292937 4971 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovsdbserver-nb-0" event={"ID":"3952f27c-fb18-4e8a-b3a4-a3fa077378c4","Type":"ContainerStarted","Data":"4d5209b0db3f6210ca0a5c2f2f8e1f491abf0445b24b4b9e24481ba434617e30"} Dec 13 07:10:45 crc kubenswrapper[4971]: I1213 07:10:45.294817 4971 generic.go:334] "Generic (PLEG): container finished" podID="1920db8d-2d52-46fa-a4cc-ad787e9886fc" containerID="e2598d64391bc83ad14b9e2b00c2cec0a72d692efb8c2d6fbb7cd6a0e5831ca5" exitCode=0 Dec 13 07:10:45 crc kubenswrapper[4971]: I1213 07:10:45.294859 4971 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-controller-ovs-jwh28" event={"ID":"1920db8d-2d52-46fa-a4cc-ad787e9886fc","Type":"ContainerDied","Data":"e2598d64391bc83ad14b9e2b00c2cec0a72d692efb8c2d6fbb7cd6a0e5831ca5"} Dec 13 07:10:45 crc kubenswrapper[4971]: I1213 07:10:45.317088 4971 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/ovn-controller-k5fz9" podStartSLOduration=27.619342644 podStartE2EDuration="31.317065671s" podCreationTimestamp="2025-12-13 07:10:14 +0000 UTC" firstStartedPulling="2025-12-13 07:10:39.973977626 +0000 UTC m=+1296.578387074" lastFinishedPulling="2025-12-13 07:10:43.671700653 +0000 UTC m=+1300.276110101" observedRunningTime="2025-12-13 07:10:45.31134641 +0000 UTC m=+1301.915755858" watchObservedRunningTime="2025-12-13 07:10:45.317065671 +0000 UTC m=+1301.921475119" Dec 13 07:10:45 crc kubenswrapper[4971]: I1213 07:10:45.352655 4971 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/openstack-galera-0" podStartSLOduration=8.839914443 podStartE2EDuration="39.352634167s" podCreationTimestamp="2025-12-13 07:10:06 +0000 UTC" firstStartedPulling="2025-12-13 07:10:09.469036514 +0000 UTC m=+1266.073445962" lastFinishedPulling="2025-12-13 07:10:39.981756228 +0000 UTC m=+1296.586165686" observedRunningTime="2025-12-13 07:10:45.346585308 +0000 UTC m=+1301.950994766" watchObservedRunningTime="2025-12-13 07:10:45.352634167 +0000 UTC m=+1301.957043615" Dec 13 07:10:46 crc kubenswrapper[4971]: I1213 07:10:46.153544 4971 patch_prober.go:28] interesting pod/machine-config-daemon-82xjz container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 13 07:10:46 crc kubenswrapper[4971]: I1213 07:10:46.153730 4971 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-82xjz" podUID="e0c4af12-edda-41d7-9b44-f87396174ef2" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 13 07:10:46 crc kubenswrapper[4971]: I1213 07:10:46.308078 4971 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-controller-ovs-jwh28" event={"ID":"1920db8d-2d52-46fa-a4cc-ad787e9886fc","Type":"ContainerStarted","Data":"64a7b8362c13f4ab3b7c9d0efa7d068fdd98576f696c1c87d5c324995f8735d7"} Dec 13 07:10:46 crc kubenswrapper[4971]: I1213 07:10:46.308422 4971 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/ovn-controller-ovs-jwh28" Dec 13 07:10:46 crc kubenswrapper[4971]: I1213 07:10:46.308453 4971 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-controller-ovs-jwh28" event={"ID":"1920db8d-2d52-46fa-a4cc-ad787e9886fc","Type":"ContainerStarted","Data":"2602cf102a68e055e12e1c8719541b9c5a1366928b538f3a91fc586eec55aba1"} Dec 13 07:10:46 crc kubenswrapper[4971]: I1213 07:10:46.344701 4971 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/openstack-cell1-galera-0" podStartSLOduration=10.153526474 podStartE2EDuration="39.344676811s" podCreationTimestamp="2025-12-13 07:10:07 +0000 UTC" firstStartedPulling="2025-12-13 07:10:10.783891995 +0000 UTC m=+1267.388301443" lastFinishedPulling="2025-12-13 07:10:39.975042332 +0000 UTC m=+1296.579451780" observedRunningTime="2025-12-13 07:10:45.37713921 +0000 UTC m=+1301.981548678" watchObservedRunningTime="2025-12-13 07:10:46.344676811 +0000 UTC m=+1302.949086259" Dec 13 07:10:46 crc kubenswrapper[4971]: I1213 07:10:46.346418 4971 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/ovn-controller-ovs-jwh28" podStartSLOduration=28.380204705 podStartE2EDuration="31.346407993s" podCreationTimestamp="2025-12-13 07:10:15 +0000 UTC" firstStartedPulling="2025-12-13 07:10:40.699526758 +0000 UTC m=+1297.303936206" lastFinishedPulling="2025-12-13 07:10:43.665730046 +0000 UTC m=+1300.270139494" observedRunningTime="2025-12-13 07:10:46.340935628 +0000 UTC m=+1302.945345086" watchObservedRunningTime="2025-12-13 07:10:46.346407993 +0000 UTC m=+1302.950817441" Dec 13 07:10:47 crc kubenswrapper[4971]: I1213 07:10:47.333082 4971 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/ovn-controller-ovs-jwh28" Dec 13 07:10:47 crc kubenswrapper[4971]: I1213 07:10:47.829103 4971 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/openstack-galera-0" Dec 13 07:10:47 crc kubenswrapper[4971]: I1213 07:10:47.829386 4971 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/openstack-galera-0" Dec 13 07:10:49 crc kubenswrapper[4971]: I1213 07:10:49.349344 4971 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovsdbserver-nb-0" event={"ID":"3952f27c-fb18-4e8a-b3a4-a3fa077378c4","Type":"ContainerStarted","Data":"b3c211afb72ca508b5ce5485e6d4bbb03261eb8755ef8eea2de2cc58277d066c"} Dec 13 07:10:49 crc kubenswrapper[4971]: I1213 07:10:49.357440 4971 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovsdbserver-sb-0" event={"ID":"a85733fd-9af5-4fc4-8096-9b368a793fe7","Type":"ContainerStarted","Data":"86f686ae1a0503d66045fcfb45dfaa00eed4705bd4a7ce78b1d4a11edf03dfff"} Dec 13 07:10:49 crc kubenswrapper[4971]: I1213 07:10:49.374149 4971 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/ovsdbserver-nb-0" podStartSLOduration=28.531115911 podStartE2EDuration="35.374127065s" podCreationTimestamp="2025-12-13 07:10:14 +0000 UTC" firstStartedPulling="2025-12-13 07:10:41.878823262 +0000 UTC m=+1298.483232710" lastFinishedPulling="2025-12-13 07:10:48.721834416 +0000 UTC m=+1305.326243864" observedRunningTime="2025-12-13 07:10:49.373030538 +0000 UTC m=+1305.977440006" watchObservedRunningTime="2025-12-13 07:10:49.374127065 +0000 UTC m=+1305.978536513" Dec 13 07:10:49 crc kubenswrapper[4971]: I1213 07:10:49.397893 4971 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/ovsdbserver-sb-0" podStartSLOduration=25.636171988 podStartE2EDuration="32.39787459s" podCreationTimestamp="2025-12-13 07:10:17 +0000 UTC" firstStartedPulling="2025-12-13 07:10:41.879465738 +0000 UTC m=+1298.483875186" lastFinishedPulling="2025-12-13 07:10:48.64116834 +0000 UTC m=+1305.245577788" observedRunningTime="2025-12-13 07:10:49.392925148 +0000 UTC m=+1305.997334606" watchObservedRunningTime="2025-12-13 07:10:49.39787459 +0000 UTC m=+1306.002284028" Dec 13 07:10:49 crc kubenswrapper[4971]: I1213 07:10:49.495673 4971 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/openstack-cell1-galera-0" Dec 13 07:10:49 crc kubenswrapper[4971]: I1213 07:10:49.495991 4971 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/openstack-cell1-galera-0" Dec 13 07:10:50 crc kubenswrapper[4971]: I1213 07:10:50.030933 4971 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/openstack-galera-0" Dec 13 07:10:50 crc kubenswrapper[4971]: I1213 07:10:50.097952 4971 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/openstack-galera-0" Dec 13 07:10:50 crc kubenswrapper[4971]: I1213 07:10:50.420134 4971 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-cell1-server-0" event={"ID":"5eed809d-c915-433c-9915-27e0b01e1ffe","Type":"ContainerStarted","Data":"0ecc7342a1a65b67ea97dd2575e3d43416e7633d3c5e40a8ec3ff6b6e6709a86"} Dec 13 07:10:50 crc kubenswrapper[4971]: I1213 07:10:50.421791 4971 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-server-0" event={"ID":"a590eb84-c8d5-4cbe-9c03-03a6ccb82073","Type":"ContainerStarted","Data":"2632986cd6a0be018d82fef579687199e356535dfe73855c9af020cc784bc892"} Dec 13 07:10:50 crc kubenswrapper[4971]: I1213 07:10:50.423616 4971 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/memcached-0" event={"ID":"e6adf8ca-c137-4b7d-bf8c-74c0766e57f0","Type":"ContainerStarted","Data":"7a7ce8c6d6226ca0f66c775c7742ec428b25c94e94e7a3b5d21e105f8e6d11c9"} Dec 13 07:10:50 crc kubenswrapper[4971]: I1213 07:10:50.424120 4971 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/memcached-0" Dec 13 07:10:50 crc kubenswrapper[4971]: I1213 07:10:50.468722 4971 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/memcached-0" podStartSLOduration=2.657242626 podStartE2EDuration="41.468698903s" podCreationTimestamp="2025-12-13 07:10:09 +0000 UTC" firstStartedPulling="2025-12-13 07:10:10.42422464 +0000 UTC m=+1267.028634088" lastFinishedPulling="2025-12-13 07:10:49.235680917 +0000 UTC m=+1305.840090365" observedRunningTime="2025-12-13 07:10:50.467817432 +0000 UTC m=+1307.072226880" watchObservedRunningTime="2025-12-13 07:10:50.468698903 +0000 UTC m=+1307.073108351" Dec 13 07:10:51 crc kubenswrapper[4971]: I1213 07:10:51.099174 4971 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/ovsdbserver-nb-0" Dec 13 07:10:51 crc kubenswrapper[4971]: I1213 07:10:51.940081 4971 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/kube-state-metrics-0" Dec 13 07:10:51 crc kubenswrapper[4971]: I1213 07:10:51.978245 4971 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/openstack-cell1-galera-0" Dec 13 07:10:52 crc kubenswrapper[4971]: I1213 07:10:52.056899 4971 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/openstack-cell1-galera-0" Dec 13 07:10:52 crc kubenswrapper[4971]: I1213 07:10:52.099102 4971 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/ovsdbserver-nb-0" Dec 13 07:10:52 crc kubenswrapper[4971]: I1213 07:10:52.136223 4971 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/ovsdbserver-nb-0" Dec 13 07:10:52 crc kubenswrapper[4971]: I1213 07:10:52.150068 4971 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/ovsdbserver-sb-0" Dec 13 07:10:52 crc kubenswrapper[4971]: I1213 07:10:52.191842 4971 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/ovsdbserver-sb-0" Dec 13 07:10:52 crc kubenswrapper[4971]: I1213 07:10:52.617551 4971 generic.go:334] "Generic (PLEG): container finished" podID="b4e999eb-ff02-4c70-b755-f85bd90dd637" containerID="6f52043e33e0ed905f8dc75a68ad19b0628da94d63ca3eaec2836a6dbde42bda" exitCode=0 Dec 13 07:10:52 crc kubenswrapper[4971]: I1213 07:10:52.617630 4971 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-744ffd65bc-zhplm" event={"ID":"b4e999eb-ff02-4c70-b755-f85bd90dd637","Type":"ContainerDied","Data":"6f52043e33e0ed905f8dc75a68ad19b0628da94d63ca3eaec2836a6dbde42bda"} Dec 13 07:10:52 crc kubenswrapper[4971]: I1213 07:10:52.618569 4971 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/ovsdbserver-sb-0" Dec 13 07:10:52 crc kubenswrapper[4971]: I1213 07:10:52.665960 4971 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/ovsdbserver-sb-0" Dec 13 07:10:52 crc kubenswrapper[4971]: I1213 07:10:52.672552 4971 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/ovsdbserver-nb-0" Dec 13 07:10:52 crc kubenswrapper[4971]: I1213 07:10:52.962776 4971 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-744ffd65bc-zhplm"] Dec 13 07:10:52 crc kubenswrapper[4971]: I1213 07:10:52.997639 4971 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-794868bd45-t92j4"] Dec 13 07:10:52 crc kubenswrapper[4971]: I1213 07:10:52.999259 4971 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-794868bd45-t92j4" Dec 13 07:10:53 crc kubenswrapper[4971]: I1213 07:10:53.004057 4971 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"ovsdbserver-sb" Dec 13 07:10:53 crc kubenswrapper[4971]: I1213 07:10:53.010766 4971 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-794868bd45-t92j4"] Dec 13 07:10:53 crc kubenswrapper[4971]: I1213 07:10:53.245735 4971 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ovn-controller-metrics-r2cbj"] Dec 13 07:10:53 crc kubenswrapper[4971]: I1213 07:10:53.246984 4971 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-controller-metrics-r2cbj" Dec 13 07:10:53 crc kubenswrapper[4971]: I1213 07:10:53.250042 4971 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"ovncontroller-metrics-config" Dec 13 07:10:53 crc kubenswrapper[4971]: I1213 07:10:53.269235 4971 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovn-controller-metrics-r2cbj"] Dec 13 07:10:53 crc kubenswrapper[4971]: I1213 07:10:53.270100 4971 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/20e72d59-fbc4-4eca-84d0-04c833aa4aab-config\") pod \"dnsmasq-dns-794868bd45-t92j4\" (UID: \"20e72d59-fbc4-4eca-84d0-04c833aa4aab\") " pod="openstack/dnsmasq-dns-794868bd45-t92j4" Dec 13 07:10:53 crc kubenswrapper[4971]: I1213 07:10:53.270167 4971 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-n652v\" (UniqueName: \"kubernetes.io/projected/20e72d59-fbc4-4eca-84d0-04c833aa4aab-kube-api-access-n652v\") pod \"dnsmasq-dns-794868bd45-t92j4\" (UID: \"20e72d59-fbc4-4eca-84d0-04c833aa4aab\") " pod="openstack/dnsmasq-dns-794868bd45-t92j4" Dec 13 07:10:53 crc kubenswrapper[4971]: I1213 07:10:53.270256 4971 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/20e72d59-fbc4-4eca-84d0-04c833aa4aab-ovsdbserver-sb\") pod \"dnsmasq-dns-794868bd45-t92j4\" (UID: \"20e72d59-fbc4-4eca-84d0-04c833aa4aab\") " pod="openstack/dnsmasq-dns-794868bd45-t92j4" Dec 13 07:10:53 crc kubenswrapper[4971]: I1213 07:10:53.270329 4971 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/20e72d59-fbc4-4eca-84d0-04c833aa4aab-dns-svc\") pod \"dnsmasq-dns-794868bd45-t92j4\" (UID: \"20e72d59-fbc4-4eca-84d0-04c833aa4aab\") " pod="openstack/dnsmasq-dns-794868bd45-t92j4" Dec 13 07:10:53 crc kubenswrapper[4971]: I1213 07:10:53.371836 4971 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovs-rundir\" (UniqueName: \"kubernetes.io/host-path/fa04788e-0c35-4205-97ed-cb1e38ec9e69-ovs-rundir\") pod \"ovn-controller-metrics-r2cbj\" (UID: \"fa04788e-0c35-4205-97ed-cb1e38ec9e69\") " pod="openstack/ovn-controller-metrics-r2cbj" Dec 13 07:10:53 crc kubenswrapper[4971]: I1213 07:10:53.371897 4971 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-ppss8\" (UniqueName: \"kubernetes.io/projected/fa04788e-0c35-4205-97ed-cb1e38ec9e69-kube-api-access-ppss8\") pod \"ovn-controller-metrics-r2cbj\" (UID: \"fa04788e-0c35-4205-97ed-cb1e38ec9e69\") " pod="openstack/ovn-controller-metrics-r2cbj" Dec 13 07:10:53 crc kubenswrapper[4971]: I1213 07:10:53.371939 4971 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/20e72d59-fbc4-4eca-84d0-04c833aa4aab-ovsdbserver-sb\") pod \"dnsmasq-dns-794868bd45-t92j4\" (UID: \"20e72d59-fbc4-4eca-84d0-04c833aa4aab\") " pod="openstack/dnsmasq-dns-794868bd45-t92j4" Dec 13 07:10:53 crc kubenswrapper[4971]: I1213 07:10:53.371994 4971 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/fa04788e-0c35-4205-97ed-cb1e38ec9e69-combined-ca-bundle\") pod \"ovn-controller-metrics-r2cbj\" (UID: \"fa04788e-0c35-4205-97ed-cb1e38ec9e69\") " pod="openstack/ovn-controller-metrics-r2cbj" Dec 13 07:10:53 crc kubenswrapper[4971]: I1213 07:10:53.372021 4971 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metrics-certs-tls-certs\" (UniqueName: \"kubernetes.io/secret/fa04788e-0c35-4205-97ed-cb1e38ec9e69-metrics-certs-tls-certs\") pod \"ovn-controller-metrics-r2cbj\" (UID: \"fa04788e-0c35-4205-97ed-cb1e38ec9e69\") " pod="openstack/ovn-controller-metrics-r2cbj" Dec 13 07:10:53 crc kubenswrapper[4971]: I1213 07:10:53.372052 4971 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/20e72d59-fbc4-4eca-84d0-04c833aa4aab-dns-svc\") pod \"dnsmasq-dns-794868bd45-t92j4\" (UID: \"20e72d59-fbc4-4eca-84d0-04c833aa4aab\") " pod="openstack/dnsmasq-dns-794868bd45-t92j4" Dec 13 07:10:53 crc kubenswrapper[4971]: I1213 07:10:53.372080 4971 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/20e72d59-fbc4-4eca-84d0-04c833aa4aab-config\") pod \"dnsmasq-dns-794868bd45-t92j4\" (UID: \"20e72d59-fbc4-4eca-84d0-04c833aa4aab\") " pod="openstack/dnsmasq-dns-794868bd45-t92j4" Dec 13 07:10:53 crc kubenswrapper[4971]: I1213 07:10:53.372103 4971 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/fa04788e-0c35-4205-97ed-cb1e38ec9e69-config\") pod \"ovn-controller-metrics-r2cbj\" (UID: \"fa04788e-0c35-4205-97ed-cb1e38ec9e69\") " pod="openstack/ovn-controller-metrics-r2cbj" Dec 13 07:10:53 crc kubenswrapper[4971]: I1213 07:10:53.372122 4971 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-n652v\" (UniqueName: \"kubernetes.io/projected/20e72d59-fbc4-4eca-84d0-04c833aa4aab-kube-api-access-n652v\") pod \"dnsmasq-dns-794868bd45-t92j4\" (UID: \"20e72d59-fbc4-4eca-84d0-04c833aa4aab\") " pod="openstack/dnsmasq-dns-794868bd45-t92j4" Dec 13 07:10:53 crc kubenswrapper[4971]: I1213 07:10:53.372148 4971 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovn-rundir\" (UniqueName: \"kubernetes.io/host-path/fa04788e-0c35-4205-97ed-cb1e38ec9e69-ovn-rundir\") pod \"ovn-controller-metrics-r2cbj\" (UID: \"fa04788e-0c35-4205-97ed-cb1e38ec9e69\") " pod="openstack/ovn-controller-metrics-r2cbj" Dec 13 07:10:53 crc kubenswrapper[4971]: I1213 07:10:53.373164 4971 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/20e72d59-fbc4-4eca-84d0-04c833aa4aab-ovsdbserver-sb\") pod \"dnsmasq-dns-794868bd45-t92j4\" (UID: \"20e72d59-fbc4-4eca-84d0-04c833aa4aab\") " pod="openstack/dnsmasq-dns-794868bd45-t92j4" Dec 13 07:10:53 crc kubenswrapper[4971]: I1213 07:10:53.373188 4971 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/20e72d59-fbc4-4eca-84d0-04c833aa4aab-dns-svc\") pod \"dnsmasq-dns-794868bd45-t92j4\" (UID: \"20e72d59-fbc4-4eca-84d0-04c833aa4aab\") " pod="openstack/dnsmasq-dns-794868bd45-t92j4" Dec 13 07:10:53 crc kubenswrapper[4971]: I1213 07:10:53.373219 4971 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/20e72d59-fbc4-4eca-84d0-04c833aa4aab-config\") pod \"dnsmasq-dns-794868bd45-t92j4\" (UID: \"20e72d59-fbc4-4eca-84d0-04c833aa4aab\") " pod="openstack/dnsmasq-dns-794868bd45-t92j4" Dec 13 07:10:53 crc kubenswrapper[4971]: I1213 07:10:53.423543 4971 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-n652v\" (UniqueName: \"kubernetes.io/projected/20e72d59-fbc4-4eca-84d0-04c833aa4aab-kube-api-access-n652v\") pod \"dnsmasq-dns-794868bd45-t92j4\" (UID: \"20e72d59-fbc4-4eca-84d0-04c833aa4aab\") " pod="openstack/dnsmasq-dns-794868bd45-t92j4" Dec 13 07:10:53 crc kubenswrapper[4971]: I1213 07:10:53.442826 4971 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-95f5f6995-qv9tw"] Dec 13 07:10:53 crc kubenswrapper[4971]: I1213 07:10:53.456309 4971 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ovn-northd-0"] Dec 13 07:10:53 crc kubenswrapper[4971]: I1213 07:10:53.457897 4971 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-northd-0" Dec 13 07:10:53 crc kubenswrapper[4971]: I1213 07:10:53.462805 4971 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"ovnnorthd-scripts" Dec 13 07:10:53 crc kubenswrapper[4971]: I1213 07:10:53.463265 4971 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"ovnnorthd-config" Dec 13 07:10:53 crc kubenswrapper[4971]: I1213 07:10:53.463370 4971 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-ovnnorthd-ovndbs" Dec 13 07:10:53 crc kubenswrapper[4971]: I1213 07:10:53.463390 4971 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ovnnorthd-ovnnorthd-dockercfg-9jc52" Dec 13 07:10:53 crc kubenswrapper[4971]: I1213 07:10:53.469790 4971 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovn-northd-0"] Dec 13 07:10:53 crc kubenswrapper[4971]: I1213 07:10:53.473572 4971 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/fa04788e-0c35-4205-97ed-cb1e38ec9e69-config\") pod \"ovn-controller-metrics-r2cbj\" (UID: \"fa04788e-0c35-4205-97ed-cb1e38ec9e69\") " pod="openstack/ovn-controller-metrics-r2cbj" Dec 13 07:10:53 crc kubenswrapper[4971]: I1213 07:10:53.473642 4971 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovn-rundir\" (UniqueName: \"kubernetes.io/host-path/fa04788e-0c35-4205-97ed-cb1e38ec9e69-ovn-rundir\") pod \"ovn-controller-metrics-r2cbj\" (UID: \"fa04788e-0c35-4205-97ed-cb1e38ec9e69\") " pod="openstack/ovn-controller-metrics-r2cbj" Dec 13 07:10:53 crc kubenswrapper[4971]: I1213 07:10:53.473724 4971 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovs-rundir\" (UniqueName: \"kubernetes.io/host-path/fa04788e-0c35-4205-97ed-cb1e38ec9e69-ovs-rundir\") pod \"ovn-controller-metrics-r2cbj\" (UID: \"fa04788e-0c35-4205-97ed-cb1e38ec9e69\") " pod="openstack/ovn-controller-metrics-r2cbj" Dec 13 07:10:53 crc kubenswrapper[4971]: I1213 07:10:53.473755 4971 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-ppss8\" (UniqueName: \"kubernetes.io/projected/fa04788e-0c35-4205-97ed-cb1e38ec9e69-kube-api-access-ppss8\") pod \"ovn-controller-metrics-r2cbj\" (UID: \"fa04788e-0c35-4205-97ed-cb1e38ec9e69\") " pod="openstack/ovn-controller-metrics-r2cbj" Dec 13 07:10:53 crc kubenswrapper[4971]: I1213 07:10:53.473826 4971 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/fa04788e-0c35-4205-97ed-cb1e38ec9e69-combined-ca-bundle\") pod \"ovn-controller-metrics-r2cbj\" (UID: \"fa04788e-0c35-4205-97ed-cb1e38ec9e69\") " pod="openstack/ovn-controller-metrics-r2cbj" Dec 13 07:10:53 crc kubenswrapper[4971]: I1213 07:10:53.473864 4971 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs-tls-certs\" (UniqueName: \"kubernetes.io/secret/fa04788e-0c35-4205-97ed-cb1e38ec9e69-metrics-certs-tls-certs\") pod \"ovn-controller-metrics-r2cbj\" (UID: \"fa04788e-0c35-4205-97ed-cb1e38ec9e69\") " pod="openstack/ovn-controller-metrics-r2cbj" Dec 13 07:10:53 crc kubenswrapper[4971]: I1213 07:10:53.475903 4971 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovs-rundir\" (UniqueName: \"kubernetes.io/host-path/fa04788e-0c35-4205-97ed-cb1e38ec9e69-ovs-rundir\") pod \"ovn-controller-metrics-r2cbj\" (UID: \"fa04788e-0c35-4205-97ed-cb1e38ec9e69\") " pod="openstack/ovn-controller-metrics-r2cbj" Dec 13 07:10:53 crc kubenswrapper[4971]: I1213 07:10:53.476204 4971 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovn-rundir\" (UniqueName: \"kubernetes.io/host-path/fa04788e-0c35-4205-97ed-cb1e38ec9e69-ovn-rundir\") pod \"ovn-controller-metrics-r2cbj\" (UID: \"fa04788e-0c35-4205-97ed-cb1e38ec9e69\") " pod="openstack/ovn-controller-metrics-r2cbj" Dec 13 07:10:53 crc kubenswrapper[4971]: I1213 07:10:53.476587 4971 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/fa04788e-0c35-4205-97ed-cb1e38ec9e69-config\") pod \"ovn-controller-metrics-r2cbj\" (UID: \"fa04788e-0c35-4205-97ed-cb1e38ec9e69\") " pod="openstack/ovn-controller-metrics-r2cbj" Dec 13 07:10:53 crc kubenswrapper[4971]: I1213 07:10:53.479721 4971 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/fa04788e-0c35-4205-97ed-cb1e38ec9e69-combined-ca-bundle\") pod \"ovn-controller-metrics-r2cbj\" (UID: \"fa04788e-0c35-4205-97ed-cb1e38ec9e69\") " pod="openstack/ovn-controller-metrics-r2cbj" Dec 13 07:10:53 crc kubenswrapper[4971]: I1213 07:10:53.485649 4971 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metrics-certs-tls-certs\" (UniqueName: \"kubernetes.io/secret/fa04788e-0c35-4205-97ed-cb1e38ec9e69-metrics-certs-tls-certs\") pod \"ovn-controller-metrics-r2cbj\" (UID: \"fa04788e-0c35-4205-97ed-cb1e38ec9e69\") " pod="openstack/ovn-controller-metrics-r2cbj" Dec 13 07:10:53 crc kubenswrapper[4971]: I1213 07:10:53.487231 4971 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-757dc6fff9-hfmv4"] Dec 13 07:10:53 crc kubenswrapper[4971]: I1213 07:10:53.488672 4971 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-757dc6fff9-hfmv4" Dec 13 07:10:53 crc kubenswrapper[4971]: I1213 07:10:53.504165 4971 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"ovsdbserver-nb" Dec 13 07:10:53 crc kubenswrapper[4971]: I1213 07:10:53.516184 4971 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-ppss8\" (UniqueName: \"kubernetes.io/projected/fa04788e-0c35-4205-97ed-cb1e38ec9e69-kube-api-access-ppss8\") pod \"ovn-controller-metrics-r2cbj\" (UID: \"fa04788e-0c35-4205-97ed-cb1e38ec9e69\") " pod="openstack/ovn-controller-metrics-r2cbj" Dec 13 07:10:53 crc kubenswrapper[4971]: I1213 07:10:53.560151 4971 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-757dc6fff9-hfmv4"] Dec 13 07:10:53 crc kubenswrapper[4971]: I1213 07:10:53.581619 4971 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metrics-certs-tls-certs\" (UniqueName: \"kubernetes.io/secret/2bbf81f1-6733-4320-9bf5-c73a00d9051a-metrics-certs-tls-certs\") pod \"ovn-northd-0\" (UID: \"2bbf81f1-6733-4320-9bf5-c73a00d9051a\") " pod="openstack/ovn-northd-0" Dec 13 07:10:53 crc kubenswrapper[4971]: I1213 07:10:53.581690 4971 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovn-northd-tls-certs\" (UniqueName: \"kubernetes.io/secret/2bbf81f1-6733-4320-9bf5-c73a00d9051a-ovn-northd-tls-certs\") pod \"ovn-northd-0\" (UID: \"2bbf81f1-6733-4320-9bf5-c73a00d9051a\") " pod="openstack/ovn-northd-0" Dec 13 07:10:53 crc kubenswrapper[4971]: I1213 07:10:53.581728 4971 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-46dxf\" (UniqueName: \"kubernetes.io/projected/2bbf81f1-6733-4320-9bf5-c73a00d9051a-kube-api-access-46dxf\") pod \"ovn-northd-0\" (UID: \"2bbf81f1-6733-4320-9bf5-c73a00d9051a\") " pod="openstack/ovn-northd-0" Dec 13 07:10:53 crc kubenswrapper[4971]: I1213 07:10:53.581762 4971 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/2bbf81f1-6733-4320-9bf5-c73a00d9051a-combined-ca-bundle\") pod \"ovn-northd-0\" (UID: \"2bbf81f1-6733-4320-9bf5-c73a00d9051a\") " pod="openstack/ovn-northd-0" Dec 13 07:10:53 crc kubenswrapper[4971]: I1213 07:10:53.581831 4971 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/2bbf81f1-6733-4320-9bf5-c73a00d9051a-scripts\") pod \"ovn-northd-0\" (UID: \"2bbf81f1-6733-4320-9bf5-c73a00d9051a\") " pod="openstack/ovn-northd-0" Dec 13 07:10:53 crc kubenswrapper[4971]: I1213 07:10:53.581895 4971 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/2bbf81f1-6733-4320-9bf5-c73a00d9051a-config\") pod \"ovn-northd-0\" (UID: \"2bbf81f1-6733-4320-9bf5-c73a00d9051a\") " pod="openstack/ovn-northd-0" Dec 13 07:10:53 crc kubenswrapper[4971]: I1213 07:10:53.581931 4971 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovn-rundir\" (UniqueName: \"kubernetes.io/empty-dir/2bbf81f1-6733-4320-9bf5-c73a00d9051a-ovn-rundir\") pod \"ovn-northd-0\" (UID: \"2bbf81f1-6733-4320-9bf5-c73a00d9051a\") " pod="openstack/ovn-northd-0" Dec 13 07:10:53 crc kubenswrapper[4971]: I1213 07:10:53.654585 4971 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-794868bd45-t92j4" Dec 13 07:10:53 crc kubenswrapper[4971]: I1213 07:10:53.655072 4971 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-controller-metrics-r2cbj" Dec 13 07:10:53 crc kubenswrapper[4971]: I1213 07:10:53.688407 4971 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/82d7cf8c-129a-483a-b190-8480e2f03418-config\") pod \"dnsmasq-dns-757dc6fff9-hfmv4\" (UID: \"82d7cf8c-129a-483a-b190-8480e2f03418\") " pod="openstack/dnsmasq-dns-757dc6fff9-hfmv4" Dec 13 07:10:53 crc kubenswrapper[4971]: I1213 07:10:53.688480 4971 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs-tls-certs\" (UniqueName: \"kubernetes.io/secret/2bbf81f1-6733-4320-9bf5-c73a00d9051a-metrics-certs-tls-certs\") pod \"ovn-northd-0\" (UID: \"2bbf81f1-6733-4320-9bf5-c73a00d9051a\") " pod="openstack/ovn-northd-0" Dec 13 07:10:53 crc kubenswrapper[4971]: I1213 07:10:53.688500 4971 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/82d7cf8c-129a-483a-b190-8480e2f03418-ovsdbserver-nb\") pod \"dnsmasq-dns-757dc6fff9-hfmv4\" (UID: \"82d7cf8c-129a-483a-b190-8480e2f03418\") " pod="openstack/dnsmasq-dns-757dc6fff9-hfmv4" Dec 13 07:10:53 crc kubenswrapper[4971]: I1213 07:10:53.696018 4971 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovn-northd-tls-certs\" (UniqueName: \"kubernetes.io/secret/2bbf81f1-6733-4320-9bf5-c73a00d9051a-ovn-northd-tls-certs\") pod \"ovn-northd-0\" (UID: \"2bbf81f1-6733-4320-9bf5-c73a00d9051a\") " pod="openstack/ovn-northd-0" Dec 13 07:10:53 crc kubenswrapper[4971]: I1213 07:10:53.696108 4971 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-46dxf\" (UniqueName: \"kubernetes.io/projected/2bbf81f1-6733-4320-9bf5-c73a00d9051a-kube-api-access-46dxf\") pod \"ovn-northd-0\" (UID: \"2bbf81f1-6733-4320-9bf5-c73a00d9051a\") " pod="openstack/ovn-northd-0" Dec 13 07:10:53 crc kubenswrapper[4971]: I1213 07:10:53.696144 4971 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/2bbf81f1-6733-4320-9bf5-c73a00d9051a-combined-ca-bundle\") pod \"ovn-northd-0\" (UID: \"2bbf81f1-6733-4320-9bf5-c73a00d9051a\") " pod="openstack/ovn-northd-0" Dec 13 07:10:53 crc kubenswrapper[4971]: I1213 07:10:53.696174 4971 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/82d7cf8c-129a-483a-b190-8480e2f03418-dns-svc\") pod \"dnsmasq-dns-757dc6fff9-hfmv4\" (UID: \"82d7cf8c-129a-483a-b190-8480e2f03418\") " pod="openstack/dnsmasq-dns-757dc6fff9-hfmv4" Dec 13 07:10:53 crc kubenswrapper[4971]: I1213 07:10:53.696221 4971 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-jjxrl\" (UniqueName: \"kubernetes.io/projected/82d7cf8c-129a-483a-b190-8480e2f03418-kube-api-access-jjxrl\") pod \"dnsmasq-dns-757dc6fff9-hfmv4\" (UID: \"82d7cf8c-129a-483a-b190-8480e2f03418\") " pod="openstack/dnsmasq-dns-757dc6fff9-hfmv4" Dec 13 07:10:53 crc kubenswrapper[4971]: I1213 07:10:53.696260 4971 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/82d7cf8c-129a-483a-b190-8480e2f03418-ovsdbserver-sb\") pod \"dnsmasq-dns-757dc6fff9-hfmv4\" (UID: \"82d7cf8c-129a-483a-b190-8480e2f03418\") " pod="openstack/dnsmasq-dns-757dc6fff9-hfmv4" Dec 13 07:10:53 crc kubenswrapper[4971]: I1213 07:10:53.696299 4971 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/2bbf81f1-6733-4320-9bf5-c73a00d9051a-scripts\") pod \"ovn-northd-0\" (UID: \"2bbf81f1-6733-4320-9bf5-c73a00d9051a\") " pod="openstack/ovn-northd-0" Dec 13 07:10:53 crc kubenswrapper[4971]: I1213 07:10:53.696381 4971 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/2bbf81f1-6733-4320-9bf5-c73a00d9051a-config\") pod \"ovn-northd-0\" (UID: \"2bbf81f1-6733-4320-9bf5-c73a00d9051a\") " pod="openstack/ovn-northd-0" Dec 13 07:10:53 crc kubenswrapper[4971]: I1213 07:10:53.696415 4971 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovn-rundir\" (UniqueName: \"kubernetes.io/empty-dir/2bbf81f1-6733-4320-9bf5-c73a00d9051a-ovn-rundir\") pod \"ovn-northd-0\" (UID: \"2bbf81f1-6733-4320-9bf5-c73a00d9051a\") " pod="openstack/ovn-northd-0" Dec 13 07:10:53 crc kubenswrapper[4971]: I1213 07:10:53.700741 4971 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovn-northd-tls-certs\" (UniqueName: \"kubernetes.io/secret/2bbf81f1-6733-4320-9bf5-c73a00d9051a-ovn-northd-tls-certs\") pod \"ovn-northd-0\" (UID: \"2bbf81f1-6733-4320-9bf5-c73a00d9051a\") " pod="openstack/ovn-northd-0" Dec 13 07:10:53 crc kubenswrapper[4971]: I1213 07:10:53.700827 4971 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/2bbf81f1-6733-4320-9bf5-c73a00d9051a-scripts\") pod \"ovn-northd-0\" (UID: \"2bbf81f1-6733-4320-9bf5-c73a00d9051a\") " pod="openstack/ovn-northd-0" Dec 13 07:10:53 crc kubenswrapper[4971]: I1213 07:10:53.701729 4971 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovn-rundir\" (UniqueName: \"kubernetes.io/empty-dir/2bbf81f1-6733-4320-9bf5-c73a00d9051a-ovn-rundir\") pod \"ovn-northd-0\" (UID: \"2bbf81f1-6733-4320-9bf5-c73a00d9051a\") " pod="openstack/ovn-northd-0" Dec 13 07:10:53 crc kubenswrapper[4971]: I1213 07:10:53.701738 4971 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/2bbf81f1-6733-4320-9bf5-c73a00d9051a-config\") pod \"ovn-northd-0\" (UID: \"2bbf81f1-6733-4320-9bf5-c73a00d9051a\") " pod="openstack/ovn-northd-0" Dec 13 07:10:53 crc kubenswrapper[4971]: I1213 07:10:53.702269 4971 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metrics-certs-tls-certs\" (UniqueName: \"kubernetes.io/secret/2bbf81f1-6733-4320-9bf5-c73a00d9051a-metrics-certs-tls-certs\") pod \"ovn-northd-0\" (UID: \"2bbf81f1-6733-4320-9bf5-c73a00d9051a\") " pod="openstack/ovn-northd-0" Dec 13 07:10:53 crc kubenswrapper[4971]: I1213 07:10:53.702469 4971 generic.go:334] "Generic (PLEG): container finished" podID="9a37323b-7693-41c2-8a8f-661291d1ef5e" containerID="a186cbeae5c8e89754919777631db2a4402c1ae0df5cae4031f8d0badb565448" exitCode=0 Dec 13 07:10:53 crc kubenswrapper[4971]: I1213 07:10:53.702623 4971 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-95f5f6995-qv9tw" event={"ID":"9a37323b-7693-41c2-8a8f-661291d1ef5e","Type":"ContainerDied","Data":"a186cbeae5c8e89754919777631db2a4402c1ae0df5cae4031f8d0badb565448"} Dec 13 07:10:53 crc kubenswrapper[4971]: I1213 07:10:53.703012 4971 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/2bbf81f1-6733-4320-9bf5-c73a00d9051a-combined-ca-bundle\") pod \"ovn-northd-0\" (UID: \"2bbf81f1-6733-4320-9bf5-c73a00d9051a\") " pod="openstack/ovn-northd-0" Dec 13 07:10:53 crc kubenswrapper[4971]: I1213 07:10:53.715640 4971 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-744ffd65bc-zhplm" event={"ID":"b4e999eb-ff02-4c70-b755-f85bd90dd637","Type":"ContainerStarted","Data":"2cc49c508c149e239b081c227ba18fe2e58780d731aaea269a8eb7b2cdf21196"} Dec 13 07:10:53 crc kubenswrapper[4971]: I1213 07:10:53.716203 4971 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/dnsmasq-dns-744ffd65bc-zhplm" podUID="b4e999eb-ff02-4c70-b755-f85bd90dd637" containerName="dnsmasq-dns" containerID="cri-o://2cc49c508c149e239b081c227ba18fe2e58780d731aaea269a8eb7b2cdf21196" gracePeriod=10 Dec 13 07:10:53 crc kubenswrapper[4971]: I1213 07:10:53.716283 4971 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/dnsmasq-dns-744ffd65bc-zhplm" Dec 13 07:10:53 crc kubenswrapper[4971]: I1213 07:10:53.724382 4971 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-46dxf\" (UniqueName: \"kubernetes.io/projected/2bbf81f1-6733-4320-9bf5-c73a00d9051a-kube-api-access-46dxf\") pod \"ovn-northd-0\" (UID: \"2bbf81f1-6733-4320-9bf5-c73a00d9051a\") " pod="openstack/ovn-northd-0" Dec 13 07:10:53 crc kubenswrapper[4971]: I1213 07:10:53.752132 4971 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/dnsmasq-dns-744ffd65bc-zhplm" podStartSLOduration=2.945370242 podStartE2EDuration="49.752114121s" podCreationTimestamp="2025-12-13 07:10:04 +0000 UTC" firstStartedPulling="2025-12-13 07:10:05.437189 +0000 UTC m=+1262.041598448" lastFinishedPulling="2025-12-13 07:10:52.243932879 +0000 UTC m=+1308.848342327" observedRunningTime="2025-12-13 07:10:53.738499415 +0000 UTC m=+1310.342908863" watchObservedRunningTime="2025-12-13 07:10:53.752114121 +0000 UTC m=+1310.356523569" Dec 13 07:10:53 crc kubenswrapper[4971]: I1213 07:10:53.801705 4971 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/82d7cf8c-129a-483a-b190-8480e2f03418-dns-svc\") pod \"dnsmasq-dns-757dc6fff9-hfmv4\" (UID: \"82d7cf8c-129a-483a-b190-8480e2f03418\") " pod="openstack/dnsmasq-dns-757dc6fff9-hfmv4" Dec 13 07:10:53 crc kubenswrapper[4971]: I1213 07:10:53.801769 4971 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-jjxrl\" (UniqueName: \"kubernetes.io/projected/82d7cf8c-129a-483a-b190-8480e2f03418-kube-api-access-jjxrl\") pod \"dnsmasq-dns-757dc6fff9-hfmv4\" (UID: \"82d7cf8c-129a-483a-b190-8480e2f03418\") " pod="openstack/dnsmasq-dns-757dc6fff9-hfmv4" Dec 13 07:10:53 crc kubenswrapper[4971]: I1213 07:10:53.801805 4971 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/82d7cf8c-129a-483a-b190-8480e2f03418-ovsdbserver-sb\") pod \"dnsmasq-dns-757dc6fff9-hfmv4\" (UID: \"82d7cf8c-129a-483a-b190-8480e2f03418\") " pod="openstack/dnsmasq-dns-757dc6fff9-hfmv4" Dec 13 07:10:53 crc kubenswrapper[4971]: I1213 07:10:53.801930 4971 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/82d7cf8c-129a-483a-b190-8480e2f03418-config\") pod \"dnsmasq-dns-757dc6fff9-hfmv4\" (UID: \"82d7cf8c-129a-483a-b190-8480e2f03418\") " pod="openstack/dnsmasq-dns-757dc6fff9-hfmv4" Dec 13 07:10:53 crc kubenswrapper[4971]: I1213 07:10:53.802010 4971 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/82d7cf8c-129a-483a-b190-8480e2f03418-ovsdbserver-nb\") pod \"dnsmasq-dns-757dc6fff9-hfmv4\" (UID: \"82d7cf8c-129a-483a-b190-8480e2f03418\") " pod="openstack/dnsmasq-dns-757dc6fff9-hfmv4" Dec 13 07:10:53 crc kubenswrapper[4971]: I1213 07:10:53.803063 4971 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/82d7cf8c-129a-483a-b190-8480e2f03418-ovsdbserver-nb\") pod \"dnsmasq-dns-757dc6fff9-hfmv4\" (UID: \"82d7cf8c-129a-483a-b190-8480e2f03418\") " pod="openstack/dnsmasq-dns-757dc6fff9-hfmv4" Dec 13 07:10:53 crc kubenswrapper[4971]: I1213 07:10:53.803409 4971 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/82d7cf8c-129a-483a-b190-8480e2f03418-dns-svc\") pod \"dnsmasq-dns-757dc6fff9-hfmv4\" (UID: \"82d7cf8c-129a-483a-b190-8480e2f03418\") " pod="openstack/dnsmasq-dns-757dc6fff9-hfmv4" Dec 13 07:10:53 crc kubenswrapper[4971]: I1213 07:10:53.804509 4971 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/82d7cf8c-129a-483a-b190-8480e2f03418-ovsdbserver-sb\") pod \"dnsmasq-dns-757dc6fff9-hfmv4\" (UID: \"82d7cf8c-129a-483a-b190-8480e2f03418\") " pod="openstack/dnsmasq-dns-757dc6fff9-hfmv4" Dec 13 07:10:53 crc kubenswrapper[4971]: I1213 07:10:53.810821 4971 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-northd-0" Dec 13 07:10:53 crc kubenswrapper[4971]: I1213 07:10:53.816364 4971 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/82d7cf8c-129a-483a-b190-8480e2f03418-config\") pod \"dnsmasq-dns-757dc6fff9-hfmv4\" (UID: \"82d7cf8c-129a-483a-b190-8480e2f03418\") " pod="openstack/dnsmasq-dns-757dc6fff9-hfmv4" Dec 13 07:10:53 crc kubenswrapper[4971]: I1213 07:10:53.841919 4971 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-jjxrl\" (UniqueName: \"kubernetes.io/projected/82d7cf8c-129a-483a-b190-8480e2f03418-kube-api-access-jjxrl\") pod \"dnsmasq-dns-757dc6fff9-hfmv4\" (UID: \"82d7cf8c-129a-483a-b190-8480e2f03418\") " pod="openstack/dnsmasq-dns-757dc6fff9-hfmv4" Dec 13 07:10:53 crc kubenswrapper[4971]: I1213 07:10:53.988359 4971 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-757dc6fff9-hfmv4" Dec 13 07:10:54 crc kubenswrapper[4971]: I1213 07:10:54.357561 4971 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-95f5f6995-qv9tw" Dec 13 07:10:54 crc kubenswrapper[4971]: I1213 07:10:54.469745 4971 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-pszmq\" (UniqueName: \"kubernetes.io/projected/9a37323b-7693-41c2-8a8f-661291d1ef5e-kube-api-access-pszmq\") pod \"9a37323b-7693-41c2-8a8f-661291d1ef5e\" (UID: \"9a37323b-7693-41c2-8a8f-661291d1ef5e\") " Dec 13 07:10:54 crc kubenswrapper[4971]: I1213 07:10:54.470162 4971 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/9a37323b-7693-41c2-8a8f-661291d1ef5e-dns-svc\") pod \"9a37323b-7693-41c2-8a8f-661291d1ef5e\" (UID: \"9a37323b-7693-41c2-8a8f-661291d1ef5e\") " Dec 13 07:10:54 crc kubenswrapper[4971]: I1213 07:10:54.470241 4971 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/9a37323b-7693-41c2-8a8f-661291d1ef5e-config\") pod \"9a37323b-7693-41c2-8a8f-661291d1ef5e\" (UID: \"9a37323b-7693-41c2-8a8f-661291d1ef5e\") " Dec 13 07:10:54 crc kubenswrapper[4971]: I1213 07:10:54.475133 4971 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/9a37323b-7693-41c2-8a8f-661291d1ef5e-kube-api-access-pszmq" (OuterVolumeSpecName: "kube-api-access-pszmq") pod "9a37323b-7693-41c2-8a8f-661291d1ef5e" (UID: "9a37323b-7693-41c2-8a8f-661291d1ef5e"). InnerVolumeSpecName "kube-api-access-pszmq". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 13 07:10:54 crc kubenswrapper[4971]: I1213 07:10:54.489488 4971 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-794868bd45-t92j4"] Dec 13 07:10:54 crc kubenswrapper[4971]: I1213 07:10:54.490396 4971 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/9a37323b-7693-41c2-8a8f-661291d1ef5e-dns-svc" (OuterVolumeSpecName: "dns-svc") pod "9a37323b-7693-41c2-8a8f-661291d1ef5e" (UID: "9a37323b-7693-41c2-8a8f-661291d1ef5e"). InnerVolumeSpecName "dns-svc". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 13 07:10:54 crc kubenswrapper[4971]: I1213 07:10:54.506109 4971 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/9a37323b-7693-41c2-8a8f-661291d1ef5e-config" (OuterVolumeSpecName: "config") pod "9a37323b-7693-41c2-8a8f-661291d1ef5e" (UID: "9a37323b-7693-41c2-8a8f-661291d1ef5e"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 13 07:10:54 crc kubenswrapper[4971]: I1213 07:10:54.640772 4971 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-pszmq\" (UniqueName: \"kubernetes.io/projected/9a37323b-7693-41c2-8a8f-661291d1ef5e-kube-api-access-pszmq\") on node \"crc\" DevicePath \"\"" Dec 13 07:10:54 crc kubenswrapper[4971]: I1213 07:10:54.640813 4971 reconciler_common.go:293] "Volume detached for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/9a37323b-7693-41c2-8a8f-661291d1ef5e-dns-svc\") on node \"crc\" DevicePath \"\"" Dec 13 07:10:54 crc kubenswrapper[4971]: I1213 07:10:54.640826 4971 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/9a37323b-7693-41c2-8a8f-661291d1ef5e-config\") on node \"crc\" DevicePath \"\"" Dec 13 07:10:54 crc kubenswrapper[4971]: I1213 07:10:54.837247 4971 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-794868bd45-t92j4" event={"ID":"20e72d59-fbc4-4eca-84d0-04c833aa4aab","Type":"ContainerStarted","Data":"4dd2d455623b372f86f60f01914be2533930044fea79e6269ca1412f898b2e9c"} Dec 13 07:10:54 crc kubenswrapper[4971]: I1213 07:10:54.849793 4971 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/memcached-0" Dec 13 07:10:54 crc kubenswrapper[4971]: I1213 07:10:54.867924 4971 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-95f5f6995-qv9tw" event={"ID":"9a37323b-7693-41c2-8a8f-661291d1ef5e","Type":"ContainerDied","Data":"85cf34c1f99a8f5cd4f170ba6f7da04e1970efebd7348563a6dc2bfe6bf6b2f9"} Dec 13 07:10:54 crc kubenswrapper[4971]: I1213 07:10:54.867984 4971 scope.go:117] "RemoveContainer" containerID="a186cbeae5c8e89754919777631db2a4402c1ae0df5cae4031f8d0badb565448" Dec 13 07:10:54 crc kubenswrapper[4971]: I1213 07:10:54.868178 4971 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-95f5f6995-qv9tw" Dec 13 07:10:54 crc kubenswrapper[4971]: I1213 07:10:54.889194 4971 generic.go:334] "Generic (PLEG): container finished" podID="b4e999eb-ff02-4c70-b755-f85bd90dd637" containerID="2cc49c508c149e239b081c227ba18fe2e58780d731aaea269a8eb7b2cdf21196" exitCode=0 Dec 13 07:10:54 crc kubenswrapper[4971]: I1213 07:10:54.890657 4971 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-757dc6fff9-hfmv4"] Dec 13 07:10:54 crc kubenswrapper[4971]: I1213 07:10:54.890697 4971 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-744ffd65bc-zhplm" event={"ID":"b4e999eb-ff02-4c70-b755-f85bd90dd637","Type":"ContainerDied","Data":"2cc49c508c149e239b081c227ba18fe2e58780d731aaea269a8eb7b2cdf21196"} Dec 13 07:10:54 crc kubenswrapper[4971]: I1213 07:10:54.936847 4971 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovn-northd-0"] Dec 13 07:10:54 crc kubenswrapper[4971]: I1213 07:10:54.965614 4971 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovn-controller-metrics-r2cbj"] Dec 13 07:10:55 crc kubenswrapper[4971]: I1213 07:10:55.046278 4971 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-744ffd65bc-zhplm" Dec 13 07:10:55 crc kubenswrapper[4971]: I1213 07:10:55.058427 4971 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-95f5f6995-qv9tw"] Dec 13 07:10:55 crc kubenswrapper[4971]: I1213 07:10:55.098840 4971 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-95f5f6995-qv9tw"] Dec 13 07:10:55 crc kubenswrapper[4971]: I1213 07:10:55.193445 4971 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/glance-db-create-9wmnw"] Dec 13 07:10:55 crc kubenswrapper[4971]: E1213 07:10:55.193934 4971 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="b4e999eb-ff02-4c70-b755-f85bd90dd637" containerName="dnsmasq-dns" Dec 13 07:10:55 crc kubenswrapper[4971]: I1213 07:10:55.193968 4971 state_mem.go:107] "Deleted CPUSet assignment" podUID="b4e999eb-ff02-4c70-b755-f85bd90dd637" containerName="dnsmasq-dns" Dec 13 07:10:55 crc kubenswrapper[4971]: E1213 07:10:55.194016 4971 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="b4e999eb-ff02-4c70-b755-f85bd90dd637" containerName="init" Dec 13 07:10:55 crc kubenswrapper[4971]: I1213 07:10:55.194026 4971 state_mem.go:107] "Deleted CPUSet assignment" podUID="b4e999eb-ff02-4c70-b755-f85bd90dd637" containerName="init" Dec 13 07:10:55 crc kubenswrapper[4971]: E1213 07:10:55.194088 4971 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="9a37323b-7693-41c2-8a8f-661291d1ef5e" containerName="init" Dec 13 07:10:55 crc kubenswrapper[4971]: I1213 07:10:55.194098 4971 state_mem.go:107] "Deleted CPUSet assignment" podUID="9a37323b-7693-41c2-8a8f-661291d1ef5e" containerName="init" Dec 13 07:10:55 crc kubenswrapper[4971]: I1213 07:10:55.194287 4971 memory_manager.go:354] "RemoveStaleState removing state" podUID="9a37323b-7693-41c2-8a8f-661291d1ef5e" containerName="init" Dec 13 07:10:55 crc kubenswrapper[4971]: I1213 07:10:55.194319 4971 memory_manager.go:354] "RemoveStaleState removing state" podUID="b4e999eb-ff02-4c70-b755-f85bd90dd637" containerName="dnsmasq-dns" Dec 13 07:10:55 crc kubenswrapper[4971]: I1213 07:10:55.194999 4971 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-db-create-9wmnw" Dec 13 07:10:55 crc kubenswrapper[4971]: I1213 07:10:55.205505 4971 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/glance-a8e5-account-create-update-7xfzr"] Dec 13 07:10:55 crc kubenswrapper[4971]: I1213 07:10:55.206598 4971 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-a8e5-account-create-update-7xfzr" Dec 13 07:10:55 crc kubenswrapper[4971]: I1213 07:10:55.210944 4971 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"glance-db-secret" Dec 13 07:10:55 crc kubenswrapper[4971]: I1213 07:10:55.226261 4971 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/b4e999eb-ff02-4c70-b755-f85bd90dd637-dns-svc\") pod \"b4e999eb-ff02-4c70-b755-f85bd90dd637\" (UID: \"b4e999eb-ff02-4c70-b755-f85bd90dd637\") " Dec 13 07:10:55 crc kubenswrapper[4971]: I1213 07:10:55.226369 4971 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/b4e999eb-ff02-4c70-b755-f85bd90dd637-config\") pod \"b4e999eb-ff02-4c70-b755-f85bd90dd637\" (UID: \"b4e999eb-ff02-4c70-b755-f85bd90dd637\") " Dec 13 07:10:55 crc kubenswrapper[4971]: I1213 07:10:55.226497 4971 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-5zfr7\" (UniqueName: \"kubernetes.io/projected/b4e999eb-ff02-4c70-b755-f85bd90dd637-kube-api-access-5zfr7\") pod \"b4e999eb-ff02-4c70-b755-f85bd90dd637\" (UID: \"b4e999eb-ff02-4c70-b755-f85bd90dd637\") " Dec 13 07:10:55 crc kubenswrapper[4971]: I1213 07:10:55.289762 4971 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/b4e999eb-ff02-4c70-b755-f85bd90dd637-kube-api-access-5zfr7" (OuterVolumeSpecName: "kube-api-access-5zfr7") pod "b4e999eb-ff02-4c70-b755-f85bd90dd637" (UID: "b4e999eb-ff02-4c70-b755-f85bd90dd637"). InnerVolumeSpecName "kube-api-access-5zfr7". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 13 07:10:55 crc kubenswrapper[4971]: I1213 07:10:55.290217 4971 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-db-create-9wmnw"] Dec 13 07:10:55 crc kubenswrapper[4971]: I1213 07:10:55.294816 4971 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-a8e5-account-create-update-7xfzr"] Dec 13 07:10:55 crc kubenswrapper[4971]: I1213 07:10:55.328503 4971 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-qrqzd\" (UniqueName: \"kubernetes.io/projected/ba68c497-330b-4e22-8ecd-d84ee54815d0-kube-api-access-qrqzd\") pod \"glance-db-create-9wmnw\" (UID: \"ba68c497-330b-4e22-8ecd-d84ee54815d0\") " pod="openstack/glance-db-create-9wmnw" Dec 13 07:10:55 crc kubenswrapper[4971]: I1213 07:10:55.328893 4971 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-zzwqb\" (UniqueName: \"kubernetes.io/projected/00df1bf7-20e4-485c-b1c2-6f00df8379ff-kube-api-access-zzwqb\") pod \"glance-a8e5-account-create-update-7xfzr\" (UID: \"00df1bf7-20e4-485c-b1c2-6f00df8379ff\") " pod="openstack/glance-a8e5-account-create-update-7xfzr" Dec 13 07:10:55 crc kubenswrapper[4971]: I1213 07:10:55.328932 4971 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/00df1bf7-20e4-485c-b1c2-6f00df8379ff-operator-scripts\") pod \"glance-a8e5-account-create-update-7xfzr\" (UID: \"00df1bf7-20e4-485c-b1c2-6f00df8379ff\") " pod="openstack/glance-a8e5-account-create-update-7xfzr" Dec 13 07:10:55 crc kubenswrapper[4971]: I1213 07:10:55.328977 4971 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/ba68c497-330b-4e22-8ecd-d84ee54815d0-operator-scripts\") pod \"glance-db-create-9wmnw\" (UID: \"ba68c497-330b-4e22-8ecd-d84ee54815d0\") " pod="openstack/glance-db-create-9wmnw" Dec 13 07:10:55 crc kubenswrapper[4971]: I1213 07:10:55.329023 4971 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-5zfr7\" (UniqueName: \"kubernetes.io/projected/b4e999eb-ff02-4c70-b755-f85bd90dd637-kube-api-access-5zfr7\") on node \"crc\" DevicePath \"\"" Dec 13 07:10:55 crc kubenswrapper[4971]: I1213 07:10:55.360121 4971 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/b4e999eb-ff02-4c70-b755-f85bd90dd637-config" (OuterVolumeSpecName: "config") pod "b4e999eb-ff02-4c70-b755-f85bd90dd637" (UID: "b4e999eb-ff02-4c70-b755-f85bd90dd637"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 13 07:10:55 crc kubenswrapper[4971]: I1213 07:10:55.371994 4971 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/b4e999eb-ff02-4c70-b755-f85bd90dd637-dns-svc" (OuterVolumeSpecName: "dns-svc") pod "b4e999eb-ff02-4c70-b755-f85bd90dd637" (UID: "b4e999eb-ff02-4c70-b755-f85bd90dd637"). InnerVolumeSpecName "dns-svc". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 13 07:10:55 crc kubenswrapper[4971]: I1213 07:10:55.430773 4971 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-qrqzd\" (UniqueName: \"kubernetes.io/projected/ba68c497-330b-4e22-8ecd-d84ee54815d0-kube-api-access-qrqzd\") pod \"glance-db-create-9wmnw\" (UID: \"ba68c497-330b-4e22-8ecd-d84ee54815d0\") " pod="openstack/glance-db-create-9wmnw" Dec 13 07:10:55 crc kubenswrapper[4971]: I1213 07:10:55.430881 4971 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-zzwqb\" (UniqueName: \"kubernetes.io/projected/00df1bf7-20e4-485c-b1c2-6f00df8379ff-kube-api-access-zzwqb\") pod \"glance-a8e5-account-create-update-7xfzr\" (UID: \"00df1bf7-20e4-485c-b1c2-6f00df8379ff\") " pod="openstack/glance-a8e5-account-create-update-7xfzr" Dec 13 07:10:55 crc kubenswrapper[4971]: I1213 07:10:55.430923 4971 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/00df1bf7-20e4-485c-b1c2-6f00df8379ff-operator-scripts\") pod \"glance-a8e5-account-create-update-7xfzr\" (UID: \"00df1bf7-20e4-485c-b1c2-6f00df8379ff\") " pod="openstack/glance-a8e5-account-create-update-7xfzr" Dec 13 07:10:55 crc kubenswrapper[4971]: I1213 07:10:55.430965 4971 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/ba68c497-330b-4e22-8ecd-d84ee54815d0-operator-scripts\") pod \"glance-db-create-9wmnw\" (UID: \"ba68c497-330b-4e22-8ecd-d84ee54815d0\") " pod="openstack/glance-db-create-9wmnw" Dec 13 07:10:55 crc kubenswrapper[4971]: I1213 07:10:55.431049 4971 reconciler_common.go:293] "Volume detached for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/b4e999eb-ff02-4c70-b755-f85bd90dd637-dns-svc\") on node \"crc\" DevicePath \"\"" Dec 13 07:10:55 crc kubenswrapper[4971]: I1213 07:10:55.431063 4971 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/b4e999eb-ff02-4c70-b755-f85bd90dd637-config\") on node \"crc\" DevicePath \"\"" Dec 13 07:10:55 crc kubenswrapper[4971]: I1213 07:10:55.431823 4971 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/ba68c497-330b-4e22-8ecd-d84ee54815d0-operator-scripts\") pod \"glance-db-create-9wmnw\" (UID: \"ba68c497-330b-4e22-8ecd-d84ee54815d0\") " pod="openstack/glance-db-create-9wmnw" Dec 13 07:10:55 crc kubenswrapper[4971]: I1213 07:10:55.432259 4971 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/00df1bf7-20e4-485c-b1c2-6f00df8379ff-operator-scripts\") pod \"glance-a8e5-account-create-update-7xfzr\" (UID: \"00df1bf7-20e4-485c-b1c2-6f00df8379ff\") " pod="openstack/glance-a8e5-account-create-update-7xfzr" Dec 13 07:10:55 crc kubenswrapper[4971]: I1213 07:10:55.449094 4971 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-qrqzd\" (UniqueName: \"kubernetes.io/projected/ba68c497-330b-4e22-8ecd-d84ee54815d0-kube-api-access-qrqzd\") pod \"glance-db-create-9wmnw\" (UID: \"ba68c497-330b-4e22-8ecd-d84ee54815d0\") " pod="openstack/glance-db-create-9wmnw" Dec 13 07:10:55 crc kubenswrapper[4971]: I1213 07:10:55.451629 4971 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-zzwqb\" (UniqueName: \"kubernetes.io/projected/00df1bf7-20e4-485c-b1c2-6f00df8379ff-kube-api-access-zzwqb\") pod \"glance-a8e5-account-create-update-7xfzr\" (UID: \"00df1bf7-20e4-485c-b1c2-6f00df8379ff\") " pod="openstack/glance-a8e5-account-create-update-7xfzr" Dec 13 07:10:55 crc kubenswrapper[4971]: I1213 07:10:55.519987 4971 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-db-create-9wmnw" Dec 13 07:10:55 crc kubenswrapper[4971]: I1213 07:10:55.748688 4971 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-a8e5-account-create-update-7xfzr" Dec 13 07:10:55 crc kubenswrapper[4971]: I1213 07:10:55.835746 4971 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="9a37323b-7693-41c2-8a8f-661291d1ef5e" path="/var/lib/kubelet/pods/9a37323b-7693-41c2-8a8f-661291d1ef5e/volumes" Dec 13 07:10:56 crc kubenswrapper[4971]: I1213 07:10:56.186021 4971 generic.go:334] "Generic (PLEG): container finished" podID="82d7cf8c-129a-483a-b190-8480e2f03418" containerID="8a0537f444717cb3bdb32c5612b6fec9aea5fc7a82996f56458c25b1ba0630a0" exitCode=0 Dec 13 07:10:56 crc kubenswrapper[4971]: I1213 07:10:56.188027 4971 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-757dc6fff9-hfmv4" event={"ID":"82d7cf8c-129a-483a-b190-8480e2f03418","Type":"ContainerDied","Data":"8a0537f444717cb3bdb32c5612b6fec9aea5fc7a82996f56458c25b1ba0630a0"} Dec 13 07:10:56 crc kubenswrapper[4971]: I1213 07:10:56.188068 4971 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-757dc6fff9-hfmv4" event={"ID":"82d7cf8c-129a-483a-b190-8480e2f03418","Type":"ContainerStarted","Data":"a9a374a27e1ac718e356aee5609c2293261a89050a92b0a3ba7f122c99934131"} Dec 13 07:10:56 crc kubenswrapper[4971]: I1213 07:10:56.206026 4971 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-controller-metrics-r2cbj" event={"ID":"fa04788e-0c35-4205-97ed-cb1e38ec9e69","Type":"ContainerStarted","Data":"f92e0cab62cc91c1b8305fece48e2b5614c0c041321214e3adc5d4987e21b463"} Dec 13 07:10:56 crc kubenswrapper[4971]: I1213 07:10:56.206080 4971 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-controller-metrics-r2cbj" event={"ID":"fa04788e-0c35-4205-97ed-cb1e38ec9e69","Type":"ContainerStarted","Data":"64d287fcd56b80b6ae7d55c30d7af1f81f14888b41d4f0bd872066767402576b"} Dec 13 07:10:56 crc kubenswrapper[4971]: I1213 07:10:56.235577 4971 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-744ffd65bc-zhplm" event={"ID":"b4e999eb-ff02-4c70-b755-f85bd90dd637","Type":"ContainerDied","Data":"fad71e9c347b23db5981bdc894b841c16cb490dcff033245da131c2ca048068a"} Dec 13 07:10:56 crc kubenswrapper[4971]: I1213 07:10:56.235639 4971 scope.go:117] "RemoveContainer" containerID="2cc49c508c149e239b081c227ba18fe2e58780d731aaea269a8eb7b2cdf21196" Dec 13 07:10:56 crc kubenswrapper[4971]: I1213 07:10:56.235823 4971 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-744ffd65bc-zhplm" Dec 13 07:10:56 crc kubenswrapper[4971]: I1213 07:10:56.242172 4971 generic.go:334] "Generic (PLEG): container finished" podID="20e72d59-fbc4-4eca-84d0-04c833aa4aab" containerID="8c57fbe679c7ff81971e834c6f85fd7c746e08bd5a9c82a44f8819f17ec47c8f" exitCode=0 Dec 13 07:10:56 crc kubenswrapper[4971]: I1213 07:10:56.242350 4971 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-794868bd45-t92j4" event={"ID":"20e72d59-fbc4-4eca-84d0-04c833aa4aab","Type":"ContainerDied","Data":"8c57fbe679c7ff81971e834c6f85fd7c746e08bd5a9c82a44f8819f17ec47c8f"} Dec 13 07:10:56 crc kubenswrapper[4971]: I1213 07:10:56.250495 4971 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-db-create-9wmnw"] Dec 13 07:10:56 crc kubenswrapper[4971]: I1213 07:10:56.269268 4971 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-northd-0" event={"ID":"2bbf81f1-6733-4320-9bf5-c73a00d9051a","Type":"ContainerStarted","Data":"21f350aac1b195063ff0bcee67634ef5440778d25424cc5deae81153c473d16f"} Dec 13 07:10:56 crc kubenswrapper[4971]: I1213 07:10:56.294879 4971 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/ovn-controller-metrics-r2cbj" podStartSLOduration=3.294856482 podStartE2EDuration="3.294856482s" podCreationTimestamp="2025-12-13 07:10:53 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-13 07:10:56.262979987 +0000 UTC m=+1312.867389435" watchObservedRunningTime="2025-12-13 07:10:56.294856482 +0000 UTC m=+1312.899265930" Dec 13 07:10:56 crc kubenswrapper[4971]: I1213 07:10:56.365878 4971 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-744ffd65bc-zhplm"] Dec 13 07:10:56 crc kubenswrapper[4971]: I1213 07:10:56.381099 4971 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-744ffd65bc-zhplm"] Dec 13 07:10:56 crc kubenswrapper[4971]: I1213 07:10:56.411958 4971 scope.go:117] "RemoveContainer" containerID="6f52043e33e0ed905f8dc75a68ad19b0628da94d63ca3eaec2836a6dbde42bda" Dec 13 07:10:56 crc kubenswrapper[4971]: I1213 07:10:56.641225 4971 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-a8e5-account-create-update-7xfzr"] Dec 13 07:10:56 crc kubenswrapper[4971]: W1213 07:10:56.660728 4971 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod00df1bf7_20e4_485c_b1c2_6f00df8379ff.slice/crio-46ee4f51ad43c512138ab8be4c7389fe370875287a2f6cd1b261a6d60607dd16 WatchSource:0}: Error finding container 46ee4f51ad43c512138ab8be4c7389fe370875287a2f6cd1b261a6d60607dd16: Status 404 returned error can't find the container with id 46ee4f51ad43c512138ab8be4c7389fe370875287a2f6cd1b261a6d60607dd16 Dec 13 07:10:57 crc kubenswrapper[4971]: I1213 07:10:57.327157 4971 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-db-create-9wmnw" event={"ID":"ba68c497-330b-4e22-8ecd-d84ee54815d0","Type":"ContainerStarted","Data":"a26e3edab88870893204ebdf9829573548f908e202436604df866f2e829e670c"} Dec 13 07:10:57 crc kubenswrapper[4971]: I1213 07:10:57.327497 4971 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-db-create-9wmnw" event={"ID":"ba68c497-330b-4e22-8ecd-d84ee54815d0","Type":"ContainerStarted","Data":"33b3bf0c5f642ba4386a2a7bd46ff2a54f43dc7d87e1ee80f9db4ccbf2b84d38"} Dec 13 07:10:57 crc kubenswrapper[4971]: I1213 07:10:57.331264 4971 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-757dc6fff9-hfmv4" event={"ID":"82d7cf8c-129a-483a-b190-8480e2f03418","Type":"ContainerStarted","Data":"45a3b8435a1a4bc04f956f428cca44d0c3e5b295322aff907527d681ce5c1f1a"} Dec 13 07:10:57 crc kubenswrapper[4971]: I1213 07:10:57.331406 4971 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/dnsmasq-dns-757dc6fff9-hfmv4" Dec 13 07:10:57 crc kubenswrapper[4971]: I1213 07:10:57.332490 4971 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-a8e5-account-create-update-7xfzr" event={"ID":"00df1bf7-20e4-485c-b1c2-6f00df8379ff","Type":"ContainerStarted","Data":"5901f03e4f1457dd53da63ac9fcae002c832ee5d16b994772d7e84ecbd94c434"} Dec 13 07:10:57 crc kubenswrapper[4971]: I1213 07:10:57.332514 4971 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-a8e5-account-create-update-7xfzr" event={"ID":"00df1bf7-20e4-485c-b1c2-6f00df8379ff","Type":"ContainerStarted","Data":"46ee4f51ad43c512138ab8be4c7389fe370875287a2f6cd1b261a6d60607dd16"} Dec 13 07:10:57 crc kubenswrapper[4971]: I1213 07:10:57.342363 4971 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-794868bd45-t92j4" event={"ID":"20e72d59-fbc4-4eca-84d0-04c833aa4aab","Type":"ContainerStarted","Data":"6f845eb132e2c1de1a88cc497e5c06a82fd897df9a62779aeccbdf2e66a02ec8"} Dec 13 07:10:57 crc kubenswrapper[4971]: I1213 07:10:57.352360 4971 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/glance-db-create-9wmnw" podStartSLOduration=2.352340967 podStartE2EDuration="2.352340967s" podCreationTimestamp="2025-12-13 07:10:55 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-13 07:10:57.350479351 +0000 UTC m=+1313.954888799" watchObservedRunningTime="2025-12-13 07:10:57.352340967 +0000 UTC m=+1313.956750415" Dec 13 07:10:57 crc kubenswrapper[4971]: I1213 07:10:57.372631 4971 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/dnsmasq-dns-794868bd45-t92j4" podStartSLOduration=5.372613256 podStartE2EDuration="5.372613256s" podCreationTimestamp="2025-12-13 07:10:52 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-13 07:10:57.370136975 +0000 UTC m=+1313.974546423" watchObservedRunningTime="2025-12-13 07:10:57.372613256 +0000 UTC m=+1313.977022704" Dec 13 07:10:57 crc kubenswrapper[4971]: I1213 07:10:57.399771 4971 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/dnsmasq-dns-757dc6fff9-hfmv4" podStartSLOduration=4.399752454 podStartE2EDuration="4.399752454s" podCreationTimestamp="2025-12-13 07:10:53 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-13 07:10:57.398462592 +0000 UTC m=+1314.002872060" watchObservedRunningTime="2025-12-13 07:10:57.399752454 +0000 UTC m=+1314.004161902" Dec 13 07:10:57 crc kubenswrapper[4971]: I1213 07:10:57.419650 4971 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/glance-a8e5-account-create-update-7xfzr" podStartSLOduration=2.419623343 podStartE2EDuration="2.419623343s" podCreationTimestamp="2025-12-13 07:10:55 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-13 07:10:57.413614556 +0000 UTC m=+1314.018024004" watchObservedRunningTime="2025-12-13 07:10:57.419623343 +0000 UTC m=+1314.024032811" Dec 13 07:10:57 crc kubenswrapper[4971]: I1213 07:10:57.779638 4971 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="b4e999eb-ff02-4c70-b755-f85bd90dd637" path="/var/lib/kubelet/pods/b4e999eb-ff02-4c70-b755-f85bd90dd637/volumes" Dec 13 07:10:58 crc kubenswrapper[4971]: I1213 07:10:58.354849 4971 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-northd-0" event={"ID":"2bbf81f1-6733-4320-9bf5-c73a00d9051a","Type":"ContainerStarted","Data":"5f731d102b23e6d2bfafdd0cd8f0015c5ae32a38dd3ae3e0fc5929d9d3fb09b5"} Dec 13 07:10:58 crc kubenswrapper[4971]: I1213 07:10:58.356832 4971 generic.go:334] "Generic (PLEG): container finished" podID="ba68c497-330b-4e22-8ecd-d84ee54815d0" containerID="a26e3edab88870893204ebdf9829573548f908e202436604df866f2e829e670c" exitCode=0 Dec 13 07:10:58 crc kubenswrapper[4971]: I1213 07:10:58.356941 4971 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-db-create-9wmnw" event={"ID":"ba68c497-330b-4e22-8ecd-d84ee54815d0","Type":"ContainerDied","Data":"a26e3edab88870893204ebdf9829573548f908e202436604df866f2e829e670c"} Dec 13 07:10:58 crc kubenswrapper[4971]: I1213 07:10:58.361008 4971 generic.go:334] "Generic (PLEG): container finished" podID="00df1bf7-20e4-485c-b1c2-6f00df8379ff" containerID="5901f03e4f1457dd53da63ac9fcae002c832ee5d16b994772d7e84ecbd94c434" exitCode=0 Dec 13 07:10:58 crc kubenswrapper[4971]: I1213 07:10:58.361116 4971 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-a8e5-account-create-update-7xfzr" event={"ID":"00df1bf7-20e4-485c-b1c2-6f00df8379ff","Type":"ContainerDied","Data":"5901f03e4f1457dd53da63ac9fcae002c832ee5d16b994772d7e84ecbd94c434"} Dec 13 07:10:58 crc kubenswrapper[4971]: I1213 07:10:58.361561 4971 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/dnsmasq-dns-794868bd45-t92j4" Dec 13 07:10:59 crc kubenswrapper[4971]: I1213 07:10:59.110673 4971 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/keystone-db-create-gzpfj"] Dec 13 07:10:59 crc kubenswrapper[4971]: I1213 07:10:59.111953 4971 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-db-create-gzpfj" Dec 13 07:10:59 crc kubenswrapper[4971]: I1213 07:10:59.127197 4971 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/keystone-db-create-gzpfj"] Dec 13 07:10:59 crc kubenswrapper[4971]: I1213 07:10:59.241745 4971 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/keystone-2ba7-account-create-update-v5s9z"] Dec 13 07:10:59 crc kubenswrapper[4971]: I1213 07:10:59.243063 4971 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-2ba7-account-create-update-v5s9z" Dec 13 07:10:59 crc kubenswrapper[4971]: I1213 07:10:59.245331 4971 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone-db-secret" Dec 13 07:10:59 crc kubenswrapper[4971]: I1213 07:10:59.248789 4971 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/keystone-2ba7-account-create-update-v5s9z"] Dec 13 07:10:59 crc kubenswrapper[4971]: I1213 07:10:59.254742 4971 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-sv8qz\" (UniqueName: \"kubernetes.io/projected/b2645870-215e-4acb-9f8c-cfa2bb0a5cf6-kube-api-access-sv8qz\") pod \"keystone-2ba7-account-create-update-v5s9z\" (UID: \"b2645870-215e-4acb-9f8c-cfa2bb0a5cf6\") " pod="openstack/keystone-2ba7-account-create-update-v5s9z" Dec 13 07:10:59 crc kubenswrapper[4971]: I1213 07:10:59.254797 4971 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-8qxrq\" (UniqueName: \"kubernetes.io/projected/e07e072c-e921-4f2d-a967-1966c35ff49b-kube-api-access-8qxrq\") pod \"keystone-db-create-gzpfj\" (UID: \"e07e072c-e921-4f2d-a967-1966c35ff49b\") " pod="openstack/keystone-db-create-gzpfj" Dec 13 07:10:59 crc kubenswrapper[4971]: I1213 07:10:59.254834 4971 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/e07e072c-e921-4f2d-a967-1966c35ff49b-operator-scripts\") pod \"keystone-db-create-gzpfj\" (UID: \"e07e072c-e921-4f2d-a967-1966c35ff49b\") " pod="openstack/keystone-db-create-gzpfj" Dec 13 07:10:59 crc kubenswrapper[4971]: I1213 07:10:59.254931 4971 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/b2645870-215e-4acb-9f8c-cfa2bb0a5cf6-operator-scripts\") pod \"keystone-2ba7-account-create-update-v5s9z\" (UID: \"b2645870-215e-4acb-9f8c-cfa2bb0a5cf6\") " pod="openstack/keystone-2ba7-account-create-update-v5s9z" Dec 13 07:10:59 crc kubenswrapper[4971]: I1213 07:10:59.355875 4971 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-sv8qz\" (UniqueName: \"kubernetes.io/projected/b2645870-215e-4acb-9f8c-cfa2bb0a5cf6-kube-api-access-sv8qz\") pod \"keystone-2ba7-account-create-update-v5s9z\" (UID: \"b2645870-215e-4acb-9f8c-cfa2bb0a5cf6\") " pod="openstack/keystone-2ba7-account-create-update-v5s9z" Dec 13 07:10:59 crc kubenswrapper[4971]: I1213 07:10:59.355939 4971 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-8qxrq\" (UniqueName: \"kubernetes.io/projected/e07e072c-e921-4f2d-a967-1966c35ff49b-kube-api-access-8qxrq\") pod \"keystone-db-create-gzpfj\" (UID: \"e07e072c-e921-4f2d-a967-1966c35ff49b\") " pod="openstack/keystone-db-create-gzpfj" Dec 13 07:10:59 crc kubenswrapper[4971]: I1213 07:10:59.355973 4971 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/e07e072c-e921-4f2d-a967-1966c35ff49b-operator-scripts\") pod \"keystone-db-create-gzpfj\" (UID: \"e07e072c-e921-4f2d-a967-1966c35ff49b\") " pod="openstack/keystone-db-create-gzpfj" Dec 13 07:10:59 crc kubenswrapper[4971]: I1213 07:10:59.356035 4971 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/b2645870-215e-4acb-9f8c-cfa2bb0a5cf6-operator-scripts\") pod \"keystone-2ba7-account-create-update-v5s9z\" (UID: \"b2645870-215e-4acb-9f8c-cfa2bb0a5cf6\") " pod="openstack/keystone-2ba7-account-create-update-v5s9z" Dec 13 07:10:59 crc kubenswrapper[4971]: I1213 07:10:59.356933 4971 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/b2645870-215e-4acb-9f8c-cfa2bb0a5cf6-operator-scripts\") pod \"keystone-2ba7-account-create-update-v5s9z\" (UID: \"b2645870-215e-4acb-9f8c-cfa2bb0a5cf6\") " pod="openstack/keystone-2ba7-account-create-update-v5s9z" Dec 13 07:10:59 crc kubenswrapper[4971]: I1213 07:10:59.358035 4971 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/e07e072c-e921-4f2d-a967-1966c35ff49b-operator-scripts\") pod \"keystone-db-create-gzpfj\" (UID: \"e07e072c-e921-4f2d-a967-1966c35ff49b\") " pod="openstack/keystone-db-create-gzpfj" Dec 13 07:10:59 crc kubenswrapper[4971]: I1213 07:10:59.371156 4971 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-northd-0" event={"ID":"2bbf81f1-6733-4320-9bf5-c73a00d9051a","Type":"ContainerStarted","Data":"f630e9d5794f95267aac564889ad67addf995113f3b603722b5331fcfceb7261"} Dec 13 07:10:59 crc kubenswrapper[4971]: I1213 07:10:59.381911 4971 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-sv8qz\" (UniqueName: \"kubernetes.io/projected/b2645870-215e-4acb-9f8c-cfa2bb0a5cf6-kube-api-access-sv8qz\") pod \"keystone-2ba7-account-create-update-v5s9z\" (UID: \"b2645870-215e-4acb-9f8c-cfa2bb0a5cf6\") " pod="openstack/keystone-2ba7-account-create-update-v5s9z" Dec 13 07:10:59 crc kubenswrapper[4971]: I1213 07:10:59.382023 4971 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-8qxrq\" (UniqueName: \"kubernetes.io/projected/e07e072c-e921-4f2d-a967-1966c35ff49b-kube-api-access-8qxrq\") pod \"keystone-db-create-gzpfj\" (UID: \"e07e072c-e921-4f2d-a967-1966c35ff49b\") " pod="openstack/keystone-db-create-gzpfj" Dec 13 07:10:59 crc kubenswrapper[4971]: I1213 07:10:59.394565 4971 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/ovn-northd-0" podStartSLOduration=3.689439166 podStartE2EDuration="6.394549285s" podCreationTimestamp="2025-12-13 07:10:53 +0000 UTC" firstStartedPulling="2025-12-13 07:10:54.929274371 +0000 UTC m=+1311.533683829" lastFinishedPulling="2025-12-13 07:10:57.6343845 +0000 UTC m=+1314.238793948" observedRunningTime="2025-12-13 07:10:59.39106797 +0000 UTC m=+1315.995477428" watchObservedRunningTime="2025-12-13 07:10:59.394549285 +0000 UTC m=+1315.998958723" Dec 13 07:10:59 crc kubenswrapper[4971]: I1213 07:10:59.734052 4971 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-db-create-gzpfj" Dec 13 07:10:59 crc kubenswrapper[4971]: I1213 07:10:59.735844 4971 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-2ba7-account-create-update-v5s9z" Dec 13 07:10:59 crc kubenswrapper[4971]: I1213 07:10:59.802994 4971 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/placement-db-create-jwvzj"] Dec 13 07:10:59 crc kubenswrapper[4971]: I1213 07:10:59.804400 4971 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/placement-db-create-jwvzj"] Dec 13 07:10:59 crc kubenswrapper[4971]: I1213 07:10:59.804498 4971 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/placement-db-create-jwvzj" Dec 13 07:10:59 crc kubenswrapper[4971]: I1213 07:10:59.837257 4971 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/d0ed4d40-1309-4641-b91c-f7dc117d6f4a-operator-scripts\") pod \"placement-db-create-jwvzj\" (UID: \"d0ed4d40-1309-4641-b91c-f7dc117d6f4a\") " pod="openstack/placement-db-create-jwvzj" Dec 13 07:10:59 crc kubenswrapper[4971]: I1213 07:10:59.837351 4971 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-gwwcf\" (UniqueName: \"kubernetes.io/projected/d0ed4d40-1309-4641-b91c-f7dc117d6f4a-kube-api-access-gwwcf\") pod \"placement-db-create-jwvzj\" (UID: \"d0ed4d40-1309-4641-b91c-f7dc117d6f4a\") " pod="openstack/placement-db-create-jwvzj" Dec 13 07:10:59 crc kubenswrapper[4971]: I1213 07:10:59.873092 4971 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/placement-e878-account-create-update-lnhvj"] Dec 13 07:10:59 crc kubenswrapper[4971]: I1213 07:10:59.874393 4971 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/placement-e878-account-create-update-lnhvj" Dec 13 07:10:59 crc kubenswrapper[4971]: I1213 07:10:59.877408 4971 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"placement-db-secret" Dec 13 07:10:59 crc kubenswrapper[4971]: I1213 07:10:59.891076 4971 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/placement-e878-account-create-update-lnhvj"] Dec 13 07:10:59 crc kubenswrapper[4971]: I1213 07:10:59.940214 4971 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-gwhdz\" (UniqueName: \"kubernetes.io/projected/292d6752-a759-4b19-9759-f0101a1b01e6-kube-api-access-gwhdz\") pod \"placement-e878-account-create-update-lnhvj\" (UID: \"292d6752-a759-4b19-9759-f0101a1b01e6\") " pod="openstack/placement-e878-account-create-update-lnhvj" Dec 13 07:10:59 crc kubenswrapper[4971]: I1213 07:10:59.940358 4971 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/d0ed4d40-1309-4641-b91c-f7dc117d6f4a-operator-scripts\") pod \"placement-db-create-jwvzj\" (UID: \"d0ed4d40-1309-4641-b91c-f7dc117d6f4a\") " pod="openstack/placement-db-create-jwvzj" Dec 13 07:10:59 crc kubenswrapper[4971]: I1213 07:10:59.940438 4971 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-gwwcf\" (UniqueName: \"kubernetes.io/projected/d0ed4d40-1309-4641-b91c-f7dc117d6f4a-kube-api-access-gwwcf\") pod \"placement-db-create-jwvzj\" (UID: \"d0ed4d40-1309-4641-b91c-f7dc117d6f4a\") " pod="openstack/placement-db-create-jwvzj" Dec 13 07:10:59 crc kubenswrapper[4971]: I1213 07:10:59.940484 4971 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/292d6752-a759-4b19-9759-f0101a1b01e6-operator-scripts\") pod \"placement-e878-account-create-update-lnhvj\" (UID: \"292d6752-a759-4b19-9759-f0101a1b01e6\") " pod="openstack/placement-e878-account-create-update-lnhvj" Dec 13 07:10:59 crc kubenswrapper[4971]: I1213 07:10:59.941434 4971 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/d0ed4d40-1309-4641-b91c-f7dc117d6f4a-operator-scripts\") pod \"placement-db-create-jwvzj\" (UID: \"d0ed4d40-1309-4641-b91c-f7dc117d6f4a\") " pod="openstack/placement-db-create-jwvzj" Dec 13 07:10:59 crc kubenswrapper[4971]: I1213 07:10:59.962107 4971 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-gwwcf\" (UniqueName: \"kubernetes.io/projected/d0ed4d40-1309-4641-b91c-f7dc117d6f4a-kube-api-access-gwwcf\") pod \"placement-db-create-jwvzj\" (UID: \"d0ed4d40-1309-4641-b91c-f7dc117d6f4a\") " pod="openstack/placement-db-create-jwvzj" Dec 13 07:11:00 crc kubenswrapper[4971]: I1213 07:11:00.041202 4971 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/292d6752-a759-4b19-9759-f0101a1b01e6-operator-scripts\") pod \"placement-e878-account-create-update-lnhvj\" (UID: \"292d6752-a759-4b19-9759-f0101a1b01e6\") " pod="openstack/placement-e878-account-create-update-lnhvj" Dec 13 07:11:00 crc kubenswrapper[4971]: I1213 07:11:00.041509 4971 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-gwhdz\" (UniqueName: \"kubernetes.io/projected/292d6752-a759-4b19-9759-f0101a1b01e6-kube-api-access-gwhdz\") pod \"placement-e878-account-create-update-lnhvj\" (UID: \"292d6752-a759-4b19-9759-f0101a1b01e6\") " pod="openstack/placement-e878-account-create-update-lnhvj" Dec 13 07:11:00 crc kubenswrapper[4971]: I1213 07:11:00.042245 4971 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/292d6752-a759-4b19-9759-f0101a1b01e6-operator-scripts\") pod \"placement-e878-account-create-update-lnhvj\" (UID: \"292d6752-a759-4b19-9759-f0101a1b01e6\") " pod="openstack/placement-e878-account-create-update-lnhvj" Dec 13 07:11:00 crc kubenswrapper[4971]: I1213 07:11:00.061591 4971 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-gwhdz\" (UniqueName: \"kubernetes.io/projected/292d6752-a759-4b19-9759-f0101a1b01e6-kube-api-access-gwhdz\") pod \"placement-e878-account-create-update-lnhvj\" (UID: \"292d6752-a759-4b19-9759-f0101a1b01e6\") " pod="openstack/placement-e878-account-create-update-lnhvj" Dec 13 07:11:00 crc kubenswrapper[4971]: I1213 07:11:00.117078 4971 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/glance-db-create-9wmnw" Dec 13 07:11:00 crc kubenswrapper[4971]: I1213 07:11:00.705288 4971 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/placement-db-create-jwvzj" Dec 13 07:11:00 crc kubenswrapper[4971]: I1213 07:11:00.710492 4971 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/ba68c497-330b-4e22-8ecd-d84ee54815d0-operator-scripts\") pod \"ba68c497-330b-4e22-8ecd-d84ee54815d0\" (UID: \"ba68c497-330b-4e22-8ecd-d84ee54815d0\") " Dec 13 07:11:00 crc kubenswrapper[4971]: I1213 07:11:00.710633 4971 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-qrqzd\" (UniqueName: \"kubernetes.io/projected/ba68c497-330b-4e22-8ecd-d84ee54815d0-kube-api-access-qrqzd\") pod \"ba68c497-330b-4e22-8ecd-d84ee54815d0\" (UID: \"ba68c497-330b-4e22-8ecd-d84ee54815d0\") " Dec 13 07:11:00 crc kubenswrapper[4971]: I1213 07:11:00.712986 4971 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/placement-e878-account-create-update-lnhvj" Dec 13 07:11:00 crc kubenswrapper[4971]: I1213 07:11:00.714352 4971 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/ba68c497-330b-4e22-8ecd-d84ee54815d0-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "ba68c497-330b-4e22-8ecd-d84ee54815d0" (UID: "ba68c497-330b-4e22-8ecd-d84ee54815d0"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 13 07:11:01 crc kubenswrapper[4971]: I1213 07:11:01.025029 4971 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-a8e5-account-create-update-7xfzr" event={"ID":"00df1bf7-20e4-485c-b1c2-6f00df8379ff","Type":"ContainerDied","Data":"46ee4f51ad43c512138ab8be4c7389fe370875287a2f6cd1b261a6d60607dd16"} Dec 13 07:11:01 crc kubenswrapper[4971]: I1213 07:11:01.025081 4971 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="46ee4f51ad43c512138ab8be4c7389fe370875287a2f6cd1b261a6d60607dd16" Dec 13 07:11:01 crc kubenswrapper[4971]: I1213 07:11:01.034419 4971 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/ba68c497-330b-4e22-8ecd-d84ee54815d0-kube-api-access-qrqzd" (OuterVolumeSpecName: "kube-api-access-qrqzd") pod "ba68c497-330b-4e22-8ecd-d84ee54815d0" (UID: "ba68c497-330b-4e22-8ecd-d84ee54815d0"). InnerVolumeSpecName "kube-api-access-qrqzd". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 13 07:11:01 crc kubenswrapper[4971]: I1213 07:11:01.035113 4971 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-qrqzd\" (UniqueName: \"kubernetes.io/projected/ba68c497-330b-4e22-8ecd-d84ee54815d0-kube-api-access-qrqzd\") pod \"ba68c497-330b-4e22-8ecd-d84ee54815d0\" (UID: \"ba68c497-330b-4e22-8ecd-d84ee54815d0\") " Dec 13 07:11:01 crc kubenswrapper[4971]: I1213 07:11:01.035881 4971 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/ba68c497-330b-4e22-8ecd-d84ee54815d0-operator-scripts\") on node \"crc\" DevicePath \"\"" Dec 13 07:11:01 crc kubenswrapper[4971]: W1213 07:11:01.036554 4971 empty_dir.go:500] Warning: Unmount skipped because path does not exist: /var/lib/kubelet/pods/ba68c497-330b-4e22-8ecd-d84ee54815d0/volumes/kubernetes.io~projected/kube-api-access-qrqzd Dec 13 07:11:01 crc kubenswrapper[4971]: I1213 07:11:01.036647 4971 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/ba68c497-330b-4e22-8ecd-d84ee54815d0-kube-api-access-qrqzd" (OuterVolumeSpecName: "kube-api-access-qrqzd") pod "ba68c497-330b-4e22-8ecd-d84ee54815d0" (UID: "ba68c497-330b-4e22-8ecd-d84ee54815d0"). InnerVolumeSpecName "kube-api-access-qrqzd". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 13 07:11:01 crc kubenswrapper[4971]: I1213 07:11:01.111307 4971 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/glance-a8e5-account-create-update-7xfzr" Dec 13 07:11:01 crc kubenswrapper[4971]: I1213 07:11:01.113170 4971 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/glance-db-create-9wmnw" Dec 13 07:11:01 crc kubenswrapper[4971]: I1213 07:11:01.114264 4971 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-db-create-9wmnw" event={"ID":"ba68c497-330b-4e22-8ecd-d84ee54815d0","Type":"ContainerDied","Data":"33b3bf0c5f642ba4386a2a7bd46ff2a54f43dc7d87e1ee80f9db4ccbf2b84d38"} Dec 13 07:11:01 crc kubenswrapper[4971]: I1213 07:11:01.114325 4971 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="33b3bf0c5f642ba4386a2a7bd46ff2a54f43dc7d87e1ee80f9db4ccbf2b84d38" Dec 13 07:11:01 crc kubenswrapper[4971]: I1213 07:11:01.114355 4971 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/ovn-northd-0" Dec 13 07:11:01 crc kubenswrapper[4971]: I1213 07:11:01.118744 4971 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/keystone-db-create-gzpfj"] Dec 13 07:11:01 crc kubenswrapper[4971]: I1213 07:11:01.141289 4971 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-qrqzd\" (UniqueName: \"kubernetes.io/projected/ba68c497-330b-4e22-8ecd-d84ee54815d0-kube-api-access-qrqzd\") on node \"crc\" DevicePath \"\"" Dec 13 07:11:01 crc kubenswrapper[4971]: I1213 07:11:01.672689 4971 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-zzwqb\" (UniqueName: \"kubernetes.io/projected/00df1bf7-20e4-485c-b1c2-6f00df8379ff-kube-api-access-zzwqb\") pod \"00df1bf7-20e4-485c-b1c2-6f00df8379ff\" (UID: \"00df1bf7-20e4-485c-b1c2-6f00df8379ff\") " Dec 13 07:11:01 crc kubenswrapper[4971]: I1213 07:11:01.673112 4971 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/00df1bf7-20e4-485c-b1c2-6f00df8379ff-operator-scripts\") pod \"00df1bf7-20e4-485c-b1c2-6f00df8379ff\" (UID: \"00df1bf7-20e4-485c-b1c2-6f00df8379ff\") " Dec 13 07:11:01 crc kubenswrapper[4971]: I1213 07:11:01.674378 4971 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/00df1bf7-20e4-485c-b1c2-6f00df8379ff-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "00df1bf7-20e4-485c-b1c2-6f00df8379ff" (UID: "00df1bf7-20e4-485c-b1c2-6f00df8379ff"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 13 07:11:01 crc kubenswrapper[4971]: I1213 07:11:01.687238 4971 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/00df1bf7-20e4-485c-b1c2-6f00df8379ff-kube-api-access-zzwqb" (OuterVolumeSpecName: "kube-api-access-zzwqb") pod "00df1bf7-20e4-485c-b1c2-6f00df8379ff" (UID: "00df1bf7-20e4-485c-b1c2-6f00df8379ff"). InnerVolumeSpecName "kube-api-access-zzwqb". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 13 07:11:01 crc kubenswrapper[4971]: I1213 07:11:01.775393 4971 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-zzwqb\" (UniqueName: \"kubernetes.io/projected/00df1bf7-20e4-485c-b1c2-6f00df8379ff-kube-api-access-zzwqb\") on node \"crc\" DevicePath \"\"" Dec 13 07:11:01 crc kubenswrapper[4971]: I1213 07:11:01.775437 4971 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/00df1bf7-20e4-485c-b1c2-6f00df8379ff-operator-scripts\") on node \"crc\" DevicePath \"\"" Dec 13 07:11:01 crc kubenswrapper[4971]: I1213 07:11:01.808391 4971 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/keystone-2ba7-account-create-update-v5s9z"] Dec 13 07:11:01 crc kubenswrapper[4971]: E1213 07:11:01.983834 4971 cadvisor_stats_provider.go:516] "Partial failure issuing cadvisor.ContainerInfoV2" err="partial failures: [\"/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podba68c497_330b_4e22_8ecd_d84ee54815d0.slice\": RecentStats: unable to find data in memory cache], [\"/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podba68c497_330b_4e22_8ecd_d84ee54815d0.slice/crio-33b3bf0c5f642ba4386a2a7bd46ff2a54f43dc7d87e1ee80f9db4ccbf2b84d38\": RecentStats: unable to find data in memory cache]" Dec 13 07:11:02 crc kubenswrapper[4971]: I1213 07:11:02.462581 4971 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/placement-db-create-jwvzj"] Dec 13 07:11:02 crc kubenswrapper[4971]: I1213 07:11:02.478586 4971 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-2ba7-account-create-update-v5s9z" event={"ID":"b2645870-215e-4acb-9f8c-cfa2bb0a5cf6","Type":"ContainerStarted","Data":"ffb1e916e3cb1e5744f54124d657c7cb2cc7aaba0a6198ee74432ef6fca98576"} Dec 13 07:11:02 crc kubenswrapper[4971]: I1213 07:11:02.491591 4971 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-db-create-gzpfj" event={"ID":"e07e072c-e921-4f2d-a967-1966c35ff49b","Type":"ContainerStarted","Data":"5c9ae72e290c3483b4208bf0b02b5d326a7928aa2cbd7159f5bd93267a3ee49d"} Dec 13 07:11:02 crc kubenswrapper[4971]: I1213 07:11:02.491663 4971 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/glance-a8e5-account-create-update-7xfzr" Dec 13 07:11:02 crc kubenswrapper[4971]: I1213 07:11:02.880911 4971 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/placement-e878-account-create-update-lnhvj"] Dec 13 07:11:02 crc kubenswrapper[4971]: W1213 07:11:02.886494 4971 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod292d6752_a759_4b19_9759_f0101a1b01e6.slice/crio-f437a19a0c612b0dfa964c1faab5f0dd1833fc33beb7f9b9a74734092d39a895 WatchSource:0}: Error finding container f437a19a0c612b0dfa964c1faab5f0dd1833fc33beb7f9b9a74734092d39a895: Status 404 returned error can't find the container with id f437a19a0c612b0dfa964c1faab5f0dd1833fc33beb7f9b9a74734092d39a895 Dec 13 07:11:03 crc kubenswrapper[4971]: I1213 07:11:03.503778 4971 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/placement-e878-account-create-update-lnhvj" event={"ID":"292d6752-a759-4b19-9759-f0101a1b01e6","Type":"ContainerStarted","Data":"f437a19a0c612b0dfa964c1faab5f0dd1833fc33beb7f9b9a74734092d39a895"} Dec 13 07:11:03 crc kubenswrapper[4971]: I1213 07:11:03.505929 4971 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/placement-db-create-jwvzj" event={"ID":"d0ed4d40-1309-4641-b91c-f7dc117d6f4a","Type":"ContainerStarted","Data":"5b61113487be9a48e236d9b3200556a7633e3c8353bdb373ab672bdbc4f2c5cf"} Dec 13 07:11:03 crc kubenswrapper[4971]: I1213 07:11:03.657751 4971 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/dnsmasq-dns-794868bd45-t92j4" Dec 13 07:11:03 crc kubenswrapper[4971]: I1213 07:11:03.990378 4971 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/dnsmasq-dns-757dc6fff9-hfmv4" Dec 13 07:11:04 crc kubenswrapper[4971]: I1213 07:11:04.086078 4971 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-794868bd45-t92j4"] Dec 13 07:11:04 crc kubenswrapper[4971]: I1213 07:11:04.512686 4971 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/dnsmasq-dns-794868bd45-t92j4" podUID="20e72d59-fbc4-4eca-84d0-04c833aa4aab" containerName="dnsmasq-dns" containerID="cri-o://6f845eb132e2c1de1a88cc497e5c06a82fd897df9a62779aeccbdf2e66a02ec8" gracePeriod=10 Dec 13 07:11:05 crc kubenswrapper[4971]: I1213 07:11:05.888302 4971 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/glance-db-sync-s24b9"] Dec 13 07:11:05 crc kubenswrapper[4971]: E1213 07:11:05.889025 4971 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="ba68c497-330b-4e22-8ecd-d84ee54815d0" containerName="mariadb-database-create" Dec 13 07:11:05 crc kubenswrapper[4971]: I1213 07:11:05.889043 4971 state_mem.go:107] "Deleted CPUSet assignment" podUID="ba68c497-330b-4e22-8ecd-d84ee54815d0" containerName="mariadb-database-create" Dec 13 07:11:05 crc kubenswrapper[4971]: E1213 07:11:05.889090 4971 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="00df1bf7-20e4-485c-b1c2-6f00df8379ff" containerName="mariadb-account-create-update" Dec 13 07:11:05 crc kubenswrapper[4971]: I1213 07:11:05.889098 4971 state_mem.go:107] "Deleted CPUSet assignment" podUID="00df1bf7-20e4-485c-b1c2-6f00df8379ff" containerName="mariadb-account-create-update" Dec 13 07:11:05 crc kubenswrapper[4971]: I1213 07:11:05.889270 4971 memory_manager.go:354] "RemoveStaleState removing state" podUID="00df1bf7-20e4-485c-b1c2-6f00df8379ff" containerName="mariadb-account-create-update" Dec 13 07:11:05 crc kubenswrapper[4971]: I1213 07:11:05.889288 4971 memory_manager.go:354] "RemoveStaleState removing state" podUID="ba68c497-330b-4e22-8ecd-d84ee54815d0" containerName="mariadb-database-create" Dec 13 07:11:05 crc kubenswrapper[4971]: I1213 07:11:05.889966 4971 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-db-sync-s24b9" Dec 13 07:11:05 crc kubenswrapper[4971]: I1213 07:11:05.892201 4971 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"glance-config-data" Dec 13 07:11:05 crc kubenswrapper[4971]: I1213 07:11:05.895841 4971 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"glance-glance-dockercfg-k864m" Dec 13 07:11:05 crc kubenswrapper[4971]: I1213 07:11:05.899381 4971 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-db-sync-s24b9"] Dec 13 07:11:05 crc kubenswrapper[4971]: I1213 07:11:05.945754 4971 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/6d483ba4-aee0-4c3b-a416-f3ee7ba19feb-config-data\") pod \"glance-db-sync-s24b9\" (UID: \"6d483ba4-aee0-4c3b-a416-f3ee7ba19feb\") " pod="openstack/glance-db-sync-s24b9" Dec 13 07:11:05 crc kubenswrapper[4971]: I1213 07:11:05.945865 4971 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-nqfn5\" (UniqueName: \"kubernetes.io/projected/6d483ba4-aee0-4c3b-a416-f3ee7ba19feb-kube-api-access-nqfn5\") pod \"glance-db-sync-s24b9\" (UID: \"6d483ba4-aee0-4c3b-a416-f3ee7ba19feb\") " pod="openstack/glance-db-sync-s24b9" Dec 13 07:11:05 crc kubenswrapper[4971]: I1213 07:11:05.946097 4971 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/6d483ba4-aee0-4c3b-a416-f3ee7ba19feb-db-sync-config-data\") pod \"glance-db-sync-s24b9\" (UID: \"6d483ba4-aee0-4c3b-a416-f3ee7ba19feb\") " pod="openstack/glance-db-sync-s24b9" Dec 13 07:11:05 crc kubenswrapper[4971]: I1213 07:11:05.946289 4971 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/6d483ba4-aee0-4c3b-a416-f3ee7ba19feb-combined-ca-bundle\") pod \"glance-db-sync-s24b9\" (UID: \"6d483ba4-aee0-4c3b-a416-f3ee7ba19feb\") " pod="openstack/glance-db-sync-s24b9" Dec 13 07:11:06 crc kubenswrapper[4971]: I1213 07:11:06.047508 4971 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-nqfn5\" (UniqueName: \"kubernetes.io/projected/6d483ba4-aee0-4c3b-a416-f3ee7ba19feb-kube-api-access-nqfn5\") pod \"glance-db-sync-s24b9\" (UID: \"6d483ba4-aee0-4c3b-a416-f3ee7ba19feb\") " pod="openstack/glance-db-sync-s24b9" Dec 13 07:11:06 crc kubenswrapper[4971]: I1213 07:11:06.047658 4971 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/6d483ba4-aee0-4c3b-a416-f3ee7ba19feb-db-sync-config-data\") pod \"glance-db-sync-s24b9\" (UID: \"6d483ba4-aee0-4c3b-a416-f3ee7ba19feb\") " pod="openstack/glance-db-sync-s24b9" Dec 13 07:11:06 crc kubenswrapper[4971]: I1213 07:11:06.047705 4971 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/6d483ba4-aee0-4c3b-a416-f3ee7ba19feb-combined-ca-bundle\") pod \"glance-db-sync-s24b9\" (UID: \"6d483ba4-aee0-4c3b-a416-f3ee7ba19feb\") " pod="openstack/glance-db-sync-s24b9" Dec 13 07:11:06 crc kubenswrapper[4971]: I1213 07:11:06.047799 4971 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/6d483ba4-aee0-4c3b-a416-f3ee7ba19feb-config-data\") pod \"glance-db-sync-s24b9\" (UID: \"6d483ba4-aee0-4c3b-a416-f3ee7ba19feb\") " pod="openstack/glance-db-sync-s24b9" Dec 13 07:11:06 crc kubenswrapper[4971]: I1213 07:11:06.054865 4971 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/6d483ba4-aee0-4c3b-a416-f3ee7ba19feb-db-sync-config-data\") pod \"glance-db-sync-s24b9\" (UID: \"6d483ba4-aee0-4c3b-a416-f3ee7ba19feb\") " pod="openstack/glance-db-sync-s24b9" Dec 13 07:11:06 crc kubenswrapper[4971]: I1213 07:11:06.056060 4971 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/6d483ba4-aee0-4c3b-a416-f3ee7ba19feb-combined-ca-bundle\") pod \"glance-db-sync-s24b9\" (UID: \"6d483ba4-aee0-4c3b-a416-f3ee7ba19feb\") " pod="openstack/glance-db-sync-s24b9" Dec 13 07:11:06 crc kubenswrapper[4971]: I1213 07:11:06.058778 4971 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/6d483ba4-aee0-4c3b-a416-f3ee7ba19feb-config-data\") pod \"glance-db-sync-s24b9\" (UID: \"6d483ba4-aee0-4c3b-a416-f3ee7ba19feb\") " pod="openstack/glance-db-sync-s24b9" Dec 13 07:11:06 crc kubenswrapper[4971]: I1213 07:11:06.069797 4971 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-nqfn5\" (UniqueName: \"kubernetes.io/projected/6d483ba4-aee0-4c3b-a416-f3ee7ba19feb-kube-api-access-nqfn5\") pod \"glance-db-sync-s24b9\" (UID: \"6d483ba4-aee0-4c3b-a416-f3ee7ba19feb\") " pod="openstack/glance-db-sync-s24b9" Dec 13 07:11:06 crc kubenswrapper[4971]: I1213 07:11:06.206228 4971 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-db-sync-s24b9" Dec 13 07:11:06 crc kubenswrapper[4971]: I1213 07:11:06.546188 4971 generic.go:334] "Generic (PLEG): container finished" podID="20e72d59-fbc4-4eca-84d0-04c833aa4aab" containerID="6f845eb132e2c1de1a88cc497e5c06a82fd897df9a62779aeccbdf2e66a02ec8" exitCode=0 Dec 13 07:11:06 crc kubenswrapper[4971]: I1213 07:11:06.546272 4971 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-794868bd45-t92j4" event={"ID":"20e72d59-fbc4-4eca-84d0-04c833aa4aab","Type":"ContainerDied","Data":"6f845eb132e2c1de1a88cc497e5c06a82fd897df9a62779aeccbdf2e66a02ec8"} Dec 13 07:11:06 crc kubenswrapper[4971]: I1213 07:11:06.549244 4971 generic.go:334] "Generic (PLEG): container finished" podID="b2645870-215e-4acb-9f8c-cfa2bb0a5cf6" containerID="9954aa1f2e47bd1129572705d83bf81bf56d842edd10741b7415b648d7ac42b9" exitCode=0 Dec 13 07:11:06 crc kubenswrapper[4971]: I1213 07:11:06.549357 4971 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-2ba7-account-create-update-v5s9z" event={"ID":"b2645870-215e-4acb-9f8c-cfa2bb0a5cf6","Type":"ContainerDied","Data":"9954aa1f2e47bd1129572705d83bf81bf56d842edd10741b7415b648d7ac42b9"} Dec 13 07:11:06 crc kubenswrapper[4971]: I1213 07:11:06.553665 4971 generic.go:334] "Generic (PLEG): container finished" podID="e07e072c-e921-4f2d-a967-1966c35ff49b" containerID="0478cc6d9e47009edd4f98ea96f3a639adea516ff1dd78509c7bfda6bffe3874" exitCode=0 Dec 13 07:11:06 crc kubenswrapper[4971]: I1213 07:11:06.553752 4971 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-db-create-gzpfj" event={"ID":"e07e072c-e921-4f2d-a967-1966c35ff49b","Type":"ContainerDied","Data":"0478cc6d9e47009edd4f98ea96f3a639adea516ff1dd78509c7bfda6bffe3874"} Dec 13 07:11:06 crc kubenswrapper[4971]: I1213 07:11:06.555979 4971 generic.go:334] "Generic (PLEG): container finished" podID="292d6752-a759-4b19-9759-f0101a1b01e6" containerID="71c48a05c5bd070dcf1c7058c0af96d1e9e2a917513a2d26b09d1778b89a2045" exitCode=0 Dec 13 07:11:06 crc kubenswrapper[4971]: I1213 07:11:06.556068 4971 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/placement-e878-account-create-update-lnhvj" event={"ID":"292d6752-a759-4b19-9759-f0101a1b01e6","Type":"ContainerDied","Data":"71c48a05c5bd070dcf1c7058c0af96d1e9e2a917513a2d26b09d1778b89a2045"} Dec 13 07:11:06 crc kubenswrapper[4971]: I1213 07:11:06.557585 4971 generic.go:334] "Generic (PLEG): container finished" podID="d0ed4d40-1309-4641-b91c-f7dc117d6f4a" containerID="3c10d92d48ac251e515ced70913bfe1054c399803e7e124888cbabe71d4abe76" exitCode=0 Dec 13 07:11:06 crc kubenswrapper[4971]: I1213 07:11:06.557613 4971 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/placement-db-create-jwvzj" event={"ID":"d0ed4d40-1309-4641-b91c-f7dc117d6f4a","Type":"ContainerDied","Data":"3c10d92d48ac251e515ced70913bfe1054c399803e7e124888cbabe71d4abe76"} Dec 13 07:11:06 crc kubenswrapper[4971]: I1213 07:11:06.728052 4971 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-db-sync-s24b9"] Dec 13 07:11:06 crc kubenswrapper[4971]: I1213 07:11:06.737047 4971 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-794868bd45-t92j4" Dec 13 07:11:06 crc kubenswrapper[4971]: I1213 07:11:06.759962 4971 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/20e72d59-fbc4-4eca-84d0-04c833aa4aab-config\") pod \"20e72d59-fbc4-4eca-84d0-04c833aa4aab\" (UID: \"20e72d59-fbc4-4eca-84d0-04c833aa4aab\") " Dec 13 07:11:06 crc kubenswrapper[4971]: I1213 07:11:06.760029 4971 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/20e72d59-fbc4-4eca-84d0-04c833aa4aab-ovsdbserver-sb\") pod \"20e72d59-fbc4-4eca-84d0-04c833aa4aab\" (UID: \"20e72d59-fbc4-4eca-84d0-04c833aa4aab\") " Dec 13 07:11:06 crc kubenswrapper[4971]: I1213 07:11:06.760075 4971 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-n652v\" (UniqueName: \"kubernetes.io/projected/20e72d59-fbc4-4eca-84d0-04c833aa4aab-kube-api-access-n652v\") pod \"20e72d59-fbc4-4eca-84d0-04c833aa4aab\" (UID: \"20e72d59-fbc4-4eca-84d0-04c833aa4aab\") " Dec 13 07:11:06 crc kubenswrapper[4971]: I1213 07:11:06.760148 4971 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/20e72d59-fbc4-4eca-84d0-04c833aa4aab-dns-svc\") pod \"20e72d59-fbc4-4eca-84d0-04c833aa4aab\" (UID: \"20e72d59-fbc4-4eca-84d0-04c833aa4aab\") " Dec 13 07:11:06 crc kubenswrapper[4971]: I1213 07:11:06.778115 4971 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/20e72d59-fbc4-4eca-84d0-04c833aa4aab-kube-api-access-n652v" (OuterVolumeSpecName: "kube-api-access-n652v") pod "20e72d59-fbc4-4eca-84d0-04c833aa4aab" (UID: "20e72d59-fbc4-4eca-84d0-04c833aa4aab"). InnerVolumeSpecName "kube-api-access-n652v". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 13 07:11:06 crc kubenswrapper[4971]: I1213 07:11:06.820901 4971 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/20e72d59-fbc4-4eca-84d0-04c833aa4aab-config" (OuterVolumeSpecName: "config") pod "20e72d59-fbc4-4eca-84d0-04c833aa4aab" (UID: "20e72d59-fbc4-4eca-84d0-04c833aa4aab"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 13 07:11:06 crc kubenswrapper[4971]: I1213 07:11:06.826849 4971 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/20e72d59-fbc4-4eca-84d0-04c833aa4aab-dns-svc" (OuterVolumeSpecName: "dns-svc") pod "20e72d59-fbc4-4eca-84d0-04c833aa4aab" (UID: "20e72d59-fbc4-4eca-84d0-04c833aa4aab"). InnerVolumeSpecName "dns-svc". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 13 07:11:06 crc kubenswrapper[4971]: I1213 07:11:06.827198 4971 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/20e72d59-fbc4-4eca-84d0-04c833aa4aab-ovsdbserver-sb" (OuterVolumeSpecName: "ovsdbserver-sb") pod "20e72d59-fbc4-4eca-84d0-04c833aa4aab" (UID: "20e72d59-fbc4-4eca-84d0-04c833aa4aab"). InnerVolumeSpecName "ovsdbserver-sb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 13 07:11:06 crc kubenswrapper[4971]: I1213 07:11:06.861506 4971 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/20e72d59-fbc4-4eca-84d0-04c833aa4aab-config\") on node \"crc\" DevicePath \"\"" Dec 13 07:11:06 crc kubenswrapper[4971]: I1213 07:11:06.861561 4971 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/20e72d59-fbc4-4eca-84d0-04c833aa4aab-ovsdbserver-sb\") on node \"crc\" DevicePath \"\"" Dec 13 07:11:06 crc kubenswrapper[4971]: I1213 07:11:06.861572 4971 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-n652v\" (UniqueName: \"kubernetes.io/projected/20e72d59-fbc4-4eca-84d0-04c833aa4aab-kube-api-access-n652v\") on node \"crc\" DevicePath \"\"" Dec 13 07:11:06 crc kubenswrapper[4971]: I1213 07:11:06.861581 4971 reconciler_common.go:293] "Volume detached for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/20e72d59-fbc4-4eca-84d0-04c833aa4aab-dns-svc\") on node \"crc\" DevicePath \"\"" Dec 13 07:11:07 crc kubenswrapper[4971]: I1213 07:11:07.567086 4971 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-db-sync-s24b9" event={"ID":"6d483ba4-aee0-4c3b-a416-f3ee7ba19feb","Type":"ContainerStarted","Data":"8708a88bcce239caff9aab875303eb33006234c20b442a2d0d6e97ee0042683a"} Dec 13 07:11:07 crc kubenswrapper[4971]: I1213 07:11:07.568991 4971 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-794868bd45-t92j4" event={"ID":"20e72d59-fbc4-4eca-84d0-04c833aa4aab","Type":"ContainerDied","Data":"4dd2d455623b372f86f60f01914be2533930044fea79e6269ca1412f898b2e9c"} Dec 13 07:11:07 crc kubenswrapper[4971]: I1213 07:11:07.569043 4971 scope.go:117] "RemoveContainer" containerID="6f845eb132e2c1de1a88cc497e5c06a82fd897df9a62779aeccbdf2e66a02ec8" Dec 13 07:11:07 crc kubenswrapper[4971]: I1213 07:11:07.569107 4971 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-794868bd45-t92j4" Dec 13 07:11:07 crc kubenswrapper[4971]: I1213 07:11:07.606609 4971 scope.go:117] "RemoveContainer" containerID="8c57fbe679c7ff81971e834c6f85fd7c746e08bd5a9c82a44f8819f17ec47c8f" Dec 13 07:11:07 crc kubenswrapper[4971]: I1213 07:11:07.614652 4971 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-794868bd45-t92j4"] Dec 13 07:11:07 crc kubenswrapper[4971]: I1213 07:11:07.617538 4971 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-794868bd45-t92j4"] Dec 13 07:11:07 crc kubenswrapper[4971]: I1213 07:11:07.789012 4971 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="20e72d59-fbc4-4eca-84d0-04c833aa4aab" path="/var/lib/kubelet/pods/20e72d59-fbc4-4eca-84d0-04c833aa4aab/volumes" Dec 13 07:11:07 crc kubenswrapper[4971]: I1213 07:11:07.976740 4971 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-db-create-gzpfj" Dec 13 07:11:07 crc kubenswrapper[4971]: I1213 07:11:07.994207 4971 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/e07e072c-e921-4f2d-a967-1966c35ff49b-operator-scripts\") pod \"e07e072c-e921-4f2d-a967-1966c35ff49b\" (UID: \"e07e072c-e921-4f2d-a967-1966c35ff49b\") " Dec 13 07:11:07 crc kubenswrapper[4971]: I1213 07:11:07.994304 4971 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-8qxrq\" (UniqueName: \"kubernetes.io/projected/e07e072c-e921-4f2d-a967-1966c35ff49b-kube-api-access-8qxrq\") pod \"e07e072c-e921-4f2d-a967-1966c35ff49b\" (UID: \"e07e072c-e921-4f2d-a967-1966c35ff49b\") " Dec 13 07:11:07 crc kubenswrapper[4971]: I1213 07:11:07.995649 4971 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/e07e072c-e921-4f2d-a967-1966c35ff49b-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "e07e072c-e921-4f2d-a967-1966c35ff49b" (UID: "e07e072c-e921-4f2d-a967-1966c35ff49b"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 13 07:11:08 crc kubenswrapper[4971]: I1213 07:11:08.024015 4971 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/e07e072c-e921-4f2d-a967-1966c35ff49b-kube-api-access-8qxrq" (OuterVolumeSpecName: "kube-api-access-8qxrq") pod "e07e072c-e921-4f2d-a967-1966c35ff49b" (UID: "e07e072c-e921-4f2d-a967-1966c35ff49b"). InnerVolumeSpecName "kube-api-access-8qxrq". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 13 07:11:08 crc kubenswrapper[4971]: I1213 07:11:08.112682 4971 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/e07e072c-e921-4f2d-a967-1966c35ff49b-operator-scripts\") on node \"crc\" DevicePath \"\"" Dec 13 07:11:08 crc kubenswrapper[4971]: I1213 07:11:08.112704 4971 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-8qxrq\" (UniqueName: \"kubernetes.io/projected/e07e072c-e921-4f2d-a967-1966c35ff49b-kube-api-access-8qxrq\") on node \"crc\" DevicePath \"\"" Dec 13 07:11:08 crc kubenswrapper[4971]: I1213 07:11:08.128637 4971 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/placement-db-create-jwvzj" Dec 13 07:11:08 crc kubenswrapper[4971]: I1213 07:11:08.136846 4971 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/placement-e878-account-create-update-lnhvj" Dec 13 07:11:08 crc kubenswrapper[4971]: I1213 07:11:08.214040 4971 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-gwwcf\" (UniqueName: \"kubernetes.io/projected/d0ed4d40-1309-4641-b91c-f7dc117d6f4a-kube-api-access-gwwcf\") pod \"d0ed4d40-1309-4641-b91c-f7dc117d6f4a\" (UID: \"d0ed4d40-1309-4641-b91c-f7dc117d6f4a\") " Dec 13 07:11:08 crc kubenswrapper[4971]: I1213 07:11:08.214121 4971 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-gwhdz\" (UniqueName: \"kubernetes.io/projected/292d6752-a759-4b19-9759-f0101a1b01e6-kube-api-access-gwhdz\") pod \"292d6752-a759-4b19-9759-f0101a1b01e6\" (UID: \"292d6752-a759-4b19-9759-f0101a1b01e6\") " Dec 13 07:11:08 crc kubenswrapper[4971]: I1213 07:11:08.214202 4971 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/292d6752-a759-4b19-9759-f0101a1b01e6-operator-scripts\") pod \"292d6752-a759-4b19-9759-f0101a1b01e6\" (UID: \"292d6752-a759-4b19-9759-f0101a1b01e6\") " Dec 13 07:11:08 crc kubenswrapper[4971]: I1213 07:11:08.214390 4971 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/d0ed4d40-1309-4641-b91c-f7dc117d6f4a-operator-scripts\") pod \"d0ed4d40-1309-4641-b91c-f7dc117d6f4a\" (UID: \"d0ed4d40-1309-4641-b91c-f7dc117d6f4a\") " Dec 13 07:11:08 crc kubenswrapper[4971]: I1213 07:11:08.215304 4971 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/292d6752-a759-4b19-9759-f0101a1b01e6-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "292d6752-a759-4b19-9759-f0101a1b01e6" (UID: "292d6752-a759-4b19-9759-f0101a1b01e6"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 13 07:11:08 crc kubenswrapper[4971]: I1213 07:11:08.215678 4971 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/d0ed4d40-1309-4641-b91c-f7dc117d6f4a-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "d0ed4d40-1309-4641-b91c-f7dc117d6f4a" (UID: "d0ed4d40-1309-4641-b91c-f7dc117d6f4a"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 13 07:11:08 crc kubenswrapper[4971]: I1213 07:11:08.218863 4971 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/d0ed4d40-1309-4641-b91c-f7dc117d6f4a-kube-api-access-gwwcf" (OuterVolumeSpecName: "kube-api-access-gwwcf") pod "d0ed4d40-1309-4641-b91c-f7dc117d6f4a" (UID: "d0ed4d40-1309-4641-b91c-f7dc117d6f4a"). InnerVolumeSpecName "kube-api-access-gwwcf". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 13 07:11:08 crc kubenswrapper[4971]: I1213 07:11:08.222331 4971 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/292d6752-a759-4b19-9759-f0101a1b01e6-kube-api-access-gwhdz" (OuterVolumeSpecName: "kube-api-access-gwhdz") pod "292d6752-a759-4b19-9759-f0101a1b01e6" (UID: "292d6752-a759-4b19-9759-f0101a1b01e6"). InnerVolumeSpecName "kube-api-access-gwhdz". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 13 07:11:08 crc kubenswrapper[4971]: I1213 07:11:08.288903 4971 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-2ba7-account-create-update-v5s9z" Dec 13 07:11:08 crc kubenswrapper[4971]: I1213 07:11:08.315988 4971 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-sv8qz\" (UniqueName: \"kubernetes.io/projected/b2645870-215e-4acb-9f8c-cfa2bb0a5cf6-kube-api-access-sv8qz\") pod \"b2645870-215e-4acb-9f8c-cfa2bb0a5cf6\" (UID: \"b2645870-215e-4acb-9f8c-cfa2bb0a5cf6\") " Dec 13 07:11:08 crc kubenswrapper[4971]: I1213 07:11:08.316568 4971 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/b2645870-215e-4acb-9f8c-cfa2bb0a5cf6-operator-scripts\") pod \"b2645870-215e-4acb-9f8c-cfa2bb0a5cf6\" (UID: \"b2645870-215e-4acb-9f8c-cfa2bb0a5cf6\") " Dec 13 07:11:08 crc kubenswrapper[4971]: I1213 07:11:08.316966 4971 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/d0ed4d40-1309-4641-b91c-f7dc117d6f4a-operator-scripts\") on node \"crc\" DevicePath \"\"" Dec 13 07:11:08 crc kubenswrapper[4971]: I1213 07:11:08.317095 4971 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-gwwcf\" (UniqueName: \"kubernetes.io/projected/d0ed4d40-1309-4641-b91c-f7dc117d6f4a-kube-api-access-gwwcf\") on node \"crc\" DevicePath \"\"" Dec 13 07:11:08 crc kubenswrapper[4971]: I1213 07:11:08.317180 4971 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-gwhdz\" (UniqueName: \"kubernetes.io/projected/292d6752-a759-4b19-9759-f0101a1b01e6-kube-api-access-gwhdz\") on node \"crc\" DevicePath \"\"" Dec 13 07:11:08 crc kubenswrapper[4971]: I1213 07:11:08.317268 4971 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/292d6752-a759-4b19-9759-f0101a1b01e6-operator-scripts\") on node \"crc\" DevicePath \"\"" Dec 13 07:11:08 crc kubenswrapper[4971]: I1213 07:11:08.317092 4971 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/b2645870-215e-4acb-9f8c-cfa2bb0a5cf6-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "b2645870-215e-4acb-9f8c-cfa2bb0a5cf6" (UID: "b2645870-215e-4acb-9f8c-cfa2bb0a5cf6"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 13 07:11:08 crc kubenswrapper[4971]: I1213 07:11:08.335290 4971 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/b2645870-215e-4acb-9f8c-cfa2bb0a5cf6-kube-api-access-sv8qz" (OuterVolumeSpecName: "kube-api-access-sv8qz") pod "b2645870-215e-4acb-9f8c-cfa2bb0a5cf6" (UID: "b2645870-215e-4acb-9f8c-cfa2bb0a5cf6"). InnerVolumeSpecName "kube-api-access-sv8qz". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 13 07:11:08 crc kubenswrapper[4971]: I1213 07:11:08.419445 4971 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/b2645870-215e-4acb-9f8c-cfa2bb0a5cf6-operator-scripts\") on node \"crc\" DevicePath \"\"" Dec 13 07:11:08 crc kubenswrapper[4971]: I1213 07:11:08.419489 4971 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-sv8qz\" (UniqueName: \"kubernetes.io/projected/b2645870-215e-4acb-9f8c-cfa2bb0a5cf6-kube-api-access-sv8qz\") on node \"crc\" DevicePath \"\"" Dec 13 07:11:08 crc kubenswrapper[4971]: I1213 07:11:08.579897 4971 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-db-create-gzpfj" Dec 13 07:11:08 crc kubenswrapper[4971]: I1213 07:11:08.579933 4971 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-db-create-gzpfj" event={"ID":"e07e072c-e921-4f2d-a967-1966c35ff49b","Type":"ContainerDied","Data":"5c9ae72e290c3483b4208bf0b02b5d326a7928aa2cbd7159f5bd93267a3ee49d"} Dec 13 07:11:08 crc kubenswrapper[4971]: I1213 07:11:08.580352 4971 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="5c9ae72e290c3483b4208bf0b02b5d326a7928aa2cbd7159f5bd93267a3ee49d" Dec 13 07:11:08 crc kubenswrapper[4971]: I1213 07:11:08.584533 4971 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/placement-e878-account-create-update-lnhvj" Dec 13 07:11:08 crc kubenswrapper[4971]: I1213 07:11:08.584533 4971 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/placement-e878-account-create-update-lnhvj" event={"ID":"292d6752-a759-4b19-9759-f0101a1b01e6","Type":"ContainerDied","Data":"f437a19a0c612b0dfa964c1faab5f0dd1833fc33beb7f9b9a74734092d39a895"} Dec 13 07:11:08 crc kubenswrapper[4971]: I1213 07:11:08.584589 4971 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="f437a19a0c612b0dfa964c1faab5f0dd1833fc33beb7f9b9a74734092d39a895" Dec 13 07:11:08 crc kubenswrapper[4971]: I1213 07:11:08.586255 4971 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/placement-db-create-jwvzj" Dec 13 07:11:08 crc kubenswrapper[4971]: I1213 07:11:08.586269 4971 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/placement-db-create-jwvzj" event={"ID":"d0ed4d40-1309-4641-b91c-f7dc117d6f4a","Type":"ContainerDied","Data":"5b61113487be9a48e236d9b3200556a7633e3c8353bdb373ab672bdbc4f2c5cf"} Dec 13 07:11:08 crc kubenswrapper[4971]: I1213 07:11:08.586327 4971 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="5b61113487be9a48e236d9b3200556a7633e3c8353bdb373ab672bdbc4f2c5cf" Dec 13 07:11:08 crc kubenswrapper[4971]: I1213 07:11:08.587772 4971 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-2ba7-account-create-update-v5s9z" event={"ID":"b2645870-215e-4acb-9f8c-cfa2bb0a5cf6","Type":"ContainerDied","Data":"ffb1e916e3cb1e5744f54124d657c7cb2cc7aaba0a6198ee74432ef6fca98576"} Dec 13 07:11:08 crc kubenswrapper[4971]: I1213 07:11:08.587803 4971 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="ffb1e916e3cb1e5744f54124d657c7cb2cc7aaba0a6198ee74432ef6fca98576" Dec 13 07:11:08 crc kubenswrapper[4971]: I1213 07:11:08.587906 4971 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-2ba7-account-create-update-v5s9z" Dec 13 07:11:08 crc kubenswrapper[4971]: I1213 07:11:08.875656 4971 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/ovn-northd-0" Dec 13 07:11:15 crc kubenswrapper[4971]: I1213 07:11:15.562361 4971 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/ovn-controller-k5fz9" podUID="c0bf2685-ca54-4508-a580-531b05473e76" containerName="ovn-controller" probeResult="failure" output=< Dec 13 07:11:15 crc kubenswrapper[4971]: ERROR - ovn-controller connection status is 'not connected', expecting 'connected' status Dec 13 07:11:15 crc kubenswrapper[4971]: > Dec 13 07:11:15 crc kubenswrapper[4971]: I1213 07:11:15.980512 4971 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/ovn-controller-ovs-jwh28" Dec 13 07:11:16 crc kubenswrapper[4971]: I1213 07:11:16.153195 4971 patch_prober.go:28] interesting pod/machine-config-daemon-82xjz container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 13 07:11:16 crc kubenswrapper[4971]: I1213 07:11:16.153492 4971 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-82xjz" podUID="e0c4af12-edda-41d7-9b44-f87396174ef2" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 13 07:11:20 crc kubenswrapper[4971]: I1213 07:11:20.537361 4971 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/ovn-controller-k5fz9" podUID="c0bf2685-ca54-4508-a580-531b05473e76" containerName="ovn-controller" probeResult="failure" output=< Dec 13 07:11:20 crc kubenswrapper[4971]: ERROR - ovn-controller connection status is 'not connected', expecting 'connected' status Dec 13 07:11:20 crc kubenswrapper[4971]: > Dec 13 07:11:20 crc kubenswrapper[4971]: I1213 07:11:20.976167 4971 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/ovn-controller-ovs-jwh28" Dec 13 07:11:21 crc kubenswrapper[4971]: I1213 07:11:21.192296 4971 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ovn-controller-k5fz9-config-gv99r"] Dec 13 07:11:21 crc kubenswrapper[4971]: E1213 07:11:21.192695 4971 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="20e72d59-fbc4-4eca-84d0-04c833aa4aab" containerName="init" Dec 13 07:11:21 crc kubenswrapper[4971]: I1213 07:11:21.192715 4971 state_mem.go:107] "Deleted CPUSet assignment" podUID="20e72d59-fbc4-4eca-84d0-04c833aa4aab" containerName="init" Dec 13 07:11:21 crc kubenswrapper[4971]: E1213 07:11:21.192735 4971 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="292d6752-a759-4b19-9759-f0101a1b01e6" containerName="mariadb-account-create-update" Dec 13 07:11:21 crc kubenswrapper[4971]: I1213 07:11:21.192744 4971 state_mem.go:107] "Deleted CPUSet assignment" podUID="292d6752-a759-4b19-9759-f0101a1b01e6" containerName="mariadb-account-create-update" Dec 13 07:11:21 crc kubenswrapper[4971]: E1213 07:11:21.192770 4971 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="b2645870-215e-4acb-9f8c-cfa2bb0a5cf6" containerName="mariadb-account-create-update" Dec 13 07:11:21 crc kubenswrapper[4971]: I1213 07:11:21.192778 4971 state_mem.go:107] "Deleted CPUSet assignment" podUID="b2645870-215e-4acb-9f8c-cfa2bb0a5cf6" containerName="mariadb-account-create-update" Dec 13 07:11:21 crc kubenswrapper[4971]: E1213 07:11:21.192793 4971 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="e07e072c-e921-4f2d-a967-1966c35ff49b" containerName="mariadb-database-create" Dec 13 07:11:21 crc kubenswrapper[4971]: I1213 07:11:21.192800 4971 state_mem.go:107] "Deleted CPUSet assignment" podUID="e07e072c-e921-4f2d-a967-1966c35ff49b" containerName="mariadb-database-create" Dec 13 07:11:21 crc kubenswrapper[4971]: E1213 07:11:21.192825 4971 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="20e72d59-fbc4-4eca-84d0-04c833aa4aab" containerName="dnsmasq-dns" Dec 13 07:11:21 crc kubenswrapper[4971]: I1213 07:11:21.192832 4971 state_mem.go:107] "Deleted CPUSet assignment" podUID="20e72d59-fbc4-4eca-84d0-04c833aa4aab" containerName="dnsmasq-dns" Dec 13 07:11:21 crc kubenswrapper[4971]: E1213 07:11:21.192850 4971 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="d0ed4d40-1309-4641-b91c-f7dc117d6f4a" containerName="mariadb-database-create" Dec 13 07:11:21 crc kubenswrapper[4971]: I1213 07:11:21.192858 4971 state_mem.go:107] "Deleted CPUSet assignment" podUID="d0ed4d40-1309-4641-b91c-f7dc117d6f4a" containerName="mariadb-database-create" Dec 13 07:11:21 crc kubenswrapper[4971]: I1213 07:11:21.193039 4971 memory_manager.go:354] "RemoveStaleState removing state" podUID="d0ed4d40-1309-4641-b91c-f7dc117d6f4a" containerName="mariadb-database-create" Dec 13 07:11:21 crc kubenswrapper[4971]: I1213 07:11:21.193059 4971 memory_manager.go:354] "RemoveStaleState removing state" podUID="20e72d59-fbc4-4eca-84d0-04c833aa4aab" containerName="dnsmasq-dns" Dec 13 07:11:21 crc kubenswrapper[4971]: I1213 07:11:21.193067 4971 memory_manager.go:354] "RemoveStaleState removing state" podUID="e07e072c-e921-4f2d-a967-1966c35ff49b" containerName="mariadb-database-create" Dec 13 07:11:21 crc kubenswrapper[4971]: I1213 07:11:21.193077 4971 memory_manager.go:354] "RemoveStaleState removing state" podUID="b2645870-215e-4acb-9f8c-cfa2bb0a5cf6" containerName="mariadb-account-create-update" Dec 13 07:11:21 crc kubenswrapper[4971]: I1213 07:11:21.193090 4971 memory_manager.go:354] "RemoveStaleState removing state" podUID="292d6752-a759-4b19-9759-f0101a1b01e6" containerName="mariadb-account-create-update" Dec 13 07:11:21 crc kubenswrapper[4971]: I1213 07:11:21.193784 4971 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-controller-k5fz9-config-gv99r" Dec 13 07:11:21 crc kubenswrapper[4971]: I1213 07:11:21.195696 4971 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"ovncontroller-extra-scripts" Dec 13 07:11:21 crc kubenswrapper[4971]: I1213 07:11:21.203646 4971 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovn-controller-k5fz9-config-gv99r"] Dec 13 07:11:21 crc kubenswrapper[4971]: I1213 07:11:21.351060 4971 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/34f6d5f6-f9d5-4b88-bb8a-ad5e520673d9-scripts\") pod \"ovn-controller-k5fz9-config-gv99r\" (UID: \"34f6d5f6-f9d5-4b88-bb8a-ad5e520673d9\") " pod="openstack/ovn-controller-k5fz9-config-gv99r" Dec 13 07:11:21 crc kubenswrapper[4971]: I1213 07:11:21.351119 4971 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-run-ovn\" (UniqueName: \"kubernetes.io/host-path/34f6d5f6-f9d5-4b88-bb8a-ad5e520673d9-var-run-ovn\") pod \"ovn-controller-k5fz9-config-gv99r\" (UID: \"34f6d5f6-f9d5-4b88-bb8a-ad5e520673d9\") " pod="openstack/ovn-controller-k5fz9-config-gv99r" Dec 13 07:11:21 crc kubenswrapper[4971]: I1213 07:11:21.351165 4971 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-log-ovn\" (UniqueName: \"kubernetes.io/host-path/34f6d5f6-f9d5-4b88-bb8a-ad5e520673d9-var-log-ovn\") pod \"ovn-controller-k5fz9-config-gv99r\" (UID: \"34f6d5f6-f9d5-4b88-bb8a-ad5e520673d9\") " pod="openstack/ovn-controller-k5fz9-config-gv99r" Dec 13 07:11:21 crc kubenswrapper[4971]: I1213 07:11:21.351205 4971 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-zltq9\" (UniqueName: \"kubernetes.io/projected/34f6d5f6-f9d5-4b88-bb8a-ad5e520673d9-kube-api-access-zltq9\") pod \"ovn-controller-k5fz9-config-gv99r\" (UID: \"34f6d5f6-f9d5-4b88-bb8a-ad5e520673d9\") " pod="openstack/ovn-controller-k5fz9-config-gv99r" Dec 13 07:11:21 crc kubenswrapper[4971]: I1213 07:11:21.351259 4971 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-run\" (UniqueName: \"kubernetes.io/host-path/34f6d5f6-f9d5-4b88-bb8a-ad5e520673d9-var-run\") pod \"ovn-controller-k5fz9-config-gv99r\" (UID: \"34f6d5f6-f9d5-4b88-bb8a-ad5e520673d9\") " pod="openstack/ovn-controller-k5fz9-config-gv99r" Dec 13 07:11:21 crc kubenswrapper[4971]: I1213 07:11:21.351280 4971 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"additional-scripts\" (UniqueName: \"kubernetes.io/configmap/34f6d5f6-f9d5-4b88-bb8a-ad5e520673d9-additional-scripts\") pod \"ovn-controller-k5fz9-config-gv99r\" (UID: \"34f6d5f6-f9d5-4b88-bb8a-ad5e520673d9\") " pod="openstack/ovn-controller-k5fz9-config-gv99r" Dec 13 07:11:21 crc kubenswrapper[4971]: I1213 07:11:21.452425 4971 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/34f6d5f6-f9d5-4b88-bb8a-ad5e520673d9-scripts\") pod \"ovn-controller-k5fz9-config-gv99r\" (UID: \"34f6d5f6-f9d5-4b88-bb8a-ad5e520673d9\") " pod="openstack/ovn-controller-k5fz9-config-gv99r" Dec 13 07:11:21 crc kubenswrapper[4971]: I1213 07:11:21.452487 4971 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-run-ovn\" (UniqueName: \"kubernetes.io/host-path/34f6d5f6-f9d5-4b88-bb8a-ad5e520673d9-var-run-ovn\") pod \"ovn-controller-k5fz9-config-gv99r\" (UID: \"34f6d5f6-f9d5-4b88-bb8a-ad5e520673d9\") " pod="openstack/ovn-controller-k5fz9-config-gv99r" Dec 13 07:11:21 crc kubenswrapper[4971]: I1213 07:11:21.452595 4971 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-log-ovn\" (UniqueName: \"kubernetes.io/host-path/34f6d5f6-f9d5-4b88-bb8a-ad5e520673d9-var-log-ovn\") pod \"ovn-controller-k5fz9-config-gv99r\" (UID: \"34f6d5f6-f9d5-4b88-bb8a-ad5e520673d9\") " pod="openstack/ovn-controller-k5fz9-config-gv99r" Dec 13 07:11:21 crc kubenswrapper[4971]: I1213 07:11:21.452649 4971 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-zltq9\" (UniqueName: \"kubernetes.io/projected/34f6d5f6-f9d5-4b88-bb8a-ad5e520673d9-kube-api-access-zltq9\") pod \"ovn-controller-k5fz9-config-gv99r\" (UID: \"34f6d5f6-f9d5-4b88-bb8a-ad5e520673d9\") " pod="openstack/ovn-controller-k5fz9-config-gv99r" Dec 13 07:11:21 crc kubenswrapper[4971]: I1213 07:11:21.453046 4971 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-run-ovn\" (UniqueName: \"kubernetes.io/host-path/34f6d5f6-f9d5-4b88-bb8a-ad5e520673d9-var-run-ovn\") pod \"ovn-controller-k5fz9-config-gv99r\" (UID: \"34f6d5f6-f9d5-4b88-bb8a-ad5e520673d9\") " pod="openstack/ovn-controller-k5fz9-config-gv99r" Dec 13 07:11:21 crc kubenswrapper[4971]: I1213 07:11:21.453078 4971 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-log-ovn\" (UniqueName: \"kubernetes.io/host-path/34f6d5f6-f9d5-4b88-bb8a-ad5e520673d9-var-log-ovn\") pod \"ovn-controller-k5fz9-config-gv99r\" (UID: \"34f6d5f6-f9d5-4b88-bb8a-ad5e520673d9\") " pod="openstack/ovn-controller-k5fz9-config-gv99r" Dec 13 07:11:21 crc kubenswrapper[4971]: I1213 07:11:21.453116 4971 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-run\" (UniqueName: \"kubernetes.io/host-path/34f6d5f6-f9d5-4b88-bb8a-ad5e520673d9-var-run\") pod \"ovn-controller-k5fz9-config-gv99r\" (UID: \"34f6d5f6-f9d5-4b88-bb8a-ad5e520673d9\") " pod="openstack/ovn-controller-k5fz9-config-gv99r" Dec 13 07:11:21 crc kubenswrapper[4971]: I1213 07:11:21.453094 4971 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-run\" (UniqueName: \"kubernetes.io/host-path/34f6d5f6-f9d5-4b88-bb8a-ad5e520673d9-var-run\") pod \"ovn-controller-k5fz9-config-gv99r\" (UID: \"34f6d5f6-f9d5-4b88-bb8a-ad5e520673d9\") " pod="openstack/ovn-controller-k5fz9-config-gv99r" Dec 13 07:11:21 crc kubenswrapper[4971]: I1213 07:11:21.453177 4971 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"additional-scripts\" (UniqueName: \"kubernetes.io/configmap/34f6d5f6-f9d5-4b88-bb8a-ad5e520673d9-additional-scripts\") pod \"ovn-controller-k5fz9-config-gv99r\" (UID: \"34f6d5f6-f9d5-4b88-bb8a-ad5e520673d9\") " pod="openstack/ovn-controller-k5fz9-config-gv99r" Dec 13 07:11:21 crc kubenswrapper[4971]: I1213 07:11:21.453900 4971 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"additional-scripts\" (UniqueName: \"kubernetes.io/configmap/34f6d5f6-f9d5-4b88-bb8a-ad5e520673d9-additional-scripts\") pod \"ovn-controller-k5fz9-config-gv99r\" (UID: \"34f6d5f6-f9d5-4b88-bb8a-ad5e520673d9\") " pod="openstack/ovn-controller-k5fz9-config-gv99r" Dec 13 07:11:21 crc kubenswrapper[4971]: I1213 07:11:21.455376 4971 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/34f6d5f6-f9d5-4b88-bb8a-ad5e520673d9-scripts\") pod \"ovn-controller-k5fz9-config-gv99r\" (UID: \"34f6d5f6-f9d5-4b88-bb8a-ad5e520673d9\") " pod="openstack/ovn-controller-k5fz9-config-gv99r" Dec 13 07:11:21 crc kubenswrapper[4971]: I1213 07:11:21.475141 4971 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-zltq9\" (UniqueName: \"kubernetes.io/projected/34f6d5f6-f9d5-4b88-bb8a-ad5e520673d9-kube-api-access-zltq9\") pod \"ovn-controller-k5fz9-config-gv99r\" (UID: \"34f6d5f6-f9d5-4b88-bb8a-ad5e520673d9\") " pod="openstack/ovn-controller-k5fz9-config-gv99r" Dec 13 07:11:21 crc kubenswrapper[4971]: I1213 07:11:21.528679 4971 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-controller-k5fz9-config-gv99r" Dec 13 07:11:22 crc kubenswrapper[4971]: E1213 07:11:22.363171 4971 cadvisor_stats_provider.go:516] "Partial failure issuing cadvisor.ContainerInfoV2" err="partial failures: [\"/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod5eed809d_c915_433c_9915_27e0b01e1ffe.slice/crio-0ecc7342a1a65b67ea97dd2575e3d43416e7633d3c5e40a8ec3ff6b6e6709a86.scope\": RecentStats: unable to find data in memory cache]" Dec 13 07:11:23 crc kubenswrapper[4971]: E1213 07:11:23.109817 4971 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying config: context canceled" image="quay.io/podified-antelope-centos9/openstack-glance-api@sha256:e4aa4ebbb1e581a12040e9ad2ae2709ac31b5d965bb64fc4252d1028b05c565f" Dec 13 07:11:23 crc kubenswrapper[4971]: E1213 07:11:23.110689 4971 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:glance-db-sync,Image:quay.io/podified-antelope-centos9/openstack-glance-api@sha256:e4aa4ebbb1e581a12040e9ad2ae2709ac31b5d965bb64fc4252d1028b05c565f,Command:[/bin/bash],Args:[-c /usr/local/bin/kolla_start],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:KOLLA_BOOTSTRAP,Value:true,ValueFrom:nil,},EnvVar{Name:KOLLA_CONFIG_STRATEGY,Value:COPY_ALWAYS,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:db-sync-config-data,ReadOnly:true,MountPath:/etc/glance/glance.conf.d,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:config-data,ReadOnly:true,MountPath:/etc/my.cnf,SubPath:my.cnf,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:config-data,ReadOnly:true,MountPath:/var/lib/kolla/config_files/config.json,SubPath:db-sync-config.json,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:combined-ca-bundle,ReadOnly:true,MountPath:/etc/pki/ca-trust/extracted/pem/tls-ca-bundle.pem,SubPath:tls-ca-bundle.pem,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-nqfn5,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[MKNOD],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*42415,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:nil,RunAsGroup:*42415,ProcMount:nil,WindowsOptions:nil,SeccompProfile:&SeccompProfile{Type:RuntimeDefault,LocalhostProfile:nil,},AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod glance-db-sync-s24b9_openstack(6d483ba4-aee0-4c3b-a416-f3ee7ba19feb): ErrImagePull: rpc error: code = Canceled desc = copying config: context canceled" logger="UnhandledError" Dec 13 07:11:23 crc kubenswrapper[4971]: E1213 07:11:23.111931 4971 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"glance-db-sync\" with ErrImagePull: \"rpc error: code = Canceled desc = copying config: context canceled\"" pod="openstack/glance-db-sync-s24b9" podUID="6d483ba4-aee0-4c3b-a416-f3ee7ba19feb" Dec 13 07:11:23 crc kubenswrapper[4971]: W1213 07:11:23.499009 4971 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod34f6d5f6_f9d5_4b88_bb8a_ad5e520673d9.slice/crio-f2fe21ad19c69cfeef6d7a24ff2a3d73950c5744333e407b485869a4ec26d067 WatchSource:0}: Error finding container f2fe21ad19c69cfeef6d7a24ff2a3d73950c5744333e407b485869a4ec26d067: Status 404 returned error can't find the container with id f2fe21ad19c69cfeef6d7a24ff2a3d73950c5744333e407b485869a4ec26d067 Dec 13 07:11:23 crc kubenswrapper[4971]: I1213 07:11:23.499230 4971 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovn-controller-k5fz9-config-gv99r"] Dec 13 07:11:23 crc kubenswrapper[4971]: I1213 07:11:23.735409 4971 generic.go:334] "Generic (PLEG): container finished" podID="5eed809d-c915-433c-9915-27e0b01e1ffe" containerID="0ecc7342a1a65b67ea97dd2575e3d43416e7633d3c5e40a8ec3ff6b6e6709a86" exitCode=0 Dec 13 07:11:23 crc kubenswrapper[4971]: I1213 07:11:23.735472 4971 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-cell1-server-0" event={"ID":"5eed809d-c915-433c-9915-27e0b01e1ffe","Type":"ContainerDied","Data":"0ecc7342a1a65b67ea97dd2575e3d43416e7633d3c5e40a8ec3ff6b6e6709a86"} Dec 13 07:11:23 crc kubenswrapper[4971]: I1213 07:11:23.738868 4971 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-controller-k5fz9-config-gv99r" event={"ID":"34f6d5f6-f9d5-4b88-bb8a-ad5e520673d9","Type":"ContainerStarted","Data":"f2fe21ad19c69cfeef6d7a24ff2a3d73950c5744333e407b485869a4ec26d067"} Dec 13 07:11:23 crc kubenswrapper[4971]: E1213 07:11:23.740397 4971 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"glance-db-sync\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/podified-antelope-centos9/openstack-glance-api@sha256:e4aa4ebbb1e581a12040e9ad2ae2709ac31b5d965bb64fc4252d1028b05c565f\\\"\"" pod="openstack/glance-db-sync-s24b9" podUID="6d483ba4-aee0-4c3b-a416-f3ee7ba19feb" Dec 13 07:11:24 crc kubenswrapper[4971]: I1213 07:11:24.749061 4971 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-cell1-server-0" event={"ID":"5eed809d-c915-433c-9915-27e0b01e1ffe","Type":"ContainerStarted","Data":"efed32724d55162da7a67edf770832ac164e124b6b49db25e2f7b8b7ec033c44"} Dec 13 07:11:24 crc kubenswrapper[4971]: I1213 07:11:24.749605 4971 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/rabbitmq-cell1-server-0" Dec 13 07:11:24 crc kubenswrapper[4971]: I1213 07:11:24.750864 4971 generic.go:334] "Generic (PLEG): container finished" podID="34f6d5f6-f9d5-4b88-bb8a-ad5e520673d9" containerID="ce0d05172c6bf7f2de3ae358f7e695533010db00a93a51ea0dcaacfe0ce9399f" exitCode=0 Dec 13 07:11:24 crc kubenswrapper[4971]: I1213 07:11:24.750904 4971 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-controller-k5fz9-config-gv99r" event={"ID":"34f6d5f6-f9d5-4b88-bb8a-ad5e520673d9","Type":"ContainerDied","Data":"ce0d05172c6bf7f2de3ae358f7e695533010db00a93a51ea0dcaacfe0ce9399f"} Dec 13 07:11:24 crc kubenswrapper[4971]: I1213 07:11:24.770347 4971 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/rabbitmq-cell1-server-0" podStartSLOduration=38.620072792 podStartE2EDuration="1m19.770329178s" podCreationTimestamp="2025-12-13 07:10:05 +0000 UTC" firstStartedPulling="2025-12-13 07:10:07.468601364 +0000 UTC m=+1264.073010812" lastFinishedPulling="2025-12-13 07:10:48.61885775 +0000 UTC m=+1305.223267198" observedRunningTime="2025-12-13 07:11:24.768096044 +0000 UTC m=+1341.372505532" watchObservedRunningTime="2025-12-13 07:11:24.770329178 +0000 UTC m=+1341.374738676" Dec 13 07:11:25 crc kubenswrapper[4971]: I1213 07:11:25.555940 4971 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/ovn-controller-k5fz9" Dec 13 07:11:26 crc kubenswrapper[4971]: I1213 07:11:26.064818 4971 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-controller-k5fz9-config-gv99r" Dec 13 07:11:26 crc kubenswrapper[4971]: I1213 07:11:26.223049 4971 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"var-log-ovn\" (UniqueName: \"kubernetes.io/host-path/34f6d5f6-f9d5-4b88-bb8a-ad5e520673d9-var-log-ovn\") pod \"34f6d5f6-f9d5-4b88-bb8a-ad5e520673d9\" (UID: \"34f6d5f6-f9d5-4b88-bb8a-ad5e520673d9\") " Dec 13 07:11:26 crc kubenswrapper[4971]: I1213 07:11:26.223115 4971 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"var-run-ovn\" (UniqueName: \"kubernetes.io/host-path/34f6d5f6-f9d5-4b88-bb8a-ad5e520673d9-var-run-ovn\") pod \"34f6d5f6-f9d5-4b88-bb8a-ad5e520673d9\" (UID: \"34f6d5f6-f9d5-4b88-bb8a-ad5e520673d9\") " Dec 13 07:11:26 crc kubenswrapper[4971]: I1213 07:11:26.223151 4971 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/34f6d5f6-f9d5-4b88-bb8a-ad5e520673d9-var-log-ovn" (OuterVolumeSpecName: "var-log-ovn") pod "34f6d5f6-f9d5-4b88-bb8a-ad5e520673d9" (UID: "34f6d5f6-f9d5-4b88-bb8a-ad5e520673d9"). InnerVolumeSpecName "var-log-ovn". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 13 07:11:26 crc kubenswrapper[4971]: I1213 07:11:26.223169 4971 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"additional-scripts\" (UniqueName: \"kubernetes.io/configmap/34f6d5f6-f9d5-4b88-bb8a-ad5e520673d9-additional-scripts\") pod \"34f6d5f6-f9d5-4b88-bb8a-ad5e520673d9\" (UID: \"34f6d5f6-f9d5-4b88-bb8a-ad5e520673d9\") " Dec 13 07:11:26 crc kubenswrapper[4971]: I1213 07:11:26.223213 4971 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/34f6d5f6-f9d5-4b88-bb8a-ad5e520673d9-var-run-ovn" (OuterVolumeSpecName: "var-run-ovn") pod "34f6d5f6-f9d5-4b88-bb8a-ad5e520673d9" (UID: "34f6d5f6-f9d5-4b88-bb8a-ad5e520673d9"). InnerVolumeSpecName "var-run-ovn". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 13 07:11:26 crc kubenswrapper[4971]: I1213 07:11:26.223359 4971 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/34f6d5f6-f9d5-4b88-bb8a-ad5e520673d9-scripts\") pod \"34f6d5f6-f9d5-4b88-bb8a-ad5e520673d9\" (UID: \"34f6d5f6-f9d5-4b88-bb8a-ad5e520673d9\") " Dec 13 07:11:26 crc kubenswrapper[4971]: I1213 07:11:26.223385 4971 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"var-run\" (UniqueName: \"kubernetes.io/host-path/34f6d5f6-f9d5-4b88-bb8a-ad5e520673d9-var-run\") pod \"34f6d5f6-f9d5-4b88-bb8a-ad5e520673d9\" (UID: \"34f6d5f6-f9d5-4b88-bb8a-ad5e520673d9\") " Dec 13 07:11:26 crc kubenswrapper[4971]: I1213 07:11:26.223490 4971 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/34f6d5f6-f9d5-4b88-bb8a-ad5e520673d9-var-run" (OuterVolumeSpecName: "var-run") pod "34f6d5f6-f9d5-4b88-bb8a-ad5e520673d9" (UID: "34f6d5f6-f9d5-4b88-bb8a-ad5e520673d9"). InnerVolumeSpecName "var-run". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 13 07:11:26 crc kubenswrapper[4971]: I1213 07:11:26.223568 4971 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-zltq9\" (UniqueName: \"kubernetes.io/projected/34f6d5f6-f9d5-4b88-bb8a-ad5e520673d9-kube-api-access-zltq9\") pod \"34f6d5f6-f9d5-4b88-bb8a-ad5e520673d9\" (UID: \"34f6d5f6-f9d5-4b88-bb8a-ad5e520673d9\") " Dec 13 07:11:26 crc kubenswrapper[4971]: I1213 07:11:26.224006 4971 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/34f6d5f6-f9d5-4b88-bb8a-ad5e520673d9-additional-scripts" (OuterVolumeSpecName: "additional-scripts") pod "34f6d5f6-f9d5-4b88-bb8a-ad5e520673d9" (UID: "34f6d5f6-f9d5-4b88-bb8a-ad5e520673d9"). InnerVolumeSpecName "additional-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 13 07:11:26 crc kubenswrapper[4971]: I1213 07:11:26.224195 4971 reconciler_common.go:293] "Volume detached for volume \"var-log-ovn\" (UniqueName: \"kubernetes.io/host-path/34f6d5f6-f9d5-4b88-bb8a-ad5e520673d9-var-log-ovn\") on node \"crc\" DevicePath \"\"" Dec 13 07:11:26 crc kubenswrapper[4971]: I1213 07:11:26.224219 4971 reconciler_common.go:293] "Volume detached for volume \"var-run-ovn\" (UniqueName: \"kubernetes.io/host-path/34f6d5f6-f9d5-4b88-bb8a-ad5e520673d9-var-run-ovn\") on node \"crc\" DevicePath \"\"" Dec 13 07:11:26 crc kubenswrapper[4971]: I1213 07:11:26.224230 4971 reconciler_common.go:293] "Volume detached for volume \"additional-scripts\" (UniqueName: \"kubernetes.io/configmap/34f6d5f6-f9d5-4b88-bb8a-ad5e520673d9-additional-scripts\") on node \"crc\" DevicePath \"\"" Dec 13 07:11:26 crc kubenswrapper[4971]: I1213 07:11:26.224245 4971 reconciler_common.go:293] "Volume detached for volume \"var-run\" (UniqueName: \"kubernetes.io/host-path/34f6d5f6-f9d5-4b88-bb8a-ad5e520673d9-var-run\") on node \"crc\" DevicePath \"\"" Dec 13 07:11:26 crc kubenswrapper[4971]: I1213 07:11:26.224279 4971 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/34f6d5f6-f9d5-4b88-bb8a-ad5e520673d9-scripts" (OuterVolumeSpecName: "scripts") pod "34f6d5f6-f9d5-4b88-bb8a-ad5e520673d9" (UID: "34f6d5f6-f9d5-4b88-bb8a-ad5e520673d9"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 13 07:11:26 crc kubenswrapper[4971]: I1213 07:11:26.228593 4971 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/34f6d5f6-f9d5-4b88-bb8a-ad5e520673d9-kube-api-access-zltq9" (OuterVolumeSpecName: "kube-api-access-zltq9") pod "34f6d5f6-f9d5-4b88-bb8a-ad5e520673d9" (UID: "34f6d5f6-f9d5-4b88-bb8a-ad5e520673d9"). InnerVolumeSpecName "kube-api-access-zltq9". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 13 07:11:26 crc kubenswrapper[4971]: I1213 07:11:26.325496 4971 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/34f6d5f6-f9d5-4b88-bb8a-ad5e520673d9-scripts\") on node \"crc\" DevicePath \"\"" Dec 13 07:11:26 crc kubenswrapper[4971]: I1213 07:11:26.325544 4971 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-zltq9\" (UniqueName: \"kubernetes.io/projected/34f6d5f6-f9d5-4b88-bb8a-ad5e520673d9-kube-api-access-zltq9\") on node \"crc\" DevicePath \"\"" Dec 13 07:11:26 crc kubenswrapper[4971]: I1213 07:11:26.773656 4971 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-controller-k5fz9-config-gv99r" event={"ID":"34f6d5f6-f9d5-4b88-bb8a-ad5e520673d9","Type":"ContainerDied","Data":"f2fe21ad19c69cfeef6d7a24ff2a3d73950c5744333e407b485869a4ec26d067"} Dec 13 07:11:26 crc kubenswrapper[4971]: I1213 07:11:26.773694 4971 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="f2fe21ad19c69cfeef6d7a24ff2a3d73950c5744333e407b485869a4ec26d067" Dec 13 07:11:26 crc kubenswrapper[4971]: I1213 07:11:26.773785 4971 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-controller-k5fz9-config-gv99r" Dec 13 07:11:27 crc kubenswrapper[4971]: I1213 07:11:27.182371 4971 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ovn-controller-k5fz9-config-gv99r"] Dec 13 07:11:27 crc kubenswrapper[4971]: I1213 07:11:27.193499 4971 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/ovn-controller-k5fz9-config-gv99r"] Dec 13 07:11:27 crc kubenswrapper[4971]: I1213 07:11:27.302915 4971 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ovn-controller-k5fz9-config-mhnrn"] Dec 13 07:11:27 crc kubenswrapper[4971]: E1213 07:11:27.303288 4971 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="34f6d5f6-f9d5-4b88-bb8a-ad5e520673d9" containerName="ovn-config" Dec 13 07:11:27 crc kubenswrapper[4971]: I1213 07:11:27.303309 4971 state_mem.go:107] "Deleted CPUSet assignment" podUID="34f6d5f6-f9d5-4b88-bb8a-ad5e520673d9" containerName="ovn-config" Dec 13 07:11:27 crc kubenswrapper[4971]: I1213 07:11:27.303548 4971 memory_manager.go:354] "RemoveStaleState removing state" podUID="34f6d5f6-f9d5-4b88-bb8a-ad5e520673d9" containerName="ovn-config" Dec 13 07:11:27 crc kubenswrapper[4971]: I1213 07:11:27.304234 4971 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-controller-k5fz9-config-mhnrn" Dec 13 07:11:27 crc kubenswrapper[4971]: I1213 07:11:27.306279 4971 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"ovncontroller-extra-scripts" Dec 13 07:11:27 crc kubenswrapper[4971]: I1213 07:11:27.320312 4971 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovn-controller-k5fz9-config-mhnrn"] Dec 13 07:11:27 crc kubenswrapper[4971]: I1213 07:11:27.444141 4971 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-run\" (UniqueName: \"kubernetes.io/host-path/6c4077ad-8789-433f-bf04-74026195e93d-var-run\") pod \"ovn-controller-k5fz9-config-mhnrn\" (UID: \"6c4077ad-8789-433f-bf04-74026195e93d\") " pod="openstack/ovn-controller-k5fz9-config-mhnrn" Dec 13 07:11:27 crc kubenswrapper[4971]: I1213 07:11:27.444206 4971 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"additional-scripts\" (UniqueName: \"kubernetes.io/configmap/6c4077ad-8789-433f-bf04-74026195e93d-additional-scripts\") pod \"ovn-controller-k5fz9-config-mhnrn\" (UID: \"6c4077ad-8789-433f-bf04-74026195e93d\") " pod="openstack/ovn-controller-k5fz9-config-mhnrn" Dec 13 07:11:27 crc kubenswrapper[4971]: I1213 07:11:27.444280 4971 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-t4ks8\" (UniqueName: \"kubernetes.io/projected/6c4077ad-8789-433f-bf04-74026195e93d-kube-api-access-t4ks8\") pod \"ovn-controller-k5fz9-config-mhnrn\" (UID: \"6c4077ad-8789-433f-bf04-74026195e93d\") " pod="openstack/ovn-controller-k5fz9-config-mhnrn" Dec 13 07:11:27 crc kubenswrapper[4971]: I1213 07:11:27.444308 4971 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-log-ovn\" (UniqueName: \"kubernetes.io/host-path/6c4077ad-8789-433f-bf04-74026195e93d-var-log-ovn\") pod \"ovn-controller-k5fz9-config-mhnrn\" (UID: \"6c4077ad-8789-433f-bf04-74026195e93d\") " pod="openstack/ovn-controller-k5fz9-config-mhnrn" Dec 13 07:11:27 crc kubenswrapper[4971]: I1213 07:11:27.444374 4971 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-run-ovn\" (UniqueName: \"kubernetes.io/host-path/6c4077ad-8789-433f-bf04-74026195e93d-var-run-ovn\") pod \"ovn-controller-k5fz9-config-mhnrn\" (UID: \"6c4077ad-8789-433f-bf04-74026195e93d\") " pod="openstack/ovn-controller-k5fz9-config-mhnrn" Dec 13 07:11:27 crc kubenswrapper[4971]: I1213 07:11:27.444419 4971 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/6c4077ad-8789-433f-bf04-74026195e93d-scripts\") pod \"ovn-controller-k5fz9-config-mhnrn\" (UID: \"6c4077ad-8789-433f-bf04-74026195e93d\") " pod="openstack/ovn-controller-k5fz9-config-mhnrn" Dec 13 07:11:27 crc kubenswrapper[4971]: I1213 07:11:27.546171 4971 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-run-ovn\" (UniqueName: \"kubernetes.io/host-path/6c4077ad-8789-433f-bf04-74026195e93d-var-run-ovn\") pod \"ovn-controller-k5fz9-config-mhnrn\" (UID: \"6c4077ad-8789-433f-bf04-74026195e93d\") " pod="openstack/ovn-controller-k5fz9-config-mhnrn" Dec 13 07:11:27 crc kubenswrapper[4971]: I1213 07:11:27.546234 4971 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/6c4077ad-8789-433f-bf04-74026195e93d-scripts\") pod \"ovn-controller-k5fz9-config-mhnrn\" (UID: \"6c4077ad-8789-433f-bf04-74026195e93d\") " pod="openstack/ovn-controller-k5fz9-config-mhnrn" Dec 13 07:11:27 crc kubenswrapper[4971]: I1213 07:11:27.546285 4971 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-run\" (UniqueName: \"kubernetes.io/host-path/6c4077ad-8789-433f-bf04-74026195e93d-var-run\") pod \"ovn-controller-k5fz9-config-mhnrn\" (UID: \"6c4077ad-8789-433f-bf04-74026195e93d\") " pod="openstack/ovn-controller-k5fz9-config-mhnrn" Dec 13 07:11:27 crc kubenswrapper[4971]: I1213 07:11:27.546308 4971 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"additional-scripts\" (UniqueName: \"kubernetes.io/configmap/6c4077ad-8789-433f-bf04-74026195e93d-additional-scripts\") pod \"ovn-controller-k5fz9-config-mhnrn\" (UID: \"6c4077ad-8789-433f-bf04-74026195e93d\") " pod="openstack/ovn-controller-k5fz9-config-mhnrn" Dec 13 07:11:27 crc kubenswrapper[4971]: I1213 07:11:27.546367 4971 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-t4ks8\" (UniqueName: \"kubernetes.io/projected/6c4077ad-8789-433f-bf04-74026195e93d-kube-api-access-t4ks8\") pod \"ovn-controller-k5fz9-config-mhnrn\" (UID: \"6c4077ad-8789-433f-bf04-74026195e93d\") " pod="openstack/ovn-controller-k5fz9-config-mhnrn" Dec 13 07:11:27 crc kubenswrapper[4971]: I1213 07:11:27.546392 4971 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-log-ovn\" (UniqueName: \"kubernetes.io/host-path/6c4077ad-8789-433f-bf04-74026195e93d-var-log-ovn\") pod \"ovn-controller-k5fz9-config-mhnrn\" (UID: \"6c4077ad-8789-433f-bf04-74026195e93d\") " pod="openstack/ovn-controller-k5fz9-config-mhnrn" Dec 13 07:11:27 crc kubenswrapper[4971]: I1213 07:11:27.546703 4971 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-log-ovn\" (UniqueName: \"kubernetes.io/host-path/6c4077ad-8789-433f-bf04-74026195e93d-var-log-ovn\") pod \"ovn-controller-k5fz9-config-mhnrn\" (UID: \"6c4077ad-8789-433f-bf04-74026195e93d\") " pod="openstack/ovn-controller-k5fz9-config-mhnrn" Dec 13 07:11:27 crc kubenswrapper[4971]: I1213 07:11:27.546746 4971 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-run\" (UniqueName: \"kubernetes.io/host-path/6c4077ad-8789-433f-bf04-74026195e93d-var-run\") pod \"ovn-controller-k5fz9-config-mhnrn\" (UID: \"6c4077ad-8789-433f-bf04-74026195e93d\") " pod="openstack/ovn-controller-k5fz9-config-mhnrn" Dec 13 07:11:27 crc kubenswrapper[4971]: I1213 07:11:27.546801 4971 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-run-ovn\" (UniqueName: \"kubernetes.io/host-path/6c4077ad-8789-433f-bf04-74026195e93d-var-run-ovn\") pod \"ovn-controller-k5fz9-config-mhnrn\" (UID: \"6c4077ad-8789-433f-bf04-74026195e93d\") " pod="openstack/ovn-controller-k5fz9-config-mhnrn" Dec 13 07:11:27 crc kubenswrapper[4971]: I1213 07:11:27.547340 4971 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"additional-scripts\" (UniqueName: \"kubernetes.io/configmap/6c4077ad-8789-433f-bf04-74026195e93d-additional-scripts\") pod \"ovn-controller-k5fz9-config-mhnrn\" (UID: \"6c4077ad-8789-433f-bf04-74026195e93d\") " pod="openstack/ovn-controller-k5fz9-config-mhnrn" Dec 13 07:11:27 crc kubenswrapper[4971]: I1213 07:11:27.549291 4971 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/6c4077ad-8789-433f-bf04-74026195e93d-scripts\") pod \"ovn-controller-k5fz9-config-mhnrn\" (UID: \"6c4077ad-8789-433f-bf04-74026195e93d\") " pod="openstack/ovn-controller-k5fz9-config-mhnrn" Dec 13 07:11:27 crc kubenswrapper[4971]: I1213 07:11:27.563266 4971 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-t4ks8\" (UniqueName: \"kubernetes.io/projected/6c4077ad-8789-433f-bf04-74026195e93d-kube-api-access-t4ks8\") pod \"ovn-controller-k5fz9-config-mhnrn\" (UID: \"6c4077ad-8789-433f-bf04-74026195e93d\") " pod="openstack/ovn-controller-k5fz9-config-mhnrn" Dec 13 07:11:27 crc kubenswrapper[4971]: I1213 07:11:27.625447 4971 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-controller-k5fz9-config-mhnrn" Dec 13 07:11:27 crc kubenswrapper[4971]: I1213 07:11:27.786182 4971 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="34f6d5f6-f9d5-4b88-bb8a-ad5e520673d9" path="/var/lib/kubelet/pods/34f6d5f6-f9d5-4b88-bb8a-ad5e520673d9/volumes" Dec 13 07:11:27 crc kubenswrapper[4971]: I1213 07:11:27.865317 4971 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovn-controller-k5fz9-config-mhnrn"] Dec 13 07:11:27 crc kubenswrapper[4971]: W1213 07:11:27.872763 4971 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod6c4077ad_8789_433f_bf04_74026195e93d.slice/crio-da73d257ce32c7d997d7be12a7784966715b1a67fb31c37195de1e5f5d148ca6 WatchSource:0}: Error finding container da73d257ce32c7d997d7be12a7784966715b1a67fb31c37195de1e5f5d148ca6: Status 404 returned error can't find the container with id da73d257ce32c7d997d7be12a7784966715b1a67fb31c37195de1e5f5d148ca6 Dec 13 07:11:28 crc kubenswrapper[4971]: I1213 07:11:28.793512 4971 generic.go:334] "Generic (PLEG): container finished" podID="6c4077ad-8789-433f-bf04-74026195e93d" containerID="804e6a307f26bd01f1e6f6fd6603421e30380f05416bb5c66264ede91de4fc73" exitCode=0 Dec 13 07:11:28 crc kubenswrapper[4971]: I1213 07:11:28.793579 4971 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-controller-k5fz9-config-mhnrn" event={"ID":"6c4077ad-8789-433f-bf04-74026195e93d","Type":"ContainerDied","Data":"804e6a307f26bd01f1e6f6fd6603421e30380f05416bb5c66264ede91de4fc73"} Dec 13 07:11:28 crc kubenswrapper[4971]: I1213 07:11:28.793770 4971 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-controller-k5fz9-config-mhnrn" event={"ID":"6c4077ad-8789-433f-bf04-74026195e93d","Type":"ContainerStarted","Data":"da73d257ce32c7d997d7be12a7784966715b1a67fb31c37195de1e5f5d148ca6"} Dec 13 07:11:30 crc kubenswrapper[4971]: I1213 07:11:30.119026 4971 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-controller-k5fz9-config-mhnrn" Dec 13 07:11:30 crc kubenswrapper[4971]: I1213 07:11:30.289650 4971 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"additional-scripts\" (UniqueName: \"kubernetes.io/configmap/6c4077ad-8789-433f-bf04-74026195e93d-additional-scripts\") pod \"6c4077ad-8789-433f-bf04-74026195e93d\" (UID: \"6c4077ad-8789-433f-bf04-74026195e93d\") " Dec 13 07:11:30 crc kubenswrapper[4971]: I1213 07:11:30.289699 4971 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-t4ks8\" (UniqueName: \"kubernetes.io/projected/6c4077ad-8789-433f-bf04-74026195e93d-kube-api-access-t4ks8\") pod \"6c4077ad-8789-433f-bf04-74026195e93d\" (UID: \"6c4077ad-8789-433f-bf04-74026195e93d\") " Dec 13 07:11:30 crc kubenswrapper[4971]: I1213 07:11:30.289748 4971 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"var-log-ovn\" (UniqueName: \"kubernetes.io/host-path/6c4077ad-8789-433f-bf04-74026195e93d-var-log-ovn\") pod \"6c4077ad-8789-433f-bf04-74026195e93d\" (UID: \"6c4077ad-8789-433f-bf04-74026195e93d\") " Dec 13 07:11:30 crc kubenswrapper[4971]: I1213 07:11:30.289777 4971 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/6c4077ad-8789-433f-bf04-74026195e93d-scripts\") pod \"6c4077ad-8789-433f-bf04-74026195e93d\" (UID: \"6c4077ad-8789-433f-bf04-74026195e93d\") " Dec 13 07:11:30 crc kubenswrapper[4971]: I1213 07:11:30.289922 4971 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/6c4077ad-8789-433f-bf04-74026195e93d-var-log-ovn" (OuterVolumeSpecName: "var-log-ovn") pod "6c4077ad-8789-433f-bf04-74026195e93d" (UID: "6c4077ad-8789-433f-bf04-74026195e93d"). InnerVolumeSpecName "var-log-ovn". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 13 07:11:30 crc kubenswrapper[4971]: I1213 07:11:30.289954 4971 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"var-run-ovn\" (UniqueName: \"kubernetes.io/host-path/6c4077ad-8789-433f-bf04-74026195e93d-var-run-ovn\") pod \"6c4077ad-8789-433f-bf04-74026195e93d\" (UID: \"6c4077ad-8789-433f-bf04-74026195e93d\") " Dec 13 07:11:30 crc kubenswrapper[4971]: I1213 07:11:30.290023 4971 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/6c4077ad-8789-433f-bf04-74026195e93d-var-run-ovn" (OuterVolumeSpecName: "var-run-ovn") pod "6c4077ad-8789-433f-bf04-74026195e93d" (UID: "6c4077ad-8789-433f-bf04-74026195e93d"). InnerVolumeSpecName "var-run-ovn". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 13 07:11:30 crc kubenswrapper[4971]: I1213 07:11:30.290042 4971 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"var-run\" (UniqueName: \"kubernetes.io/host-path/6c4077ad-8789-433f-bf04-74026195e93d-var-run\") pod \"6c4077ad-8789-433f-bf04-74026195e93d\" (UID: \"6c4077ad-8789-433f-bf04-74026195e93d\") " Dec 13 07:11:30 crc kubenswrapper[4971]: I1213 07:11:30.290120 4971 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/6c4077ad-8789-433f-bf04-74026195e93d-var-run" (OuterVolumeSpecName: "var-run") pod "6c4077ad-8789-433f-bf04-74026195e93d" (UID: "6c4077ad-8789-433f-bf04-74026195e93d"). InnerVolumeSpecName "var-run". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 13 07:11:30 crc kubenswrapper[4971]: I1213 07:11:30.290359 4971 reconciler_common.go:293] "Volume detached for volume \"var-log-ovn\" (UniqueName: \"kubernetes.io/host-path/6c4077ad-8789-433f-bf04-74026195e93d-var-log-ovn\") on node \"crc\" DevicePath \"\"" Dec 13 07:11:30 crc kubenswrapper[4971]: I1213 07:11:30.290372 4971 reconciler_common.go:293] "Volume detached for volume \"var-run-ovn\" (UniqueName: \"kubernetes.io/host-path/6c4077ad-8789-433f-bf04-74026195e93d-var-run-ovn\") on node \"crc\" DevicePath \"\"" Dec 13 07:11:30 crc kubenswrapper[4971]: I1213 07:11:30.290380 4971 reconciler_common.go:293] "Volume detached for volume \"var-run\" (UniqueName: \"kubernetes.io/host-path/6c4077ad-8789-433f-bf04-74026195e93d-var-run\") on node \"crc\" DevicePath \"\"" Dec 13 07:11:30 crc kubenswrapper[4971]: I1213 07:11:30.290679 4971 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/6c4077ad-8789-433f-bf04-74026195e93d-additional-scripts" (OuterVolumeSpecName: "additional-scripts") pod "6c4077ad-8789-433f-bf04-74026195e93d" (UID: "6c4077ad-8789-433f-bf04-74026195e93d"). InnerVolumeSpecName "additional-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 13 07:11:30 crc kubenswrapper[4971]: I1213 07:11:30.290975 4971 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/6c4077ad-8789-433f-bf04-74026195e93d-scripts" (OuterVolumeSpecName: "scripts") pod "6c4077ad-8789-433f-bf04-74026195e93d" (UID: "6c4077ad-8789-433f-bf04-74026195e93d"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 13 07:11:30 crc kubenswrapper[4971]: I1213 07:11:30.297739 4971 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/6c4077ad-8789-433f-bf04-74026195e93d-kube-api-access-t4ks8" (OuterVolumeSpecName: "kube-api-access-t4ks8") pod "6c4077ad-8789-433f-bf04-74026195e93d" (UID: "6c4077ad-8789-433f-bf04-74026195e93d"). InnerVolumeSpecName "kube-api-access-t4ks8". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 13 07:11:30 crc kubenswrapper[4971]: I1213 07:11:30.391805 4971 reconciler_common.go:293] "Volume detached for volume \"additional-scripts\" (UniqueName: \"kubernetes.io/configmap/6c4077ad-8789-433f-bf04-74026195e93d-additional-scripts\") on node \"crc\" DevicePath \"\"" Dec 13 07:11:30 crc kubenswrapper[4971]: I1213 07:11:30.391859 4971 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-t4ks8\" (UniqueName: \"kubernetes.io/projected/6c4077ad-8789-433f-bf04-74026195e93d-kube-api-access-t4ks8\") on node \"crc\" DevicePath \"\"" Dec 13 07:11:30 crc kubenswrapper[4971]: I1213 07:11:30.391881 4971 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/6c4077ad-8789-433f-bf04-74026195e93d-scripts\") on node \"crc\" DevicePath \"\"" Dec 13 07:11:30 crc kubenswrapper[4971]: I1213 07:11:30.810224 4971 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-controller-k5fz9-config-mhnrn" event={"ID":"6c4077ad-8789-433f-bf04-74026195e93d","Type":"ContainerDied","Data":"da73d257ce32c7d997d7be12a7784966715b1a67fb31c37195de1e5f5d148ca6"} Dec 13 07:11:30 crc kubenswrapper[4971]: I1213 07:11:30.810479 4971 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="da73d257ce32c7d997d7be12a7784966715b1a67fb31c37195de1e5f5d148ca6" Dec 13 07:11:30 crc kubenswrapper[4971]: I1213 07:11:30.810312 4971 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-controller-k5fz9-config-mhnrn" Dec 13 07:11:31 crc kubenswrapper[4971]: I1213 07:11:31.197105 4971 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ovn-controller-k5fz9-config-mhnrn"] Dec 13 07:11:31 crc kubenswrapper[4971]: I1213 07:11:31.205248 4971 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/ovn-controller-k5fz9-config-mhnrn"] Dec 13 07:11:31 crc kubenswrapper[4971]: I1213 07:11:31.787001 4971 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="6c4077ad-8789-433f-bf04-74026195e93d" path="/var/lib/kubelet/pods/6c4077ad-8789-433f-bf04-74026195e93d/volumes" Dec 13 07:11:31 crc kubenswrapper[4971]: I1213 07:11:31.819080 4971 generic.go:334] "Generic (PLEG): container finished" podID="a590eb84-c8d5-4cbe-9c03-03a6ccb82073" containerID="2632986cd6a0be018d82fef579687199e356535dfe73855c9af020cc784bc892" exitCode=0 Dec 13 07:11:31 crc kubenswrapper[4971]: I1213 07:11:31.819140 4971 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-server-0" event={"ID":"a590eb84-c8d5-4cbe-9c03-03a6ccb82073","Type":"ContainerDied","Data":"2632986cd6a0be018d82fef579687199e356535dfe73855c9af020cc784bc892"} Dec 13 07:11:32 crc kubenswrapper[4971]: I1213 07:11:32.830863 4971 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-server-0" event={"ID":"a590eb84-c8d5-4cbe-9c03-03a6ccb82073","Type":"ContainerStarted","Data":"75710b914a6fcd71bf7b0cc53696f512ac865a4da1eb06734a6db6ff35c892ad"} Dec 13 07:11:32 crc kubenswrapper[4971]: I1213 07:11:32.833123 4971 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/rabbitmq-server-0" Dec 13 07:11:32 crc kubenswrapper[4971]: I1213 07:11:32.862440 4971 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/rabbitmq-server-0" podStartSLOduration=47.536787939999996 podStartE2EDuration="1m28.862420444s" podCreationTimestamp="2025-12-13 07:10:04 +0000 UTC" firstStartedPulling="2025-12-13 07:10:07.317279368 +0000 UTC m=+1263.921688816" lastFinishedPulling="2025-12-13 07:10:48.642911872 +0000 UTC m=+1305.247321320" observedRunningTime="2025-12-13 07:11:32.856054837 +0000 UTC m=+1349.460464305" watchObservedRunningTime="2025-12-13 07:11:32.862420444 +0000 UTC m=+1349.466829892" Dec 13 07:11:35 crc kubenswrapper[4971]: I1213 07:11:35.852407 4971 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-db-sync-s24b9" event={"ID":"6d483ba4-aee0-4c3b-a416-f3ee7ba19feb","Type":"ContainerStarted","Data":"74efeeab58249e168c5ec8b783a7a55a261ee15a5da946cae38e1d8ca17d36bc"} Dec 13 07:11:35 crc kubenswrapper[4971]: I1213 07:11:35.869820 4971 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/glance-db-sync-s24b9" podStartSLOduration=2.333045929 podStartE2EDuration="30.869796915s" podCreationTimestamp="2025-12-13 07:11:05 +0000 UTC" firstStartedPulling="2025-12-13 07:11:06.729791757 +0000 UTC m=+1323.334201205" lastFinishedPulling="2025-12-13 07:11:35.266542743 +0000 UTC m=+1351.870952191" observedRunningTime="2025-12-13 07:11:35.869328843 +0000 UTC m=+1352.473738291" watchObservedRunningTime="2025-12-13 07:11:35.869796915 +0000 UTC m=+1352.474206363" Dec 13 07:11:36 crc kubenswrapper[4971]: I1213 07:11:36.747693 4971 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/rabbitmq-cell1-server-0" Dec 13 07:11:42 crc kubenswrapper[4971]: I1213 07:11:42.921151 4971 generic.go:334] "Generic (PLEG): container finished" podID="6d483ba4-aee0-4c3b-a416-f3ee7ba19feb" containerID="74efeeab58249e168c5ec8b783a7a55a261ee15a5da946cae38e1d8ca17d36bc" exitCode=0 Dec 13 07:11:42 crc kubenswrapper[4971]: I1213 07:11:42.921416 4971 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-db-sync-s24b9" event={"ID":"6d483ba4-aee0-4c3b-a416-f3ee7ba19feb","Type":"ContainerDied","Data":"74efeeab58249e168c5ec8b783a7a55a261ee15a5da946cae38e1d8ca17d36bc"} Dec 13 07:11:44 crc kubenswrapper[4971]: I1213 07:11:44.307632 4971 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/glance-db-sync-s24b9" Dec 13 07:11:44 crc kubenswrapper[4971]: I1213 07:11:44.411021 4971 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/6d483ba4-aee0-4c3b-a416-f3ee7ba19feb-combined-ca-bundle\") pod \"6d483ba4-aee0-4c3b-a416-f3ee7ba19feb\" (UID: \"6d483ba4-aee0-4c3b-a416-f3ee7ba19feb\") " Dec 13 07:11:44 crc kubenswrapper[4971]: I1213 07:11:44.411088 4971 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-nqfn5\" (UniqueName: \"kubernetes.io/projected/6d483ba4-aee0-4c3b-a416-f3ee7ba19feb-kube-api-access-nqfn5\") pod \"6d483ba4-aee0-4c3b-a416-f3ee7ba19feb\" (UID: \"6d483ba4-aee0-4c3b-a416-f3ee7ba19feb\") " Dec 13 07:11:44 crc kubenswrapper[4971]: I1213 07:11:44.411143 4971 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/6d483ba4-aee0-4c3b-a416-f3ee7ba19feb-db-sync-config-data\") pod \"6d483ba4-aee0-4c3b-a416-f3ee7ba19feb\" (UID: \"6d483ba4-aee0-4c3b-a416-f3ee7ba19feb\") " Dec 13 07:11:44 crc kubenswrapper[4971]: I1213 07:11:44.411190 4971 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/6d483ba4-aee0-4c3b-a416-f3ee7ba19feb-config-data\") pod \"6d483ba4-aee0-4c3b-a416-f3ee7ba19feb\" (UID: \"6d483ba4-aee0-4c3b-a416-f3ee7ba19feb\") " Dec 13 07:11:44 crc kubenswrapper[4971]: I1213 07:11:44.417044 4971 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/6d483ba4-aee0-4c3b-a416-f3ee7ba19feb-kube-api-access-nqfn5" (OuterVolumeSpecName: "kube-api-access-nqfn5") pod "6d483ba4-aee0-4c3b-a416-f3ee7ba19feb" (UID: "6d483ba4-aee0-4c3b-a416-f3ee7ba19feb"). InnerVolumeSpecName "kube-api-access-nqfn5". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 13 07:11:44 crc kubenswrapper[4971]: I1213 07:11:44.417116 4971 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/6d483ba4-aee0-4c3b-a416-f3ee7ba19feb-db-sync-config-data" (OuterVolumeSpecName: "db-sync-config-data") pod "6d483ba4-aee0-4c3b-a416-f3ee7ba19feb" (UID: "6d483ba4-aee0-4c3b-a416-f3ee7ba19feb"). InnerVolumeSpecName "db-sync-config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 13 07:11:44 crc kubenswrapper[4971]: I1213 07:11:44.435730 4971 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/6d483ba4-aee0-4c3b-a416-f3ee7ba19feb-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "6d483ba4-aee0-4c3b-a416-f3ee7ba19feb" (UID: "6d483ba4-aee0-4c3b-a416-f3ee7ba19feb"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 13 07:11:44 crc kubenswrapper[4971]: I1213 07:11:44.453916 4971 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/6d483ba4-aee0-4c3b-a416-f3ee7ba19feb-config-data" (OuterVolumeSpecName: "config-data") pod "6d483ba4-aee0-4c3b-a416-f3ee7ba19feb" (UID: "6d483ba4-aee0-4c3b-a416-f3ee7ba19feb"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 13 07:11:44 crc kubenswrapper[4971]: I1213 07:11:44.513254 4971 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/6d483ba4-aee0-4c3b-a416-f3ee7ba19feb-config-data\") on node \"crc\" DevicePath \"\"" Dec 13 07:11:44 crc kubenswrapper[4971]: I1213 07:11:44.513290 4971 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/6d483ba4-aee0-4c3b-a416-f3ee7ba19feb-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 13 07:11:44 crc kubenswrapper[4971]: I1213 07:11:44.513305 4971 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-nqfn5\" (UniqueName: \"kubernetes.io/projected/6d483ba4-aee0-4c3b-a416-f3ee7ba19feb-kube-api-access-nqfn5\") on node \"crc\" DevicePath \"\"" Dec 13 07:11:44 crc kubenswrapper[4971]: I1213 07:11:44.513319 4971 reconciler_common.go:293] "Volume detached for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/6d483ba4-aee0-4c3b-a416-f3ee7ba19feb-db-sync-config-data\") on node \"crc\" DevicePath \"\"" Dec 13 07:11:44 crc kubenswrapper[4971]: I1213 07:11:44.940365 4971 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-db-sync-s24b9" event={"ID":"6d483ba4-aee0-4c3b-a416-f3ee7ba19feb","Type":"ContainerDied","Data":"8708a88bcce239caff9aab875303eb33006234c20b442a2d0d6e97ee0042683a"} Dec 13 07:11:44 crc kubenswrapper[4971]: I1213 07:11:44.940402 4971 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/glance-db-sync-s24b9" Dec 13 07:11:44 crc kubenswrapper[4971]: I1213 07:11:44.940421 4971 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="8708a88bcce239caff9aab875303eb33006234c20b442a2d0d6e97ee0042683a" Dec 13 07:11:45 crc kubenswrapper[4971]: I1213 07:11:45.202016 4971 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-57768dd7b5-86hlw"] Dec 13 07:11:45 crc kubenswrapper[4971]: E1213 07:11:45.203148 4971 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="6d483ba4-aee0-4c3b-a416-f3ee7ba19feb" containerName="glance-db-sync" Dec 13 07:11:45 crc kubenswrapper[4971]: I1213 07:11:45.203197 4971 state_mem.go:107] "Deleted CPUSet assignment" podUID="6d483ba4-aee0-4c3b-a416-f3ee7ba19feb" containerName="glance-db-sync" Dec 13 07:11:45 crc kubenswrapper[4971]: E1213 07:11:45.203228 4971 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="6c4077ad-8789-433f-bf04-74026195e93d" containerName="ovn-config" Dec 13 07:11:45 crc kubenswrapper[4971]: I1213 07:11:45.203237 4971 state_mem.go:107] "Deleted CPUSet assignment" podUID="6c4077ad-8789-433f-bf04-74026195e93d" containerName="ovn-config" Dec 13 07:11:45 crc kubenswrapper[4971]: I1213 07:11:45.207378 4971 memory_manager.go:354] "RemoveStaleState removing state" podUID="6c4077ad-8789-433f-bf04-74026195e93d" containerName="ovn-config" Dec 13 07:11:45 crc kubenswrapper[4971]: I1213 07:11:45.207435 4971 memory_manager.go:354] "RemoveStaleState removing state" podUID="6d483ba4-aee0-4c3b-a416-f3ee7ba19feb" containerName="glance-db-sync" Dec 13 07:11:45 crc kubenswrapper[4971]: I1213 07:11:45.208565 4971 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-57768dd7b5-86hlw" Dec 13 07:11:45 crc kubenswrapper[4971]: I1213 07:11:45.214610 4971 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-57768dd7b5-86hlw"] Dec 13 07:11:45 crc kubenswrapper[4971]: I1213 07:11:45.327833 4971 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/c9e30081-d603-4c9d-81ce-7bcff6848243-config\") pod \"dnsmasq-dns-57768dd7b5-86hlw\" (UID: \"c9e30081-d603-4c9d-81ce-7bcff6848243\") " pod="openstack/dnsmasq-dns-57768dd7b5-86hlw" Dec 13 07:11:45 crc kubenswrapper[4971]: I1213 07:11:45.327898 4971 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/c9e30081-d603-4c9d-81ce-7bcff6848243-dns-svc\") pod \"dnsmasq-dns-57768dd7b5-86hlw\" (UID: \"c9e30081-d603-4c9d-81ce-7bcff6848243\") " pod="openstack/dnsmasq-dns-57768dd7b5-86hlw" Dec 13 07:11:45 crc kubenswrapper[4971]: I1213 07:11:45.327985 4971 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/c9e30081-d603-4c9d-81ce-7bcff6848243-ovsdbserver-nb\") pod \"dnsmasq-dns-57768dd7b5-86hlw\" (UID: \"c9e30081-d603-4c9d-81ce-7bcff6848243\") " pod="openstack/dnsmasq-dns-57768dd7b5-86hlw" Dec 13 07:11:45 crc kubenswrapper[4971]: I1213 07:11:45.328037 4971 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-ccks5\" (UniqueName: \"kubernetes.io/projected/c9e30081-d603-4c9d-81ce-7bcff6848243-kube-api-access-ccks5\") pod \"dnsmasq-dns-57768dd7b5-86hlw\" (UID: \"c9e30081-d603-4c9d-81ce-7bcff6848243\") " pod="openstack/dnsmasq-dns-57768dd7b5-86hlw" Dec 13 07:11:45 crc kubenswrapper[4971]: I1213 07:11:45.328123 4971 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/c9e30081-d603-4c9d-81ce-7bcff6848243-ovsdbserver-sb\") pod \"dnsmasq-dns-57768dd7b5-86hlw\" (UID: \"c9e30081-d603-4c9d-81ce-7bcff6848243\") " pod="openstack/dnsmasq-dns-57768dd7b5-86hlw" Dec 13 07:11:45 crc kubenswrapper[4971]: I1213 07:11:45.429449 4971 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/c9e30081-d603-4c9d-81ce-7bcff6848243-config\") pod \"dnsmasq-dns-57768dd7b5-86hlw\" (UID: \"c9e30081-d603-4c9d-81ce-7bcff6848243\") " pod="openstack/dnsmasq-dns-57768dd7b5-86hlw" Dec 13 07:11:45 crc kubenswrapper[4971]: I1213 07:11:45.429512 4971 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/c9e30081-d603-4c9d-81ce-7bcff6848243-dns-svc\") pod \"dnsmasq-dns-57768dd7b5-86hlw\" (UID: \"c9e30081-d603-4c9d-81ce-7bcff6848243\") " pod="openstack/dnsmasq-dns-57768dd7b5-86hlw" Dec 13 07:11:45 crc kubenswrapper[4971]: I1213 07:11:45.429571 4971 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/c9e30081-d603-4c9d-81ce-7bcff6848243-ovsdbserver-nb\") pod \"dnsmasq-dns-57768dd7b5-86hlw\" (UID: \"c9e30081-d603-4c9d-81ce-7bcff6848243\") " pod="openstack/dnsmasq-dns-57768dd7b5-86hlw" Dec 13 07:11:45 crc kubenswrapper[4971]: I1213 07:11:45.429612 4971 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-ccks5\" (UniqueName: \"kubernetes.io/projected/c9e30081-d603-4c9d-81ce-7bcff6848243-kube-api-access-ccks5\") pod \"dnsmasq-dns-57768dd7b5-86hlw\" (UID: \"c9e30081-d603-4c9d-81ce-7bcff6848243\") " pod="openstack/dnsmasq-dns-57768dd7b5-86hlw" Dec 13 07:11:45 crc kubenswrapper[4971]: I1213 07:11:45.429648 4971 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/c9e30081-d603-4c9d-81ce-7bcff6848243-ovsdbserver-sb\") pod \"dnsmasq-dns-57768dd7b5-86hlw\" (UID: \"c9e30081-d603-4c9d-81ce-7bcff6848243\") " pod="openstack/dnsmasq-dns-57768dd7b5-86hlw" Dec 13 07:11:45 crc kubenswrapper[4971]: I1213 07:11:45.430802 4971 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/c9e30081-d603-4c9d-81ce-7bcff6848243-ovsdbserver-nb\") pod \"dnsmasq-dns-57768dd7b5-86hlw\" (UID: \"c9e30081-d603-4c9d-81ce-7bcff6848243\") " pod="openstack/dnsmasq-dns-57768dd7b5-86hlw" Dec 13 07:11:45 crc kubenswrapper[4971]: I1213 07:11:45.430820 4971 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/c9e30081-d603-4c9d-81ce-7bcff6848243-dns-svc\") pod \"dnsmasq-dns-57768dd7b5-86hlw\" (UID: \"c9e30081-d603-4c9d-81ce-7bcff6848243\") " pod="openstack/dnsmasq-dns-57768dd7b5-86hlw" Dec 13 07:11:45 crc kubenswrapper[4971]: I1213 07:11:45.430848 4971 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/c9e30081-d603-4c9d-81ce-7bcff6848243-config\") pod \"dnsmasq-dns-57768dd7b5-86hlw\" (UID: \"c9e30081-d603-4c9d-81ce-7bcff6848243\") " pod="openstack/dnsmasq-dns-57768dd7b5-86hlw" Dec 13 07:11:45 crc kubenswrapper[4971]: I1213 07:11:45.430824 4971 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/c9e30081-d603-4c9d-81ce-7bcff6848243-ovsdbserver-sb\") pod \"dnsmasq-dns-57768dd7b5-86hlw\" (UID: \"c9e30081-d603-4c9d-81ce-7bcff6848243\") " pod="openstack/dnsmasq-dns-57768dd7b5-86hlw" Dec 13 07:11:45 crc kubenswrapper[4971]: I1213 07:11:45.447404 4971 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-ccks5\" (UniqueName: \"kubernetes.io/projected/c9e30081-d603-4c9d-81ce-7bcff6848243-kube-api-access-ccks5\") pod \"dnsmasq-dns-57768dd7b5-86hlw\" (UID: \"c9e30081-d603-4c9d-81ce-7bcff6848243\") " pod="openstack/dnsmasq-dns-57768dd7b5-86hlw" Dec 13 07:11:45 crc kubenswrapper[4971]: I1213 07:11:45.577830 4971 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-57768dd7b5-86hlw" Dec 13 07:11:46 crc kubenswrapper[4971]: I1213 07:11:46.029797 4971 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-57768dd7b5-86hlw"] Dec 13 07:11:46 crc kubenswrapper[4971]: I1213 07:11:46.153183 4971 patch_prober.go:28] interesting pod/machine-config-daemon-82xjz container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 13 07:11:46 crc kubenswrapper[4971]: I1213 07:11:46.153422 4971 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-82xjz" podUID="e0c4af12-edda-41d7-9b44-f87396174ef2" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 13 07:11:46 crc kubenswrapper[4971]: I1213 07:11:46.153473 4971 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-82xjz" Dec 13 07:11:46 crc kubenswrapper[4971]: I1213 07:11:46.154124 4971 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"6543005a61441f4a045d5ad167cc1798a7b51524223dc500db29c8f9518b38f1"} pod="openshift-machine-config-operator/machine-config-daemon-82xjz" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Dec 13 07:11:46 crc kubenswrapper[4971]: I1213 07:11:46.154199 4971 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-82xjz" podUID="e0c4af12-edda-41d7-9b44-f87396174ef2" containerName="machine-config-daemon" containerID="cri-o://6543005a61441f4a045d5ad167cc1798a7b51524223dc500db29c8f9518b38f1" gracePeriod=600 Dec 13 07:11:46 crc kubenswrapper[4971]: I1213 07:11:46.317753 4971 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/rabbitmq-server-0" Dec 13 07:11:46 crc kubenswrapper[4971]: I1213 07:11:46.677010 4971 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/cinder-db-create-8sbpv"] Dec 13 07:11:46 crc kubenswrapper[4971]: I1213 07:11:46.678246 4971 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-db-create-8sbpv" Dec 13 07:11:46 crc kubenswrapper[4971]: I1213 07:11:46.690387 4971 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cinder-db-create-8sbpv"] Dec 13 07:11:46 crc kubenswrapper[4971]: I1213 07:11:46.802909 4971 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/barbican-5251-account-create-update-dr8tv"] Dec 13 07:11:46 crc kubenswrapper[4971]: I1213 07:11:46.804010 4971 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-5251-account-create-update-dr8tv" Dec 13 07:11:46 crc kubenswrapper[4971]: I1213 07:11:46.816933 4971 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/barbican-db-create-5czp4"] Dec 13 07:11:46 crc kubenswrapper[4971]: I1213 07:11:46.817758 4971 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"barbican-db-secret" Dec 13 07:11:46 crc kubenswrapper[4971]: I1213 07:11:46.817879 4971 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-db-create-5czp4" Dec 13 07:11:46 crc kubenswrapper[4971]: I1213 07:11:46.829386 4971 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/barbican-db-create-5czp4"] Dec 13 07:11:46 crc kubenswrapper[4971]: I1213 07:11:46.843337 4971 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/barbican-5251-account-create-update-dr8tv"] Dec 13 07:11:46 crc kubenswrapper[4971]: I1213 07:11:46.857977 4971 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-4zz7m\" (UniqueName: \"kubernetes.io/projected/4c7ef6d3-1b68-4fb6-9784-0ee0bb154a8d-kube-api-access-4zz7m\") pod \"cinder-db-create-8sbpv\" (UID: \"4c7ef6d3-1b68-4fb6-9784-0ee0bb154a8d\") " pod="openstack/cinder-db-create-8sbpv" Dec 13 07:11:46 crc kubenswrapper[4971]: I1213 07:11:46.858069 4971 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/4c7ef6d3-1b68-4fb6-9784-0ee0bb154a8d-operator-scripts\") pod \"cinder-db-create-8sbpv\" (UID: \"4c7ef6d3-1b68-4fb6-9784-0ee0bb154a8d\") " pod="openstack/cinder-db-create-8sbpv" Dec 13 07:11:46 crc kubenswrapper[4971]: I1213 07:11:46.910541 4971 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/cinder-51d6-account-create-update-hm8kr"] Dec 13 07:11:46 crc kubenswrapper[4971]: I1213 07:11:46.911744 4971 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-51d6-account-create-update-hm8kr" Dec 13 07:11:46 crc kubenswrapper[4971]: I1213 07:11:46.917203 4971 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cinder-db-secret" Dec 13 07:11:46 crc kubenswrapper[4971]: I1213 07:11:46.934760 4971 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cinder-51d6-account-create-update-hm8kr"] Dec 13 07:11:46 crc kubenswrapper[4971]: I1213 07:11:46.959477 4971 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/48a5104a-2e2e-4828-bb54-bf40553d7ee0-operator-scripts\") pod \"barbican-5251-account-create-update-dr8tv\" (UID: \"48a5104a-2e2e-4828-bb54-bf40553d7ee0\") " pod="openstack/barbican-5251-account-create-update-dr8tv" Dec 13 07:11:46 crc kubenswrapper[4971]: I1213 07:11:46.961222 4971 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-h7b8r\" (UniqueName: \"kubernetes.io/projected/f136ffc5-3a49-4807-8025-10356e78ee46-kube-api-access-h7b8r\") pod \"barbican-db-create-5czp4\" (UID: \"f136ffc5-3a49-4807-8025-10356e78ee46\") " pod="openstack/barbican-db-create-5czp4" Dec 13 07:11:46 crc kubenswrapper[4971]: I1213 07:11:46.961292 4971 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/f136ffc5-3a49-4807-8025-10356e78ee46-operator-scripts\") pod \"barbican-db-create-5czp4\" (UID: \"f136ffc5-3a49-4807-8025-10356e78ee46\") " pod="openstack/barbican-db-create-5czp4" Dec 13 07:11:46 crc kubenswrapper[4971]: I1213 07:11:46.961349 4971 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-4zz7m\" (UniqueName: \"kubernetes.io/projected/4c7ef6d3-1b68-4fb6-9784-0ee0bb154a8d-kube-api-access-4zz7m\") pod \"cinder-db-create-8sbpv\" (UID: \"4c7ef6d3-1b68-4fb6-9784-0ee0bb154a8d\") " pod="openstack/cinder-db-create-8sbpv" Dec 13 07:11:46 crc kubenswrapper[4971]: I1213 07:11:46.961465 4971 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/4c7ef6d3-1b68-4fb6-9784-0ee0bb154a8d-operator-scripts\") pod \"cinder-db-create-8sbpv\" (UID: \"4c7ef6d3-1b68-4fb6-9784-0ee0bb154a8d\") " pod="openstack/cinder-db-create-8sbpv" Dec 13 07:11:46 crc kubenswrapper[4971]: I1213 07:11:46.961505 4971 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-kp2s8\" (UniqueName: \"kubernetes.io/projected/48a5104a-2e2e-4828-bb54-bf40553d7ee0-kube-api-access-kp2s8\") pod \"barbican-5251-account-create-update-dr8tv\" (UID: \"48a5104a-2e2e-4828-bb54-bf40553d7ee0\") " pod="openstack/barbican-5251-account-create-update-dr8tv" Dec 13 07:11:46 crc kubenswrapper[4971]: I1213 07:11:46.962808 4971 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/4c7ef6d3-1b68-4fb6-9784-0ee0bb154a8d-operator-scripts\") pod \"cinder-db-create-8sbpv\" (UID: \"4c7ef6d3-1b68-4fb6-9784-0ee0bb154a8d\") " pod="openstack/cinder-db-create-8sbpv" Dec 13 07:11:46 crc kubenswrapper[4971]: I1213 07:11:46.962894 4971 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/keystone-db-sync-b2cmm"] Dec 13 07:11:46 crc kubenswrapper[4971]: I1213 07:11:46.963828 4971 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-db-sync-b2cmm" Dec 13 07:11:46 crc kubenswrapper[4971]: I1213 07:11:46.976675 4971 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone-config-data" Dec 13 07:11:46 crc kubenswrapper[4971]: I1213 07:11:46.976704 4971 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone-scripts" Dec 13 07:11:46 crc kubenswrapper[4971]: I1213 07:11:46.976878 4971 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone" Dec 13 07:11:46 crc kubenswrapper[4971]: I1213 07:11:46.976968 4971 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone-keystone-dockercfg-56xxb" Dec 13 07:11:46 crc kubenswrapper[4971]: I1213 07:11:46.990790 4971 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/keystone-db-sync-b2cmm"] Dec 13 07:11:46 crc kubenswrapper[4971]: I1213 07:11:46.996785 4971 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-4zz7m\" (UniqueName: \"kubernetes.io/projected/4c7ef6d3-1b68-4fb6-9784-0ee0bb154a8d-kube-api-access-4zz7m\") pod \"cinder-db-create-8sbpv\" (UID: \"4c7ef6d3-1b68-4fb6-9784-0ee0bb154a8d\") " pod="openstack/cinder-db-create-8sbpv" Dec 13 07:11:46 crc kubenswrapper[4971]: I1213 07:11:46.998204 4971 generic.go:334] "Generic (PLEG): container finished" podID="e0c4af12-edda-41d7-9b44-f87396174ef2" containerID="6543005a61441f4a045d5ad167cc1798a7b51524223dc500db29c8f9518b38f1" exitCode=0 Dec 13 07:11:46 crc kubenswrapper[4971]: I1213 07:11:46.998283 4971 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-82xjz" event={"ID":"e0c4af12-edda-41d7-9b44-f87396174ef2","Type":"ContainerDied","Data":"6543005a61441f4a045d5ad167cc1798a7b51524223dc500db29c8f9518b38f1"} Dec 13 07:11:46 crc kubenswrapper[4971]: I1213 07:11:46.998334 4971 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-82xjz" event={"ID":"e0c4af12-edda-41d7-9b44-f87396174ef2","Type":"ContainerStarted","Data":"e8beee43d15719dccb6892d6f5711674492425300efa098be89033bc65977d14"} Dec 13 07:11:46 crc kubenswrapper[4971]: I1213 07:11:46.998351 4971 scope.go:117] "RemoveContainer" containerID="21a30fd79dbe0c677372d180191864bac6880bbd1aaac9bae997b690cf3d4a99" Dec 13 07:11:47 crc kubenswrapper[4971]: I1213 07:11:47.015017 4971 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-db-create-8sbpv" Dec 13 07:11:47 crc kubenswrapper[4971]: I1213 07:11:47.023192 4971 generic.go:334] "Generic (PLEG): container finished" podID="c9e30081-d603-4c9d-81ce-7bcff6848243" containerID="c707864d3136ddbe0a0fb0f1eb6f1a2bdcc6c2f0651b9e51d1d68857e280fb88" exitCode=0 Dec 13 07:11:47 crc kubenswrapper[4971]: I1213 07:11:47.023250 4971 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-57768dd7b5-86hlw" event={"ID":"c9e30081-d603-4c9d-81ce-7bcff6848243","Type":"ContainerDied","Data":"c707864d3136ddbe0a0fb0f1eb6f1a2bdcc6c2f0651b9e51d1d68857e280fb88"} Dec 13 07:11:47 crc kubenswrapper[4971]: I1213 07:11:47.023295 4971 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-57768dd7b5-86hlw" event={"ID":"c9e30081-d603-4c9d-81ce-7bcff6848243","Type":"ContainerStarted","Data":"93dcd74806ab03a60c4e4da3b8b953c968fc6cc7e8ce91566c8874b613715598"} Dec 13 07:11:47 crc kubenswrapper[4971]: I1213 07:11:47.064187 4971 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-kp2s8\" (UniqueName: \"kubernetes.io/projected/48a5104a-2e2e-4828-bb54-bf40553d7ee0-kube-api-access-kp2s8\") pod \"barbican-5251-account-create-update-dr8tv\" (UID: \"48a5104a-2e2e-4828-bb54-bf40553d7ee0\") " pod="openstack/barbican-5251-account-create-update-dr8tv" Dec 13 07:11:47 crc kubenswrapper[4971]: I1213 07:11:47.064289 4971 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d6b11c79-8c13-42d7-8dce-c434dfab1912-combined-ca-bundle\") pod \"keystone-db-sync-b2cmm\" (UID: \"d6b11c79-8c13-42d7-8dce-c434dfab1912\") " pod="openstack/keystone-db-sync-b2cmm" Dec 13 07:11:47 crc kubenswrapper[4971]: I1213 07:11:47.064358 4971 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-gw7wr\" (UniqueName: \"kubernetes.io/projected/ad253e3a-226e-464a-9772-46f4b0081314-kube-api-access-gw7wr\") pod \"cinder-51d6-account-create-update-hm8kr\" (UID: \"ad253e3a-226e-464a-9772-46f4b0081314\") " pod="openstack/cinder-51d6-account-create-update-hm8kr" Dec 13 07:11:47 crc kubenswrapper[4971]: I1213 07:11:47.064445 4971 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-wr98f\" (UniqueName: \"kubernetes.io/projected/d6b11c79-8c13-42d7-8dce-c434dfab1912-kube-api-access-wr98f\") pod \"keystone-db-sync-b2cmm\" (UID: \"d6b11c79-8c13-42d7-8dce-c434dfab1912\") " pod="openstack/keystone-db-sync-b2cmm" Dec 13 07:11:47 crc kubenswrapper[4971]: I1213 07:11:47.064553 4971 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/48a5104a-2e2e-4828-bb54-bf40553d7ee0-operator-scripts\") pod \"barbican-5251-account-create-update-dr8tv\" (UID: \"48a5104a-2e2e-4828-bb54-bf40553d7ee0\") " pod="openstack/barbican-5251-account-create-update-dr8tv" Dec 13 07:11:47 crc kubenswrapper[4971]: I1213 07:11:47.064590 4971 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/d6b11c79-8c13-42d7-8dce-c434dfab1912-config-data\") pod \"keystone-db-sync-b2cmm\" (UID: \"d6b11c79-8c13-42d7-8dce-c434dfab1912\") " pod="openstack/keystone-db-sync-b2cmm" Dec 13 07:11:47 crc kubenswrapper[4971]: I1213 07:11:47.065407 4971 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/48a5104a-2e2e-4828-bb54-bf40553d7ee0-operator-scripts\") pod \"barbican-5251-account-create-update-dr8tv\" (UID: \"48a5104a-2e2e-4828-bb54-bf40553d7ee0\") " pod="openstack/barbican-5251-account-create-update-dr8tv" Dec 13 07:11:47 crc kubenswrapper[4971]: I1213 07:11:47.065476 4971 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-h7b8r\" (UniqueName: \"kubernetes.io/projected/f136ffc5-3a49-4807-8025-10356e78ee46-kube-api-access-h7b8r\") pod \"barbican-db-create-5czp4\" (UID: \"f136ffc5-3a49-4807-8025-10356e78ee46\") " pod="openstack/barbican-db-create-5czp4" Dec 13 07:11:47 crc kubenswrapper[4971]: I1213 07:11:47.065759 4971 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/f136ffc5-3a49-4807-8025-10356e78ee46-operator-scripts\") pod \"barbican-db-create-5czp4\" (UID: \"f136ffc5-3a49-4807-8025-10356e78ee46\") " pod="openstack/barbican-db-create-5czp4" Dec 13 07:11:47 crc kubenswrapper[4971]: I1213 07:11:47.065920 4971 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/ad253e3a-226e-464a-9772-46f4b0081314-operator-scripts\") pod \"cinder-51d6-account-create-update-hm8kr\" (UID: \"ad253e3a-226e-464a-9772-46f4b0081314\") " pod="openstack/cinder-51d6-account-create-update-hm8kr" Dec 13 07:11:47 crc kubenswrapper[4971]: I1213 07:11:47.066784 4971 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/f136ffc5-3a49-4807-8025-10356e78ee46-operator-scripts\") pod \"barbican-db-create-5czp4\" (UID: \"f136ffc5-3a49-4807-8025-10356e78ee46\") " pod="openstack/barbican-db-create-5czp4" Dec 13 07:11:47 crc kubenswrapper[4971]: I1213 07:11:47.088552 4971 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-kp2s8\" (UniqueName: \"kubernetes.io/projected/48a5104a-2e2e-4828-bb54-bf40553d7ee0-kube-api-access-kp2s8\") pod \"barbican-5251-account-create-update-dr8tv\" (UID: \"48a5104a-2e2e-4828-bb54-bf40553d7ee0\") " pod="openstack/barbican-5251-account-create-update-dr8tv" Dec 13 07:11:47 crc kubenswrapper[4971]: I1213 07:11:47.103113 4971 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-h7b8r\" (UniqueName: \"kubernetes.io/projected/f136ffc5-3a49-4807-8025-10356e78ee46-kube-api-access-h7b8r\") pod \"barbican-db-create-5czp4\" (UID: \"f136ffc5-3a49-4807-8025-10356e78ee46\") " pod="openstack/barbican-db-create-5czp4" Dec 13 07:11:47 crc kubenswrapper[4971]: I1213 07:11:47.109724 4971 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/neutron-2db2-account-create-update-t6l5w"] Dec 13 07:11:47 crc kubenswrapper[4971]: I1213 07:11:47.111165 4971 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-2db2-account-create-update-t6l5w" Dec 13 07:11:47 crc kubenswrapper[4971]: I1213 07:11:47.119930 4971 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"neutron-db-secret" Dec 13 07:11:47 crc kubenswrapper[4971]: I1213 07:11:47.120864 4971 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-5251-account-create-update-dr8tv" Dec 13 07:11:47 crc kubenswrapper[4971]: I1213 07:11:47.128474 4971 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/neutron-2db2-account-create-update-t6l5w"] Dec 13 07:11:47 crc kubenswrapper[4971]: I1213 07:11:47.138449 4971 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-db-create-5czp4" Dec 13 07:11:47 crc kubenswrapper[4971]: I1213 07:11:47.175495 4971 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/ad253e3a-226e-464a-9772-46f4b0081314-operator-scripts\") pod \"cinder-51d6-account-create-update-hm8kr\" (UID: \"ad253e3a-226e-464a-9772-46f4b0081314\") " pod="openstack/cinder-51d6-account-create-update-hm8kr" Dec 13 07:11:47 crc kubenswrapper[4971]: I1213 07:11:47.175724 4971 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d6b11c79-8c13-42d7-8dce-c434dfab1912-combined-ca-bundle\") pod \"keystone-db-sync-b2cmm\" (UID: \"d6b11c79-8c13-42d7-8dce-c434dfab1912\") " pod="openstack/keystone-db-sync-b2cmm" Dec 13 07:11:47 crc kubenswrapper[4971]: I1213 07:11:47.175788 4971 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-gw7wr\" (UniqueName: \"kubernetes.io/projected/ad253e3a-226e-464a-9772-46f4b0081314-kube-api-access-gw7wr\") pod \"cinder-51d6-account-create-update-hm8kr\" (UID: \"ad253e3a-226e-464a-9772-46f4b0081314\") " pod="openstack/cinder-51d6-account-create-update-hm8kr" Dec 13 07:11:47 crc kubenswrapper[4971]: I1213 07:11:47.175870 4971 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-wr98f\" (UniqueName: \"kubernetes.io/projected/d6b11c79-8c13-42d7-8dce-c434dfab1912-kube-api-access-wr98f\") pod \"keystone-db-sync-b2cmm\" (UID: \"d6b11c79-8c13-42d7-8dce-c434dfab1912\") " pod="openstack/keystone-db-sync-b2cmm" Dec 13 07:11:47 crc kubenswrapper[4971]: I1213 07:11:47.176016 4971 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/d6b11c79-8c13-42d7-8dce-c434dfab1912-config-data\") pod \"keystone-db-sync-b2cmm\" (UID: \"d6b11c79-8c13-42d7-8dce-c434dfab1912\") " pod="openstack/keystone-db-sync-b2cmm" Dec 13 07:11:47 crc kubenswrapper[4971]: I1213 07:11:47.177899 4971 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/ad253e3a-226e-464a-9772-46f4b0081314-operator-scripts\") pod \"cinder-51d6-account-create-update-hm8kr\" (UID: \"ad253e3a-226e-464a-9772-46f4b0081314\") " pod="openstack/cinder-51d6-account-create-update-hm8kr" Dec 13 07:11:47 crc kubenswrapper[4971]: I1213 07:11:47.180450 4971 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/d6b11c79-8c13-42d7-8dce-c434dfab1912-config-data\") pod \"keystone-db-sync-b2cmm\" (UID: \"d6b11c79-8c13-42d7-8dce-c434dfab1912\") " pod="openstack/keystone-db-sync-b2cmm" Dec 13 07:11:47 crc kubenswrapper[4971]: I1213 07:11:47.196155 4971 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d6b11c79-8c13-42d7-8dce-c434dfab1912-combined-ca-bundle\") pod \"keystone-db-sync-b2cmm\" (UID: \"d6b11c79-8c13-42d7-8dce-c434dfab1912\") " pod="openstack/keystone-db-sync-b2cmm" Dec 13 07:11:47 crc kubenswrapper[4971]: I1213 07:11:47.213190 4971 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-gw7wr\" (UniqueName: \"kubernetes.io/projected/ad253e3a-226e-464a-9772-46f4b0081314-kube-api-access-gw7wr\") pod \"cinder-51d6-account-create-update-hm8kr\" (UID: \"ad253e3a-226e-464a-9772-46f4b0081314\") " pod="openstack/cinder-51d6-account-create-update-hm8kr" Dec 13 07:11:47 crc kubenswrapper[4971]: I1213 07:11:47.213340 4971 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-wr98f\" (UniqueName: \"kubernetes.io/projected/d6b11c79-8c13-42d7-8dce-c434dfab1912-kube-api-access-wr98f\") pod \"keystone-db-sync-b2cmm\" (UID: \"d6b11c79-8c13-42d7-8dce-c434dfab1912\") " pod="openstack/keystone-db-sync-b2cmm" Dec 13 07:11:47 crc kubenswrapper[4971]: I1213 07:11:47.230742 4971 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-51d6-account-create-update-hm8kr" Dec 13 07:11:47 crc kubenswrapper[4971]: I1213 07:11:47.234501 4971 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/neutron-db-create-s42n4"] Dec 13 07:11:47 crc kubenswrapper[4971]: I1213 07:11:47.235901 4971 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-db-create-s42n4" Dec 13 07:11:47 crc kubenswrapper[4971]: I1213 07:11:47.249396 4971 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/neutron-db-create-s42n4"] Dec 13 07:11:47 crc kubenswrapper[4971]: I1213 07:11:47.277292 4971 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-5zg9f\" (UniqueName: \"kubernetes.io/projected/349ffc46-502d-4c10-9d0c-69bcfdb46fe7-kube-api-access-5zg9f\") pod \"neutron-2db2-account-create-update-t6l5w\" (UID: \"349ffc46-502d-4c10-9d0c-69bcfdb46fe7\") " pod="openstack/neutron-2db2-account-create-update-t6l5w" Dec 13 07:11:47 crc kubenswrapper[4971]: I1213 07:11:47.277676 4971 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/349ffc46-502d-4c10-9d0c-69bcfdb46fe7-operator-scripts\") pod \"neutron-2db2-account-create-update-t6l5w\" (UID: \"349ffc46-502d-4c10-9d0c-69bcfdb46fe7\") " pod="openstack/neutron-2db2-account-create-update-t6l5w" Dec 13 07:11:47 crc kubenswrapper[4971]: I1213 07:11:47.379328 4971 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/349ffc46-502d-4c10-9d0c-69bcfdb46fe7-operator-scripts\") pod \"neutron-2db2-account-create-update-t6l5w\" (UID: \"349ffc46-502d-4c10-9d0c-69bcfdb46fe7\") " pod="openstack/neutron-2db2-account-create-update-t6l5w" Dec 13 07:11:47 crc kubenswrapper[4971]: I1213 07:11:47.379392 4971 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/20b5092f-e5a2-467f-8dbc-e083156b0e20-operator-scripts\") pod \"neutron-db-create-s42n4\" (UID: \"20b5092f-e5a2-467f-8dbc-e083156b0e20\") " pod="openstack/neutron-db-create-s42n4" Dec 13 07:11:47 crc kubenswrapper[4971]: I1213 07:11:47.379432 4971 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-cspql\" (UniqueName: \"kubernetes.io/projected/20b5092f-e5a2-467f-8dbc-e083156b0e20-kube-api-access-cspql\") pod \"neutron-db-create-s42n4\" (UID: \"20b5092f-e5a2-467f-8dbc-e083156b0e20\") " pod="openstack/neutron-db-create-s42n4" Dec 13 07:11:47 crc kubenswrapper[4971]: I1213 07:11:47.379492 4971 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-5zg9f\" (UniqueName: \"kubernetes.io/projected/349ffc46-502d-4c10-9d0c-69bcfdb46fe7-kube-api-access-5zg9f\") pod \"neutron-2db2-account-create-update-t6l5w\" (UID: \"349ffc46-502d-4c10-9d0c-69bcfdb46fe7\") " pod="openstack/neutron-2db2-account-create-update-t6l5w" Dec 13 07:11:47 crc kubenswrapper[4971]: I1213 07:11:47.380684 4971 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/349ffc46-502d-4c10-9d0c-69bcfdb46fe7-operator-scripts\") pod \"neutron-2db2-account-create-update-t6l5w\" (UID: \"349ffc46-502d-4c10-9d0c-69bcfdb46fe7\") " pod="openstack/neutron-2db2-account-create-update-t6l5w" Dec 13 07:11:47 crc kubenswrapper[4971]: I1213 07:11:47.414346 4971 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-5zg9f\" (UniqueName: \"kubernetes.io/projected/349ffc46-502d-4c10-9d0c-69bcfdb46fe7-kube-api-access-5zg9f\") pod \"neutron-2db2-account-create-update-t6l5w\" (UID: \"349ffc46-502d-4c10-9d0c-69bcfdb46fe7\") " pod="openstack/neutron-2db2-account-create-update-t6l5w" Dec 13 07:11:47 crc kubenswrapper[4971]: I1213 07:11:47.491871 4971 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/20b5092f-e5a2-467f-8dbc-e083156b0e20-operator-scripts\") pod \"neutron-db-create-s42n4\" (UID: \"20b5092f-e5a2-467f-8dbc-e083156b0e20\") " pod="openstack/neutron-db-create-s42n4" Dec 13 07:11:47 crc kubenswrapper[4971]: I1213 07:11:47.491915 4971 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cspql\" (UniqueName: \"kubernetes.io/projected/20b5092f-e5a2-467f-8dbc-e083156b0e20-kube-api-access-cspql\") pod \"neutron-db-create-s42n4\" (UID: \"20b5092f-e5a2-467f-8dbc-e083156b0e20\") " pod="openstack/neutron-db-create-s42n4" Dec 13 07:11:47 crc kubenswrapper[4971]: I1213 07:11:47.492039 4971 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-db-sync-b2cmm" Dec 13 07:11:47 crc kubenswrapper[4971]: I1213 07:11:47.492490 4971 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-2db2-account-create-update-t6l5w" Dec 13 07:11:47 crc kubenswrapper[4971]: I1213 07:11:47.493031 4971 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/20b5092f-e5a2-467f-8dbc-e083156b0e20-operator-scripts\") pod \"neutron-db-create-s42n4\" (UID: \"20b5092f-e5a2-467f-8dbc-e083156b0e20\") " pod="openstack/neutron-db-create-s42n4" Dec 13 07:11:47 crc kubenswrapper[4971]: I1213 07:11:47.550895 4971 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-cspql\" (UniqueName: \"kubernetes.io/projected/20b5092f-e5a2-467f-8dbc-e083156b0e20-kube-api-access-cspql\") pod \"neutron-db-create-s42n4\" (UID: \"20b5092f-e5a2-467f-8dbc-e083156b0e20\") " pod="openstack/neutron-db-create-s42n4" Dec 13 07:11:47 crc kubenswrapper[4971]: I1213 07:11:47.571884 4971 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-db-create-s42n4" Dec 13 07:11:47 crc kubenswrapper[4971]: I1213 07:11:47.615623 4971 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/barbican-db-create-5czp4"] Dec 13 07:11:47 crc kubenswrapper[4971]: W1213 07:11:47.628091 4971 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podf136ffc5_3a49_4807_8025_10356e78ee46.slice/crio-07b1a26df63fcd01efa637c27f207240bab73cefdc2df6a7ed81fe0ab6e53e6a WatchSource:0}: Error finding container 07b1a26df63fcd01efa637c27f207240bab73cefdc2df6a7ed81fe0ab6e53e6a: Status 404 returned error can't find the container with id 07b1a26df63fcd01efa637c27f207240bab73cefdc2df6a7ed81fe0ab6e53e6a Dec 13 07:11:47 crc kubenswrapper[4971]: I1213 07:11:47.873017 4971 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cinder-db-create-8sbpv"] Dec 13 07:11:47 crc kubenswrapper[4971]: W1213 07:11:47.891682 4971 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod4c7ef6d3_1b68_4fb6_9784_0ee0bb154a8d.slice/crio-2a3d4554c671509259d66b4f503025f8d64fdf766a4559582876a0a5449cef89 WatchSource:0}: Error finding container 2a3d4554c671509259d66b4f503025f8d64fdf766a4559582876a0a5449cef89: Status 404 returned error can't find the container with id 2a3d4554c671509259d66b4f503025f8d64fdf766a4559582876a0a5449cef89 Dec 13 07:11:47 crc kubenswrapper[4971]: I1213 07:11:47.895957 4971 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/barbican-5251-account-create-update-dr8tv"] Dec 13 07:11:48 crc kubenswrapper[4971]: I1213 07:11:48.035918 4971 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-57768dd7b5-86hlw" event={"ID":"c9e30081-d603-4c9d-81ce-7bcff6848243","Type":"ContainerStarted","Data":"303ee14cece58dbf062bf3b1d7ce80d4f37ced4fdb8e6c483677f2b786064dd4"} Dec 13 07:11:48 crc kubenswrapper[4971]: I1213 07:11:48.036289 4971 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/dnsmasq-dns-57768dd7b5-86hlw" Dec 13 07:11:48 crc kubenswrapper[4971]: I1213 07:11:48.040941 4971 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-db-create-5czp4" event={"ID":"f136ffc5-3a49-4807-8025-10356e78ee46","Type":"ContainerStarted","Data":"e620d182d52f7c29766f7d0a12dbd3f37b8c5f4a578281d86ab99036ece89ae9"} Dec 13 07:11:48 crc kubenswrapper[4971]: I1213 07:11:48.040973 4971 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-db-create-5czp4" event={"ID":"f136ffc5-3a49-4807-8025-10356e78ee46","Type":"ContainerStarted","Data":"07b1a26df63fcd01efa637c27f207240bab73cefdc2df6a7ed81fe0ab6e53e6a"} Dec 13 07:11:48 crc kubenswrapper[4971]: I1213 07:11:48.042667 4971 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-5251-account-create-update-dr8tv" event={"ID":"48a5104a-2e2e-4828-bb54-bf40553d7ee0","Type":"ContainerStarted","Data":"b783904e1f84f1d7bad0d373f6e98e2e2ac2c1af86ccbccb0a67b46357c25dca"} Dec 13 07:11:48 crc kubenswrapper[4971]: I1213 07:11:48.042702 4971 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-5251-account-create-update-dr8tv" event={"ID":"48a5104a-2e2e-4828-bb54-bf40553d7ee0","Type":"ContainerStarted","Data":"12a3dac2703a556e3c7b3995c67a43e453c1df9641175ed491ddfadd0f61dd53"} Dec 13 07:11:48 crc kubenswrapper[4971]: I1213 07:11:48.044898 4971 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-db-create-8sbpv" event={"ID":"4c7ef6d3-1b68-4fb6-9784-0ee0bb154a8d","Type":"ContainerStarted","Data":"465494d76081e5489304d7756f8b20158debd5991405b16dba4332284efc2ab6"} Dec 13 07:11:48 crc kubenswrapper[4971]: I1213 07:11:48.044943 4971 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-db-create-8sbpv" event={"ID":"4c7ef6d3-1b68-4fb6-9784-0ee0bb154a8d","Type":"ContainerStarted","Data":"2a3d4554c671509259d66b4f503025f8d64fdf766a4559582876a0a5449cef89"} Dec 13 07:11:48 crc kubenswrapper[4971]: I1213 07:11:48.056573 4971 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cinder-51d6-account-create-update-hm8kr"] Dec 13 07:11:48 crc kubenswrapper[4971]: W1213 07:11:48.059259 4971 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podad253e3a_226e_464a_9772_46f4b0081314.slice/crio-1859e240a6237c4af9923d6b13b027f3e2a782ef7658f7b26d2cd667efd8e33e WatchSource:0}: Error finding container 1859e240a6237c4af9923d6b13b027f3e2a782ef7658f7b26d2cd667efd8e33e: Status 404 returned error can't find the container with id 1859e240a6237c4af9923d6b13b027f3e2a782ef7658f7b26d2cd667efd8e33e Dec 13 07:11:48 crc kubenswrapper[4971]: I1213 07:11:48.060785 4971 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/dnsmasq-dns-57768dd7b5-86hlw" podStartSLOduration=3.060769554 podStartE2EDuration="3.060769554s" podCreationTimestamp="2025-12-13 07:11:45 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-13 07:11:48.052062409 +0000 UTC m=+1364.656471867" watchObservedRunningTime="2025-12-13 07:11:48.060769554 +0000 UTC m=+1364.665179002" Dec 13 07:11:48 crc kubenswrapper[4971]: I1213 07:11:48.080430 4971 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/cinder-db-create-8sbpv" podStartSLOduration=2.080409928 podStartE2EDuration="2.080409928s" podCreationTimestamp="2025-12-13 07:11:46 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-13 07:11:48.070141424 +0000 UTC m=+1364.674550872" watchObservedRunningTime="2025-12-13 07:11:48.080409928 +0000 UTC m=+1364.684819376" Dec 13 07:11:48 crc kubenswrapper[4971]: I1213 07:11:48.084268 4971 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/barbican-db-create-5czp4" podStartSLOduration=2.084256952 podStartE2EDuration="2.084256952s" podCreationTimestamp="2025-12-13 07:11:46 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-13 07:11:48.082826697 +0000 UTC m=+1364.687236145" watchObservedRunningTime="2025-12-13 07:11:48.084256952 +0000 UTC m=+1364.688666400" Dec 13 07:11:48 crc kubenswrapper[4971]: I1213 07:11:48.102431 4971 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/barbican-5251-account-create-update-dr8tv" podStartSLOduration=2.102412289 podStartE2EDuration="2.102412289s" podCreationTimestamp="2025-12-13 07:11:46 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-13 07:11:48.100876531 +0000 UTC m=+1364.705285979" watchObservedRunningTime="2025-12-13 07:11:48.102412289 +0000 UTC m=+1364.706821747" Dec 13 07:11:48 crc kubenswrapper[4971]: W1213 07:11:48.144623 4971 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podd6b11c79_8c13_42d7_8dce_c434dfab1912.slice/crio-9286a985004d24f6ef97375b220cf0c6ff19b1d7ee0e228a33ea94c2e589ee2a WatchSource:0}: Error finding container 9286a985004d24f6ef97375b220cf0c6ff19b1d7ee0e228a33ea94c2e589ee2a: Status 404 returned error can't find the container with id 9286a985004d24f6ef97375b220cf0c6ff19b1d7ee0e228a33ea94c2e589ee2a Dec 13 07:11:48 crc kubenswrapper[4971]: I1213 07:11:48.157670 4971 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/keystone-db-sync-b2cmm"] Dec 13 07:11:48 crc kubenswrapper[4971]: I1213 07:11:48.171320 4971 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/neutron-db-create-s42n4"] Dec 13 07:11:48 crc kubenswrapper[4971]: W1213 07:11:48.179163 4971 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod349ffc46_502d_4c10_9d0c_69bcfdb46fe7.slice/crio-5f16b7e6d43e48e80a9baf888420ddb436d0544386e2d58172e0ef242371c4df WatchSource:0}: Error finding container 5f16b7e6d43e48e80a9baf888420ddb436d0544386e2d58172e0ef242371c4df: Status 404 returned error can't find the container with id 5f16b7e6d43e48e80a9baf888420ddb436d0544386e2d58172e0ef242371c4df Dec 13 07:11:48 crc kubenswrapper[4971]: I1213 07:11:48.182449 4971 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/neutron-2db2-account-create-update-t6l5w"] Dec 13 07:11:49 crc kubenswrapper[4971]: I1213 07:11:49.057162 4971 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-db-sync-b2cmm" event={"ID":"d6b11c79-8c13-42d7-8dce-c434dfab1912","Type":"ContainerStarted","Data":"9286a985004d24f6ef97375b220cf0c6ff19b1d7ee0e228a33ea94c2e589ee2a"} Dec 13 07:11:49 crc kubenswrapper[4971]: I1213 07:11:49.060128 4971 generic.go:334] "Generic (PLEG): container finished" podID="4c7ef6d3-1b68-4fb6-9784-0ee0bb154a8d" containerID="465494d76081e5489304d7756f8b20158debd5991405b16dba4332284efc2ab6" exitCode=0 Dec 13 07:11:49 crc kubenswrapper[4971]: I1213 07:11:49.060206 4971 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-db-create-8sbpv" event={"ID":"4c7ef6d3-1b68-4fb6-9784-0ee0bb154a8d","Type":"ContainerDied","Data":"465494d76081e5489304d7756f8b20158debd5991405b16dba4332284efc2ab6"} Dec 13 07:11:49 crc kubenswrapper[4971]: I1213 07:11:49.063657 4971 generic.go:334] "Generic (PLEG): container finished" podID="ad253e3a-226e-464a-9772-46f4b0081314" containerID="360bfe834cf511f784e167d260a057730d73c2f25a0284f0622fe6be6d7b48d2" exitCode=0 Dec 13 07:11:49 crc kubenswrapper[4971]: I1213 07:11:49.063708 4971 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-51d6-account-create-update-hm8kr" event={"ID":"ad253e3a-226e-464a-9772-46f4b0081314","Type":"ContainerDied","Data":"360bfe834cf511f784e167d260a057730d73c2f25a0284f0622fe6be6d7b48d2"} Dec 13 07:11:49 crc kubenswrapper[4971]: I1213 07:11:49.063728 4971 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-51d6-account-create-update-hm8kr" event={"ID":"ad253e3a-226e-464a-9772-46f4b0081314","Type":"ContainerStarted","Data":"1859e240a6237c4af9923d6b13b027f3e2a782ef7658f7b26d2cd667efd8e33e"} Dec 13 07:11:49 crc kubenswrapper[4971]: I1213 07:11:49.066669 4971 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-2db2-account-create-update-t6l5w" event={"ID":"349ffc46-502d-4c10-9d0c-69bcfdb46fe7","Type":"ContainerStarted","Data":"3f7c5ba1c6814836f2d82610a709511d47a2e71f2dcf2e913651c6519d7acf4a"} Dec 13 07:11:49 crc kubenswrapper[4971]: I1213 07:11:49.066703 4971 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-2db2-account-create-update-t6l5w" event={"ID":"349ffc46-502d-4c10-9d0c-69bcfdb46fe7","Type":"ContainerStarted","Data":"5f16b7e6d43e48e80a9baf888420ddb436d0544386e2d58172e0ef242371c4df"} Dec 13 07:11:49 crc kubenswrapper[4971]: I1213 07:11:49.070720 4971 generic.go:334] "Generic (PLEG): container finished" podID="f136ffc5-3a49-4807-8025-10356e78ee46" containerID="e620d182d52f7c29766f7d0a12dbd3f37b8c5f4a578281d86ab99036ece89ae9" exitCode=0 Dec 13 07:11:49 crc kubenswrapper[4971]: I1213 07:11:49.070776 4971 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-db-create-5czp4" event={"ID":"f136ffc5-3a49-4807-8025-10356e78ee46","Type":"ContainerDied","Data":"e620d182d52f7c29766f7d0a12dbd3f37b8c5f4a578281d86ab99036ece89ae9"} Dec 13 07:11:49 crc kubenswrapper[4971]: I1213 07:11:49.076448 4971 generic.go:334] "Generic (PLEG): container finished" podID="20b5092f-e5a2-467f-8dbc-e083156b0e20" containerID="3c9811af77b10cff3571a757300c8a4590fe34ec9a412c81599a4241a91ba697" exitCode=0 Dec 13 07:11:49 crc kubenswrapper[4971]: I1213 07:11:49.076537 4971 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-db-create-s42n4" event={"ID":"20b5092f-e5a2-467f-8dbc-e083156b0e20","Type":"ContainerDied","Data":"3c9811af77b10cff3571a757300c8a4590fe34ec9a412c81599a4241a91ba697"} Dec 13 07:11:49 crc kubenswrapper[4971]: I1213 07:11:49.076570 4971 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-db-create-s42n4" event={"ID":"20b5092f-e5a2-467f-8dbc-e083156b0e20","Type":"ContainerStarted","Data":"a111e1aaf75a42314bc9ff00482aaba602b34c7f382cdc816e6521b77a20e916"} Dec 13 07:11:49 crc kubenswrapper[4971]: I1213 07:11:49.078929 4971 generic.go:334] "Generic (PLEG): container finished" podID="48a5104a-2e2e-4828-bb54-bf40553d7ee0" containerID="b783904e1f84f1d7bad0d373f6e98e2e2ac2c1af86ccbccb0a67b46357c25dca" exitCode=0 Dec 13 07:11:49 crc kubenswrapper[4971]: I1213 07:11:49.079752 4971 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-5251-account-create-update-dr8tv" event={"ID":"48a5104a-2e2e-4828-bb54-bf40553d7ee0","Type":"ContainerDied","Data":"b783904e1f84f1d7bad0d373f6e98e2e2ac2c1af86ccbccb0a67b46357c25dca"} Dec 13 07:11:49 crc kubenswrapper[4971]: I1213 07:11:49.140746 4971 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/neutron-2db2-account-create-update-t6l5w" podStartSLOduration=2.140729202 podStartE2EDuration="2.140729202s" podCreationTimestamp="2025-12-13 07:11:47 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-13 07:11:49.136378205 +0000 UTC m=+1365.740787653" watchObservedRunningTime="2025-12-13 07:11:49.140729202 +0000 UTC m=+1365.745138650" Dec 13 07:11:50 crc kubenswrapper[4971]: I1213 07:11:50.088126 4971 generic.go:334] "Generic (PLEG): container finished" podID="349ffc46-502d-4c10-9d0c-69bcfdb46fe7" containerID="3f7c5ba1c6814836f2d82610a709511d47a2e71f2dcf2e913651c6519d7acf4a" exitCode=0 Dec 13 07:11:50 crc kubenswrapper[4971]: I1213 07:11:50.088643 4971 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-2db2-account-create-update-t6l5w" event={"ID":"349ffc46-502d-4c10-9d0c-69bcfdb46fe7","Type":"ContainerDied","Data":"3f7c5ba1c6814836f2d82610a709511d47a2e71f2dcf2e913651c6519d7acf4a"} Dec 13 07:11:50 crc kubenswrapper[4971]: I1213 07:11:50.541293 4971 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-db-create-8sbpv" Dec 13 07:11:50 crc kubenswrapper[4971]: I1213 07:11:50.648767 4971 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-4zz7m\" (UniqueName: \"kubernetes.io/projected/4c7ef6d3-1b68-4fb6-9784-0ee0bb154a8d-kube-api-access-4zz7m\") pod \"4c7ef6d3-1b68-4fb6-9784-0ee0bb154a8d\" (UID: \"4c7ef6d3-1b68-4fb6-9784-0ee0bb154a8d\") " Dec 13 07:11:50 crc kubenswrapper[4971]: I1213 07:11:50.648896 4971 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/4c7ef6d3-1b68-4fb6-9784-0ee0bb154a8d-operator-scripts\") pod \"4c7ef6d3-1b68-4fb6-9784-0ee0bb154a8d\" (UID: \"4c7ef6d3-1b68-4fb6-9784-0ee0bb154a8d\") " Dec 13 07:11:50 crc kubenswrapper[4971]: I1213 07:11:50.650034 4971 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/4c7ef6d3-1b68-4fb6-9784-0ee0bb154a8d-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "4c7ef6d3-1b68-4fb6-9784-0ee0bb154a8d" (UID: "4c7ef6d3-1b68-4fb6-9784-0ee0bb154a8d"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 13 07:11:50 crc kubenswrapper[4971]: I1213 07:11:50.656384 4971 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/4c7ef6d3-1b68-4fb6-9784-0ee0bb154a8d-kube-api-access-4zz7m" (OuterVolumeSpecName: "kube-api-access-4zz7m") pod "4c7ef6d3-1b68-4fb6-9784-0ee0bb154a8d" (UID: "4c7ef6d3-1b68-4fb6-9784-0ee0bb154a8d"). InnerVolumeSpecName "kube-api-access-4zz7m". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 13 07:11:50 crc kubenswrapper[4971]: I1213 07:11:50.729043 4971 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-db-create-s42n4" Dec 13 07:11:50 crc kubenswrapper[4971]: I1213 07:11:50.741194 4971 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-51d6-account-create-update-hm8kr" Dec 13 07:11:50 crc kubenswrapper[4971]: I1213 07:11:50.750348 4971 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-4zz7m\" (UniqueName: \"kubernetes.io/projected/4c7ef6d3-1b68-4fb6-9784-0ee0bb154a8d-kube-api-access-4zz7m\") on node \"crc\" DevicePath \"\"" Dec 13 07:11:50 crc kubenswrapper[4971]: I1213 07:11:50.750380 4971 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/4c7ef6d3-1b68-4fb6-9784-0ee0bb154a8d-operator-scripts\") on node \"crc\" DevicePath \"\"" Dec 13 07:11:50 crc kubenswrapper[4971]: I1213 07:11:50.751406 4971 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-db-create-5czp4" Dec 13 07:11:50 crc kubenswrapper[4971]: I1213 07:11:50.761863 4971 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-5251-account-create-update-dr8tv" Dec 13 07:11:50 crc kubenswrapper[4971]: I1213 07:11:50.851629 4971 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/20b5092f-e5a2-467f-8dbc-e083156b0e20-operator-scripts\") pod \"20b5092f-e5a2-467f-8dbc-e083156b0e20\" (UID: \"20b5092f-e5a2-467f-8dbc-e083156b0e20\") " Dec 13 07:11:50 crc kubenswrapper[4971]: I1213 07:11:50.851698 4971 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-h7b8r\" (UniqueName: \"kubernetes.io/projected/f136ffc5-3a49-4807-8025-10356e78ee46-kube-api-access-h7b8r\") pod \"f136ffc5-3a49-4807-8025-10356e78ee46\" (UID: \"f136ffc5-3a49-4807-8025-10356e78ee46\") " Dec 13 07:11:50 crc kubenswrapper[4971]: I1213 07:11:50.851795 4971 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-gw7wr\" (UniqueName: \"kubernetes.io/projected/ad253e3a-226e-464a-9772-46f4b0081314-kube-api-access-gw7wr\") pod \"ad253e3a-226e-464a-9772-46f4b0081314\" (UID: \"ad253e3a-226e-464a-9772-46f4b0081314\") " Dec 13 07:11:50 crc kubenswrapper[4971]: I1213 07:11:50.852145 4971 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/f136ffc5-3a49-4807-8025-10356e78ee46-operator-scripts\") pod \"f136ffc5-3a49-4807-8025-10356e78ee46\" (UID: \"f136ffc5-3a49-4807-8025-10356e78ee46\") " Dec 13 07:11:50 crc kubenswrapper[4971]: I1213 07:11:50.852209 4971 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-cspql\" (UniqueName: \"kubernetes.io/projected/20b5092f-e5a2-467f-8dbc-e083156b0e20-kube-api-access-cspql\") pod \"20b5092f-e5a2-467f-8dbc-e083156b0e20\" (UID: \"20b5092f-e5a2-467f-8dbc-e083156b0e20\") " Dec 13 07:11:50 crc kubenswrapper[4971]: I1213 07:11:50.852301 4971 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/ad253e3a-226e-464a-9772-46f4b0081314-operator-scripts\") pod \"ad253e3a-226e-464a-9772-46f4b0081314\" (UID: \"ad253e3a-226e-464a-9772-46f4b0081314\") " Dec 13 07:11:50 crc kubenswrapper[4971]: I1213 07:11:50.856750 4971 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/20b5092f-e5a2-467f-8dbc-e083156b0e20-kube-api-access-cspql" (OuterVolumeSpecName: "kube-api-access-cspql") pod "20b5092f-e5a2-467f-8dbc-e083156b0e20" (UID: "20b5092f-e5a2-467f-8dbc-e083156b0e20"). InnerVolumeSpecName "kube-api-access-cspql". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 13 07:11:50 crc kubenswrapper[4971]: I1213 07:11:50.856814 4971 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/ad253e3a-226e-464a-9772-46f4b0081314-kube-api-access-gw7wr" (OuterVolumeSpecName: "kube-api-access-gw7wr") pod "ad253e3a-226e-464a-9772-46f4b0081314" (UID: "ad253e3a-226e-464a-9772-46f4b0081314"). InnerVolumeSpecName "kube-api-access-gw7wr". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 13 07:11:50 crc kubenswrapper[4971]: I1213 07:11:50.857708 4971 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/f136ffc5-3a49-4807-8025-10356e78ee46-kube-api-access-h7b8r" (OuterVolumeSpecName: "kube-api-access-h7b8r") pod "f136ffc5-3a49-4807-8025-10356e78ee46" (UID: "f136ffc5-3a49-4807-8025-10356e78ee46"). InnerVolumeSpecName "kube-api-access-h7b8r". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 13 07:11:50 crc kubenswrapper[4971]: I1213 07:11:50.930774 4971 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/f136ffc5-3a49-4807-8025-10356e78ee46-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "f136ffc5-3a49-4807-8025-10356e78ee46" (UID: "f136ffc5-3a49-4807-8025-10356e78ee46"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 13 07:11:50 crc kubenswrapper[4971]: I1213 07:11:50.932681 4971 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/20b5092f-e5a2-467f-8dbc-e083156b0e20-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "20b5092f-e5a2-467f-8dbc-e083156b0e20" (UID: "20b5092f-e5a2-467f-8dbc-e083156b0e20"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 13 07:11:50 crc kubenswrapper[4971]: I1213 07:11:50.932773 4971 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/ad253e3a-226e-464a-9772-46f4b0081314-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "ad253e3a-226e-464a-9772-46f4b0081314" (UID: "ad253e3a-226e-464a-9772-46f4b0081314"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 13 07:11:50 crc kubenswrapper[4971]: I1213 07:11:50.953742 4971 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/48a5104a-2e2e-4828-bb54-bf40553d7ee0-operator-scripts\") pod \"48a5104a-2e2e-4828-bb54-bf40553d7ee0\" (UID: \"48a5104a-2e2e-4828-bb54-bf40553d7ee0\") " Dec 13 07:11:50 crc kubenswrapper[4971]: I1213 07:11:50.953912 4971 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-kp2s8\" (UniqueName: \"kubernetes.io/projected/48a5104a-2e2e-4828-bb54-bf40553d7ee0-kube-api-access-kp2s8\") pod \"48a5104a-2e2e-4828-bb54-bf40553d7ee0\" (UID: \"48a5104a-2e2e-4828-bb54-bf40553d7ee0\") " Dec 13 07:11:50 crc kubenswrapper[4971]: I1213 07:11:50.954445 4971 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-gw7wr\" (UniqueName: \"kubernetes.io/projected/ad253e3a-226e-464a-9772-46f4b0081314-kube-api-access-gw7wr\") on node \"crc\" DevicePath \"\"" Dec 13 07:11:50 crc kubenswrapper[4971]: I1213 07:11:50.954470 4971 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/f136ffc5-3a49-4807-8025-10356e78ee46-operator-scripts\") on node \"crc\" DevicePath \"\"" Dec 13 07:11:50 crc kubenswrapper[4971]: I1213 07:11:50.954484 4971 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-cspql\" (UniqueName: \"kubernetes.io/projected/20b5092f-e5a2-467f-8dbc-e083156b0e20-kube-api-access-cspql\") on node \"crc\" DevicePath \"\"" Dec 13 07:11:50 crc kubenswrapper[4971]: I1213 07:11:50.954495 4971 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/ad253e3a-226e-464a-9772-46f4b0081314-operator-scripts\") on node \"crc\" DevicePath \"\"" Dec 13 07:11:50 crc kubenswrapper[4971]: I1213 07:11:50.954505 4971 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/20b5092f-e5a2-467f-8dbc-e083156b0e20-operator-scripts\") on node \"crc\" DevicePath \"\"" Dec 13 07:11:50 crc kubenswrapper[4971]: I1213 07:11:50.954559 4971 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-h7b8r\" (UniqueName: \"kubernetes.io/projected/f136ffc5-3a49-4807-8025-10356e78ee46-kube-api-access-h7b8r\") on node \"crc\" DevicePath \"\"" Dec 13 07:11:50 crc kubenswrapper[4971]: I1213 07:11:50.954758 4971 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/48a5104a-2e2e-4828-bb54-bf40553d7ee0-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "48a5104a-2e2e-4828-bb54-bf40553d7ee0" (UID: "48a5104a-2e2e-4828-bb54-bf40553d7ee0"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 13 07:11:50 crc kubenswrapper[4971]: I1213 07:11:50.958841 4971 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/48a5104a-2e2e-4828-bb54-bf40553d7ee0-kube-api-access-kp2s8" (OuterVolumeSpecName: "kube-api-access-kp2s8") pod "48a5104a-2e2e-4828-bb54-bf40553d7ee0" (UID: "48a5104a-2e2e-4828-bb54-bf40553d7ee0"). InnerVolumeSpecName "kube-api-access-kp2s8". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 13 07:11:51 crc kubenswrapper[4971]: I1213 07:11:51.056029 4971 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/48a5104a-2e2e-4828-bb54-bf40553d7ee0-operator-scripts\") on node \"crc\" DevicePath \"\"" Dec 13 07:11:51 crc kubenswrapper[4971]: I1213 07:11:51.056097 4971 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-kp2s8\" (UniqueName: \"kubernetes.io/projected/48a5104a-2e2e-4828-bb54-bf40553d7ee0-kube-api-access-kp2s8\") on node \"crc\" DevicePath \"\"" Dec 13 07:11:51 crc kubenswrapper[4971]: I1213 07:11:51.097530 4971 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-db-create-5czp4" event={"ID":"f136ffc5-3a49-4807-8025-10356e78ee46","Type":"ContainerDied","Data":"07b1a26df63fcd01efa637c27f207240bab73cefdc2df6a7ed81fe0ab6e53e6a"} Dec 13 07:11:51 crc kubenswrapper[4971]: I1213 07:11:51.097576 4971 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="07b1a26df63fcd01efa637c27f207240bab73cefdc2df6a7ed81fe0ab6e53e6a" Dec 13 07:11:51 crc kubenswrapper[4971]: I1213 07:11:51.097587 4971 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-db-create-5czp4" Dec 13 07:11:51 crc kubenswrapper[4971]: I1213 07:11:51.099714 4971 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-db-create-s42n4" Dec 13 07:11:51 crc kubenswrapper[4971]: I1213 07:11:51.099715 4971 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-db-create-s42n4" event={"ID":"20b5092f-e5a2-467f-8dbc-e083156b0e20","Type":"ContainerDied","Data":"a111e1aaf75a42314bc9ff00482aaba602b34c7f382cdc816e6521b77a20e916"} Dec 13 07:11:51 crc kubenswrapper[4971]: I1213 07:11:51.099828 4971 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="a111e1aaf75a42314bc9ff00482aaba602b34c7f382cdc816e6521b77a20e916" Dec 13 07:11:51 crc kubenswrapper[4971]: I1213 07:11:51.101723 4971 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-5251-account-create-update-dr8tv" event={"ID":"48a5104a-2e2e-4828-bb54-bf40553d7ee0","Type":"ContainerDied","Data":"12a3dac2703a556e3c7b3995c67a43e453c1df9641175ed491ddfadd0f61dd53"} Dec 13 07:11:51 crc kubenswrapper[4971]: I1213 07:11:51.101913 4971 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="12a3dac2703a556e3c7b3995c67a43e453c1df9641175ed491ddfadd0f61dd53" Dec 13 07:11:51 crc kubenswrapper[4971]: I1213 07:11:51.102015 4971 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-5251-account-create-update-dr8tv" Dec 13 07:11:51 crc kubenswrapper[4971]: I1213 07:11:51.103462 4971 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-db-create-8sbpv" event={"ID":"4c7ef6d3-1b68-4fb6-9784-0ee0bb154a8d","Type":"ContainerDied","Data":"2a3d4554c671509259d66b4f503025f8d64fdf766a4559582876a0a5449cef89"} Dec 13 07:11:51 crc kubenswrapper[4971]: I1213 07:11:51.103491 4971 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="2a3d4554c671509259d66b4f503025f8d64fdf766a4559582876a0a5449cef89" Dec 13 07:11:51 crc kubenswrapper[4971]: I1213 07:11:51.103565 4971 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-db-create-8sbpv" Dec 13 07:11:51 crc kubenswrapper[4971]: I1213 07:11:51.106607 4971 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-51d6-account-create-update-hm8kr" event={"ID":"ad253e3a-226e-464a-9772-46f4b0081314","Type":"ContainerDied","Data":"1859e240a6237c4af9923d6b13b027f3e2a782ef7658f7b26d2cd667efd8e33e"} Dec 13 07:11:51 crc kubenswrapper[4971]: I1213 07:11:51.106658 4971 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="1859e240a6237c4af9923d6b13b027f3e2a782ef7658f7b26d2cd667efd8e33e" Dec 13 07:11:51 crc kubenswrapper[4971]: I1213 07:11:51.106662 4971 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-51d6-account-create-update-hm8kr" Dec 13 07:11:51 crc kubenswrapper[4971]: I1213 07:11:51.349944 4971 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-2db2-account-create-update-t6l5w" Dec 13 07:11:51 crc kubenswrapper[4971]: I1213 07:11:51.463077 4971 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-5zg9f\" (UniqueName: \"kubernetes.io/projected/349ffc46-502d-4c10-9d0c-69bcfdb46fe7-kube-api-access-5zg9f\") pod \"349ffc46-502d-4c10-9d0c-69bcfdb46fe7\" (UID: \"349ffc46-502d-4c10-9d0c-69bcfdb46fe7\") " Dec 13 07:11:51 crc kubenswrapper[4971]: I1213 07:11:51.463150 4971 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/349ffc46-502d-4c10-9d0c-69bcfdb46fe7-operator-scripts\") pod \"349ffc46-502d-4c10-9d0c-69bcfdb46fe7\" (UID: \"349ffc46-502d-4c10-9d0c-69bcfdb46fe7\") " Dec 13 07:11:51 crc kubenswrapper[4971]: I1213 07:11:51.463790 4971 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/349ffc46-502d-4c10-9d0c-69bcfdb46fe7-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "349ffc46-502d-4c10-9d0c-69bcfdb46fe7" (UID: "349ffc46-502d-4c10-9d0c-69bcfdb46fe7"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 13 07:11:51 crc kubenswrapper[4971]: I1213 07:11:51.466920 4971 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/349ffc46-502d-4c10-9d0c-69bcfdb46fe7-kube-api-access-5zg9f" (OuterVolumeSpecName: "kube-api-access-5zg9f") pod "349ffc46-502d-4c10-9d0c-69bcfdb46fe7" (UID: "349ffc46-502d-4c10-9d0c-69bcfdb46fe7"). InnerVolumeSpecName "kube-api-access-5zg9f". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 13 07:11:51 crc kubenswrapper[4971]: I1213 07:11:51.565119 4971 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-5zg9f\" (UniqueName: \"kubernetes.io/projected/349ffc46-502d-4c10-9d0c-69bcfdb46fe7-kube-api-access-5zg9f\") on node \"crc\" DevicePath \"\"" Dec 13 07:11:51 crc kubenswrapper[4971]: I1213 07:11:51.565160 4971 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/349ffc46-502d-4c10-9d0c-69bcfdb46fe7-operator-scripts\") on node \"crc\" DevicePath \"\"" Dec 13 07:11:52 crc kubenswrapper[4971]: I1213 07:11:52.119550 4971 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-2db2-account-create-update-t6l5w" event={"ID":"349ffc46-502d-4c10-9d0c-69bcfdb46fe7","Type":"ContainerDied","Data":"5f16b7e6d43e48e80a9baf888420ddb436d0544386e2d58172e0ef242371c4df"} Dec 13 07:11:52 crc kubenswrapper[4971]: I1213 07:11:52.119919 4971 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="5f16b7e6d43e48e80a9baf888420ddb436d0544386e2d58172e0ef242371c4df" Dec 13 07:11:52 crc kubenswrapper[4971]: I1213 07:11:52.120053 4971 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-2db2-account-create-update-t6l5w" Dec 13 07:11:55 crc kubenswrapper[4971]: I1213 07:11:55.579608 4971 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/dnsmasq-dns-57768dd7b5-86hlw" Dec 13 07:11:55 crc kubenswrapper[4971]: I1213 07:11:55.637709 4971 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-757dc6fff9-hfmv4"] Dec 13 07:11:55 crc kubenswrapper[4971]: I1213 07:11:55.638031 4971 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/dnsmasq-dns-757dc6fff9-hfmv4" podUID="82d7cf8c-129a-483a-b190-8480e2f03418" containerName="dnsmasq-dns" containerID="cri-o://45a3b8435a1a4bc04f956f428cca44d0c3e5b295322aff907527d681ce5c1f1a" gracePeriod=10 Dec 13 07:11:56 crc kubenswrapper[4971]: I1213 07:11:56.202166 4971 generic.go:334] "Generic (PLEG): container finished" podID="82d7cf8c-129a-483a-b190-8480e2f03418" containerID="45a3b8435a1a4bc04f956f428cca44d0c3e5b295322aff907527d681ce5c1f1a" exitCode=0 Dec 13 07:11:56 crc kubenswrapper[4971]: I1213 07:11:56.202481 4971 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-757dc6fff9-hfmv4" event={"ID":"82d7cf8c-129a-483a-b190-8480e2f03418","Type":"ContainerDied","Data":"45a3b8435a1a4bc04f956f428cca44d0c3e5b295322aff907527d681ce5c1f1a"} Dec 13 07:11:58 crc kubenswrapper[4971]: I1213 07:11:57.614838 4971 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-757dc6fff9-hfmv4" Dec 13 07:11:58 crc kubenswrapper[4971]: I1213 07:11:57.756892 4971 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/82d7cf8c-129a-483a-b190-8480e2f03418-ovsdbserver-sb\") pod \"82d7cf8c-129a-483a-b190-8480e2f03418\" (UID: \"82d7cf8c-129a-483a-b190-8480e2f03418\") " Dec 13 07:11:58 crc kubenswrapper[4971]: I1213 07:11:57.757378 4971 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/82d7cf8c-129a-483a-b190-8480e2f03418-ovsdbserver-nb\") pod \"82d7cf8c-129a-483a-b190-8480e2f03418\" (UID: \"82d7cf8c-129a-483a-b190-8480e2f03418\") " Dec 13 07:11:58 crc kubenswrapper[4971]: I1213 07:11:57.757450 4971 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-jjxrl\" (UniqueName: \"kubernetes.io/projected/82d7cf8c-129a-483a-b190-8480e2f03418-kube-api-access-jjxrl\") pod \"82d7cf8c-129a-483a-b190-8480e2f03418\" (UID: \"82d7cf8c-129a-483a-b190-8480e2f03418\") " Dec 13 07:11:58 crc kubenswrapper[4971]: I1213 07:11:57.757483 4971 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/82d7cf8c-129a-483a-b190-8480e2f03418-dns-svc\") pod \"82d7cf8c-129a-483a-b190-8480e2f03418\" (UID: \"82d7cf8c-129a-483a-b190-8480e2f03418\") " Dec 13 07:11:58 crc kubenswrapper[4971]: I1213 07:11:57.757558 4971 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/82d7cf8c-129a-483a-b190-8480e2f03418-config\") pod \"82d7cf8c-129a-483a-b190-8480e2f03418\" (UID: \"82d7cf8c-129a-483a-b190-8480e2f03418\") " Dec 13 07:11:58 crc kubenswrapper[4971]: I1213 07:11:57.770733 4971 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/82d7cf8c-129a-483a-b190-8480e2f03418-kube-api-access-jjxrl" (OuterVolumeSpecName: "kube-api-access-jjxrl") pod "82d7cf8c-129a-483a-b190-8480e2f03418" (UID: "82d7cf8c-129a-483a-b190-8480e2f03418"). InnerVolumeSpecName "kube-api-access-jjxrl". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 13 07:11:58 crc kubenswrapper[4971]: I1213 07:11:57.804679 4971 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/82d7cf8c-129a-483a-b190-8480e2f03418-ovsdbserver-nb" (OuterVolumeSpecName: "ovsdbserver-nb") pod "82d7cf8c-129a-483a-b190-8480e2f03418" (UID: "82d7cf8c-129a-483a-b190-8480e2f03418"). InnerVolumeSpecName "ovsdbserver-nb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 13 07:11:58 crc kubenswrapper[4971]: I1213 07:11:57.806772 4971 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/82d7cf8c-129a-483a-b190-8480e2f03418-config" (OuterVolumeSpecName: "config") pod "82d7cf8c-129a-483a-b190-8480e2f03418" (UID: "82d7cf8c-129a-483a-b190-8480e2f03418"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 13 07:11:58 crc kubenswrapper[4971]: I1213 07:11:57.809709 4971 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/82d7cf8c-129a-483a-b190-8480e2f03418-dns-svc" (OuterVolumeSpecName: "dns-svc") pod "82d7cf8c-129a-483a-b190-8480e2f03418" (UID: "82d7cf8c-129a-483a-b190-8480e2f03418"). InnerVolumeSpecName "dns-svc". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 13 07:11:58 crc kubenswrapper[4971]: I1213 07:11:57.812451 4971 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/82d7cf8c-129a-483a-b190-8480e2f03418-ovsdbserver-sb" (OuterVolumeSpecName: "ovsdbserver-sb") pod "82d7cf8c-129a-483a-b190-8480e2f03418" (UID: "82d7cf8c-129a-483a-b190-8480e2f03418"). InnerVolumeSpecName "ovsdbserver-sb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 13 07:11:58 crc kubenswrapper[4971]: I1213 07:11:57.859964 4971 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/82d7cf8c-129a-483a-b190-8480e2f03418-ovsdbserver-nb\") on node \"crc\" DevicePath \"\"" Dec 13 07:11:58 crc kubenswrapper[4971]: I1213 07:11:57.860025 4971 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-jjxrl\" (UniqueName: \"kubernetes.io/projected/82d7cf8c-129a-483a-b190-8480e2f03418-kube-api-access-jjxrl\") on node \"crc\" DevicePath \"\"" Dec 13 07:11:58 crc kubenswrapper[4971]: I1213 07:11:57.860085 4971 reconciler_common.go:293] "Volume detached for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/82d7cf8c-129a-483a-b190-8480e2f03418-dns-svc\") on node \"crc\" DevicePath \"\"" Dec 13 07:11:58 crc kubenswrapper[4971]: I1213 07:11:57.860098 4971 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/82d7cf8c-129a-483a-b190-8480e2f03418-config\") on node \"crc\" DevicePath \"\"" Dec 13 07:11:58 crc kubenswrapper[4971]: I1213 07:11:57.860110 4971 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/82d7cf8c-129a-483a-b190-8480e2f03418-ovsdbserver-sb\") on node \"crc\" DevicePath \"\"" Dec 13 07:11:58 crc kubenswrapper[4971]: I1213 07:11:58.220370 4971 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-db-sync-b2cmm" event={"ID":"d6b11c79-8c13-42d7-8dce-c434dfab1912","Type":"ContainerStarted","Data":"9850b81303e2106176d4957d891cc8a2a2779060307683a31e2a544ab57a390b"} Dec 13 07:11:58 crc kubenswrapper[4971]: I1213 07:11:58.224745 4971 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-757dc6fff9-hfmv4" event={"ID":"82d7cf8c-129a-483a-b190-8480e2f03418","Type":"ContainerDied","Data":"a9a374a27e1ac718e356aee5609c2293261a89050a92b0a3ba7f122c99934131"} Dec 13 07:11:58 crc kubenswrapper[4971]: I1213 07:11:58.224813 4971 scope.go:117] "RemoveContainer" containerID="45a3b8435a1a4bc04f956f428cca44d0c3e5b295322aff907527d681ce5c1f1a" Dec 13 07:11:58 crc kubenswrapper[4971]: I1213 07:11:58.224954 4971 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-757dc6fff9-hfmv4" Dec 13 07:11:58 crc kubenswrapper[4971]: I1213 07:11:58.250695 4971 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/keystone-db-sync-b2cmm" podStartSLOduration=3.180806668 podStartE2EDuration="12.250672677s" podCreationTimestamp="2025-12-13 07:11:46 +0000 UTC" firstStartedPulling="2025-12-13 07:11:48.148280918 +0000 UTC m=+1364.752690366" lastFinishedPulling="2025-12-13 07:11:57.218146927 +0000 UTC m=+1373.822556375" observedRunningTime="2025-12-13 07:11:58.241304847 +0000 UTC m=+1374.845714295" watchObservedRunningTime="2025-12-13 07:11:58.250672677 +0000 UTC m=+1374.855082125" Dec 13 07:11:58 crc kubenswrapper[4971]: I1213 07:11:58.278844 4971 scope.go:117] "RemoveContainer" containerID="8a0537f444717cb3bdb32c5612b6fec9aea5fc7a82996f56458c25b1ba0630a0" Dec 13 07:11:58 crc kubenswrapper[4971]: I1213 07:11:58.283179 4971 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-757dc6fff9-hfmv4"] Dec 13 07:11:58 crc kubenswrapper[4971]: I1213 07:11:58.292345 4971 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-757dc6fff9-hfmv4"] Dec 13 07:11:59 crc kubenswrapper[4971]: I1213 07:11:59.781644 4971 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="82d7cf8c-129a-483a-b190-8480e2f03418" path="/var/lib/kubelet/pods/82d7cf8c-129a-483a-b190-8480e2f03418/volumes" Dec 13 07:12:01 crc kubenswrapper[4971]: I1213 07:12:01.262753 4971 generic.go:334] "Generic (PLEG): container finished" podID="d6b11c79-8c13-42d7-8dce-c434dfab1912" containerID="9850b81303e2106176d4957d891cc8a2a2779060307683a31e2a544ab57a390b" exitCode=0 Dec 13 07:12:01 crc kubenswrapper[4971]: I1213 07:12:01.262928 4971 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-db-sync-b2cmm" event={"ID":"d6b11c79-8c13-42d7-8dce-c434dfab1912","Type":"ContainerDied","Data":"9850b81303e2106176d4957d891cc8a2a2779060307683a31e2a544ab57a390b"} Dec 13 07:12:02 crc kubenswrapper[4971]: I1213 07:12:02.706939 4971 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-db-sync-b2cmm" Dec 13 07:12:02 crc kubenswrapper[4971]: I1213 07:12:02.845195 4971 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/d6b11c79-8c13-42d7-8dce-c434dfab1912-config-data\") pod \"d6b11c79-8c13-42d7-8dce-c434dfab1912\" (UID: \"d6b11c79-8c13-42d7-8dce-c434dfab1912\") " Dec 13 07:12:02 crc kubenswrapper[4971]: I1213 07:12:02.845282 4971 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d6b11c79-8c13-42d7-8dce-c434dfab1912-combined-ca-bundle\") pod \"d6b11c79-8c13-42d7-8dce-c434dfab1912\" (UID: \"d6b11c79-8c13-42d7-8dce-c434dfab1912\") " Dec 13 07:12:02 crc kubenswrapper[4971]: I1213 07:12:02.845371 4971 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-wr98f\" (UniqueName: \"kubernetes.io/projected/d6b11c79-8c13-42d7-8dce-c434dfab1912-kube-api-access-wr98f\") pod \"d6b11c79-8c13-42d7-8dce-c434dfab1912\" (UID: \"d6b11c79-8c13-42d7-8dce-c434dfab1912\") " Dec 13 07:12:02 crc kubenswrapper[4971]: I1213 07:12:02.970727 4971 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/d6b11c79-8c13-42d7-8dce-c434dfab1912-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "d6b11c79-8c13-42d7-8dce-c434dfab1912" (UID: "d6b11c79-8c13-42d7-8dce-c434dfab1912"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 13 07:12:02 crc kubenswrapper[4971]: I1213 07:12:02.975670 4971 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/d6b11c79-8c13-42d7-8dce-c434dfab1912-kube-api-access-wr98f" (OuterVolumeSpecName: "kube-api-access-wr98f") pod "d6b11c79-8c13-42d7-8dce-c434dfab1912" (UID: "d6b11c79-8c13-42d7-8dce-c434dfab1912"). InnerVolumeSpecName "kube-api-access-wr98f". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 13 07:12:02 crc kubenswrapper[4971]: I1213 07:12:02.983921 4971 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/d6b11c79-8c13-42d7-8dce-c434dfab1912-config-data" (OuterVolumeSpecName: "config-data") pod "d6b11c79-8c13-42d7-8dce-c434dfab1912" (UID: "d6b11c79-8c13-42d7-8dce-c434dfab1912"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 13 07:12:03 crc kubenswrapper[4971]: I1213 07:12:03.052901 4971 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d6b11c79-8c13-42d7-8dce-c434dfab1912-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 13 07:12:03 crc kubenswrapper[4971]: I1213 07:12:03.053270 4971 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-wr98f\" (UniqueName: \"kubernetes.io/projected/d6b11c79-8c13-42d7-8dce-c434dfab1912-kube-api-access-wr98f\") on node \"crc\" DevicePath \"\"" Dec 13 07:12:03 crc kubenswrapper[4971]: I1213 07:12:03.053284 4971 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/d6b11c79-8c13-42d7-8dce-c434dfab1912-config-data\") on node \"crc\" DevicePath \"\"" Dec 13 07:12:03 crc kubenswrapper[4971]: I1213 07:12:03.283217 4971 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-db-sync-b2cmm" event={"ID":"d6b11c79-8c13-42d7-8dce-c434dfab1912","Type":"ContainerDied","Data":"9286a985004d24f6ef97375b220cf0c6ff19b1d7ee0e228a33ea94c2e589ee2a"} Dec 13 07:12:03 crc kubenswrapper[4971]: I1213 07:12:03.283257 4971 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="9286a985004d24f6ef97375b220cf0c6ff19b1d7ee0e228a33ea94c2e589ee2a" Dec 13 07:12:03 crc kubenswrapper[4971]: I1213 07:12:03.283311 4971 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-db-sync-b2cmm" Dec 13 07:12:03 crc kubenswrapper[4971]: I1213 07:12:03.698541 4971 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/keystone-bootstrap-xp9z4"] Dec 13 07:12:03 crc kubenswrapper[4971]: E1213 07:12:03.698939 4971 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="4c7ef6d3-1b68-4fb6-9784-0ee0bb154a8d" containerName="mariadb-database-create" Dec 13 07:12:03 crc kubenswrapper[4971]: I1213 07:12:03.698955 4971 state_mem.go:107] "Deleted CPUSet assignment" podUID="4c7ef6d3-1b68-4fb6-9784-0ee0bb154a8d" containerName="mariadb-database-create" Dec 13 07:12:03 crc kubenswrapper[4971]: E1213 07:12:03.698967 4971 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="20b5092f-e5a2-467f-8dbc-e083156b0e20" containerName="mariadb-database-create" Dec 13 07:12:03 crc kubenswrapper[4971]: I1213 07:12:03.698974 4971 state_mem.go:107] "Deleted CPUSet assignment" podUID="20b5092f-e5a2-467f-8dbc-e083156b0e20" containerName="mariadb-database-create" Dec 13 07:12:03 crc kubenswrapper[4971]: E1213 07:12:03.698984 4971 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="82d7cf8c-129a-483a-b190-8480e2f03418" containerName="dnsmasq-dns" Dec 13 07:12:03 crc kubenswrapper[4971]: I1213 07:12:03.698992 4971 state_mem.go:107] "Deleted CPUSet assignment" podUID="82d7cf8c-129a-483a-b190-8480e2f03418" containerName="dnsmasq-dns" Dec 13 07:12:03 crc kubenswrapper[4971]: E1213 07:12:03.699006 4971 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="82d7cf8c-129a-483a-b190-8480e2f03418" containerName="init" Dec 13 07:12:03 crc kubenswrapper[4971]: I1213 07:12:03.699012 4971 state_mem.go:107] "Deleted CPUSet assignment" podUID="82d7cf8c-129a-483a-b190-8480e2f03418" containerName="init" Dec 13 07:12:03 crc kubenswrapper[4971]: E1213 07:12:03.699034 4971 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="d6b11c79-8c13-42d7-8dce-c434dfab1912" containerName="keystone-db-sync" Dec 13 07:12:03 crc kubenswrapper[4971]: I1213 07:12:03.699042 4971 state_mem.go:107] "Deleted CPUSet assignment" podUID="d6b11c79-8c13-42d7-8dce-c434dfab1912" containerName="keystone-db-sync" Dec 13 07:12:03 crc kubenswrapper[4971]: E1213 07:12:03.699054 4971 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="349ffc46-502d-4c10-9d0c-69bcfdb46fe7" containerName="mariadb-account-create-update" Dec 13 07:12:03 crc kubenswrapper[4971]: I1213 07:12:03.699062 4971 state_mem.go:107] "Deleted CPUSet assignment" podUID="349ffc46-502d-4c10-9d0c-69bcfdb46fe7" containerName="mariadb-account-create-update" Dec 13 07:12:03 crc kubenswrapper[4971]: E1213 07:12:03.699072 4971 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="48a5104a-2e2e-4828-bb54-bf40553d7ee0" containerName="mariadb-account-create-update" Dec 13 07:12:03 crc kubenswrapper[4971]: I1213 07:12:03.699080 4971 state_mem.go:107] "Deleted CPUSet assignment" podUID="48a5104a-2e2e-4828-bb54-bf40553d7ee0" containerName="mariadb-account-create-update" Dec 13 07:12:03 crc kubenswrapper[4971]: E1213 07:12:03.699092 4971 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="ad253e3a-226e-464a-9772-46f4b0081314" containerName="mariadb-account-create-update" Dec 13 07:12:03 crc kubenswrapper[4971]: I1213 07:12:03.699101 4971 state_mem.go:107] "Deleted CPUSet assignment" podUID="ad253e3a-226e-464a-9772-46f4b0081314" containerName="mariadb-account-create-update" Dec 13 07:12:03 crc kubenswrapper[4971]: E1213 07:12:03.699115 4971 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f136ffc5-3a49-4807-8025-10356e78ee46" containerName="mariadb-database-create" Dec 13 07:12:03 crc kubenswrapper[4971]: I1213 07:12:03.699122 4971 state_mem.go:107] "Deleted CPUSet assignment" podUID="f136ffc5-3a49-4807-8025-10356e78ee46" containerName="mariadb-database-create" Dec 13 07:12:03 crc kubenswrapper[4971]: I1213 07:12:03.699296 4971 memory_manager.go:354] "RemoveStaleState removing state" podUID="48a5104a-2e2e-4828-bb54-bf40553d7ee0" containerName="mariadb-account-create-update" Dec 13 07:12:03 crc kubenswrapper[4971]: I1213 07:12:03.699312 4971 memory_manager.go:354] "RemoveStaleState removing state" podUID="ad253e3a-226e-464a-9772-46f4b0081314" containerName="mariadb-account-create-update" Dec 13 07:12:03 crc kubenswrapper[4971]: I1213 07:12:03.699329 4971 memory_manager.go:354] "RemoveStaleState removing state" podUID="20b5092f-e5a2-467f-8dbc-e083156b0e20" containerName="mariadb-database-create" Dec 13 07:12:03 crc kubenswrapper[4971]: I1213 07:12:03.699345 4971 memory_manager.go:354] "RemoveStaleState removing state" podUID="f136ffc5-3a49-4807-8025-10356e78ee46" containerName="mariadb-database-create" Dec 13 07:12:03 crc kubenswrapper[4971]: I1213 07:12:03.699355 4971 memory_manager.go:354] "RemoveStaleState removing state" podUID="349ffc46-502d-4c10-9d0c-69bcfdb46fe7" containerName="mariadb-account-create-update" Dec 13 07:12:03 crc kubenswrapper[4971]: I1213 07:12:03.699364 4971 memory_manager.go:354] "RemoveStaleState removing state" podUID="82d7cf8c-129a-483a-b190-8480e2f03418" containerName="dnsmasq-dns" Dec 13 07:12:03 crc kubenswrapper[4971]: I1213 07:12:03.699374 4971 memory_manager.go:354] "RemoveStaleState removing state" podUID="4c7ef6d3-1b68-4fb6-9784-0ee0bb154a8d" containerName="mariadb-database-create" Dec 13 07:12:03 crc kubenswrapper[4971]: I1213 07:12:03.699385 4971 memory_manager.go:354] "RemoveStaleState removing state" podUID="d6b11c79-8c13-42d7-8dce-c434dfab1912" containerName="keystone-db-sync" Dec 13 07:12:03 crc kubenswrapper[4971]: I1213 07:12:03.700024 4971 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-bootstrap-xp9z4" Dec 13 07:12:03 crc kubenswrapper[4971]: I1213 07:12:03.702055 4971 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone-scripts" Dec 13 07:12:03 crc kubenswrapper[4971]: I1213 07:12:03.708470 4971 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"osp-secret" Dec 13 07:12:03 crc kubenswrapper[4971]: I1213 07:12:03.708662 4971 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone" Dec 13 07:12:03 crc kubenswrapper[4971]: I1213 07:12:03.708731 4971 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone-keystone-dockercfg-56xxb" Dec 13 07:12:03 crc kubenswrapper[4971]: I1213 07:12:03.708959 4971 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone-config-data" Dec 13 07:12:03 crc kubenswrapper[4971]: I1213 07:12:03.718662 4971 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-78fbc4bbf-6tztc"] Dec 13 07:12:03 crc kubenswrapper[4971]: I1213 07:12:03.720908 4971 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-78fbc4bbf-6tztc" Dec 13 07:12:03 crc kubenswrapper[4971]: I1213 07:12:03.734625 4971 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/keystone-bootstrap-xp9z4"] Dec 13 07:12:03 crc kubenswrapper[4971]: I1213 07:12:03.835426 4971 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/1b1d2613-64f2-4b48-aaf5-0681ff6cba2d-config\") pod \"dnsmasq-dns-78fbc4bbf-6tztc\" (UID: \"1b1d2613-64f2-4b48-aaf5-0681ff6cba2d\") " pod="openstack/dnsmasq-dns-78fbc4bbf-6tztc" Dec 13 07:12:03 crc kubenswrapper[4971]: I1213 07:12:03.835582 4971 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/1b1d2613-64f2-4b48-aaf5-0681ff6cba2d-ovsdbserver-sb\") pod \"dnsmasq-dns-78fbc4bbf-6tztc\" (UID: \"1b1d2613-64f2-4b48-aaf5-0681ff6cba2d\") " pod="openstack/dnsmasq-dns-78fbc4bbf-6tztc" Dec 13 07:12:03 crc kubenswrapper[4971]: I1213 07:12:03.835776 4971 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/1b1d2613-64f2-4b48-aaf5-0681ff6cba2d-ovsdbserver-nb\") pod \"dnsmasq-dns-78fbc4bbf-6tztc\" (UID: \"1b1d2613-64f2-4b48-aaf5-0681ff6cba2d\") " pod="openstack/dnsmasq-dns-78fbc4bbf-6tztc" Dec 13 07:12:03 crc kubenswrapper[4971]: I1213 07:12:03.835856 4971 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-4q2wg\" (UniqueName: \"kubernetes.io/projected/1b1d2613-64f2-4b48-aaf5-0681ff6cba2d-kube-api-access-4q2wg\") pod \"dnsmasq-dns-78fbc4bbf-6tztc\" (UID: \"1b1d2613-64f2-4b48-aaf5-0681ff6cba2d\") " pod="openstack/dnsmasq-dns-78fbc4bbf-6tztc" Dec 13 07:12:03 crc kubenswrapper[4971]: I1213 07:12:03.835885 4971 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/1b1d2613-64f2-4b48-aaf5-0681ff6cba2d-dns-svc\") pod \"dnsmasq-dns-78fbc4bbf-6tztc\" (UID: \"1b1d2613-64f2-4b48-aaf5-0681ff6cba2d\") " pod="openstack/dnsmasq-dns-78fbc4bbf-6tztc" Dec 13 07:12:04 crc kubenswrapper[4971]: I1213 07:12:04.092711 4971 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-78fbc4bbf-6tztc"] Dec 13 07:12:04 crc kubenswrapper[4971]: I1213 07:12:04.109131 4971 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/1b1d2613-64f2-4b48-aaf5-0681ff6cba2d-ovsdbserver-sb\") pod \"dnsmasq-dns-78fbc4bbf-6tztc\" (UID: \"1b1d2613-64f2-4b48-aaf5-0681ff6cba2d\") " pod="openstack/dnsmasq-dns-78fbc4bbf-6tztc" Dec 13 07:12:04 crc kubenswrapper[4971]: I1213 07:12:04.109289 4971 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/5b9e78cc-aa6e-4369-ab8a-5e89b86359c1-config-data\") pod \"keystone-bootstrap-xp9z4\" (UID: \"5b9e78cc-aa6e-4369-ab8a-5e89b86359c1\") " pod="openstack/keystone-bootstrap-xp9z4" Dec 13 07:12:04 crc kubenswrapper[4971]: I1213 07:12:04.109342 4971 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/5b9e78cc-aa6e-4369-ab8a-5e89b86359c1-scripts\") pod \"keystone-bootstrap-xp9z4\" (UID: \"5b9e78cc-aa6e-4369-ab8a-5e89b86359c1\") " pod="openstack/keystone-bootstrap-xp9z4" Dec 13 07:12:04 crc kubenswrapper[4971]: I1213 07:12:04.109365 4971 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/1b1d2613-64f2-4b48-aaf5-0681ff6cba2d-ovsdbserver-nb\") pod \"dnsmasq-dns-78fbc4bbf-6tztc\" (UID: \"1b1d2613-64f2-4b48-aaf5-0681ff6cba2d\") " pod="openstack/dnsmasq-dns-78fbc4bbf-6tztc" Dec 13 07:12:04 crc kubenswrapper[4971]: I1213 07:12:04.109384 4971 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/5b9e78cc-aa6e-4369-ab8a-5e89b86359c1-fernet-keys\") pod \"keystone-bootstrap-xp9z4\" (UID: \"5b9e78cc-aa6e-4369-ab8a-5e89b86359c1\") " pod="openstack/keystone-bootstrap-xp9z4" Dec 13 07:12:04 crc kubenswrapper[4971]: I1213 07:12:04.109434 4971 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-ds44p\" (UniqueName: \"kubernetes.io/projected/5b9e78cc-aa6e-4369-ab8a-5e89b86359c1-kube-api-access-ds44p\") pod \"keystone-bootstrap-xp9z4\" (UID: \"5b9e78cc-aa6e-4369-ab8a-5e89b86359c1\") " pod="openstack/keystone-bootstrap-xp9z4" Dec 13 07:12:04 crc kubenswrapper[4971]: I1213 07:12:04.109471 4971 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"credential-keys\" (UniqueName: \"kubernetes.io/secret/5b9e78cc-aa6e-4369-ab8a-5e89b86359c1-credential-keys\") pod \"keystone-bootstrap-xp9z4\" (UID: \"5b9e78cc-aa6e-4369-ab8a-5e89b86359c1\") " pod="openstack/keystone-bootstrap-xp9z4" Dec 13 07:12:04 crc kubenswrapper[4971]: I1213 07:12:04.109494 4971 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-4q2wg\" (UniqueName: \"kubernetes.io/projected/1b1d2613-64f2-4b48-aaf5-0681ff6cba2d-kube-api-access-4q2wg\") pod \"dnsmasq-dns-78fbc4bbf-6tztc\" (UID: \"1b1d2613-64f2-4b48-aaf5-0681ff6cba2d\") " pod="openstack/dnsmasq-dns-78fbc4bbf-6tztc" Dec 13 07:12:04 crc kubenswrapper[4971]: I1213 07:12:04.109533 4971 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/1b1d2613-64f2-4b48-aaf5-0681ff6cba2d-dns-svc\") pod \"dnsmasq-dns-78fbc4bbf-6tztc\" (UID: \"1b1d2613-64f2-4b48-aaf5-0681ff6cba2d\") " pod="openstack/dnsmasq-dns-78fbc4bbf-6tztc" Dec 13 07:12:04 crc kubenswrapper[4971]: I1213 07:12:04.109566 4971 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/5b9e78cc-aa6e-4369-ab8a-5e89b86359c1-combined-ca-bundle\") pod \"keystone-bootstrap-xp9z4\" (UID: \"5b9e78cc-aa6e-4369-ab8a-5e89b86359c1\") " pod="openstack/keystone-bootstrap-xp9z4" Dec 13 07:12:04 crc kubenswrapper[4971]: I1213 07:12:04.109621 4971 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/1b1d2613-64f2-4b48-aaf5-0681ff6cba2d-config\") pod \"dnsmasq-dns-78fbc4bbf-6tztc\" (UID: \"1b1d2613-64f2-4b48-aaf5-0681ff6cba2d\") " pod="openstack/dnsmasq-dns-78fbc4bbf-6tztc" Dec 13 07:12:04 crc kubenswrapper[4971]: I1213 07:12:04.110558 4971 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/1b1d2613-64f2-4b48-aaf5-0681ff6cba2d-config\") pod \"dnsmasq-dns-78fbc4bbf-6tztc\" (UID: \"1b1d2613-64f2-4b48-aaf5-0681ff6cba2d\") " pod="openstack/dnsmasq-dns-78fbc4bbf-6tztc" Dec 13 07:12:04 crc kubenswrapper[4971]: I1213 07:12:04.111226 4971 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/1b1d2613-64f2-4b48-aaf5-0681ff6cba2d-ovsdbserver-sb\") pod \"dnsmasq-dns-78fbc4bbf-6tztc\" (UID: \"1b1d2613-64f2-4b48-aaf5-0681ff6cba2d\") " pod="openstack/dnsmasq-dns-78fbc4bbf-6tztc" Dec 13 07:12:04 crc kubenswrapper[4971]: I1213 07:12:04.112062 4971 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/1b1d2613-64f2-4b48-aaf5-0681ff6cba2d-ovsdbserver-nb\") pod \"dnsmasq-dns-78fbc4bbf-6tztc\" (UID: \"1b1d2613-64f2-4b48-aaf5-0681ff6cba2d\") " pod="openstack/dnsmasq-dns-78fbc4bbf-6tztc" Dec 13 07:12:04 crc kubenswrapper[4971]: I1213 07:12:04.112590 4971 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/1b1d2613-64f2-4b48-aaf5-0681ff6cba2d-dns-svc\") pod \"dnsmasq-dns-78fbc4bbf-6tztc\" (UID: \"1b1d2613-64f2-4b48-aaf5-0681ff6cba2d\") " pod="openstack/dnsmasq-dns-78fbc4bbf-6tztc" Dec 13 07:12:04 crc kubenswrapper[4971]: I1213 07:12:04.124891 4971 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/cinder-db-sync-26k47"] Dec 13 07:12:04 crc kubenswrapper[4971]: I1213 07:12:04.126052 4971 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-db-sync-26k47" Dec 13 07:12:04 crc kubenswrapper[4971]: I1213 07:12:04.140088 4971 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cinder-config-data" Dec 13 07:12:04 crc kubenswrapper[4971]: I1213 07:12:04.140326 4971 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cinder-scripts" Dec 13 07:12:04 crc kubenswrapper[4971]: I1213 07:12:04.140429 4971 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cinder-cinder-dockercfg-dxkxj" Dec 13 07:12:04 crc kubenswrapper[4971]: I1213 07:12:04.159169 4971 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cinder-db-sync-26k47"] Dec 13 07:12:04 crc kubenswrapper[4971]: I1213 07:12:04.160260 4971 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-4q2wg\" (UniqueName: \"kubernetes.io/projected/1b1d2613-64f2-4b48-aaf5-0681ff6cba2d-kube-api-access-4q2wg\") pod \"dnsmasq-dns-78fbc4bbf-6tztc\" (UID: \"1b1d2613-64f2-4b48-aaf5-0681ff6cba2d\") " pod="openstack/dnsmasq-dns-78fbc4bbf-6tztc" Dec 13 07:12:04 crc kubenswrapper[4971]: I1213 07:12:04.173374 4971 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/horizon-bcdd7767-j7d7f"] Dec 13 07:12:04 crc kubenswrapper[4971]: I1213 07:12:04.175547 4971 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/horizon-bcdd7767-j7d7f" Dec 13 07:12:04 crc kubenswrapper[4971]: I1213 07:12:04.179168 4971 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"horizon-scripts" Dec 13 07:12:04 crc kubenswrapper[4971]: I1213 07:12:04.179408 4971 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"horizon-horizon-dockercfg-9bbfr" Dec 13 07:12:04 crc kubenswrapper[4971]: I1213 07:12:04.179544 4971 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"horizon" Dec 13 07:12:04 crc kubenswrapper[4971]: I1213 07:12:04.179659 4971 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"horizon-config-data" Dec 13 07:12:04 crc kubenswrapper[4971]: I1213 07:12:04.180000 4971 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ceilometer-0"] Dec 13 07:12:04 crc kubenswrapper[4971]: I1213 07:12:04.181954 4971 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Dec 13 07:12:04 crc kubenswrapper[4971]: I1213 07:12:04.201415 4971 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceilometer-config-data" Dec 13 07:12:04 crc kubenswrapper[4971]: I1213 07:12:04.201871 4971 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceilometer-scripts" Dec 13 07:12:04 crc kubenswrapper[4971]: I1213 07:12:04.210805 4971 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"credential-keys\" (UniqueName: \"kubernetes.io/secret/5b9e78cc-aa6e-4369-ab8a-5e89b86359c1-credential-keys\") pod \"keystone-bootstrap-xp9z4\" (UID: \"5b9e78cc-aa6e-4369-ab8a-5e89b86359c1\") " pod="openstack/keystone-bootstrap-xp9z4" Dec 13 07:12:04 crc kubenswrapper[4971]: I1213 07:12:04.210855 4971 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/56e56ce1-4dd1-4ae5-b868-e98d27d5ad94-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"56e56ce1-4dd1-4ae5-b868-e98d27d5ad94\") " pod="openstack/ceilometer-0" Dec 13 07:12:04 crc kubenswrapper[4971]: I1213 07:12:04.210872 4971 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/ff35e760-13e7-4145-9970-742b73a36176-logs\") pod \"horizon-bcdd7767-j7d7f\" (UID: \"ff35e760-13e7-4145-9970-742b73a36176\") " pod="openstack/horizon-bcdd7767-j7d7f" Dec 13 07:12:04 crc kubenswrapper[4971]: I1213 07:12:04.210887 4971 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/5b9e78cc-aa6e-4369-ab8a-5e89b86359c1-combined-ca-bundle\") pod \"keystone-bootstrap-xp9z4\" (UID: \"5b9e78cc-aa6e-4369-ab8a-5e89b86359c1\") " pod="openstack/keystone-bootstrap-xp9z4" Dec 13 07:12:04 crc kubenswrapper[4971]: I1213 07:12:04.210909 4971 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-hw5mk\" (UniqueName: \"kubernetes.io/projected/56e56ce1-4dd1-4ae5-b868-e98d27d5ad94-kube-api-access-hw5mk\") pod \"ceilometer-0\" (UID: \"56e56ce1-4dd1-4ae5-b868-e98d27d5ad94\") " pod="openstack/ceilometer-0" Dec 13 07:12:04 crc kubenswrapper[4971]: I1213 07:12:04.210924 4971 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/56e56ce1-4dd1-4ae5-b868-e98d27d5ad94-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"56e56ce1-4dd1-4ae5-b868-e98d27d5ad94\") " pod="openstack/ceilometer-0" Dec 13 07:12:04 crc kubenswrapper[4971]: I1213 07:12:04.210945 4971 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-r4lhp\" (UniqueName: \"kubernetes.io/projected/c1d819d0-339f-4e7e-9180-63d2f7c0042e-kube-api-access-r4lhp\") pod \"cinder-db-sync-26k47\" (UID: \"c1d819d0-339f-4e7e-9180-63d2f7c0042e\") " pod="openstack/cinder-db-sync-26k47" Dec 13 07:12:04 crc kubenswrapper[4971]: I1213 07:12:04.210961 4971 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-cpxzk\" (UniqueName: \"kubernetes.io/projected/ff35e760-13e7-4145-9970-742b73a36176-kube-api-access-cpxzk\") pod \"horizon-bcdd7767-j7d7f\" (UID: \"ff35e760-13e7-4145-9970-742b73a36176\") " pod="openstack/horizon-bcdd7767-j7d7f" Dec 13 07:12:04 crc kubenswrapper[4971]: I1213 07:12:04.210974 4971 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/56e56ce1-4dd1-4ae5-b868-e98d27d5ad94-log-httpd\") pod \"ceilometer-0\" (UID: \"56e56ce1-4dd1-4ae5-b868-e98d27d5ad94\") " pod="openstack/ceilometer-0" Dec 13 07:12:04 crc kubenswrapper[4971]: I1213 07:12:04.210988 4971 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/ff35e760-13e7-4145-9970-742b73a36176-config-data\") pod \"horizon-bcdd7767-j7d7f\" (UID: \"ff35e760-13e7-4145-9970-742b73a36176\") " pod="openstack/horizon-bcdd7767-j7d7f" Dec 13 07:12:04 crc kubenswrapper[4971]: I1213 07:12:04.211041 4971 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/56e56ce1-4dd1-4ae5-b868-e98d27d5ad94-config-data\") pod \"ceilometer-0\" (UID: \"56e56ce1-4dd1-4ae5-b868-e98d27d5ad94\") " pod="openstack/ceilometer-0" Dec 13 07:12:04 crc kubenswrapper[4971]: I1213 07:12:04.211060 4971 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"horizon-secret-key\" (UniqueName: \"kubernetes.io/secret/ff35e760-13e7-4145-9970-742b73a36176-horizon-secret-key\") pod \"horizon-bcdd7767-j7d7f\" (UID: \"ff35e760-13e7-4145-9970-742b73a36176\") " pod="openstack/horizon-bcdd7767-j7d7f" Dec 13 07:12:04 crc kubenswrapper[4971]: I1213 07:12:04.211091 4971 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/c1d819d0-339f-4e7e-9180-63d2f7c0042e-config-data\") pod \"cinder-db-sync-26k47\" (UID: \"c1d819d0-339f-4e7e-9180-63d2f7c0042e\") " pod="openstack/cinder-db-sync-26k47" Dec 13 07:12:04 crc kubenswrapper[4971]: I1213 07:12:04.211117 4971 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/c1d819d0-339f-4e7e-9180-63d2f7c0042e-scripts\") pod \"cinder-db-sync-26k47\" (UID: \"c1d819d0-339f-4e7e-9180-63d2f7c0042e\") " pod="openstack/cinder-db-sync-26k47" Dec 13 07:12:04 crc kubenswrapper[4971]: I1213 07:12:04.211133 4971 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/ff35e760-13e7-4145-9970-742b73a36176-scripts\") pod \"horizon-bcdd7767-j7d7f\" (UID: \"ff35e760-13e7-4145-9970-742b73a36176\") " pod="openstack/horizon-bcdd7767-j7d7f" Dec 13 07:12:04 crc kubenswrapper[4971]: I1213 07:12:04.211152 4971 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/c1d819d0-339f-4e7e-9180-63d2f7c0042e-etc-machine-id\") pod \"cinder-db-sync-26k47\" (UID: \"c1d819d0-339f-4e7e-9180-63d2f7c0042e\") " pod="openstack/cinder-db-sync-26k47" Dec 13 07:12:04 crc kubenswrapper[4971]: I1213 07:12:04.211169 4971 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/c1d819d0-339f-4e7e-9180-63d2f7c0042e-combined-ca-bundle\") pod \"cinder-db-sync-26k47\" (UID: \"c1d819d0-339f-4e7e-9180-63d2f7c0042e\") " pod="openstack/cinder-db-sync-26k47" Dec 13 07:12:04 crc kubenswrapper[4971]: I1213 07:12:04.211190 4971 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/5b9e78cc-aa6e-4369-ab8a-5e89b86359c1-config-data\") pod \"keystone-bootstrap-xp9z4\" (UID: \"5b9e78cc-aa6e-4369-ab8a-5e89b86359c1\") " pod="openstack/keystone-bootstrap-xp9z4" Dec 13 07:12:04 crc kubenswrapper[4971]: I1213 07:12:04.211205 4971 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/56e56ce1-4dd1-4ae5-b868-e98d27d5ad94-run-httpd\") pod \"ceilometer-0\" (UID: \"56e56ce1-4dd1-4ae5-b868-e98d27d5ad94\") " pod="openstack/ceilometer-0" Dec 13 07:12:04 crc kubenswrapper[4971]: I1213 07:12:04.211228 4971 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/c1d819d0-339f-4e7e-9180-63d2f7c0042e-db-sync-config-data\") pod \"cinder-db-sync-26k47\" (UID: \"c1d819d0-339f-4e7e-9180-63d2f7c0042e\") " pod="openstack/cinder-db-sync-26k47" Dec 13 07:12:04 crc kubenswrapper[4971]: I1213 07:12:04.211251 4971 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/5b9e78cc-aa6e-4369-ab8a-5e89b86359c1-scripts\") pod \"keystone-bootstrap-xp9z4\" (UID: \"5b9e78cc-aa6e-4369-ab8a-5e89b86359c1\") " pod="openstack/keystone-bootstrap-xp9z4" Dec 13 07:12:04 crc kubenswrapper[4971]: I1213 07:12:04.211267 4971 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/56e56ce1-4dd1-4ae5-b868-e98d27d5ad94-scripts\") pod \"ceilometer-0\" (UID: \"56e56ce1-4dd1-4ae5-b868-e98d27d5ad94\") " pod="openstack/ceilometer-0" Dec 13 07:12:04 crc kubenswrapper[4971]: I1213 07:12:04.211285 4971 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/5b9e78cc-aa6e-4369-ab8a-5e89b86359c1-fernet-keys\") pod \"keystone-bootstrap-xp9z4\" (UID: \"5b9e78cc-aa6e-4369-ab8a-5e89b86359c1\") " pod="openstack/keystone-bootstrap-xp9z4" Dec 13 07:12:04 crc kubenswrapper[4971]: I1213 07:12:04.211303 4971 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-ds44p\" (UniqueName: \"kubernetes.io/projected/5b9e78cc-aa6e-4369-ab8a-5e89b86359c1-kube-api-access-ds44p\") pod \"keystone-bootstrap-xp9z4\" (UID: \"5b9e78cc-aa6e-4369-ab8a-5e89b86359c1\") " pod="openstack/keystone-bootstrap-xp9z4" Dec 13 07:12:04 crc kubenswrapper[4971]: I1213 07:12:04.224968 4971 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/5b9e78cc-aa6e-4369-ab8a-5e89b86359c1-scripts\") pod \"keystone-bootstrap-xp9z4\" (UID: \"5b9e78cc-aa6e-4369-ab8a-5e89b86359c1\") " pod="openstack/keystone-bootstrap-xp9z4" Dec 13 07:12:04 crc kubenswrapper[4971]: I1213 07:12:04.231313 4971 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/5b9e78cc-aa6e-4369-ab8a-5e89b86359c1-fernet-keys\") pod \"keystone-bootstrap-xp9z4\" (UID: \"5b9e78cc-aa6e-4369-ab8a-5e89b86359c1\") " pod="openstack/keystone-bootstrap-xp9z4" Dec 13 07:12:04 crc kubenswrapper[4971]: I1213 07:12:04.232069 4971 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/5b9e78cc-aa6e-4369-ab8a-5e89b86359c1-combined-ca-bundle\") pod \"keystone-bootstrap-xp9z4\" (UID: \"5b9e78cc-aa6e-4369-ab8a-5e89b86359c1\") " pod="openstack/keystone-bootstrap-xp9z4" Dec 13 07:12:04 crc kubenswrapper[4971]: I1213 07:12:04.247795 4971 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/horizon-bcdd7767-j7d7f"] Dec 13 07:12:04 crc kubenswrapper[4971]: I1213 07:12:04.249184 4971 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"credential-keys\" (UniqueName: \"kubernetes.io/secret/5b9e78cc-aa6e-4369-ab8a-5e89b86359c1-credential-keys\") pod \"keystone-bootstrap-xp9z4\" (UID: \"5b9e78cc-aa6e-4369-ab8a-5e89b86359c1\") " pod="openstack/keystone-bootstrap-xp9z4" Dec 13 07:12:04 crc kubenswrapper[4971]: I1213 07:12:04.251806 4971 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/5b9e78cc-aa6e-4369-ab8a-5e89b86359c1-config-data\") pod \"keystone-bootstrap-xp9z4\" (UID: \"5b9e78cc-aa6e-4369-ab8a-5e89b86359c1\") " pod="openstack/keystone-bootstrap-xp9z4" Dec 13 07:12:04 crc kubenswrapper[4971]: I1213 07:12:04.254967 4971 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ceilometer-0"] Dec 13 07:12:04 crc kubenswrapper[4971]: I1213 07:12:04.275283 4971 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-ds44p\" (UniqueName: \"kubernetes.io/projected/5b9e78cc-aa6e-4369-ab8a-5e89b86359c1-kube-api-access-ds44p\") pod \"keystone-bootstrap-xp9z4\" (UID: \"5b9e78cc-aa6e-4369-ab8a-5e89b86359c1\") " pod="openstack/keystone-bootstrap-xp9z4" Dec 13 07:12:04 crc kubenswrapper[4971]: I1213 07:12:04.288753 4971 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/neutron-db-sync-8z2nm"] Dec 13 07:12:04 crc kubenswrapper[4971]: I1213 07:12:04.292536 4971 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-db-sync-8z2nm" Dec 13 07:12:04 crc kubenswrapper[4971]: I1213 07:12:04.298259 4971 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"neutron-httpd-config" Dec 13 07:12:04 crc kubenswrapper[4971]: I1213 07:12:04.298494 4971 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"neutron-config" Dec 13 07:12:04 crc kubenswrapper[4971]: I1213 07:12:04.298655 4971 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"neutron-neutron-dockercfg-tz9f5" Dec 13 07:12:04 crc kubenswrapper[4971]: I1213 07:12:04.311765 4971 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/c1d819d0-339f-4e7e-9180-63d2f7c0042e-config-data\") pod \"cinder-db-sync-26k47\" (UID: \"c1d819d0-339f-4e7e-9180-63d2f7c0042e\") " pod="openstack/cinder-db-sync-26k47" Dec 13 07:12:04 crc kubenswrapper[4971]: I1213 07:12:04.311812 4971 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/c1d819d0-339f-4e7e-9180-63d2f7c0042e-scripts\") pod \"cinder-db-sync-26k47\" (UID: \"c1d819d0-339f-4e7e-9180-63d2f7c0042e\") " pod="openstack/cinder-db-sync-26k47" Dec 13 07:12:04 crc kubenswrapper[4971]: I1213 07:12:04.311832 4971 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/ff35e760-13e7-4145-9970-742b73a36176-scripts\") pod \"horizon-bcdd7767-j7d7f\" (UID: \"ff35e760-13e7-4145-9970-742b73a36176\") " pod="openstack/horizon-bcdd7767-j7d7f" Dec 13 07:12:04 crc kubenswrapper[4971]: I1213 07:12:04.311851 4971 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/c1d819d0-339f-4e7e-9180-63d2f7c0042e-etc-machine-id\") pod \"cinder-db-sync-26k47\" (UID: \"c1d819d0-339f-4e7e-9180-63d2f7c0042e\") " pod="openstack/cinder-db-sync-26k47" Dec 13 07:12:04 crc kubenswrapper[4971]: I1213 07:12:04.311869 4971 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/c1d819d0-339f-4e7e-9180-63d2f7c0042e-combined-ca-bundle\") pod \"cinder-db-sync-26k47\" (UID: \"c1d819d0-339f-4e7e-9180-63d2f7c0042e\") " pod="openstack/cinder-db-sync-26k47" Dec 13 07:12:04 crc kubenswrapper[4971]: I1213 07:12:04.311889 4971 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/56e56ce1-4dd1-4ae5-b868-e98d27d5ad94-run-httpd\") pod \"ceilometer-0\" (UID: \"56e56ce1-4dd1-4ae5-b868-e98d27d5ad94\") " pod="openstack/ceilometer-0" Dec 13 07:12:04 crc kubenswrapper[4971]: I1213 07:12:04.311910 4971 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/c1d819d0-339f-4e7e-9180-63d2f7c0042e-db-sync-config-data\") pod \"cinder-db-sync-26k47\" (UID: \"c1d819d0-339f-4e7e-9180-63d2f7c0042e\") " pod="openstack/cinder-db-sync-26k47" Dec 13 07:12:04 crc kubenswrapper[4971]: I1213 07:12:04.311934 4971 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/secret/0813350d-430f-4564-a09a-062a3239ba2f-config\") pod \"neutron-db-sync-8z2nm\" (UID: \"0813350d-430f-4564-a09a-062a3239ba2f\") " pod="openstack/neutron-db-sync-8z2nm" Dec 13 07:12:04 crc kubenswrapper[4971]: I1213 07:12:04.311951 4971 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/56e56ce1-4dd1-4ae5-b868-e98d27d5ad94-scripts\") pod \"ceilometer-0\" (UID: \"56e56ce1-4dd1-4ae5-b868-e98d27d5ad94\") " pod="openstack/ceilometer-0" Dec 13 07:12:04 crc kubenswrapper[4971]: I1213 07:12:04.311987 4971 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/56e56ce1-4dd1-4ae5-b868-e98d27d5ad94-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"56e56ce1-4dd1-4ae5-b868-e98d27d5ad94\") " pod="openstack/ceilometer-0" Dec 13 07:12:04 crc kubenswrapper[4971]: I1213 07:12:04.312007 4971 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/ff35e760-13e7-4145-9970-742b73a36176-logs\") pod \"horizon-bcdd7767-j7d7f\" (UID: \"ff35e760-13e7-4145-9970-742b73a36176\") " pod="openstack/horizon-bcdd7767-j7d7f" Dec 13 07:12:04 crc kubenswrapper[4971]: I1213 07:12:04.312024 4971 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-hw5mk\" (UniqueName: \"kubernetes.io/projected/56e56ce1-4dd1-4ae5-b868-e98d27d5ad94-kube-api-access-hw5mk\") pod \"ceilometer-0\" (UID: \"56e56ce1-4dd1-4ae5-b868-e98d27d5ad94\") " pod="openstack/ceilometer-0" Dec 13 07:12:04 crc kubenswrapper[4971]: I1213 07:12:04.312040 4971 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/56e56ce1-4dd1-4ae5-b868-e98d27d5ad94-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"56e56ce1-4dd1-4ae5-b868-e98d27d5ad94\") " pod="openstack/ceilometer-0" Dec 13 07:12:04 crc kubenswrapper[4971]: I1213 07:12:04.312090 4971 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-r4lhp\" (UniqueName: \"kubernetes.io/projected/c1d819d0-339f-4e7e-9180-63d2f7c0042e-kube-api-access-r4lhp\") pod \"cinder-db-sync-26k47\" (UID: \"c1d819d0-339f-4e7e-9180-63d2f7c0042e\") " pod="openstack/cinder-db-sync-26k47" Dec 13 07:12:04 crc kubenswrapper[4971]: I1213 07:12:04.312108 4971 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cpxzk\" (UniqueName: \"kubernetes.io/projected/ff35e760-13e7-4145-9970-742b73a36176-kube-api-access-cpxzk\") pod \"horizon-bcdd7767-j7d7f\" (UID: \"ff35e760-13e7-4145-9970-742b73a36176\") " pod="openstack/horizon-bcdd7767-j7d7f" Dec 13 07:12:04 crc kubenswrapper[4971]: I1213 07:12:04.312123 4971 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/56e56ce1-4dd1-4ae5-b868-e98d27d5ad94-log-httpd\") pod \"ceilometer-0\" (UID: \"56e56ce1-4dd1-4ae5-b868-e98d27d5ad94\") " pod="openstack/ceilometer-0" Dec 13 07:12:04 crc kubenswrapper[4971]: I1213 07:12:04.312136 4971 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/ff35e760-13e7-4145-9970-742b73a36176-config-data\") pod \"horizon-bcdd7767-j7d7f\" (UID: \"ff35e760-13e7-4145-9970-742b73a36176\") " pod="openstack/horizon-bcdd7767-j7d7f" Dec 13 07:12:04 crc kubenswrapper[4971]: I1213 07:12:04.312152 4971 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/56e56ce1-4dd1-4ae5-b868-e98d27d5ad94-config-data\") pod \"ceilometer-0\" (UID: \"56e56ce1-4dd1-4ae5-b868-e98d27d5ad94\") " pod="openstack/ceilometer-0" Dec 13 07:12:04 crc kubenswrapper[4971]: I1213 07:12:04.312169 4971 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"horizon-secret-key\" (UniqueName: \"kubernetes.io/secret/ff35e760-13e7-4145-9970-742b73a36176-horizon-secret-key\") pod \"horizon-bcdd7767-j7d7f\" (UID: \"ff35e760-13e7-4145-9970-742b73a36176\") " pod="openstack/horizon-bcdd7767-j7d7f" Dec 13 07:12:04 crc kubenswrapper[4971]: I1213 07:12:04.312184 4971 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/0813350d-430f-4564-a09a-062a3239ba2f-combined-ca-bundle\") pod \"neutron-db-sync-8z2nm\" (UID: \"0813350d-430f-4564-a09a-062a3239ba2f\") " pod="openstack/neutron-db-sync-8z2nm" Dec 13 07:12:04 crc kubenswrapper[4971]: I1213 07:12:04.312208 4971 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-ktmdv\" (UniqueName: \"kubernetes.io/projected/0813350d-430f-4564-a09a-062a3239ba2f-kube-api-access-ktmdv\") pod \"neutron-db-sync-8z2nm\" (UID: \"0813350d-430f-4564-a09a-062a3239ba2f\") " pod="openstack/neutron-db-sync-8z2nm" Dec 13 07:12:04 crc kubenswrapper[4971]: I1213 07:12:04.314468 4971 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/ff35e760-13e7-4145-9970-742b73a36176-logs\") pod \"horizon-bcdd7767-j7d7f\" (UID: \"ff35e760-13e7-4145-9970-742b73a36176\") " pod="openstack/horizon-bcdd7767-j7d7f" Dec 13 07:12:04 crc kubenswrapper[4971]: I1213 07:12:04.317699 4971 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/56e56ce1-4dd1-4ae5-b868-e98d27d5ad94-log-httpd\") pod \"ceilometer-0\" (UID: \"56e56ce1-4dd1-4ae5-b868-e98d27d5ad94\") " pod="openstack/ceilometer-0" Dec 13 07:12:04 crc kubenswrapper[4971]: I1213 07:12:04.324259 4971 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/56e56ce1-4dd1-4ae5-b868-e98d27d5ad94-run-httpd\") pod \"ceilometer-0\" (UID: \"56e56ce1-4dd1-4ae5-b868-e98d27d5ad94\") " pod="openstack/ceilometer-0" Dec 13 07:12:04 crc kubenswrapper[4971]: I1213 07:12:04.327329 4971 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/56e56ce1-4dd1-4ae5-b868-e98d27d5ad94-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"56e56ce1-4dd1-4ae5-b868-e98d27d5ad94\") " pod="openstack/ceilometer-0" Dec 13 07:12:04 crc kubenswrapper[4971]: I1213 07:12:04.327575 4971 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/c1d819d0-339f-4e7e-9180-63d2f7c0042e-config-data\") pod \"cinder-db-sync-26k47\" (UID: \"c1d819d0-339f-4e7e-9180-63d2f7c0042e\") " pod="openstack/cinder-db-sync-26k47" Dec 13 07:12:04 crc kubenswrapper[4971]: I1213 07:12:04.329198 4971 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/c1d819d0-339f-4e7e-9180-63d2f7c0042e-etc-machine-id\") pod \"cinder-db-sync-26k47\" (UID: \"c1d819d0-339f-4e7e-9180-63d2f7c0042e\") " pod="openstack/cinder-db-sync-26k47" Dec 13 07:12:04 crc kubenswrapper[4971]: I1213 07:12:04.330832 4971 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/ff35e760-13e7-4145-9970-742b73a36176-config-data\") pod \"horizon-bcdd7767-j7d7f\" (UID: \"ff35e760-13e7-4145-9970-742b73a36176\") " pod="openstack/horizon-bcdd7767-j7d7f" Dec 13 07:12:04 crc kubenswrapper[4971]: I1213 07:12:04.339141 4971 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/56e56ce1-4dd1-4ae5-b868-e98d27d5ad94-config-data\") pod \"ceilometer-0\" (UID: \"56e56ce1-4dd1-4ae5-b868-e98d27d5ad94\") " pod="openstack/ceilometer-0" Dec 13 07:12:04 crc kubenswrapper[4971]: I1213 07:12:04.340478 4971 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/56e56ce1-4dd1-4ae5-b868-e98d27d5ad94-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"56e56ce1-4dd1-4ae5-b868-e98d27d5ad94\") " pod="openstack/ceilometer-0" Dec 13 07:12:04 crc kubenswrapper[4971]: I1213 07:12:04.344182 4971 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/ff35e760-13e7-4145-9970-742b73a36176-scripts\") pod \"horizon-bcdd7767-j7d7f\" (UID: \"ff35e760-13e7-4145-9970-742b73a36176\") " pod="openstack/horizon-bcdd7767-j7d7f" Dec 13 07:12:04 crc kubenswrapper[4971]: I1213 07:12:04.353790 4971 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/c1d819d0-339f-4e7e-9180-63d2f7c0042e-combined-ca-bundle\") pod \"cinder-db-sync-26k47\" (UID: \"c1d819d0-339f-4e7e-9180-63d2f7c0042e\") " pod="openstack/cinder-db-sync-26k47" Dec 13 07:12:04 crc kubenswrapper[4971]: I1213 07:12:04.359216 4971 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-hw5mk\" (UniqueName: \"kubernetes.io/projected/56e56ce1-4dd1-4ae5-b868-e98d27d5ad94-kube-api-access-hw5mk\") pod \"ceilometer-0\" (UID: \"56e56ce1-4dd1-4ae5-b868-e98d27d5ad94\") " pod="openstack/ceilometer-0" Dec 13 07:12:04 crc kubenswrapper[4971]: I1213 07:12:04.361268 4971 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"horizon-secret-key\" (UniqueName: \"kubernetes.io/secret/ff35e760-13e7-4145-9970-742b73a36176-horizon-secret-key\") pod \"horizon-bcdd7767-j7d7f\" (UID: \"ff35e760-13e7-4145-9970-742b73a36176\") " pod="openstack/horizon-bcdd7767-j7d7f" Dec 13 07:12:04 crc kubenswrapper[4971]: I1213 07:12:04.362119 4971 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-r4lhp\" (UniqueName: \"kubernetes.io/projected/c1d819d0-339f-4e7e-9180-63d2f7c0042e-kube-api-access-r4lhp\") pod \"cinder-db-sync-26k47\" (UID: \"c1d819d0-339f-4e7e-9180-63d2f7c0042e\") " pod="openstack/cinder-db-sync-26k47" Dec 13 07:12:04 crc kubenswrapper[4971]: I1213 07:12:04.365061 4971 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/c1d819d0-339f-4e7e-9180-63d2f7c0042e-db-sync-config-data\") pod \"cinder-db-sync-26k47\" (UID: \"c1d819d0-339f-4e7e-9180-63d2f7c0042e\") " pod="openstack/cinder-db-sync-26k47" Dec 13 07:12:04 crc kubenswrapper[4971]: I1213 07:12:04.365767 4971 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/56e56ce1-4dd1-4ae5-b868-e98d27d5ad94-scripts\") pod \"ceilometer-0\" (UID: \"56e56ce1-4dd1-4ae5-b868-e98d27d5ad94\") " pod="openstack/ceilometer-0" Dec 13 07:12:04 crc kubenswrapper[4971]: I1213 07:12:04.367119 4971 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/c1d819d0-339f-4e7e-9180-63d2f7c0042e-scripts\") pod \"cinder-db-sync-26k47\" (UID: \"c1d819d0-339f-4e7e-9180-63d2f7c0042e\") " pod="openstack/cinder-db-sync-26k47" Dec 13 07:12:04 crc kubenswrapper[4971]: I1213 07:12:04.368442 4971 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-cpxzk\" (UniqueName: \"kubernetes.io/projected/ff35e760-13e7-4145-9970-742b73a36176-kube-api-access-cpxzk\") pod \"horizon-bcdd7767-j7d7f\" (UID: \"ff35e760-13e7-4145-9970-742b73a36176\") " pod="openstack/horizon-bcdd7767-j7d7f" Dec 13 07:12:04 crc kubenswrapper[4971]: I1213 07:12:04.377347 4971 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/barbican-db-sync-dvchm"] Dec 13 07:12:04 crc kubenswrapper[4971]: I1213 07:12:04.378404 4971 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-db-sync-dvchm" Dec 13 07:12:04 crc kubenswrapper[4971]: I1213 07:12:04.380888 4971 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"barbican-barbican-dockercfg-6zl24" Dec 13 07:12:04 crc kubenswrapper[4971]: I1213 07:12:04.383601 4971 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/neutron-db-sync-8z2nm"] Dec 13 07:12:04 crc kubenswrapper[4971]: I1213 07:12:04.385046 4971 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"barbican-config-data" Dec 13 07:12:04 crc kubenswrapper[4971]: I1213 07:12:04.397576 4971 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/barbican-db-sync-dvchm"] Dec 13 07:12:04 crc kubenswrapper[4971]: I1213 07:12:04.409341 4971 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-bootstrap-xp9z4" Dec 13 07:12:04 crc kubenswrapper[4971]: I1213 07:12:04.412936 4971 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-w84mh\" (UniqueName: \"kubernetes.io/projected/d7cf1b55-0dae-4a17-857a-6336fe71ff24-kube-api-access-w84mh\") pod \"barbican-db-sync-dvchm\" (UID: \"d7cf1b55-0dae-4a17-857a-6336fe71ff24\") " pod="openstack/barbican-db-sync-dvchm" Dec 13 07:12:04 crc kubenswrapper[4971]: I1213 07:12:04.412965 4971 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/0813350d-430f-4564-a09a-062a3239ba2f-combined-ca-bundle\") pod \"neutron-db-sync-8z2nm\" (UID: \"0813350d-430f-4564-a09a-062a3239ba2f\") " pod="openstack/neutron-db-sync-8z2nm" Dec 13 07:12:04 crc kubenswrapper[4971]: I1213 07:12:04.412990 4971 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-ktmdv\" (UniqueName: \"kubernetes.io/projected/0813350d-430f-4564-a09a-062a3239ba2f-kube-api-access-ktmdv\") pod \"neutron-db-sync-8z2nm\" (UID: \"0813350d-430f-4564-a09a-062a3239ba2f\") " pod="openstack/neutron-db-sync-8z2nm" Dec 13 07:12:04 crc kubenswrapper[4971]: I1213 07:12:04.413011 4971 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/d7cf1b55-0dae-4a17-857a-6336fe71ff24-db-sync-config-data\") pod \"barbican-db-sync-dvchm\" (UID: \"d7cf1b55-0dae-4a17-857a-6336fe71ff24\") " pod="openstack/barbican-db-sync-dvchm" Dec 13 07:12:04 crc kubenswrapper[4971]: I1213 07:12:04.413046 4971 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d7cf1b55-0dae-4a17-857a-6336fe71ff24-combined-ca-bundle\") pod \"barbican-db-sync-dvchm\" (UID: \"d7cf1b55-0dae-4a17-857a-6336fe71ff24\") " pod="openstack/barbican-db-sync-dvchm" Dec 13 07:12:04 crc kubenswrapper[4971]: I1213 07:12:04.413089 4971 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/secret/0813350d-430f-4564-a09a-062a3239ba2f-config\") pod \"neutron-db-sync-8z2nm\" (UID: \"0813350d-430f-4564-a09a-062a3239ba2f\") " pod="openstack/neutron-db-sync-8z2nm" Dec 13 07:12:04 crc kubenswrapper[4971]: I1213 07:12:04.424377 4971 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/secret/0813350d-430f-4564-a09a-062a3239ba2f-config\") pod \"neutron-db-sync-8z2nm\" (UID: \"0813350d-430f-4564-a09a-062a3239ba2f\") " pod="openstack/neutron-db-sync-8z2nm" Dec 13 07:12:04 crc kubenswrapper[4971]: I1213 07:12:04.425280 4971 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-78fbc4bbf-6tztc" Dec 13 07:12:04 crc kubenswrapper[4971]: I1213 07:12:04.428394 4971 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/0813350d-430f-4564-a09a-062a3239ba2f-combined-ca-bundle\") pod \"neutron-db-sync-8z2nm\" (UID: \"0813350d-430f-4564-a09a-062a3239ba2f\") " pod="openstack/neutron-db-sync-8z2nm" Dec 13 07:12:04 crc kubenswrapper[4971]: I1213 07:12:04.451609 4971 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-78fbc4bbf-6tztc"] Dec 13 07:12:04 crc kubenswrapper[4971]: I1213 07:12:04.676383 4971 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-db-sync-26k47" Dec 13 07:12:04 crc kubenswrapper[4971]: I1213 07:12:04.677621 4971 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d7cf1b55-0dae-4a17-857a-6336fe71ff24-combined-ca-bundle\") pod \"barbican-db-sync-dvchm\" (UID: \"d7cf1b55-0dae-4a17-857a-6336fe71ff24\") " pod="openstack/barbican-db-sync-dvchm" Dec 13 07:12:04 crc kubenswrapper[4971]: I1213 07:12:04.677768 4971 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-w84mh\" (UniqueName: \"kubernetes.io/projected/d7cf1b55-0dae-4a17-857a-6336fe71ff24-kube-api-access-w84mh\") pod \"barbican-db-sync-dvchm\" (UID: \"d7cf1b55-0dae-4a17-857a-6336fe71ff24\") " pod="openstack/barbican-db-sync-dvchm" Dec 13 07:12:04 crc kubenswrapper[4971]: I1213 07:12:04.677813 4971 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/d7cf1b55-0dae-4a17-857a-6336fe71ff24-db-sync-config-data\") pod \"barbican-db-sync-dvchm\" (UID: \"d7cf1b55-0dae-4a17-857a-6336fe71ff24\") " pod="openstack/barbican-db-sync-dvchm" Dec 13 07:12:04 crc kubenswrapper[4971]: I1213 07:12:04.683243 4971 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-ktmdv\" (UniqueName: \"kubernetes.io/projected/0813350d-430f-4564-a09a-062a3239ba2f-kube-api-access-ktmdv\") pod \"neutron-db-sync-8z2nm\" (UID: \"0813350d-430f-4564-a09a-062a3239ba2f\") " pod="openstack/neutron-db-sync-8z2nm" Dec 13 07:12:04 crc kubenswrapper[4971]: I1213 07:12:04.685777 4971 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/d7cf1b55-0dae-4a17-857a-6336fe71ff24-db-sync-config-data\") pod \"barbican-db-sync-dvchm\" (UID: \"d7cf1b55-0dae-4a17-857a-6336fe71ff24\") " pod="openstack/barbican-db-sync-dvchm" Dec 13 07:12:04 crc kubenswrapper[4971]: I1213 07:12:04.687146 4971 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/placement-db-sync-954vn"] Dec 13 07:12:04 crc kubenswrapper[4971]: I1213 07:12:04.688322 4971 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/placement-db-sync-954vn" Dec 13 07:12:04 crc kubenswrapper[4971]: I1213 07:12:04.690328 4971 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d7cf1b55-0dae-4a17-857a-6336fe71ff24-combined-ca-bundle\") pod \"barbican-db-sync-dvchm\" (UID: \"d7cf1b55-0dae-4a17-857a-6336fe71ff24\") " pod="openstack/barbican-db-sync-dvchm" Dec 13 07:12:04 crc kubenswrapper[4971]: I1213 07:12:04.690559 4971 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"placement-placement-dockercfg-tt6pj" Dec 13 07:12:04 crc kubenswrapper[4971]: I1213 07:12:04.691054 4971 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"placement-scripts" Dec 13 07:12:04 crc kubenswrapper[4971]: I1213 07:12:04.691064 4971 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"placement-config-data" Dec 13 07:12:04 crc kubenswrapper[4971]: I1213 07:12:04.696363 4971 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/placement-db-sync-954vn"] Dec 13 07:12:04 crc kubenswrapper[4971]: I1213 07:12:04.711448 4971 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-w84mh\" (UniqueName: \"kubernetes.io/projected/d7cf1b55-0dae-4a17-857a-6336fe71ff24-kube-api-access-w84mh\") pod \"barbican-db-sync-dvchm\" (UID: \"d7cf1b55-0dae-4a17-857a-6336fe71ff24\") " pod="openstack/barbican-db-sync-dvchm" Dec 13 07:12:04 crc kubenswrapper[4971]: I1213 07:12:04.717565 4971 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/horizon-bcdd7767-j7d7f" Dec 13 07:12:04 crc kubenswrapper[4971]: I1213 07:12:04.736858 4971 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Dec 13 07:12:04 crc kubenswrapper[4971]: I1213 07:12:04.741335 4971 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-5d87b7c6dc-sxklp"] Dec 13 07:12:04 crc kubenswrapper[4971]: I1213 07:12:04.742585 4971 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-5d87b7c6dc-sxklp" Dec 13 07:12:04 crc kubenswrapper[4971]: I1213 07:12:04.767681 4971 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-5d87b7c6dc-sxklp"] Dec 13 07:12:04 crc kubenswrapper[4971]: I1213 07:12:04.785259 4971 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-db-sync-8z2nm" Dec 13 07:12:04 crc kubenswrapper[4971]: I1213 07:12:04.800263 4971 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/horizon-77c4585f77-kkx2c"] Dec 13 07:12:04 crc kubenswrapper[4971]: I1213 07:12:04.801924 4971 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/horizon-77c4585f77-kkx2c" Dec 13 07:12:04 crc kubenswrapper[4971]: I1213 07:12:04.813721 4971 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/horizon-77c4585f77-kkx2c"] Dec 13 07:12:04 crc kubenswrapper[4971]: I1213 07:12:04.822167 4971 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-db-sync-dvchm" Dec 13 07:12:04 crc kubenswrapper[4971]: I1213 07:12:04.884423 4971 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/166612b7-15b4-4a04-b7bb-62603363de80-config-data\") pod \"placement-db-sync-954vn\" (UID: \"166612b7-15b4-4a04-b7bb-62603363de80\") " pod="openstack/placement-db-sync-954vn" Dec 13 07:12:04 crc kubenswrapper[4971]: I1213 07:12:04.884497 4971 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/2430f20e-0768-47e9-89de-fc6ac2615707-ovsdbserver-nb\") pod \"dnsmasq-dns-5d87b7c6dc-sxklp\" (UID: \"2430f20e-0768-47e9-89de-fc6ac2615707\") " pod="openstack/dnsmasq-dns-5d87b7c6dc-sxklp" Dec 13 07:12:04 crc kubenswrapper[4971]: I1213 07:12:04.884585 4971 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/2430f20e-0768-47e9-89de-fc6ac2615707-dns-svc\") pod \"dnsmasq-dns-5d87b7c6dc-sxklp\" (UID: \"2430f20e-0768-47e9-89de-fc6ac2615707\") " pod="openstack/dnsmasq-dns-5d87b7c6dc-sxklp" Dec 13 07:12:04 crc kubenswrapper[4971]: I1213 07:12:04.884633 4971 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/2430f20e-0768-47e9-89de-fc6ac2615707-config\") pod \"dnsmasq-dns-5d87b7c6dc-sxklp\" (UID: \"2430f20e-0768-47e9-89de-fc6ac2615707\") " pod="openstack/dnsmasq-dns-5d87b7c6dc-sxklp" Dec 13 07:12:04 crc kubenswrapper[4971]: I1213 07:12:04.884649 4971 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-97xnq\" (UniqueName: \"kubernetes.io/projected/2430f20e-0768-47e9-89de-fc6ac2615707-kube-api-access-97xnq\") pod \"dnsmasq-dns-5d87b7c6dc-sxklp\" (UID: \"2430f20e-0768-47e9-89de-fc6ac2615707\") " pod="openstack/dnsmasq-dns-5d87b7c6dc-sxklp" Dec 13 07:12:04 crc kubenswrapper[4971]: I1213 07:12:04.884681 4971 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/166612b7-15b4-4a04-b7bb-62603363de80-combined-ca-bundle\") pod \"placement-db-sync-954vn\" (UID: \"166612b7-15b4-4a04-b7bb-62603363de80\") " pod="openstack/placement-db-sync-954vn" Dec 13 07:12:04 crc kubenswrapper[4971]: I1213 07:12:04.884751 4971 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/166612b7-15b4-4a04-b7bb-62603363de80-scripts\") pod \"placement-db-sync-954vn\" (UID: \"166612b7-15b4-4a04-b7bb-62603363de80\") " pod="openstack/placement-db-sync-954vn" Dec 13 07:12:04 crc kubenswrapper[4971]: I1213 07:12:04.884771 4971 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-5lxdd\" (UniqueName: \"kubernetes.io/projected/166612b7-15b4-4a04-b7bb-62603363de80-kube-api-access-5lxdd\") pod \"placement-db-sync-954vn\" (UID: \"166612b7-15b4-4a04-b7bb-62603363de80\") " pod="openstack/placement-db-sync-954vn" Dec 13 07:12:04 crc kubenswrapper[4971]: I1213 07:12:04.884823 4971 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/2430f20e-0768-47e9-89de-fc6ac2615707-ovsdbserver-sb\") pod \"dnsmasq-dns-5d87b7c6dc-sxklp\" (UID: \"2430f20e-0768-47e9-89de-fc6ac2615707\") " pod="openstack/dnsmasq-dns-5d87b7c6dc-sxklp" Dec 13 07:12:04 crc kubenswrapper[4971]: I1213 07:12:04.884843 4971 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/166612b7-15b4-4a04-b7bb-62603363de80-logs\") pod \"placement-db-sync-954vn\" (UID: \"166612b7-15b4-4a04-b7bb-62603363de80\") " pod="openstack/placement-db-sync-954vn" Dec 13 07:12:05 crc kubenswrapper[4971]: I1213 07:12:05.189964 4971 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/2430f20e-0768-47e9-89de-fc6ac2615707-ovsdbserver-nb\") pod \"dnsmasq-dns-5d87b7c6dc-sxklp\" (UID: \"2430f20e-0768-47e9-89de-fc6ac2615707\") " pod="openstack/dnsmasq-dns-5d87b7c6dc-sxklp" Dec 13 07:12:05 crc kubenswrapper[4971]: I1213 07:12:05.190381 4971 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/2430f20e-0768-47e9-89de-fc6ac2615707-dns-svc\") pod \"dnsmasq-dns-5d87b7c6dc-sxklp\" (UID: \"2430f20e-0768-47e9-89de-fc6ac2615707\") " pod="openstack/dnsmasq-dns-5d87b7c6dc-sxklp" Dec 13 07:12:05 crc kubenswrapper[4971]: I1213 07:12:05.190464 4971 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/2430f20e-0768-47e9-89de-fc6ac2615707-config\") pod \"dnsmasq-dns-5d87b7c6dc-sxklp\" (UID: \"2430f20e-0768-47e9-89de-fc6ac2615707\") " pod="openstack/dnsmasq-dns-5d87b7c6dc-sxklp" Dec 13 07:12:05 crc kubenswrapper[4971]: I1213 07:12:05.190548 4971 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/b314f107-1cb6-49f3-aa51-3f607ca79f6b-scripts\") pod \"horizon-77c4585f77-kkx2c\" (UID: \"b314f107-1cb6-49f3-aa51-3f607ca79f6b\") " pod="openstack/horizon-77c4585f77-kkx2c" Dec 13 07:12:05 crc kubenswrapper[4971]: I1213 07:12:05.190624 4971 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-97xnq\" (UniqueName: \"kubernetes.io/projected/2430f20e-0768-47e9-89de-fc6ac2615707-kube-api-access-97xnq\") pod \"dnsmasq-dns-5d87b7c6dc-sxklp\" (UID: \"2430f20e-0768-47e9-89de-fc6ac2615707\") " pod="openstack/dnsmasq-dns-5d87b7c6dc-sxklp" Dec 13 07:12:05 crc kubenswrapper[4971]: I1213 07:12:05.190710 4971 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/166612b7-15b4-4a04-b7bb-62603363de80-combined-ca-bundle\") pod \"placement-db-sync-954vn\" (UID: \"166612b7-15b4-4a04-b7bb-62603363de80\") " pod="openstack/placement-db-sync-954vn" Dec 13 07:12:05 crc kubenswrapper[4971]: I1213 07:12:05.190831 4971 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/b314f107-1cb6-49f3-aa51-3f607ca79f6b-logs\") pod \"horizon-77c4585f77-kkx2c\" (UID: \"b314f107-1cb6-49f3-aa51-3f607ca79f6b\") " pod="openstack/horizon-77c4585f77-kkx2c" Dec 13 07:12:05 crc kubenswrapper[4971]: I1213 07:12:05.190912 4971 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/166612b7-15b4-4a04-b7bb-62603363de80-scripts\") pod \"placement-db-sync-954vn\" (UID: \"166612b7-15b4-4a04-b7bb-62603363de80\") " pod="openstack/placement-db-sync-954vn" Dec 13 07:12:05 crc kubenswrapper[4971]: I1213 07:12:05.190995 4971 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-5lxdd\" (UniqueName: \"kubernetes.io/projected/166612b7-15b4-4a04-b7bb-62603363de80-kube-api-access-5lxdd\") pod \"placement-db-sync-954vn\" (UID: \"166612b7-15b4-4a04-b7bb-62603363de80\") " pod="openstack/placement-db-sync-954vn" Dec 13 07:12:05 crc kubenswrapper[4971]: I1213 07:12:05.191059 4971 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/b314f107-1cb6-49f3-aa51-3f607ca79f6b-config-data\") pod \"horizon-77c4585f77-kkx2c\" (UID: \"b314f107-1cb6-49f3-aa51-3f607ca79f6b\") " pod="openstack/horizon-77c4585f77-kkx2c" Dec 13 07:12:05 crc kubenswrapper[4971]: I1213 07:12:05.191225 4971 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/2430f20e-0768-47e9-89de-fc6ac2615707-ovsdbserver-sb\") pod \"dnsmasq-dns-5d87b7c6dc-sxklp\" (UID: \"2430f20e-0768-47e9-89de-fc6ac2615707\") " pod="openstack/dnsmasq-dns-5d87b7c6dc-sxklp" Dec 13 07:12:05 crc kubenswrapper[4971]: I1213 07:12:05.191304 4971 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/166612b7-15b4-4a04-b7bb-62603363de80-logs\") pod \"placement-db-sync-954vn\" (UID: \"166612b7-15b4-4a04-b7bb-62603363de80\") " pod="openstack/placement-db-sync-954vn" Dec 13 07:12:05 crc kubenswrapper[4971]: I1213 07:12:05.191383 4971 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"horizon-secret-key\" (UniqueName: \"kubernetes.io/secret/b314f107-1cb6-49f3-aa51-3f607ca79f6b-horizon-secret-key\") pod \"horizon-77c4585f77-kkx2c\" (UID: \"b314f107-1cb6-49f3-aa51-3f607ca79f6b\") " pod="openstack/horizon-77c4585f77-kkx2c" Dec 13 07:12:05 crc kubenswrapper[4971]: I1213 07:12:05.191449 4971 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/166612b7-15b4-4a04-b7bb-62603363de80-config-data\") pod \"placement-db-sync-954vn\" (UID: \"166612b7-15b4-4a04-b7bb-62603363de80\") " pod="openstack/placement-db-sync-954vn" Dec 13 07:12:05 crc kubenswrapper[4971]: I1213 07:12:05.191536 4971 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-tghr9\" (UniqueName: \"kubernetes.io/projected/b314f107-1cb6-49f3-aa51-3f607ca79f6b-kube-api-access-tghr9\") pod \"horizon-77c4585f77-kkx2c\" (UID: \"b314f107-1cb6-49f3-aa51-3f607ca79f6b\") " pod="openstack/horizon-77c4585f77-kkx2c" Dec 13 07:12:05 crc kubenswrapper[4971]: I1213 07:12:05.192539 4971 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/2430f20e-0768-47e9-89de-fc6ac2615707-ovsdbserver-nb\") pod \"dnsmasq-dns-5d87b7c6dc-sxklp\" (UID: \"2430f20e-0768-47e9-89de-fc6ac2615707\") " pod="openstack/dnsmasq-dns-5d87b7c6dc-sxklp" Dec 13 07:12:05 crc kubenswrapper[4971]: I1213 07:12:05.213548 4971 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/166612b7-15b4-4a04-b7bb-62603363de80-logs\") pod \"placement-db-sync-954vn\" (UID: \"166612b7-15b4-4a04-b7bb-62603363de80\") " pod="openstack/placement-db-sync-954vn" Dec 13 07:12:05 crc kubenswrapper[4971]: I1213 07:12:05.213898 4971 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/2430f20e-0768-47e9-89de-fc6ac2615707-ovsdbserver-sb\") pod \"dnsmasq-dns-5d87b7c6dc-sxklp\" (UID: \"2430f20e-0768-47e9-89de-fc6ac2615707\") " pod="openstack/dnsmasq-dns-5d87b7c6dc-sxklp" Dec 13 07:12:05 crc kubenswrapper[4971]: I1213 07:12:05.228037 4971 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/2430f20e-0768-47e9-89de-fc6ac2615707-config\") pod \"dnsmasq-dns-5d87b7c6dc-sxklp\" (UID: \"2430f20e-0768-47e9-89de-fc6ac2615707\") " pod="openstack/dnsmasq-dns-5d87b7c6dc-sxklp" Dec 13 07:12:05 crc kubenswrapper[4971]: I1213 07:12:05.229085 4971 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/2430f20e-0768-47e9-89de-fc6ac2615707-dns-svc\") pod \"dnsmasq-dns-5d87b7c6dc-sxklp\" (UID: \"2430f20e-0768-47e9-89de-fc6ac2615707\") " pod="openstack/dnsmasq-dns-5d87b7c6dc-sxklp" Dec 13 07:12:05 crc kubenswrapper[4971]: I1213 07:12:05.237585 4971 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/166612b7-15b4-4a04-b7bb-62603363de80-config-data\") pod \"placement-db-sync-954vn\" (UID: \"166612b7-15b4-4a04-b7bb-62603363de80\") " pod="openstack/placement-db-sync-954vn" Dec 13 07:12:05 crc kubenswrapper[4971]: I1213 07:12:05.242112 4971 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/166612b7-15b4-4a04-b7bb-62603363de80-combined-ca-bundle\") pod \"placement-db-sync-954vn\" (UID: \"166612b7-15b4-4a04-b7bb-62603363de80\") " pod="openstack/placement-db-sync-954vn" Dec 13 07:12:05 crc kubenswrapper[4971]: I1213 07:12:05.242316 4971 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/166612b7-15b4-4a04-b7bb-62603363de80-scripts\") pod \"placement-db-sync-954vn\" (UID: \"166612b7-15b4-4a04-b7bb-62603363de80\") " pod="openstack/placement-db-sync-954vn" Dec 13 07:12:05 crc kubenswrapper[4971]: I1213 07:12:05.260182 4971 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-97xnq\" (UniqueName: \"kubernetes.io/projected/2430f20e-0768-47e9-89de-fc6ac2615707-kube-api-access-97xnq\") pod \"dnsmasq-dns-5d87b7c6dc-sxklp\" (UID: \"2430f20e-0768-47e9-89de-fc6ac2615707\") " pod="openstack/dnsmasq-dns-5d87b7c6dc-sxklp" Dec 13 07:12:05 crc kubenswrapper[4971]: I1213 07:12:05.268132 4971 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-5lxdd\" (UniqueName: \"kubernetes.io/projected/166612b7-15b4-4a04-b7bb-62603363de80-kube-api-access-5lxdd\") pod \"placement-db-sync-954vn\" (UID: \"166612b7-15b4-4a04-b7bb-62603363de80\") " pod="openstack/placement-db-sync-954vn" Dec 13 07:12:05 crc kubenswrapper[4971]: I1213 07:12:05.293858 4971 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/b314f107-1cb6-49f3-aa51-3f607ca79f6b-logs\") pod \"horizon-77c4585f77-kkx2c\" (UID: \"b314f107-1cb6-49f3-aa51-3f607ca79f6b\") " pod="openstack/horizon-77c4585f77-kkx2c" Dec 13 07:12:05 crc kubenswrapper[4971]: I1213 07:12:05.293903 4971 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/b314f107-1cb6-49f3-aa51-3f607ca79f6b-config-data\") pod \"horizon-77c4585f77-kkx2c\" (UID: \"b314f107-1cb6-49f3-aa51-3f607ca79f6b\") " pod="openstack/horizon-77c4585f77-kkx2c" Dec 13 07:12:05 crc kubenswrapper[4971]: I1213 07:12:05.293959 4971 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"horizon-secret-key\" (UniqueName: \"kubernetes.io/secret/b314f107-1cb6-49f3-aa51-3f607ca79f6b-horizon-secret-key\") pod \"horizon-77c4585f77-kkx2c\" (UID: \"b314f107-1cb6-49f3-aa51-3f607ca79f6b\") " pod="openstack/horizon-77c4585f77-kkx2c" Dec 13 07:12:05 crc kubenswrapper[4971]: I1213 07:12:05.293980 4971 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-tghr9\" (UniqueName: \"kubernetes.io/projected/b314f107-1cb6-49f3-aa51-3f607ca79f6b-kube-api-access-tghr9\") pod \"horizon-77c4585f77-kkx2c\" (UID: \"b314f107-1cb6-49f3-aa51-3f607ca79f6b\") " pod="openstack/horizon-77c4585f77-kkx2c" Dec 13 07:12:05 crc kubenswrapper[4971]: I1213 07:12:05.294018 4971 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/b314f107-1cb6-49f3-aa51-3f607ca79f6b-scripts\") pod \"horizon-77c4585f77-kkx2c\" (UID: \"b314f107-1cb6-49f3-aa51-3f607ca79f6b\") " pod="openstack/horizon-77c4585f77-kkx2c" Dec 13 07:12:05 crc kubenswrapper[4971]: I1213 07:12:05.294867 4971 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/b314f107-1cb6-49f3-aa51-3f607ca79f6b-scripts\") pod \"horizon-77c4585f77-kkx2c\" (UID: \"b314f107-1cb6-49f3-aa51-3f607ca79f6b\") " pod="openstack/horizon-77c4585f77-kkx2c" Dec 13 07:12:05 crc kubenswrapper[4971]: I1213 07:12:05.295075 4971 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/b314f107-1cb6-49f3-aa51-3f607ca79f6b-logs\") pod \"horizon-77c4585f77-kkx2c\" (UID: \"b314f107-1cb6-49f3-aa51-3f607ca79f6b\") " pod="openstack/horizon-77c4585f77-kkx2c" Dec 13 07:12:05 crc kubenswrapper[4971]: I1213 07:12:05.316975 4971 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"horizon-secret-key\" (UniqueName: \"kubernetes.io/secret/b314f107-1cb6-49f3-aa51-3f607ca79f6b-horizon-secret-key\") pod \"horizon-77c4585f77-kkx2c\" (UID: \"b314f107-1cb6-49f3-aa51-3f607ca79f6b\") " pod="openstack/horizon-77c4585f77-kkx2c" Dec 13 07:12:05 crc kubenswrapper[4971]: I1213 07:12:05.322944 4971 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/b314f107-1cb6-49f3-aa51-3f607ca79f6b-config-data\") pod \"horizon-77c4585f77-kkx2c\" (UID: \"b314f107-1cb6-49f3-aa51-3f607ca79f6b\") " pod="openstack/horizon-77c4585f77-kkx2c" Dec 13 07:12:05 crc kubenswrapper[4971]: I1213 07:12:05.346100 4971 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-tghr9\" (UniqueName: \"kubernetes.io/projected/b314f107-1cb6-49f3-aa51-3f607ca79f6b-kube-api-access-tghr9\") pod \"horizon-77c4585f77-kkx2c\" (UID: \"b314f107-1cb6-49f3-aa51-3f607ca79f6b\") " pod="openstack/horizon-77c4585f77-kkx2c" Dec 13 07:12:05 crc kubenswrapper[4971]: I1213 07:12:05.442364 4971 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/placement-db-sync-954vn" Dec 13 07:12:05 crc kubenswrapper[4971]: I1213 07:12:05.676866 4971 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/horizon-77c4585f77-kkx2c" Dec 13 07:12:05 crc kubenswrapper[4971]: I1213 07:12:05.677234 4971 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-5d87b7c6dc-sxklp" Dec 13 07:12:06 crc kubenswrapper[4971]: I1213 07:12:06.104340 4971 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/keystone-bootstrap-xp9z4"] Dec 13 07:12:06 crc kubenswrapper[4971]: I1213 07:12:06.222405 4971 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cinder-db-sync-26k47"] Dec 13 07:12:06 crc kubenswrapper[4971]: I1213 07:12:06.240841 4971 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-78fbc4bbf-6tztc"] Dec 13 07:12:06 crc kubenswrapper[4971]: I1213 07:12:06.355759 4971 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-bootstrap-xp9z4" event={"ID":"5b9e78cc-aa6e-4369-ab8a-5e89b86359c1","Type":"ContainerStarted","Data":"4191d30758fbfe0bd41086142577cea859c9a2795fba391926a355d657e0dc0d"} Dec 13 07:12:06 crc kubenswrapper[4971]: I1213 07:12:06.359613 4971 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-db-sync-26k47" event={"ID":"c1d819d0-339f-4e7e-9180-63d2f7c0042e","Type":"ContainerStarted","Data":"4a0005fe1401e2d15dee97c0cd5b2ea49f753e03b159110ac32c15aa76fd93e2"} Dec 13 07:12:06 crc kubenswrapper[4971]: I1213 07:12:06.361878 4971 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-78fbc4bbf-6tztc" event={"ID":"1b1d2613-64f2-4b48-aaf5-0681ff6cba2d","Type":"ContainerStarted","Data":"298778dec3373fc52010fcb45dcfa064cf54fd8eb6f0ed7e782ff8e010f8d850"} Dec 13 07:12:06 crc kubenswrapper[4971]: I1213 07:12:06.421047 4971 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/horizon-bcdd7767-j7d7f"] Dec 13 07:12:06 crc kubenswrapper[4971]: W1213 07:12:06.435700 4971 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podff35e760_13e7_4145_9970_742b73a36176.slice/crio-46b26a551dc80a0e0a26e6ba29b9a6a2b3c363d24f8e1ca63552331408259a59 WatchSource:0}: Error finding container 46b26a551dc80a0e0a26e6ba29b9a6a2b3c363d24f8e1ca63552331408259a59: Status 404 returned error can't find the container with id 46b26a551dc80a0e0a26e6ba29b9a6a2b3c363d24f8e1ca63552331408259a59 Dec 13 07:12:06 crc kubenswrapper[4971]: I1213 07:12:06.442667 4971 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ceilometer-0"] Dec 13 07:12:06 crc kubenswrapper[4971]: I1213 07:12:06.449049 4971 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/barbican-db-sync-dvchm"] Dec 13 07:12:06 crc kubenswrapper[4971]: W1213 07:12:06.457290 4971 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod56e56ce1_4dd1_4ae5_b868_e98d27d5ad94.slice/crio-8e88d8b440b648a680026f56322ec6eb8ebd3ef0483daa3a94e143feacc7200b WatchSource:0}: Error finding container 8e88d8b440b648a680026f56322ec6eb8ebd3ef0483daa3a94e143feacc7200b: Status 404 returned error can't find the container with id 8e88d8b440b648a680026f56322ec6eb8ebd3ef0483daa3a94e143feacc7200b Dec 13 07:12:06 crc kubenswrapper[4971]: W1213 07:12:06.464153 4971 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podd7cf1b55_0dae_4a17_857a_6336fe71ff24.slice/crio-4d91578739ace0c00325d2663283e687dd7acfed3e767ee9d529e92f7ceaf861 WatchSource:0}: Error finding container 4d91578739ace0c00325d2663283e687dd7acfed3e767ee9d529e92f7ceaf861: Status 404 returned error can't find the container with id 4d91578739ace0c00325d2663283e687dd7acfed3e767ee9d529e92f7ceaf861 Dec 13 07:12:06 crc kubenswrapper[4971]: I1213 07:12:06.562768 4971 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/placement-db-sync-954vn"] Dec 13 07:12:06 crc kubenswrapper[4971]: W1213 07:12:06.571209 4971 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod166612b7_15b4_4a04_b7bb_62603363de80.slice/crio-c55f47b9cb113dd1b8560c9119be9a982a21f03ef350e168081b073d93064aaf WatchSource:0}: Error finding container c55f47b9cb113dd1b8560c9119be9a982a21f03ef350e168081b073d93064aaf: Status 404 returned error can't find the container with id c55f47b9cb113dd1b8560c9119be9a982a21f03ef350e168081b073d93064aaf Dec 13 07:12:06 crc kubenswrapper[4971]: I1213 07:12:06.605067 4971 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/neutron-db-sync-8z2nm"] Dec 13 07:12:06 crc kubenswrapper[4971]: W1213 07:12:06.605066 4971 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podb314f107_1cb6_49f3_aa51_3f607ca79f6b.slice/crio-8daaeddcce34e5e2f6d2ef32ac8d2a6c22737e4346ddf888ec4a8f83769d6f9d WatchSource:0}: Error finding container 8daaeddcce34e5e2f6d2ef32ac8d2a6c22737e4346ddf888ec4a8f83769d6f9d: Status 404 returned error can't find the container with id 8daaeddcce34e5e2f6d2ef32ac8d2a6c22737e4346ddf888ec4a8f83769d6f9d Dec 13 07:12:06 crc kubenswrapper[4971]: W1213 07:12:06.609764 4971 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod0813350d_430f_4564_a09a_062a3239ba2f.slice/crio-f8b36cdbe233f6b61c0a47a1a93c6169e7f2fae7dbfe2b7a99ce9ae4fd061d40 WatchSource:0}: Error finding container f8b36cdbe233f6b61c0a47a1a93c6169e7f2fae7dbfe2b7a99ce9ae4fd061d40: Status 404 returned error can't find the container with id f8b36cdbe233f6b61c0a47a1a93c6169e7f2fae7dbfe2b7a99ce9ae4fd061d40 Dec 13 07:12:06 crc kubenswrapper[4971]: W1213 07:12:06.612292 4971 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod2430f20e_0768_47e9_89de_fc6ac2615707.slice/crio-e87b3742a8d0d0c1e72b0732bb0b473a60d040294def9f4ace6b98cb9c23f2cf WatchSource:0}: Error finding container e87b3742a8d0d0c1e72b0732bb0b473a60d040294def9f4ace6b98cb9c23f2cf: Status 404 returned error can't find the container with id e87b3742a8d0d0c1e72b0732bb0b473a60d040294def9f4ace6b98cb9c23f2cf Dec 13 07:12:06 crc kubenswrapper[4971]: I1213 07:12:06.618029 4971 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/horizon-77c4585f77-kkx2c"] Dec 13 07:12:06 crc kubenswrapper[4971]: I1213 07:12:06.626662 4971 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-5d87b7c6dc-sxklp"] Dec 13 07:12:07 crc kubenswrapper[4971]: I1213 07:12:07.033581 4971 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/horizon-77c4585f77-kkx2c"] Dec 13 07:12:07 crc kubenswrapper[4971]: I1213 07:12:07.046728 4971 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ceilometer-0"] Dec 13 07:12:07 crc kubenswrapper[4971]: I1213 07:12:07.082034 4971 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/horizon-fbfd486df-xmvls"] Dec 13 07:12:07 crc kubenswrapper[4971]: I1213 07:12:07.094173 4971 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/horizon-fbfd486df-xmvls" Dec 13 07:12:07 crc kubenswrapper[4971]: I1213 07:12:07.108590 4971 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/horizon-fbfd486df-xmvls"] Dec 13 07:12:07 crc kubenswrapper[4971]: I1213 07:12:07.206915 4971 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/f9bd96be-6d7f-415e-b67c-5bd02b05c536-scripts\") pod \"horizon-fbfd486df-xmvls\" (UID: \"f9bd96be-6d7f-415e-b67c-5bd02b05c536\") " pod="openstack/horizon-fbfd486df-xmvls" Dec 13 07:12:07 crc kubenswrapper[4971]: I1213 07:12:07.206974 4971 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-kjwdg\" (UniqueName: \"kubernetes.io/projected/f9bd96be-6d7f-415e-b67c-5bd02b05c536-kube-api-access-kjwdg\") pod \"horizon-fbfd486df-xmvls\" (UID: \"f9bd96be-6d7f-415e-b67c-5bd02b05c536\") " pod="openstack/horizon-fbfd486df-xmvls" Dec 13 07:12:07 crc kubenswrapper[4971]: I1213 07:12:07.207161 4971 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/f9bd96be-6d7f-415e-b67c-5bd02b05c536-config-data\") pod \"horizon-fbfd486df-xmvls\" (UID: \"f9bd96be-6d7f-415e-b67c-5bd02b05c536\") " pod="openstack/horizon-fbfd486df-xmvls" Dec 13 07:12:07 crc kubenswrapper[4971]: I1213 07:12:07.207217 4971 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/f9bd96be-6d7f-415e-b67c-5bd02b05c536-logs\") pod \"horizon-fbfd486df-xmvls\" (UID: \"f9bd96be-6d7f-415e-b67c-5bd02b05c536\") " pod="openstack/horizon-fbfd486df-xmvls" Dec 13 07:12:07 crc kubenswrapper[4971]: I1213 07:12:07.207255 4971 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"horizon-secret-key\" (UniqueName: \"kubernetes.io/secret/f9bd96be-6d7f-415e-b67c-5bd02b05c536-horizon-secret-key\") pod \"horizon-fbfd486df-xmvls\" (UID: \"f9bd96be-6d7f-415e-b67c-5bd02b05c536\") " pod="openstack/horizon-fbfd486df-xmvls" Dec 13 07:12:07 crc kubenswrapper[4971]: I1213 07:12:07.309230 4971 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/f9bd96be-6d7f-415e-b67c-5bd02b05c536-config-data\") pod \"horizon-fbfd486df-xmvls\" (UID: \"f9bd96be-6d7f-415e-b67c-5bd02b05c536\") " pod="openstack/horizon-fbfd486df-xmvls" Dec 13 07:12:07 crc kubenswrapper[4971]: I1213 07:12:07.309792 4971 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/f9bd96be-6d7f-415e-b67c-5bd02b05c536-logs\") pod \"horizon-fbfd486df-xmvls\" (UID: \"f9bd96be-6d7f-415e-b67c-5bd02b05c536\") " pod="openstack/horizon-fbfd486df-xmvls" Dec 13 07:12:07 crc kubenswrapper[4971]: I1213 07:12:07.309816 4971 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"horizon-secret-key\" (UniqueName: \"kubernetes.io/secret/f9bd96be-6d7f-415e-b67c-5bd02b05c536-horizon-secret-key\") pod \"horizon-fbfd486df-xmvls\" (UID: \"f9bd96be-6d7f-415e-b67c-5bd02b05c536\") " pod="openstack/horizon-fbfd486df-xmvls" Dec 13 07:12:07 crc kubenswrapper[4971]: I1213 07:12:07.309895 4971 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/f9bd96be-6d7f-415e-b67c-5bd02b05c536-scripts\") pod \"horizon-fbfd486df-xmvls\" (UID: \"f9bd96be-6d7f-415e-b67c-5bd02b05c536\") " pod="openstack/horizon-fbfd486df-xmvls" Dec 13 07:12:07 crc kubenswrapper[4971]: I1213 07:12:07.309925 4971 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-kjwdg\" (UniqueName: \"kubernetes.io/projected/f9bd96be-6d7f-415e-b67c-5bd02b05c536-kube-api-access-kjwdg\") pod \"horizon-fbfd486df-xmvls\" (UID: \"f9bd96be-6d7f-415e-b67c-5bd02b05c536\") " pod="openstack/horizon-fbfd486df-xmvls" Dec 13 07:12:07 crc kubenswrapper[4971]: I1213 07:12:07.311797 4971 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/f9bd96be-6d7f-415e-b67c-5bd02b05c536-logs\") pod \"horizon-fbfd486df-xmvls\" (UID: \"f9bd96be-6d7f-415e-b67c-5bd02b05c536\") " pod="openstack/horizon-fbfd486df-xmvls" Dec 13 07:12:07 crc kubenswrapper[4971]: I1213 07:12:07.312182 4971 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/f9bd96be-6d7f-415e-b67c-5bd02b05c536-scripts\") pod \"horizon-fbfd486df-xmvls\" (UID: \"f9bd96be-6d7f-415e-b67c-5bd02b05c536\") " pod="openstack/horizon-fbfd486df-xmvls" Dec 13 07:12:07 crc kubenswrapper[4971]: I1213 07:12:07.313584 4971 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/f9bd96be-6d7f-415e-b67c-5bd02b05c536-config-data\") pod \"horizon-fbfd486df-xmvls\" (UID: \"f9bd96be-6d7f-415e-b67c-5bd02b05c536\") " pod="openstack/horizon-fbfd486df-xmvls" Dec 13 07:12:07 crc kubenswrapper[4971]: I1213 07:12:07.319956 4971 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"horizon-secret-key\" (UniqueName: \"kubernetes.io/secret/f9bd96be-6d7f-415e-b67c-5bd02b05c536-horizon-secret-key\") pod \"horizon-fbfd486df-xmvls\" (UID: \"f9bd96be-6d7f-415e-b67c-5bd02b05c536\") " pod="openstack/horizon-fbfd486df-xmvls" Dec 13 07:12:07 crc kubenswrapper[4971]: I1213 07:12:07.333092 4971 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-kjwdg\" (UniqueName: \"kubernetes.io/projected/f9bd96be-6d7f-415e-b67c-5bd02b05c536-kube-api-access-kjwdg\") pod \"horizon-fbfd486df-xmvls\" (UID: \"f9bd96be-6d7f-415e-b67c-5bd02b05c536\") " pod="openstack/horizon-fbfd486df-xmvls" Dec 13 07:12:07 crc kubenswrapper[4971]: I1213 07:12:07.382938 4971 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/placement-db-sync-954vn" event={"ID":"166612b7-15b4-4a04-b7bb-62603363de80","Type":"ContainerStarted","Data":"c55f47b9cb113dd1b8560c9119be9a982a21f03ef350e168081b073d93064aaf"} Dec 13 07:12:07 crc kubenswrapper[4971]: I1213 07:12:07.385475 4971 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/horizon-77c4585f77-kkx2c" event={"ID":"b314f107-1cb6-49f3-aa51-3f607ca79f6b","Type":"ContainerStarted","Data":"8daaeddcce34e5e2f6d2ef32ac8d2a6c22737e4346ddf888ec4a8f83769d6f9d"} Dec 13 07:12:07 crc kubenswrapper[4971]: I1213 07:12:07.388076 4971 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-db-sync-dvchm" event={"ID":"d7cf1b55-0dae-4a17-857a-6336fe71ff24","Type":"ContainerStarted","Data":"4d91578739ace0c00325d2663283e687dd7acfed3e767ee9d529e92f7ceaf861"} Dec 13 07:12:07 crc kubenswrapper[4971]: I1213 07:12:07.389945 4971 generic.go:334] "Generic (PLEG): container finished" podID="1b1d2613-64f2-4b48-aaf5-0681ff6cba2d" containerID="ba8b92c4feec80f2fa98a863f27ba987f3e38266fe6f9a3aca0fd5f97e01a75e" exitCode=0 Dec 13 07:12:07 crc kubenswrapper[4971]: I1213 07:12:07.390028 4971 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-78fbc4bbf-6tztc" event={"ID":"1b1d2613-64f2-4b48-aaf5-0681ff6cba2d","Type":"ContainerDied","Data":"ba8b92c4feec80f2fa98a863f27ba987f3e38266fe6f9a3aca0fd5f97e01a75e"} Dec 13 07:12:07 crc kubenswrapper[4971]: I1213 07:12:07.408433 4971 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"56e56ce1-4dd1-4ae5-b868-e98d27d5ad94","Type":"ContainerStarted","Data":"8e88d8b440b648a680026f56322ec6eb8ebd3ef0483daa3a94e143feacc7200b"} Dec 13 07:12:07 crc kubenswrapper[4971]: I1213 07:12:07.412567 4971 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/horizon-bcdd7767-j7d7f" event={"ID":"ff35e760-13e7-4145-9970-742b73a36176","Type":"ContainerStarted","Data":"46b26a551dc80a0e0a26e6ba29b9a6a2b3c363d24f8e1ca63552331408259a59"} Dec 13 07:12:07 crc kubenswrapper[4971]: I1213 07:12:07.415208 4971 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-bootstrap-xp9z4" event={"ID":"5b9e78cc-aa6e-4369-ab8a-5e89b86359c1","Type":"ContainerStarted","Data":"32818695048dd3362ff1e26ded731fa65a905cf3d8e201439acefbf1750b86da"} Dec 13 07:12:07 crc kubenswrapper[4971]: I1213 07:12:07.422076 4971 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-db-sync-8z2nm" event={"ID":"0813350d-430f-4564-a09a-062a3239ba2f","Type":"ContainerStarted","Data":"79444a316a054f8a31c3874025055b3501be54e3988c3b3a02890b8643e62d72"} Dec 13 07:12:07 crc kubenswrapper[4971]: I1213 07:12:07.422128 4971 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-db-sync-8z2nm" event={"ID":"0813350d-430f-4564-a09a-062a3239ba2f","Type":"ContainerStarted","Data":"f8b36cdbe233f6b61c0a47a1a93c6169e7f2fae7dbfe2b7a99ce9ae4fd061d40"} Dec 13 07:12:07 crc kubenswrapper[4971]: I1213 07:12:07.425849 4971 generic.go:334] "Generic (PLEG): container finished" podID="2430f20e-0768-47e9-89de-fc6ac2615707" containerID="4f6c4e975a707094499c418eb050018f9723b4eb8f77c6b7f44577b34b9736cd" exitCode=0 Dec 13 07:12:07 crc kubenswrapper[4971]: I1213 07:12:07.425882 4971 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-5d87b7c6dc-sxklp" event={"ID":"2430f20e-0768-47e9-89de-fc6ac2615707","Type":"ContainerDied","Data":"4f6c4e975a707094499c418eb050018f9723b4eb8f77c6b7f44577b34b9736cd"} Dec 13 07:12:07 crc kubenswrapper[4971]: I1213 07:12:07.425901 4971 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-5d87b7c6dc-sxklp" event={"ID":"2430f20e-0768-47e9-89de-fc6ac2615707","Type":"ContainerStarted","Data":"e87b3742a8d0d0c1e72b0732bb0b473a60d040294def9f4ace6b98cb9c23f2cf"} Dec 13 07:12:07 crc kubenswrapper[4971]: I1213 07:12:07.434456 4971 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/horizon-fbfd486df-xmvls" Dec 13 07:12:07 crc kubenswrapper[4971]: I1213 07:12:07.451019 4971 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/keystone-bootstrap-xp9z4" podStartSLOduration=4.450999526 podStartE2EDuration="4.450999526s" podCreationTimestamp="2025-12-13 07:12:03 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-13 07:12:07.437946205 +0000 UTC m=+1384.042355653" watchObservedRunningTime="2025-12-13 07:12:07.450999526 +0000 UTC m=+1384.055408974" Dec 13 07:12:07 crc kubenswrapper[4971]: I1213 07:12:07.535862 4971 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/neutron-db-sync-8z2nm" podStartSLOduration=3.535838545 podStartE2EDuration="3.535838545s" podCreationTimestamp="2025-12-13 07:12:04 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-13 07:12:07.490679494 +0000 UTC m=+1384.095088952" watchObservedRunningTime="2025-12-13 07:12:07.535838545 +0000 UTC m=+1384.140247983" Dec 13 07:12:07 crc kubenswrapper[4971]: I1213 07:12:07.843250 4971 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-78fbc4bbf-6tztc" Dec 13 07:12:07 crc kubenswrapper[4971]: I1213 07:12:07.926156 4971 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/1b1d2613-64f2-4b48-aaf5-0681ff6cba2d-dns-svc\") pod \"1b1d2613-64f2-4b48-aaf5-0681ff6cba2d\" (UID: \"1b1d2613-64f2-4b48-aaf5-0681ff6cba2d\") " Dec 13 07:12:07 crc kubenswrapper[4971]: I1213 07:12:07.926250 4971 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/1b1d2613-64f2-4b48-aaf5-0681ff6cba2d-ovsdbserver-nb\") pod \"1b1d2613-64f2-4b48-aaf5-0681ff6cba2d\" (UID: \"1b1d2613-64f2-4b48-aaf5-0681ff6cba2d\") " Dec 13 07:12:07 crc kubenswrapper[4971]: I1213 07:12:07.926337 4971 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-4q2wg\" (UniqueName: \"kubernetes.io/projected/1b1d2613-64f2-4b48-aaf5-0681ff6cba2d-kube-api-access-4q2wg\") pod \"1b1d2613-64f2-4b48-aaf5-0681ff6cba2d\" (UID: \"1b1d2613-64f2-4b48-aaf5-0681ff6cba2d\") " Dec 13 07:12:07 crc kubenswrapper[4971]: I1213 07:12:07.926363 4971 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/1b1d2613-64f2-4b48-aaf5-0681ff6cba2d-ovsdbserver-sb\") pod \"1b1d2613-64f2-4b48-aaf5-0681ff6cba2d\" (UID: \"1b1d2613-64f2-4b48-aaf5-0681ff6cba2d\") " Dec 13 07:12:07 crc kubenswrapper[4971]: I1213 07:12:07.926392 4971 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/1b1d2613-64f2-4b48-aaf5-0681ff6cba2d-config\") pod \"1b1d2613-64f2-4b48-aaf5-0681ff6cba2d\" (UID: \"1b1d2613-64f2-4b48-aaf5-0681ff6cba2d\") " Dec 13 07:12:07 crc kubenswrapper[4971]: I1213 07:12:07.946071 4971 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/1b1d2613-64f2-4b48-aaf5-0681ff6cba2d-kube-api-access-4q2wg" (OuterVolumeSpecName: "kube-api-access-4q2wg") pod "1b1d2613-64f2-4b48-aaf5-0681ff6cba2d" (UID: "1b1d2613-64f2-4b48-aaf5-0681ff6cba2d"). InnerVolumeSpecName "kube-api-access-4q2wg". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 13 07:12:07 crc kubenswrapper[4971]: I1213 07:12:07.955381 4971 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/1b1d2613-64f2-4b48-aaf5-0681ff6cba2d-config" (OuterVolumeSpecName: "config") pod "1b1d2613-64f2-4b48-aaf5-0681ff6cba2d" (UID: "1b1d2613-64f2-4b48-aaf5-0681ff6cba2d"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 13 07:12:07 crc kubenswrapper[4971]: I1213 07:12:07.957953 4971 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/1b1d2613-64f2-4b48-aaf5-0681ff6cba2d-ovsdbserver-nb" (OuterVolumeSpecName: "ovsdbserver-nb") pod "1b1d2613-64f2-4b48-aaf5-0681ff6cba2d" (UID: "1b1d2613-64f2-4b48-aaf5-0681ff6cba2d"). InnerVolumeSpecName "ovsdbserver-nb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 13 07:12:07 crc kubenswrapper[4971]: I1213 07:12:07.961545 4971 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/1b1d2613-64f2-4b48-aaf5-0681ff6cba2d-dns-svc" (OuterVolumeSpecName: "dns-svc") pod "1b1d2613-64f2-4b48-aaf5-0681ff6cba2d" (UID: "1b1d2613-64f2-4b48-aaf5-0681ff6cba2d"). InnerVolumeSpecName "dns-svc". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 13 07:12:07 crc kubenswrapper[4971]: I1213 07:12:07.972122 4971 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/1b1d2613-64f2-4b48-aaf5-0681ff6cba2d-ovsdbserver-sb" (OuterVolumeSpecName: "ovsdbserver-sb") pod "1b1d2613-64f2-4b48-aaf5-0681ff6cba2d" (UID: "1b1d2613-64f2-4b48-aaf5-0681ff6cba2d"). InnerVolumeSpecName "ovsdbserver-sb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 13 07:12:08 crc kubenswrapper[4971]: I1213 07:12:08.032503 4971 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/1b1d2613-64f2-4b48-aaf5-0681ff6cba2d-ovsdbserver-nb\") on node \"crc\" DevicePath \"\"" Dec 13 07:12:08 crc kubenswrapper[4971]: I1213 07:12:08.032560 4971 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-4q2wg\" (UniqueName: \"kubernetes.io/projected/1b1d2613-64f2-4b48-aaf5-0681ff6cba2d-kube-api-access-4q2wg\") on node \"crc\" DevicePath \"\"" Dec 13 07:12:08 crc kubenswrapper[4971]: I1213 07:12:08.032571 4971 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/1b1d2613-64f2-4b48-aaf5-0681ff6cba2d-ovsdbserver-sb\") on node \"crc\" DevicePath \"\"" Dec 13 07:12:08 crc kubenswrapper[4971]: I1213 07:12:08.032580 4971 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/1b1d2613-64f2-4b48-aaf5-0681ff6cba2d-config\") on node \"crc\" DevicePath \"\"" Dec 13 07:12:08 crc kubenswrapper[4971]: I1213 07:12:08.032588 4971 reconciler_common.go:293] "Volume detached for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/1b1d2613-64f2-4b48-aaf5-0681ff6cba2d-dns-svc\") on node \"crc\" DevicePath \"\"" Dec 13 07:12:08 crc kubenswrapper[4971]: I1213 07:12:08.052418 4971 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/horizon-fbfd486df-xmvls"] Dec 13 07:12:08 crc kubenswrapper[4971]: W1213 07:12:08.075790 4971 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podf9bd96be_6d7f_415e_b67c_5bd02b05c536.slice/crio-729d1ba01c5742e471d188d542eb28bae9f7b38d079a5cc82bda8adc7c17ada6 WatchSource:0}: Error finding container 729d1ba01c5742e471d188d542eb28bae9f7b38d079a5cc82bda8adc7c17ada6: Status 404 returned error can't find the container with id 729d1ba01c5742e471d188d542eb28bae9f7b38d079a5cc82bda8adc7c17ada6 Dec 13 07:12:08 crc kubenswrapper[4971]: I1213 07:12:08.441996 4971 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-78fbc4bbf-6tztc" event={"ID":"1b1d2613-64f2-4b48-aaf5-0681ff6cba2d","Type":"ContainerDied","Data":"298778dec3373fc52010fcb45dcfa064cf54fd8eb6f0ed7e782ff8e010f8d850"} Dec 13 07:12:08 crc kubenswrapper[4971]: I1213 07:12:08.442444 4971 scope.go:117] "RemoveContainer" containerID="ba8b92c4feec80f2fa98a863f27ba987f3e38266fe6f9a3aca0fd5f97e01a75e" Dec 13 07:12:08 crc kubenswrapper[4971]: I1213 07:12:08.442225 4971 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-78fbc4bbf-6tztc" Dec 13 07:12:08 crc kubenswrapper[4971]: I1213 07:12:08.447809 4971 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/horizon-fbfd486df-xmvls" event={"ID":"f9bd96be-6d7f-415e-b67c-5bd02b05c536","Type":"ContainerStarted","Data":"729d1ba01c5742e471d188d542eb28bae9f7b38d079a5cc82bda8adc7c17ada6"} Dec 13 07:12:08 crc kubenswrapper[4971]: I1213 07:12:08.487044 4971 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-5d87b7c6dc-sxklp" event={"ID":"2430f20e-0768-47e9-89de-fc6ac2615707","Type":"ContainerStarted","Data":"c70e97bcd1c39a27d0495531710de58ef3511f93ac0bb637c42dd96a63642cc8"} Dec 13 07:12:08 crc kubenswrapper[4971]: I1213 07:12:08.510397 4971 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/dnsmasq-dns-5d87b7c6dc-sxklp" podStartSLOduration=4.510380478 podStartE2EDuration="4.510380478s" podCreationTimestamp="2025-12-13 07:12:04 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-13 07:12:08.504417762 +0000 UTC m=+1385.108827220" watchObservedRunningTime="2025-12-13 07:12:08.510380478 +0000 UTC m=+1385.114789926" Dec 13 07:12:08 crc kubenswrapper[4971]: I1213 07:12:08.561454 4971 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-78fbc4bbf-6tztc"] Dec 13 07:12:08 crc kubenswrapper[4971]: I1213 07:12:08.578676 4971 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-78fbc4bbf-6tztc"] Dec 13 07:12:09 crc kubenswrapper[4971]: I1213 07:12:09.521151 4971 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/dnsmasq-dns-5d87b7c6dc-sxklp" Dec 13 07:12:09 crc kubenswrapper[4971]: I1213 07:12:09.784806 4971 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="1b1d2613-64f2-4b48-aaf5-0681ff6cba2d" path="/var/lib/kubelet/pods/1b1d2613-64f2-4b48-aaf5-0681ff6cba2d/volumes" Dec 13 07:12:10 crc kubenswrapper[4971]: I1213 07:12:10.529914 4971 generic.go:334] "Generic (PLEG): container finished" podID="5b9e78cc-aa6e-4369-ab8a-5e89b86359c1" containerID="32818695048dd3362ff1e26ded731fa65a905cf3d8e201439acefbf1750b86da" exitCode=0 Dec 13 07:12:10 crc kubenswrapper[4971]: I1213 07:12:10.530750 4971 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-bootstrap-xp9z4" event={"ID":"5b9e78cc-aa6e-4369-ab8a-5e89b86359c1","Type":"ContainerDied","Data":"32818695048dd3362ff1e26ded731fa65a905cf3d8e201439acefbf1750b86da"} Dec 13 07:12:13 crc kubenswrapper[4971]: I1213 07:12:13.659220 4971 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/horizon-bcdd7767-j7d7f"] Dec 13 07:12:13 crc kubenswrapper[4971]: I1213 07:12:13.711314 4971 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/horizon-8699dc887d-sl2h6"] Dec 13 07:12:13 crc kubenswrapper[4971]: E1213 07:12:13.711704 4971 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="1b1d2613-64f2-4b48-aaf5-0681ff6cba2d" containerName="init" Dec 13 07:12:13 crc kubenswrapper[4971]: I1213 07:12:13.711744 4971 state_mem.go:107] "Deleted CPUSet assignment" podUID="1b1d2613-64f2-4b48-aaf5-0681ff6cba2d" containerName="init" Dec 13 07:12:13 crc kubenswrapper[4971]: I1213 07:12:13.711888 4971 memory_manager.go:354] "RemoveStaleState removing state" podUID="1b1d2613-64f2-4b48-aaf5-0681ff6cba2d" containerName="init" Dec 13 07:12:13 crc kubenswrapper[4971]: I1213 07:12:13.712732 4971 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/horizon-8699dc887d-sl2h6" Dec 13 07:12:13 crc kubenswrapper[4971]: I1213 07:12:13.745265 4971 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-horizon-svc" Dec 13 07:12:13 crc kubenswrapper[4971]: I1213 07:12:13.745415 4971 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/horizon-8699dc887d-sl2h6"] Dec 13 07:12:13 crc kubenswrapper[4971]: I1213 07:12:13.792176 4971 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/horizon-fbfd486df-xmvls"] Dec 13 07:12:13 crc kubenswrapper[4971]: I1213 07:12:13.812032 4971 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/horizon-797dfcfc58-wlch7"] Dec 13 07:12:13 crc kubenswrapper[4971]: I1213 07:12:13.813365 4971 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/horizon-797dfcfc58-wlch7" Dec 13 07:12:13 crc kubenswrapper[4971]: I1213 07:12:13.843072 4971 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/horizon-797dfcfc58-wlch7"] Dec 13 07:12:13 crc kubenswrapper[4971]: I1213 07:12:13.856399 4971 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-6sss5\" (UniqueName: \"kubernetes.io/projected/772eedab-7c62-454b-b722-ee453b39e89f-kube-api-access-6sss5\") pod \"horizon-8699dc887d-sl2h6\" (UID: \"772eedab-7c62-454b-b722-ee453b39e89f\") " pod="openstack/horizon-8699dc887d-sl2h6" Dec 13 07:12:13 crc kubenswrapper[4971]: I1213 07:12:13.856479 4971 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/772eedab-7c62-454b-b722-ee453b39e89f-config-data\") pod \"horizon-8699dc887d-sl2h6\" (UID: \"772eedab-7c62-454b-b722-ee453b39e89f\") " pod="openstack/horizon-8699dc887d-sl2h6" Dec 13 07:12:13 crc kubenswrapper[4971]: I1213 07:12:13.856671 4971 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/772eedab-7c62-454b-b722-ee453b39e89f-combined-ca-bundle\") pod \"horizon-8699dc887d-sl2h6\" (UID: \"772eedab-7c62-454b-b722-ee453b39e89f\") " pod="openstack/horizon-8699dc887d-sl2h6" Dec 13 07:12:13 crc kubenswrapper[4971]: I1213 07:12:13.856713 4971 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"horizon-tls-certs\" (UniqueName: \"kubernetes.io/secret/772eedab-7c62-454b-b722-ee453b39e89f-horizon-tls-certs\") pod \"horizon-8699dc887d-sl2h6\" (UID: \"772eedab-7c62-454b-b722-ee453b39e89f\") " pod="openstack/horizon-8699dc887d-sl2h6" Dec 13 07:12:13 crc kubenswrapper[4971]: I1213 07:12:13.856746 4971 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/772eedab-7c62-454b-b722-ee453b39e89f-logs\") pod \"horizon-8699dc887d-sl2h6\" (UID: \"772eedab-7c62-454b-b722-ee453b39e89f\") " pod="openstack/horizon-8699dc887d-sl2h6" Dec 13 07:12:13 crc kubenswrapper[4971]: I1213 07:12:13.856779 4971 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"horizon-secret-key\" (UniqueName: \"kubernetes.io/secret/772eedab-7c62-454b-b722-ee453b39e89f-horizon-secret-key\") pod \"horizon-8699dc887d-sl2h6\" (UID: \"772eedab-7c62-454b-b722-ee453b39e89f\") " pod="openstack/horizon-8699dc887d-sl2h6" Dec 13 07:12:13 crc kubenswrapper[4971]: I1213 07:12:13.856842 4971 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/772eedab-7c62-454b-b722-ee453b39e89f-scripts\") pod \"horizon-8699dc887d-sl2h6\" (UID: \"772eedab-7c62-454b-b722-ee453b39e89f\") " pod="openstack/horizon-8699dc887d-sl2h6" Dec 13 07:12:13 crc kubenswrapper[4971]: I1213 07:12:13.959678 4971 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/772eedab-7c62-454b-b722-ee453b39e89f-config-data\") pod \"horizon-8699dc887d-sl2h6\" (UID: \"772eedab-7c62-454b-b722-ee453b39e89f\") " pod="openstack/horizon-8699dc887d-sl2h6" Dec 13 07:12:13 crc kubenswrapper[4971]: I1213 07:12:13.959962 4971 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/9a4736be-5537-4638-a203-d36db1ae35cc-combined-ca-bundle\") pod \"horizon-797dfcfc58-wlch7\" (UID: \"9a4736be-5537-4638-a203-d36db1ae35cc\") " pod="openstack/horizon-797dfcfc58-wlch7" Dec 13 07:12:13 crc kubenswrapper[4971]: I1213 07:12:13.959997 4971 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"horizon-secret-key\" (UniqueName: \"kubernetes.io/secret/9a4736be-5537-4638-a203-d36db1ae35cc-horizon-secret-key\") pod \"horizon-797dfcfc58-wlch7\" (UID: \"9a4736be-5537-4638-a203-d36db1ae35cc\") " pod="openstack/horizon-797dfcfc58-wlch7" Dec 13 07:12:13 crc kubenswrapper[4971]: I1213 07:12:13.960025 4971 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"horizon-tls-certs\" (UniqueName: \"kubernetes.io/secret/9a4736be-5537-4638-a203-d36db1ae35cc-horizon-tls-certs\") pod \"horizon-797dfcfc58-wlch7\" (UID: \"9a4736be-5537-4638-a203-d36db1ae35cc\") " pod="openstack/horizon-797dfcfc58-wlch7" Dec 13 07:12:13 crc kubenswrapper[4971]: I1213 07:12:13.960052 4971 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/9a4736be-5537-4638-a203-d36db1ae35cc-config-data\") pod \"horizon-797dfcfc58-wlch7\" (UID: \"9a4736be-5537-4638-a203-d36db1ae35cc\") " pod="openstack/horizon-797dfcfc58-wlch7" Dec 13 07:12:13 crc kubenswrapper[4971]: I1213 07:12:13.960076 4971 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-m2558\" (UniqueName: \"kubernetes.io/projected/9a4736be-5537-4638-a203-d36db1ae35cc-kube-api-access-m2558\") pod \"horizon-797dfcfc58-wlch7\" (UID: \"9a4736be-5537-4638-a203-d36db1ae35cc\") " pod="openstack/horizon-797dfcfc58-wlch7" Dec 13 07:12:13 crc kubenswrapper[4971]: I1213 07:12:13.960103 4971 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/772eedab-7c62-454b-b722-ee453b39e89f-combined-ca-bundle\") pod \"horizon-8699dc887d-sl2h6\" (UID: \"772eedab-7c62-454b-b722-ee453b39e89f\") " pod="openstack/horizon-8699dc887d-sl2h6" Dec 13 07:12:13 crc kubenswrapper[4971]: I1213 07:12:13.960121 4971 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"horizon-tls-certs\" (UniqueName: \"kubernetes.io/secret/772eedab-7c62-454b-b722-ee453b39e89f-horizon-tls-certs\") pod \"horizon-8699dc887d-sl2h6\" (UID: \"772eedab-7c62-454b-b722-ee453b39e89f\") " pod="openstack/horizon-8699dc887d-sl2h6" Dec 13 07:12:13 crc kubenswrapper[4971]: I1213 07:12:13.960149 4971 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/772eedab-7c62-454b-b722-ee453b39e89f-logs\") pod \"horizon-8699dc887d-sl2h6\" (UID: \"772eedab-7c62-454b-b722-ee453b39e89f\") " pod="openstack/horizon-8699dc887d-sl2h6" Dec 13 07:12:13 crc kubenswrapper[4971]: I1213 07:12:13.960168 4971 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"horizon-secret-key\" (UniqueName: \"kubernetes.io/secret/772eedab-7c62-454b-b722-ee453b39e89f-horizon-secret-key\") pod \"horizon-8699dc887d-sl2h6\" (UID: \"772eedab-7c62-454b-b722-ee453b39e89f\") " pod="openstack/horizon-8699dc887d-sl2h6" Dec 13 07:12:13 crc kubenswrapper[4971]: I1213 07:12:13.960199 4971 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/9a4736be-5537-4638-a203-d36db1ae35cc-scripts\") pod \"horizon-797dfcfc58-wlch7\" (UID: \"9a4736be-5537-4638-a203-d36db1ae35cc\") " pod="openstack/horizon-797dfcfc58-wlch7" Dec 13 07:12:13 crc kubenswrapper[4971]: I1213 07:12:13.960221 4971 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/772eedab-7c62-454b-b722-ee453b39e89f-scripts\") pod \"horizon-8699dc887d-sl2h6\" (UID: \"772eedab-7c62-454b-b722-ee453b39e89f\") " pod="openstack/horizon-8699dc887d-sl2h6" Dec 13 07:12:13 crc kubenswrapper[4971]: I1213 07:12:13.960257 4971 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-6sss5\" (UniqueName: \"kubernetes.io/projected/772eedab-7c62-454b-b722-ee453b39e89f-kube-api-access-6sss5\") pod \"horizon-8699dc887d-sl2h6\" (UID: \"772eedab-7c62-454b-b722-ee453b39e89f\") " pod="openstack/horizon-8699dc887d-sl2h6" Dec 13 07:12:13 crc kubenswrapper[4971]: I1213 07:12:13.960281 4971 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/9a4736be-5537-4638-a203-d36db1ae35cc-logs\") pod \"horizon-797dfcfc58-wlch7\" (UID: \"9a4736be-5537-4638-a203-d36db1ae35cc\") " pod="openstack/horizon-797dfcfc58-wlch7" Dec 13 07:12:13 crc kubenswrapper[4971]: I1213 07:12:13.962058 4971 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/772eedab-7c62-454b-b722-ee453b39e89f-config-data\") pod \"horizon-8699dc887d-sl2h6\" (UID: \"772eedab-7c62-454b-b722-ee453b39e89f\") " pod="openstack/horizon-8699dc887d-sl2h6" Dec 13 07:12:13 crc kubenswrapper[4971]: I1213 07:12:13.962492 4971 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/772eedab-7c62-454b-b722-ee453b39e89f-logs\") pod \"horizon-8699dc887d-sl2h6\" (UID: \"772eedab-7c62-454b-b722-ee453b39e89f\") " pod="openstack/horizon-8699dc887d-sl2h6" Dec 13 07:12:13 crc kubenswrapper[4971]: I1213 07:12:13.963285 4971 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/772eedab-7c62-454b-b722-ee453b39e89f-scripts\") pod \"horizon-8699dc887d-sl2h6\" (UID: \"772eedab-7c62-454b-b722-ee453b39e89f\") " pod="openstack/horizon-8699dc887d-sl2h6" Dec 13 07:12:13 crc kubenswrapper[4971]: I1213 07:12:13.976417 4971 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"horizon-tls-certs\" (UniqueName: \"kubernetes.io/secret/772eedab-7c62-454b-b722-ee453b39e89f-horizon-tls-certs\") pod \"horizon-8699dc887d-sl2h6\" (UID: \"772eedab-7c62-454b-b722-ee453b39e89f\") " pod="openstack/horizon-8699dc887d-sl2h6" Dec 13 07:12:13 crc kubenswrapper[4971]: I1213 07:12:13.976646 4971 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/772eedab-7c62-454b-b722-ee453b39e89f-combined-ca-bundle\") pod \"horizon-8699dc887d-sl2h6\" (UID: \"772eedab-7c62-454b-b722-ee453b39e89f\") " pod="openstack/horizon-8699dc887d-sl2h6" Dec 13 07:12:13 crc kubenswrapper[4971]: I1213 07:12:13.978222 4971 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"horizon-secret-key\" (UniqueName: \"kubernetes.io/secret/772eedab-7c62-454b-b722-ee453b39e89f-horizon-secret-key\") pod \"horizon-8699dc887d-sl2h6\" (UID: \"772eedab-7c62-454b-b722-ee453b39e89f\") " pod="openstack/horizon-8699dc887d-sl2h6" Dec 13 07:12:13 crc kubenswrapper[4971]: I1213 07:12:13.980502 4971 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-6sss5\" (UniqueName: \"kubernetes.io/projected/772eedab-7c62-454b-b722-ee453b39e89f-kube-api-access-6sss5\") pod \"horizon-8699dc887d-sl2h6\" (UID: \"772eedab-7c62-454b-b722-ee453b39e89f\") " pod="openstack/horizon-8699dc887d-sl2h6" Dec 13 07:12:14 crc kubenswrapper[4971]: I1213 07:12:14.034457 4971 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/horizon-8699dc887d-sl2h6" Dec 13 07:12:14 crc kubenswrapper[4971]: I1213 07:12:14.061675 4971 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/9a4736be-5537-4638-a203-d36db1ae35cc-combined-ca-bundle\") pod \"horizon-797dfcfc58-wlch7\" (UID: \"9a4736be-5537-4638-a203-d36db1ae35cc\") " pod="openstack/horizon-797dfcfc58-wlch7" Dec 13 07:12:14 crc kubenswrapper[4971]: I1213 07:12:14.061736 4971 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"horizon-secret-key\" (UniqueName: \"kubernetes.io/secret/9a4736be-5537-4638-a203-d36db1ae35cc-horizon-secret-key\") pod \"horizon-797dfcfc58-wlch7\" (UID: \"9a4736be-5537-4638-a203-d36db1ae35cc\") " pod="openstack/horizon-797dfcfc58-wlch7" Dec 13 07:12:14 crc kubenswrapper[4971]: I1213 07:12:14.061763 4971 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"horizon-tls-certs\" (UniqueName: \"kubernetes.io/secret/9a4736be-5537-4638-a203-d36db1ae35cc-horizon-tls-certs\") pod \"horizon-797dfcfc58-wlch7\" (UID: \"9a4736be-5537-4638-a203-d36db1ae35cc\") " pod="openstack/horizon-797dfcfc58-wlch7" Dec 13 07:12:14 crc kubenswrapper[4971]: I1213 07:12:14.061790 4971 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/9a4736be-5537-4638-a203-d36db1ae35cc-config-data\") pod \"horizon-797dfcfc58-wlch7\" (UID: \"9a4736be-5537-4638-a203-d36db1ae35cc\") " pod="openstack/horizon-797dfcfc58-wlch7" Dec 13 07:12:14 crc kubenswrapper[4971]: I1213 07:12:14.061810 4971 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-m2558\" (UniqueName: \"kubernetes.io/projected/9a4736be-5537-4638-a203-d36db1ae35cc-kube-api-access-m2558\") pod \"horizon-797dfcfc58-wlch7\" (UID: \"9a4736be-5537-4638-a203-d36db1ae35cc\") " pod="openstack/horizon-797dfcfc58-wlch7" Dec 13 07:12:14 crc kubenswrapper[4971]: I1213 07:12:14.061857 4971 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/9a4736be-5537-4638-a203-d36db1ae35cc-scripts\") pod \"horizon-797dfcfc58-wlch7\" (UID: \"9a4736be-5537-4638-a203-d36db1ae35cc\") " pod="openstack/horizon-797dfcfc58-wlch7" Dec 13 07:12:14 crc kubenswrapper[4971]: I1213 07:12:14.061901 4971 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/9a4736be-5537-4638-a203-d36db1ae35cc-logs\") pod \"horizon-797dfcfc58-wlch7\" (UID: \"9a4736be-5537-4638-a203-d36db1ae35cc\") " pod="openstack/horizon-797dfcfc58-wlch7" Dec 13 07:12:14 crc kubenswrapper[4971]: I1213 07:12:14.063178 4971 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/9a4736be-5537-4638-a203-d36db1ae35cc-config-data\") pod \"horizon-797dfcfc58-wlch7\" (UID: \"9a4736be-5537-4638-a203-d36db1ae35cc\") " pod="openstack/horizon-797dfcfc58-wlch7" Dec 13 07:12:14 crc kubenswrapper[4971]: I1213 07:12:14.063877 4971 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/9a4736be-5537-4638-a203-d36db1ae35cc-logs\") pod \"horizon-797dfcfc58-wlch7\" (UID: \"9a4736be-5537-4638-a203-d36db1ae35cc\") " pod="openstack/horizon-797dfcfc58-wlch7" Dec 13 07:12:14 crc kubenswrapper[4971]: I1213 07:12:14.064788 4971 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"horizon-tls-certs\" (UniqueName: \"kubernetes.io/secret/9a4736be-5537-4638-a203-d36db1ae35cc-horizon-tls-certs\") pod \"horizon-797dfcfc58-wlch7\" (UID: \"9a4736be-5537-4638-a203-d36db1ae35cc\") " pod="openstack/horizon-797dfcfc58-wlch7" Dec 13 07:12:14 crc kubenswrapper[4971]: I1213 07:12:14.064873 4971 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/9a4736be-5537-4638-a203-d36db1ae35cc-scripts\") pod \"horizon-797dfcfc58-wlch7\" (UID: \"9a4736be-5537-4638-a203-d36db1ae35cc\") " pod="openstack/horizon-797dfcfc58-wlch7" Dec 13 07:12:14 crc kubenswrapper[4971]: I1213 07:12:14.066387 4971 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/9a4736be-5537-4638-a203-d36db1ae35cc-combined-ca-bundle\") pod \"horizon-797dfcfc58-wlch7\" (UID: \"9a4736be-5537-4638-a203-d36db1ae35cc\") " pod="openstack/horizon-797dfcfc58-wlch7" Dec 13 07:12:14 crc kubenswrapper[4971]: I1213 07:12:14.076876 4971 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"horizon-secret-key\" (UniqueName: \"kubernetes.io/secret/9a4736be-5537-4638-a203-d36db1ae35cc-horizon-secret-key\") pod \"horizon-797dfcfc58-wlch7\" (UID: \"9a4736be-5537-4638-a203-d36db1ae35cc\") " pod="openstack/horizon-797dfcfc58-wlch7" Dec 13 07:12:14 crc kubenswrapper[4971]: I1213 07:12:14.079551 4971 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-m2558\" (UniqueName: \"kubernetes.io/projected/9a4736be-5537-4638-a203-d36db1ae35cc-kube-api-access-m2558\") pod \"horizon-797dfcfc58-wlch7\" (UID: \"9a4736be-5537-4638-a203-d36db1ae35cc\") " pod="openstack/horizon-797dfcfc58-wlch7" Dec 13 07:12:14 crc kubenswrapper[4971]: I1213 07:12:14.163312 4971 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/horizon-797dfcfc58-wlch7" Dec 13 07:12:15 crc kubenswrapper[4971]: I1213 07:12:15.678748 4971 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/dnsmasq-dns-5d87b7c6dc-sxklp" Dec 13 07:12:15 crc kubenswrapper[4971]: I1213 07:12:15.725187 4971 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-57768dd7b5-86hlw"] Dec 13 07:12:15 crc kubenswrapper[4971]: I1213 07:12:15.725488 4971 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/dnsmasq-dns-57768dd7b5-86hlw" podUID="c9e30081-d603-4c9d-81ce-7bcff6848243" containerName="dnsmasq-dns" containerID="cri-o://303ee14cece58dbf062bf3b1d7ce80d4f37ced4fdb8e6c483677f2b786064dd4" gracePeriod=10 Dec 13 07:12:16 crc kubenswrapper[4971]: I1213 07:12:16.578266 4971 generic.go:334] "Generic (PLEG): container finished" podID="c9e30081-d603-4c9d-81ce-7bcff6848243" containerID="303ee14cece58dbf062bf3b1d7ce80d4f37ced4fdb8e6c483677f2b786064dd4" exitCode=0 Dec 13 07:12:16 crc kubenswrapper[4971]: I1213 07:12:16.578338 4971 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-57768dd7b5-86hlw" event={"ID":"c9e30081-d603-4c9d-81ce-7bcff6848243","Type":"ContainerDied","Data":"303ee14cece58dbf062bf3b1d7ce80d4f37ced4fdb8e6c483677f2b786064dd4"} Dec 13 07:12:20 crc kubenswrapper[4971]: I1213 07:12:20.578473 4971 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/dnsmasq-dns-57768dd7b5-86hlw" podUID="c9e30081-d603-4c9d-81ce-7bcff6848243" containerName="dnsmasq-dns" probeResult="failure" output="dial tcp 10.217.0.121:5353: connect: connection refused" Dec 13 07:12:25 crc kubenswrapper[4971]: I1213 07:12:25.579089 4971 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/dnsmasq-dns-57768dd7b5-86hlw" podUID="c9e30081-d603-4c9d-81ce-7bcff6848243" containerName="dnsmasq-dns" probeResult="failure" output="dial tcp 10.217.0.121:5353: connect: connection refused" Dec 13 07:12:26 crc kubenswrapper[4971]: E1213 07:12:26.468391 4971 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying config: context canceled" image="quay.io/podified-antelope-centos9/openstack-horizon@sha256:dd7600bc5278c663cfcfecafd3fb051a2cd2ddc3c1efb07738bf09512aa23ae7" Dec 13 07:12:26 crc kubenswrapper[4971]: E1213 07:12:26.468715 4971 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:horizon-log,Image:quay.io/podified-antelope-centos9/openstack-horizon@sha256:dd7600bc5278c663cfcfecafd3fb051a2cd2ddc3c1efb07738bf09512aa23ae7,Command:[/bin/bash],Args:[-c tail -n+1 -F /var/log/horizon/horizon.log],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:CONFIG_HASH,Value:n8dh558h578h654h89h654h659h5d5h68h559h5dh5c9h7bhfch57dhddhdh55bh77h76h58fh576hd7h5cdh6bh547h589h568h7fh5dfh565h5bcq,ValueFrom:nil,},EnvVar{Name:ENABLE_DESIGNATE,Value:yes,ValueFrom:nil,},EnvVar{Name:ENABLE_HEAT,Value:yes,ValueFrom:nil,},EnvVar{Name:ENABLE_IRONIC,Value:yes,ValueFrom:nil,},EnvVar{Name:ENABLE_MANILA,Value:yes,ValueFrom:nil,},EnvVar{Name:ENABLE_OCTAVIA,Value:yes,ValueFrom:nil,},EnvVar{Name:ENABLE_WATCHER,Value:no,ValueFrom:nil,},EnvVar{Name:KOLLA_CONFIG_STRATEGY,Value:COPY_ALWAYS,ValueFrom:nil,},EnvVar{Name:UNPACK_THEME,Value:true,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:logs,ReadOnly:false,MountPath:/var/log/horizon,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-cpxzk,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[MKNOD],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*48,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*true,RunAsGroup:*42400,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod horizon-bcdd7767-j7d7f_openstack(ff35e760-13e7-4145-9970-742b73a36176): ErrImagePull: rpc error: code = Canceled desc = copying config: context canceled" logger="UnhandledError" Dec 13 07:12:26 crc kubenswrapper[4971]: E1213 07:12:26.501091 4971 pod_workers.go:1301] "Error syncing pod, skipping" err="[failed to \"StartContainer\" for \"horizon-log\" with ErrImagePull: \"rpc error: code = Canceled desc = copying config: context canceled\", failed to \"StartContainer\" for \"horizon\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/podified-antelope-centos9/openstack-horizon@sha256:dd7600bc5278c663cfcfecafd3fb051a2cd2ddc3c1efb07738bf09512aa23ae7\\\"\"]" pod="openstack/horizon-bcdd7767-j7d7f" podUID="ff35e760-13e7-4145-9970-742b73a36176" Dec 13 07:12:28 crc kubenswrapper[4971]: E1213 07:12:28.172162 4971 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying config: context canceled" image="quay.io/podified-antelope-centos9/openstack-horizon@sha256:dd7600bc5278c663cfcfecafd3fb051a2cd2ddc3c1efb07738bf09512aa23ae7" Dec 13 07:12:28 crc kubenswrapper[4971]: E1213 07:12:28.172739 4971 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:horizon-log,Image:quay.io/podified-antelope-centos9/openstack-horizon@sha256:dd7600bc5278c663cfcfecafd3fb051a2cd2ddc3c1efb07738bf09512aa23ae7,Command:[/bin/bash],Args:[-c tail -n+1 -F /var/log/horizon/horizon.log],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:CONFIG_HASH,Value:n5d4h5d8h5cfh58bh8fhfch7dhdbh57fh698h595h685h645h569h588h556h57ch5cfh5b7h5dbh5cbh6fh585h566h64h5ch5d5h577h664h558h659h577q,ValueFrom:nil,},EnvVar{Name:ENABLE_DESIGNATE,Value:yes,ValueFrom:nil,},EnvVar{Name:ENABLE_HEAT,Value:yes,ValueFrom:nil,},EnvVar{Name:ENABLE_IRONIC,Value:yes,ValueFrom:nil,},EnvVar{Name:ENABLE_MANILA,Value:yes,ValueFrom:nil,},EnvVar{Name:ENABLE_OCTAVIA,Value:yes,ValueFrom:nil,},EnvVar{Name:ENABLE_WATCHER,Value:no,ValueFrom:nil,},EnvVar{Name:KOLLA_CONFIG_STRATEGY,Value:COPY_ALWAYS,ValueFrom:nil,},EnvVar{Name:UNPACK_THEME,Value:true,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:logs,ReadOnly:false,MountPath:/var/log/horizon,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-tghr9,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[MKNOD],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*48,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*true,RunAsGroup:*42400,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod horizon-77c4585f77-kkx2c_openstack(b314f107-1cb6-49f3-aa51-3f607ca79f6b): ErrImagePull: rpc error: code = Canceled desc = copying config: context canceled" logger="UnhandledError" Dec 13 07:12:28 crc kubenswrapper[4971]: E1213 07:12:28.174952 4971 pod_workers.go:1301] "Error syncing pod, skipping" err="[failed to \"StartContainer\" for \"horizon-log\" with ErrImagePull: \"rpc error: code = Canceled desc = copying config: context canceled\", failed to \"StartContainer\" for \"horizon\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/podified-antelope-centos9/openstack-horizon@sha256:dd7600bc5278c663cfcfecafd3fb051a2cd2ddc3c1efb07738bf09512aa23ae7\\\"\"]" pod="openstack/horizon-77c4585f77-kkx2c" podUID="b314f107-1cb6-49f3-aa51-3f607ca79f6b" Dec 13 07:12:28 crc kubenswrapper[4971]: E1213 07:12:28.190793 4971 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying config: context canceled" image="quay.io/podified-antelope-centos9/openstack-horizon@sha256:dd7600bc5278c663cfcfecafd3fb051a2cd2ddc3c1efb07738bf09512aa23ae7" Dec 13 07:12:28 crc kubenswrapper[4971]: E1213 07:12:28.190981 4971 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:horizon-log,Image:quay.io/podified-antelope-centos9/openstack-horizon@sha256:dd7600bc5278c663cfcfecafd3fb051a2cd2ddc3c1efb07738bf09512aa23ae7,Command:[/bin/bash],Args:[-c tail -n+1 -F /var/log/horizon/horizon.log],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:CONFIG_HASH,Value:nb5h57bh58h5d6h578hfh696h98h585h59fh8h5fh649h65h5dfh54dh5d5h5bbh54fh665h644h5b8h68h5cch7bhfhb8h8dh64bh6bh687hb9q,ValueFrom:nil,},EnvVar{Name:ENABLE_DESIGNATE,Value:yes,ValueFrom:nil,},EnvVar{Name:ENABLE_HEAT,Value:yes,ValueFrom:nil,},EnvVar{Name:ENABLE_IRONIC,Value:yes,ValueFrom:nil,},EnvVar{Name:ENABLE_MANILA,Value:yes,ValueFrom:nil,},EnvVar{Name:ENABLE_OCTAVIA,Value:yes,ValueFrom:nil,},EnvVar{Name:ENABLE_WATCHER,Value:no,ValueFrom:nil,},EnvVar{Name:KOLLA_CONFIG_STRATEGY,Value:COPY_ALWAYS,ValueFrom:nil,},EnvVar{Name:UNPACK_THEME,Value:true,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:logs,ReadOnly:false,MountPath:/var/log/horizon,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-kjwdg,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[MKNOD],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*48,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*true,RunAsGroup:*42400,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod horizon-fbfd486df-xmvls_openstack(f9bd96be-6d7f-415e-b67c-5bd02b05c536): ErrImagePull: rpc error: code = Canceled desc = copying config: context canceled" logger="UnhandledError" Dec 13 07:12:28 crc kubenswrapper[4971]: E1213 07:12:28.200990 4971 pod_workers.go:1301] "Error syncing pod, skipping" err="[failed to \"StartContainer\" for \"horizon-log\" with ErrImagePull: \"rpc error: code = Canceled desc = copying config: context canceled\", failed to \"StartContainer\" for \"horizon\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/podified-antelope-centos9/openstack-horizon@sha256:dd7600bc5278c663cfcfecafd3fb051a2cd2ddc3c1efb07738bf09512aa23ae7\\\"\"]" pod="openstack/horizon-fbfd486df-xmvls" podUID="f9bd96be-6d7f-415e-b67c-5bd02b05c536" Dec 13 07:12:28 crc kubenswrapper[4971]: I1213 07:12:28.243148 4971 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-bootstrap-xp9z4" Dec 13 07:12:28 crc kubenswrapper[4971]: I1213 07:12:28.427968 4971 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/5b9e78cc-aa6e-4369-ab8a-5e89b86359c1-scripts\") pod \"5b9e78cc-aa6e-4369-ab8a-5e89b86359c1\" (UID: \"5b9e78cc-aa6e-4369-ab8a-5e89b86359c1\") " Dec 13 07:12:28 crc kubenswrapper[4971]: I1213 07:12:28.428030 4971 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/5b9e78cc-aa6e-4369-ab8a-5e89b86359c1-config-data\") pod \"5b9e78cc-aa6e-4369-ab8a-5e89b86359c1\" (UID: \"5b9e78cc-aa6e-4369-ab8a-5e89b86359c1\") " Dec 13 07:12:28 crc kubenswrapper[4971]: I1213 07:12:28.428060 4971 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/5b9e78cc-aa6e-4369-ab8a-5e89b86359c1-combined-ca-bundle\") pod \"5b9e78cc-aa6e-4369-ab8a-5e89b86359c1\" (UID: \"5b9e78cc-aa6e-4369-ab8a-5e89b86359c1\") " Dec 13 07:12:28 crc kubenswrapper[4971]: I1213 07:12:28.428116 4971 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/5b9e78cc-aa6e-4369-ab8a-5e89b86359c1-fernet-keys\") pod \"5b9e78cc-aa6e-4369-ab8a-5e89b86359c1\" (UID: \"5b9e78cc-aa6e-4369-ab8a-5e89b86359c1\") " Dec 13 07:12:28 crc kubenswrapper[4971]: I1213 07:12:28.428187 4971 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"credential-keys\" (UniqueName: \"kubernetes.io/secret/5b9e78cc-aa6e-4369-ab8a-5e89b86359c1-credential-keys\") pod \"5b9e78cc-aa6e-4369-ab8a-5e89b86359c1\" (UID: \"5b9e78cc-aa6e-4369-ab8a-5e89b86359c1\") " Dec 13 07:12:28 crc kubenswrapper[4971]: I1213 07:12:28.428233 4971 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-ds44p\" (UniqueName: \"kubernetes.io/projected/5b9e78cc-aa6e-4369-ab8a-5e89b86359c1-kube-api-access-ds44p\") pod \"5b9e78cc-aa6e-4369-ab8a-5e89b86359c1\" (UID: \"5b9e78cc-aa6e-4369-ab8a-5e89b86359c1\") " Dec 13 07:12:28 crc kubenswrapper[4971]: I1213 07:12:28.434434 4971 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/5b9e78cc-aa6e-4369-ab8a-5e89b86359c1-kube-api-access-ds44p" (OuterVolumeSpecName: "kube-api-access-ds44p") pod "5b9e78cc-aa6e-4369-ab8a-5e89b86359c1" (UID: "5b9e78cc-aa6e-4369-ab8a-5e89b86359c1"). InnerVolumeSpecName "kube-api-access-ds44p". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 13 07:12:28 crc kubenswrapper[4971]: I1213 07:12:28.434670 4971 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/5b9e78cc-aa6e-4369-ab8a-5e89b86359c1-scripts" (OuterVolumeSpecName: "scripts") pod "5b9e78cc-aa6e-4369-ab8a-5e89b86359c1" (UID: "5b9e78cc-aa6e-4369-ab8a-5e89b86359c1"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 13 07:12:28 crc kubenswrapper[4971]: I1213 07:12:28.435228 4971 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/5b9e78cc-aa6e-4369-ab8a-5e89b86359c1-fernet-keys" (OuterVolumeSpecName: "fernet-keys") pod "5b9e78cc-aa6e-4369-ab8a-5e89b86359c1" (UID: "5b9e78cc-aa6e-4369-ab8a-5e89b86359c1"). InnerVolumeSpecName "fernet-keys". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 13 07:12:28 crc kubenswrapper[4971]: I1213 07:12:28.435602 4971 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/5b9e78cc-aa6e-4369-ab8a-5e89b86359c1-credential-keys" (OuterVolumeSpecName: "credential-keys") pod "5b9e78cc-aa6e-4369-ab8a-5e89b86359c1" (UID: "5b9e78cc-aa6e-4369-ab8a-5e89b86359c1"). InnerVolumeSpecName "credential-keys". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 13 07:12:28 crc kubenswrapper[4971]: I1213 07:12:28.453550 4971 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/5b9e78cc-aa6e-4369-ab8a-5e89b86359c1-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "5b9e78cc-aa6e-4369-ab8a-5e89b86359c1" (UID: "5b9e78cc-aa6e-4369-ab8a-5e89b86359c1"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 13 07:12:28 crc kubenswrapper[4971]: I1213 07:12:28.462676 4971 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/5b9e78cc-aa6e-4369-ab8a-5e89b86359c1-config-data" (OuterVolumeSpecName: "config-data") pod "5b9e78cc-aa6e-4369-ab8a-5e89b86359c1" (UID: "5b9e78cc-aa6e-4369-ab8a-5e89b86359c1"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 13 07:12:28 crc kubenswrapper[4971]: I1213 07:12:28.530453 4971 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/5b9e78cc-aa6e-4369-ab8a-5e89b86359c1-scripts\") on node \"crc\" DevicePath \"\"" Dec 13 07:12:28 crc kubenswrapper[4971]: I1213 07:12:28.530483 4971 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/5b9e78cc-aa6e-4369-ab8a-5e89b86359c1-config-data\") on node \"crc\" DevicePath \"\"" Dec 13 07:12:28 crc kubenswrapper[4971]: I1213 07:12:28.530494 4971 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/5b9e78cc-aa6e-4369-ab8a-5e89b86359c1-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 13 07:12:28 crc kubenswrapper[4971]: I1213 07:12:28.530505 4971 reconciler_common.go:293] "Volume detached for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/5b9e78cc-aa6e-4369-ab8a-5e89b86359c1-fernet-keys\") on node \"crc\" DevicePath \"\"" Dec 13 07:12:28 crc kubenswrapper[4971]: I1213 07:12:28.530526 4971 reconciler_common.go:293] "Volume detached for volume \"credential-keys\" (UniqueName: \"kubernetes.io/secret/5b9e78cc-aa6e-4369-ab8a-5e89b86359c1-credential-keys\") on node \"crc\" DevicePath \"\"" Dec 13 07:12:28 crc kubenswrapper[4971]: I1213 07:12:28.530534 4971 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-ds44p\" (UniqueName: \"kubernetes.io/projected/5b9e78cc-aa6e-4369-ab8a-5e89b86359c1-kube-api-access-ds44p\") on node \"crc\" DevicePath \"\"" Dec 13 07:12:28 crc kubenswrapper[4971]: I1213 07:12:28.692474 4971 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-bootstrap-xp9z4" event={"ID":"5b9e78cc-aa6e-4369-ab8a-5e89b86359c1","Type":"ContainerDied","Data":"4191d30758fbfe0bd41086142577cea859c9a2795fba391926a355d657e0dc0d"} Dec 13 07:12:28 crc kubenswrapper[4971]: I1213 07:12:28.692533 4971 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="4191d30758fbfe0bd41086142577cea859c9a2795fba391926a355d657e0dc0d" Dec 13 07:12:28 crc kubenswrapper[4971]: I1213 07:12:28.692609 4971 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-bootstrap-xp9z4" Dec 13 07:12:29 crc kubenswrapper[4971]: I1213 07:12:29.327655 4971 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/keystone-bootstrap-xp9z4"] Dec 13 07:12:29 crc kubenswrapper[4971]: I1213 07:12:29.338054 4971 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/keystone-bootstrap-xp9z4"] Dec 13 07:12:29 crc kubenswrapper[4971]: I1213 07:12:29.430620 4971 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/keystone-bootstrap-dn9h9"] Dec 13 07:12:29 crc kubenswrapper[4971]: E1213 07:12:29.430995 4971 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="5b9e78cc-aa6e-4369-ab8a-5e89b86359c1" containerName="keystone-bootstrap" Dec 13 07:12:29 crc kubenswrapper[4971]: I1213 07:12:29.431013 4971 state_mem.go:107] "Deleted CPUSet assignment" podUID="5b9e78cc-aa6e-4369-ab8a-5e89b86359c1" containerName="keystone-bootstrap" Dec 13 07:12:29 crc kubenswrapper[4971]: I1213 07:12:29.431259 4971 memory_manager.go:354] "RemoveStaleState removing state" podUID="5b9e78cc-aa6e-4369-ab8a-5e89b86359c1" containerName="keystone-bootstrap" Dec 13 07:12:29 crc kubenswrapper[4971]: I1213 07:12:29.431804 4971 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-bootstrap-dn9h9" Dec 13 07:12:29 crc kubenswrapper[4971]: I1213 07:12:29.433778 4971 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone-scripts" Dec 13 07:12:29 crc kubenswrapper[4971]: I1213 07:12:29.434382 4971 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone-config-data" Dec 13 07:12:29 crc kubenswrapper[4971]: I1213 07:12:29.434442 4971 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone" Dec 13 07:12:29 crc kubenswrapper[4971]: I1213 07:12:29.435998 4971 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone-keystone-dockercfg-56xxb" Dec 13 07:12:29 crc kubenswrapper[4971]: I1213 07:12:29.436141 4971 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"osp-secret" Dec 13 07:12:29 crc kubenswrapper[4971]: I1213 07:12:29.451416 4971 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/keystone-bootstrap-dn9h9"] Dec 13 07:12:29 crc kubenswrapper[4971]: I1213 07:12:29.552046 4971 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/ee3afc2f-79c6-444c-b6a3-2d4c475c3aac-fernet-keys\") pod \"keystone-bootstrap-dn9h9\" (UID: \"ee3afc2f-79c6-444c-b6a3-2d4c475c3aac\") " pod="openstack/keystone-bootstrap-dn9h9" Dec 13 07:12:29 crc kubenswrapper[4971]: I1213 07:12:29.552099 4971 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/ee3afc2f-79c6-444c-b6a3-2d4c475c3aac-scripts\") pod \"keystone-bootstrap-dn9h9\" (UID: \"ee3afc2f-79c6-444c-b6a3-2d4c475c3aac\") " pod="openstack/keystone-bootstrap-dn9h9" Dec 13 07:12:29 crc kubenswrapper[4971]: I1213 07:12:29.552123 4971 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-mdl24\" (UniqueName: \"kubernetes.io/projected/ee3afc2f-79c6-444c-b6a3-2d4c475c3aac-kube-api-access-mdl24\") pod \"keystone-bootstrap-dn9h9\" (UID: \"ee3afc2f-79c6-444c-b6a3-2d4c475c3aac\") " pod="openstack/keystone-bootstrap-dn9h9" Dec 13 07:12:29 crc kubenswrapper[4971]: I1213 07:12:29.552180 4971 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"credential-keys\" (UniqueName: \"kubernetes.io/secret/ee3afc2f-79c6-444c-b6a3-2d4c475c3aac-credential-keys\") pod \"keystone-bootstrap-dn9h9\" (UID: \"ee3afc2f-79c6-444c-b6a3-2d4c475c3aac\") " pod="openstack/keystone-bootstrap-dn9h9" Dec 13 07:12:29 crc kubenswrapper[4971]: I1213 07:12:29.552231 4971 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/ee3afc2f-79c6-444c-b6a3-2d4c475c3aac-combined-ca-bundle\") pod \"keystone-bootstrap-dn9h9\" (UID: \"ee3afc2f-79c6-444c-b6a3-2d4c475c3aac\") " pod="openstack/keystone-bootstrap-dn9h9" Dec 13 07:12:29 crc kubenswrapper[4971]: I1213 07:12:29.552260 4971 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/ee3afc2f-79c6-444c-b6a3-2d4c475c3aac-config-data\") pod \"keystone-bootstrap-dn9h9\" (UID: \"ee3afc2f-79c6-444c-b6a3-2d4c475c3aac\") " pod="openstack/keystone-bootstrap-dn9h9" Dec 13 07:12:29 crc kubenswrapper[4971]: I1213 07:12:29.653683 4971 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/ee3afc2f-79c6-444c-b6a3-2d4c475c3aac-fernet-keys\") pod \"keystone-bootstrap-dn9h9\" (UID: \"ee3afc2f-79c6-444c-b6a3-2d4c475c3aac\") " pod="openstack/keystone-bootstrap-dn9h9" Dec 13 07:12:29 crc kubenswrapper[4971]: I1213 07:12:29.653743 4971 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/ee3afc2f-79c6-444c-b6a3-2d4c475c3aac-scripts\") pod \"keystone-bootstrap-dn9h9\" (UID: \"ee3afc2f-79c6-444c-b6a3-2d4c475c3aac\") " pod="openstack/keystone-bootstrap-dn9h9" Dec 13 07:12:29 crc kubenswrapper[4971]: I1213 07:12:29.653783 4971 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-mdl24\" (UniqueName: \"kubernetes.io/projected/ee3afc2f-79c6-444c-b6a3-2d4c475c3aac-kube-api-access-mdl24\") pod \"keystone-bootstrap-dn9h9\" (UID: \"ee3afc2f-79c6-444c-b6a3-2d4c475c3aac\") " pod="openstack/keystone-bootstrap-dn9h9" Dec 13 07:12:29 crc kubenswrapper[4971]: I1213 07:12:29.653853 4971 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"credential-keys\" (UniqueName: \"kubernetes.io/secret/ee3afc2f-79c6-444c-b6a3-2d4c475c3aac-credential-keys\") pod \"keystone-bootstrap-dn9h9\" (UID: \"ee3afc2f-79c6-444c-b6a3-2d4c475c3aac\") " pod="openstack/keystone-bootstrap-dn9h9" Dec 13 07:12:29 crc kubenswrapper[4971]: I1213 07:12:29.653927 4971 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/ee3afc2f-79c6-444c-b6a3-2d4c475c3aac-combined-ca-bundle\") pod \"keystone-bootstrap-dn9h9\" (UID: \"ee3afc2f-79c6-444c-b6a3-2d4c475c3aac\") " pod="openstack/keystone-bootstrap-dn9h9" Dec 13 07:12:29 crc kubenswrapper[4971]: I1213 07:12:29.653959 4971 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/ee3afc2f-79c6-444c-b6a3-2d4c475c3aac-config-data\") pod \"keystone-bootstrap-dn9h9\" (UID: \"ee3afc2f-79c6-444c-b6a3-2d4c475c3aac\") " pod="openstack/keystone-bootstrap-dn9h9" Dec 13 07:12:29 crc kubenswrapper[4971]: I1213 07:12:29.659460 4971 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/ee3afc2f-79c6-444c-b6a3-2d4c475c3aac-config-data\") pod \"keystone-bootstrap-dn9h9\" (UID: \"ee3afc2f-79c6-444c-b6a3-2d4c475c3aac\") " pod="openstack/keystone-bootstrap-dn9h9" Dec 13 07:12:29 crc kubenswrapper[4971]: I1213 07:12:29.659465 4971 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"credential-keys\" (UniqueName: \"kubernetes.io/secret/ee3afc2f-79c6-444c-b6a3-2d4c475c3aac-credential-keys\") pod \"keystone-bootstrap-dn9h9\" (UID: \"ee3afc2f-79c6-444c-b6a3-2d4c475c3aac\") " pod="openstack/keystone-bootstrap-dn9h9" Dec 13 07:12:29 crc kubenswrapper[4971]: I1213 07:12:29.674426 4971 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/ee3afc2f-79c6-444c-b6a3-2d4c475c3aac-fernet-keys\") pod \"keystone-bootstrap-dn9h9\" (UID: \"ee3afc2f-79c6-444c-b6a3-2d4c475c3aac\") " pod="openstack/keystone-bootstrap-dn9h9" Dec 13 07:12:29 crc kubenswrapper[4971]: I1213 07:12:29.674808 4971 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/ee3afc2f-79c6-444c-b6a3-2d4c475c3aac-combined-ca-bundle\") pod \"keystone-bootstrap-dn9h9\" (UID: \"ee3afc2f-79c6-444c-b6a3-2d4c475c3aac\") " pod="openstack/keystone-bootstrap-dn9h9" Dec 13 07:12:29 crc kubenswrapper[4971]: I1213 07:12:29.677932 4971 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/ee3afc2f-79c6-444c-b6a3-2d4c475c3aac-scripts\") pod \"keystone-bootstrap-dn9h9\" (UID: \"ee3afc2f-79c6-444c-b6a3-2d4c475c3aac\") " pod="openstack/keystone-bootstrap-dn9h9" Dec 13 07:12:29 crc kubenswrapper[4971]: I1213 07:12:29.678568 4971 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-mdl24\" (UniqueName: \"kubernetes.io/projected/ee3afc2f-79c6-444c-b6a3-2d4c475c3aac-kube-api-access-mdl24\") pod \"keystone-bootstrap-dn9h9\" (UID: \"ee3afc2f-79c6-444c-b6a3-2d4c475c3aac\") " pod="openstack/keystone-bootstrap-dn9h9" Dec 13 07:12:29 crc kubenswrapper[4971]: I1213 07:12:29.752604 4971 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-bootstrap-dn9h9" Dec 13 07:12:29 crc kubenswrapper[4971]: I1213 07:12:29.783095 4971 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="5b9e78cc-aa6e-4369-ab8a-5e89b86359c1" path="/var/lib/kubelet/pods/5b9e78cc-aa6e-4369-ab8a-5e89b86359c1/volumes" Dec 13 07:12:35 crc kubenswrapper[4971]: I1213 07:12:35.579288 4971 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/dnsmasq-dns-57768dd7b5-86hlw" podUID="c9e30081-d603-4c9d-81ce-7bcff6848243" containerName="dnsmasq-dns" probeResult="failure" output="dial tcp 10.217.0.121:5353: i/o timeout" Dec 13 07:12:35 crc kubenswrapper[4971]: I1213 07:12:35.580369 4971 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/dnsmasq-dns-57768dd7b5-86hlw" Dec 13 07:12:36 crc kubenswrapper[4971]: I1213 07:12:36.275976 4971 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-57768dd7b5-86hlw" Dec 13 07:12:36 crc kubenswrapper[4971]: I1213 07:12:36.383690 4971 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/c9e30081-d603-4c9d-81ce-7bcff6848243-ovsdbserver-nb\") pod \"c9e30081-d603-4c9d-81ce-7bcff6848243\" (UID: \"c9e30081-d603-4c9d-81ce-7bcff6848243\") " Dec 13 07:12:36 crc kubenswrapper[4971]: I1213 07:12:36.383781 4971 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/c9e30081-d603-4c9d-81ce-7bcff6848243-ovsdbserver-sb\") pod \"c9e30081-d603-4c9d-81ce-7bcff6848243\" (UID: \"c9e30081-d603-4c9d-81ce-7bcff6848243\") " Dec 13 07:12:36 crc kubenswrapper[4971]: I1213 07:12:36.383891 4971 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-ccks5\" (UniqueName: \"kubernetes.io/projected/c9e30081-d603-4c9d-81ce-7bcff6848243-kube-api-access-ccks5\") pod \"c9e30081-d603-4c9d-81ce-7bcff6848243\" (UID: \"c9e30081-d603-4c9d-81ce-7bcff6848243\") " Dec 13 07:12:36 crc kubenswrapper[4971]: I1213 07:12:36.384063 4971 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/c9e30081-d603-4c9d-81ce-7bcff6848243-config\") pod \"c9e30081-d603-4c9d-81ce-7bcff6848243\" (UID: \"c9e30081-d603-4c9d-81ce-7bcff6848243\") " Dec 13 07:12:36 crc kubenswrapper[4971]: I1213 07:12:36.384119 4971 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/c9e30081-d603-4c9d-81ce-7bcff6848243-dns-svc\") pod \"c9e30081-d603-4c9d-81ce-7bcff6848243\" (UID: \"c9e30081-d603-4c9d-81ce-7bcff6848243\") " Dec 13 07:12:36 crc kubenswrapper[4971]: I1213 07:12:36.405153 4971 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/c9e30081-d603-4c9d-81ce-7bcff6848243-kube-api-access-ccks5" (OuterVolumeSpecName: "kube-api-access-ccks5") pod "c9e30081-d603-4c9d-81ce-7bcff6848243" (UID: "c9e30081-d603-4c9d-81ce-7bcff6848243"). InnerVolumeSpecName "kube-api-access-ccks5". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 13 07:12:36 crc kubenswrapper[4971]: I1213 07:12:36.426720 4971 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/c9e30081-d603-4c9d-81ce-7bcff6848243-config" (OuterVolumeSpecName: "config") pod "c9e30081-d603-4c9d-81ce-7bcff6848243" (UID: "c9e30081-d603-4c9d-81ce-7bcff6848243"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 13 07:12:36 crc kubenswrapper[4971]: I1213 07:12:36.429926 4971 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/c9e30081-d603-4c9d-81ce-7bcff6848243-ovsdbserver-nb" (OuterVolumeSpecName: "ovsdbserver-nb") pod "c9e30081-d603-4c9d-81ce-7bcff6848243" (UID: "c9e30081-d603-4c9d-81ce-7bcff6848243"). InnerVolumeSpecName "ovsdbserver-nb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 13 07:12:36 crc kubenswrapper[4971]: I1213 07:12:36.430229 4971 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/c9e30081-d603-4c9d-81ce-7bcff6848243-ovsdbserver-sb" (OuterVolumeSpecName: "ovsdbserver-sb") pod "c9e30081-d603-4c9d-81ce-7bcff6848243" (UID: "c9e30081-d603-4c9d-81ce-7bcff6848243"). InnerVolumeSpecName "ovsdbserver-sb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 13 07:12:36 crc kubenswrapper[4971]: I1213 07:12:36.443715 4971 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/c9e30081-d603-4c9d-81ce-7bcff6848243-dns-svc" (OuterVolumeSpecName: "dns-svc") pod "c9e30081-d603-4c9d-81ce-7bcff6848243" (UID: "c9e30081-d603-4c9d-81ce-7bcff6848243"). InnerVolumeSpecName "dns-svc". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 13 07:12:36 crc kubenswrapper[4971]: I1213 07:12:36.486169 4971 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-ccks5\" (UniqueName: \"kubernetes.io/projected/c9e30081-d603-4c9d-81ce-7bcff6848243-kube-api-access-ccks5\") on node \"crc\" DevicePath \"\"" Dec 13 07:12:36 crc kubenswrapper[4971]: I1213 07:12:36.486216 4971 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/c9e30081-d603-4c9d-81ce-7bcff6848243-config\") on node \"crc\" DevicePath \"\"" Dec 13 07:12:36 crc kubenswrapper[4971]: I1213 07:12:36.486230 4971 reconciler_common.go:293] "Volume detached for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/c9e30081-d603-4c9d-81ce-7bcff6848243-dns-svc\") on node \"crc\" DevicePath \"\"" Dec 13 07:12:36 crc kubenswrapper[4971]: I1213 07:12:36.486246 4971 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/c9e30081-d603-4c9d-81ce-7bcff6848243-ovsdbserver-nb\") on node \"crc\" DevicePath \"\"" Dec 13 07:12:36 crc kubenswrapper[4971]: I1213 07:12:36.486257 4971 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/c9e30081-d603-4c9d-81ce-7bcff6848243-ovsdbserver-sb\") on node \"crc\" DevicePath \"\"" Dec 13 07:12:36 crc kubenswrapper[4971]: I1213 07:12:36.759570 4971 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-57768dd7b5-86hlw" event={"ID":"c9e30081-d603-4c9d-81ce-7bcff6848243","Type":"ContainerDied","Data":"93dcd74806ab03a60c4e4da3b8b953c968fc6cc7e8ce91566c8874b613715598"} Dec 13 07:12:36 crc kubenswrapper[4971]: I1213 07:12:36.759638 4971 scope.go:117] "RemoveContainer" containerID="303ee14cece58dbf062bf3b1d7ce80d4f37ced4fdb8e6c483677f2b786064dd4" Dec 13 07:12:36 crc kubenswrapper[4971]: I1213 07:12:36.759667 4971 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-57768dd7b5-86hlw" Dec 13 07:12:36 crc kubenswrapper[4971]: E1213 07:12:36.763864 4971 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying config: context canceled" image="quay.io/podified-antelope-centos9/openstack-barbican-api@sha256:fe32d3ea620f0c7ecfdde9bbf28417fde03bc18c6f60b1408fa8da24d8188f16" Dec 13 07:12:36 crc kubenswrapper[4971]: E1213 07:12:36.764056 4971 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:barbican-db-sync,Image:quay.io/podified-antelope-centos9/openstack-barbican-api@sha256:fe32d3ea620f0c7ecfdde9bbf28417fde03bc18c6f60b1408fa8da24d8188f16,Command:[/bin/bash],Args:[-c barbican-manage db upgrade],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:KOLLA_BOOTSTRAP,Value:TRUE,ValueFrom:nil,},EnvVar{Name:KOLLA_CONFIG_STRATEGY,Value:COPY_ALWAYS,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:db-sync-config-data,ReadOnly:true,MountPath:/etc/barbican/barbican.conf.d,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:combined-ca-bundle,ReadOnly:true,MountPath:/etc/pki/ca-trust/extracted/pem/tls-ca-bundle.pem,SubPath:tls-ca-bundle.pem,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-w84mh,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[MKNOD],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*42403,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:nil,RunAsGroup:*42403,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod barbican-db-sync-dvchm_openstack(d7cf1b55-0dae-4a17-857a-6336fe71ff24): ErrImagePull: rpc error: code = Canceled desc = copying config: context canceled" logger="UnhandledError" Dec 13 07:12:36 crc kubenswrapper[4971]: I1213 07:12:36.764298 4971 generic.go:334] "Generic (PLEG): container finished" podID="0813350d-430f-4564-a09a-062a3239ba2f" containerID="79444a316a054f8a31c3874025055b3501be54e3988c3b3a02890b8643e62d72" exitCode=0 Dec 13 07:12:36 crc kubenswrapper[4971]: I1213 07:12:36.764340 4971 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-db-sync-8z2nm" event={"ID":"0813350d-430f-4564-a09a-062a3239ba2f","Type":"ContainerDied","Data":"79444a316a054f8a31c3874025055b3501be54e3988c3b3a02890b8643e62d72"} Dec 13 07:12:36 crc kubenswrapper[4971]: E1213 07:12:36.765137 4971 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"barbican-db-sync\" with ErrImagePull: \"rpc error: code = Canceled desc = copying config: context canceled\"" pod="openstack/barbican-db-sync-dvchm" podUID="d7cf1b55-0dae-4a17-857a-6336fe71ff24" Dec 13 07:12:36 crc kubenswrapper[4971]: I1213 07:12:36.769094 4971 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/horizon-bcdd7767-j7d7f" Dec 13 07:12:36 crc kubenswrapper[4971]: I1213 07:12:36.853410 4971 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/horizon-fbfd486df-xmvls" Dec 13 07:12:36 crc kubenswrapper[4971]: I1213 07:12:36.866322 4971 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/horizon-77c4585f77-kkx2c" Dec 13 07:12:36 crc kubenswrapper[4971]: I1213 07:12:36.873470 4971 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-57768dd7b5-86hlw"] Dec 13 07:12:36 crc kubenswrapper[4971]: I1213 07:12:36.882236 4971 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-57768dd7b5-86hlw"] Dec 13 07:12:36 crc kubenswrapper[4971]: I1213 07:12:36.891769 4971 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/ff35e760-13e7-4145-9970-742b73a36176-scripts\") pod \"ff35e760-13e7-4145-9970-742b73a36176\" (UID: \"ff35e760-13e7-4145-9970-742b73a36176\") " Dec 13 07:12:36 crc kubenswrapper[4971]: I1213 07:12:36.891870 4971 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/ff35e760-13e7-4145-9970-742b73a36176-logs\") pod \"ff35e760-13e7-4145-9970-742b73a36176\" (UID: \"ff35e760-13e7-4145-9970-742b73a36176\") " Dec 13 07:12:36 crc kubenswrapper[4971]: I1213 07:12:36.892039 4971 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/ff35e760-13e7-4145-9970-742b73a36176-config-data\") pod \"ff35e760-13e7-4145-9970-742b73a36176\" (UID: \"ff35e760-13e7-4145-9970-742b73a36176\") " Dec 13 07:12:36 crc kubenswrapper[4971]: I1213 07:12:36.892115 4971 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-cpxzk\" (UniqueName: \"kubernetes.io/projected/ff35e760-13e7-4145-9970-742b73a36176-kube-api-access-cpxzk\") pod \"ff35e760-13e7-4145-9970-742b73a36176\" (UID: \"ff35e760-13e7-4145-9970-742b73a36176\") " Dec 13 07:12:36 crc kubenswrapper[4971]: I1213 07:12:36.892165 4971 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"horizon-secret-key\" (UniqueName: \"kubernetes.io/secret/ff35e760-13e7-4145-9970-742b73a36176-horizon-secret-key\") pod \"ff35e760-13e7-4145-9970-742b73a36176\" (UID: \"ff35e760-13e7-4145-9970-742b73a36176\") " Dec 13 07:12:36 crc kubenswrapper[4971]: I1213 07:12:36.892321 4971 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/ff35e760-13e7-4145-9970-742b73a36176-logs" (OuterVolumeSpecName: "logs") pod "ff35e760-13e7-4145-9970-742b73a36176" (UID: "ff35e760-13e7-4145-9970-742b73a36176"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 13 07:12:36 crc kubenswrapper[4971]: I1213 07:12:36.892695 4971 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/ff35e760-13e7-4145-9970-742b73a36176-logs\") on node \"crc\" DevicePath \"\"" Dec 13 07:12:36 crc kubenswrapper[4971]: I1213 07:12:36.893101 4971 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/ff35e760-13e7-4145-9970-742b73a36176-scripts" (OuterVolumeSpecName: "scripts") pod "ff35e760-13e7-4145-9970-742b73a36176" (UID: "ff35e760-13e7-4145-9970-742b73a36176"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 13 07:12:36 crc kubenswrapper[4971]: I1213 07:12:36.893455 4971 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/ff35e760-13e7-4145-9970-742b73a36176-config-data" (OuterVolumeSpecName: "config-data") pod "ff35e760-13e7-4145-9970-742b73a36176" (UID: "ff35e760-13e7-4145-9970-742b73a36176"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 13 07:12:36 crc kubenswrapper[4971]: I1213 07:12:36.896868 4971 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/ff35e760-13e7-4145-9970-742b73a36176-kube-api-access-cpxzk" (OuterVolumeSpecName: "kube-api-access-cpxzk") pod "ff35e760-13e7-4145-9970-742b73a36176" (UID: "ff35e760-13e7-4145-9970-742b73a36176"). InnerVolumeSpecName "kube-api-access-cpxzk". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 13 07:12:36 crc kubenswrapper[4971]: I1213 07:12:36.897508 4971 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/ff35e760-13e7-4145-9970-742b73a36176-horizon-secret-key" (OuterVolumeSpecName: "horizon-secret-key") pod "ff35e760-13e7-4145-9970-742b73a36176" (UID: "ff35e760-13e7-4145-9970-742b73a36176"). InnerVolumeSpecName "horizon-secret-key". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 13 07:12:36 crc kubenswrapper[4971]: I1213 07:12:36.993993 4971 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"horizon-secret-key\" (UniqueName: \"kubernetes.io/secret/f9bd96be-6d7f-415e-b67c-5bd02b05c536-horizon-secret-key\") pod \"f9bd96be-6d7f-415e-b67c-5bd02b05c536\" (UID: \"f9bd96be-6d7f-415e-b67c-5bd02b05c536\") " Dec 13 07:12:36 crc kubenswrapper[4971]: I1213 07:12:36.994118 4971 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-kjwdg\" (UniqueName: \"kubernetes.io/projected/f9bd96be-6d7f-415e-b67c-5bd02b05c536-kube-api-access-kjwdg\") pod \"f9bd96be-6d7f-415e-b67c-5bd02b05c536\" (UID: \"f9bd96be-6d7f-415e-b67c-5bd02b05c536\") " Dec 13 07:12:36 crc kubenswrapper[4971]: I1213 07:12:36.994159 4971 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/f9bd96be-6d7f-415e-b67c-5bd02b05c536-config-data\") pod \"f9bd96be-6d7f-415e-b67c-5bd02b05c536\" (UID: \"f9bd96be-6d7f-415e-b67c-5bd02b05c536\") " Dec 13 07:12:36 crc kubenswrapper[4971]: I1213 07:12:36.994197 4971 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/b314f107-1cb6-49f3-aa51-3f607ca79f6b-config-data\") pod \"b314f107-1cb6-49f3-aa51-3f607ca79f6b\" (UID: \"b314f107-1cb6-49f3-aa51-3f607ca79f6b\") " Dec 13 07:12:36 crc kubenswrapper[4971]: I1213 07:12:36.994247 4971 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/f9bd96be-6d7f-415e-b67c-5bd02b05c536-scripts\") pod \"f9bd96be-6d7f-415e-b67c-5bd02b05c536\" (UID: \"f9bd96be-6d7f-415e-b67c-5bd02b05c536\") " Dec 13 07:12:36 crc kubenswrapper[4971]: I1213 07:12:36.994304 4971 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/f9bd96be-6d7f-415e-b67c-5bd02b05c536-logs\") pod \"f9bd96be-6d7f-415e-b67c-5bd02b05c536\" (UID: \"f9bd96be-6d7f-415e-b67c-5bd02b05c536\") " Dec 13 07:12:36 crc kubenswrapper[4971]: I1213 07:12:36.994811 4971 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/f9bd96be-6d7f-415e-b67c-5bd02b05c536-logs" (OuterVolumeSpecName: "logs") pod "f9bd96be-6d7f-415e-b67c-5bd02b05c536" (UID: "f9bd96be-6d7f-415e-b67c-5bd02b05c536"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 13 07:12:36 crc kubenswrapper[4971]: I1213 07:12:36.994856 4971 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/f9bd96be-6d7f-415e-b67c-5bd02b05c536-scripts" (OuterVolumeSpecName: "scripts") pod "f9bd96be-6d7f-415e-b67c-5bd02b05c536" (UID: "f9bd96be-6d7f-415e-b67c-5bd02b05c536"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 13 07:12:36 crc kubenswrapper[4971]: I1213 07:12:36.995067 4971 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/b314f107-1cb6-49f3-aa51-3f607ca79f6b-logs\") pod \"b314f107-1cb6-49f3-aa51-3f607ca79f6b\" (UID: \"b314f107-1cb6-49f3-aa51-3f607ca79f6b\") " Dec 13 07:12:36 crc kubenswrapper[4971]: I1213 07:12:36.995133 4971 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"horizon-secret-key\" (UniqueName: \"kubernetes.io/secret/b314f107-1cb6-49f3-aa51-3f607ca79f6b-horizon-secret-key\") pod \"b314f107-1cb6-49f3-aa51-3f607ca79f6b\" (UID: \"b314f107-1cb6-49f3-aa51-3f607ca79f6b\") " Dec 13 07:12:36 crc kubenswrapper[4971]: I1213 07:12:36.995061 4971 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/b314f107-1cb6-49f3-aa51-3f607ca79f6b-config-data" (OuterVolumeSpecName: "config-data") pod "b314f107-1cb6-49f3-aa51-3f607ca79f6b" (UID: "b314f107-1cb6-49f3-aa51-3f607ca79f6b"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 13 07:12:36 crc kubenswrapper[4971]: I1213 07:12:36.995157 4971 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/b314f107-1cb6-49f3-aa51-3f607ca79f6b-scripts\") pod \"b314f107-1cb6-49f3-aa51-3f607ca79f6b\" (UID: \"b314f107-1cb6-49f3-aa51-3f607ca79f6b\") " Dec 13 07:12:36 crc kubenswrapper[4971]: I1213 07:12:36.995443 4971 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/b314f107-1cb6-49f3-aa51-3f607ca79f6b-logs" (OuterVolumeSpecName: "logs") pod "b314f107-1cb6-49f3-aa51-3f607ca79f6b" (UID: "b314f107-1cb6-49f3-aa51-3f607ca79f6b"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 13 07:12:36 crc kubenswrapper[4971]: I1213 07:12:36.995466 4971 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-tghr9\" (UniqueName: \"kubernetes.io/projected/b314f107-1cb6-49f3-aa51-3f607ca79f6b-kube-api-access-tghr9\") pod \"b314f107-1cb6-49f3-aa51-3f607ca79f6b\" (UID: \"b314f107-1cb6-49f3-aa51-3f607ca79f6b\") " Dec 13 07:12:36 crc kubenswrapper[4971]: I1213 07:12:36.995817 4971 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/b314f107-1cb6-49f3-aa51-3f607ca79f6b-scripts" (OuterVolumeSpecName: "scripts") pod "b314f107-1cb6-49f3-aa51-3f607ca79f6b" (UID: "b314f107-1cb6-49f3-aa51-3f607ca79f6b"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 13 07:12:36 crc kubenswrapper[4971]: I1213 07:12:36.996000 4971 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/f9bd96be-6d7f-415e-b67c-5bd02b05c536-config-data" (OuterVolumeSpecName: "config-data") pod "f9bd96be-6d7f-415e-b67c-5bd02b05c536" (UID: "f9bd96be-6d7f-415e-b67c-5bd02b05c536"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 13 07:12:36 crc kubenswrapper[4971]: I1213 07:12:36.996062 4971 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/f9bd96be-6d7f-415e-b67c-5bd02b05c536-scripts\") on node \"crc\" DevicePath \"\"" Dec 13 07:12:36 crc kubenswrapper[4971]: I1213 07:12:36.996076 4971 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/f9bd96be-6d7f-415e-b67c-5bd02b05c536-logs\") on node \"crc\" DevicePath \"\"" Dec 13 07:12:36 crc kubenswrapper[4971]: I1213 07:12:36.996087 4971 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/b314f107-1cb6-49f3-aa51-3f607ca79f6b-logs\") on node \"crc\" DevicePath \"\"" Dec 13 07:12:36 crc kubenswrapper[4971]: I1213 07:12:36.996096 4971 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/ff35e760-13e7-4145-9970-742b73a36176-config-data\") on node \"crc\" DevicePath \"\"" Dec 13 07:12:36 crc kubenswrapper[4971]: I1213 07:12:36.996105 4971 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/b314f107-1cb6-49f3-aa51-3f607ca79f6b-scripts\") on node \"crc\" DevicePath \"\"" Dec 13 07:12:36 crc kubenswrapper[4971]: I1213 07:12:36.996115 4971 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-cpxzk\" (UniqueName: \"kubernetes.io/projected/ff35e760-13e7-4145-9970-742b73a36176-kube-api-access-cpxzk\") on node \"crc\" DevicePath \"\"" Dec 13 07:12:36 crc kubenswrapper[4971]: I1213 07:12:36.996124 4971 reconciler_common.go:293] "Volume detached for volume \"horizon-secret-key\" (UniqueName: \"kubernetes.io/secret/ff35e760-13e7-4145-9970-742b73a36176-horizon-secret-key\") on node \"crc\" DevicePath \"\"" Dec 13 07:12:36 crc kubenswrapper[4971]: I1213 07:12:36.996132 4971 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/ff35e760-13e7-4145-9970-742b73a36176-scripts\") on node \"crc\" DevicePath \"\"" Dec 13 07:12:36 crc kubenswrapper[4971]: I1213 07:12:36.996140 4971 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/b314f107-1cb6-49f3-aa51-3f607ca79f6b-config-data\") on node \"crc\" DevicePath \"\"" Dec 13 07:12:36 crc kubenswrapper[4971]: I1213 07:12:36.996967 4971 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/f9bd96be-6d7f-415e-b67c-5bd02b05c536-horizon-secret-key" (OuterVolumeSpecName: "horizon-secret-key") pod "f9bd96be-6d7f-415e-b67c-5bd02b05c536" (UID: "f9bd96be-6d7f-415e-b67c-5bd02b05c536"). InnerVolumeSpecName "horizon-secret-key". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 13 07:12:36 crc kubenswrapper[4971]: I1213 07:12:36.997280 4971 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/b314f107-1cb6-49f3-aa51-3f607ca79f6b-horizon-secret-key" (OuterVolumeSpecName: "horizon-secret-key") pod "b314f107-1cb6-49f3-aa51-3f607ca79f6b" (UID: "b314f107-1cb6-49f3-aa51-3f607ca79f6b"). InnerVolumeSpecName "horizon-secret-key". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 13 07:12:36 crc kubenswrapper[4971]: I1213 07:12:36.998757 4971 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/b314f107-1cb6-49f3-aa51-3f607ca79f6b-kube-api-access-tghr9" (OuterVolumeSpecName: "kube-api-access-tghr9") pod "b314f107-1cb6-49f3-aa51-3f607ca79f6b" (UID: "b314f107-1cb6-49f3-aa51-3f607ca79f6b"). InnerVolumeSpecName "kube-api-access-tghr9". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 13 07:12:36 crc kubenswrapper[4971]: I1213 07:12:36.999825 4971 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/f9bd96be-6d7f-415e-b67c-5bd02b05c536-kube-api-access-kjwdg" (OuterVolumeSpecName: "kube-api-access-kjwdg") pod "f9bd96be-6d7f-415e-b67c-5bd02b05c536" (UID: "f9bd96be-6d7f-415e-b67c-5bd02b05c536"). InnerVolumeSpecName "kube-api-access-kjwdg". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 13 07:12:37 crc kubenswrapper[4971]: I1213 07:12:37.098386 4971 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-kjwdg\" (UniqueName: \"kubernetes.io/projected/f9bd96be-6d7f-415e-b67c-5bd02b05c536-kube-api-access-kjwdg\") on node \"crc\" DevicePath \"\"" Dec 13 07:12:37 crc kubenswrapper[4971]: I1213 07:12:37.098427 4971 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/f9bd96be-6d7f-415e-b67c-5bd02b05c536-config-data\") on node \"crc\" DevicePath \"\"" Dec 13 07:12:37 crc kubenswrapper[4971]: I1213 07:12:37.098447 4971 reconciler_common.go:293] "Volume detached for volume \"horizon-secret-key\" (UniqueName: \"kubernetes.io/secret/b314f107-1cb6-49f3-aa51-3f607ca79f6b-horizon-secret-key\") on node \"crc\" DevicePath \"\"" Dec 13 07:12:37 crc kubenswrapper[4971]: I1213 07:12:37.098459 4971 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-tghr9\" (UniqueName: \"kubernetes.io/projected/b314f107-1cb6-49f3-aa51-3f607ca79f6b-kube-api-access-tghr9\") on node \"crc\" DevicePath \"\"" Dec 13 07:12:37 crc kubenswrapper[4971]: I1213 07:12:37.098471 4971 reconciler_common.go:293] "Volume detached for volume \"horizon-secret-key\" (UniqueName: \"kubernetes.io/secret/f9bd96be-6d7f-415e-b67c-5bd02b05c536-horizon-secret-key\") on node \"crc\" DevicePath \"\"" Dec 13 07:12:37 crc kubenswrapper[4971]: I1213 07:12:37.785092 4971 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/horizon-fbfd486df-xmvls" Dec 13 07:12:37 crc kubenswrapper[4971]: I1213 07:12:37.786568 4971 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/horizon-77c4585f77-kkx2c" Dec 13 07:12:37 crc kubenswrapper[4971]: I1213 07:12:37.790974 4971 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/horizon-bcdd7767-j7d7f" Dec 13 07:12:37 crc kubenswrapper[4971]: I1213 07:12:37.791618 4971 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="c9e30081-d603-4c9d-81ce-7bcff6848243" path="/var/lib/kubelet/pods/c9e30081-d603-4c9d-81ce-7bcff6848243/volumes" Dec 13 07:12:37 crc kubenswrapper[4971]: I1213 07:12:37.797337 4971 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/horizon-fbfd486df-xmvls" event={"ID":"f9bd96be-6d7f-415e-b67c-5bd02b05c536","Type":"ContainerDied","Data":"729d1ba01c5742e471d188d542eb28bae9f7b38d079a5cc82bda8adc7c17ada6"} Dec 13 07:12:37 crc kubenswrapper[4971]: I1213 07:12:37.797387 4971 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/horizon-77c4585f77-kkx2c" event={"ID":"b314f107-1cb6-49f3-aa51-3f607ca79f6b","Type":"ContainerDied","Data":"8daaeddcce34e5e2f6d2ef32ac8d2a6c22737e4346ddf888ec4a8f83769d6f9d"} Dec 13 07:12:37 crc kubenswrapper[4971]: I1213 07:12:37.797409 4971 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/horizon-bcdd7767-j7d7f" event={"ID":"ff35e760-13e7-4145-9970-742b73a36176","Type":"ContainerDied","Data":"46b26a551dc80a0e0a26e6ba29b9a6a2b3c363d24f8e1ca63552331408259a59"} Dec 13 07:12:37 crc kubenswrapper[4971]: E1213 07:12:37.800392 4971 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"barbican-db-sync\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/podified-antelope-centos9/openstack-barbican-api@sha256:fe32d3ea620f0c7ecfdde9bbf28417fde03bc18c6f60b1408fa8da24d8188f16\\\"\"" pod="openstack/barbican-db-sync-dvchm" podUID="d7cf1b55-0dae-4a17-857a-6336fe71ff24" Dec 13 07:12:37 crc kubenswrapper[4971]: I1213 07:12:37.845749 4971 scope.go:117] "RemoveContainer" containerID="c707864d3136ddbe0a0fb0f1eb6f1a2bdcc6c2f0651b9e51d1d68857e280fb88" Dec 13 07:12:37 crc kubenswrapper[4971]: E1213 07:12:37.874614 4971 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying config: context canceled" image="quay.io/podified-antelope-centos9/openstack-cinder-api@sha256:b59b7445e581cc720038107e421371c86c5765b2967e77d884ef29b1d9fd0f49" Dec 13 07:12:37 crc kubenswrapper[4971]: E1213 07:12:37.874796 4971 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:cinder-db-sync,Image:quay.io/podified-antelope-centos9/openstack-cinder-api@sha256:b59b7445e581cc720038107e421371c86c5765b2967e77d884ef29b1d9fd0f49,Command:[/bin/bash],Args:[-c /usr/local/bin/kolla_set_configs && /usr/local/bin/kolla_start],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:KOLLA_BOOTSTRAP,Value:TRUE,ValueFrom:nil,},EnvVar{Name:KOLLA_CONFIG_STRATEGY,Value:COPY_ALWAYS,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:etc-machine-id,ReadOnly:true,MountPath:/etc/machine-id,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:scripts,ReadOnly:true,MountPath:/usr/local/bin/container-scripts,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:config-data,ReadOnly:true,MountPath:/var/lib/config-data/merged,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:config-data,ReadOnly:true,MountPath:/etc/my.cnf,SubPath:my.cnf,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:db-sync-config-data,ReadOnly:true,MountPath:/etc/cinder/cinder.conf.d,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:config-data,ReadOnly:true,MountPath:/var/lib/kolla/config_files/config.json,SubPath:db-sync-config.json,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:combined-ca-bundle,ReadOnly:true,MountPath:/etc/pki/ca-trust/extracted/pem/tls-ca-bundle.pem,SubPath:tls-ca-bundle.pem,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-r4lhp,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:nil,Privileged:nil,SELinuxOptions:nil,RunAsUser:*0,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:nil,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod cinder-db-sync-26k47_openstack(c1d819d0-339f-4e7e-9180-63d2f7c0042e): ErrImagePull: rpc error: code = Canceled desc = copying config: context canceled" logger="UnhandledError" Dec 13 07:12:37 crc kubenswrapper[4971]: E1213 07:12:37.876245 4971 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"cinder-db-sync\" with ErrImagePull: \"rpc error: code = Canceled desc = copying config: context canceled\"" pod="openstack/cinder-db-sync-26k47" podUID="c1d819d0-339f-4e7e-9180-63d2f7c0042e" Dec 13 07:12:38 crc kubenswrapper[4971]: I1213 07:12:38.142606 4971 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/horizon-bcdd7767-j7d7f"] Dec 13 07:12:38 crc kubenswrapper[4971]: I1213 07:12:38.153682 4971 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/horizon-bcdd7767-j7d7f"] Dec 13 07:12:38 crc kubenswrapper[4971]: I1213 07:12:38.168029 4971 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/horizon-77c4585f77-kkx2c"] Dec 13 07:12:38 crc kubenswrapper[4971]: I1213 07:12:38.174231 4971 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/horizon-77c4585f77-kkx2c"] Dec 13 07:12:38 crc kubenswrapper[4971]: I1213 07:12:38.333702 4971 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-db-sync-8z2nm" Dec 13 07:12:38 crc kubenswrapper[4971]: I1213 07:12:38.463100 4971 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/0813350d-430f-4564-a09a-062a3239ba2f-combined-ca-bundle\") pod \"0813350d-430f-4564-a09a-062a3239ba2f\" (UID: \"0813350d-430f-4564-a09a-062a3239ba2f\") " Dec 13 07:12:38 crc kubenswrapper[4971]: I1213 07:12:38.463235 4971 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/secret/0813350d-430f-4564-a09a-062a3239ba2f-config\") pod \"0813350d-430f-4564-a09a-062a3239ba2f\" (UID: \"0813350d-430f-4564-a09a-062a3239ba2f\") " Dec 13 07:12:38 crc kubenswrapper[4971]: I1213 07:12:38.463293 4971 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-ktmdv\" (UniqueName: \"kubernetes.io/projected/0813350d-430f-4564-a09a-062a3239ba2f-kube-api-access-ktmdv\") pod \"0813350d-430f-4564-a09a-062a3239ba2f\" (UID: \"0813350d-430f-4564-a09a-062a3239ba2f\") " Dec 13 07:12:38 crc kubenswrapper[4971]: I1213 07:12:38.505943 4971 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/0813350d-430f-4564-a09a-062a3239ba2f-kube-api-access-ktmdv" (OuterVolumeSpecName: "kube-api-access-ktmdv") pod "0813350d-430f-4564-a09a-062a3239ba2f" (UID: "0813350d-430f-4564-a09a-062a3239ba2f"). InnerVolumeSpecName "kube-api-access-ktmdv". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 13 07:12:38 crc kubenswrapper[4971]: I1213 07:12:38.510966 4971 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/0813350d-430f-4564-a09a-062a3239ba2f-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "0813350d-430f-4564-a09a-062a3239ba2f" (UID: "0813350d-430f-4564-a09a-062a3239ba2f"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 13 07:12:38 crc kubenswrapper[4971]: I1213 07:12:38.523777 4971 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/horizon-8699dc887d-sl2h6"] Dec 13 07:12:38 crc kubenswrapper[4971]: I1213 07:12:38.531594 4971 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/0813350d-430f-4564-a09a-062a3239ba2f-config" (OuterVolumeSpecName: "config") pod "0813350d-430f-4564-a09a-062a3239ba2f" (UID: "0813350d-430f-4564-a09a-062a3239ba2f"). InnerVolumeSpecName "config". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 13 07:12:38 crc kubenswrapper[4971]: I1213 07:12:38.566149 4971 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/0813350d-430f-4564-a09a-062a3239ba2f-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 13 07:12:38 crc kubenswrapper[4971]: I1213 07:12:38.566191 4971 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/secret/0813350d-430f-4564-a09a-062a3239ba2f-config\") on node \"crc\" DevicePath \"\"" Dec 13 07:12:38 crc kubenswrapper[4971]: I1213 07:12:38.566203 4971 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-ktmdv\" (UniqueName: \"kubernetes.io/projected/0813350d-430f-4564-a09a-062a3239ba2f-kube-api-access-ktmdv\") on node \"crc\" DevicePath \"\"" Dec 13 07:12:38 crc kubenswrapper[4971]: I1213 07:12:38.576138 4971 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/keystone-bootstrap-dn9h9"] Dec 13 07:12:38 crc kubenswrapper[4971]: W1213 07:12:38.578600 4971 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod9a4736be_5537_4638_a203_d36db1ae35cc.slice/crio-90a7ac52aff84d312a6639f807172d8d14d9c2b632657f90fe22dcd2b08e245b WatchSource:0}: Error finding container 90a7ac52aff84d312a6639f807172d8d14d9c2b632657f90fe22dcd2b08e245b: Status 404 returned error can't find the container with id 90a7ac52aff84d312a6639f807172d8d14d9c2b632657f90fe22dcd2b08e245b Dec 13 07:12:38 crc kubenswrapper[4971]: I1213 07:12:38.586862 4971 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/horizon-797dfcfc58-wlch7"] Dec 13 07:12:38 crc kubenswrapper[4971]: I1213 07:12:38.799017 4971 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/horizon-8699dc887d-sl2h6" event={"ID":"772eedab-7c62-454b-b722-ee453b39e89f","Type":"ContainerStarted","Data":"e1c30d274e8d692e18e387041e360c7599ed69b5331a9256cef5045f972a15e3"} Dec 13 07:12:38 crc kubenswrapper[4971]: I1213 07:12:38.804831 4971 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-db-sync-8z2nm" Dec 13 07:12:38 crc kubenswrapper[4971]: I1213 07:12:38.804972 4971 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-db-sync-8z2nm" event={"ID":"0813350d-430f-4564-a09a-062a3239ba2f","Type":"ContainerDied","Data":"f8b36cdbe233f6b61c0a47a1a93c6169e7f2fae7dbfe2b7a99ce9ae4fd061d40"} Dec 13 07:12:38 crc kubenswrapper[4971]: I1213 07:12:38.805081 4971 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="f8b36cdbe233f6b61c0a47a1a93c6169e7f2fae7dbfe2b7a99ce9ae4fd061d40" Dec 13 07:12:38 crc kubenswrapper[4971]: I1213 07:12:38.806994 4971 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-bootstrap-dn9h9" event={"ID":"ee3afc2f-79c6-444c-b6a3-2d4c475c3aac","Type":"ContainerStarted","Data":"9d3fdcdb191f5a67a4fa2a8f1e92401ee6d44779f8a7c3fead7ebd96791a001b"} Dec 13 07:12:38 crc kubenswrapper[4971]: I1213 07:12:38.807075 4971 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-bootstrap-dn9h9" event={"ID":"ee3afc2f-79c6-444c-b6a3-2d4c475c3aac","Type":"ContainerStarted","Data":"ac05c4b6e27e476542a19f5f67b2b66dacc449e38088c1b8d13ff803b99435e7"} Dec 13 07:12:38 crc kubenswrapper[4971]: I1213 07:12:38.809241 4971 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"56e56ce1-4dd1-4ae5-b868-e98d27d5ad94","Type":"ContainerStarted","Data":"5d8a06b64cc9778b600eb7ee56a36445340d0fb5dafba3ee09865a0b6e2343ea"} Dec 13 07:12:38 crc kubenswrapper[4971]: I1213 07:12:38.810090 4971 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/horizon-797dfcfc58-wlch7" event={"ID":"9a4736be-5537-4638-a203-d36db1ae35cc","Type":"ContainerStarted","Data":"90a7ac52aff84d312a6639f807172d8d14d9c2b632657f90fe22dcd2b08e245b"} Dec 13 07:12:38 crc kubenswrapper[4971]: I1213 07:12:38.816574 4971 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/placement-db-sync-954vn" event={"ID":"166612b7-15b4-4a04-b7bb-62603363de80","Type":"ContainerStarted","Data":"8c01f00505b9ed9c18e1f1f83e29164af0bbff0a698dc2db5d4e5ba5e0a656f9"} Dec 13 07:12:38 crc kubenswrapper[4971]: E1213 07:12:38.818439 4971 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"cinder-db-sync\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/podified-antelope-centos9/openstack-cinder-api@sha256:b59b7445e581cc720038107e421371c86c5765b2967e77d884ef29b1d9fd0f49\\\"\"" pod="openstack/cinder-db-sync-26k47" podUID="c1d819d0-339f-4e7e-9180-63d2f7c0042e" Dec 13 07:12:38 crc kubenswrapper[4971]: I1213 07:12:38.849893 4971 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/keystone-bootstrap-dn9h9" podStartSLOduration=9.849869699 podStartE2EDuration="9.849869699s" podCreationTimestamp="2025-12-13 07:12:29 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-13 07:12:38.837421872 +0000 UTC m=+1415.441831390" watchObservedRunningTime="2025-12-13 07:12:38.849869699 +0000 UTC m=+1415.454279147" Dec 13 07:12:38 crc kubenswrapper[4971]: I1213 07:12:38.882405 4971 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/placement-db-sync-954vn" podStartSLOduration=4.677810401 podStartE2EDuration="34.882385909s" podCreationTimestamp="2025-12-13 07:12:04 +0000 UTC" firstStartedPulling="2025-12-13 07:12:06.574879667 +0000 UTC m=+1383.179289115" lastFinishedPulling="2025-12-13 07:12:36.779455185 +0000 UTC m=+1413.383864623" observedRunningTime="2025-12-13 07:12:38.879951079 +0000 UTC m=+1415.484360597" watchObservedRunningTime="2025-12-13 07:12:38.882385909 +0000 UTC m=+1415.486795357" Dec 13 07:12:39 crc kubenswrapper[4971]: I1213 07:12:39.054156 4971 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-548894858c-j5bhs"] Dec 13 07:12:39 crc kubenswrapper[4971]: E1213 07:12:39.054512 4971 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="c9e30081-d603-4c9d-81ce-7bcff6848243" containerName="dnsmasq-dns" Dec 13 07:12:39 crc kubenswrapper[4971]: I1213 07:12:39.054543 4971 state_mem.go:107] "Deleted CPUSet assignment" podUID="c9e30081-d603-4c9d-81ce-7bcff6848243" containerName="dnsmasq-dns" Dec 13 07:12:39 crc kubenswrapper[4971]: E1213 07:12:39.054556 4971 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="0813350d-430f-4564-a09a-062a3239ba2f" containerName="neutron-db-sync" Dec 13 07:12:39 crc kubenswrapper[4971]: I1213 07:12:39.054562 4971 state_mem.go:107] "Deleted CPUSet assignment" podUID="0813350d-430f-4564-a09a-062a3239ba2f" containerName="neutron-db-sync" Dec 13 07:12:39 crc kubenswrapper[4971]: E1213 07:12:39.054573 4971 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="c9e30081-d603-4c9d-81ce-7bcff6848243" containerName="init" Dec 13 07:12:39 crc kubenswrapper[4971]: I1213 07:12:39.054579 4971 state_mem.go:107] "Deleted CPUSet assignment" podUID="c9e30081-d603-4c9d-81ce-7bcff6848243" containerName="init" Dec 13 07:12:39 crc kubenswrapper[4971]: I1213 07:12:39.054734 4971 memory_manager.go:354] "RemoveStaleState removing state" podUID="c9e30081-d603-4c9d-81ce-7bcff6848243" containerName="dnsmasq-dns" Dec 13 07:12:39 crc kubenswrapper[4971]: I1213 07:12:39.054752 4971 memory_manager.go:354] "RemoveStaleState removing state" podUID="0813350d-430f-4564-a09a-062a3239ba2f" containerName="neutron-db-sync" Dec 13 07:12:39 crc kubenswrapper[4971]: I1213 07:12:39.055508 4971 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-548894858c-j5bhs" Dec 13 07:12:39 crc kubenswrapper[4971]: I1213 07:12:39.072711 4971 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/cd7d111d-0b3e-4bd7-ba07-10e296208a7f-ovsdbserver-nb\") pod \"dnsmasq-dns-548894858c-j5bhs\" (UID: \"cd7d111d-0b3e-4bd7-ba07-10e296208a7f\") " pod="openstack/dnsmasq-dns-548894858c-j5bhs" Dec 13 07:12:39 crc kubenswrapper[4971]: I1213 07:12:39.072808 4971 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/cd7d111d-0b3e-4bd7-ba07-10e296208a7f-config\") pod \"dnsmasq-dns-548894858c-j5bhs\" (UID: \"cd7d111d-0b3e-4bd7-ba07-10e296208a7f\") " pod="openstack/dnsmasq-dns-548894858c-j5bhs" Dec 13 07:12:39 crc kubenswrapper[4971]: I1213 07:12:39.072862 4971 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/cd7d111d-0b3e-4bd7-ba07-10e296208a7f-ovsdbserver-sb\") pod \"dnsmasq-dns-548894858c-j5bhs\" (UID: \"cd7d111d-0b3e-4bd7-ba07-10e296208a7f\") " pod="openstack/dnsmasq-dns-548894858c-j5bhs" Dec 13 07:12:39 crc kubenswrapper[4971]: I1213 07:12:39.072880 4971 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-7hqhn\" (UniqueName: \"kubernetes.io/projected/cd7d111d-0b3e-4bd7-ba07-10e296208a7f-kube-api-access-7hqhn\") pod \"dnsmasq-dns-548894858c-j5bhs\" (UID: \"cd7d111d-0b3e-4bd7-ba07-10e296208a7f\") " pod="openstack/dnsmasq-dns-548894858c-j5bhs" Dec 13 07:12:39 crc kubenswrapper[4971]: I1213 07:12:39.072899 4971 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/cd7d111d-0b3e-4bd7-ba07-10e296208a7f-dns-svc\") pod \"dnsmasq-dns-548894858c-j5bhs\" (UID: \"cd7d111d-0b3e-4bd7-ba07-10e296208a7f\") " pod="openstack/dnsmasq-dns-548894858c-j5bhs" Dec 13 07:12:39 crc kubenswrapper[4971]: I1213 07:12:39.093643 4971 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-548894858c-j5bhs"] Dec 13 07:12:39 crc kubenswrapper[4971]: I1213 07:12:39.176493 4971 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/cd7d111d-0b3e-4bd7-ba07-10e296208a7f-ovsdbserver-sb\") pod \"dnsmasq-dns-548894858c-j5bhs\" (UID: \"cd7d111d-0b3e-4bd7-ba07-10e296208a7f\") " pod="openstack/dnsmasq-dns-548894858c-j5bhs" Dec 13 07:12:39 crc kubenswrapper[4971]: I1213 07:12:39.178430 4971 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-7hqhn\" (UniqueName: \"kubernetes.io/projected/cd7d111d-0b3e-4bd7-ba07-10e296208a7f-kube-api-access-7hqhn\") pod \"dnsmasq-dns-548894858c-j5bhs\" (UID: \"cd7d111d-0b3e-4bd7-ba07-10e296208a7f\") " pod="openstack/dnsmasq-dns-548894858c-j5bhs" Dec 13 07:12:39 crc kubenswrapper[4971]: I1213 07:12:39.178482 4971 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/cd7d111d-0b3e-4bd7-ba07-10e296208a7f-dns-svc\") pod \"dnsmasq-dns-548894858c-j5bhs\" (UID: \"cd7d111d-0b3e-4bd7-ba07-10e296208a7f\") " pod="openstack/dnsmasq-dns-548894858c-j5bhs" Dec 13 07:12:39 crc kubenswrapper[4971]: I1213 07:12:39.178642 4971 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/cd7d111d-0b3e-4bd7-ba07-10e296208a7f-ovsdbserver-nb\") pod \"dnsmasq-dns-548894858c-j5bhs\" (UID: \"cd7d111d-0b3e-4bd7-ba07-10e296208a7f\") " pod="openstack/dnsmasq-dns-548894858c-j5bhs" Dec 13 07:12:39 crc kubenswrapper[4971]: I1213 07:12:39.178776 4971 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/cd7d111d-0b3e-4bd7-ba07-10e296208a7f-config\") pod \"dnsmasq-dns-548894858c-j5bhs\" (UID: \"cd7d111d-0b3e-4bd7-ba07-10e296208a7f\") " pod="openstack/dnsmasq-dns-548894858c-j5bhs" Dec 13 07:12:39 crc kubenswrapper[4971]: I1213 07:12:39.181337 4971 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/cd7d111d-0b3e-4bd7-ba07-10e296208a7f-config\") pod \"dnsmasq-dns-548894858c-j5bhs\" (UID: \"cd7d111d-0b3e-4bd7-ba07-10e296208a7f\") " pod="openstack/dnsmasq-dns-548894858c-j5bhs" Dec 13 07:12:39 crc kubenswrapper[4971]: I1213 07:12:39.181415 4971 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/cd7d111d-0b3e-4bd7-ba07-10e296208a7f-ovsdbserver-sb\") pod \"dnsmasq-dns-548894858c-j5bhs\" (UID: \"cd7d111d-0b3e-4bd7-ba07-10e296208a7f\") " pod="openstack/dnsmasq-dns-548894858c-j5bhs" Dec 13 07:12:39 crc kubenswrapper[4971]: I1213 07:12:39.182081 4971 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/cd7d111d-0b3e-4bd7-ba07-10e296208a7f-dns-svc\") pod \"dnsmasq-dns-548894858c-j5bhs\" (UID: \"cd7d111d-0b3e-4bd7-ba07-10e296208a7f\") " pod="openstack/dnsmasq-dns-548894858c-j5bhs" Dec 13 07:12:39 crc kubenswrapper[4971]: I1213 07:12:39.182767 4971 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/cd7d111d-0b3e-4bd7-ba07-10e296208a7f-ovsdbserver-nb\") pod \"dnsmasq-dns-548894858c-j5bhs\" (UID: \"cd7d111d-0b3e-4bd7-ba07-10e296208a7f\") " pod="openstack/dnsmasq-dns-548894858c-j5bhs" Dec 13 07:12:39 crc kubenswrapper[4971]: I1213 07:12:39.203888 4971 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/neutron-6597fdddd4-q5zr5"] Dec 13 07:12:39 crc kubenswrapper[4971]: I1213 07:12:39.209917 4971 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-7hqhn\" (UniqueName: \"kubernetes.io/projected/cd7d111d-0b3e-4bd7-ba07-10e296208a7f-kube-api-access-7hqhn\") pod \"dnsmasq-dns-548894858c-j5bhs\" (UID: \"cd7d111d-0b3e-4bd7-ba07-10e296208a7f\") " pod="openstack/dnsmasq-dns-548894858c-j5bhs" Dec 13 07:12:39 crc kubenswrapper[4971]: I1213 07:12:39.213100 4971 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/neutron-6597fdddd4-q5zr5"] Dec 13 07:12:39 crc kubenswrapper[4971]: I1213 07:12:39.213345 4971 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-6597fdddd4-q5zr5" Dec 13 07:12:39 crc kubenswrapper[4971]: I1213 07:12:39.216104 4971 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"neutron-httpd-config" Dec 13 07:12:39 crc kubenswrapper[4971]: I1213 07:12:39.216626 4971 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"neutron-config" Dec 13 07:12:39 crc kubenswrapper[4971]: I1213 07:12:39.216842 4971 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"neutron-neutron-dockercfg-tz9f5" Dec 13 07:12:39 crc kubenswrapper[4971]: I1213 07:12:39.216965 4971 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-neutron-ovndbs" Dec 13 07:12:39 crc kubenswrapper[4971]: I1213 07:12:39.381704 4971 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"httpd-config\" (UniqueName: \"kubernetes.io/secret/a844da97-38e3-4bc5-981f-af3fa9f76483-httpd-config\") pod \"neutron-6597fdddd4-q5zr5\" (UID: \"a844da97-38e3-4bc5-981f-af3fa9f76483\") " pod="openstack/neutron-6597fdddd4-q5zr5" Dec 13 07:12:39 crc kubenswrapper[4971]: I1213 07:12:39.381750 4971 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/a844da97-38e3-4bc5-981f-af3fa9f76483-combined-ca-bundle\") pod \"neutron-6597fdddd4-q5zr5\" (UID: \"a844da97-38e3-4bc5-981f-af3fa9f76483\") " pod="openstack/neutron-6597fdddd4-q5zr5" Dec 13 07:12:39 crc kubenswrapper[4971]: I1213 07:12:39.381780 4971 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovndb-tls-certs\" (UniqueName: \"kubernetes.io/secret/a844da97-38e3-4bc5-981f-af3fa9f76483-ovndb-tls-certs\") pod \"neutron-6597fdddd4-q5zr5\" (UID: \"a844da97-38e3-4bc5-981f-af3fa9f76483\") " pod="openstack/neutron-6597fdddd4-q5zr5" Dec 13 07:12:39 crc kubenswrapper[4971]: I1213 07:12:39.381805 4971 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-knv8t\" (UniqueName: \"kubernetes.io/projected/a844da97-38e3-4bc5-981f-af3fa9f76483-kube-api-access-knv8t\") pod \"neutron-6597fdddd4-q5zr5\" (UID: \"a844da97-38e3-4bc5-981f-af3fa9f76483\") " pod="openstack/neutron-6597fdddd4-q5zr5" Dec 13 07:12:39 crc kubenswrapper[4971]: I1213 07:12:39.381898 4971 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/secret/a844da97-38e3-4bc5-981f-af3fa9f76483-config\") pod \"neutron-6597fdddd4-q5zr5\" (UID: \"a844da97-38e3-4bc5-981f-af3fa9f76483\") " pod="openstack/neutron-6597fdddd4-q5zr5" Dec 13 07:12:39 crc kubenswrapper[4971]: I1213 07:12:39.386069 4971 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-548894858c-j5bhs" Dec 13 07:12:39 crc kubenswrapper[4971]: I1213 07:12:39.484419 4971 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"httpd-config\" (UniqueName: \"kubernetes.io/secret/a844da97-38e3-4bc5-981f-af3fa9f76483-httpd-config\") pod \"neutron-6597fdddd4-q5zr5\" (UID: \"a844da97-38e3-4bc5-981f-af3fa9f76483\") " pod="openstack/neutron-6597fdddd4-q5zr5" Dec 13 07:12:39 crc kubenswrapper[4971]: I1213 07:12:39.484719 4971 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/a844da97-38e3-4bc5-981f-af3fa9f76483-combined-ca-bundle\") pod \"neutron-6597fdddd4-q5zr5\" (UID: \"a844da97-38e3-4bc5-981f-af3fa9f76483\") " pod="openstack/neutron-6597fdddd4-q5zr5" Dec 13 07:12:39 crc kubenswrapper[4971]: I1213 07:12:39.484748 4971 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovndb-tls-certs\" (UniqueName: \"kubernetes.io/secret/a844da97-38e3-4bc5-981f-af3fa9f76483-ovndb-tls-certs\") pod \"neutron-6597fdddd4-q5zr5\" (UID: \"a844da97-38e3-4bc5-981f-af3fa9f76483\") " pod="openstack/neutron-6597fdddd4-q5zr5" Dec 13 07:12:39 crc kubenswrapper[4971]: I1213 07:12:39.484892 4971 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-knv8t\" (UniqueName: \"kubernetes.io/projected/a844da97-38e3-4bc5-981f-af3fa9f76483-kube-api-access-knv8t\") pod \"neutron-6597fdddd4-q5zr5\" (UID: \"a844da97-38e3-4bc5-981f-af3fa9f76483\") " pod="openstack/neutron-6597fdddd4-q5zr5" Dec 13 07:12:39 crc kubenswrapper[4971]: I1213 07:12:39.484982 4971 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/secret/a844da97-38e3-4bc5-981f-af3fa9f76483-config\") pod \"neutron-6597fdddd4-q5zr5\" (UID: \"a844da97-38e3-4bc5-981f-af3fa9f76483\") " pod="openstack/neutron-6597fdddd4-q5zr5" Dec 13 07:12:39 crc kubenswrapper[4971]: I1213 07:12:39.489172 4971 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/secret/a844da97-38e3-4bc5-981f-af3fa9f76483-config\") pod \"neutron-6597fdddd4-q5zr5\" (UID: \"a844da97-38e3-4bc5-981f-af3fa9f76483\") " pod="openstack/neutron-6597fdddd4-q5zr5" Dec 13 07:12:39 crc kubenswrapper[4971]: I1213 07:12:39.489479 4971 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/a844da97-38e3-4bc5-981f-af3fa9f76483-combined-ca-bundle\") pod \"neutron-6597fdddd4-q5zr5\" (UID: \"a844da97-38e3-4bc5-981f-af3fa9f76483\") " pod="openstack/neutron-6597fdddd4-q5zr5" Dec 13 07:12:39 crc kubenswrapper[4971]: I1213 07:12:39.489947 4971 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovndb-tls-certs\" (UniqueName: \"kubernetes.io/secret/a844da97-38e3-4bc5-981f-af3fa9f76483-ovndb-tls-certs\") pod \"neutron-6597fdddd4-q5zr5\" (UID: \"a844da97-38e3-4bc5-981f-af3fa9f76483\") " pod="openstack/neutron-6597fdddd4-q5zr5" Dec 13 07:12:39 crc kubenswrapper[4971]: I1213 07:12:39.496512 4971 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"httpd-config\" (UniqueName: \"kubernetes.io/secret/a844da97-38e3-4bc5-981f-af3fa9f76483-httpd-config\") pod \"neutron-6597fdddd4-q5zr5\" (UID: \"a844da97-38e3-4bc5-981f-af3fa9f76483\") " pod="openstack/neutron-6597fdddd4-q5zr5" Dec 13 07:12:39 crc kubenswrapper[4971]: I1213 07:12:39.507298 4971 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-knv8t\" (UniqueName: \"kubernetes.io/projected/a844da97-38e3-4bc5-981f-af3fa9f76483-kube-api-access-knv8t\") pod \"neutron-6597fdddd4-q5zr5\" (UID: \"a844da97-38e3-4bc5-981f-af3fa9f76483\") " pod="openstack/neutron-6597fdddd4-q5zr5" Dec 13 07:12:39 crc kubenswrapper[4971]: I1213 07:12:39.550899 4971 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-6597fdddd4-q5zr5" Dec 13 07:12:39 crc kubenswrapper[4971]: I1213 07:12:39.784726 4971 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="b314f107-1cb6-49f3-aa51-3f607ca79f6b" path="/var/lib/kubelet/pods/b314f107-1cb6-49f3-aa51-3f607ca79f6b/volumes" Dec 13 07:12:39 crc kubenswrapper[4971]: I1213 07:12:39.785405 4971 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="ff35e760-13e7-4145-9970-742b73a36176" path="/var/lib/kubelet/pods/ff35e760-13e7-4145-9970-742b73a36176/volumes" Dec 13 07:12:39 crc kubenswrapper[4971]: I1213 07:12:39.847411 4971 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/horizon-8699dc887d-sl2h6" event={"ID":"772eedab-7c62-454b-b722-ee453b39e89f","Type":"ContainerStarted","Data":"4a148882fad66a5e1444a664e27875de5e47f0f5a778fde8428a33e1ba8f489f"} Dec 13 07:12:39 crc kubenswrapper[4971]: I1213 07:12:39.847459 4971 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/horizon-8699dc887d-sl2h6" event={"ID":"772eedab-7c62-454b-b722-ee453b39e89f","Type":"ContainerStarted","Data":"2074da306f3c927fa0e63c982fab7574e67b1f06c38e55ba7dc406c0d32df42a"} Dec 13 07:12:39 crc kubenswrapper[4971]: I1213 07:12:39.851686 4971 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/horizon-797dfcfc58-wlch7" event={"ID":"9a4736be-5537-4638-a203-d36db1ae35cc","Type":"ContainerStarted","Data":"a727e323f7c912f7f04bc539cc15e694592384ce4a73b6fe60ccb0d0378fdb79"} Dec 13 07:12:39 crc kubenswrapper[4971]: I1213 07:12:39.851726 4971 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/horizon-797dfcfc58-wlch7" event={"ID":"9a4736be-5537-4638-a203-d36db1ae35cc","Type":"ContainerStarted","Data":"4cf1487fee38b8790b515ad9b20f42309df31cb004b7ca14d1783d04a0edcf24"} Dec 13 07:12:39 crc kubenswrapper[4971]: I1213 07:12:39.884145 4971 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/horizon-8699dc887d-sl2h6" podStartSLOduration=26.250870251 podStartE2EDuration="26.884121581s" podCreationTimestamp="2025-12-13 07:12:13 +0000 UTC" firstStartedPulling="2025-12-13 07:12:38.537444457 +0000 UTC m=+1415.141853905" lastFinishedPulling="2025-12-13 07:12:39.170695787 +0000 UTC m=+1415.775105235" observedRunningTime="2025-12-13 07:12:39.866849766 +0000 UTC m=+1416.471259214" watchObservedRunningTime="2025-12-13 07:12:39.884121581 +0000 UTC m=+1416.488531029" Dec 13 07:12:39 crc kubenswrapper[4971]: I1213 07:12:39.895486 4971 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/horizon-797dfcfc58-wlch7" podStartSLOduration=26.443688258 podStartE2EDuration="26.895467231s" podCreationTimestamp="2025-12-13 07:12:13 +0000 UTC" firstStartedPulling="2025-12-13 07:12:38.581961752 +0000 UTC m=+1415.186371200" lastFinishedPulling="2025-12-13 07:12:39.033740725 +0000 UTC m=+1415.638150173" observedRunningTime="2025-12-13 07:12:39.891642627 +0000 UTC m=+1416.496052085" watchObservedRunningTime="2025-12-13 07:12:39.895467231 +0000 UTC m=+1416.499876689" Dec 13 07:12:39 crc kubenswrapper[4971]: I1213 07:12:39.935059 4971 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-548894858c-j5bhs"] Dec 13 07:12:40 crc kubenswrapper[4971]: I1213 07:12:40.062951 4971 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/neutron-6597fdddd4-q5zr5"] Dec 13 07:12:40 crc kubenswrapper[4971]: W1213 07:12:40.064565 4971 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-poda844da97_38e3_4bc5_981f_af3fa9f76483.slice/crio-93faa59d0bd7c5a3f465196c197b6c4c5cf492443b1bf4cbadf37610a3d42122 WatchSource:0}: Error finding container 93faa59d0bd7c5a3f465196c197b6c4c5cf492443b1bf4cbadf37610a3d42122: Status 404 returned error can't find the container with id 93faa59d0bd7c5a3f465196c197b6c4c5cf492443b1bf4cbadf37610a3d42122 Dec 13 07:12:40 crc kubenswrapper[4971]: I1213 07:12:40.580408 4971 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/dnsmasq-dns-57768dd7b5-86hlw" podUID="c9e30081-d603-4c9d-81ce-7bcff6848243" containerName="dnsmasq-dns" probeResult="failure" output="dial tcp 10.217.0.121:5353: i/o timeout" Dec 13 07:12:40 crc kubenswrapper[4971]: I1213 07:12:40.868257 4971 generic.go:334] "Generic (PLEG): container finished" podID="166612b7-15b4-4a04-b7bb-62603363de80" containerID="8c01f00505b9ed9c18e1f1f83e29164af0bbff0a698dc2db5d4e5ba5e0a656f9" exitCode=0 Dec 13 07:12:40 crc kubenswrapper[4971]: I1213 07:12:40.868359 4971 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/placement-db-sync-954vn" event={"ID":"166612b7-15b4-4a04-b7bb-62603363de80","Type":"ContainerDied","Data":"8c01f00505b9ed9c18e1f1f83e29164af0bbff0a698dc2db5d4e5ba5e0a656f9"} Dec 13 07:12:40 crc kubenswrapper[4971]: I1213 07:12:40.871597 4971 generic.go:334] "Generic (PLEG): container finished" podID="cd7d111d-0b3e-4bd7-ba07-10e296208a7f" containerID="5f0a1e401c9c7ec2cb4fb7ef1c38076d6e666c4f5c4d211000b63a3fb564c157" exitCode=0 Dec 13 07:12:40 crc kubenswrapper[4971]: I1213 07:12:40.871682 4971 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-548894858c-j5bhs" event={"ID":"cd7d111d-0b3e-4bd7-ba07-10e296208a7f","Type":"ContainerDied","Data":"5f0a1e401c9c7ec2cb4fb7ef1c38076d6e666c4f5c4d211000b63a3fb564c157"} Dec 13 07:12:40 crc kubenswrapper[4971]: I1213 07:12:40.871726 4971 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-548894858c-j5bhs" event={"ID":"cd7d111d-0b3e-4bd7-ba07-10e296208a7f","Type":"ContainerStarted","Data":"786637934b9af6b6e09bd09c84cb2ce4d124a73c928f00884fe07b7200c727e1"} Dec 13 07:12:40 crc kubenswrapper[4971]: I1213 07:12:40.874898 4971 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-6597fdddd4-q5zr5" event={"ID":"a844da97-38e3-4bc5-981f-af3fa9f76483","Type":"ContainerStarted","Data":"93faa59d0bd7c5a3f465196c197b6c4c5cf492443b1bf4cbadf37610a3d42122"} Dec 13 07:12:41 crc kubenswrapper[4971]: I1213 07:12:41.571850 4971 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/neutron-7956479f45-8g7zz"] Dec 13 07:12:41 crc kubenswrapper[4971]: I1213 07:12:41.573103 4971 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-7956479f45-8g7zz" Dec 13 07:12:41 crc kubenswrapper[4971]: I1213 07:12:41.575363 4971 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-neutron-public-svc" Dec 13 07:12:41 crc kubenswrapper[4971]: I1213 07:12:41.579858 4971 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-neutron-internal-svc" Dec 13 07:12:41 crc kubenswrapper[4971]: I1213 07:12:41.586646 4971 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/neutron-7956479f45-8g7zz"] Dec 13 07:12:41 crc kubenswrapper[4971]: I1213 07:12:41.727227 4971 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovndb-tls-certs\" (UniqueName: \"kubernetes.io/secret/4426fcd4-0b89-4e70-8f2f-43b056896953-ovndb-tls-certs\") pod \"neutron-7956479f45-8g7zz\" (UID: \"4426fcd4-0b89-4e70-8f2f-43b056896953\") " pod="openstack/neutron-7956479f45-8g7zz" Dec 13 07:12:41 crc kubenswrapper[4971]: I1213 07:12:41.727490 4971 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"httpd-config\" (UniqueName: \"kubernetes.io/secret/4426fcd4-0b89-4e70-8f2f-43b056896953-httpd-config\") pod \"neutron-7956479f45-8g7zz\" (UID: \"4426fcd4-0b89-4e70-8f2f-43b056896953\") " pod="openstack/neutron-7956479f45-8g7zz" Dec 13 07:12:41 crc kubenswrapper[4971]: I1213 07:12:41.727540 4971 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-dlbqv\" (UniqueName: \"kubernetes.io/projected/4426fcd4-0b89-4e70-8f2f-43b056896953-kube-api-access-dlbqv\") pod \"neutron-7956479f45-8g7zz\" (UID: \"4426fcd4-0b89-4e70-8f2f-43b056896953\") " pod="openstack/neutron-7956479f45-8g7zz" Dec 13 07:12:41 crc kubenswrapper[4971]: I1213 07:12:41.727671 4971 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/secret/4426fcd4-0b89-4e70-8f2f-43b056896953-config\") pod \"neutron-7956479f45-8g7zz\" (UID: \"4426fcd4-0b89-4e70-8f2f-43b056896953\") " pod="openstack/neutron-7956479f45-8g7zz" Dec 13 07:12:41 crc kubenswrapper[4971]: I1213 07:12:41.727757 4971 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/4426fcd4-0b89-4e70-8f2f-43b056896953-internal-tls-certs\") pod \"neutron-7956479f45-8g7zz\" (UID: \"4426fcd4-0b89-4e70-8f2f-43b056896953\") " pod="openstack/neutron-7956479f45-8g7zz" Dec 13 07:12:41 crc kubenswrapper[4971]: I1213 07:12:41.727846 4971 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/4426fcd4-0b89-4e70-8f2f-43b056896953-combined-ca-bundle\") pod \"neutron-7956479f45-8g7zz\" (UID: \"4426fcd4-0b89-4e70-8f2f-43b056896953\") " pod="openstack/neutron-7956479f45-8g7zz" Dec 13 07:12:41 crc kubenswrapper[4971]: I1213 07:12:41.727889 4971 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/4426fcd4-0b89-4e70-8f2f-43b056896953-public-tls-certs\") pod \"neutron-7956479f45-8g7zz\" (UID: \"4426fcd4-0b89-4e70-8f2f-43b056896953\") " pod="openstack/neutron-7956479f45-8g7zz" Dec 13 07:12:41 crc kubenswrapper[4971]: I1213 07:12:41.829694 4971 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovndb-tls-certs\" (UniqueName: \"kubernetes.io/secret/4426fcd4-0b89-4e70-8f2f-43b056896953-ovndb-tls-certs\") pod \"neutron-7956479f45-8g7zz\" (UID: \"4426fcd4-0b89-4e70-8f2f-43b056896953\") " pod="openstack/neutron-7956479f45-8g7zz" Dec 13 07:12:41 crc kubenswrapper[4971]: I1213 07:12:41.829763 4971 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"httpd-config\" (UniqueName: \"kubernetes.io/secret/4426fcd4-0b89-4e70-8f2f-43b056896953-httpd-config\") pod \"neutron-7956479f45-8g7zz\" (UID: \"4426fcd4-0b89-4e70-8f2f-43b056896953\") " pod="openstack/neutron-7956479f45-8g7zz" Dec 13 07:12:41 crc kubenswrapper[4971]: I1213 07:12:41.829816 4971 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-dlbqv\" (UniqueName: \"kubernetes.io/projected/4426fcd4-0b89-4e70-8f2f-43b056896953-kube-api-access-dlbqv\") pod \"neutron-7956479f45-8g7zz\" (UID: \"4426fcd4-0b89-4e70-8f2f-43b056896953\") " pod="openstack/neutron-7956479f45-8g7zz" Dec 13 07:12:41 crc kubenswrapper[4971]: I1213 07:12:41.829890 4971 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/secret/4426fcd4-0b89-4e70-8f2f-43b056896953-config\") pod \"neutron-7956479f45-8g7zz\" (UID: \"4426fcd4-0b89-4e70-8f2f-43b056896953\") " pod="openstack/neutron-7956479f45-8g7zz" Dec 13 07:12:41 crc kubenswrapper[4971]: I1213 07:12:41.829922 4971 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/4426fcd4-0b89-4e70-8f2f-43b056896953-internal-tls-certs\") pod \"neutron-7956479f45-8g7zz\" (UID: \"4426fcd4-0b89-4e70-8f2f-43b056896953\") " pod="openstack/neutron-7956479f45-8g7zz" Dec 13 07:12:41 crc kubenswrapper[4971]: I1213 07:12:41.829969 4971 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/4426fcd4-0b89-4e70-8f2f-43b056896953-combined-ca-bundle\") pod \"neutron-7956479f45-8g7zz\" (UID: \"4426fcd4-0b89-4e70-8f2f-43b056896953\") " pod="openstack/neutron-7956479f45-8g7zz" Dec 13 07:12:41 crc kubenswrapper[4971]: I1213 07:12:41.830001 4971 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/4426fcd4-0b89-4e70-8f2f-43b056896953-public-tls-certs\") pod \"neutron-7956479f45-8g7zz\" (UID: \"4426fcd4-0b89-4e70-8f2f-43b056896953\") " pod="openstack/neutron-7956479f45-8g7zz" Dec 13 07:12:41 crc kubenswrapper[4971]: I1213 07:12:41.836412 4971 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/4426fcd4-0b89-4e70-8f2f-43b056896953-public-tls-certs\") pod \"neutron-7956479f45-8g7zz\" (UID: \"4426fcd4-0b89-4e70-8f2f-43b056896953\") " pod="openstack/neutron-7956479f45-8g7zz" Dec 13 07:12:41 crc kubenswrapper[4971]: I1213 07:12:41.836669 4971 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/secret/4426fcd4-0b89-4e70-8f2f-43b056896953-config\") pod \"neutron-7956479f45-8g7zz\" (UID: \"4426fcd4-0b89-4e70-8f2f-43b056896953\") " pod="openstack/neutron-7956479f45-8g7zz" Dec 13 07:12:41 crc kubenswrapper[4971]: I1213 07:12:41.838989 4971 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"httpd-config\" (UniqueName: \"kubernetes.io/secret/4426fcd4-0b89-4e70-8f2f-43b056896953-httpd-config\") pod \"neutron-7956479f45-8g7zz\" (UID: \"4426fcd4-0b89-4e70-8f2f-43b056896953\") " pod="openstack/neutron-7956479f45-8g7zz" Dec 13 07:12:41 crc kubenswrapper[4971]: I1213 07:12:41.841405 4971 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovndb-tls-certs\" (UniqueName: \"kubernetes.io/secret/4426fcd4-0b89-4e70-8f2f-43b056896953-ovndb-tls-certs\") pod \"neutron-7956479f45-8g7zz\" (UID: \"4426fcd4-0b89-4e70-8f2f-43b056896953\") " pod="openstack/neutron-7956479f45-8g7zz" Dec 13 07:12:41 crc kubenswrapper[4971]: I1213 07:12:41.843331 4971 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/4426fcd4-0b89-4e70-8f2f-43b056896953-internal-tls-certs\") pod \"neutron-7956479f45-8g7zz\" (UID: \"4426fcd4-0b89-4e70-8f2f-43b056896953\") " pod="openstack/neutron-7956479f45-8g7zz" Dec 13 07:12:41 crc kubenswrapper[4971]: I1213 07:12:41.844258 4971 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/4426fcd4-0b89-4e70-8f2f-43b056896953-combined-ca-bundle\") pod \"neutron-7956479f45-8g7zz\" (UID: \"4426fcd4-0b89-4e70-8f2f-43b056896953\") " pod="openstack/neutron-7956479f45-8g7zz" Dec 13 07:12:41 crc kubenswrapper[4971]: I1213 07:12:41.871139 4971 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-dlbqv\" (UniqueName: \"kubernetes.io/projected/4426fcd4-0b89-4e70-8f2f-43b056896953-kube-api-access-dlbqv\") pod \"neutron-7956479f45-8g7zz\" (UID: \"4426fcd4-0b89-4e70-8f2f-43b056896953\") " pod="openstack/neutron-7956479f45-8g7zz" Dec 13 07:12:41 crc kubenswrapper[4971]: I1213 07:12:41.884938 4971 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-548894858c-j5bhs" event={"ID":"cd7d111d-0b3e-4bd7-ba07-10e296208a7f","Type":"ContainerStarted","Data":"aebcdf008ab057d1a88376abfdad8b3a2c538ea0af5a431a1b81134323b14d9c"} Dec 13 07:12:41 crc kubenswrapper[4971]: I1213 07:12:41.885258 4971 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/dnsmasq-dns-548894858c-j5bhs" Dec 13 07:12:41 crc kubenswrapper[4971]: I1213 07:12:41.887493 4971 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-6597fdddd4-q5zr5" event={"ID":"a844da97-38e3-4bc5-981f-af3fa9f76483","Type":"ContainerStarted","Data":"ff17f9b1aa76599719c611ab63925edefc199a526338bcd7c68bdf8db15d11b8"} Dec 13 07:12:41 crc kubenswrapper[4971]: I1213 07:12:41.887557 4971 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-6597fdddd4-q5zr5" event={"ID":"a844da97-38e3-4bc5-981f-af3fa9f76483","Type":"ContainerStarted","Data":"bc6784a5d2347f8bea8ff074080f8d5331cf74f51b2062a8e0fa3b8e1dc2fdaf"} Dec 13 07:12:41 crc kubenswrapper[4971]: I1213 07:12:41.887632 4971 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/neutron-6597fdddd4-q5zr5" Dec 13 07:12:41 crc kubenswrapper[4971]: I1213 07:12:41.889580 4971 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"56e56ce1-4dd1-4ae5-b868-e98d27d5ad94","Type":"ContainerStarted","Data":"d208a7c037f088ceda5804583d1d1761594f3fbc2ad4ae265ebc6e2bf875c418"} Dec 13 07:12:41 crc kubenswrapper[4971]: I1213 07:12:41.930538 4971 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/dnsmasq-dns-548894858c-j5bhs" podStartSLOduration=2.930503483 podStartE2EDuration="2.930503483s" podCreationTimestamp="2025-12-13 07:12:39 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-13 07:12:41.909972867 +0000 UTC m=+1418.514382315" watchObservedRunningTime="2025-12-13 07:12:41.930503483 +0000 UTC m=+1418.534912931" Dec 13 07:12:41 crc kubenswrapper[4971]: I1213 07:12:41.939399 4971 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-7956479f45-8g7zz" Dec 13 07:12:41 crc kubenswrapper[4971]: I1213 07:12:41.948270 4971 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/neutron-6597fdddd4-q5zr5" podStartSLOduration=2.94825056 podStartE2EDuration="2.94825056s" podCreationTimestamp="2025-12-13 07:12:39 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-13 07:12:41.940798556 +0000 UTC m=+1418.545208014" watchObservedRunningTime="2025-12-13 07:12:41.94825056 +0000 UTC m=+1418.552660008" Dec 13 07:12:44 crc kubenswrapper[4971]: I1213 07:12:44.034645 4971 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/horizon-8699dc887d-sl2h6" Dec 13 07:12:44 crc kubenswrapper[4971]: I1213 07:12:44.035147 4971 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/horizon-8699dc887d-sl2h6" Dec 13 07:12:44 crc kubenswrapper[4971]: I1213 07:12:44.163917 4971 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/horizon-797dfcfc58-wlch7" Dec 13 07:12:44 crc kubenswrapper[4971]: I1213 07:12:44.164236 4971 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/horizon-797dfcfc58-wlch7" Dec 13 07:12:45 crc kubenswrapper[4971]: I1213 07:12:45.489975 4971 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/placement-db-sync-954vn" Dec 13 07:12:45 crc kubenswrapper[4971]: I1213 07:12:45.618166 4971 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/166612b7-15b4-4a04-b7bb-62603363de80-scripts\") pod \"166612b7-15b4-4a04-b7bb-62603363de80\" (UID: \"166612b7-15b4-4a04-b7bb-62603363de80\") " Dec 13 07:12:45 crc kubenswrapper[4971]: I1213 07:12:45.618492 4971 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/166612b7-15b4-4a04-b7bb-62603363de80-combined-ca-bundle\") pod \"166612b7-15b4-4a04-b7bb-62603363de80\" (UID: \"166612b7-15b4-4a04-b7bb-62603363de80\") " Dec 13 07:12:45 crc kubenswrapper[4971]: I1213 07:12:45.618687 4971 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-5lxdd\" (UniqueName: \"kubernetes.io/projected/166612b7-15b4-4a04-b7bb-62603363de80-kube-api-access-5lxdd\") pod \"166612b7-15b4-4a04-b7bb-62603363de80\" (UID: \"166612b7-15b4-4a04-b7bb-62603363de80\") " Dec 13 07:12:45 crc kubenswrapper[4971]: I1213 07:12:45.618730 4971 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/166612b7-15b4-4a04-b7bb-62603363de80-config-data\") pod \"166612b7-15b4-4a04-b7bb-62603363de80\" (UID: \"166612b7-15b4-4a04-b7bb-62603363de80\") " Dec 13 07:12:45 crc kubenswrapper[4971]: I1213 07:12:45.618760 4971 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/166612b7-15b4-4a04-b7bb-62603363de80-logs\") pod \"166612b7-15b4-4a04-b7bb-62603363de80\" (UID: \"166612b7-15b4-4a04-b7bb-62603363de80\") " Dec 13 07:12:45 crc kubenswrapper[4971]: I1213 07:12:45.619355 4971 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/166612b7-15b4-4a04-b7bb-62603363de80-logs" (OuterVolumeSpecName: "logs") pod "166612b7-15b4-4a04-b7bb-62603363de80" (UID: "166612b7-15b4-4a04-b7bb-62603363de80"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 13 07:12:45 crc kubenswrapper[4971]: I1213 07:12:45.628415 4971 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/166612b7-15b4-4a04-b7bb-62603363de80-scripts" (OuterVolumeSpecName: "scripts") pod "166612b7-15b4-4a04-b7bb-62603363de80" (UID: "166612b7-15b4-4a04-b7bb-62603363de80"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 13 07:12:45 crc kubenswrapper[4971]: I1213 07:12:45.660728 4971 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/166612b7-15b4-4a04-b7bb-62603363de80-config-data" (OuterVolumeSpecName: "config-data") pod "166612b7-15b4-4a04-b7bb-62603363de80" (UID: "166612b7-15b4-4a04-b7bb-62603363de80"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 13 07:12:45 crc kubenswrapper[4971]: I1213 07:12:45.684470 4971 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/166612b7-15b4-4a04-b7bb-62603363de80-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "166612b7-15b4-4a04-b7bb-62603363de80" (UID: "166612b7-15b4-4a04-b7bb-62603363de80"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 13 07:12:45 crc kubenswrapper[4971]: I1213 07:12:45.720665 4971 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/166612b7-15b4-4a04-b7bb-62603363de80-config-data\") on node \"crc\" DevicePath \"\"" Dec 13 07:12:45 crc kubenswrapper[4971]: I1213 07:12:45.720703 4971 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/166612b7-15b4-4a04-b7bb-62603363de80-logs\") on node \"crc\" DevicePath \"\"" Dec 13 07:12:45 crc kubenswrapper[4971]: I1213 07:12:45.720712 4971 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/166612b7-15b4-4a04-b7bb-62603363de80-scripts\") on node \"crc\" DevicePath \"\"" Dec 13 07:12:45 crc kubenswrapper[4971]: I1213 07:12:45.720724 4971 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/166612b7-15b4-4a04-b7bb-62603363de80-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 13 07:12:45 crc kubenswrapper[4971]: I1213 07:12:45.809710 4971 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/166612b7-15b4-4a04-b7bb-62603363de80-kube-api-access-5lxdd" (OuterVolumeSpecName: "kube-api-access-5lxdd") pod "166612b7-15b4-4a04-b7bb-62603363de80" (UID: "166612b7-15b4-4a04-b7bb-62603363de80"). InnerVolumeSpecName "kube-api-access-5lxdd". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 13 07:12:45 crc kubenswrapper[4971]: I1213 07:12:45.825231 4971 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-5lxdd\" (UniqueName: \"kubernetes.io/projected/166612b7-15b4-4a04-b7bb-62603363de80-kube-api-access-5lxdd\") on node \"crc\" DevicePath \"\"" Dec 13 07:12:45 crc kubenswrapper[4971]: I1213 07:12:45.913883 4971 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/neutron-7956479f45-8g7zz"] Dec 13 07:12:45 crc kubenswrapper[4971]: W1213 07:12:45.927886 4971 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod4426fcd4_0b89_4e70_8f2f_43b056896953.slice/crio-4ee40cdf950a04cf85b6489a37ab1d5ca19cad3184189c4cfd2efc000b1e71ad WatchSource:0}: Error finding container 4ee40cdf950a04cf85b6489a37ab1d5ca19cad3184189c4cfd2efc000b1e71ad: Status 404 returned error can't find the container with id 4ee40cdf950a04cf85b6489a37ab1d5ca19cad3184189c4cfd2efc000b1e71ad Dec 13 07:12:45 crc kubenswrapper[4971]: I1213 07:12:45.947990 4971 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-7956479f45-8g7zz" event={"ID":"4426fcd4-0b89-4e70-8f2f-43b056896953","Type":"ContainerStarted","Data":"4ee40cdf950a04cf85b6489a37ab1d5ca19cad3184189c4cfd2efc000b1e71ad"} Dec 13 07:12:45 crc kubenswrapper[4971]: I1213 07:12:45.951342 4971 generic.go:334] "Generic (PLEG): container finished" podID="ee3afc2f-79c6-444c-b6a3-2d4c475c3aac" containerID="9d3fdcdb191f5a67a4fa2a8f1e92401ee6d44779f8a7c3fead7ebd96791a001b" exitCode=0 Dec 13 07:12:45 crc kubenswrapper[4971]: I1213 07:12:45.951427 4971 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-bootstrap-dn9h9" event={"ID":"ee3afc2f-79c6-444c-b6a3-2d4c475c3aac","Type":"ContainerDied","Data":"9d3fdcdb191f5a67a4fa2a8f1e92401ee6d44779f8a7c3fead7ebd96791a001b"} Dec 13 07:12:45 crc kubenswrapper[4971]: I1213 07:12:45.969598 4971 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/placement-db-sync-954vn" event={"ID":"166612b7-15b4-4a04-b7bb-62603363de80","Type":"ContainerDied","Data":"c55f47b9cb113dd1b8560c9119be9a982a21f03ef350e168081b073d93064aaf"} Dec 13 07:12:45 crc kubenswrapper[4971]: I1213 07:12:45.969638 4971 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="c55f47b9cb113dd1b8560c9119be9a982a21f03ef350e168081b073d93064aaf" Dec 13 07:12:45 crc kubenswrapper[4971]: I1213 07:12:45.969699 4971 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/placement-db-sync-954vn" Dec 13 07:12:46 crc kubenswrapper[4971]: I1213 07:12:46.594388 4971 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/placement-569b6fd56b-w7xmm"] Dec 13 07:12:46 crc kubenswrapper[4971]: E1213 07:12:46.597244 4971 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="166612b7-15b4-4a04-b7bb-62603363de80" containerName="placement-db-sync" Dec 13 07:12:46 crc kubenswrapper[4971]: I1213 07:12:46.597264 4971 state_mem.go:107] "Deleted CPUSet assignment" podUID="166612b7-15b4-4a04-b7bb-62603363de80" containerName="placement-db-sync" Dec 13 07:12:46 crc kubenswrapper[4971]: I1213 07:12:46.597456 4971 memory_manager.go:354] "RemoveStaleState removing state" podUID="166612b7-15b4-4a04-b7bb-62603363de80" containerName="placement-db-sync" Dec 13 07:12:46 crc kubenswrapper[4971]: I1213 07:12:46.598322 4971 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/placement-569b6fd56b-w7xmm" Dec 13 07:12:46 crc kubenswrapper[4971]: I1213 07:12:46.604559 4971 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"placement-config-data" Dec 13 07:12:46 crc kubenswrapper[4971]: I1213 07:12:46.604702 4971 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"placement-scripts" Dec 13 07:12:46 crc kubenswrapper[4971]: I1213 07:12:46.604800 4971 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-placement-internal-svc" Dec 13 07:12:46 crc kubenswrapper[4971]: I1213 07:12:46.604943 4971 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"placement-placement-dockercfg-tt6pj" Dec 13 07:12:46 crc kubenswrapper[4971]: I1213 07:12:46.605223 4971 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-placement-public-svc" Dec 13 07:12:46 crc kubenswrapper[4971]: I1213 07:12:46.632772 4971 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/placement-569b6fd56b-w7xmm"] Dec 13 07:12:46 crc kubenswrapper[4971]: I1213 07:12:46.742365 4971 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/abb2fa2a-6153-4b45-83b9-2c34265c5e7f-scripts\") pod \"placement-569b6fd56b-w7xmm\" (UID: \"abb2fa2a-6153-4b45-83b9-2c34265c5e7f\") " pod="openstack/placement-569b6fd56b-w7xmm" Dec 13 07:12:46 crc kubenswrapper[4971]: I1213 07:12:46.743144 4971 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-v9jrv\" (UniqueName: \"kubernetes.io/projected/abb2fa2a-6153-4b45-83b9-2c34265c5e7f-kube-api-access-v9jrv\") pod \"placement-569b6fd56b-w7xmm\" (UID: \"abb2fa2a-6153-4b45-83b9-2c34265c5e7f\") " pod="openstack/placement-569b6fd56b-w7xmm" Dec 13 07:12:46 crc kubenswrapper[4971]: I1213 07:12:46.743251 4971 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/abb2fa2a-6153-4b45-83b9-2c34265c5e7f-logs\") pod \"placement-569b6fd56b-w7xmm\" (UID: \"abb2fa2a-6153-4b45-83b9-2c34265c5e7f\") " pod="openstack/placement-569b6fd56b-w7xmm" Dec 13 07:12:46 crc kubenswrapper[4971]: I1213 07:12:46.743294 4971 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/abb2fa2a-6153-4b45-83b9-2c34265c5e7f-public-tls-certs\") pod \"placement-569b6fd56b-w7xmm\" (UID: \"abb2fa2a-6153-4b45-83b9-2c34265c5e7f\") " pod="openstack/placement-569b6fd56b-w7xmm" Dec 13 07:12:46 crc kubenswrapper[4971]: I1213 07:12:46.743368 4971 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/abb2fa2a-6153-4b45-83b9-2c34265c5e7f-combined-ca-bundle\") pod \"placement-569b6fd56b-w7xmm\" (UID: \"abb2fa2a-6153-4b45-83b9-2c34265c5e7f\") " pod="openstack/placement-569b6fd56b-w7xmm" Dec 13 07:12:46 crc kubenswrapper[4971]: I1213 07:12:46.743393 4971 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/abb2fa2a-6153-4b45-83b9-2c34265c5e7f-internal-tls-certs\") pod \"placement-569b6fd56b-w7xmm\" (UID: \"abb2fa2a-6153-4b45-83b9-2c34265c5e7f\") " pod="openstack/placement-569b6fd56b-w7xmm" Dec 13 07:12:46 crc kubenswrapper[4971]: I1213 07:12:46.743413 4971 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/abb2fa2a-6153-4b45-83b9-2c34265c5e7f-config-data\") pod \"placement-569b6fd56b-w7xmm\" (UID: \"abb2fa2a-6153-4b45-83b9-2c34265c5e7f\") " pod="openstack/placement-569b6fd56b-w7xmm" Dec 13 07:12:46 crc kubenswrapper[4971]: I1213 07:12:46.845065 4971 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/abb2fa2a-6153-4b45-83b9-2c34265c5e7f-scripts\") pod \"placement-569b6fd56b-w7xmm\" (UID: \"abb2fa2a-6153-4b45-83b9-2c34265c5e7f\") " pod="openstack/placement-569b6fd56b-w7xmm" Dec 13 07:12:46 crc kubenswrapper[4971]: I1213 07:12:46.845114 4971 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-v9jrv\" (UniqueName: \"kubernetes.io/projected/abb2fa2a-6153-4b45-83b9-2c34265c5e7f-kube-api-access-v9jrv\") pod \"placement-569b6fd56b-w7xmm\" (UID: \"abb2fa2a-6153-4b45-83b9-2c34265c5e7f\") " pod="openstack/placement-569b6fd56b-w7xmm" Dec 13 07:12:46 crc kubenswrapper[4971]: I1213 07:12:46.845178 4971 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/abb2fa2a-6153-4b45-83b9-2c34265c5e7f-logs\") pod \"placement-569b6fd56b-w7xmm\" (UID: \"abb2fa2a-6153-4b45-83b9-2c34265c5e7f\") " pod="openstack/placement-569b6fd56b-w7xmm" Dec 13 07:12:46 crc kubenswrapper[4971]: I1213 07:12:46.845205 4971 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/abb2fa2a-6153-4b45-83b9-2c34265c5e7f-public-tls-certs\") pod \"placement-569b6fd56b-w7xmm\" (UID: \"abb2fa2a-6153-4b45-83b9-2c34265c5e7f\") " pod="openstack/placement-569b6fd56b-w7xmm" Dec 13 07:12:46 crc kubenswrapper[4971]: I1213 07:12:46.845271 4971 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/abb2fa2a-6153-4b45-83b9-2c34265c5e7f-combined-ca-bundle\") pod \"placement-569b6fd56b-w7xmm\" (UID: \"abb2fa2a-6153-4b45-83b9-2c34265c5e7f\") " pod="openstack/placement-569b6fd56b-w7xmm" Dec 13 07:12:46 crc kubenswrapper[4971]: I1213 07:12:46.845290 4971 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/abb2fa2a-6153-4b45-83b9-2c34265c5e7f-internal-tls-certs\") pod \"placement-569b6fd56b-w7xmm\" (UID: \"abb2fa2a-6153-4b45-83b9-2c34265c5e7f\") " pod="openstack/placement-569b6fd56b-w7xmm" Dec 13 07:12:46 crc kubenswrapper[4971]: I1213 07:12:46.845312 4971 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/abb2fa2a-6153-4b45-83b9-2c34265c5e7f-config-data\") pod \"placement-569b6fd56b-w7xmm\" (UID: \"abb2fa2a-6153-4b45-83b9-2c34265c5e7f\") " pod="openstack/placement-569b6fd56b-w7xmm" Dec 13 07:12:46 crc kubenswrapper[4971]: I1213 07:12:46.846020 4971 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/abb2fa2a-6153-4b45-83b9-2c34265c5e7f-logs\") pod \"placement-569b6fd56b-w7xmm\" (UID: \"abb2fa2a-6153-4b45-83b9-2c34265c5e7f\") " pod="openstack/placement-569b6fd56b-w7xmm" Dec 13 07:12:46 crc kubenswrapper[4971]: I1213 07:12:46.850724 4971 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/abb2fa2a-6153-4b45-83b9-2c34265c5e7f-public-tls-certs\") pod \"placement-569b6fd56b-w7xmm\" (UID: \"abb2fa2a-6153-4b45-83b9-2c34265c5e7f\") " pod="openstack/placement-569b6fd56b-w7xmm" Dec 13 07:12:46 crc kubenswrapper[4971]: I1213 07:12:46.853663 4971 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/abb2fa2a-6153-4b45-83b9-2c34265c5e7f-internal-tls-certs\") pod \"placement-569b6fd56b-w7xmm\" (UID: \"abb2fa2a-6153-4b45-83b9-2c34265c5e7f\") " pod="openstack/placement-569b6fd56b-w7xmm" Dec 13 07:12:46 crc kubenswrapper[4971]: I1213 07:12:46.854343 4971 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/abb2fa2a-6153-4b45-83b9-2c34265c5e7f-config-data\") pod \"placement-569b6fd56b-w7xmm\" (UID: \"abb2fa2a-6153-4b45-83b9-2c34265c5e7f\") " pod="openstack/placement-569b6fd56b-w7xmm" Dec 13 07:12:46 crc kubenswrapper[4971]: I1213 07:12:46.855952 4971 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/abb2fa2a-6153-4b45-83b9-2c34265c5e7f-scripts\") pod \"placement-569b6fd56b-w7xmm\" (UID: \"abb2fa2a-6153-4b45-83b9-2c34265c5e7f\") " pod="openstack/placement-569b6fd56b-w7xmm" Dec 13 07:12:46 crc kubenswrapper[4971]: I1213 07:12:46.864476 4971 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/abb2fa2a-6153-4b45-83b9-2c34265c5e7f-combined-ca-bundle\") pod \"placement-569b6fd56b-w7xmm\" (UID: \"abb2fa2a-6153-4b45-83b9-2c34265c5e7f\") " pod="openstack/placement-569b6fd56b-w7xmm" Dec 13 07:12:46 crc kubenswrapper[4971]: I1213 07:12:46.866704 4971 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-v9jrv\" (UniqueName: \"kubernetes.io/projected/abb2fa2a-6153-4b45-83b9-2c34265c5e7f-kube-api-access-v9jrv\") pod \"placement-569b6fd56b-w7xmm\" (UID: \"abb2fa2a-6153-4b45-83b9-2c34265c5e7f\") " pod="openstack/placement-569b6fd56b-w7xmm" Dec 13 07:12:46 crc kubenswrapper[4971]: I1213 07:12:46.923135 4971 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/placement-569b6fd56b-w7xmm" Dec 13 07:12:46 crc kubenswrapper[4971]: I1213 07:12:46.990884 4971 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-7956479f45-8g7zz" event={"ID":"4426fcd4-0b89-4e70-8f2f-43b056896953","Type":"ContainerStarted","Data":"c45e392e62ed479628ba5dbd140d2bfaf321334557863ff2f5f87e19b7115c87"} Dec 13 07:12:46 crc kubenswrapper[4971]: I1213 07:12:46.990946 4971 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-7956479f45-8g7zz" event={"ID":"4426fcd4-0b89-4e70-8f2f-43b056896953","Type":"ContainerStarted","Data":"c4b30f0daac5d2ffec441f8915af3b0948797d66aaea773751f784a440a69156"} Dec 13 07:12:46 crc kubenswrapper[4971]: I1213 07:12:46.991158 4971 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/neutron-7956479f45-8g7zz" Dec 13 07:12:47 crc kubenswrapper[4971]: I1213 07:12:47.016752 4971 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/neutron-7956479f45-8g7zz" podStartSLOduration=6.016731755 podStartE2EDuration="6.016731755s" podCreationTimestamp="2025-12-13 07:12:41 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-13 07:12:47.011219319 +0000 UTC m=+1423.615628767" watchObservedRunningTime="2025-12-13 07:12:47.016731755 +0000 UTC m=+1423.621141203" Dec 13 07:12:47 crc kubenswrapper[4971]: I1213 07:12:47.564648 4971 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-bootstrap-dn9h9" Dec 13 07:12:47 crc kubenswrapper[4971]: I1213 07:12:47.662728 4971 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/ee3afc2f-79c6-444c-b6a3-2d4c475c3aac-combined-ca-bundle\") pod \"ee3afc2f-79c6-444c-b6a3-2d4c475c3aac\" (UID: \"ee3afc2f-79c6-444c-b6a3-2d4c475c3aac\") " Dec 13 07:12:47 crc kubenswrapper[4971]: I1213 07:12:47.662954 4971 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-mdl24\" (UniqueName: \"kubernetes.io/projected/ee3afc2f-79c6-444c-b6a3-2d4c475c3aac-kube-api-access-mdl24\") pod \"ee3afc2f-79c6-444c-b6a3-2d4c475c3aac\" (UID: \"ee3afc2f-79c6-444c-b6a3-2d4c475c3aac\") " Dec 13 07:12:47 crc kubenswrapper[4971]: I1213 07:12:47.662994 4971 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"credential-keys\" (UniqueName: \"kubernetes.io/secret/ee3afc2f-79c6-444c-b6a3-2d4c475c3aac-credential-keys\") pod \"ee3afc2f-79c6-444c-b6a3-2d4c475c3aac\" (UID: \"ee3afc2f-79c6-444c-b6a3-2d4c475c3aac\") " Dec 13 07:12:47 crc kubenswrapper[4971]: I1213 07:12:47.663018 4971 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/ee3afc2f-79c6-444c-b6a3-2d4c475c3aac-scripts\") pod \"ee3afc2f-79c6-444c-b6a3-2d4c475c3aac\" (UID: \"ee3afc2f-79c6-444c-b6a3-2d4c475c3aac\") " Dec 13 07:12:47 crc kubenswrapper[4971]: I1213 07:12:47.663077 4971 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/ee3afc2f-79c6-444c-b6a3-2d4c475c3aac-config-data\") pod \"ee3afc2f-79c6-444c-b6a3-2d4c475c3aac\" (UID: \"ee3afc2f-79c6-444c-b6a3-2d4c475c3aac\") " Dec 13 07:12:47 crc kubenswrapper[4971]: I1213 07:12:47.663184 4971 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/ee3afc2f-79c6-444c-b6a3-2d4c475c3aac-fernet-keys\") pod \"ee3afc2f-79c6-444c-b6a3-2d4c475c3aac\" (UID: \"ee3afc2f-79c6-444c-b6a3-2d4c475c3aac\") " Dec 13 07:12:47 crc kubenswrapper[4971]: I1213 07:12:47.678433 4971 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/ee3afc2f-79c6-444c-b6a3-2d4c475c3aac-kube-api-access-mdl24" (OuterVolumeSpecName: "kube-api-access-mdl24") pod "ee3afc2f-79c6-444c-b6a3-2d4c475c3aac" (UID: "ee3afc2f-79c6-444c-b6a3-2d4c475c3aac"). InnerVolumeSpecName "kube-api-access-mdl24". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 13 07:12:47 crc kubenswrapper[4971]: I1213 07:12:47.683710 4971 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/ee3afc2f-79c6-444c-b6a3-2d4c475c3aac-credential-keys" (OuterVolumeSpecName: "credential-keys") pod "ee3afc2f-79c6-444c-b6a3-2d4c475c3aac" (UID: "ee3afc2f-79c6-444c-b6a3-2d4c475c3aac"). InnerVolumeSpecName "credential-keys". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 13 07:12:47 crc kubenswrapper[4971]: I1213 07:12:47.691793 4971 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/ee3afc2f-79c6-444c-b6a3-2d4c475c3aac-scripts" (OuterVolumeSpecName: "scripts") pod "ee3afc2f-79c6-444c-b6a3-2d4c475c3aac" (UID: "ee3afc2f-79c6-444c-b6a3-2d4c475c3aac"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 13 07:12:47 crc kubenswrapper[4971]: I1213 07:12:47.695811 4971 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/placement-569b6fd56b-w7xmm"] Dec 13 07:12:47 crc kubenswrapper[4971]: I1213 07:12:47.696435 4971 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/ee3afc2f-79c6-444c-b6a3-2d4c475c3aac-fernet-keys" (OuterVolumeSpecName: "fernet-keys") pod "ee3afc2f-79c6-444c-b6a3-2d4c475c3aac" (UID: "ee3afc2f-79c6-444c-b6a3-2d4c475c3aac"). InnerVolumeSpecName "fernet-keys". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 13 07:12:47 crc kubenswrapper[4971]: I1213 07:12:47.700340 4971 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/ee3afc2f-79c6-444c-b6a3-2d4c475c3aac-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "ee3afc2f-79c6-444c-b6a3-2d4c475c3aac" (UID: "ee3afc2f-79c6-444c-b6a3-2d4c475c3aac"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 13 07:12:47 crc kubenswrapper[4971]: W1213 07:12:47.705015 4971 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podabb2fa2a_6153_4b45_83b9_2c34265c5e7f.slice/crio-b012f20c9a1a93937ab3316c741adad6ca07ec72958a3ded5a9f1a93c6927e15 WatchSource:0}: Error finding container b012f20c9a1a93937ab3316c741adad6ca07ec72958a3ded5a9f1a93c6927e15: Status 404 returned error can't find the container with id b012f20c9a1a93937ab3316c741adad6ca07ec72958a3ded5a9f1a93c6927e15 Dec 13 07:12:47 crc kubenswrapper[4971]: I1213 07:12:47.708976 4971 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/ee3afc2f-79c6-444c-b6a3-2d4c475c3aac-config-data" (OuterVolumeSpecName: "config-data") pod "ee3afc2f-79c6-444c-b6a3-2d4c475c3aac" (UID: "ee3afc2f-79c6-444c-b6a3-2d4c475c3aac"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 13 07:12:47 crc kubenswrapper[4971]: I1213 07:12:47.766291 4971 reconciler_common.go:293] "Volume detached for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/ee3afc2f-79c6-444c-b6a3-2d4c475c3aac-fernet-keys\") on node \"crc\" DevicePath \"\"" Dec 13 07:12:47 crc kubenswrapper[4971]: I1213 07:12:47.766337 4971 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/ee3afc2f-79c6-444c-b6a3-2d4c475c3aac-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 13 07:12:47 crc kubenswrapper[4971]: I1213 07:12:47.766353 4971 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-mdl24\" (UniqueName: \"kubernetes.io/projected/ee3afc2f-79c6-444c-b6a3-2d4c475c3aac-kube-api-access-mdl24\") on node \"crc\" DevicePath \"\"" Dec 13 07:12:47 crc kubenswrapper[4971]: I1213 07:12:47.766364 4971 reconciler_common.go:293] "Volume detached for volume \"credential-keys\" (UniqueName: \"kubernetes.io/secret/ee3afc2f-79c6-444c-b6a3-2d4c475c3aac-credential-keys\") on node \"crc\" DevicePath \"\"" Dec 13 07:12:47 crc kubenswrapper[4971]: I1213 07:12:47.766382 4971 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/ee3afc2f-79c6-444c-b6a3-2d4c475c3aac-scripts\") on node \"crc\" DevicePath \"\"" Dec 13 07:12:47 crc kubenswrapper[4971]: I1213 07:12:47.766395 4971 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/ee3afc2f-79c6-444c-b6a3-2d4c475c3aac-config-data\") on node \"crc\" DevicePath \"\"" Dec 13 07:12:48 crc kubenswrapper[4971]: I1213 07:12:48.036558 4971 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-bootstrap-dn9h9" event={"ID":"ee3afc2f-79c6-444c-b6a3-2d4c475c3aac","Type":"ContainerDied","Data":"ac05c4b6e27e476542a19f5f67b2b66dacc449e38088c1b8d13ff803b99435e7"} Dec 13 07:12:48 crc kubenswrapper[4971]: I1213 07:12:48.036900 4971 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="ac05c4b6e27e476542a19f5f67b2b66dacc449e38088c1b8d13ff803b99435e7" Dec 13 07:12:48 crc kubenswrapper[4971]: I1213 07:12:48.036637 4971 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-bootstrap-dn9h9" Dec 13 07:12:48 crc kubenswrapper[4971]: I1213 07:12:48.040448 4971 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/placement-569b6fd56b-w7xmm" event={"ID":"abb2fa2a-6153-4b45-83b9-2c34265c5e7f","Type":"ContainerStarted","Data":"e3f4545f96d1db7e0ca549ae5f8264e88b80af8e47c11d042b98a5fe5f270439"} Dec 13 07:12:48 crc kubenswrapper[4971]: I1213 07:12:48.040539 4971 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/placement-569b6fd56b-w7xmm" event={"ID":"abb2fa2a-6153-4b45-83b9-2c34265c5e7f","Type":"ContainerStarted","Data":"b012f20c9a1a93937ab3316c741adad6ca07ec72958a3ded5a9f1a93c6927e15"} Dec 13 07:12:48 crc kubenswrapper[4971]: I1213 07:12:48.083142 4971 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/keystone-7969997c86-5t8c8"] Dec 13 07:12:48 crc kubenswrapper[4971]: E1213 07:12:48.083628 4971 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="ee3afc2f-79c6-444c-b6a3-2d4c475c3aac" containerName="keystone-bootstrap" Dec 13 07:12:48 crc kubenswrapper[4971]: I1213 07:12:48.083653 4971 state_mem.go:107] "Deleted CPUSet assignment" podUID="ee3afc2f-79c6-444c-b6a3-2d4c475c3aac" containerName="keystone-bootstrap" Dec 13 07:12:48 crc kubenswrapper[4971]: I1213 07:12:48.083863 4971 memory_manager.go:354] "RemoveStaleState removing state" podUID="ee3afc2f-79c6-444c-b6a3-2d4c475c3aac" containerName="keystone-bootstrap" Dec 13 07:12:48 crc kubenswrapper[4971]: I1213 07:12:48.084566 4971 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-7969997c86-5t8c8" Dec 13 07:12:48 crc kubenswrapper[4971]: I1213 07:12:48.088896 4971 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone-scripts" Dec 13 07:12:48 crc kubenswrapper[4971]: I1213 07:12:48.089193 4971 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone-keystone-dockercfg-56xxb" Dec 13 07:12:48 crc kubenswrapper[4971]: I1213 07:12:48.089394 4971 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-keystone-public-svc" Dec 13 07:12:48 crc kubenswrapper[4971]: I1213 07:12:48.089403 4971 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone-config-data" Dec 13 07:12:48 crc kubenswrapper[4971]: I1213 07:12:48.089574 4971 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-keystone-internal-svc" Dec 13 07:12:48 crc kubenswrapper[4971]: I1213 07:12:48.089635 4971 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone" Dec 13 07:12:48 crc kubenswrapper[4971]: I1213 07:12:48.100582 4971 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/keystone-7969997c86-5t8c8"] Dec 13 07:12:48 crc kubenswrapper[4971]: I1213 07:12:48.179773 4971 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"credential-keys\" (UniqueName: \"kubernetes.io/secret/8e1b68c0-d925-4c7c-8904-d71b9f6ef0e1-credential-keys\") pod \"keystone-7969997c86-5t8c8\" (UID: \"8e1b68c0-d925-4c7c-8904-d71b9f6ef0e1\") " pod="openstack/keystone-7969997c86-5t8c8" Dec 13 07:12:48 crc kubenswrapper[4971]: I1213 07:12:48.179832 4971 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/8e1b68c0-d925-4c7c-8904-d71b9f6ef0e1-public-tls-certs\") pod \"keystone-7969997c86-5t8c8\" (UID: \"8e1b68c0-d925-4c7c-8904-d71b9f6ef0e1\") " pod="openstack/keystone-7969997c86-5t8c8" Dec 13 07:12:48 crc kubenswrapper[4971]: I1213 07:12:48.179868 4971 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/8e1b68c0-d925-4c7c-8904-d71b9f6ef0e1-combined-ca-bundle\") pod \"keystone-7969997c86-5t8c8\" (UID: \"8e1b68c0-d925-4c7c-8904-d71b9f6ef0e1\") " pod="openstack/keystone-7969997c86-5t8c8" Dec 13 07:12:48 crc kubenswrapper[4971]: I1213 07:12:48.179893 4971 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/8e1b68c0-d925-4c7c-8904-d71b9f6ef0e1-internal-tls-certs\") pod \"keystone-7969997c86-5t8c8\" (UID: \"8e1b68c0-d925-4c7c-8904-d71b9f6ef0e1\") " pod="openstack/keystone-7969997c86-5t8c8" Dec 13 07:12:48 crc kubenswrapper[4971]: I1213 07:12:48.179920 4971 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-tqjsk\" (UniqueName: \"kubernetes.io/projected/8e1b68c0-d925-4c7c-8904-d71b9f6ef0e1-kube-api-access-tqjsk\") pod \"keystone-7969997c86-5t8c8\" (UID: \"8e1b68c0-d925-4c7c-8904-d71b9f6ef0e1\") " pod="openstack/keystone-7969997c86-5t8c8" Dec 13 07:12:48 crc kubenswrapper[4971]: I1213 07:12:48.179985 4971 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/8e1b68c0-d925-4c7c-8904-d71b9f6ef0e1-scripts\") pod \"keystone-7969997c86-5t8c8\" (UID: \"8e1b68c0-d925-4c7c-8904-d71b9f6ef0e1\") " pod="openstack/keystone-7969997c86-5t8c8" Dec 13 07:12:48 crc kubenswrapper[4971]: I1213 07:12:48.180036 4971 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/8e1b68c0-d925-4c7c-8904-d71b9f6ef0e1-config-data\") pod \"keystone-7969997c86-5t8c8\" (UID: \"8e1b68c0-d925-4c7c-8904-d71b9f6ef0e1\") " pod="openstack/keystone-7969997c86-5t8c8" Dec 13 07:12:48 crc kubenswrapper[4971]: I1213 07:12:48.180083 4971 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/8e1b68c0-d925-4c7c-8904-d71b9f6ef0e1-fernet-keys\") pod \"keystone-7969997c86-5t8c8\" (UID: \"8e1b68c0-d925-4c7c-8904-d71b9f6ef0e1\") " pod="openstack/keystone-7969997c86-5t8c8" Dec 13 07:12:48 crc kubenswrapper[4971]: I1213 07:12:48.281187 4971 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"credential-keys\" (UniqueName: \"kubernetes.io/secret/8e1b68c0-d925-4c7c-8904-d71b9f6ef0e1-credential-keys\") pod \"keystone-7969997c86-5t8c8\" (UID: \"8e1b68c0-d925-4c7c-8904-d71b9f6ef0e1\") " pod="openstack/keystone-7969997c86-5t8c8" Dec 13 07:12:48 crc kubenswrapper[4971]: I1213 07:12:48.281246 4971 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/8e1b68c0-d925-4c7c-8904-d71b9f6ef0e1-public-tls-certs\") pod \"keystone-7969997c86-5t8c8\" (UID: \"8e1b68c0-d925-4c7c-8904-d71b9f6ef0e1\") " pod="openstack/keystone-7969997c86-5t8c8" Dec 13 07:12:48 crc kubenswrapper[4971]: I1213 07:12:48.281281 4971 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/8e1b68c0-d925-4c7c-8904-d71b9f6ef0e1-combined-ca-bundle\") pod \"keystone-7969997c86-5t8c8\" (UID: \"8e1b68c0-d925-4c7c-8904-d71b9f6ef0e1\") " pod="openstack/keystone-7969997c86-5t8c8" Dec 13 07:12:48 crc kubenswrapper[4971]: I1213 07:12:48.281304 4971 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/8e1b68c0-d925-4c7c-8904-d71b9f6ef0e1-internal-tls-certs\") pod \"keystone-7969997c86-5t8c8\" (UID: \"8e1b68c0-d925-4c7c-8904-d71b9f6ef0e1\") " pod="openstack/keystone-7969997c86-5t8c8" Dec 13 07:12:48 crc kubenswrapper[4971]: I1213 07:12:48.281329 4971 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-tqjsk\" (UniqueName: \"kubernetes.io/projected/8e1b68c0-d925-4c7c-8904-d71b9f6ef0e1-kube-api-access-tqjsk\") pod \"keystone-7969997c86-5t8c8\" (UID: \"8e1b68c0-d925-4c7c-8904-d71b9f6ef0e1\") " pod="openstack/keystone-7969997c86-5t8c8" Dec 13 07:12:48 crc kubenswrapper[4971]: I1213 07:12:48.281395 4971 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/8e1b68c0-d925-4c7c-8904-d71b9f6ef0e1-scripts\") pod \"keystone-7969997c86-5t8c8\" (UID: \"8e1b68c0-d925-4c7c-8904-d71b9f6ef0e1\") " pod="openstack/keystone-7969997c86-5t8c8" Dec 13 07:12:48 crc kubenswrapper[4971]: I1213 07:12:48.281442 4971 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/8e1b68c0-d925-4c7c-8904-d71b9f6ef0e1-config-data\") pod \"keystone-7969997c86-5t8c8\" (UID: \"8e1b68c0-d925-4c7c-8904-d71b9f6ef0e1\") " pod="openstack/keystone-7969997c86-5t8c8" Dec 13 07:12:48 crc kubenswrapper[4971]: I1213 07:12:48.281490 4971 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/8e1b68c0-d925-4c7c-8904-d71b9f6ef0e1-fernet-keys\") pod \"keystone-7969997c86-5t8c8\" (UID: \"8e1b68c0-d925-4c7c-8904-d71b9f6ef0e1\") " pod="openstack/keystone-7969997c86-5t8c8" Dec 13 07:12:48 crc kubenswrapper[4971]: I1213 07:12:48.298327 4971 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/8e1b68c0-d925-4c7c-8904-d71b9f6ef0e1-scripts\") pod \"keystone-7969997c86-5t8c8\" (UID: \"8e1b68c0-d925-4c7c-8904-d71b9f6ef0e1\") " pod="openstack/keystone-7969997c86-5t8c8" Dec 13 07:12:48 crc kubenswrapper[4971]: I1213 07:12:48.298359 4971 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"credential-keys\" (UniqueName: \"kubernetes.io/secret/8e1b68c0-d925-4c7c-8904-d71b9f6ef0e1-credential-keys\") pod \"keystone-7969997c86-5t8c8\" (UID: \"8e1b68c0-d925-4c7c-8904-d71b9f6ef0e1\") " pod="openstack/keystone-7969997c86-5t8c8" Dec 13 07:12:48 crc kubenswrapper[4971]: I1213 07:12:48.298591 4971 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/8e1b68c0-d925-4c7c-8904-d71b9f6ef0e1-internal-tls-certs\") pod \"keystone-7969997c86-5t8c8\" (UID: \"8e1b68c0-d925-4c7c-8904-d71b9f6ef0e1\") " pod="openstack/keystone-7969997c86-5t8c8" Dec 13 07:12:48 crc kubenswrapper[4971]: I1213 07:12:48.299842 4971 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/8e1b68c0-d925-4c7c-8904-d71b9f6ef0e1-combined-ca-bundle\") pod \"keystone-7969997c86-5t8c8\" (UID: \"8e1b68c0-d925-4c7c-8904-d71b9f6ef0e1\") " pod="openstack/keystone-7969997c86-5t8c8" Dec 13 07:12:48 crc kubenswrapper[4971]: I1213 07:12:48.300262 4971 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/8e1b68c0-d925-4c7c-8904-d71b9f6ef0e1-fernet-keys\") pod \"keystone-7969997c86-5t8c8\" (UID: \"8e1b68c0-d925-4c7c-8904-d71b9f6ef0e1\") " pod="openstack/keystone-7969997c86-5t8c8" Dec 13 07:12:48 crc kubenswrapper[4971]: I1213 07:12:48.301077 4971 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/8e1b68c0-d925-4c7c-8904-d71b9f6ef0e1-public-tls-certs\") pod \"keystone-7969997c86-5t8c8\" (UID: \"8e1b68c0-d925-4c7c-8904-d71b9f6ef0e1\") " pod="openstack/keystone-7969997c86-5t8c8" Dec 13 07:12:48 crc kubenswrapper[4971]: I1213 07:12:48.301345 4971 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/8e1b68c0-d925-4c7c-8904-d71b9f6ef0e1-config-data\") pod \"keystone-7969997c86-5t8c8\" (UID: \"8e1b68c0-d925-4c7c-8904-d71b9f6ef0e1\") " pod="openstack/keystone-7969997c86-5t8c8" Dec 13 07:12:48 crc kubenswrapper[4971]: I1213 07:12:48.313252 4971 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-tqjsk\" (UniqueName: \"kubernetes.io/projected/8e1b68c0-d925-4c7c-8904-d71b9f6ef0e1-kube-api-access-tqjsk\") pod \"keystone-7969997c86-5t8c8\" (UID: \"8e1b68c0-d925-4c7c-8904-d71b9f6ef0e1\") " pod="openstack/keystone-7969997c86-5t8c8" Dec 13 07:12:48 crc kubenswrapper[4971]: I1213 07:12:48.408072 4971 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-7969997c86-5t8c8" Dec 13 07:12:49 crc kubenswrapper[4971]: I1213 07:12:49.388752 4971 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/dnsmasq-dns-548894858c-j5bhs" Dec 13 07:12:49 crc kubenswrapper[4971]: I1213 07:12:49.476109 4971 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-5d87b7c6dc-sxklp"] Dec 13 07:12:49 crc kubenswrapper[4971]: I1213 07:12:49.476338 4971 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/dnsmasq-dns-5d87b7c6dc-sxklp" podUID="2430f20e-0768-47e9-89de-fc6ac2615707" containerName="dnsmasq-dns" containerID="cri-o://c70e97bcd1c39a27d0495531710de58ef3511f93ac0bb637c42dd96a63642cc8" gracePeriod=10 Dec 13 07:12:50 crc kubenswrapper[4971]: I1213 07:12:50.067115 4971 generic.go:334] "Generic (PLEG): container finished" podID="2430f20e-0768-47e9-89de-fc6ac2615707" containerID="c70e97bcd1c39a27d0495531710de58ef3511f93ac0bb637c42dd96a63642cc8" exitCode=0 Dec 13 07:12:50 crc kubenswrapper[4971]: I1213 07:12:50.067156 4971 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-5d87b7c6dc-sxklp" event={"ID":"2430f20e-0768-47e9-89de-fc6ac2615707","Type":"ContainerDied","Data":"c70e97bcd1c39a27d0495531710de58ef3511f93ac0bb637c42dd96a63642cc8"} Dec 13 07:12:52 crc kubenswrapper[4971]: I1213 07:12:52.660065 4971 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-5d87b7c6dc-sxklp" Dec 13 07:12:52 crc kubenswrapper[4971]: I1213 07:12:52.787197 4971 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/2430f20e-0768-47e9-89de-fc6ac2615707-ovsdbserver-nb\") pod \"2430f20e-0768-47e9-89de-fc6ac2615707\" (UID: \"2430f20e-0768-47e9-89de-fc6ac2615707\") " Dec 13 07:12:52 crc kubenswrapper[4971]: I1213 07:12:52.787240 4971 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/2430f20e-0768-47e9-89de-fc6ac2615707-config\") pod \"2430f20e-0768-47e9-89de-fc6ac2615707\" (UID: \"2430f20e-0768-47e9-89de-fc6ac2615707\") " Dec 13 07:12:52 crc kubenswrapper[4971]: I1213 07:12:52.787283 4971 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/2430f20e-0768-47e9-89de-fc6ac2615707-dns-svc\") pod \"2430f20e-0768-47e9-89de-fc6ac2615707\" (UID: \"2430f20e-0768-47e9-89de-fc6ac2615707\") " Dec 13 07:12:52 crc kubenswrapper[4971]: I1213 07:12:52.787344 4971 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/2430f20e-0768-47e9-89de-fc6ac2615707-ovsdbserver-sb\") pod \"2430f20e-0768-47e9-89de-fc6ac2615707\" (UID: \"2430f20e-0768-47e9-89de-fc6ac2615707\") " Dec 13 07:12:52 crc kubenswrapper[4971]: I1213 07:12:52.787562 4971 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-97xnq\" (UniqueName: \"kubernetes.io/projected/2430f20e-0768-47e9-89de-fc6ac2615707-kube-api-access-97xnq\") pod \"2430f20e-0768-47e9-89de-fc6ac2615707\" (UID: \"2430f20e-0768-47e9-89de-fc6ac2615707\") " Dec 13 07:12:52 crc kubenswrapper[4971]: I1213 07:12:52.793971 4971 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/2430f20e-0768-47e9-89de-fc6ac2615707-kube-api-access-97xnq" (OuterVolumeSpecName: "kube-api-access-97xnq") pod "2430f20e-0768-47e9-89de-fc6ac2615707" (UID: "2430f20e-0768-47e9-89de-fc6ac2615707"). InnerVolumeSpecName "kube-api-access-97xnq". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 13 07:12:52 crc kubenswrapper[4971]: I1213 07:12:52.843490 4971 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/2430f20e-0768-47e9-89de-fc6ac2615707-ovsdbserver-sb" (OuterVolumeSpecName: "ovsdbserver-sb") pod "2430f20e-0768-47e9-89de-fc6ac2615707" (UID: "2430f20e-0768-47e9-89de-fc6ac2615707"). InnerVolumeSpecName "ovsdbserver-sb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 13 07:12:52 crc kubenswrapper[4971]: I1213 07:12:52.863486 4971 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/2430f20e-0768-47e9-89de-fc6ac2615707-config" (OuterVolumeSpecName: "config") pod "2430f20e-0768-47e9-89de-fc6ac2615707" (UID: "2430f20e-0768-47e9-89de-fc6ac2615707"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 13 07:12:52 crc kubenswrapper[4971]: I1213 07:12:52.864668 4971 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/2430f20e-0768-47e9-89de-fc6ac2615707-dns-svc" (OuterVolumeSpecName: "dns-svc") pod "2430f20e-0768-47e9-89de-fc6ac2615707" (UID: "2430f20e-0768-47e9-89de-fc6ac2615707"). InnerVolumeSpecName "dns-svc". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 13 07:12:52 crc kubenswrapper[4971]: I1213 07:12:52.888296 4971 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/2430f20e-0768-47e9-89de-fc6ac2615707-ovsdbserver-nb" (OuterVolumeSpecName: "ovsdbserver-nb") pod "2430f20e-0768-47e9-89de-fc6ac2615707" (UID: "2430f20e-0768-47e9-89de-fc6ac2615707"). InnerVolumeSpecName "ovsdbserver-nb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 13 07:12:52 crc kubenswrapper[4971]: I1213 07:12:52.890034 4971 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-97xnq\" (UniqueName: \"kubernetes.io/projected/2430f20e-0768-47e9-89de-fc6ac2615707-kube-api-access-97xnq\") on node \"crc\" DevicePath \"\"" Dec 13 07:12:52 crc kubenswrapper[4971]: I1213 07:12:52.890061 4971 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/2430f20e-0768-47e9-89de-fc6ac2615707-ovsdbserver-nb\") on node \"crc\" DevicePath \"\"" Dec 13 07:12:52 crc kubenswrapper[4971]: I1213 07:12:52.890072 4971 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/2430f20e-0768-47e9-89de-fc6ac2615707-config\") on node \"crc\" DevicePath \"\"" Dec 13 07:12:52 crc kubenswrapper[4971]: I1213 07:12:52.890081 4971 reconciler_common.go:293] "Volume detached for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/2430f20e-0768-47e9-89de-fc6ac2615707-dns-svc\") on node \"crc\" DevicePath \"\"" Dec 13 07:12:52 crc kubenswrapper[4971]: I1213 07:12:52.890089 4971 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/2430f20e-0768-47e9-89de-fc6ac2615707-ovsdbserver-sb\") on node \"crc\" DevicePath \"\"" Dec 13 07:12:53 crc kubenswrapper[4971]: I1213 07:12:53.074123 4971 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/keystone-7969997c86-5t8c8"] Dec 13 07:12:53 crc kubenswrapper[4971]: W1213 07:12:53.101010 4971 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod8e1b68c0_d925_4c7c_8904_d71b9f6ef0e1.slice/crio-47089d2c08cf243bff22c10b9c00292813f4154a3a859b1e7c0f63d830006397 WatchSource:0}: Error finding container 47089d2c08cf243bff22c10b9c00292813f4154a3a859b1e7c0f63d830006397: Status 404 returned error can't find the container with id 47089d2c08cf243bff22c10b9c00292813f4154a3a859b1e7c0f63d830006397 Dec 13 07:12:53 crc kubenswrapper[4971]: I1213 07:12:53.101997 4971 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-5d87b7c6dc-sxklp" event={"ID":"2430f20e-0768-47e9-89de-fc6ac2615707","Type":"ContainerDied","Data":"e87b3742a8d0d0c1e72b0732bb0b473a60d040294def9f4ace6b98cb9c23f2cf"} Dec 13 07:12:53 crc kubenswrapper[4971]: I1213 07:12:53.102210 4971 scope.go:117] "RemoveContainer" containerID="c70e97bcd1c39a27d0495531710de58ef3511f93ac0bb637c42dd96a63642cc8" Dec 13 07:12:53 crc kubenswrapper[4971]: I1213 07:12:53.102066 4971 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-5d87b7c6dc-sxklp" Dec 13 07:12:53 crc kubenswrapper[4971]: I1213 07:12:53.157809 4971 scope.go:117] "RemoveContainer" containerID="4f6c4e975a707094499c418eb050018f9723b4eb8f77c6b7f44577b34b9736cd" Dec 13 07:12:53 crc kubenswrapper[4971]: I1213 07:12:53.171170 4971 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-5d87b7c6dc-sxklp"] Dec 13 07:12:53 crc kubenswrapper[4971]: I1213 07:12:53.187649 4971 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-5d87b7c6dc-sxklp"] Dec 13 07:12:53 crc kubenswrapper[4971]: I1213 07:12:53.885591 4971 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="2430f20e-0768-47e9-89de-fc6ac2615707" path="/var/lib/kubelet/pods/2430f20e-0768-47e9-89de-fc6ac2615707/volumes" Dec 13 07:12:54 crc kubenswrapper[4971]: I1213 07:12:54.039396 4971 prober.go:107] "Probe failed" probeType="Startup" pod="openstack/horizon-8699dc887d-sl2h6" podUID="772eedab-7c62-454b-b722-ee453b39e89f" containerName="horizon" probeResult="failure" output="Get \"https://10.217.0.140:8443/dashboard/auth/login/?next=/dashboard/\": dial tcp 10.217.0.140:8443: connect: connection refused" Dec 13 07:12:54 crc kubenswrapper[4971]: I1213 07:12:54.119075 4971 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-db-sync-26k47" event={"ID":"c1d819d0-339f-4e7e-9180-63d2f7c0042e","Type":"ContainerStarted","Data":"3397bd79f11b055f37e31637c81e296356926fc8aaa643a9a2f444c08a08c661"} Dec 13 07:12:54 crc kubenswrapper[4971]: I1213 07:12:54.121991 4971 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/placement-569b6fd56b-w7xmm" event={"ID":"abb2fa2a-6153-4b45-83b9-2c34265c5e7f","Type":"ContainerStarted","Data":"748cea00783fe58eba1b86f9c4178bc2df2668c5d9c778525a425a5c3932ecc1"} Dec 13 07:12:54 crc kubenswrapper[4971]: I1213 07:12:54.122805 4971 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/placement-569b6fd56b-w7xmm" Dec 13 07:12:54 crc kubenswrapper[4971]: I1213 07:12:54.122845 4971 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/placement-569b6fd56b-w7xmm" Dec 13 07:12:54 crc kubenswrapper[4971]: I1213 07:12:54.130202 4971 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-db-sync-dvchm" event={"ID":"d7cf1b55-0dae-4a17-857a-6336fe71ff24","Type":"ContainerStarted","Data":"0b05501843e82c1e08bcb4071dfc72aa660772984720ae7dcfa4a26088cbd9be"} Dec 13 07:12:54 crc kubenswrapper[4971]: I1213 07:12:54.143198 4971 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"56e56ce1-4dd1-4ae5-b868-e98d27d5ad94","Type":"ContainerStarted","Data":"a0e5a1ecd136567279050709089de36451185ac24d99f2e15b755c1d0658ad99"} Dec 13 07:12:54 crc kubenswrapper[4971]: I1213 07:12:54.145245 4971 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-7969997c86-5t8c8" event={"ID":"8e1b68c0-d925-4c7c-8904-d71b9f6ef0e1","Type":"ContainerStarted","Data":"1f84bf0903ab0e455d6f38dcd80a1d6a3a5718ea8ce53a789c7291cbfc7fb6cf"} Dec 13 07:12:54 crc kubenswrapper[4971]: I1213 07:12:54.145284 4971 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-7969997c86-5t8c8" event={"ID":"8e1b68c0-d925-4c7c-8904-d71b9f6ef0e1","Type":"ContainerStarted","Data":"47089d2c08cf243bff22c10b9c00292813f4154a3a859b1e7c0f63d830006397"} Dec 13 07:12:54 crc kubenswrapper[4971]: I1213 07:12:54.145322 4971 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/cinder-db-sync-26k47" podStartSLOduration=4.683439528 podStartE2EDuration="51.145304258s" podCreationTimestamp="2025-12-13 07:12:03 +0000 UTC" firstStartedPulling="2025-12-13 07:12:06.242178685 +0000 UTC m=+1382.846588133" lastFinishedPulling="2025-12-13 07:12:52.704043415 +0000 UTC m=+1429.308452863" observedRunningTime="2025-12-13 07:12:54.138227894 +0000 UTC m=+1430.742637362" watchObservedRunningTime="2025-12-13 07:12:54.145304258 +0000 UTC m=+1430.749713706" Dec 13 07:12:54 crc kubenswrapper[4971]: I1213 07:12:54.145864 4971 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/keystone-7969997c86-5t8c8" Dec 13 07:12:54 crc kubenswrapper[4971]: I1213 07:12:54.167270 4971 prober.go:107] "Probe failed" probeType="Startup" pod="openstack/horizon-797dfcfc58-wlch7" podUID="9a4736be-5537-4638-a203-d36db1ae35cc" containerName="horizon" probeResult="failure" output="Get \"https://10.217.0.141:8443/dashboard/auth/login/?next=/dashboard/\": dial tcp 10.217.0.141:8443: connect: connection refused" Dec 13 07:12:54 crc kubenswrapper[4971]: I1213 07:12:54.177765 4971 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/placement-569b6fd56b-w7xmm" podStartSLOduration=8.177720016 podStartE2EDuration="8.177720016s" podCreationTimestamp="2025-12-13 07:12:46 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-13 07:12:54.162562982 +0000 UTC m=+1430.766972430" watchObservedRunningTime="2025-12-13 07:12:54.177720016 +0000 UTC m=+1430.782129464" Dec 13 07:12:54 crc kubenswrapper[4971]: I1213 07:12:54.179236 4971 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/barbican-db-sync-dvchm" podStartSLOduration=3.920334221 podStartE2EDuration="50.179225543s" podCreationTimestamp="2025-12-13 07:12:04 +0000 UTC" firstStartedPulling="2025-12-13 07:12:06.466036717 +0000 UTC m=+1383.070446165" lastFinishedPulling="2025-12-13 07:12:52.724928039 +0000 UTC m=+1429.329337487" observedRunningTime="2025-12-13 07:12:54.177632634 +0000 UTC m=+1430.782042152" watchObservedRunningTime="2025-12-13 07:12:54.179225543 +0000 UTC m=+1430.783635001" Dec 13 07:12:54 crc kubenswrapper[4971]: I1213 07:12:54.248836 4971 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/keystone-7969997c86-5t8c8" podStartSLOduration=6.248815126 podStartE2EDuration="6.248815126s" podCreationTimestamp="2025-12-13 07:12:48 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-13 07:12:54.21687032 +0000 UTC m=+1430.821279778" watchObservedRunningTime="2025-12-13 07:12:54.248815126 +0000 UTC m=+1430.853224574" Dec 13 07:12:55 crc kubenswrapper[4971]: I1213 07:12:55.380835 4971 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/placement-569b6fd56b-w7xmm" Dec 13 07:12:55 crc kubenswrapper[4971]: I1213 07:12:55.680877 4971 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/dnsmasq-dns-5d87b7c6dc-sxklp" podUID="2430f20e-0768-47e9-89de-fc6ac2615707" containerName="dnsmasq-dns" probeResult="failure" output="dial tcp 10.217.0.137:5353: i/o timeout" Dec 13 07:12:57 crc kubenswrapper[4971]: I1213 07:12:57.289935 4971 generic.go:334] "Generic (PLEG): container finished" podID="d7cf1b55-0dae-4a17-857a-6336fe71ff24" containerID="0b05501843e82c1e08bcb4071dfc72aa660772984720ae7dcfa4a26088cbd9be" exitCode=0 Dec 13 07:12:57 crc kubenswrapper[4971]: I1213 07:12:57.290033 4971 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-db-sync-dvchm" event={"ID":"d7cf1b55-0dae-4a17-857a-6336fe71ff24","Type":"ContainerDied","Data":"0b05501843e82c1e08bcb4071dfc72aa660772984720ae7dcfa4a26088cbd9be"} Dec 13 07:12:59 crc kubenswrapper[4971]: I1213 07:12:59.320570 4971 generic.go:334] "Generic (PLEG): container finished" podID="c1d819d0-339f-4e7e-9180-63d2f7c0042e" containerID="3397bd79f11b055f37e31637c81e296356926fc8aaa643a9a2f444c08a08c661" exitCode=0 Dec 13 07:12:59 crc kubenswrapper[4971]: I1213 07:12:59.320674 4971 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-db-sync-26k47" event={"ID":"c1d819d0-339f-4e7e-9180-63d2f7c0042e","Type":"ContainerDied","Data":"3397bd79f11b055f37e31637c81e296356926fc8aaa643a9a2f444c08a08c661"} Dec 13 07:13:03 crc kubenswrapper[4971]: I1213 07:13:03.714269 4971 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-db-sync-26k47" Dec 13 07:13:03 crc kubenswrapper[4971]: I1213 07:13:03.719207 4971 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-db-sync-dvchm" Dec 13 07:13:03 crc kubenswrapper[4971]: I1213 07:13:03.877499 4971 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/d7cf1b55-0dae-4a17-857a-6336fe71ff24-db-sync-config-data\") pod \"d7cf1b55-0dae-4a17-857a-6336fe71ff24\" (UID: \"d7cf1b55-0dae-4a17-857a-6336fe71ff24\") " Dec 13 07:13:03 crc kubenswrapper[4971]: I1213 07:13:03.877604 4971 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/c1d819d0-339f-4e7e-9180-63d2f7c0042e-etc-machine-id\") pod \"c1d819d0-339f-4e7e-9180-63d2f7c0042e\" (UID: \"c1d819d0-339f-4e7e-9180-63d2f7c0042e\") " Dec 13 07:13:03 crc kubenswrapper[4971]: I1213 07:13:03.877633 4971 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/c1d819d0-339f-4e7e-9180-63d2f7c0042e-config-data\") pod \"c1d819d0-339f-4e7e-9180-63d2f7c0042e\" (UID: \"c1d819d0-339f-4e7e-9180-63d2f7c0042e\") " Dec 13 07:13:03 crc kubenswrapper[4971]: I1213 07:13:03.877676 4971 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-r4lhp\" (UniqueName: \"kubernetes.io/projected/c1d819d0-339f-4e7e-9180-63d2f7c0042e-kube-api-access-r4lhp\") pod \"c1d819d0-339f-4e7e-9180-63d2f7c0042e\" (UID: \"c1d819d0-339f-4e7e-9180-63d2f7c0042e\") " Dec 13 07:13:03 crc kubenswrapper[4971]: I1213 07:13:03.877714 4971 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/c1d819d0-339f-4e7e-9180-63d2f7c0042e-scripts\") pod \"c1d819d0-339f-4e7e-9180-63d2f7c0042e\" (UID: \"c1d819d0-339f-4e7e-9180-63d2f7c0042e\") " Dec 13 07:13:03 crc kubenswrapper[4971]: I1213 07:13:03.877752 4971 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d7cf1b55-0dae-4a17-857a-6336fe71ff24-combined-ca-bundle\") pod \"d7cf1b55-0dae-4a17-857a-6336fe71ff24\" (UID: \"d7cf1b55-0dae-4a17-857a-6336fe71ff24\") " Dec 13 07:13:03 crc kubenswrapper[4971]: I1213 07:13:03.877824 4971 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/c1d819d0-339f-4e7e-9180-63d2f7c0042e-db-sync-config-data\") pod \"c1d819d0-339f-4e7e-9180-63d2f7c0042e\" (UID: \"c1d819d0-339f-4e7e-9180-63d2f7c0042e\") " Dec 13 07:13:03 crc kubenswrapper[4971]: I1213 07:13:03.877851 4971 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-w84mh\" (UniqueName: \"kubernetes.io/projected/d7cf1b55-0dae-4a17-857a-6336fe71ff24-kube-api-access-w84mh\") pod \"d7cf1b55-0dae-4a17-857a-6336fe71ff24\" (UID: \"d7cf1b55-0dae-4a17-857a-6336fe71ff24\") " Dec 13 07:13:03 crc kubenswrapper[4971]: I1213 07:13:03.878013 4971 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/c1d819d0-339f-4e7e-9180-63d2f7c0042e-combined-ca-bundle\") pod \"c1d819d0-339f-4e7e-9180-63d2f7c0042e\" (UID: \"c1d819d0-339f-4e7e-9180-63d2f7c0042e\") " Dec 13 07:13:03 crc kubenswrapper[4971]: I1213 07:13:03.880504 4971 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/c1d819d0-339f-4e7e-9180-63d2f7c0042e-etc-machine-id" (OuterVolumeSpecName: "etc-machine-id") pod "c1d819d0-339f-4e7e-9180-63d2f7c0042e" (UID: "c1d819d0-339f-4e7e-9180-63d2f7c0042e"). InnerVolumeSpecName "etc-machine-id". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 13 07:13:03 crc kubenswrapper[4971]: I1213 07:13:03.884346 4971 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/d7cf1b55-0dae-4a17-857a-6336fe71ff24-db-sync-config-data" (OuterVolumeSpecName: "db-sync-config-data") pod "d7cf1b55-0dae-4a17-857a-6336fe71ff24" (UID: "d7cf1b55-0dae-4a17-857a-6336fe71ff24"). InnerVolumeSpecName "db-sync-config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 13 07:13:03 crc kubenswrapper[4971]: I1213 07:13:03.887112 4971 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/c1d819d0-339f-4e7e-9180-63d2f7c0042e-kube-api-access-r4lhp" (OuterVolumeSpecName: "kube-api-access-r4lhp") pod "c1d819d0-339f-4e7e-9180-63d2f7c0042e" (UID: "c1d819d0-339f-4e7e-9180-63d2f7c0042e"). InnerVolumeSpecName "kube-api-access-r4lhp". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 13 07:13:03 crc kubenswrapper[4971]: I1213 07:13:03.888425 4971 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/c1d819d0-339f-4e7e-9180-63d2f7c0042e-scripts" (OuterVolumeSpecName: "scripts") pod "c1d819d0-339f-4e7e-9180-63d2f7c0042e" (UID: "c1d819d0-339f-4e7e-9180-63d2f7c0042e"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 13 07:13:03 crc kubenswrapper[4971]: I1213 07:13:03.891126 4971 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/d7cf1b55-0dae-4a17-857a-6336fe71ff24-kube-api-access-w84mh" (OuterVolumeSpecName: "kube-api-access-w84mh") pod "d7cf1b55-0dae-4a17-857a-6336fe71ff24" (UID: "d7cf1b55-0dae-4a17-857a-6336fe71ff24"). InnerVolumeSpecName "kube-api-access-w84mh". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 13 07:13:03 crc kubenswrapper[4971]: I1213 07:13:03.895724 4971 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/c1d819d0-339f-4e7e-9180-63d2f7c0042e-db-sync-config-data" (OuterVolumeSpecName: "db-sync-config-data") pod "c1d819d0-339f-4e7e-9180-63d2f7c0042e" (UID: "c1d819d0-339f-4e7e-9180-63d2f7c0042e"). InnerVolumeSpecName "db-sync-config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 13 07:13:03 crc kubenswrapper[4971]: I1213 07:13:03.906155 4971 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/c1d819d0-339f-4e7e-9180-63d2f7c0042e-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "c1d819d0-339f-4e7e-9180-63d2f7c0042e" (UID: "c1d819d0-339f-4e7e-9180-63d2f7c0042e"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 13 07:13:03 crc kubenswrapper[4971]: I1213 07:13:03.906587 4971 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/d7cf1b55-0dae-4a17-857a-6336fe71ff24-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "d7cf1b55-0dae-4a17-857a-6336fe71ff24" (UID: "d7cf1b55-0dae-4a17-857a-6336fe71ff24"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 13 07:13:03 crc kubenswrapper[4971]: I1213 07:13:03.929125 4971 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/c1d819d0-339f-4e7e-9180-63d2f7c0042e-config-data" (OuterVolumeSpecName: "config-data") pod "c1d819d0-339f-4e7e-9180-63d2f7c0042e" (UID: "c1d819d0-339f-4e7e-9180-63d2f7c0042e"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 13 07:13:04 crc kubenswrapper[4971]: I1213 07:13:04.012765 4971 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/c1d819d0-339f-4e7e-9180-63d2f7c0042e-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 13 07:13:04 crc kubenswrapper[4971]: I1213 07:13:04.012799 4971 reconciler_common.go:293] "Volume detached for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/d7cf1b55-0dae-4a17-857a-6336fe71ff24-db-sync-config-data\") on node \"crc\" DevicePath \"\"" Dec 13 07:13:04 crc kubenswrapper[4971]: I1213 07:13:04.012808 4971 reconciler_common.go:293] "Volume detached for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/c1d819d0-339f-4e7e-9180-63d2f7c0042e-etc-machine-id\") on node \"crc\" DevicePath \"\"" Dec 13 07:13:04 crc kubenswrapper[4971]: I1213 07:13:04.012818 4971 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/c1d819d0-339f-4e7e-9180-63d2f7c0042e-config-data\") on node \"crc\" DevicePath \"\"" Dec 13 07:13:04 crc kubenswrapper[4971]: I1213 07:13:04.012828 4971 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-r4lhp\" (UniqueName: \"kubernetes.io/projected/c1d819d0-339f-4e7e-9180-63d2f7c0042e-kube-api-access-r4lhp\") on node \"crc\" DevicePath \"\"" Dec 13 07:13:04 crc kubenswrapper[4971]: I1213 07:13:04.012840 4971 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/c1d819d0-339f-4e7e-9180-63d2f7c0042e-scripts\") on node \"crc\" DevicePath \"\"" Dec 13 07:13:04 crc kubenswrapper[4971]: I1213 07:13:04.012848 4971 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d7cf1b55-0dae-4a17-857a-6336fe71ff24-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 13 07:13:04 crc kubenswrapper[4971]: I1213 07:13:04.012856 4971 reconciler_common.go:293] "Volume detached for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/c1d819d0-339f-4e7e-9180-63d2f7c0042e-db-sync-config-data\") on node \"crc\" DevicePath \"\"" Dec 13 07:13:04 crc kubenswrapper[4971]: I1213 07:13:04.012865 4971 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-w84mh\" (UniqueName: \"kubernetes.io/projected/d7cf1b55-0dae-4a17-857a-6336fe71ff24-kube-api-access-w84mh\") on node \"crc\" DevicePath \"\"" Dec 13 07:13:04 crc kubenswrapper[4971]: I1213 07:13:04.367332 4971 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-db-sync-26k47" Dec 13 07:13:04 crc kubenswrapper[4971]: I1213 07:13:04.367351 4971 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-db-sync-26k47" event={"ID":"c1d819d0-339f-4e7e-9180-63d2f7c0042e","Type":"ContainerDied","Data":"4a0005fe1401e2d15dee97c0cd5b2ea49f753e03b159110ac32c15aa76fd93e2"} Dec 13 07:13:04 crc kubenswrapper[4971]: I1213 07:13:04.367765 4971 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="4a0005fe1401e2d15dee97c0cd5b2ea49f753e03b159110ac32c15aa76fd93e2" Dec 13 07:13:04 crc kubenswrapper[4971]: I1213 07:13:04.369761 4971 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-db-sync-dvchm" event={"ID":"d7cf1b55-0dae-4a17-857a-6336fe71ff24","Type":"ContainerDied","Data":"4d91578739ace0c00325d2663283e687dd7acfed3e767ee9d529e92f7ceaf861"} Dec 13 07:13:04 crc kubenswrapper[4971]: I1213 07:13:04.369789 4971 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="4d91578739ace0c00325d2663283e687dd7acfed3e767ee9d529e92f7ceaf861" Dec 13 07:13:04 crc kubenswrapper[4971]: I1213 07:13:04.369850 4971 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-db-sync-dvchm" Dec 13 07:13:05 crc kubenswrapper[4971]: I1213 07:13:05.258069 4971 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/barbican-worker-c7c6b44fc-8hd9n"] Dec 13 07:13:05 crc kubenswrapper[4971]: E1213 07:13:05.258855 4971 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="c1d819d0-339f-4e7e-9180-63d2f7c0042e" containerName="cinder-db-sync" Dec 13 07:13:05 crc kubenswrapper[4971]: I1213 07:13:05.258877 4971 state_mem.go:107] "Deleted CPUSet assignment" podUID="c1d819d0-339f-4e7e-9180-63d2f7c0042e" containerName="cinder-db-sync" Dec 13 07:13:05 crc kubenswrapper[4971]: E1213 07:13:05.258903 4971 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="2430f20e-0768-47e9-89de-fc6ac2615707" containerName="dnsmasq-dns" Dec 13 07:13:05 crc kubenswrapper[4971]: I1213 07:13:05.258911 4971 state_mem.go:107] "Deleted CPUSet assignment" podUID="2430f20e-0768-47e9-89de-fc6ac2615707" containerName="dnsmasq-dns" Dec 13 07:13:05 crc kubenswrapper[4971]: E1213 07:13:05.258930 4971 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="2430f20e-0768-47e9-89de-fc6ac2615707" containerName="init" Dec 13 07:13:05 crc kubenswrapper[4971]: I1213 07:13:05.258938 4971 state_mem.go:107] "Deleted CPUSet assignment" podUID="2430f20e-0768-47e9-89de-fc6ac2615707" containerName="init" Dec 13 07:13:05 crc kubenswrapper[4971]: E1213 07:13:05.258959 4971 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="d7cf1b55-0dae-4a17-857a-6336fe71ff24" containerName="barbican-db-sync" Dec 13 07:13:05 crc kubenswrapper[4971]: I1213 07:13:05.258967 4971 state_mem.go:107] "Deleted CPUSet assignment" podUID="d7cf1b55-0dae-4a17-857a-6336fe71ff24" containerName="barbican-db-sync" Dec 13 07:13:05 crc kubenswrapper[4971]: I1213 07:13:05.259179 4971 memory_manager.go:354] "RemoveStaleState removing state" podUID="c1d819d0-339f-4e7e-9180-63d2f7c0042e" containerName="cinder-db-sync" Dec 13 07:13:05 crc kubenswrapper[4971]: I1213 07:13:05.259206 4971 memory_manager.go:354] "RemoveStaleState removing state" podUID="d7cf1b55-0dae-4a17-857a-6336fe71ff24" containerName="barbican-db-sync" Dec 13 07:13:05 crc kubenswrapper[4971]: I1213 07:13:05.259219 4971 memory_manager.go:354] "RemoveStaleState removing state" podUID="2430f20e-0768-47e9-89de-fc6ac2615707" containerName="dnsmasq-dns" Dec 13 07:13:05 crc kubenswrapper[4971]: I1213 07:13:05.260340 4971 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-worker-c7c6b44fc-8hd9n" Dec 13 07:13:05 crc kubenswrapper[4971]: I1213 07:13:05.266494 4971 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"barbican-barbican-dockercfg-6zl24" Dec 13 07:13:05 crc kubenswrapper[4971]: I1213 07:13:05.266630 4971 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"barbican-worker-config-data" Dec 13 07:13:05 crc kubenswrapper[4971]: I1213 07:13:05.266772 4971 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"barbican-config-data" Dec 13 07:13:05 crc kubenswrapper[4971]: I1213 07:13:05.280119 4971 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/barbican-keystone-listener-6959d6f94b-tq6hh"] Dec 13 07:13:05 crc kubenswrapper[4971]: I1213 07:13:05.282075 4971 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-keystone-listener-6959d6f94b-tq6hh" Dec 13 07:13:05 crc kubenswrapper[4971]: I1213 07:13:05.294879 4971 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"barbican-keystone-listener-config-data" Dec 13 07:13:05 crc kubenswrapper[4971]: I1213 07:13:05.300488 4971 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/barbican-worker-c7c6b44fc-8hd9n"] Dec 13 07:13:05 crc kubenswrapper[4971]: I1213 07:13:05.312111 4971 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/12a4a604-178e-4ad8-a0f7-e2e3d47590bd-config-data-custom\") pod \"barbican-keystone-listener-6959d6f94b-tq6hh\" (UID: \"12a4a604-178e-4ad8-a0f7-e2e3d47590bd\") " pod="openstack/barbican-keystone-listener-6959d6f94b-tq6hh" Dec 13 07:13:05 crc kubenswrapper[4971]: I1213 07:13:05.312161 4971 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/fb068297-28dd-4a70-acbf-650b9edf380e-config-data-custom\") pod \"barbican-worker-c7c6b44fc-8hd9n\" (UID: \"fb068297-28dd-4a70-acbf-650b9edf380e\") " pod="openstack/barbican-worker-c7c6b44fc-8hd9n" Dec 13 07:13:05 crc kubenswrapper[4971]: I1213 07:13:05.312187 4971 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/fb068297-28dd-4a70-acbf-650b9edf380e-combined-ca-bundle\") pod \"barbican-worker-c7c6b44fc-8hd9n\" (UID: \"fb068297-28dd-4a70-acbf-650b9edf380e\") " pod="openstack/barbican-worker-c7c6b44fc-8hd9n" Dec 13 07:13:05 crc kubenswrapper[4971]: I1213 07:13:05.312207 4971 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/fb068297-28dd-4a70-acbf-650b9edf380e-config-data\") pod \"barbican-worker-c7c6b44fc-8hd9n\" (UID: \"fb068297-28dd-4a70-acbf-650b9edf380e\") " pod="openstack/barbican-worker-c7c6b44fc-8hd9n" Dec 13 07:13:05 crc kubenswrapper[4971]: I1213 07:13:05.312242 4971 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/12a4a604-178e-4ad8-a0f7-e2e3d47590bd-combined-ca-bundle\") pod \"barbican-keystone-listener-6959d6f94b-tq6hh\" (UID: \"12a4a604-178e-4ad8-a0f7-e2e3d47590bd\") " pod="openstack/barbican-keystone-listener-6959d6f94b-tq6hh" Dec 13 07:13:05 crc kubenswrapper[4971]: I1213 07:13:05.312269 4971 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-dmwr7\" (UniqueName: \"kubernetes.io/projected/12a4a604-178e-4ad8-a0f7-e2e3d47590bd-kube-api-access-dmwr7\") pod \"barbican-keystone-listener-6959d6f94b-tq6hh\" (UID: \"12a4a604-178e-4ad8-a0f7-e2e3d47590bd\") " pod="openstack/barbican-keystone-listener-6959d6f94b-tq6hh" Dec 13 07:13:05 crc kubenswrapper[4971]: I1213 07:13:05.312296 4971 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-lwtmk\" (UniqueName: \"kubernetes.io/projected/fb068297-28dd-4a70-acbf-650b9edf380e-kube-api-access-lwtmk\") pod \"barbican-worker-c7c6b44fc-8hd9n\" (UID: \"fb068297-28dd-4a70-acbf-650b9edf380e\") " pod="openstack/barbican-worker-c7c6b44fc-8hd9n" Dec 13 07:13:05 crc kubenswrapper[4971]: I1213 07:13:05.312322 4971 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/12a4a604-178e-4ad8-a0f7-e2e3d47590bd-logs\") pod \"barbican-keystone-listener-6959d6f94b-tq6hh\" (UID: \"12a4a604-178e-4ad8-a0f7-e2e3d47590bd\") " pod="openstack/barbican-keystone-listener-6959d6f94b-tq6hh" Dec 13 07:13:05 crc kubenswrapper[4971]: I1213 07:13:05.312336 4971 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/fb068297-28dd-4a70-acbf-650b9edf380e-logs\") pod \"barbican-worker-c7c6b44fc-8hd9n\" (UID: \"fb068297-28dd-4a70-acbf-650b9edf380e\") " pod="openstack/barbican-worker-c7c6b44fc-8hd9n" Dec 13 07:13:05 crc kubenswrapper[4971]: I1213 07:13:05.312369 4971 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/12a4a604-178e-4ad8-a0f7-e2e3d47590bd-config-data\") pod \"barbican-keystone-listener-6959d6f94b-tq6hh\" (UID: \"12a4a604-178e-4ad8-a0f7-e2e3d47590bd\") " pod="openstack/barbican-keystone-listener-6959d6f94b-tq6hh" Dec 13 07:13:05 crc kubenswrapper[4971]: I1213 07:13:05.324158 4971 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/barbican-keystone-listener-6959d6f94b-tq6hh"] Dec 13 07:13:05 crc kubenswrapper[4971]: I1213 07:13:05.416299 4971 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/12a4a604-178e-4ad8-a0f7-e2e3d47590bd-config-data-custom\") pod \"barbican-keystone-listener-6959d6f94b-tq6hh\" (UID: \"12a4a604-178e-4ad8-a0f7-e2e3d47590bd\") " pod="openstack/barbican-keystone-listener-6959d6f94b-tq6hh" Dec 13 07:13:05 crc kubenswrapper[4971]: I1213 07:13:05.416370 4971 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/fb068297-28dd-4a70-acbf-650b9edf380e-config-data-custom\") pod \"barbican-worker-c7c6b44fc-8hd9n\" (UID: \"fb068297-28dd-4a70-acbf-650b9edf380e\") " pod="openstack/barbican-worker-c7c6b44fc-8hd9n" Dec 13 07:13:05 crc kubenswrapper[4971]: I1213 07:13:05.416405 4971 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/fb068297-28dd-4a70-acbf-650b9edf380e-combined-ca-bundle\") pod \"barbican-worker-c7c6b44fc-8hd9n\" (UID: \"fb068297-28dd-4a70-acbf-650b9edf380e\") " pod="openstack/barbican-worker-c7c6b44fc-8hd9n" Dec 13 07:13:05 crc kubenswrapper[4971]: I1213 07:13:05.416444 4971 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/fb068297-28dd-4a70-acbf-650b9edf380e-config-data\") pod \"barbican-worker-c7c6b44fc-8hd9n\" (UID: \"fb068297-28dd-4a70-acbf-650b9edf380e\") " pod="openstack/barbican-worker-c7c6b44fc-8hd9n" Dec 13 07:13:05 crc kubenswrapper[4971]: I1213 07:13:05.416503 4971 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/12a4a604-178e-4ad8-a0f7-e2e3d47590bd-combined-ca-bundle\") pod \"barbican-keystone-listener-6959d6f94b-tq6hh\" (UID: \"12a4a604-178e-4ad8-a0f7-e2e3d47590bd\") " pod="openstack/barbican-keystone-listener-6959d6f94b-tq6hh" Dec 13 07:13:05 crc kubenswrapper[4971]: I1213 07:13:05.625835 4971 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-dmwr7\" (UniqueName: \"kubernetes.io/projected/12a4a604-178e-4ad8-a0f7-e2e3d47590bd-kube-api-access-dmwr7\") pod \"barbican-keystone-listener-6959d6f94b-tq6hh\" (UID: \"12a4a604-178e-4ad8-a0f7-e2e3d47590bd\") " pod="openstack/barbican-keystone-listener-6959d6f94b-tq6hh" Dec 13 07:13:05 crc kubenswrapper[4971]: I1213 07:13:05.625942 4971 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-lwtmk\" (UniqueName: \"kubernetes.io/projected/fb068297-28dd-4a70-acbf-650b9edf380e-kube-api-access-lwtmk\") pod \"barbican-worker-c7c6b44fc-8hd9n\" (UID: \"fb068297-28dd-4a70-acbf-650b9edf380e\") " pod="openstack/barbican-worker-c7c6b44fc-8hd9n" Dec 13 07:13:05 crc kubenswrapper[4971]: I1213 07:13:05.626006 4971 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/12a4a604-178e-4ad8-a0f7-e2e3d47590bd-logs\") pod \"barbican-keystone-listener-6959d6f94b-tq6hh\" (UID: \"12a4a604-178e-4ad8-a0f7-e2e3d47590bd\") " pod="openstack/barbican-keystone-listener-6959d6f94b-tq6hh" Dec 13 07:13:05 crc kubenswrapper[4971]: I1213 07:13:05.626031 4971 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/fb068297-28dd-4a70-acbf-650b9edf380e-logs\") pod \"barbican-worker-c7c6b44fc-8hd9n\" (UID: \"fb068297-28dd-4a70-acbf-650b9edf380e\") " pod="openstack/barbican-worker-c7c6b44fc-8hd9n" Dec 13 07:13:05 crc kubenswrapper[4971]: I1213 07:13:05.626084 4971 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/12a4a604-178e-4ad8-a0f7-e2e3d47590bd-config-data\") pod \"barbican-keystone-listener-6959d6f94b-tq6hh\" (UID: \"12a4a604-178e-4ad8-a0f7-e2e3d47590bd\") " pod="openstack/barbican-keystone-listener-6959d6f94b-tq6hh" Dec 13 07:13:05 crc kubenswrapper[4971]: I1213 07:13:05.516532 4971 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/fb068297-28dd-4a70-acbf-650b9edf380e-combined-ca-bundle\") pod \"barbican-worker-c7c6b44fc-8hd9n\" (UID: \"fb068297-28dd-4a70-acbf-650b9edf380e\") " pod="openstack/barbican-worker-c7c6b44fc-8hd9n" Dec 13 07:13:05 crc kubenswrapper[4971]: I1213 07:13:05.508825 4971 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/fb068297-28dd-4a70-acbf-650b9edf380e-config-data\") pod \"barbican-worker-c7c6b44fc-8hd9n\" (UID: \"fb068297-28dd-4a70-acbf-650b9edf380e\") " pod="openstack/barbican-worker-c7c6b44fc-8hd9n" Dec 13 07:13:05 crc kubenswrapper[4971]: I1213 07:13:05.525073 4971 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/12a4a604-178e-4ad8-a0f7-e2e3d47590bd-combined-ca-bundle\") pod \"barbican-keystone-listener-6959d6f94b-tq6hh\" (UID: \"12a4a604-178e-4ad8-a0f7-e2e3d47590bd\") " pod="openstack/barbican-keystone-listener-6959d6f94b-tq6hh" Dec 13 07:13:05 crc kubenswrapper[4971]: I1213 07:13:05.515841 4971 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/12a4a604-178e-4ad8-a0f7-e2e3d47590bd-config-data-custom\") pod \"barbican-keystone-listener-6959d6f94b-tq6hh\" (UID: \"12a4a604-178e-4ad8-a0f7-e2e3d47590bd\") " pod="openstack/barbican-keystone-listener-6959d6f94b-tq6hh" Dec 13 07:13:05 crc kubenswrapper[4971]: I1213 07:13:05.617050 4971 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/fb068297-28dd-4a70-acbf-650b9edf380e-config-data-custom\") pod \"barbican-worker-c7c6b44fc-8hd9n\" (UID: \"fb068297-28dd-4a70-acbf-650b9edf380e\") " pod="openstack/barbican-worker-c7c6b44fc-8hd9n" Dec 13 07:13:05 crc kubenswrapper[4971]: I1213 07:13:05.630130 4971 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/12a4a604-178e-4ad8-a0f7-e2e3d47590bd-logs\") pod \"barbican-keystone-listener-6959d6f94b-tq6hh\" (UID: \"12a4a604-178e-4ad8-a0f7-e2e3d47590bd\") " pod="openstack/barbican-keystone-listener-6959d6f94b-tq6hh" Dec 13 07:13:05 crc kubenswrapper[4971]: I1213 07:13:05.630321 4971 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/fb068297-28dd-4a70-acbf-650b9edf380e-logs\") pod \"barbican-worker-c7c6b44fc-8hd9n\" (UID: \"fb068297-28dd-4a70-acbf-650b9edf380e\") " pod="openstack/barbican-worker-c7c6b44fc-8hd9n" Dec 13 07:13:05 crc kubenswrapper[4971]: I1213 07:13:05.640089 4971 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/12a4a604-178e-4ad8-a0f7-e2e3d47590bd-config-data\") pod \"barbican-keystone-listener-6959d6f94b-tq6hh\" (UID: \"12a4a604-178e-4ad8-a0f7-e2e3d47590bd\") " pod="openstack/barbican-keystone-listener-6959d6f94b-tq6hh" Dec 13 07:13:05 crc kubenswrapper[4971]: I1213 07:13:05.646591 4971 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"56e56ce1-4dd1-4ae5-b868-e98d27d5ad94","Type":"ContainerStarted","Data":"1e636e556657fdb168cfe90bbd03b3e59853b99c24a8b8151d7adedc027197ad"} Dec 13 07:13:05 crc kubenswrapper[4971]: I1213 07:13:05.646909 4971 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="56e56ce1-4dd1-4ae5-b868-e98d27d5ad94" containerName="ceilometer-central-agent" containerID="cri-o://5d8a06b64cc9778b600eb7ee56a36445340d0fb5dafba3ee09865a0b6e2343ea" gracePeriod=30 Dec 13 07:13:05 crc kubenswrapper[4971]: I1213 07:13:05.647244 4971 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/ceilometer-0" Dec 13 07:13:05 crc kubenswrapper[4971]: I1213 07:13:05.648364 4971 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="56e56ce1-4dd1-4ae5-b868-e98d27d5ad94" containerName="proxy-httpd" containerID="cri-o://1e636e556657fdb168cfe90bbd03b3e59853b99c24a8b8151d7adedc027197ad" gracePeriod=30 Dec 13 07:13:05 crc kubenswrapper[4971]: I1213 07:13:05.648451 4971 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="56e56ce1-4dd1-4ae5-b868-e98d27d5ad94" containerName="sg-core" containerID="cri-o://a0e5a1ecd136567279050709089de36451185ac24d99f2e15b755c1d0658ad99" gracePeriod=30 Dec 13 07:13:05 crc kubenswrapper[4971]: I1213 07:13:05.648558 4971 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="56e56ce1-4dd1-4ae5-b868-e98d27d5ad94" containerName="ceilometer-notification-agent" containerID="cri-o://d208a7c037f088ceda5804583d1d1761594f3fbc2ad4ae265ebc6e2bf875c418" gracePeriod=30 Dec 13 07:13:05 crc kubenswrapper[4971]: I1213 07:13:05.680751 4971 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-lwtmk\" (UniqueName: \"kubernetes.io/projected/fb068297-28dd-4a70-acbf-650b9edf380e-kube-api-access-lwtmk\") pod \"barbican-worker-c7c6b44fc-8hd9n\" (UID: \"fb068297-28dd-4a70-acbf-650b9edf380e\") " pod="openstack/barbican-worker-c7c6b44fc-8hd9n" Dec 13 07:13:05 crc kubenswrapper[4971]: I1213 07:13:05.680817 4971 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-dmwr7\" (UniqueName: \"kubernetes.io/projected/12a4a604-178e-4ad8-a0f7-e2e3d47590bd-kube-api-access-dmwr7\") pod \"barbican-keystone-listener-6959d6f94b-tq6hh\" (UID: \"12a4a604-178e-4ad8-a0f7-e2e3d47590bd\") " pod="openstack/barbican-keystone-listener-6959d6f94b-tq6hh" Dec 13 07:13:05 crc kubenswrapper[4971]: I1213 07:13:05.686858 4971 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-744f57c8cc-w8rwt"] Dec 13 07:13:05 crc kubenswrapper[4971]: I1213 07:13:05.689305 4971 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-744f57c8cc-w8rwt" Dec 13 07:13:05 crc kubenswrapper[4971]: I1213 07:13:05.745658 4971 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-744f57c8cc-w8rwt"] Dec 13 07:13:05 crc kubenswrapper[4971]: I1213 07:13:05.954638 4971 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-worker-c7c6b44fc-8hd9n" Dec 13 07:13:05 crc kubenswrapper[4971]: I1213 07:13:05.965348 4971 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/ceilometer-0" podStartSLOduration=3.870645197 podStartE2EDuration="1m1.965325784s" podCreationTimestamp="2025-12-13 07:12:04 +0000 UTC" firstStartedPulling="2025-12-13 07:12:06.466102868 +0000 UTC m=+1383.070512306" lastFinishedPulling="2025-12-13 07:13:04.560783445 +0000 UTC m=+1441.165192893" observedRunningTime="2025-12-13 07:13:05.728738549 +0000 UTC m=+1442.333148007" watchObservedRunningTime="2025-12-13 07:13:05.965325784 +0000 UTC m=+1442.569735232" Dec 13 07:13:06 crc kubenswrapper[4971]: I1213 07:13:06.055419 4971 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-keystone-listener-6959d6f94b-tq6hh" Dec 13 07:13:06 crc kubenswrapper[4971]: I1213 07:13:06.068230 4971 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/cinder-scheduler-0"] Dec 13 07:13:06 crc kubenswrapper[4971]: I1213 07:13:06.069924 4971 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cinder-scheduler-0"] Dec 13 07:13:06 crc kubenswrapper[4971]: I1213 07:13:06.069955 4971 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-744f57c8cc-w8rwt"] Dec 13 07:13:06 crc kubenswrapper[4971]: I1213 07:13:06.069978 4971 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/barbican-api-5c88bd59bd-2qlbj"] Dec 13 07:13:06 crc kubenswrapper[4971]: I1213 07:13:06.074988 4971 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/barbican-api-5c88bd59bd-2qlbj"] Dec 13 07:13:06 crc kubenswrapper[4971]: I1213 07:13:06.075035 4971 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-7474d577dc-q7d6q"] Dec 13 07:13:06 crc kubenswrapper[4971]: I1213 07:13:06.077944 4971 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-scheduler-0" Dec 13 07:13:06 crc kubenswrapper[4971]: I1213 07:13:06.079270 4971 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-7474d577dc-q7d6q" Dec 13 07:13:06 crc kubenswrapper[4971]: I1213 07:13:06.080009 4971 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-7474d577dc-q7d6q"] Dec 13 07:13:06 crc kubenswrapper[4971]: I1213 07:13:06.081260 4971 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-api-5c88bd59bd-2qlbj" Dec 13 07:13:06 crc kubenswrapper[4971]: I1213 07:13:06.086258 4971 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"barbican-api-config-data" Dec 13 07:13:06 crc kubenswrapper[4971]: I1213 07:13:06.086391 4971 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cinder-cinder-dockercfg-dxkxj" Dec 13 07:13:06 crc kubenswrapper[4971]: I1213 07:13:06.086448 4971 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cinder-scheduler-config-data" Dec 13 07:13:06 crc kubenswrapper[4971]: I1213 07:13:06.086683 4971 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cinder-config-data" Dec 13 07:13:06 crc kubenswrapper[4971]: I1213 07:13:06.089961 4971 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cinder-scripts" Dec 13 07:13:06 crc kubenswrapper[4971]: E1213 07:13:06.142805 4971 pod_workers.go:1301] "Error syncing pod, skipping" err="unmounted volumes=[config dns-svc kube-api-access-blhpn ovsdbserver-nb ovsdbserver-sb], unattached volumes=[], failed to process volumes=[]: context canceled" pod="openstack/dnsmasq-dns-744f57c8cc-w8rwt" podUID="a709a2b8-324a-468f-bca8-330676b24a85" Dec 13 07:13:06 crc kubenswrapper[4971]: I1213 07:13:06.143526 4971 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/a709a2b8-324a-468f-bca8-330676b24a85-ovsdbserver-sb\") pod \"dnsmasq-dns-744f57c8cc-w8rwt\" (UID: \"a709a2b8-324a-468f-bca8-330676b24a85\") " pod="openstack/dnsmasq-dns-744f57c8cc-w8rwt" Dec 13 07:13:06 crc kubenswrapper[4971]: I1213 07:13:06.143627 4971 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-blhpn\" (UniqueName: \"kubernetes.io/projected/a709a2b8-324a-468f-bca8-330676b24a85-kube-api-access-blhpn\") pod \"dnsmasq-dns-744f57c8cc-w8rwt\" (UID: \"a709a2b8-324a-468f-bca8-330676b24a85\") " pod="openstack/dnsmasq-dns-744f57c8cc-w8rwt" Dec 13 07:13:06 crc kubenswrapper[4971]: I1213 07:13:06.143726 4971 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/a709a2b8-324a-468f-bca8-330676b24a85-ovsdbserver-nb\") pod \"dnsmasq-dns-744f57c8cc-w8rwt\" (UID: \"a709a2b8-324a-468f-bca8-330676b24a85\") " pod="openstack/dnsmasq-dns-744f57c8cc-w8rwt" Dec 13 07:13:06 crc kubenswrapper[4971]: I1213 07:13:06.143806 4971 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/a709a2b8-324a-468f-bca8-330676b24a85-config\") pod \"dnsmasq-dns-744f57c8cc-w8rwt\" (UID: \"a709a2b8-324a-468f-bca8-330676b24a85\") " pod="openstack/dnsmasq-dns-744f57c8cc-w8rwt" Dec 13 07:13:06 crc kubenswrapper[4971]: I1213 07:13:06.143879 4971 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/a709a2b8-324a-468f-bca8-330676b24a85-dns-svc\") pod \"dnsmasq-dns-744f57c8cc-w8rwt\" (UID: \"a709a2b8-324a-468f-bca8-330676b24a85\") " pod="openstack/dnsmasq-dns-744f57c8cc-w8rwt" Dec 13 07:13:06 crc kubenswrapper[4971]: I1213 07:13:06.174260 4971 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/cinder-api-0"] Dec 13 07:13:06 crc kubenswrapper[4971]: I1213 07:13:06.175725 4971 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-api-0" Dec 13 07:13:06 crc kubenswrapper[4971]: I1213 07:13:06.180888 4971 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cinder-api-config-data" Dec 13 07:13:06 crc kubenswrapper[4971]: I1213 07:13:06.249599 4971 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cinder-api-0"] Dec 13 07:13:06 crc kubenswrapper[4971]: I1213 07:13:06.250731 4971 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/e2b4e2e7-cfa5-41d9-833e-b5afc95c91fe-combined-ca-bundle\") pod \"cinder-api-0\" (UID: \"e2b4e2e7-cfa5-41d9-833e-b5afc95c91fe\") " pod="openstack/cinder-api-0" Dec 13 07:13:06 crc kubenswrapper[4971]: I1213 07:13:06.250811 4971 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/a709a2b8-324a-468f-bca8-330676b24a85-dns-svc\") pod \"dnsmasq-dns-744f57c8cc-w8rwt\" (UID: \"a709a2b8-324a-468f-bca8-330676b24a85\") " pod="openstack/dnsmasq-dns-744f57c8cc-w8rwt" Dec 13 07:13:06 crc kubenswrapper[4971]: I1213 07:13:06.250839 4971 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/eb706473-627a-4570-9dfc-0e8fe99d59e1-scripts\") pod \"cinder-scheduler-0\" (UID: \"eb706473-627a-4570-9dfc-0e8fe99d59e1\") " pod="openstack/cinder-scheduler-0" Dec 13 07:13:06 crc kubenswrapper[4971]: I1213 07:13:06.250862 4971 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/a709a2b8-324a-468f-bca8-330676b24a85-ovsdbserver-sb\") pod \"dnsmasq-dns-744f57c8cc-w8rwt\" (UID: \"a709a2b8-324a-468f-bca8-330676b24a85\") " pod="openstack/dnsmasq-dns-744f57c8cc-w8rwt" Dec 13 07:13:06 crc kubenswrapper[4971]: I1213 07:13:06.250910 4971 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/e2b4e2e7-cfa5-41d9-833e-b5afc95c91fe-config-data\") pod \"cinder-api-0\" (UID: \"e2b4e2e7-cfa5-41d9-833e-b5afc95c91fe\") " pod="openstack/cinder-api-0" Dec 13 07:13:06 crc kubenswrapper[4971]: I1213 07:13:06.250928 4971 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-9bwlc\" (UniqueName: \"kubernetes.io/projected/e2b4e2e7-cfa5-41d9-833e-b5afc95c91fe-kube-api-access-9bwlc\") pod \"cinder-api-0\" (UID: \"e2b4e2e7-cfa5-41d9-833e-b5afc95c91fe\") " pod="openstack/cinder-api-0" Dec 13 07:13:06 crc kubenswrapper[4971]: I1213 07:13:06.250943 4971 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/02578e71-2865-4a4b-96e9-5bf6c625183c-logs\") pod \"barbican-api-5c88bd59bd-2qlbj\" (UID: \"02578e71-2865-4a4b-96e9-5bf6c625183c\") " pod="openstack/barbican-api-5c88bd59bd-2qlbj" Dec 13 07:13:06 crc kubenswrapper[4971]: I1213 07:13:06.250965 4971 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-blhpn\" (UniqueName: \"kubernetes.io/projected/a709a2b8-324a-468f-bca8-330676b24a85-kube-api-access-blhpn\") pod \"dnsmasq-dns-744f57c8cc-w8rwt\" (UID: \"a709a2b8-324a-468f-bca8-330676b24a85\") " pod="openstack/dnsmasq-dns-744f57c8cc-w8rwt" Dec 13 07:13:06 crc kubenswrapper[4971]: I1213 07:13:06.250991 4971 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/eb706473-627a-4570-9dfc-0e8fe99d59e1-config-data-custom\") pod \"cinder-scheduler-0\" (UID: \"eb706473-627a-4570-9dfc-0e8fe99d59e1\") " pod="openstack/cinder-scheduler-0" Dec 13 07:13:06 crc kubenswrapper[4971]: I1213 07:13:06.251026 4971 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/e2b4e2e7-cfa5-41d9-833e-b5afc95c91fe-etc-machine-id\") pod \"cinder-api-0\" (UID: \"e2b4e2e7-cfa5-41d9-833e-b5afc95c91fe\") " pod="openstack/cinder-api-0" Dec 13 07:13:06 crc kubenswrapper[4971]: I1213 07:13:06.251042 4971 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/eb706473-627a-4570-9dfc-0e8fe99d59e1-config-data\") pod \"cinder-scheduler-0\" (UID: \"eb706473-627a-4570-9dfc-0e8fe99d59e1\") " pod="openstack/cinder-scheduler-0" Dec 13 07:13:06 crc kubenswrapper[4971]: I1213 07:13:06.251057 4971 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/4c00defc-fd2b-43e6-ba62-ccc9fd2267fc-config\") pod \"dnsmasq-dns-7474d577dc-q7d6q\" (UID: \"4c00defc-fd2b-43e6-ba62-ccc9fd2267fc\") " pod="openstack/dnsmasq-dns-7474d577dc-q7d6q" Dec 13 07:13:06 crc kubenswrapper[4971]: I1213 07:13:06.251080 4971 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/4c00defc-fd2b-43e6-ba62-ccc9fd2267fc-ovsdbserver-nb\") pod \"dnsmasq-dns-7474d577dc-q7d6q\" (UID: \"4c00defc-fd2b-43e6-ba62-ccc9fd2267fc\") " pod="openstack/dnsmasq-dns-7474d577dc-q7d6q" Dec 13 07:13:06 crc kubenswrapper[4971]: I1213 07:13:06.251104 4971 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/02578e71-2865-4a4b-96e9-5bf6c625183c-combined-ca-bundle\") pod \"barbican-api-5c88bd59bd-2qlbj\" (UID: \"02578e71-2865-4a4b-96e9-5bf6c625183c\") " pod="openstack/barbican-api-5c88bd59bd-2qlbj" Dec 13 07:13:06 crc kubenswrapper[4971]: I1213 07:13:06.251127 4971 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/e2b4e2e7-cfa5-41d9-833e-b5afc95c91fe-scripts\") pod \"cinder-api-0\" (UID: \"e2b4e2e7-cfa5-41d9-833e-b5afc95c91fe\") " pod="openstack/cinder-api-0" Dec 13 07:13:06 crc kubenswrapper[4971]: I1213 07:13:06.251170 4971 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-kk4n8\" (UniqueName: \"kubernetes.io/projected/02578e71-2865-4a4b-96e9-5bf6c625183c-kube-api-access-kk4n8\") pod \"barbican-api-5c88bd59bd-2qlbj\" (UID: \"02578e71-2865-4a4b-96e9-5bf6c625183c\") " pod="openstack/barbican-api-5c88bd59bd-2qlbj" Dec 13 07:13:06 crc kubenswrapper[4971]: I1213 07:13:06.251188 4971 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/02578e71-2865-4a4b-96e9-5bf6c625183c-config-data-custom\") pod \"barbican-api-5c88bd59bd-2qlbj\" (UID: \"02578e71-2865-4a4b-96e9-5bf6c625183c\") " pod="openstack/barbican-api-5c88bd59bd-2qlbj" Dec 13 07:13:06 crc kubenswrapper[4971]: I1213 07:13:06.253581 4971 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/eb706473-627a-4570-9dfc-0e8fe99d59e1-combined-ca-bundle\") pod \"cinder-scheduler-0\" (UID: \"eb706473-627a-4570-9dfc-0e8fe99d59e1\") " pod="openstack/cinder-scheduler-0" Dec 13 07:13:06 crc kubenswrapper[4971]: I1213 07:13:06.253613 4971 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/a709a2b8-324a-468f-bca8-330676b24a85-ovsdbserver-nb\") pod \"dnsmasq-dns-744f57c8cc-w8rwt\" (UID: \"a709a2b8-324a-468f-bca8-330676b24a85\") " pod="openstack/dnsmasq-dns-744f57c8cc-w8rwt" Dec 13 07:13:06 crc kubenswrapper[4971]: I1213 07:13:06.253645 4971 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-s22xv\" (UniqueName: \"kubernetes.io/projected/eb706473-627a-4570-9dfc-0e8fe99d59e1-kube-api-access-s22xv\") pod \"cinder-scheduler-0\" (UID: \"eb706473-627a-4570-9dfc-0e8fe99d59e1\") " pod="openstack/cinder-scheduler-0" Dec 13 07:13:06 crc kubenswrapper[4971]: I1213 07:13:06.254710 4971 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/a709a2b8-324a-468f-bca8-330676b24a85-dns-svc\") pod \"dnsmasq-dns-744f57c8cc-w8rwt\" (UID: \"a709a2b8-324a-468f-bca8-330676b24a85\") " pod="openstack/dnsmasq-dns-744f57c8cc-w8rwt" Dec 13 07:13:06 crc kubenswrapper[4971]: I1213 07:13:06.255338 4971 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/a709a2b8-324a-468f-bca8-330676b24a85-ovsdbserver-sb\") pod \"dnsmasq-dns-744f57c8cc-w8rwt\" (UID: \"a709a2b8-324a-468f-bca8-330676b24a85\") " pod="openstack/dnsmasq-dns-744f57c8cc-w8rwt" Dec 13 07:13:06 crc kubenswrapper[4971]: I1213 07:13:06.257291 4971 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/a709a2b8-324a-468f-bca8-330676b24a85-ovsdbserver-nb\") pod \"dnsmasq-dns-744f57c8cc-w8rwt\" (UID: \"a709a2b8-324a-468f-bca8-330676b24a85\") " pod="openstack/dnsmasq-dns-744f57c8cc-w8rwt" Dec 13 07:13:06 crc kubenswrapper[4971]: I1213 07:13:06.266703 4971 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-h96b2\" (UniqueName: \"kubernetes.io/projected/4c00defc-fd2b-43e6-ba62-ccc9fd2267fc-kube-api-access-h96b2\") pod \"dnsmasq-dns-7474d577dc-q7d6q\" (UID: \"4c00defc-fd2b-43e6-ba62-ccc9fd2267fc\") " pod="openstack/dnsmasq-dns-7474d577dc-q7d6q" Dec 13 07:13:06 crc kubenswrapper[4971]: I1213 07:13:06.266767 4971 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/eb706473-627a-4570-9dfc-0e8fe99d59e1-etc-machine-id\") pod \"cinder-scheduler-0\" (UID: \"eb706473-627a-4570-9dfc-0e8fe99d59e1\") " pod="openstack/cinder-scheduler-0" Dec 13 07:13:06 crc kubenswrapper[4971]: I1213 07:13:06.266800 4971 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/4c00defc-fd2b-43e6-ba62-ccc9fd2267fc-dns-svc\") pod \"dnsmasq-dns-7474d577dc-q7d6q\" (UID: \"4c00defc-fd2b-43e6-ba62-ccc9fd2267fc\") " pod="openstack/dnsmasq-dns-7474d577dc-q7d6q" Dec 13 07:13:06 crc kubenswrapper[4971]: I1213 07:13:06.266819 4971 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/e2b4e2e7-cfa5-41d9-833e-b5afc95c91fe-logs\") pod \"cinder-api-0\" (UID: \"e2b4e2e7-cfa5-41d9-833e-b5afc95c91fe\") " pod="openstack/cinder-api-0" Dec 13 07:13:06 crc kubenswrapper[4971]: I1213 07:13:06.266843 4971 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/e2b4e2e7-cfa5-41d9-833e-b5afc95c91fe-config-data-custom\") pod \"cinder-api-0\" (UID: \"e2b4e2e7-cfa5-41d9-833e-b5afc95c91fe\") " pod="openstack/cinder-api-0" Dec 13 07:13:06 crc kubenswrapper[4971]: I1213 07:13:06.266871 4971 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/4c00defc-fd2b-43e6-ba62-ccc9fd2267fc-ovsdbserver-sb\") pod \"dnsmasq-dns-7474d577dc-q7d6q\" (UID: \"4c00defc-fd2b-43e6-ba62-ccc9fd2267fc\") " pod="openstack/dnsmasq-dns-7474d577dc-q7d6q" Dec 13 07:13:06 crc kubenswrapper[4971]: I1213 07:13:06.266954 4971 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/02578e71-2865-4a4b-96e9-5bf6c625183c-config-data\") pod \"barbican-api-5c88bd59bd-2qlbj\" (UID: \"02578e71-2865-4a4b-96e9-5bf6c625183c\") " pod="openstack/barbican-api-5c88bd59bd-2qlbj" Dec 13 07:13:06 crc kubenswrapper[4971]: I1213 07:13:06.267021 4971 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/a709a2b8-324a-468f-bca8-330676b24a85-config\") pod \"dnsmasq-dns-744f57c8cc-w8rwt\" (UID: \"a709a2b8-324a-468f-bca8-330676b24a85\") " pod="openstack/dnsmasq-dns-744f57c8cc-w8rwt" Dec 13 07:13:06 crc kubenswrapper[4971]: I1213 07:13:06.267943 4971 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/a709a2b8-324a-468f-bca8-330676b24a85-config\") pod \"dnsmasq-dns-744f57c8cc-w8rwt\" (UID: \"a709a2b8-324a-468f-bca8-330676b24a85\") " pod="openstack/dnsmasq-dns-744f57c8cc-w8rwt" Dec 13 07:13:06 crc kubenswrapper[4971]: I1213 07:13:06.293342 4971 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-blhpn\" (UniqueName: \"kubernetes.io/projected/a709a2b8-324a-468f-bca8-330676b24a85-kube-api-access-blhpn\") pod \"dnsmasq-dns-744f57c8cc-w8rwt\" (UID: \"a709a2b8-324a-468f-bca8-330676b24a85\") " pod="openstack/dnsmasq-dns-744f57c8cc-w8rwt" Dec 13 07:13:06 crc kubenswrapper[4971]: I1213 07:13:06.464789 4971 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/e2b4e2e7-cfa5-41d9-833e-b5afc95c91fe-config-data\") pod \"cinder-api-0\" (UID: \"e2b4e2e7-cfa5-41d9-833e-b5afc95c91fe\") " pod="openstack/cinder-api-0" Dec 13 07:13:06 crc kubenswrapper[4971]: I1213 07:13:06.465095 4971 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-9bwlc\" (UniqueName: \"kubernetes.io/projected/e2b4e2e7-cfa5-41d9-833e-b5afc95c91fe-kube-api-access-9bwlc\") pod \"cinder-api-0\" (UID: \"e2b4e2e7-cfa5-41d9-833e-b5afc95c91fe\") " pod="openstack/cinder-api-0" Dec 13 07:13:06 crc kubenswrapper[4971]: I1213 07:13:06.465121 4971 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/02578e71-2865-4a4b-96e9-5bf6c625183c-logs\") pod \"barbican-api-5c88bd59bd-2qlbj\" (UID: \"02578e71-2865-4a4b-96e9-5bf6c625183c\") " pod="openstack/barbican-api-5c88bd59bd-2qlbj" Dec 13 07:13:06 crc kubenswrapper[4971]: I1213 07:13:06.465147 4971 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/eb706473-627a-4570-9dfc-0e8fe99d59e1-config-data-custom\") pod \"cinder-scheduler-0\" (UID: \"eb706473-627a-4570-9dfc-0e8fe99d59e1\") " pod="openstack/cinder-scheduler-0" Dec 13 07:13:06 crc kubenswrapper[4971]: I1213 07:13:06.465176 4971 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/e2b4e2e7-cfa5-41d9-833e-b5afc95c91fe-etc-machine-id\") pod \"cinder-api-0\" (UID: \"e2b4e2e7-cfa5-41d9-833e-b5afc95c91fe\") " pod="openstack/cinder-api-0" Dec 13 07:13:06 crc kubenswrapper[4971]: I1213 07:13:06.465198 4971 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/eb706473-627a-4570-9dfc-0e8fe99d59e1-config-data\") pod \"cinder-scheduler-0\" (UID: \"eb706473-627a-4570-9dfc-0e8fe99d59e1\") " pod="openstack/cinder-scheduler-0" Dec 13 07:13:06 crc kubenswrapper[4971]: I1213 07:13:06.465217 4971 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/4c00defc-fd2b-43e6-ba62-ccc9fd2267fc-config\") pod \"dnsmasq-dns-7474d577dc-q7d6q\" (UID: \"4c00defc-fd2b-43e6-ba62-ccc9fd2267fc\") " pod="openstack/dnsmasq-dns-7474d577dc-q7d6q" Dec 13 07:13:06 crc kubenswrapper[4971]: I1213 07:13:06.465235 4971 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/4c00defc-fd2b-43e6-ba62-ccc9fd2267fc-ovsdbserver-nb\") pod \"dnsmasq-dns-7474d577dc-q7d6q\" (UID: \"4c00defc-fd2b-43e6-ba62-ccc9fd2267fc\") " pod="openstack/dnsmasq-dns-7474d577dc-q7d6q" Dec 13 07:13:06 crc kubenswrapper[4971]: I1213 07:13:06.465253 4971 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/02578e71-2865-4a4b-96e9-5bf6c625183c-combined-ca-bundle\") pod \"barbican-api-5c88bd59bd-2qlbj\" (UID: \"02578e71-2865-4a4b-96e9-5bf6c625183c\") " pod="openstack/barbican-api-5c88bd59bd-2qlbj" Dec 13 07:13:06 crc kubenswrapper[4971]: I1213 07:13:06.465269 4971 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/e2b4e2e7-cfa5-41d9-833e-b5afc95c91fe-scripts\") pod \"cinder-api-0\" (UID: \"e2b4e2e7-cfa5-41d9-833e-b5afc95c91fe\") " pod="openstack/cinder-api-0" Dec 13 07:13:06 crc kubenswrapper[4971]: I1213 07:13:06.465299 4971 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-kk4n8\" (UniqueName: \"kubernetes.io/projected/02578e71-2865-4a4b-96e9-5bf6c625183c-kube-api-access-kk4n8\") pod \"barbican-api-5c88bd59bd-2qlbj\" (UID: \"02578e71-2865-4a4b-96e9-5bf6c625183c\") " pod="openstack/barbican-api-5c88bd59bd-2qlbj" Dec 13 07:13:06 crc kubenswrapper[4971]: I1213 07:13:06.465318 4971 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/02578e71-2865-4a4b-96e9-5bf6c625183c-config-data-custom\") pod \"barbican-api-5c88bd59bd-2qlbj\" (UID: \"02578e71-2865-4a4b-96e9-5bf6c625183c\") " pod="openstack/barbican-api-5c88bd59bd-2qlbj" Dec 13 07:13:06 crc kubenswrapper[4971]: I1213 07:13:06.465332 4971 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/eb706473-627a-4570-9dfc-0e8fe99d59e1-combined-ca-bundle\") pod \"cinder-scheduler-0\" (UID: \"eb706473-627a-4570-9dfc-0e8fe99d59e1\") " pod="openstack/cinder-scheduler-0" Dec 13 07:13:06 crc kubenswrapper[4971]: I1213 07:13:06.465351 4971 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-s22xv\" (UniqueName: \"kubernetes.io/projected/eb706473-627a-4570-9dfc-0e8fe99d59e1-kube-api-access-s22xv\") pod \"cinder-scheduler-0\" (UID: \"eb706473-627a-4570-9dfc-0e8fe99d59e1\") " pod="openstack/cinder-scheduler-0" Dec 13 07:13:06 crc kubenswrapper[4971]: I1213 07:13:06.465373 4971 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-h96b2\" (UniqueName: \"kubernetes.io/projected/4c00defc-fd2b-43e6-ba62-ccc9fd2267fc-kube-api-access-h96b2\") pod \"dnsmasq-dns-7474d577dc-q7d6q\" (UID: \"4c00defc-fd2b-43e6-ba62-ccc9fd2267fc\") " pod="openstack/dnsmasq-dns-7474d577dc-q7d6q" Dec 13 07:13:06 crc kubenswrapper[4971]: I1213 07:13:06.465390 4971 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/eb706473-627a-4570-9dfc-0e8fe99d59e1-etc-machine-id\") pod \"cinder-scheduler-0\" (UID: \"eb706473-627a-4570-9dfc-0e8fe99d59e1\") " pod="openstack/cinder-scheduler-0" Dec 13 07:13:06 crc kubenswrapper[4971]: I1213 07:13:06.465409 4971 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/e2b4e2e7-cfa5-41d9-833e-b5afc95c91fe-logs\") pod \"cinder-api-0\" (UID: \"e2b4e2e7-cfa5-41d9-833e-b5afc95c91fe\") " pod="openstack/cinder-api-0" Dec 13 07:13:06 crc kubenswrapper[4971]: I1213 07:13:06.465425 4971 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/4c00defc-fd2b-43e6-ba62-ccc9fd2267fc-dns-svc\") pod \"dnsmasq-dns-7474d577dc-q7d6q\" (UID: \"4c00defc-fd2b-43e6-ba62-ccc9fd2267fc\") " pod="openstack/dnsmasq-dns-7474d577dc-q7d6q" Dec 13 07:13:06 crc kubenswrapper[4971]: I1213 07:13:06.465438 4971 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/e2b4e2e7-cfa5-41d9-833e-b5afc95c91fe-config-data-custom\") pod \"cinder-api-0\" (UID: \"e2b4e2e7-cfa5-41d9-833e-b5afc95c91fe\") " pod="openstack/cinder-api-0" Dec 13 07:13:06 crc kubenswrapper[4971]: I1213 07:13:06.465469 4971 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/4c00defc-fd2b-43e6-ba62-ccc9fd2267fc-ovsdbserver-sb\") pod \"dnsmasq-dns-7474d577dc-q7d6q\" (UID: \"4c00defc-fd2b-43e6-ba62-ccc9fd2267fc\") " pod="openstack/dnsmasq-dns-7474d577dc-q7d6q" Dec 13 07:13:06 crc kubenswrapper[4971]: I1213 07:13:06.465485 4971 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/02578e71-2865-4a4b-96e9-5bf6c625183c-config-data\") pod \"barbican-api-5c88bd59bd-2qlbj\" (UID: \"02578e71-2865-4a4b-96e9-5bf6c625183c\") " pod="openstack/barbican-api-5c88bd59bd-2qlbj" Dec 13 07:13:06 crc kubenswrapper[4971]: I1213 07:13:06.465533 4971 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/e2b4e2e7-cfa5-41d9-833e-b5afc95c91fe-combined-ca-bundle\") pod \"cinder-api-0\" (UID: \"e2b4e2e7-cfa5-41d9-833e-b5afc95c91fe\") " pod="openstack/cinder-api-0" Dec 13 07:13:06 crc kubenswrapper[4971]: I1213 07:13:06.465566 4971 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/eb706473-627a-4570-9dfc-0e8fe99d59e1-scripts\") pod \"cinder-scheduler-0\" (UID: \"eb706473-627a-4570-9dfc-0e8fe99d59e1\") " pod="openstack/cinder-scheduler-0" Dec 13 07:13:06 crc kubenswrapper[4971]: I1213 07:13:06.473049 4971 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/e2b4e2e7-cfa5-41d9-833e-b5afc95c91fe-logs\") pod \"cinder-api-0\" (UID: \"e2b4e2e7-cfa5-41d9-833e-b5afc95c91fe\") " pod="openstack/cinder-api-0" Dec 13 07:13:06 crc kubenswrapper[4971]: I1213 07:13:06.473347 4971 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/eb706473-627a-4570-9dfc-0e8fe99d59e1-etc-machine-id\") pod \"cinder-scheduler-0\" (UID: \"eb706473-627a-4570-9dfc-0e8fe99d59e1\") " pod="openstack/cinder-scheduler-0" Dec 13 07:13:06 crc kubenswrapper[4971]: I1213 07:13:06.474268 4971 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/4c00defc-fd2b-43e6-ba62-ccc9fd2267fc-ovsdbserver-sb\") pod \"dnsmasq-dns-7474d577dc-q7d6q\" (UID: \"4c00defc-fd2b-43e6-ba62-ccc9fd2267fc\") " pod="openstack/dnsmasq-dns-7474d577dc-q7d6q" Dec 13 07:13:06 crc kubenswrapper[4971]: I1213 07:13:06.475277 4971 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/4c00defc-fd2b-43e6-ba62-ccc9fd2267fc-ovsdbserver-nb\") pod \"dnsmasq-dns-7474d577dc-q7d6q\" (UID: \"4c00defc-fd2b-43e6-ba62-ccc9fd2267fc\") " pod="openstack/dnsmasq-dns-7474d577dc-q7d6q" Dec 13 07:13:06 crc kubenswrapper[4971]: I1213 07:13:06.475307 4971 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/4c00defc-fd2b-43e6-ba62-ccc9fd2267fc-config\") pod \"dnsmasq-dns-7474d577dc-q7d6q\" (UID: \"4c00defc-fd2b-43e6-ba62-ccc9fd2267fc\") " pod="openstack/dnsmasq-dns-7474d577dc-q7d6q" Dec 13 07:13:06 crc kubenswrapper[4971]: I1213 07:13:06.475718 4971 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/02578e71-2865-4a4b-96e9-5bf6c625183c-logs\") pod \"barbican-api-5c88bd59bd-2qlbj\" (UID: \"02578e71-2865-4a4b-96e9-5bf6c625183c\") " pod="openstack/barbican-api-5c88bd59bd-2qlbj" Dec 13 07:13:06 crc kubenswrapper[4971]: I1213 07:13:06.476579 4971 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/4c00defc-fd2b-43e6-ba62-ccc9fd2267fc-dns-svc\") pod \"dnsmasq-dns-7474d577dc-q7d6q\" (UID: \"4c00defc-fd2b-43e6-ba62-ccc9fd2267fc\") " pod="openstack/dnsmasq-dns-7474d577dc-q7d6q" Dec 13 07:13:06 crc kubenswrapper[4971]: I1213 07:13:06.478434 4971 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/e2b4e2e7-cfa5-41d9-833e-b5afc95c91fe-config-data\") pod \"cinder-api-0\" (UID: \"e2b4e2e7-cfa5-41d9-833e-b5afc95c91fe\") " pod="openstack/cinder-api-0" Dec 13 07:13:06 crc kubenswrapper[4971]: I1213 07:13:06.479244 4971 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/eb706473-627a-4570-9dfc-0e8fe99d59e1-config-data-custom\") pod \"cinder-scheduler-0\" (UID: \"eb706473-627a-4570-9dfc-0e8fe99d59e1\") " pod="openstack/cinder-scheduler-0" Dec 13 07:13:06 crc kubenswrapper[4971]: I1213 07:13:06.479370 4971 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/e2b4e2e7-cfa5-41d9-833e-b5afc95c91fe-etc-machine-id\") pod \"cinder-api-0\" (UID: \"e2b4e2e7-cfa5-41d9-833e-b5afc95c91fe\") " pod="openstack/cinder-api-0" Dec 13 07:13:06 crc kubenswrapper[4971]: I1213 07:13:06.483337 4971 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/eb706473-627a-4570-9dfc-0e8fe99d59e1-scripts\") pod \"cinder-scheduler-0\" (UID: \"eb706473-627a-4570-9dfc-0e8fe99d59e1\") " pod="openstack/cinder-scheduler-0" Dec 13 07:13:06 crc kubenswrapper[4971]: I1213 07:13:06.484417 4971 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/02578e71-2865-4a4b-96e9-5bf6c625183c-combined-ca-bundle\") pod \"barbican-api-5c88bd59bd-2qlbj\" (UID: \"02578e71-2865-4a4b-96e9-5bf6c625183c\") " pod="openstack/barbican-api-5c88bd59bd-2qlbj" Dec 13 07:13:06 crc kubenswrapper[4971]: I1213 07:13:06.487849 4971 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/e2b4e2e7-cfa5-41d9-833e-b5afc95c91fe-scripts\") pod \"cinder-api-0\" (UID: \"e2b4e2e7-cfa5-41d9-833e-b5afc95c91fe\") " pod="openstack/cinder-api-0" Dec 13 07:13:06 crc kubenswrapper[4971]: I1213 07:13:06.491765 4971 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/eb706473-627a-4570-9dfc-0e8fe99d59e1-config-data\") pod \"cinder-scheduler-0\" (UID: \"eb706473-627a-4570-9dfc-0e8fe99d59e1\") " pod="openstack/cinder-scheduler-0" Dec 13 07:13:06 crc kubenswrapper[4971]: I1213 07:13:06.492298 4971 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/02578e71-2865-4a4b-96e9-5bf6c625183c-config-data-custom\") pod \"barbican-api-5c88bd59bd-2qlbj\" (UID: \"02578e71-2865-4a4b-96e9-5bf6c625183c\") " pod="openstack/barbican-api-5c88bd59bd-2qlbj" Dec 13 07:13:06 crc kubenswrapper[4971]: I1213 07:13:06.495614 4971 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/02578e71-2865-4a4b-96e9-5bf6c625183c-config-data\") pod \"barbican-api-5c88bd59bd-2qlbj\" (UID: \"02578e71-2865-4a4b-96e9-5bf6c625183c\") " pod="openstack/barbican-api-5c88bd59bd-2qlbj" Dec 13 07:13:06 crc kubenswrapper[4971]: I1213 07:13:06.502026 4971 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/e2b4e2e7-cfa5-41d9-833e-b5afc95c91fe-config-data-custom\") pod \"cinder-api-0\" (UID: \"e2b4e2e7-cfa5-41d9-833e-b5afc95c91fe\") " pod="openstack/cinder-api-0" Dec 13 07:13:06 crc kubenswrapper[4971]: I1213 07:13:06.503225 4971 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/eb706473-627a-4570-9dfc-0e8fe99d59e1-combined-ca-bundle\") pod \"cinder-scheduler-0\" (UID: \"eb706473-627a-4570-9dfc-0e8fe99d59e1\") " pod="openstack/cinder-scheduler-0" Dec 13 07:13:06 crc kubenswrapper[4971]: I1213 07:13:06.503405 4971 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/e2b4e2e7-cfa5-41d9-833e-b5afc95c91fe-combined-ca-bundle\") pod \"cinder-api-0\" (UID: \"e2b4e2e7-cfa5-41d9-833e-b5afc95c91fe\") " pod="openstack/cinder-api-0" Dec 13 07:13:06 crc kubenswrapper[4971]: I1213 07:13:06.507346 4971 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-kk4n8\" (UniqueName: \"kubernetes.io/projected/02578e71-2865-4a4b-96e9-5bf6c625183c-kube-api-access-kk4n8\") pod \"barbican-api-5c88bd59bd-2qlbj\" (UID: \"02578e71-2865-4a4b-96e9-5bf6c625183c\") " pod="openstack/barbican-api-5c88bd59bd-2qlbj" Dec 13 07:13:06 crc kubenswrapper[4971]: I1213 07:13:06.507419 4971 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-9bwlc\" (UniqueName: \"kubernetes.io/projected/e2b4e2e7-cfa5-41d9-833e-b5afc95c91fe-kube-api-access-9bwlc\") pod \"cinder-api-0\" (UID: \"e2b4e2e7-cfa5-41d9-833e-b5afc95c91fe\") " pod="openstack/cinder-api-0" Dec 13 07:13:06 crc kubenswrapper[4971]: I1213 07:13:06.511447 4971 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-h96b2\" (UniqueName: \"kubernetes.io/projected/4c00defc-fd2b-43e6-ba62-ccc9fd2267fc-kube-api-access-h96b2\") pod \"dnsmasq-dns-7474d577dc-q7d6q\" (UID: \"4c00defc-fd2b-43e6-ba62-ccc9fd2267fc\") " pod="openstack/dnsmasq-dns-7474d577dc-q7d6q" Dec 13 07:13:06 crc kubenswrapper[4971]: I1213 07:13:06.516013 4971 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-s22xv\" (UniqueName: \"kubernetes.io/projected/eb706473-627a-4570-9dfc-0e8fe99d59e1-kube-api-access-s22xv\") pod \"cinder-scheduler-0\" (UID: \"eb706473-627a-4570-9dfc-0e8fe99d59e1\") " pod="openstack/cinder-scheduler-0" Dec 13 07:13:06 crc kubenswrapper[4971]: I1213 07:13:06.548777 4971 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-7474d577dc-q7d6q" Dec 13 07:13:06 crc kubenswrapper[4971]: I1213 07:13:06.569773 4971 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-api-5c88bd59bd-2qlbj" Dec 13 07:13:06 crc kubenswrapper[4971]: I1213 07:13:06.584170 4971 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-api-0" Dec 13 07:13:06 crc kubenswrapper[4971]: I1213 07:13:06.696100 4971 generic.go:334] "Generic (PLEG): container finished" podID="56e56ce1-4dd1-4ae5-b868-e98d27d5ad94" containerID="a0e5a1ecd136567279050709089de36451185ac24d99f2e15b755c1d0658ad99" exitCode=2 Dec 13 07:13:06 crc kubenswrapper[4971]: I1213 07:13:06.696462 4971 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-744f57c8cc-w8rwt" Dec 13 07:13:06 crc kubenswrapper[4971]: I1213 07:13:06.696251 4971 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"56e56ce1-4dd1-4ae5-b868-e98d27d5ad94","Type":"ContainerDied","Data":"a0e5a1ecd136567279050709089de36451185ac24d99f2e15b755c1d0658ad99"} Dec 13 07:13:06 crc kubenswrapper[4971]: I1213 07:13:06.782469 4971 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-scheduler-0" Dec 13 07:13:06 crc kubenswrapper[4971]: I1213 07:13:06.872054 4971 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/barbican-worker-c7c6b44fc-8hd9n"] Dec 13 07:13:06 crc kubenswrapper[4971]: I1213 07:13:06.897241 4971 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-744f57c8cc-w8rwt" Dec 13 07:13:06 crc kubenswrapper[4971]: I1213 07:13:06.988727 4971 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/a709a2b8-324a-468f-bca8-330676b24a85-ovsdbserver-sb\") pod \"a709a2b8-324a-468f-bca8-330676b24a85\" (UID: \"a709a2b8-324a-468f-bca8-330676b24a85\") " Dec 13 07:13:06 crc kubenswrapper[4971]: I1213 07:13:06.988802 4971 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/a709a2b8-324a-468f-bca8-330676b24a85-config\") pod \"a709a2b8-324a-468f-bca8-330676b24a85\" (UID: \"a709a2b8-324a-468f-bca8-330676b24a85\") " Dec 13 07:13:06 crc kubenswrapper[4971]: I1213 07:13:06.988930 4971 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/a709a2b8-324a-468f-bca8-330676b24a85-dns-svc\") pod \"a709a2b8-324a-468f-bca8-330676b24a85\" (UID: \"a709a2b8-324a-468f-bca8-330676b24a85\") " Dec 13 07:13:06 crc kubenswrapper[4971]: I1213 07:13:06.988995 4971 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/a709a2b8-324a-468f-bca8-330676b24a85-ovsdbserver-nb\") pod \"a709a2b8-324a-468f-bca8-330676b24a85\" (UID: \"a709a2b8-324a-468f-bca8-330676b24a85\") " Dec 13 07:13:06 crc kubenswrapper[4971]: I1213 07:13:06.989071 4971 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-blhpn\" (UniqueName: \"kubernetes.io/projected/a709a2b8-324a-468f-bca8-330676b24a85-kube-api-access-blhpn\") pod \"a709a2b8-324a-468f-bca8-330676b24a85\" (UID: \"a709a2b8-324a-468f-bca8-330676b24a85\") " Dec 13 07:13:06 crc kubenswrapper[4971]: I1213 07:13:06.990924 4971 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/a709a2b8-324a-468f-bca8-330676b24a85-config" (OuterVolumeSpecName: "config") pod "a709a2b8-324a-468f-bca8-330676b24a85" (UID: "a709a2b8-324a-468f-bca8-330676b24a85"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 13 07:13:06 crc kubenswrapper[4971]: I1213 07:13:06.991321 4971 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/a709a2b8-324a-468f-bca8-330676b24a85-ovsdbserver-sb" (OuterVolumeSpecName: "ovsdbserver-sb") pod "a709a2b8-324a-468f-bca8-330676b24a85" (UID: "a709a2b8-324a-468f-bca8-330676b24a85"). InnerVolumeSpecName "ovsdbserver-sb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 13 07:13:06 crc kubenswrapper[4971]: I1213 07:13:06.991713 4971 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/a709a2b8-324a-468f-bca8-330676b24a85-dns-svc" (OuterVolumeSpecName: "dns-svc") pod "a709a2b8-324a-468f-bca8-330676b24a85" (UID: "a709a2b8-324a-468f-bca8-330676b24a85"). InnerVolumeSpecName "dns-svc". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 13 07:13:06 crc kubenswrapper[4971]: I1213 07:13:06.992128 4971 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/a709a2b8-324a-468f-bca8-330676b24a85-ovsdbserver-nb" (OuterVolumeSpecName: "ovsdbserver-nb") pod "a709a2b8-324a-468f-bca8-330676b24a85" (UID: "a709a2b8-324a-468f-bca8-330676b24a85"). InnerVolumeSpecName "ovsdbserver-nb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 13 07:13:06 crc kubenswrapper[4971]: I1213 07:13:06.995167 4971 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/a709a2b8-324a-468f-bca8-330676b24a85-kube-api-access-blhpn" (OuterVolumeSpecName: "kube-api-access-blhpn") pod "a709a2b8-324a-468f-bca8-330676b24a85" (UID: "a709a2b8-324a-468f-bca8-330676b24a85"). InnerVolumeSpecName "kube-api-access-blhpn". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 13 07:13:07 crc kubenswrapper[4971]: I1213 07:13:07.040790 4971 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/barbican-keystone-listener-6959d6f94b-tq6hh"] Dec 13 07:13:07 crc kubenswrapper[4971]: I1213 07:13:07.092625 4971 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/a709a2b8-324a-468f-bca8-330676b24a85-config\") on node \"crc\" DevicePath \"\"" Dec 13 07:13:07 crc kubenswrapper[4971]: I1213 07:13:07.092653 4971 reconciler_common.go:293] "Volume detached for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/a709a2b8-324a-468f-bca8-330676b24a85-dns-svc\") on node \"crc\" DevicePath \"\"" Dec 13 07:13:07 crc kubenswrapper[4971]: I1213 07:13:07.092663 4971 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/a709a2b8-324a-468f-bca8-330676b24a85-ovsdbserver-nb\") on node \"crc\" DevicePath \"\"" Dec 13 07:13:07 crc kubenswrapper[4971]: I1213 07:13:07.092674 4971 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-blhpn\" (UniqueName: \"kubernetes.io/projected/a709a2b8-324a-468f-bca8-330676b24a85-kube-api-access-blhpn\") on node \"crc\" DevicePath \"\"" Dec 13 07:13:07 crc kubenswrapper[4971]: I1213 07:13:07.092683 4971 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/a709a2b8-324a-468f-bca8-330676b24a85-ovsdbserver-sb\") on node \"crc\" DevicePath \"\"" Dec 13 07:13:07 crc kubenswrapper[4971]: I1213 07:13:07.142719 4971 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-7474d577dc-q7d6q"] Dec 13 07:13:07 crc kubenswrapper[4971]: I1213 07:13:07.350683 4971 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cinder-api-0"] Dec 13 07:13:07 crc kubenswrapper[4971]: W1213 07:13:07.357211 4971 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod02578e71_2865_4a4b_96e9_5bf6c625183c.slice/crio-255762a06cafe12675758dacf9268763d0b8bf0ed85d57ca20e28c70a0253280 WatchSource:0}: Error finding container 255762a06cafe12675758dacf9268763d0b8bf0ed85d57ca20e28c70a0253280: Status 404 returned error can't find the container with id 255762a06cafe12675758dacf9268763d0b8bf0ed85d57ca20e28c70a0253280 Dec 13 07:13:07 crc kubenswrapper[4971]: I1213 07:13:07.357966 4971 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/barbican-api-5c88bd59bd-2qlbj"] Dec 13 07:13:07 crc kubenswrapper[4971]: I1213 07:13:07.710268 4971 generic.go:334] "Generic (PLEG): container finished" podID="56e56ce1-4dd1-4ae5-b868-e98d27d5ad94" containerID="1e636e556657fdb168cfe90bbd03b3e59853b99c24a8b8151d7adedc027197ad" exitCode=0 Dec 13 07:13:07 crc kubenswrapper[4971]: I1213 07:13:07.710323 4971 generic.go:334] "Generic (PLEG): container finished" podID="56e56ce1-4dd1-4ae5-b868-e98d27d5ad94" containerID="5d8a06b64cc9778b600eb7ee56a36445340d0fb5dafba3ee09865a0b6e2343ea" exitCode=0 Dec 13 07:13:07 crc kubenswrapper[4971]: I1213 07:13:07.710308 4971 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"56e56ce1-4dd1-4ae5-b868-e98d27d5ad94","Type":"ContainerDied","Data":"1e636e556657fdb168cfe90bbd03b3e59853b99c24a8b8151d7adedc027197ad"} Dec 13 07:13:07 crc kubenswrapper[4971]: I1213 07:13:07.710377 4971 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"56e56ce1-4dd1-4ae5-b868-e98d27d5ad94","Type":"ContainerDied","Data":"5d8a06b64cc9778b600eb7ee56a36445340d0fb5dafba3ee09865a0b6e2343ea"} Dec 13 07:13:07 crc kubenswrapper[4971]: I1213 07:13:07.711717 4971 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-worker-c7c6b44fc-8hd9n" event={"ID":"fb068297-28dd-4a70-acbf-650b9edf380e","Type":"ContainerStarted","Data":"04b2b33460fd24bd9f451736bfa41f71d58d85302bd4bb9d1d80b20ed348d529"} Dec 13 07:13:07 crc kubenswrapper[4971]: I1213 07:13:07.713205 4971 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-7474d577dc-q7d6q" event={"ID":"4c00defc-fd2b-43e6-ba62-ccc9fd2267fc","Type":"ContainerStarted","Data":"5461c8ff70588f04523e12e1d7c4bd641e9422a3c5001d3ae5943e8399800888"} Dec 13 07:13:07 crc kubenswrapper[4971]: I1213 07:13:07.715123 4971 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-api-0" event={"ID":"e2b4e2e7-cfa5-41d9-833e-b5afc95c91fe","Type":"ContainerStarted","Data":"f3b9f07fbc7f8d5f03029bb89ed518d701ae9188dd45027ac37a6c8b1e9c35e4"} Dec 13 07:13:07 crc kubenswrapper[4971]: I1213 07:13:07.716627 4971 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-api-5c88bd59bd-2qlbj" event={"ID":"02578e71-2865-4a4b-96e9-5bf6c625183c","Type":"ContainerStarted","Data":"255762a06cafe12675758dacf9268763d0b8bf0ed85d57ca20e28c70a0253280"} Dec 13 07:13:07 crc kubenswrapper[4971]: I1213 07:13:07.718908 4971 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-744f57c8cc-w8rwt" Dec 13 07:13:07 crc kubenswrapper[4971]: I1213 07:13:07.718921 4971 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-keystone-listener-6959d6f94b-tq6hh" event={"ID":"12a4a604-178e-4ad8-a0f7-e2e3d47590bd","Type":"ContainerStarted","Data":"595beec21a03d60d1463f6440b17cadbe52e6df088831f60f82335060756a6b4"} Dec 13 07:13:07 crc kubenswrapper[4971]: I1213 07:13:07.807403 4971 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cinder-scheduler-0"] Dec 13 07:13:07 crc kubenswrapper[4971]: I1213 07:13:07.877750 4971 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-744f57c8cc-w8rwt"] Dec 13 07:13:07 crc kubenswrapper[4971]: I1213 07:13:07.885715 4971 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-744f57c8cc-w8rwt"] Dec 13 07:13:07 crc kubenswrapper[4971]: I1213 07:13:07.950893 4971 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/horizon-8699dc887d-sl2h6" Dec 13 07:13:07 crc kubenswrapper[4971]: I1213 07:13:07.970306 4971 pod_container_manager_linux.go:210] "Failed to delete cgroup paths" cgroupName=["kubepods","besteffort","podf9bd96be-6d7f-415e-b67c-5bd02b05c536"] err="unable to destroy cgroup paths for cgroup [kubepods besteffort podf9bd96be-6d7f-415e-b67c-5bd02b05c536] : Timed out while waiting for systemd to remove kubepods-besteffort-podf9bd96be_6d7f_415e_b67c_5bd02b05c536.slice" Dec 13 07:13:07 crc kubenswrapper[4971]: E1213 07:13:07.970364 4971 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to delete cgroup paths for [kubepods besteffort podf9bd96be-6d7f-415e-b67c-5bd02b05c536] : unable to destroy cgroup paths for cgroup [kubepods besteffort podf9bd96be-6d7f-415e-b67c-5bd02b05c536] : Timed out while waiting for systemd to remove kubepods-besteffort-podf9bd96be_6d7f_415e_b67c_5bd02b05c536.slice" pod="openstack/horizon-fbfd486df-xmvls" podUID="f9bd96be-6d7f-415e-b67c-5bd02b05c536" Dec 13 07:13:07 crc kubenswrapper[4971]: I1213 07:13:07.982980 4971 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/horizon-797dfcfc58-wlch7" Dec 13 07:13:08 crc kubenswrapper[4971]: I1213 07:13:08.044156 4971 pod_container_manager_linux.go:210] "Failed to delete cgroup paths" cgroupName=["kubepods","besteffort","podff35e760-13e7-4145-9970-742b73a36176"] err="unable to destroy cgroup paths for cgroup [kubepods besteffort podff35e760-13e7-4145-9970-742b73a36176] : Timed out while waiting for systemd to remove kubepods-besteffort-podff35e760_13e7_4145_9970_742b73a36176.slice" Dec 13 07:13:08 crc kubenswrapper[4971]: I1213 07:13:08.729885 4971 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/horizon-fbfd486df-xmvls" Dec 13 07:13:08 crc kubenswrapper[4971]: I1213 07:13:08.730837 4971 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-scheduler-0" event={"ID":"eb706473-627a-4570-9dfc-0e8fe99d59e1","Type":"ContainerStarted","Data":"76b4096e1c2bbcba6f9d6fba5708ba0d182ba2eaad08a0ee5e7e30f8217e05ac"} Dec 13 07:13:08 crc kubenswrapper[4971]: I1213 07:13:08.790024 4971 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/horizon-fbfd486df-xmvls"] Dec 13 07:13:08 crc kubenswrapper[4971]: I1213 07:13:08.802673 4971 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/horizon-fbfd486df-xmvls"] Dec 13 07:13:09 crc kubenswrapper[4971]: I1213 07:13:09.019868 4971 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/cinder-api-0"] Dec 13 07:13:09 crc kubenswrapper[4971]: I1213 07:13:09.579782 4971 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/neutron-6597fdddd4-q5zr5" Dec 13 07:13:09 crc kubenswrapper[4971]: I1213 07:13:09.748025 4971 generic.go:334] "Generic (PLEG): container finished" podID="4c00defc-fd2b-43e6-ba62-ccc9fd2267fc" containerID="5763ec7db6e3b9278c34e8c29379aa4998a4685bb3b627ac91904b0d0f9e6a67" exitCode=0 Dec 13 07:13:09 crc kubenswrapper[4971]: I1213 07:13:09.748121 4971 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-7474d577dc-q7d6q" event={"ID":"4c00defc-fd2b-43e6-ba62-ccc9fd2267fc","Type":"ContainerDied","Data":"5763ec7db6e3b9278c34e8c29379aa4998a4685bb3b627ac91904b0d0f9e6a67"} Dec 13 07:13:09 crc kubenswrapper[4971]: I1213 07:13:09.755791 4971 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-api-0" event={"ID":"e2b4e2e7-cfa5-41d9-833e-b5afc95c91fe","Type":"ContainerStarted","Data":"1627e8372a698405e5dd80f505feda5b356d74b9735ba8519bf519633e0519f1"} Dec 13 07:13:09 crc kubenswrapper[4971]: I1213 07:13:09.762418 4971 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-api-5c88bd59bd-2qlbj" event={"ID":"02578e71-2865-4a4b-96e9-5bf6c625183c","Type":"ContainerStarted","Data":"75fc3d56284952d3e64a99eccdb0f35e87027803ee64f33d79a9a487157aad17"} Dec 13 07:13:09 crc kubenswrapper[4971]: I1213 07:13:09.762456 4971 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-api-5c88bd59bd-2qlbj" event={"ID":"02578e71-2865-4a4b-96e9-5bf6c625183c","Type":"ContainerStarted","Data":"5c2237976dc29579a8c63c4fbffbcb524f6e0ae5cbf13be0ece76df573636e9c"} Dec 13 07:13:09 crc kubenswrapper[4971]: I1213 07:13:09.816431 4971 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="a709a2b8-324a-468f-bca8-330676b24a85" path="/var/lib/kubelet/pods/a709a2b8-324a-468f-bca8-330676b24a85/volumes" Dec 13 07:13:09 crc kubenswrapper[4971]: I1213 07:13:09.817296 4971 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="f9bd96be-6d7f-415e-b67c-5bd02b05c536" path="/var/lib/kubelet/pods/f9bd96be-6d7f-415e-b67c-5bd02b05c536/volumes" Dec 13 07:13:10 crc kubenswrapper[4971]: I1213 07:13:10.148496 4971 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/horizon-797dfcfc58-wlch7" Dec 13 07:13:10 crc kubenswrapper[4971]: I1213 07:13:10.250795 4971 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/horizon-8699dc887d-sl2h6"] Dec 13 07:13:10 crc kubenswrapper[4971]: I1213 07:13:10.251359 4971 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/horizon-8699dc887d-sl2h6" podUID="772eedab-7c62-454b-b722-ee453b39e89f" containerName="horizon-log" containerID="cri-o://2074da306f3c927fa0e63c982fab7574e67b1f06c38e55ba7dc406c0d32df42a" gracePeriod=30 Dec 13 07:13:10 crc kubenswrapper[4971]: I1213 07:13:10.251864 4971 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/horizon-8699dc887d-sl2h6" podUID="772eedab-7c62-454b-b722-ee453b39e89f" containerName="horizon" containerID="cri-o://4a148882fad66a5e1444a664e27875de5e47f0f5a778fde8428a33e1ba8f489f" gracePeriod=30 Dec 13 07:13:10 crc kubenswrapper[4971]: I1213 07:13:10.277142 4971 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/horizon-8699dc887d-sl2h6" podUID="772eedab-7c62-454b-b722-ee453b39e89f" containerName="horizon" probeResult="failure" output="Get \"https://10.217.0.140:8443/dashboard/auth/login/?next=/dashboard/\": EOF" Dec 13 07:13:10 crc kubenswrapper[4971]: I1213 07:13:10.299349 4971 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/horizon-8699dc887d-sl2h6" podUID="772eedab-7c62-454b-b722-ee453b39e89f" containerName="horizon" probeResult="failure" output="Get \"https://10.217.0.140:8443/dashboard/auth/login/?next=/dashboard/\": read tcp 10.217.0.2:41920->10.217.0.140:8443: read: connection reset by peer" Dec 13 07:13:10 crc kubenswrapper[4971]: I1213 07:13:10.781433 4971 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-7474d577dc-q7d6q" event={"ID":"4c00defc-fd2b-43e6-ba62-ccc9fd2267fc","Type":"ContainerStarted","Data":"9647f002123326031aafc6c3c41875e56907e06cbf016e592145c84346eae3de"} Dec 13 07:13:10 crc kubenswrapper[4971]: I1213 07:13:10.781871 4971 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/dnsmasq-dns-7474d577dc-q7d6q" Dec 13 07:13:10 crc kubenswrapper[4971]: I1213 07:13:10.801694 4971 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-api-0" event={"ID":"e2b4e2e7-cfa5-41d9-833e-b5afc95c91fe","Type":"ContainerStarted","Data":"cffdf5b9bc9d0261388ccb6fc29ca51169e8bc553aec7ae953ebc5ee2bd00652"} Dec 13 07:13:10 crc kubenswrapper[4971]: I1213 07:13:10.801844 4971 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/cinder-api-0" podUID="e2b4e2e7-cfa5-41d9-833e-b5afc95c91fe" containerName="cinder-api-log" containerID="cri-o://1627e8372a698405e5dd80f505feda5b356d74b9735ba8519bf519633e0519f1" gracePeriod=30 Dec 13 07:13:10 crc kubenswrapper[4971]: I1213 07:13:10.801876 4971 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/cinder-api-0" podUID="e2b4e2e7-cfa5-41d9-833e-b5afc95c91fe" containerName="cinder-api" containerID="cri-o://cffdf5b9bc9d0261388ccb6fc29ca51169e8bc553aec7ae953ebc5ee2bd00652" gracePeriod=30 Dec 13 07:13:10 crc kubenswrapper[4971]: I1213 07:13:10.801997 4971 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/barbican-api-5c88bd59bd-2qlbj" Dec 13 07:13:10 crc kubenswrapper[4971]: I1213 07:13:10.802026 4971 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/barbican-api-5c88bd59bd-2qlbj" Dec 13 07:13:10 crc kubenswrapper[4971]: I1213 07:13:10.824823 4971 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/dnsmasq-dns-7474d577dc-q7d6q" podStartSLOduration=5.824784633 podStartE2EDuration="5.824784633s" podCreationTimestamp="2025-12-13 07:13:05 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-13 07:13:10.811391553 +0000 UTC m=+1447.415801001" watchObservedRunningTime="2025-12-13 07:13:10.824784633 +0000 UTC m=+1447.429194081" Dec 13 07:13:10 crc kubenswrapper[4971]: I1213 07:13:10.849559 4971 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/cinder-api-0" podStartSLOduration=4.849539612 podStartE2EDuration="4.849539612s" podCreationTimestamp="2025-12-13 07:13:06 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-13 07:13:10.842006617 +0000 UTC m=+1447.446416075" watchObservedRunningTime="2025-12-13 07:13:10.849539612 +0000 UTC m=+1447.453949050" Dec 13 07:13:10 crc kubenswrapper[4971]: I1213 07:13:10.865472 4971 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/barbican-api-5c88bd59bd-2qlbj" podStartSLOduration=5.865456344 podStartE2EDuration="5.865456344s" podCreationTimestamp="2025-12-13 07:13:05 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-13 07:13:10.861090397 +0000 UTC m=+1447.465499855" watchObservedRunningTime="2025-12-13 07:13:10.865456344 +0000 UTC m=+1447.469865792" Dec 13 07:13:11 crc kubenswrapper[4971]: I1213 07:13:11.585191 4971 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/cinder-api-0" Dec 13 07:13:11 crc kubenswrapper[4971]: I1213 07:13:11.828612 4971 generic.go:334] "Generic (PLEG): container finished" podID="e2b4e2e7-cfa5-41d9-833e-b5afc95c91fe" containerID="cffdf5b9bc9d0261388ccb6fc29ca51169e8bc553aec7ae953ebc5ee2bd00652" exitCode=0 Dec 13 07:13:11 crc kubenswrapper[4971]: I1213 07:13:11.828944 4971 generic.go:334] "Generic (PLEG): container finished" podID="e2b4e2e7-cfa5-41d9-833e-b5afc95c91fe" containerID="1627e8372a698405e5dd80f505feda5b356d74b9735ba8519bf519633e0519f1" exitCode=143 Dec 13 07:13:11 crc kubenswrapper[4971]: I1213 07:13:11.828704 4971 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-api-0" event={"ID":"e2b4e2e7-cfa5-41d9-833e-b5afc95c91fe","Type":"ContainerDied","Data":"cffdf5b9bc9d0261388ccb6fc29ca51169e8bc553aec7ae953ebc5ee2bd00652"} Dec 13 07:13:11 crc kubenswrapper[4971]: I1213 07:13:11.829027 4971 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-api-0" event={"ID":"e2b4e2e7-cfa5-41d9-833e-b5afc95c91fe","Type":"ContainerDied","Data":"1627e8372a698405e5dd80f505feda5b356d74b9735ba8519bf519633e0519f1"} Dec 13 07:13:11 crc kubenswrapper[4971]: I1213 07:13:11.831693 4971 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-keystone-listener-6959d6f94b-tq6hh" event={"ID":"12a4a604-178e-4ad8-a0f7-e2e3d47590bd","Type":"ContainerStarted","Data":"a09926b331bbe6ba1c4c5a57c6c0985c1c4eae915a2971673ad5afba157c158f"} Dec 13 07:13:11 crc kubenswrapper[4971]: I1213 07:13:11.833683 4971 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Dec 13 07:13:11 crc kubenswrapper[4971]: I1213 07:13:11.836482 4971 generic.go:334] "Generic (PLEG): container finished" podID="56e56ce1-4dd1-4ae5-b868-e98d27d5ad94" containerID="d208a7c037f088ceda5804583d1d1761594f3fbc2ad4ae265ebc6e2bf875c418" exitCode=0 Dec 13 07:13:11 crc kubenswrapper[4971]: I1213 07:13:11.836560 4971 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"56e56ce1-4dd1-4ae5-b868-e98d27d5ad94","Type":"ContainerDied","Data":"d208a7c037f088ceda5804583d1d1761594f3fbc2ad4ae265ebc6e2bf875c418"} Dec 13 07:13:11 crc kubenswrapper[4971]: I1213 07:13:11.836596 4971 scope.go:117] "RemoveContainer" containerID="1e636e556657fdb168cfe90bbd03b3e59853b99c24a8b8151d7adedc027197ad" Dec 13 07:13:11 crc kubenswrapper[4971]: I1213 07:13:11.840752 4971 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-scheduler-0" event={"ID":"eb706473-627a-4570-9dfc-0e8fe99d59e1","Type":"ContainerStarted","Data":"c2bb3a07998a3cffd09385e18ab96ef8b7a2067d9333b9e5fe773a03866bf0a5"} Dec 13 07:13:11 crc kubenswrapper[4971]: I1213 07:13:11.845290 4971 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-worker-c7c6b44fc-8hd9n" event={"ID":"fb068297-28dd-4a70-acbf-650b9edf380e","Type":"ContainerStarted","Data":"eba61d2020512c5673543203140717ffeae6a97793e60dd1db966dbbda7df3ee"} Dec 13 07:13:11 crc kubenswrapper[4971]: I1213 07:13:11.910104 4971 scope.go:117] "RemoveContainer" containerID="a0e5a1ecd136567279050709089de36451185ac24d99f2e15b755c1d0658ad99" Dec 13 07:13:11 crc kubenswrapper[4971]: I1213 07:13:11.950405 4971 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-api-0" Dec 13 07:13:11 crc kubenswrapper[4971]: I1213 07:13:11.951688 4971 scope.go:117] "RemoveContainer" containerID="d208a7c037f088ceda5804583d1d1761594f3fbc2ad4ae265ebc6e2bf875c418" Dec 13 07:13:11 crc kubenswrapper[4971]: I1213 07:13:11.961854 4971 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/56e56ce1-4dd1-4ae5-b868-e98d27d5ad94-run-httpd\") pod \"56e56ce1-4dd1-4ae5-b868-e98d27d5ad94\" (UID: \"56e56ce1-4dd1-4ae5-b868-e98d27d5ad94\") " Dec 13 07:13:11 crc kubenswrapper[4971]: I1213 07:13:11.961932 4971 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/56e56ce1-4dd1-4ae5-b868-e98d27d5ad94-log-httpd\") pod \"56e56ce1-4dd1-4ae5-b868-e98d27d5ad94\" (UID: \"56e56ce1-4dd1-4ae5-b868-e98d27d5ad94\") " Dec 13 07:13:11 crc kubenswrapper[4971]: I1213 07:13:11.961983 4971 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/56e56ce1-4dd1-4ae5-b868-e98d27d5ad94-sg-core-conf-yaml\") pod \"56e56ce1-4dd1-4ae5-b868-e98d27d5ad94\" (UID: \"56e56ce1-4dd1-4ae5-b868-e98d27d5ad94\") " Dec 13 07:13:11 crc kubenswrapper[4971]: I1213 07:13:11.962006 4971 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/56e56ce1-4dd1-4ae5-b868-e98d27d5ad94-combined-ca-bundle\") pod \"56e56ce1-4dd1-4ae5-b868-e98d27d5ad94\" (UID: \"56e56ce1-4dd1-4ae5-b868-e98d27d5ad94\") " Dec 13 07:13:11 crc kubenswrapper[4971]: I1213 07:13:11.962056 4971 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/56e56ce1-4dd1-4ae5-b868-e98d27d5ad94-scripts\") pod \"56e56ce1-4dd1-4ae5-b868-e98d27d5ad94\" (UID: \"56e56ce1-4dd1-4ae5-b868-e98d27d5ad94\") " Dec 13 07:13:11 crc kubenswrapper[4971]: I1213 07:13:11.962129 4971 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-hw5mk\" (UniqueName: \"kubernetes.io/projected/56e56ce1-4dd1-4ae5-b868-e98d27d5ad94-kube-api-access-hw5mk\") pod \"56e56ce1-4dd1-4ae5-b868-e98d27d5ad94\" (UID: \"56e56ce1-4dd1-4ae5-b868-e98d27d5ad94\") " Dec 13 07:13:11 crc kubenswrapper[4971]: I1213 07:13:11.962194 4971 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/56e56ce1-4dd1-4ae5-b868-e98d27d5ad94-config-data\") pod \"56e56ce1-4dd1-4ae5-b868-e98d27d5ad94\" (UID: \"56e56ce1-4dd1-4ae5-b868-e98d27d5ad94\") " Dec 13 07:13:11 crc kubenswrapper[4971]: I1213 07:13:11.962711 4971 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/56e56ce1-4dd1-4ae5-b868-e98d27d5ad94-log-httpd" (OuterVolumeSpecName: "log-httpd") pod "56e56ce1-4dd1-4ae5-b868-e98d27d5ad94" (UID: "56e56ce1-4dd1-4ae5-b868-e98d27d5ad94"). InnerVolumeSpecName "log-httpd". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 13 07:13:11 crc kubenswrapper[4971]: I1213 07:13:11.962835 4971 reconciler_common.go:293] "Volume detached for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/56e56ce1-4dd1-4ae5-b868-e98d27d5ad94-log-httpd\") on node \"crc\" DevicePath \"\"" Dec 13 07:13:11 crc kubenswrapper[4971]: I1213 07:13:11.963562 4971 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/neutron-7956479f45-8g7zz" Dec 13 07:13:11 crc kubenswrapper[4971]: I1213 07:13:11.963576 4971 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/56e56ce1-4dd1-4ae5-b868-e98d27d5ad94-run-httpd" (OuterVolumeSpecName: "run-httpd") pod "56e56ce1-4dd1-4ae5-b868-e98d27d5ad94" (UID: "56e56ce1-4dd1-4ae5-b868-e98d27d5ad94"). InnerVolumeSpecName "run-httpd". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 13 07:13:12 crc kubenswrapper[4971]: I1213 07:13:12.009074 4971 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/56e56ce1-4dd1-4ae5-b868-e98d27d5ad94-kube-api-access-hw5mk" (OuterVolumeSpecName: "kube-api-access-hw5mk") pod "56e56ce1-4dd1-4ae5-b868-e98d27d5ad94" (UID: "56e56ce1-4dd1-4ae5-b868-e98d27d5ad94"). InnerVolumeSpecName "kube-api-access-hw5mk". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 13 07:13:12 crc kubenswrapper[4971]: I1213 07:13:12.013064 4971 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/56e56ce1-4dd1-4ae5-b868-e98d27d5ad94-scripts" (OuterVolumeSpecName: "scripts") pod "56e56ce1-4dd1-4ae5-b868-e98d27d5ad94" (UID: "56e56ce1-4dd1-4ae5-b868-e98d27d5ad94"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 13 07:13:12 crc kubenswrapper[4971]: I1213 07:13:12.029748 4971 scope.go:117] "RemoveContainer" containerID="5d8a06b64cc9778b600eb7ee56a36445340d0fb5dafba3ee09865a0b6e2343ea" Dec 13 07:13:12 crc kubenswrapper[4971]: I1213 07:13:12.067562 4971 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/e2b4e2e7-cfa5-41d9-833e-b5afc95c91fe-etc-machine-id\") pod \"e2b4e2e7-cfa5-41d9-833e-b5afc95c91fe\" (UID: \"e2b4e2e7-cfa5-41d9-833e-b5afc95c91fe\") " Dec 13 07:13:12 crc kubenswrapper[4971]: I1213 07:13:12.067764 4971 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/e2b4e2e7-cfa5-41d9-833e-b5afc95c91fe-etc-machine-id" (OuterVolumeSpecName: "etc-machine-id") pod "e2b4e2e7-cfa5-41d9-833e-b5afc95c91fe" (UID: "e2b4e2e7-cfa5-41d9-833e-b5afc95c91fe"). InnerVolumeSpecName "etc-machine-id". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 13 07:13:12 crc kubenswrapper[4971]: I1213 07:13:12.067977 4971 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-9bwlc\" (UniqueName: \"kubernetes.io/projected/e2b4e2e7-cfa5-41d9-833e-b5afc95c91fe-kube-api-access-9bwlc\") pod \"e2b4e2e7-cfa5-41d9-833e-b5afc95c91fe\" (UID: \"e2b4e2e7-cfa5-41d9-833e-b5afc95c91fe\") " Dec 13 07:13:12 crc kubenswrapper[4971]: I1213 07:13:12.073581 4971 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/e2b4e2e7-cfa5-41d9-833e-b5afc95c91fe-config-data-custom\") pod \"e2b4e2e7-cfa5-41d9-833e-b5afc95c91fe\" (UID: \"e2b4e2e7-cfa5-41d9-833e-b5afc95c91fe\") " Dec 13 07:13:12 crc kubenswrapper[4971]: I1213 07:13:12.073657 4971 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/e2b4e2e7-cfa5-41d9-833e-b5afc95c91fe-config-data\") pod \"e2b4e2e7-cfa5-41d9-833e-b5afc95c91fe\" (UID: \"e2b4e2e7-cfa5-41d9-833e-b5afc95c91fe\") " Dec 13 07:13:12 crc kubenswrapper[4971]: I1213 07:13:12.073697 4971 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/e2b4e2e7-cfa5-41d9-833e-b5afc95c91fe-scripts\") pod \"e2b4e2e7-cfa5-41d9-833e-b5afc95c91fe\" (UID: \"e2b4e2e7-cfa5-41d9-833e-b5afc95c91fe\") " Dec 13 07:13:12 crc kubenswrapper[4971]: I1213 07:13:12.073741 4971 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/e2b4e2e7-cfa5-41d9-833e-b5afc95c91fe-logs\") pod \"e2b4e2e7-cfa5-41d9-833e-b5afc95c91fe\" (UID: \"e2b4e2e7-cfa5-41d9-833e-b5afc95c91fe\") " Dec 13 07:13:12 crc kubenswrapper[4971]: I1213 07:13:12.073832 4971 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/e2b4e2e7-cfa5-41d9-833e-b5afc95c91fe-combined-ca-bundle\") pod \"e2b4e2e7-cfa5-41d9-833e-b5afc95c91fe\" (UID: \"e2b4e2e7-cfa5-41d9-833e-b5afc95c91fe\") " Dec 13 07:13:12 crc kubenswrapper[4971]: I1213 07:13:12.074860 4971 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/e2b4e2e7-cfa5-41d9-833e-b5afc95c91fe-logs" (OuterVolumeSpecName: "logs") pod "e2b4e2e7-cfa5-41d9-833e-b5afc95c91fe" (UID: "e2b4e2e7-cfa5-41d9-833e-b5afc95c91fe"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 13 07:13:12 crc kubenswrapper[4971]: I1213 07:13:12.074976 4971 reconciler_common.go:293] "Volume detached for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/56e56ce1-4dd1-4ae5-b868-e98d27d5ad94-run-httpd\") on node \"crc\" DevicePath \"\"" Dec 13 07:13:12 crc kubenswrapper[4971]: I1213 07:13:12.075024 4971 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/56e56ce1-4dd1-4ae5-b868-e98d27d5ad94-scripts\") on node \"crc\" DevicePath \"\"" Dec 13 07:13:12 crc kubenswrapper[4971]: I1213 07:13:12.075040 4971 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-hw5mk\" (UniqueName: \"kubernetes.io/projected/56e56ce1-4dd1-4ae5-b868-e98d27d5ad94-kube-api-access-hw5mk\") on node \"crc\" DevicePath \"\"" Dec 13 07:13:12 crc kubenswrapper[4971]: I1213 07:13:12.075054 4971 reconciler_common.go:293] "Volume detached for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/e2b4e2e7-cfa5-41d9-833e-b5afc95c91fe-etc-machine-id\") on node \"crc\" DevicePath \"\"" Dec 13 07:13:12 crc kubenswrapper[4971]: I1213 07:13:12.081894 4971 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/neutron-6597fdddd4-q5zr5"] Dec 13 07:13:12 crc kubenswrapper[4971]: I1213 07:13:12.082301 4971 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/neutron-6597fdddd4-q5zr5" podUID="a844da97-38e3-4bc5-981f-af3fa9f76483" containerName="neutron-api" containerID="cri-o://ff17f9b1aa76599719c611ab63925edefc199a526338bcd7c68bdf8db15d11b8" gracePeriod=30 Dec 13 07:13:12 crc kubenswrapper[4971]: I1213 07:13:12.082475 4971 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/neutron-6597fdddd4-q5zr5" podUID="a844da97-38e3-4bc5-981f-af3fa9f76483" containerName="neutron-httpd" containerID="cri-o://bc6784a5d2347f8bea8ff074080f8d5331cf74f51b2062a8e0fa3b8e1dc2fdaf" gracePeriod=30 Dec 13 07:13:12 crc kubenswrapper[4971]: I1213 07:13:12.121828 4971 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/e2b4e2e7-cfa5-41d9-833e-b5afc95c91fe-kube-api-access-9bwlc" (OuterVolumeSpecName: "kube-api-access-9bwlc") pod "e2b4e2e7-cfa5-41d9-833e-b5afc95c91fe" (UID: "e2b4e2e7-cfa5-41d9-833e-b5afc95c91fe"). InnerVolumeSpecName "kube-api-access-9bwlc". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 13 07:13:12 crc kubenswrapper[4971]: I1213 07:13:12.129697 4971 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/e2b4e2e7-cfa5-41d9-833e-b5afc95c91fe-config-data-custom" (OuterVolumeSpecName: "config-data-custom") pod "e2b4e2e7-cfa5-41d9-833e-b5afc95c91fe" (UID: "e2b4e2e7-cfa5-41d9-833e-b5afc95c91fe"). InnerVolumeSpecName "config-data-custom". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 13 07:13:12 crc kubenswrapper[4971]: I1213 07:13:12.130838 4971 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/e2b4e2e7-cfa5-41d9-833e-b5afc95c91fe-scripts" (OuterVolumeSpecName: "scripts") pod "e2b4e2e7-cfa5-41d9-833e-b5afc95c91fe" (UID: "e2b4e2e7-cfa5-41d9-833e-b5afc95c91fe"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 13 07:13:12 crc kubenswrapper[4971]: I1213 07:13:12.176234 4971 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-9bwlc\" (UniqueName: \"kubernetes.io/projected/e2b4e2e7-cfa5-41d9-833e-b5afc95c91fe-kube-api-access-9bwlc\") on node \"crc\" DevicePath \"\"" Dec 13 07:13:12 crc kubenswrapper[4971]: I1213 07:13:12.176266 4971 reconciler_common.go:293] "Volume detached for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/e2b4e2e7-cfa5-41d9-833e-b5afc95c91fe-config-data-custom\") on node \"crc\" DevicePath \"\"" Dec 13 07:13:12 crc kubenswrapper[4971]: I1213 07:13:12.176275 4971 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/e2b4e2e7-cfa5-41d9-833e-b5afc95c91fe-scripts\") on node \"crc\" DevicePath \"\"" Dec 13 07:13:12 crc kubenswrapper[4971]: I1213 07:13:12.176284 4971 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/e2b4e2e7-cfa5-41d9-833e-b5afc95c91fe-logs\") on node \"crc\" DevicePath \"\"" Dec 13 07:13:12 crc kubenswrapper[4971]: I1213 07:13:12.252637 4971 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/barbican-api-55549fd796-4tztx"] Dec 13 07:13:12 crc kubenswrapper[4971]: E1213 07:13:12.253051 4971 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="56e56ce1-4dd1-4ae5-b868-e98d27d5ad94" containerName="proxy-httpd" Dec 13 07:13:12 crc kubenswrapper[4971]: I1213 07:13:12.253071 4971 state_mem.go:107] "Deleted CPUSet assignment" podUID="56e56ce1-4dd1-4ae5-b868-e98d27d5ad94" containerName="proxy-httpd" Dec 13 07:13:12 crc kubenswrapper[4971]: E1213 07:13:12.253087 4971 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="e2b4e2e7-cfa5-41d9-833e-b5afc95c91fe" containerName="cinder-api-log" Dec 13 07:13:12 crc kubenswrapper[4971]: I1213 07:13:12.253097 4971 state_mem.go:107] "Deleted CPUSet assignment" podUID="e2b4e2e7-cfa5-41d9-833e-b5afc95c91fe" containerName="cinder-api-log" Dec 13 07:13:12 crc kubenswrapper[4971]: E1213 07:13:12.253113 4971 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="56e56ce1-4dd1-4ae5-b868-e98d27d5ad94" containerName="ceilometer-central-agent" Dec 13 07:13:12 crc kubenswrapper[4971]: I1213 07:13:12.253121 4971 state_mem.go:107] "Deleted CPUSet assignment" podUID="56e56ce1-4dd1-4ae5-b868-e98d27d5ad94" containerName="ceilometer-central-agent" Dec 13 07:13:12 crc kubenswrapper[4971]: E1213 07:13:12.253136 4971 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="e2b4e2e7-cfa5-41d9-833e-b5afc95c91fe" containerName="cinder-api" Dec 13 07:13:12 crc kubenswrapper[4971]: I1213 07:13:12.253143 4971 state_mem.go:107] "Deleted CPUSet assignment" podUID="e2b4e2e7-cfa5-41d9-833e-b5afc95c91fe" containerName="cinder-api" Dec 13 07:13:12 crc kubenswrapper[4971]: E1213 07:13:12.253180 4971 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="56e56ce1-4dd1-4ae5-b868-e98d27d5ad94" containerName="sg-core" Dec 13 07:13:12 crc kubenswrapper[4971]: I1213 07:13:12.253237 4971 state_mem.go:107] "Deleted CPUSet assignment" podUID="56e56ce1-4dd1-4ae5-b868-e98d27d5ad94" containerName="sg-core" Dec 13 07:13:12 crc kubenswrapper[4971]: E1213 07:13:12.253253 4971 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="56e56ce1-4dd1-4ae5-b868-e98d27d5ad94" containerName="ceilometer-notification-agent" Dec 13 07:13:12 crc kubenswrapper[4971]: I1213 07:13:12.253260 4971 state_mem.go:107] "Deleted CPUSet assignment" podUID="56e56ce1-4dd1-4ae5-b868-e98d27d5ad94" containerName="ceilometer-notification-agent" Dec 13 07:13:12 crc kubenswrapper[4971]: I1213 07:13:12.253448 4971 memory_manager.go:354] "RemoveStaleState removing state" podUID="56e56ce1-4dd1-4ae5-b868-e98d27d5ad94" containerName="proxy-httpd" Dec 13 07:13:12 crc kubenswrapper[4971]: I1213 07:13:12.253464 4971 memory_manager.go:354] "RemoveStaleState removing state" podUID="e2b4e2e7-cfa5-41d9-833e-b5afc95c91fe" containerName="cinder-api" Dec 13 07:13:12 crc kubenswrapper[4971]: I1213 07:13:12.253479 4971 memory_manager.go:354] "RemoveStaleState removing state" podUID="56e56ce1-4dd1-4ae5-b868-e98d27d5ad94" containerName="ceilometer-notification-agent" Dec 13 07:13:12 crc kubenswrapper[4971]: I1213 07:13:12.253489 4971 memory_manager.go:354] "RemoveStaleState removing state" podUID="56e56ce1-4dd1-4ae5-b868-e98d27d5ad94" containerName="sg-core" Dec 13 07:13:12 crc kubenswrapper[4971]: I1213 07:13:12.253504 4971 memory_manager.go:354] "RemoveStaleState removing state" podUID="56e56ce1-4dd1-4ae5-b868-e98d27d5ad94" containerName="ceilometer-central-agent" Dec 13 07:13:12 crc kubenswrapper[4971]: I1213 07:13:12.253512 4971 memory_manager.go:354] "RemoveStaleState removing state" podUID="e2b4e2e7-cfa5-41d9-833e-b5afc95c91fe" containerName="cinder-api-log" Dec 13 07:13:12 crc kubenswrapper[4971]: I1213 07:13:12.267023 4971 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/56e56ce1-4dd1-4ae5-b868-e98d27d5ad94-sg-core-conf-yaml" (OuterVolumeSpecName: "sg-core-conf-yaml") pod "56e56ce1-4dd1-4ae5-b868-e98d27d5ad94" (UID: "56e56ce1-4dd1-4ae5-b868-e98d27d5ad94"). InnerVolumeSpecName "sg-core-conf-yaml". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 13 07:13:12 crc kubenswrapper[4971]: I1213 07:13:12.267839 4971 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-api-55549fd796-4tztx" Dec 13 07:13:12 crc kubenswrapper[4971]: I1213 07:13:12.270353 4971 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-barbican-internal-svc" Dec 13 07:13:12 crc kubenswrapper[4971]: I1213 07:13:12.270556 4971 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-barbican-public-svc" Dec 13 07:13:12 crc kubenswrapper[4971]: I1213 07:13:12.272553 4971 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/barbican-api-55549fd796-4tztx"] Dec 13 07:13:12 crc kubenswrapper[4971]: I1213 07:13:12.278305 4971 reconciler_common.go:293] "Volume detached for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/56e56ce1-4dd1-4ae5-b868-e98d27d5ad94-sg-core-conf-yaml\") on node \"crc\" DevicePath \"\"" Dec 13 07:13:12 crc kubenswrapper[4971]: I1213 07:13:12.353272 4971 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/e2b4e2e7-cfa5-41d9-833e-b5afc95c91fe-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "e2b4e2e7-cfa5-41d9-833e-b5afc95c91fe" (UID: "e2b4e2e7-cfa5-41d9-833e-b5afc95c91fe"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 13 07:13:12 crc kubenswrapper[4971]: I1213 07:13:12.379834 4971 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/d553c6ea-f592-4890-bccd-f970616d23fe-config-data-custom\") pod \"barbican-api-55549fd796-4tztx\" (UID: \"d553c6ea-f592-4890-bccd-f970616d23fe\") " pod="openstack/barbican-api-55549fd796-4tztx" Dec 13 07:13:12 crc kubenswrapper[4971]: I1213 07:13:12.380119 4971 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/d553c6ea-f592-4890-bccd-f970616d23fe-logs\") pod \"barbican-api-55549fd796-4tztx\" (UID: \"d553c6ea-f592-4890-bccd-f970616d23fe\") " pod="openstack/barbican-api-55549fd796-4tztx" Dec 13 07:13:12 crc kubenswrapper[4971]: I1213 07:13:12.380741 4971 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d553c6ea-f592-4890-bccd-f970616d23fe-combined-ca-bundle\") pod \"barbican-api-55549fd796-4tztx\" (UID: \"d553c6ea-f592-4890-bccd-f970616d23fe\") " pod="openstack/barbican-api-55549fd796-4tztx" Dec 13 07:13:12 crc kubenswrapper[4971]: I1213 07:13:12.380940 4971 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/d553c6ea-f592-4890-bccd-f970616d23fe-config-data\") pod \"barbican-api-55549fd796-4tztx\" (UID: \"d553c6ea-f592-4890-bccd-f970616d23fe\") " pod="openstack/barbican-api-55549fd796-4tztx" Dec 13 07:13:12 crc kubenswrapper[4971]: I1213 07:13:12.381045 4971 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/d553c6ea-f592-4890-bccd-f970616d23fe-public-tls-certs\") pod \"barbican-api-55549fd796-4tztx\" (UID: \"d553c6ea-f592-4890-bccd-f970616d23fe\") " pod="openstack/barbican-api-55549fd796-4tztx" Dec 13 07:13:12 crc kubenswrapper[4971]: I1213 07:13:12.381181 4971 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-c6xbr\" (UniqueName: \"kubernetes.io/projected/d553c6ea-f592-4890-bccd-f970616d23fe-kube-api-access-c6xbr\") pod \"barbican-api-55549fd796-4tztx\" (UID: \"d553c6ea-f592-4890-bccd-f970616d23fe\") " pod="openstack/barbican-api-55549fd796-4tztx" Dec 13 07:13:12 crc kubenswrapper[4971]: I1213 07:13:12.381274 4971 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/d553c6ea-f592-4890-bccd-f970616d23fe-internal-tls-certs\") pod \"barbican-api-55549fd796-4tztx\" (UID: \"d553c6ea-f592-4890-bccd-f970616d23fe\") " pod="openstack/barbican-api-55549fd796-4tztx" Dec 13 07:13:12 crc kubenswrapper[4971]: I1213 07:13:12.381393 4971 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/e2b4e2e7-cfa5-41d9-833e-b5afc95c91fe-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 13 07:13:12 crc kubenswrapper[4971]: I1213 07:13:12.441008 4971 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/56e56ce1-4dd1-4ae5-b868-e98d27d5ad94-config-data" (OuterVolumeSpecName: "config-data") pod "56e56ce1-4dd1-4ae5-b868-e98d27d5ad94" (UID: "56e56ce1-4dd1-4ae5-b868-e98d27d5ad94"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 13 07:13:12 crc kubenswrapper[4971]: I1213 07:13:12.441410 4971 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/56e56ce1-4dd1-4ae5-b868-e98d27d5ad94-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "56e56ce1-4dd1-4ae5-b868-e98d27d5ad94" (UID: "56e56ce1-4dd1-4ae5-b868-e98d27d5ad94"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 13 07:13:12 crc kubenswrapper[4971]: I1213 07:13:12.444668 4971 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/e2b4e2e7-cfa5-41d9-833e-b5afc95c91fe-config-data" (OuterVolumeSpecName: "config-data") pod "e2b4e2e7-cfa5-41d9-833e-b5afc95c91fe" (UID: "e2b4e2e7-cfa5-41d9-833e-b5afc95c91fe"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 13 07:13:12 crc kubenswrapper[4971]: I1213 07:13:12.482885 4971 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/d553c6ea-f592-4890-bccd-f970616d23fe-config-data\") pod \"barbican-api-55549fd796-4tztx\" (UID: \"d553c6ea-f592-4890-bccd-f970616d23fe\") " pod="openstack/barbican-api-55549fd796-4tztx" Dec 13 07:13:12 crc kubenswrapper[4971]: I1213 07:13:12.482946 4971 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/d553c6ea-f592-4890-bccd-f970616d23fe-public-tls-certs\") pod \"barbican-api-55549fd796-4tztx\" (UID: \"d553c6ea-f592-4890-bccd-f970616d23fe\") " pod="openstack/barbican-api-55549fd796-4tztx" Dec 13 07:13:12 crc kubenswrapper[4971]: I1213 07:13:12.482989 4971 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-c6xbr\" (UniqueName: \"kubernetes.io/projected/d553c6ea-f592-4890-bccd-f970616d23fe-kube-api-access-c6xbr\") pod \"barbican-api-55549fd796-4tztx\" (UID: \"d553c6ea-f592-4890-bccd-f970616d23fe\") " pod="openstack/barbican-api-55549fd796-4tztx" Dec 13 07:13:12 crc kubenswrapper[4971]: I1213 07:13:12.483034 4971 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/d553c6ea-f592-4890-bccd-f970616d23fe-internal-tls-certs\") pod \"barbican-api-55549fd796-4tztx\" (UID: \"d553c6ea-f592-4890-bccd-f970616d23fe\") " pod="openstack/barbican-api-55549fd796-4tztx" Dec 13 07:13:12 crc kubenswrapper[4971]: I1213 07:13:12.483055 4971 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/d553c6ea-f592-4890-bccd-f970616d23fe-config-data-custom\") pod \"barbican-api-55549fd796-4tztx\" (UID: \"d553c6ea-f592-4890-bccd-f970616d23fe\") " pod="openstack/barbican-api-55549fd796-4tztx" Dec 13 07:13:12 crc kubenswrapper[4971]: I1213 07:13:12.483099 4971 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/d553c6ea-f592-4890-bccd-f970616d23fe-logs\") pod \"barbican-api-55549fd796-4tztx\" (UID: \"d553c6ea-f592-4890-bccd-f970616d23fe\") " pod="openstack/barbican-api-55549fd796-4tztx" Dec 13 07:13:12 crc kubenswrapper[4971]: I1213 07:13:12.483121 4971 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d553c6ea-f592-4890-bccd-f970616d23fe-combined-ca-bundle\") pod \"barbican-api-55549fd796-4tztx\" (UID: \"d553c6ea-f592-4890-bccd-f970616d23fe\") " pod="openstack/barbican-api-55549fd796-4tztx" Dec 13 07:13:12 crc kubenswrapper[4971]: I1213 07:13:12.483172 4971 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/56e56ce1-4dd1-4ae5-b868-e98d27d5ad94-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 13 07:13:12 crc kubenswrapper[4971]: I1213 07:13:12.483183 4971 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/e2b4e2e7-cfa5-41d9-833e-b5afc95c91fe-config-data\") on node \"crc\" DevicePath \"\"" Dec 13 07:13:12 crc kubenswrapper[4971]: I1213 07:13:12.483195 4971 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/56e56ce1-4dd1-4ae5-b868-e98d27d5ad94-config-data\") on node \"crc\" DevicePath \"\"" Dec 13 07:13:12 crc kubenswrapper[4971]: I1213 07:13:12.485488 4971 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/d553c6ea-f592-4890-bccd-f970616d23fe-logs\") pod \"barbican-api-55549fd796-4tztx\" (UID: \"d553c6ea-f592-4890-bccd-f970616d23fe\") " pod="openstack/barbican-api-55549fd796-4tztx" Dec 13 07:13:12 crc kubenswrapper[4971]: I1213 07:13:12.487481 4971 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d553c6ea-f592-4890-bccd-f970616d23fe-combined-ca-bundle\") pod \"barbican-api-55549fd796-4tztx\" (UID: \"d553c6ea-f592-4890-bccd-f970616d23fe\") " pod="openstack/barbican-api-55549fd796-4tztx" Dec 13 07:13:12 crc kubenswrapper[4971]: I1213 07:13:12.487657 4971 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/d553c6ea-f592-4890-bccd-f970616d23fe-internal-tls-certs\") pod \"barbican-api-55549fd796-4tztx\" (UID: \"d553c6ea-f592-4890-bccd-f970616d23fe\") " pod="openstack/barbican-api-55549fd796-4tztx" Dec 13 07:13:12 crc kubenswrapper[4971]: I1213 07:13:12.488472 4971 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/d553c6ea-f592-4890-bccd-f970616d23fe-config-data-custom\") pod \"barbican-api-55549fd796-4tztx\" (UID: \"d553c6ea-f592-4890-bccd-f970616d23fe\") " pod="openstack/barbican-api-55549fd796-4tztx" Dec 13 07:13:12 crc kubenswrapper[4971]: I1213 07:13:12.488712 4971 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/d553c6ea-f592-4890-bccd-f970616d23fe-public-tls-certs\") pod \"barbican-api-55549fd796-4tztx\" (UID: \"d553c6ea-f592-4890-bccd-f970616d23fe\") " pod="openstack/barbican-api-55549fd796-4tztx" Dec 13 07:13:12 crc kubenswrapper[4971]: I1213 07:13:12.488962 4971 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/d553c6ea-f592-4890-bccd-f970616d23fe-config-data\") pod \"barbican-api-55549fd796-4tztx\" (UID: \"d553c6ea-f592-4890-bccd-f970616d23fe\") " pod="openstack/barbican-api-55549fd796-4tztx" Dec 13 07:13:12 crc kubenswrapper[4971]: I1213 07:13:12.503331 4971 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-c6xbr\" (UniqueName: \"kubernetes.io/projected/d553c6ea-f592-4890-bccd-f970616d23fe-kube-api-access-c6xbr\") pod \"barbican-api-55549fd796-4tztx\" (UID: \"d553c6ea-f592-4890-bccd-f970616d23fe\") " pod="openstack/barbican-api-55549fd796-4tztx" Dec 13 07:13:12 crc kubenswrapper[4971]: I1213 07:13:12.633343 4971 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-api-55549fd796-4tztx" Dec 13 07:13:12 crc kubenswrapper[4971]: I1213 07:13:12.860803 4971 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-api-0" event={"ID":"e2b4e2e7-cfa5-41d9-833e-b5afc95c91fe","Type":"ContainerDied","Data":"f3b9f07fbc7f8d5f03029bb89ed518d701ae9188dd45027ac37a6c8b1e9c35e4"} Dec 13 07:13:12 crc kubenswrapper[4971]: I1213 07:13:12.861077 4971 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-api-0" Dec 13 07:13:12 crc kubenswrapper[4971]: I1213 07:13:12.861115 4971 scope.go:117] "RemoveContainer" containerID="cffdf5b9bc9d0261388ccb6fc29ca51169e8bc553aec7ae953ebc5ee2bd00652" Dec 13 07:13:12 crc kubenswrapper[4971]: I1213 07:13:12.874390 4971 generic.go:334] "Generic (PLEG): container finished" podID="a844da97-38e3-4bc5-981f-af3fa9f76483" containerID="bc6784a5d2347f8bea8ff074080f8d5331cf74f51b2062a8e0fa3b8e1dc2fdaf" exitCode=0 Dec 13 07:13:12 crc kubenswrapper[4971]: I1213 07:13:12.874480 4971 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-6597fdddd4-q5zr5" event={"ID":"a844da97-38e3-4bc5-981f-af3fa9f76483","Type":"ContainerDied","Data":"bc6784a5d2347f8bea8ff074080f8d5331cf74f51b2062a8e0fa3b8e1dc2fdaf"} Dec 13 07:13:12 crc kubenswrapper[4971]: I1213 07:13:12.878098 4971 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"56e56ce1-4dd1-4ae5-b868-e98d27d5ad94","Type":"ContainerDied","Data":"8e88d8b440b648a680026f56322ec6eb8ebd3ef0483daa3a94e143feacc7200b"} Dec 13 07:13:12 crc kubenswrapper[4971]: I1213 07:13:12.878123 4971 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Dec 13 07:13:12 crc kubenswrapper[4971]: I1213 07:13:12.899188 4971 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-worker-c7c6b44fc-8hd9n" event={"ID":"fb068297-28dd-4a70-acbf-650b9edf380e","Type":"ContainerStarted","Data":"59ab25ded4584b498640cd1a96179289fd3688a02db45e4bf4b9da6556965cde"} Dec 13 07:13:12 crc kubenswrapper[4971]: I1213 07:13:12.960578 4971 scope.go:117] "RemoveContainer" containerID="1627e8372a698405e5dd80f505feda5b356d74b9735ba8519bf519633e0519f1" Dec 13 07:13:12 crc kubenswrapper[4971]: I1213 07:13:12.977750 4971 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ceilometer-0"] Dec 13 07:13:13 crc kubenswrapper[4971]: I1213 07:13:13.012058 4971 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/ceilometer-0"] Dec 13 07:13:13 crc kubenswrapper[4971]: I1213 07:13:13.022257 4971 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ceilometer-0"] Dec 13 07:13:13 crc kubenswrapper[4971]: I1213 07:13:13.025076 4971 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Dec 13 07:13:13 crc kubenswrapper[4971]: I1213 07:13:13.031147 4971 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceilometer-config-data" Dec 13 07:13:13 crc kubenswrapper[4971]: I1213 07:13:13.031386 4971 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceilometer-scripts" Dec 13 07:13:13 crc kubenswrapper[4971]: I1213 07:13:13.033607 4971 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ceilometer-0"] Dec 13 07:13:13 crc kubenswrapper[4971]: I1213 07:13:13.048633 4971 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/cinder-api-0"] Dec 13 07:13:13 crc kubenswrapper[4971]: I1213 07:13:13.054275 4971 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/cinder-api-0"] Dec 13 07:13:13 crc kubenswrapper[4971]: I1213 07:13:13.062581 4971 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/cinder-api-0"] Dec 13 07:13:13 crc kubenswrapper[4971]: I1213 07:13:13.070854 4971 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cinder-api-0"] Dec 13 07:13:13 crc kubenswrapper[4971]: I1213 07:13:13.070977 4971 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-api-0" Dec 13 07:13:13 crc kubenswrapper[4971]: I1213 07:13:13.076067 4971 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-cinder-public-svc" Dec 13 07:13:13 crc kubenswrapper[4971]: I1213 07:13:13.076313 4971 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-cinder-internal-svc" Dec 13 07:13:13 crc kubenswrapper[4971]: I1213 07:13:13.077131 4971 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cinder-api-config-data" Dec 13 07:13:13 crc kubenswrapper[4971]: I1213 07:13:13.101847 4971 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/a8767723-1c4f-41c3-ac4e-988a1d93d56a-log-httpd\") pod \"ceilometer-0\" (UID: \"a8767723-1c4f-41c3-ac4e-988a1d93d56a\") " pod="openstack/ceilometer-0" Dec 13 07:13:13 crc kubenswrapper[4971]: I1213 07:13:13.101904 4971 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/a8767723-1c4f-41c3-ac4e-988a1d93d56a-scripts\") pod \"ceilometer-0\" (UID: \"a8767723-1c4f-41c3-ac4e-988a1d93d56a\") " pod="openstack/ceilometer-0" Dec 13 07:13:13 crc kubenswrapper[4971]: I1213 07:13:13.101950 4971 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/a8767723-1c4f-41c3-ac4e-988a1d93d56a-run-httpd\") pod \"ceilometer-0\" (UID: \"a8767723-1c4f-41c3-ac4e-988a1d93d56a\") " pod="openstack/ceilometer-0" Dec 13 07:13:13 crc kubenswrapper[4971]: I1213 07:13:13.102000 4971 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/a8767723-1c4f-41c3-ac4e-988a1d93d56a-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"a8767723-1c4f-41c3-ac4e-988a1d93d56a\") " pod="openstack/ceilometer-0" Dec 13 07:13:13 crc kubenswrapper[4971]: I1213 07:13:13.102051 4971 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/a8767723-1c4f-41c3-ac4e-988a1d93d56a-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"a8767723-1c4f-41c3-ac4e-988a1d93d56a\") " pod="openstack/ceilometer-0" Dec 13 07:13:13 crc kubenswrapper[4971]: I1213 07:13:13.102073 4971 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/a8767723-1c4f-41c3-ac4e-988a1d93d56a-config-data\") pod \"ceilometer-0\" (UID: \"a8767723-1c4f-41c3-ac4e-988a1d93d56a\") " pod="openstack/ceilometer-0" Dec 13 07:13:13 crc kubenswrapper[4971]: I1213 07:13:13.102103 4971 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-6hh4c\" (UniqueName: \"kubernetes.io/projected/a8767723-1c4f-41c3-ac4e-988a1d93d56a-kube-api-access-6hh4c\") pod \"ceilometer-0\" (UID: \"a8767723-1c4f-41c3-ac4e-988a1d93d56a\") " pod="openstack/ceilometer-0" Dec 13 07:13:13 crc kubenswrapper[4971]: I1213 07:13:13.157046 4971 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/barbican-api-55549fd796-4tztx"] Dec 13 07:13:13 crc kubenswrapper[4971]: I1213 07:13:13.206400 4971 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-6hh4c\" (UniqueName: \"kubernetes.io/projected/a8767723-1c4f-41c3-ac4e-988a1d93d56a-kube-api-access-6hh4c\") pod \"ceilometer-0\" (UID: \"a8767723-1c4f-41c3-ac4e-988a1d93d56a\") " pod="openstack/ceilometer-0" Dec 13 07:13:13 crc kubenswrapper[4971]: I1213 07:13:13.206460 4971 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/7cbba320-ebcc-40d7-88ea-ba668d549245-internal-tls-certs\") pod \"cinder-api-0\" (UID: \"7cbba320-ebcc-40d7-88ea-ba668d549245\") " pod="openstack/cinder-api-0" Dec 13 07:13:13 crc kubenswrapper[4971]: I1213 07:13:13.206491 4971 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/7cbba320-ebcc-40d7-88ea-ba668d549245-public-tls-certs\") pod \"cinder-api-0\" (UID: \"7cbba320-ebcc-40d7-88ea-ba668d549245\") " pod="openstack/cinder-api-0" Dec 13 07:13:13 crc kubenswrapper[4971]: I1213 07:13:13.206532 4971 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/a8767723-1c4f-41c3-ac4e-988a1d93d56a-log-httpd\") pod \"ceilometer-0\" (UID: \"a8767723-1c4f-41c3-ac4e-988a1d93d56a\") " pod="openstack/ceilometer-0" Dec 13 07:13:13 crc kubenswrapper[4971]: I1213 07:13:13.206557 4971 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/7cbba320-ebcc-40d7-88ea-ba668d549245-config-data-custom\") pod \"cinder-api-0\" (UID: \"7cbba320-ebcc-40d7-88ea-ba668d549245\") " pod="openstack/cinder-api-0" Dec 13 07:13:13 crc kubenswrapper[4971]: I1213 07:13:13.206582 4971 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/a8767723-1c4f-41c3-ac4e-988a1d93d56a-scripts\") pod \"ceilometer-0\" (UID: \"a8767723-1c4f-41c3-ac4e-988a1d93d56a\") " pod="openstack/ceilometer-0" Dec 13 07:13:13 crc kubenswrapper[4971]: I1213 07:13:13.206596 4971 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/7cbba320-ebcc-40d7-88ea-ba668d549245-combined-ca-bundle\") pod \"cinder-api-0\" (UID: \"7cbba320-ebcc-40d7-88ea-ba668d549245\") " pod="openstack/cinder-api-0" Dec 13 07:13:13 crc kubenswrapper[4971]: I1213 07:13:13.206614 4971 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/7cbba320-ebcc-40d7-88ea-ba668d549245-logs\") pod \"cinder-api-0\" (UID: \"7cbba320-ebcc-40d7-88ea-ba668d549245\") " pod="openstack/cinder-api-0" Dec 13 07:13:13 crc kubenswrapper[4971]: I1213 07:13:13.206635 4971 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-wxxrz\" (UniqueName: \"kubernetes.io/projected/7cbba320-ebcc-40d7-88ea-ba668d549245-kube-api-access-wxxrz\") pod \"cinder-api-0\" (UID: \"7cbba320-ebcc-40d7-88ea-ba668d549245\") " pod="openstack/cinder-api-0" Dec 13 07:13:13 crc kubenswrapper[4971]: I1213 07:13:13.206653 4971 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/a8767723-1c4f-41c3-ac4e-988a1d93d56a-run-httpd\") pod \"ceilometer-0\" (UID: \"a8767723-1c4f-41c3-ac4e-988a1d93d56a\") " pod="openstack/ceilometer-0" Dec 13 07:13:13 crc kubenswrapper[4971]: I1213 07:13:13.206677 4971 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/7cbba320-ebcc-40d7-88ea-ba668d549245-etc-machine-id\") pod \"cinder-api-0\" (UID: \"7cbba320-ebcc-40d7-88ea-ba668d549245\") " pod="openstack/cinder-api-0" Dec 13 07:13:13 crc kubenswrapper[4971]: I1213 07:13:13.206699 4971 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/7cbba320-ebcc-40d7-88ea-ba668d549245-config-data\") pod \"cinder-api-0\" (UID: \"7cbba320-ebcc-40d7-88ea-ba668d549245\") " pod="openstack/cinder-api-0" Dec 13 07:13:13 crc kubenswrapper[4971]: I1213 07:13:13.206720 4971 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/7cbba320-ebcc-40d7-88ea-ba668d549245-scripts\") pod \"cinder-api-0\" (UID: \"7cbba320-ebcc-40d7-88ea-ba668d549245\") " pod="openstack/cinder-api-0" Dec 13 07:13:13 crc kubenswrapper[4971]: I1213 07:13:13.206737 4971 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/a8767723-1c4f-41c3-ac4e-988a1d93d56a-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"a8767723-1c4f-41c3-ac4e-988a1d93d56a\") " pod="openstack/ceilometer-0" Dec 13 07:13:13 crc kubenswrapper[4971]: I1213 07:13:13.206769 4971 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/a8767723-1c4f-41c3-ac4e-988a1d93d56a-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"a8767723-1c4f-41c3-ac4e-988a1d93d56a\") " pod="openstack/ceilometer-0" Dec 13 07:13:13 crc kubenswrapper[4971]: I1213 07:13:13.206787 4971 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/a8767723-1c4f-41c3-ac4e-988a1d93d56a-config-data\") pod \"ceilometer-0\" (UID: \"a8767723-1c4f-41c3-ac4e-988a1d93d56a\") " pod="openstack/ceilometer-0" Dec 13 07:13:13 crc kubenswrapper[4971]: I1213 07:13:13.209243 4971 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/a8767723-1c4f-41c3-ac4e-988a1d93d56a-log-httpd\") pod \"ceilometer-0\" (UID: \"a8767723-1c4f-41c3-ac4e-988a1d93d56a\") " pod="openstack/ceilometer-0" Dec 13 07:13:13 crc kubenswrapper[4971]: I1213 07:13:13.209898 4971 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/a8767723-1c4f-41c3-ac4e-988a1d93d56a-run-httpd\") pod \"ceilometer-0\" (UID: \"a8767723-1c4f-41c3-ac4e-988a1d93d56a\") " pod="openstack/ceilometer-0" Dec 13 07:13:13 crc kubenswrapper[4971]: I1213 07:13:13.226399 4971 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/a8767723-1c4f-41c3-ac4e-988a1d93d56a-scripts\") pod \"ceilometer-0\" (UID: \"a8767723-1c4f-41c3-ac4e-988a1d93d56a\") " pod="openstack/ceilometer-0" Dec 13 07:13:13 crc kubenswrapper[4971]: I1213 07:13:13.232177 4971 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/a8767723-1c4f-41c3-ac4e-988a1d93d56a-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"a8767723-1c4f-41c3-ac4e-988a1d93d56a\") " pod="openstack/ceilometer-0" Dec 13 07:13:13 crc kubenswrapper[4971]: I1213 07:13:13.234579 4971 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/a8767723-1c4f-41c3-ac4e-988a1d93d56a-config-data\") pod \"ceilometer-0\" (UID: \"a8767723-1c4f-41c3-ac4e-988a1d93d56a\") " pod="openstack/ceilometer-0" Dec 13 07:13:13 crc kubenswrapper[4971]: I1213 07:13:13.236857 4971 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/a8767723-1c4f-41c3-ac4e-988a1d93d56a-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"a8767723-1c4f-41c3-ac4e-988a1d93d56a\") " pod="openstack/ceilometer-0" Dec 13 07:13:13 crc kubenswrapper[4971]: I1213 07:13:13.254599 4971 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-6hh4c\" (UniqueName: \"kubernetes.io/projected/a8767723-1c4f-41c3-ac4e-988a1d93d56a-kube-api-access-6hh4c\") pod \"ceilometer-0\" (UID: \"a8767723-1c4f-41c3-ac4e-988a1d93d56a\") " pod="openstack/ceilometer-0" Dec 13 07:13:13 crc kubenswrapper[4971]: I1213 07:13:13.308276 4971 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/7cbba320-ebcc-40d7-88ea-ba668d549245-config-data-custom\") pod \"cinder-api-0\" (UID: \"7cbba320-ebcc-40d7-88ea-ba668d549245\") " pod="openstack/cinder-api-0" Dec 13 07:13:13 crc kubenswrapper[4971]: I1213 07:13:13.308346 4971 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/7cbba320-ebcc-40d7-88ea-ba668d549245-combined-ca-bundle\") pod \"cinder-api-0\" (UID: \"7cbba320-ebcc-40d7-88ea-ba668d549245\") " pod="openstack/cinder-api-0" Dec 13 07:13:13 crc kubenswrapper[4971]: I1213 07:13:13.308377 4971 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/7cbba320-ebcc-40d7-88ea-ba668d549245-logs\") pod \"cinder-api-0\" (UID: \"7cbba320-ebcc-40d7-88ea-ba668d549245\") " pod="openstack/cinder-api-0" Dec 13 07:13:13 crc kubenswrapper[4971]: I1213 07:13:13.308412 4971 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-wxxrz\" (UniqueName: \"kubernetes.io/projected/7cbba320-ebcc-40d7-88ea-ba668d549245-kube-api-access-wxxrz\") pod \"cinder-api-0\" (UID: \"7cbba320-ebcc-40d7-88ea-ba668d549245\") " pod="openstack/cinder-api-0" Dec 13 07:13:13 crc kubenswrapper[4971]: I1213 07:13:13.308481 4971 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/7cbba320-ebcc-40d7-88ea-ba668d549245-etc-machine-id\") pod \"cinder-api-0\" (UID: \"7cbba320-ebcc-40d7-88ea-ba668d549245\") " pod="openstack/cinder-api-0" Dec 13 07:13:13 crc kubenswrapper[4971]: I1213 07:13:13.308510 4971 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/7cbba320-ebcc-40d7-88ea-ba668d549245-config-data\") pod \"cinder-api-0\" (UID: \"7cbba320-ebcc-40d7-88ea-ba668d549245\") " pod="openstack/cinder-api-0" Dec 13 07:13:13 crc kubenswrapper[4971]: I1213 07:13:13.308560 4971 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/7cbba320-ebcc-40d7-88ea-ba668d549245-scripts\") pod \"cinder-api-0\" (UID: \"7cbba320-ebcc-40d7-88ea-ba668d549245\") " pod="openstack/cinder-api-0" Dec 13 07:13:13 crc kubenswrapper[4971]: I1213 07:13:13.308644 4971 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/7cbba320-ebcc-40d7-88ea-ba668d549245-internal-tls-certs\") pod \"cinder-api-0\" (UID: \"7cbba320-ebcc-40d7-88ea-ba668d549245\") " pod="openstack/cinder-api-0" Dec 13 07:13:13 crc kubenswrapper[4971]: I1213 07:13:13.308681 4971 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/7cbba320-ebcc-40d7-88ea-ba668d549245-public-tls-certs\") pod \"cinder-api-0\" (UID: \"7cbba320-ebcc-40d7-88ea-ba668d549245\") " pod="openstack/cinder-api-0" Dec 13 07:13:13 crc kubenswrapper[4971]: I1213 07:13:13.310387 4971 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/7cbba320-ebcc-40d7-88ea-ba668d549245-logs\") pod \"cinder-api-0\" (UID: \"7cbba320-ebcc-40d7-88ea-ba668d549245\") " pod="openstack/cinder-api-0" Dec 13 07:13:13 crc kubenswrapper[4971]: I1213 07:13:13.311360 4971 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/7cbba320-ebcc-40d7-88ea-ba668d549245-etc-machine-id\") pod \"cinder-api-0\" (UID: \"7cbba320-ebcc-40d7-88ea-ba668d549245\") " pod="openstack/cinder-api-0" Dec 13 07:13:13 crc kubenswrapper[4971]: I1213 07:13:13.333075 4971 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/7cbba320-ebcc-40d7-88ea-ba668d549245-scripts\") pod \"cinder-api-0\" (UID: \"7cbba320-ebcc-40d7-88ea-ba668d549245\") " pod="openstack/cinder-api-0" Dec 13 07:13:13 crc kubenswrapper[4971]: I1213 07:13:13.333589 4971 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/7cbba320-ebcc-40d7-88ea-ba668d549245-public-tls-certs\") pod \"cinder-api-0\" (UID: \"7cbba320-ebcc-40d7-88ea-ba668d549245\") " pod="openstack/cinder-api-0" Dec 13 07:13:13 crc kubenswrapper[4971]: I1213 07:13:13.335192 4971 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/7cbba320-ebcc-40d7-88ea-ba668d549245-combined-ca-bundle\") pod \"cinder-api-0\" (UID: \"7cbba320-ebcc-40d7-88ea-ba668d549245\") " pod="openstack/cinder-api-0" Dec 13 07:13:13 crc kubenswrapper[4971]: I1213 07:13:13.336444 4971 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-wxxrz\" (UniqueName: \"kubernetes.io/projected/7cbba320-ebcc-40d7-88ea-ba668d549245-kube-api-access-wxxrz\") pod \"cinder-api-0\" (UID: \"7cbba320-ebcc-40d7-88ea-ba668d549245\") " pod="openstack/cinder-api-0" Dec 13 07:13:13 crc kubenswrapper[4971]: I1213 07:13:13.336982 4971 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/7cbba320-ebcc-40d7-88ea-ba668d549245-config-data-custom\") pod \"cinder-api-0\" (UID: \"7cbba320-ebcc-40d7-88ea-ba668d549245\") " pod="openstack/cinder-api-0" Dec 13 07:13:13 crc kubenswrapper[4971]: I1213 07:13:13.359114 4971 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/7cbba320-ebcc-40d7-88ea-ba668d549245-internal-tls-certs\") pod \"cinder-api-0\" (UID: \"7cbba320-ebcc-40d7-88ea-ba668d549245\") " pod="openstack/cinder-api-0" Dec 13 07:13:13 crc kubenswrapper[4971]: I1213 07:13:13.360334 4971 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/7cbba320-ebcc-40d7-88ea-ba668d549245-config-data\") pod \"cinder-api-0\" (UID: \"7cbba320-ebcc-40d7-88ea-ba668d549245\") " pod="openstack/cinder-api-0" Dec 13 07:13:13 crc kubenswrapper[4971]: I1213 07:13:13.388960 4971 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Dec 13 07:13:13 crc kubenswrapper[4971]: I1213 07:13:13.404179 4971 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-api-0" Dec 13 07:13:13 crc kubenswrapper[4971]: I1213 07:13:13.798434 4971 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="56e56ce1-4dd1-4ae5-b868-e98d27d5ad94" path="/var/lib/kubelet/pods/56e56ce1-4dd1-4ae5-b868-e98d27d5ad94/volumes" Dec 13 07:13:13 crc kubenswrapper[4971]: I1213 07:13:13.800742 4971 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="e2b4e2e7-cfa5-41d9-833e-b5afc95c91fe" path="/var/lib/kubelet/pods/e2b4e2e7-cfa5-41d9-833e-b5afc95c91fe/volumes" Dec 13 07:13:13 crc kubenswrapper[4971]: I1213 07:13:13.942215 4971 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-api-55549fd796-4tztx" event={"ID":"d553c6ea-f592-4890-bccd-f970616d23fe","Type":"ContainerStarted","Data":"1b5a745d97953388271d123979d658fba54a7237dcd06688ae47ee01e3d53a12"} Dec 13 07:13:13 crc kubenswrapper[4971]: I1213 07:13:13.942251 4971 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-api-55549fd796-4tztx" event={"ID":"d553c6ea-f592-4890-bccd-f970616d23fe","Type":"ContainerStarted","Data":"73429ab3a762ccac4ab69d15ecf52272f248d70a8a02fa27043cec8c9af2e8af"} Dec 13 07:13:13 crc kubenswrapper[4971]: I1213 07:13:13.942261 4971 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-api-55549fd796-4tztx" event={"ID":"d553c6ea-f592-4890-bccd-f970616d23fe","Type":"ContainerStarted","Data":"f907966f2f2e2c5a6c24fb9310f4db3eb8b2ef1767e40224c72cde6742456b97"} Dec 13 07:13:13 crc kubenswrapper[4971]: I1213 07:13:13.943235 4971 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/barbican-api-55549fd796-4tztx" Dec 13 07:13:13 crc kubenswrapper[4971]: I1213 07:13:13.943257 4971 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/barbican-api-55549fd796-4tztx" Dec 13 07:13:13 crc kubenswrapper[4971]: I1213 07:13:13.950319 4971 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-keystone-listener-6959d6f94b-tq6hh" event={"ID":"12a4a604-178e-4ad8-a0f7-e2e3d47590bd","Type":"ContainerStarted","Data":"8959deeda7f517366eef12df47c1e2cb115d5a69769dcd5677e63f063cd68093"} Dec 13 07:13:13 crc kubenswrapper[4971]: I1213 07:13:13.956597 4971 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ceilometer-0"] Dec 13 07:13:13 crc kubenswrapper[4971]: I1213 07:13:13.972192 4971 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-scheduler-0" event={"ID":"eb706473-627a-4570-9dfc-0e8fe99d59e1","Type":"ContainerStarted","Data":"543df5cddbd9e5462a1d07304489859d33809a103e27e5250da5a046ea3a4bc7"} Dec 13 07:13:13 crc kubenswrapper[4971]: I1213 07:13:13.976277 4971 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/barbican-api-55549fd796-4tztx" podStartSLOduration=1.976259311 podStartE2EDuration="1.976259311s" podCreationTimestamp="2025-12-13 07:13:12 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-13 07:13:13.971077784 +0000 UTC m=+1450.575487232" watchObservedRunningTime="2025-12-13 07:13:13.976259311 +0000 UTC m=+1450.580668759" Dec 13 07:13:13 crc kubenswrapper[4971]: I1213 07:13:13.990837 4971 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/barbican-keystone-listener-6959d6f94b-tq6hh" podStartSLOduration=4.563724556 podStartE2EDuration="8.99081847s" podCreationTimestamp="2025-12-13 07:13:05 +0000 UTC" firstStartedPulling="2025-12-13 07:13:07.044039621 +0000 UTC m=+1443.648449069" lastFinishedPulling="2025-12-13 07:13:11.471133535 +0000 UTC m=+1448.075542983" observedRunningTime="2025-12-13 07:13:13.987975 +0000 UTC m=+1450.592384448" watchObservedRunningTime="2025-12-13 07:13:13.99081847 +0000 UTC m=+1450.595227918" Dec 13 07:13:14 crc kubenswrapper[4971]: I1213 07:13:14.025774 4971 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/cinder-scheduler-0" podStartSLOduration=6.666553481 podStartE2EDuration="9.02575468s" podCreationTimestamp="2025-12-13 07:13:05 +0000 UTC" firstStartedPulling="2025-12-13 07:13:07.782553268 +0000 UTC m=+1444.386962716" lastFinishedPulling="2025-12-13 07:13:10.141754467 +0000 UTC m=+1446.746163915" observedRunningTime="2025-12-13 07:13:14.013409876 +0000 UTC m=+1450.617819334" watchObservedRunningTime="2025-12-13 07:13:14.02575468 +0000 UTC m=+1450.630164128" Dec 13 07:13:14 crc kubenswrapper[4971]: I1213 07:13:14.035915 4971 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/horizon-8699dc887d-sl2h6" podUID="772eedab-7c62-454b-b722-ee453b39e89f" containerName="horizon" probeResult="failure" output="Get \"https://10.217.0.140:8443/dashboard/auth/login/?next=/dashboard/\": dial tcp 10.217.0.140:8443: connect: connection refused" Dec 13 07:13:14 crc kubenswrapper[4971]: I1213 07:13:14.045264 4971 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/barbican-worker-c7c6b44fc-8hd9n" podStartSLOduration=4.462664618 podStartE2EDuration="9.045235879s" podCreationTimestamp="2025-12-13 07:13:05 +0000 UTC" firstStartedPulling="2025-12-13 07:13:06.880164467 +0000 UTC m=+1443.484573905" lastFinishedPulling="2025-12-13 07:13:11.462735718 +0000 UTC m=+1448.067145166" observedRunningTime="2025-12-13 07:13:14.027474152 +0000 UTC m=+1450.631883610" watchObservedRunningTime="2025-12-13 07:13:14.045235879 +0000 UTC m=+1450.649645327" Dec 13 07:13:14 crc kubenswrapper[4971]: I1213 07:13:14.139860 4971 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cinder-api-0"] Dec 13 07:13:14 crc kubenswrapper[4971]: I1213 07:13:14.819554 4971 provider.go:102] Refreshing cache for provider: *credentialprovider.defaultDockerConfigProvider Dec 13 07:13:15 crc kubenswrapper[4971]: I1213 07:13:15.017092 4971 generic.go:334] "Generic (PLEG): container finished" podID="a844da97-38e3-4bc5-981f-af3fa9f76483" containerID="ff17f9b1aa76599719c611ab63925edefc199a526338bcd7c68bdf8db15d11b8" exitCode=0 Dec 13 07:13:15 crc kubenswrapper[4971]: I1213 07:13:15.017335 4971 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-6597fdddd4-q5zr5" event={"ID":"a844da97-38e3-4bc5-981f-af3fa9f76483","Type":"ContainerDied","Data":"ff17f9b1aa76599719c611ab63925edefc199a526338bcd7c68bdf8db15d11b8"} Dec 13 07:13:15 crc kubenswrapper[4971]: I1213 07:13:15.027228 4971 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"a8767723-1c4f-41c3-ac4e-988a1d93d56a","Type":"ContainerStarted","Data":"9ac009ffef2b0f0f983c49e9d333c68f5bc4ca696b14089108a670ec9c8ae630"} Dec 13 07:13:15 crc kubenswrapper[4971]: I1213 07:13:15.027275 4971 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"a8767723-1c4f-41c3-ac4e-988a1d93d56a","Type":"ContainerStarted","Data":"8c4779bbd343458c9371720b54076b50eb9297ef3e09c1c5758959afd81be4d5"} Dec 13 07:13:15 crc kubenswrapper[4971]: I1213 07:13:15.031966 4971 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-api-0" event={"ID":"7cbba320-ebcc-40d7-88ea-ba668d549245","Type":"ContainerStarted","Data":"93f4c1211f1986aee29817f212a0d3aac7cb492b71fe8c23662a926c47107d7d"} Dec 13 07:13:15 crc kubenswrapper[4971]: I1213 07:13:15.032024 4971 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-api-0" event={"ID":"7cbba320-ebcc-40d7-88ea-ba668d549245","Type":"ContainerStarted","Data":"2c084b4413f680321e6378737ce26c1ccee2d6190b61345cb64da0a4241f25ac"} Dec 13 07:13:15 crc kubenswrapper[4971]: I1213 07:13:15.039942 4971 generic.go:334] "Generic (PLEG): container finished" podID="772eedab-7c62-454b-b722-ee453b39e89f" containerID="4a148882fad66a5e1444a664e27875de5e47f0f5a778fde8428a33e1ba8f489f" exitCode=0 Dec 13 07:13:15 crc kubenswrapper[4971]: I1213 07:13:15.040099 4971 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/horizon-8699dc887d-sl2h6" event={"ID":"772eedab-7c62-454b-b722-ee453b39e89f","Type":"ContainerDied","Data":"4a148882fad66a5e1444a664e27875de5e47f0f5a778fde8428a33e1ba8f489f"} Dec 13 07:13:15 crc kubenswrapper[4971]: I1213 07:13:15.190936 4971 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-6597fdddd4-q5zr5" Dec 13 07:13:15 crc kubenswrapper[4971]: I1213 07:13:15.259215 4971 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-knv8t\" (UniqueName: \"kubernetes.io/projected/a844da97-38e3-4bc5-981f-af3fa9f76483-kube-api-access-knv8t\") pod \"a844da97-38e3-4bc5-981f-af3fa9f76483\" (UID: \"a844da97-38e3-4bc5-981f-af3fa9f76483\") " Dec 13 07:13:15 crc kubenswrapper[4971]: I1213 07:13:15.259421 4971 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovndb-tls-certs\" (UniqueName: \"kubernetes.io/secret/a844da97-38e3-4bc5-981f-af3fa9f76483-ovndb-tls-certs\") pod \"a844da97-38e3-4bc5-981f-af3fa9f76483\" (UID: \"a844da97-38e3-4bc5-981f-af3fa9f76483\") " Dec 13 07:13:15 crc kubenswrapper[4971]: I1213 07:13:15.259503 4971 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"httpd-config\" (UniqueName: \"kubernetes.io/secret/a844da97-38e3-4bc5-981f-af3fa9f76483-httpd-config\") pod \"a844da97-38e3-4bc5-981f-af3fa9f76483\" (UID: \"a844da97-38e3-4bc5-981f-af3fa9f76483\") " Dec 13 07:13:15 crc kubenswrapper[4971]: I1213 07:13:15.259564 4971 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/secret/a844da97-38e3-4bc5-981f-af3fa9f76483-config\") pod \"a844da97-38e3-4bc5-981f-af3fa9f76483\" (UID: \"a844da97-38e3-4bc5-981f-af3fa9f76483\") " Dec 13 07:13:15 crc kubenswrapper[4971]: I1213 07:13:15.259849 4971 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/a844da97-38e3-4bc5-981f-af3fa9f76483-combined-ca-bundle\") pod \"a844da97-38e3-4bc5-981f-af3fa9f76483\" (UID: \"a844da97-38e3-4bc5-981f-af3fa9f76483\") " Dec 13 07:13:15 crc kubenswrapper[4971]: I1213 07:13:15.264436 4971 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/a844da97-38e3-4bc5-981f-af3fa9f76483-httpd-config" (OuterVolumeSpecName: "httpd-config") pod "a844da97-38e3-4bc5-981f-af3fa9f76483" (UID: "a844da97-38e3-4bc5-981f-af3fa9f76483"). InnerVolumeSpecName "httpd-config". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 13 07:13:15 crc kubenswrapper[4971]: I1213 07:13:15.265070 4971 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/a844da97-38e3-4bc5-981f-af3fa9f76483-kube-api-access-knv8t" (OuterVolumeSpecName: "kube-api-access-knv8t") pod "a844da97-38e3-4bc5-981f-af3fa9f76483" (UID: "a844da97-38e3-4bc5-981f-af3fa9f76483"). InnerVolumeSpecName "kube-api-access-knv8t". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 13 07:13:15 crc kubenswrapper[4971]: I1213 07:13:15.320331 4971 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/a844da97-38e3-4bc5-981f-af3fa9f76483-config" (OuterVolumeSpecName: "config") pod "a844da97-38e3-4bc5-981f-af3fa9f76483" (UID: "a844da97-38e3-4bc5-981f-af3fa9f76483"). InnerVolumeSpecName "config". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 13 07:13:15 crc kubenswrapper[4971]: I1213 07:13:15.322295 4971 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/a844da97-38e3-4bc5-981f-af3fa9f76483-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "a844da97-38e3-4bc5-981f-af3fa9f76483" (UID: "a844da97-38e3-4bc5-981f-af3fa9f76483"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 13 07:13:15 crc kubenswrapper[4971]: I1213 07:13:15.341863 4971 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/a844da97-38e3-4bc5-981f-af3fa9f76483-ovndb-tls-certs" (OuterVolumeSpecName: "ovndb-tls-certs") pod "a844da97-38e3-4bc5-981f-af3fa9f76483" (UID: "a844da97-38e3-4bc5-981f-af3fa9f76483"). InnerVolumeSpecName "ovndb-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 13 07:13:15 crc kubenswrapper[4971]: I1213 07:13:15.363101 4971 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/a844da97-38e3-4bc5-981f-af3fa9f76483-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 13 07:13:15 crc kubenswrapper[4971]: I1213 07:13:15.363131 4971 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-knv8t\" (UniqueName: \"kubernetes.io/projected/a844da97-38e3-4bc5-981f-af3fa9f76483-kube-api-access-knv8t\") on node \"crc\" DevicePath \"\"" Dec 13 07:13:15 crc kubenswrapper[4971]: I1213 07:13:15.363141 4971 reconciler_common.go:293] "Volume detached for volume \"ovndb-tls-certs\" (UniqueName: \"kubernetes.io/secret/a844da97-38e3-4bc5-981f-af3fa9f76483-ovndb-tls-certs\") on node \"crc\" DevicePath \"\"" Dec 13 07:13:15 crc kubenswrapper[4971]: I1213 07:13:15.363149 4971 reconciler_common.go:293] "Volume detached for volume \"httpd-config\" (UniqueName: \"kubernetes.io/secret/a844da97-38e3-4bc5-981f-af3fa9f76483-httpd-config\") on node \"crc\" DevicePath \"\"" Dec 13 07:13:15 crc kubenswrapper[4971]: I1213 07:13:15.363160 4971 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/secret/a844da97-38e3-4bc5-981f-af3fa9f76483-config\") on node \"crc\" DevicePath \"\"" Dec 13 07:13:16 crc kubenswrapper[4971]: I1213 07:13:16.051438 4971 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-6597fdddd4-q5zr5" event={"ID":"a844da97-38e3-4bc5-981f-af3fa9f76483","Type":"ContainerDied","Data":"93faa59d0bd7c5a3f465196c197b6c4c5cf492443b1bf4cbadf37610a3d42122"} Dec 13 07:13:16 crc kubenswrapper[4971]: I1213 07:13:16.051943 4971 scope.go:117] "RemoveContainer" containerID="bc6784a5d2347f8bea8ff074080f8d5331cf74f51b2062a8e0fa3b8e1dc2fdaf" Dec 13 07:13:16 crc kubenswrapper[4971]: I1213 07:13:16.051486 4971 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-6597fdddd4-q5zr5" Dec 13 07:13:16 crc kubenswrapper[4971]: I1213 07:13:16.054625 4971 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"a8767723-1c4f-41c3-ac4e-988a1d93d56a","Type":"ContainerStarted","Data":"3ea374af6edd7bef76b4de25745f39b1b6be9ae48de86284b377702d49b7b32a"} Dec 13 07:13:16 crc kubenswrapper[4971]: I1213 07:13:16.062276 4971 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-api-0" event={"ID":"7cbba320-ebcc-40d7-88ea-ba668d549245","Type":"ContainerStarted","Data":"0db3983c96c1da605287474a0081d42658961a2689e2707f5fb130fce03e62e8"} Dec 13 07:13:16 crc kubenswrapper[4971]: I1213 07:13:16.062321 4971 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/cinder-api-0" Dec 13 07:13:16 crc kubenswrapper[4971]: I1213 07:13:16.079886 4971 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/neutron-6597fdddd4-q5zr5"] Dec 13 07:13:16 crc kubenswrapper[4971]: I1213 07:13:16.086614 4971 scope.go:117] "RemoveContainer" containerID="ff17f9b1aa76599719c611ab63925edefc199a526338bcd7c68bdf8db15d11b8" Dec 13 07:13:16 crc kubenswrapper[4971]: I1213 07:13:16.088700 4971 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/neutron-6597fdddd4-q5zr5"] Dec 13 07:13:16 crc kubenswrapper[4971]: I1213 07:13:16.119285 4971 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/cinder-api-0" podStartSLOduration=4.119260611 podStartE2EDuration="4.119260611s" podCreationTimestamp="2025-12-13 07:13:12 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-13 07:13:16.107337958 +0000 UTC m=+1452.711747426" watchObservedRunningTime="2025-12-13 07:13:16.119260611 +0000 UTC m=+1452.723670059" Dec 13 07:13:16 crc kubenswrapper[4971]: I1213 07:13:16.550717 4971 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/dnsmasq-dns-7474d577dc-q7d6q" Dec 13 07:13:16 crc kubenswrapper[4971]: I1213 07:13:16.638076 4971 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-548894858c-j5bhs"] Dec 13 07:13:16 crc kubenswrapper[4971]: I1213 07:13:16.638297 4971 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/dnsmasq-dns-548894858c-j5bhs" podUID="cd7d111d-0b3e-4bd7-ba07-10e296208a7f" containerName="dnsmasq-dns" containerID="cri-o://aebcdf008ab057d1a88376abfdad8b3a2c538ea0af5a431a1b81134323b14d9c" gracePeriod=10 Dec 13 07:13:16 crc kubenswrapper[4971]: I1213 07:13:16.790429 4971 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/cinder-scheduler-0" Dec 13 07:13:17 crc kubenswrapper[4971]: I1213 07:13:17.003976 4971 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/placement-569b6fd56b-w7xmm" Dec 13 07:13:17 crc kubenswrapper[4971]: I1213 07:13:17.079256 4971 generic.go:334] "Generic (PLEG): container finished" podID="cd7d111d-0b3e-4bd7-ba07-10e296208a7f" containerID="aebcdf008ab057d1a88376abfdad8b3a2c538ea0af5a431a1b81134323b14d9c" exitCode=0 Dec 13 07:13:17 crc kubenswrapper[4971]: I1213 07:13:17.079323 4971 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-548894858c-j5bhs" event={"ID":"cd7d111d-0b3e-4bd7-ba07-10e296208a7f","Type":"ContainerDied","Data":"aebcdf008ab057d1a88376abfdad8b3a2c538ea0af5a431a1b81134323b14d9c"} Dec 13 07:13:17 crc kubenswrapper[4971]: I1213 07:13:17.187626 4971 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/cinder-scheduler-0" Dec 13 07:13:17 crc kubenswrapper[4971]: I1213 07:13:17.254853 4971 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/cinder-scheduler-0"] Dec 13 07:13:17 crc kubenswrapper[4971]: I1213 07:13:17.788674 4971 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="a844da97-38e3-4bc5-981f-af3fa9f76483" path="/var/lib/kubelet/pods/a844da97-38e3-4bc5-981f-af3fa9f76483/volumes" Dec 13 07:13:17 crc kubenswrapper[4971]: I1213 07:13:17.854973 4971 prober.go:107] "Probe failed" probeType="Liveness" pod="openstack/barbican-api-5c88bd59bd-2qlbj" podUID="02578e71-2865-4a4b-96e9-5bf6c625183c" containerName="barbican-api" probeResult="failure" output="HTTP probe failed with statuscode: 500" Dec 13 07:13:18 crc kubenswrapper[4971]: I1213 07:13:18.094325 4971 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"a8767723-1c4f-41c3-ac4e-988a1d93d56a","Type":"ContainerStarted","Data":"714d0b6157ede690f30644b9c368321eb015d8734735a3cf3e7f11de2e94d6bb"} Dec 13 07:13:18 crc kubenswrapper[4971]: I1213 07:13:18.096375 4971 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/cinder-scheduler-0" podUID="eb706473-627a-4570-9dfc-0e8fe99d59e1" containerName="cinder-scheduler" containerID="cri-o://c2bb3a07998a3cffd09385e18ab96ef8b7a2067d9333b9e5fe773a03866bf0a5" gracePeriod=30 Dec 13 07:13:18 crc kubenswrapper[4971]: I1213 07:13:18.096673 4971 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-548894858c-j5bhs" event={"ID":"cd7d111d-0b3e-4bd7-ba07-10e296208a7f","Type":"ContainerDied","Data":"786637934b9af6b6e09bd09c84cb2ce4d124a73c928f00884fe07b7200c727e1"} Dec 13 07:13:18 crc kubenswrapper[4971]: I1213 07:13:18.096696 4971 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="786637934b9af6b6e09bd09c84cb2ce4d124a73c928f00884fe07b7200c727e1" Dec 13 07:13:18 crc kubenswrapper[4971]: I1213 07:13:18.097060 4971 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/cinder-scheduler-0" podUID="eb706473-627a-4570-9dfc-0e8fe99d59e1" containerName="probe" containerID="cri-o://543df5cddbd9e5462a1d07304489859d33809a103e27e5250da5a046ea3a4bc7" gracePeriod=30 Dec 13 07:13:18 crc kubenswrapper[4971]: I1213 07:13:18.178408 4971 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-548894858c-j5bhs" Dec 13 07:13:18 crc kubenswrapper[4971]: I1213 07:13:18.240710 4971 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/cd7d111d-0b3e-4bd7-ba07-10e296208a7f-ovsdbserver-sb\") pod \"cd7d111d-0b3e-4bd7-ba07-10e296208a7f\" (UID: \"cd7d111d-0b3e-4bd7-ba07-10e296208a7f\") " Dec 13 07:13:18 crc kubenswrapper[4971]: I1213 07:13:18.240885 4971 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/cd7d111d-0b3e-4bd7-ba07-10e296208a7f-config\") pod \"cd7d111d-0b3e-4bd7-ba07-10e296208a7f\" (UID: \"cd7d111d-0b3e-4bd7-ba07-10e296208a7f\") " Dec 13 07:13:18 crc kubenswrapper[4971]: I1213 07:13:18.240932 4971 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-7hqhn\" (UniqueName: \"kubernetes.io/projected/cd7d111d-0b3e-4bd7-ba07-10e296208a7f-kube-api-access-7hqhn\") pod \"cd7d111d-0b3e-4bd7-ba07-10e296208a7f\" (UID: \"cd7d111d-0b3e-4bd7-ba07-10e296208a7f\") " Dec 13 07:13:18 crc kubenswrapper[4971]: I1213 07:13:18.241010 4971 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/cd7d111d-0b3e-4bd7-ba07-10e296208a7f-ovsdbserver-nb\") pod \"cd7d111d-0b3e-4bd7-ba07-10e296208a7f\" (UID: \"cd7d111d-0b3e-4bd7-ba07-10e296208a7f\") " Dec 13 07:13:18 crc kubenswrapper[4971]: I1213 07:13:18.241088 4971 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/cd7d111d-0b3e-4bd7-ba07-10e296208a7f-dns-svc\") pod \"cd7d111d-0b3e-4bd7-ba07-10e296208a7f\" (UID: \"cd7d111d-0b3e-4bd7-ba07-10e296208a7f\") " Dec 13 07:13:18 crc kubenswrapper[4971]: I1213 07:13:18.259353 4971 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/cd7d111d-0b3e-4bd7-ba07-10e296208a7f-kube-api-access-7hqhn" (OuterVolumeSpecName: "kube-api-access-7hqhn") pod "cd7d111d-0b3e-4bd7-ba07-10e296208a7f" (UID: "cd7d111d-0b3e-4bd7-ba07-10e296208a7f"). InnerVolumeSpecName "kube-api-access-7hqhn". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 13 07:13:18 crc kubenswrapper[4971]: I1213 07:13:18.302000 4971 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/cd7d111d-0b3e-4bd7-ba07-10e296208a7f-ovsdbserver-sb" (OuterVolumeSpecName: "ovsdbserver-sb") pod "cd7d111d-0b3e-4bd7-ba07-10e296208a7f" (UID: "cd7d111d-0b3e-4bd7-ba07-10e296208a7f"). InnerVolumeSpecName "ovsdbserver-sb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 13 07:13:18 crc kubenswrapper[4971]: I1213 07:13:18.305304 4971 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/cd7d111d-0b3e-4bd7-ba07-10e296208a7f-ovsdbserver-nb" (OuterVolumeSpecName: "ovsdbserver-nb") pod "cd7d111d-0b3e-4bd7-ba07-10e296208a7f" (UID: "cd7d111d-0b3e-4bd7-ba07-10e296208a7f"). InnerVolumeSpecName "ovsdbserver-nb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 13 07:13:18 crc kubenswrapper[4971]: I1213 07:13:18.308747 4971 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/cd7d111d-0b3e-4bd7-ba07-10e296208a7f-dns-svc" (OuterVolumeSpecName: "dns-svc") pod "cd7d111d-0b3e-4bd7-ba07-10e296208a7f" (UID: "cd7d111d-0b3e-4bd7-ba07-10e296208a7f"). InnerVolumeSpecName "dns-svc". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 13 07:13:18 crc kubenswrapper[4971]: I1213 07:13:18.311330 4971 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/cd7d111d-0b3e-4bd7-ba07-10e296208a7f-config" (OuterVolumeSpecName: "config") pod "cd7d111d-0b3e-4bd7-ba07-10e296208a7f" (UID: "cd7d111d-0b3e-4bd7-ba07-10e296208a7f"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 13 07:13:18 crc kubenswrapper[4971]: I1213 07:13:18.351566 4971 reconciler_common.go:293] "Volume detached for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/cd7d111d-0b3e-4bd7-ba07-10e296208a7f-dns-svc\") on node \"crc\" DevicePath \"\"" Dec 13 07:13:18 crc kubenswrapper[4971]: I1213 07:13:18.351598 4971 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/cd7d111d-0b3e-4bd7-ba07-10e296208a7f-ovsdbserver-sb\") on node \"crc\" DevicePath \"\"" Dec 13 07:13:18 crc kubenswrapper[4971]: I1213 07:13:18.351619 4971 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/cd7d111d-0b3e-4bd7-ba07-10e296208a7f-config\") on node \"crc\" DevicePath \"\"" Dec 13 07:13:18 crc kubenswrapper[4971]: I1213 07:13:18.351628 4971 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-7hqhn\" (UniqueName: \"kubernetes.io/projected/cd7d111d-0b3e-4bd7-ba07-10e296208a7f-kube-api-access-7hqhn\") on node \"crc\" DevicePath \"\"" Dec 13 07:13:18 crc kubenswrapper[4971]: I1213 07:13:18.351637 4971 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/cd7d111d-0b3e-4bd7-ba07-10e296208a7f-ovsdbserver-nb\") on node \"crc\" DevicePath \"\"" Dec 13 07:13:18 crc kubenswrapper[4971]: I1213 07:13:18.927562 4971 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/barbican-api-5c88bd59bd-2qlbj" Dec 13 07:13:19 crc kubenswrapper[4971]: I1213 07:13:19.087289 4971 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/barbican-api-5c88bd59bd-2qlbj" Dec 13 07:13:19 crc kubenswrapper[4971]: I1213 07:13:19.122394 4971 generic.go:334] "Generic (PLEG): container finished" podID="eb706473-627a-4570-9dfc-0e8fe99d59e1" containerID="543df5cddbd9e5462a1d07304489859d33809a103e27e5250da5a046ea3a4bc7" exitCode=0 Dec 13 07:13:19 crc kubenswrapper[4971]: I1213 07:13:19.122431 4971 generic.go:334] "Generic (PLEG): container finished" podID="eb706473-627a-4570-9dfc-0e8fe99d59e1" containerID="c2bb3a07998a3cffd09385e18ab96ef8b7a2067d9333b9e5fe773a03866bf0a5" exitCode=0 Dec 13 07:13:19 crc kubenswrapper[4971]: I1213 07:13:19.122505 4971 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-scheduler-0" event={"ID":"eb706473-627a-4570-9dfc-0e8fe99d59e1","Type":"ContainerDied","Data":"543df5cddbd9e5462a1d07304489859d33809a103e27e5250da5a046ea3a4bc7"} Dec 13 07:13:19 crc kubenswrapper[4971]: I1213 07:13:19.122554 4971 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-scheduler-0" event={"ID":"eb706473-627a-4570-9dfc-0e8fe99d59e1","Type":"ContainerDied","Data":"c2bb3a07998a3cffd09385e18ab96ef8b7a2067d9333b9e5fe773a03866bf0a5"} Dec 13 07:13:19 crc kubenswrapper[4971]: I1213 07:13:19.139777 4971 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"a8767723-1c4f-41c3-ac4e-988a1d93d56a","Type":"ContainerStarted","Data":"3eb143d88d13c5a55a04a05582effc70650cefc1337e365a514f4fabdcf9888f"} Dec 13 07:13:19 crc kubenswrapper[4971]: I1213 07:13:19.141013 4971 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-548894858c-j5bhs" Dec 13 07:13:19 crc kubenswrapper[4971]: I1213 07:13:19.197654 4971 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/ceilometer-0" podStartSLOduration=2.660740444 podStartE2EDuration="7.19763428s" podCreationTimestamp="2025-12-13 07:13:12 +0000 UTC" firstStartedPulling="2025-12-13 07:13:13.954323961 +0000 UTC m=+1450.558733409" lastFinishedPulling="2025-12-13 07:13:18.491217797 +0000 UTC m=+1455.095627245" observedRunningTime="2025-12-13 07:13:19.183156843 +0000 UTC m=+1455.787566291" watchObservedRunningTime="2025-12-13 07:13:19.19763428 +0000 UTC m=+1455.802043728" Dec 13 07:13:19 crc kubenswrapper[4971]: I1213 07:13:19.220887 4971 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-548894858c-j5bhs"] Dec 13 07:13:19 crc kubenswrapper[4971]: I1213 07:13:19.234371 4971 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-548894858c-j5bhs"] Dec 13 07:13:19 crc kubenswrapper[4971]: I1213 07:13:19.588817 4971 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-scheduler-0" Dec 13 07:13:19 crc kubenswrapper[4971]: I1213 07:13:19.695078 4971 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/eb706473-627a-4570-9dfc-0e8fe99d59e1-etc-machine-id\") pod \"eb706473-627a-4570-9dfc-0e8fe99d59e1\" (UID: \"eb706473-627a-4570-9dfc-0e8fe99d59e1\") " Dec 13 07:13:19 crc kubenswrapper[4971]: I1213 07:13:19.695478 4971 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-s22xv\" (UniqueName: \"kubernetes.io/projected/eb706473-627a-4570-9dfc-0e8fe99d59e1-kube-api-access-s22xv\") pod \"eb706473-627a-4570-9dfc-0e8fe99d59e1\" (UID: \"eb706473-627a-4570-9dfc-0e8fe99d59e1\") " Dec 13 07:13:19 crc kubenswrapper[4971]: I1213 07:13:19.695231 4971 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/eb706473-627a-4570-9dfc-0e8fe99d59e1-etc-machine-id" (OuterVolumeSpecName: "etc-machine-id") pod "eb706473-627a-4570-9dfc-0e8fe99d59e1" (UID: "eb706473-627a-4570-9dfc-0e8fe99d59e1"). InnerVolumeSpecName "etc-machine-id". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 13 07:13:19 crc kubenswrapper[4971]: I1213 07:13:19.695540 4971 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/eb706473-627a-4570-9dfc-0e8fe99d59e1-config-data\") pod \"eb706473-627a-4570-9dfc-0e8fe99d59e1\" (UID: \"eb706473-627a-4570-9dfc-0e8fe99d59e1\") " Dec 13 07:13:19 crc kubenswrapper[4971]: I1213 07:13:19.695588 4971 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/eb706473-627a-4570-9dfc-0e8fe99d59e1-scripts\") pod \"eb706473-627a-4570-9dfc-0e8fe99d59e1\" (UID: \"eb706473-627a-4570-9dfc-0e8fe99d59e1\") " Dec 13 07:13:19 crc kubenswrapper[4971]: I1213 07:13:19.695686 4971 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/eb706473-627a-4570-9dfc-0e8fe99d59e1-config-data-custom\") pod \"eb706473-627a-4570-9dfc-0e8fe99d59e1\" (UID: \"eb706473-627a-4570-9dfc-0e8fe99d59e1\") " Dec 13 07:13:19 crc kubenswrapper[4971]: I1213 07:13:19.695770 4971 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/eb706473-627a-4570-9dfc-0e8fe99d59e1-combined-ca-bundle\") pod \"eb706473-627a-4570-9dfc-0e8fe99d59e1\" (UID: \"eb706473-627a-4570-9dfc-0e8fe99d59e1\") " Dec 13 07:13:19 crc kubenswrapper[4971]: I1213 07:13:19.696094 4971 reconciler_common.go:293] "Volume detached for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/eb706473-627a-4570-9dfc-0e8fe99d59e1-etc-machine-id\") on node \"crc\" DevicePath \"\"" Dec 13 07:13:19 crc kubenswrapper[4971]: I1213 07:13:19.705749 4971 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/eb706473-627a-4570-9dfc-0e8fe99d59e1-scripts" (OuterVolumeSpecName: "scripts") pod "eb706473-627a-4570-9dfc-0e8fe99d59e1" (UID: "eb706473-627a-4570-9dfc-0e8fe99d59e1"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 13 07:13:19 crc kubenswrapper[4971]: I1213 07:13:19.719546 4971 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/eb706473-627a-4570-9dfc-0e8fe99d59e1-config-data-custom" (OuterVolumeSpecName: "config-data-custom") pod "eb706473-627a-4570-9dfc-0e8fe99d59e1" (UID: "eb706473-627a-4570-9dfc-0e8fe99d59e1"). InnerVolumeSpecName "config-data-custom". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 13 07:13:19 crc kubenswrapper[4971]: I1213 07:13:19.736724 4971 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/eb706473-627a-4570-9dfc-0e8fe99d59e1-kube-api-access-s22xv" (OuterVolumeSpecName: "kube-api-access-s22xv") pod "eb706473-627a-4570-9dfc-0e8fe99d59e1" (UID: "eb706473-627a-4570-9dfc-0e8fe99d59e1"). InnerVolumeSpecName "kube-api-access-s22xv". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 13 07:13:19 crc kubenswrapper[4971]: I1213 07:13:19.778754 4971 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/eb706473-627a-4570-9dfc-0e8fe99d59e1-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "eb706473-627a-4570-9dfc-0e8fe99d59e1" (UID: "eb706473-627a-4570-9dfc-0e8fe99d59e1"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 13 07:13:19 crc kubenswrapper[4971]: I1213 07:13:19.791317 4971 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="cd7d111d-0b3e-4bd7-ba07-10e296208a7f" path="/var/lib/kubelet/pods/cd7d111d-0b3e-4bd7-ba07-10e296208a7f/volumes" Dec 13 07:13:19 crc kubenswrapper[4971]: I1213 07:13:19.799647 4971 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/eb706473-627a-4570-9dfc-0e8fe99d59e1-scripts\") on node \"crc\" DevicePath \"\"" Dec 13 07:13:19 crc kubenswrapper[4971]: I1213 07:13:19.799803 4971 reconciler_common.go:293] "Volume detached for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/eb706473-627a-4570-9dfc-0e8fe99d59e1-config-data-custom\") on node \"crc\" DevicePath \"\"" Dec 13 07:13:19 crc kubenswrapper[4971]: I1213 07:13:19.799825 4971 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/eb706473-627a-4570-9dfc-0e8fe99d59e1-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 13 07:13:19 crc kubenswrapper[4971]: I1213 07:13:19.799835 4971 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-s22xv\" (UniqueName: \"kubernetes.io/projected/eb706473-627a-4570-9dfc-0e8fe99d59e1-kube-api-access-s22xv\") on node \"crc\" DevicePath \"\"" Dec 13 07:13:19 crc kubenswrapper[4971]: I1213 07:13:19.844829 4971 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/eb706473-627a-4570-9dfc-0e8fe99d59e1-config-data" (OuterVolumeSpecName: "config-data") pod "eb706473-627a-4570-9dfc-0e8fe99d59e1" (UID: "eb706473-627a-4570-9dfc-0e8fe99d59e1"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 13 07:13:19 crc kubenswrapper[4971]: I1213 07:13:19.901578 4971 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/eb706473-627a-4570-9dfc-0e8fe99d59e1-config-data\") on node \"crc\" DevicePath \"\"" Dec 13 07:13:20 crc kubenswrapper[4971]: I1213 07:13:20.150795 4971 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-scheduler-0" event={"ID":"eb706473-627a-4570-9dfc-0e8fe99d59e1","Type":"ContainerDied","Data":"76b4096e1c2bbcba6f9d6fba5708ba0d182ba2eaad08a0ee5e7e30f8217e05ac"} Dec 13 07:13:20 crc kubenswrapper[4971]: I1213 07:13:20.150854 4971 scope.go:117] "RemoveContainer" containerID="543df5cddbd9e5462a1d07304489859d33809a103e27e5250da5a046ea3a4bc7" Dec 13 07:13:20 crc kubenswrapper[4971]: I1213 07:13:20.150819 4971 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-scheduler-0" Dec 13 07:13:20 crc kubenswrapper[4971]: I1213 07:13:20.151024 4971 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/ceilometer-0" Dec 13 07:13:20 crc kubenswrapper[4971]: I1213 07:13:20.184885 4971 scope.go:117] "RemoveContainer" containerID="c2bb3a07998a3cffd09385e18ab96ef8b7a2067d9333b9e5fe773a03866bf0a5" Dec 13 07:13:20 crc kubenswrapper[4971]: I1213 07:13:20.199152 4971 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/cinder-scheduler-0"] Dec 13 07:13:20 crc kubenswrapper[4971]: I1213 07:13:20.211210 4971 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/cinder-scheduler-0"] Dec 13 07:13:20 crc kubenswrapper[4971]: I1213 07:13:20.223882 4971 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/cinder-scheduler-0"] Dec 13 07:13:20 crc kubenswrapper[4971]: E1213 07:13:20.224261 4971 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="a844da97-38e3-4bc5-981f-af3fa9f76483" containerName="neutron-httpd" Dec 13 07:13:20 crc kubenswrapper[4971]: I1213 07:13:20.224279 4971 state_mem.go:107] "Deleted CPUSet assignment" podUID="a844da97-38e3-4bc5-981f-af3fa9f76483" containerName="neutron-httpd" Dec 13 07:13:20 crc kubenswrapper[4971]: E1213 07:13:20.224293 4971 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="cd7d111d-0b3e-4bd7-ba07-10e296208a7f" containerName="dnsmasq-dns" Dec 13 07:13:20 crc kubenswrapper[4971]: I1213 07:13:20.224299 4971 state_mem.go:107] "Deleted CPUSet assignment" podUID="cd7d111d-0b3e-4bd7-ba07-10e296208a7f" containerName="dnsmasq-dns" Dec 13 07:13:20 crc kubenswrapper[4971]: E1213 07:13:20.224324 4971 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="eb706473-627a-4570-9dfc-0e8fe99d59e1" containerName="cinder-scheduler" Dec 13 07:13:20 crc kubenswrapper[4971]: I1213 07:13:20.224330 4971 state_mem.go:107] "Deleted CPUSet assignment" podUID="eb706473-627a-4570-9dfc-0e8fe99d59e1" containerName="cinder-scheduler" Dec 13 07:13:20 crc kubenswrapper[4971]: E1213 07:13:20.224340 4971 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="cd7d111d-0b3e-4bd7-ba07-10e296208a7f" containerName="init" Dec 13 07:13:20 crc kubenswrapper[4971]: I1213 07:13:20.224347 4971 state_mem.go:107] "Deleted CPUSet assignment" podUID="cd7d111d-0b3e-4bd7-ba07-10e296208a7f" containerName="init" Dec 13 07:13:20 crc kubenswrapper[4971]: E1213 07:13:20.224357 4971 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="eb706473-627a-4570-9dfc-0e8fe99d59e1" containerName="probe" Dec 13 07:13:20 crc kubenswrapper[4971]: I1213 07:13:20.224362 4971 state_mem.go:107] "Deleted CPUSet assignment" podUID="eb706473-627a-4570-9dfc-0e8fe99d59e1" containerName="probe" Dec 13 07:13:20 crc kubenswrapper[4971]: E1213 07:13:20.224378 4971 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="a844da97-38e3-4bc5-981f-af3fa9f76483" containerName="neutron-api" Dec 13 07:13:20 crc kubenswrapper[4971]: I1213 07:13:20.224384 4971 state_mem.go:107] "Deleted CPUSet assignment" podUID="a844da97-38e3-4bc5-981f-af3fa9f76483" containerName="neutron-api" Dec 13 07:13:20 crc kubenswrapper[4971]: I1213 07:13:20.224573 4971 memory_manager.go:354] "RemoveStaleState removing state" podUID="a844da97-38e3-4bc5-981f-af3fa9f76483" containerName="neutron-httpd" Dec 13 07:13:20 crc kubenswrapper[4971]: I1213 07:13:20.224582 4971 memory_manager.go:354] "RemoveStaleState removing state" podUID="eb706473-627a-4570-9dfc-0e8fe99d59e1" containerName="cinder-scheduler" Dec 13 07:13:20 crc kubenswrapper[4971]: I1213 07:13:20.224590 4971 memory_manager.go:354] "RemoveStaleState removing state" podUID="a844da97-38e3-4bc5-981f-af3fa9f76483" containerName="neutron-api" Dec 13 07:13:20 crc kubenswrapper[4971]: I1213 07:13:20.224602 4971 memory_manager.go:354] "RemoveStaleState removing state" podUID="eb706473-627a-4570-9dfc-0e8fe99d59e1" containerName="probe" Dec 13 07:13:20 crc kubenswrapper[4971]: I1213 07:13:20.224614 4971 memory_manager.go:354] "RemoveStaleState removing state" podUID="cd7d111d-0b3e-4bd7-ba07-10e296208a7f" containerName="dnsmasq-dns" Dec 13 07:13:20 crc kubenswrapper[4971]: I1213 07:13:20.225472 4971 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-scheduler-0" Dec 13 07:13:20 crc kubenswrapper[4971]: I1213 07:13:20.227801 4971 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cinder-scheduler-config-data" Dec 13 07:13:20 crc kubenswrapper[4971]: I1213 07:13:20.236375 4971 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cinder-scheduler-0"] Dec 13 07:13:20 crc kubenswrapper[4971]: I1213 07:13:20.310018 4971 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/4c666e54-0053-4ecc-b8bd-6a2771ebbad1-combined-ca-bundle\") pod \"cinder-scheduler-0\" (UID: \"4c666e54-0053-4ecc-b8bd-6a2771ebbad1\") " pod="openstack/cinder-scheduler-0" Dec 13 07:13:20 crc kubenswrapper[4971]: I1213 07:13:20.310106 4971 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/4c666e54-0053-4ecc-b8bd-6a2771ebbad1-config-data\") pod \"cinder-scheduler-0\" (UID: \"4c666e54-0053-4ecc-b8bd-6a2771ebbad1\") " pod="openstack/cinder-scheduler-0" Dec 13 07:13:20 crc kubenswrapper[4971]: I1213 07:13:20.310148 4971 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/4c666e54-0053-4ecc-b8bd-6a2771ebbad1-etc-machine-id\") pod \"cinder-scheduler-0\" (UID: \"4c666e54-0053-4ecc-b8bd-6a2771ebbad1\") " pod="openstack/cinder-scheduler-0" Dec 13 07:13:20 crc kubenswrapper[4971]: I1213 07:13:20.310188 4971 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/4c666e54-0053-4ecc-b8bd-6a2771ebbad1-config-data-custom\") pod \"cinder-scheduler-0\" (UID: \"4c666e54-0053-4ecc-b8bd-6a2771ebbad1\") " pod="openstack/cinder-scheduler-0" Dec 13 07:13:20 crc kubenswrapper[4971]: I1213 07:13:20.310212 4971 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-j57pt\" (UniqueName: \"kubernetes.io/projected/4c666e54-0053-4ecc-b8bd-6a2771ebbad1-kube-api-access-j57pt\") pod \"cinder-scheduler-0\" (UID: \"4c666e54-0053-4ecc-b8bd-6a2771ebbad1\") " pod="openstack/cinder-scheduler-0" Dec 13 07:13:20 crc kubenswrapper[4971]: I1213 07:13:20.310261 4971 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/4c666e54-0053-4ecc-b8bd-6a2771ebbad1-scripts\") pod \"cinder-scheduler-0\" (UID: \"4c666e54-0053-4ecc-b8bd-6a2771ebbad1\") " pod="openstack/cinder-scheduler-0" Dec 13 07:13:20 crc kubenswrapper[4971]: I1213 07:13:20.412302 4971 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/4c666e54-0053-4ecc-b8bd-6a2771ebbad1-scripts\") pod \"cinder-scheduler-0\" (UID: \"4c666e54-0053-4ecc-b8bd-6a2771ebbad1\") " pod="openstack/cinder-scheduler-0" Dec 13 07:13:20 crc kubenswrapper[4971]: I1213 07:13:20.412433 4971 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/4c666e54-0053-4ecc-b8bd-6a2771ebbad1-combined-ca-bundle\") pod \"cinder-scheduler-0\" (UID: \"4c666e54-0053-4ecc-b8bd-6a2771ebbad1\") " pod="openstack/cinder-scheduler-0" Dec 13 07:13:20 crc kubenswrapper[4971]: I1213 07:13:20.412495 4971 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/4c666e54-0053-4ecc-b8bd-6a2771ebbad1-config-data\") pod \"cinder-scheduler-0\" (UID: \"4c666e54-0053-4ecc-b8bd-6a2771ebbad1\") " pod="openstack/cinder-scheduler-0" Dec 13 07:13:20 crc kubenswrapper[4971]: I1213 07:13:20.412643 4971 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/4c666e54-0053-4ecc-b8bd-6a2771ebbad1-etc-machine-id\") pod \"cinder-scheduler-0\" (UID: \"4c666e54-0053-4ecc-b8bd-6a2771ebbad1\") " pod="openstack/cinder-scheduler-0" Dec 13 07:13:20 crc kubenswrapper[4971]: I1213 07:13:20.412679 4971 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/4c666e54-0053-4ecc-b8bd-6a2771ebbad1-config-data-custom\") pod \"cinder-scheduler-0\" (UID: \"4c666e54-0053-4ecc-b8bd-6a2771ebbad1\") " pod="openstack/cinder-scheduler-0" Dec 13 07:13:20 crc kubenswrapper[4971]: I1213 07:13:20.412709 4971 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-j57pt\" (UniqueName: \"kubernetes.io/projected/4c666e54-0053-4ecc-b8bd-6a2771ebbad1-kube-api-access-j57pt\") pod \"cinder-scheduler-0\" (UID: \"4c666e54-0053-4ecc-b8bd-6a2771ebbad1\") " pod="openstack/cinder-scheduler-0" Dec 13 07:13:20 crc kubenswrapper[4971]: I1213 07:13:20.412792 4971 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/4c666e54-0053-4ecc-b8bd-6a2771ebbad1-etc-machine-id\") pod \"cinder-scheduler-0\" (UID: \"4c666e54-0053-4ecc-b8bd-6a2771ebbad1\") " pod="openstack/cinder-scheduler-0" Dec 13 07:13:20 crc kubenswrapper[4971]: I1213 07:13:20.418133 4971 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/4c666e54-0053-4ecc-b8bd-6a2771ebbad1-scripts\") pod \"cinder-scheduler-0\" (UID: \"4c666e54-0053-4ecc-b8bd-6a2771ebbad1\") " pod="openstack/cinder-scheduler-0" Dec 13 07:13:20 crc kubenswrapper[4971]: I1213 07:13:20.419032 4971 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/4c666e54-0053-4ecc-b8bd-6a2771ebbad1-config-data-custom\") pod \"cinder-scheduler-0\" (UID: \"4c666e54-0053-4ecc-b8bd-6a2771ebbad1\") " pod="openstack/cinder-scheduler-0" Dec 13 07:13:20 crc kubenswrapper[4971]: I1213 07:13:20.423394 4971 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/4c666e54-0053-4ecc-b8bd-6a2771ebbad1-config-data\") pod \"cinder-scheduler-0\" (UID: \"4c666e54-0053-4ecc-b8bd-6a2771ebbad1\") " pod="openstack/cinder-scheduler-0" Dec 13 07:13:20 crc kubenswrapper[4971]: I1213 07:13:20.426297 4971 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/4c666e54-0053-4ecc-b8bd-6a2771ebbad1-combined-ca-bundle\") pod \"cinder-scheduler-0\" (UID: \"4c666e54-0053-4ecc-b8bd-6a2771ebbad1\") " pod="openstack/cinder-scheduler-0" Dec 13 07:13:20 crc kubenswrapper[4971]: I1213 07:13:20.464269 4971 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-j57pt\" (UniqueName: \"kubernetes.io/projected/4c666e54-0053-4ecc-b8bd-6a2771ebbad1-kube-api-access-j57pt\") pod \"cinder-scheduler-0\" (UID: \"4c666e54-0053-4ecc-b8bd-6a2771ebbad1\") " pod="openstack/cinder-scheduler-0" Dec 13 07:13:20 crc kubenswrapper[4971]: I1213 07:13:20.552393 4971 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-scheduler-0" Dec 13 07:13:20 crc kubenswrapper[4971]: I1213 07:13:20.928547 4971 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/keystone-7969997c86-5t8c8" Dec 13 07:13:21 crc kubenswrapper[4971]: I1213 07:13:21.046370 4971 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cinder-scheduler-0"] Dec 13 07:13:21 crc kubenswrapper[4971]: I1213 07:13:21.160950 4971 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-scheduler-0" event={"ID":"4c666e54-0053-4ecc-b8bd-6a2771ebbad1","Type":"ContainerStarted","Data":"5b69fb405dbb8aca8c5320201b7d7d20e2003bc01e23ed9e559e8e86ccfde888"} Dec 13 07:13:21 crc kubenswrapper[4971]: I1213 07:13:21.296795 4971 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/openstackclient"] Dec 13 07:13:21 crc kubenswrapper[4971]: I1213 07:13:21.298050 4971 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/openstackclient" Dec 13 07:13:21 crc kubenswrapper[4971]: I1213 07:13:21.300261 4971 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstackclient-openstackclient-dockercfg-rpzxh" Dec 13 07:13:21 crc kubenswrapper[4971]: I1213 07:13:21.300399 4971 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-config" Dec 13 07:13:21 crc kubenswrapper[4971]: I1213 07:13:21.303609 4971 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-config-secret" Dec 13 07:13:21 crc kubenswrapper[4971]: I1213 07:13:21.308358 4971 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/openstackclient"] Dec 13 07:13:21 crc kubenswrapper[4971]: I1213 07:13:21.437009 4971 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"openstack-config-secret\" (UniqueName: \"kubernetes.io/secret/b21860ea-6663-4bc0-ac29-b91d640fad5a-openstack-config-secret\") pod \"openstackclient\" (UID: \"b21860ea-6663-4bc0-ac29-b91d640fad5a\") " pod="openstack/openstackclient" Dec 13 07:13:21 crc kubenswrapper[4971]: I1213 07:13:21.437189 4971 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/b21860ea-6663-4bc0-ac29-b91d640fad5a-combined-ca-bundle\") pod \"openstackclient\" (UID: \"b21860ea-6663-4bc0-ac29-b91d640fad5a\") " pod="openstack/openstackclient" Dec 13 07:13:21 crc kubenswrapper[4971]: I1213 07:13:21.437270 4971 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-9ctbl\" (UniqueName: \"kubernetes.io/projected/b21860ea-6663-4bc0-ac29-b91d640fad5a-kube-api-access-9ctbl\") pod \"openstackclient\" (UID: \"b21860ea-6663-4bc0-ac29-b91d640fad5a\") " pod="openstack/openstackclient" Dec 13 07:13:21 crc kubenswrapper[4971]: I1213 07:13:21.437476 4971 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"openstack-config\" (UniqueName: \"kubernetes.io/configmap/b21860ea-6663-4bc0-ac29-b91d640fad5a-openstack-config\") pod \"openstackclient\" (UID: \"b21860ea-6663-4bc0-ac29-b91d640fad5a\") " pod="openstack/openstackclient" Dec 13 07:13:21 crc kubenswrapper[4971]: I1213 07:13:21.539656 4971 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"openstack-config-secret\" (UniqueName: \"kubernetes.io/secret/b21860ea-6663-4bc0-ac29-b91d640fad5a-openstack-config-secret\") pod \"openstackclient\" (UID: \"b21860ea-6663-4bc0-ac29-b91d640fad5a\") " pod="openstack/openstackclient" Dec 13 07:13:21 crc kubenswrapper[4971]: I1213 07:13:21.539799 4971 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/b21860ea-6663-4bc0-ac29-b91d640fad5a-combined-ca-bundle\") pod \"openstackclient\" (UID: \"b21860ea-6663-4bc0-ac29-b91d640fad5a\") " pod="openstack/openstackclient" Dec 13 07:13:21 crc kubenswrapper[4971]: I1213 07:13:21.539843 4971 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-9ctbl\" (UniqueName: \"kubernetes.io/projected/b21860ea-6663-4bc0-ac29-b91d640fad5a-kube-api-access-9ctbl\") pod \"openstackclient\" (UID: \"b21860ea-6663-4bc0-ac29-b91d640fad5a\") " pod="openstack/openstackclient" Dec 13 07:13:21 crc kubenswrapper[4971]: I1213 07:13:21.539933 4971 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"openstack-config\" (UniqueName: \"kubernetes.io/configmap/b21860ea-6663-4bc0-ac29-b91d640fad5a-openstack-config\") pod \"openstackclient\" (UID: \"b21860ea-6663-4bc0-ac29-b91d640fad5a\") " pod="openstack/openstackclient" Dec 13 07:13:21 crc kubenswrapper[4971]: I1213 07:13:21.541030 4971 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"openstack-config\" (UniqueName: \"kubernetes.io/configmap/b21860ea-6663-4bc0-ac29-b91d640fad5a-openstack-config\") pod \"openstackclient\" (UID: \"b21860ea-6663-4bc0-ac29-b91d640fad5a\") " pod="openstack/openstackclient" Dec 13 07:13:21 crc kubenswrapper[4971]: I1213 07:13:21.544557 4971 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"openstack-config-secret\" (UniqueName: \"kubernetes.io/secret/b21860ea-6663-4bc0-ac29-b91d640fad5a-openstack-config-secret\") pod \"openstackclient\" (UID: \"b21860ea-6663-4bc0-ac29-b91d640fad5a\") " pod="openstack/openstackclient" Dec 13 07:13:21 crc kubenswrapper[4971]: I1213 07:13:21.544640 4971 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/b21860ea-6663-4bc0-ac29-b91d640fad5a-combined-ca-bundle\") pod \"openstackclient\" (UID: \"b21860ea-6663-4bc0-ac29-b91d640fad5a\") " pod="openstack/openstackclient" Dec 13 07:13:21 crc kubenswrapper[4971]: I1213 07:13:21.564631 4971 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-9ctbl\" (UniqueName: \"kubernetes.io/projected/b21860ea-6663-4bc0-ac29-b91d640fad5a-kube-api-access-9ctbl\") pod \"openstackclient\" (UID: \"b21860ea-6663-4bc0-ac29-b91d640fad5a\") " pod="openstack/openstackclient" Dec 13 07:13:21 crc kubenswrapper[4971]: I1213 07:13:21.617467 4971 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/openstackclient" Dec 13 07:13:21 crc kubenswrapper[4971]: I1213 07:13:21.786114 4971 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="eb706473-627a-4570-9dfc-0e8fe99d59e1" path="/var/lib/kubelet/pods/eb706473-627a-4570-9dfc-0e8fe99d59e1/volumes" Dec 13 07:13:22 crc kubenswrapper[4971]: I1213 07:13:22.126055 4971 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/openstackclient"] Dec 13 07:13:22 crc kubenswrapper[4971]: W1213 07:13:22.143967 4971 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podb21860ea_6663_4bc0_ac29_b91d640fad5a.slice/crio-dc9d092c4ae185a274c070449219810d56c31ccee9b23509c53a9c0f398f79de WatchSource:0}: Error finding container dc9d092c4ae185a274c070449219810d56c31ccee9b23509c53a9c0f398f79de: Status 404 returned error can't find the container with id dc9d092c4ae185a274c070449219810d56c31ccee9b23509c53a9c0f398f79de Dec 13 07:13:22 crc kubenswrapper[4971]: I1213 07:13:22.192003 4971 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-scheduler-0" event={"ID":"4c666e54-0053-4ecc-b8bd-6a2771ebbad1","Type":"ContainerStarted","Data":"b76b1bf857ff861eead7d61968905f6b60bf21a5872fb391c56da19922b2bbb7"} Dec 13 07:13:22 crc kubenswrapper[4971]: I1213 07:13:22.197564 4971 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/openstackclient" event={"ID":"b21860ea-6663-4bc0-ac29-b91d640fad5a","Type":"ContainerStarted","Data":"dc9d092c4ae185a274c070449219810d56c31ccee9b23509c53a9c0f398f79de"} Dec 13 07:13:23 crc kubenswrapper[4971]: I1213 07:13:23.206387 4971 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-scheduler-0" event={"ID":"4c666e54-0053-4ecc-b8bd-6a2771ebbad1","Type":"ContainerStarted","Data":"4416b3c12f088c3895454fbc2ff69831badde87198030a2944a60810fe880786"} Dec 13 07:13:23 crc kubenswrapper[4971]: I1213 07:13:23.225557 4971 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/cinder-scheduler-0" podStartSLOduration=3.225540026 podStartE2EDuration="3.225540026s" podCreationTimestamp="2025-12-13 07:13:20 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-13 07:13:23.224372247 +0000 UTC m=+1459.828781695" watchObservedRunningTime="2025-12-13 07:13:23.225540026 +0000 UTC m=+1459.829949474" Dec 13 07:13:24 crc kubenswrapper[4971]: I1213 07:13:24.035758 4971 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/horizon-8699dc887d-sl2h6" podUID="772eedab-7c62-454b-b722-ee453b39e89f" containerName="horizon" probeResult="failure" output="Get \"https://10.217.0.140:8443/dashboard/auth/login/?next=/dashboard/\": dial tcp 10.217.0.140:8443: connect: connection refused" Dec 13 07:13:24 crc kubenswrapper[4971]: I1213 07:13:24.747827 4971 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/barbican-api-55549fd796-4tztx" Dec 13 07:13:25 crc kubenswrapper[4971]: I1213 07:13:25.553115 4971 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/cinder-scheduler-0" Dec 13 07:13:25 crc kubenswrapper[4971]: I1213 07:13:25.582883 4971 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/barbican-api-55549fd796-4tztx" Dec 13 07:13:25 crc kubenswrapper[4971]: I1213 07:13:25.674158 4971 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/barbican-api-5c88bd59bd-2qlbj"] Dec 13 07:13:25 crc kubenswrapper[4971]: I1213 07:13:25.674425 4971 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/barbican-api-5c88bd59bd-2qlbj" podUID="02578e71-2865-4a4b-96e9-5bf6c625183c" containerName="barbican-api-log" containerID="cri-o://5c2237976dc29579a8c63c4fbffbcb524f6e0ae5cbf13be0ece76df573636e9c" gracePeriod=30 Dec 13 07:13:25 crc kubenswrapper[4971]: I1213 07:13:25.674578 4971 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/barbican-api-5c88bd59bd-2qlbj" podUID="02578e71-2865-4a4b-96e9-5bf6c625183c" containerName="barbican-api" containerID="cri-o://75fc3d56284952d3e64a99eccdb0f35e87027803ee64f33d79a9a487157aad17" gracePeriod=30 Dec 13 07:13:26 crc kubenswrapper[4971]: I1213 07:13:26.251175 4971 generic.go:334] "Generic (PLEG): container finished" podID="02578e71-2865-4a4b-96e9-5bf6c625183c" containerID="5c2237976dc29579a8c63c4fbffbcb524f6e0ae5cbf13be0ece76df573636e9c" exitCode=143 Dec 13 07:13:26 crc kubenswrapper[4971]: I1213 07:13:26.251217 4971 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-api-5c88bd59bd-2qlbj" event={"ID":"02578e71-2865-4a4b-96e9-5bf6c625183c","Type":"ContainerDied","Data":"5c2237976dc29579a8c63c4fbffbcb524f6e0ae5cbf13be0ece76df573636e9c"} Dec 13 07:13:27 crc kubenswrapper[4971]: I1213 07:13:27.302205 4971 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/cinder-api-0" Dec 13 07:13:28 crc kubenswrapper[4971]: I1213 07:13:28.858277 4971 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/barbican-api-5c88bd59bd-2qlbj" podUID="02578e71-2865-4a4b-96e9-5bf6c625183c" containerName="barbican-api-log" probeResult="failure" output="Get \"http://10.217.0.152:9311/healthcheck\": read tcp 10.217.0.2:50046->10.217.0.152:9311: read: connection reset by peer" Dec 13 07:13:28 crc kubenswrapper[4971]: I1213 07:13:28.858322 4971 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/barbican-api-5c88bd59bd-2qlbj" podUID="02578e71-2865-4a4b-96e9-5bf6c625183c" containerName="barbican-api" probeResult="failure" output="Get \"http://10.217.0.152:9311/healthcheck\": read tcp 10.217.0.2:50030->10.217.0.152:9311: read: connection reset by peer" Dec 13 07:13:29 crc kubenswrapper[4971]: I1213 07:13:29.243933 4971 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-api-5c88bd59bd-2qlbj" Dec 13 07:13:29 crc kubenswrapper[4971]: I1213 07:13:29.286609 4971 generic.go:334] "Generic (PLEG): container finished" podID="02578e71-2865-4a4b-96e9-5bf6c625183c" containerID="75fc3d56284952d3e64a99eccdb0f35e87027803ee64f33d79a9a487157aad17" exitCode=0 Dec 13 07:13:29 crc kubenswrapper[4971]: I1213 07:13:29.286957 4971 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-api-5c88bd59bd-2qlbj" event={"ID":"02578e71-2865-4a4b-96e9-5bf6c625183c","Type":"ContainerDied","Data":"75fc3d56284952d3e64a99eccdb0f35e87027803ee64f33d79a9a487157aad17"} Dec 13 07:13:29 crc kubenswrapper[4971]: I1213 07:13:29.286980 4971 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-api-5c88bd59bd-2qlbj" Dec 13 07:13:29 crc kubenswrapper[4971]: I1213 07:13:29.287003 4971 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-api-5c88bd59bd-2qlbj" event={"ID":"02578e71-2865-4a4b-96e9-5bf6c625183c","Type":"ContainerDied","Data":"255762a06cafe12675758dacf9268763d0b8bf0ed85d57ca20e28c70a0253280"} Dec 13 07:13:29 crc kubenswrapper[4971]: I1213 07:13:29.287025 4971 scope.go:117] "RemoveContainer" containerID="75fc3d56284952d3e64a99eccdb0f35e87027803ee64f33d79a9a487157aad17" Dec 13 07:13:29 crc kubenswrapper[4971]: I1213 07:13:29.329302 4971 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-kk4n8\" (UniqueName: \"kubernetes.io/projected/02578e71-2865-4a4b-96e9-5bf6c625183c-kube-api-access-kk4n8\") pod \"02578e71-2865-4a4b-96e9-5bf6c625183c\" (UID: \"02578e71-2865-4a4b-96e9-5bf6c625183c\") " Dec 13 07:13:29 crc kubenswrapper[4971]: I1213 07:13:29.329390 4971 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/02578e71-2865-4a4b-96e9-5bf6c625183c-logs\") pod \"02578e71-2865-4a4b-96e9-5bf6c625183c\" (UID: \"02578e71-2865-4a4b-96e9-5bf6c625183c\") " Dec 13 07:13:29 crc kubenswrapper[4971]: I1213 07:13:29.329448 4971 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/02578e71-2865-4a4b-96e9-5bf6c625183c-combined-ca-bundle\") pod \"02578e71-2865-4a4b-96e9-5bf6c625183c\" (UID: \"02578e71-2865-4a4b-96e9-5bf6c625183c\") " Dec 13 07:13:29 crc kubenswrapper[4971]: I1213 07:13:29.329546 4971 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/02578e71-2865-4a4b-96e9-5bf6c625183c-config-data\") pod \"02578e71-2865-4a4b-96e9-5bf6c625183c\" (UID: \"02578e71-2865-4a4b-96e9-5bf6c625183c\") " Dec 13 07:13:29 crc kubenswrapper[4971]: I1213 07:13:29.329585 4971 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/02578e71-2865-4a4b-96e9-5bf6c625183c-config-data-custom\") pod \"02578e71-2865-4a4b-96e9-5bf6c625183c\" (UID: \"02578e71-2865-4a4b-96e9-5bf6c625183c\") " Dec 13 07:13:29 crc kubenswrapper[4971]: I1213 07:13:29.333151 4971 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/02578e71-2865-4a4b-96e9-5bf6c625183c-logs" (OuterVolumeSpecName: "logs") pod "02578e71-2865-4a4b-96e9-5bf6c625183c" (UID: "02578e71-2865-4a4b-96e9-5bf6c625183c"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 13 07:13:29 crc kubenswrapper[4971]: I1213 07:13:29.343717 4971 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/02578e71-2865-4a4b-96e9-5bf6c625183c-config-data-custom" (OuterVolumeSpecName: "config-data-custom") pod "02578e71-2865-4a4b-96e9-5bf6c625183c" (UID: "02578e71-2865-4a4b-96e9-5bf6c625183c"). InnerVolumeSpecName "config-data-custom". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 13 07:13:29 crc kubenswrapper[4971]: I1213 07:13:29.346767 4971 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/02578e71-2865-4a4b-96e9-5bf6c625183c-kube-api-access-kk4n8" (OuterVolumeSpecName: "kube-api-access-kk4n8") pod "02578e71-2865-4a4b-96e9-5bf6c625183c" (UID: "02578e71-2865-4a4b-96e9-5bf6c625183c"). InnerVolumeSpecName "kube-api-access-kk4n8". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 13 07:13:29 crc kubenswrapper[4971]: I1213 07:13:29.373804 4971 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/02578e71-2865-4a4b-96e9-5bf6c625183c-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "02578e71-2865-4a4b-96e9-5bf6c625183c" (UID: "02578e71-2865-4a4b-96e9-5bf6c625183c"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 13 07:13:29 crc kubenswrapper[4971]: I1213 07:13:29.433254 4971 reconciler_common.go:293] "Volume detached for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/02578e71-2865-4a4b-96e9-5bf6c625183c-config-data-custom\") on node \"crc\" DevicePath \"\"" Dec 13 07:13:29 crc kubenswrapper[4971]: I1213 07:13:29.433306 4971 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-kk4n8\" (UniqueName: \"kubernetes.io/projected/02578e71-2865-4a4b-96e9-5bf6c625183c-kube-api-access-kk4n8\") on node \"crc\" DevicePath \"\"" Dec 13 07:13:29 crc kubenswrapper[4971]: I1213 07:13:29.433320 4971 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/02578e71-2865-4a4b-96e9-5bf6c625183c-logs\") on node \"crc\" DevicePath \"\"" Dec 13 07:13:29 crc kubenswrapper[4971]: I1213 07:13:29.433331 4971 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/02578e71-2865-4a4b-96e9-5bf6c625183c-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 13 07:13:29 crc kubenswrapper[4971]: I1213 07:13:29.433458 4971 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/02578e71-2865-4a4b-96e9-5bf6c625183c-config-data" (OuterVolumeSpecName: "config-data") pod "02578e71-2865-4a4b-96e9-5bf6c625183c" (UID: "02578e71-2865-4a4b-96e9-5bf6c625183c"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 13 07:13:29 crc kubenswrapper[4971]: I1213 07:13:29.436587 4971 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-api-db-create-ndmqx"] Dec 13 07:13:29 crc kubenswrapper[4971]: E1213 07:13:29.437022 4971 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="02578e71-2865-4a4b-96e9-5bf6c625183c" containerName="barbican-api" Dec 13 07:13:29 crc kubenswrapper[4971]: I1213 07:13:29.437046 4971 state_mem.go:107] "Deleted CPUSet assignment" podUID="02578e71-2865-4a4b-96e9-5bf6c625183c" containerName="barbican-api" Dec 13 07:13:29 crc kubenswrapper[4971]: E1213 07:13:29.437070 4971 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="02578e71-2865-4a4b-96e9-5bf6c625183c" containerName="barbican-api-log" Dec 13 07:13:29 crc kubenswrapper[4971]: I1213 07:13:29.437079 4971 state_mem.go:107] "Deleted CPUSet assignment" podUID="02578e71-2865-4a4b-96e9-5bf6c625183c" containerName="barbican-api-log" Dec 13 07:13:29 crc kubenswrapper[4971]: I1213 07:13:29.437290 4971 memory_manager.go:354] "RemoveStaleState removing state" podUID="02578e71-2865-4a4b-96e9-5bf6c625183c" containerName="barbican-api" Dec 13 07:13:29 crc kubenswrapper[4971]: I1213 07:13:29.437310 4971 memory_manager.go:354] "RemoveStaleState removing state" podUID="02578e71-2865-4a4b-96e9-5bf6c625183c" containerName="barbican-api-log" Dec 13 07:13:29 crc kubenswrapper[4971]: I1213 07:13:29.438743 4971 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-db-create-ndmqx" Dec 13 07:13:29 crc kubenswrapper[4971]: I1213 07:13:29.453169 4971 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-api-db-create-ndmqx"] Dec 13 07:13:29 crc kubenswrapper[4971]: I1213 07:13:29.536722 4971 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-8nxsg\" (UniqueName: \"kubernetes.io/projected/8b60a7ee-8d5e-4bf2-8d64-161aeced2357-kube-api-access-8nxsg\") pod \"nova-api-db-create-ndmqx\" (UID: \"8b60a7ee-8d5e-4bf2-8d64-161aeced2357\") " pod="openstack/nova-api-db-create-ndmqx" Dec 13 07:13:29 crc kubenswrapper[4971]: I1213 07:13:29.536781 4971 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/8b60a7ee-8d5e-4bf2-8d64-161aeced2357-operator-scripts\") pod \"nova-api-db-create-ndmqx\" (UID: \"8b60a7ee-8d5e-4bf2-8d64-161aeced2357\") " pod="openstack/nova-api-db-create-ndmqx" Dec 13 07:13:29 crc kubenswrapper[4971]: I1213 07:13:29.541442 4971 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/02578e71-2865-4a4b-96e9-5bf6c625183c-config-data\") on node \"crc\" DevicePath \"\"" Dec 13 07:13:29 crc kubenswrapper[4971]: I1213 07:13:29.551956 4971 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-cell0-db-create-jsdsv"] Dec 13 07:13:29 crc kubenswrapper[4971]: I1213 07:13:29.555412 4971 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-db-create-jsdsv" Dec 13 07:13:29 crc kubenswrapper[4971]: I1213 07:13:29.584554 4971 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell0-db-create-jsdsv"] Dec 13 07:13:29 crc kubenswrapper[4971]: I1213 07:13:29.628149 4971 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/barbican-api-5c88bd59bd-2qlbj"] Dec 13 07:13:29 crc kubenswrapper[4971]: I1213 07:13:29.637815 4971 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/barbican-api-5c88bd59bd-2qlbj"] Dec 13 07:13:29 crc kubenswrapper[4971]: I1213 07:13:29.644669 4971 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-8nxsg\" (UniqueName: \"kubernetes.io/projected/8b60a7ee-8d5e-4bf2-8d64-161aeced2357-kube-api-access-8nxsg\") pod \"nova-api-db-create-ndmqx\" (UID: \"8b60a7ee-8d5e-4bf2-8d64-161aeced2357\") " pod="openstack/nova-api-db-create-ndmqx" Dec 13 07:13:29 crc kubenswrapper[4971]: I1213 07:13:29.644735 4971 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/8b60a7ee-8d5e-4bf2-8d64-161aeced2357-operator-scripts\") pod \"nova-api-db-create-ndmqx\" (UID: \"8b60a7ee-8d5e-4bf2-8d64-161aeced2357\") " pod="openstack/nova-api-db-create-ndmqx" Dec 13 07:13:29 crc kubenswrapper[4971]: I1213 07:13:29.645753 4971 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/8b60a7ee-8d5e-4bf2-8d64-161aeced2357-operator-scripts\") pod \"nova-api-db-create-ndmqx\" (UID: \"8b60a7ee-8d5e-4bf2-8d64-161aeced2357\") " pod="openstack/nova-api-db-create-ndmqx" Dec 13 07:13:29 crc kubenswrapper[4971]: I1213 07:13:29.658611 4971 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-cell1-db-create-4j49j"] Dec 13 07:13:29 crc kubenswrapper[4971]: I1213 07:13:29.659923 4971 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-db-create-4j49j" Dec 13 07:13:29 crc kubenswrapper[4971]: I1213 07:13:29.669424 4971 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-api-048e-account-create-update-kggpx"] Dec 13 07:13:29 crc kubenswrapper[4971]: I1213 07:13:29.670418 4971 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-8nxsg\" (UniqueName: \"kubernetes.io/projected/8b60a7ee-8d5e-4bf2-8d64-161aeced2357-kube-api-access-8nxsg\") pod \"nova-api-db-create-ndmqx\" (UID: \"8b60a7ee-8d5e-4bf2-8d64-161aeced2357\") " pod="openstack/nova-api-db-create-ndmqx" Dec 13 07:13:29 crc kubenswrapper[4971]: I1213 07:13:29.683472 4971 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-api-048e-account-create-update-kggpx"] Dec 13 07:13:29 crc kubenswrapper[4971]: I1213 07:13:29.683605 4971 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-048e-account-create-update-kggpx" Dec 13 07:13:29 crc kubenswrapper[4971]: I1213 07:13:29.685571 4971 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-api-db-secret" Dec 13 07:13:29 crc kubenswrapper[4971]: I1213 07:13:29.692941 4971 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell1-db-create-4j49j"] Dec 13 07:13:29 crc kubenswrapper[4971]: I1213 07:13:29.746680 4971 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-tvh6r\" (UniqueName: \"kubernetes.io/projected/0ed1405a-8326-4fe9-8fb4-1d97b9e8befb-kube-api-access-tvh6r\") pod \"nova-cell0-db-create-jsdsv\" (UID: \"0ed1405a-8326-4fe9-8fb4-1d97b9e8befb\") " pod="openstack/nova-cell0-db-create-jsdsv" Dec 13 07:13:29 crc kubenswrapper[4971]: I1213 07:13:29.747491 4971 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/0ed1405a-8326-4fe9-8fb4-1d97b9e8befb-operator-scripts\") pod \"nova-cell0-db-create-jsdsv\" (UID: \"0ed1405a-8326-4fe9-8fb4-1d97b9e8befb\") " pod="openstack/nova-cell0-db-create-jsdsv" Dec 13 07:13:29 crc kubenswrapper[4971]: I1213 07:13:29.770591 4971 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-db-create-ndmqx" Dec 13 07:13:29 crc kubenswrapper[4971]: I1213 07:13:29.782243 4971 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="02578e71-2865-4a4b-96e9-5bf6c625183c" path="/var/lib/kubelet/pods/02578e71-2865-4a4b-96e9-5bf6c625183c/volumes" Dec 13 07:13:29 crc kubenswrapper[4971]: I1213 07:13:29.851461 4971 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-cell0-ed10-account-create-update-tth4x"] Dec 13 07:13:29 crc kubenswrapper[4971]: I1213 07:13:29.851624 4971 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/5ba87867-bb90-4587-9012-6db9d2901b10-operator-scripts\") pod \"nova-cell1-db-create-4j49j\" (UID: \"5ba87867-bb90-4587-9012-6db9d2901b10\") " pod="openstack/nova-cell1-db-create-4j49j" Dec 13 07:13:29 crc kubenswrapper[4971]: I1213 07:13:29.851677 4971 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-frnhv\" (UniqueName: \"kubernetes.io/projected/5ba87867-bb90-4587-9012-6db9d2901b10-kube-api-access-frnhv\") pod \"nova-cell1-db-create-4j49j\" (UID: \"5ba87867-bb90-4587-9012-6db9d2901b10\") " pod="openstack/nova-cell1-db-create-4j49j" Dec 13 07:13:29 crc kubenswrapper[4971]: I1213 07:13:29.851759 4971 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-4w6kr\" (UniqueName: \"kubernetes.io/projected/edac9227-16d2-48b0-ac2b-3f71c67d6cf8-kube-api-access-4w6kr\") pod \"nova-api-048e-account-create-update-kggpx\" (UID: \"edac9227-16d2-48b0-ac2b-3f71c67d6cf8\") " pod="openstack/nova-api-048e-account-create-update-kggpx" Dec 13 07:13:29 crc kubenswrapper[4971]: I1213 07:13:29.851797 4971 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-tvh6r\" (UniqueName: \"kubernetes.io/projected/0ed1405a-8326-4fe9-8fb4-1d97b9e8befb-kube-api-access-tvh6r\") pod \"nova-cell0-db-create-jsdsv\" (UID: \"0ed1405a-8326-4fe9-8fb4-1d97b9e8befb\") " pod="openstack/nova-cell0-db-create-jsdsv" Dec 13 07:13:29 crc kubenswrapper[4971]: I1213 07:13:29.851824 4971 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/0ed1405a-8326-4fe9-8fb4-1d97b9e8befb-operator-scripts\") pod \"nova-cell0-db-create-jsdsv\" (UID: \"0ed1405a-8326-4fe9-8fb4-1d97b9e8befb\") " pod="openstack/nova-cell0-db-create-jsdsv" Dec 13 07:13:29 crc kubenswrapper[4971]: I1213 07:13:29.851881 4971 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/edac9227-16d2-48b0-ac2b-3f71c67d6cf8-operator-scripts\") pod \"nova-api-048e-account-create-update-kggpx\" (UID: \"edac9227-16d2-48b0-ac2b-3f71c67d6cf8\") " pod="openstack/nova-api-048e-account-create-update-kggpx" Dec 13 07:13:29 crc kubenswrapper[4971]: I1213 07:13:29.852970 4971 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-ed10-account-create-update-tth4x" Dec 13 07:13:29 crc kubenswrapper[4971]: I1213 07:13:29.853105 4971 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/0ed1405a-8326-4fe9-8fb4-1d97b9e8befb-operator-scripts\") pod \"nova-cell0-db-create-jsdsv\" (UID: \"0ed1405a-8326-4fe9-8fb4-1d97b9e8befb\") " pod="openstack/nova-cell0-db-create-jsdsv" Dec 13 07:13:29 crc kubenswrapper[4971]: I1213 07:13:29.858024 4971 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-cell0-db-secret" Dec 13 07:13:29 crc kubenswrapper[4971]: I1213 07:13:29.864868 4971 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell0-ed10-account-create-update-tth4x"] Dec 13 07:13:29 crc kubenswrapper[4971]: I1213 07:13:29.893372 4971 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-tvh6r\" (UniqueName: \"kubernetes.io/projected/0ed1405a-8326-4fe9-8fb4-1d97b9e8befb-kube-api-access-tvh6r\") pod \"nova-cell0-db-create-jsdsv\" (UID: \"0ed1405a-8326-4fe9-8fb4-1d97b9e8befb\") " pod="openstack/nova-cell0-db-create-jsdsv" Dec 13 07:13:29 crc kubenswrapper[4971]: I1213 07:13:29.957649 4971 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-4w6kr\" (UniqueName: \"kubernetes.io/projected/edac9227-16d2-48b0-ac2b-3f71c67d6cf8-kube-api-access-4w6kr\") pod \"nova-api-048e-account-create-update-kggpx\" (UID: \"edac9227-16d2-48b0-ac2b-3f71c67d6cf8\") " pod="openstack/nova-api-048e-account-create-update-kggpx" Dec 13 07:13:29 crc kubenswrapper[4971]: I1213 07:13:29.957733 4971 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-mmc9b\" (UniqueName: \"kubernetes.io/projected/5ffa477d-1f7b-445f-9a97-65a7889e44d4-kube-api-access-mmc9b\") pod \"nova-cell0-ed10-account-create-update-tth4x\" (UID: \"5ffa477d-1f7b-445f-9a97-65a7889e44d4\") " pod="openstack/nova-cell0-ed10-account-create-update-tth4x" Dec 13 07:13:29 crc kubenswrapper[4971]: I1213 07:13:29.957768 4971 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/edac9227-16d2-48b0-ac2b-3f71c67d6cf8-operator-scripts\") pod \"nova-api-048e-account-create-update-kggpx\" (UID: \"edac9227-16d2-48b0-ac2b-3f71c67d6cf8\") " pod="openstack/nova-api-048e-account-create-update-kggpx" Dec 13 07:13:29 crc kubenswrapper[4971]: I1213 07:13:29.957792 4971 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/5ffa477d-1f7b-445f-9a97-65a7889e44d4-operator-scripts\") pod \"nova-cell0-ed10-account-create-update-tth4x\" (UID: \"5ffa477d-1f7b-445f-9a97-65a7889e44d4\") " pod="openstack/nova-cell0-ed10-account-create-update-tth4x" Dec 13 07:13:29 crc kubenswrapper[4971]: I1213 07:13:29.957889 4971 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/5ba87867-bb90-4587-9012-6db9d2901b10-operator-scripts\") pod \"nova-cell1-db-create-4j49j\" (UID: \"5ba87867-bb90-4587-9012-6db9d2901b10\") " pod="openstack/nova-cell1-db-create-4j49j" Dec 13 07:13:29 crc kubenswrapper[4971]: I1213 07:13:29.957918 4971 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-frnhv\" (UniqueName: \"kubernetes.io/projected/5ba87867-bb90-4587-9012-6db9d2901b10-kube-api-access-frnhv\") pod \"nova-cell1-db-create-4j49j\" (UID: \"5ba87867-bb90-4587-9012-6db9d2901b10\") " pod="openstack/nova-cell1-db-create-4j49j" Dec 13 07:13:29 crc kubenswrapper[4971]: I1213 07:13:29.959407 4971 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/5ba87867-bb90-4587-9012-6db9d2901b10-operator-scripts\") pod \"nova-cell1-db-create-4j49j\" (UID: \"5ba87867-bb90-4587-9012-6db9d2901b10\") " pod="openstack/nova-cell1-db-create-4j49j" Dec 13 07:13:29 crc kubenswrapper[4971]: I1213 07:13:29.959436 4971 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/edac9227-16d2-48b0-ac2b-3f71c67d6cf8-operator-scripts\") pod \"nova-api-048e-account-create-update-kggpx\" (UID: \"edac9227-16d2-48b0-ac2b-3f71c67d6cf8\") " pod="openstack/nova-api-048e-account-create-update-kggpx" Dec 13 07:13:30 crc kubenswrapper[4971]: I1213 07:13:30.006204 4971 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-frnhv\" (UniqueName: \"kubernetes.io/projected/5ba87867-bb90-4587-9012-6db9d2901b10-kube-api-access-frnhv\") pod \"nova-cell1-db-create-4j49j\" (UID: \"5ba87867-bb90-4587-9012-6db9d2901b10\") " pod="openstack/nova-cell1-db-create-4j49j" Dec 13 07:13:30 crc kubenswrapper[4971]: I1213 07:13:30.011135 4971 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-4w6kr\" (UniqueName: \"kubernetes.io/projected/edac9227-16d2-48b0-ac2b-3f71c67d6cf8-kube-api-access-4w6kr\") pod \"nova-api-048e-account-create-update-kggpx\" (UID: \"edac9227-16d2-48b0-ac2b-3f71c67d6cf8\") " pod="openstack/nova-api-048e-account-create-update-kggpx" Dec 13 07:13:30 crc kubenswrapper[4971]: I1213 07:13:30.060623 4971 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-mmc9b\" (UniqueName: \"kubernetes.io/projected/5ffa477d-1f7b-445f-9a97-65a7889e44d4-kube-api-access-mmc9b\") pod \"nova-cell0-ed10-account-create-update-tth4x\" (UID: \"5ffa477d-1f7b-445f-9a97-65a7889e44d4\") " pod="openstack/nova-cell0-ed10-account-create-update-tth4x" Dec 13 07:13:30 crc kubenswrapper[4971]: I1213 07:13:30.060686 4971 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/5ffa477d-1f7b-445f-9a97-65a7889e44d4-operator-scripts\") pod \"nova-cell0-ed10-account-create-update-tth4x\" (UID: \"5ffa477d-1f7b-445f-9a97-65a7889e44d4\") " pod="openstack/nova-cell0-ed10-account-create-update-tth4x" Dec 13 07:13:30 crc kubenswrapper[4971]: I1213 07:13:30.061624 4971 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/5ffa477d-1f7b-445f-9a97-65a7889e44d4-operator-scripts\") pod \"nova-cell0-ed10-account-create-update-tth4x\" (UID: \"5ffa477d-1f7b-445f-9a97-65a7889e44d4\") " pod="openstack/nova-cell0-ed10-account-create-update-tth4x" Dec 13 07:13:30 crc kubenswrapper[4971]: I1213 07:13:30.081131 4971 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-cell1-61ad-account-create-update-4ckp9"] Dec 13 07:13:30 crc kubenswrapper[4971]: I1213 07:13:30.082365 4971 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-61ad-account-create-update-4ckp9" Dec 13 07:13:30 crc kubenswrapper[4971]: I1213 07:13:30.086017 4971 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-cell1-db-secret" Dec 13 07:13:30 crc kubenswrapper[4971]: I1213 07:13:30.102725 4971 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell1-61ad-account-create-update-4ckp9"] Dec 13 07:13:30 crc kubenswrapper[4971]: I1213 07:13:30.111138 4971 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-mmc9b\" (UniqueName: \"kubernetes.io/projected/5ffa477d-1f7b-445f-9a97-65a7889e44d4-kube-api-access-mmc9b\") pod \"nova-cell0-ed10-account-create-update-tth4x\" (UID: \"5ffa477d-1f7b-445f-9a97-65a7889e44d4\") " pod="openstack/nova-cell0-ed10-account-create-update-tth4x" Dec 13 07:13:30 crc kubenswrapper[4971]: I1213 07:13:30.174162 4971 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-ed10-account-create-update-tth4x" Dec 13 07:13:30 crc kubenswrapper[4971]: I1213 07:13:30.181063 4971 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-db-create-jsdsv" Dec 13 07:13:30 crc kubenswrapper[4971]: I1213 07:13:30.265924 4971 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-ffkdr\" (UniqueName: \"kubernetes.io/projected/e0e9f238-3cb4-43ff-bac3-f1c283d0915f-kube-api-access-ffkdr\") pod \"nova-cell1-61ad-account-create-update-4ckp9\" (UID: \"e0e9f238-3cb4-43ff-bac3-f1c283d0915f\") " pod="openstack/nova-cell1-61ad-account-create-update-4ckp9" Dec 13 07:13:30 crc kubenswrapper[4971]: I1213 07:13:30.266047 4971 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/e0e9f238-3cb4-43ff-bac3-f1c283d0915f-operator-scripts\") pod \"nova-cell1-61ad-account-create-update-4ckp9\" (UID: \"e0e9f238-3cb4-43ff-bac3-f1c283d0915f\") " pod="openstack/nova-cell1-61ad-account-create-update-4ckp9" Dec 13 07:13:30 crc kubenswrapper[4971]: I1213 07:13:30.274268 4971 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-db-create-4j49j" Dec 13 07:13:30 crc kubenswrapper[4971]: I1213 07:13:30.298661 4971 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-048e-account-create-update-kggpx" Dec 13 07:13:30 crc kubenswrapper[4971]: I1213 07:13:30.368095 4971 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/e0e9f238-3cb4-43ff-bac3-f1c283d0915f-operator-scripts\") pod \"nova-cell1-61ad-account-create-update-4ckp9\" (UID: \"e0e9f238-3cb4-43ff-bac3-f1c283d0915f\") " pod="openstack/nova-cell1-61ad-account-create-update-4ckp9" Dec 13 07:13:30 crc kubenswrapper[4971]: I1213 07:13:30.368207 4971 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-ffkdr\" (UniqueName: \"kubernetes.io/projected/e0e9f238-3cb4-43ff-bac3-f1c283d0915f-kube-api-access-ffkdr\") pod \"nova-cell1-61ad-account-create-update-4ckp9\" (UID: \"e0e9f238-3cb4-43ff-bac3-f1c283d0915f\") " pod="openstack/nova-cell1-61ad-account-create-update-4ckp9" Dec 13 07:13:30 crc kubenswrapper[4971]: I1213 07:13:30.368819 4971 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/e0e9f238-3cb4-43ff-bac3-f1c283d0915f-operator-scripts\") pod \"nova-cell1-61ad-account-create-update-4ckp9\" (UID: \"e0e9f238-3cb4-43ff-bac3-f1c283d0915f\") " pod="openstack/nova-cell1-61ad-account-create-update-4ckp9" Dec 13 07:13:30 crc kubenswrapper[4971]: I1213 07:13:30.394047 4971 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-ffkdr\" (UniqueName: \"kubernetes.io/projected/e0e9f238-3cb4-43ff-bac3-f1c283d0915f-kube-api-access-ffkdr\") pod \"nova-cell1-61ad-account-create-update-4ckp9\" (UID: \"e0e9f238-3cb4-43ff-bac3-f1c283d0915f\") " pod="openstack/nova-cell1-61ad-account-create-update-4ckp9" Dec 13 07:13:30 crc kubenswrapper[4971]: I1213 07:13:30.493092 4971 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-61ad-account-create-update-4ckp9" Dec 13 07:13:30 crc kubenswrapper[4971]: I1213 07:13:30.778814 4971 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/cinder-scheduler-0" Dec 13 07:13:34 crc kubenswrapper[4971]: I1213 07:13:34.037004 4971 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/horizon-8699dc887d-sl2h6" podUID="772eedab-7c62-454b-b722-ee453b39e89f" containerName="horizon" probeResult="failure" output="Get \"https://10.217.0.140:8443/dashboard/auth/login/?next=/dashboard/\": dial tcp 10.217.0.140:8443: connect: connection refused" Dec 13 07:13:34 crc kubenswrapper[4971]: I1213 07:13:34.805127 4971 scope.go:117] "RemoveContainer" containerID="5c2237976dc29579a8c63c4fbffbcb524f6e0ae5cbf13be0ece76df573636e9c" Dec 13 07:13:34 crc kubenswrapper[4971]: I1213 07:13:34.936036 4971 scope.go:117] "RemoveContainer" containerID="75fc3d56284952d3e64a99eccdb0f35e87027803ee64f33d79a9a487157aad17" Dec 13 07:13:34 crc kubenswrapper[4971]: E1213 07:13:34.936873 4971 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"75fc3d56284952d3e64a99eccdb0f35e87027803ee64f33d79a9a487157aad17\": container with ID starting with 75fc3d56284952d3e64a99eccdb0f35e87027803ee64f33d79a9a487157aad17 not found: ID does not exist" containerID="75fc3d56284952d3e64a99eccdb0f35e87027803ee64f33d79a9a487157aad17" Dec 13 07:13:34 crc kubenswrapper[4971]: I1213 07:13:34.936904 4971 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"75fc3d56284952d3e64a99eccdb0f35e87027803ee64f33d79a9a487157aad17"} err="failed to get container status \"75fc3d56284952d3e64a99eccdb0f35e87027803ee64f33d79a9a487157aad17\": rpc error: code = NotFound desc = could not find container \"75fc3d56284952d3e64a99eccdb0f35e87027803ee64f33d79a9a487157aad17\": container with ID starting with 75fc3d56284952d3e64a99eccdb0f35e87027803ee64f33d79a9a487157aad17 not found: ID does not exist" Dec 13 07:13:34 crc kubenswrapper[4971]: I1213 07:13:34.936924 4971 scope.go:117] "RemoveContainer" containerID="5c2237976dc29579a8c63c4fbffbcb524f6e0ae5cbf13be0ece76df573636e9c" Dec 13 07:13:34 crc kubenswrapper[4971]: E1213 07:13:34.937719 4971 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"5c2237976dc29579a8c63c4fbffbcb524f6e0ae5cbf13be0ece76df573636e9c\": container with ID starting with 5c2237976dc29579a8c63c4fbffbcb524f6e0ae5cbf13be0ece76df573636e9c not found: ID does not exist" containerID="5c2237976dc29579a8c63c4fbffbcb524f6e0ae5cbf13be0ece76df573636e9c" Dec 13 07:13:34 crc kubenswrapper[4971]: I1213 07:13:34.937745 4971 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"5c2237976dc29579a8c63c4fbffbcb524f6e0ae5cbf13be0ece76df573636e9c"} err="failed to get container status \"5c2237976dc29579a8c63c4fbffbcb524f6e0ae5cbf13be0ece76df573636e9c\": rpc error: code = NotFound desc = could not find container \"5c2237976dc29579a8c63c4fbffbcb524f6e0ae5cbf13be0ece76df573636e9c\": container with ID starting with 5c2237976dc29579a8c63c4fbffbcb524f6e0ae5cbf13be0ece76df573636e9c not found: ID does not exist" Dec 13 07:13:35 crc kubenswrapper[4971]: I1213 07:13:35.348959 4971 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/openstackclient" event={"ID":"b21860ea-6663-4bc0-ac29-b91d640fad5a","Type":"ContainerStarted","Data":"bcfe84e451c8d0e0080cf3ddc74cc227b2c9f973de80fe94391d90a802d99d7f"} Dec 13 07:13:35 crc kubenswrapper[4971]: W1213 07:13:35.388950 4971 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod0ed1405a_8326_4fe9_8fb4_1d97b9e8befb.slice/crio-289a384ea2f17437068706371c810954cd10f7b1ef6be88d66070f02da680ba1 WatchSource:0}: Error finding container 289a384ea2f17437068706371c810954cd10f7b1ef6be88d66070f02da680ba1: Status 404 returned error can't find the container with id 289a384ea2f17437068706371c810954cd10f7b1ef6be88d66070f02da680ba1 Dec 13 07:13:35 crc kubenswrapper[4971]: I1213 07:13:35.392070 4971 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/openstackclient" podStartSLOduration=1.644928332 podStartE2EDuration="14.392047502s" podCreationTimestamp="2025-12-13 07:13:21 +0000 UTC" firstStartedPulling="2025-12-13 07:13:22.151178615 +0000 UTC m=+1458.755588073" lastFinishedPulling="2025-12-13 07:13:34.898297795 +0000 UTC m=+1471.502707243" observedRunningTime="2025-12-13 07:13:35.373309031 +0000 UTC m=+1471.977718499" watchObservedRunningTime="2025-12-13 07:13:35.392047502 +0000 UTC m=+1471.996456950" Dec 13 07:13:35 crc kubenswrapper[4971]: I1213 07:13:35.395899 4971 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell0-db-create-jsdsv"] Dec 13 07:13:35 crc kubenswrapper[4971]: W1213 07:13:35.596205 4971 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod5ffa477d_1f7b_445f_9a97_65a7889e44d4.slice/crio-f99713dddd1de9283ed6e5a394e191b9286980670bc295b43df3341c98e115ce WatchSource:0}: Error finding container f99713dddd1de9283ed6e5a394e191b9286980670bc295b43df3341c98e115ce: Status 404 returned error can't find the container with id f99713dddd1de9283ed6e5a394e191b9286980670bc295b43df3341c98e115ce Dec 13 07:13:35 crc kubenswrapper[4971]: I1213 07:13:35.599025 4971 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell0-ed10-account-create-update-tth4x"] Dec 13 07:13:35 crc kubenswrapper[4971]: I1213 07:13:35.615750 4971 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-api-db-create-ndmqx"] Dec 13 07:13:35 crc kubenswrapper[4971]: I1213 07:13:35.628888 4971 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell1-db-create-4j49j"] Dec 13 07:13:35 crc kubenswrapper[4971]: I1213 07:13:35.780844 4971 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell1-61ad-account-create-update-4ckp9"] Dec 13 07:13:35 crc kubenswrapper[4971]: I1213 07:13:35.784880 4971 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-api-048e-account-create-update-kggpx"] Dec 13 07:13:35 crc kubenswrapper[4971]: W1213 07:13:35.789196 4971 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pode0e9f238_3cb4_43ff_bac3_f1c283d0915f.slice/crio-5d39b4ff17b9db11b25d6aa3b2f1f431f60bc23995cc2c13c034737bbc56bf9d WatchSource:0}: Error finding container 5d39b4ff17b9db11b25d6aa3b2f1f431f60bc23995cc2c13c034737bbc56bf9d: Status 404 returned error can't find the container with id 5d39b4ff17b9db11b25d6aa3b2f1f431f60bc23995cc2c13c034737bbc56bf9d Dec 13 07:13:35 crc kubenswrapper[4971]: W1213 07:13:35.792856 4971 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podedac9227_16d2_48b0_ac2b_3f71c67d6cf8.slice/crio-e2c7368fb200f793c809c627d576ade857994f23b8e99d2eaff75471789826ce WatchSource:0}: Error finding container e2c7368fb200f793c809c627d576ade857994f23b8e99d2eaff75471789826ce: Status 404 returned error can't find the container with id e2c7368fb200f793c809c627d576ade857994f23b8e99d2eaff75471789826ce Dec 13 07:13:35 crc kubenswrapper[4971]: I1213 07:13:35.963704 4971 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ceilometer-0"] Dec 13 07:13:35 crc kubenswrapper[4971]: I1213 07:13:35.964266 4971 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="a8767723-1c4f-41c3-ac4e-988a1d93d56a" containerName="ceilometer-central-agent" containerID="cri-o://9ac009ffef2b0f0f983c49e9d333c68f5bc4ca696b14089108a670ec9c8ae630" gracePeriod=30 Dec 13 07:13:35 crc kubenswrapper[4971]: I1213 07:13:35.964392 4971 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="a8767723-1c4f-41c3-ac4e-988a1d93d56a" containerName="proxy-httpd" containerID="cri-o://3eb143d88d13c5a55a04a05582effc70650cefc1337e365a514f4fabdcf9888f" gracePeriod=30 Dec 13 07:13:35 crc kubenswrapper[4971]: I1213 07:13:35.964432 4971 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="a8767723-1c4f-41c3-ac4e-988a1d93d56a" containerName="sg-core" containerID="cri-o://714d0b6157ede690f30644b9c368321eb015d8734735a3cf3e7f11de2e94d6bb" gracePeriod=30 Dec 13 07:13:35 crc kubenswrapper[4971]: I1213 07:13:35.964467 4971 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="a8767723-1c4f-41c3-ac4e-988a1d93d56a" containerName="ceilometer-notification-agent" containerID="cri-o://3ea374af6edd7bef76b4de25745f39b1b6be9ae48de86284b377702d49b7b32a" gracePeriod=30 Dec 13 07:13:35 crc kubenswrapper[4971]: I1213 07:13:35.975152 4971 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/ceilometer-0" Dec 13 07:13:36 crc kubenswrapper[4971]: I1213 07:13:36.365130 4971 generic.go:334] "Generic (PLEG): container finished" podID="a8767723-1c4f-41c3-ac4e-988a1d93d56a" containerID="714d0b6157ede690f30644b9c368321eb015d8734735a3cf3e7f11de2e94d6bb" exitCode=2 Dec 13 07:13:36 crc kubenswrapper[4971]: I1213 07:13:36.365193 4971 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"a8767723-1c4f-41c3-ac4e-988a1d93d56a","Type":"ContainerDied","Data":"714d0b6157ede690f30644b9c368321eb015d8734735a3cf3e7f11de2e94d6bb"} Dec 13 07:13:36 crc kubenswrapper[4971]: I1213 07:13:36.367198 4971 generic.go:334] "Generic (PLEG): container finished" podID="0ed1405a-8326-4fe9-8fb4-1d97b9e8befb" containerID="cb21d277a6ad833347a698fee9c98b7876b7a07a2de9d2e3630e0994cda10e94" exitCode=0 Dec 13 07:13:36 crc kubenswrapper[4971]: I1213 07:13:36.367218 4971 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-db-create-jsdsv" event={"ID":"0ed1405a-8326-4fe9-8fb4-1d97b9e8befb","Type":"ContainerDied","Data":"cb21d277a6ad833347a698fee9c98b7876b7a07a2de9d2e3630e0994cda10e94"} Dec 13 07:13:36 crc kubenswrapper[4971]: I1213 07:13:36.367251 4971 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-db-create-jsdsv" event={"ID":"0ed1405a-8326-4fe9-8fb4-1d97b9e8befb","Type":"ContainerStarted","Data":"289a384ea2f17437068706371c810954cd10f7b1ef6be88d66070f02da680ba1"} Dec 13 07:13:36 crc kubenswrapper[4971]: I1213 07:13:36.369539 4971 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-db-create-4j49j" event={"ID":"5ba87867-bb90-4587-9012-6db9d2901b10","Type":"ContainerStarted","Data":"3dc5bacbb497492ceb200a8f4079f1fe1919ec2f11414de66931136305d0fa67"} Dec 13 07:13:36 crc kubenswrapper[4971]: I1213 07:13:36.369564 4971 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-db-create-4j49j" event={"ID":"5ba87867-bb90-4587-9012-6db9d2901b10","Type":"ContainerStarted","Data":"d9859141baa96b81e15d05b43318ff87cca36560f34d29225fe32e109f0af4c3"} Dec 13 07:13:36 crc kubenswrapper[4971]: I1213 07:13:36.371670 4971 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-ed10-account-create-update-tth4x" event={"ID":"5ffa477d-1f7b-445f-9a97-65a7889e44d4","Type":"ContainerStarted","Data":"44cc99feda7a93b0ad3cddca4023a75dbf6b44a85ac3a407d60f674701f50cab"} Dec 13 07:13:36 crc kubenswrapper[4971]: I1213 07:13:36.371695 4971 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-ed10-account-create-update-tth4x" event={"ID":"5ffa477d-1f7b-445f-9a97-65a7889e44d4","Type":"ContainerStarted","Data":"f99713dddd1de9283ed6e5a394e191b9286980670bc295b43df3341c98e115ce"} Dec 13 07:13:36 crc kubenswrapper[4971]: I1213 07:13:36.374255 4971 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-61ad-account-create-update-4ckp9" event={"ID":"e0e9f238-3cb4-43ff-bac3-f1c283d0915f","Type":"ContainerStarted","Data":"ab8949e28c39670644fadb5e1c3465a991ef49397ef4b090a05d46af40d1b3a9"} Dec 13 07:13:36 crc kubenswrapper[4971]: I1213 07:13:36.374280 4971 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-61ad-account-create-update-4ckp9" event={"ID":"e0e9f238-3cb4-43ff-bac3-f1c283d0915f","Type":"ContainerStarted","Data":"5d39b4ff17b9db11b25d6aa3b2f1f431f60bc23995cc2c13c034737bbc56bf9d"} Dec 13 07:13:36 crc kubenswrapper[4971]: I1213 07:13:36.375948 4971 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-048e-account-create-update-kggpx" event={"ID":"edac9227-16d2-48b0-ac2b-3f71c67d6cf8","Type":"ContainerStarted","Data":"9509f0f8e726726fa1500e026105ca7cecd8caa43de3dcf3443517afda368ef9"} Dec 13 07:13:36 crc kubenswrapper[4971]: I1213 07:13:36.375971 4971 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-048e-account-create-update-kggpx" event={"ID":"edac9227-16d2-48b0-ac2b-3f71c67d6cf8","Type":"ContainerStarted","Data":"e2c7368fb200f793c809c627d576ade857994f23b8e99d2eaff75471789826ce"} Dec 13 07:13:36 crc kubenswrapper[4971]: I1213 07:13:36.378131 4971 generic.go:334] "Generic (PLEG): container finished" podID="8b60a7ee-8d5e-4bf2-8d64-161aeced2357" containerID="0035d5bc04fb0f44fabb4f7127f141f9fb4c42242812ceedc881646754b08540" exitCode=0 Dec 13 07:13:36 crc kubenswrapper[4971]: I1213 07:13:36.378576 4971 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-db-create-ndmqx" event={"ID":"8b60a7ee-8d5e-4bf2-8d64-161aeced2357","Type":"ContainerDied","Data":"0035d5bc04fb0f44fabb4f7127f141f9fb4c42242812ceedc881646754b08540"} Dec 13 07:13:36 crc kubenswrapper[4971]: I1213 07:13:36.378598 4971 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-db-create-ndmqx" event={"ID":"8b60a7ee-8d5e-4bf2-8d64-161aeced2357","Type":"ContainerStarted","Data":"4181baad9f77aa467e6a67724ba935b9d2e419ac082f1d4a6df59d844dd5269f"} Dec 13 07:13:36 crc kubenswrapper[4971]: I1213 07:13:36.412672 4971 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-cell1-61ad-account-create-update-4ckp9" podStartSLOduration=6.412652459 podStartE2EDuration="6.412652459s" podCreationTimestamp="2025-12-13 07:13:30 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-13 07:13:36.411424289 +0000 UTC m=+1473.015833737" watchObservedRunningTime="2025-12-13 07:13:36.412652459 +0000 UTC m=+1473.017061907" Dec 13 07:13:36 crc kubenswrapper[4971]: I1213 07:13:36.430757 4971 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-cell0-ed10-account-create-update-tth4x" podStartSLOduration=7.430736474 podStartE2EDuration="7.430736474s" podCreationTimestamp="2025-12-13 07:13:29 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-13 07:13:36.426590472 +0000 UTC m=+1473.030999920" watchObservedRunningTime="2025-12-13 07:13:36.430736474 +0000 UTC m=+1473.035145922" Dec 13 07:13:36 crc kubenswrapper[4971]: I1213 07:13:36.469694 4971 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-api-048e-account-create-update-kggpx" podStartSLOduration=7.469675803 podStartE2EDuration="7.469675803s" podCreationTimestamp="2025-12-13 07:13:29 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-13 07:13:36.462070385 +0000 UTC m=+1473.066479833" watchObservedRunningTime="2025-12-13 07:13:36.469675803 +0000 UTC m=+1473.074085251" Dec 13 07:13:37 crc kubenswrapper[4971]: I1213 07:13:37.388884 4971 generic.go:334] "Generic (PLEG): container finished" podID="edac9227-16d2-48b0-ac2b-3f71c67d6cf8" containerID="9509f0f8e726726fa1500e026105ca7cecd8caa43de3dcf3443517afda368ef9" exitCode=0 Dec 13 07:13:37 crc kubenswrapper[4971]: I1213 07:13:37.390014 4971 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-048e-account-create-update-kggpx" event={"ID":"edac9227-16d2-48b0-ac2b-3f71c67d6cf8","Type":"ContainerDied","Data":"9509f0f8e726726fa1500e026105ca7cecd8caa43de3dcf3443517afda368ef9"} Dec 13 07:13:37 crc kubenswrapper[4971]: I1213 07:13:37.393632 4971 generic.go:334] "Generic (PLEG): container finished" podID="a8767723-1c4f-41c3-ac4e-988a1d93d56a" containerID="3eb143d88d13c5a55a04a05582effc70650cefc1337e365a514f4fabdcf9888f" exitCode=0 Dec 13 07:13:37 crc kubenswrapper[4971]: I1213 07:13:37.393775 4971 generic.go:334] "Generic (PLEG): container finished" podID="a8767723-1c4f-41c3-ac4e-988a1d93d56a" containerID="9ac009ffef2b0f0f983c49e9d333c68f5bc4ca696b14089108a670ec9c8ae630" exitCode=0 Dec 13 07:13:37 crc kubenswrapper[4971]: I1213 07:13:37.393720 4971 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"a8767723-1c4f-41c3-ac4e-988a1d93d56a","Type":"ContainerDied","Data":"3eb143d88d13c5a55a04a05582effc70650cefc1337e365a514f4fabdcf9888f"} Dec 13 07:13:37 crc kubenswrapper[4971]: I1213 07:13:37.393990 4971 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"a8767723-1c4f-41c3-ac4e-988a1d93d56a","Type":"ContainerDied","Data":"9ac009ffef2b0f0f983c49e9d333c68f5bc4ca696b14089108a670ec9c8ae630"} Dec 13 07:13:37 crc kubenswrapper[4971]: I1213 07:13:37.395702 4971 generic.go:334] "Generic (PLEG): container finished" podID="5ba87867-bb90-4587-9012-6db9d2901b10" containerID="3dc5bacbb497492ceb200a8f4079f1fe1919ec2f11414de66931136305d0fa67" exitCode=0 Dec 13 07:13:37 crc kubenswrapper[4971]: I1213 07:13:37.395868 4971 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-db-create-4j49j" event={"ID":"5ba87867-bb90-4587-9012-6db9d2901b10","Type":"ContainerDied","Data":"3dc5bacbb497492ceb200a8f4079f1fe1919ec2f11414de66931136305d0fa67"} Dec 13 07:13:37 crc kubenswrapper[4971]: I1213 07:13:37.398834 4971 generic.go:334] "Generic (PLEG): container finished" podID="5ffa477d-1f7b-445f-9a97-65a7889e44d4" containerID="44cc99feda7a93b0ad3cddca4023a75dbf6b44a85ac3a407d60f674701f50cab" exitCode=0 Dec 13 07:13:37 crc kubenswrapper[4971]: I1213 07:13:37.398987 4971 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-ed10-account-create-update-tth4x" event={"ID":"5ffa477d-1f7b-445f-9a97-65a7889e44d4","Type":"ContainerDied","Data":"44cc99feda7a93b0ad3cddca4023a75dbf6b44a85ac3a407d60f674701f50cab"} Dec 13 07:13:37 crc kubenswrapper[4971]: I1213 07:13:37.402490 4971 generic.go:334] "Generic (PLEG): container finished" podID="e0e9f238-3cb4-43ff-bac3-f1c283d0915f" containerID="ab8949e28c39670644fadb5e1c3465a991ef49397ef4b090a05d46af40d1b3a9" exitCode=0 Dec 13 07:13:37 crc kubenswrapper[4971]: I1213 07:13:37.402870 4971 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-61ad-account-create-update-4ckp9" event={"ID":"e0e9f238-3cb4-43ff-bac3-f1c283d0915f","Type":"ContainerDied","Data":"ab8949e28c39670644fadb5e1c3465a991ef49397ef4b090a05d46af40d1b3a9"} Dec 13 07:13:37 crc kubenswrapper[4971]: I1213 07:13:37.856686 4971 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-db-create-4j49j" Dec 13 07:13:37 crc kubenswrapper[4971]: I1213 07:13:37.946739 4971 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/5ba87867-bb90-4587-9012-6db9d2901b10-operator-scripts\") pod \"5ba87867-bb90-4587-9012-6db9d2901b10\" (UID: \"5ba87867-bb90-4587-9012-6db9d2901b10\") " Dec 13 07:13:37 crc kubenswrapper[4971]: I1213 07:13:37.946928 4971 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-frnhv\" (UniqueName: \"kubernetes.io/projected/5ba87867-bb90-4587-9012-6db9d2901b10-kube-api-access-frnhv\") pod \"5ba87867-bb90-4587-9012-6db9d2901b10\" (UID: \"5ba87867-bb90-4587-9012-6db9d2901b10\") " Dec 13 07:13:37 crc kubenswrapper[4971]: I1213 07:13:37.947207 4971 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/5ba87867-bb90-4587-9012-6db9d2901b10-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "5ba87867-bb90-4587-9012-6db9d2901b10" (UID: "5ba87867-bb90-4587-9012-6db9d2901b10"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 13 07:13:37 crc kubenswrapper[4971]: I1213 07:13:37.947452 4971 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/5ba87867-bb90-4587-9012-6db9d2901b10-operator-scripts\") on node \"crc\" DevicePath \"\"" Dec 13 07:13:37 crc kubenswrapper[4971]: I1213 07:13:37.952270 4971 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/5ba87867-bb90-4587-9012-6db9d2901b10-kube-api-access-frnhv" (OuterVolumeSpecName: "kube-api-access-frnhv") pod "5ba87867-bb90-4587-9012-6db9d2901b10" (UID: "5ba87867-bb90-4587-9012-6db9d2901b10"). InnerVolumeSpecName "kube-api-access-frnhv". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 13 07:13:38 crc kubenswrapper[4971]: I1213 07:13:38.046913 4971 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-db-create-jsdsv" Dec 13 07:13:38 crc kubenswrapper[4971]: I1213 07:13:38.049053 4971 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-frnhv\" (UniqueName: \"kubernetes.io/projected/5ba87867-bb90-4587-9012-6db9d2901b10-kube-api-access-frnhv\") on node \"crc\" DevicePath \"\"" Dec 13 07:13:38 crc kubenswrapper[4971]: I1213 07:13:38.052557 4971 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-db-create-ndmqx" Dec 13 07:13:38 crc kubenswrapper[4971]: I1213 07:13:38.150072 4971 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-8nxsg\" (UniqueName: \"kubernetes.io/projected/8b60a7ee-8d5e-4bf2-8d64-161aeced2357-kube-api-access-8nxsg\") pod \"8b60a7ee-8d5e-4bf2-8d64-161aeced2357\" (UID: \"8b60a7ee-8d5e-4bf2-8d64-161aeced2357\") " Dec 13 07:13:38 crc kubenswrapper[4971]: I1213 07:13:38.150278 4971 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/8b60a7ee-8d5e-4bf2-8d64-161aeced2357-operator-scripts\") pod \"8b60a7ee-8d5e-4bf2-8d64-161aeced2357\" (UID: \"8b60a7ee-8d5e-4bf2-8d64-161aeced2357\") " Dec 13 07:13:38 crc kubenswrapper[4971]: I1213 07:13:38.150348 4971 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/0ed1405a-8326-4fe9-8fb4-1d97b9e8befb-operator-scripts\") pod \"0ed1405a-8326-4fe9-8fb4-1d97b9e8befb\" (UID: \"0ed1405a-8326-4fe9-8fb4-1d97b9e8befb\") " Dec 13 07:13:38 crc kubenswrapper[4971]: I1213 07:13:38.150423 4971 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-tvh6r\" (UniqueName: \"kubernetes.io/projected/0ed1405a-8326-4fe9-8fb4-1d97b9e8befb-kube-api-access-tvh6r\") pod \"0ed1405a-8326-4fe9-8fb4-1d97b9e8befb\" (UID: \"0ed1405a-8326-4fe9-8fb4-1d97b9e8befb\") " Dec 13 07:13:38 crc kubenswrapper[4971]: I1213 07:13:38.150928 4971 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/8b60a7ee-8d5e-4bf2-8d64-161aeced2357-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "8b60a7ee-8d5e-4bf2-8d64-161aeced2357" (UID: "8b60a7ee-8d5e-4bf2-8d64-161aeced2357"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 13 07:13:38 crc kubenswrapper[4971]: I1213 07:13:38.150927 4971 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/0ed1405a-8326-4fe9-8fb4-1d97b9e8befb-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "0ed1405a-8326-4fe9-8fb4-1d97b9e8befb" (UID: "0ed1405a-8326-4fe9-8fb4-1d97b9e8befb"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 13 07:13:38 crc kubenswrapper[4971]: I1213 07:13:38.151411 4971 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/0ed1405a-8326-4fe9-8fb4-1d97b9e8befb-operator-scripts\") on node \"crc\" DevicePath \"\"" Dec 13 07:13:38 crc kubenswrapper[4971]: I1213 07:13:38.151436 4971 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/8b60a7ee-8d5e-4bf2-8d64-161aeced2357-operator-scripts\") on node \"crc\" DevicePath \"\"" Dec 13 07:13:38 crc kubenswrapper[4971]: I1213 07:13:38.154840 4971 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/0ed1405a-8326-4fe9-8fb4-1d97b9e8befb-kube-api-access-tvh6r" (OuterVolumeSpecName: "kube-api-access-tvh6r") pod "0ed1405a-8326-4fe9-8fb4-1d97b9e8befb" (UID: "0ed1405a-8326-4fe9-8fb4-1d97b9e8befb"). InnerVolumeSpecName "kube-api-access-tvh6r". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 13 07:13:38 crc kubenswrapper[4971]: I1213 07:13:38.154988 4971 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/8b60a7ee-8d5e-4bf2-8d64-161aeced2357-kube-api-access-8nxsg" (OuterVolumeSpecName: "kube-api-access-8nxsg") pod "8b60a7ee-8d5e-4bf2-8d64-161aeced2357" (UID: "8b60a7ee-8d5e-4bf2-8d64-161aeced2357"). InnerVolumeSpecName "kube-api-access-8nxsg". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 13 07:13:38 crc kubenswrapper[4971]: I1213 07:13:38.253637 4971 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-8nxsg\" (UniqueName: \"kubernetes.io/projected/8b60a7ee-8d5e-4bf2-8d64-161aeced2357-kube-api-access-8nxsg\") on node \"crc\" DevicePath \"\"" Dec 13 07:13:38 crc kubenswrapper[4971]: I1213 07:13:38.253690 4971 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-tvh6r\" (UniqueName: \"kubernetes.io/projected/0ed1405a-8326-4fe9-8fb4-1d97b9e8befb-kube-api-access-tvh6r\") on node \"crc\" DevicePath \"\"" Dec 13 07:13:38 crc kubenswrapper[4971]: I1213 07:13:38.412651 4971 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-db-create-4j49j" event={"ID":"5ba87867-bb90-4587-9012-6db9d2901b10","Type":"ContainerDied","Data":"d9859141baa96b81e15d05b43318ff87cca36560f34d29225fe32e109f0af4c3"} Dec 13 07:13:38 crc kubenswrapper[4971]: I1213 07:13:38.412700 4971 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="d9859141baa96b81e15d05b43318ff87cca36560f34d29225fe32e109f0af4c3" Dec 13 07:13:38 crc kubenswrapper[4971]: I1213 07:13:38.412762 4971 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-db-create-4j49j" Dec 13 07:13:38 crc kubenswrapper[4971]: I1213 07:13:38.416741 4971 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-db-create-ndmqx" event={"ID":"8b60a7ee-8d5e-4bf2-8d64-161aeced2357","Type":"ContainerDied","Data":"4181baad9f77aa467e6a67724ba935b9d2e419ac082f1d4a6df59d844dd5269f"} Dec 13 07:13:38 crc kubenswrapper[4971]: I1213 07:13:38.416790 4971 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="4181baad9f77aa467e6a67724ba935b9d2e419ac082f1d4a6df59d844dd5269f" Dec 13 07:13:38 crc kubenswrapper[4971]: I1213 07:13:38.416852 4971 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-db-create-ndmqx" Dec 13 07:13:38 crc kubenswrapper[4971]: I1213 07:13:38.425862 4971 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-db-create-jsdsv" Dec 13 07:13:38 crc kubenswrapper[4971]: I1213 07:13:38.429751 4971 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-db-create-jsdsv" event={"ID":"0ed1405a-8326-4fe9-8fb4-1d97b9e8befb","Type":"ContainerDied","Data":"289a384ea2f17437068706371c810954cd10f7b1ef6be88d66070f02da680ba1"} Dec 13 07:13:38 crc kubenswrapper[4971]: I1213 07:13:38.429799 4971 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="289a384ea2f17437068706371c810954cd10f7b1ef6be88d66070f02da680ba1" Dec 13 07:13:38 crc kubenswrapper[4971]: I1213 07:13:38.753423 4971 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-048e-account-create-update-kggpx" Dec 13 07:13:38 crc kubenswrapper[4971]: I1213 07:13:38.866794 4971 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/edac9227-16d2-48b0-ac2b-3f71c67d6cf8-operator-scripts\") pod \"edac9227-16d2-48b0-ac2b-3f71c67d6cf8\" (UID: \"edac9227-16d2-48b0-ac2b-3f71c67d6cf8\") " Dec 13 07:13:38 crc kubenswrapper[4971]: I1213 07:13:38.866959 4971 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-4w6kr\" (UniqueName: \"kubernetes.io/projected/edac9227-16d2-48b0-ac2b-3f71c67d6cf8-kube-api-access-4w6kr\") pod \"edac9227-16d2-48b0-ac2b-3f71c67d6cf8\" (UID: \"edac9227-16d2-48b0-ac2b-3f71c67d6cf8\") " Dec 13 07:13:38 crc kubenswrapper[4971]: I1213 07:13:38.867447 4971 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/edac9227-16d2-48b0-ac2b-3f71c67d6cf8-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "edac9227-16d2-48b0-ac2b-3f71c67d6cf8" (UID: "edac9227-16d2-48b0-ac2b-3f71c67d6cf8"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 13 07:13:38 crc kubenswrapper[4971]: I1213 07:13:38.874106 4971 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/edac9227-16d2-48b0-ac2b-3f71c67d6cf8-kube-api-access-4w6kr" (OuterVolumeSpecName: "kube-api-access-4w6kr") pod "edac9227-16d2-48b0-ac2b-3f71c67d6cf8" (UID: "edac9227-16d2-48b0-ac2b-3f71c67d6cf8"). InnerVolumeSpecName "kube-api-access-4w6kr". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 13 07:13:38 crc kubenswrapper[4971]: I1213 07:13:38.941543 4971 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-ed10-account-create-update-tth4x" Dec 13 07:13:38 crc kubenswrapper[4971]: I1213 07:13:38.948211 4971 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-61ad-account-create-update-4ckp9" Dec 13 07:13:38 crc kubenswrapper[4971]: I1213 07:13:38.971134 4971 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-mmc9b\" (UniqueName: \"kubernetes.io/projected/5ffa477d-1f7b-445f-9a97-65a7889e44d4-kube-api-access-mmc9b\") pod \"5ffa477d-1f7b-445f-9a97-65a7889e44d4\" (UID: \"5ffa477d-1f7b-445f-9a97-65a7889e44d4\") " Dec 13 07:13:38 crc kubenswrapper[4971]: I1213 07:13:38.971259 4971 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/5ffa477d-1f7b-445f-9a97-65a7889e44d4-operator-scripts\") pod \"5ffa477d-1f7b-445f-9a97-65a7889e44d4\" (UID: \"5ffa477d-1f7b-445f-9a97-65a7889e44d4\") " Dec 13 07:13:38 crc kubenswrapper[4971]: I1213 07:13:38.971307 4971 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-ffkdr\" (UniqueName: \"kubernetes.io/projected/e0e9f238-3cb4-43ff-bac3-f1c283d0915f-kube-api-access-ffkdr\") pod \"e0e9f238-3cb4-43ff-bac3-f1c283d0915f\" (UID: \"e0e9f238-3cb4-43ff-bac3-f1c283d0915f\") " Dec 13 07:13:38 crc kubenswrapper[4971]: I1213 07:13:38.971345 4971 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/e0e9f238-3cb4-43ff-bac3-f1c283d0915f-operator-scripts\") pod \"e0e9f238-3cb4-43ff-bac3-f1c283d0915f\" (UID: \"e0e9f238-3cb4-43ff-bac3-f1c283d0915f\") " Dec 13 07:13:38 crc kubenswrapper[4971]: I1213 07:13:38.971898 4971 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-4w6kr\" (UniqueName: \"kubernetes.io/projected/edac9227-16d2-48b0-ac2b-3f71c67d6cf8-kube-api-access-4w6kr\") on node \"crc\" DevicePath \"\"" Dec 13 07:13:38 crc kubenswrapper[4971]: I1213 07:13:38.971905 4971 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/5ffa477d-1f7b-445f-9a97-65a7889e44d4-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "5ffa477d-1f7b-445f-9a97-65a7889e44d4" (UID: "5ffa477d-1f7b-445f-9a97-65a7889e44d4"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 13 07:13:38 crc kubenswrapper[4971]: I1213 07:13:38.971924 4971 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/edac9227-16d2-48b0-ac2b-3f71c67d6cf8-operator-scripts\") on node \"crc\" DevicePath \"\"" Dec 13 07:13:38 crc kubenswrapper[4971]: I1213 07:13:38.972193 4971 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/e0e9f238-3cb4-43ff-bac3-f1c283d0915f-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "e0e9f238-3cb4-43ff-bac3-f1c283d0915f" (UID: "e0e9f238-3cb4-43ff-bac3-f1c283d0915f"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 13 07:13:38 crc kubenswrapper[4971]: I1213 07:13:38.974703 4971 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/e0e9f238-3cb4-43ff-bac3-f1c283d0915f-kube-api-access-ffkdr" (OuterVolumeSpecName: "kube-api-access-ffkdr") pod "e0e9f238-3cb4-43ff-bac3-f1c283d0915f" (UID: "e0e9f238-3cb4-43ff-bac3-f1c283d0915f"). InnerVolumeSpecName "kube-api-access-ffkdr". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 13 07:13:38 crc kubenswrapper[4971]: I1213 07:13:38.982042 4971 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/5ffa477d-1f7b-445f-9a97-65a7889e44d4-kube-api-access-mmc9b" (OuterVolumeSpecName: "kube-api-access-mmc9b") pod "5ffa477d-1f7b-445f-9a97-65a7889e44d4" (UID: "5ffa477d-1f7b-445f-9a97-65a7889e44d4"). InnerVolumeSpecName "kube-api-access-mmc9b". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 13 07:13:39 crc kubenswrapper[4971]: I1213 07:13:39.073360 4971 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-mmc9b\" (UniqueName: \"kubernetes.io/projected/5ffa477d-1f7b-445f-9a97-65a7889e44d4-kube-api-access-mmc9b\") on node \"crc\" DevicePath \"\"" Dec 13 07:13:39 crc kubenswrapper[4971]: I1213 07:13:39.073395 4971 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/5ffa477d-1f7b-445f-9a97-65a7889e44d4-operator-scripts\") on node \"crc\" DevicePath \"\"" Dec 13 07:13:39 crc kubenswrapper[4971]: I1213 07:13:39.073405 4971 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-ffkdr\" (UniqueName: \"kubernetes.io/projected/e0e9f238-3cb4-43ff-bac3-f1c283d0915f-kube-api-access-ffkdr\") on node \"crc\" DevicePath \"\"" Dec 13 07:13:39 crc kubenswrapper[4971]: I1213 07:13:39.073416 4971 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/e0e9f238-3cb4-43ff-bac3-f1c283d0915f-operator-scripts\") on node \"crc\" DevicePath \"\"" Dec 13 07:13:39 crc kubenswrapper[4971]: I1213 07:13:39.440972 4971 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-048e-account-create-update-kggpx" Dec 13 07:13:39 crc kubenswrapper[4971]: I1213 07:13:39.440974 4971 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-048e-account-create-update-kggpx" event={"ID":"edac9227-16d2-48b0-ac2b-3f71c67d6cf8","Type":"ContainerDied","Data":"e2c7368fb200f793c809c627d576ade857994f23b8e99d2eaff75471789826ce"} Dec 13 07:13:39 crc kubenswrapper[4971]: I1213 07:13:39.441403 4971 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="e2c7368fb200f793c809c627d576ade857994f23b8e99d2eaff75471789826ce" Dec 13 07:13:39 crc kubenswrapper[4971]: I1213 07:13:39.443595 4971 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-ed10-account-create-update-tth4x" event={"ID":"5ffa477d-1f7b-445f-9a97-65a7889e44d4","Type":"ContainerDied","Data":"f99713dddd1de9283ed6e5a394e191b9286980670bc295b43df3341c98e115ce"} Dec 13 07:13:39 crc kubenswrapper[4971]: I1213 07:13:39.443642 4971 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="f99713dddd1de9283ed6e5a394e191b9286980670bc295b43df3341c98e115ce" Dec 13 07:13:39 crc kubenswrapper[4971]: I1213 07:13:39.443645 4971 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-ed10-account-create-update-tth4x" Dec 13 07:13:39 crc kubenswrapper[4971]: I1213 07:13:39.445610 4971 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-61ad-account-create-update-4ckp9" event={"ID":"e0e9f238-3cb4-43ff-bac3-f1c283d0915f","Type":"ContainerDied","Data":"5d39b4ff17b9db11b25d6aa3b2f1f431f60bc23995cc2c13c034737bbc56bf9d"} Dec 13 07:13:39 crc kubenswrapper[4971]: I1213 07:13:39.445629 4971 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="5d39b4ff17b9db11b25d6aa3b2f1f431f60bc23995cc2c13c034737bbc56bf9d" Dec 13 07:13:39 crc kubenswrapper[4971]: I1213 07:13:39.445660 4971 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-61ad-account-create-update-4ckp9" Dec 13 07:13:40 crc kubenswrapper[4971]: I1213 07:13:40.462163 4971 generic.go:334] "Generic (PLEG): container finished" podID="772eedab-7c62-454b-b722-ee453b39e89f" containerID="2074da306f3c927fa0e63c982fab7574e67b1f06c38e55ba7dc406c0d32df42a" exitCode=137 Dec 13 07:13:40 crc kubenswrapper[4971]: I1213 07:13:40.462260 4971 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/horizon-8699dc887d-sl2h6" event={"ID":"772eedab-7c62-454b-b722-ee453b39e89f","Type":"ContainerDied","Data":"2074da306f3c927fa0e63c982fab7574e67b1f06c38e55ba7dc406c0d32df42a"} Dec 13 07:13:40 crc kubenswrapper[4971]: I1213 07:13:40.655555 4971 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/horizon-8699dc887d-sl2h6" Dec 13 07:13:40 crc kubenswrapper[4971]: I1213 07:13:40.699814 4971 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"horizon-secret-key\" (UniqueName: \"kubernetes.io/secret/772eedab-7c62-454b-b722-ee453b39e89f-horizon-secret-key\") pod \"772eedab-7c62-454b-b722-ee453b39e89f\" (UID: \"772eedab-7c62-454b-b722-ee453b39e89f\") " Dec 13 07:13:40 crc kubenswrapper[4971]: I1213 07:13:40.700116 4971 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-6sss5\" (UniqueName: \"kubernetes.io/projected/772eedab-7c62-454b-b722-ee453b39e89f-kube-api-access-6sss5\") pod \"772eedab-7c62-454b-b722-ee453b39e89f\" (UID: \"772eedab-7c62-454b-b722-ee453b39e89f\") " Dec 13 07:13:40 crc kubenswrapper[4971]: I1213 07:13:40.700307 4971 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/772eedab-7c62-454b-b722-ee453b39e89f-combined-ca-bundle\") pod \"772eedab-7c62-454b-b722-ee453b39e89f\" (UID: \"772eedab-7c62-454b-b722-ee453b39e89f\") " Dec 13 07:13:40 crc kubenswrapper[4971]: I1213 07:13:40.700482 4971 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"horizon-tls-certs\" (UniqueName: \"kubernetes.io/secret/772eedab-7c62-454b-b722-ee453b39e89f-horizon-tls-certs\") pod \"772eedab-7c62-454b-b722-ee453b39e89f\" (UID: \"772eedab-7c62-454b-b722-ee453b39e89f\") " Dec 13 07:13:40 crc kubenswrapper[4971]: I1213 07:13:40.700640 4971 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/772eedab-7c62-454b-b722-ee453b39e89f-config-data\") pod \"772eedab-7c62-454b-b722-ee453b39e89f\" (UID: \"772eedab-7c62-454b-b722-ee453b39e89f\") " Dec 13 07:13:40 crc kubenswrapper[4971]: I1213 07:13:40.700854 4971 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/772eedab-7c62-454b-b722-ee453b39e89f-scripts\") pod \"772eedab-7c62-454b-b722-ee453b39e89f\" (UID: \"772eedab-7c62-454b-b722-ee453b39e89f\") " Dec 13 07:13:40 crc kubenswrapper[4971]: I1213 07:13:40.700985 4971 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/772eedab-7c62-454b-b722-ee453b39e89f-logs\") pod \"772eedab-7c62-454b-b722-ee453b39e89f\" (UID: \"772eedab-7c62-454b-b722-ee453b39e89f\") " Dec 13 07:13:40 crc kubenswrapper[4971]: I1213 07:13:40.701425 4971 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/772eedab-7c62-454b-b722-ee453b39e89f-logs" (OuterVolumeSpecName: "logs") pod "772eedab-7c62-454b-b722-ee453b39e89f" (UID: "772eedab-7c62-454b-b722-ee453b39e89f"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 13 07:13:40 crc kubenswrapper[4971]: I1213 07:13:40.701676 4971 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/772eedab-7c62-454b-b722-ee453b39e89f-logs\") on node \"crc\" DevicePath \"\"" Dec 13 07:13:40 crc kubenswrapper[4971]: I1213 07:13:40.712677 4971 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/772eedab-7c62-454b-b722-ee453b39e89f-horizon-secret-key" (OuterVolumeSpecName: "horizon-secret-key") pod "772eedab-7c62-454b-b722-ee453b39e89f" (UID: "772eedab-7c62-454b-b722-ee453b39e89f"). InnerVolumeSpecName "horizon-secret-key". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 13 07:13:40 crc kubenswrapper[4971]: I1213 07:13:40.723091 4971 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/772eedab-7c62-454b-b722-ee453b39e89f-kube-api-access-6sss5" (OuterVolumeSpecName: "kube-api-access-6sss5") pod "772eedab-7c62-454b-b722-ee453b39e89f" (UID: "772eedab-7c62-454b-b722-ee453b39e89f"). InnerVolumeSpecName "kube-api-access-6sss5". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 13 07:13:40 crc kubenswrapper[4971]: I1213 07:13:40.723940 4971 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/772eedab-7c62-454b-b722-ee453b39e89f-scripts" (OuterVolumeSpecName: "scripts") pod "772eedab-7c62-454b-b722-ee453b39e89f" (UID: "772eedab-7c62-454b-b722-ee453b39e89f"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 13 07:13:40 crc kubenswrapper[4971]: I1213 07:13:40.727462 4971 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/772eedab-7c62-454b-b722-ee453b39e89f-config-data" (OuterVolumeSpecName: "config-data") pod "772eedab-7c62-454b-b722-ee453b39e89f" (UID: "772eedab-7c62-454b-b722-ee453b39e89f"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 13 07:13:40 crc kubenswrapper[4971]: I1213 07:13:40.748093 4971 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/772eedab-7c62-454b-b722-ee453b39e89f-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "772eedab-7c62-454b-b722-ee453b39e89f" (UID: "772eedab-7c62-454b-b722-ee453b39e89f"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 13 07:13:40 crc kubenswrapper[4971]: I1213 07:13:40.754714 4971 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/772eedab-7c62-454b-b722-ee453b39e89f-horizon-tls-certs" (OuterVolumeSpecName: "horizon-tls-certs") pod "772eedab-7c62-454b-b722-ee453b39e89f" (UID: "772eedab-7c62-454b-b722-ee453b39e89f"). InnerVolumeSpecName "horizon-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 13 07:13:40 crc kubenswrapper[4971]: I1213 07:13:40.803253 4971 reconciler_common.go:293] "Volume detached for volume \"horizon-tls-certs\" (UniqueName: \"kubernetes.io/secret/772eedab-7c62-454b-b722-ee453b39e89f-horizon-tls-certs\") on node \"crc\" DevicePath \"\"" Dec 13 07:13:40 crc kubenswrapper[4971]: I1213 07:13:40.803286 4971 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/772eedab-7c62-454b-b722-ee453b39e89f-config-data\") on node \"crc\" DevicePath \"\"" Dec 13 07:13:40 crc kubenswrapper[4971]: I1213 07:13:40.803298 4971 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/772eedab-7c62-454b-b722-ee453b39e89f-scripts\") on node \"crc\" DevicePath \"\"" Dec 13 07:13:40 crc kubenswrapper[4971]: I1213 07:13:40.803308 4971 reconciler_common.go:293] "Volume detached for volume \"horizon-secret-key\" (UniqueName: \"kubernetes.io/secret/772eedab-7c62-454b-b722-ee453b39e89f-horizon-secret-key\") on node \"crc\" DevicePath \"\"" Dec 13 07:13:40 crc kubenswrapper[4971]: I1213 07:13:40.803319 4971 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-6sss5\" (UniqueName: \"kubernetes.io/projected/772eedab-7c62-454b-b722-ee453b39e89f-kube-api-access-6sss5\") on node \"crc\" DevicePath \"\"" Dec 13 07:13:40 crc kubenswrapper[4971]: I1213 07:13:40.803331 4971 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/772eedab-7c62-454b-b722-ee453b39e89f-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 13 07:13:41 crc kubenswrapper[4971]: I1213 07:13:41.260256 4971 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/kube-state-metrics-0"] Dec 13 07:13:41 crc kubenswrapper[4971]: I1213 07:13:41.260761 4971 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/kube-state-metrics-0" podUID="2bab8c0a-bd76-4265-8355-8df0c6fce6c3" containerName="kube-state-metrics" containerID="cri-o://7a965470ccb73744ddf8839144cea274fc2632c1baaa2c97193772f8c5e14440" gracePeriod=30 Dec 13 07:13:41 crc kubenswrapper[4971]: I1213 07:13:41.423274 4971 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Dec 13 07:13:41 crc kubenswrapper[4971]: I1213 07:13:41.479131 4971 generic.go:334] "Generic (PLEG): container finished" podID="a8767723-1c4f-41c3-ac4e-988a1d93d56a" containerID="3ea374af6edd7bef76b4de25745f39b1b6be9ae48de86284b377702d49b7b32a" exitCode=0 Dec 13 07:13:41 crc kubenswrapper[4971]: I1213 07:13:41.479240 4971 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"a8767723-1c4f-41c3-ac4e-988a1d93d56a","Type":"ContainerDied","Data":"3ea374af6edd7bef76b4de25745f39b1b6be9ae48de86284b377702d49b7b32a"} Dec 13 07:13:41 crc kubenswrapper[4971]: I1213 07:13:41.479277 4971 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"a8767723-1c4f-41c3-ac4e-988a1d93d56a","Type":"ContainerDied","Data":"8c4779bbd343458c9371720b54076b50eb9297ef3e09c1c5758959afd81be4d5"} Dec 13 07:13:41 crc kubenswrapper[4971]: I1213 07:13:41.479312 4971 scope.go:117] "RemoveContainer" containerID="3eb143d88d13c5a55a04a05582effc70650cefc1337e365a514f4fabdcf9888f" Dec 13 07:13:41 crc kubenswrapper[4971]: I1213 07:13:41.479485 4971 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Dec 13 07:13:41 crc kubenswrapper[4971]: I1213 07:13:41.483571 4971 generic.go:334] "Generic (PLEG): container finished" podID="2bab8c0a-bd76-4265-8355-8df0c6fce6c3" containerID="7a965470ccb73744ddf8839144cea274fc2632c1baaa2c97193772f8c5e14440" exitCode=2 Dec 13 07:13:41 crc kubenswrapper[4971]: I1213 07:13:41.483660 4971 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/kube-state-metrics-0" event={"ID":"2bab8c0a-bd76-4265-8355-8df0c6fce6c3","Type":"ContainerDied","Data":"7a965470ccb73744ddf8839144cea274fc2632c1baaa2c97193772f8c5e14440"} Dec 13 07:13:41 crc kubenswrapper[4971]: I1213 07:13:41.490471 4971 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/horizon-8699dc887d-sl2h6" event={"ID":"772eedab-7c62-454b-b722-ee453b39e89f","Type":"ContainerDied","Data":"e1c30d274e8d692e18e387041e360c7599ed69b5331a9256cef5045f972a15e3"} Dec 13 07:13:41 crc kubenswrapper[4971]: I1213 07:13:41.490604 4971 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/horizon-8699dc887d-sl2h6" Dec 13 07:13:41 crc kubenswrapper[4971]: I1213 07:13:41.500482 4971 scope.go:117] "RemoveContainer" containerID="714d0b6157ede690f30644b9c368321eb015d8734735a3cf3e7f11de2e94d6bb" Dec 13 07:13:41 crc kubenswrapper[4971]: I1213 07:13:41.517050 4971 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/a8767723-1c4f-41c3-ac4e-988a1d93d56a-run-httpd\") pod \"a8767723-1c4f-41c3-ac4e-988a1d93d56a\" (UID: \"a8767723-1c4f-41c3-ac4e-988a1d93d56a\") " Dec 13 07:13:41 crc kubenswrapper[4971]: I1213 07:13:41.517097 4971 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/a8767723-1c4f-41c3-ac4e-988a1d93d56a-combined-ca-bundle\") pod \"a8767723-1c4f-41c3-ac4e-988a1d93d56a\" (UID: \"a8767723-1c4f-41c3-ac4e-988a1d93d56a\") " Dec 13 07:13:41 crc kubenswrapper[4971]: I1213 07:13:41.517423 4971 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/a8767723-1c4f-41c3-ac4e-988a1d93d56a-log-httpd\") pod \"a8767723-1c4f-41c3-ac4e-988a1d93d56a\" (UID: \"a8767723-1c4f-41c3-ac4e-988a1d93d56a\") " Dec 13 07:13:41 crc kubenswrapper[4971]: I1213 07:13:41.517456 4971 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-6hh4c\" (UniqueName: \"kubernetes.io/projected/a8767723-1c4f-41c3-ac4e-988a1d93d56a-kube-api-access-6hh4c\") pod \"a8767723-1c4f-41c3-ac4e-988a1d93d56a\" (UID: \"a8767723-1c4f-41c3-ac4e-988a1d93d56a\") " Dec 13 07:13:41 crc kubenswrapper[4971]: I1213 07:13:41.517484 4971 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/a8767723-1c4f-41c3-ac4e-988a1d93d56a-scripts\") pod \"a8767723-1c4f-41c3-ac4e-988a1d93d56a\" (UID: \"a8767723-1c4f-41c3-ac4e-988a1d93d56a\") " Dec 13 07:13:41 crc kubenswrapper[4971]: I1213 07:13:41.517500 4971 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/a8767723-1c4f-41c3-ac4e-988a1d93d56a-config-data\") pod \"a8767723-1c4f-41c3-ac4e-988a1d93d56a\" (UID: \"a8767723-1c4f-41c3-ac4e-988a1d93d56a\") " Dec 13 07:13:41 crc kubenswrapper[4971]: I1213 07:13:41.517705 4971 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/a8767723-1c4f-41c3-ac4e-988a1d93d56a-sg-core-conf-yaml\") pod \"a8767723-1c4f-41c3-ac4e-988a1d93d56a\" (UID: \"a8767723-1c4f-41c3-ac4e-988a1d93d56a\") " Dec 13 07:13:41 crc kubenswrapper[4971]: I1213 07:13:41.521880 4971 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/a8767723-1c4f-41c3-ac4e-988a1d93d56a-log-httpd" (OuterVolumeSpecName: "log-httpd") pod "a8767723-1c4f-41c3-ac4e-988a1d93d56a" (UID: "a8767723-1c4f-41c3-ac4e-988a1d93d56a"). InnerVolumeSpecName "log-httpd". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 13 07:13:41 crc kubenswrapper[4971]: I1213 07:13:41.522182 4971 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/a8767723-1c4f-41c3-ac4e-988a1d93d56a-run-httpd" (OuterVolumeSpecName: "run-httpd") pod "a8767723-1c4f-41c3-ac4e-988a1d93d56a" (UID: "a8767723-1c4f-41c3-ac4e-988a1d93d56a"). InnerVolumeSpecName "run-httpd". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 13 07:13:41 crc kubenswrapper[4971]: I1213 07:13:41.525596 4971 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/horizon-8699dc887d-sl2h6"] Dec 13 07:13:41 crc kubenswrapper[4971]: I1213 07:13:41.525746 4971 scope.go:117] "RemoveContainer" containerID="3ea374af6edd7bef76b4de25745f39b1b6be9ae48de86284b377702d49b7b32a" Dec 13 07:13:41 crc kubenswrapper[4971]: I1213 07:13:41.536598 4971 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/a8767723-1c4f-41c3-ac4e-988a1d93d56a-scripts" (OuterVolumeSpecName: "scripts") pod "a8767723-1c4f-41c3-ac4e-988a1d93d56a" (UID: "a8767723-1c4f-41c3-ac4e-988a1d93d56a"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 13 07:13:41 crc kubenswrapper[4971]: I1213 07:13:41.538833 4971 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/a8767723-1c4f-41c3-ac4e-988a1d93d56a-kube-api-access-6hh4c" (OuterVolumeSpecName: "kube-api-access-6hh4c") pod "a8767723-1c4f-41c3-ac4e-988a1d93d56a" (UID: "a8767723-1c4f-41c3-ac4e-988a1d93d56a"). InnerVolumeSpecName "kube-api-access-6hh4c". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 13 07:13:41 crc kubenswrapper[4971]: I1213 07:13:41.547534 4971 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/horizon-8699dc887d-sl2h6"] Dec 13 07:13:41 crc kubenswrapper[4971]: I1213 07:13:41.567728 4971 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/a8767723-1c4f-41c3-ac4e-988a1d93d56a-sg-core-conf-yaml" (OuterVolumeSpecName: "sg-core-conf-yaml") pod "a8767723-1c4f-41c3-ac4e-988a1d93d56a" (UID: "a8767723-1c4f-41c3-ac4e-988a1d93d56a"). InnerVolumeSpecName "sg-core-conf-yaml". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 13 07:13:41 crc kubenswrapper[4971]: I1213 07:13:41.619651 4971 reconciler_common.go:293] "Volume detached for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/a8767723-1c4f-41c3-ac4e-988a1d93d56a-sg-core-conf-yaml\") on node \"crc\" DevicePath \"\"" Dec 13 07:13:41 crc kubenswrapper[4971]: I1213 07:13:41.619680 4971 reconciler_common.go:293] "Volume detached for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/a8767723-1c4f-41c3-ac4e-988a1d93d56a-run-httpd\") on node \"crc\" DevicePath \"\"" Dec 13 07:13:41 crc kubenswrapper[4971]: I1213 07:13:41.619691 4971 reconciler_common.go:293] "Volume detached for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/a8767723-1c4f-41c3-ac4e-988a1d93d56a-log-httpd\") on node \"crc\" DevicePath \"\"" Dec 13 07:13:41 crc kubenswrapper[4971]: I1213 07:13:41.619704 4971 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-6hh4c\" (UniqueName: \"kubernetes.io/projected/a8767723-1c4f-41c3-ac4e-988a1d93d56a-kube-api-access-6hh4c\") on node \"crc\" DevicePath \"\"" Dec 13 07:13:41 crc kubenswrapper[4971]: I1213 07:13:41.619886 4971 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/a8767723-1c4f-41c3-ac4e-988a1d93d56a-scripts\") on node \"crc\" DevicePath \"\"" Dec 13 07:13:41 crc kubenswrapper[4971]: I1213 07:13:41.638658 4971 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/a8767723-1c4f-41c3-ac4e-988a1d93d56a-config-data" (OuterVolumeSpecName: "config-data") pod "a8767723-1c4f-41c3-ac4e-988a1d93d56a" (UID: "a8767723-1c4f-41c3-ac4e-988a1d93d56a"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 13 07:13:41 crc kubenswrapper[4971]: I1213 07:13:41.638992 4971 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/a8767723-1c4f-41c3-ac4e-988a1d93d56a-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "a8767723-1c4f-41c3-ac4e-988a1d93d56a" (UID: "a8767723-1c4f-41c3-ac4e-988a1d93d56a"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 13 07:13:41 crc kubenswrapper[4971]: I1213 07:13:41.670731 4971 scope.go:117] "RemoveContainer" containerID="9ac009ffef2b0f0f983c49e9d333c68f5bc4ca696b14089108a670ec9c8ae630" Dec 13 07:13:41 crc kubenswrapper[4971]: I1213 07:13:41.685573 4971 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/kube-state-metrics-0" Dec 13 07:13:41 crc kubenswrapper[4971]: I1213 07:13:41.693330 4971 scope.go:117] "RemoveContainer" containerID="3eb143d88d13c5a55a04a05582effc70650cefc1337e365a514f4fabdcf9888f" Dec 13 07:13:41 crc kubenswrapper[4971]: E1213 07:13:41.693847 4971 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"3eb143d88d13c5a55a04a05582effc70650cefc1337e365a514f4fabdcf9888f\": container with ID starting with 3eb143d88d13c5a55a04a05582effc70650cefc1337e365a514f4fabdcf9888f not found: ID does not exist" containerID="3eb143d88d13c5a55a04a05582effc70650cefc1337e365a514f4fabdcf9888f" Dec 13 07:13:41 crc kubenswrapper[4971]: I1213 07:13:41.693895 4971 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"3eb143d88d13c5a55a04a05582effc70650cefc1337e365a514f4fabdcf9888f"} err="failed to get container status \"3eb143d88d13c5a55a04a05582effc70650cefc1337e365a514f4fabdcf9888f\": rpc error: code = NotFound desc = could not find container \"3eb143d88d13c5a55a04a05582effc70650cefc1337e365a514f4fabdcf9888f\": container with ID starting with 3eb143d88d13c5a55a04a05582effc70650cefc1337e365a514f4fabdcf9888f not found: ID does not exist" Dec 13 07:13:41 crc kubenswrapper[4971]: I1213 07:13:41.693921 4971 scope.go:117] "RemoveContainer" containerID="714d0b6157ede690f30644b9c368321eb015d8734735a3cf3e7f11de2e94d6bb" Dec 13 07:13:41 crc kubenswrapper[4971]: E1213 07:13:41.694187 4971 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"714d0b6157ede690f30644b9c368321eb015d8734735a3cf3e7f11de2e94d6bb\": container with ID starting with 714d0b6157ede690f30644b9c368321eb015d8734735a3cf3e7f11de2e94d6bb not found: ID does not exist" containerID="714d0b6157ede690f30644b9c368321eb015d8734735a3cf3e7f11de2e94d6bb" Dec 13 07:13:41 crc kubenswrapper[4971]: I1213 07:13:41.694218 4971 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"714d0b6157ede690f30644b9c368321eb015d8734735a3cf3e7f11de2e94d6bb"} err="failed to get container status \"714d0b6157ede690f30644b9c368321eb015d8734735a3cf3e7f11de2e94d6bb\": rpc error: code = NotFound desc = could not find container \"714d0b6157ede690f30644b9c368321eb015d8734735a3cf3e7f11de2e94d6bb\": container with ID starting with 714d0b6157ede690f30644b9c368321eb015d8734735a3cf3e7f11de2e94d6bb not found: ID does not exist" Dec 13 07:13:41 crc kubenswrapper[4971]: I1213 07:13:41.694235 4971 scope.go:117] "RemoveContainer" containerID="3ea374af6edd7bef76b4de25745f39b1b6be9ae48de86284b377702d49b7b32a" Dec 13 07:13:41 crc kubenswrapper[4971]: E1213 07:13:41.694622 4971 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"3ea374af6edd7bef76b4de25745f39b1b6be9ae48de86284b377702d49b7b32a\": container with ID starting with 3ea374af6edd7bef76b4de25745f39b1b6be9ae48de86284b377702d49b7b32a not found: ID does not exist" containerID="3ea374af6edd7bef76b4de25745f39b1b6be9ae48de86284b377702d49b7b32a" Dec 13 07:13:41 crc kubenswrapper[4971]: I1213 07:13:41.694667 4971 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"3ea374af6edd7bef76b4de25745f39b1b6be9ae48de86284b377702d49b7b32a"} err="failed to get container status \"3ea374af6edd7bef76b4de25745f39b1b6be9ae48de86284b377702d49b7b32a\": rpc error: code = NotFound desc = could not find container \"3ea374af6edd7bef76b4de25745f39b1b6be9ae48de86284b377702d49b7b32a\": container with ID starting with 3ea374af6edd7bef76b4de25745f39b1b6be9ae48de86284b377702d49b7b32a not found: ID does not exist" Dec 13 07:13:41 crc kubenswrapper[4971]: I1213 07:13:41.694694 4971 scope.go:117] "RemoveContainer" containerID="9ac009ffef2b0f0f983c49e9d333c68f5bc4ca696b14089108a670ec9c8ae630" Dec 13 07:13:41 crc kubenswrapper[4971]: E1213 07:13:41.695496 4971 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"9ac009ffef2b0f0f983c49e9d333c68f5bc4ca696b14089108a670ec9c8ae630\": container with ID starting with 9ac009ffef2b0f0f983c49e9d333c68f5bc4ca696b14089108a670ec9c8ae630 not found: ID does not exist" containerID="9ac009ffef2b0f0f983c49e9d333c68f5bc4ca696b14089108a670ec9c8ae630" Dec 13 07:13:41 crc kubenswrapper[4971]: I1213 07:13:41.695549 4971 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"9ac009ffef2b0f0f983c49e9d333c68f5bc4ca696b14089108a670ec9c8ae630"} err="failed to get container status \"9ac009ffef2b0f0f983c49e9d333c68f5bc4ca696b14089108a670ec9c8ae630\": rpc error: code = NotFound desc = could not find container \"9ac009ffef2b0f0f983c49e9d333c68f5bc4ca696b14089108a670ec9c8ae630\": container with ID starting with 9ac009ffef2b0f0f983c49e9d333c68f5bc4ca696b14089108a670ec9c8ae630 not found: ID does not exist" Dec 13 07:13:41 crc kubenswrapper[4971]: I1213 07:13:41.695566 4971 scope.go:117] "RemoveContainer" containerID="4a148882fad66a5e1444a664e27875de5e47f0f5a778fde8428a33e1ba8f489f" Dec 13 07:13:41 crc kubenswrapper[4971]: I1213 07:13:41.720991 4971 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-c58tt\" (UniqueName: \"kubernetes.io/projected/2bab8c0a-bd76-4265-8355-8df0c6fce6c3-kube-api-access-c58tt\") pod \"2bab8c0a-bd76-4265-8355-8df0c6fce6c3\" (UID: \"2bab8c0a-bd76-4265-8355-8df0c6fce6c3\") " Dec 13 07:13:41 crc kubenswrapper[4971]: I1213 07:13:41.721602 4971 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/a8767723-1c4f-41c3-ac4e-988a1d93d56a-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 13 07:13:41 crc kubenswrapper[4971]: I1213 07:13:41.721634 4971 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/a8767723-1c4f-41c3-ac4e-988a1d93d56a-config-data\") on node \"crc\" DevicePath \"\"" Dec 13 07:13:41 crc kubenswrapper[4971]: I1213 07:13:41.724110 4971 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/2bab8c0a-bd76-4265-8355-8df0c6fce6c3-kube-api-access-c58tt" (OuterVolumeSpecName: "kube-api-access-c58tt") pod "2bab8c0a-bd76-4265-8355-8df0c6fce6c3" (UID: "2bab8c0a-bd76-4265-8355-8df0c6fce6c3"). InnerVolumeSpecName "kube-api-access-c58tt". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 13 07:13:41 crc kubenswrapper[4971]: I1213 07:13:41.781162 4971 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="772eedab-7c62-454b-b722-ee453b39e89f" path="/var/lib/kubelet/pods/772eedab-7c62-454b-b722-ee453b39e89f/volumes" Dec 13 07:13:41 crc kubenswrapper[4971]: I1213 07:13:41.825389 4971 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-c58tt\" (UniqueName: \"kubernetes.io/projected/2bab8c0a-bd76-4265-8355-8df0c6fce6c3-kube-api-access-c58tt\") on node \"crc\" DevicePath \"\"" Dec 13 07:13:41 crc kubenswrapper[4971]: I1213 07:13:41.828143 4971 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ceilometer-0"] Dec 13 07:13:41 crc kubenswrapper[4971]: I1213 07:13:41.844934 4971 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/ceilometer-0"] Dec 13 07:13:41 crc kubenswrapper[4971]: I1213 07:13:41.871355 4971 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ceilometer-0"] Dec 13 07:13:41 crc kubenswrapper[4971]: E1213 07:13:41.871811 4971 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="a8767723-1c4f-41c3-ac4e-988a1d93d56a" containerName="ceilometer-notification-agent" Dec 13 07:13:41 crc kubenswrapper[4971]: I1213 07:13:41.871837 4971 state_mem.go:107] "Deleted CPUSet assignment" podUID="a8767723-1c4f-41c3-ac4e-988a1d93d56a" containerName="ceilometer-notification-agent" Dec 13 07:13:41 crc kubenswrapper[4971]: E1213 07:13:41.871853 4971 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="0ed1405a-8326-4fe9-8fb4-1d97b9e8befb" containerName="mariadb-database-create" Dec 13 07:13:41 crc kubenswrapper[4971]: I1213 07:13:41.871862 4971 state_mem.go:107] "Deleted CPUSet assignment" podUID="0ed1405a-8326-4fe9-8fb4-1d97b9e8befb" containerName="mariadb-database-create" Dec 13 07:13:41 crc kubenswrapper[4971]: E1213 07:13:41.871881 4971 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="5ffa477d-1f7b-445f-9a97-65a7889e44d4" containerName="mariadb-account-create-update" Dec 13 07:13:41 crc kubenswrapper[4971]: I1213 07:13:41.871890 4971 state_mem.go:107] "Deleted CPUSet assignment" podUID="5ffa477d-1f7b-445f-9a97-65a7889e44d4" containerName="mariadb-account-create-update" Dec 13 07:13:41 crc kubenswrapper[4971]: E1213 07:13:41.871900 4971 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="e0e9f238-3cb4-43ff-bac3-f1c283d0915f" containerName="mariadb-account-create-update" Dec 13 07:13:41 crc kubenswrapper[4971]: I1213 07:13:41.871910 4971 state_mem.go:107] "Deleted CPUSet assignment" podUID="e0e9f238-3cb4-43ff-bac3-f1c283d0915f" containerName="mariadb-account-create-update" Dec 13 07:13:41 crc kubenswrapper[4971]: E1213 07:13:41.871919 4971 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="5ba87867-bb90-4587-9012-6db9d2901b10" containerName="mariadb-database-create" Dec 13 07:13:41 crc kubenswrapper[4971]: I1213 07:13:41.871927 4971 state_mem.go:107] "Deleted CPUSet assignment" podUID="5ba87867-bb90-4587-9012-6db9d2901b10" containerName="mariadb-database-create" Dec 13 07:13:41 crc kubenswrapper[4971]: E1213 07:13:41.871937 4971 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="a8767723-1c4f-41c3-ac4e-988a1d93d56a" containerName="sg-core" Dec 13 07:13:41 crc kubenswrapper[4971]: I1213 07:13:41.871946 4971 state_mem.go:107] "Deleted CPUSet assignment" podUID="a8767723-1c4f-41c3-ac4e-988a1d93d56a" containerName="sg-core" Dec 13 07:13:41 crc kubenswrapper[4971]: E1213 07:13:41.871962 4971 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="8b60a7ee-8d5e-4bf2-8d64-161aeced2357" containerName="mariadb-database-create" Dec 13 07:13:41 crc kubenswrapper[4971]: I1213 07:13:41.871971 4971 state_mem.go:107] "Deleted CPUSet assignment" podUID="8b60a7ee-8d5e-4bf2-8d64-161aeced2357" containerName="mariadb-database-create" Dec 13 07:13:41 crc kubenswrapper[4971]: E1213 07:13:41.871983 4971 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="a8767723-1c4f-41c3-ac4e-988a1d93d56a" containerName="proxy-httpd" Dec 13 07:13:41 crc kubenswrapper[4971]: I1213 07:13:41.871991 4971 state_mem.go:107] "Deleted CPUSet assignment" podUID="a8767723-1c4f-41c3-ac4e-988a1d93d56a" containerName="proxy-httpd" Dec 13 07:13:41 crc kubenswrapper[4971]: E1213 07:13:41.872020 4971 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="772eedab-7c62-454b-b722-ee453b39e89f" containerName="horizon-log" Dec 13 07:13:41 crc kubenswrapper[4971]: I1213 07:13:41.872028 4971 state_mem.go:107] "Deleted CPUSet assignment" podUID="772eedab-7c62-454b-b722-ee453b39e89f" containerName="horizon-log" Dec 13 07:13:41 crc kubenswrapper[4971]: E1213 07:13:41.872038 4971 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="edac9227-16d2-48b0-ac2b-3f71c67d6cf8" containerName="mariadb-account-create-update" Dec 13 07:13:41 crc kubenswrapper[4971]: I1213 07:13:41.872047 4971 state_mem.go:107] "Deleted CPUSet assignment" podUID="edac9227-16d2-48b0-ac2b-3f71c67d6cf8" containerName="mariadb-account-create-update" Dec 13 07:13:41 crc kubenswrapper[4971]: E1213 07:13:41.872063 4971 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="2bab8c0a-bd76-4265-8355-8df0c6fce6c3" containerName="kube-state-metrics" Dec 13 07:13:41 crc kubenswrapper[4971]: I1213 07:13:41.872071 4971 state_mem.go:107] "Deleted CPUSet assignment" podUID="2bab8c0a-bd76-4265-8355-8df0c6fce6c3" containerName="kube-state-metrics" Dec 13 07:13:41 crc kubenswrapper[4971]: E1213 07:13:41.872092 4971 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="a8767723-1c4f-41c3-ac4e-988a1d93d56a" containerName="ceilometer-central-agent" Dec 13 07:13:41 crc kubenswrapper[4971]: I1213 07:13:41.872101 4971 state_mem.go:107] "Deleted CPUSet assignment" podUID="a8767723-1c4f-41c3-ac4e-988a1d93d56a" containerName="ceilometer-central-agent" Dec 13 07:13:41 crc kubenswrapper[4971]: E1213 07:13:41.872116 4971 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="772eedab-7c62-454b-b722-ee453b39e89f" containerName="horizon" Dec 13 07:13:41 crc kubenswrapper[4971]: I1213 07:13:41.872126 4971 state_mem.go:107] "Deleted CPUSet assignment" podUID="772eedab-7c62-454b-b722-ee453b39e89f" containerName="horizon" Dec 13 07:13:41 crc kubenswrapper[4971]: I1213 07:13:41.872329 4971 memory_manager.go:354] "RemoveStaleState removing state" podUID="5ba87867-bb90-4587-9012-6db9d2901b10" containerName="mariadb-database-create" Dec 13 07:13:41 crc kubenswrapper[4971]: I1213 07:13:41.872371 4971 memory_manager.go:354] "RemoveStaleState removing state" podUID="2bab8c0a-bd76-4265-8355-8df0c6fce6c3" containerName="kube-state-metrics" Dec 13 07:13:41 crc kubenswrapper[4971]: I1213 07:13:41.872388 4971 memory_manager.go:354] "RemoveStaleState removing state" podUID="edac9227-16d2-48b0-ac2b-3f71c67d6cf8" containerName="mariadb-account-create-update" Dec 13 07:13:41 crc kubenswrapper[4971]: I1213 07:13:41.872411 4971 memory_manager.go:354] "RemoveStaleState removing state" podUID="a8767723-1c4f-41c3-ac4e-988a1d93d56a" containerName="sg-core" Dec 13 07:13:41 crc kubenswrapper[4971]: I1213 07:13:41.872421 4971 memory_manager.go:354] "RemoveStaleState removing state" podUID="a8767723-1c4f-41c3-ac4e-988a1d93d56a" containerName="ceilometer-notification-agent" Dec 13 07:13:41 crc kubenswrapper[4971]: I1213 07:13:41.872433 4971 memory_manager.go:354] "RemoveStaleState removing state" podUID="e0e9f238-3cb4-43ff-bac3-f1c283d0915f" containerName="mariadb-account-create-update" Dec 13 07:13:41 crc kubenswrapper[4971]: I1213 07:13:41.872449 4971 memory_manager.go:354] "RemoveStaleState removing state" podUID="772eedab-7c62-454b-b722-ee453b39e89f" containerName="horizon" Dec 13 07:13:41 crc kubenswrapper[4971]: I1213 07:13:41.872472 4971 memory_manager.go:354] "RemoveStaleState removing state" podUID="8b60a7ee-8d5e-4bf2-8d64-161aeced2357" containerName="mariadb-database-create" Dec 13 07:13:41 crc kubenswrapper[4971]: I1213 07:13:41.872484 4971 memory_manager.go:354] "RemoveStaleState removing state" podUID="a8767723-1c4f-41c3-ac4e-988a1d93d56a" containerName="ceilometer-central-agent" Dec 13 07:13:41 crc kubenswrapper[4971]: I1213 07:13:41.872503 4971 memory_manager.go:354] "RemoveStaleState removing state" podUID="a8767723-1c4f-41c3-ac4e-988a1d93d56a" containerName="proxy-httpd" Dec 13 07:13:41 crc kubenswrapper[4971]: I1213 07:13:41.872529 4971 memory_manager.go:354] "RemoveStaleState removing state" podUID="772eedab-7c62-454b-b722-ee453b39e89f" containerName="horizon-log" Dec 13 07:13:41 crc kubenswrapper[4971]: I1213 07:13:41.872551 4971 memory_manager.go:354] "RemoveStaleState removing state" podUID="0ed1405a-8326-4fe9-8fb4-1d97b9e8befb" containerName="mariadb-database-create" Dec 13 07:13:41 crc kubenswrapper[4971]: I1213 07:13:41.872570 4971 memory_manager.go:354] "RemoveStaleState removing state" podUID="5ffa477d-1f7b-445f-9a97-65a7889e44d4" containerName="mariadb-account-create-update" Dec 13 07:13:41 crc kubenswrapper[4971]: I1213 07:13:41.874385 4971 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Dec 13 07:13:41 crc kubenswrapper[4971]: I1213 07:13:41.879012 4971 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceilometer-scripts" Dec 13 07:13:41 crc kubenswrapper[4971]: I1213 07:13:41.879199 4971 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceilometer-config-data" Dec 13 07:13:41 crc kubenswrapper[4971]: I1213 07:13:41.902828 4971 scope.go:117] "RemoveContainer" containerID="2074da306f3c927fa0e63c982fab7574e67b1f06c38e55ba7dc406c0d32df42a" Dec 13 07:13:41 crc kubenswrapper[4971]: I1213 07:13:41.908566 4971 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ceilometer-0"] Dec 13 07:13:41 crc kubenswrapper[4971]: I1213 07:13:41.928488 4971 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/1cd47626-171c-4c51-a572-3e9880154889-run-httpd\") pod \"ceilometer-0\" (UID: \"1cd47626-171c-4c51-a572-3e9880154889\") " pod="openstack/ceilometer-0" Dec 13 07:13:41 crc kubenswrapper[4971]: I1213 07:13:41.928547 4971 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/1cd47626-171c-4c51-a572-3e9880154889-scripts\") pod \"ceilometer-0\" (UID: \"1cd47626-171c-4c51-a572-3e9880154889\") " pod="openstack/ceilometer-0" Dec 13 07:13:41 crc kubenswrapper[4971]: I1213 07:13:41.928646 4971 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/1cd47626-171c-4c51-a572-3e9880154889-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"1cd47626-171c-4c51-a572-3e9880154889\") " pod="openstack/ceilometer-0" Dec 13 07:13:41 crc kubenswrapper[4971]: I1213 07:13:41.928686 4971 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-54rhn\" (UniqueName: \"kubernetes.io/projected/1cd47626-171c-4c51-a572-3e9880154889-kube-api-access-54rhn\") pod \"ceilometer-0\" (UID: \"1cd47626-171c-4c51-a572-3e9880154889\") " pod="openstack/ceilometer-0" Dec 13 07:13:41 crc kubenswrapper[4971]: I1213 07:13:41.928711 4971 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/1cd47626-171c-4c51-a572-3e9880154889-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"1cd47626-171c-4c51-a572-3e9880154889\") " pod="openstack/ceilometer-0" Dec 13 07:13:41 crc kubenswrapper[4971]: I1213 07:13:41.928737 4971 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/1cd47626-171c-4c51-a572-3e9880154889-config-data\") pod \"ceilometer-0\" (UID: \"1cd47626-171c-4c51-a572-3e9880154889\") " pod="openstack/ceilometer-0" Dec 13 07:13:41 crc kubenswrapper[4971]: I1213 07:13:41.928910 4971 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/1cd47626-171c-4c51-a572-3e9880154889-log-httpd\") pod \"ceilometer-0\" (UID: \"1cd47626-171c-4c51-a572-3e9880154889\") " pod="openstack/ceilometer-0" Dec 13 07:13:42 crc kubenswrapper[4971]: I1213 07:13:42.030197 4971 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/1cd47626-171c-4c51-a572-3e9880154889-run-httpd\") pod \"ceilometer-0\" (UID: \"1cd47626-171c-4c51-a572-3e9880154889\") " pod="openstack/ceilometer-0" Dec 13 07:13:42 crc kubenswrapper[4971]: I1213 07:13:42.030237 4971 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/1cd47626-171c-4c51-a572-3e9880154889-scripts\") pod \"ceilometer-0\" (UID: \"1cd47626-171c-4c51-a572-3e9880154889\") " pod="openstack/ceilometer-0" Dec 13 07:13:42 crc kubenswrapper[4971]: I1213 07:13:42.030295 4971 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/1cd47626-171c-4c51-a572-3e9880154889-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"1cd47626-171c-4c51-a572-3e9880154889\") " pod="openstack/ceilometer-0" Dec 13 07:13:42 crc kubenswrapper[4971]: I1213 07:13:42.030329 4971 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-54rhn\" (UniqueName: \"kubernetes.io/projected/1cd47626-171c-4c51-a572-3e9880154889-kube-api-access-54rhn\") pod \"ceilometer-0\" (UID: \"1cd47626-171c-4c51-a572-3e9880154889\") " pod="openstack/ceilometer-0" Dec 13 07:13:42 crc kubenswrapper[4971]: I1213 07:13:42.030362 4971 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/1cd47626-171c-4c51-a572-3e9880154889-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"1cd47626-171c-4c51-a572-3e9880154889\") " pod="openstack/ceilometer-0" Dec 13 07:13:42 crc kubenswrapper[4971]: I1213 07:13:42.030383 4971 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/1cd47626-171c-4c51-a572-3e9880154889-config-data\") pod \"ceilometer-0\" (UID: \"1cd47626-171c-4c51-a572-3e9880154889\") " pod="openstack/ceilometer-0" Dec 13 07:13:42 crc kubenswrapper[4971]: I1213 07:13:42.030888 4971 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/1cd47626-171c-4c51-a572-3e9880154889-run-httpd\") pod \"ceilometer-0\" (UID: \"1cd47626-171c-4c51-a572-3e9880154889\") " pod="openstack/ceilometer-0" Dec 13 07:13:42 crc kubenswrapper[4971]: I1213 07:13:42.030971 4971 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/1cd47626-171c-4c51-a572-3e9880154889-log-httpd\") pod \"ceilometer-0\" (UID: \"1cd47626-171c-4c51-a572-3e9880154889\") " pod="openstack/ceilometer-0" Dec 13 07:13:42 crc kubenswrapper[4971]: I1213 07:13:42.031273 4971 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/1cd47626-171c-4c51-a572-3e9880154889-log-httpd\") pod \"ceilometer-0\" (UID: \"1cd47626-171c-4c51-a572-3e9880154889\") " pod="openstack/ceilometer-0" Dec 13 07:13:42 crc kubenswrapper[4971]: I1213 07:13:42.037617 4971 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/1cd47626-171c-4c51-a572-3e9880154889-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"1cd47626-171c-4c51-a572-3e9880154889\") " pod="openstack/ceilometer-0" Dec 13 07:13:42 crc kubenswrapper[4971]: I1213 07:13:42.041169 4971 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/1cd47626-171c-4c51-a572-3e9880154889-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"1cd47626-171c-4c51-a572-3e9880154889\") " pod="openstack/ceilometer-0" Dec 13 07:13:42 crc kubenswrapper[4971]: I1213 07:13:42.041246 4971 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/1cd47626-171c-4c51-a572-3e9880154889-scripts\") pod \"ceilometer-0\" (UID: \"1cd47626-171c-4c51-a572-3e9880154889\") " pod="openstack/ceilometer-0" Dec 13 07:13:42 crc kubenswrapper[4971]: I1213 07:13:42.041638 4971 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/1cd47626-171c-4c51-a572-3e9880154889-config-data\") pod \"ceilometer-0\" (UID: \"1cd47626-171c-4c51-a572-3e9880154889\") " pod="openstack/ceilometer-0" Dec 13 07:13:42 crc kubenswrapper[4971]: I1213 07:13:42.052343 4971 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-54rhn\" (UniqueName: \"kubernetes.io/projected/1cd47626-171c-4c51-a572-3e9880154889-kube-api-access-54rhn\") pod \"ceilometer-0\" (UID: \"1cd47626-171c-4c51-a572-3e9880154889\") " pod="openstack/ceilometer-0" Dec 13 07:13:42 crc kubenswrapper[4971]: I1213 07:13:42.191865 4971 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Dec 13 07:13:42 crc kubenswrapper[4971]: I1213 07:13:42.503477 4971 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/kube-state-metrics-0" event={"ID":"2bab8c0a-bd76-4265-8355-8df0c6fce6c3","Type":"ContainerDied","Data":"76129f35a3b0ed0f3d63fd9275ae2264a1ef0588120ed9ab9858a169a60b5f88"} Dec 13 07:13:42 crc kubenswrapper[4971]: I1213 07:13:42.503555 4971 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/kube-state-metrics-0" Dec 13 07:13:42 crc kubenswrapper[4971]: I1213 07:13:42.503879 4971 scope.go:117] "RemoveContainer" containerID="7a965470ccb73744ddf8839144cea274fc2632c1baaa2c97193772f8c5e14440" Dec 13 07:13:42 crc kubenswrapper[4971]: I1213 07:13:42.530951 4971 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/kube-state-metrics-0"] Dec 13 07:13:42 crc kubenswrapper[4971]: I1213 07:13:42.541669 4971 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/kube-state-metrics-0"] Dec 13 07:13:42 crc kubenswrapper[4971]: I1213 07:13:42.556541 4971 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/kube-state-metrics-0"] Dec 13 07:13:42 crc kubenswrapper[4971]: I1213 07:13:42.557840 4971 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/kube-state-metrics-0" Dec 13 07:13:42 crc kubenswrapper[4971]: I1213 07:13:42.560159 4971 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"kube-state-metrics-tls-config" Dec 13 07:13:42 crc kubenswrapper[4971]: I1213 07:13:42.561842 4971 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-kube-state-metrics-svc" Dec 13 07:13:42 crc kubenswrapper[4971]: I1213 07:13:42.568891 4971 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/kube-state-metrics-0"] Dec 13 07:13:42 crc kubenswrapper[4971]: I1213 07:13:42.644788 4971 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/8314d6ed-4778-4709-a494-0642e0ce0617-combined-ca-bundle\") pod \"kube-state-metrics-0\" (UID: \"8314d6ed-4778-4709-a494-0642e0ce0617\") " pod="openstack/kube-state-metrics-0" Dec 13 07:13:42 crc kubenswrapper[4971]: I1213 07:13:42.644914 4971 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-state-metrics-tls-certs\" (UniqueName: \"kubernetes.io/secret/8314d6ed-4778-4709-a494-0642e0ce0617-kube-state-metrics-tls-certs\") pod \"kube-state-metrics-0\" (UID: \"8314d6ed-4778-4709-a494-0642e0ce0617\") " pod="openstack/kube-state-metrics-0" Dec 13 07:13:42 crc kubenswrapper[4971]: I1213 07:13:42.644941 4971 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-b6z6f\" (UniqueName: \"kubernetes.io/projected/8314d6ed-4778-4709-a494-0642e0ce0617-kube-api-access-b6z6f\") pod \"kube-state-metrics-0\" (UID: \"8314d6ed-4778-4709-a494-0642e0ce0617\") " pod="openstack/kube-state-metrics-0" Dec 13 07:13:42 crc kubenswrapper[4971]: I1213 07:13:42.644990 4971 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-state-metrics-tls-config\" (UniqueName: \"kubernetes.io/secret/8314d6ed-4778-4709-a494-0642e0ce0617-kube-state-metrics-tls-config\") pod \"kube-state-metrics-0\" (UID: \"8314d6ed-4778-4709-a494-0642e0ce0617\") " pod="openstack/kube-state-metrics-0" Dec 13 07:13:42 crc kubenswrapper[4971]: I1213 07:13:42.662871 4971 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ceilometer-0"] Dec 13 07:13:42 crc kubenswrapper[4971]: I1213 07:13:42.727259 4971 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ceilometer-0"] Dec 13 07:13:42 crc kubenswrapper[4971]: I1213 07:13:42.747121 4971 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-state-metrics-tls-certs\" (UniqueName: \"kubernetes.io/secret/8314d6ed-4778-4709-a494-0642e0ce0617-kube-state-metrics-tls-certs\") pod \"kube-state-metrics-0\" (UID: \"8314d6ed-4778-4709-a494-0642e0ce0617\") " pod="openstack/kube-state-metrics-0" Dec 13 07:13:42 crc kubenswrapper[4971]: I1213 07:13:42.747162 4971 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-b6z6f\" (UniqueName: \"kubernetes.io/projected/8314d6ed-4778-4709-a494-0642e0ce0617-kube-api-access-b6z6f\") pod \"kube-state-metrics-0\" (UID: \"8314d6ed-4778-4709-a494-0642e0ce0617\") " pod="openstack/kube-state-metrics-0" Dec 13 07:13:42 crc kubenswrapper[4971]: I1213 07:13:42.747214 4971 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-state-metrics-tls-config\" (UniqueName: \"kubernetes.io/secret/8314d6ed-4778-4709-a494-0642e0ce0617-kube-state-metrics-tls-config\") pod \"kube-state-metrics-0\" (UID: \"8314d6ed-4778-4709-a494-0642e0ce0617\") " pod="openstack/kube-state-metrics-0" Dec 13 07:13:42 crc kubenswrapper[4971]: I1213 07:13:42.747283 4971 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/8314d6ed-4778-4709-a494-0642e0ce0617-combined-ca-bundle\") pod \"kube-state-metrics-0\" (UID: \"8314d6ed-4778-4709-a494-0642e0ce0617\") " pod="openstack/kube-state-metrics-0" Dec 13 07:13:42 crc kubenswrapper[4971]: I1213 07:13:42.752249 4971 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-state-metrics-tls-config\" (UniqueName: \"kubernetes.io/secret/8314d6ed-4778-4709-a494-0642e0ce0617-kube-state-metrics-tls-config\") pod \"kube-state-metrics-0\" (UID: \"8314d6ed-4778-4709-a494-0642e0ce0617\") " pod="openstack/kube-state-metrics-0" Dec 13 07:13:42 crc kubenswrapper[4971]: I1213 07:13:42.752278 4971 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-state-metrics-tls-certs\" (UniqueName: \"kubernetes.io/secret/8314d6ed-4778-4709-a494-0642e0ce0617-kube-state-metrics-tls-certs\") pod \"kube-state-metrics-0\" (UID: \"8314d6ed-4778-4709-a494-0642e0ce0617\") " pod="openstack/kube-state-metrics-0" Dec 13 07:13:42 crc kubenswrapper[4971]: I1213 07:13:42.754676 4971 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/8314d6ed-4778-4709-a494-0642e0ce0617-combined-ca-bundle\") pod \"kube-state-metrics-0\" (UID: \"8314d6ed-4778-4709-a494-0642e0ce0617\") " pod="openstack/kube-state-metrics-0" Dec 13 07:13:42 crc kubenswrapper[4971]: I1213 07:13:42.766161 4971 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-b6z6f\" (UniqueName: \"kubernetes.io/projected/8314d6ed-4778-4709-a494-0642e0ce0617-kube-api-access-b6z6f\") pod \"kube-state-metrics-0\" (UID: \"8314d6ed-4778-4709-a494-0642e0ce0617\") " pod="openstack/kube-state-metrics-0" Dec 13 07:13:42 crc kubenswrapper[4971]: I1213 07:13:42.874778 4971 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/kube-state-metrics-0" Dec 13 07:13:43 crc kubenswrapper[4971]: I1213 07:13:43.310115 4971 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/kube-state-metrics-0"] Dec 13 07:13:43 crc kubenswrapper[4971]: I1213 07:13:43.516245 4971 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/kube-state-metrics-0" event={"ID":"8314d6ed-4778-4709-a494-0642e0ce0617","Type":"ContainerStarted","Data":"356b830fce08357f19db427db379d8a44db0d93cbbc40ecceea11cc850a8a8bd"} Dec 13 07:13:43 crc kubenswrapper[4971]: I1213 07:13:43.518211 4971 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"1cd47626-171c-4c51-a572-3e9880154889","Type":"ContainerStarted","Data":"df0910877e0c6048d15e192c1497a2330564ff6f2f4f3b2dc531118f87e55a02"} Dec 13 07:13:43 crc kubenswrapper[4971]: I1213 07:13:43.518250 4971 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"1cd47626-171c-4c51-a572-3e9880154889","Type":"ContainerStarted","Data":"50fb0c503e99053d2c7f20d771dec4bac9aabb054953ccb12bff0a6fdcdbf963"} Dec 13 07:13:43 crc kubenswrapper[4971]: I1213 07:13:43.802178 4971 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="2bab8c0a-bd76-4265-8355-8df0c6fce6c3" path="/var/lib/kubelet/pods/2bab8c0a-bd76-4265-8355-8df0c6fce6c3/volumes" Dec 13 07:13:43 crc kubenswrapper[4971]: I1213 07:13:43.811468 4971 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="a8767723-1c4f-41c3-ac4e-988a1d93d56a" path="/var/lib/kubelet/pods/a8767723-1c4f-41c3-ac4e-988a1d93d56a/volumes" Dec 13 07:13:44 crc kubenswrapper[4971]: I1213 07:13:44.528413 4971 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/kube-state-metrics-0" event={"ID":"8314d6ed-4778-4709-a494-0642e0ce0617","Type":"ContainerStarted","Data":"0e6fc824ea3d4ed056ea5180bf6134fd93922e74728dc4d652b1c50c455a3839"} Dec 13 07:13:44 crc kubenswrapper[4971]: I1213 07:13:44.528809 4971 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/kube-state-metrics-0" Dec 13 07:13:44 crc kubenswrapper[4971]: I1213 07:13:44.530884 4971 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"1cd47626-171c-4c51-a572-3e9880154889","Type":"ContainerStarted","Data":"2433437f27cac2201085434dc4550db63cdb41eb32c3f5ba0b33f5940980f2ae"} Dec 13 07:13:44 crc kubenswrapper[4971]: I1213 07:13:44.548112 4971 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/kube-state-metrics-0" podStartSLOduration=2.170622557 podStartE2EDuration="2.548092881s" podCreationTimestamp="2025-12-13 07:13:42 +0000 UTC" firstStartedPulling="2025-12-13 07:13:43.319680038 +0000 UTC m=+1479.924089486" lastFinishedPulling="2025-12-13 07:13:43.697150362 +0000 UTC m=+1480.301559810" observedRunningTime="2025-12-13 07:13:44.547921167 +0000 UTC m=+1481.152330635" watchObservedRunningTime="2025-12-13 07:13:44.548092881 +0000 UTC m=+1481.152502329" Dec 13 07:13:45 crc kubenswrapper[4971]: I1213 07:13:45.117137 4971 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-cell0-conductor-db-sync-g9t27"] Dec 13 07:13:45 crc kubenswrapper[4971]: I1213 07:13:45.118289 4971 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-conductor-db-sync-g9t27" Dec 13 07:13:45 crc kubenswrapper[4971]: I1213 07:13:45.121725 4971 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-nova-dockercfg-dsf9t" Dec 13 07:13:45 crc kubenswrapper[4971]: I1213 07:13:45.121906 4971 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-cell0-conductor-config-data" Dec 13 07:13:45 crc kubenswrapper[4971]: I1213 07:13:45.122315 4971 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-cell0-conductor-scripts" Dec 13 07:13:45 crc kubenswrapper[4971]: I1213 07:13:45.129227 4971 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell0-conductor-db-sync-g9t27"] Dec 13 07:13:45 crc kubenswrapper[4971]: I1213 07:13:45.290461 4971 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/31dddcf3-380b-4003-a766-88d4f5bc00de-combined-ca-bundle\") pod \"nova-cell0-conductor-db-sync-g9t27\" (UID: \"31dddcf3-380b-4003-a766-88d4f5bc00de\") " pod="openstack/nova-cell0-conductor-db-sync-g9t27" Dec 13 07:13:45 crc kubenswrapper[4971]: I1213 07:13:45.290664 4971 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/31dddcf3-380b-4003-a766-88d4f5bc00de-scripts\") pod \"nova-cell0-conductor-db-sync-g9t27\" (UID: \"31dddcf3-380b-4003-a766-88d4f5bc00de\") " pod="openstack/nova-cell0-conductor-db-sync-g9t27" Dec 13 07:13:45 crc kubenswrapper[4971]: I1213 07:13:45.290702 4971 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/31dddcf3-380b-4003-a766-88d4f5bc00de-config-data\") pod \"nova-cell0-conductor-db-sync-g9t27\" (UID: \"31dddcf3-380b-4003-a766-88d4f5bc00de\") " pod="openstack/nova-cell0-conductor-db-sync-g9t27" Dec 13 07:13:45 crc kubenswrapper[4971]: I1213 07:13:45.290932 4971 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-t8b6j\" (UniqueName: \"kubernetes.io/projected/31dddcf3-380b-4003-a766-88d4f5bc00de-kube-api-access-t8b6j\") pod \"nova-cell0-conductor-db-sync-g9t27\" (UID: \"31dddcf3-380b-4003-a766-88d4f5bc00de\") " pod="openstack/nova-cell0-conductor-db-sync-g9t27" Dec 13 07:13:45 crc kubenswrapper[4971]: I1213 07:13:45.392731 4971 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/31dddcf3-380b-4003-a766-88d4f5bc00de-combined-ca-bundle\") pod \"nova-cell0-conductor-db-sync-g9t27\" (UID: \"31dddcf3-380b-4003-a766-88d4f5bc00de\") " pod="openstack/nova-cell0-conductor-db-sync-g9t27" Dec 13 07:13:45 crc kubenswrapper[4971]: I1213 07:13:45.392987 4971 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/31dddcf3-380b-4003-a766-88d4f5bc00de-scripts\") pod \"nova-cell0-conductor-db-sync-g9t27\" (UID: \"31dddcf3-380b-4003-a766-88d4f5bc00de\") " pod="openstack/nova-cell0-conductor-db-sync-g9t27" Dec 13 07:13:45 crc kubenswrapper[4971]: I1213 07:13:45.393013 4971 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/31dddcf3-380b-4003-a766-88d4f5bc00de-config-data\") pod \"nova-cell0-conductor-db-sync-g9t27\" (UID: \"31dddcf3-380b-4003-a766-88d4f5bc00de\") " pod="openstack/nova-cell0-conductor-db-sync-g9t27" Dec 13 07:13:45 crc kubenswrapper[4971]: I1213 07:13:45.393088 4971 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-t8b6j\" (UniqueName: \"kubernetes.io/projected/31dddcf3-380b-4003-a766-88d4f5bc00de-kube-api-access-t8b6j\") pod \"nova-cell0-conductor-db-sync-g9t27\" (UID: \"31dddcf3-380b-4003-a766-88d4f5bc00de\") " pod="openstack/nova-cell0-conductor-db-sync-g9t27" Dec 13 07:13:45 crc kubenswrapper[4971]: I1213 07:13:45.398441 4971 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/31dddcf3-380b-4003-a766-88d4f5bc00de-config-data\") pod \"nova-cell0-conductor-db-sync-g9t27\" (UID: \"31dddcf3-380b-4003-a766-88d4f5bc00de\") " pod="openstack/nova-cell0-conductor-db-sync-g9t27" Dec 13 07:13:45 crc kubenswrapper[4971]: I1213 07:13:45.399121 4971 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/31dddcf3-380b-4003-a766-88d4f5bc00de-scripts\") pod \"nova-cell0-conductor-db-sync-g9t27\" (UID: \"31dddcf3-380b-4003-a766-88d4f5bc00de\") " pod="openstack/nova-cell0-conductor-db-sync-g9t27" Dec 13 07:13:45 crc kubenswrapper[4971]: I1213 07:13:45.399820 4971 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/31dddcf3-380b-4003-a766-88d4f5bc00de-combined-ca-bundle\") pod \"nova-cell0-conductor-db-sync-g9t27\" (UID: \"31dddcf3-380b-4003-a766-88d4f5bc00de\") " pod="openstack/nova-cell0-conductor-db-sync-g9t27" Dec 13 07:13:45 crc kubenswrapper[4971]: I1213 07:13:45.411194 4971 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-t8b6j\" (UniqueName: \"kubernetes.io/projected/31dddcf3-380b-4003-a766-88d4f5bc00de-kube-api-access-t8b6j\") pod \"nova-cell0-conductor-db-sync-g9t27\" (UID: \"31dddcf3-380b-4003-a766-88d4f5bc00de\") " pod="openstack/nova-cell0-conductor-db-sync-g9t27" Dec 13 07:13:45 crc kubenswrapper[4971]: I1213 07:13:45.434072 4971 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-conductor-db-sync-g9t27" Dec 13 07:13:45 crc kubenswrapper[4971]: I1213 07:13:45.560540 4971 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"1cd47626-171c-4c51-a572-3e9880154889","Type":"ContainerStarted","Data":"628fca53bd17ddb88f713fab9756df4bfda8271a5f16a4176b53807106991e75"} Dec 13 07:13:45 crc kubenswrapper[4971]: I1213 07:13:45.904274 4971 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell0-conductor-db-sync-g9t27"] Dec 13 07:13:46 crc kubenswrapper[4971]: I1213 07:13:46.153092 4971 patch_prober.go:28] interesting pod/machine-config-daemon-82xjz container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 13 07:13:46 crc kubenswrapper[4971]: I1213 07:13:46.153401 4971 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-82xjz" podUID="e0c4af12-edda-41d7-9b44-f87396174ef2" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 13 07:13:46 crc kubenswrapper[4971]: I1213 07:13:46.569553 4971 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-conductor-db-sync-g9t27" event={"ID":"31dddcf3-380b-4003-a766-88d4f5bc00de","Type":"ContainerStarted","Data":"035cb215e2733c570e904c2f01af5581d55b5cbba9a858b412367ebbe1ce286e"} Dec 13 07:13:46 crc kubenswrapper[4971]: I1213 07:13:46.572494 4971 generic.go:334] "Generic (PLEG): container finished" podID="1cd47626-171c-4c51-a572-3e9880154889" containerID="f2699dc331abef3e19d68cf1f4e9135033ee0c7ac8385176e02eea94bea1cdef" exitCode=1 Dec 13 07:13:46 crc kubenswrapper[4971]: I1213 07:13:46.572571 4971 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"1cd47626-171c-4c51-a572-3e9880154889","Type":"ContainerDied","Data":"f2699dc331abef3e19d68cf1f4e9135033ee0c7ac8385176e02eea94bea1cdef"} Dec 13 07:13:46 crc kubenswrapper[4971]: I1213 07:13:46.572695 4971 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="1cd47626-171c-4c51-a572-3e9880154889" containerName="ceilometer-central-agent" containerID="cri-o://df0910877e0c6048d15e192c1497a2330564ff6f2f4f3b2dc531118f87e55a02" gracePeriod=30 Dec 13 07:13:46 crc kubenswrapper[4971]: I1213 07:13:46.572821 4971 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="1cd47626-171c-4c51-a572-3e9880154889" containerName="sg-core" containerID="cri-o://628fca53bd17ddb88f713fab9756df4bfda8271a5f16a4176b53807106991e75" gracePeriod=30 Dec 13 07:13:46 crc kubenswrapper[4971]: I1213 07:13:46.572842 4971 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="1cd47626-171c-4c51-a572-3e9880154889" containerName="ceilometer-notification-agent" containerID="cri-o://2433437f27cac2201085434dc4550db63cdb41eb32c3f5ba0b33f5940980f2ae" gracePeriod=30 Dec 13 07:13:47 crc kubenswrapper[4971]: I1213 07:13:47.589376 4971 generic.go:334] "Generic (PLEG): container finished" podID="1cd47626-171c-4c51-a572-3e9880154889" containerID="628fca53bd17ddb88f713fab9756df4bfda8271a5f16a4176b53807106991e75" exitCode=2 Dec 13 07:13:47 crc kubenswrapper[4971]: I1213 07:13:47.589730 4971 generic.go:334] "Generic (PLEG): container finished" podID="1cd47626-171c-4c51-a572-3e9880154889" containerID="2433437f27cac2201085434dc4550db63cdb41eb32c3f5ba0b33f5940980f2ae" exitCode=0 Dec 13 07:13:47 crc kubenswrapper[4971]: I1213 07:13:47.589445 4971 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"1cd47626-171c-4c51-a572-3e9880154889","Type":"ContainerDied","Data":"628fca53bd17ddb88f713fab9756df4bfda8271a5f16a4176b53807106991e75"} Dec 13 07:13:47 crc kubenswrapper[4971]: I1213 07:13:47.589770 4971 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"1cd47626-171c-4c51-a572-3e9880154889","Type":"ContainerDied","Data":"2433437f27cac2201085434dc4550db63cdb41eb32c3f5ba0b33f5940980f2ae"} Dec 13 07:13:49 crc kubenswrapper[4971]: I1213 07:13:49.177476 4971 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Dec 13 07:13:49 crc kubenswrapper[4971]: I1213 07:13:49.292040 4971 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/1cd47626-171c-4c51-a572-3e9880154889-sg-core-conf-yaml\") pod \"1cd47626-171c-4c51-a572-3e9880154889\" (UID: \"1cd47626-171c-4c51-a572-3e9880154889\") " Dec 13 07:13:49 crc kubenswrapper[4971]: I1213 07:13:49.292442 4971 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/1cd47626-171c-4c51-a572-3e9880154889-run-httpd\") pod \"1cd47626-171c-4c51-a572-3e9880154889\" (UID: \"1cd47626-171c-4c51-a572-3e9880154889\") " Dec 13 07:13:49 crc kubenswrapper[4971]: I1213 07:13:49.292493 4971 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/1cd47626-171c-4c51-a572-3e9880154889-scripts\") pod \"1cd47626-171c-4c51-a572-3e9880154889\" (UID: \"1cd47626-171c-4c51-a572-3e9880154889\") " Dec 13 07:13:49 crc kubenswrapper[4971]: I1213 07:13:49.292552 4971 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-54rhn\" (UniqueName: \"kubernetes.io/projected/1cd47626-171c-4c51-a572-3e9880154889-kube-api-access-54rhn\") pod \"1cd47626-171c-4c51-a572-3e9880154889\" (UID: \"1cd47626-171c-4c51-a572-3e9880154889\") " Dec 13 07:13:49 crc kubenswrapper[4971]: I1213 07:13:49.292610 4971 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/1cd47626-171c-4c51-a572-3e9880154889-log-httpd\") pod \"1cd47626-171c-4c51-a572-3e9880154889\" (UID: \"1cd47626-171c-4c51-a572-3e9880154889\") " Dec 13 07:13:49 crc kubenswrapper[4971]: I1213 07:13:49.292647 4971 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/1cd47626-171c-4c51-a572-3e9880154889-config-data\") pod \"1cd47626-171c-4c51-a572-3e9880154889\" (UID: \"1cd47626-171c-4c51-a572-3e9880154889\") " Dec 13 07:13:49 crc kubenswrapper[4971]: I1213 07:13:49.292666 4971 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/1cd47626-171c-4c51-a572-3e9880154889-combined-ca-bundle\") pod \"1cd47626-171c-4c51-a572-3e9880154889\" (UID: \"1cd47626-171c-4c51-a572-3e9880154889\") " Dec 13 07:13:49 crc kubenswrapper[4971]: I1213 07:13:49.293109 4971 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/1cd47626-171c-4c51-a572-3e9880154889-run-httpd" (OuterVolumeSpecName: "run-httpd") pod "1cd47626-171c-4c51-a572-3e9880154889" (UID: "1cd47626-171c-4c51-a572-3e9880154889"). InnerVolumeSpecName "run-httpd". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 13 07:13:49 crc kubenswrapper[4971]: I1213 07:13:49.293833 4971 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/1cd47626-171c-4c51-a572-3e9880154889-log-httpd" (OuterVolumeSpecName: "log-httpd") pod "1cd47626-171c-4c51-a572-3e9880154889" (UID: "1cd47626-171c-4c51-a572-3e9880154889"). InnerVolumeSpecName "log-httpd". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 13 07:13:49 crc kubenswrapper[4971]: I1213 07:13:49.300119 4971 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/1cd47626-171c-4c51-a572-3e9880154889-kube-api-access-54rhn" (OuterVolumeSpecName: "kube-api-access-54rhn") pod "1cd47626-171c-4c51-a572-3e9880154889" (UID: "1cd47626-171c-4c51-a572-3e9880154889"). InnerVolumeSpecName "kube-api-access-54rhn". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 13 07:13:49 crc kubenswrapper[4971]: I1213 07:13:49.301961 4971 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/1cd47626-171c-4c51-a572-3e9880154889-scripts" (OuterVolumeSpecName: "scripts") pod "1cd47626-171c-4c51-a572-3e9880154889" (UID: "1cd47626-171c-4c51-a572-3e9880154889"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 13 07:13:49 crc kubenswrapper[4971]: I1213 07:13:49.350996 4971 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/1cd47626-171c-4c51-a572-3e9880154889-sg-core-conf-yaml" (OuterVolumeSpecName: "sg-core-conf-yaml") pod "1cd47626-171c-4c51-a572-3e9880154889" (UID: "1cd47626-171c-4c51-a572-3e9880154889"). InnerVolumeSpecName "sg-core-conf-yaml". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 13 07:13:49 crc kubenswrapper[4971]: I1213 07:13:49.395419 4971 reconciler_common.go:293] "Volume detached for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/1cd47626-171c-4c51-a572-3e9880154889-log-httpd\") on node \"crc\" DevicePath \"\"" Dec 13 07:13:49 crc kubenswrapper[4971]: I1213 07:13:49.395454 4971 reconciler_common.go:293] "Volume detached for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/1cd47626-171c-4c51-a572-3e9880154889-sg-core-conf-yaml\") on node \"crc\" DevicePath \"\"" Dec 13 07:13:49 crc kubenswrapper[4971]: I1213 07:13:49.395470 4971 reconciler_common.go:293] "Volume detached for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/1cd47626-171c-4c51-a572-3e9880154889-run-httpd\") on node \"crc\" DevicePath \"\"" Dec 13 07:13:49 crc kubenswrapper[4971]: I1213 07:13:49.395480 4971 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/1cd47626-171c-4c51-a572-3e9880154889-scripts\") on node \"crc\" DevicePath \"\"" Dec 13 07:13:49 crc kubenswrapper[4971]: I1213 07:13:49.395491 4971 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-54rhn\" (UniqueName: \"kubernetes.io/projected/1cd47626-171c-4c51-a572-3e9880154889-kube-api-access-54rhn\") on node \"crc\" DevicePath \"\"" Dec 13 07:13:49 crc kubenswrapper[4971]: I1213 07:13:49.406923 4971 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/1cd47626-171c-4c51-a572-3e9880154889-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "1cd47626-171c-4c51-a572-3e9880154889" (UID: "1cd47626-171c-4c51-a572-3e9880154889"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 13 07:13:49 crc kubenswrapper[4971]: I1213 07:13:49.426819 4971 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/1cd47626-171c-4c51-a572-3e9880154889-config-data" (OuterVolumeSpecName: "config-data") pod "1cd47626-171c-4c51-a572-3e9880154889" (UID: "1cd47626-171c-4c51-a572-3e9880154889"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 13 07:13:49 crc kubenswrapper[4971]: I1213 07:13:49.497084 4971 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/1cd47626-171c-4c51-a572-3e9880154889-config-data\") on node \"crc\" DevicePath \"\"" Dec 13 07:13:49 crc kubenswrapper[4971]: I1213 07:13:49.497123 4971 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/1cd47626-171c-4c51-a572-3e9880154889-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 13 07:13:49 crc kubenswrapper[4971]: I1213 07:13:49.617172 4971 generic.go:334] "Generic (PLEG): container finished" podID="1cd47626-171c-4c51-a572-3e9880154889" containerID="df0910877e0c6048d15e192c1497a2330564ff6f2f4f3b2dc531118f87e55a02" exitCode=0 Dec 13 07:13:49 crc kubenswrapper[4971]: I1213 07:13:49.617215 4971 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"1cd47626-171c-4c51-a572-3e9880154889","Type":"ContainerDied","Data":"df0910877e0c6048d15e192c1497a2330564ff6f2f4f3b2dc531118f87e55a02"} Dec 13 07:13:49 crc kubenswrapper[4971]: I1213 07:13:49.617245 4971 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"1cd47626-171c-4c51-a572-3e9880154889","Type":"ContainerDied","Data":"50fb0c503e99053d2c7f20d771dec4bac9aabb054953ccb12bff0a6fdcdbf963"} Dec 13 07:13:49 crc kubenswrapper[4971]: I1213 07:13:49.617307 4971 scope.go:117] "RemoveContainer" containerID="f2699dc331abef3e19d68cf1f4e9135033ee0c7ac8385176e02eea94bea1cdef" Dec 13 07:13:49 crc kubenswrapper[4971]: I1213 07:13:49.617474 4971 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Dec 13 07:13:49 crc kubenswrapper[4971]: I1213 07:13:49.642211 4971 scope.go:117] "RemoveContainer" containerID="628fca53bd17ddb88f713fab9756df4bfda8271a5f16a4176b53807106991e75" Dec 13 07:13:49 crc kubenswrapper[4971]: I1213 07:13:49.663721 4971 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ceilometer-0"] Dec 13 07:13:49 crc kubenswrapper[4971]: I1213 07:13:49.676557 4971 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/ceilometer-0"] Dec 13 07:13:49 crc kubenswrapper[4971]: I1213 07:13:49.677714 4971 scope.go:117] "RemoveContainer" containerID="2433437f27cac2201085434dc4550db63cdb41eb32c3f5ba0b33f5940980f2ae" Dec 13 07:13:49 crc kubenswrapper[4971]: I1213 07:13:49.690361 4971 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ceilometer-0"] Dec 13 07:13:49 crc kubenswrapper[4971]: E1213 07:13:49.690784 4971 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="1cd47626-171c-4c51-a572-3e9880154889" containerName="ceilometer-notification-agent" Dec 13 07:13:49 crc kubenswrapper[4971]: I1213 07:13:49.690799 4971 state_mem.go:107] "Deleted CPUSet assignment" podUID="1cd47626-171c-4c51-a572-3e9880154889" containerName="ceilometer-notification-agent" Dec 13 07:13:49 crc kubenswrapper[4971]: E1213 07:13:49.690815 4971 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="1cd47626-171c-4c51-a572-3e9880154889" containerName="proxy-httpd" Dec 13 07:13:49 crc kubenswrapper[4971]: I1213 07:13:49.690821 4971 state_mem.go:107] "Deleted CPUSet assignment" podUID="1cd47626-171c-4c51-a572-3e9880154889" containerName="proxy-httpd" Dec 13 07:13:49 crc kubenswrapper[4971]: E1213 07:13:49.690834 4971 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="1cd47626-171c-4c51-a572-3e9880154889" containerName="ceilometer-central-agent" Dec 13 07:13:49 crc kubenswrapper[4971]: I1213 07:13:49.690842 4971 state_mem.go:107] "Deleted CPUSet assignment" podUID="1cd47626-171c-4c51-a572-3e9880154889" containerName="ceilometer-central-agent" Dec 13 07:13:49 crc kubenswrapper[4971]: E1213 07:13:49.690857 4971 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="1cd47626-171c-4c51-a572-3e9880154889" containerName="sg-core" Dec 13 07:13:49 crc kubenswrapper[4971]: I1213 07:13:49.690862 4971 state_mem.go:107] "Deleted CPUSet assignment" podUID="1cd47626-171c-4c51-a572-3e9880154889" containerName="sg-core" Dec 13 07:13:49 crc kubenswrapper[4971]: I1213 07:13:49.691032 4971 memory_manager.go:354] "RemoveStaleState removing state" podUID="1cd47626-171c-4c51-a572-3e9880154889" containerName="proxy-httpd" Dec 13 07:13:49 crc kubenswrapper[4971]: I1213 07:13:49.691047 4971 memory_manager.go:354] "RemoveStaleState removing state" podUID="1cd47626-171c-4c51-a572-3e9880154889" containerName="sg-core" Dec 13 07:13:49 crc kubenswrapper[4971]: I1213 07:13:49.691062 4971 memory_manager.go:354] "RemoveStaleState removing state" podUID="1cd47626-171c-4c51-a572-3e9880154889" containerName="ceilometer-central-agent" Dec 13 07:13:49 crc kubenswrapper[4971]: I1213 07:13:49.691072 4971 memory_manager.go:354] "RemoveStaleState removing state" podUID="1cd47626-171c-4c51-a572-3e9880154889" containerName="ceilometer-notification-agent" Dec 13 07:13:49 crc kubenswrapper[4971]: I1213 07:13:49.695204 4971 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Dec 13 07:13:49 crc kubenswrapper[4971]: I1213 07:13:49.697578 4971 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-ceilometer-internal-svc" Dec 13 07:13:49 crc kubenswrapper[4971]: I1213 07:13:49.697644 4971 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceilometer-config-data" Dec 13 07:13:49 crc kubenswrapper[4971]: I1213 07:13:49.697821 4971 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceilometer-scripts" Dec 13 07:13:49 crc kubenswrapper[4971]: I1213 07:13:49.713845 4971 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ceilometer-0"] Dec 13 07:13:49 crc kubenswrapper[4971]: I1213 07:13:49.753071 4971 scope.go:117] "RemoveContainer" containerID="df0910877e0c6048d15e192c1497a2330564ff6f2f4f3b2dc531118f87e55a02" Dec 13 07:13:49 crc kubenswrapper[4971]: I1213 07:13:49.781436 4971 scope.go:117] "RemoveContainer" containerID="f2699dc331abef3e19d68cf1f4e9135033ee0c7ac8385176e02eea94bea1cdef" Dec 13 07:13:49 crc kubenswrapper[4971]: E1213 07:13:49.781831 4971 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"f2699dc331abef3e19d68cf1f4e9135033ee0c7ac8385176e02eea94bea1cdef\": container with ID starting with f2699dc331abef3e19d68cf1f4e9135033ee0c7ac8385176e02eea94bea1cdef not found: ID does not exist" containerID="f2699dc331abef3e19d68cf1f4e9135033ee0c7ac8385176e02eea94bea1cdef" Dec 13 07:13:49 crc kubenswrapper[4971]: I1213 07:13:49.781859 4971 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"f2699dc331abef3e19d68cf1f4e9135033ee0c7ac8385176e02eea94bea1cdef"} err="failed to get container status \"f2699dc331abef3e19d68cf1f4e9135033ee0c7ac8385176e02eea94bea1cdef\": rpc error: code = NotFound desc = could not find container \"f2699dc331abef3e19d68cf1f4e9135033ee0c7ac8385176e02eea94bea1cdef\": container with ID starting with f2699dc331abef3e19d68cf1f4e9135033ee0c7ac8385176e02eea94bea1cdef not found: ID does not exist" Dec 13 07:13:49 crc kubenswrapper[4971]: I1213 07:13:49.781883 4971 scope.go:117] "RemoveContainer" containerID="628fca53bd17ddb88f713fab9756df4bfda8271a5f16a4176b53807106991e75" Dec 13 07:13:49 crc kubenswrapper[4971]: E1213 07:13:49.782090 4971 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"628fca53bd17ddb88f713fab9756df4bfda8271a5f16a4176b53807106991e75\": container with ID starting with 628fca53bd17ddb88f713fab9756df4bfda8271a5f16a4176b53807106991e75 not found: ID does not exist" containerID="628fca53bd17ddb88f713fab9756df4bfda8271a5f16a4176b53807106991e75" Dec 13 07:13:49 crc kubenswrapper[4971]: I1213 07:13:49.782121 4971 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"628fca53bd17ddb88f713fab9756df4bfda8271a5f16a4176b53807106991e75"} err="failed to get container status \"628fca53bd17ddb88f713fab9756df4bfda8271a5f16a4176b53807106991e75\": rpc error: code = NotFound desc = could not find container \"628fca53bd17ddb88f713fab9756df4bfda8271a5f16a4176b53807106991e75\": container with ID starting with 628fca53bd17ddb88f713fab9756df4bfda8271a5f16a4176b53807106991e75 not found: ID does not exist" Dec 13 07:13:49 crc kubenswrapper[4971]: I1213 07:13:49.782137 4971 scope.go:117] "RemoveContainer" containerID="2433437f27cac2201085434dc4550db63cdb41eb32c3f5ba0b33f5940980f2ae" Dec 13 07:13:49 crc kubenswrapper[4971]: E1213 07:13:49.782391 4971 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"2433437f27cac2201085434dc4550db63cdb41eb32c3f5ba0b33f5940980f2ae\": container with ID starting with 2433437f27cac2201085434dc4550db63cdb41eb32c3f5ba0b33f5940980f2ae not found: ID does not exist" containerID="2433437f27cac2201085434dc4550db63cdb41eb32c3f5ba0b33f5940980f2ae" Dec 13 07:13:49 crc kubenswrapper[4971]: I1213 07:13:49.782412 4971 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"2433437f27cac2201085434dc4550db63cdb41eb32c3f5ba0b33f5940980f2ae"} err="failed to get container status \"2433437f27cac2201085434dc4550db63cdb41eb32c3f5ba0b33f5940980f2ae\": rpc error: code = NotFound desc = could not find container \"2433437f27cac2201085434dc4550db63cdb41eb32c3f5ba0b33f5940980f2ae\": container with ID starting with 2433437f27cac2201085434dc4550db63cdb41eb32c3f5ba0b33f5940980f2ae not found: ID does not exist" Dec 13 07:13:49 crc kubenswrapper[4971]: I1213 07:13:49.782424 4971 scope.go:117] "RemoveContainer" containerID="df0910877e0c6048d15e192c1497a2330564ff6f2f4f3b2dc531118f87e55a02" Dec 13 07:13:49 crc kubenswrapper[4971]: E1213 07:13:49.782734 4971 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"df0910877e0c6048d15e192c1497a2330564ff6f2f4f3b2dc531118f87e55a02\": container with ID starting with df0910877e0c6048d15e192c1497a2330564ff6f2f4f3b2dc531118f87e55a02 not found: ID does not exist" containerID="df0910877e0c6048d15e192c1497a2330564ff6f2f4f3b2dc531118f87e55a02" Dec 13 07:13:49 crc kubenswrapper[4971]: I1213 07:13:49.782767 4971 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"df0910877e0c6048d15e192c1497a2330564ff6f2f4f3b2dc531118f87e55a02"} err="failed to get container status \"df0910877e0c6048d15e192c1497a2330564ff6f2f4f3b2dc531118f87e55a02\": rpc error: code = NotFound desc = could not find container \"df0910877e0c6048d15e192c1497a2330564ff6f2f4f3b2dc531118f87e55a02\": container with ID starting with df0910877e0c6048d15e192c1497a2330564ff6f2f4f3b2dc531118f87e55a02 not found: ID does not exist" Dec 13 07:13:49 crc kubenswrapper[4971]: I1213 07:13:49.791380 4971 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="1cd47626-171c-4c51-a572-3e9880154889" path="/var/lib/kubelet/pods/1cd47626-171c-4c51-a572-3e9880154889/volumes" Dec 13 07:13:49 crc kubenswrapper[4971]: I1213 07:13:49.805914 4971 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-gzbh4\" (UniqueName: \"kubernetes.io/projected/80fb658f-1dbc-42a8-a888-11afedfd3bba-kube-api-access-gzbh4\") pod \"ceilometer-0\" (UID: \"80fb658f-1dbc-42a8-a888-11afedfd3bba\") " pod="openstack/ceilometer-0" Dec 13 07:13:49 crc kubenswrapper[4971]: I1213 07:13:49.805978 4971 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/80fb658f-1dbc-42a8-a888-11afedfd3bba-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"80fb658f-1dbc-42a8-a888-11afedfd3bba\") " pod="openstack/ceilometer-0" Dec 13 07:13:49 crc kubenswrapper[4971]: I1213 07:13:49.806018 4971 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/80fb658f-1dbc-42a8-a888-11afedfd3bba-log-httpd\") pod \"ceilometer-0\" (UID: \"80fb658f-1dbc-42a8-a888-11afedfd3bba\") " pod="openstack/ceilometer-0" Dec 13 07:13:49 crc kubenswrapper[4971]: I1213 07:13:49.806040 4971 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/80fb658f-1dbc-42a8-a888-11afedfd3bba-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"80fb658f-1dbc-42a8-a888-11afedfd3bba\") " pod="openstack/ceilometer-0" Dec 13 07:13:49 crc kubenswrapper[4971]: I1213 07:13:49.806059 4971 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ceilometer-tls-certs\" (UniqueName: \"kubernetes.io/secret/80fb658f-1dbc-42a8-a888-11afedfd3bba-ceilometer-tls-certs\") pod \"ceilometer-0\" (UID: \"80fb658f-1dbc-42a8-a888-11afedfd3bba\") " pod="openstack/ceilometer-0" Dec 13 07:13:49 crc kubenswrapper[4971]: I1213 07:13:49.806100 4971 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/80fb658f-1dbc-42a8-a888-11afedfd3bba-scripts\") pod \"ceilometer-0\" (UID: \"80fb658f-1dbc-42a8-a888-11afedfd3bba\") " pod="openstack/ceilometer-0" Dec 13 07:13:49 crc kubenswrapper[4971]: I1213 07:13:49.806130 4971 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/80fb658f-1dbc-42a8-a888-11afedfd3bba-run-httpd\") pod \"ceilometer-0\" (UID: \"80fb658f-1dbc-42a8-a888-11afedfd3bba\") " pod="openstack/ceilometer-0" Dec 13 07:13:49 crc kubenswrapper[4971]: I1213 07:13:49.806144 4971 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/80fb658f-1dbc-42a8-a888-11afedfd3bba-config-data\") pod \"ceilometer-0\" (UID: \"80fb658f-1dbc-42a8-a888-11afedfd3bba\") " pod="openstack/ceilometer-0" Dec 13 07:13:49 crc kubenswrapper[4971]: I1213 07:13:49.907449 4971 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-gzbh4\" (UniqueName: \"kubernetes.io/projected/80fb658f-1dbc-42a8-a888-11afedfd3bba-kube-api-access-gzbh4\") pod \"ceilometer-0\" (UID: \"80fb658f-1dbc-42a8-a888-11afedfd3bba\") " pod="openstack/ceilometer-0" Dec 13 07:13:49 crc kubenswrapper[4971]: I1213 07:13:49.907540 4971 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/80fb658f-1dbc-42a8-a888-11afedfd3bba-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"80fb658f-1dbc-42a8-a888-11afedfd3bba\") " pod="openstack/ceilometer-0" Dec 13 07:13:49 crc kubenswrapper[4971]: I1213 07:13:49.907626 4971 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/80fb658f-1dbc-42a8-a888-11afedfd3bba-log-httpd\") pod \"ceilometer-0\" (UID: \"80fb658f-1dbc-42a8-a888-11afedfd3bba\") " pod="openstack/ceilometer-0" Dec 13 07:13:49 crc kubenswrapper[4971]: I1213 07:13:49.907659 4971 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/80fb658f-1dbc-42a8-a888-11afedfd3bba-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"80fb658f-1dbc-42a8-a888-11afedfd3bba\") " pod="openstack/ceilometer-0" Dec 13 07:13:49 crc kubenswrapper[4971]: I1213 07:13:49.907684 4971 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ceilometer-tls-certs\" (UniqueName: \"kubernetes.io/secret/80fb658f-1dbc-42a8-a888-11afedfd3bba-ceilometer-tls-certs\") pod \"ceilometer-0\" (UID: \"80fb658f-1dbc-42a8-a888-11afedfd3bba\") " pod="openstack/ceilometer-0" Dec 13 07:13:49 crc kubenswrapper[4971]: I1213 07:13:49.907764 4971 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/80fb658f-1dbc-42a8-a888-11afedfd3bba-scripts\") pod \"ceilometer-0\" (UID: \"80fb658f-1dbc-42a8-a888-11afedfd3bba\") " pod="openstack/ceilometer-0" Dec 13 07:13:49 crc kubenswrapper[4971]: I1213 07:13:49.907842 4971 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/80fb658f-1dbc-42a8-a888-11afedfd3bba-config-data\") pod \"ceilometer-0\" (UID: \"80fb658f-1dbc-42a8-a888-11afedfd3bba\") " pod="openstack/ceilometer-0" Dec 13 07:13:49 crc kubenswrapper[4971]: I1213 07:13:49.907864 4971 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/80fb658f-1dbc-42a8-a888-11afedfd3bba-run-httpd\") pod \"ceilometer-0\" (UID: \"80fb658f-1dbc-42a8-a888-11afedfd3bba\") " pod="openstack/ceilometer-0" Dec 13 07:13:49 crc kubenswrapper[4971]: I1213 07:13:49.911193 4971 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/80fb658f-1dbc-42a8-a888-11afedfd3bba-run-httpd\") pod \"ceilometer-0\" (UID: \"80fb658f-1dbc-42a8-a888-11afedfd3bba\") " pod="openstack/ceilometer-0" Dec 13 07:13:49 crc kubenswrapper[4971]: I1213 07:13:49.911251 4971 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/80fb658f-1dbc-42a8-a888-11afedfd3bba-log-httpd\") pod \"ceilometer-0\" (UID: \"80fb658f-1dbc-42a8-a888-11afedfd3bba\") " pod="openstack/ceilometer-0" Dec 13 07:13:49 crc kubenswrapper[4971]: I1213 07:13:49.916401 4971 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/80fb658f-1dbc-42a8-a888-11afedfd3bba-scripts\") pod \"ceilometer-0\" (UID: \"80fb658f-1dbc-42a8-a888-11afedfd3bba\") " pod="openstack/ceilometer-0" Dec 13 07:13:49 crc kubenswrapper[4971]: I1213 07:13:49.916739 4971 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/80fb658f-1dbc-42a8-a888-11afedfd3bba-config-data\") pod \"ceilometer-0\" (UID: \"80fb658f-1dbc-42a8-a888-11afedfd3bba\") " pod="openstack/ceilometer-0" Dec 13 07:13:49 crc kubenswrapper[4971]: I1213 07:13:49.920104 4971 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ceilometer-tls-certs\" (UniqueName: \"kubernetes.io/secret/80fb658f-1dbc-42a8-a888-11afedfd3bba-ceilometer-tls-certs\") pod \"ceilometer-0\" (UID: \"80fb658f-1dbc-42a8-a888-11afedfd3bba\") " pod="openstack/ceilometer-0" Dec 13 07:13:49 crc kubenswrapper[4971]: I1213 07:13:49.920841 4971 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/80fb658f-1dbc-42a8-a888-11afedfd3bba-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"80fb658f-1dbc-42a8-a888-11afedfd3bba\") " pod="openstack/ceilometer-0" Dec 13 07:13:49 crc kubenswrapper[4971]: I1213 07:13:49.928916 4971 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/80fb658f-1dbc-42a8-a888-11afedfd3bba-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"80fb658f-1dbc-42a8-a888-11afedfd3bba\") " pod="openstack/ceilometer-0" Dec 13 07:13:49 crc kubenswrapper[4971]: I1213 07:13:49.931714 4971 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-gzbh4\" (UniqueName: \"kubernetes.io/projected/80fb658f-1dbc-42a8-a888-11afedfd3bba-kube-api-access-gzbh4\") pod \"ceilometer-0\" (UID: \"80fb658f-1dbc-42a8-a888-11afedfd3bba\") " pod="openstack/ceilometer-0" Dec 13 07:13:50 crc kubenswrapper[4971]: I1213 07:13:50.081123 4971 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Dec 13 07:13:52 crc kubenswrapper[4971]: I1213 07:13:52.925254 4971 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/kube-state-metrics-0" Dec 13 07:13:53 crc kubenswrapper[4971]: I1213 07:13:53.389645 4971 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ceilometer-0"] Dec 13 07:14:00 crc kubenswrapper[4971]: E1213 07:14:00.345168 4971 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying config: context canceled" image="quay.io/podified-antelope-centos9/openstack-nova-conductor@sha256:22f097cb86b28ac48dc670ed7e0e841280bef1608f11b2b4536fbc2d2a6a90be" Dec 13 07:14:00 crc kubenswrapper[4971]: E1213 07:14:00.346323 4971 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:nova-cell0-conductor-db-sync,Image:quay.io/podified-antelope-centos9/openstack-nova-conductor@sha256:22f097cb86b28ac48dc670ed7e0e841280bef1608f11b2b4536fbc2d2a6a90be,Command:[/bin/bash],Args:[-c /usr/local/bin/kolla_start],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:CELL_NAME,Value:cell0,ValueFrom:nil,},EnvVar{Name:KOLLA_BOOTSTRAP,Value:true,ValueFrom:nil,},EnvVar{Name:KOLLA_CONFIG_STRATEGY,Value:COPY_ALWAYS,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:config-data,ReadOnly:false,MountPath:/var/lib/openstack/config,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:scripts,ReadOnly:false,MountPath:/var/lib/openstack/bin,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:config-data,ReadOnly:false,MountPath:/var/lib/kolla/config_files/config.json,SubPath:nova-conductor-dbsync-config.json,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:combined-ca-bundle,ReadOnly:true,MountPath:/etc/pki/ca-trust/extracted/pem/tls-ca-bundle.pem,SubPath:tls-ca-bundle.pem,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-t8b6j,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[MKNOD],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*42436,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:nil,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod nova-cell0-conductor-db-sync-g9t27_openstack(31dddcf3-380b-4003-a766-88d4f5bc00de): ErrImagePull: rpc error: code = Canceled desc = copying config: context canceled" logger="UnhandledError" Dec 13 07:14:00 crc kubenswrapper[4971]: E1213 07:14:00.348398 4971 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"nova-cell0-conductor-db-sync\" with ErrImagePull: \"rpc error: code = Canceled desc = copying config: context canceled\"" pod="openstack/nova-cell0-conductor-db-sync-g9t27" podUID="31dddcf3-380b-4003-a766-88d4f5bc00de" Dec 13 07:14:00 crc kubenswrapper[4971]: E1213 07:14:00.717138 4971 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"nova-cell0-conductor-db-sync\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/podified-antelope-centos9/openstack-nova-conductor@sha256:22f097cb86b28ac48dc670ed7e0e841280bef1608f11b2b4536fbc2d2a6a90be\\\"\"" pod="openstack/nova-cell0-conductor-db-sync-g9t27" podUID="31dddcf3-380b-4003-a766-88d4f5bc00de" Dec 13 07:14:00 crc kubenswrapper[4971]: I1213 07:14:00.765222 4971 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ceilometer-0"] Dec 13 07:14:01 crc kubenswrapper[4971]: I1213 07:14:01.723182 4971 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"80fb658f-1dbc-42a8-a888-11afedfd3bba","Type":"ContainerStarted","Data":"eb8a1becb80414dc0f287dd72df7dfb79b4fb5711921166a579e34af18c3ca7d"} Dec 13 07:14:02 crc kubenswrapper[4971]: I1213 07:14:02.732986 4971 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"80fb658f-1dbc-42a8-a888-11afedfd3bba","Type":"ContainerStarted","Data":"f119ef3c07258d4dee06c12c543c7b23e4386958785adfb6524ba0662f06b280"} Dec 13 07:14:03 crc kubenswrapper[4971]: I1213 07:14:03.744122 4971 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"80fb658f-1dbc-42a8-a888-11afedfd3bba","Type":"ContainerStarted","Data":"9009ff428ac67061fac12bd7f9c52bdb707cbe1f8d74772a4533bb05661e33c7"} Dec 13 07:14:05 crc kubenswrapper[4971]: I1213 07:14:05.762392 4971 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"80fb658f-1dbc-42a8-a888-11afedfd3bba","Type":"ContainerStarted","Data":"4aa2ebd15a1835190cee6f8d1b95e42af2d3d12050eeac717d8e111cd9295a2a"} Dec 13 07:14:07 crc kubenswrapper[4971]: I1213 07:14:07.786891 4971 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"80fb658f-1dbc-42a8-a888-11afedfd3bba","Type":"ContainerStarted","Data":"71fad303cda2bd73dcc3083f468961d98f8172537240aeca396bb177d8650254"} Dec 13 07:14:07 crc kubenswrapper[4971]: I1213 07:14:07.787390 4971 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/ceilometer-0" Dec 13 07:14:07 crc kubenswrapper[4971]: I1213 07:14:07.787054 4971 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="80fb658f-1dbc-42a8-a888-11afedfd3bba" containerName="ceilometer-central-agent" containerID="cri-o://f119ef3c07258d4dee06c12c543c7b23e4386958785adfb6524ba0662f06b280" gracePeriod=30 Dec 13 07:14:07 crc kubenswrapper[4971]: I1213 07:14:07.787073 4971 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="80fb658f-1dbc-42a8-a888-11afedfd3bba" containerName="proxy-httpd" containerID="cri-o://71fad303cda2bd73dcc3083f468961d98f8172537240aeca396bb177d8650254" gracePeriod=30 Dec 13 07:14:07 crc kubenswrapper[4971]: I1213 07:14:07.787146 4971 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="80fb658f-1dbc-42a8-a888-11afedfd3bba" containerName="ceilometer-notification-agent" containerID="cri-o://9009ff428ac67061fac12bd7f9c52bdb707cbe1f8d74772a4533bb05661e33c7" gracePeriod=30 Dec 13 07:14:07 crc kubenswrapper[4971]: I1213 07:14:07.791030 4971 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="80fb658f-1dbc-42a8-a888-11afedfd3bba" containerName="sg-core" containerID="cri-o://4aa2ebd15a1835190cee6f8d1b95e42af2d3d12050eeac717d8e111cd9295a2a" gracePeriod=30 Dec 13 07:14:07 crc kubenswrapper[4971]: I1213 07:14:07.823204 4971 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/ceilometer-0" podStartSLOduration=12.339756298 podStartE2EDuration="18.823183597s" podCreationTimestamp="2025-12-13 07:13:49 +0000 UTC" firstStartedPulling="2025-12-13 07:14:00.771926227 +0000 UTC m=+1497.376335675" lastFinishedPulling="2025-12-13 07:14:07.255353536 +0000 UTC m=+1503.859762974" observedRunningTime="2025-12-13 07:14:07.81562587 +0000 UTC m=+1504.420035318" watchObservedRunningTime="2025-12-13 07:14:07.823183597 +0000 UTC m=+1504.427593055" Dec 13 07:14:08 crc kubenswrapper[4971]: I1213 07:14:08.801724 4971 generic.go:334] "Generic (PLEG): container finished" podID="80fb658f-1dbc-42a8-a888-11afedfd3bba" containerID="71fad303cda2bd73dcc3083f468961d98f8172537240aeca396bb177d8650254" exitCode=0 Dec 13 07:14:08 crc kubenswrapper[4971]: I1213 07:14:08.801987 4971 generic.go:334] "Generic (PLEG): container finished" podID="80fb658f-1dbc-42a8-a888-11afedfd3bba" containerID="4aa2ebd15a1835190cee6f8d1b95e42af2d3d12050eeac717d8e111cd9295a2a" exitCode=2 Dec 13 07:14:08 crc kubenswrapper[4971]: I1213 07:14:08.801813 4971 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"80fb658f-1dbc-42a8-a888-11afedfd3bba","Type":"ContainerDied","Data":"71fad303cda2bd73dcc3083f468961d98f8172537240aeca396bb177d8650254"} Dec 13 07:14:08 crc kubenswrapper[4971]: I1213 07:14:08.802031 4971 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"80fb658f-1dbc-42a8-a888-11afedfd3bba","Type":"ContainerDied","Data":"4aa2ebd15a1835190cee6f8d1b95e42af2d3d12050eeac717d8e111cd9295a2a"} Dec 13 07:14:09 crc kubenswrapper[4971]: I1213 07:14:09.829304 4971 generic.go:334] "Generic (PLEG): container finished" podID="80fb658f-1dbc-42a8-a888-11afedfd3bba" containerID="9009ff428ac67061fac12bd7f9c52bdb707cbe1f8d74772a4533bb05661e33c7" exitCode=0 Dec 13 07:14:09 crc kubenswrapper[4971]: I1213 07:14:09.829360 4971 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"80fb658f-1dbc-42a8-a888-11afedfd3bba","Type":"ContainerDied","Data":"9009ff428ac67061fac12bd7f9c52bdb707cbe1f8d74772a4533bb05661e33c7"} Dec 13 07:14:10 crc kubenswrapper[4971]: I1213 07:14:10.842493 4971 generic.go:334] "Generic (PLEG): container finished" podID="80fb658f-1dbc-42a8-a888-11afedfd3bba" containerID="f119ef3c07258d4dee06c12c543c7b23e4386958785adfb6524ba0662f06b280" exitCode=0 Dec 13 07:14:10 crc kubenswrapper[4971]: I1213 07:14:10.842641 4971 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"80fb658f-1dbc-42a8-a888-11afedfd3bba","Type":"ContainerDied","Data":"f119ef3c07258d4dee06c12c543c7b23e4386958785adfb6524ba0662f06b280"} Dec 13 07:14:11 crc kubenswrapper[4971]: I1213 07:14:11.360830 4971 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Dec 13 07:14:11 crc kubenswrapper[4971]: I1213 07:14:11.506565 4971 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/80fb658f-1dbc-42a8-a888-11afedfd3bba-log-httpd\") pod \"80fb658f-1dbc-42a8-a888-11afedfd3bba\" (UID: \"80fb658f-1dbc-42a8-a888-11afedfd3bba\") " Dec 13 07:14:11 crc kubenswrapper[4971]: I1213 07:14:11.506692 4971 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/80fb658f-1dbc-42a8-a888-11afedfd3bba-combined-ca-bundle\") pod \"80fb658f-1dbc-42a8-a888-11afedfd3bba\" (UID: \"80fb658f-1dbc-42a8-a888-11afedfd3bba\") " Dec 13 07:14:11 crc kubenswrapper[4971]: I1213 07:14:11.506797 4971 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-gzbh4\" (UniqueName: \"kubernetes.io/projected/80fb658f-1dbc-42a8-a888-11afedfd3bba-kube-api-access-gzbh4\") pod \"80fb658f-1dbc-42a8-a888-11afedfd3bba\" (UID: \"80fb658f-1dbc-42a8-a888-11afedfd3bba\") " Dec 13 07:14:11 crc kubenswrapper[4971]: I1213 07:14:11.506821 4971 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ceilometer-tls-certs\" (UniqueName: \"kubernetes.io/secret/80fb658f-1dbc-42a8-a888-11afedfd3bba-ceilometer-tls-certs\") pod \"80fb658f-1dbc-42a8-a888-11afedfd3bba\" (UID: \"80fb658f-1dbc-42a8-a888-11afedfd3bba\") " Dec 13 07:14:11 crc kubenswrapper[4971]: I1213 07:14:11.506903 4971 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/80fb658f-1dbc-42a8-a888-11afedfd3bba-config-data\") pod \"80fb658f-1dbc-42a8-a888-11afedfd3bba\" (UID: \"80fb658f-1dbc-42a8-a888-11afedfd3bba\") " Dec 13 07:14:11 crc kubenswrapper[4971]: I1213 07:14:11.506977 4971 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/80fb658f-1dbc-42a8-a888-11afedfd3bba-sg-core-conf-yaml\") pod \"80fb658f-1dbc-42a8-a888-11afedfd3bba\" (UID: \"80fb658f-1dbc-42a8-a888-11afedfd3bba\") " Dec 13 07:14:11 crc kubenswrapper[4971]: I1213 07:14:11.507144 4971 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/80fb658f-1dbc-42a8-a888-11afedfd3bba-log-httpd" (OuterVolumeSpecName: "log-httpd") pod "80fb658f-1dbc-42a8-a888-11afedfd3bba" (UID: "80fb658f-1dbc-42a8-a888-11afedfd3bba"). InnerVolumeSpecName "log-httpd". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 13 07:14:11 crc kubenswrapper[4971]: I1213 07:14:11.507720 4971 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/80fb658f-1dbc-42a8-a888-11afedfd3bba-scripts\") pod \"80fb658f-1dbc-42a8-a888-11afedfd3bba\" (UID: \"80fb658f-1dbc-42a8-a888-11afedfd3bba\") " Dec 13 07:14:11 crc kubenswrapper[4971]: I1213 07:14:11.507766 4971 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/80fb658f-1dbc-42a8-a888-11afedfd3bba-run-httpd\") pod \"80fb658f-1dbc-42a8-a888-11afedfd3bba\" (UID: \"80fb658f-1dbc-42a8-a888-11afedfd3bba\") " Dec 13 07:14:11 crc kubenswrapper[4971]: I1213 07:14:11.508124 4971 reconciler_common.go:293] "Volume detached for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/80fb658f-1dbc-42a8-a888-11afedfd3bba-log-httpd\") on node \"crc\" DevicePath \"\"" Dec 13 07:14:11 crc kubenswrapper[4971]: I1213 07:14:11.508196 4971 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/80fb658f-1dbc-42a8-a888-11afedfd3bba-run-httpd" (OuterVolumeSpecName: "run-httpd") pod "80fb658f-1dbc-42a8-a888-11afedfd3bba" (UID: "80fb658f-1dbc-42a8-a888-11afedfd3bba"). InnerVolumeSpecName "run-httpd". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 13 07:14:11 crc kubenswrapper[4971]: I1213 07:14:11.513717 4971 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/80fb658f-1dbc-42a8-a888-11afedfd3bba-scripts" (OuterVolumeSpecName: "scripts") pod "80fb658f-1dbc-42a8-a888-11afedfd3bba" (UID: "80fb658f-1dbc-42a8-a888-11afedfd3bba"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 13 07:14:11 crc kubenswrapper[4971]: I1213 07:14:11.513870 4971 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/80fb658f-1dbc-42a8-a888-11afedfd3bba-kube-api-access-gzbh4" (OuterVolumeSpecName: "kube-api-access-gzbh4") pod "80fb658f-1dbc-42a8-a888-11afedfd3bba" (UID: "80fb658f-1dbc-42a8-a888-11afedfd3bba"). InnerVolumeSpecName "kube-api-access-gzbh4". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 13 07:14:11 crc kubenswrapper[4971]: I1213 07:14:11.538218 4971 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/80fb658f-1dbc-42a8-a888-11afedfd3bba-sg-core-conf-yaml" (OuterVolumeSpecName: "sg-core-conf-yaml") pod "80fb658f-1dbc-42a8-a888-11afedfd3bba" (UID: "80fb658f-1dbc-42a8-a888-11afedfd3bba"). InnerVolumeSpecName "sg-core-conf-yaml". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 13 07:14:11 crc kubenswrapper[4971]: I1213 07:14:11.560613 4971 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/80fb658f-1dbc-42a8-a888-11afedfd3bba-ceilometer-tls-certs" (OuterVolumeSpecName: "ceilometer-tls-certs") pod "80fb658f-1dbc-42a8-a888-11afedfd3bba" (UID: "80fb658f-1dbc-42a8-a888-11afedfd3bba"). InnerVolumeSpecName "ceilometer-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 13 07:14:11 crc kubenswrapper[4971]: I1213 07:14:11.580235 4971 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/80fb658f-1dbc-42a8-a888-11afedfd3bba-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "80fb658f-1dbc-42a8-a888-11afedfd3bba" (UID: "80fb658f-1dbc-42a8-a888-11afedfd3bba"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 13 07:14:11 crc kubenswrapper[4971]: I1213 07:14:11.602921 4971 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/80fb658f-1dbc-42a8-a888-11afedfd3bba-config-data" (OuterVolumeSpecName: "config-data") pod "80fb658f-1dbc-42a8-a888-11afedfd3bba" (UID: "80fb658f-1dbc-42a8-a888-11afedfd3bba"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 13 07:14:11 crc kubenswrapper[4971]: I1213 07:14:11.610052 4971 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/80fb658f-1dbc-42a8-a888-11afedfd3bba-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 13 07:14:11 crc kubenswrapper[4971]: I1213 07:14:11.610090 4971 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-gzbh4\" (UniqueName: \"kubernetes.io/projected/80fb658f-1dbc-42a8-a888-11afedfd3bba-kube-api-access-gzbh4\") on node \"crc\" DevicePath \"\"" Dec 13 07:14:11 crc kubenswrapper[4971]: I1213 07:14:11.610103 4971 reconciler_common.go:293] "Volume detached for volume \"ceilometer-tls-certs\" (UniqueName: \"kubernetes.io/secret/80fb658f-1dbc-42a8-a888-11afedfd3bba-ceilometer-tls-certs\") on node \"crc\" DevicePath \"\"" Dec 13 07:14:11 crc kubenswrapper[4971]: I1213 07:14:11.610112 4971 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/80fb658f-1dbc-42a8-a888-11afedfd3bba-config-data\") on node \"crc\" DevicePath \"\"" Dec 13 07:14:11 crc kubenswrapper[4971]: I1213 07:14:11.610121 4971 reconciler_common.go:293] "Volume detached for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/80fb658f-1dbc-42a8-a888-11afedfd3bba-sg-core-conf-yaml\") on node \"crc\" DevicePath \"\"" Dec 13 07:14:11 crc kubenswrapper[4971]: I1213 07:14:11.610129 4971 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/80fb658f-1dbc-42a8-a888-11afedfd3bba-scripts\") on node \"crc\" DevicePath \"\"" Dec 13 07:14:11 crc kubenswrapper[4971]: I1213 07:14:11.610138 4971 reconciler_common.go:293] "Volume detached for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/80fb658f-1dbc-42a8-a888-11afedfd3bba-run-httpd\") on node \"crc\" DevicePath \"\"" Dec 13 07:14:11 crc kubenswrapper[4971]: I1213 07:14:11.855392 4971 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"80fb658f-1dbc-42a8-a888-11afedfd3bba","Type":"ContainerDied","Data":"eb8a1becb80414dc0f287dd72df7dfb79b4fb5711921166a579e34af18c3ca7d"} Dec 13 07:14:11 crc kubenswrapper[4971]: I1213 07:14:11.855505 4971 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Dec 13 07:14:11 crc kubenswrapper[4971]: I1213 07:14:11.855902 4971 scope.go:117] "RemoveContainer" containerID="71fad303cda2bd73dcc3083f468961d98f8172537240aeca396bb177d8650254" Dec 13 07:14:11 crc kubenswrapper[4971]: I1213 07:14:11.884657 4971 scope.go:117] "RemoveContainer" containerID="4aa2ebd15a1835190cee6f8d1b95e42af2d3d12050eeac717d8e111cd9295a2a" Dec 13 07:14:11 crc kubenswrapper[4971]: I1213 07:14:11.887939 4971 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ceilometer-0"] Dec 13 07:14:11 crc kubenswrapper[4971]: I1213 07:14:11.898566 4971 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/ceilometer-0"] Dec 13 07:14:11 crc kubenswrapper[4971]: I1213 07:14:11.913446 4971 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ceilometer-0"] Dec 13 07:14:11 crc kubenswrapper[4971]: E1213 07:14:11.913952 4971 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="80fb658f-1dbc-42a8-a888-11afedfd3bba" containerName="ceilometer-central-agent" Dec 13 07:14:11 crc kubenswrapper[4971]: I1213 07:14:11.913973 4971 state_mem.go:107] "Deleted CPUSet assignment" podUID="80fb658f-1dbc-42a8-a888-11afedfd3bba" containerName="ceilometer-central-agent" Dec 13 07:14:11 crc kubenswrapper[4971]: E1213 07:14:11.913984 4971 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="80fb658f-1dbc-42a8-a888-11afedfd3bba" containerName="ceilometer-notification-agent" Dec 13 07:14:11 crc kubenswrapper[4971]: I1213 07:14:11.914005 4971 state_mem.go:107] "Deleted CPUSet assignment" podUID="80fb658f-1dbc-42a8-a888-11afedfd3bba" containerName="ceilometer-notification-agent" Dec 13 07:14:11 crc kubenswrapper[4971]: E1213 07:14:11.914017 4971 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="80fb658f-1dbc-42a8-a888-11afedfd3bba" containerName="proxy-httpd" Dec 13 07:14:11 crc kubenswrapper[4971]: I1213 07:14:11.914024 4971 state_mem.go:107] "Deleted CPUSet assignment" podUID="80fb658f-1dbc-42a8-a888-11afedfd3bba" containerName="proxy-httpd" Dec 13 07:14:11 crc kubenswrapper[4971]: E1213 07:14:11.914037 4971 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="80fb658f-1dbc-42a8-a888-11afedfd3bba" containerName="sg-core" Dec 13 07:14:11 crc kubenswrapper[4971]: I1213 07:14:11.914043 4971 state_mem.go:107] "Deleted CPUSet assignment" podUID="80fb658f-1dbc-42a8-a888-11afedfd3bba" containerName="sg-core" Dec 13 07:14:11 crc kubenswrapper[4971]: I1213 07:14:11.914282 4971 memory_manager.go:354] "RemoveStaleState removing state" podUID="80fb658f-1dbc-42a8-a888-11afedfd3bba" containerName="ceilometer-central-agent" Dec 13 07:14:11 crc kubenswrapper[4971]: I1213 07:14:11.914308 4971 memory_manager.go:354] "RemoveStaleState removing state" podUID="80fb658f-1dbc-42a8-a888-11afedfd3bba" containerName="ceilometer-notification-agent" Dec 13 07:14:11 crc kubenswrapper[4971]: I1213 07:14:11.914332 4971 memory_manager.go:354] "RemoveStaleState removing state" podUID="80fb658f-1dbc-42a8-a888-11afedfd3bba" containerName="sg-core" Dec 13 07:14:11 crc kubenswrapper[4971]: I1213 07:14:11.914346 4971 memory_manager.go:354] "RemoveStaleState removing state" podUID="80fb658f-1dbc-42a8-a888-11afedfd3bba" containerName="proxy-httpd" Dec 13 07:14:11 crc kubenswrapper[4971]: I1213 07:14:11.916093 4971 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Dec 13 07:14:12 crc kubenswrapper[4971]: I1213 07:14:12.214082 4971 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceilometer-scripts" Dec 13 07:14:12 crc kubenswrapper[4971]: I1213 07:14:12.214451 4971 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceilometer-config-data" Dec 13 07:14:12 crc kubenswrapper[4971]: I1213 07:14:12.214809 4971 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-ceilometer-internal-svc" Dec 13 07:14:12 crc kubenswrapper[4971]: I1213 07:14:12.215776 4971 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ceilometer-0"] Dec 13 07:14:12 crc kubenswrapper[4971]: I1213 07:14:12.215950 4971 scope.go:117] "RemoveContainer" containerID="9009ff428ac67061fac12bd7f9c52bdb707cbe1f8d74772a4533bb05661e33c7" Dec 13 07:14:12 crc kubenswrapper[4971]: I1213 07:14:12.237951 4971 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/18ea016d-26bf-46a9-a8c3-2f555643e967-scripts\") pod \"ceilometer-0\" (UID: \"18ea016d-26bf-46a9-a8c3-2f555643e967\") " pod="openstack/ceilometer-0" Dec 13 07:14:12 crc kubenswrapper[4971]: I1213 07:14:12.238541 4971 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ceilometer-tls-certs\" (UniqueName: \"kubernetes.io/secret/18ea016d-26bf-46a9-a8c3-2f555643e967-ceilometer-tls-certs\") pod \"ceilometer-0\" (UID: \"18ea016d-26bf-46a9-a8c3-2f555643e967\") " pod="openstack/ceilometer-0" Dec 13 07:14:12 crc kubenswrapper[4971]: I1213 07:14:12.238577 4971 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/18ea016d-26bf-46a9-a8c3-2f555643e967-log-httpd\") pod \"ceilometer-0\" (UID: \"18ea016d-26bf-46a9-a8c3-2f555643e967\") " pod="openstack/ceilometer-0" Dec 13 07:14:12 crc kubenswrapper[4971]: I1213 07:14:12.238630 4971 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/18ea016d-26bf-46a9-a8c3-2f555643e967-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"18ea016d-26bf-46a9-a8c3-2f555643e967\") " pod="openstack/ceilometer-0" Dec 13 07:14:12 crc kubenswrapper[4971]: I1213 07:14:12.238654 4971 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/18ea016d-26bf-46a9-a8c3-2f555643e967-config-data\") pod \"ceilometer-0\" (UID: \"18ea016d-26bf-46a9-a8c3-2f555643e967\") " pod="openstack/ceilometer-0" Dec 13 07:14:12 crc kubenswrapper[4971]: I1213 07:14:12.238703 4971 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/18ea016d-26bf-46a9-a8c3-2f555643e967-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"18ea016d-26bf-46a9-a8c3-2f555643e967\") " pod="openstack/ceilometer-0" Dec 13 07:14:12 crc kubenswrapper[4971]: I1213 07:14:12.239906 4971 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/18ea016d-26bf-46a9-a8c3-2f555643e967-run-httpd\") pod \"ceilometer-0\" (UID: \"18ea016d-26bf-46a9-a8c3-2f555643e967\") " pod="openstack/ceilometer-0" Dec 13 07:14:12 crc kubenswrapper[4971]: I1213 07:14:12.240750 4971 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-9qhqh\" (UniqueName: \"kubernetes.io/projected/18ea016d-26bf-46a9-a8c3-2f555643e967-kube-api-access-9qhqh\") pod \"ceilometer-0\" (UID: \"18ea016d-26bf-46a9-a8c3-2f555643e967\") " pod="openstack/ceilometer-0" Dec 13 07:14:12 crc kubenswrapper[4971]: I1213 07:14:12.255756 4971 scope.go:117] "RemoveContainer" containerID="f119ef3c07258d4dee06c12c543c7b23e4386958785adfb6524ba0662f06b280" Dec 13 07:14:12 crc kubenswrapper[4971]: I1213 07:14:12.342734 4971 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-9qhqh\" (UniqueName: \"kubernetes.io/projected/18ea016d-26bf-46a9-a8c3-2f555643e967-kube-api-access-9qhqh\") pod \"ceilometer-0\" (UID: \"18ea016d-26bf-46a9-a8c3-2f555643e967\") " pod="openstack/ceilometer-0" Dec 13 07:14:12 crc kubenswrapper[4971]: I1213 07:14:12.342822 4971 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/18ea016d-26bf-46a9-a8c3-2f555643e967-scripts\") pod \"ceilometer-0\" (UID: \"18ea016d-26bf-46a9-a8c3-2f555643e967\") " pod="openstack/ceilometer-0" Dec 13 07:14:12 crc kubenswrapper[4971]: I1213 07:14:12.342857 4971 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ceilometer-tls-certs\" (UniqueName: \"kubernetes.io/secret/18ea016d-26bf-46a9-a8c3-2f555643e967-ceilometer-tls-certs\") pod \"ceilometer-0\" (UID: \"18ea016d-26bf-46a9-a8c3-2f555643e967\") " pod="openstack/ceilometer-0" Dec 13 07:14:12 crc kubenswrapper[4971]: I1213 07:14:12.342885 4971 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/18ea016d-26bf-46a9-a8c3-2f555643e967-log-httpd\") pod \"ceilometer-0\" (UID: \"18ea016d-26bf-46a9-a8c3-2f555643e967\") " pod="openstack/ceilometer-0" Dec 13 07:14:12 crc kubenswrapper[4971]: I1213 07:14:12.342933 4971 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/18ea016d-26bf-46a9-a8c3-2f555643e967-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"18ea016d-26bf-46a9-a8c3-2f555643e967\") " pod="openstack/ceilometer-0" Dec 13 07:14:12 crc kubenswrapper[4971]: I1213 07:14:12.343009 4971 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/18ea016d-26bf-46a9-a8c3-2f555643e967-config-data\") pod \"ceilometer-0\" (UID: \"18ea016d-26bf-46a9-a8c3-2f555643e967\") " pod="openstack/ceilometer-0" Dec 13 07:14:12 crc kubenswrapper[4971]: I1213 07:14:12.343059 4971 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/18ea016d-26bf-46a9-a8c3-2f555643e967-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"18ea016d-26bf-46a9-a8c3-2f555643e967\") " pod="openstack/ceilometer-0" Dec 13 07:14:12 crc kubenswrapper[4971]: I1213 07:14:12.343105 4971 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/18ea016d-26bf-46a9-a8c3-2f555643e967-run-httpd\") pod \"ceilometer-0\" (UID: \"18ea016d-26bf-46a9-a8c3-2f555643e967\") " pod="openstack/ceilometer-0" Dec 13 07:14:12 crc kubenswrapper[4971]: I1213 07:14:12.343885 4971 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/18ea016d-26bf-46a9-a8c3-2f555643e967-run-httpd\") pod \"ceilometer-0\" (UID: \"18ea016d-26bf-46a9-a8c3-2f555643e967\") " pod="openstack/ceilometer-0" Dec 13 07:14:12 crc kubenswrapper[4971]: I1213 07:14:12.343915 4971 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/18ea016d-26bf-46a9-a8c3-2f555643e967-log-httpd\") pod \"ceilometer-0\" (UID: \"18ea016d-26bf-46a9-a8c3-2f555643e967\") " pod="openstack/ceilometer-0" Dec 13 07:14:12 crc kubenswrapper[4971]: I1213 07:14:12.348203 4971 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ceilometer-tls-certs\" (UniqueName: \"kubernetes.io/secret/18ea016d-26bf-46a9-a8c3-2f555643e967-ceilometer-tls-certs\") pod \"ceilometer-0\" (UID: \"18ea016d-26bf-46a9-a8c3-2f555643e967\") " pod="openstack/ceilometer-0" Dec 13 07:14:12 crc kubenswrapper[4971]: I1213 07:14:12.348228 4971 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/18ea016d-26bf-46a9-a8c3-2f555643e967-config-data\") pod \"ceilometer-0\" (UID: \"18ea016d-26bf-46a9-a8c3-2f555643e967\") " pod="openstack/ceilometer-0" Dec 13 07:14:12 crc kubenswrapper[4971]: I1213 07:14:12.352692 4971 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/18ea016d-26bf-46a9-a8c3-2f555643e967-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"18ea016d-26bf-46a9-a8c3-2f555643e967\") " pod="openstack/ceilometer-0" Dec 13 07:14:12 crc kubenswrapper[4971]: I1213 07:14:12.354815 4971 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/18ea016d-26bf-46a9-a8c3-2f555643e967-scripts\") pod \"ceilometer-0\" (UID: \"18ea016d-26bf-46a9-a8c3-2f555643e967\") " pod="openstack/ceilometer-0" Dec 13 07:14:12 crc kubenswrapper[4971]: I1213 07:14:12.356484 4971 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/18ea016d-26bf-46a9-a8c3-2f555643e967-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"18ea016d-26bf-46a9-a8c3-2f555643e967\") " pod="openstack/ceilometer-0" Dec 13 07:14:12 crc kubenswrapper[4971]: I1213 07:14:12.362761 4971 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-9qhqh\" (UniqueName: \"kubernetes.io/projected/18ea016d-26bf-46a9-a8c3-2f555643e967-kube-api-access-9qhqh\") pod \"ceilometer-0\" (UID: \"18ea016d-26bf-46a9-a8c3-2f555643e967\") " pod="openstack/ceilometer-0" Dec 13 07:14:12 crc kubenswrapper[4971]: I1213 07:14:12.540395 4971 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Dec 13 07:14:12 crc kubenswrapper[4971]: I1213 07:14:12.878684 4971 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-conductor-db-sync-g9t27" event={"ID":"31dddcf3-380b-4003-a766-88d4f5bc00de","Type":"ContainerStarted","Data":"153628e66d4982d16e87f162d44c475a8042e5c3ad6fd999fa22c399d1af2b9f"} Dec 13 07:14:12 crc kubenswrapper[4971]: I1213 07:14:12.901818 4971 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-cell0-conductor-db-sync-g9t27" podStartSLOduration=1.407271854 podStartE2EDuration="27.9017908s" podCreationTimestamp="2025-12-13 07:13:45 +0000 UTC" firstStartedPulling="2025-12-13 07:13:45.913638551 +0000 UTC m=+1482.518048009" lastFinishedPulling="2025-12-13 07:14:12.408157497 +0000 UTC m=+1509.012566955" observedRunningTime="2025-12-13 07:14:12.895562678 +0000 UTC m=+1509.499972136" watchObservedRunningTime="2025-12-13 07:14:12.9017908 +0000 UTC m=+1509.506200258" Dec 13 07:14:12 crc kubenswrapper[4971]: I1213 07:14:12.974797 4971 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ceilometer-0"] Dec 13 07:14:12 crc kubenswrapper[4971]: W1213 07:14:12.979267 4971 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod18ea016d_26bf_46a9_a8c3_2f555643e967.slice/crio-198989c780b92cb74ce23e480469c4dddb72d91548c9d3208ceb3d1b12175bc3 WatchSource:0}: Error finding container 198989c780b92cb74ce23e480469c4dddb72d91548c9d3208ceb3d1b12175bc3: Status 404 returned error can't find the container with id 198989c780b92cb74ce23e480469c4dddb72d91548c9d3208ceb3d1b12175bc3 Dec 13 07:14:13 crc kubenswrapper[4971]: I1213 07:14:13.781099 4971 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="80fb658f-1dbc-42a8-a888-11afedfd3bba" path="/var/lib/kubelet/pods/80fb658f-1dbc-42a8-a888-11afedfd3bba/volumes" Dec 13 07:14:13 crc kubenswrapper[4971]: I1213 07:14:13.890256 4971 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"18ea016d-26bf-46a9-a8c3-2f555643e967","Type":"ContainerStarted","Data":"fb55d70eea8aa0f44521717d475539a24d67f5b76a554632aa7eea6670d42a3d"} Dec 13 07:14:13 crc kubenswrapper[4971]: I1213 07:14:13.890310 4971 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"18ea016d-26bf-46a9-a8c3-2f555643e967","Type":"ContainerStarted","Data":"198989c780b92cb74ce23e480469c4dddb72d91548c9d3208ceb3d1b12175bc3"} Dec 13 07:14:14 crc kubenswrapper[4971]: I1213 07:14:14.901766 4971 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"18ea016d-26bf-46a9-a8c3-2f555643e967","Type":"ContainerStarted","Data":"e6e2bf1726974affec4fe8a70d50136677acd3da1008bacecd28f89bc4e2586c"} Dec 13 07:14:15 crc kubenswrapper[4971]: I1213 07:14:15.998412 4971 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"18ea016d-26bf-46a9-a8c3-2f555643e967","Type":"ContainerStarted","Data":"2d7101e713a8a556bd60407784326637a7599d4c2fe8201a04275e953f9be12f"} Dec 13 07:14:16 crc kubenswrapper[4971]: I1213 07:14:16.154268 4971 patch_prober.go:28] interesting pod/machine-config-daemon-82xjz container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 13 07:14:16 crc kubenswrapper[4971]: I1213 07:14:16.154349 4971 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-82xjz" podUID="e0c4af12-edda-41d7-9b44-f87396174ef2" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 13 07:14:17 crc kubenswrapper[4971]: I1213 07:14:17.009652 4971 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"18ea016d-26bf-46a9-a8c3-2f555643e967","Type":"ContainerStarted","Data":"9894e8db28cfff1cb5f97d08464004fdc4d8d729765e4dfce4f19b760d82bc9c"} Dec 13 07:14:17 crc kubenswrapper[4971]: I1213 07:14:17.011417 4971 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/ceilometer-0" Dec 13 07:14:25 crc kubenswrapper[4971]: I1213 07:14:25.444599 4971 generic.go:334] "Generic (PLEG): container finished" podID="31dddcf3-380b-4003-a766-88d4f5bc00de" containerID="153628e66d4982d16e87f162d44c475a8042e5c3ad6fd999fa22c399d1af2b9f" exitCode=0 Dec 13 07:14:25 crc kubenswrapper[4971]: I1213 07:14:25.444688 4971 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-conductor-db-sync-g9t27" event={"ID":"31dddcf3-380b-4003-a766-88d4f5bc00de","Type":"ContainerDied","Data":"153628e66d4982d16e87f162d44c475a8042e5c3ad6fd999fa22c399d1af2b9f"} Dec 13 07:14:25 crc kubenswrapper[4971]: I1213 07:14:25.473010 4971 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/ceilometer-0" podStartSLOduration=10.967893307 podStartE2EDuration="14.472984531s" podCreationTimestamp="2025-12-13 07:14:11 +0000 UTC" firstStartedPulling="2025-12-13 07:14:12.981551724 +0000 UTC m=+1509.585961172" lastFinishedPulling="2025-12-13 07:14:16.486642948 +0000 UTC m=+1513.091052396" observedRunningTime="2025-12-13 07:14:17.302192587 +0000 UTC m=+1513.906602095" watchObservedRunningTime="2025-12-13 07:14:25.472984531 +0000 UTC m=+1522.077393979" Dec 13 07:14:26 crc kubenswrapper[4971]: I1213 07:14:26.827928 4971 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-conductor-db-sync-g9t27" Dec 13 07:14:26 crc kubenswrapper[4971]: I1213 07:14:26.908655 4971 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/31dddcf3-380b-4003-a766-88d4f5bc00de-scripts\") pod \"31dddcf3-380b-4003-a766-88d4f5bc00de\" (UID: \"31dddcf3-380b-4003-a766-88d4f5bc00de\") " Dec 13 07:14:26 crc kubenswrapper[4971]: I1213 07:14:26.908714 4971 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/31dddcf3-380b-4003-a766-88d4f5bc00de-config-data\") pod \"31dddcf3-380b-4003-a766-88d4f5bc00de\" (UID: \"31dddcf3-380b-4003-a766-88d4f5bc00de\") " Dec 13 07:14:26 crc kubenswrapper[4971]: I1213 07:14:26.908744 4971 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-t8b6j\" (UniqueName: \"kubernetes.io/projected/31dddcf3-380b-4003-a766-88d4f5bc00de-kube-api-access-t8b6j\") pod \"31dddcf3-380b-4003-a766-88d4f5bc00de\" (UID: \"31dddcf3-380b-4003-a766-88d4f5bc00de\") " Dec 13 07:14:26 crc kubenswrapper[4971]: I1213 07:14:26.908835 4971 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/31dddcf3-380b-4003-a766-88d4f5bc00de-combined-ca-bundle\") pod \"31dddcf3-380b-4003-a766-88d4f5bc00de\" (UID: \"31dddcf3-380b-4003-a766-88d4f5bc00de\") " Dec 13 07:14:26 crc kubenswrapper[4971]: I1213 07:14:26.915904 4971 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/31dddcf3-380b-4003-a766-88d4f5bc00de-kube-api-access-t8b6j" (OuterVolumeSpecName: "kube-api-access-t8b6j") pod "31dddcf3-380b-4003-a766-88d4f5bc00de" (UID: "31dddcf3-380b-4003-a766-88d4f5bc00de"). InnerVolumeSpecName "kube-api-access-t8b6j". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 13 07:14:26 crc kubenswrapper[4971]: I1213 07:14:26.915927 4971 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/31dddcf3-380b-4003-a766-88d4f5bc00de-scripts" (OuterVolumeSpecName: "scripts") pod "31dddcf3-380b-4003-a766-88d4f5bc00de" (UID: "31dddcf3-380b-4003-a766-88d4f5bc00de"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 13 07:14:26 crc kubenswrapper[4971]: I1213 07:14:26.938649 4971 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/31dddcf3-380b-4003-a766-88d4f5bc00de-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "31dddcf3-380b-4003-a766-88d4f5bc00de" (UID: "31dddcf3-380b-4003-a766-88d4f5bc00de"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 13 07:14:26 crc kubenswrapper[4971]: I1213 07:14:26.946722 4971 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/31dddcf3-380b-4003-a766-88d4f5bc00de-config-data" (OuterVolumeSpecName: "config-data") pod "31dddcf3-380b-4003-a766-88d4f5bc00de" (UID: "31dddcf3-380b-4003-a766-88d4f5bc00de"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 13 07:14:27 crc kubenswrapper[4971]: I1213 07:14:27.010560 4971 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/31dddcf3-380b-4003-a766-88d4f5bc00de-config-data\") on node \"crc\" DevicePath \"\"" Dec 13 07:14:27 crc kubenswrapper[4971]: I1213 07:14:27.010599 4971 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/31dddcf3-380b-4003-a766-88d4f5bc00de-scripts\") on node \"crc\" DevicePath \"\"" Dec 13 07:14:27 crc kubenswrapper[4971]: I1213 07:14:27.010611 4971 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-t8b6j\" (UniqueName: \"kubernetes.io/projected/31dddcf3-380b-4003-a766-88d4f5bc00de-kube-api-access-t8b6j\") on node \"crc\" DevicePath \"\"" Dec 13 07:14:27 crc kubenswrapper[4971]: I1213 07:14:27.010624 4971 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/31dddcf3-380b-4003-a766-88d4f5bc00de-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 13 07:14:27 crc kubenswrapper[4971]: I1213 07:14:27.462596 4971 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-conductor-db-sync-g9t27" event={"ID":"31dddcf3-380b-4003-a766-88d4f5bc00de","Type":"ContainerDied","Data":"035cb215e2733c570e904c2f01af5581d55b5cbba9a858b412367ebbe1ce286e"} Dec 13 07:14:27 crc kubenswrapper[4971]: I1213 07:14:27.462645 4971 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="035cb215e2733c570e904c2f01af5581d55b5cbba9a858b412367ebbe1ce286e" Dec 13 07:14:27 crc kubenswrapper[4971]: I1213 07:14:27.462698 4971 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-conductor-db-sync-g9t27" Dec 13 07:14:27 crc kubenswrapper[4971]: I1213 07:14:27.587801 4971 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-cell0-conductor-0"] Dec 13 07:14:27 crc kubenswrapper[4971]: E1213 07:14:27.588171 4971 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="31dddcf3-380b-4003-a766-88d4f5bc00de" containerName="nova-cell0-conductor-db-sync" Dec 13 07:14:27 crc kubenswrapper[4971]: I1213 07:14:27.588185 4971 state_mem.go:107] "Deleted CPUSet assignment" podUID="31dddcf3-380b-4003-a766-88d4f5bc00de" containerName="nova-cell0-conductor-db-sync" Dec 13 07:14:27 crc kubenswrapper[4971]: I1213 07:14:27.588396 4971 memory_manager.go:354] "RemoveStaleState removing state" podUID="31dddcf3-380b-4003-a766-88d4f5bc00de" containerName="nova-cell0-conductor-db-sync" Dec 13 07:14:27 crc kubenswrapper[4971]: I1213 07:14:27.588964 4971 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-conductor-0" Dec 13 07:14:27 crc kubenswrapper[4971]: I1213 07:14:27.590906 4971 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-nova-dockercfg-dsf9t" Dec 13 07:14:27 crc kubenswrapper[4971]: I1213 07:14:27.591476 4971 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-cell0-conductor-config-data" Dec 13 07:14:27 crc kubenswrapper[4971]: I1213 07:14:27.603067 4971 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell0-conductor-0"] Dec 13 07:14:27 crc kubenswrapper[4971]: I1213 07:14:27.622159 4971 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/5220d566-7310-4e2f-b5b0-4e0c5e826aa9-config-data\") pod \"nova-cell0-conductor-0\" (UID: \"5220d566-7310-4e2f-b5b0-4e0c5e826aa9\") " pod="openstack/nova-cell0-conductor-0" Dec 13 07:14:27 crc kubenswrapper[4971]: I1213 07:14:27.622205 4971 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-lxpln\" (UniqueName: \"kubernetes.io/projected/5220d566-7310-4e2f-b5b0-4e0c5e826aa9-kube-api-access-lxpln\") pod \"nova-cell0-conductor-0\" (UID: \"5220d566-7310-4e2f-b5b0-4e0c5e826aa9\") " pod="openstack/nova-cell0-conductor-0" Dec 13 07:14:27 crc kubenswrapper[4971]: I1213 07:14:27.622329 4971 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/5220d566-7310-4e2f-b5b0-4e0c5e826aa9-combined-ca-bundle\") pod \"nova-cell0-conductor-0\" (UID: \"5220d566-7310-4e2f-b5b0-4e0c5e826aa9\") " pod="openstack/nova-cell0-conductor-0" Dec 13 07:14:27 crc kubenswrapper[4971]: I1213 07:14:27.723855 4971 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/5220d566-7310-4e2f-b5b0-4e0c5e826aa9-combined-ca-bundle\") pod \"nova-cell0-conductor-0\" (UID: \"5220d566-7310-4e2f-b5b0-4e0c5e826aa9\") " pod="openstack/nova-cell0-conductor-0" Dec 13 07:14:27 crc kubenswrapper[4971]: I1213 07:14:27.723918 4971 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/5220d566-7310-4e2f-b5b0-4e0c5e826aa9-config-data\") pod \"nova-cell0-conductor-0\" (UID: \"5220d566-7310-4e2f-b5b0-4e0c5e826aa9\") " pod="openstack/nova-cell0-conductor-0" Dec 13 07:14:27 crc kubenswrapper[4971]: I1213 07:14:27.723939 4971 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-lxpln\" (UniqueName: \"kubernetes.io/projected/5220d566-7310-4e2f-b5b0-4e0c5e826aa9-kube-api-access-lxpln\") pod \"nova-cell0-conductor-0\" (UID: \"5220d566-7310-4e2f-b5b0-4e0c5e826aa9\") " pod="openstack/nova-cell0-conductor-0" Dec 13 07:14:27 crc kubenswrapper[4971]: I1213 07:14:27.729466 4971 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/5220d566-7310-4e2f-b5b0-4e0c5e826aa9-config-data\") pod \"nova-cell0-conductor-0\" (UID: \"5220d566-7310-4e2f-b5b0-4e0c5e826aa9\") " pod="openstack/nova-cell0-conductor-0" Dec 13 07:14:27 crc kubenswrapper[4971]: I1213 07:14:27.733939 4971 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/5220d566-7310-4e2f-b5b0-4e0c5e826aa9-combined-ca-bundle\") pod \"nova-cell0-conductor-0\" (UID: \"5220d566-7310-4e2f-b5b0-4e0c5e826aa9\") " pod="openstack/nova-cell0-conductor-0" Dec 13 07:14:27 crc kubenswrapper[4971]: I1213 07:14:27.740091 4971 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-lxpln\" (UniqueName: \"kubernetes.io/projected/5220d566-7310-4e2f-b5b0-4e0c5e826aa9-kube-api-access-lxpln\") pod \"nova-cell0-conductor-0\" (UID: \"5220d566-7310-4e2f-b5b0-4e0c5e826aa9\") " pod="openstack/nova-cell0-conductor-0" Dec 13 07:14:27 crc kubenswrapper[4971]: I1213 07:14:27.905652 4971 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-conductor-0" Dec 13 07:14:28 crc kubenswrapper[4971]: I1213 07:14:28.401099 4971 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell0-conductor-0"] Dec 13 07:14:28 crc kubenswrapper[4971]: I1213 07:14:28.471673 4971 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-conductor-0" event={"ID":"5220d566-7310-4e2f-b5b0-4e0c5e826aa9","Type":"ContainerStarted","Data":"c5c3612c5a3338c84c580459a40336413ab553d068c7c77f6682e7b4838a75a9"} Dec 13 07:14:29 crc kubenswrapper[4971]: I1213 07:14:29.481653 4971 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-conductor-0" event={"ID":"5220d566-7310-4e2f-b5b0-4e0c5e826aa9","Type":"ContainerStarted","Data":"55730b25465f98776ac9fd8dc732cc50876ab4f918a4adae5825d334323881f5"} Dec 13 07:14:29 crc kubenswrapper[4971]: I1213 07:14:29.483145 4971 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-cell0-conductor-0" Dec 13 07:14:29 crc kubenswrapper[4971]: I1213 07:14:29.511065 4971 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-cell0-conductor-0" podStartSLOduration=2.511042017 podStartE2EDuration="2.511042017s" podCreationTimestamp="2025-12-13 07:14:27 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-13 07:14:29.500095297 +0000 UTC m=+1526.104504755" watchObservedRunningTime="2025-12-13 07:14:29.511042017 +0000 UTC m=+1526.115451465" Dec 13 07:14:29 crc kubenswrapper[4971]: I1213 07:14:29.532005 4971 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-marketplace-8rlw2"] Dec 13 07:14:29 crc kubenswrapper[4971]: I1213 07:14:29.534399 4971 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-8rlw2" Dec 13 07:14:29 crc kubenswrapper[4971]: I1213 07:14:29.548068 4971 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-8rlw2"] Dec 13 07:14:29 crc kubenswrapper[4971]: I1213 07:14:29.662110 4971 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/d5c250a4-d784-4508-9433-d684c6b01531-catalog-content\") pod \"redhat-marketplace-8rlw2\" (UID: \"d5c250a4-d784-4508-9433-d684c6b01531\") " pod="openshift-marketplace/redhat-marketplace-8rlw2" Dec 13 07:14:29 crc kubenswrapper[4971]: I1213 07:14:29.662151 4971 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/d5c250a4-d784-4508-9433-d684c6b01531-utilities\") pod \"redhat-marketplace-8rlw2\" (UID: \"d5c250a4-d784-4508-9433-d684c6b01531\") " pod="openshift-marketplace/redhat-marketplace-8rlw2" Dec 13 07:14:29 crc kubenswrapper[4971]: I1213 07:14:29.662183 4971 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-cl57q\" (UniqueName: \"kubernetes.io/projected/d5c250a4-d784-4508-9433-d684c6b01531-kube-api-access-cl57q\") pod \"redhat-marketplace-8rlw2\" (UID: \"d5c250a4-d784-4508-9433-d684c6b01531\") " pod="openshift-marketplace/redhat-marketplace-8rlw2" Dec 13 07:14:29 crc kubenswrapper[4971]: I1213 07:14:29.763864 4971 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/d5c250a4-d784-4508-9433-d684c6b01531-catalog-content\") pod \"redhat-marketplace-8rlw2\" (UID: \"d5c250a4-d784-4508-9433-d684c6b01531\") " pod="openshift-marketplace/redhat-marketplace-8rlw2" Dec 13 07:14:29 crc kubenswrapper[4971]: I1213 07:14:29.763907 4971 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/d5c250a4-d784-4508-9433-d684c6b01531-utilities\") pod \"redhat-marketplace-8rlw2\" (UID: \"d5c250a4-d784-4508-9433-d684c6b01531\") " pod="openshift-marketplace/redhat-marketplace-8rlw2" Dec 13 07:14:29 crc kubenswrapper[4971]: I1213 07:14:29.763942 4971 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cl57q\" (UniqueName: \"kubernetes.io/projected/d5c250a4-d784-4508-9433-d684c6b01531-kube-api-access-cl57q\") pod \"redhat-marketplace-8rlw2\" (UID: \"d5c250a4-d784-4508-9433-d684c6b01531\") " pod="openshift-marketplace/redhat-marketplace-8rlw2" Dec 13 07:14:29 crc kubenswrapper[4971]: I1213 07:14:29.764671 4971 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/d5c250a4-d784-4508-9433-d684c6b01531-catalog-content\") pod \"redhat-marketplace-8rlw2\" (UID: \"d5c250a4-d784-4508-9433-d684c6b01531\") " pod="openshift-marketplace/redhat-marketplace-8rlw2" Dec 13 07:14:29 crc kubenswrapper[4971]: I1213 07:14:29.764944 4971 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/d5c250a4-d784-4508-9433-d684c6b01531-utilities\") pod \"redhat-marketplace-8rlw2\" (UID: \"d5c250a4-d784-4508-9433-d684c6b01531\") " pod="openshift-marketplace/redhat-marketplace-8rlw2" Dec 13 07:14:29 crc kubenswrapper[4971]: I1213 07:14:29.801037 4971 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-cl57q\" (UniqueName: \"kubernetes.io/projected/d5c250a4-d784-4508-9433-d684c6b01531-kube-api-access-cl57q\") pod \"redhat-marketplace-8rlw2\" (UID: \"d5c250a4-d784-4508-9433-d684c6b01531\") " pod="openshift-marketplace/redhat-marketplace-8rlw2" Dec 13 07:14:29 crc kubenswrapper[4971]: I1213 07:14:29.865336 4971 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-8rlw2" Dec 13 07:14:30 crc kubenswrapper[4971]: I1213 07:14:30.339814 4971 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-8rlw2"] Dec 13 07:14:30 crc kubenswrapper[4971]: I1213 07:14:30.491069 4971 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-8rlw2" event={"ID":"d5c250a4-d784-4508-9433-d684c6b01531","Type":"ContainerStarted","Data":"bb8d6547555de0ffa2354fe79b4ef9710a8a1fce0aa02b49ccc3c9e52c8ef897"} Dec 13 07:14:31 crc kubenswrapper[4971]: I1213 07:14:31.502397 4971 generic.go:334] "Generic (PLEG): container finished" podID="d5c250a4-d784-4508-9433-d684c6b01531" containerID="7d89c39b751696f0853323248c94bdc2f2bd46d2ecf99447674a6e1e32eff850" exitCode=0 Dec 13 07:14:31 crc kubenswrapper[4971]: I1213 07:14:31.502479 4971 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-8rlw2" event={"ID":"d5c250a4-d784-4508-9433-d684c6b01531","Type":"ContainerDied","Data":"7d89c39b751696f0853323248c94bdc2f2bd46d2ecf99447674a6e1e32eff850"} Dec 13 07:14:33 crc kubenswrapper[4971]: I1213 07:14:33.524285 4971 generic.go:334] "Generic (PLEG): container finished" podID="d5c250a4-d784-4508-9433-d684c6b01531" containerID="693a818b1ae2abfe18c3d6a10996082200f8677d5087a58ec65f4b2d7dac24ed" exitCode=0 Dec 13 07:14:33 crc kubenswrapper[4971]: I1213 07:14:33.524337 4971 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-8rlw2" event={"ID":"d5c250a4-d784-4508-9433-d684c6b01531","Type":"ContainerDied","Data":"693a818b1ae2abfe18c3d6a10996082200f8677d5087a58ec65f4b2d7dac24ed"} Dec 13 07:14:34 crc kubenswrapper[4971]: I1213 07:14:34.533755 4971 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-8rlw2" event={"ID":"d5c250a4-d784-4508-9433-d684c6b01531","Type":"ContainerStarted","Data":"ae179455b245343d250e77a8751d8b15e9f429baa9c54c6b6b56737c04ca7cdf"} Dec 13 07:14:34 crc kubenswrapper[4971]: I1213 07:14:34.561329 4971 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-marketplace-8rlw2" podStartSLOduration=2.979632793 podStartE2EDuration="5.561308123s" podCreationTimestamp="2025-12-13 07:14:29 +0000 UTC" firstStartedPulling="2025-12-13 07:14:31.504664898 +0000 UTC m=+1528.109074356" lastFinishedPulling="2025-12-13 07:14:34.086340238 +0000 UTC m=+1530.690749686" observedRunningTime="2025-12-13 07:14:34.556150455 +0000 UTC m=+1531.160559913" watchObservedRunningTime="2025-12-13 07:14:34.561308123 +0000 UTC m=+1531.165717581" Dec 13 07:14:37 crc kubenswrapper[4971]: I1213 07:14:37.942066 4971 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/nova-cell0-conductor-0" Dec 13 07:14:38 crc kubenswrapper[4971]: I1213 07:14:38.595638 4971 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-cell0-cell-mapping-rgjbg"] Dec 13 07:14:38 crc kubenswrapper[4971]: I1213 07:14:38.602735 4971 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-cell-mapping-rgjbg" Dec 13 07:14:38 crc kubenswrapper[4971]: I1213 07:14:38.605135 4971 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-cell0-manage-scripts" Dec 13 07:14:38 crc kubenswrapper[4971]: I1213 07:14:38.605326 4971 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-cell0-manage-config-data" Dec 13 07:14:38 crc kubenswrapper[4971]: I1213 07:14:38.606393 4971 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell0-cell-mapping-rgjbg"] Dec 13 07:14:38 crc kubenswrapper[4971]: I1213 07:14:38.728674 4971 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-6htvt\" (UniqueName: \"kubernetes.io/projected/74fcba46-f364-4c8a-a4e3-ace14451d1f7-kube-api-access-6htvt\") pod \"nova-cell0-cell-mapping-rgjbg\" (UID: \"74fcba46-f364-4c8a-a4e3-ace14451d1f7\") " pod="openstack/nova-cell0-cell-mapping-rgjbg" Dec 13 07:14:38 crc kubenswrapper[4971]: I1213 07:14:38.728775 4971 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/74fcba46-f364-4c8a-a4e3-ace14451d1f7-config-data\") pod \"nova-cell0-cell-mapping-rgjbg\" (UID: \"74fcba46-f364-4c8a-a4e3-ace14451d1f7\") " pod="openstack/nova-cell0-cell-mapping-rgjbg" Dec 13 07:14:38 crc kubenswrapper[4971]: I1213 07:14:38.728839 4971 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/74fcba46-f364-4c8a-a4e3-ace14451d1f7-combined-ca-bundle\") pod \"nova-cell0-cell-mapping-rgjbg\" (UID: \"74fcba46-f364-4c8a-a4e3-ace14451d1f7\") " pod="openstack/nova-cell0-cell-mapping-rgjbg" Dec 13 07:14:38 crc kubenswrapper[4971]: I1213 07:14:38.728916 4971 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/74fcba46-f364-4c8a-a4e3-ace14451d1f7-scripts\") pod \"nova-cell0-cell-mapping-rgjbg\" (UID: \"74fcba46-f364-4c8a-a4e3-ace14451d1f7\") " pod="openstack/nova-cell0-cell-mapping-rgjbg" Dec 13 07:14:38 crc kubenswrapper[4971]: I1213 07:14:38.832638 4971 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/74fcba46-f364-4c8a-a4e3-ace14451d1f7-config-data\") pod \"nova-cell0-cell-mapping-rgjbg\" (UID: \"74fcba46-f364-4c8a-a4e3-ace14451d1f7\") " pod="openstack/nova-cell0-cell-mapping-rgjbg" Dec 13 07:14:38 crc kubenswrapper[4971]: I1213 07:14:38.833175 4971 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/74fcba46-f364-4c8a-a4e3-ace14451d1f7-combined-ca-bundle\") pod \"nova-cell0-cell-mapping-rgjbg\" (UID: \"74fcba46-f364-4c8a-a4e3-ace14451d1f7\") " pod="openstack/nova-cell0-cell-mapping-rgjbg" Dec 13 07:14:38 crc kubenswrapper[4971]: I1213 07:14:38.833314 4971 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/74fcba46-f364-4c8a-a4e3-ace14451d1f7-scripts\") pod \"nova-cell0-cell-mapping-rgjbg\" (UID: \"74fcba46-f364-4c8a-a4e3-ace14451d1f7\") " pod="openstack/nova-cell0-cell-mapping-rgjbg" Dec 13 07:14:38 crc kubenswrapper[4971]: I1213 07:14:38.833485 4971 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-6htvt\" (UniqueName: \"kubernetes.io/projected/74fcba46-f364-4c8a-a4e3-ace14451d1f7-kube-api-access-6htvt\") pod \"nova-cell0-cell-mapping-rgjbg\" (UID: \"74fcba46-f364-4c8a-a4e3-ace14451d1f7\") " pod="openstack/nova-cell0-cell-mapping-rgjbg" Dec 13 07:14:38 crc kubenswrapper[4971]: I1213 07:14:38.841082 4971 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/74fcba46-f364-4c8a-a4e3-ace14451d1f7-scripts\") pod \"nova-cell0-cell-mapping-rgjbg\" (UID: \"74fcba46-f364-4c8a-a4e3-ace14451d1f7\") " pod="openstack/nova-cell0-cell-mapping-rgjbg" Dec 13 07:14:38 crc kubenswrapper[4971]: I1213 07:14:38.848415 4971 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/74fcba46-f364-4c8a-a4e3-ace14451d1f7-config-data\") pod \"nova-cell0-cell-mapping-rgjbg\" (UID: \"74fcba46-f364-4c8a-a4e3-ace14451d1f7\") " pod="openstack/nova-cell0-cell-mapping-rgjbg" Dec 13 07:14:38 crc kubenswrapper[4971]: I1213 07:14:38.891141 4971 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/74fcba46-f364-4c8a-a4e3-ace14451d1f7-combined-ca-bundle\") pod \"nova-cell0-cell-mapping-rgjbg\" (UID: \"74fcba46-f364-4c8a-a4e3-ace14451d1f7\") " pod="openstack/nova-cell0-cell-mapping-rgjbg" Dec 13 07:14:38 crc kubenswrapper[4971]: I1213 07:14:38.893249 4971 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-6htvt\" (UniqueName: \"kubernetes.io/projected/74fcba46-f364-4c8a-a4e3-ace14451d1f7-kube-api-access-6htvt\") pod \"nova-cell0-cell-mapping-rgjbg\" (UID: \"74fcba46-f364-4c8a-a4e3-ace14451d1f7\") " pod="openstack/nova-cell0-cell-mapping-rgjbg" Dec 13 07:14:38 crc kubenswrapper[4971]: I1213 07:14:38.919576 4971 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-cell1-novncproxy-0"] Dec 13 07:14:38 crc kubenswrapper[4971]: I1213 07:14:38.920909 4971 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-novncproxy-0" Dec 13 07:14:38 crc kubenswrapper[4971]: I1213 07:14:38.921885 4971 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-cell-mapping-rgjbg" Dec 13 07:14:38 crc kubenswrapper[4971]: I1213 07:14:38.945060 4971 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-cell1-novncproxy-config-data" Dec 13 07:14:38 crc kubenswrapper[4971]: I1213 07:14:38.971384 4971 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell1-novncproxy-0"] Dec 13 07:14:38 crc kubenswrapper[4971]: I1213 07:14:38.984328 4971 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-metadata-0"] Dec 13 07:14:38 crc kubenswrapper[4971]: I1213 07:14:38.991588 4971 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-metadata-0" Dec 13 07:14:39 crc kubenswrapper[4971]: I1213 07:14:39.007546 4971 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-metadata-config-data" Dec 13 07:14:39 crc kubenswrapper[4971]: I1213 07:14:39.047530 4971 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-sgnlr\" (UniqueName: \"kubernetes.io/projected/9bb1801a-532f-4b09-9ae3-d4b61e8a9c97-kube-api-access-sgnlr\") pod \"nova-cell1-novncproxy-0\" (UID: \"9bb1801a-532f-4b09-9ae3-d4b61e8a9c97\") " pod="openstack/nova-cell1-novncproxy-0" Dec 13 07:14:39 crc kubenswrapper[4971]: I1213 07:14:39.047642 4971 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/feaf3daf-1c1f-493f-8358-8e9afb515229-combined-ca-bundle\") pod \"nova-metadata-0\" (UID: \"feaf3daf-1c1f-493f-8358-8e9afb515229\") " pod="openstack/nova-metadata-0" Dec 13 07:14:39 crc kubenswrapper[4971]: I1213 07:14:39.047698 4971 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/9bb1801a-532f-4b09-9ae3-d4b61e8a9c97-combined-ca-bundle\") pod \"nova-cell1-novncproxy-0\" (UID: \"9bb1801a-532f-4b09-9ae3-d4b61e8a9c97\") " pod="openstack/nova-cell1-novncproxy-0" Dec 13 07:14:39 crc kubenswrapper[4971]: I1213 07:14:39.047730 4971 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-6fdl2\" (UniqueName: \"kubernetes.io/projected/feaf3daf-1c1f-493f-8358-8e9afb515229-kube-api-access-6fdl2\") pod \"nova-metadata-0\" (UID: \"feaf3daf-1c1f-493f-8358-8e9afb515229\") " pod="openstack/nova-metadata-0" Dec 13 07:14:39 crc kubenswrapper[4971]: I1213 07:14:39.047764 4971 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/feaf3daf-1c1f-493f-8358-8e9afb515229-logs\") pod \"nova-metadata-0\" (UID: \"feaf3daf-1c1f-493f-8358-8e9afb515229\") " pod="openstack/nova-metadata-0" Dec 13 07:14:39 crc kubenswrapper[4971]: I1213 07:14:39.047786 4971 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/9bb1801a-532f-4b09-9ae3-d4b61e8a9c97-config-data\") pod \"nova-cell1-novncproxy-0\" (UID: \"9bb1801a-532f-4b09-9ae3-d4b61e8a9c97\") " pod="openstack/nova-cell1-novncproxy-0" Dec 13 07:14:39 crc kubenswrapper[4971]: I1213 07:14:39.047836 4971 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/feaf3daf-1c1f-493f-8358-8e9afb515229-config-data\") pod \"nova-metadata-0\" (UID: \"feaf3daf-1c1f-493f-8358-8e9afb515229\") " pod="openstack/nova-metadata-0" Dec 13 07:14:39 crc kubenswrapper[4971]: I1213 07:14:39.050602 4971 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-metadata-0"] Dec 13 07:14:39 crc kubenswrapper[4971]: I1213 07:14:39.098137 4971 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-scheduler-0"] Dec 13 07:14:39 crc kubenswrapper[4971]: I1213 07:14:39.099303 4971 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-scheduler-0" Dec 13 07:14:39 crc kubenswrapper[4971]: I1213 07:14:39.102279 4971 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-scheduler-config-data" Dec 13 07:14:39 crc kubenswrapper[4971]: I1213 07:14:39.111597 4971 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-scheduler-0"] Dec 13 07:14:39 crc kubenswrapper[4971]: I1213 07:14:39.121376 4971 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-59fd54bbff-hnxpb"] Dec 13 07:14:39 crc kubenswrapper[4971]: I1213 07:14:39.125957 4971 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-59fd54bbff-hnxpb" Dec 13 07:14:39 crc kubenswrapper[4971]: I1213 07:14:39.149748 4971 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/feaf3daf-1c1f-493f-8358-8e9afb515229-combined-ca-bundle\") pod \"nova-metadata-0\" (UID: \"feaf3daf-1c1f-493f-8358-8e9afb515229\") " pod="openstack/nova-metadata-0" Dec 13 07:14:39 crc kubenswrapper[4971]: I1213 07:14:39.150039 4971 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/156254b1-933a-49b3-ae26-f0ba7f1418b8-config-data\") pod \"nova-scheduler-0\" (UID: \"156254b1-933a-49b3-ae26-f0ba7f1418b8\") " pod="openstack/nova-scheduler-0" Dec 13 07:14:39 crc kubenswrapper[4971]: I1213 07:14:39.150159 4971 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/9bb1801a-532f-4b09-9ae3-d4b61e8a9c97-combined-ca-bundle\") pod \"nova-cell1-novncproxy-0\" (UID: \"9bb1801a-532f-4b09-9ae3-d4b61e8a9c97\") " pod="openstack/nova-cell1-novncproxy-0" Dec 13 07:14:39 crc kubenswrapper[4971]: I1213 07:14:39.150264 4971 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-6fdl2\" (UniqueName: \"kubernetes.io/projected/feaf3daf-1c1f-493f-8358-8e9afb515229-kube-api-access-6fdl2\") pod \"nova-metadata-0\" (UID: \"feaf3daf-1c1f-493f-8358-8e9afb515229\") " pod="openstack/nova-metadata-0" Dec 13 07:14:39 crc kubenswrapper[4971]: I1213 07:14:39.150363 4971 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/feaf3daf-1c1f-493f-8358-8e9afb515229-logs\") pod \"nova-metadata-0\" (UID: \"feaf3daf-1c1f-493f-8358-8e9afb515229\") " pod="openstack/nova-metadata-0" Dec 13 07:14:39 crc kubenswrapper[4971]: I1213 07:14:39.151137 4971 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/feaf3daf-1c1f-493f-8358-8e9afb515229-logs\") pod \"nova-metadata-0\" (UID: \"feaf3daf-1c1f-493f-8358-8e9afb515229\") " pod="openstack/nova-metadata-0" Dec 13 07:14:39 crc kubenswrapper[4971]: I1213 07:14:39.151612 4971 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-2k4fc\" (UniqueName: \"kubernetes.io/projected/156254b1-933a-49b3-ae26-f0ba7f1418b8-kube-api-access-2k4fc\") pod \"nova-scheduler-0\" (UID: \"156254b1-933a-49b3-ae26-f0ba7f1418b8\") " pod="openstack/nova-scheduler-0" Dec 13 07:14:39 crc kubenswrapper[4971]: I1213 07:14:39.151752 4971 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/9bb1801a-532f-4b09-9ae3-d4b61e8a9c97-config-data\") pod \"nova-cell1-novncproxy-0\" (UID: \"9bb1801a-532f-4b09-9ae3-d4b61e8a9c97\") " pod="openstack/nova-cell1-novncproxy-0" Dec 13 07:14:39 crc kubenswrapper[4971]: I1213 07:14:39.152100 4971 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/156254b1-933a-49b3-ae26-f0ba7f1418b8-combined-ca-bundle\") pod \"nova-scheduler-0\" (UID: \"156254b1-933a-49b3-ae26-f0ba7f1418b8\") " pod="openstack/nova-scheduler-0" Dec 13 07:14:39 crc kubenswrapper[4971]: I1213 07:14:39.152234 4971 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/feaf3daf-1c1f-493f-8358-8e9afb515229-config-data\") pod \"nova-metadata-0\" (UID: \"feaf3daf-1c1f-493f-8358-8e9afb515229\") " pod="openstack/nova-metadata-0" Dec 13 07:14:39 crc kubenswrapper[4971]: I1213 07:14:39.152369 4971 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-sgnlr\" (UniqueName: \"kubernetes.io/projected/9bb1801a-532f-4b09-9ae3-d4b61e8a9c97-kube-api-access-sgnlr\") pod \"nova-cell1-novncproxy-0\" (UID: \"9bb1801a-532f-4b09-9ae3-d4b61e8a9c97\") " pod="openstack/nova-cell1-novncproxy-0" Dec 13 07:14:39 crc kubenswrapper[4971]: I1213 07:14:39.162835 4971 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-59fd54bbff-hnxpb"] Dec 13 07:14:39 crc kubenswrapper[4971]: I1213 07:14:39.166020 4971 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/feaf3daf-1c1f-493f-8358-8e9afb515229-combined-ca-bundle\") pod \"nova-metadata-0\" (UID: \"feaf3daf-1c1f-493f-8358-8e9afb515229\") " pod="openstack/nova-metadata-0" Dec 13 07:14:39 crc kubenswrapper[4971]: I1213 07:14:39.174279 4971 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/9bb1801a-532f-4b09-9ae3-d4b61e8a9c97-combined-ca-bundle\") pod \"nova-cell1-novncproxy-0\" (UID: \"9bb1801a-532f-4b09-9ae3-d4b61e8a9c97\") " pod="openstack/nova-cell1-novncproxy-0" Dec 13 07:14:39 crc kubenswrapper[4971]: I1213 07:14:39.189227 4971 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/9bb1801a-532f-4b09-9ae3-d4b61e8a9c97-config-data\") pod \"nova-cell1-novncproxy-0\" (UID: \"9bb1801a-532f-4b09-9ae3-d4b61e8a9c97\") " pod="openstack/nova-cell1-novncproxy-0" Dec 13 07:14:39 crc kubenswrapper[4971]: I1213 07:14:39.195420 4971 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-sgnlr\" (UniqueName: \"kubernetes.io/projected/9bb1801a-532f-4b09-9ae3-d4b61e8a9c97-kube-api-access-sgnlr\") pod \"nova-cell1-novncproxy-0\" (UID: \"9bb1801a-532f-4b09-9ae3-d4b61e8a9c97\") " pod="openstack/nova-cell1-novncproxy-0" Dec 13 07:14:39 crc kubenswrapper[4971]: I1213 07:14:39.207724 4971 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/feaf3daf-1c1f-493f-8358-8e9afb515229-config-data\") pod \"nova-metadata-0\" (UID: \"feaf3daf-1c1f-493f-8358-8e9afb515229\") " pod="openstack/nova-metadata-0" Dec 13 07:14:39 crc kubenswrapper[4971]: I1213 07:14:39.220202 4971 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-6fdl2\" (UniqueName: \"kubernetes.io/projected/feaf3daf-1c1f-493f-8358-8e9afb515229-kube-api-access-6fdl2\") pod \"nova-metadata-0\" (UID: \"feaf3daf-1c1f-493f-8358-8e9afb515229\") " pod="openstack/nova-metadata-0" Dec 13 07:14:39 crc kubenswrapper[4971]: I1213 07:14:39.243602 4971 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-api-0"] Dec 13 07:14:39 crc kubenswrapper[4971]: I1213 07:14:39.244990 4971 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-0" Dec 13 07:14:39 crc kubenswrapper[4971]: I1213 07:14:39.254563 4971 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-api-config-data" Dec 13 07:14:39 crc kubenswrapper[4971]: I1213 07:14:39.254645 4971 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/156254b1-933a-49b3-ae26-f0ba7f1418b8-config-data\") pod \"nova-scheduler-0\" (UID: \"156254b1-933a-49b3-ae26-f0ba7f1418b8\") " pod="openstack/nova-scheduler-0" Dec 13 07:14:39 crc kubenswrapper[4971]: I1213 07:14:39.254729 4971 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-2k4fc\" (UniqueName: \"kubernetes.io/projected/156254b1-933a-49b3-ae26-f0ba7f1418b8-kube-api-access-2k4fc\") pod \"nova-scheduler-0\" (UID: \"156254b1-933a-49b3-ae26-f0ba7f1418b8\") " pod="openstack/nova-scheduler-0" Dec 13 07:14:39 crc kubenswrapper[4971]: I1213 07:14:39.254755 4971 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/5e843ecc-f9c0-492f-bcd8-0d34447ae18a-ovsdbserver-sb\") pod \"dnsmasq-dns-59fd54bbff-hnxpb\" (UID: \"5e843ecc-f9c0-492f-bcd8-0d34447ae18a\") " pod="openstack/dnsmasq-dns-59fd54bbff-hnxpb" Dec 13 07:14:39 crc kubenswrapper[4971]: I1213 07:14:39.254800 4971 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/156254b1-933a-49b3-ae26-f0ba7f1418b8-combined-ca-bundle\") pod \"nova-scheduler-0\" (UID: \"156254b1-933a-49b3-ae26-f0ba7f1418b8\") " pod="openstack/nova-scheduler-0" Dec 13 07:14:39 crc kubenswrapper[4971]: I1213 07:14:39.254827 4971 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/5e843ecc-f9c0-492f-bcd8-0d34447ae18a-config\") pod \"dnsmasq-dns-59fd54bbff-hnxpb\" (UID: \"5e843ecc-f9c0-492f-bcd8-0d34447ae18a\") " pod="openstack/dnsmasq-dns-59fd54bbff-hnxpb" Dec 13 07:14:39 crc kubenswrapper[4971]: I1213 07:14:39.254901 4971 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/5e843ecc-f9c0-492f-bcd8-0d34447ae18a-ovsdbserver-nb\") pod \"dnsmasq-dns-59fd54bbff-hnxpb\" (UID: \"5e843ecc-f9c0-492f-bcd8-0d34447ae18a\") " pod="openstack/dnsmasq-dns-59fd54bbff-hnxpb" Dec 13 07:14:39 crc kubenswrapper[4971]: I1213 07:14:39.254929 4971 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-api-0"] Dec 13 07:14:39 crc kubenswrapper[4971]: I1213 07:14:39.254940 4971 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/5e843ecc-f9c0-492f-bcd8-0d34447ae18a-dns-svc\") pod \"dnsmasq-dns-59fd54bbff-hnxpb\" (UID: \"5e843ecc-f9c0-492f-bcd8-0d34447ae18a\") " pod="openstack/dnsmasq-dns-59fd54bbff-hnxpb" Dec 13 07:14:39 crc kubenswrapper[4971]: I1213 07:14:39.255016 4971 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-qmmxb\" (UniqueName: \"kubernetes.io/projected/5e843ecc-f9c0-492f-bcd8-0d34447ae18a-kube-api-access-qmmxb\") pod \"dnsmasq-dns-59fd54bbff-hnxpb\" (UID: \"5e843ecc-f9c0-492f-bcd8-0d34447ae18a\") " pod="openstack/dnsmasq-dns-59fd54bbff-hnxpb" Dec 13 07:14:39 crc kubenswrapper[4971]: I1213 07:14:39.264446 4971 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/156254b1-933a-49b3-ae26-f0ba7f1418b8-combined-ca-bundle\") pod \"nova-scheduler-0\" (UID: \"156254b1-933a-49b3-ae26-f0ba7f1418b8\") " pod="openstack/nova-scheduler-0" Dec 13 07:14:39 crc kubenswrapper[4971]: I1213 07:14:39.264539 4971 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/156254b1-933a-49b3-ae26-f0ba7f1418b8-config-data\") pod \"nova-scheduler-0\" (UID: \"156254b1-933a-49b3-ae26-f0ba7f1418b8\") " pod="openstack/nova-scheduler-0" Dec 13 07:14:39 crc kubenswrapper[4971]: I1213 07:14:39.303546 4971 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-2k4fc\" (UniqueName: \"kubernetes.io/projected/156254b1-933a-49b3-ae26-f0ba7f1418b8-kube-api-access-2k4fc\") pod \"nova-scheduler-0\" (UID: \"156254b1-933a-49b3-ae26-f0ba7f1418b8\") " pod="openstack/nova-scheduler-0" Dec 13 07:14:39 crc kubenswrapper[4971]: I1213 07:14:39.356544 4971 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/b99cc168-f32c-49c0-b150-e4e730e9eacb-combined-ca-bundle\") pod \"nova-api-0\" (UID: \"b99cc168-f32c-49c0-b150-e4e730e9eacb\") " pod="openstack/nova-api-0" Dec 13 07:14:39 crc kubenswrapper[4971]: I1213 07:14:39.356894 4971 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/5e843ecc-f9c0-492f-bcd8-0d34447ae18a-ovsdbserver-nb\") pod \"dnsmasq-dns-59fd54bbff-hnxpb\" (UID: \"5e843ecc-f9c0-492f-bcd8-0d34447ae18a\") " pod="openstack/dnsmasq-dns-59fd54bbff-hnxpb" Dec 13 07:14:39 crc kubenswrapper[4971]: I1213 07:14:39.356944 4971 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/5e843ecc-f9c0-492f-bcd8-0d34447ae18a-dns-svc\") pod \"dnsmasq-dns-59fd54bbff-hnxpb\" (UID: \"5e843ecc-f9c0-492f-bcd8-0d34447ae18a\") " pod="openstack/dnsmasq-dns-59fd54bbff-hnxpb" Dec 13 07:14:39 crc kubenswrapper[4971]: I1213 07:14:39.356967 4971 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-qmmxb\" (UniqueName: \"kubernetes.io/projected/5e843ecc-f9c0-492f-bcd8-0d34447ae18a-kube-api-access-qmmxb\") pod \"dnsmasq-dns-59fd54bbff-hnxpb\" (UID: \"5e843ecc-f9c0-492f-bcd8-0d34447ae18a\") " pod="openstack/dnsmasq-dns-59fd54bbff-hnxpb" Dec 13 07:14:39 crc kubenswrapper[4971]: I1213 07:14:39.357026 4971 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/b99cc168-f32c-49c0-b150-e4e730e9eacb-config-data\") pod \"nova-api-0\" (UID: \"b99cc168-f32c-49c0-b150-e4e730e9eacb\") " pod="openstack/nova-api-0" Dec 13 07:14:39 crc kubenswrapper[4971]: I1213 07:14:39.357057 4971 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-4kxf9\" (UniqueName: \"kubernetes.io/projected/b99cc168-f32c-49c0-b150-e4e730e9eacb-kube-api-access-4kxf9\") pod \"nova-api-0\" (UID: \"b99cc168-f32c-49c0-b150-e4e730e9eacb\") " pod="openstack/nova-api-0" Dec 13 07:14:39 crc kubenswrapper[4971]: I1213 07:14:39.357087 4971 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/5e843ecc-f9c0-492f-bcd8-0d34447ae18a-ovsdbserver-sb\") pod \"dnsmasq-dns-59fd54bbff-hnxpb\" (UID: \"5e843ecc-f9c0-492f-bcd8-0d34447ae18a\") " pod="openstack/dnsmasq-dns-59fd54bbff-hnxpb" Dec 13 07:14:39 crc kubenswrapper[4971]: I1213 07:14:39.357132 4971 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/5e843ecc-f9c0-492f-bcd8-0d34447ae18a-config\") pod \"dnsmasq-dns-59fd54bbff-hnxpb\" (UID: \"5e843ecc-f9c0-492f-bcd8-0d34447ae18a\") " pod="openstack/dnsmasq-dns-59fd54bbff-hnxpb" Dec 13 07:14:39 crc kubenswrapper[4971]: I1213 07:14:39.357160 4971 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/b99cc168-f32c-49c0-b150-e4e730e9eacb-logs\") pod \"nova-api-0\" (UID: \"b99cc168-f32c-49c0-b150-e4e730e9eacb\") " pod="openstack/nova-api-0" Dec 13 07:14:39 crc kubenswrapper[4971]: I1213 07:14:39.358022 4971 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/5e843ecc-f9c0-492f-bcd8-0d34447ae18a-ovsdbserver-nb\") pod \"dnsmasq-dns-59fd54bbff-hnxpb\" (UID: \"5e843ecc-f9c0-492f-bcd8-0d34447ae18a\") " pod="openstack/dnsmasq-dns-59fd54bbff-hnxpb" Dec 13 07:14:39 crc kubenswrapper[4971]: I1213 07:14:39.358854 4971 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/5e843ecc-f9c0-492f-bcd8-0d34447ae18a-dns-svc\") pod \"dnsmasq-dns-59fd54bbff-hnxpb\" (UID: \"5e843ecc-f9c0-492f-bcd8-0d34447ae18a\") " pod="openstack/dnsmasq-dns-59fd54bbff-hnxpb" Dec 13 07:14:39 crc kubenswrapper[4971]: I1213 07:14:39.359338 4971 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/5e843ecc-f9c0-492f-bcd8-0d34447ae18a-ovsdbserver-sb\") pod \"dnsmasq-dns-59fd54bbff-hnxpb\" (UID: \"5e843ecc-f9c0-492f-bcd8-0d34447ae18a\") " pod="openstack/dnsmasq-dns-59fd54bbff-hnxpb" Dec 13 07:14:39 crc kubenswrapper[4971]: I1213 07:14:39.359344 4971 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/5e843ecc-f9c0-492f-bcd8-0d34447ae18a-config\") pod \"dnsmasq-dns-59fd54bbff-hnxpb\" (UID: \"5e843ecc-f9c0-492f-bcd8-0d34447ae18a\") " pod="openstack/dnsmasq-dns-59fd54bbff-hnxpb" Dec 13 07:14:39 crc kubenswrapper[4971]: I1213 07:14:39.392573 4971 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-qmmxb\" (UniqueName: \"kubernetes.io/projected/5e843ecc-f9c0-492f-bcd8-0d34447ae18a-kube-api-access-qmmxb\") pod \"dnsmasq-dns-59fd54bbff-hnxpb\" (UID: \"5e843ecc-f9c0-492f-bcd8-0d34447ae18a\") " pod="openstack/dnsmasq-dns-59fd54bbff-hnxpb" Dec 13 07:14:39 crc kubenswrapper[4971]: I1213 07:14:39.450079 4971 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-novncproxy-0" Dec 13 07:14:39 crc kubenswrapper[4971]: I1213 07:14:39.459058 4971 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/b99cc168-f32c-49c0-b150-e4e730e9eacb-config-data\") pod \"nova-api-0\" (UID: \"b99cc168-f32c-49c0-b150-e4e730e9eacb\") " pod="openstack/nova-api-0" Dec 13 07:14:39 crc kubenswrapper[4971]: I1213 07:14:39.459121 4971 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-4kxf9\" (UniqueName: \"kubernetes.io/projected/b99cc168-f32c-49c0-b150-e4e730e9eacb-kube-api-access-4kxf9\") pod \"nova-api-0\" (UID: \"b99cc168-f32c-49c0-b150-e4e730e9eacb\") " pod="openstack/nova-api-0" Dec 13 07:14:39 crc kubenswrapper[4971]: I1213 07:14:39.459175 4971 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/b99cc168-f32c-49c0-b150-e4e730e9eacb-logs\") pod \"nova-api-0\" (UID: \"b99cc168-f32c-49c0-b150-e4e730e9eacb\") " pod="openstack/nova-api-0" Dec 13 07:14:39 crc kubenswrapper[4971]: I1213 07:14:39.459200 4971 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/b99cc168-f32c-49c0-b150-e4e730e9eacb-combined-ca-bundle\") pod \"nova-api-0\" (UID: \"b99cc168-f32c-49c0-b150-e4e730e9eacb\") " pod="openstack/nova-api-0" Dec 13 07:14:39 crc kubenswrapper[4971]: I1213 07:14:39.471116 4971 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/b99cc168-f32c-49c0-b150-e4e730e9eacb-logs\") pod \"nova-api-0\" (UID: \"b99cc168-f32c-49c0-b150-e4e730e9eacb\") " pod="openstack/nova-api-0" Dec 13 07:14:39 crc kubenswrapper[4971]: I1213 07:14:39.471670 4971 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/b99cc168-f32c-49c0-b150-e4e730e9eacb-combined-ca-bundle\") pod \"nova-api-0\" (UID: \"b99cc168-f32c-49c0-b150-e4e730e9eacb\") " pod="openstack/nova-api-0" Dec 13 07:14:39 crc kubenswrapper[4971]: I1213 07:14:39.473954 4971 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/b99cc168-f32c-49c0-b150-e4e730e9eacb-config-data\") pod \"nova-api-0\" (UID: \"b99cc168-f32c-49c0-b150-e4e730e9eacb\") " pod="openstack/nova-api-0" Dec 13 07:14:39 crc kubenswrapper[4971]: I1213 07:14:39.491334 4971 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-metadata-0" Dec 13 07:14:39 crc kubenswrapper[4971]: I1213 07:14:39.498996 4971 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-4kxf9\" (UniqueName: \"kubernetes.io/projected/b99cc168-f32c-49c0-b150-e4e730e9eacb-kube-api-access-4kxf9\") pod \"nova-api-0\" (UID: \"b99cc168-f32c-49c0-b150-e4e730e9eacb\") " pod="openstack/nova-api-0" Dec 13 07:14:39 crc kubenswrapper[4971]: I1213 07:14:39.512242 4971 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell0-cell-mapping-rgjbg"] Dec 13 07:14:39 crc kubenswrapper[4971]: I1213 07:14:39.520426 4971 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-scheduler-0" Dec 13 07:14:39 crc kubenswrapper[4971]: I1213 07:14:39.538267 4971 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-59fd54bbff-hnxpb" Dec 13 07:14:39 crc kubenswrapper[4971]: I1213 07:14:39.578825 4971 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-0" Dec 13 07:14:39 crc kubenswrapper[4971]: I1213 07:14:39.611506 4971 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-cell-mapping-rgjbg" event={"ID":"74fcba46-f364-4c8a-a4e3-ace14451d1f7","Type":"ContainerStarted","Data":"e2f608fc6357e7c23334d2f79d959e45ce3e234e164b5586305683b5663bb302"} Dec 13 07:14:39 crc kubenswrapper[4971]: I1213 07:14:39.758851 4971 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell1-novncproxy-0"] Dec 13 07:14:39 crc kubenswrapper[4971]: I1213 07:14:39.866773 4971 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-marketplace-8rlw2" Dec 13 07:14:39 crc kubenswrapper[4971]: I1213 07:14:39.870003 4971 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-marketplace-8rlw2" Dec 13 07:14:39 crc kubenswrapper[4971]: I1213 07:14:39.952379 4971 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-marketplace-8rlw2" Dec 13 07:14:40 crc kubenswrapper[4971]: I1213 07:14:40.129590 4971 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-cell1-conductor-db-sync-kvw2t"] Dec 13 07:14:40 crc kubenswrapper[4971]: I1213 07:14:40.130868 4971 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-conductor-db-sync-kvw2t" Dec 13 07:14:40 crc kubenswrapper[4971]: I1213 07:14:40.132460 4971 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-cell1-conductor-scripts" Dec 13 07:14:40 crc kubenswrapper[4971]: I1213 07:14:40.132680 4971 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-cell1-conductor-config-data" Dec 13 07:14:40 crc kubenswrapper[4971]: I1213 07:14:40.141835 4971 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell1-conductor-db-sync-kvw2t"] Dec 13 07:14:40 crc kubenswrapper[4971]: I1213 07:14:40.172456 4971 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/6c6829bb-25ee-4102-8335-b747bdbe6121-scripts\") pod \"nova-cell1-conductor-db-sync-kvw2t\" (UID: \"6c6829bb-25ee-4102-8335-b747bdbe6121\") " pod="openstack/nova-cell1-conductor-db-sync-kvw2t" Dec 13 07:14:40 crc kubenswrapper[4971]: I1213 07:14:40.172549 4971 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-qczdv\" (UniqueName: \"kubernetes.io/projected/6c6829bb-25ee-4102-8335-b747bdbe6121-kube-api-access-qczdv\") pod \"nova-cell1-conductor-db-sync-kvw2t\" (UID: \"6c6829bb-25ee-4102-8335-b747bdbe6121\") " pod="openstack/nova-cell1-conductor-db-sync-kvw2t" Dec 13 07:14:40 crc kubenswrapper[4971]: I1213 07:14:40.172678 4971 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/6c6829bb-25ee-4102-8335-b747bdbe6121-combined-ca-bundle\") pod \"nova-cell1-conductor-db-sync-kvw2t\" (UID: \"6c6829bb-25ee-4102-8335-b747bdbe6121\") " pod="openstack/nova-cell1-conductor-db-sync-kvw2t" Dec 13 07:14:40 crc kubenswrapper[4971]: I1213 07:14:40.172759 4971 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/6c6829bb-25ee-4102-8335-b747bdbe6121-config-data\") pod \"nova-cell1-conductor-db-sync-kvw2t\" (UID: \"6c6829bb-25ee-4102-8335-b747bdbe6121\") " pod="openstack/nova-cell1-conductor-db-sync-kvw2t" Dec 13 07:14:40 crc kubenswrapper[4971]: I1213 07:14:40.207052 4971 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-metadata-0"] Dec 13 07:14:40 crc kubenswrapper[4971]: I1213 07:14:40.234923 4971 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-scheduler-0"] Dec 13 07:14:40 crc kubenswrapper[4971]: I1213 07:14:40.275789 4971 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/6c6829bb-25ee-4102-8335-b747bdbe6121-scripts\") pod \"nova-cell1-conductor-db-sync-kvw2t\" (UID: \"6c6829bb-25ee-4102-8335-b747bdbe6121\") " pod="openstack/nova-cell1-conductor-db-sync-kvw2t" Dec 13 07:14:40 crc kubenswrapper[4971]: I1213 07:14:40.275846 4971 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-qczdv\" (UniqueName: \"kubernetes.io/projected/6c6829bb-25ee-4102-8335-b747bdbe6121-kube-api-access-qczdv\") pod \"nova-cell1-conductor-db-sync-kvw2t\" (UID: \"6c6829bb-25ee-4102-8335-b747bdbe6121\") " pod="openstack/nova-cell1-conductor-db-sync-kvw2t" Dec 13 07:14:40 crc kubenswrapper[4971]: I1213 07:14:40.277314 4971 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/6c6829bb-25ee-4102-8335-b747bdbe6121-combined-ca-bundle\") pod \"nova-cell1-conductor-db-sync-kvw2t\" (UID: \"6c6829bb-25ee-4102-8335-b747bdbe6121\") " pod="openstack/nova-cell1-conductor-db-sync-kvw2t" Dec 13 07:14:40 crc kubenswrapper[4971]: I1213 07:14:40.277466 4971 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/6c6829bb-25ee-4102-8335-b747bdbe6121-config-data\") pod \"nova-cell1-conductor-db-sync-kvw2t\" (UID: \"6c6829bb-25ee-4102-8335-b747bdbe6121\") " pod="openstack/nova-cell1-conductor-db-sync-kvw2t" Dec 13 07:14:40 crc kubenswrapper[4971]: I1213 07:14:40.294250 4971 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/6c6829bb-25ee-4102-8335-b747bdbe6121-scripts\") pod \"nova-cell1-conductor-db-sync-kvw2t\" (UID: \"6c6829bb-25ee-4102-8335-b747bdbe6121\") " pod="openstack/nova-cell1-conductor-db-sync-kvw2t" Dec 13 07:14:40 crc kubenswrapper[4971]: I1213 07:14:40.294404 4971 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/6c6829bb-25ee-4102-8335-b747bdbe6121-combined-ca-bundle\") pod \"nova-cell1-conductor-db-sync-kvw2t\" (UID: \"6c6829bb-25ee-4102-8335-b747bdbe6121\") " pod="openstack/nova-cell1-conductor-db-sync-kvw2t" Dec 13 07:14:40 crc kubenswrapper[4971]: I1213 07:14:40.296597 4971 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/6c6829bb-25ee-4102-8335-b747bdbe6121-config-data\") pod \"nova-cell1-conductor-db-sync-kvw2t\" (UID: \"6c6829bb-25ee-4102-8335-b747bdbe6121\") " pod="openstack/nova-cell1-conductor-db-sync-kvw2t" Dec 13 07:14:40 crc kubenswrapper[4971]: I1213 07:14:40.304460 4971 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-qczdv\" (UniqueName: \"kubernetes.io/projected/6c6829bb-25ee-4102-8335-b747bdbe6121-kube-api-access-qczdv\") pod \"nova-cell1-conductor-db-sync-kvw2t\" (UID: \"6c6829bb-25ee-4102-8335-b747bdbe6121\") " pod="openstack/nova-cell1-conductor-db-sync-kvw2t" Dec 13 07:14:40 crc kubenswrapper[4971]: I1213 07:14:40.309462 4971 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-api-0"] Dec 13 07:14:40 crc kubenswrapper[4971]: I1213 07:14:40.317315 4971 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-59fd54bbff-hnxpb"] Dec 13 07:14:40 crc kubenswrapper[4971]: W1213 07:14:40.329640 4971 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod5e843ecc_f9c0_492f_bcd8_0d34447ae18a.slice/crio-acf3b65a85b7c3fafbce46e07c02d107a4b68ebb1f81a0419c069fe8f6987e36 WatchSource:0}: Error finding container acf3b65a85b7c3fafbce46e07c02d107a4b68ebb1f81a0419c069fe8f6987e36: Status 404 returned error can't find the container with id acf3b65a85b7c3fafbce46e07c02d107a4b68ebb1f81a0419c069fe8f6987e36 Dec 13 07:14:40 crc kubenswrapper[4971]: I1213 07:14:40.462179 4971 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-conductor-db-sync-kvw2t" Dec 13 07:14:40 crc kubenswrapper[4971]: I1213 07:14:40.645465 4971 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"b99cc168-f32c-49c0-b150-e4e730e9eacb","Type":"ContainerStarted","Data":"8dc63f657d1685554d528cb2fe27443e7b84fc13541695786e7ea8bcaa4fc57c"} Dec 13 07:14:40 crc kubenswrapper[4971]: I1213 07:14:40.649749 4971 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-cell-mapping-rgjbg" event={"ID":"74fcba46-f364-4c8a-a4e3-ace14451d1f7","Type":"ContainerStarted","Data":"76f58f854276b22f5878baf9dc07e632f05e6335bb6779edc28b4315b134c334"} Dec 13 07:14:40 crc kubenswrapper[4971]: I1213 07:14:40.657654 4971 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"feaf3daf-1c1f-493f-8358-8e9afb515229","Type":"ContainerStarted","Data":"01ca0cf2c1e267330b506796c2bcae569f599c9b122130ed5284728f73474b3e"} Dec 13 07:14:40 crc kubenswrapper[4971]: I1213 07:14:40.661929 4971 generic.go:334] "Generic (PLEG): container finished" podID="5e843ecc-f9c0-492f-bcd8-0d34447ae18a" containerID="7ef782e1bad569b3a550d4733a33011d066fce8a39433023866c5c83e6ff80ec" exitCode=0 Dec 13 07:14:40 crc kubenswrapper[4971]: I1213 07:14:40.662000 4971 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-59fd54bbff-hnxpb" event={"ID":"5e843ecc-f9c0-492f-bcd8-0d34447ae18a","Type":"ContainerDied","Data":"7ef782e1bad569b3a550d4733a33011d066fce8a39433023866c5c83e6ff80ec"} Dec 13 07:14:40 crc kubenswrapper[4971]: I1213 07:14:40.662032 4971 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-59fd54bbff-hnxpb" event={"ID":"5e843ecc-f9c0-492f-bcd8-0d34447ae18a","Type":"ContainerStarted","Data":"acf3b65a85b7c3fafbce46e07c02d107a4b68ebb1f81a0419c069fe8f6987e36"} Dec 13 07:14:40 crc kubenswrapper[4971]: I1213 07:14:40.670288 4971 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-cell0-cell-mapping-rgjbg" podStartSLOduration=2.670267363 podStartE2EDuration="2.670267363s" podCreationTimestamp="2025-12-13 07:14:38 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-13 07:14:40.66324142 +0000 UTC m=+1537.267650878" watchObservedRunningTime="2025-12-13 07:14:40.670267363 +0000 UTC m=+1537.274676811" Dec 13 07:14:40 crc kubenswrapper[4971]: I1213 07:14:40.670612 4971 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-novncproxy-0" event={"ID":"9bb1801a-532f-4b09-9ae3-d4b61e8a9c97","Type":"ContainerStarted","Data":"35cabbbbe703889299039bf63469bfd0df4dceb8cc391e80c4911395615172c5"} Dec 13 07:14:40 crc kubenswrapper[4971]: I1213 07:14:40.678860 4971 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-scheduler-0" event={"ID":"156254b1-933a-49b3-ae26-f0ba7f1418b8","Type":"ContainerStarted","Data":"5d89258312c0366c0d0d89e9bd9e968716f95d94345a8555090415c78567768f"} Dec 13 07:14:40 crc kubenswrapper[4971]: I1213 07:14:40.774456 4971 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-marketplace-8rlw2" Dec 13 07:14:40 crc kubenswrapper[4971]: I1213 07:14:40.833729 4971 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-8rlw2"] Dec 13 07:14:41 crc kubenswrapper[4971]: I1213 07:14:41.007934 4971 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell1-conductor-db-sync-kvw2t"] Dec 13 07:14:41 crc kubenswrapper[4971]: I1213 07:14:41.702481 4971 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-59fd54bbff-hnxpb" event={"ID":"5e843ecc-f9c0-492f-bcd8-0d34447ae18a","Type":"ContainerStarted","Data":"1d77379c2441ec0288702f3e7f47f5164ec26f2f604e3edec19699d040133047"} Dec 13 07:14:41 crc kubenswrapper[4971]: I1213 07:14:41.704013 4971 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/dnsmasq-dns-59fd54bbff-hnxpb" Dec 13 07:14:41 crc kubenswrapper[4971]: I1213 07:14:41.726602 4971 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/dnsmasq-dns-59fd54bbff-hnxpb" podStartSLOduration=2.7265823989999998 podStartE2EDuration="2.726582399s" podCreationTimestamp="2025-12-13 07:14:39 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-13 07:14:41.725096813 +0000 UTC m=+1538.329506261" watchObservedRunningTime="2025-12-13 07:14:41.726582399 +0000 UTC m=+1538.330991847" Dec 13 07:14:42 crc kubenswrapper[4971]: I1213 07:14:42.600981 4971 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/ceilometer-0" Dec 13 07:14:42 crc kubenswrapper[4971]: I1213 07:14:42.715227 4971 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-conductor-db-sync-kvw2t" event={"ID":"6c6829bb-25ee-4102-8335-b747bdbe6121","Type":"ContainerStarted","Data":"0495140fdf836519f2c1ebcfffbac483fcc73ca5e8f9532225253c5f477758a7"} Dec 13 07:14:42 crc kubenswrapper[4971]: I1213 07:14:42.716575 4971 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-marketplace-8rlw2" podUID="d5c250a4-d784-4508-9433-d684c6b01531" containerName="registry-server" containerID="cri-o://ae179455b245343d250e77a8751d8b15e9f429baa9c54c6b6b56737c04ca7cdf" gracePeriod=2 Dec 13 07:14:42 crc kubenswrapper[4971]: I1213 07:14:42.866246 4971 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-cell1-novncproxy-0"] Dec 13 07:14:42 crc kubenswrapper[4971]: I1213 07:14:42.875862 4971 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-metadata-0"] Dec 13 07:14:43 crc kubenswrapper[4971]: I1213 07:14:43.732302 4971 generic.go:334] "Generic (PLEG): container finished" podID="d5c250a4-d784-4508-9433-d684c6b01531" containerID="ae179455b245343d250e77a8751d8b15e9f429baa9c54c6b6b56737c04ca7cdf" exitCode=0 Dec 13 07:14:43 crc kubenswrapper[4971]: I1213 07:14:43.732503 4971 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-8rlw2" event={"ID":"d5c250a4-d784-4508-9433-d684c6b01531","Type":"ContainerDied","Data":"ae179455b245343d250e77a8751d8b15e9f429baa9c54c6b6b56737c04ca7cdf"} Dec 13 07:14:43 crc kubenswrapper[4971]: I1213 07:14:43.732713 4971 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-8rlw2" event={"ID":"d5c250a4-d784-4508-9433-d684c6b01531","Type":"ContainerDied","Data":"bb8d6547555de0ffa2354fe79b4ef9710a8a1fce0aa02b49ccc3c9e52c8ef897"} Dec 13 07:14:43 crc kubenswrapper[4971]: I1213 07:14:43.732734 4971 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="bb8d6547555de0ffa2354fe79b4ef9710a8a1fce0aa02b49ccc3c9e52c8ef897" Dec 13 07:14:43 crc kubenswrapper[4971]: I1213 07:14:43.734619 4971 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-conductor-db-sync-kvw2t" event={"ID":"6c6829bb-25ee-4102-8335-b747bdbe6121","Type":"ContainerStarted","Data":"cc91778c66ca205f06be892439fa7bd0075527b3d7db3226d2d4b560aac675ab"} Dec 13 07:14:43 crc kubenswrapper[4971]: I1213 07:14:43.739968 4971 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"feaf3daf-1c1f-493f-8358-8e9afb515229","Type":"ContainerStarted","Data":"0fe09b92dde246aeee405678c37a73bac16ad892f9fbfd5624b317116b86880b"} Dec 13 07:14:43 crc kubenswrapper[4971]: I1213 07:14:43.741753 4971 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-novncproxy-0" event={"ID":"9bb1801a-532f-4b09-9ae3-d4b61e8a9c97","Type":"ContainerStarted","Data":"f48369c49c58c454eabe532f1f904b7c0c3a079ce470fb0caf73489fc5a83168"} Dec 13 07:14:43 crc kubenswrapper[4971]: I1213 07:14:43.741890 4971 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-cell1-novncproxy-0" podUID="9bb1801a-532f-4b09-9ae3-d4b61e8a9c97" containerName="nova-cell1-novncproxy-novncproxy" containerID="cri-o://f48369c49c58c454eabe532f1f904b7c0c3a079ce470fb0caf73489fc5a83168" gracePeriod=30 Dec 13 07:14:43 crc kubenswrapper[4971]: I1213 07:14:43.750257 4971 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"b99cc168-f32c-49c0-b150-e4e730e9eacb","Type":"ContainerStarted","Data":"28b75774392e9b16f160e9594f4ecede802f1fec9d3a898016414cd66da31587"} Dec 13 07:14:43 crc kubenswrapper[4971]: I1213 07:14:43.753122 4971 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-cell1-conductor-db-sync-kvw2t" podStartSLOduration=3.7531024410000002 podStartE2EDuration="3.753102441s" podCreationTimestamp="2025-12-13 07:14:40 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-13 07:14:43.751964654 +0000 UTC m=+1540.356374102" watchObservedRunningTime="2025-12-13 07:14:43.753102441 +0000 UTC m=+1540.357511889" Dec 13 07:14:43 crc kubenswrapper[4971]: I1213 07:14:43.807772 4971 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-cell1-novncproxy-0" podStartSLOduration=2.270602943 podStartE2EDuration="5.807752696s" podCreationTimestamp="2025-12-13 07:14:38 +0000 UTC" firstStartedPulling="2025-12-13 07:14:39.78871037 +0000 UTC m=+1536.393119808" lastFinishedPulling="2025-12-13 07:14:43.325860113 +0000 UTC m=+1539.930269561" observedRunningTime="2025-12-13 07:14:43.770634223 +0000 UTC m=+1540.375043681" watchObservedRunningTime="2025-12-13 07:14:43.807752696 +0000 UTC m=+1540.412162144" Dec 13 07:14:43 crc kubenswrapper[4971]: I1213 07:14:43.895574 4971 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-8rlw2" Dec 13 07:14:43 crc kubenswrapper[4971]: I1213 07:14:43.976800 4971 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-cl57q\" (UniqueName: \"kubernetes.io/projected/d5c250a4-d784-4508-9433-d684c6b01531-kube-api-access-cl57q\") pod \"d5c250a4-d784-4508-9433-d684c6b01531\" (UID: \"d5c250a4-d784-4508-9433-d684c6b01531\") " Dec 13 07:14:43 crc kubenswrapper[4971]: I1213 07:14:43.976940 4971 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/d5c250a4-d784-4508-9433-d684c6b01531-catalog-content\") pod \"d5c250a4-d784-4508-9433-d684c6b01531\" (UID: \"d5c250a4-d784-4508-9433-d684c6b01531\") " Dec 13 07:14:43 crc kubenswrapper[4971]: I1213 07:14:43.977160 4971 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/d5c250a4-d784-4508-9433-d684c6b01531-utilities\") pod \"d5c250a4-d784-4508-9433-d684c6b01531\" (UID: \"d5c250a4-d784-4508-9433-d684c6b01531\") " Dec 13 07:14:43 crc kubenswrapper[4971]: I1213 07:14:43.978501 4971 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/d5c250a4-d784-4508-9433-d684c6b01531-utilities" (OuterVolumeSpecName: "utilities") pod "d5c250a4-d784-4508-9433-d684c6b01531" (UID: "d5c250a4-d784-4508-9433-d684c6b01531"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 13 07:14:44 crc kubenswrapper[4971]: I1213 07:14:44.002753 4971 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/d5c250a4-d784-4508-9433-d684c6b01531-kube-api-access-cl57q" (OuterVolumeSpecName: "kube-api-access-cl57q") pod "d5c250a4-d784-4508-9433-d684c6b01531" (UID: "d5c250a4-d784-4508-9433-d684c6b01531"). InnerVolumeSpecName "kube-api-access-cl57q". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 13 07:14:44 crc kubenswrapper[4971]: I1213 07:14:44.003926 4971 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/d5c250a4-d784-4508-9433-d684c6b01531-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "d5c250a4-d784-4508-9433-d684c6b01531" (UID: "d5c250a4-d784-4508-9433-d684c6b01531"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 13 07:14:44 crc kubenswrapper[4971]: I1213 07:14:44.079301 4971 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/d5c250a4-d784-4508-9433-d684c6b01531-utilities\") on node \"crc\" DevicePath \"\"" Dec 13 07:14:44 crc kubenswrapper[4971]: I1213 07:14:44.079344 4971 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-cl57q\" (UniqueName: \"kubernetes.io/projected/d5c250a4-d784-4508-9433-d684c6b01531-kube-api-access-cl57q\") on node \"crc\" DevicePath \"\"" Dec 13 07:14:44 crc kubenswrapper[4971]: I1213 07:14:44.079357 4971 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/d5c250a4-d784-4508-9433-d684c6b01531-catalog-content\") on node \"crc\" DevicePath \"\"" Dec 13 07:14:44 crc kubenswrapper[4971]: I1213 07:14:44.451397 4971 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-cell1-novncproxy-0" Dec 13 07:14:44 crc kubenswrapper[4971]: I1213 07:14:44.758307 4971 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-scheduler-0" event={"ID":"156254b1-933a-49b3-ae26-f0ba7f1418b8","Type":"ContainerStarted","Data":"97dd6db8c1ca363f1d2b43537619b59bc570a33e67bb41c9d1803485ae0de49e"} Dec 13 07:14:44 crc kubenswrapper[4971]: I1213 07:14:44.760258 4971 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"b99cc168-f32c-49c0-b150-e4e730e9eacb","Type":"ContainerStarted","Data":"de4bd8652eb387c905466a27a8920b3fcec43795ba4ca6babe7f7fc4a4f3f1bd"} Dec 13 07:14:44 crc kubenswrapper[4971]: I1213 07:14:44.762432 4971 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-8rlw2" Dec 13 07:14:44 crc kubenswrapper[4971]: I1213 07:14:44.762504 4971 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-metadata-0" podUID="feaf3daf-1c1f-493f-8358-8e9afb515229" containerName="nova-metadata-log" containerID="cri-o://0fe09b92dde246aeee405678c37a73bac16ad892f9fbfd5624b317116b86880b" gracePeriod=30 Dec 13 07:14:44 crc kubenswrapper[4971]: I1213 07:14:44.762573 4971 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-metadata-0" podUID="feaf3daf-1c1f-493f-8358-8e9afb515229" containerName="nova-metadata-metadata" containerID="cri-o://0b638189bdde08ddc7e54408f42820725c6476fdf517c971e2ae85109f691e2f" gracePeriod=30 Dec 13 07:14:44 crc kubenswrapper[4971]: I1213 07:14:44.762433 4971 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"feaf3daf-1c1f-493f-8358-8e9afb515229","Type":"ContainerStarted","Data":"0b638189bdde08ddc7e54408f42820725c6476fdf517c971e2ae85109f691e2f"} Dec 13 07:14:44 crc kubenswrapper[4971]: I1213 07:14:44.795766 4971 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-scheduler-0" podStartSLOduration=2.656005461 podStartE2EDuration="5.795745011s" podCreationTimestamp="2025-12-13 07:14:39 +0000 UTC" firstStartedPulling="2025-12-13 07:14:40.24120672 +0000 UTC m=+1536.845616168" lastFinishedPulling="2025-12-13 07:14:43.380946269 +0000 UTC m=+1539.985355718" observedRunningTime="2025-12-13 07:14:44.782021703 +0000 UTC m=+1541.386431161" watchObservedRunningTime="2025-12-13 07:14:44.795745011 +0000 UTC m=+1541.400154469" Dec 13 07:14:44 crc kubenswrapper[4971]: I1213 07:14:44.807094 4971 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-api-0" podStartSLOduration=2.746217173 podStartE2EDuration="5.80707303s" podCreationTimestamp="2025-12-13 07:14:39 +0000 UTC" firstStartedPulling="2025-12-13 07:14:40.31919389 +0000 UTC m=+1536.923603348" lastFinishedPulling="2025-12-13 07:14:43.380049757 +0000 UTC m=+1539.984459205" observedRunningTime="2025-12-13 07:14:44.804834675 +0000 UTC m=+1541.409244123" watchObservedRunningTime="2025-12-13 07:14:44.80707303 +0000 UTC m=+1541.411482478" Dec 13 07:14:44 crc kubenswrapper[4971]: I1213 07:14:44.826019 4971 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-metadata-0" podStartSLOduration=3.694090799 podStartE2EDuration="6.825985516s" podCreationTimestamp="2025-12-13 07:14:38 +0000 UTC" firstStartedPulling="2025-12-13 07:14:40.229350068 +0000 UTC m=+1536.833759506" lastFinishedPulling="2025-12-13 07:14:43.361244775 +0000 UTC m=+1539.965654223" observedRunningTime="2025-12-13 07:14:44.824191411 +0000 UTC m=+1541.428600859" watchObservedRunningTime="2025-12-13 07:14:44.825985516 +0000 UTC m=+1541.430394954" Dec 13 07:14:44 crc kubenswrapper[4971]: I1213 07:14:44.845940 4971 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-8rlw2"] Dec 13 07:14:44 crc kubenswrapper[4971]: I1213 07:14:44.883975 4971 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-marketplace-8rlw2"] Dec 13 07:14:45 crc kubenswrapper[4971]: I1213 07:14:45.362042 4971 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-metadata-0" Dec 13 07:14:45 crc kubenswrapper[4971]: I1213 07:14:45.505909 4971 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/feaf3daf-1c1f-493f-8358-8e9afb515229-config-data\") pod \"feaf3daf-1c1f-493f-8358-8e9afb515229\" (UID: \"feaf3daf-1c1f-493f-8358-8e9afb515229\") " Dec 13 07:14:45 crc kubenswrapper[4971]: I1213 07:14:45.506056 4971 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/feaf3daf-1c1f-493f-8358-8e9afb515229-logs\") pod \"feaf3daf-1c1f-493f-8358-8e9afb515229\" (UID: \"feaf3daf-1c1f-493f-8358-8e9afb515229\") " Dec 13 07:14:45 crc kubenswrapper[4971]: I1213 07:14:45.506088 4971 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-6fdl2\" (UniqueName: \"kubernetes.io/projected/feaf3daf-1c1f-493f-8358-8e9afb515229-kube-api-access-6fdl2\") pod \"feaf3daf-1c1f-493f-8358-8e9afb515229\" (UID: \"feaf3daf-1c1f-493f-8358-8e9afb515229\") " Dec 13 07:14:45 crc kubenswrapper[4971]: I1213 07:14:45.506369 4971 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/feaf3daf-1c1f-493f-8358-8e9afb515229-combined-ca-bundle\") pod \"feaf3daf-1c1f-493f-8358-8e9afb515229\" (UID: \"feaf3daf-1c1f-493f-8358-8e9afb515229\") " Dec 13 07:14:45 crc kubenswrapper[4971]: I1213 07:14:45.507887 4971 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/feaf3daf-1c1f-493f-8358-8e9afb515229-logs" (OuterVolumeSpecName: "logs") pod "feaf3daf-1c1f-493f-8358-8e9afb515229" (UID: "feaf3daf-1c1f-493f-8358-8e9afb515229"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 13 07:14:45 crc kubenswrapper[4971]: I1213 07:14:45.511884 4971 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/feaf3daf-1c1f-493f-8358-8e9afb515229-kube-api-access-6fdl2" (OuterVolumeSpecName: "kube-api-access-6fdl2") pod "feaf3daf-1c1f-493f-8358-8e9afb515229" (UID: "feaf3daf-1c1f-493f-8358-8e9afb515229"). InnerVolumeSpecName "kube-api-access-6fdl2". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 13 07:14:45 crc kubenswrapper[4971]: I1213 07:14:45.539080 4971 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/feaf3daf-1c1f-493f-8358-8e9afb515229-config-data" (OuterVolumeSpecName: "config-data") pod "feaf3daf-1c1f-493f-8358-8e9afb515229" (UID: "feaf3daf-1c1f-493f-8358-8e9afb515229"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 13 07:14:45 crc kubenswrapper[4971]: I1213 07:14:45.542134 4971 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/feaf3daf-1c1f-493f-8358-8e9afb515229-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "feaf3daf-1c1f-493f-8358-8e9afb515229" (UID: "feaf3daf-1c1f-493f-8358-8e9afb515229"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 13 07:14:45 crc kubenswrapper[4971]: I1213 07:14:45.608306 4971 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/feaf3daf-1c1f-493f-8358-8e9afb515229-logs\") on node \"crc\" DevicePath \"\"" Dec 13 07:14:45 crc kubenswrapper[4971]: I1213 07:14:45.608342 4971 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-6fdl2\" (UniqueName: \"kubernetes.io/projected/feaf3daf-1c1f-493f-8358-8e9afb515229-kube-api-access-6fdl2\") on node \"crc\" DevicePath \"\"" Dec 13 07:14:45 crc kubenswrapper[4971]: I1213 07:14:45.608352 4971 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/feaf3daf-1c1f-493f-8358-8e9afb515229-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 13 07:14:45 crc kubenswrapper[4971]: I1213 07:14:45.608362 4971 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/feaf3daf-1c1f-493f-8358-8e9afb515229-config-data\") on node \"crc\" DevicePath \"\"" Dec 13 07:14:45 crc kubenswrapper[4971]: I1213 07:14:45.773672 4971 generic.go:334] "Generic (PLEG): container finished" podID="feaf3daf-1c1f-493f-8358-8e9afb515229" containerID="0b638189bdde08ddc7e54408f42820725c6476fdf517c971e2ae85109f691e2f" exitCode=0 Dec 13 07:14:45 crc kubenswrapper[4971]: I1213 07:14:45.773714 4971 generic.go:334] "Generic (PLEG): container finished" podID="feaf3daf-1c1f-493f-8358-8e9afb515229" containerID="0fe09b92dde246aeee405678c37a73bac16ad892f9fbfd5624b317116b86880b" exitCode=143 Dec 13 07:14:45 crc kubenswrapper[4971]: I1213 07:14:45.774021 4971 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-metadata-0" Dec 13 07:14:45 crc kubenswrapper[4971]: I1213 07:14:45.782105 4971 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="d5c250a4-d784-4508-9433-d684c6b01531" path="/var/lib/kubelet/pods/d5c250a4-d784-4508-9433-d684c6b01531/volumes" Dec 13 07:14:45 crc kubenswrapper[4971]: I1213 07:14:45.783101 4971 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"feaf3daf-1c1f-493f-8358-8e9afb515229","Type":"ContainerDied","Data":"0b638189bdde08ddc7e54408f42820725c6476fdf517c971e2ae85109f691e2f"} Dec 13 07:14:45 crc kubenswrapper[4971]: I1213 07:14:45.783130 4971 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"feaf3daf-1c1f-493f-8358-8e9afb515229","Type":"ContainerDied","Data":"0fe09b92dde246aeee405678c37a73bac16ad892f9fbfd5624b317116b86880b"} Dec 13 07:14:45 crc kubenswrapper[4971]: I1213 07:14:45.783143 4971 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"feaf3daf-1c1f-493f-8358-8e9afb515229","Type":"ContainerDied","Data":"01ca0cf2c1e267330b506796c2bcae569f599c9b122130ed5284728f73474b3e"} Dec 13 07:14:45 crc kubenswrapper[4971]: I1213 07:14:45.783162 4971 scope.go:117] "RemoveContainer" containerID="0b638189bdde08ddc7e54408f42820725c6476fdf517c971e2ae85109f691e2f" Dec 13 07:14:45 crc kubenswrapper[4971]: I1213 07:14:45.819832 4971 scope.go:117] "RemoveContainer" containerID="0fe09b92dde246aeee405678c37a73bac16ad892f9fbfd5624b317116b86880b" Dec 13 07:14:45 crc kubenswrapper[4971]: I1213 07:14:45.820110 4971 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-metadata-0"] Dec 13 07:14:45 crc kubenswrapper[4971]: I1213 07:14:45.845452 4971 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-metadata-0"] Dec 13 07:14:45 crc kubenswrapper[4971]: I1213 07:14:45.848726 4971 scope.go:117] "RemoveContainer" containerID="0b638189bdde08ddc7e54408f42820725c6476fdf517c971e2ae85109f691e2f" Dec 13 07:14:45 crc kubenswrapper[4971]: E1213 07:14:45.849196 4971 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"0b638189bdde08ddc7e54408f42820725c6476fdf517c971e2ae85109f691e2f\": container with ID starting with 0b638189bdde08ddc7e54408f42820725c6476fdf517c971e2ae85109f691e2f not found: ID does not exist" containerID="0b638189bdde08ddc7e54408f42820725c6476fdf517c971e2ae85109f691e2f" Dec 13 07:14:45 crc kubenswrapper[4971]: I1213 07:14:45.849227 4971 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"0b638189bdde08ddc7e54408f42820725c6476fdf517c971e2ae85109f691e2f"} err="failed to get container status \"0b638189bdde08ddc7e54408f42820725c6476fdf517c971e2ae85109f691e2f\": rpc error: code = NotFound desc = could not find container \"0b638189bdde08ddc7e54408f42820725c6476fdf517c971e2ae85109f691e2f\": container with ID starting with 0b638189bdde08ddc7e54408f42820725c6476fdf517c971e2ae85109f691e2f not found: ID does not exist" Dec 13 07:14:45 crc kubenswrapper[4971]: I1213 07:14:45.849249 4971 scope.go:117] "RemoveContainer" containerID="0fe09b92dde246aeee405678c37a73bac16ad892f9fbfd5624b317116b86880b" Dec 13 07:14:45 crc kubenswrapper[4971]: E1213 07:14:45.849484 4971 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"0fe09b92dde246aeee405678c37a73bac16ad892f9fbfd5624b317116b86880b\": container with ID starting with 0fe09b92dde246aeee405678c37a73bac16ad892f9fbfd5624b317116b86880b not found: ID does not exist" containerID="0fe09b92dde246aeee405678c37a73bac16ad892f9fbfd5624b317116b86880b" Dec 13 07:14:45 crc kubenswrapper[4971]: I1213 07:14:45.849503 4971 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"0fe09b92dde246aeee405678c37a73bac16ad892f9fbfd5624b317116b86880b"} err="failed to get container status \"0fe09b92dde246aeee405678c37a73bac16ad892f9fbfd5624b317116b86880b\": rpc error: code = NotFound desc = could not find container \"0fe09b92dde246aeee405678c37a73bac16ad892f9fbfd5624b317116b86880b\": container with ID starting with 0fe09b92dde246aeee405678c37a73bac16ad892f9fbfd5624b317116b86880b not found: ID does not exist" Dec 13 07:14:45 crc kubenswrapper[4971]: I1213 07:14:45.849529 4971 scope.go:117] "RemoveContainer" containerID="0b638189bdde08ddc7e54408f42820725c6476fdf517c971e2ae85109f691e2f" Dec 13 07:14:45 crc kubenswrapper[4971]: I1213 07:14:45.849779 4971 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"0b638189bdde08ddc7e54408f42820725c6476fdf517c971e2ae85109f691e2f"} err="failed to get container status \"0b638189bdde08ddc7e54408f42820725c6476fdf517c971e2ae85109f691e2f\": rpc error: code = NotFound desc = could not find container \"0b638189bdde08ddc7e54408f42820725c6476fdf517c971e2ae85109f691e2f\": container with ID starting with 0b638189bdde08ddc7e54408f42820725c6476fdf517c971e2ae85109f691e2f not found: ID does not exist" Dec 13 07:14:45 crc kubenswrapper[4971]: I1213 07:14:45.849806 4971 scope.go:117] "RemoveContainer" containerID="0fe09b92dde246aeee405678c37a73bac16ad892f9fbfd5624b317116b86880b" Dec 13 07:14:45 crc kubenswrapper[4971]: I1213 07:14:45.850003 4971 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"0fe09b92dde246aeee405678c37a73bac16ad892f9fbfd5624b317116b86880b"} err="failed to get container status \"0fe09b92dde246aeee405678c37a73bac16ad892f9fbfd5624b317116b86880b\": rpc error: code = NotFound desc = could not find container \"0fe09b92dde246aeee405678c37a73bac16ad892f9fbfd5624b317116b86880b\": container with ID starting with 0fe09b92dde246aeee405678c37a73bac16ad892f9fbfd5624b317116b86880b not found: ID does not exist" Dec 13 07:14:45 crc kubenswrapper[4971]: I1213 07:14:45.853883 4971 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-metadata-0"] Dec 13 07:14:45 crc kubenswrapper[4971]: E1213 07:14:45.859458 4971 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="d5c250a4-d784-4508-9433-d684c6b01531" containerName="extract-utilities" Dec 13 07:14:45 crc kubenswrapper[4971]: I1213 07:14:45.859504 4971 state_mem.go:107] "Deleted CPUSet assignment" podUID="d5c250a4-d784-4508-9433-d684c6b01531" containerName="extract-utilities" Dec 13 07:14:45 crc kubenswrapper[4971]: E1213 07:14:45.859562 4971 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="d5c250a4-d784-4508-9433-d684c6b01531" containerName="extract-content" Dec 13 07:14:45 crc kubenswrapper[4971]: I1213 07:14:45.859573 4971 state_mem.go:107] "Deleted CPUSet assignment" podUID="d5c250a4-d784-4508-9433-d684c6b01531" containerName="extract-content" Dec 13 07:14:45 crc kubenswrapper[4971]: E1213 07:14:45.859587 4971 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="feaf3daf-1c1f-493f-8358-8e9afb515229" containerName="nova-metadata-metadata" Dec 13 07:14:45 crc kubenswrapper[4971]: I1213 07:14:45.859596 4971 state_mem.go:107] "Deleted CPUSet assignment" podUID="feaf3daf-1c1f-493f-8358-8e9afb515229" containerName="nova-metadata-metadata" Dec 13 07:14:45 crc kubenswrapper[4971]: E1213 07:14:45.859613 4971 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="feaf3daf-1c1f-493f-8358-8e9afb515229" containerName="nova-metadata-log" Dec 13 07:14:45 crc kubenswrapper[4971]: I1213 07:14:45.859621 4971 state_mem.go:107] "Deleted CPUSet assignment" podUID="feaf3daf-1c1f-493f-8358-8e9afb515229" containerName="nova-metadata-log" Dec 13 07:14:45 crc kubenswrapper[4971]: E1213 07:14:45.859651 4971 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="d5c250a4-d784-4508-9433-d684c6b01531" containerName="registry-server" Dec 13 07:14:45 crc kubenswrapper[4971]: I1213 07:14:45.859660 4971 state_mem.go:107] "Deleted CPUSet assignment" podUID="d5c250a4-d784-4508-9433-d684c6b01531" containerName="registry-server" Dec 13 07:14:45 crc kubenswrapper[4971]: I1213 07:14:45.860853 4971 memory_manager.go:354] "RemoveStaleState removing state" podUID="feaf3daf-1c1f-493f-8358-8e9afb515229" containerName="nova-metadata-metadata" Dec 13 07:14:45 crc kubenswrapper[4971]: I1213 07:14:45.860902 4971 memory_manager.go:354] "RemoveStaleState removing state" podUID="feaf3daf-1c1f-493f-8358-8e9afb515229" containerName="nova-metadata-log" Dec 13 07:14:45 crc kubenswrapper[4971]: I1213 07:14:45.860917 4971 memory_manager.go:354] "RemoveStaleState removing state" podUID="d5c250a4-d784-4508-9433-d684c6b01531" containerName="registry-server" Dec 13 07:14:45 crc kubenswrapper[4971]: I1213 07:14:45.862928 4971 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-metadata-0"] Dec 13 07:14:45 crc kubenswrapper[4971]: I1213 07:14:45.863037 4971 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-metadata-0" Dec 13 07:14:45 crc kubenswrapper[4971]: I1213 07:14:45.865727 4971 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-nova-metadata-internal-svc" Dec 13 07:14:45 crc kubenswrapper[4971]: I1213 07:14:45.868313 4971 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-metadata-config-data" Dec 13 07:14:46 crc kubenswrapper[4971]: I1213 07:14:46.014713 4971 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/3eaa9674-9cd0-497f-80ca-27067fa54ab1-combined-ca-bundle\") pod \"nova-metadata-0\" (UID: \"3eaa9674-9cd0-497f-80ca-27067fa54ab1\") " pod="openstack/nova-metadata-0" Dec 13 07:14:46 crc kubenswrapper[4971]: I1213 07:14:46.014765 4971 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-4wcdz\" (UniqueName: \"kubernetes.io/projected/3eaa9674-9cd0-497f-80ca-27067fa54ab1-kube-api-access-4wcdz\") pod \"nova-metadata-0\" (UID: \"3eaa9674-9cd0-497f-80ca-27067fa54ab1\") " pod="openstack/nova-metadata-0" Dec 13 07:14:46 crc kubenswrapper[4971]: I1213 07:14:46.014792 4971 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/3eaa9674-9cd0-497f-80ca-27067fa54ab1-config-data\") pod \"nova-metadata-0\" (UID: \"3eaa9674-9cd0-497f-80ca-27067fa54ab1\") " pod="openstack/nova-metadata-0" Dec 13 07:14:46 crc kubenswrapper[4971]: I1213 07:14:46.014807 4971 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"nova-metadata-tls-certs\" (UniqueName: \"kubernetes.io/secret/3eaa9674-9cd0-497f-80ca-27067fa54ab1-nova-metadata-tls-certs\") pod \"nova-metadata-0\" (UID: \"3eaa9674-9cd0-497f-80ca-27067fa54ab1\") " pod="openstack/nova-metadata-0" Dec 13 07:14:46 crc kubenswrapper[4971]: I1213 07:14:46.014939 4971 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/3eaa9674-9cd0-497f-80ca-27067fa54ab1-logs\") pod \"nova-metadata-0\" (UID: \"3eaa9674-9cd0-497f-80ca-27067fa54ab1\") " pod="openstack/nova-metadata-0" Dec 13 07:14:46 crc kubenswrapper[4971]: I1213 07:14:46.117450 4971 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/3eaa9674-9cd0-497f-80ca-27067fa54ab1-combined-ca-bundle\") pod \"nova-metadata-0\" (UID: \"3eaa9674-9cd0-497f-80ca-27067fa54ab1\") " pod="openstack/nova-metadata-0" Dec 13 07:14:46 crc kubenswrapper[4971]: I1213 07:14:46.117510 4971 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-4wcdz\" (UniqueName: \"kubernetes.io/projected/3eaa9674-9cd0-497f-80ca-27067fa54ab1-kube-api-access-4wcdz\") pod \"nova-metadata-0\" (UID: \"3eaa9674-9cd0-497f-80ca-27067fa54ab1\") " pod="openstack/nova-metadata-0" Dec 13 07:14:46 crc kubenswrapper[4971]: I1213 07:14:46.117552 4971 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/3eaa9674-9cd0-497f-80ca-27067fa54ab1-config-data\") pod \"nova-metadata-0\" (UID: \"3eaa9674-9cd0-497f-80ca-27067fa54ab1\") " pod="openstack/nova-metadata-0" Dec 13 07:14:46 crc kubenswrapper[4971]: I1213 07:14:46.117565 4971 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nova-metadata-tls-certs\" (UniqueName: \"kubernetes.io/secret/3eaa9674-9cd0-497f-80ca-27067fa54ab1-nova-metadata-tls-certs\") pod \"nova-metadata-0\" (UID: \"3eaa9674-9cd0-497f-80ca-27067fa54ab1\") " pod="openstack/nova-metadata-0" Dec 13 07:14:46 crc kubenswrapper[4971]: I1213 07:14:46.117690 4971 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/3eaa9674-9cd0-497f-80ca-27067fa54ab1-logs\") pod \"nova-metadata-0\" (UID: \"3eaa9674-9cd0-497f-80ca-27067fa54ab1\") " pod="openstack/nova-metadata-0" Dec 13 07:14:46 crc kubenswrapper[4971]: I1213 07:14:46.118368 4971 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/3eaa9674-9cd0-497f-80ca-27067fa54ab1-logs\") pod \"nova-metadata-0\" (UID: \"3eaa9674-9cd0-497f-80ca-27067fa54ab1\") " pod="openstack/nova-metadata-0" Dec 13 07:14:46 crc kubenswrapper[4971]: I1213 07:14:46.122369 4971 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"nova-metadata-tls-certs\" (UniqueName: \"kubernetes.io/secret/3eaa9674-9cd0-497f-80ca-27067fa54ab1-nova-metadata-tls-certs\") pod \"nova-metadata-0\" (UID: \"3eaa9674-9cd0-497f-80ca-27067fa54ab1\") " pod="openstack/nova-metadata-0" Dec 13 07:14:46 crc kubenswrapper[4971]: I1213 07:14:46.122853 4971 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/3eaa9674-9cd0-497f-80ca-27067fa54ab1-config-data\") pod \"nova-metadata-0\" (UID: \"3eaa9674-9cd0-497f-80ca-27067fa54ab1\") " pod="openstack/nova-metadata-0" Dec 13 07:14:46 crc kubenswrapper[4971]: I1213 07:14:46.138349 4971 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/3eaa9674-9cd0-497f-80ca-27067fa54ab1-combined-ca-bundle\") pod \"nova-metadata-0\" (UID: \"3eaa9674-9cd0-497f-80ca-27067fa54ab1\") " pod="openstack/nova-metadata-0" Dec 13 07:14:46 crc kubenswrapper[4971]: I1213 07:14:46.144885 4971 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-4wcdz\" (UniqueName: \"kubernetes.io/projected/3eaa9674-9cd0-497f-80ca-27067fa54ab1-kube-api-access-4wcdz\") pod \"nova-metadata-0\" (UID: \"3eaa9674-9cd0-497f-80ca-27067fa54ab1\") " pod="openstack/nova-metadata-0" Dec 13 07:14:46 crc kubenswrapper[4971]: I1213 07:14:46.154144 4971 patch_prober.go:28] interesting pod/machine-config-daemon-82xjz container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 13 07:14:46 crc kubenswrapper[4971]: I1213 07:14:46.154242 4971 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-82xjz" podUID="e0c4af12-edda-41d7-9b44-f87396174ef2" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 13 07:14:46 crc kubenswrapper[4971]: I1213 07:14:46.154293 4971 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-82xjz" Dec 13 07:14:46 crc kubenswrapper[4971]: I1213 07:14:46.155165 4971 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"e8beee43d15719dccb6892d6f5711674492425300efa098be89033bc65977d14"} pod="openshift-machine-config-operator/machine-config-daemon-82xjz" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Dec 13 07:14:46 crc kubenswrapper[4971]: I1213 07:14:46.155234 4971 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-82xjz" podUID="e0c4af12-edda-41d7-9b44-f87396174ef2" containerName="machine-config-daemon" containerID="cri-o://e8beee43d15719dccb6892d6f5711674492425300efa098be89033bc65977d14" gracePeriod=600 Dec 13 07:14:46 crc kubenswrapper[4971]: I1213 07:14:46.241981 4971 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-metadata-0" Dec 13 07:14:46 crc kubenswrapper[4971]: E1213 07:14:46.295016 4971 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-82xjz_openshift-machine-config-operator(e0c4af12-edda-41d7-9b44-f87396174ef2)\"" pod="openshift-machine-config-operator/machine-config-daemon-82xjz" podUID="e0c4af12-edda-41d7-9b44-f87396174ef2" Dec 13 07:14:46 crc kubenswrapper[4971]: W1213 07:14:46.739105 4971 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod3eaa9674_9cd0_497f_80ca_27067fa54ab1.slice/crio-20f7d35f0207ae552234a98ea7f7d0e168645a8a743b0e2760b4572155d45ce9 WatchSource:0}: Error finding container 20f7d35f0207ae552234a98ea7f7d0e168645a8a743b0e2760b4572155d45ce9: Status 404 returned error can't find the container with id 20f7d35f0207ae552234a98ea7f7d0e168645a8a743b0e2760b4572155d45ce9 Dec 13 07:14:46 crc kubenswrapper[4971]: I1213 07:14:46.739471 4971 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-metadata-0"] Dec 13 07:14:46 crc kubenswrapper[4971]: I1213 07:14:46.784441 4971 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"3eaa9674-9cd0-497f-80ca-27067fa54ab1","Type":"ContainerStarted","Data":"20f7d35f0207ae552234a98ea7f7d0e168645a8a743b0e2760b4572155d45ce9"} Dec 13 07:14:46 crc kubenswrapper[4971]: I1213 07:14:46.794735 4971 generic.go:334] "Generic (PLEG): container finished" podID="e0c4af12-edda-41d7-9b44-f87396174ef2" containerID="e8beee43d15719dccb6892d6f5711674492425300efa098be89033bc65977d14" exitCode=0 Dec 13 07:14:46 crc kubenswrapper[4971]: I1213 07:14:46.794797 4971 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-82xjz" event={"ID":"e0c4af12-edda-41d7-9b44-f87396174ef2","Type":"ContainerDied","Data":"e8beee43d15719dccb6892d6f5711674492425300efa098be89033bc65977d14"} Dec 13 07:14:46 crc kubenswrapper[4971]: I1213 07:14:46.794852 4971 scope.go:117] "RemoveContainer" containerID="6543005a61441f4a045d5ad167cc1798a7b51524223dc500db29c8f9518b38f1" Dec 13 07:14:46 crc kubenswrapper[4971]: I1213 07:14:46.795402 4971 scope.go:117] "RemoveContainer" containerID="e8beee43d15719dccb6892d6f5711674492425300efa098be89033bc65977d14" Dec 13 07:14:46 crc kubenswrapper[4971]: E1213 07:14:46.795674 4971 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-82xjz_openshift-machine-config-operator(e0c4af12-edda-41d7-9b44-f87396174ef2)\"" pod="openshift-machine-config-operator/machine-config-daemon-82xjz" podUID="e0c4af12-edda-41d7-9b44-f87396174ef2" Dec 13 07:14:47 crc kubenswrapper[4971]: I1213 07:14:47.779426 4971 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="feaf3daf-1c1f-493f-8358-8e9afb515229" path="/var/lib/kubelet/pods/feaf3daf-1c1f-493f-8358-8e9afb515229/volumes" Dec 13 07:14:47 crc kubenswrapper[4971]: I1213 07:14:47.806200 4971 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"3eaa9674-9cd0-497f-80ca-27067fa54ab1","Type":"ContainerStarted","Data":"7e17af08c9ceb2f3f5caa6eca044f666d09c9392bf5b4ccddaa534ea7d9b47e3"} Dec 13 07:14:47 crc kubenswrapper[4971]: I1213 07:14:47.806789 4971 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"3eaa9674-9cd0-497f-80ca-27067fa54ab1","Type":"ContainerStarted","Data":"8d56124da0d9c36bc614406a67461bde4aa04e5b647a3ccf10450e4a6d15413d"} Dec 13 07:14:47 crc kubenswrapper[4971]: I1213 07:14:47.834603 4971 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-metadata-0" podStartSLOduration=2.834582196 podStartE2EDuration="2.834582196s" podCreationTimestamp="2025-12-13 07:14:45 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-13 07:14:47.833948711 +0000 UTC m=+1544.438358169" watchObservedRunningTime="2025-12-13 07:14:47.834582196 +0000 UTC m=+1544.438991664" Dec 13 07:14:48 crc kubenswrapper[4971]: I1213 07:14:48.817211 4971 generic.go:334] "Generic (PLEG): container finished" podID="74fcba46-f364-4c8a-a4e3-ace14451d1f7" containerID="76f58f854276b22f5878baf9dc07e632f05e6335bb6779edc28b4315b134c334" exitCode=0 Dec 13 07:14:48 crc kubenswrapper[4971]: I1213 07:14:48.818397 4971 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-cell-mapping-rgjbg" event={"ID":"74fcba46-f364-4c8a-a4e3-ace14451d1f7","Type":"ContainerDied","Data":"76f58f854276b22f5878baf9dc07e632f05e6335bb6779edc28b4315b134c334"} Dec 13 07:14:49 crc kubenswrapper[4971]: I1213 07:14:49.521121 4971 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-scheduler-0" Dec 13 07:14:49 crc kubenswrapper[4971]: I1213 07:14:49.521181 4971 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/nova-scheduler-0" Dec 13 07:14:49 crc kubenswrapper[4971]: I1213 07:14:49.540798 4971 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/dnsmasq-dns-59fd54bbff-hnxpb" Dec 13 07:14:49 crc kubenswrapper[4971]: I1213 07:14:49.550332 4971 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/nova-scheduler-0" Dec 13 07:14:49 crc kubenswrapper[4971]: I1213 07:14:49.579562 4971 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/nova-api-0" Dec 13 07:14:49 crc kubenswrapper[4971]: I1213 07:14:49.579979 4971 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/nova-api-0" Dec 13 07:14:49 crc kubenswrapper[4971]: I1213 07:14:49.619255 4971 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-7474d577dc-q7d6q"] Dec 13 07:14:49 crc kubenswrapper[4971]: I1213 07:14:49.619494 4971 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/dnsmasq-dns-7474d577dc-q7d6q" podUID="4c00defc-fd2b-43e6-ba62-ccc9fd2267fc" containerName="dnsmasq-dns" containerID="cri-o://9647f002123326031aafc6c3c41875e56907e06cbf016e592145c84346eae3de" gracePeriod=10 Dec 13 07:14:49 crc kubenswrapper[4971]: I1213 07:14:49.828998 4971 generic.go:334] "Generic (PLEG): container finished" podID="4c00defc-fd2b-43e6-ba62-ccc9fd2267fc" containerID="9647f002123326031aafc6c3c41875e56907e06cbf016e592145c84346eae3de" exitCode=0 Dec 13 07:14:49 crc kubenswrapper[4971]: I1213 07:14:49.829123 4971 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-7474d577dc-q7d6q" event={"ID":"4c00defc-fd2b-43e6-ba62-ccc9fd2267fc","Type":"ContainerDied","Data":"9647f002123326031aafc6c3c41875e56907e06cbf016e592145c84346eae3de"} Dec 13 07:14:49 crc kubenswrapper[4971]: I1213 07:14:49.903487 4971 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/nova-scheduler-0" Dec 13 07:14:50 crc kubenswrapper[4971]: I1213 07:14:50.216589 4971 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-7474d577dc-q7d6q" Dec 13 07:14:50 crc kubenswrapper[4971]: I1213 07:14:50.305324 4971 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/4c00defc-fd2b-43e6-ba62-ccc9fd2267fc-ovsdbserver-sb\") pod \"4c00defc-fd2b-43e6-ba62-ccc9fd2267fc\" (UID: \"4c00defc-fd2b-43e6-ba62-ccc9fd2267fc\") " Dec 13 07:14:50 crc kubenswrapper[4971]: I1213 07:14:50.305382 4971 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/4c00defc-fd2b-43e6-ba62-ccc9fd2267fc-ovsdbserver-nb\") pod \"4c00defc-fd2b-43e6-ba62-ccc9fd2267fc\" (UID: \"4c00defc-fd2b-43e6-ba62-ccc9fd2267fc\") " Dec 13 07:14:50 crc kubenswrapper[4971]: I1213 07:14:50.305454 4971 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/4c00defc-fd2b-43e6-ba62-ccc9fd2267fc-dns-svc\") pod \"4c00defc-fd2b-43e6-ba62-ccc9fd2267fc\" (UID: \"4c00defc-fd2b-43e6-ba62-ccc9fd2267fc\") " Dec 13 07:14:50 crc kubenswrapper[4971]: I1213 07:14:50.305592 4971 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-h96b2\" (UniqueName: \"kubernetes.io/projected/4c00defc-fd2b-43e6-ba62-ccc9fd2267fc-kube-api-access-h96b2\") pod \"4c00defc-fd2b-43e6-ba62-ccc9fd2267fc\" (UID: \"4c00defc-fd2b-43e6-ba62-ccc9fd2267fc\") " Dec 13 07:14:50 crc kubenswrapper[4971]: I1213 07:14:50.305619 4971 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/4c00defc-fd2b-43e6-ba62-ccc9fd2267fc-config\") pod \"4c00defc-fd2b-43e6-ba62-ccc9fd2267fc\" (UID: \"4c00defc-fd2b-43e6-ba62-ccc9fd2267fc\") " Dec 13 07:14:50 crc kubenswrapper[4971]: I1213 07:14:50.311795 4971 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/4c00defc-fd2b-43e6-ba62-ccc9fd2267fc-kube-api-access-h96b2" (OuterVolumeSpecName: "kube-api-access-h96b2") pod "4c00defc-fd2b-43e6-ba62-ccc9fd2267fc" (UID: "4c00defc-fd2b-43e6-ba62-ccc9fd2267fc"). InnerVolumeSpecName "kube-api-access-h96b2". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 13 07:14:50 crc kubenswrapper[4971]: I1213 07:14:50.346585 4971 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-cell-mapping-rgjbg" Dec 13 07:14:50 crc kubenswrapper[4971]: I1213 07:14:50.369259 4971 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/4c00defc-fd2b-43e6-ba62-ccc9fd2267fc-config" (OuterVolumeSpecName: "config") pod "4c00defc-fd2b-43e6-ba62-ccc9fd2267fc" (UID: "4c00defc-fd2b-43e6-ba62-ccc9fd2267fc"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 13 07:14:50 crc kubenswrapper[4971]: I1213 07:14:50.375461 4971 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/4c00defc-fd2b-43e6-ba62-ccc9fd2267fc-dns-svc" (OuterVolumeSpecName: "dns-svc") pod "4c00defc-fd2b-43e6-ba62-ccc9fd2267fc" (UID: "4c00defc-fd2b-43e6-ba62-ccc9fd2267fc"). InnerVolumeSpecName "dns-svc". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 13 07:14:50 crc kubenswrapper[4971]: I1213 07:14:50.399681 4971 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/4c00defc-fd2b-43e6-ba62-ccc9fd2267fc-ovsdbserver-sb" (OuterVolumeSpecName: "ovsdbserver-sb") pod "4c00defc-fd2b-43e6-ba62-ccc9fd2267fc" (UID: "4c00defc-fd2b-43e6-ba62-ccc9fd2267fc"). InnerVolumeSpecName "ovsdbserver-sb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 13 07:14:50 crc kubenswrapper[4971]: I1213 07:14:50.407797 4971 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-6htvt\" (UniqueName: \"kubernetes.io/projected/74fcba46-f364-4c8a-a4e3-ace14451d1f7-kube-api-access-6htvt\") pod \"74fcba46-f364-4c8a-a4e3-ace14451d1f7\" (UID: \"74fcba46-f364-4c8a-a4e3-ace14451d1f7\") " Dec 13 07:14:50 crc kubenswrapper[4971]: I1213 07:14:50.407872 4971 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/74fcba46-f364-4c8a-a4e3-ace14451d1f7-combined-ca-bundle\") pod \"74fcba46-f364-4c8a-a4e3-ace14451d1f7\" (UID: \"74fcba46-f364-4c8a-a4e3-ace14451d1f7\") " Dec 13 07:14:50 crc kubenswrapper[4971]: I1213 07:14:50.407960 4971 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/74fcba46-f364-4c8a-a4e3-ace14451d1f7-config-data\") pod \"74fcba46-f364-4c8a-a4e3-ace14451d1f7\" (UID: \"74fcba46-f364-4c8a-a4e3-ace14451d1f7\") " Dec 13 07:14:50 crc kubenswrapper[4971]: I1213 07:14:50.408088 4971 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/74fcba46-f364-4c8a-a4e3-ace14451d1f7-scripts\") pod \"74fcba46-f364-4c8a-a4e3-ace14451d1f7\" (UID: \"74fcba46-f364-4c8a-a4e3-ace14451d1f7\") " Dec 13 07:14:50 crc kubenswrapper[4971]: I1213 07:14:50.408431 4971 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-h96b2\" (UniqueName: \"kubernetes.io/projected/4c00defc-fd2b-43e6-ba62-ccc9fd2267fc-kube-api-access-h96b2\") on node \"crc\" DevicePath \"\"" Dec 13 07:14:50 crc kubenswrapper[4971]: I1213 07:14:50.408450 4971 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/4c00defc-fd2b-43e6-ba62-ccc9fd2267fc-config\") on node \"crc\" DevicePath \"\"" Dec 13 07:14:50 crc kubenswrapper[4971]: I1213 07:14:50.408460 4971 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/4c00defc-fd2b-43e6-ba62-ccc9fd2267fc-ovsdbserver-sb\") on node \"crc\" DevicePath \"\"" Dec 13 07:14:50 crc kubenswrapper[4971]: I1213 07:14:50.408470 4971 reconciler_common.go:293] "Volume detached for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/4c00defc-fd2b-43e6-ba62-ccc9fd2267fc-dns-svc\") on node \"crc\" DevicePath \"\"" Dec 13 07:14:50 crc kubenswrapper[4971]: I1213 07:14:50.412686 4971 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/74fcba46-f364-4c8a-a4e3-ace14451d1f7-scripts" (OuterVolumeSpecName: "scripts") pod "74fcba46-f364-4c8a-a4e3-ace14451d1f7" (UID: "74fcba46-f364-4c8a-a4e3-ace14451d1f7"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 13 07:14:50 crc kubenswrapper[4971]: I1213 07:14:50.413119 4971 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/74fcba46-f364-4c8a-a4e3-ace14451d1f7-kube-api-access-6htvt" (OuterVolumeSpecName: "kube-api-access-6htvt") pod "74fcba46-f364-4c8a-a4e3-ace14451d1f7" (UID: "74fcba46-f364-4c8a-a4e3-ace14451d1f7"). InnerVolumeSpecName "kube-api-access-6htvt". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 13 07:14:50 crc kubenswrapper[4971]: I1213 07:14:50.423148 4971 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/4c00defc-fd2b-43e6-ba62-ccc9fd2267fc-ovsdbserver-nb" (OuterVolumeSpecName: "ovsdbserver-nb") pod "4c00defc-fd2b-43e6-ba62-ccc9fd2267fc" (UID: "4c00defc-fd2b-43e6-ba62-ccc9fd2267fc"). InnerVolumeSpecName "ovsdbserver-nb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 13 07:14:50 crc kubenswrapper[4971]: I1213 07:14:50.438918 4971 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/74fcba46-f364-4c8a-a4e3-ace14451d1f7-config-data" (OuterVolumeSpecName: "config-data") pod "74fcba46-f364-4c8a-a4e3-ace14451d1f7" (UID: "74fcba46-f364-4c8a-a4e3-ace14451d1f7"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 13 07:14:50 crc kubenswrapper[4971]: I1213 07:14:50.460498 4971 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/74fcba46-f364-4c8a-a4e3-ace14451d1f7-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "74fcba46-f364-4c8a-a4e3-ace14451d1f7" (UID: "74fcba46-f364-4c8a-a4e3-ace14451d1f7"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 13 07:14:50 crc kubenswrapper[4971]: I1213 07:14:50.509556 4971 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/74fcba46-f364-4c8a-a4e3-ace14451d1f7-scripts\") on node \"crc\" DevicePath \"\"" Dec 13 07:14:50 crc kubenswrapper[4971]: I1213 07:14:50.509592 4971 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-6htvt\" (UniqueName: \"kubernetes.io/projected/74fcba46-f364-4c8a-a4e3-ace14451d1f7-kube-api-access-6htvt\") on node \"crc\" DevicePath \"\"" Dec 13 07:14:50 crc kubenswrapper[4971]: I1213 07:14:50.509602 4971 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/74fcba46-f364-4c8a-a4e3-ace14451d1f7-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 13 07:14:50 crc kubenswrapper[4971]: I1213 07:14:50.509611 4971 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/4c00defc-fd2b-43e6-ba62-ccc9fd2267fc-ovsdbserver-nb\") on node \"crc\" DevicePath \"\"" Dec 13 07:14:50 crc kubenswrapper[4971]: I1213 07:14:50.509621 4971 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/74fcba46-f364-4c8a-a4e3-ace14451d1f7-config-data\") on node \"crc\" DevicePath \"\"" Dec 13 07:14:50 crc kubenswrapper[4971]: I1213 07:14:50.663682 4971 prober.go:107] "Probe failed" probeType="Startup" pod="openstack/nova-api-0" podUID="b99cc168-f32c-49c0-b150-e4e730e9eacb" containerName="nova-api-api" probeResult="failure" output="Get \"http://10.217.0.178:8774/\": context deadline exceeded (Client.Timeout exceeded while awaiting headers)" Dec 13 07:14:50 crc kubenswrapper[4971]: I1213 07:14:50.663726 4971 prober.go:107] "Probe failed" probeType="Startup" pod="openstack/nova-api-0" podUID="b99cc168-f32c-49c0-b150-e4e730e9eacb" containerName="nova-api-log" probeResult="failure" output="Get \"http://10.217.0.178:8774/\": context deadline exceeded (Client.Timeout exceeded while awaiting headers)" Dec 13 07:14:50 crc kubenswrapper[4971]: I1213 07:14:50.840104 4971 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-7474d577dc-q7d6q" event={"ID":"4c00defc-fd2b-43e6-ba62-ccc9fd2267fc","Type":"ContainerDied","Data":"5461c8ff70588f04523e12e1d7c4bd641e9422a3c5001d3ae5943e8399800888"} Dec 13 07:14:50 crc kubenswrapper[4971]: I1213 07:14:50.840176 4971 scope.go:117] "RemoveContainer" containerID="9647f002123326031aafc6c3c41875e56907e06cbf016e592145c84346eae3de" Dec 13 07:14:50 crc kubenswrapper[4971]: I1213 07:14:50.840435 4971 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-7474d577dc-q7d6q" Dec 13 07:14:50 crc kubenswrapper[4971]: I1213 07:14:50.841976 4971 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-cell-mapping-rgjbg" event={"ID":"74fcba46-f364-4c8a-a4e3-ace14451d1f7","Type":"ContainerDied","Data":"e2f608fc6357e7c23334d2f79d959e45ce3e234e164b5586305683b5663bb302"} Dec 13 07:14:50 crc kubenswrapper[4971]: I1213 07:14:50.842016 4971 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="e2f608fc6357e7c23334d2f79d959e45ce3e234e164b5586305683b5663bb302" Dec 13 07:14:50 crc kubenswrapper[4971]: I1213 07:14:50.842184 4971 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-cell-mapping-rgjbg" Dec 13 07:14:50 crc kubenswrapper[4971]: I1213 07:14:50.868418 4971 scope.go:117] "RemoveContainer" containerID="5763ec7db6e3b9278c34e8c29379aa4998a4685bb3b627ac91904b0d0f9e6a67" Dec 13 07:14:50 crc kubenswrapper[4971]: I1213 07:14:50.888916 4971 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-7474d577dc-q7d6q"] Dec 13 07:14:50 crc kubenswrapper[4971]: I1213 07:14:50.898710 4971 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-7474d577dc-q7d6q"] Dec 13 07:14:51 crc kubenswrapper[4971]: I1213 07:14:51.035918 4971 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-api-0"] Dec 13 07:14:51 crc kubenswrapper[4971]: I1213 07:14:51.036200 4971 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-api-0" podUID="b99cc168-f32c-49c0-b150-e4e730e9eacb" containerName="nova-api-log" containerID="cri-o://28b75774392e9b16f160e9594f4ecede802f1fec9d3a898016414cd66da31587" gracePeriod=30 Dec 13 07:14:51 crc kubenswrapper[4971]: I1213 07:14:51.036281 4971 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-api-0" podUID="b99cc168-f32c-49c0-b150-e4e730e9eacb" containerName="nova-api-api" containerID="cri-o://de4bd8652eb387c905466a27a8920b3fcec43795ba4ca6babe7f7fc4a4f3f1bd" gracePeriod=30 Dec 13 07:14:51 crc kubenswrapper[4971]: I1213 07:14:51.080182 4971 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-scheduler-0"] Dec 13 07:14:51 crc kubenswrapper[4971]: I1213 07:14:51.093164 4971 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-metadata-0"] Dec 13 07:14:51 crc kubenswrapper[4971]: I1213 07:14:51.093454 4971 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-metadata-0" podUID="3eaa9674-9cd0-497f-80ca-27067fa54ab1" containerName="nova-metadata-log" containerID="cri-o://8d56124da0d9c36bc614406a67461bde4aa04e5b647a3ccf10450e4a6d15413d" gracePeriod=30 Dec 13 07:14:51 crc kubenswrapper[4971]: I1213 07:14:51.093771 4971 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-metadata-0" podUID="3eaa9674-9cd0-497f-80ca-27067fa54ab1" containerName="nova-metadata-metadata" containerID="cri-o://7e17af08c9ceb2f3f5caa6eca044f666d09c9392bf5b4ccddaa534ea7d9b47e3" gracePeriod=30 Dec 13 07:14:51 crc kubenswrapper[4971]: I1213 07:14:51.242997 4971 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-metadata-0" Dec 13 07:14:51 crc kubenswrapper[4971]: I1213 07:14:51.243043 4971 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-metadata-0" Dec 13 07:14:51 crc kubenswrapper[4971]: I1213 07:14:51.654368 4971 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-metadata-0" Dec 13 07:14:51 crc kubenswrapper[4971]: I1213 07:14:51.746079 4971 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/3eaa9674-9cd0-497f-80ca-27067fa54ab1-combined-ca-bundle\") pod \"3eaa9674-9cd0-497f-80ca-27067fa54ab1\" (UID: \"3eaa9674-9cd0-497f-80ca-27067fa54ab1\") " Dec 13 07:14:51 crc kubenswrapper[4971]: I1213 07:14:51.746232 4971 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-4wcdz\" (UniqueName: \"kubernetes.io/projected/3eaa9674-9cd0-497f-80ca-27067fa54ab1-kube-api-access-4wcdz\") pod \"3eaa9674-9cd0-497f-80ca-27067fa54ab1\" (UID: \"3eaa9674-9cd0-497f-80ca-27067fa54ab1\") " Dec 13 07:14:51 crc kubenswrapper[4971]: I1213 07:14:51.746277 4971 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/3eaa9674-9cd0-497f-80ca-27067fa54ab1-logs\") pod \"3eaa9674-9cd0-497f-80ca-27067fa54ab1\" (UID: \"3eaa9674-9cd0-497f-80ca-27067fa54ab1\") " Dec 13 07:14:51 crc kubenswrapper[4971]: I1213 07:14:51.746361 4971 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/3eaa9674-9cd0-497f-80ca-27067fa54ab1-config-data\") pod \"3eaa9674-9cd0-497f-80ca-27067fa54ab1\" (UID: \"3eaa9674-9cd0-497f-80ca-27067fa54ab1\") " Dec 13 07:14:51 crc kubenswrapper[4971]: I1213 07:14:51.746441 4971 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"nova-metadata-tls-certs\" (UniqueName: \"kubernetes.io/secret/3eaa9674-9cd0-497f-80ca-27067fa54ab1-nova-metadata-tls-certs\") pod \"3eaa9674-9cd0-497f-80ca-27067fa54ab1\" (UID: \"3eaa9674-9cd0-497f-80ca-27067fa54ab1\") " Dec 13 07:14:51 crc kubenswrapper[4971]: I1213 07:14:51.746974 4971 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/3eaa9674-9cd0-497f-80ca-27067fa54ab1-logs" (OuterVolumeSpecName: "logs") pod "3eaa9674-9cd0-497f-80ca-27067fa54ab1" (UID: "3eaa9674-9cd0-497f-80ca-27067fa54ab1"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 13 07:14:51 crc kubenswrapper[4971]: I1213 07:14:51.747350 4971 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/3eaa9674-9cd0-497f-80ca-27067fa54ab1-logs\") on node \"crc\" DevicePath \"\"" Dec 13 07:14:51 crc kubenswrapper[4971]: I1213 07:14:51.764766 4971 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/3eaa9674-9cd0-497f-80ca-27067fa54ab1-kube-api-access-4wcdz" (OuterVolumeSpecName: "kube-api-access-4wcdz") pod "3eaa9674-9cd0-497f-80ca-27067fa54ab1" (UID: "3eaa9674-9cd0-497f-80ca-27067fa54ab1"). InnerVolumeSpecName "kube-api-access-4wcdz". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 13 07:14:51 crc kubenswrapper[4971]: I1213 07:14:51.800469 4971 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/3eaa9674-9cd0-497f-80ca-27067fa54ab1-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "3eaa9674-9cd0-497f-80ca-27067fa54ab1" (UID: "3eaa9674-9cd0-497f-80ca-27067fa54ab1"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 13 07:14:51 crc kubenswrapper[4971]: I1213 07:14:51.801210 4971 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="4c00defc-fd2b-43e6-ba62-ccc9fd2267fc" path="/var/lib/kubelet/pods/4c00defc-fd2b-43e6-ba62-ccc9fd2267fc/volumes" Dec 13 07:14:51 crc kubenswrapper[4971]: I1213 07:14:51.848446 4971 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-4wcdz\" (UniqueName: \"kubernetes.io/projected/3eaa9674-9cd0-497f-80ca-27067fa54ab1-kube-api-access-4wcdz\") on node \"crc\" DevicePath \"\"" Dec 13 07:14:51 crc kubenswrapper[4971]: I1213 07:14:51.848475 4971 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/3eaa9674-9cd0-497f-80ca-27067fa54ab1-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 13 07:14:51 crc kubenswrapper[4971]: I1213 07:14:51.849640 4971 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/3eaa9674-9cd0-497f-80ca-27067fa54ab1-config-data" (OuterVolumeSpecName: "config-data") pod "3eaa9674-9cd0-497f-80ca-27067fa54ab1" (UID: "3eaa9674-9cd0-497f-80ca-27067fa54ab1"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 13 07:14:51 crc kubenswrapper[4971]: I1213 07:14:51.899643 4971 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/3eaa9674-9cd0-497f-80ca-27067fa54ab1-nova-metadata-tls-certs" (OuterVolumeSpecName: "nova-metadata-tls-certs") pod "3eaa9674-9cd0-497f-80ca-27067fa54ab1" (UID: "3eaa9674-9cd0-497f-80ca-27067fa54ab1"). InnerVolumeSpecName "nova-metadata-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 13 07:14:51 crc kubenswrapper[4971]: I1213 07:14:51.899860 4971 generic.go:334] "Generic (PLEG): container finished" podID="3eaa9674-9cd0-497f-80ca-27067fa54ab1" containerID="7e17af08c9ceb2f3f5caa6eca044f666d09c9392bf5b4ccddaa534ea7d9b47e3" exitCode=0 Dec 13 07:14:51 crc kubenswrapper[4971]: I1213 07:14:51.899882 4971 generic.go:334] "Generic (PLEG): container finished" podID="3eaa9674-9cd0-497f-80ca-27067fa54ab1" containerID="8d56124da0d9c36bc614406a67461bde4aa04e5b647a3ccf10450e4a6d15413d" exitCode=143 Dec 13 07:14:51 crc kubenswrapper[4971]: I1213 07:14:51.899978 4971 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-metadata-0" Dec 13 07:14:51 crc kubenswrapper[4971]: I1213 07:14:51.911003 4971 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"3eaa9674-9cd0-497f-80ca-27067fa54ab1","Type":"ContainerDied","Data":"7e17af08c9ceb2f3f5caa6eca044f666d09c9392bf5b4ccddaa534ea7d9b47e3"} Dec 13 07:14:51 crc kubenswrapper[4971]: I1213 07:14:51.911267 4971 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"3eaa9674-9cd0-497f-80ca-27067fa54ab1","Type":"ContainerDied","Data":"8d56124da0d9c36bc614406a67461bde4aa04e5b647a3ccf10450e4a6d15413d"} Dec 13 07:14:51 crc kubenswrapper[4971]: I1213 07:14:51.911358 4971 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"3eaa9674-9cd0-497f-80ca-27067fa54ab1","Type":"ContainerDied","Data":"20f7d35f0207ae552234a98ea7f7d0e168645a8a743b0e2760b4572155d45ce9"} Dec 13 07:14:51 crc kubenswrapper[4971]: I1213 07:14:51.911486 4971 scope.go:117] "RemoveContainer" containerID="7e17af08c9ceb2f3f5caa6eca044f666d09c9392bf5b4ccddaa534ea7d9b47e3" Dec 13 07:14:51 crc kubenswrapper[4971]: I1213 07:14:51.921433 4971 generic.go:334] "Generic (PLEG): container finished" podID="b99cc168-f32c-49c0-b150-e4e730e9eacb" containerID="28b75774392e9b16f160e9594f4ecede802f1fec9d3a898016414cd66da31587" exitCode=143 Dec 13 07:14:51 crc kubenswrapper[4971]: I1213 07:14:51.921961 4971 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"b99cc168-f32c-49c0-b150-e4e730e9eacb","Type":"ContainerDied","Data":"28b75774392e9b16f160e9594f4ecede802f1fec9d3a898016414cd66da31587"} Dec 13 07:14:51 crc kubenswrapper[4971]: I1213 07:14:51.936976 4971 generic.go:334] "Generic (PLEG): container finished" podID="6c6829bb-25ee-4102-8335-b747bdbe6121" containerID="cc91778c66ca205f06be892439fa7bd0075527b3d7db3226d2d4b560aac675ab" exitCode=0 Dec 13 07:14:51 crc kubenswrapper[4971]: I1213 07:14:51.937046 4971 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-conductor-db-sync-kvw2t" event={"ID":"6c6829bb-25ee-4102-8335-b747bdbe6121","Type":"ContainerDied","Data":"cc91778c66ca205f06be892439fa7bd0075527b3d7db3226d2d4b560aac675ab"} Dec 13 07:14:51 crc kubenswrapper[4971]: I1213 07:14:51.938349 4971 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-scheduler-0" podUID="156254b1-933a-49b3-ae26-f0ba7f1418b8" containerName="nova-scheduler-scheduler" containerID="cri-o://97dd6db8c1ca363f1d2b43537619b59bc570a33e67bb41c9d1803485ae0de49e" gracePeriod=30 Dec 13 07:14:51 crc kubenswrapper[4971]: I1213 07:14:51.954101 4971 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/3eaa9674-9cd0-497f-80ca-27067fa54ab1-config-data\") on node \"crc\" DevicePath \"\"" Dec 13 07:14:51 crc kubenswrapper[4971]: I1213 07:14:51.954137 4971 reconciler_common.go:293] "Volume detached for volume \"nova-metadata-tls-certs\" (UniqueName: \"kubernetes.io/secret/3eaa9674-9cd0-497f-80ca-27067fa54ab1-nova-metadata-tls-certs\") on node \"crc\" DevicePath \"\"" Dec 13 07:14:51 crc kubenswrapper[4971]: I1213 07:14:51.973570 4971 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-metadata-0"] Dec 13 07:14:51 crc kubenswrapper[4971]: I1213 07:14:51.982506 4971 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-metadata-0"] Dec 13 07:14:51 crc kubenswrapper[4971]: I1213 07:14:51.983091 4971 scope.go:117] "RemoveContainer" containerID="8d56124da0d9c36bc614406a67461bde4aa04e5b647a3ccf10450e4a6d15413d" Dec 13 07:14:52 crc kubenswrapper[4971]: I1213 07:14:52.000576 4971 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-metadata-0"] Dec 13 07:14:52 crc kubenswrapper[4971]: E1213 07:14:52.000902 4971 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="3eaa9674-9cd0-497f-80ca-27067fa54ab1" containerName="nova-metadata-metadata" Dec 13 07:14:52 crc kubenswrapper[4971]: I1213 07:14:52.000919 4971 state_mem.go:107] "Deleted CPUSet assignment" podUID="3eaa9674-9cd0-497f-80ca-27067fa54ab1" containerName="nova-metadata-metadata" Dec 13 07:14:52 crc kubenswrapper[4971]: E1213 07:14:52.000928 4971 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="4c00defc-fd2b-43e6-ba62-ccc9fd2267fc" containerName="init" Dec 13 07:14:52 crc kubenswrapper[4971]: I1213 07:14:52.000935 4971 state_mem.go:107] "Deleted CPUSet assignment" podUID="4c00defc-fd2b-43e6-ba62-ccc9fd2267fc" containerName="init" Dec 13 07:14:52 crc kubenswrapper[4971]: E1213 07:14:52.000951 4971 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="4c00defc-fd2b-43e6-ba62-ccc9fd2267fc" containerName="dnsmasq-dns" Dec 13 07:14:52 crc kubenswrapper[4971]: I1213 07:14:52.000957 4971 state_mem.go:107] "Deleted CPUSet assignment" podUID="4c00defc-fd2b-43e6-ba62-ccc9fd2267fc" containerName="dnsmasq-dns" Dec 13 07:14:52 crc kubenswrapper[4971]: E1213 07:14:52.000968 4971 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="74fcba46-f364-4c8a-a4e3-ace14451d1f7" containerName="nova-manage" Dec 13 07:14:52 crc kubenswrapper[4971]: I1213 07:14:52.000973 4971 state_mem.go:107] "Deleted CPUSet assignment" podUID="74fcba46-f364-4c8a-a4e3-ace14451d1f7" containerName="nova-manage" Dec 13 07:14:52 crc kubenswrapper[4971]: E1213 07:14:52.000982 4971 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="3eaa9674-9cd0-497f-80ca-27067fa54ab1" containerName="nova-metadata-log" Dec 13 07:14:52 crc kubenswrapper[4971]: I1213 07:14:52.000987 4971 state_mem.go:107] "Deleted CPUSet assignment" podUID="3eaa9674-9cd0-497f-80ca-27067fa54ab1" containerName="nova-metadata-log" Dec 13 07:14:52 crc kubenswrapper[4971]: I1213 07:14:52.001137 4971 memory_manager.go:354] "RemoveStaleState removing state" podUID="3eaa9674-9cd0-497f-80ca-27067fa54ab1" containerName="nova-metadata-metadata" Dec 13 07:14:52 crc kubenswrapper[4971]: I1213 07:14:52.001152 4971 memory_manager.go:354] "RemoveStaleState removing state" podUID="4c00defc-fd2b-43e6-ba62-ccc9fd2267fc" containerName="dnsmasq-dns" Dec 13 07:14:52 crc kubenswrapper[4971]: I1213 07:14:52.001160 4971 memory_manager.go:354] "RemoveStaleState removing state" podUID="74fcba46-f364-4c8a-a4e3-ace14451d1f7" containerName="nova-manage" Dec 13 07:14:52 crc kubenswrapper[4971]: I1213 07:14:52.001177 4971 memory_manager.go:354] "RemoveStaleState removing state" podUID="3eaa9674-9cd0-497f-80ca-27067fa54ab1" containerName="nova-metadata-log" Dec 13 07:14:52 crc kubenswrapper[4971]: I1213 07:14:52.004397 4971 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-metadata-0" Dec 13 07:14:52 crc kubenswrapper[4971]: I1213 07:14:52.008048 4971 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-nova-metadata-internal-svc" Dec 13 07:14:52 crc kubenswrapper[4971]: I1213 07:14:52.008280 4971 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-metadata-config-data" Dec 13 07:14:52 crc kubenswrapper[4971]: I1213 07:14:52.019083 4971 scope.go:117] "RemoveContainer" containerID="7e17af08c9ceb2f3f5caa6eca044f666d09c9392bf5b4ccddaa534ea7d9b47e3" Dec 13 07:14:52 crc kubenswrapper[4971]: E1213 07:14:52.019636 4971 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"7e17af08c9ceb2f3f5caa6eca044f666d09c9392bf5b4ccddaa534ea7d9b47e3\": container with ID starting with 7e17af08c9ceb2f3f5caa6eca044f666d09c9392bf5b4ccddaa534ea7d9b47e3 not found: ID does not exist" containerID="7e17af08c9ceb2f3f5caa6eca044f666d09c9392bf5b4ccddaa534ea7d9b47e3" Dec 13 07:14:52 crc kubenswrapper[4971]: I1213 07:14:52.019689 4971 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"7e17af08c9ceb2f3f5caa6eca044f666d09c9392bf5b4ccddaa534ea7d9b47e3"} err="failed to get container status \"7e17af08c9ceb2f3f5caa6eca044f666d09c9392bf5b4ccddaa534ea7d9b47e3\": rpc error: code = NotFound desc = could not find container \"7e17af08c9ceb2f3f5caa6eca044f666d09c9392bf5b4ccddaa534ea7d9b47e3\": container with ID starting with 7e17af08c9ceb2f3f5caa6eca044f666d09c9392bf5b4ccddaa534ea7d9b47e3 not found: ID does not exist" Dec 13 07:14:52 crc kubenswrapper[4971]: I1213 07:14:52.019719 4971 scope.go:117] "RemoveContainer" containerID="8d56124da0d9c36bc614406a67461bde4aa04e5b647a3ccf10450e4a6d15413d" Dec 13 07:14:52 crc kubenswrapper[4971]: E1213 07:14:52.020687 4971 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"8d56124da0d9c36bc614406a67461bde4aa04e5b647a3ccf10450e4a6d15413d\": container with ID starting with 8d56124da0d9c36bc614406a67461bde4aa04e5b647a3ccf10450e4a6d15413d not found: ID does not exist" containerID="8d56124da0d9c36bc614406a67461bde4aa04e5b647a3ccf10450e4a6d15413d" Dec 13 07:14:52 crc kubenswrapper[4971]: I1213 07:14:52.020722 4971 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"8d56124da0d9c36bc614406a67461bde4aa04e5b647a3ccf10450e4a6d15413d"} err="failed to get container status \"8d56124da0d9c36bc614406a67461bde4aa04e5b647a3ccf10450e4a6d15413d\": rpc error: code = NotFound desc = could not find container \"8d56124da0d9c36bc614406a67461bde4aa04e5b647a3ccf10450e4a6d15413d\": container with ID starting with 8d56124da0d9c36bc614406a67461bde4aa04e5b647a3ccf10450e4a6d15413d not found: ID does not exist" Dec 13 07:14:52 crc kubenswrapper[4971]: I1213 07:14:52.020744 4971 scope.go:117] "RemoveContainer" containerID="7e17af08c9ceb2f3f5caa6eca044f666d09c9392bf5b4ccddaa534ea7d9b47e3" Dec 13 07:14:52 crc kubenswrapper[4971]: I1213 07:14:52.026774 4971 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"7e17af08c9ceb2f3f5caa6eca044f666d09c9392bf5b4ccddaa534ea7d9b47e3"} err="failed to get container status \"7e17af08c9ceb2f3f5caa6eca044f666d09c9392bf5b4ccddaa534ea7d9b47e3\": rpc error: code = NotFound desc = could not find container \"7e17af08c9ceb2f3f5caa6eca044f666d09c9392bf5b4ccddaa534ea7d9b47e3\": container with ID starting with 7e17af08c9ceb2f3f5caa6eca044f666d09c9392bf5b4ccddaa534ea7d9b47e3 not found: ID does not exist" Dec 13 07:14:52 crc kubenswrapper[4971]: I1213 07:14:52.026888 4971 scope.go:117] "RemoveContainer" containerID="8d56124da0d9c36bc614406a67461bde4aa04e5b647a3ccf10450e4a6d15413d" Dec 13 07:14:52 crc kubenswrapper[4971]: I1213 07:14:52.034663 4971 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"8d56124da0d9c36bc614406a67461bde4aa04e5b647a3ccf10450e4a6d15413d"} err="failed to get container status \"8d56124da0d9c36bc614406a67461bde4aa04e5b647a3ccf10450e4a6d15413d\": rpc error: code = NotFound desc = could not find container \"8d56124da0d9c36bc614406a67461bde4aa04e5b647a3ccf10450e4a6d15413d\": container with ID starting with 8d56124da0d9c36bc614406a67461bde4aa04e5b647a3ccf10450e4a6d15413d not found: ID does not exist" Dec 13 07:14:52 crc kubenswrapper[4971]: I1213 07:14:52.054441 4971 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-metadata-0"] Dec 13 07:14:52 crc kubenswrapper[4971]: I1213 07:14:52.157140 4971 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/ddbf55d9-778f-4601-bb0d-05a3d394de49-config-data\") pod \"nova-metadata-0\" (UID: \"ddbf55d9-778f-4601-bb0d-05a3d394de49\") " pod="openstack/nova-metadata-0" Dec 13 07:14:52 crc kubenswrapper[4971]: I1213 07:14:52.157381 4971 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/ddbf55d9-778f-4601-bb0d-05a3d394de49-logs\") pod \"nova-metadata-0\" (UID: \"ddbf55d9-778f-4601-bb0d-05a3d394de49\") " pod="openstack/nova-metadata-0" Dec 13 07:14:52 crc kubenswrapper[4971]: I1213 07:14:52.157433 4971 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"nova-metadata-tls-certs\" (UniqueName: \"kubernetes.io/secret/ddbf55d9-778f-4601-bb0d-05a3d394de49-nova-metadata-tls-certs\") pod \"nova-metadata-0\" (UID: \"ddbf55d9-778f-4601-bb0d-05a3d394de49\") " pod="openstack/nova-metadata-0" Dec 13 07:14:52 crc kubenswrapper[4971]: I1213 07:14:52.157473 4971 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/ddbf55d9-778f-4601-bb0d-05a3d394de49-combined-ca-bundle\") pod \"nova-metadata-0\" (UID: \"ddbf55d9-778f-4601-bb0d-05a3d394de49\") " pod="openstack/nova-metadata-0" Dec 13 07:14:52 crc kubenswrapper[4971]: I1213 07:14:52.157574 4971 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-rrdbn\" (UniqueName: \"kubernetes.io/projected/ddbf55d9-778f-4601-bb0d-05a3d394de49-kube-api-access-rrdbn\") pod \"nova-metadata-0\" (UID: \"ddbf55d9-778f-4601-bb0d-05a3d394de49\") " pod="openstack/nova-metadata-0" Dec 13 07:14:52 crc kubenswrapper[4971]: I1213 07:14:52.259630 4971 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/ddbf55d9-778f-4601-bb0d-05a3d394de49-logs\") pod \"nova-metadata-0\" (UID: \"ddbf55d9-778f-4601-bb0d-05a3d394de49\") " pod="openstack/nova-metadata-0" Dec 13 07:14:52 crc kubenswrapper[4971]: I1213 07:14:52.259682 4971 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nova-metadata-tls-certs\" (UniqueName: \"kubernetes.io/secret/ddbf55d9-778f-4601-bb0d-05a3d394de49-nova-metadata-tls-certs\") pod \"nova-metadata-0\" (UID: \"ddbf55d9-778f-4601-bb0d-05a3d394de49\") " pod="openstack/nova-metadata-0" Dec 13 07:14:52 crc kubenswrapper[4971]: I1213 07:14:52.259704 4971 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/ddbf55d9-778f-4601-bb0d-05a3d394de49-combined-ca-bundle\") pod \"nova-metadata-0\" (UID: \"ddbf55d9-778f-4601-bb0d-05a3d394de49\") " pod="openstack/nova-metadata-0" Dec 13 07:14:52 crc kubenswrapper[4971]: I1213 07:14:52.259741 4971 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-rrdbn\" (UniqueName: \"kubernetes.io/projected/ddbf55d9-778f-4601-bb0d-05a3d394de49-kube-api-access-rrdbn\") pod \"nova-metadata-0\" (UID: \"ddbf55d9-778f-4601-bb0d-05a3d394de49\") " pod="openstack/nova-metadata-0" Dec 13 07:14:52 crc kubenswrapper[4971]: I1213 07:14:52.259800 4971 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/ddbf55d9-778f-4601-bb0d-05a3d394de49-config-data\") pod \"nova-metadata-0\" (UID: \"ddbf55d9-778f-4601-bb0d-05a3d394de49\") " pod="openstack/nova-metadata-0" Dec 13 07:14:52 crc kubenswrapper[4971]: I1213 07:14:52.260736 4971 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/ddbf55d9-778f-4601-bb0d-05a3d394de49-logs\") pod \"nova-metadata-0\" (UID: \"ddbf55d9-778f-4601-bb0d-05a3d394de49\") " pod="openstack/nova-metadata-0" Dec 13 07:14:52 crc kubenswrapper[4971]: I1213 07:14:52.265253 4971 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/ddbf55d9-778f-4601-bb0d-05a3d394de49-combined-ca-bundle\") pod \"nova-metadata-0\" (UID: \"ddbf55d9-778f-4601-bb0d-05a3d394de49\") " pod="openstack/nova-metadata-0" Dec 13 07:14:52 crc kubenswrapper[4971]: I1213 07:14:52.265268 4971 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"nova-metadata-tls-certs\" (UniqueName: \"kubernetes.io/secret/ddbf55d9-778f-4601-bb0d-05a3d394de49-nova-metadata-tls-certs\") pod \"nova-metadata-0\" (UID: \"ddbf55d9-778f-4601-bb0d-05a3d394de49\") " pod="openstack/nova-metadata-0" Dec 13 07:14:52 crc kubenswrapper[4971]: I1213 07:14:52.265778 4971 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/ddbf55d9-778f-4601-bb0d-05a3d394de49-config-data\") pod \"nova-metadata-0\" (UID: \"ddbf55d9-778f-4601-bb0d-05a3d394de49\") " pod="openstack/nova-metadata-0" Dec 13 07:14:52 crc kubenswrapper[4971]: I1213 07:14:52.279086 4971 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-rrdbn\" (UniqueName: \"kubernetes.io/projected/ddbf55d9-778f-4601-bb0d-05a3d394de49-kube-api-access-rrdbn\") pod \"nova-metadata-0\" (UID: \"ddbf55d9-778f-4601-bb0d-05a3d394de49\") " pod="openstack/nova-metadata-0" Dec 13 07:14:52 crc kubenswrapper[4971]: I1213 07:14:52.323479 4971 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-metadata-0" Dec 13 07:14:52 crc kubenswrapper[4971]: I1213 07:14:52.752912 4971 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-metadata-0"] Dec 13 07:14:52 crc kubenswrapper[4971]: W1213 07:14:52.761442 4971 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podddbf55d9_778f_4601_bb0d_05a3d394de49.slice/crio-ae222c3d269cebe9893d00ef044b83c2fb5d71db6f454b39a4ca176b353d4e3e WatchSource:0}: Error finding container ae222c3d269cebe9893d00ef044b83c2fb5d71db6f454b39a4ca176b353d4e3e: Status 404 returned error can't find the container with id ae222c3d269cebe9893d00ef044b83c2fb5d71db6f454b39a4ca176b353d4e3e Dec 13 07:14:52 crc kubenswrapper[4971]: I1213 07:14:52.950303 4971 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"ddbf55d9-778f-4601-bb0d-05a3d394de49","Type":"ContainerStarted","Data":"ae222c3d269cebe9893d00ef044b83c2fb5d71db6f454b39a4ca176b353d4e3e"} Dec 13 07:14:53 crc kubenswrapper[4971]: I1213 07:14:53.242124 4971 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-conductor-db-sync-kvw2t" Dec 13 07:14:53 crc kubenswrapper[4971]: I1213 07:14:53.377565 4971 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/6c6829bb-25ee-4102-8335-b747bdbe6121-scripts\") pod \"6c6829bb-25ee-4102-8335-b747bdbe6121\" (UID: \"6c6829bb-25ee-4102-8335-b747bdbe6121\") " Dec 13 07:14:53 crc kubenswrapper[4971]: I1213 07:14:53.377674 4971 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/6c6829bb-25ee-4102-8335-b747bdbe6121-config-data\") pod \"6c6829bb-25ee-4102-8335-b747bdbe6121\" (UID: \"6c6829bb-25ee-4102-8335-b747bdbe6121\") " Dec 13 07:14:53 crc kubenswrapper[4971]: I1213 07:14:53.377750 4971 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-qczdv\" (UniqueName: \"kubernetes.io/projected/6c6829bb-25ee-4102-8335-b747bdbe6121-kube-api-access-qczdv\") pod \"6c6829bb-25ee-4102-8335-b747bdbe6121\" (UID: \"6c6829bb-25ee-4102-8335-b747bdbe6121\") " Dec 13 07:14:53 crc kubenswrapper[4971]: I1213 07:14:53.377840 4971 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/6c6829bb-25ee-4102-8335-b747bdbe6121-combined-ca-bundle\") pod \"6c6829bb-25ee-4102-8335-b747bdbe6121\" (UID: \"6c6829bb-25ee-4102-8335-b747bdbe6121\") " Dec 13 07:14:53 crc kubenswrapper[4971]: I1213 07:14:53.381652 4971 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/6c6829bb-25ee-4102-8335-b747bdbe6121-scripts" (OuterVolumeSpecName: "scripts") pod "6c6829bb-25ee-4102-8335-b747bdbe6121" (UID: "6c6829bb-25ee-4102-8335-b747bdbe6121"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 13 07:14:53 crc kubenswrapper[4971]: I1213 07:14:53.381725 4971 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/6c6829bb-25ee-4102-8335-b747bdbe6121-kube-api-access-qczdv" (OuterVolumeSpecName: "kube-api-access-qczdv") pod "6c6829bb-25ee-4102-8335-b747bdbe6121" (UID: "6c6829bb-25ee-4102-8335-b747bdbe6121"). InnerVolumeSpecName "kube-api-access-qczdv". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 13 07:14:53 crc kubenswrapper[4971]: I1213 07:14:53.403879 4971 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/6c6829bb-25ee-4102-8335-b747bdbe6121-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "6c6829bb-25ee-4102-8335-b747bdbe6121" (UID: "6c6829bb-25ee-4102-8335-b747bdbe6121"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 13 07:14:53 crc kubenswrapper[4971]: I1213 07:14:53.405715 4971 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/6c6829bb-25ee-4102-8335-b747bdbe6121-config-data" (OuterVolumeSpecName: "config-data") pod "6c6829bb-25ee-4102-8335-b747bdbe6121" (UID: "6c6829bb-25ee-4102-8335-b747bdbe6121"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 13 07:14:53 crc kubenswrapper[4971]: I1213 07:14:53.479446 4971 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/6c6829bb-25ee-4102-8335-b747bdbe6121-scripts\") on node \"crc\" DevicePath \"\"" Dec 13 07:14:53 crc kubenswrapper[4971]: I1213 07:14:53.479479 4971 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/6c6829bb-25ee-4102-8335-b747bdbe6121-config-data\") on node \"crc\" DevicePath \"\"" Dec 13 07:14:53 crc kubenswrapper[4971]: I1213 07:14:53.479492 4971 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-qczdv\" (UniqueName: \"kubernetes.io/projected/6c6829bb-25ee-4102-8335-b747bdbe6121-kube-api-access-qczdv\") on node \"crc\" DevicePath \"\"" Dec 13 07:14:53 crc kubenswrapper[4971]: I1213 07:14:53.479504 4971 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/6c6829bb-25ee-4102-8335-b747bdbe6121-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 13 07:14:53 crc kubenswrapper[4971]: I1213 07:14:53.784485 4971 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="3eaa9674-9cd0-497f-80ca-27067fa54ab1" path="/var/lib/kubelet/pods/3eaa9674-9cd0-497f-80ca-27067fa54ab1/volumes" Dec 13 07:14:53 crc kubenswrapper[4971]: I1213 07:14:53.962452 4971 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"ddbf55d9-778f-4601-bb0d-05a3d394de49","Type":"ContainerStarted","Data":"edd3479a90674b9026df96971415f53d846bd1e1419434fa30680a8c2a2007ad"} Dec 13 07:14:53 crc kubenswrapper[4971]: I1213 07:14:53.962502 4971 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"ddbf55d9-778f-4601-bb0d-05a3d394de49","Type":"ContainerStarted","Data":"0902dafe9cf0d3f386fd5aab39ce9ea0516cc49c2c673963ec8413690dd887b6"} Dec 13 07:14:53 crc kubenswrapper[4971]: I1213 07:14:53.964321 4971 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-conductor-db-sync-kvw2t" event={"ID":"6c6829bb-25ee-4102-8335-b747bdbe6121","Type":"ContainerDied","Data":"0495140fdf836519f2c1ebcfffbac483fcc73ca5e8f9532225253c5f477758a7"} Dec 13 07:14:53 crc kubenswrapper[4971]: I1213 07:14:53.964376 4971 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-conductor-db-sync-kvw2t" Dec 13 07:14:53 crc kubenswrapper[4971]: I1213 07:14:53.964380 4971 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="0495140fdf836519f2c1ebcfffbac483fcc73ca5e8f9532225253c5f477758a7" Dec 13 07:14:53 crc kubenswrapper[4971]: I1213 07:14:53.979935 4971 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-metadata-0" podStartSLOduration=2.979918783 podStartE2EDuration="2.979918783s" podCreationTimestamp="2025-12-13 07:14:51 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-13 07:14:53.977153305 +0000 UTC m=+1550.581562753" watchObservedRunningTime="2025-12-13 07:14:53.979918783 +0000 UTC m=+1550.584328231" Dec 13 07:14:54 crc kubenswrapper[4971]: I1213 07:14:54.249482 4971 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-cell1-conductor-0"] Dec 13 07:14:54 crc kubenswrapper[4971]: E1213 07:14:54.249925 4971 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="6c6829bb-25ee-4102-8335-b747bdbe6121" containerName="nova-cell1-conductor-db-sync" Dec 13 07:14:54 crc kubenswrapper[4971]: I1213 07:14:54.249946 4971 state_mem.go:107] "Deleted CPUSet assignment" podUID="6c6829bb-25ee-4102-8335-b747bdbe6121" containerName="nova-cell1-conductor-db-sync" Dec 13 07:14:54 crc kubenswrapper[4971]: I1213 07:14:54.250201 4971 memory_manager.go:354] "RemoveStaleState removing state" podUID="6c6829bb-25ee-4102-8335-b747bdbe6121" containerName="nova-cell1-conductor-db-sync" Dec 13 07:14:54 crc kubenswrapper[4971]: I1213 07:14:54.251530 4971 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-conductor-0" Dec 13 07:14:54 crc kubenswrapper[4971]: I1213 07:14:54.253926 4971 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-cell1-conductor-config-data" Dec 13 07:14:54 crc kubenswrapper[4971]: I1213 07:14:54.265961 4971 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell1-conductor-0"] Dec 13 07:14:54 crc kubenswrapper[4971]: I1213 07:14:54.395850 4971 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/c4b4d448-bf03-4d5c-9017-3b2257922796-combined-ca-bundle\") pod \"nova-cell1-conductor-0\" (UID: \"c4b4d448-bf03-4d5c-9017-3b2257922796\") " pod="openstack/nova-cell1-conductor-0" Dec 13 07:14:54 crc kubenswrapper[4971]: I1213 07:14:54.395923 4971 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-qr4d4\" (UniqueName: \"kubernetes.io/projected/c4b4d448-bf03-4d5c-9017-3b2257922796-kube-api-access-qr4d4\") pod \"nova-cell1-conductor-0\" (UID: \"c4b4d448-bf03-4d5c-9017-3b2257922796\") " pod="openstack/nova-cell1-conductor-0" Dec 13 07:14:54 crc kubenswrapper[4971]: I1213 07:14:54.396002 4971 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/c4b4d448-bf03-4d5c-9017-3b2257922796-config-data\") pod \"nova-cell1-conductor-0\" (UID: \"c4b4d448-bf03-4d5c-9017-3b2257922796\") " pod="openstack/nova-cell1-conductor-0" Dec 13 07:14:54 crc kubenswrapper[4971]: I1213 07:14:54.497756 4971 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-qr4d4\" (UniqueName: \"kubernetes.io/projected/c4b4d448-bf03-4d5c-9017-3b2257922796-kube-api-access-qr4d4\") pod \"nova-cell1-conductor-0\" (UID: \"c4b4d448-bf03-4d5c-9017-3b2257922796\") " pod="openstack/nova-cell1-conductor-0" Dec 13 07:14:54 crc kubenswrapper[4971]: I1213 07:14:54.497893 4971 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/c4b4d448-bf03-4d5c-9017-3b2257922796-config-data\") pod \"nova-cell1-conductor-0\" (UID: \"c4b4d448-bf03-4d5c-9017-3b2257922796\") " pod="openstack/nova-cell1-conductor-0" Dec 13 07:14:54 crc kubenswrapper[4971]: I1213 07:14:54.497973 4971 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/c4b4d448-bf03-4d5c-9017-3b2257922796-combined-ca-bundle\") pod \"nova-cell1-conductor-0\" (UID: \"c4b4d448-bf03-4d5c-9017-3b2257922796\") " pod="openstack/nova-cell1-conductor-0" Dec 13 07:14:54 crc kubenswrapper[4971]: I1213 07:14:54.503610 4971 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/c4b4d448-bf03-4d5c-9017-3b2257922796-combined-ca-bundle\") pod \"nova-cell1-conductor-0\" (UID: \"c4b4d448-bf03-4d5c-9017-3b2257922796\") " pod="openstack/nova-cell1-conductor-0" Dec 13 07:14:54 crc kubenswrapper[4971]: I1213 07:14:54.504646 4971 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/c4b4d448-bf03-4d5c-9017-3b2257922796-config-data\") pod \"nova-cell1-conductor-0\" (UID: \"c4b4d448-bf03-4d5c-9017-3b2257922796\") " pod="openstack/nova-cell1-conductor-0" Dec 13 07:14:54 crc kubenswrapper[4971]: I1213 07:14:54.515903 4971 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-qr4d4\" (UniqueName: \"kubernetes.io/projected/c4b4d448-bf03-4d5c-9017-3b2257922796-kube-api-access-qr4d4\") pod \"nova-cell1-conductor-0\" (UID: \"c4b4d448-bf03-4d5c-9017-3b2257922796\") " pod="openstack/nova-cell1-conductor-0" Dec 13 07:14:54 crc kubenswrapper[4971]: E1213 07:14:54.523609 4971 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" containerID="97dd6db8c1ca363f1d2b43537619b59bc570a33e67bb41c9d1803485ae0de49e" cmd=["/usr/bin/pgrep","-r","DRST","nova-scheduler"] Dec 13 07:14:54 crc kubenswrapper[4971]: E1213 07:14:54.524936 4971 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" containerID="97dd6db8c1ca363f1d2b43537619b59bc570a33e67bb41c9d1803485ae0de49e" cmd=["/usr/bin/pgrep","-r","DRST","nova-scheduler"] Dec 13 07:14:54 crc kubenswrapper[4971]: E1213 07:14:54.526354 4971 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" containerID="97dd6db8c1ca363f1d2b43537619b59bc570a33e67bb41c9d1803485ae0de49e" cmd=["/usr/bin/pgrep","-r","DRST","nova-scheduler"] Dec 13 07:14:54 crc kubenswrapper[4971]: E1213 07:14:54.526395 4971 prober.go:104] "Probe errored" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" probeType="Readiness" pod="openstack/nova-scheduler-0" podUID="156254b1-933a-49b3-ae26-f0ba7f1418b8" containerName="nova-scheduler-scheduler" Dec 13 07:14:54 crc kubenswrapper[4971]: I1213 07:14:54.577937 4971 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-conductor-0" Dec 13 07:14:55 crc kubenswrapper[4971]: I1213 07:14:55.034381 4971 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell1-conductor-0"] Dec 13 07:14:55 crc kubenswrapper[4971]: I1213 07:14:55.665922 4971 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-scheduler-0" Dec 13 07:14:55 crc kubenswrapper[4971]: I1213 07:14:55.827434 4971 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/156254b1-933a-49b3-ae26-f0ba7f1418b8-combined-ca-bundle\") pod \"156254b1-933a-49b3-ae26-f0ba7f1418b8\" (UID: \"156254b1-933a-49b3-ae26-f0ba7f1418b8\") " Dec 13 07:14:55 crc kubenswrapper[4971]: I1213 07:14:55.827492 4971 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-2k4fc\" (UniqueName: \"kubernetes.io/projected/156254b1-933a-49b3-ae26-f0ba7f1418b8-kube-api-access-2k4fc\") pod \"156254b1-933a-49b3-ae26-f0ba7f1418b8\" (UID: \"156254b1-933a-49b3-ae26-f0ba7f1418b8\") " Dec 13 07:14:55 crc kubenswrapper[4971]: I1213 07:14:55.827554 4971 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/156254b1-933a-49b3-ae26-f0ba7f1418b8-config-data\") pod \"156254b1-933a-49b3-ae26-f0ba7f1418b8\" (UID: \"156254b1-933a-49b3-ae26-f0ba7f1418b8\") " Dec 13 07:14:55 crc kubenswrapper[4971]: I1213 07:14:55.837538 4971 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/156254b1-933a-49b3-ae26-f0ba7f1418b8-kube-api-access-2k4fc" (OuterVolumeSpecName: "kube-api-access-2k4fc") pod "156254b1-933a-49b3-ae26-f0ba7f1418b8" (UID: "156254b1-933a-49b3-ae26-f0ba7f1418b8"). InnerVolumeSpecName "kube-api-access-2k4fc". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 13 07:14:55 crc kubenswrapper[4971]: I1213 07:14:55.870864 4971 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/156254b1-933a-49b3-ae26-f0ba7f1418b8-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "156254b1-933a-49b3-ae26-f0ba7f1418b8" (UID: "156254b1-933a-49b3-ae26-f0ba7f1418b8"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 13 07:14:55 crc kubenswrapper[4971]: I1213 07:14:55.899741 4971 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/156254b1-933a-49b3-ae26-f0ba7f1418b8-config-data" (OuterVolumeSpecName: "config-data") pod "156254b1-933a-49b3-ae26-f0ba7f1418b8" (UID: "156254b1-933a-49b3-ae26-f0ba7f1418b8"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 13 07:14:55 crc kubenswrapper[4971]: I1213 07:14:55.929639 4971 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/156254b1-933a-49b3-ae26-f0ba7f1418b8-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 13 07:14:55 crc kubenswrapper[4971]: I1213 07:14:55.929685 4971 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-2k4fc\" (UniqueName: \"kubernetes.io/projected/156254b1-933a-49b3-ae26-f0ba7f1418b8-kube-api-access-2k4fc\") on node \"crc\" DevicePath \"\"" Dec 13 07:14:55 crc kubenswrapper[4971]: I1213 07:14:55.929699 4971 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/156254b1-933a-49b3-ae26-f0ba7f1418b8-config-data\") on node \"crc\" DevicePath \"\"" Dec 13 07:14:55 crc kubenswrapper[4971]: I1213 07:14:55.985417 4971 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-conductor-0" event={"ID":"c4b4d448-bf03-4d5c-9017-3b2257922796","Type":"ContainerStarted","Data":"d0768eae8da769260c7eaccede4ebeaefc22664b59a97dc9bdd199a17d3f021d"} Dec 13 07:14:55 crc kubenswrapper[4971]: I1213 07:14:55.985462 4971 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-conductor-0" event={"ID":"c4b4d448-bf03-4d5c-9017-3b2257922796","Type":"ContainerStarted","Data":"5ccb85c66a3c9a76f77f8ccc4f126f45932182a1773e91d1cc4e8477e205093d"} Dec 13 07:14:55 crc kubenswrapper[4971]: I1213 07:14:55.986617 4971 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-cell1-conductor-0" Dec 13 07:14:55 crc kubenswrapper[4971]: I1213 07:14:55.988716 4971 generic.go:334] "Generic (PLEG): container finished" podID="156254b1-933a-49b3-ae26-f0ba7f1418b8" containerID="97dd6db8c1ca363f1d2b43537619b59bc570a33e67bb41c9d1803485ae0de49e" exitCode=0 Dec 13 07:14:55 crc kubenswrapper[4971]: I1213 07:14:55.988751 4971 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-scheduler-0" event={"ID":"156254b1-933a-49b3-ae26-f0ba7f1418b8","Type":"ContainerDied","Data":"97dd6db8c1ca363f1d2b43537619b59bc570a33e67bb41c9d1803485ae0de49e"} Dec 13 07:14:55 crc kubenswrapper[4971]: I1213 07:14:55.988772 4971 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-scheduler-0" event={"ID":"156254b1-933a-49b3-ae26-f0ba7f1418b8","Type":"ContainerDied","Data":"5d89258312c0366c0d0d89e9bd9e968716f95d94345a8555090415c78567768f"} Dec 13 07:14:55 crc kubenswrapper[4971]: I1213 07:14:55.988788 4971 scope.go:117] "RemoveContainer" containerID="97dd6db8c1ca363f1d2b43537619b59bc570a33e67bb41c9d1803485ae0de49e" Dec 13 07:14:55 crc kubenswrapper[4971]: I1213 07:14:55.988881 4971 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-scheduler-0" Dec 13 07:14:56 crc kubenswrapper[4971]: I1213 07:14:56.006745 4971 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-cell1-conductor-0" podStartSLOduration=2.006724612 podStartE2EDuration="2.006724612s" podCreationTimestamp="2025-12-13 07:14:54 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-13 07:14:56.002087878 +0000 UTC m=+1552.606497326" watchObservedRunningTime="2025-12-13 07:14:56.006724612 +0000 UTC m=+1552.611134070" Dec 13 07:14:56 crc kubenswrapper[4971]: I1213 07:14:56.029276 4971 scope.go:117] "RemoveContainer" containerID="97dd6db8c1ca363f1d2b43537619b59bc570a33e67bb41c9d1803485ae0de49e" Dec 13 07:14:56 crc kubenswrapper[4971]: E1213 07:14:56.029801 4971 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"97dd6db8c1ca363f1d2b43537619b59bc570a33e67bb41c9d1803485ae0de49e\": container with ID starting with 97dd6db8c1ca363f1d2b43537619b59bc570a33e67bb41c9d1803485ae0de49e not found: ID does not exist" containerID="97dd6db8c1ca363f1d2b43537619b59bc570a33e67bb41c9d1803485ae0de49e" Dec 13 07:14:56 crc kubenswrapper[4971]: I1213 07:14:56.029852 4971 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"97dd6db8c1ca363f1d2b43537619b59bc570a33e67bb41c9d1803485ae0de49e"} err="failed to get container status \"97dd6db8c1ca363f1d2b43537619b59bc570a33e67bb41c9d1803485ae0de49e\": rpc error: code = NotFound desc = could not find container \"97dd6db8c1ca363f1d2b43537619b59bc570a33e67bb41c9d1803485ae0de49e\": container with ID starting with 97dd6db8c1ca363f1d2b43537619b59bc570a33e67bb41c9d1803485ae0de49e not found: ID does not exist" Dec 13 07:14:56 crc kubenswrapper[4971]: I1213 07:14:56.044256 4971 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-scheduler-0"] Dec 13 07:14:56 crc kubenswrapper[4971]: I1213 07:14:56.054039 4971 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-scheduler-0"] Dec 13 07:14:56 crc kubenswrapper[4971]: I1213 07:14:56.071425 4971 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-scheduler-0"] Dec 13 07:14:56 crc kubenswrapper[4971]: E1213 07:14:56.071816 4971 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="156254b1-933a-49b3-ae26-f0ba7f1418b8" containerName="nova-scheduler-scheduler" Dec 13 07:14:56 crc kubenswrapper[4971]: I1213 07:14:56.071828 4971 state_mem.go:107] "Deleted CPUSet assignment" podUID="156254b1-933a-49b3-ae26-f0ba7f1418b8" containerName="nova-scheduler-scheduler" Dec 13 07:14:56 crc kubenswrapper[4971]: I1213 07:14:56.072009 4971 memory_manager.go:354] "RemoveStaleState removing state" podUID="156254b1-933a-49b3-ae26-f0ba7f1418b8" containerName="nova-scheduler-scheduler" Dec 13 07:14:56 crc kubenswrapper[4971]: I1213 07:14:56.072660 4971 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-scheduler-0" Dec 13 07:14:56 crc kubenswrapper[4971]: I1213 07:14:56.076890 4971 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-scheduler-config-data" Dec 13 07:14:56 crc kubenswrapper[4971]: I1213 07:14:56.082868 4971 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-scheduler-0"] Dec 13 07:14:56 crc kubenswrapper[4971]: I1213 07:14:56.138476 4971 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/6ae42bbd-1447-4a68-ad90-ae94e2e0b6e5-config-data\") pod \"nova-scheduler-0\" (UID: \"6ae42bbd-1447-4a68-ad90-ae94e2e0b6e5\") " pod="openstack/nova-scheduler-0" Dec 13 07:14:56 crc kubenswrapper[4971]: I1213 07:14:56.138572 4971 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-fdz56\" (UniqueName: \"kubernetes.io/projected/6ae42bbd-1447-4a68-ad90-ae94e2e0b6e5-kube-api-access-fdz56\") pod \"nova-scheduler-0\" (UID: \"6ae42bbd-1447-4a68-ad90-ae94e2e0b6e5\") " pod="openstack/nova-scheduler-0" Dec 13 07:14:56 crc kubenswrapper[4971]: I1213 07:14:56.138641 4971 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/6ae42bbd-1447-4a68-ad90-ae94e2e0b6e5-combined-ca-bundle\") pod \"nova-scheduler-0\" (UID: \"6ae42bbd-1447-4a68-ad90-ae94e2e0b6e5\") " pod="openstack/nova-scheduler-0" Dec 13 07:14:56 crc kubenswrapper[4971]: I1213 07:14:56.241159 4971 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-fdz56\" (UniqueName: \"kubernetes.io/projected/6ae42bbd-1447-4a68-ad90-ae94e2e0b6e5-kube-api-access-fdz56\") pod \"nova-scheduler-0\" (UID: \"6ae42bbd-1447-4a68-ad90-ae94e2e0b6e5\") " pod="openstack/nova-scheduler-0" Dec 13 07:14:56 crc kubenswrapper[4971]: I1213 07:14:56.241733 4971 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/6ae42bbd-1447-4a68-ad90-ae94e2e0b6e5-combined-ca-bundle\") pod \"nova-scheduler-0\" (UID: \"6ae42bbd-1447-4a68-ad90-ae94e2e0b6e5\") " pod="openstack/nova-scheduler-0" Dec 13 07:14:56 crc kubenswrapper[4971]: I1213 07:14:56.241892 4971 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/6ae42bbd-1447-4a68-ad90-ae94e2e0b6e5-config-data\") pod \"nova-scheduler-0\" (UID: \"6ae42bbd-1447-4a68-ad90-ae94e2e0b6e5\") " pod="openstack/nova-scheduler-0" Dec 13 07:14:56 crc kubenswrapper[4971]: I1213 07:14:56.246175 4971 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/6ae42bbd-1447-4a68-ad90-ae94e2e0b6e5-combined-ca-bundle\") pod \"nova-scheduler-0\" (UID: \"6ae42bbd-1447-4a68-ad90-ae94e2e0b6e5\") " pod="openstack/nova-scheduler-0" Dec 13 07:14:56 crc kubenswrapper[4971]: I1213 07:14:56.255099 4971 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/6ae42bbd-1447-4a68-ad90-ae94e2e0b6e5-config-data\") pod \"nova-scheduler-0\" (UID: \"6ae42bbd-1447-4a68-ad90-ae94e2e0b6e5\") " pod="openstack/nova-scheduler-0" Dec 13 07:14:56 crc kubenswrapper[4971]: I1213 07:14:56.262716 4971 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-fdz56\" (UniqueName: \"kubernetes.io/projected/6ae42bbd-1447-4a68-ad90-ae94e2e0b6e5-kube-api-access-fdz56\") pod \"nova-scheduler-0\" (UID: \"6ae42bbd-1447-4a68-ad90-ae94e2e0b6e5\") " pod="openstack/nova-scheduler-0" Dec 13 07:14:56 crc kubenswrapper[4971]: I1213 07:14:56.407127 4971 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-scheduler-0" Dec 13 07:14:56 crc kubenswrapper[4971]: I1213 07:14:56.880625 4971 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-scheduler-0"] Dec 13 07:14:56 crc kubenswrapper[4971]: W1213 07:14:56.886617 4971 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod6ae42bbd_1447_4a68_ad90_ae94e2e0b6e5.slice/crio-d5a99e398b2e8cf2c59612c292e5994dae1a2e739a25eebb15d08ea8cd0b1d46 WatchSource:0}: Error finding container d5a99e398b2e8cf2c59612c292e5994dae1a2e739a25eebb15d08ea8cd0b1d46: Status 404 returned error can't find the container with id d5a99e398b2e8cf2c59612c292e5994dae1a2e739a25eebb15d08ea8cd0b1d46 Dec 13 07:14:56 crc kubenswrapper[4971]: I1213 07:14:56.933132 4971 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-0" Dec 13 07:14:57 crc kubenswrapper[4971]: I1213 07:14:57.003156 4971 generic.go:334] "Generic (PLEG): container finished" podID="b99cc168-f32c-49c0-b150-e4e730e9eacb" containerID="de4bd8652eb387c905466a27a8920b3fcec43795ba4ca6babe7f7fc4a4f3f1bd" exitCode=0 Dec 13 07:14:57 crc kubenswrapper[4971]: I1213 07:14:57.003254 4971 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"b99cc168-f32c-49c0-b150-e4e730e9eacb","Type":"ContainerDied","Data":"de4bd8652eb387c905466a27a8920b3fcec43795ba4ca6babe7f7fc4a4f3f1bd"} Dec 13 07:14:57 crc kubenswrapper[4971]: I1213 07:14:57.003276 4971 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"b99cc168-f32c-49c0-b150-e4e730e9eacb","Type":"ContainerDied","Data":"8dc63f657d1685554d528cb2fe27443e7b84fc13541695786e7ea8bcaa4fc57c"} Dec 13 07:14:57 crc kubenswrapper[4971]: I1213 07:14:57.003292 4971 scope.go:117] "RemoveContainer" containerID="de4bd8652eb387c905466a27a8920b3fcec43795ba4ca6babe7f7fc4a4f3f1bd" Dec 13 07:14:57 crc kubenswrapper[4971]: I1213 07:14:57.003372 4971 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-0" Dec 13 07:14:57 crc kubenswrapper[4971]: I1213 07:14:57.005539 4971 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-scheduler-0" event={"ID":"6ae42bbd-1447-4a68-ad90-ae94e2e0b6e5","Type":"ContainerStarted","Data":"d5a99e398b2e8cf2c59612c292e5994dae1a2e739a25eebb15d08ea8cd0b1d46"} Dec 13 07:14:57 crc kubenswrapper[4971]: I1213 07:14:57.058652 4971 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-4kxf9\" (UniqueName: \"kubernetes.io/projected/b99cc168-f32c-49c0-b150-e4e730e9eacb-kube-api-access-4kxf9\") pod \"b99cc168-f32c-49c0-b150-e4e730e9eacb\" (UID: \"b99cc168-f32c-49c0-b150-e4e730e9eacb\") " Dec 13 07:14:57 crc kubenswrapper[4971]: I1213 07:14:57.058710 4971 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/b99cc168-f32c-49c0-b150-e4e730e9eacb-combined-ca-bundle\") pod \"b99cc168-f32c-49c0-b150-e4e730e9eacb\" (UID: \"b99cc168-f32c-49c0-b150-e4e730e9eacb\") " Dec 13 07:14:57 crc kubenswrapper[4971]: I1213 07:14:57.058790 4971 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/b99cc168-f32c-49c0-b150-e4e730e9eacb-config-data\") pod \"b99cc168-f32c-49c0-b150-e4e730e9eacb\" (UID: \"b99cc168-f32c-49c0-b150-e4e730e9eacb\") " Dec 13 07:14:57 crc kubenswrapper[4971]: I1213 07:14:57.058875 4971 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/b99cc168-f32c-49c0-b150-e4e730e9eacb-logs\") pod \"b99cc168-f32c-49c0-b150-e4e730e9eacb\" (UID: \"b99cc168-f32c-49c0-b150-e4e730e9eacb\") " Dec 13 07:14:57 crc kubenswrapper[4971]: I1213 07:14:57.059479 4971 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/b99cc168-f32c-49c0-b150-e4e730e9eacb-logs" (OuterVolumeSpecName: "logs") pod "b99cc168-f32c-49c0-b150-e4e730e9eacb" (UID: "b99cc168-f32c-49c0-b150-e4e730e9eacb"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 13 07:14:57 crc kubenswrapper[4971]: I1213 07:14:57.059791 4971 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/b99cc168-f32c-49c0-b150-e4e730e9eacb-logs\") on node \"crc\" DevicePath \"\"" Dec 13 07:14:57 crc kubenswrapper[4971]: I1213 07:14:57.064389 4971 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/b99cc168-f32c-49c0-b150-e4e730e9eacb-kube-api-access-4kxf9" (OuterVolumeSpecName: "kube-api-access-4kxf9") pod "b99cc168-f32c-49c0-b150-e4e730e9eacb" (UID: "b99cc168-f32c-49c0-b150-e4e730e9eacb"). InnerVolumeSpecName "kube-api-access-4kxf9". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 13 07:14:57 crc kubenswrapper[4971]: I1213 07:14:57.105704 4971 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/b99cc168-f32c-49c0-b150-e4e730e9eacb-config-data" (OuterVolumeSpecName: "config-data") pod "b99cc168-f32c-49c0-b150-e4e730e9eacb" (UID: "b99cc168-f32c-49c0-b150-e4e730e9eacb"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 13 07:14:57 crc kubenswrapper[4971]: I1213 07:14:57.108462 4971 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/b99cc168-f32c-49c0-b150-e4e730e9eacb-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "b99cc168-f32c-49c0-b150-e4e730e9eacb" (UID: "b99cc168-f32c-49c0-b150-e4e730e9eacb"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 13 07:14:57 crc kubenswrapper[4971]: I1213 07:14:57.161891 4971 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-4kxf9\" (UniqueName: \"kubernetes.io/projected/b99cc168-f32c-49c0-b150-e4e730e9eacb-kube-api-access-4kxf9\") on node \"crc\" DevicePath \"\"" Dec 13 07:14:57 crc kubenswrapper[4971]: I1213 07:14:57.161932 4971 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/b99cc168-f32c-49c0-b150-e4e730e9eacb-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 13 07:14:57 crc kubenswrapper[4971]: I1213 07:14:57.161945 4971 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/b99cc168-f32c-49c0-b150-e4e730e9eacb-config-data\") on node \"crc\" DevicePath \"\"" Dec 13 07:14:57 crc kubenswrapper[4971]: I1213 07:14:57.192100 4971 scope.go:117] "RemoveContainer" containerID="28b75774392e9b16f160e9594f4ecede802f1fec9d3a898016414cd66da31587" Dec 13 07:14:57 crc kubenswrapper[4971]: I1213 07:14:57.213887 4971 scope.go:117] "RemoveContainer" containerID="de4bd8652eb387c905466a27a8920b3fcec43795ba4ca6babe7f7fc4a4f3f1bd" Dec 13 07:14:57 crc kubenswrapper[4971]: E1213 07:14:57.214432 4971 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"de4bd8652eb387c905466a27a8920b3fcec43795ba4ca6babe7f7fc4a4f3f1bd\": container with ID starting with de4bd8652eb387c905466a27a8920b3fcec43795ba4ca6babe7f7fc4a4f3f1bd not found: ID does not exist" containerID="de4bd8652eb387c905466a27a8920b3fcec43795ba4ca6babe7f7fc4a4f3f1bd" Dec 13 07:14:57 crc kubenswrapper[4971]: I1213 07:14:57.214480 4971 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"de4bd8652eb387c905466a27a8920b3fcec43795ba4ca6babe7f7fc4a4f3f1bd"} err="failed to get container status \"de4bd8652eb387c905466a27a8920b3fcec43795ba4ca6babe7f7fc4a4f3f1bd\": rpc error: code = NotFound desc = could not find container \"de4bd8652eb387c905466a27a8920b3fcec43795ba4ca6babe7f7fc4a4f3f1bd\": container with ID starting with de4bd8652eb387c905466a27a8920b3fcec43795ba4ca6babe7f7fc4a4f3f1bd not found: ID does not exist" Dec 13 07:14:57 crc kubenswrapper[4971]: I1213 07:14:57.214508 4971 scope.go:117] "RemoveContainer" containerID="28b75774392e9b16f160e9594f4ecede802f1fec9d3a898016414cd66da31587" Dec 13 07:14:57 crc kubenswrapper[4971]: E1213 07:14:57.214956 4971 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"28b75774392e9b16f160e9594f4ecede802f1fec9d3a898016414cd66da31587\": container with ID starting with 28b75774392e9b16f160e9594f4ecede802f1fec9d3a898016414cd66da31587 not found: ID does not exist" containerID="28b75774392e9b16f160e9594f4ecede802f1fec9d3a898016414cd66da31587" Dec 13 07:14:57 crc kubenswrapper[4971]: I1213 07:14:57.214984 4971 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"28b75774392e9b16f160e9594f4ecede802f1fec9d3a898016414cd66da31587"} err="failed to get container status \"28b75774392e9b16f160e9594f4ecede802f1fec9d3a898016414cd66da31587\": rpc error: code = NotFound desc = could not find container \"28b75774392e9b16f160e9594f4ecede802f1fec9d3a898016414cd66da31587\": container with ID starting with 28b75774392e9b16f160e9594f4ecede802f1fec9d3a898016414cd66da31587 not found: ID does not exist" Dec 13 07:14:57 crc kubenswrapper[4971]: I1213 07:14:57.323585 4971 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-metadata-0" Dec 13 07:14:57 crc kubenswrapper[4971]: I1213 07:14:57.324777 4971 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-metadata-0" Dec 13 07:14:57 crc kubenswrapper[4971]: I1213 07:14:57.369100 4971 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-api-0"] Dec 13 07:14:57 crc kubenswrapper[4971]: I1213 07:14:57.399656 4971 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-api-0"] Dec 13 07:14:57 crc kubenswrapper[4971]: I1213 07:14:57.411911 4971 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-api-0"] Dec 13 07:14:57 crc kubenswrapper[4971]: E1213 07:14:57.412385 4971 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="b99cc168-f32c-49c0-b150-e4e730e9eacb" containerName="nova-api-log" Dec 13 07:14:57 crc kubenswrapper[4971]: I1213 07:14:57.412410 4971 state_mem.go:107] "Deleted CPUSet assignment" podUID="b99cc168-f32c-49c0-b150-e4e730e9eacb" containerName="nova-api-log" Dec 13 07:14:57 crc kubenswrapper[4971]: E1213 07:14:57.412446 4971 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="b99cc168-f32c-49c0-b150-e4e730e9eacb" containerName="nova-api-api" Dec 13 07:14:57 crc kubenswrapper[4971]: I1213 07:14:57.412455 4971 state_mem.go:107] "Deleted CPUSet assignment" podUID="b99cc168-f32c-49c0-b150-e4e730e9eacb" containerName="nova-api-api" Dec 13 07:14:57 crc kubenswrapper[4971]: I1213 07:14:57.412698 4971 memory_manager.go:354] "RemoveStaleState removing state" podUID="b99cc168-f32c-49c0-b150-e4e730e9eacb" containerName="nova-api-api" Dec 13 07:14:57 crc kubenswrapper[4971]: I1213 07:14:57.412726 4971 memory_manager.go:354] "RemoveStaleState removing state" podUID="b99cc168-f32c-49c0-b150-e4e730e9eacb" containerName="nova-api-log" Dec 13 07:14:57 crc kubenswrapper[4971]: I1213 07:14:57.413964 4971 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-0" Dec 13 07:14:57 crc kubenswrapper[4971]: I1213 07:14:57.416232 4971 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-api-config-data" Dec 13 07:14:57 crc kubenswrapper[4971]: I1213 07:14:57.422186 4971 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-api-0"] Dec 13 07:14:57 crc kubenswrapper[4971]: I1213 07:14:57.571353 4971 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/57f1d9a5-cc22-419a-a944-ec9581785de2-config-data\") pod \"nova-api-0\" (UID: \"57f1d9a5-cc22-419a-a944-ec9581785de2\") " pod="openstack/nova-api-0" Dec 13 07:14:57 crc kubenswrapper[4971]: I1213 07:14:57.571396 4971 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/57f1d9a5-cc22-419a-a944-ec9581785de2-logs\") pod \"nova-api-0\" (UID: \"57f1d9a5-cc22-419a-a944-ec9581785de2\") " pod="openstack/nova-api-0" Dec 13 07:14:57 crc kubenswrapper[4971]: I1213 07:14:57.571423 4971 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-2r8wv\" (UniqueName: \"kubernetes.io/projected/57f1d9a5-cc22-419a-a944-ec9581785de2-kube-api-access-2r8wv\") pod \"nova-api-0\" (UID: \"57f1d9a5-cc22-419a-a944-ec9581785de2\") " pod="openstack/nova-api-0" Dec 13 07:14:57 crc kubenswrapper[4971]: I1213 07:14:57.571445 4971 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/57f1d9a5-cc22-419a-a944-ec9581785de2-combined-ca-bundle\") pod \"nova-api-0\" (UID: \"57f1d9a5-cc22-419a-a944-ec9581785de2\") " pod="openstack/nova-api-0" Dec 13 07:14:57 crc kubenswrapper[4971]: I1213 07:14:57.672951 4971 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/57f1d9a5-cc22-419a-a944-ec9581785de2-config-data\") pod \"nova-api-0\" (UID: \"57f1d9a5-cc22-419a-a944-ec9581785de2\") " pod="openstack/nova-api-0" Dec 13 07:14:57 crc kubenswrapper[4971]: I1213 07:14:57.673062 4971 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/57f1d9a5-cc22-419a-a944-ec9581785de2-logs\") pod \"nova-api-0\" (UID: \"57f1d9a5-cc22-419a-a944-ec9581785de2\") " pod="openstack/nova-api-0" Dec 13 07:14:57 crc kubenswrapper[4971]: I1213 07:14:57.673096 4971 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-2r8wv\" (UniqueName: \"kubernetes.io/projected/57f1d9a5-cc22-419a-a944-ec9581785de2-kube-api-access-2r8wv\") pod \"nova-api-0\" (UID: \"57f1d9a5-cc22-419a-a944-ec9581785de2\") " pod="openstack/nova-api-0" Dec 13 07:14:57 crc kubenswrapper[4971]: I1213 07:14:57.673641 4971 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/57f1d9a5-cc22-419a-a944-ec9581785de2-combined-ca-bundle\") pod \"nova-api-0\" (UID: \"57f1d9a5-cc22-419a-a944-ec9581785de2\") " pod="openstack/nova-api-0" Dec 13 07:14:57 crc kubenswrapper[4971]: I1213 07:14:57.673583 4971 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/57f1d9a5-cc22-419a-a944-ec9581785de2-logs\") pod \"nova-api-0\" (UID: \"57f1d9a5-cc22-419a-a944-ec9581785de2\") " pod="openstack/nova-api-0" Dec 13 07:14:57 crc kubenswrapper[4971]: I1213 07:14:57.686887 4971 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/57f1d9a5-cc22-419a-a944-ec9581785de2-config-data\") pod \"nova-api-0\" (UID: \"57f1d9a5-cc22-419a-a944-ec9581785de2\") " pod="openstack/nova-api-0" Dec 13 07:14:57 crc kubenswrapper[4971]: I1213 07:14:57.691611 4971 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-2r8wv\" (UniqueName: \"kubernetes.io/projected/57f1d9a5-cc22-419a-a944-ec9581785de2-kube-api-access-2r8wv\") pod \"nova-api-0\" (UID: \"57f1d9a5-cc22-419a-a944-ec9581785de2\") " pod="openstack/nova-api-0" Dec 13 07:14:57 crc kubenswrapper[4971]: I1213 07:14:57.706627 4971 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/57f1d9a5-cc22-419a-a944-ec9581785de2-combined-ca-bundle\") pod \"nova-api-0\" (UID: \"57f1d9a5-cc22-419a-a944-ec9581785de2\") " pod="openstack/nova-api-0" Dec 13 07:14:57 crc kubenswrapper[4971]: I1213 07:14:57.733194 4971 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-0" Dec 13 07:14:57 crc kubenswrapper[4971]: I1213 07:14:57.797929 4971 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="156254b1-933a-49b3-ae26-f0ba7f1418b8" path="/var/lib/kubelet/pods/156254b1-933a-49b3-ae26-f0ba7f1418b8/volumes" Dec 13 07:14:57 crc kubenswrapper[4971]: I1213 07:14:57.799111 4971 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="b99cc168-f32c-49c0-b150-e4e730e9eacb" path="/var/lib/kubelet/pods/b99cc168-f32c-49c0-b150-e4e730e9eacb/volumes" Dec 13 07:14:58 crc kubenswrapper[4971]: I1213 07:14:58.015676 4971 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-scheduler-0" event={"ID":"6ae42bbd-1447-4a68-ad90-ae94e2e0b6e5","Type":"ContainerStarted","Data":"932b370c20d3f4525fe2074e697c4a9bcc071571b0a037c1b415fc03f7a65e0a"} Dec 13 07:14:58 crc kubenswrapper[4971]: I1213 07:14:58.043951 4971 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-scheduler-0" podStartSLOduration=2.043928877 podStartE2EDuration="2.043928877s" podCreationTimestamp="2025-12-13 07:14:56 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-13 07:14:58.037364975 +0000 UTC m=+1554.641774423" watchObservedRunningTime="2025-12-13 07:14:58.043928877 +0000 UTC m=+1554.648338325" Dec 13 07:14:58 crc kubenswrapper[4971]: I1213 07:14:58.226779 4971 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-api-0"] Dec 13 07:14:59 crc kubenswrapper[4971]: I1213 07:14:59.030846 4971 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"57f1d9a5-cc22-419a-a944-ec9581785de2","Type":"ContainerStarted","Data":"c67d150ca7039f2ee1dcb3af173c2b626da09aa4a1cc47532a65a415b3fc9f05"} Dec 13 07:14:59 crc kubenswrapper[4971]: I1213 07:14:59.031571 4971 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"57f1d9a5-cc22-419a-a944-ec9581785de2","Type":"ContainerStarted","Data":"3679a59578f7ed86541756cdb8f478aa439e64cf5a03668c1353a452c6e1b989"} Dec 13 07:14:59 crc kubenswrapper[4971]: I1213 07:14:59.031592 4971 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"57f1d9a5-cc22-419a-a944-ec9581785de2","Type":"ContainerStarted","Data":"f36843c1785b95b0e6501feea2b88005e62fc386d4de84c81d1b3a896e580731"} Dec 13 07:14:59 crc kubenswrapper[4971]: I1213 07:14:59.768796 4971 scope.go:117] "RemoveContainer" containerID="e8beee43d15719dccb6892d6f5711674492425300efa098be89033bc65977d14" Dec 13 07:14:59 crc kubenswrapper[4971]: E1213 07:14:59.769031 4971 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-82xjz_openshift-machine-config-operator(e0c4af12-edda-41d7-9b44-f87396174ef2)\"" pod="openshift-machine-config-operator/machine-config-daemon-82xjz" podUID="e0c4af12-edda-41d7-9b44-f87396174ef2" Dec 13 07:15:00 crc kubenswrapper[4971]: I1213 07:15:00.136415 4971 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-api-0" podStartSLOduration=3.136396464 podStartE2EDuration="3.136396464s" podCreationTimestamp="2025-12-13 07:14:57 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-13 07:14:59.055370989 +0000 UTC m=+1555.659780457" watchObservedRunningTime="2025-12-13 07:15:00.136396464 +0000 UTC m=+1556.740805912" Dec 13 07:15:00 crc kubenswrapper[4971]: I1213 07:15:00.138195 4971 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29426835-d7bs5"] Dec 13 07:15:00 crc kubenswrapper[4971]: I1213 07:15:00.139497 4971 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29426835-d7bs5" Dec 13 07:15:00 crc kubenswrapper[4971]: I1213 07:15:00.142342 4971 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-operator-lifecycle-manager"/"collect-profiles-config" Dec 13 07:15:00 crc kubenswrapper[4971]: I1213 07:15:00.142468 4971 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"collect-profiles-dockercfg-kzf4t" Dec 13 07:15:00 crc kubenswrapper[4971]: I1213 07:15:00.153656 4971 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29426835-d7bs5"] Dec 13 07:15:00 crc kubenswrapper[4971]: I1213 07:15:00.225675 4971 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/2054169f-b228-4471-ba23-2a74cb0a126c-config-volume\") pod \"collect-profiles-29426835-d7bs5\" (UID: \"2054169f-b228-4471-ba23-2a74cb0a126c\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29426835-d7bs5" Dec 13 07:15:00 crc kubenswrapper[4971]: I1213 07:15:00.226024 4971 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/2054169f-b228-4471-ba23-2a74cb0a126c-secret-volume\") pod \"collect-profiles-29426835-d7bs5\" (UID: \"2054169f-b228-4471-ba23-2a74cb0a126c\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29426835-d7bs5" Dec 13 07:15:00 crc kubenswrapper[4971]: I1213 07:15:00.226584 4971 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-n77jz\" (UniqueName: \"kubernetes.io/projected/2054169f-b228-4471-ba23-2a74cb0a126c-kube-api-access-n77jz\") pod \"collect-profiles-29426835-d7bs5\" (UID: \"2054169f-b228-4471-ba23-2a74cb0a126c\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29426835-d7bs5" Dec 13 07:15:00 crc kubenswrapper[4971]: I1213 07:15:00.328147 4971 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/2054169f-b228-4471-ba23-2a74cb0a126c-config-volume\") pod \"collect-profiles-29426835-d7bs5\" (UID: \"2054169f-b228-4471-ba23-2a74cb0a126c\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29426835-d7bs5" Dec 13 07:15:00 crc kubenswrapper[4971]: I1213 07:15:00.328264 4971 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/2054169f-b228-4471-ba23-2a74cb0a126c-secret-volume\") pod \"collect-profiles-29426835-d7bs5\" (UID: \"2054169f-b228-4471-ba23-2a74cb0a126c\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29426835-d7bs5" Dec 13 07:15:00 crc kubenswrapper[4971]: I1213 07:15:00.328392 4971 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-n77jz\" (UniqueName: \"kubernetes.io/projected/2054169f-b228-4471-ba23-2a74cb0a126c-kube-api-access-n77jz\") pod \"collect-profiles-29426835-d7bs5\" (UID: \"2054169f-b228-4471-ba23-2a74cb0a126c\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29426835-d7bs5" Dec 13 07:15:00 crc kubenswrapper[4971]: I1213 07:15:00.329815 4971 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/2054169f-b228-4471-ba23-2a74cb0a126c-config-volume\") pod \"collect-profiles-29426835-d7bs5\" (UID: \"2054169f-b228-4471-ba23-2a74cb0a126c\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29426835-d7bs5" Dec 13 07:15:00 crc kubenswrapper[4971]: I1213 07:15:00.341216 4971 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/2054169f-b228-4471-ba23-2a74cb0a126c-secret-volume\") pod \"collect-profiles-29426835-d7bs5\" (UID: \"2054169f-b228-4471-ba23-2a74cb0a126c\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29426835-d7bs5" Dec 13 07:15:00 crc kubenswrapper[4971]: I1213 07:15:00.350833 4971 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-n77jz\" (UniqueName: \"kubernetes.io/projected/2054169f-b228-4471-ba23-2a74cb0a126c-kube-api-access-n77jz\") pod \"collect-profiles-29426835-d7bs5\" (UID: \"2054169f-b228-4471-ba23-2a74cb0a126c\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29426835-d7bs5" Dec 13 07:15:00 crc kubenswrapper[4971]: I1213 07:15:00.458593 4971 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29426835-d7bs5" Dec 13 07:15:00 crc kubenswrapper[4971]: I1213 07:15:00.893618 4971 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29426835-d7bs5"] Dec 13 07:15:01 crc kubenswrapper[4971]: I1213 07:15:01.052374 4971 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29426835-d7bs5" event={"ID":"2054169f-b228-4471-ba23-2a74cb0a126c","Type":"ContainerStarted","Data":"6334e16788385b038665f26911754d614f73fb16854784b3497d817139e836da"} Dec 13 07:15:01 crc kubenswrapper[4971]: I1213 07:15:01.052412 4971 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29426835-d7bs5" event={"ID":"2054169f-b228-4471-ba23-2a74cb0a126c","Type":"ContainerStarted","Data":"2287c350a74ca1cafcc6f49a32f5385c258ae82f6954a48cedf95a2f432a22be"} Dec 13 07:15:01 crc kubenswrapper[4971]: I1213 07:15:01.071457 4971 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-operator-lifecycle-manager/collect-profiles-29426835-d7bs5" podStartSLOduration=1.071439794 podStartE2EDuration="1.071439794s" podCreationTimestamp="2025-12-13 07:15:00 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-13 07:15:01.070168962 +0000 UTC m=+1557.674578410" watchObservedRunningTime="2025-12-13 07:15:01.071439794 +0000 UTC m=+1557.675849242" Dec 13 07:15:01 crc kubenswrapper[4971]: I1213 07:15:01.407769 4971 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-scheduler-0" Dec 13 07:15:02 crc kubenswrapper[4971]: I1213 07:15:02.064100 4971 generic.go:334] "Generic (PLEG): container finished" podID="2054169f-b228-4471-ba23-2a74cb0a126c" containerID="6334e16788385b038665f26911754d614f73fb16854784b3497d817139e836da" exitCode=0 Dec 13 07:15:02 crc kubenswrapper[4971]: I1213 07:15:02.064188 4971 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29426835-d7bs5" event={"ID":"2054169f-b228-4471-ba23-2a74cb0a126c","Type":"ContainerDied","Data":"6334e16788385b038665f26911754d614f73fb16854784b3497d817139e836da"} Dec 13 07:15:02 crc kubenswrapper[4971]: I1213 07:15:02.323739 4971 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/nova-metadata-0" Dec 13 07:15:02 crc kubenswrapper[4971]: I1213 07:15:02.324097 4971 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/nova-metadata-0" Dec 13 07:15:03 crc kubenswrapper[4971]: I1213 07:15:03.339767 4971 prober.go:107] "Probe failed" probeType="Startup" pod="openstack/nova-metadata-0" podUID="ddbf55d9-778f-4601-bb0d-05a3d394de49" containerName="nova-metadata-metadata" probeResult="failure" output="Get \"https://10.217.0.181:8775/\": net/http: request canceled (Client.Timeout exceeded while awaiting headers)" Dec 13 07:15:03 crc kubenswrapper[4971]: I1213 07:15:03.339770 4971 prober.go:107] "Probe failed" probeType="Startup" pod="openstack/nova-metadata-0" podUID="ddbf55d9-778f-4601-bb0d-05a3d394de49" containerName="nova-metadata-log" probeResult="failure" output="Get \"https://10.217.0.181:8775/\": net/http: request canceled (Client.Timeout exceeded while awaiting headers)" Dec 13 07:15:03 crc kubenswrapper[4971]: I1213 07:15:03.518038 4971 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29426835-d7bs5" Dec 13 07:15:03 crc kubenswrapper[4971]: I1213 07:15:03.600921 4971 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/2054169f-b228-4471-ba23-2a74cb0a126c-config-volume\") pod \"2054169f-b228-4471-ba23-2a74cb0a126c\" (UID: \"2054169f-b228-4471-ba23-2a74cb0a126c\") " Dec 13 07:15:03 crc kubenswrapper[4971]: I1213 07:15:03.601323 4971 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/2054169f-b228-4471-ba23-2a74cb0a126c-secret-volume\") pod \"2054169f-b228-4471-ba23-2a74cb0a126c\" (UID: \"2054169f-b228-4471-ba23-2a74cb0a126c\") " Dec 13 07:15:03 crc kubenswrapper[4971]: I1213 07:15:03.601487 4971 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-n77jz\" (UniqueName: \"kubernetes.io/projected/2054169f-b228-4471-ba23-2a74cb0a126c-kube-api-access-n77jz\") pod \"2054169f-b228-4471-ba23-2a74cb0a126c\" (UID: \"2054169f-b228-4471-ba23-2a74cb0a126c\") " Dec 13 07:15:03 crc kubenswrapper[4971]: I1213 07:15:03.603502 4971 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/2054169f-b228-4471-ba23-2a74cb0a126c-config-volume" (OuterVolumeSpecName: "config-volume") pod "2054169f-b228-4471-ba23-2a74cb0a126c" (UID: "2054169f-b228-4471-ba23-2a74cb0a126c"). InnerVolumeSpecName "config-volume". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 13 07:15:03 crc kubenswrapper[4971]: I1213 07:15:03.611341 4971 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/2054169f-b228-4471-ba23-2a74cb0a126c-kube-api-access-n77jz" (OuterVolumeSpecName: "kube-api-access-n77jz") pod "2054169f-b228-4471-ba23-2a74cb0a126c" (UID: "2054169f-b228-4471-ba23-2a74cb0a126c"). InnerVolumeSpecName "kube-api-access-n77jz". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 13 07:15:03 crc kubenswrapper[4971]: I1213 07:15:03.614686 4971 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/2054169f-b228-4471-ba23-2a74cb0a126c-secret-volume" (OuterVolumeSpecName: "secret-volume") pod "2054169f-b228-4471-ba23-2a74cb0a126c" (UID: "2054169f-b228-4471-ba23-2a74cb0a126c"). InnerVolumeSpecName "secret-volume". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 13 07:15:03 crc kubenswrapper[4971]: I1213 07:15:03.703501 4971 reconciler_common.go:293] "Volume detached for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/2054169f-b228-4471-ba23-2a74cb0a126c-secret-volume\") on node \"crc\" DevicePath \"\"" Dec 13 07:15:03 crc kubenswrapper[4971]: I1213 07:15:03.703821 4971 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-n77jz\" (UniqueName: \"kubernetes.io/projected/2054169f-b228-4471-ba23-2a74cb0a126c-kube-api-access-n77jz\") on node \"crc\" DevicePath \"\"" Dec 13 07:15:03 crc kubenswrapper[4971]: I1213 07:15:03.703835 4971 reconciler_common.go:293] "Volume detached for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/2054169f-b228-4471-ba23-2a74cb0a126c-config-volume\") on node \"crc\" DevicePath \"\"" Dec 13 07:15:04 crc kubenswrapper[4971]: I1213 07:15:04.085801 4971 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29426835-d7bs5" event={"ID":"2054169f-b228-4471-ba23-2a74cb0a126c","Type":"ContainerDied","Data":"2287c350a74ca1cafcc6f49a32f5385c258ae82f6954a48cedf95a2f432a22be"} Dec 13 07:15:04 crc kubenswrapper[4971]: I1213 07:15:04.086060 4971 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="2287c350a74ca1cafcc6f49a32f5385c258ae82f6954a48cedf95a2f432a22be" Dec 13 07:15:04 crc kubenswrapper[4971]: I1213 07:15:04.085877 4971 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29426835-d7bs5" Dec 13 07:15:04 crc kubenswrapper[4971]: I1213 07:15:04.611483 4971 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/nova-cell1-conductor-0" Dec 13 07:15:06 crc kubenswrapper[4971]: I1213 07:15:06.408318 4971 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/nova-scheduler-0" Dec 13 07:15:06 crc kubenswrapper[4971]: I1213 07:15:06.435727 4971 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/nova-scheduler-0" Dec 13 07:15:07 crc kubenswrapper[4971]: I1213 07:15:07.139658 4971 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/nova-scheduler-0" Dec 13 07:15:07 crc kubenswrapper[4971]: I1213 07:15:07.735434 4971 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/nova-api-0" Dec 13 07:15:07 crc kubenswrapper[4971]: I1213 07:15:07.736701 4971 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/nova-api-0" Dec 13 07:15:08 crc kubenswrapper[4971]: I1213 07:15:08.817706 4971 prober.go:107] "Probe failed" probeType="Startup" pod="openstack/nova-api-0" podUID="57f1d9a5-cc22-419a-a944-ec9581785de2" containerName="nova-api-api" probeResult="failure" output="Get \"http://10.217.0.184:8774/\": context deadline exceeded (Client.Timeout exceeded while awaiting headers)" Dec 13 07:15:08 crc kubenswrapper[4971]: I1213 07:15:08.817755 4971 prober.go:107] "Probe failed" probeType="Startup" pod="openstack/nova-api-0" podUID="57f1d9a5-cc22-419a-a944-ec9581785de2" containerName="nova-api-log" probeResult="failure" output="Get \"http://10.217.0.184:8774/\": context deadline exceeded (Client.Timeout exceeded while awaiting headers)" Dec 13 07:15:12 crc kubenswrapper[4971]: I1213 07:15:12.334738 4971 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/nova-metadata-0" Dec 13 07:15:12 crc kubenswrapper[4971]: I1213 07:15:12.338859 4971 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/nova-metadata-0" Dec 13 07:15:12 crc kubenswrapper[4971]: I1213 07:15:12.344613 4971 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/nova-metadata-0" Dec 13 07:15:13 crc kubenswrapper[4971]: I1213 07:15:13.169244 4971 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/nova-metadata-0" Dec 13 07:15:14 crc kubenswrapper[4971]: I1213 07:15:14.169212 4971 generic.go:334] "Generic (PLEG): container finished" podID="9bb1801a-532f-4b09-9ae3-d4b61e8a9c97" containerID="f48369c49c58c454eabe532f1f904b7c0c3a079ce470fb0caf73489fc5a83168" exitCode=137 Dec 13 07:15:14 crc kubenswrapper[4971]: I1213 07:15:14.169249 4971 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-novncproxy-0" event={"ID":"9bb1801a-532f-4b09-9ae3-d4b61e8a9c97","Type":"ContainerDied","Data":"f48369c49c58c454eabe532f1f904b7c0c3a079ce470fb0caf73489fc5a83168"} Dec 13 07:15:14 crc kubenswrapper[4971]: I1213 07:15:14.169472 4971 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-novncproxy-0" event={"ID":"9bb1801a-532f-4b09-9ae3-d4b61e8a9c97","Type":"ContainerDied","Data":"35cabbbbe703889299039bf63469bfd0df4dceb8cc391e80c4911395615172c5"} Dec 13 07:15:14 crc kubenswrapper[4971]: I1213 07:15:14.169490 4971 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="35cabbbbe703889299039bf63469bfd0df4dceb8cc391e80c4911395615172c5" Dec 13 07:15:14 crc kubenswrapper[4971]: I1213 07:15:14.174782 4971 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-novncproxy-0" Dec 13 07:15:14 crc kubenswrapper[4971]: I1213 07:15:14.312590 4971 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/9bb1801a-532f-4b09-9ae3-d4b61e8a9c97-combined-ca-bundle\") pod \"9bb1801a-532f-4b09-9ae3-d4b61e8a9c97\" (UID: \"9bb1801a-532f-4b09-9ae3-d4b61e8a9c97\") " Dec 13 07:15:14 crc kubenswrapper[4971]: I1213 07:15:14.312944 4971 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-sgnlr\" (UniqueName: \"kubernetes.io/projected/9bb1801a-532f-4b09-9ae3-d4b61e8a9c97-kube-api-access-sgnlr\") pod \"9bb1801a-532f-4b09-9ae3-d4b61e8a9c97\" (UID: \"9bb1801a-532f-4b09-9ae3-d4b61e8a9c97\") " Dec 13 07:15:14 crc kubenswrapper[4971]: I1213 07:15:14.312992 4971 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/9bb1801a-532f-4b09-9ae3-d4b61e8a9c97-config-data\") pod \"9bb1801a-532f-4b09-9ae3-d4b61e8a9c97\" (UID: \"9bb1801a-532f-4b09-9ae3-d4b61e8a9c97\") " Dec 13 07:15:14 crc kubenswrapper[4971]: I1213 07:15:14.321255 4971 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/9bb1801a-532f-4b09-9ae3-d4b61e8a9c97-kube-api-access-sgnlr" (OuterVolumeSpecName: "kube-api-access-sgnlr") pod "9bb1801a-532f-4b09-9ae3-d4b61e8a9c97" (UID: "9bb1801a-532f-4b09-9ae3-d4b61e8a9c97"). InnerVolumeSpecName "kube-api-access-sgnlr". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 13 07:15:14 crc kubenswrapper[4971]: I1213 07:15:14.340484 4971 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/9bb1801a-532f-4b09-9ae3-d4b61e8a9c97-config-data" (OuterVolumeSpecName: "config-data") pod "9bb1801a-532f-4b09-9ae3-d4b61e8a9c97" (UID: "9bb1801a-532f-4b09-9ae3-d4b61e8a9c97"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 13 07:15:14 crc kubenswrapper[4971]: I1213 07:15:14.358289 4971 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/9bb1801a-532f-4b09-9ae3-d4b61e8a9c97-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "9bb1801a-532f-4b09-9ae3-d4b61e8a9c97" (UID: "9bb1801a-532f-4b09-9ae3-d4b61e8a9c97"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 13 07:15:14 crc kubenswrapper[4971]: I1213 07:15:14.416863 4971 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-sgnlr\" (UniqueName: \"kubernetes.io/projected/9bb1801a-532f-4b09-9ae3-d4b61e8a9c97-kube-api-access-sgnlr\") on node \"crc\" DevicePath \"\"" Dec 13 07:15:14 crc kubenswrapper[4971]: I1213 07:15:14.416913 4971 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/9bb1801a-532f-4b09-9ae3-d4b61e8a9c97-config-data\") on node \"crc\" DevicePath \"\"" Dec 13 07:15:14 crc kubenswrapper[4971]: I1213 07:15:14.416931 4971 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/9bb1801a-532f-4b09-9ae3-d4b61e8a9c97-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 13 07:15:14 crc kubenswrapper[4971]: I1213 07:15:14.768408 4971 scope.go:117] "RemoveContainer" containerID="e8beee43d15719dccb6892d6f5711674492425300efa098be89033bc65977d14" Dec 13 07:15:14 crc kubenswrapper[4971]: E1213 07:15:14.768863 4971 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-82xjz_openshift-machine-config-operator(e0c4af12-edda-41d7-9b44-f87396174ef2)\"" pod="openshift-machine-config-operator/machine-config-daemon-82xjz" podUID="e0c4af12-edda-41d7-9b44-f87396174ef2" Dec 13 07:15:15 crc kubenswrapper[4971]: I1213 07:15:15.179393 4971 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-novncproxy-0" Dec 13 07:15:15 crc kubenswrapper[4971]: I1213 07:15:15.213440 4971 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-cell1-novncproxy-0"] Dec 13 07:15:15 crc kubenswrapper[4971]: I1213 07:15:15.240729 4971 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-cell1-novncproxy-0"] Dec 13 07:15:15 crc kubenswrapper[4971]: I1213 07:15:15.247651 4971 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-cell1-novncproxy-0"] Dec 13 07:15:15 crc kubenswrapper[4971]: E1213 07:15:15.248031 4971 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="9bb1801a-532f-4b09-9ae3-d4b61e8a9c97" containerName="nova-cell1-novncproxy-novncproxy" Dec 13 07:15:15 crc kubenswrapper[4971]: I1213 07:15:15.248049 4971 state_mem.go:107] "Deleted CPUSet assignment" podUID="9bb1801a-532f-4b09-9ae3-d4b61e8a9c97" containerName="nova-cell1-novncproxy-novncproxy" Dec 13 07:15:15 crc kubenswrapper[4971]: E1213 07:15:15.248085 4971 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="2054169f-b228-4471-ba23-2a74cb0a126c" containerName="collect-profiles" Dec 13 07:15:15 crc kubenswrapper[4971]: I1213 07:15:15.248092 4971 state_mem.go:107] "Deleted CPUSet assignment" podUID="2054169f-b228-4471-ba23-2a74cb0a126c" containerName="collect-profiles" Dec 13 07:15:15 crc kubenswrapper[4971]: I1213 07:15:15.248246 4971 memory_manager.go:354] "RemoveStaleState removing state" podUID="9bb1801a-532f-4b09-9ae3-d4b61e8a9c97" containerName="nova-cell1-novncproxy-novncproxy" Dec 13 07:15:15 crc kubenswrapper[4971]: I1213 07:15:15.248271 4971 memory_manager.go:354] "RemoveStaleState removing state" podUID="2054169f-b228-4471-ba23-2a74cb0a126c" containerName="collect-profiles" Dec 13 07:15:15 crc kubenswrapper[4971]: I1213 07:15:15.248882 4971 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-novncproxy-0" Dec 13 07:15:15 crc kubenswrapper[4971]: I1213 07:15:15.251893 4971 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-nova-novncproxy-cell1-vencrypt" Dec 13 07:15:15 crc kubenswrapper[4971]: I1213 07:15:15.252551 4971 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-cell1-novncproxy-config-data" Dec 13 07:15:15 crc kubenswrapper[4971]: I1213 07:15:15.252784 4971 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-nova-novncproxy-cell1-public-svc" Dec 13 07:15:15 crc kubenswrapper[4971]: I1213 07:15:15.267652 4971 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell1-novncproxy-0"] Dec 13 07:15:15 crc kubenswrapper[4971]: I1213 07:15:15.332122 4971 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/38d62929-7b21-4bf7-a2ac-f31047134d54-config-data\") pod \"nova-cell1-novncproxy-0\" (UID: \"38d62929-7b21-4bf7-a2ac-f31047134d54\") " pod="openstack/nova-cell1-novncproxy-0" Dec 13 07:15:15 crc kubenswrapper[4971]: I1213 07:15:15.332176 4971 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/38d62929-7b21-4bf7-a2ac-f31047134d54-combined-ca-bundle\") pod \"nova-cell1-novncproxy-0\" (UID: \"38d62929-7b21-4bf7-a2ac-f31047134d54\") " pod="openstack/nova-cell1-novncproxy-0" Dec 13 07:15:15 crc kubenswrapper[4971]: I1213 07:15:15.332217 4971 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-hmn5x\" (UniqueName: \"kubernetes.io/projected/38d62929-7b21-4bf7-a2ac-f31047134d54-kube-api-access-hmn5x\") pod \"nova-cell1-novncproxy-0\" (UID: \"38d62929-7b21-4bf7-a2ac-f31047134d54\") " pod="openstack/nova-cell1-novncproxy-0" Dec 13 07:15:15 crc kubenswrapper[4971]: I1213 07:15:15.332268 4971 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"vencrypt-tls-certs\" (UniqueName: \"kubernetes.io/secret/38d62929-7b21-4bf7-a2ac-f31047134d54-vencrypt-tls-certs\") pod \"nova-cell1-novncproxy-0\" (UID: \"38d62929-7b21-4bf7-a2ac-f31047134d54\") " pod="openstack/nova-cell1-novncproxy-0" Dec 13 07:15:15 crc kubenswrapper[4971]: I1213 07:15:15.332290 4971 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"nova-novncproxy-tls-certs\" (UniqueName: \"kubernetes.io/secret/38d62929-7b21-4bf7-a2ac-f31047134d54-nova-novncproxy-tls-certs\") pod \"nova-cell1-novncproxy-0\" (UID: \"38d62929-7b21-4bf7-a2ac-f31047134d54\") " pod="openstack/nova-cell1-novncproxy-0" Dec 13 07:15:15 crc kubenswrapper[4971]: I1213 07:15:15.433646 4971 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/38d62929-7b21-4bf7-a2ac-f31047134d54-config-data\") pod \"nova-cell1-novncproxy-0\" (UID: \"38d62929-7b21-4bf7-a2ac-f31047134d54\") " pod="openstack/nova-cell1-novncproxy-0" Dec 13 07:15:15 crc kubenswrapper[4971]: I1213 07:15:15.433858 4971 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/38d62929-7b21-4bf7-a2ac-f31047134d54-combined-ca-bundle\") pod \"nova-cell1-novncproxy-0\" (UID: \"38d62929-7b21-4bf7-a2ac-f31047134d54\") " pod="openstack/nova-cell1-novncproxy-0" Dec 13 07:15:15 crc kubenswrapper[4971]: I1213 07:15:15.433970 4971 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-hmn5x\" (UniqueName: \"kubernetes.io/projected/38d62929-7b21-4bf7-a2ac-f31047134d54-kube-api-access-hmn5x\") pod \"nova-cell1-novncproxy-0\" (UID: \"38d62929-7b21-4bf7-a2ac-f31047134d54\") " pod="openstack/nova-cell1-novncproxy-0" Dec 13 07:15:15 crc kubenswrapper[4971]: I1213 07:15:15.434102 4971 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"vencrypt-tls-certs\" (UniqueName: \"kubernetes.io/secret/38d62929-7b21-4bf7-a2ac-f31047134d54-vencrypt-tls-certs\") pod \"nova-cell1-novncproxy-0\" (UID: \"38d62929-7b21-4bf7-a2ac-f31047134d54\") " pod="openstack/nova-cell1-novncproxy-0" Dec 13 07:15:15 crc kubenswrapper[4971]: I1213 07:15:15.434140 4971 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nova-novncproxy-tls-certs\" (UniqueName: \"kubernetes.io/secret/38d62929-7b21-4bf7-a2ac-f31047134d54-nova-novncproxy-tls-certs\") pod \"nova-cell1-novncproxy-0\" (UID: \"38d62929-7b21-4bf7-a2ac-f31047134d54\") " pod="openstack/nova-cell1-novncproxy-0" Dec 13 07:15:15 crc kubenswrapper[4971]: I1213 07:15:15.439541 4971 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"vencrypt-tls-certs\" (UniqueName: \"kubernetes.io/secret/38d62929-7b21-4bf7-a2ac-f31047134d54-vencrypt-tls-certs\") pod \"nova-cell1-novncproxy-0\" (UID: \"38d62929-7b21-4bf7-a2ac-f31047134d54\") " pod="openstack/nova-cell1-novncproxy-0" Dec 13 07:15:15 crc kubenswrapper[4971]: I1213 07:15:15.440079 4971 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/38d62929-7b21-4bf7-a2ac-f31047134d54-combined-ca-bundle\") pod \"nova-cell1-novncproxy-0\" (UID: \"38d62929-7b21-4bf7-a2ac-f31047134d54\") " pod="openstack/nova-cell1-novncproxy-0" Dec 13 07:15:15 crc kubenswrapper[4971]: I1213 07:15:15.447315 4971 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"nova-novncproxy-tls-certs\" (UniqueName: \"kubernetes.io/secret/38d62929-7b21-4bf7-a2ac-f31047134d54-nova-novncproxy-tls-certs\") pod \"nova-cell1-novncproxy-0\" (UID: \"38d62929-7b21-4bf7-a2ac-f31047134d54\") " pod="openstack/nova-cell1-novncproxy-0" Dec 13 07:15:15 crc kubenswrapper[4971]: I1213 07:15:15.451438 4971 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/38d62929-7b21-4bf7-a2ac-f31047134d54-config-data\") pod \"nova-cell1-novncproxy-0\" (UID: \"38d62929-7b21-4bf7-a2ac-f31047134d54\") " pod="openstack/nova-cell1-novncproxy-0" Dec 13 07:15:15 crc kubenswrapper[4971]: I1213 07:15:15.452590 4971 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-hmn5x\" (UniqueName: \"kubernetes.io/projected/38d62929-7b21-4bf7-a2ac-f31047134d54-kube-api-access-hmn5x\") pod \"nova-cell1-novncproxy-0\" (UID: \"38d62929-7b21-4bf7-a2ac-f31047134d54\") " pod="openstack/nova-cell1-novncproxy-0" Dec 13 07:15:15 crc kubenswrapper[4971]: I1213 07:15:15.576487 4971 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-novncproxy-0" Dec 13 07:15:15 crc kubenswrapper[4971]: I1213 07:15:15.782989 4971 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="9bb1801a-532f-4b09-9ae3-d4b61e8a9c97" path="/var/lib/kubelet/pods/9bb1801a-532f-4b09-9ae3-d4b61e8a9c97/volumes" Dec 13 07:15:15 crc kubenswrapper[4971]: I1213 07:15:15.849801 4971 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell1-novncproxy-0"] Dec 13 07:15:15 crc kubenswrapper[4971]: W1213 07:15:15.852214 4971 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod38d62929_7b21_4bf7_a2ac_f31047134d54.slice/crio-41424a9f2163b3f449dbc3d951b1668e6501957ea2795f476a6ae7a3ef8007d8 WatchSource:0}: Error finding container 41424a9f2163b3f449dbc3d951b1668e6501957ea2795f476a6ae7a3ef8007d8: Status 404 returned error can't find the container with id 41424a9f2163b3f449dbc3d951b1668e6501957ea2795f476a6ae7a3ef8007d8 Dec 13 07:15:16 crc kubenswrapper[4971]: I1213 07:15:16.191407 4971 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-novncproxy-0" event={"ID":"38d62929-7b21-4bf7-a2ac-f31047134d54","Type":"ContainerStarted","Data":"dce01ab60f91aa07750de20c40f43a4a6d48e5a0cc4cbbeb8efe31e20a159399"} Dec 13 07:15:16 crc kubenswrapper[4971]: I1213 07:15:16.191742 4971 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-novncproxy-0" event={"ID":"38d62929-7b21-4bf7-a2ac-f31047134d54","Type":"ContainerStarted","Data":"41424a9f2163b3f449dbc3d951b1668e6501957ea2795f476a6ae7a3ef8007d8"} Dec 13 07:15:17 crc kubenswrapper[4971]: I1213 07:15:17.739830 4971 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/nova-api-0" Dec 13 07:15:17 crc kubenswrapper[4971]: I1213 07:15:17.740916 4971 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-api-0" Dec 13 07:15:17 crc kubenswrapper[4971]: I1213 07:15:17.741610 4971 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/nova-api-0" Dec 13 07:15:17 crc kubenswrapper[4971]: I1213 07:15:17.744749 4971 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/nova-api-0" Dec 13 07:15:17 crc kubenswrapper[4971]: I1213 07:15:17.781284 4971 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-cell1-novncproxy-0" podStartSLOduration=2.781252066 podStartE2EDuration="2.781252066s" podCreationTimestamp="2025-12-13 07:15:15 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-13 07:15:16.214481173 +0000 UTC m=+1572.818890611" watchObservedRunningTime="2025-12-13 07:15:17.781252066 +0000 UTC m=+1574.385661554" Dec 13 07:15:18 crc kubenswrapper[4971]: I1213 07:15:18.212807 4971 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-api-0" Dec 13 07:15:18 crc kubenswrapper[4971]: I1213 07:15:18.216788 4971 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/nova-api-0" Dec 13 07:15:18 crc kubenswrapper[4971]: I1213 07:15:18.420470 4971 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-78c596d7cf-9jdzl"] Dec 13 07:15:18 crc kubenswrapper[4971]: I1213 07:15:18.422846 4971 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-78c596d7cf-9jdzl" Dec 13 07:15:18 crc kubenswrapper[4971]: I1213 07:15:18.447069 4971 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-78c596d7cf-9jdzl"] Dec 13 07:15:18 crc kubenswrapper[4971]: I1213 07:15:18.525313 4971 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/aadeb784-643a-4e78-909c-4a167bfe11eb-dns-svc\") pod \"dnsmasq-dns-78c596d7cf-9jdzl\" (UID: \"aadeb784-643a-4e78-909c-4a167bfe11eb\") " pod="openstack/dnsmasq-dns-78c596d7cf-9jdzl" Dec 13 07:15:18 crc kubenswrapper[4971]: I1213 07:15:18.525368 4971 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-7jkrp\" (UniqueName: \"kubernetes.io/projected/aadeb784-643a-4e78-909c-4a167bfe11eb-kube-api-access-7jkrp\") pod \"dnsmasq-dns-78c596d7cf-9jdzl\" (UID: \"aadeb784-643a-4e78-909c-4a167bfe11eb\") " pod="openstack/dnsmasq-dns-78c596d7cf-9jdzl" Dec 13 07:15:18 crc kubenswrapper[4971]: I1213 07:15:18.525411 4971 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/aadeb784-643a-4e78-909c-4a167bfe11eb-ovsdbserver-sb\") pod \"dnsmasq-dns-78c596d7cf-9jdzl\" (UID: \"aadeb784-643a-4e78-909c-4a167bfe11eb\") " pod="openstack/dnsmasq-dns-78c596d7cf-9jdzl" Dec 13 07:15:18 crc kubenswrapper[4971]: I1213 07:15:18.525501 4971 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/aadeb784-643a-4e78-909c-4a167bfe11eb-config\") pod \"dnsmasq-dns-78c596d7cf-9jdzl\" (UID: \"aadeb784-643a-4e78-909c-4a167bfe11eb\") " pod="openstack/dnsmasq-dns-78c596d7cf-9jdzl" Dec 13 07:15:18 crc kubenswrapper[4971]: I1213 07:15:18.525595 4971 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/aadeb784-643a-4e78-909c-4a167bfe11eb-ovsdbserver-nb\") pod \"dnsmasq-dns-78c596d7cf-9jdzl\" (UID: \"aadeb784-643a-4e78-909c-4a167bfe11eb\") " pod="openstack/dnsmasq-dns-78c596d7cf-9jdzl" Dec 13 07:15:18 crc kubenswrapper[4971]: I1213 07:15:18.627266 4971 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/aadeb784-643a-4e78-909c-4a167bfe11eb-dns-svc\") pod \"dnsmasq-dns-78c596d7cf-9jdzl\" (UID: \"aadeb784-643a-4e78-909c-4a167bfe11eb\") " pod="openstack/dnsmasq-dns-78c596d7cf-9jdzl" Dec 13 07:15:18 crc kubenswrapper[4971]: I1213 07:15:18.627307 4971 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-7jkrp\" (UniqueName: \"kubernetes.io/projected/aadeb784-643a-4e78-909c-4a167bfe11eb-kube-api-access-7jkrp\") pod \"dnsmasq-dns-78c596d7cf-9jdzl\" (UID: \"aadeb784-643a-4e78-909c-4a167bfe11eb\") " pod="openstack/dnsmasq-dns-78c596d7cf-9jdzl" Dec 13 07:15:18 crc kubenswrapper[4971]: I1213 07:15:18.627346 4971 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/aadeb784-643a-4e78-909c-4a167bfe11eb-ovsdbserver-sb\") pod \"dnsmasq-dns-78c596d7cf-9jdzl\" (UID: \"aadeb784-643a-4e78-909c-4a167bfe11eb\") " pod="openstack/dnsmasq-dns-78c596d7cf-9jdzl" Dec 13 07:15:18 crc kubenswrapper[4971]: I1213 07:15:18.627412 4971 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/aadeb784-643a-4e78-909c-4a167bfe11eb-config\") pod \"dnsmasq-dns-78c596d7cf-9jdzl\" (UID: \"aadeb784-643a-4e78-909c-4a167bfe11eb\") " pod="openstack/dnsmasq-dns-78c596d7cf-9jdzl" Dec 13 07:15:18 crc kubenswrapper[4971]: I1213 07:15:18.627466 4971 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/aadeb784-643a-4e78-909c-4a167bfe11eb-ovsdbserver-nb\") pod \"dnsmasq-dns-78c596d7cf-9jdzl\" (UID: \"aadeb784-643a-4e78-909c-4a167bfe11eb\") " pod="openstack/dnsmasq-dns-78c596d7cf-9jdzl" Dec 13 07:15:18 crc kubenswrapper[4971]: I1213 07:15:18.628602 4971 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/aadeb784-643a-4e78-909c-4a167bfe11eb-ovsdbserver-sb\") pod \"dnsmasq-dns-78c596d7cf-9jdzl\" (UID: \"aadeb784-643a-4e78-909c-4a167bfe11eb\") " pod="openstack/dnsmasq-dns-78c596d7cf-9jdzl" Dec 13 07:15:18 crc kubenswrapper[4971]: I1213 07:15:18.628643 4971 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/aadeb784-643a-4e78-909c-4a167bfe11eb-ovsdbserver-nb\") pod \"dnsmasq-dns-78c596d7cf-9jdzl\" (UID: \"aadeb784-643a-4e78-909c-4a167bfe11eb\") " pod="openstack/dnsmasq-dns-78c596d7cf-9jdzl" Dec 13 07:15:18 crc kubenswrapper[4971]: I1213 07:15:18.628650 4971 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/aadeb784-643a-4e78-909c-4a167bfe11eb-config\") pod \"dnsmasq-dns-78c596d7cf-9jdzl\" (UID: \"aadeb784-643a-4e78-909c-4a167bfe11eb\") " pod="openstack/dnsmasq-dns-78c596d7cf-9jdzl" Dec 13 07:15:18 crc kubenswrapper[4971]: I1213 07:15:18.629237 4971 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/aadeb784-643a-4e78-909c-4a167bfe11eb-dns-svc\") pod \"dnsmasq-dns-78c596d7cf-9jdzl\" (UID: \"aadeb784-643a-4e78-909c-4a167bfe11eb\") " pod="openstack/dnsmasq-dns-78c596d7cf-9jdzl" Dec 13 07:15:18 crc kubenswrapper[4971]: I1213 07:15:18.652221 4971 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-7jkrp\" (UniqueName: \"kubernetes.io/projected/aadeb784-643a-4e78-909c-4a167bfe11eb-kube-api-access-7jkrp\") pod \"dnsmasq-dns-78c596d7cf-9jdzl\" (UID: \"aadeb784-643a-4e78-909c-4a167bfe11eb\") " pod="openstack/dnsmasq-dns-78c596d7cf-9jdzl" Dec 13 07:15:18 crc kubenswrapper[4971]: I1213 07:15:18.770341 4971 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-78c596d7cf-9jdzl" Dec 13 07:15:19 crc kubenswrapper[4971]: I1213 07:15:19.252480 4971 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-78c596d7cf-9jdzl"] Dec 13 07:15:19 crc kubenswrapper[4971]: W1213 07:15:19.255193 4971 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podaadeb784_643a_4e78_909c_4a167bfe11eb.slice/crio-39b7203f122b7ff1e610263e4bb3c7a4f458ce9c2a599f2d6c594aff255c3a77 WatchSource:0}: Error finding container 39b7203f122b7ff1e610263e4bb3c7a4f458ce9c2a599f2d6c594aff255c3a77: Status 404 returned error can't find the container with id 39b7203f122b7ff1e610263e4bb3c7a4f458ce9c2a599f2d6c594aff255c3a77 Dec 13 07:15:20 crc kubenswrapper[4971]: I1213 07:15:20.229788 4971 generic.go:334] "Generic (PLEG): container finished" podID="aadeb784-643a-4e78-909c-4a167bfe11eb" containerID="7326cf828fc3f8902b2798804319ee20d2885e7a77300034607853426b6231d1" exitCode=0 Dec 13 07:15:20 crc kubenswrapper[4971]: I1213 07:15:20.229899 4971 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-78c596d7cf-9jdzl" event={"ID":"aadeb784-643a-4e78-909c-4a167bfe11eb","Type":"ContainerDied","Data":"7326cf828fc3f8902b2798804319ee20d2885e7a77300034607853426b6231d1"} Dec 13 07:15:20 crc kubenswrapper[4971]: I1213 07:15:20.230407 4971 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-78c596d7cf-9jdzl" event={"ID":"aadeb784-643a-4e78-909c-4a167bfe11eb","Type":"ContainerStarted","Data":"39b7203f122b7ff1e610263e4bb3c7a4f458ce9c2a599f2d6c594aff255c3a77"} Dec 13 07:15:20 crc kubenswrapper[4971]: I1213 07:15:20.576887 4971 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-cell1-novncproxy-0" Dec 13 07:15:20 crc kubenswrapper[4971]: I1213 07:15:20.705741 4971 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ceilometer-0"] Dec 13 07:15:20 crc kubenswrapper[4971]: I1213 07:15:20.706082 4971 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="18ea016d-26bf-46a9-a8c3-2f555643e967" containerName="ceilometer-central-agent" containerID="cri-o://fb55d70eea8aa0f44521717d475539a24d67f5b76a554632aa7eea6670d42a3d" gracePeriod=30 Dec 13 07:15:20 crc kubenswrapper[4971]: I1213 07:15:20.706182 4971 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="18ea016d-26bf-46a9-a8c3-2f555643e967" containerName="sg-core" containerID="cri-o://2d7101e713a8a556bd60407784326637a7599d4c2fe8201a04275e953f9be12f" gracePeriod=30 Dec 13 07:15:20 crc kubenswrapper[4971]: I1213 07:15:20.706298 4971 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="18ea016d-26bf-46a9-a8c3-2f555643e967" containerName="proxy-httpd" containerID="cri-o://9894e8db28cfff1cb5f97d08464004fdc4d8d729765e4dfce4f19b760d82bc9c" gracePeriod=30 Dec 13 07:15:20 crc kubenswrapper[4971]: I1213 07:15:20.706328 4971 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="18ea016d-26bf-46a9-a8c3-2f555643e967" containerName="ceilometer-notification-agent" containerID="cri-o://e6e2bf1726974affec4fe8a70d50136677acd3da1008bacecd28f89bc4e2586c" gracePeriod=30 Dec 13 07:15:20 crc kubenswrapper[4971]: I1213 07:15:20.888485 4971 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-api-0"] Dec 13 07:15:21 crc kubenswrapper[4971]: I1213 07:15:21.242013 4971 generic.go:334] "Generic (PLEG): container finished" podID="18ea016d-26bf-46a9-a8c3-2f555643e967" containerID="9894e8db28cfff1cb5f97d08464004fdc4d8d729765e4dfce4f19b760d82bc9c" exitCode=0 Dec 13 07:15:21 crc kubenswrapper[4971]: I1213 07:15:21.242045 4971 generic.go:334] "Generic (PLEG): container finished" podID="18ea016d-26bf-46a9-a8c3-2f555643e967" containerID="2d7101e713a8a556bd60407784326637a7599d4c2fe8201a04275e953f9be12f" exitCode=2 Dec 13 07:15:21 crc kubenswrapper[4971]: I1213 07:15:21.242055 4971 generic.go:334] "Generic (PLEG): container finished" podID="18ea016d-26bf-46a9-a8c3-2f555643e967" containerID="fb55d70eea8aa0f44521717d475539a24d67f5b76a554632aa7eea6670d42a3d" exitCode=0 Dec 13 07:15:21 crc kubenswrapper[4971]: I1213 07:15:21.242116 4971 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"18ea016d-26bf-46a9-a8c3-2f555643e967","Type":"ContainerDied","Data":"9894e8db28cfff1cb5f97d08464004fdc4d8d729765e4dfce4f19b760d82bc9c"} Dec 13 07:15:21 crc kubenswrapper[4971]: I1213 07:15:21.242180 4971 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"18ea016d-26bf-46a9-a8c3-2f555643e967","Type":"ContainerDied","Data":"2d7101e713a8a556bd60407784326637a7599d4c2fe8201a04275e953f9be12f"} Dec 13 07:15:21 crc kubenswrapper[4971]: I1213 07:15:21.242204 4971 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"18ea016d-26bf-46a9-a8c3-2f555643e967","Type":"ContainerDied","Data":"fb55d70eea8aa0f44521717d475539a24d67f5b76a554632aa7eea6670d42a3d"} Dec 13 07:15:21 crc kubenswrapper[4971]: I1213 07:15:21.246731 4971 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-78c596d7cf-9jdzl" event={"ID":"aadeb784-643a-4e78-909c-4a167bfe11eb","Type":"ContainerStarted","Data":"5de93625f93b3f1510893ddd803bdb0ed482aa9dedb3ae4499ae9c0394002778"} Dec 13 07:15:21 crc kubenswrapper[4971]: I1213 07:15:21.246853 4971 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-api-0" podUID="57f1d9a5-cc22-419a-a944-ec9581785de2" containerName="nova-api-log" containerID="cri-o://3679a59578f7ed86541756cdb8f478aa439e64cf5a03668c1353a452c6e1b989" gracePeriod=30 Dec 13 07:15:21 crc kubenswrapper[4971]: I1213 07:15:21.246937 4971 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/dnsmasq-dns-78c596d7cf-9jdzl" Dec 13 07:15:21 crc kubenswrapper[4971]: I1213 07:15:21.247040 4971 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-api-0" podUID="57f1d9a5-cc22-419a-a944-ec9581785de2" containerName="nova-api-api" containerID="cri-o://c67d150ca7039f2ee1dcb3af173c2b626da09aa4a1cc47532a65a415b3fc9f05" gracePeriod=30 Dec 13 07:15:21 crc kubenswrapper[4971]: I1213 07:15:21.271778 4971 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/dnsmasq-dns-78c596d7cf-9jdzl" podStartSLOduration=3.271760272 podStartE2EDuration="3.271760272s" podCreationTimestamp="2025-12-13 07:15:18 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-13 07:15:21.271035934 +0000 UTC m=+1577.875445382" watchObservedRunningTime="2025-12-13 07:15:21.271760272 +0000 UTC m=+1577.876169730" Dec 13 07:15:22 crc kubenswrapper[4971]: I1213 07:15:22.258335 4971 generic.go:334] "Generic (PLEG): container finished" podID="57f1d9a5-cc22-419a-a944-ec9581785de2" containerID="3679a59578f7ed86541756cdb8f478aa439e64cf5a03668c1353a452c6e1b989" exitCode=143 Dec 13 07:15:22 crc kubenswrapper[4971]: I1213 07:15:22.259554 4971 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"57f1d9a5-cc22-419a-a944-ec9581785de2","Type":"ContainerDied","Data":"3679a59578f7ed86541756cdb8f478aa439e64cf5a03668c1353a452c6e1b989"} Dec 13 07:15:24 crc kubenswrapper[4971]: I1213 07:15:24.857330 4971 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-0" Dec 13 07:15:24 crc kubenswrapper[4971]: I1213 07:15:24.959838 4971 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-2r8wv\" (UniqueName: \"kubernetes.io/projected/57f1d9a5-cc22-419a-a944-ec9581785de2-kube-api-access-2r8wv\") pod \"57f1d9a5-cc22-419a-a944-ec9581785de2\" (UID: \"57f1d9a5-cc22-419a-a944-ec9581785de2\") " Dec 13 07:15:24 crc kubenswrapper[4971]: I1213 07:15:24.959940 4971 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/57f1d9a5-cc22-419a-a944-ec9581785de2-combined-ca-bundle\") pod \"57f1d9a5-cc22-419a-a944-ec9581785de2\" (UID: \"57f1d9a5-cc22-419a-a944-ec9581785de2\") " Dec 13 07:15:24 crc kubenswrapper[4971]: I1213 07:15:24.960550 4971 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/57f1d9a5-cc22-419a-a944-ec9581785de2-logs" (OuterVolumeSpecName: "logs") pod "57f1d9a5-cc22-419a-a944-ec9581785de2" (UID: "57f1d9a5-cc22-419a-a944-ec9581785de2"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 13 07:15:24 crc kubenswrapper[4971]: I1213 07:15:24.959976 4971 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/57f1d9a5-cc22-419a-a944-ec9581785de2-logs\") pod \"57f1d9a5-cc22-419a-a944-ec9581785de2\" (UID: \"57f1d9a5-cc22-419a-a944-ec9581785de2\") " Dec 13 07:15:24 crc kubenswrapper[4971]: I1213 07:15:24.960902 4971 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/57f1d9a5-cc22-419a-a944-ec9581785de2-config-data\") pod \"57f1d9a5-cc22-419a-a944-ec9581785de2\" (UID: \"57f1d9a5-cc22-419a-a944-ec9581785de2\") " Dec 13 07:15:24 crc kubenswrapper[4971]: I1213 07:15:24.961713 4971 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/57f1d9a5-cc22-419a-a944-ec9581785de2-logs\") on node \"crc\" DevicePath \"\"" Dec 13 07:15:24 crc kubenswrapper[4971]: I1213 07:15:24.966032 4971 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/57f1d9a5-cc22-419a-a944-ec9581785de2-kube-api-access-2r8wv" (OuterVolumeSpecName: "kube-api-access-2r8wv") pod "57f1d9a5-cc22-419a-a944-ec9581785de2" (UID: "57f1d9a5-cc22-419a-a944-ec9581785de2"). InnerVolumeSpecName "kube-api-access-2r8wv". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 13 07:15:24 crc kubenswrapper[4971]: I1213 07:15:24.992772 4971 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/57f1d9a5-cc22-419a-a944-ec9581785de2-config-data" (OuterVolumeSpecName: "config-data") pod "57f1d9a5-cc22-419a-a944-ec9581785de2" (UID: "57f1d9a5-cc22-419a-a944-ec9581785de2"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 13 07:15:25 crc kubenswrapper[4971]: I1213 07:15:25.010578 4971 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/57f1d9a5-cc22-419a-a944-ec9581785de2-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "57f1d9a5-cc22-419a-a944-ec9581785de2" (UID: "57f1d9a5-cc22-419a-a944-ec9581785de2"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 13 07:15:25 crc kubenswrapper[4971]: I1213 07:15:25.064229 4971 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-2r8wv\" (UniqueName: \"kubernetes.io/projected/57f1d9a5-cc22-419a-a944-ec9581785de2-kube-api-access-2r8wv\") on node \"crc\" DevicePath \"\"" Dec 13 07:15:25 crc kubenswrapper[4971]: I1213 07:15:25.155377 4971 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/57f1d9a5-cc22-419a-a944-ec9581785de2-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 13 07:15:25 crc kubenswrapper[4971]: I1213 07:15:25.155762 4971 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/57f1d9a5-cc22-419a-a944-ec9581785de2-config-data\") on node \"crc\" DevicePath \"\"" Dec 13 07:15:25 crc kubenswrapper[4971]: I1213 07:15:25.284876 4971 generic.go:334] "Generic (PLEG): container finished" podID="57f1d9a5-cc22-419a-a944-ec9581785de2" containerID="c67d150ca7039f2ee1dcb3af173c2b626da09aa4a1cc47532a65a415b3fc9f05" exitCode=0 Dec 13 07:15:25 crc kubenswrapper[4971]: I1213 07:15:25.284916 4971 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"57f1d9a5-cc22-419a-a944-ec9581785de2","Type":"ContainerDied","Data":"c67d150ca7039f2ee1dcb3af173c2b626da09aa4a1cc47532a65a415b3fc9f05"} Dec 13 07:15:25 crc kubenswrapper[4971]: I1213 07:15:25.284938 4971 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-0" Dec 13 07:15:25 crc kubenswrapper[4971]: I1213 07:15:25.284952 4971 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"57f1d9a5-cc22-419a-a944-ec9581785de2","Type":"ContainerDied","Data":"f36843c1785b95b0e6501feea2b88005e62fc386d4de84c81d1b3a896e580731"} Dec 13 07:15:25 crc kubenswrapper[4971]: I1213 07:15:25.284999 4971 scope.go:117] "RemoveContainer" containerID="c67d150ca7039f2ee1dcb3af173c2b626da09aa4a1cc47532a65a415b3fc9f05" Dec 13 07:15:25 crc kubenswrapper[4971]: I1213 07:15:25.319236 4971 scope.go:117] "RemoveContainer" containerID="3679a59578f7ed86541756cdb8f478aa439e64cf5a03668c1353a452c6e1b989" Dec 13 07:15:25 crc kubenswrapper[4971]: I1213 07:15:25.326846 4971 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-api-0"] Dec 13 07:15:25 crc kubenswrapper[4971]: I1213 07:15:25.341556 4971 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-api-0"] Dec 13 07:15:25 crc kubenswrapper[4971]: I1213 07:15:25.344157 4971 scope.go:117] "RemoveContainer" containerID="c67d150ca7039f2ee1dcb3af173c2b626da09aa4a1cc47532a65a415b3fc9f05" Dec 13 07:15:25 crc kubenswrapper[4971]: E1213 07:15:25.344694 4971 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"c67d150ca7039f2ee1dcb3af173c2b626da09aa4a1cc47532a65a415b3fc9f05\": container with ID starting with c67d150ca7039f2ee1dcb3af173c2b626da09aa4a1cc47532a65a415b3fc9f05 not found: ID does not exist" containerID="c67d150ca7039f2ee1dcb3af173c2b626da09aa4a1cc47532a65a415b3fc9f05" Dec 13 07:15:25 crc kubenswrapper[4971]: I1213 07:15:25.344733 4971 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"c67d150ca7039f2ee1dcb3af173c2b626da09aa4a1cc47532a65a415b3fc9f05"} err="failed to get container status \"c67d150ca7039f2ee1dcb3af173c2b626da09aa4a1cc47532a65a415b3fc9f05\": rpc error: code = NotFound desc = could not find container \"c67d150ca7039f2ee1dcb3af173c2b626da09aa4a1cc47532a65a415b3fc9f05\": container with ID starting with c67d150ca7039f2ee1dcb3af173c2b626da09aa4a1cc47532a65a415b3fc9f05 not found: ID does not exist" Dec 13 07:15:25 crc kubenswrapper[4971]: I1213 07:15:25.344760 4971 scope.go:117] "RemoveContainer" containerID="3679a59578f7ed86541756cdb8f478aa439e64cf5a03668c1353a452c6e1b989" Dec 13 07:15:25 crc kubenswrapper[4971]: E1213 07:15:25.344999 4971 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"3679a59578f7ed86541756cdb8f478aa439e64cf5a03668c1353a452c6e1b989\": container with ID starting with 3679a59578f7ed86541756cdb8f478aa439e64cf5a03668c1353a452c6e1b989 not found: ID does not exist" containerID="3679a59578f7ed86541756cdb8f478aa439e64cf5a03668c1353a452c6e1b989" Dec 13 07:15:25 crc kubenswrapper[4971]: I1213 07:15:25.345035 4971 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"3679a59578f7ed86541756cdb8f478aa439e64cf5a03668c1353a452c6e1b989"} err="failed to get container status \"3679a59578f7ed86541756cdb8f478aa439e64cf5a03668c1353a452c6e1b989\": rpc error: code = NotFound desc = could not find container \"3679a59578f7ed86541756cdb8f478aa439e64cf5a03668c1353a452c6e1b989\": container with ID starting with 3679a59578f7ed86541756cdb8f478aa439e64cf5a03668c1353a452c6e1b989 not found: ID does not exist" Dec 13 07:15:25 crc kubenswrapper[4971]: I1213 07:15:25.351577 4971 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-api-0"] Dec 13 07:15:25 crc kubenswrapper[4971]: E1213 07:15:25.351979 4971 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="57f1d9a5-cc22-419a-a944-ec9581785de2" containerName="nova-api-api" Dec 13 07:15:25 crc kubenswrapper[4971]: I1213 07:15:25.351997 4971 state_mem.go:107] "Deleted CPUSet assignment" podUID="57f1d9a5-cc22-419a-a944-ec9581785de2" containerName="nova-api-api" Dec 13 07:15:25 crc kubenswrapper[4971]: E1213 07:15:25.352024 4971 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="57f1d9a5-cc22-419a-a944-ec9581785de2" containerName="nova-api-log" Dec 13 07:15:25 crc kubenswrapper[4971]: I1213 07:15:25.352030 4971 state_mem.go:107] "Deleted CPUSet assignment" podUID="57f1d9a5-cc22-419a-a944-ec9581785de2" containerName="nova-api-log" Dec 13 07:15:25 crc kubenswrapper[4971]: I1213 07:15:25.352174 4971 memory_manager.go:354] "RemoveStaleState removing state" podUID="57f1d9a5-cc22-419a-a944-ec9581785de2" containerName="nova-api-log" Dec 13 07:15:25 crc kubenswrapper[4971]: I1213 07:15:25.352201 4971 memory_manager.go:354] "RemoveStaleState removing state" podUID="57f1d9a5-cc22-419a-a944-ec9581785de2" containerName="nova-api-api" Dec 13 07:15:25 crc kubenswrapper[4971]: I1213 07:15:25.353219 4971 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-0" Dec 13 07:15:25 crc kubenswrapper[4971]: I1213 07:15:25.356012 4971 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-nova-internal-svc" Dec 13 07:15:25 crc kubenswrapper[4971]: I1213 07:15:25.356056 4971 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-nova-public-svc" Dec 13 07:15:25 crc kubenswrapper[4971]: I1213 07:15:25.356340 4971 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-api-config-data" Dec 13 07:15:25 crc kubenswrapper[4971]: I1213 07:15:25.359953 4971 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-api-0"] Dec 13 07:15:25 crc kubenswrapper[4971]: I1213 07:15:25.461133 4971 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-t747v\" (UniqueName: \"kubernetes.io/projected/ce91d492-58fe-48a0-baa1-e26e736ab49a-kube-api-access-t747v\") pod \"nova-api-0\" (UID: \"ce91d492-58fe-48a0-baa1-e26e736ab49a\") " pod="openstack/nova-api-0" Dec 13 07:15:25 crc kubenswrapper[4971]: I1213 07:15:25.461193 4971 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/ce91d492-58fe-48a0-baa1-e26e736ab49a-internal-tls-certs\") pod \"nova-api-0\" (UID: \"ce91d492-58fe-48a0-baa1-e26e736ab49a\") " pod="openstack/nova-api-0" Dec 13 07:15:25 crc kubenswrapper[4971]: I1213 07:15:25.461215 4971 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/ce91d492-58fe-48a0-baa1-e26e736ab49a-logs\") pod \"nova-api-0\" (UID: \"ce91d492-58fe-48a0-baa1-e26e736ab49a\") " pod="openstack/nova-api-0" Dec 13 07:15:25 crc kubenswrapper[4971]: I1213 07:15:25.461247 4971 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/ce91d492-58fe-48a0-baa1-e26e736ab49a-combined-ca-bundle\") pod \"nova-api-0\" (UID: \"ce91d492-58fe-48a0-baa1-e26e736ab49a\") " pod="openstack/nova-api-0" Dec 13 07:15:25 crc kubenswrapper[4971]: I1213 07:15:25.461428 4971 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/ce91d492-58fe-48a0-baa1-e26e736ab49a-public-tls-certs\") pod \"nova-api-0\" (UID: \"ce91d492-58fe-48a0-baa1-e26e736ab49a\") " pod="openstack/nova-api-0" Dec 13 07:15:25 crc kubenswrapper[4971]: I1213 07:15:25.461541 4971 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/ce91d492-58fe-48a0-baa1-e26e736ab49a-config-data\") pod \"nova-api-0\" (UID: \"ce91d492-58fe-48a0-baa1-e26e736ab49a\") " pod="openstack/nova-api-0" Dec 13 07:15:25 crc kubenswrapper[4971]: I1213 07:15:25.628709 4971 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/nova-cell1-novncproxy-0" Dec 13 07:15:25 crc kubenswrapper[4971]: I1213 07:15:25.629215 4971 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/ce91d492-58fe-48a0-baa1-e26e736ab49a-public-tls-certs\") pod \"nova-api-0\" (UID: \"ce91d492-58fe-48a0-baa1-e26e736ab49a\") " pod="openstack/nova-api-0" Dec 13 07:15:25 crc kubenswrapper[4971]: I1213 07:15:25.629294 4971 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/ce91d492-58fe-48a0-baa1-e26e736ab49a-config-data\") pod \"nova-api-0\" (UID: \"ce91d492-58fe-48a0-baa1-e26e736ab49a\") " pod="openstack/nova-api-0" Dec 13 07:15:25 crc kubenswrapper[4971]: I1213 07:15:25.629368 4971 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-t747v\" (UniqueName: \"kubernetes.io/projected/ce91d492-58fe-48a0-baa1-e26e736ab49a-kube-api-access-t747v\") pod \"nova-api-0\" (UID: \"ce91d492-58fe-48a0-baa1-e26e736ab49a\") " pod="openstack/nova-api-0" Dec 13 07:15:25 crc kubenswrapper[4971]: I1213 07:15:25.629419 4971 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/ce91d492-58fe-48a0-baa1-e26e736ab49a-internal-tls-certs\") pod \"nova-api-0\" (UID: \"ce91d492-58fe-48a0-baa1-e26e736ab49a\") " pod="openstack/nova-api-0" Dec 13 07:15:25 crc kubenswrapper[4971]: I1213 07:15:25.629443 4971 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/ce91d492-58fe-48a0-baa1-e26e736ab49a-logs\") pod \"nova-api-0\" (UID: \"ce91d492-58fe-48a0-baa1-e26e736ab49a\") " pod="openstack/nova-api-0" Dec 13 07:15:25 crc kubenswrapper[4971]: I1213 07:15:25.629476 4971 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/ce91d492-58fe-48a0-baa1-e26e736ab49a-combined-ca-bundle\") pod \"nova-api-0\" (UID: \"ce91d492-58fe-48a0-baa1-e26e736ab49a\") " pod="openstack/nova-api-0" Dec 13 07:15:25 crc kubenswrapper[4971]: I1213 07:15:25.630319 4971 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/ce91d492-58fe-48a0-baa1-e26e736ab49a-logs\") pod \"nova-api-0\" (UID: \"ce91d492-58fe-48a0-baa1-e26e736ab49a\") " pod="openstack/nova-api-0" Dec 13 07:15:25 crc kubenswrapper[4971]: I1213 07:15:25.634463 4971 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/ce91d492-58fe-48a0-baa1-e26e736ab49a-combined-ca-bundle\") pod \"nova-api-0\" (UID: \"ce91d492-58fe-48a0-baa1-e26e736ab49a\") " pod="openstack/nova-api-0" Dec 13 07:15:25 crc kubenswrapper[4971]: I1213 07:15:25.634894 4971 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/ce91d492-58fe-48a0-baa1-e26e736ab49a-public-tls-certs\") pod \"nova-api-0\" (UID: \"ce91d492-58fe-48a0-baa1-e26e736ab49a\") " pod="openstack/nova-api-0" Dec 13 07:15:25 crc kubenswrapper[4971]: I1213 07:15:25.635869 4971 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/ce91d492-58fe-48a0-baa1-e26e736ab49a-internal-tls-certs\") pod \"nova-api-0\" (UID: \"ce91d492-58fe-48a0-baa1-e26e736ab49a\") " pod="openstack/nova-api-0" Dec 13 07:15:25 crc kubenswrapper[4971]: I1213 07:15:25.640308 4971 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/ce91d492-58fe-48a0-baa1-e26e736ab49a-config-data\") pod \"nova-api-0\" (UID: \"ce91d492-58fe-48a0-baa1-e26e736ab49a\") " pod="openstack/nova-api-0" Dec 13 07:15:25 crc kubenswrapper[4971]: I1213 07:15:25.651068 4971 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-t747v\" (UniqueName: \"kubernetes.io/projected/ce91d492-58fe-48a0-baa1-e26e736ab49a-kube-api-access-t747v\") pod \"nova-api-0\" (UID: \"ce91d492-58fe-48a0-baa1-e26e736ab49a\") " pod="openstack/nova-api-0" Dec 13 07:15:25 crc kubenswrapper[4971]: I1213 07:15:25.654945 4971 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/nova-cell1-novncproxy-0" Dec 13 07:15:25 crc kubenswrapper[4971]: I1213 07:15:25.708488 4971 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-0" Dec 13 07:15:25 crc kubenswrapper[4971]: I1213 07:15:25.773827 4971 scope.go:117] "RemoveContainer" containerID="e8beee43d15719dccb6892d6f5711674492425300efa098be89033bc65977d14" Dec 13 07:15:25 crc kubenswrapper[4971]: E1213 07:15:25.774121 4971 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-82xjz_openshift-machine-config-operator(e0c4af12-edda-41d7-9b44-f87396174ef2)\"" pod="openshift-machine-config-operator/machine-config-daemon-82xjz" podUID="e0c4af12-edda-41d7-9b44-f87396174ef2" Dec 13 07:15:25 crc kubenswrapper[4971]: I1213 07:15:25.786892 4971 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="57f1d9a5-cc22-419a-a944-ec9581785de2" path="/var/lib/kubelet/pods/57f1d9a5-cc22-419a-a944-ec9581785de2/volumes" Dec 13 07:15:26 crc kubenswrapper[4971]: I1213 07:15:26.251971 4971 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-api-0"] Dec 13 07:15:26 crc kubenswrapper[4971]: I1213 07:15:26.293767 4971 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"ce91d492-58fe-48a0-baa1-e26e736ab49a","Type":"ContainerStarted","Data":"765fd8ecec45e22b725ddb6c03cc9e5def2390806e4b6f5cc5802184b5c0c0bc"} Dec 13 07:15:26 crc kubenswrapper[4971]: I1213 07:15:26.311274 4971 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/nova-cell1-novncproxy-0" Dec 13 07:15:26 crc kubenswrapper[4971]: I1213 07:15:26.484580 4971 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-cell1-cell-mapping-kqkzw"] Dec 13 07:15:26 crc kubenswrapper[4971]: I1213 07:15:26.486733 4971 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-cell-mapping-kqkzw" Dec 13 07:15:26 crc kubenswrapper[4971]: I1213 07:15:26.494880 4971 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell1-cell-mapping-kqkzw"] Dec 13 07:15:26 crc kubenswrapper[4971]: I1213 07:15:26.542402 4971 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-cell1-manage-scripts" Dec 13 07:15:26 crc kubenswrapper[4971]: I1213 07:15:26.542721 4971 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-cell1-manage-config-data" Dec 13 07:15:26 crc kubenswrapper[4971]: I1213 07:15:26.645676 4971 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-hczqd\" (UniqueName: \"kubernetes.io/projected/20d00522-0b07-416b-90b4-0b588f99c41f-kube-api-access-hczqd\") pod \"nova-cell1-cell-mapping-kqkzw\" (UID: \"20d00522-0b07-416b-90b4-0b588f99c41f\") " pod="openstack/nova-cell1-cell-mapping-kqkzw" Dec 13 07:15:26 crc kubenswrapper[4971]: I1213 07:15:26.645914 4971 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/20d00522-0b07-416b-90b4-0b588f99c41f-combined-ca-bundle\") pod \"nova-cell1-cell-mapping-kqkzw\" (UID: \"20d00522-0b07-416b-90b4-0b588f99c41f\") " pod="openstack/nova-cell1-cell-mapping-kqkzw" Dec 13 07:15:26 crc kubenswrapper[4971]: I1213 07:15:26.646075 4971 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/20d00522-0b07-416b-90b4-0b588f99c41f-scripts\") pod \"nova-cell1-cell-mapping-kqkzw\" (UID: \"20d00522-0b07-416b-90b4-0b588f99c41f\") " pod="openstack/nova-cell1-cell-mapping-kqkzw" Dec 13 07:15:26 crc kubenswrapper[4971]: I1213 07:15:26.646216 4971 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/20d00522-0b07-416b-90b4-0b588f99c41f-config-data\") pod \"nova-cell1-cell-mapping-kqkzw\" (UID: \"20d00522-0b07-416b-90b4-0b588f99c41f\") " pod="openstack/nova-cell1-cell-mapping-kqkzw" Dec 13 07:15:26 crc kubenswrapper[4971]: I1213 07:15:26.747782 4971 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-hczqd\" (UniqueName: \"kubernetes.io/projected/20d00522-0b07-416b-90b4-0b588f99c41f-kube-api-access-hczqd\") pod \"nova-cell1-cell-mapping-kqkzw\" (UID: \"20d00522-0b07-416b-90b4-0b588f99c41f\") " pod="openstack/nova-cell1-cell-mapping-kqkzw" Dec 13 07:15:26 crc kubenswrapper[4971]: I1213 07:15:26.748212 4971 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/20d00522-0b07-416b-90b4-0b588f99c41f-combined-ca-bundle\") pod \"nova-cell1-cell-mapping-kqkzw\" (UID: \"20d00522-0b07-416b-90b4-0b588f99c41f\") " pod="openstack/nova-cell1-cell-mapping-kqkzw" Dec 13 07:15:26 crc kubenswrapper[4971]: I1213 07:15:26.748319 4971 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/20d00522-0b07-416b-90b4-0b588f99c41f-scripts\") pod \"nova-cell1-cell-mapping-kqkzw\" (UID: \"20d00522-0b07-416b-90b4-0b588f99c41f\") " pod="openstack/nova-cell1-cell-mapping-kqkzw" Dec 13 07:15:26 crc kubenswrapper[4971]: I1213 07:15:26.748477 4971 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/20d00522-0b07-416b-90b4-0b588f99c41f-config-data\") pod \"nova-cell1-cell-mapping-kqkzw\" (UID: \"20d00522-0b07-416b-90b4-0b588f99c41f\") " pod="openstack/nova-cell1-cell-mapping-kqkzw" Dec 13 07:15:26 crc kubenswrapper[4971]: I1213 07:15:26.753131 4971 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/20d00522-0b07-416b-90b4-0b588f99c41f-scripts\") pod \"nova-cell1-cell-mapping-kqkzw\" (UID: \"20d00522-0b07-416b-90b4-0b588f99c41f\") " pod="openstack/nova-cell1-cell-mapping-kqkzw" Dec 13 07:15:26 crc kubenswrapper[4971]: I1213 07:15:26.753176 4971 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/20d00522-0b07-416b-90b4-0b588f99c41f-config-data\") pod \"nova-cell1-cell-mapping-kqkzw\" (UID: \"20d00522-0b07-416b-90b4-0b588f99c41f\") " pod="openstack/nova-cell1-cell-mapping-kqkzw" Dec 13 07:15:26 crc kubenswrapper[4971]: I1213 07:15:26.755489 4971 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/20d00522-0b07-416b-90b4-0b588f99c41f-combined-ca-bundle\") pod \"nova-cell1-cell-mapping-kqkzw\" (UID: \"20d00522-0b07-416b-90b4-0b588f99c41f\") " pod="openstack/nova-cell1-cell-mapping-kqkzw" Dec 13 07:15:26 crc kubenswrapper[4971]: I1213 07:15:26.769285 4971 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-hczqd\" (UniqueName: \"kubernetes.io/projected/20d00522-0b07-416b-90b4-0b588f99c41f-kube-api-access-hczqd\") pod \"nova-cell1-cell-mapping-kqkzw\" (UID: \"20d00522-0b07-416b-90b4-0b588f99c41f\") " pod="openstack/nova-cell1-cell-mapping-kqkzw" Dec 13 07:15:26 crc kubenswrapper[4971]: I1213 07:15:26.877541 4971 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-cell-mapping-kqkzw" Dec 13 07:15:27 crc kubenswrapper[4971]: I1213 07:15:27.424709 4971 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"ce91d492-58fe-48a0-baa1-e26e736ab49a","Type":"ContainerStarted","Data":"092c2442d88a2d6edbfffc8f033cfd7a1abeb2f3a58b906d2f18d96ae471ca6b"} Dec 13 07:15:27 crc kubenswrapper[4971]: I1213 07:15:27.424964 4971 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"ce91d492-58fe-48a0-baa1-e26e736ab49a","Type":"ContainerStarted","Data":"3869966112be8a7f6f89d5ea0ceb4a3a1dea2af5d13e82953998769f10e53115"} Dec 13 07:15:27 crc kubenswrapper[4971]: I1213 07:15:27.447375 4971 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-api-0" podStartSLOduration=2.447354485 podStartE2EDuration="2.447354485s" podCreationTimestamp="2025-12-13 07:15:25 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-13 07:15:27.440099957 +0000 UTC m=+1584.044509405" watchObservedRunningTime="2025-12-13 07:15:27.447354485 +0000 UTC m=+1584.051763933" Dec 13 07:15:27 crc kubenswrapper[4971]: W1213 07:15:27.461482 4971 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod20d00522_0b07_416b_90b4_0b588f99c41f.slice/crio-7a4d4193c4014c785b7952870ad36e4c6f7bc3d235e38bb4016d323802cc4f45 WatchSource:0}: Error finding container 7a4d4193c4014c785b7952870ad36e4c6f7bc3d235e38bb4016d323802cc4f45: Status 404 returned error can't find the container with id 7a4d4193c4014c785b7952870ad36e4c6f7bc3d235e38bb4016d323802cc4f45 Dec 13 07:15:27 crc kubenswrapper[4971]: I1213 07:15:27.462368 4971 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell1-cell-mapping-kqkzw"] Dec 13 07:15:28 crc kubenswrapper[4971]: I1213 07:15:28.446766 4971 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-cell-mapping-kqkzw" event={"ID":"20d00522-0b07-416b-90b4-0b588f99c41f","Type":"ContainerStarted","Data":"c993f4cf7451d6815046085857a794dcbdad51b484925bf88562a78935bb1b7f"} Dec 13 07:15:28 crc kubenswrapper[4971]: I1213 07:15:28.447239 4971 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-cell-mapping-kqkzw" event={"ID":"20d00522-0b07-416b-90b4-0b588f99c41f","Type":"ContainerStarted","Data":"7a4d4193c4014c785b7952870ad36e4c6f7bc3d235e38bb4016d323802cc4f45"} Dec 13 07:15:28 crc kubenswrapper[4971]: I1213 07:15:28.473017 4971 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-cell1-cell-mapping-kqkzw" podStartSLOduration=2.472996954 podStartE2EDuration="2.472996954s" podCreationTimestamp="2025-12-13 07:15:26 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-13 07:15:28.469943029 +0000 UTC m=+1585.074352477" watchObservedRunningTime="2025-12-13 07:15:28.472996954 +0000 UTC m=+1585.077406402" Dec 13 07:15:28 crc kubenswrapper[4971]: I1213 07:15:28.772686 4971 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/dnsmasq-dns-78c596d7cf-9jdzl" Dec 13 07:15:29 crc kubenswrapper[4971]: I1213 07:15:29.016714 4971 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-59fd54bbff-hnxpb"] Dec 13 07:15:29 crc kubenswrapper[4971]: I1213 07:15:29.016983 4971 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/dnsmasq-dns-59fd54bbff-hnxpb" podUID="5e843ecc-f9c0-492f-bcd8-0d34447ae18a" containerName="dnsmasq-dns" containerID="cri-o://1d77379c2441ec0288702f3e7f47f5164ec26f2f604e3edec19699d040133047" gracePeriod=10 Dec 13 07:15:29 crc kubenswrapper[4971]: I1213 07:15:29.619843 4971 generic.go:334] "Generic (PLEG): container finished" podID="18ea016d-26bf-46a9-a8c3-2f555643e967" containerID="e6e2bf1726974affec4fe8a70d50136677acd3da1008bacecd28f89bc4e2586c" exitCode=0 Dec 13 07:15:29 crc kubenswrapper[4971]: I1213 07:15:29.619899 4971 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"18ea016d-26bf-46a9-a8c3-2f555643e967","Type":"ContainerDied","Data":"e6e2bf1726974affec4fe8a70d50136677acd3da1008bacecd28f89bc4e2586c"} Dec 13 07:15:29 crc kubenswrapper[4971]: I1213 07:15:29.648418 4971 generic.go:334] "Generic (PLEG): container finished" podID="5e843ecc-f9c0-492f-bcd8-0d34447ae18a" containerID="1d77379c2441ec0288702f3e7f47f5164ec26f2f604e3edec19699d040133047" exitCode=0 Dec 13 07:15:29 crc kubenswrapper[4971]: I1213 07:15:29.649442 4971 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-59fd54bbff-hnxpb" event={"ID":"5e843ecc-f9c0-492f-bcd8-0d34447ae18a","Type":"ContainerDied","Data":"1d77379c2441ec0288702f3e7f47f5164ec26f2f604e3edec19699d040133047"} Dec 13 07:15:29 crc kubenswrapper[4971]: I1213 07:15:29.785048 4971 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-59fd54bbff-hnxpb" Dec 13 07:15:29 crc kubenswrapper[4971]: I1213 07:15:29.901720 4971 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/5e843ecc-f9c0-492f-bcd8-0d34447ae18a-ovsdbserver-sb\") pod \"5e843ecc-f9c0-492f-bcd8-0d34447ae18a\" (UID: \"5e843ecc-f9c0-492f-bcd8-0d34447ae18a\") " Dec 13 07:15:29 crc kubenswrapper[4971]: I1213 07:15:29.901895 4971 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/5e843ecc-f9c0-492f-bcd8-0d34447ae18a-config\") pod \"5e843ecc-f9c0-492f-bcd8-0d34447ae18a\" (UID: \"5e843ecc-f9c0-492f-bcd8-0d34447ae18a\") " Dec 13 07:15:29 crc kubenswrapper[4971]: I1213 07:15:29.902032 4971 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-qmmxb\" (UniqueName: \"kubernetes.io/projected/5e843ecc-f9c0-492f-bcd8-0d34447ae18a-kube-api-access-qmmxb\") pod \"5e843ecc-f9c0-492f-bcd8-0d34447ae18a\" (UID: \"5e843ecc-f9c0-492f-bcd8-0d34447ae18a\") " Dec 13 07:15:29 crc kubenswrapper[4971]: I1213 07:15:29.902104 4971 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/5e843ecc-f9c0-492f-bcd8-0d34447ae18a-ovsdbserver-nb\") pod \"5e843ecc-f9c0-492f-bcd8-0d34447ae18a\" (UID: \"5e843ecc-f9c0-492f-bcd8-0d34447ae18a\") " Dec 13 07:15:29 crc kubenswrapper[4971]: I1213 07:15:29.902150 4971 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/5e843ecc-f9c0-492f-bcd8-0d34447ae18a-dns-svc\") pod \"5e843ecc-f9c0-492f-bcd8-0d34447ae18a\" (UID: \"5e843ecc-f9c0-492f-bcd8-0d34447ae18a\") " Dec 13 07:15:29 crc kubenswrapper[4971]: I1213 07:15:29.926091 4971 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/5e843ecc-f9c0-492f-bcd8-0d34447ae18a-kube-api-access-qmmxb" (OuterVolumeSpecName: "kube-api-access-qmmxb") pod "5e843ecc-f9c0-492f-bcd8-0d34447ae18a" (UID: "5e843ecc-f9c0-492f-bcd8-0d34447ae18a"). InnerVolumeSpecName "kube-api-access-qmmxb". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 13 07:15:29 crc kubenswrapper[4971]: I1213 07:15:29.938186 4971 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Dec 13 07:15:29 crc kubenswrapper[4971]: I1213 07:15:29.956732 4971 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/5e843ecc-f9c0-492f-bcd8-0d34447ae18a-dns-svc" (OuterVolumeSpecName: "dns-svc") pod "5e843ecc-f9c0-492f-bcd8-0d34447ae18a" (UID: "5e843ecc-f9c0-492f-bcd8-0d34447ae18a"). InnerVolumeSpecName "dns-svc". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 13 07:15:30 crc kubenswrapper[4971]: I1213 07:15:30.150848 4971 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-qmmxb\" (UniqueName: \"kubernetes.io/projected/5e843ecc-f9c0-492f-bcd8-0d34447ae18a-kube-api-access-qmmxb\") on node \"crc\" DevicePath \"\"" Dec 13 07:15:30 crc kubenswrapper[4971]: I1213 07:15:30.151011 4971 reconciler_common.go:293] "Volume detached for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/5e843ecc-f9c0-492f-bcd8-0d34447ae18a-dns-svc\") on node \"crc\" DevicePath \"\"" Dec 13 07:15:30 crc kubenswrapper[4971]: I1213 07:15:30.158480 4971 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/5e843ecc-f9c0-492f-bcd8-0d34447ae18a-ovsdbserver-nb" (OuterVolumeSpecName: "ovsdbserver-nb") pod "5e843ecc-f9c0-492f-bcd8-0d34447ae18a" (UID: "5e843ecc-f9c0-492f-bcd8-0d34447ae18a"). InnerVolumeSpecName "ovsdbserver-nb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 13 07:15:30 crc kubenswrapper[4971]: I1213 07:15:30.168909 4971 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/5e843ecc-f9c0-492f-bcd8-0d34447ae18a-ovsdbserver-sb" (OuterVolumeSpecName: "ovsdbserver-sb") pod "5e843ecc-f9c0-492f-bcd8-0d34447ae18a" (UID: "5e843ecc-f9c0-492f-bcd8-0d34447ae18a"). InnerVolumeSpecName "ovsdbserver-sb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 13 07:15:30 crc kubenswrapper[4971]: I1213 07:15:30.210023 4971 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/5e843ecc-f9c0-492f-bcd8-0d34447ae18a-config" (OuterVolumeSpecName: "config") pod "5e843ecc-f9c0-492f-bcd8-0d34447ae18a" (UID: "5e843ecc-f9c0-492f-bcd8-0d34447ae18a"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 13 07:15:30 crc kubenswrapper[4971]: I1213 07:15:30.253026 4971 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ceilometer-tls-certs\" (UniqueName: \"kubernetes.io/secret/18ea016d-26bf-46a9-a8c3-2f555643e967-ceilometer-tls-certs\") pod \"18ea016d-26bf-46a9-a8c3-2f555643e967\" (UID: \"18ea016d-26bf-46a9-a8c3-2f555643e967\") " Dec 13 07:15:30 crc kubenswrapper[4971]: I1213 07:15:30.253107 4971 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/18ea016d-26bf-46a9-a8c3-2f555643e967-run-httpd\") pod \"18ea016d-26bf-46a9-a8c3-2f555643e967\" (UID: \"18ea016d-26bf-46a9-a8c3-2f555643e967\") " Dec 13 07:15:30 crc kubenswrapper[4971]: I1213 07:15:30.253146 4971 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/18ea016d-26bf-46a9-a8c3-2f555643e967-log-httpd\") pod \"18ea016d-26bf-46a9-a8c3-2f555643e967\" (UID: \"18ea016d-26bf-46a9-a8c3-2f555643e967\") " Dec 13 07:15:30 crc kubenswrapper[4971]: I1213 07:15:30.253172 4971 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-9qhqh\" (UniqueName: \"kubernetes.io/projected/18ea016d-26bf-46a9-a8c3-2f555643e967-kube-api-access-9qhqh\") pod \"18ea016d-26bf-46a9-a8c3-2f555643e967\" (UID: \"18ea016d-26bf-46a9-a8c3-2f555643e967\") " Dec 13 07:15:30 crc kubenswrapper[4971]: I1213 07:15:30.253194 4971 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/18ea016d-26bf-46a9-a8c3-2f555643e967-sg-core-conf-yaml\") pod \"18ea016d-26bf-46a9-a8c3-2f555643e967\" (UID: \"18ea016d-26bf-46a9-a8c3-2f555643e967\") " Dec 13 07:15:30 crc kubenswrapper[4971]: I1213 07:15:30.253229 4971 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/18ea016d-26bf-46a9-a8c3-2f555643e967-config-data\") pod \"18ea016d-26bf-46a9-a8c3-2f555643e967\" (UID: \"18ea016d-26bf-46a9-a8c3-2f555643e967\") " Dec 13 07:15:30 crc kubenswrapper[4971]: I1213 07:15:30.253254 4971 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/18ea016d-26bf-46a9-a8c3-2f555643e967-combined-ca-bundle\") pod \"18ea016d-26bf-46a9-a8c3-2f555643e967\" (UID: \"18ea016d-26bf-46a9-a8c3-2f555643e967\") " Dec 13 07:15:30 crc kubenswrapper[4971]: I1213 07:15:30.253271 4971 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/18ea016d-26bf-46a9-a8c3-2f555643e967-scripts\") pod \"18ea016d-26bf-46a9-a8c3-2f555643e967\" (UID: \"18ea016d-26bf-46a9-a8c3-2f555643e967\") " Dec 13 07:15:30 crc kubenswrapper[4971]: I1213 07:15:30.253529 4971 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/5e843ecc-f9c0-492f-bcd8-0d34447ae18a-ovsdbserver-nb\") on node \"crc\" DevicePath \"\"" Dec 13 07:15:30 crc kubenswrapper[4971]: I1213 07:15:30.253541 4971 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/5e843ecc-f9c0-492f-bcd8-0d34447ae18a-ovsdbserver-sb\") on node \"crc\" DevicePath \"\"" Dec 13 07:15:30 crc kubenswrapper[4971]: I1213 07:15:30.253552 4971 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/5e843ecc-f9c0-492f-bcd8-0d34447ae18a-config\") on node \"crc\" DevicePath \"\"" Dec 13 07:15:30 crc kubenswrapper[4971]: I1213 07:15:30.254832 4971 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/18ea016d-26bf-46a9-a8c3-2f555643e967-run-httpd" (OuterVolumeSpecName: "run-httpd") pod "18ea016d-26bf-46a9-a8c3-2f555643e967" (UID: "18ea016d-26bf-46a9-a8c3-2f555643e967"). InnerVolumeSpecName "run-httpd". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 13 07:15:30 crc kubenswrapper[4971]: I1213 07:15:30.272234 4971 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/18ea016d-26bf-46a9-a8c3-2f555643e967-kube-api-access-9qhqh" (OuterVolumeSpecName: "kube-api-access-9qhqh") pod "18ea016d-26bf-46a9-a8c3-2f555643e967" (UID: "18ea016d-26bf-46a9-a8c3-2f555643e967"). InnerVolumeSpecName "kube-api-access-9qhqh". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 13 07:15:30 crc kubenswrapper[4971]: I1213 07:15:30.272658 4971 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/18ea016d-26bf-46a9-a8c3-2f555643e967-log-httpd" (OuterVolumeSpecName: "log-httpd") pod "18ea016d-26bf-46a9-a8c3-2f555643e967" (UID: "18ea016d-26bf-46a9-a8c3-2f555643e967"). InnerVolumeSpecName "log-httpd". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 13 07:15:30 crc kubenswrapper[4971]: I1213 07:15:30.293333 4971 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/18ea016d-26bf-46a9-a8c3-2f555643e967-scripts" (OuterVolumeSpecName: "scripts") pod "18ea016d-26bf-46a9-a8c3-2f555643e967" (UID: "18ea016d-26bf-46a9-a8c3-2f555643e967"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 13 07:15:30 crc kubenswrapper[4971]: I1213 07:15:30.355818 4971 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/18ea016d-26bf-46a9-a8c3-2f555643e967-scripts\") on node \"crc\" DevicePath \"\"" Dec 13 07:15:30 crc kubenswrapper[4971]: I1213 07:15:30.355849 4971 reconciler_common.go:293] "Volume detached for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/18ea016d-26bf-46a9-a8c3-2f555643e967-run-httpd\") on node \"crc\" DevicePath \"\"" Dec 13 07:15:30 crc kubenswrapper[4971]: I1213 07:15:30.355859 4971 reconciler_common.go:293] "Volume detached for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/18ea016d-26bf-46a9-a8c3-2f555643e967-log-httpd\") on node \"crc\" DevicePath \"\"" Dec 13 07:15:30 crc kubenswrapper[4971]: I1213 07:15:30.355867 4971 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-9qhqh\" (UniqueName: \"kubernetes.io/projected/18ea016d-26bf-46a9-a8c3-2f555643e967-kube-api-access-9qhqh\") on node \"crc\" DevicePath \"\"" Dec 13 07:15:30 crc kubenswrapper[4971]: I1213 07:15:30.357854 4971 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/18ea016d-26bf-46a9-a8c3-2f555643e967-sg-core-conf-yaml" (OuterVolumeSpecName: "sg-core-conf-yaml") pod "18ea016d-26bf-46a9-a8c3-2f555643e967" (UID: "18ea016d-26bf-46a9-a8c3-2f555643e967"). InnerVolumeSpecName "sg-core-conf-yaml". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 13 07:15:30 crc kubenswrapper[4971]: I1213 07:15:30.377688 4971 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/18ea016d-26bf-46a9-a8c3-2f555643e967-ceilometer-tls-certs" (OuterVolumeSpecName: "ceilometer-tls-certs") pod "18ea016d-26bf-46a9-a8c3-2f555643e967" (UID: "18ea016d-26bf-46a9-a8c3-2f555643e967"). InnerVolumeSpecName "ceilometer-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 13 07:15:30 crc kubenswrapper[4971]: I1213 07:15:30.434734 4971 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/18ea016d-26bf-46a9-a8c3-2f555643e967-config-data" (OuterVolumeSpecName: "config-data") pod "18ea016d-26bf-46a9-a8c3-2f555643e967" (UID: "18ea016d-26bf-46a9-a8c3-2f555643e967"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 13 07:15:30 crc kubenswrapper[4971]: I1213 07:15:30.450673 4971 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/18ea016d-26bf-46a9-a8c3-2f555643e967-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "18ea016d-26bf-46a9-a8c3-2f555643e967" (UID: "18ea016d-26bf-46a9-a8c3-2f555643e967"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 13 07:15:30 crc kubenswrapper[4971]: I1213 07:15:30.457316 4971 reconciler_common.go:293] "Volume detached for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/18ea016d-26bf-46a9-a8c3-2f555643e967-sg-core-conf-yaml\") on node \"crc\" DevicePath \"\"" Dec 13 07:15:30 crc kubenswrapper[4971]: I1213 07:15:30.457474 4971 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/18ea016d-26bf-46a9-a8c3-2f555643e967-config-data\") on node \"crc\" DevicePath \"\"" Dec 13 07:15:30 crc kubenswrapper[4971]: I1213 07:15:30.457567 4971 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/18ea016d-26bf-46a9-a8c3-2f555643e967-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 13 07:15:30 crc kubenswrapper[4971]: I1213 07:15:30.457673 4971 reconciler_common.go:293] "Volume detached for volume \"ceilometer-tls-certs\" (UniqueName: \"kubernetes.io/secret/18ea016d-26bf-46a9-a8c3-2f555643e967-ceilometer-tls-certs\") on node \"crc\" DevicePath \"\"" Dec 13 07:15:30 crc kubenswrapper[4971]: I1213 07:15:30.693593 4971 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-59fd54bbff-hnxpb" event={"ID":"5e843ecc-f9c0-492f-bcd8-0d34447ae18a","Type":"ContainerDied","Data":"acf3b65a85b7c3fafbce46e07c02d107a4b68ebb1f81a0419c069fe8f6987e36"} Dec 13 07:15:30 crc kubenswrapper[4971]: I1213 07:15:30.693885 4971 scope.go:117] "RemoveContainer" containerID="1d77379c2441ec0288702f3e7f47f5164ec26f2f604e3edec19699d040133047" Dec 13 07:15:30 crc kubenswrapper[4971]: I1213 07:15:30.694063 4971 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-59fd54bbff-hnxpb" Dec 13 07:15:30 crc kubenswrapper[4971]: I1213 07:15:30.714700 4971 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"18ea016d-26bf-46a9-a8c3-2f555643e967","Type":"ContainerDied","Data":"198989c780b92cb74ce23e480469c4dddb72d91548c9d3208ceb3d1b12175bc3"} Dec 13 07:15:30 crc kubenswrapper[4971]: I1213 07:15:30.714800 4971 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Dec 13 07:15:30 crc kubenswrapper[4971]: I1213 07:15:30.733682 4971 scope.go:117] "RemoveContainer" containerID="7ef782e1bad569b3a550d4733a33011d066fce8a39433023866c5c83e6ff80ec" Dec 13 07:15:30 crc kubenswrapper[4971]: I1213 07:15:30.738853 4971 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-59fd54bbff-hnxpb"] Dec 13 07:15:30 crc kubenswrapper[4971]: I1213 07:15:30.748733 4971 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-59fd54bbff-hnxpb"] Dec 13 07:15:30 crc kubenswrapper[4971]: I1213 07:15:30.772684 4971 scope.go:117] "RemoveContainer" containerID="9894e8db28cfff1cb5f97d08464004fdc4d8d729765e4dfce4f19b760d82bc9c" Dec 13 07:15:30 crc kubenswrapper[4971]: I1213 07:15:30.791384 4971 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ceilometer-0"] Dec 13 07:15:30 crc kubenswrapper[4971]: I1213 07:15:30.802307 4971 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/ceilometer-0"] Dec 13 07:15:30 crc kubenswrapper[4971]: I1213 07:15:30.805289 4971 scope.go:117] "RemoveContainer" containerID="2d7101e713a8a556bd60407784326637a7599d4c2fe8201a04275e953f9be12f" Dec 13 07:15:30 crc kubenswrapper[4971]: I1213 07:15:30.860919 4971 scope.go:117] "RemoveContainer" containerID="e6e2bf1726974affec4fe8a70d50136677acd3da1008bacecd28f89bc4e2586c" Dec 13 07:15:30 crc kubenswrapper[4971]: I1213 07:15:30.865484 4971 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ceilometer-0"] Dec 13 07:15:30 crc kubenswrapper[4971]: E1213 07:15:30.866065 4971 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="5e843ecc-f9c0-492f-bcd8-0d34447ae18a" containerName="init" Dec 13 07:15:30 crc kubenswrapper[4971]: I1213 07:15:30.866197 4971 state_mem.go:107] "Deleted CPUSet assignment" podUID="5e843ecc-f9c0-492f-bcd8-0d34447ae18a" containerName="init" Dec 13 07:15:30 crc kubenswrapper[4971]: E1213 07:15:30.866302 4971 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="5e843ecc-f9c0-492f-bcd8-0d34447ae18a" containerName="dnsmasq-dns" Dec 13 07:15:30 crc kubenswrapper[4971]: I1213 07:15:30.866382 4971 state_mem.go:107] "Deleted CPUSet assignment" podUID="5e843ecc-f9c0-492f-bcd8-0d34447ae18a" containerName="dnsmasq-dns" Dec 13 07:15:30 crc kubenswrapper[4971]: E1213 07:15:30.866473 4971 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="18ea016d-26bf-46a9-a8c3-2f555643e967" containerName="sg-core" Dec 13 07:15:30 crc kubenswrapper[4971]: I1213 07:15:30.866605 4971 state_mem.go:107] "Deleted CPUSet assignment" podUID="18ea016d-26bf-46a9-a8c3-2f555643e967" containerName="sg-core" Dec 13 07:15:30 crc kubenswrapper[4971]: E1213 07:15:30.866767 4971 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="18ea016d-26bf-46a9-a8c3-2f555643e967" containerName="ceilometer-notification-agent" Dec 13 07:15:30 crc kubenswrapper[4971]: I1213 07:15:30.866848 4971 state_mem.go:107] "Deleted CPUSet assignment" podUID="18ea016d-26bf-46a9-a8c3-2f555643e967" containerName="ceilometer-notification-agent" Dec 13 07:15:30 crc kubenswrapper[4971]: E1213 07:15:30.866929 4971 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="18ea016d-26bf-46a9-a8c3-2f555643e967" containerName="ceilometer-central-agent" Dec 13 07:15:30 crc kubenswrapper[4971]: I1213 07:15:30.866998 4971 state_mem.go:107] "Deleted CPUSet assignment" podUID="18ea016d-26bf-46a9-a8c3-2f555643e967" containerName="ceilometer-central-agent" Dec 13 07:15:30 crc kubenswrapper[4971]: E1213 07:15:30.867063 4971 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="18ea016d-26bf-46a9-a8c3-2f555643e967" containerName="proxy-httpd" Dec 13 07:15:30 crc kubenswrapper[4971]: I1213 07:15:30.867119 4971 state_mem.go:107] "Deleted CPUSet assignment" podUID="18ea016d-26bf-46a9-a8c3-2f555643e967" containerName="proxy-httpd" Dec 13 07:15:30 crc kubenswrapper[4971]: I1213 07:15:30.867404 4971 memory_manager.go:354] "RemoveStaleState removing state" podUID="5e843ecc-f9c0-492f-bcd8-0d34447ae18a" containerName="dnsmasq-dns" Dec 13 07:15:30 crc kubenswrapper[4971]: I1213 07:15:30.867475 4971 memory_manager.go:354] "RemoveStaleState removing state" podUID="18ea016d-26bf-46a9-a8c3-2f555643e967" containerName="ceilometer-central-agent" Dec 13 07:15:30 crc kubenswrapper[4971]: I1213 07:15:30.867586 4971 memory_manager.go:354] "RemoveStaleState removing state" podUID="18ea016d-26bf-46a9-a8c3-2f555643e967" containerName="sg-core" Dec 13 07:15:30 crc kubenswrapper[4971]: I1213 07:15:30.867694 4971 memory_manager.go:354] "RemoveStaleState removing state" podUID="18ea016d-26bf-46a9-a8c3-2f555643e967" containerName="proxy-httpd" Dec 13 07:15:30 crc kubenswrapper[4971]: I1213 07:15:30.867783 4971 memory_manager.go:354] "RemoveStaleState removing state" podUID="18ea016d-26bf-46a9-a8c3-2f555643e967" containerName="ceilometer-notification-agent" Dec 13 07:15:30 crc kubenswrapper[4971]: I1213 07:15:30.875696 4971 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ceilometer-0"] Dec 13 07:15:30 crc kubenswrapper[4971]: I1213 07:15:30.875793 4971 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Dec 13 07:15:30 crc kubenswrapper[4971]: I1213 07:15:30.877999 4971 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceilometer-config-data" Dec 13 07:15:30 crc kubenswrapper[4971]: I1213 07:15:30.878401 4971 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-ceilometer-internal-svc" Dec 13 07:15:30 crc kubenswrapper[4971]: I1213 07:15:30.878563 4971 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceilometer-scripts" Dec 13 07:15:30 crc kubenswrapper[4971]: I1213 07:15:30.886068 4971 scope.go:117] "RemoveContainer" containerID="fb55d70eea8aa0f44521717d475539a24d67f5b76a554632aa7eea6670d42a3d" Dec 13 07:15:30 crc kubenswrapper[4971]: I1213 07:15:30.975384 4971 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/7cfb4dbf-cc55-48b7-b50c-65d6d4ae64ba-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"7cfb4dbf-cc55-48b7-b50c-65d6d4ae64ba\") " pod="openstack/ceilometer-0" Dec 13 07:15:30 crc kubenswrapper[4971]: I1213 07:15:30.975463 4971 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/7cfb4dbf-cc55-48b7-b50c-65d6d4ae64ba-scripts\") pod \"ceilometer-0\" (UID: \"7cfb4dbf-cc55-48b7-b50c-65d6d4ae64ba\") " pod="openstack/ceilometer-0" Dec 13 07:15:30 crc kubenswrapper[4971]: I1213 07:15:30.975490 4971 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/7cfb4dbf-cc55-48b7-b50c-65d6d4ae64ba-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"7cfb4dbf-cc55-48b7-b50c-65d6d4ae64ba\") " pod="openstack/ceilometer-0" Dec 13 07:15:30 crc kubenswrapper[4971]: I1213 07:15:30.975564 4971 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/7cfb4dbf-cc55-48b7-b50c-65d6d4ae64ba-run-httpd\") pod \"ceilometer-0\" (UID: \"7cfb4dbf-cc55-48b7-b50c-65d6d4ae64ba\") " pod="openstack/ceilometer-0" Dec 13 07:15:30 crc kubenswrapper[4971]: I1213 07:15:30.975589 4971 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-2tqrx\" (UniqueName: \"kubernetes.io/projected/7cfb4dbf-cc55-48b7-b50c-65d6d4ae64ba-kube-api-access-2tqrx\") pod \"ceilometer-0\" (UID: \"7cfb4dbf-cc55-48b7-b50c-65d6d4ae64ba\") " pod="openstack/ceilometer-0" Dec 13 07:15:30 crc kubenswrapper[4971]: I1213 07:15:30.975610 4971 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/7cfb4dbf-cc55-48b7-b50c-65d6d4ae64ba-config-data\") pod \"ceilometer-0\" (UID: \"7cfb4dbf-cc55-48b7-b50c-65d6d4ae64ba\") " pod="openstack/ceilometer-0" Dec 13 07:15:30 crc kubenswrapper[4971]: I1213 07:15:30.975803 4971 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/7cfb4dbf-cc55-48b7-b50c-65d6d4ae64ba-log-httpd\") pod \"ceilometer-0\" (UID: \"7cfb4dbf-cc55-48b7-b50c-65d6d4ae64ba\") " pod="openstack/ceilometer-0" Dec 13 07:15:30 crc kubenswrapper[4971]: I1213 07:15:30.975878 4971 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ceilometer-tls-certs\" (UniqueName: \"kubernetes.io/secret/7cfb4dbf-cc55-48b7-b50c-65d6d4ae64ba-ceilometer-tls-certs\") pod \"ceilometer-0\" (UID: \"7cfb4dbf-cc55-48b7-b50c-65d6d4ae64ba\") " pod="openstack/ceilometer-0" Dec 13 07:15:31 crc kubenswrapper[4971]: I1213 07:15:31.077508 4971 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/7cfb4dbf-cc55-48b7-b50c-65d6d4ae64ba-scripts\") pod \"ceilometer-0\" (UID: \"7cfb4dbf-cc55-48b7-b50c-65d6d4ae64ba\") " pod="openstack/ceilometer-0" Dec 13 07:15:31 crc kubenswrapper[4971]: I1213 07:15:31.077570 4971 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/7cfb4dbf-cc55-48b7-b50c-65d6d4ae64ba-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"7cfb4dbf-cc55-48b7-b50c-65d6d4ae64ba\") " pod="openstack/ceilometer-0" Dec 13 07:15:31 crc kubenswrapper[4971]: I1213 07:15:31.077634 4971 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/7cfb4dbf-cc55-48b7-b50c-65d6d4ae64ba-run-httpd\") pod \"ceilometer-0\" (UID: \"7cfb4dbf-cc55-48b7-b50c-65d6d4ae64ba\") " pod="openstack/ceilometer-0" Dec 13 07:15:31 crc kubenswrapper[4971]: I1213 07:15:31.077668 4971 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-2tqrx\" (UniqueName: \"kubernetes.io/projected/7cfb4dbf-cc55-48b7-b50c-65d6d4ae64ba-kube-api-access-2tqrx\") pod \"ceilometer-0\" (UID: \"7cfb4dbf-cc55-48b7-b50c-65d6d4ae64ba\") " pod="openstack/ceilometer-0" Dec 13 07:15:31 crc kubenswrapper[4971]: I1213 07:15:31.077695 4971 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/7cfb4dbf-cc55-48b7-b50c-65d6d4ae64ba-config-data\") pod \"ceilometer-0\" (UID: \"7cfb4dbf-cc55-48b7-b50c-65d6d4ae64ba\") " pod="openstack/ceilometer-0" Dec 13 07:15:31 crc kubenswrapper[4971]: I1213 07:15:31.077763 4971 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/7cfb4dbf-cc55-48b7-b50c-65d6d4ae64ba-log-httpd\") pod \"ceilometer-0\" (UID: \"7cfb4dbf-cc55-48b7-b50c-65d6d4ae64ba\") " pod="openstack/ceilometer-0" Dec 13 07:15:31 crc kubenswrapper[4971]: I1213 07:15:31.077821 4971 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ceilometer-tls-certs\" (UniqueName: \"kubernetes.io/secret/7cfb4dbf-cc55-48b7-b50c-65d6d4ae64ba-ceilometer-tls-certs\") pod \"ceilometer-0\" (UID: \"7cfb4dbf-cc55-48b7-b50c-65d6d4ae64ba\") " pod="openstack/ceilometer-0" Dec 13 07:15:31 crc kubenswrapper[4971]: I1213 07:15:31.077852 4971 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/7cfb4dbf-cc55-48b7-b50c-65d6d4ae64ba-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"7cfb4dbf-cc55-48b7-b50c-65d6d4ae64ba\") " pod="openstack/ceilometer-0" Dec 13 07:15:31 crc kubenswrapper[4971]: I1213 07:15:31.078497 4971 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/7cfb4dbf-cc55-48b7-b50c-65d6d4ae64ba-run-httpd\") pod \"ceilometer-0\" (UID: \"7cfb4dbf-cc55-48b7-b50c-65d6d4ae64ba\") " pod="openstack/ceilometer-0" Dec 13 07:15:31 crc kubenswrapper[4971]: I1213 07:15:31.078573 4971 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/7cfb4dbf-cc55-48b7-b50c-65d6d4ae64ba-log-httpd\") pod \"ceilometer-0\" (UID: \"7cfb4dbf-cc55-48b7-b50c-65d6d4ae64ba\") " pod="openstack/ceilometer-0" Dec 13 07:15:31 crc kubenswrapper[4971]: I1213 07:15:31.082431 4971 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/7cfb4dbf-cc55-48b7-b50c-65d6d4ae64ba-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"7cfb4dbf-cc55-48b7-b50c-65d6d4ae64ba\") " pod="openstack/ceilometer-0" Dec 13 07:15:31 crc kubenswrapper[4971]: I1213 07:15:31.082850 4971 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/7cfb4dbf-cc55-48b7-b50c-65d6d4ae64ba-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"7cfb4dbf-cc55-48b7-b50c-65d6d4ae64ba\") " pod="openstack/ceilometer-0" Dec 13 07:15:31 crc kubenswrapper[4971]: I1213 07:15:31.082984 4971 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ceilometer-tls-certs\" (UniqueName: \"kubernetes.io/secret/7cfb4dbf-cc55-48b7-b50c-65d6d4ae64ba-ceilometer-tls-certs\") pod \"ceilometer-0\" (UID: \"7cfb4dbf-cc55-48b7-b50c-65d6d4ae64ba\") " pod="openstack/ceilometer-0" Dec 13 07:15:31 crc kubenswrapper[4971]: I1213 07:15:31.083829 4971 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/7cfb4dbf-cc55-48b7-b50c-65d6d4ae64ba-config-data\") pod \"ceilometer-0\" (UID: \"7cfb4dbf-cc55-48b7-b50c-65d6d4ae64ba\") " pod="openstack/ceilometer-0" Dec 13 07:15:31 crc kubenswrapper[4971]: I1213 07:15:31.084717 4971 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/7cfb4dbf-cc55-48b7-b50c-65d6d4ae64ba-scripts\") pod \"ceilometer-0\" (UID: \"7cfb4dbf-cc55-48b7-b50c-65d6d4ae64ba\") " pod="openstack/ceilometer-0" Dec 13 07:15:31 crc kubenswrapper[4971]: I1213 07:15:31.097059 4971 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-2tqrx\" (UniqueName: \"kubernetes.io/projected/7cfb4dbf-cc55-48b7-b50c-65d6d4ae64ba-kube-api-access-2tqrx\") pod \"ceilometer-0\" (UID: \"7cfb4dbf-cc55-48b7-b50c-65d6d4ae64ba\") " pod="openstack/ceilometer-0" Dec 13 07:15:31 crc kubenswrapper[4971]: I1213 07:15:31.200837 4971 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Dec 13 07:15:31 crc kubenswrapper[4971]: I1213 07:15:31.783032 4971 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="18ea016d-26bf-46a9-a8c3-2f555643e967" path="/var/lib/kubelet/pods/18ea016d-26bf-46a9-a8c3-2f555643e967/volumes" Dec 13 07:15:31 crc kubenswrapper[4971]: I1213 07:15:31.784727 4971 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="5e843ecc-f9c0-492f-bcd8-0d34447ae18a" path="/var/lib/kubelet/pods/5e843ecc-f9c0-492f-bcd8-0d34447ae18a/volumes" Dec 13 07:15:31 crc kubenswrapper[4971]: I1213 07:15:31.897105 4971 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ceilometer-0"] Dec 13 07:15:32 crc kubenswrapper[4971]: I1213 07:15:32.738461 4971 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"7cfb4dbf-cc55-48b7-b50c-65d6d4ae64ba","Type":"ContainerStarted","Data":"7d58386baf345d60c820859eeb823a8fa1776e41a1099901668394ee34ba7a40"} Dec 13 07:15:33 crc kubenswrapper[4971]: I1213 07:15:33.753791 4971 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"7cfb4dbf-cc55-48b7-b50c-65d6d4ae64ba","Type":"ContainerStarted","Data":"5d071b2ffdcfc4bb8c866abc518564c7b5dac13fb9b40ac923cc1c5a8d52504f"} Dec 13 07:15:34 crc kubenswrapper[4971]: I1213 07:15:34.596990 4971 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/dnsmasq-dns-59fd54bbff-hnxpb" podUID="5e843ecc-f9c0-492f-bcd8-0d34447ae18a" containerName="dnsmasq-dns" probeResult="failure" output="dial tcp 10.217.0.177:5353: i/o timeout" Dec 13 07:15:34 crc kubenswrapper[4971]: I1213 07:15:34.764799 4971 generic.go:334] "Generic (PLEG): container finished" podID="20d00522-0b07-416b-90b4-0b588f99c41f" containerID="c993f4cf7451d6815046085857a794dcbdad51b484925bf88562a78935bb1b7f" exitCode=0 Dec 13 07:15:34 crc kubenswrapper[4971]: I1213 07:15:34.764859 4971 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-cell-mapping-kqkzw" event={"ID":"20d00522-0b07-416b-90b4-0b588f99c41f","Type":"ContainerDied","Data":"c993f4cf7451d6815046085857a794dcbdad51b484925bf88562a78935bb1b7f"} Dec 13 07:15:34 crc kubenswrapper[4971]: I1213 07:15:34.768497 4971 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"7cfb4dbf-cc55-48b7-b50c-65d6d4ae64ba","Type":"ContainerStarted","Data":"aa8dbe1b33fab32fe56308f8a63d03f9d6b4e85ef1f5a3bacd7f345b960f96d7"} Dec 13 07:15:35 crc kubenswrapper[4971]: I1213 07:15:35.709599 4971 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/nova-api-0" Dec 13 07:15:35 crc kubenswrapper[4971]: I1213 07:15:35.709998 4971 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/nova-api-0" Dec 13 07:15:35 crc kubenswrapper[4971]: I1213 07:15:35.786001 4971 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"7cfb4dbf-cc55-48b7-b50c-65d6d4ae64ba","Type":"ContainerStarted","Data":"fcb14c60d29a4cc2cb8ee4e92bbaef4cab76cb5fd4e53060369519333e285056"} Dec 13 07:15:36 crc kubenswrapper[4971]: I1213 07:15:36.361672 4971 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-cell-mapping-kqkzw" Dec 13 07:15:36 crc kubenswrapper[4971]: I1213 07:15:36.528726 4971 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-hczqd\" (UniqueName: \"kubernetes.io/projected/20d00522-0b07-416b-90b4-0b588f99c41f-kube-api-access-hczqd\") pod \"20d00522-0b07-416b-90b4-0b588f99c41f\" (UID: \"20d00522-0b07-416b-90b4-0b588f99c41f\") " Dec 13 07:15:36 crc kubenswrapper[4971]: I1213 07:15:36.528868 4971 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/20d00522-0b07-416b-90b4-0b588f99c41f-combined-ca-bundle\") pod \"20d00522-0b07-416b-90b4-0b588f99c41f\" (UID: \"20d00522-0b07-416b-90b4-0b588f99c41f\") " Dec 13 07:15:36 crc kubenswrapper[4971]: I1213 07:15:36.528982 4971 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/20d00522-0b07-416b-90b4-0b588f99c41f-config-data\") pod \"20d00522-0b07-416b-90b4-0b588f99c41f\" (UID: \"20d00522-0b07-416b-90b4-0b588f99c41f\") " Dec 13 07:15:36 crc kubenswrapper[4971]: I1213 07:15:36.529135 4971 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/20d00522-0b07-416b-90b4-0b588f99c41f-scripts\") pod \"20d00522-0b07-416b-90b4-0b588f99c41f\" (UID: \"20d00522-0b07-416b-90b4-0b588f99c41f\") " Dec 13 07:15:36 crc kubenswrapper[4971]: I1213 07:15:36.535089 4971 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/20d00522-0b07-416b-90b4-0b588f99c41f-kube-api-access-hczqd" (OuterVolumeSpecName: "kube-api-access-hczqd") pod "20d00522-0b07-416b-90b4-0b588f99c41f" (UID: "20d00522-0b07-416b-90b4-0b588f99c41f"). InnerVolumeSpecName "kube-api-access-hczqd". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 13 07:15:36 crc kubenswrapper[4971]: I1213 07:15:36.535301 4971 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/20d00522-0b07-416b-90b4-0b588f99c41f-scripts" (OuterVolumeSpecName: "scripts") pod "20d00522-0b07-416b-90b4-0b588f99c41f" (UID: "20d00522-0b07-416b-90b4-0b588f99c41f"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 13 07:15:36 crc kubenswrapper[4971]: I1213 07:15:36.710719 4971 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/20d00522-0b07-416b-90b4-0b588f99c41f-scripts\") on node \"crc\" DevicePath \"\"" Dec 13 07:15:36 crc kubenswrapper[4971]: I1213 07:15:36.710899 4971 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-hczqd\" (UniqueName: \"kubernetes.io/projected/20d00522-0b07-416b-90b4-0b588f99c41f-kube-api-access-hczqd\") on node \"crc\" DevicePath \"\"" Dec 13 07:15:36 crc kubenswrapper[4971]: I1213 07:15:36.717929 4971 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/20d00522-0b07-416b-90b4-0b588f99c41f-config-data" (OuterVolumeSpecName: "config-data") pod "20d00522-0b07-416b-90b4-0b588f99c41f" (UID: "20d00522-0b07-416b-90b4-0b588f99c41f"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 13 07:15:36 crc kubenswrapper[4971]: I1213 07:15:36.723691 4971 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/20d00522-0b07-416b-90b4-0b588f99c41f-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "20d00522-0b07-416b-90b4-0b588f99c41f" (UID: "20d00522-0b07-416b-90b4-0b588f99c41f"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 13 07:15:36 crc kubenswrapper[4971]: I1213 07:15:36.728633 4971 prober.go:107] "Probe failed" probeType="Startup" pod="openstack/nova-api-0" podUID="ce91d492-58fe-48a0-baa1-e26e736ab49a" containerName="nova-api-log" probeResult="failure" output="Get \"https://10.217.0.188:8774/\": net/http: request canceled (Client.Timeout exceeded while awaiting headers)" Dec 13 07:15:36 crc kubenswrapper[4971]: I1213 07:15:36.728937 4971 prober.go:107] "Probe failed" probeType="Startup" pod="openstack/nova-api-0" podUID="ce91d492-58fe-48a0-baa1-e26e736ab49a" containerName="nova-api-api" probeResult="failure" output="Get \"https://10.217.0.188:8774/\": net/http: request canceled (Client.Timeout exceeded while awaiting headers)" Dec 13 07:15:36 crc kubenswrapper[4971]: I1213 07:15:36.811915 4971 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-cell-mapping-kqkzw" event={"ID":"20d00522-0b07-416b-90b4-0b588f99c41f","Type":"ContainerDied","Data":"7a4d4193c4014c785b7952870ad36e4c6f7bc3d235e38bb4016d323802cc4f45"} Dec 13 07:15:36 crc kubenswrapper[4971]: I1213 07:15:36.811955 4971 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="7a4d4193c4014c785b7952870ad36e4c6f7bc3d235e38bb4016d323802cc4f45" Dec 13 07:15:36 crc kubenswrapper[4971]: I1213 07:15:36.812029 4971 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-cell-mapping-kqkzw" Dec 13 07:15:36 crc kubenswrapper[4971]: I1213 07:15:36.812842 4971 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/20d00522-0b07-416b-90b4-0b588f99c41f-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 13 07:15:36 crc kubenswrapper[4971]: I1213 07:15:36.812875 4971 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/20d00522-0b07-416b-90b4-0b588f99c41f-config-data\") on node \"crc\" DevicePath \"\"" Dec 13 07:15:36 crc kubenswrapper[4971]: I1213 07:15:36.828879 4971 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"7cfb4dbf-cc55-48b7-b50c-65d6d4ae64ba","Type":"ContainerStarted","Data":"9392bd2978cae265c91fac659c8e2f215ca3c33f3a347e415f189d6226655a93"} Dec 13 07:15:36 crc kubenswrapper[4971]: I1213 07:15:36.829740 4971 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/ceilometer-0" Dec 13 07:15:36 crc kubenswrapper[4971]: I1213 07:15:36.850496 4971 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/ceilometer-0" podStartSLOduration=3.12196233 podStartE2EDuration="6.850478415s" podCreationTimestamp="2025-12-13 07:15:30 +0000 UTC" firstStartedPulling="2025-12-13 07:15:32.045309533 +0000 UTC m=+1588.649718981" lastFinishedPulling="2025-12-13 07:15:35.773825618 +0000 UTC m=+1592.378235066" observedRunningTime="2025-12-13 07:15:36.847398659 +0000 UTC m=+1593.451808117" watchObservedRunningTime="2025-12-13 07:15:36.850478415 +0000 UTC m=+1593.454887863" Dec 13 07:15:36 crc kubenswrapper[4971]: I1213 07:15:36.974909 4971 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-api-0"] Dec 13 07:15:36 crc kubenswrapper[4971]: I1213 07:15:36.975153 4971 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-api-0" podUID="ce91d492-58fe-48a0-baa1-e26e736ab49a" containerName="nova-api-log" containerID="cri-o://3869966112be8a7f6f89d5ea0ceb4a3a1dea2af5d13e82953998769f10e53115" gracePeriod=30 Dec 13 07:15:36 crc kubenswrapper[4971]: I1213 07:15:36.975367 4971 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-api-0" podUID="ce91d492-58fe-48a0-baa1-e26e736ab49a" containerName="nova-api-api" containerID="cri-o://092c2442d88a2d6edbfffc8f033cfd7a1abeb2f3a58b906d2f18d96ae471ca6b" gracePeriod=30 Dec 13 07:15:36 crc kubenswrapper[4971]: I1213 07:15:36.996774 4971 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-scheduler-0"] Dec 13 07:15:36 crc kubenswrapper[4971]: I1213 07:15:36.997193 4971 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-scheduler-0" podUID="6ae42bbd-1447-4a68-ad90-ae94e2e0b6e5" containerName="nova-scheduler-scheduler" containerID="cri-o://932b370c20d3f4525fe2074e697c4a9bcc071571b0a037c1b415fc03f7a65e0a" gracePeriod=30 Dec 13 07:15:37 crc kubenswrapper[4971]: I1213 07:15:37.072060 4971 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-metadata-0"] Dec 13 07:15:37 crc kubenswrapper[4971]: I1213 07:15:37.072348 4971 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-metadata-0" podUID="ddbf55d9-778f-4601-bb0d-05a3d394de49" containerName="nova-metadata-log" containerID="cri-o://0902dafe9cf0d3f386fd5aab39ce9ea0516cc49c2c673963ec8413690dd887b6" gracePeriod=30 Dec 13 07:15:37 crc kubenswrapper[4971]: I1213 07:15:37.072397 4971 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-metadata-0" podUID="ddbf55d9-778f-4601-bb0d-05a3d394de49" containerName="nova-metadata-metadata" containerID="cri-o://edd3479a90674b9026df96971415f53d846bd1e1419434fa30680a8c2a2007ad" gracePeriod=30 Dec 13 07:15:37 crc kubenswrapper[4971]: I1213 07:15:37.806440 4971 scope.go:117] "RemoveContainer" containerID="e8beee43d15719dccb6892d6f5711674492425300efa098be89033bc65977d14" Dec 13 07:15:37 crc kubenswrapper[4971]: E1213 07:15:37.806765 4971 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-82xjz_openshift-machine-config-operator(e0c4af12-edda-41d7-9b44-f87396174ef2)\"" pod="openshift-machine-config-operator/machine-config-daemon-82xjz" podUID="e0c4af12-edda-41d7-9b44-f87396174ef2" Dec 13 07:15:37 crc kubenswrapper[4971]: I1213 07:15:37.838806 4971 generic.go:334] "Generic (PLEG): container finished" podID="ddbf55d9-778f-4601-bb0d-05a3d394de49" containerID="0902dafe9cf0d3f386fd5aab39ce9ea0516cc49c2c673963ec8413690dd887b6" exitCode=143 Dec 13 07:15:37 crc kubenswrapper[4971]: I1213 07:15:37.838869 4971 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"ddbf55d9-778f-4601-bb0d-05a3d394de49","Type":"ContainerDied","Data":"0902dafe9cf0d3f386fd5aab39ce9ea0516cc49c2c673963ec8413690dd887b6"} Dec 13 07:15:37 crc kubenswrapper[4971]: I1213 07:15:37.840444 4971 generic.go:334] "Generic (PLEG): container finished" podID="ce91d492-58fe-48a0-baa1-e26e736ab49a" containerID="3869966112be8a7f6f89d5ea0ceb4a3a1dea2af5d13e82953998769f10e53115" exitCode=143 Dec 13 07:15:37 crc kubenswrapper[4971]: I1213 07:15:37.841370 4971 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"ce91d492-58fe-48a0-baa1-e26e736ab49a","Type":"ContainerDied","Data":"3869966112be8a7f6f89d5ea0ceb4a3a1dea2af5d13e82953998769f10e53115"} Dec 13 07:15:40 crc kubenswrapper[4971]: I1213 07:15:40.476131 4971 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/nova-metadata-0" podUID="ddbf55d9-778f-4601-bb0d-05a3d394de49" containerName="nova-metadata-metadata" probeResult="failure" output="Get \"https://10.217.0.181:8775/\": read tcp 10.217.0.2:48404->10.217.0.181:8775: read: connection reset by peer" Dec 13 07:15:40 crc kubenswrapper[4971]: I1213 07:15:40.476163 4971 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/nova-metadata-0" podUID="ddbf55d9-778f-4601-bb0d-05a3d394de49" containerName="nova-metadata-log" probeResult="failure" output="Get \"https://10.217.0.181:8775/\": read tcp 10.217.0.2:48416->10.217.0.181:8775: read: connection reset by peer" Dec 13 07:15:40 crc kubenswrapper[4971]: I1213 07:15:40.891494 4971 generic.go:334] "Generic (PLEG): container finished" podID="ddbf55d9-778f-4601-bb0d-05a3d394de49" containerID="edd3479a90674b9026df96971415f53d846bd1e1419434fa30680a8c2a2007ad" exitCode=0 Dec 13 07:15:40 crc kubenswrapper[4971]: I1213 07:15:40.891618 4971 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"ddbf55d9-778f-4601-bb0d-05a3d394de49","Type":"ContainerDied","Data":"edd3479a90674b9026df96971415f53d846bd1e1419434fa30680a8c2a2007ad"} Dec 13 07:15:40 crc kubenswrapper[4971]: I1213 07:15:40.893897 4971 generic.go:334] "Generic (PLEG): container finished" podID="6ae42bbd-1447-4a68-ad90-ae94e2e0b6e5" containerID="932b370c20d3f4525fe2074e697c4a9bcc071571b0a037c1b415fc03f7a65e0a" exitCode=0 Dec 13 07:15:40 crc kubenswrapper[4971]: I1213 07:15:40.893936 4971 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-scheduler-0" event={"ID":"6ae42bbd-1447-4a68-ad90-ae94e2e0b6e5","Type":"ContainerDied","Data":"932b370c20d3f4525fe2074e697c4a9bcc071571b0a037c1b415fc03f7a65e0a"} Dec 13 07:15:40 crc kubenswrapper[4971]: I1213 07:15:40.960861 4971 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-metadata-0" Dec 13 07:15:41 crc kubenswrapper[4971]: I1213 07:15:41.006862 4971 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-rrdbn\" (UniqueName: \"kubernetes.io/projected/ddbf55d9-778f-4601-bb0d-05a3d394de49-kube-api-access-rrdbn\") pod \"ddbf55d9-778f-4601-bb0d-05a3d394de49\" (UID: \"ddbf55d9-778f-4601-bb0d-05a3d394de49\") " Dec 13 07:15:41 crc kubenswrapper[4971]: I1213 07:15:41.006971 4971 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/ddbf55d9-778f-4601-bb0d-05a3d394de49-config-data\") pod \"ddbf55d9-778f-4601-bb0d-05a3d394de49\" (UID: \"ddbf55d9-778f-4601-bb0d-05a3d394de49\") " Dec 13 07:15:41 crc kubenswrapper[4971]: I1213 07:15:41.006999 4971 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/ddbf55d9-778f-4601-bb0d-05a3d394de49-combined-ca-bundle\") pod \"ddbf55d9-778f-4601-bb0d-05a3d394de49\" (UID: \"ddbf55d9-778f-4601-bb0d-05a3d394de49\") " Dec 13 07:15:41 crc kubenswrapper[4971]: I1213 07:15:41.007133 4971 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/ddbf55d9-778f-4601-bb0d-05a3d394de49-logs\") pod \"ddbf55d9-778f-4601-bb0d-05a3d394de49\" (UID: \"ddbf55d9-778f-4601-bb0d-05a3d394de49\") " Dec 13 07:15:41 crc kubenswrapper[4971]: I1213 07:15:41.007224 4971 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"nova-metadata-tls-certs\" (UniqueName: \"kubernetes.io/secret/ddbf55d9-778f-4601-bb0d-05a3d394de49-nova-metadata-tls-certs\") pod \"ddbf55d9-778f-4601-bb0d-05a3d394de49\" (UID: \"ddbf55d9-778f-4601-bb0d-05a3d394de49\") " Dec 13 07:15:41 crc kubenswrapper[4971]: I1213 07:15:41.010168 4971 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/ddbf55d9-778f-4601-bb0d-05a3d394de49-logs" (OuterVolumeSpecName: "logs") pod "ddbf55d9-778f-4601-bb0d-05a3d394de49" (UID: "ddbf55d9-778f-4601-bb0d-05a3d394de49"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 13 07:15:41 crc kubenswrapper[4971]: I1213 07:15:41.015675 4971 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/ddbf55d9-778f-4601-bb0d-05a3d394de49-kube-api-access-rrdbn" (OuterVolumeSpecName: "kube-api-access-rrdbn") pod "ddbf55d9-778f-4601-bb0d-05a3d394de49" (UID: "ddbf55d9-778f-4601-bb0d-05a3d394de49"). InnerVolumeSpecName "kube-api-access-rrdbn". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 13 07:15:41 crc kubenswrapper[4971]: I1213 07:15:41.051995 4971 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/ddbf55d9-778f-4601-bb0d-05a3d394de49-config-data" (OuterVolumeSpecName: "config-data") pod "ddbf55d9-778f-4601-bb0d-05a3d394de49" (UID: "ddbf55d9-778f-4601-bb0d-05a3d394de49"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 13 07:15:41 crc kubenswrapper[4971]: I1213 07:15:41.063983 4971 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-scheduler-0" Dec 13 07:15:41 crc kubenswrapper[4971]: I1213 07:15:41.089854 4971 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/ddbf55d9-778f-4601-bb0d-05a3d394de49-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "ddbf55d9-778f-4601-bb0d-05a3d394de49" (UID: "ddbf55d9-778f-4601-bb0d-05a3d394de49"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 13 07:15:41 crc kubenswrapper[4971]: I1213 07:15:41.110711 4971 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-fdz56\" (UniqueName: \"kubernetes.io/projected/6ae42bbd-1447-4a68-ad90-ae94e2e0b6e5-kube-api-access-fdz56\") pod \"6ae42bbd-1447-4a68-ad90-ae94e2e0b6e5\" (UID: \"6ae42bbd-1447-4a68-ad90-ae94e2e0b6e5\") " Dec 13 07:15:41 crc kubenswrapper[4971]: I1213 07:15:41.110793 4971 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/6ae42bbd-1447-4a68-ad90-ae94e2e0b6e5-config-data\") pod \"6ae42bbd-1447-4a68-ad90-ae94e2e0b6e5\" (UID: \"6ae42bbd-1447-4a68-ad90-ae94e2e0b6e5\") " Dec 13 07:15:41 crc kubenswrapper[4971]: I1213 07:15:41.110890 4971 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/6ae42bbd-1447-4a68-ad90-ae94e2e0b6e5-combined-ca-bundle\") pod \"6ae42bbd-1447-4a68-ad90-ae94e2e0b6e5\" (UID: \"6ae42bbd-1447-4a68-ad90-ae94e2e0b6e5\") " Dec 13 07:15:41 crc kubenswrapper[4971]: I1213 07:15:41.111360 4971 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/ddbf55d9-778f-4601-bb0d-05a3d394de49-logs\") on node \"crc\" DevicePath \"\"" Dec 13 07:15:41 crc kubenswrapper[4971]: I1213 07:15:41.111371 4971 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-rrdbn\" (UniqueName: \"kubernetes.io/projected/ddbf55d9-778f-4601-bb0d-05a3d394de49-kube-api-access-rrdbn\") on node \"crc\" DevicePath \"\"" Dec 13 07:15:41 crc kubenswrapper[4971]: I1213 07:15:41.111384 4971 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/ddbf55d9-778f-4601-bb0d-05a3d394de49-config-data\") on node \"crc\" DevicePath \"\"" Dec 13 07:15:41 crc kubenswrapper[4971]: I1213 07:15:41.111391 4971 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/ddbf55d9-778f-4601-bb0d-05a3d394de49-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 13 07:15:41 crc kubenswrapper[4971]: I1213 07:15:41.119349 4971 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/ddbf55d9-778f-4601-bb0d-05a3d394de49-nova-metadata-tls-certs" (OuterVolumeSpecName: "nova-metadata-tls-certs") pod "ddbf55d9-778f-4601-bb0d-05a3d394de49" (UID: "ddbf55d9-778f-4601-bb0d-05a3d394de49"). InnerVolumeSpecName "nova-metadata-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 13 07:15:41 crc kubenswrapper[4971]: I1213 07:15:41.119735 4971 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/6ae42bbd-1447-4a68-ad90-ae94e2e0b6e5-kube-api-access-fdz56" (OuterVolumeSpecName: "kube-api-access-fdz56") pod "6ae42bbd-1447-4a68-ad90-ae94e2e0b6e5" (UID: "6ae42bbd-1447-4a68-ad90-ae94e2e0b6e5"). InnerVolumeSpecName "kube-api-access-fdz56". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 13 07:15:41 crc kubenswrapper[4971]: I1213 07:15:41.149672 4971 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/6ae42bbd-1447-4a68-ad90-ae94e2e0b6e5-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "6ae42bbd-1447-4a68-ad90-ae94e2e0b6e5" (UID: "6ae42bbd-1447-4a68-ad90-ae94e2e0b6e5"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 13 07:15:41 crc kubenswrapper[4971]: I1213 07:15:41.157496 4971 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/6ae42bbd-1447-4a68-ad90-ae94e2e0b6e5-config-data" (OuterVolumeSpecName: "config-data") pod "6ae42bbd-1447-4a68-ad90-ae94e2e0b6e5" (UID: "6ae42bbd-1447-4a68-ad90-ae94e2e0b6e5"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 13 07:15:41 crc kubenswrapper[4971]: I1213 07:15:41.213334 4971 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-fdz56\" (UniqueName: \"kubernetes.io/projected/6ae42bbd-1447-4a68-ad90-ae94e2e0b6e5-kube-api-access-fdz56\") on node \"crc\" DevicePath \"\"" Dec 13 07:15:41 crc kubenswrapper[4971]: I1213 07:15:41.213378 4971 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/6ae42bbd-1447-4a68-ad90-ae94e2e0b6e5-config-data\") on node \"crc\" DevicePath \"\"" Dec 13 07:15:41 crc kubenswrapper[4971]: I1213 07:15:41.213392 4971 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/6ae42bbd-1447-4a68-ad90-ae94e2e0b6e5-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 13 07:15:41 crc kubenswrapper[4971]: I1213 07:15:41.213403 4971 reconciler_common.go:293] "Volume detached for volume \"nova-metadata-tls-certs\" (UniqueName: \"kubernetes.io/secret/ddbf55d9-778f-4601-bb0d-05a3d394de49-nova-metadata-tls-certs\") on node \"crc\" DevicePath \"\"" Dec 13 07:15:41 crc kubenswrapper[4971]: I1213 07:15:41.904486 4971 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-scheduler-0" event={"ID":"6ae42bbd-1447-4a68-ad90-ae94e2e0b6e5","Type":"ContainerDied","Data":"d5a99e398b2e8cf2c59612c292e5994dae1a2e739a25eebb15d08ea8cd0b1d46"} Dec 13 07:15:41 crc kubenswrapper[4971]: I1213 07:15:41.904755 4971 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-scheduler-0" Dec 13 07:15:41 crc kubenswrapper[4971]: I1213 07:15:41.904855 4971 scope.go:117] "RemoveContainer" containerID="932b370c20d3f4525fe2074e697c4a9bcc071571b0a037c1b415fc03f7a65e0a" Dec 13 07:15:41 crc kubenswrapper[4971]: I1213 07:15:41.907152 4971 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"ddbf55d9-778f-4601-bb0d-05a3d394de49","Type":"ContainerDied","Data":"ae222c3d269cebe9893d00ef044b83c2fb5d71db6f454b39a4ca176b353d4e3e"} Dec 13 07:15:41 crc kubenswrapper[4971]: I1213 07:15:41.907216 4971 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-metadata-0" Dec 13 07:15:41 crc kubenswrapper[4971]: I1213 07:15:41.934848 4971 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-metadata-0"] Dec 13 07:15:41 crc kubenswrapper[4971]: I1213 07:15:41.937316 4971 scope.go:117] "RemoveContainer" containerID="edd3479a90674b9026df96971415f53d846bd1e1419434fa30680a8c2a2007ad" Dec 13 07:15:41 crc kubenswrapper[4971]: I1213 07:15:41.944644 4971 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-metadata-0"] Dec 13 07:15:41 crc kubenswrapper[4971]: I1213 07:15:41.956498 4971 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-scheduler-0"] Dec 13 07:15:41 crc kubenswrapper[4971]: I1213 07:15:41.968936 4971 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-metadata-0"] Dec 13 07:15:41 crc kubenswrapper[4971]: E1213 07:15:41.969471 4971 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="ddbf55d9-778f-4601-bb0d-05a3d394de49" containerName="nova-metadata-log" Dec 13 07:15:41 crc kubenswrapper[4971]: I1213 07:15:41.969492 4971 state_mem.go:107] "Deleted CPUSet assignment" podUID="ddbf55d9-778f-4601-bb0d-05a3d394de49" containerName="nova-metadata-log" Dec 13 07:15:41 crc kubenswrapper[4971]: E1213 07:15:41.969542 4971 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="20d00522-0b07-416b-90b4-0b588f99c41f" containerName="nova-manage" Dec 13 07:15:41 crc kubenswrapper[4971]: I1213 07:15:41.969549 4971 state_mem.go:107] "Deleted CPUSet assignment" podUID="20d00522-0b07-416b-90b4-0b588f99c41f" containerName="nova-manage" Dec 13 07:15:41 crc kubenswrapper[4971]: E1213 07:15:41.969567 4971 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="ddbf55d9-778f-4601-bb0d-05a3d394de49" containerName="nova-metadata-metadata" Dec 13 07:15:41 crc kubenswrapper[4971]: I1213 07:15:41.969573 4971 state_mem.go:107] "Deleted CPUSet assignment" podUID="ddbf55d9-778f-4601-bb0d-05a3d394de49" containerName="nova-metadata-metadata" Dec 13 07:15:41 crc kubenswrapper[4971]: E1213 07:15:41.969587 4971 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="6ae42bbd-1447-4a68-ad90-ae94e2e0b6e5" containerName="nova-scheduler-scheduler" Dec 13 07:15:41 crc kubenswrapper[4971]: I1213 07:15:41.969593 4971 state_mem.go:107] "Deleted CPUSet assignment" podUID="6ae42bbd-1447-4a68-ad90-ae94e2e0b6e5" containerName="nova-scheduler-scheduler" Dec 13 07:15:41 crc kubenswrapper[4971]: I1213 07:15:41.969753 4971 memory_manager.go:354] "RemoveStaleState removing state" podUID="ddbf55d9-778f-4601-bb0d-05a3d394de49" containerName="nova-metadata-log" Dec 13 07:15:41 crc kubenswrapper[4971]: I1213 07:15:41.969778 4971 memory_manager.go:354] "RemoveStaleState removing state" podUID="6ae42bbd-1447-4a68-ad90-ae94e2e0b6e5" containerName="nova-scheduler-scheduler" Dec 13 07:15:41 crc kubenswrapper[4971]: I1213 07:15:41.969788 4971 memory_manager.go:354] "RemoveStaleState removing state" podUID="20d00522-0b07-416b-90b4-0b588f99c41f" containerName="nova-manage" Dec 13 07:15:41 crc kubenswrapper[4971]: I1213 07:15:41.969797 4971 memory_manager.go:354] "RemoveStaleState removing state" podUID="ddbf55d9-778f-4601-bb0d-05a3d394de49" containerName="nova-metadata-metadata" Dec 13 07:15:41 crc kubenswrapper[4971]: I1213 07:15:41.971121 4971 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-metadata-0" Dec 13 07:15:41 crc kubenswrapper[4971]: I1213 07:15:41.975945 4971 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-nova-metadata-internal-svc" Dec 13 07:15:41 crc kubenswrapper[4971]: I1213 07:15:41.976931 4971 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-metadata-config-data" Dec 13 07:15:41 crc kubenswrapper[4971]: I1213 07:15:41.989569 4971 scope.go:117] "RemoveContainer" containerID="0902dafe9cf0d3f386fd5aab39ce9ea0516cc49c2c673963ec8413690dd887b6" Dec 13 07:15:41 crc kubenswrapper[4971]: I1213 07:15:41.994888 4971 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-scheduler-0"] Dec 13 07:15:42 crc kubenswrapper[4971]: I1213 07:15:42.005883 4971 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-metadata-0"] Dec 13 07:15:42 crc kubenswrapper[4971]: I1213 07:15:42.020312 4971 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-scheduler-0"] Dec 13 07:15:42 crc kubenswrapper[4971]: I1213 07:15:42.022778 4971 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-scheduler-0" Dec 13 07:15:42 crc kubenswrapper[4971]: I1213 07:15:42.024884 4971 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-scheduler-config-data" Dec 13 07:15:42 crc kubenswrapper[4971]: I1213 07:15:42.031386 4971 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-scheduler-0"] Dec 13 07:15:42 crc kubenswrapper[4971]: I1213 07:15:42.032485 4971 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d4bb47ae-3205-452b-b596-563d3833ee82-combined-ca-bundle\") pod \"nova-metadata-0\" (UID: \"d4bb47ae-3205-452b-b596-563d3833ee82\") " pod="openstack/nova-metadata-0" Dec 13 07:15:42 crc kubenswrapper[4971]: I1213 07:15:42.032603 4971 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/d4bb47ae-3205-452b-b596-563d3833ee82-logs\") pod \"nova-metadata-0\" (UID: \"d4bb47ae-3205-452b-b596-563d3833ee82\") " pod="openstack/nova-metadata-0" Dec 13 07:15:42 crc kubenswrapper[4971]: I1213 07:15:42.032679 4971 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/d4bb47ae-3205-452b-b596-563d3833ee82-config-data\") pod \"nova-metadata-0\" (UID: \"d4bb47ae-3205-452b-b596-563d3833ee82\") " pod="openstack/nova-metadata-0" Dec 13 07:15:42 crc kubenswrapper[4971]: I1213 07:15:42.032740 4971 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"nova-metadata-tls-certs\" (UniqueName: \"kubernetes.io/secret/d4bb47ae-3205-452b-b596-563d3833ee82-nova-metadata-tls-certs\") pod \"nova-metadata-0\" (UID: \"d4bb47ae-3205-452b-b596-563d3833ee82\") " pod="openstack/nova-metadata-0" Dec 13 07:15:42 crc kubenswrapper[4971]: I1213 07:15:42.032768 4971 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-67hhb\" (UniqueName: \"kubernetes.io/projected/d4bb47ae-3205-452b-b596-563d3833ee82-kube-api-access-67hhb\") pod \"nova-metadata-0\" (UID: \"d4bb47ae-3205-452b-b596-563d3833ee82\") " pod="openstack/nova-metadata-0" Dec 13 07:15:42 crc kubenswrapper[4971]: I1213 07:15:42.134794 4971 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/14465eba-34c5-43d3-ad9d-06b454f79963-config-data\") pod \"nova-scheduler-0\" (UID: \"14465eba-34c5-43d3-ad9d-06b454f79963\") " pod="openstack/nova-scheduler-0" Dec 13 07:15:42 crc kubenswrapper[4971]: I1213 07:15:42.134874 4971 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/d4bb47ae-3205-452b-b596-563d3833ee82-config-data\") pod \"nova-metadata-0\" (UID: \"d4bb47ae-3205-452b-b596-563d3833ee82\") " pod="openstack/nova-metadata-0" Dec 13 07:15:42 crc kubenswrapper[4971]: I1213 07:15:42.134924 4971 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nova-metadata-tls-certs\" (UniqueName: \"kubernetes.io/secret/d4bb47ae-3205-452b-b596-563d3833ee82-nova-metadata-tls-certs\") pod \"nova-metadata-0\" (UID: \"d4bb47ae-3205-452b-b596-563d3833ee82\") " pod="openstack/nova-metadata-0" Dec 13 07:15:42 crc kubenswrapper[4971]: I1213 07:15:42.134949 4971 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-67hhb\" (UniqueName: \"kubernetes.io/projected/d4bb47ae-3205-452b-b596-563d3833ee82-kube-api-access-67hhb\") pod \"nova-metadata-0\" (UID: \"d4bb47ae-3205-452b-b596-563d3833ee82\") " pod="openstack/nova-metadata-0" Dec 13 07:15:42 crc kubenswrapper[4971]: I1213 07:15:42.134991 4971 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/14465eba-34c5-43d3-ad9d-06b454f79963-combined-ca-bundle\") pod \"nova-scheduler-0\" (UID: \"14465eba-34c5-43d3-ad9d-06b454f79963\") " pod="openstack/nova-scheduler-0" Dec 13 07:15:42 crc kubenswrapper[4971]: I1213 07:15:42.135049 4971 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-6kkbd\" (UniqueName: \"kubernetes.io/projected/14465eba-34c5-43d3-ad9d-06b454f79963-kube-api-access-6kkbd\") pod \"nova-scheduler-0\" (UID: \"14465eba-34c5-43d3-ad9d-06b454f79963\") " pod="openstack/nova-scheduler-0" Dec 13 07:15:42 crc kubenswrapper[4971]: I1213 07:15:42.135077 4971 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d4bb47ae-3205-452b-b596-563d3833ee82-combined-ca-bundle\") pod \"nova-metadata-0\" (UID: \"d4bb47ae-3205-452b-b596-563d3833ee82\") " pod="openstack/nova-metadata-0" Dec 13 07:15:42 crc kubenswrapper[4971]: I1213 07:15:42.135143 4971 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/d4bb47ae-3205-452b-b596-563d3833ee82-logs\") pod \"nova-metadata-0\" (UID: \"d4bb47ae-3205-452b-b596-563d3833ee82\") " pod="openstack/nova-metadata-0" Dec 13 07:15:42 crc kubenswrapper[4971]: I1213 07:15:42.135786 4971 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/d4bb47ae-3205-452b-b596-563d3833ee82-logs\") pod \"nova-metadata-0\" (UID: \"d4bb47ae-3205-452b-b596-563d3833ee82\") " pod="openstack/nova-metadata-0" Dec 13 07:15:42 crc kubenswrapper[4971]: I1213 07:15:42.142672 4971 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"nova-metadata-tls-certs\" (UniqueName: \"kubernetes.io/secret/d4bb47ae-3205-452b-b596-563d3833ee82-nova-metadata-tls-certs\") pod \"nova-metadata-0\" (UID: \"d4bb47ae-3205-452b-b596-563d3833ee82\") " pod="openstack/nova-metadata-0" Dec 13 07:15:42 crc kubenswrapper[4971]: I1213 07:15:42.146828 4971 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d4bb47ae-3205-452b-b596-563d3833ee82-combined-ca-bundle\") pod \"nova-metadata-0\" (UID: \"d4bb47ae-3205-452b-b596-563d3833ee82\") " pod="openstack/nova-metadata-0" Dec 13 07:15:42 crc kubenswrapper[4971]: I1213 07:15:42.148253 4971 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/d4bb47ae-3205-452b-b596-563d3833ee82-config-data\") pod \"nova-metadata-0\" (UID: \"d4bb47ae-3205-452b-b596-563d3833ee82\") " pod="openstack/nova-metadata-0" Dec 13 07:15:42 crc kubenswrapper[4971]: I1213 07:15:42.187081 4971 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-67hhb\" (UniqueName: \"kubernetes.io/projected/d4bb47ae-3205-452b-b596-563d3833ee82-kube-api-access-67hhb\") pod \"nova-metadata-0\" (UID: \"d4bb47ae-3205-452b-b596-563d3833ee82\") " pod="openstack/nova-metadata-0" Dec 13 07:15:42 crc kubenswrapper[4971]: I1213 07:15:42.240524 4971 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/14465eba-34c5-43d3-ad9d-06b454f79963-config-data\") pod \"nova-scheduler-0\" (UID: \"14465eba-34c5-43d3-ad9d-06b454f79963\") " pod="openstack/nova-scheduler-0" Dec 13 07:15:42 crc kubenswrapper[4971]: I1213 07:15:42.240602 4971 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/14465eba-34c5-43d3-ad9d-06b454f79963-combined-ca-bundle\") pod \"nova-scheduler-0\" (UID: \"14465eba-34c5-43d3-ad9d-06b454f79963\") " pod="openstack/nova-scheduler-0" Dec 13 07:15:42 crc kubenswrapper[4971]: I1213 07:15:42.240644 4971 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-6kkbd\" (UniqueName: \"kubernetes.io/projected/14465eba-34c5-43d3-ad9d-06b454f79963-kube-api-access-6kkbd\") pod \"nova-scheduler-0\" (UID: \"14465eba-34c5-43d3-ad9d-06b454f79963\") " pod="openstack/nova-scheduler-0" Dec 13 07:15:42 crc kubenswrapper[4971]: I1213 07:15:42.243924 4971 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/14465eba-34c5-43d3-ad9d-06b454f79963-config-data\") pod \"nova-scheduler-0\" (UID: \"14465eba-34c5-43d3-ad9d-06b454f79963\") " pod="openstack/nova-scheduler-0" Dec 13 07:15:42 crc kubenswrapper[4971]: I1213 07:15:42.247225 4971 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/14465eba-34c5-43d3-ad9d-06b454f79963-combined-ca-bundle\") pod \"nova-scheduler-0\" (UID: \"14465eba-34c5-43d3-ad9d-06b454f79963\") " pod="openstack/nova-scheduler-0" Dec 13 07:15:42 crc kubenswrapper[4971]: I1213 07:15:42.273231 4971 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-6kkbd\" (UniqueName: \"kubernetes.io/projected/14465eba-34c5-43d3-ad9d-06b454f79963-kube-api-access-6kkbd\") pod \"nova-scheduler-0\" (UID: \"14465eba-34c5-43d3-ad9d-06b454f79963\") " pod="openstack/nova-scheduler-0" Dec 13 07:15:42 crc kubenswrapper[4971]: I1213 07:15:42.298201 4971 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-metadata-0" Dec 13 07:15:42 crc kubenswrapper[4971]: I1213 07:15:42.352496 4971 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-scheduler-0" Dec 13 07:15:42 crc kubenswrapper[4971]: I1213 07:15:42.796432 4971 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-metadata-0"] Dec 13 07:15:42 crc kubenswrapper[4971]: I1213 07:15:42.932500 4971 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"d4bb47ae-3205-452b-b596-563d3833ee82","Type":"ContainerStarted","Data":"f55fde368c522703cac763eff07483aea4580a4c8d6d4e6b9fd25ad190a56da1"} Dec 13 07:15:42 crc kubenswrapper[4971]: W1213 07:15:42.956104 4971 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod14465eba_34c5_43d3_ad9d_06b454f79963.slice/crio-4aed11d9a4901cdc913329541233f95173d5b87e2f41bb40b8defff64460aa48 WatchSource:0}: Error finding container 4aed11d9a4901cdc913329541233f95173d5b87e2f41bb40b8defff64460aa48: Status 404 returned error can't find the container with id 4aed11d9a4901cdc913329541233f95173d5b87e2f41bb40b8defff64460aa48 Dec 13 07:15:42 crc kubenswrapper[4971]: I1213 07:15:42.964377 4971 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-scheduler-0"] Dec 13 07:15:43 crc kubenswrapper[4971]: I1213 07:15:43.786848 4971 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="6ae42bbd-1447-4a68-ad90-ae94e2e0b6e5" path="/var/lib/kubelet/pods/6ae42bbd-1447-4a68-ad90-ae94e2e0b6e5/volumes" Dec 13 07:15:43 crc kubenswrapper[4971]: I1213 07:15:43.788212 4971 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="ddbf55d9-778f-4601-bb0d-05a3d394de49" path="/var/lib/kubelet/pods/ddbf55d9-778f-4601-bb0d-05a3d394de49/volumes" Dec 13 07:15:43 crc kubenswrapper[4971]: I1213 07:15:43.954305 4971 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-scheduler-0" event={"ID":"14465eba-34c5-43d3-ad9d-06b454f79963","Type":"ContainerStarted","Data":"6124ee47111872411457d81e80008d760fb1621b3c620790be4eb4b7b5d1f174"} Dec 13 07:15:43 crc kubenswrapper[4971]: I1213 07:15:43.954558 4971 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-scheduler-0" event={"ID":"14465eba-34c5-43d3-ad9d-06b454f79963","Type":"ContainerStarted","Data":"4aed11d9a4901cdc913329541233f95173d5b87e2f41bb40b8defff64460aa48"} Dec 13 07:15:43 crc kubenswrapper[4971]: I1213 07:15:43.962992 4971 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"d4bb47ae-3205-452b-b596-563d3833ee82","Type":"ContainerStarted","Data":"172a79519fbdbf90cf67903e099f96305fd5364468e257e0dfc1842f85d89021"} Dec 13 07:15:43 crc kubenswrapper[4971]: I1213 07:15:43.963029 4971 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"d4bb47ae-3205-452b-b596-563d3833ee82","Type":"ContainerStarted","Data":"269d92f4f9f8d2c3092b010593eaeb4fafa93f43a00a450a44da25861575815c"} Dec 13 07:15:43 crc kubenswrapper[4971]: I1213 07:15:43.968120 4971 generic.go:334] "Generic (PLEG): container finished" podID="ce91d492-58fe-48a0-baa1-e26e736ab49a" containerID="092c2442d88a2d6edbfffc8f033cfd7a1abeb2f3a58b906d2f18d96ae471ca6b" exitCode=0 Dec 13 07:15:43 crc kubenswrapper[4971]: I1213 07:15:43.968173 4971 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"ce91d492-58fe-48a0-baa1-e26e736ab49a","Type":"ContainerDied","Data":"092c2442d88a2d6edbfffc8f033cfd7a1abeb2f3a58b906d2f18d96ae471ca6b"} Dec 13 07:15:43 crc kubenswrapper[4971]: I1213 07:15:43.968210 4971 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"ce91d492-58fe-48a0-baa1-e26e736ab49a","Type":"ContainerDied","Data":"765fd8ecec45e22b725ddb6c03cc9e5def2390806e4b6f5cc5802184b5c0c0bc"} Dec 13 07:15:43 crc kubenswrapper[4971]: I1213 07:15:43.968223 4971 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="765fd8ecec45e22b725ddb6c03cc9e5def2390806e4b6f5cc5802184b5c0c0bc" Dec 13 07:15:43 crc kubenswrapper[4971]: I1213 07:15:43.976760 4971 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-scheduler-0" podStartSLOduration=2.976743612 podStartE2EDuration="2.976743612s" podCreationTimestamp="2025-12-13 07:15:41 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-13 07:15:43.975503212 +0000 UTC m=+1600.579912670" watchObservedRunningTime="2025-12-13 07:15:43.976743612 +0000 UTC m=+1600.581153050" Dec 13 07:15:43 crc kubenswrapper[4971]: I1213 07:15:43.999643 4971 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-metadata-0" podStartSLOduration=2.999616955 podStartE2EDuration="2.999616955s" podCreationTimestamp="2025-12-13 07:15:41 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-13 07:15:43.994186351 +0000 UTC m=+1600.598595799" watchObservedRunningTime="2025-12-13 07:15:43.999616955 +0000 UTC m=+1600.604026403" Dec 13 07:15:44 crc kubenswrapper[4971]: I1213 07:15:44.022273 4971 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-0" Dec 13 07:15:44 crc kubenswrapper[4971]: I1213 07:15:44.077279 4971 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-t747v\" (UniqueName: \"kubernetes.io/projected/ce91d492-58fe-48a0-baa1-e26e736ab49a-kube-api-access-t747v\") pod \"ce91d492-58fe-48a0-baa1-e26e736ab49a\" (UID: \"ce91d492-58fe-48a0-baa1-e26e736ab49a\") " Dec 13 07:15:44 crc kubenswrapper[4971]: I1213 07:15:44.077354 4971 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/ce91d492-58fe-48a0-baa1-e26e736ab49a-internal-tls-certs\") pod \"ce91d492-58fe-48a0-baa1-e26e736ab49a\" (UID: \"ce91d492-58fe-48a0-baa1-e26e736ab49a\") " Dec 13 07:15:44 crc kubenswrapper[4971]: I1213 07:15:44.077395 4971 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/ce91d492-58fe-48a0-baa1-e26e736ab49a-public-tls-certs\") pod \"ce91d492-58fe-48a0-baa1-e26e736ab49a\" (UID: \"ce91d492-58fe-48a0-baa1-e26e736ab49a\") " Dec 13 07:15:44 crc kubenswrapper[4971]: I1213 07:15:44.077449 4971 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/ce91d492-58fe-48a0-baa1-e26e736ab49a-logs\") pod \"ce91d492-58fe-48a0-baa1-e26e736ab49a\" (UID: \"ce91d492-58fe-48a0-baa1-e26e736ab49a\") " Dec 13 07:15:44 crc kubenswrapper[4971]: I1213 07:15:44.077664 4971 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/ce91d492-58fe-48a0-baa1-e26e736ab49a-config-data\") pod \"ce91d492-58fe-48a0-baa1-e26e736ab49a\" (UID: \"ce91d492-58fe-48a0-baa1-e26e736ab49a\") " Dec 13 07:15:44 crc kubenswrapper[4971]: I1213 07:15:44.077716 4971 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/ce91d492-58fe-48a0-baa1-e26e736ab49a-combined-ca-bundle\") pod \"ce91d492-58fe-48a0-baa1-e26e736ab49a\" (UID: \"ce91d492-58fe-48a0-baa1-e26e736ab49a\") " Dec 13 07:15:44 crc kubenswrapper[4971]: I1213 07:15:44.095170 4971 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/ce91d492-58fe-48a0-baa1-e26e736ab49a-logs" (OuterVolumeSpecName: "logs") pod "ce91d492-58fe-48a0-baa1-e26e736ab49a" (UID: "ce91d492-58fe-48a0-baa1-e26e736ab49a"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 13 07:15:44 crc kubenswrapper[4971]: I1213 07:15:44.101801 4971 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/ce91d492-58fe-48a0-baa1-e26e736ab49a-kube-api-access-t747v" (OuterVolumeSpecName: "kube-api-access-t747v") pod "ce91d492-58fe-48a0-baa1-e26e736ab49a" (UID: "ce91d492-58fe-48a0-baa1-e26e736ab49a"). InnerVolumeSpecName "kube-api-access-t747v". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 13 07:15:44 crc kubenswrapper[4971]: I1213 07:15:44.141384 4971 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/ce91d492-58fe-48a0-baa1-e26e736ab49a-public-tls-certs" (OuterVolumeSpecName: "public-tls-certs") pod "ce91d492-58fe-48a0-baa1-e26e736ab49a" (UID: "ce91d492-58fe-48a0-baa1-e26e736ab49a"). InnerVolumeSpecName "public-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 13 07:15:44 crc kubenswrapper[4971]: I1213 07:15:44.156724 4971 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/ce91d492-58fe-48a0-baa1-e26e736ab49a-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "ce91d492-58fe-48a0-baa1-e26e736ab49a" (UID: "ce91d492-58fe-48a0-baa1-e26e736ab49a"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 13 07:15:44 crc kubenswrapper[4971]: I1213 07:15:44.171083 4971 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/ce91d492-58fe-48a0-baa1-e26e736ab49a-config-data" (OuterVolumeSpecName: "config-data") pod "ce91d492-58fe-48a0-baa1-e26e736ab49a" (UID: "ce91d492-58fe-48a0-baa1-e26e736ab49a"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 13 07:15:44 crc kubenswrapper[4971]: I1213 07:15:44.178934 4971 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/ce91d492-58fe-48a0-baa1-e26e736ab49a-internal-tls-certs" (OuterVolumeSpecName: "internal-tls-certs") pod "ce91d492-58fe-48a0-baa1-e26e736ab49a" (UID: "ce91d492-58fe-48a0-baa1-e26e736ab49a"). InnerVolumeSpecName "internal-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 13 07:15:44 crc kubenswrapper[4971]: I1213 07:15:44.179958 4971 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/ce91d492-58fe-48a0-baa1-e26e736ab49a-config-data\") on node \"crc\" DevicePath \"\"" Dec 13 07:15:44 crc kubenswrapper[4971]: I1213 07:15:44.179986 4971 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/ce91d492-58fe-48a0-baa1-e26e736ab49a-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 13 07:15:44 crc kubenswrapper[4971]: I1213 07:15:44.179996 4971 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-t747v\" (UniqueName: \"kubernetes.io/projected/ce91d492-58fe-48a0-baa1-e26e736ab49a-kube-api-access-t747v\") on node \"crc\" DevicePath \"\"" Dec 13 07:15:44 crc kubenswrapper[4971]: I1213 07:15:44.180006 4971 reconciler_common.go:293] "Volume detached for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/ce91d492-58fe-48a0-baa1-e26e736ab49a-internal-tls-certs\") on node \"crc\" DevicePath \"\"" Dec 13 07:15:44 crc kubenswrapper[4971]: I1213 07:15:44.180014 4971 reconciler_common.go:293] "Volume detached for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/ce91d492-58fe-48a0-baa1-e26e736ab49a-public-tls-certs\") on node \"crc\" DevicePath \"\"" Dec 13 07:15:44 crc kubenswrapper[4971]: I1213 07:15:44.180022 4971 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/ce91d492-58fe-48a0-baa1-e26e736ab49a-logs\") on node \"crc\" DevicePath \"\"" Dec 13 07:15:44 crc kubenswrapper[4971]: I1213 07:15:44.977612 4971 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-0" Dec 13 07:15:45 crc kubenswrapper[4971]: I1213 07:15:45.017911 4971 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-api-0"] Dec 13 07:15:45 crc kubenswrapper[4971]: I1213 07:15:45.034004 4971 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-api-0"] Dec 13 07:15:45 crc kubenswrapper[4971]: I1213 07:15:45.044713 4971 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-api-0"] Dec 13 07:15:45 crc kubenswrapper[4971]: E1213 07:15:45.045106 4971 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="ce91d492-58fe-48a0-baa1-e26e736ab49a" containerName="nova-api-log" Dec 13 07:15:45 crc kubenswrapper[4971]: I1213 07:15:45.045125 4971 state_mem.go:107] "Deleted CPUSet assignment" podUID="ce91d492-58fe-48a0-baa1-e26e736ab49a" containerName="nova-api-log" Dec 13 07:15:45 crc kubenswrapper[4971]: E1213 07:15:45.045232 4971 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="ce91d492-58fe-48a0-baa1-e26e736ab49a" containerName="nova-api-api" Dec 13 07:15:45 crc kubenswrapper[4971]: I1213 07:15:45.045241 4971 state_mem.go:107] "Deleted CPUSet assignment" podUID="ce91d492-58fe-48a0-baa1-e26e736ab49a" containerName="nova-api-api" Dec 13 07:15:45 crc kubenswrapper[4971]: I1213 07:15:45.045409 4971 memory_manager.go:354] "RemoveStaleState removing state" podUID="ce91d492-58fe-48a0-baa1-e26e736ab49a" containerName="nova-api-api" Dec 13 07:15:45 crc kubenswrapper[4971]: I1213 07:15:45.045439 4971 memory_manager.go:354] "RemoveStaleState removing state" podUID="ce91d492-58fe-48a0-baa1-e26e736ab49a" containerName="nova-api-log" Dec 13 07:15:45 crc kubenswrapper[4971]: I1213 07:15:45.046350 4971 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-0" Dec 13 07:15:45 crc kubenswrapper[4971]: I1213 07:15:45.050163 4971 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-nova-public-svc" Dec 13 07:15:45 crc kubenswrapper[4971]: I1213 07:15:45.050411 4971 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-nova-internal-svc" Dec 13 07:15:45 crc kubenswrapper[4971]: I1213 07:15:45.051349 4971 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-api-config-data" Dec 13 07:15:45 crc kubenswrapper[4971]: I1213 07:15:45.058091 4971 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-api-0"] Dec 13 07:15:45 crc kubenswrapper[4971]: I1213 07:15:45.093210 4971 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/d66b7a63-bb42-46dc-8dbb-94452070e5b2-public-tls-certs\") pod \"nova-api-0\" (UID: \"d66b7a63-bb42-46dc-8dbb-94452070e5b2\") " pod="openstack/nova-api-0" Dec 13 07:15:45 crc kubenswrapper[4971]: I1213 07:15:45.093259 4971 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d66b7a63-bb42-46dc-8dbb-94452070e5b2-combined-ca-bundle\") pod \"nova-api-0\" (UID: \"d66b7a63-bb42-46dc-8dbb-94452070e5b2\") " pod="openstack/nova-api-0" Dec 13 07:15:45 crc kubenswrapper[4971]: I1213 07:15:45.093334 4971 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-pt2t4\" (UniqueName: \"kubernetes.io/projected/d66b7a63-bb42-46dc-8dbb-94452070e5b2-kube-api-access-pt2t4\") pod \"nova-api-0\" (UID: \"d66b7a63-bb42-46dc-8dbb-94452070e5b2\") " pod="openstack/nova-api-0" Dec 13 07:15:45 crc kubenswrapper[4971]: I1213 07:15:45.093384 4971 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/d66b7a63-bb42-46dc-8dbb-94452070e5b2-internal-tls-certs\") pod \"nova-api-0\" (UID: \"d66b7a63-bb42-46dc-8dbb-94452070e5b2\") " pod="openstack/nova-api-0" Dec 13 07:15:45 crc kubenswrapper[4971]: I1213 07:15:45.093502 4971 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/d66b7a63-bb42-46dc-8dbb-94452070e5b2-config-data\") pod \"nova-api-0\" (UID: \"d66b7a63-bb42-46dc-8dbb-94452070e5b2\") " pod="openstack/nova-api-0" Dec 13 07:15:45 crc kubenswrapper[4971]: I1213 07:15:45.093609 4971 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/d66b7a63-bb42-46dc-8dbb-94452070e5b2-logs\") pod \"nova-api-0\" (UID: \"d66b7a63-bb42-46dc-8dbb-94452070e5b2\") " pod="openstack/nova-api-0" Dec 13 07:15:45 crc kubenswrapper[4971]: I1213 07:15:45.194329 4971 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-pt2t4\" (UniqueName: \"kubernetes.io/projected/d66b7a63-bb42-46dc-8dbb-94452070e5b2-kube-api-access-pt2t4\") pod \"nova-api-0\" (UID: \"d66b7a63-bb42-46dc-8dbb-94452070e5b2\") " pod="openstack/nova-api-0" Dec 13 07:15:45 crc kubenswrapper[4971]: I1213 07:15:45.194418 4971 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/d66b7a63-bb42-46dc-8dbb-94452070e5b2-internal-tls-certs\") pod \"nova-api-0\" (UID: \"d66b7a63-bb42-46dc-8dbb-94452070e5b2\") " pod="openstack/nova-api-0" Dec 13 07:15:45 crc kubenswrapper[4971]: I1213 07:15:45.194476 4971 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/d66b7a63-bb42-46dc-8dbb-94452070e5b2-config-data\") pod \"nova-api-0\" (UID: \"d66b7a63-bb42-46dc-8dbb-94452070e5b2\") " pod="openstack/nova-api-0" Dec 13 07:15:45 crc kubenswrapper[4971]: I1213 07:15:45.194501 4971 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/d66b7a63-bb42-46dc-8dbb-94452070e5b2-logs\") pod \"nova-api-0\" (UID: \"d66b7a63-bb42-46dc-8dbb-94452070e5b2\") " pod="openstack/nova-api-0" Dec 13 07:15:45 crc kubenswrapper[4971]: I1213 07:15:45.194591 4971 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/d66b7a63-bb42-46dc-8dbb-94452070e5b2-public-tls-certs\") pod \"nova-api-0\" (UID: \"d66b7a63-bb42-46dc-8dbb-94452070e5b2\") " pod="openstack/nova-api-0" Dec 13 07:15:45 crc kubenswrapper[4971]: I1213 07:15:45.194628 4971 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d66b7a63-bb42-46dc-8dbb-94452070e5b2-combined-ca-bundle\") pod \"nova-api-0\" (UID: \"d66b7a63-bb42-46dc-8dbb-94452070e5b2\") " pod="openstack/nova-api-0" Dec 13 07:15:45 crc kubenswrapper[4971]: I1213 07:15:45.195153 4971 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/d66b7a63-bb42-46dc-8dbb-94452070e5b2-logs\") pod \"nova-api-0\" (UID: \"d66b7a63-bb42-46dc-8dbb-94452070e5b2\") " pod="openstack/nova-api-0" Dec 13 07:15:45 crc kubenswrapper[4971]: I1213 07:15:45.200041 4971 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/d66b7a63-bb42-46dc-8dbb-94452070e5b2-internal-tls-certs\") pod \"nova-api-0\" (UID: \"d66b7a63-bb42-46dc-8dbb-94452070e5b2\") " pod="openstack/nova-api-0" Dec 13 07:15:45 crc kubenswrapper[4971]: I1213 07:15:45.200805 4971 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d66b7a63-bb42-46dc-8dbb-94452070e5b2-combined-ca-bundle\") pod \"nova-api-0\" (UID: \"d66b7a63-bb42-46dc-8dbb-94452070e5b2\") " pod="openstack/nova-api-0" Dec 13 07:15:45 crc kubenswrapper[4971]: I1213 07:15:45.205146 4971 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/d66b7a63-bb42-46dc-8dbb-94452070e5b2-config-data\") pod \"nova-api-0\" (UID: \"d66b7a63-bb42-46dc-8dbb-94452070e5b2\") " pod="openstack/nova-api-0" Dec 13 07:15:45 crc kubenswrapper[4971]: I1213 07:15:45.211629 4971 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/d66b7a63-bb42-46dc-8dbb-94452070e5b2-public-tls-certs\") pod \"nova-api-0\" (UID: \"d66b7a63-bb42-46dc-8dbb-94452070e5b2\") " pod="openstack/nova-api-0" Dec 13 07:15:45 crc kubenswrapper[4971]: I1213 07:15:45.214846 4971 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-pt2t4\" (UniqueName: \"kubernetes.io/projected/d66b7a63-bb42-46dc-8dbb-94452070e5b2-kube-api-access-pt2t4\") pod \"nova-api-0\" (UID: \"d66b7a63-bb42-46dc-8dbb-94452070e5b2\") " pod="openstack/nova-api-0" Dec 13 07:15:45 crc kubenswrapper[4971]: I1213 07:15:45.364639 4971 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-0" Dec 13 07:15:45 crc kubenswrapper[4971]: I1213 07:15:45.790742 4971 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="ce91d492-58fe-48a0-baa1-e26e736ab49a" path="/var/lib/kubelet/pods/ce91d492-58fe-48a0-baa1-e26e736ab49a/volumes" Dec 13 07:15:45 crc kubenswrapper[4971]: I1213 07:15:45.834274 4971 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-api-0"] Dec 13 07:15:45 crc kubenswrapper[4971]: I1213 07:15:45.988908 4971 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"d66b7a63-bb42-46dc-8dbb-94452070e5b2","Type":"ContainerStarted","Data":"404a281959757c3f208549d427be0be8f00ffb562dcaa3b8bf855ceae4bc06d9"} Dec 13 07:15:47 crc kubenswrapper[4971]: I1213 07:15:47.018051 4971 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"d66b7a63-bb42-46dc-8dbb-94452070e5b2","Type":"ContainerStarted","Data":"3bbc919f671094cbe2512f9b0e5d01af35ee6ecf99fc44b519c9343dde56cd76"} Dec 13 07:15:47 crc kubenswrapper[4971]: I1213 07:15:47.018660 4971 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"d66b7a63-bb42-46dc-8dbb-94452070e5b2","Type":"ContainerStarted","Data":"66af8a32babf42aa4d69220a59ccbebe1a9b53b3e0089da6069305208149be12"} Dec 13 07:15:47 crc kubenswrapper[4971]: I1213 07:15:47.299021 4971 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-metadata-0" Dec 13 07:15:47 crc kubenswrapper[4971]: I1213 07:15:47.299103 4971 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-metadata-0" Dec 13 07:15:47 crc kubenswrapper[4971]: I1213 07:15:47.352785 4971 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-scheduler-0" Dec 13 07:15:49 crc kubenswrapper[4971]: I1213 07:15:49.769593 4971 scope.go:117] "RemoveContainer" containerID="e8beee43d15719dccb6892d6f5711674492425300efa098be89033bc65977d14" Dec 13 07:15:49 crc kubenswrapper[4971]: E1213 07:15:49.770377 4971 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-82xjz_openshift-machine-config-operator(e0c4af12-edda-41d7-9b44-f87396174ef2)\"" pod="openshift-machine-config-operator/machine-config-daemon-82xjz" podUID="e0c4af12-edda-41d7-9b44-f87396174ef2" Dec 13 07:15:52 crc kubenswrapper[4971]: I1213 07:15:52.298618 4971 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/nova-metadata-0" Dec 13 07:15:52 crc kubenswrapper[4971]: I1213 07:15:52.299368 4971 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/nova-metadata-0" Dec 13 07:15:52 crc kubenswrapper[4971]: I1213 07:15:52.353465 4971 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/nova-scheduler-0" Dec 13 07:15:52 crc kubenswrapper[4971]: I1213 07:15:52.378271 4971 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/nova-scheduler-0" Dec 13 07:15:52 crc kubenswrapper[4971]: I1213 07:15:52.393955 4971 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-api-0" podStartSLOduration=7.393940813 podStartE2EDuration="7.393940813s" podCreationTimestamp="2025-12-13 07:15:45 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-13 07:15:47.046213382 +0000 UTC m=+1603.650622830" watchObservedRunningTime="2025-12-13 07:15:52.393940813 +0000 UTC m=+1608.998350261" Dec 13 07:15:53 crc kubenswrapper[4971]: I1213 07:15:53.134734 4971 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/nova-scheduler-0" Dec 13 07:15:53 crc kubenswrapper[4971]: I1213 07:15:53.317713 4971 prober.go:107] "Probe failed" probeType="Startup" pod="openstack/nova-metadata-0" podUID="d4bb47ae-3205-452b-b596-563d3833ee82" containerName="nova-metadata-log" probeResult="failure" output="Get \"https://10.217.0.191:8775/\": net/http: request canceled (Client.Timeout exceeded while awaiting headers)" Dec 13 07:15:53 crc kubenswrapper[4971]: I1213 07:15:53.317719 4971 prober.go:107] "Probe failed" probeType="Startup" pod="openstack/nova-metadata-0" podUID="d4bb47ae-3205-452b-b596-563d3833ee82" containerName="nova-metadata-metadata" probeResult="failure" output="Get \"https://10.217.0.191:8775/\": net/http: request canceled (Client.Timeout exceeded while awaiting headers)" Dec 13 07:15:55 crc kubenswrapper[4971]: I1213 07:15:55.365057 4971 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/nova-api-0" Dec 13 07:15:55 crc kubenswrapper[4971]: I1213 07:15:55.365108 4971 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/nova-api-0" Dec 13 07:15:56 crc kubenswrapper[4971]: I1213 07:15:56.379874 4971 prober.go:107] "Probe failed" probeType="Startup" pod="openstack/nova-api-0" podUID="d66b7a63-bb42-46dc-8dbb-94452070e5b2" containerName="nova-api-log" probeResult="failure" output="Get \"https://10.217.0.193:8774/\": net/http: request canceled (Client.Timeout exceeded while awaiting headers)" Dec 13 07:15:56 crc kubenswrapper[4971]: I1213 07:15:56.379865 4971 prober.go:107] "Probe failed" probeType="Startup" pod="openstack/nova-api-0" podUID="d66b7a63-bb42-46dc-8dbb-94452070e5b2" containerName="nova-api-api" probeResult="failure" output="Get \"https://10.217.0.193:8774/\": net/http: request canceled (Client.Timeout exceeded while awaiting headers)" Dec 13 07:16:01 crc kubenswrapper[4971]: I1213 07:16:01.212793 4971 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/ceilometer-0" Dec 13 07:16:01 crc kubenswrapper[4971]: I1213 07:16:01.768696 4971 scope.go:117] "RemoveContainer" containerID="e8beee43d15719dccb6892d6f5711674492425300efa098be89033bc65977d14" Dec 13 07:16:01 crc kubenswrapper[4971]: E1213 07:16:01.769013 4971 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-82xjz_openshift-machine-config-operator(e0c4af12-edda-41d7-9b44-f87396174ef2)\"" pod="openshift-machine-config-operator/machine-config-daemon-82xjz" podUID="e0c4af12-edda-41d7-9b44-f87396174ef2" Dec 13 07:16:02 crc kubenswrapper[4971]: I1213 07:16:02.304036 4971 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/nova-metadata-0" Dec 13 07:16:02 crc kubenswrapper[4971]: I1213 07:16:02.306621 4971 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/nova-metadata-0" Dec 13 07:16:02 crc kubenswrapper[4971]: I1213 07:16:02.312647 4971 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/nova-metadata-0" Dec 13 07:16:03 crc kubenswrapper[4971]: I1213 07:16:03.283650 4971 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/nova-metadata-0" Dec 13 07:16:05 crc kubenswrapper[4971]: I1213 07:16:05.371634 4971 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/nova-api-0" Dec 13 07:16:05 crc kubenswrapper[4971]: I1213 07:16:05.373156 4971 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-api-0" Dec 13 07:16:05 crc kubenswrapper[4971]: I1213 07:16:05.374003 4971 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/nova-api-0" Dec 13 07:16:05 crc kubenswrapper[4971]: I1213 07:16:05.380284 4971 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/nova-api-0" Dec 13 07:16:06 crc kubenswrapper[4971]: I1213 07:16:06.235430 4971 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-api-0" Dec 13 07:16:06 crc kubenswrapper[4971]: I1213 07:16:06.249172 4971 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/nova-api-0" Dec 13 07:16:15 crc kubenswrapper[4971]: I1213 07:16:15.080831 4971 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/rabbitmq-server-0"] Dec 13 07:16:16 crc kubenswrapper[4971]: I1213 07:16:16.100195 4971 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/rabbitmq-cell1-server-0"] Dec 13 07:16:16 crc kubenswrapper[4971]: I1213 07:16:16.769934 4971 scope.go:117] "RemoveContainer" containerID="e8beee43d15719dccb6892d6f5711674492425300efa098be89033bc65977d14" Dec 13 07:16:16 crc kubenswrapper[4971]: E1213 07:16:16.770681 4971 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-82xjz_openshift-machine-config-operator(e0c4af12-edda-41d7-9b44-f87396174ef2)\"" pod="openshift-machine-config-operator/machine-config-daemon-82xjz" podUID="e0c4af12-edda-41d7-9b44-f87396174ef2" Dec 13 07:16:19 crc kubenswrapper[4971]: I1213 07:16:19.425088 4971 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/rabbitmq-server-0" podUID="a590eb84-c8d5-4cbe-9c03-03a6ccb82073" containerName="rabbitmq" containerID="cri-o://75710b914a6fcd71bf7b0cc53696f512ac865a4da1eb06734a6db6ff35c892ad" gracePeriod=604796 Dec 13 07:16:20 crc kubenswrapper[4971]: I1213 07:16:20.380372 4971 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/rabbitmq-cell1-server-0" podUID="5eed809d-c915-433c-9915-27e0b01e1ffe" containerName="rabbitmq" containerID="cri-o://efed32724d55162da7a67edf770832ac164e124b6b49db25e2f7b8b7ec033c44" gracePeriod=604796 Dec 13 07:16:25 crc kubenswrapper[4971]: I1213 07:16:25.979281 4971 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/rabbitmq-server-0" Dec 13 07:16:25 crc kubenswrapper[4971]: I1213 07:16:25.980995 4971 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"rabbitmq-erlang-cookie\" (UniqueName: \"kubernetes.io/empty-dir/a590eb84-c8d5-4cbe-9c03-03a6ccb82073-rabbitmq-erlang-cookie\") pod \"a590eb84-c8d5-4cbe-9c03-03a6ccb82073\" (UID: \"a590eb84-c8d5-4cbe-9c03-03a6ccb82073\") " Dec 13 07:16:25 crc kubenswrapper[4971]: I1213 07:16:25.981040 4971 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"rabbitmq-confd\" (UniqueName: \"kubernetes.io/projected/a590eb84-c8d5-4cbe-9c03-03a6ccb82073-rabbitmq-confd\") pod \"a590eb84-c8d5-4cbe-9c03-03a6ccb82073\" (UID: \"a590eb84-c8d5-4cbe-9c03-03a6ccb82073\") " Dec 13 07:16:25 crc kubenswrapper[4971]: I1213 07:16:25.981088 4971 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"rabbitmq-plugins\" (UniqueName: \"kubernetes.io/empty-dir/a590eb84-c8d5-4cbe-9c03-03a6ccb82073-rabbitmq-plugins\") pod \"a590eb84-c8d5-4cbe-9c03-03a6ccb82073\" (UID: \"a590eb84-c8d5-4cbe-9c03-03a6ccb82073\") " Dec 13 07:16:25 crc kubenswrapper[4971]: I1213 07:16:25.981124 4971 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"rabbitmq-tls\" (UniqueName: \"kubernetes.io/projected/a590eb84-c8d5-4cbe-9c03-03a6ccb82073-rabbitmq-tls\") pod \"a590eb84-c8d5-4cbe-9c03-03a6ccb82073\" (UID: \"a590eb84-c8d5-4cbe-9c03-03a6ccb82073\") " Dec 13 07:16:25 crc kubenswrapper[4971]: I1213 07:16:25.981763 4971 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/a590eb84-c8d5-4cbe-9c03-03a6ccb82073-rabbitmq-erlang-cookie" (OuterVolumeSpecName: "rabbitmq-erlang-cookie") pod "a590eb84-c8d5-4cbe-9c03-03a6ccb82073" (UID: "a590eb84-c8d5-4cbe-9c03-03a6ccb82073"). InnerVolumeSpecName "rabbitmq-erlang-cookie". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 13 07:16:25 crc kubenswrapper[4971]: I1213 07:16:25.981872 4971 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/a590eb84-c8d5-4cbe-9c03-03a6ccb82073-rabbitmq-plugins" (OuterVolumeSpecName: "rabbitmq-plugins") pod "a590eb84-c8d5-4cbe-9c03-03a6ccb82073" (UID: "a590eb84-c8d5-4cbe-9c03-03a6ccb82073"). InnerVolumeSpecName "rabbitmq-plugins". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 13 07:16:25 crc kubenswrapper[4971]: I1213 07:16:25.986208 4971 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/a590eb84-c8d5-4cbe-9c03-03a6ccb82073-rabbitmq-tls" (OuterVolumeSpecName: "rabbitmq-tls") pod "a590eb84-c8d5-4cbe-9c03-03a6ccb82073" (UID: "a590eb84-c8d5-4cbe-9c03-03a6ccb82073"). InnerVolumeSpecName "rabbitmq-tls". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 13 07:16:26 crc kubenswrapper[4971]: I1213 07:16:26.082562 4971 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-wfzdp\" (UniqueName: \"kubernetes.io/projected/a590eb84-c8d5-4cbe-9c03-03a6ccb82073-kube-api-access-wfzdp\") pod \"a590eb84-c8d5-4cbe-9c03-03a6ccb82073\" (UID: \"a590eb84-c8d5-4cbe-9c03-03a6ccb82073\") " Dec 13 07:16:26 crc kubenswrapper[4971]: I1213 07:16:26.082650 4971 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"plugins-conf\" (UniqueName: \"kubernetes.io/configmap/a590eb84-c8d5-4cbe-9c03-03a6ccb82073-plugins-conf\") pod \"a590eb84-c8d5-4cbe-9c03-03a6ccb82073\" (UID: \"a590eb84-c8d5-4cbe-9c03-03a6ccb82073\") " Dec 13 07:16:26 crc kubenswrapper[4971]: I1213 07:16:26.082698 4971 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pod-info\" (UniqueName: \"kubernetes.io/downward-api/a590eb84-c8d5-4cbe-9c03-03a6ccb82073-pod-info\") pod \"a590eb84-c8d5-4cbe-9c03-03a6ccb82073\" (UID: \"a590eb84-c8d5-4cbe-9c03-03a6ccb82073\") " Dec 13 07:16:26 crc kubenswrapper[4971]: I1213 07:16:26.082742 4971 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"erlang-cookie-secret\" (UniqueName: \"kubernetes.io/secret/a590eb84-c8d5-4cbe-9c03-03a6ccb82073-erlang-cookie-secret\") pod \"a590eb84-c8d5-4cbe-9c03-03a6ccb82073\" (UID: \"a590eb84-c8d5-4cbe-9c03-03a6ccb82073\") " Dec 13 07:16:26 crc kubenswrapper[4971]: I1213 07:16:26.082770 4971 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/a590eb84-c8d5-4cbe-9c03-03a6ccb82073-config-data\") pod \"a590eb84-c8d5-4cbe-9c03-03a6ccb82073\" (UID: \"a590eb84-c8d5-4cbe-9c03-03a6ccb82073\") " Dec 13 07:16:26 crc kubenswrapper[4971]: I1213 07:16:26.082827 4971 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"persistence\" (UniqueName: \"kubernetes.io/local-volume/local-storage06-crc\") pod \"a590eb84-c8d5-4cbe-9c03-03a6ccb82073\" (UID: \"a590eb84-c8d5-4cbe-9c03-03a6ccb82073\") " Dec 13 07:16:26 crc kubenswrapper[4971]: I1213 07:16:26.082875 4971 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"server-conf\" (UniqueName: \"kubernetes.io/configmap/a590eb84-c8d5-4cbe-9c03-03a6ccb82073-server-conf\") pod \"a590eb84-c8d5-4cbe-9c03-03a6ccb82073\" (UID: \"a590eb84-c8d5-4cbe-9c03-03a6ccb82073\") " Dec 13 07:16:26 crc kubenswrapper[4971]: I1213 07:16:26.083348 4971 reconciler_common.go:293] "Volume detached for volume \"rabbitmq-erlang-cookie\" (UniqueName: \"kubernetes.io/empty-dir/a590eb84-c8d5-4cbe-9c03-03a6ccb82073-rabbitmq-erlang-cookie\") on node \"crc\" DevicePath \"\"" Dec 13 07:16:26 crc kubenswrapper[4971]: I1213 07:16:26.083374 4971 reconciler_common.go:293] "Volume detached for volume \"rabbitmq-plugins\" (UniqueName: \"kubernetes.io/empty-dir/a590eb84-c8d5-4cbe-9c03-03a6ccb82073-rabbitmq-plugins\") on node \"crc\" DevicePath \"\"" Dec 13 07:16:26 crc kubenswrapper[4971]: I1213 07:16:26.083386 4971 reconciler_common.go:293] "Volume detached for volume \"rabbitmq-tls\" (UniqueName: \"kubernetes.io/projected/a590eb84-c8d5-4cbe-9c03-03a6ccb82073-rabbitmq-tls\") on node \"crc\" DevicePath \"\"" Dec 13 07:16:26 crc kubenswrapper[4971]: I1213 07:16:26.086753 4971 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/a590eb84-c8d5-4cbe-9c03-03a6ccb82073-erlang-cookie-secret" (OuterVolumeSpecName: "erlang-cookie-secret") pod "a590eb84-c8d5-4cbe-9c03-03a6ccb82073" (UID: "a590eb84-c8d5-4cbe-9c03-03a6ccb82073"). InnerVolumeSpecName "erlang-cookie-secret". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 13 07:16:26 crc kubenswrapper[4971]: I1213 07:16:26.090114 4971 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/a590eb84-c8d5-4cbe-9c03-03a6ccb82073-plugins-conf" (OuterVolumeSpecName: "plugins-conf") pod "a590eb84-c8d5-4cbe-9c03-03a6ccb82073" (UID: "a590eb84-c8d5-4cbe-9c03-03a6ccb82073"). InnerVolumeSpecName "plugins-conf". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 13 07:16:26 crc kubenswrapper[4971]: I1213 07:16:26.102716 4971 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/local-volume/local-storage06-crc" (OuterVolumeSpecName: "persistence") pod "a590eb84-c8d5-4cbe-9c03-03a6ccb82073" (UID: "a590eb84-c8d5-4cbe-9c03-03a6ccb82073"). InnerVolumeSpecName "local-storage06-crc". PluginName "kubernetes.io/local-volume", VolumeGidValue "" Dec 13 07:16:26 crc kubenswrapper[4971]: I1213 07:16:26.110828 4971 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/a590eb84-c8d5-4cbe-9c03-03a6ccb82073-kube-api-access-wfzdp" (OuterVolumeSpecName: "kube-api-access-wfzdp") pod "a590eb84-c8d5-4cbe-9c03-03a6ccb82073" (UID: "a590eb84-c8d5-4cbe-9c03-03a6ccb82073"). InnerVolumeSpecName "kube-api-access-wfzdp". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 13 07:16:26 crc kubenswrapper[4971]: I1213 07:16:26.112830 4971 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/downward-api/a590eb84-c8d5-4cbe-9c03-03a6ccb82073-pod-info" (OuterVolumeSpecName: "pod-info") pod "a590eb84-c8d5-4cbe-9c03-03a6ccb82073" (UID: "a590eb84-c8d5-4cbe-9c03-03a6ccb82073"). InnerVolumeSpecName "pod-info". PluginName "kubernetes.io/downward-api", VolumeGidValue "" Dec 13 07:16:26 crc kubenswrapper[4971]: I1213 07:16:26.134776 4971 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/a590eb84-c8d5-4cbe-9c03-03a6ccb82073-rabbitmq-confd" (OuterVolumeSpecName: "rabbitmq-confd") pod "a590eb84-c8d5-4cbe-9c03-03a6ccb82073" (UID: "a590eb84-c8d5-4cbe-9c03-03a6ccb82073"). InnerVolumeSpecName "rabbitmq-confd". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 13 07:16:26 crc kubenswrapper[4971]: I1213 07:16:26.171203 4971 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/a590eb84-c8d5-4cbe-9c03-03a6ccb82073-config-data" (OuterVolumeSpecName: "config-data") pod "a590eb84-c8d5-4cbe-9c03-03a6ccb82073" (UID: "a590eb84-c8d5-4cbe-9c03-03a6ccb82073"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 13 07:16:26 crc kubenswrapper[4971]: I1213 07:16:26.194068 4971 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-wfzdp\" (UniqueName: \"kubernetes.io/projected/a590eb84-c8d5-4cbe-9c03-03a6ccb82073-kube-api-access-wfzdp\") on node \"crc\" DevicePath \"\"" Dec 13 07:16:26 crc kubenswrapper[4971]: I1213 07:16:26.194112 4971 reconciler_common.go:293] "Volume detached for volume \"plugins-conf\" (UniqueName: \"kubernetes.io/configmap/a590eb84-c8d5-4cbe-9c03-03a6ccb82073-plugins-conf\") on node \"crc\" DevicePath \"\"" Dec 13 07:16:26 crc kubenswrapper[4971]: I1213 07:16:26.194128 4971 reconciler_common.go:293] "Volume detached for volume \"pod-info\" (UniqueName: \"kubernetes.io/downward-api/a590eb84-c8d5-4cbe-9c03-03a6ccb82073-pod-info\") on node \"crc\" DevicePath \"\"" Dec 13 07:16:26 crc kubenswrapper[4971]: I1213 07:16:26.194138 4971 reconciler_common.go:293] "Volume detached for volume \"erlang-cookie-secret\" (UniqueName: \"kubernetes.io/secret/a590eb84-c8d5-4cbe-9c03-03a6ccb82073-erlang-cookie-secret\") on node \"crc\" DevicePath \"\"" Dec 13 07:16:26 crc kubenswrapper[4971]: I1213 07:16:26.194150 4971 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/a590eb84-c8d5-4cbe-9c03-03a6ccb82073-config-data\") on node \"crc\" DevicePath \"\"" Dec 13 07:16:26 crc kubenswrapper[4971]: I1213 07:16:26.194176 4971 reconciler_common.go:286] "operationExecutor.UnmountDevice started for volume \"local-storage06-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage06-crc\") on node \"crc\" " Dec 13 07:16:26 crc kubenswrapper[4971]: I1213 07:16:26.194188 4971 reconciler_common.go:293] "Volume detached for volume \"rabbitmq-confd\" (UniqueName: \"kubernetes.io/projected/a590eb84-c8d5-4cbe-9c03-03a6ccb82073-rabbitmq-confd\") on node \"crc\" DevicePath \"\"" Dec 13 07:16:26 crc kubenswrapper[4971]: I1213 07:16:26.251632 4971 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/a590eb84-c8d5-4cbe-9c03-03a6ccb82073-server-conf" (OuterVolumeSpecName: "server-conf") pod "a590eb84-c8d5-4cbe-9c03-03a6ccb82073" (UID: "a590eb84-c8d5-4cbe-9c03-03a6ccb82073"). InnerVolumeSpecName "server-conf". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 13 07:16:26 crc kubenswrapper[4971]: I1213 07:16:26.260344 4971 operation_generator.go:917] UnmountDevice succeeded for volume "local-storage06-crc" (UniqueName: "kubernetes.io/local-volume/local-storage06-crc") on node "crc" Dec 13 07:16:26 crc kubenswrapper[4971]: I1213 07:16:26.295473 4971 reconciler_common.go:293] "Volume detached for volume \"local-storage06-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage06-crc\") on node \"crc\" DevicePath \"\"" Dec 13 07:16:26 crc kubenswrapper[4971]: I1213 07:16:26.295508 4971 reconciler_common.go:293] "Volume detached for volume \"server-conf\" (UniqueName: \"kubernetes.io/configmap/a590eb84-c8d5-4cbe-9c03-03a6ccb82073-server-conf\") on node \"crc\" DevicePath \"\"" Dec 13 07:16:26 crc kubenswrapper[4971]: I1213 07:16:26.444530 4971 generic.go:334] "Generic (PLEG): container finished" podID="a590eb84-c8d5-4cbe-9c03-03a6ccb82073" containerID="75710b914a6fcd71bf7b0cc53696f512ac865a4da1eb06734a6db6ff35c892ad" exitCode=0 Dec 13 07:16:26 crc kubenswrapper[4971]: I1213 07:16:26.444577 4971 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-server-0" event={"ID":"a590eb84-c8d5-4cbe-9c03-03a6ccb82073","Type":"ContainerDied","Data":"75710b914a6fcd71bf7b0cc53696f512ac865a4da1eb06734a6db6ff35c892ad"} Dec 13 07:16:26 crc kubenswrapper[4971]: I1213 07:16:26.444593 4971 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/rabbitmq-server-0" Dec 13 07:16:26 crc kubenswrapper[4971]: I1213 07:16:26.444605 4971 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-server-0" event={"ID":"a590eb84-c8d5-4cbe-9c03-03a6ccb82073","Type":"ContainerDied","Data":"b3d06f90258a1549aba7fc0a8fdb6a60a2048a29be25f55fa06f83a5ee5cd52c"} Dec 13 07:16:26 crc kubenswrapper[4971]: I1213 07:16:26.444623 4971 scope.go:117] "RemoveContainer" containerID="75710b914a6fcd71bf7b0cc53696f512ac865a4da1eb06734a6db6ff35c892ad" Dec 13 07:16:26 crc kubenswrapper[4971]: I1213 07:16:26.468307 4971 scope.go:117] "RemoveContainer" containerID="2632986cd6a0be018d82fef579687199e356535dfe73855c9af020cc784bc892" Dec 13 07:16:26 crc kubenswrapper[4971]: I1213 07:16:26.486170 4971 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/rabbitmq-server-0"] Dec 13 07:16:26 crc kubenswrapper[4971]: I1213 07:16:26.493092 4971 scope.go:117] "RemoveContainer" containerID="75710b914a6fcd71bf7b0cc53696f512ac865a4da1eb06734a6db6ff35c892ad" Dec 13 07:16:26 crc kubenswrapper[4971]: E1213 07:16:26.494768 4971 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"75710b914a6fcd71bf7b0cc53696f512ac865a4da1eb06734a6db6ff35c892ad\": container with ID starting with 75710b914a6fcd71bf7b0cc53696f512ac865a4da1eb06734a6db6ff35c892ad not found: ID does not exist" containerID="75710b914a6fcd71bf7b0cc53696f512ac865a4da1eb06734a6db6ff35c892ad" Dec 13 07:16:26 crc kubenswrapper[4971]: I1213 07:16:26.494831 4971 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"75710b914a6fcd71bf7b0cc53696f512ac865a4da1eb06734a6db6ff35c892ad"} err="failed to get container status \"75710b914a6fcd71bf7b0cc53696f512ac865a4da1eb06734a6db6ff35c892ad\": rpc error: code = NotFound desc = could not find container \"75710b914a6fcd71bf7b0cc53696f512ac865a4da1eb06734a6db6ff35c892ad\": container with ID starting with 75710b914a6fcd71bf7b0cc53696f512ac865a4da1eb06734a6db6ff35c892ad not found: ID does not exist" Dec 13 07:16:26 crc kubenswrapper[4971]: I1213 07:16:26.494863 4971 scope.go:117] "RemoveContainer" containerID="2632986cd6a0be018d82fef579687199e356535dfe73855c9af020cc784bc892" Dec 13 07:16:26 crc kubenswrapper[4971]: E1213 07:16:26.495222 4971 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"2632986cd6a0be018d82fef579687199e356535dfe73855c9af020cc784bc892\": container with ID starting with 2632986cd6a0be018d82fef579687199e356535dfe73855c9af020cc784bc892 not found: ID does not exist" containerID="2632986cd6a0be018d82fef579687199e356535dfe73855c9af020cc784bc892" Dec 13 07:16:26 crc kubenswrapper[4971]: I1213 07:16:26.495275 4971 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"2632986cd6a0be018d82fef579687199e356535dfe73855c9af020cc784bc892"} err="failed to get container status \"2632986cd6a0be018d82fef579687199e356535dfe73855c9af020cc784bc892\": rpc error: code = NotFound desc = could not find container \"2632986cd6a0be018d82fef579687199e356535dfe73855c9af020cc784bc892\": container with ID starting with 2632986cd6a0be018d82fef579687199e356535dfe73855c9af020cc784bc892 not found: ID does not exist" Dec 13 07:16:26 crc kubenswrapper[4971]: I1213 07:16:26.504160 4971 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/rabbitmq-server-0"] Dec 13 07:16:26 crc kubenswrapper[4971]: I1213 07:16:26.515572 4971 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/rabbitmq-server-0"] Dec 13 07:16:26 crc kubenswrapper[4971]: E1213 07:16:26.516011 4971 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="a590eb84-c8d5-4cbe-9c03-03a6ccb82073" containerName="rabbitmq" Dec 13 07:16:26 crc kubenswrapper[4971]: I1213 07:16:26.516041 4971 state_mem.go:107] "Deleted CPUSet assignment" podUID="a590eb84-c8d5-4cbe-9c03-03a6ccb82073" containerName="rabbitmq" Dec 13 07:16:26 crc kubenswrapper[4971]: E1213 07:16:26.516063 4971 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="a590eb84-c8d5-4cbe-9c03-03a6ccb82073" containerName="setup-container" Dec 13 07:16:26 crc kubenswrapper[4971]: I1213 07:16:26.516077 4971 state_mem.go:107] "Deleted CPUSet assignment" podUID="a590eb84-c8d5-4cbe-9c03-03a6ccb82073" containerName="setup-container" Dec 13 07:16:26 crc kubenswrapper[4971]: I1213 07:16:26.516356 4971 memory_manager.go:354] "RemoveStaleState removing state" podUID="a590eb84-c8d5-4cbe-9c03-03a6ccb82073" containerName="rabbitmq" Dec 13 07:16:26 crc kubenswrapper[4971]: I1213 07:16:26.517604 4971 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/rabbitmq-server-0" Dec 13 07:16:26 crc kubenswrapper[4971]: I1213 07:16:26.521427 4971 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"rabbitmq-plugins-conf" Dec 13 07:16:26 crc kubenswrapper[4971]: I1213 07:16:26.521485 4971 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"rabbitmq-config-data" Dec 13 07:16:26 crc kubenswrapper[4971]: I1213 07:16:26.521956 4971 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-rabbitmq-svc" Dec 13 07:16:26 crc kubenswrapper[4971]: I1213 07:16:26.522056 4971 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"rabbitmq-server-conf" Dec 13 07:16:26 crc kubenswrapper[4971]: I1213 07:16:26.522145 4971 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"rabbitmq-erlang-cookie" Dec 13 07:16:26 crc kubenswrapper[4971]: I1213 07:16:26.522153 4971 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"rabbitmq-server-dockercfg-vzbd9" Dec 13 07:16:26 crc kubenswrapper[4971]: I1213 07:16:26.522994 4971 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"rabbitmq-default-user" Dec 13 07:16:26 crc kubenswrapper[4971]: I1213 07:16:26.533134 4971 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/rabbitmq-server-0"] Dec 13 07:16:26 crc kubenswrapper[4971]: I1213 07:16:26.601742 4971 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"server-conf\" (UniqueName: \"kubernetes.io/configmap/42eb5011-6651-43cb-a562-c89cc6df2c53-server-conf\") pod \"rabbitmq-server-0\" (UID: \"42eb5011-6651-43cb-a562-c89cc6df2c53\") " pod="openstack/rabbitmq-server-0" Dec 13 07:16:26 crc kubenswrapper[4971]: I1213 07:16:26.601814 4971 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rabbitmq-erlang-cookie\" (UniqueName: \"kubernetes.io/empty-dir/42eb5011-6651-43cb-a562-c89cc6df2c53-rabbitmq-erlang-cookie\") pod \"rabbitmq-server-0\" (UID: \"42eb5011-6651-43cb-a562-c89cc6df2c53\") " pod="openstack/rabbitmq-server-0" Dec 13 07:16:26 crc kubenswrapper[4971]: I1213 07:16:26.601835 4971 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rabbitmq-plugins\" (UniqueName: \"kubernetes.io/empty-dir/42eb5011-6651-43cb-a562-c89cc6df2c53-rabbitmq-plugins\") pod \"rabbitmq-server-0\" (UID: \"42eb5011-6651-43cb-a562-c89cc6df2c53\") " pod="openstack/rabbitmq-server-0" Dec 13 07:16:26 crc kubenswrapper[4971]: I1213 07:16:26.601961 4971 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"local-storage06-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage06-crc\") pod \"rabbitmq-server-0\" (UID: \"42eb5011-6651-43cb-a562-c89cc6df2c53\") " pod="openstack/rabbitmq-server-0" Dec 13 07:16:26 crc kubenswrapper[4971]: I1213 07:16:26.602015 4971 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"plugins-conf\" (UniqueName: \"kubernetes.io/configmap/42eb5011-6651-43cb-a562-c89cc6df2c53-plugins-conf\") pod \"rabbitmq-server-0\" (UID: \"42eb5011-6651-43cb-a562-c89cc6df2c53\") " pod="openstack/rabbitmq-server-0" Dec 13 07:16:26 crc kubenswrapper[4971]: I1213 07:16:26.602061 4971 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rabbitmq-confd\" (UniqueName: \"kubernetes.io/projected/42eb5011-6651-43cb-a562-c89cc6df2c53-rabbitmq-confd\") pod \"rabbitmq-server-0\" (UID: \"42eb5011-6651-43cb-a562-c89cc6df2c53\") " pod="openstack/rabbitmq-server-0" Dec 13 07:16:26 crc kubenswrapper[4971]: I1213 07:16:26.602090 4971 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rabbitmq-tls\" (UniqueName: \"kubernetes.io/projected/42eb5011-6651-43cb-a562-c89cc6df2c53-rabbitmq-tls\") pod \"rabbitmq-server-0\" (UID: \"42eb5011-6651-43cb-a562-c89cc6df2c53\") " pod="openstack/rabbitmq-server-0" Dec 13 07:16:26 crc kubenswrapper[4971]: I1213 07:16:26.602271 4971 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"pod-info\" (UniqueName: \"kubernetes.io/downward-api/42eb5011-6651-43cb-a562-c89cc6df2c53-pod-info\") pod \"rabbitmq-server-0\" (UID: \"42eb5011-6651-43cb-a562-c89cc6df2c53\") " pod="openstack/rabbitmq-server-0" Dec 13 07:16:26 crc kubenswrapper[4971]: I1213 07:16:26.602315 4971 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-2s5b6\" (UniqueName: \"kubernetes.io/projected/42eb5011-6651-43cb-a562-c89cc6df2c53-kube-api-access-2s5b6\") pod \"rabbitmq-server-0\" (UID: \"42eb5011-6651-43cb-a562-c89cc6df2c53\") " pod="openstack/rabbitmq-server-0" Dec 13 07:16:26 crc kubenswrapper[4971]: I1213 07:16:26.602578 4971 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"erlang-cookie-secret\" (UniqueName: \"kubernetes.io/secret/42eb5011-6651-43cb-a562-c89cc6df2c53-erlang-cookie-secret\") pod \"rabbitmq-server-0\" (UID: \"42eb5011-6651-43cb-a562-c89cc6df2c53\") " pod="openstack/rabbitmq-server-0" Dec 13 07:16:26 crc kubenswrapper[4971]: I1213 07:16:26.602649 4971 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/42eb5011-6651-43cb-a562-c89cc6df2c53-config-data\") pod \"rabbitmq-server-0\" (UID: \"42eb5011-6651-43cb-a562-c89cc6df2c53\") " pod="openstack/rabbitmq-server-0" Dec 13 07:16:26 crc kubenswrapper[4971]: I1213 07:16:26.703617 4971 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"erlang-cookie-secret\" (UniqueName: \"kubernetes.io/secret/42eb5011-6651-43cb-a562-c89cc6df2c53-erlang-cookie-secret\") pod \"rabbitmq-server-0\" (UID: \"42eb5011-6651-43cb-a562-c89cc6df2c53\") " pod="openstack/rabbitmq-server-0" Dec 13 07:16:26 crc kubenswrapper[4971]: I1213 07:16:26.703650 4971 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/42eb5011-6651-43cb-a562-c89cc6df2c53-config-data\") pod \"rabbitmq-server-0\" (UID: \"42eb5011-6651-43cb-a562-c89cc6df2c53\") " pod="openstack/rabbitmq-server-0" Dec 13 07:16:26 crc kubenswrapper[4971]: I1213 07:16:26.703679 4971 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"server-conf\" (UniqueName: \"kubernetes.io/configmap/42eb5011-6651-43cb-a562-c89cc6df2c53-server-conf\") pod \"rabbitmq-server-0\" (UID: \"42eb5011-6651-43cb-a562-c89cc6df2c53\") " pod="openstack/rabbitmq-server-0" Dec 13 07:16:26 crc kubenswrapper[4971]: I1213 07:16:26.703727 4971 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rabbitmq-erlang-cookie\" (UniqueName: \"kubernetes.io/empty-dir/42eb5011-6651-43cb-a562-c89cc6df2c53-rabbitmq-erlang-cookie\") pod \"rabbitmq-server-0\" (UID: \"42eb5011-6651-43cb-a562-c89cc6df2c53\") " pod="openstack/rabbitmq-server-0" Dec 13 07:16:26 crc kubenswrapper[4971]: I1213 07:16:26.703745 4971 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rabbitmq-plugins\" (UniqueName: \"kubernetes.io/empty-dir/42eb5011-6651-43cb-a562-c89cc6df2c53-rabbitmq-plugins\") pod \"rabbitmq-server-0\" (UID: \"42eb5011-6651-43cb-a562-c89cc6df2c53\") " pod="openstack/rabbitmq-server-0" Dec 13 07:16:26 crc kubenswrapper[4971]: I1213 07:16:26.703791 4971 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"local-storage06-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage06-crc\") pod \"rabbitmq-server-0\" (UID: \"42eb5011-6651-43cb-a562-c89cc6df2c53\") " pod="openstack/rabbitmq-server-0" Dec 13 07:16:26 crc kubenswrapper[4971]: I1213 07:16:26.703813 4971 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"plugins-conf\" (UniqueName: \"kubernetes.io/configmap/42eb5011-6651-43cb-a562-c89cc6df2c53-plugins-conf\") pod \"rabbitmq-server-0\" (UID: \"42eb5011-6651-43cb-a562-c89cc6df2c53\") " pod="openstack/rabbitmq-server-0" Dec 13 07:16:26 crc kubenswrapper[4971]: I1213 07:16:26.703832 4971 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rabbitmq-confd\" (UniqueName: \"kubernetes.io/projected/42eb5011-6651-43cb-a562-c89cc6df2c53-rabbitmq-confd\") pod \"rabbitmq-server-0\" (UID: \"42eb5011-6651-43cb-a562-c89cc6df2c53\") " pod="openstack/rabbitmq-server-0" Dec 13 07:16:26 crc kubenswrapper[4971]: I1213 07:16:26.703849 4971 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rabbitmq-tls\" (UniqueName: \"kubernetes.io/projected/42eb5011-6651-43cb-a562-c89cc6df2c53-rabbitmq-tls\") pod \"rabbitmq-server-0\" (UID: \"42eb5011-6651-43cb-a562-c89cc6df2c53\") " pod="openstack/rabbitmq-server-0" Dec 13 07:16:26 crc kubenswrapper[4971]: I1213 07:16:26.703866 4971 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pod-info\" (UniqueName: \"kubernetes.io/downward-api/42eb5011-6651-43cb-a562-c89cc6df2c53-pod-info\") pod \"rabbitmq-server-0\" (UID: \"42eb5011-6651-43cb-a562-c89cc6df2c53\") " pod="openstack/rabbitmq-server-0" Dec 13 07:16:26 crc kubenswrapper[4971]: I1213 07:16:26.703881 4971 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-2s5b6\" (UniqueName: \"kubernetes.io/projected/42eb5011-6651-43cb-a562-c89cc6df2c53-kube-api-access-2s5b6\") pod \"rabbitmq-server-0\" (UID: \"42eb5011-6651-43cb-a562-c89cc6df2c53\") " pod="openstack/rabbitmq-server-0" Dec 13 07:16:26 crc kubenswrapper[4971]: I1213 07:16:26.704221 4971 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"local-storage06-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage06-crc\") pod \"rabbitmq-server-0\" (UID: \"42eb5011-6651-43cb-a562-c89cc6df2c53\") device mount path \"/mnt/openstack/pv06\"" pod="openstack/rabbitmq-server-0" Dec 13 07:16:26 crc kubenswrapper[4971]: I1213 07:16:26.704463 4971 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rabbitmq-plugins\" (UniqueName: \"kubernetes.io/empty-dir/42eb5011-6651-43cb-a562-c89cc6df2c53-rabbitmq-plugins\") pod \"rabbitmq-server-0\" (UID: \"42eb5011-6651-43cb-a562-c89cc6df2c53\") " pod="openstack/rabbitmq-server-0" Dec 13 07:16:26 crc kubenswrapper[4971]: I1213 07:16:26.704604 4971 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rabbitmq-erlang-cookie\" (UniqueName: \"kubernetes.io/empty-dir/42eb5011-6651-43cb-a562-c89cc6df2c53-rabbitmq-erlang-cookie\") pod \"rabbitmq-server-0\" (UID: \"42eb5011-6651-43cb-a562-c89cc6df2c53\") " pod="openstack/rabbitmq-server-0" Dec 13 07:16:26 crc kubenswrapper[4971]: I1213 07:16:26.704755 4971 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/42eb5011-6651-43cb-a562-c89cc6df2c53-config-data\") pod \"rabbitmq-server-0\" (UID: \"42eb5011-6651-43cb-a562-c89cc6df2c53\") " pod="openstack/rabbitmq-server-0" Dec 13 07:16:26 crc kubenswrapper[4971]: I1213 07:16:26.704988 4971 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"plugins-conf\" (UniqueName: \"kubernetes.io/configmap/42eb5011-6651-43cb-a562-c89cc6df2c53-plugins-conf\") pod \"rabbitmq-server-0\" (UID: \"42eb5011-6651-43cb-a562-c89cc6df2c53\") " pod="openstack/rabbitmq-server-0" Dec 13 07:16:26 crc kubenswrapper[4971]: I1213 07:16:26.705237 4971 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"server-conf\" (UniqueName: \"kubernetes.io/configmap/42eb5011-6651-43cb-a562-c89cc6df2c53-server-conf\") pod \"rabbitmq-server-0\" (UID: \"42eb5011-6651-43cb-a562-c89cc6df2c53\") " pod="openstack/rabbitmq-server-0" Dec 13 07:16:26 crc kubenswrapper[4971]: I1213 07:16:26.708582 4971 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rabbitmq-confd\" (UniqueName: \"kubernetes.io/projected/42eb5011-6651-43cb-a562-c89cc6df2c53-rabbitmq-confd\") pod \"rabbitmq-server-0\" (UID: \"42eb5011-6651-43cb-a562-c89cc6df2c53\") " pod="openstack/rabbitmq-server-0" Dec 13 07:16:26 crc kubenswrapper[4971]: I1213 07:16:26.709021 4971 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"erlang-cookie-secret\" (UniqueName: \"kubernetes.io/secret/42eb5011-6651-43cb-a562-c89cc6df2c53-erlang-cookie-secret\") pod \"rabbitmq-server-0\" (UID: \"42eb5011-6651-43cb-a562-c89cc6df2c53\") " pod="openstack/rabbitmq-server-0" Dec 13 07:16:26 crc kubenswrapper[4971]: I1213 07:16:26.709997 4971 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rabbitmq-tls\" (UniqueName: \"kubernetes.io/projected/42eb5011-6651-43cb-a562-c89cc6df2c53-rabbitmq-tls\") pod \"rabbitmq-server-0\" (UID: \"42eb5011-6651-43cb-a562-c89cc6df2c53\") " pod="openstack/rabbitmq-server-0" Dec 13 07:16:26 crc kubenswrapper[4971]: I1213 07:16:26.713030 4971 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"pod-info\" (UniqueName: \"kubernetes.io/downward-api/42eb5011-6651-43cb-a562-c89cc6df2c53-pod-info\") pod \"rabbitmq-server-0\" (UID: \"42eb5011-6651-43cb-a562-c89cc6df2c53\") " pod="openstack/rabbitmq-server-0" Dec 13 07:16:26 crc kubenswrapper[4971]: I1213 07:16:26.725628 4971 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-2s5b6\" (UniqueName: \"kubernetes.io/projected/42eb5011-6651-43cb-a562-c89cc6df2c53-kube-api-access-2s5b6\") pod \"rabbitmq-server-0\" (UID: \"42eb5011-6651-43cb-a562-c89cc6df2c53\") " pod="openstack/rabbitmq-server-0" Dec 13 07:16:26 crc kubenswrapper[4971]: I1213 07:16:26.742882 4971 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"local-storage06-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage06-crc\") pod \"rabbitmq-server-0\" (UID: \"42eb5011-6651-43cb-a562-c89cc6df2c53\") " pod="openstack/rabbitmq-server-0" Dec 13 07:16:26 crc kubenswrapper[4971]: I1213 07:16:26.743928 4971 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/rabbitmq-cell1-server-0" podUID="5eed809d-c915-433c-9915-27e0b01e1ffe" containerName="rabbitmq" probeResult="failure" output="dial tcp 10.217.0.99:5671: connect: connection refused" Dec 13 07:16:26 crc kubenswrapper[4971]: I1213 07:16:26.860062 4971 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/rabbitmq-server-0" Dec 13 07:16:26 crc kubenswrapper[4971]: I1213 07:16:26.971953 4971 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/rabbitmq-cell1-server-0" Dec 13 07:16:27 crc kubenswrapper[4971]: I1213 07:16:27.011139 4971 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"persistence\" (UniqueName: \"kubernetes.io/local-volume/local-storage10-crc\") pod \"5eed809d-c915-433c-9915-27e0b01e1ffe\" (UID: \"5eed809d-c915-433c-9915-27e0b01e1ffe\") " Dec 13 07:16:27 crc kubenswrapper[4971]: I1213 07:16:27.011206 4971 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pod-info\" (UniqueName: \"kubernetes.io/downward-api/5eed809d-c915-433c-9915-27e0b01e1ffe-pod-info\") pod \"5eed809d-c915-433c-9915-27e0b01e1ffe\" (UID: \"5eed809d-c915-433c-9915-27e0b01e1ffe\") " Dec 13 07:16:27 crc kubenswrapper[4971]: I1213 07:16:27.011239 4971 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"erlang-cookie-secret\" (UniqueName: \"kubernetes.io/secret/5eed809d-c915-433c-9915-27e0b01e1ffe-erlang-cookie-secret\") pod \"5eed809d-c915-433c-9915-27e0b01e1ffe\" (UID: \"5eed809d-c915-433c-9915-27e0b01e1ffe\") " Dec 13 07:16:27 crc kubenswrapper[4971]: I1213 07:16:27.011274 4971 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"rabbitmq-tls\" (UniqueName: \"kubernetes.io/projected/5eed809d-c915-433c-9915-27e0b01e1ffe-rabbitmq-tls\") pod \"5eed809d-c915-433c-9915-27e0b01e1ffe\" (UID: \"5eed809d-c915-433c-9915-27e0b01e1ffe\") " Dec 13 07:16:27 crc kubenswrapper[4971]: I1213 07:16:27.011368 4971 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"server-conf\" (UniqueName: \"kubernetes.io/configmap/5eed809d-c915-433c-9915-27e0b01e1ffe-server-conf\") pod \"5eed809d-c915-433c-9915-27e0b01e1ffe\" (UID: \"5eed809d-c915-433c-9915-27e0b01e1ffe\") " Dec 13 07:16:27 crc kubenswrapper[4971]: I1213 07:16:27.011385 4971 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"plugins-conf\" (UniqueName: \"kubernetes.io/configmap/5eed809d-c915-433c-9915-27e0b01e1ffe-plugins-conf\") pod \"5eed809d-c915-433c-9915-27e0b01e1ffe\" (UID: \"5eed809d-c915-433c-9915-27e0b01e1ffe\") " Dec 13 07:16:27 crc kubenswrapper[4971]: I1213 07:16:27.011435 4971 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"rabbitmq-plugins\" (UniqueName: \"kubernetes.io/empty-dir/5eed809d-c915-433c-9915-27e0b01e1ffe-rabbitmq-plugins\") pod \"5eed809d-c915-433c-9915-27e0b01e1ffe\" (UID: \"5eed809d-c915-433c-9915-27e0b01e1ffe\") " Dec 13 07:16:27 crc kubenswrapper[4971]: I1213 07:16:27.011462 4971 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/5eed809d-c915-433c-9915-27e0b01e1ffe-config-data\") pod \"5eed809d-c915-433c-9915-27e0b01e1ffe\" (UID: \"5eed809d-c915-433c-9915-27e0b01e1ffe\") " Dec 13 07:16:27 crc kubenswrapper[4971]: I1213 07:16:27.011596 4971 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"rabbitmq-confd\" (UniqueName: \"kubernetes.io/projected/5eed809d-c915-433c-9915-27e0b01e1ffe-rabbitmq-confd\") pod \"5eed809d-c915-433c-9915-27e0b01e1ffe\" (UID: \"5eed809d-c915-433c-9915-27e0b01e1ffe\") " Dec 13 07:16:27 crc kubenswrapper[4971]: I1213 07:16:27.011647 4971 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-jpcrh\" (UniqueName: \"kubernetes.io/projected/5eed809d-c915-433c-9915-27e0b01e1ffe-kube-api-access-jpcrh\") pod \"5eed809d-c915-433c-9915-27e0b01e1ffe\" (UID: \"5eed809d-c915-433c-9915-27e0b01e1ffe\") " Dec 13 07:16:27 crc kubenswrapper[4971]: I1213 07:16:27.011761 4971 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"rabbitmq-erlang-cookie\" (UniqueName: \"kubernetes.io/empty-dir/5eed809d-c915-433c-9915-27e0b01e1ffe-rabbitmq-erlang-cookie\") pod \"5eed809d-c915-433c-9915-27e0b01e1ffe\" (UID: \"5eed809d-c915-433c-9915-27e0b01e1ffe\") " Dec 13 07:16:27 crc kubenswrapper[4971]: I1213 07:16:27.012676 4971 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/5eed809d-c915-433c-9915-27e0b01e1ffe-rabbitmq-erlang-cookie" (OuterVolumeSpecName: "rabbitmq-erlang-cookie") pod "5eed809d-c915-433c-9915-27e0b01e1ffe" (UID: "5eed809d-c915-433c-9915-27e0b01e1ffe"). InnerVolumeSpecName "rabbitmq-erlang-cookie". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 13 07:16:27 crc kubenswrapper[4971]: I1213 07:16:27.015040 4971 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/5eed809d-c915-433c-9915-27e0b01e1ffe-plugins-conf" (OuterVolumeSpecName: "plugins-conf") pod "5eed809d-c915-433c-9915-27e0b01e1ffe" (UID: "5eed809d-c915-433c-9915-27e0b01e1ffe"). InnerVolumeSpecName "plugins-conf". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 13 07:16:27 crc kubenswrapper[4971]: I1213 07:16:27.016653 4971 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/local-volume/local-storage10-crc" (OuterVolumeSpecName: "persistence") pod "5eed809d-c915-433c-9915-27e0b01e1ffe" (UID: "5eed809d-c915-433c-9915-27e0b01e1ffe"). InnerVolumeSpecName "local-storage10-crc". PluginName "kubernetes.io/local-volume", VolumeGidValue "" Dec 13 07:16:27 crc kubenswrapper[4971]: I1213 07:16:27.017070 4971 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/5eed809d-c915-433c-9915-27e0b01e1ffe-erlang-cookie-secret" (OuterVolumeSpecName: "erlang-cookie-secret") pod "5eed809d-c915-433c-9915-27e0b01e1ffe" (UID: "5eed809d-c915-433c-9915-27e0b01e1ffe"). InnerVolumeSpecName "erlang-cookie-secret". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 13 07:16:27 crc kubenswrapper[4971]: I1213 07:16:27.018180 4971 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/5eed809d-c915-433c-9915-27e0b01e1ffe-rabbitmq-plugins" (OuterVolumeSpecName: "rabbitmq-plugins") pod "5eed809d-c915-433c-9915-27e0b01e1ffe" (UID: "5eed809d-c915-433c-9915-27e0b01e1ffe"). InnerVolumeSpecName "rabbitmq-plugins". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 13 07:16:27 crc kubenswrapper[4971]: I1213 07:16:27.019646 4971 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/5eed809d-c915-433c-9915-27e0b01e1ffe-kube-api-access-jpcrh" (OuterVolumeSpecName: "kube-api-access-jpcrh") pod "5eed809d-c915-433c-9915-27e0b01e1ffe" (UID: "5eed809d-c915-433c-9915-27e0b01e1ffe"). InnerVolumeSpecName "kube-api-access-jpcrh". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 13 07:16:27 crc kubenswrapper[4971]: I1213 07:16:27.021307 4971 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/downward-api/5eed809d-c915-433c-9915-27e0b01e1ffe-pod-info" (OuterVolumeSpecName: "pod-info") pod "5eed809d-c915-433c-9915-27e0b01e1ffe" (UID: "5eed809d-c915-433c-9915-27e0b01e1ffe"). InnerVolumeSpecName "pod-info". PluginName "kubernetes.io/downward-api", VolumeGidValue "" Dec 13 07:16:27 crc kubenswrapper[4971]: I1213 07:16:27.021830 4971 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/5eed809d-c915-433c-9915-27e0b01e1ffe-rabbitmq-tls" (OuterVolumeSpecName: "rabbitmq-tls") pod "5eed809d-c915-433c-9915-27e0b01e1ffe" (UID: "5eed809d-c915-433c-9915-27e0b01e1ffe"). InnerVolumeSpecName "rabbitmq-tls". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 13 07:16:27 crc kubenswrapper[4971]: I1213 07:16:27.039856 4971 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/5eed809d-c915-433c-9915-27e0b01e1ffe-config-data" (OuterVolumeSpecName: "config-data") pod "5eed809d-c915-433c-9915-27e0b01e1ffe" (UID: "5eed809d-c915-433c-9915-27e0b01e1ffe"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 13 07:16:27 crc kubenswrapper[4971]: I1213 07:16:27.091052 4971 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/5eed809d-c915-433c-9915-27e0b01e1ffe-server-conf" (OuterVolumeSpecName: "server-conf") pod "5eed809d-c915-433c-9915-27e0b01e1ffe" (UID: "5eed809d-c915-433c-9915-27e0b01e1ffe"). InnerVolumeSpecName "server-conf". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 13 07:16:27 crc kubenswrapper[4971]: I1213 07:16:27.112924 4971 reconciler_common.go:293] "Volume detached for volume \"rabbitmq-erlang-cookie\" (UniqueName: \"kubernetes.io/empty-dir/5eed809d-c915-433c-9915-27e0b01e1ffe-rabbitmq-erlang-cookie\") on node \"crc\" DevicePath \"\"" Dec 13 07:16:27 crc kubenswrapper[4971]: I1213 07:16:27.112990 4971 reconciler_common.go:286] "operationExecutor.UnmountDevice started for volume \"local-storage10-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage10-crc\") on node \"crc\" " Dec 13 07:16:27 crc kubenswrapper[4971]: I1213 07:16:27.113004 4971 reconciler_common.go:293] "Volume detached for volume \"pod-info\" (UniqueName: \"kubernetes.io/downward-api/5eed809d-c915-433c-9915-27e0b01e1ffe-pod-info\") on node \"crc\" DevicePath \"\"" Dec 13 07:16:27 crc kubenswrapper[4971]: I1213 07:16:27.113018 4971 reconciler_common.go:293] "Volume detached for volume \"erlang-cookie-secret\" (UniqueName: \"kubernetes.io/secret/5eed809d-c915-433c-9915-27e0b01e1ffe-erlang-cookie-secret\") on node \"crc\" DevicePath \"\"" Dec 13 07:16:27 crc kubenswrapper[4971]: I1213 07:16:27.113029 4971 reconciler_common.go:293] "Volume detached for volume \"rabbitmq-tls\" (UniqueName: \"kubernetes.io/projected/5eed809d-c915-433c-9915-27e0b01e1ffe-rabbitmq-tls\") on node \"crc\" DevicePath \"\"" Dec 13 07:16:27 crc kubenswrapper[4971]: I1213 07:16:27.113037 4971 reconciler_common.go:293] "Volume detached for volume \"server-conf\" (UniqueName: \"kubernetes.io/configmap/5eed809d-c915-433c-9915-27e0b01e1ffe-server-conf\") on node \"crc\" DevicePath \"\"" Dec 13 07:16:27 crc kubenswrapper[4971]: I1213 07:16:27.113046 4971 reconciler_common.go:293] "Volume detached for volume \"plugins-conf\" (UniqueName: \"kubernetes.io/configmap/5eed809d-c915-433c-9915-27e0b01e1ffe-plugins-conf\") on node \"crc\" DevicePath \"\"" Dec 13 07:16:27 crc kubenswrapper[4971]: I1213 07:16:27.113054 4971 reconciler_common.go:293] "Volume detached for volume \"rabbitmq-plugins\" (UniqueName: \"kubernetes.io/empty-dir/5eed809d-c915-433c-9915-27e0b01e1ffe-rabbitmq-plugins\") on node \"crc\" DevicePath \"\"" Dec 13 07:16:27 crc kubenswrapper[4971]: I1213 07:16:27.113061 4971 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/5eed809d-c915-433c-9915-27e0b01e1ffe-config-data\") on node \"crc\" DevicePath \"\"" Dec 13 07:16:27 crc kubenswrapper[4971]: I1213 07:16:27.113071 4971 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-jpcrh\" (UniqueName: \"kubernetes.io/projected/5eed809d-c915-433c-9915-27e0b01e1ffe-kube-api-access-jpcrh\") on node \"crc\" DevicePath \"\"" Dec 13 07:16:27 crc kubenswrapper[4971]: I1213 07:16:27.130539 4971 operation_generator.go:917] UnmountDevice succeeded for volume "local-storage10-crc" (UniqueName: "kubernetes.io/local-volume/local-storage10-crc") on node "crc" Dec 13 07:16:27 crc kubenswrapper[4971]: I1213 07:16:27.136212 4971 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/5eed809d-c915-433c-9915-27e0b01e1ffe-rabbitmq-confd" (OuterVolumeSpecName: "rabbitmq-confd") pod "5eed809d-c915-433c-9915-27e0b01e1ffe" (UID: "5eed809d-c915-433c-9915-27e0b01e1ffe"). InnerVolumeSpecName "rabbitmq-confd". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 13 07:16:27 crc kubenswrapper[4971]: I1213 07:16:27.214457 4971 reconciler_common.go:293] "Volume detached for volume \"local-storage10-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage10-crc\") on node \"crc\" DevicePath \"\"" Dec 13 07:16:27 crc kubenswrapper[4971]: I1213 07:16:27.214494 4971 reconciler_common.go:293] "Volume detached for volume \"rabbitmq-confd\" (UniqueName: \"kubernetes.io/projected/5eed809d-c915-433c-9915-27e0b01e1ffe-rabbitmq-confd\") on node \"crc\" DevicePath \"\"" Dec 13 07:16:27 crc kubenswrapper[4971]: I1213 07:16:27.338172 4971 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/rabbitmq-server-0"] Dec 13 07:16:27 crc kubenswrapper[4971]: W1213 07:16:27.347554 4971 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod42eb5011_6651_43cb_a562_c89cc6df2c53.slice/crio-6e0ec1c065ce32830aa4507b7e7f050efb4a47322319ba777b53fef8da9ead75 WatchSource:0}: Error finding container 6e0ec1c065ce32830aa4507b7e7f050efb4a47322319ba777b53fef8da9ead75: Status 404 returned error can't find the container with id 6e0ec1c065ce32830aa4507b7e7f050efb4a47322319ba777b53fef8da9ead75 Dec 13 07:16:27 crc kubenswrapper[4971]: I1213 07:16:27.454974 4971 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-server-0" event={"ID":"42eb5011-6651-43cb-a562-c89cc6df2c53","Type":"ContainerStarted","Data":"6e0ec1c065ce32830aa4507b7e7f050efb4a47322319ba777b53fef8da9ead75"} Dec 13 07:16:27 crc kubenswrapper[4971]: I1213 07:16:27.462215 4971 generic.go:334] "Generic (PLEG): container finished" podID="5eed809d-c915-433c-9915-27e0b01e1ffe" containerID="efed32724d55162da7a67edf770832ac164e124b6b49db25e2f7b8b7ec033c44" exitCode=0 Dec 13 07:16:27 crc kubenswrapper[4971]: I1213 07:16:27.462269 4971 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-cell1-server-0" event={"ID":"5eed809d-c915-433c-9915-27e0b01e1ffe","Type":"ContainerDied","Data":"efed32724d55162da7a67edf770832ac164e124b6b49db25e2f7b8b7ec033c44"} Dec 13 07:16:27 crc kubenswrapper[4971]: I1213 07:16:27.462312 4971 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-cell1-server-0" event={"ID":"5eed809d-c915-433c-9915-27e0b01e1ffe","Type":"ContainerDied","Data":"532feca2c32e4b9e2c2fe06c7c12a79ff8d63c0dbe5f42443e2e3c61337e69b7"} Dec 13 07:16:27 crc kubenswrapper[4971]: I1213 07:16:27.462335 4971 scope.go:117] "RemoveContainer" containerID="efed32724d55162da7a67edf770832ac164e124b6b49db25e2f7b8b7ec033c44" Dec 13 07:16:27 crc kubenswrapper[4971]: I1213 07:16:27.462680 4971 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/rabbitmq-cell1-server-0" Dec 13 07:16:27 crc kubenswrapper[4971]: I1213 07:16:27.484535 4971 scope.go:117] "RemoveContainer" containerID="0ecc7342a1a65b67ea97dd2575e3d43416e7633d3c5e40a8ec3ff6b6e6709a86" Dec 13 07:16:27 crc kubenswrapper[4971]: I1213 07:16:27.502279 4971 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/rabbitmq-cell1-server-0"] Dec 13 07:16:27 crc kubenswrapper[4971]: I1213 07:16:27.516198 4971 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/rabbitmq-cell1-server-0"] Dec 13 07:16:27 crc kubenswrapper[4971]: I1213 07:16:27.542607 4971 scope.go:117] "RemoveContainer" containerID="efed32724d55162da7a67edf770832ac164e124b6b49db25e2f7b8b7ec033c44" Dec 13 07:16:27 crc kubenswrapper[4971]: E1213 07:16:27.543708 4971 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"efed32724d55162da7a67edf770832ac164e124b6b49db25e2f7b8b7ec033c44\": container with ID starting with efed32724d55162da7a67edf770832ac164e124b6b49db25e2f7b8b7ec033c44 not found: ID does not exist" containerID="efed32724d55162da7a67edf770832ac164e124b6b49db25e2f7b8b7ec033c44" Dec 13 07:16:27 crc kubenswrapper[4971]: I1213 07:16:27.543754 4971 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"efed32724d55162da7a67edf770832ac164e124b6b49db25e2f7b8b7ec033c44"} err="failed to get container status \"efed32724d55162da7a67edf770832ac164e124b6b49db25e2f7b8b7ec033c44\": rpc error: code = NotFound desc = could not find container \"efed32724d55162da7a67edf770832ac164e124b6b49db25e2f7b8b7ec033c44\": container with ID starting with efed32724d55162da7a67edf770832ac164e124b6b49db25e2f7b8b7ec033c44 not found: ID does not exist" Dec 13 07:16:27 crc kubenswrapper[4971]: I1213 07:16:27.543794 4971 scope.go:117] "RemoveContainer" containerID="0ecc7342a1a65b67ea97dd2575e3d43416e7633d3c5e40a8ec3ff6b6e6709a86" Dec 13 07:16:27 crc kubenswrapper[4971]: E1213 07:16:27.544079 4971 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"0ecc7342a1a65b67ea97dd2575e3d43416e7633d3c5e40a8ec3ff6b6e6709a86\": container with ID starting with 0ecc7342a1a65b67ea97dd2575e3d43416e7633d3c5e40a8ec3ff6b6e6709a86 not found: ID does not exist" containerID="0ecc7342a1a65b67ea97dd2575e3d43416e7633d3c5e40a8ec3ff6b6e6709a86" Dec 13 07:16:27 crc kubenswrapper[4971]: I1213 07:16:27.544109 4971 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"0ecc7342a1a65b67ea97dd2575e3d43416e7633d3c5e40a8ec3ff6b6e6709a86"} err="failed to get container status \"0ecc7342a1a65b67ea97dd2575e3d43416e7633d3c5e40a8ec3ff6b6e6709a86\": rpc error: code = NotFound desc = could not find container \"0ecc7342a1a65b67ea97dd2575e3d43416e7633d3c5e40a8ec3ff6b6e6709a86\": container with ID starting with 0ecc7342a1a65b67ea97dd2575e3d43416e7633d3c5e40a8ec3ff6b6e6709a86 not found: ID does not exist" Dec 13 07:16:27 crc kubenswrapper[4971]: I1213 07:16:27.551643 4971 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/rabbitmq-cell1-server-0"] Dec 13 07:16:27 crc kubenswrapper[4971]: E1213 07:16:27.552016 4971 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="5eed809d-c915-433c-9915-27e0b01e1ffe" containerName="setup-container" Dec 13 07:16:27 crc kubenswrapper[4971]: I1213 07:16:27.552030 4971 state_mem.go:107] "Deleted CPUSet assignment" podUID="5eed809d-c915-433c-9915-27e0b01e1ffe" containerName="setup-container" Dec 13 07:16:27 crc kubenswrapper[4971]: E1213 07:16:27.552051 4971 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="5eed809d-c915-433c-9915-27e0b01e1ffe" containerName="rabbitmq" Dec 13 07:16:27 crc kubenswrapper[4971]: I1213 07:16:27.552057 4971 state_mem.go:107] "Deleted CPUSet assignment" podUID="5eed809d-c915-433c-9915-27e0b01e1ffe" containerName="rabbitmq" Dec 13 07:16:27 crc kubenswrapper[4971]: I1213 07:16:27.552224 4971 memory_manager.go:354] "RemoveStaleState removing state" podUID="5eed809d-c915-433c-9915-27e0b01e1ffe" containerName="rabbitmq" Dec 13 07:16:27 crc kubenswrapper[4971]: I1213 07:16:27.553120 4971 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/rabbitmq-cell1-server-0" Dec 13 07:16:27 crc kubenswrapper[4971]: I1213 07:16:27.556758 4971 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"rabbitmq-cell1-server-conf" Dec 13 07:16:27 crc kubenswrapper[4971]: I1213 07:16:27.556820 4971 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"rabbitmq-cell1-erlang-cookie" Dec 13 07:16:27 crc kubenswrapper[4971]: I1213 07:16:27.557069 4971 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"rabbitmq-cell1-server-dockercfg-5725c" Dec 13 07:16:27 crc kubenswrapper[4971]: I1213 07:16:27.557174 4971 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"rabbitmq-cell1-config-data" Dec 13 07:16:27 crc kubenswrapper[4971]: I1213 07:16:27.557282 4971 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"rabbitmq-cell1-plugins-conf" Dec 13 07:16:27 crc kubenswrapper[4971]: I1213 07:16:27.557409 4971 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-rabbitmq-cell1-svc" Dec 13 07:16:27 crc kubenswrapper[4971]: I1213 07:16:27.557533 4971 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"rabbitmq-cell1-default-user" Dec 13 07:16:27 crc kubenswrapper[4971]: I1213 07:16:27.567719 4971 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/rabbitmq-cell1-server-0"] Dec 13 07:16:27 crc kubenswrapper[4971]: I1213 07:16:27.738788 4971 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/299317ec-a652-40f2-b0c5-e25f099204c4-config-data\") pod \"rabbitmq-cell1-server-0\" (UID: \"299317ec-a652-40f2-b0c5-e25f099204c4\") " pod="openstack/rabbitmq-cell1-server-0" Dec 13 07:16:27 crc kubenswrapper[4971]: I1213 07:16:27.739302 4971 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"pod-info\" (UniqueName: \"kubernetes.io/downward-api/299317ec-a652-40f2-b0c5-e25f099204c4-pod-info\") pod \"rabbitmq-cell1-server-0\" (UID: \"299317ec-a652-40f2-b0c5-e25f099204c4\") " pod="openstack/rabbitmq-cell1-server-0" Dec 13 07:16:27 crc kubenswrapper[4971]: I1213 07:16:27.739427 4971 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rabbitmq-erlang-cookie\" (UniqueName: \"kubernetes.io/empty-dir/299317ec-a652-40f2-b0c5-e25f099204c4-rabbitmq-erlang-cookie\") pod \"rabbitmq-cell1-server-0\" (UID: \"299317ec-a652-40f2-b0c5-e25f099204c4\") " pod="openstack/rabbitmq-cell1-server-0" Dec 13 07:16:27 crc kubenswrapper[4971]: I1213 07:16:27.739573 4971 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rabbitmq-tls\" (UniqueName: \"kubernetes.io/projected/299317ec-a652-40f2-b0c5-e25f099204c4-rabbitmq-tls\") pod \"rabbitmq-cell1-server-0\" (UID: \"299317ec-a652-40f2-b0c5-e25f099204c4\") " pod="openstack/rabbitmq-cell1-server-0" Dec 13 07:16:27 crc kubenswrapper[4971]: I1213 07:16:27.739640 4971 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"erlang-cookie-secret\" (UniqueName: \"kubernetes.io/secret/299317ec-a652-40f2-b0c5-e25f099204c4-erlang-cookie-secret\") pod \"rabbitmq-cell1-server-0\" (UID: \"299317ec-a652-40f2-b0c5-e25f099204c4\") " pod="openstack/rabbitmq-cell1-server-0" Dec 13 07:16:27 crc kubenswrapper[4971]: I1213 07:16:27.739774 4971 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-bjfzv\" (UniqueName: \"kubernetes.io/projected/299317ec-a652-40f2-b0c5-e25f099204c4-kube-api-access-bjfzv\") pod \"rabbitmq-cell1-server-0\" (UID: \"299317ec-a652-40f2-b0c5-e25f099204c4\") " pod="openstack/rabbitmq-cell1-server-0" Dec 13 07:16:27 crc kubenswrapper[4971]: I1213 07:16:27.739847 4971 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"plugins-conf\" (UniqueName: \"kubernetes.io/configmap/299317ec-a652-40f2-b0c5-e25f099204c4-plugins-conf\") pod \"rabbitmq-cell1-server-0\" (UID: \"299317ec-a652-40f2-b0c5-e25f099204c4\") " pod="openstack/rabbitmq-cell1-server-0" Dec 13 07:16:27 crc kubenswrapper[4971]: I1213 07:16:27.739937 4971 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rabbitmq-plugins\" (UniqueName: \"kubernetes.io/empty-dir/299317ec-a652-40f2-b0c5-e25f099204c4-rabbitmq-plugins\") pod \"rabbitmq-cell1-server-0\" (UID: \"299317ec-a652-40f2-b0c5-e25f099204c4\") " pod="openstack/rabbitmq-cell1-server-0" Dec 13 07:16:27 crc kubenswrapper[4971]: I1213 07:16:27.740007 4971 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"local-storage10-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage10-crc\") pod \"rabbitmq-cell1-server-0\" (UID: \"299317ec-a652-40f2-b0c5-e25f099204c4\") " pod="openstack/rabbitmq-cell1-server-0" Dec 13 07:16:27 crc kubenswrapper[4971]: I1213 07:16:27.740055 4971 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"server-conf\" (UniqueName: \"kubernetes.io/configmap/299317ec-a652-40f2-b0c5-e25f099204c4-server-conf\") pod \"rabbitmq-cell1-server-0\" (UID: \"299317ec-a652-40f2-b0c5-e25f099204c4\") " pod="openstack/rabbitmq-cell1-server-0" Dec 13 07:16:27 crc kubenswrapper[4971]: I1213 07:16:27.740094 4971 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rabbitmq-confd\" (UniqueName: \"kubernetes.io/projected/299317ec-a652-40f2-b0c5-e25f099204c4-rabbitmq-confd\") pod \"rabbitmq-cell1-server-0\" (UID: \"299317ec-a652-40f2-b0c5-e25f099204c4\") " pod="openstack/rabbitmq-cell1-server-0" Dec 13 07:16:27 crc kubenswrapper[4971]: I1213 07:16:27.786672 4971 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="5eed809d-c915-433c-9915-27e0b01e1ffe" path="/var/lib/kubelet/pods/5eed809d-c915-433c-9915-27e0b01e1ffe/volumes" Dec 13 07:16:27 crc kubenswrapper[4971]: I1213 07:16:27.788783 4971 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="a590eb84-c8d5-4cbe-9c03-03a6ccb82073" path="/var/lib/kubelet/pods/a590eb84-c8d5-4cbe-9c03-03a6ccb82073/volumes" Dec 13 07:16:27 crc kubenswrapper[4971]: I1213 07:16:27.842049 4971 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-bjfzv\" (UniqueName: \"kubernetes.io/projected/299317ec-a652-40f2-b0c5-e25f099204c4-kube-api-access-bjfzv\") pod \"rabbitmq-cell1-server-0\" (UID: \"299317ec-a652-40f2-b0c5-e25f099204c4\") " pod="openstack/rabbitmq-cell1-server-0" Dec 13 07:16:27 crc kubenswrapper[4971]: I1213 07:16:27.842102 4971 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"plugins-conf\" (UniqueName: \"kubernetes.io/configmap/299317ec-a652-40f2-b0c5-e25f099204c4-plugins-conf\") pod \"rabbitmq-cell1-server-0\" (UID: \"299317ec-a652-40f2-b0c5-e25f099204c4\") " pod="openstack/rabbitmq-cell1-server-0" Dec 13 07:16:27 crc kubenswrapper[4971]: I1213 07:16:27.842142 4971 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rabbitmq-plugins\" (UniqueName: \"kubernetes.io/empty-dir/299317ec-a652-40f2-b0c5-e25f099204c4-rabbitmq-plugins\") pod \"rabbitmq-cell1-server-0\" (UID: \"299317ec-a652-40f2-b0c5-e25f099204c4\") " pod="openstack/rabbitmq-cell1-server-0" Dec 13 07:16:27 crc kubenswrapper[4971]: I1213 07:16:27.842178 4971 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"local-storage10-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage10-crc\") pod \"rabbitmq-cell1-server-0\" (UID: \"299317ec-a652-40f2-b0c5-e25f099204c4\") " pod="openstack/rabbitmq-cell1-server-0" Dec 13 07:16:27 crc kubenswrapper[4971]: I1213 07:16:27.842203 4971 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"server-conf\" (UniqueName: \"kubernetes.io/configmap/299317ec-a652-40f2-b0c5-e25f099204c4-server-conf\") pod \"rabbitmq-cell1-server-0\" (UID: \"299317ec-a652-40f2-b0c5-e25f099204c4\") " pod="openstack/rabbitmq-cell1-server-0" Dec 13 07:16:27 crc kubenswrapper[4971]: I1213 07:16:27.842230 4971 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rabbitmq-confd\" (UniqueName: \"kubernetes.io/projected/299317ec-a652-40f2-b0c5-e25f099204c4-rabbitmq-confd\") pod \"rabbitmq-cell1-server-0\" (UID: \"299317ec-a652-40f2-b0c5-e25f099204c4\") " pod="openstack/rabbitmq-cell1-server-0" Dec 13 07:16:27 crc kubenswrapper[4971]: I1213 07:16:27.842277 4971 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/299317ec-a652-40f2-b0c5-e25f099204c4-config-data\") pod \"rabbitmq-cell1-server-0\" (UID: \"299317ec-a652-40f2-b0c5-e25f099204c4\") " pod="openstack/rabbitmq-cell1-server-0" Dec 13 07:16:27 crc kubenswrapper[4971]: I1213 07:16:27.842296 4971 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pod-info\" (UniqueName: \"kubernetes.io/downward-api/299317ec-a652-40f2-b0c5-e25f099204c4-pod-info\") pod \"rabbitmq-cell1-server-0\" (UID: \"299317ec-a652-40f2-b0c5-e25f099204c4\") " pod="openstack/rabbitmq-cell1-server-0" Dec 13 07:16:27 crc kubenswrapper[4971]: I1213 07:16:27.842340 4971 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rabbitmq-erlang-cookie\" (UniqueName: \"kubernetes.io/empty-dir/299317ec-a652-40f2-b0c5-e25f099204c4-rabbitmq-erlang-cookie\") pod \"rabbitmq-cell1-server-0\" (UID: \"299317ec-a652-40f2-b0c5-e25f099204c4\") " pod="openstack/rabbitmq-cell1-server-0" Dec 13 07:16:27 crc kubenswrapper[4971]: I1213 07:16:27.842373 4971 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rabbitmq-tls\" (UniqueName: \"kubernetes.io/projected/299317ec-a652-40f2-b0c5-e25f099204c4-rabbitmq-tls\") pod \"rabbitmq-cell1-server-0\" (UID: \"299317ec-a652-40f2-b0c5-e25f099204c4\") " pod="openstack/rabbitmq-cell1-server-0" Dec 13 07:16:27 crc kubenswrapper[4971]: I1213 07:16:27.842397 4971 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"erlang-cookie-secret\" (UniqueName: \"kubernetes.io/secret/299317ec-a652-40f2-b0c5-e25f099204c4-erlang-cookie-secret\") pod \"rabbitmq-cell1-server-0\" (UID: \"299317ec-a652-40f2-b0c5-e25f099204c4\") " pod="openstack/rabbitmq-cell1-server-0" Dec 13 07:16:27 crc kubenswrapper[4971]: I1213 07:16:27.842730 4971 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"local-storage10-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage10-crc\") pod \"rabbitmq-cell1-server-0\" (UID: \"299317ec-a652-40f2-b0c5-e25f099204c4\") device mount path \"/mnt/openstack/pv10\"" pod="openstack/rabbitmq-cell1-server-0" Dec 13 07:16:27 crc kubenswrapper[4971]: I1213 07:16:27.843538 4971 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rabbitmq-erlang-cookie\" (UniqueName: \"kubernetes.io/empty-dir/299317ec-a652-40f2-b0c5-e25f099204c4-rabbitmq-erlang-cookie\") pod \"rabbitmq-cell1-server-0\" (UID: \"299317ec-a652-40f2-b0c5-e25f099204c4\") " pod="openstack/rabbitmq-cell1-server-0" Dec 13 07:16:27 crc kubenswrapper[4971]: I1213 07:16:27.844506 4971 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rabbitmq-plugins\" (UniqueName: \"kubernetes.io/empty-dir/299317ec-a652-40f2-b0c5-e25f099204c4-rabbitmq-plugins\") pod \"rabbitmq-cell1-server-0\" (UID: \"299317ec-a652-40f2-b0c5-e25f099204c4\") " pod="openstack/rabbitmq-cell1-server-0" Dec 13 07:16:27 crc kubenswrapper[4971]: I1213 07:16:27.844704 4971 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/299317ec-a652-40f2-b0c5-e25f099204c4-config-data\") pod \"rabbitmq-cell1-server-0\" (UID: \"299317ec-a652-40f2-b0c5-e25f099204c4\") " pod="openstack/rabbitmq-cell1-server-0" Dec 13 07:16:27 crc kubenswrapper[4971]: I1213 07:16:27.846229 4971 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"server-conf\" (UniqueName: \"kubernetes.io/configmap/299317ec-a652-40f2-b0c5-e25f099204c4-server-conf\") pod \"rabbitmq-cell1-server-0\" (UID: \"299317ec-a652-40f2-b0c5-e25f099204c4\") " pod="openstack/rabbitmq-cell1-server-0" Dec 13 07:16:27 crc kubenswrapper[4971]: I1213 07:16:27.846288 4971 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"plugins-conf\" (UniqueName: \"kubernetes.io/configmap/299317ec-a652-40f2-b0c5-e25f099204c4-plugins-conf\") pod \"rabbitmq-cell1-server-0\" (UID: \"299317ec-a652-40f2-b0c5-e25f099204c4\") " pod="openstack/rabbitmq-cell1-server-0" Dec 13 07:16:27 crc kubenswrapper[4971]: I1213 07:16:27.848473 4971 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rabbitmq-tls\" (UniqueName: \"kubernetes.io/projected/299317ec-a652-40f2-b0c5-e25f099204c4-rabbitmq-tls\") pod \"rabbitmq-cell1-server-0\" (UID: \"299317ec-a652-40f2-b0c5-e25f099204c4\") " pod="openstack/rabbitmq-cell1-server-0" Dec 13 07:16:27 crc kubenswrapper[4971]: I1213 07:16:27.850990 4971 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"erlang-cookie-secret\" (UniqueName: \"kubernetes.io/secret/299317ec-a652-40f2-b0c5-e25f099204c4-erlang-cookie-secret\") pod \"rabbitmq-cell1-server-0\" (UID: \"299317ec-a652-40f2-b0c5-e25f099204c4\") " pod="openstack/rabbitmq-cell1-server-0" Dec 13 07:16:27 crc kubenswrapper[4971]: I1213 07:16:27.861358 4971 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"pod-info\" (UniqueName: \"kubernetes.io/downward-api/299317ec-a652-40f2-b0c5-e25f099204c4-pod-info\") pod \"rabbitmq-cell1-server-0\" (UID: \"299317ec-a652-40f2-b0c5-e25f099204c4\") " pod="openstack/rabbitmq-cell1-server-0" Dec 13 07:16:27 crc kubenswrapper[4971]: I1213 07:16:27.862911 4971 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-bjfzv\" (UniqueName: \"kubernetes.io/projected/299317ec-a652-40f2-b0c5-e25f099204c4-kube-api-access-bjfzv\") pod \"rabbitmq-cell1-server-0\" (UID: \"299317ec-a652-40f2-b0c5-e25f099204c4\") " pod="openstack/rabbitmq-cell1-server-0" Dec 13 07:16:27 crc kubenswrapper[4971]: I1213 07:16:27.889364 4971 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rabbitmq-confd\" (UniqueName: \"kubernetes.io/projected/299317ec-a652-40f2-b0c5-e25f099204c4-rabbitmq-confd\") pod \"rabbitmq-cell1-server-0\" (UID: \"299317ec-a652-40f2-b0c5-e25f099204c4\") " pod="openstack/rabbitmq-cell1-server-0" Dec 13 07:16:27 crc kubenswrapper[4971]: I1213 07:16:27.896860 4971 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"local-storage10-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage10-crc\") pod \"rabbitmq-cell1-server-0\" (UID: \"299317ec-a652-40f2-b0c5-e25f099204c4\") " pod="openstack/rabbitmq-cell1-server-0" Dec 13 07:16:28 crc kubenswrapper[4971]: I1213 07:16:28.176857 4971 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/rabbitmq-cell1-server-0" Dec 13 07:16:28 crc kubenswrapper[4971]: I1213 07:16:28.663885 4971 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/rabbitmq-cell1-server-0"] Dec 13 07:16:28 crc kubenswrapper[4971]: W1213 07:16:28.669277 4971 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod299317ec_a652_40f2_b0c5_e25f099204c4.slice/crio-0e010208e66976761e32c751a7b5589f524cea9437eaaf20b4b2ce748c6547a9 WatchSource:0}: Error finding container 0e010208e66976761e32c751a7b5589f524cea9437eaaf20b4b2ce748c6547a9: Status 404 returned error can't find the container with id 0e010208e66976761e32c751a7b5589f524cea9437eaaf20b4b2ce748c6547a9 Dec 13 07:16:29 crc kubenswrapper[4971]: I1213 07:16:29.481157 4971 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-cell1-server-0" event={"ID":"299317ec-a652-40f2-b0c5-e25f099204c4","Type":"ContainerStarted","Data":"0e010208e66976761e32c751a7b5589f524cea9437eaaf20b4b2ce748c6547a9"} Dec 13 07:16:29 crc kubenswrapper[4971]: I1213 07:16:29.483453 4971 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-server-0" event={"ID":"42eb5011-6651-43cb-a562-c89cc6df2c53","Type":"ContainerStarted","Data":"61c8ef8fd4dd958d475d5f013ee5bfe47c0dda61b86af7a1955684dfb85be60b"} Dec 13 07:16:30 crc kubenswrapper[4971]: I1213 07:16:30.494979 4971 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-cell1-server-0" event={"ID":"299317ec-a652-40f2-b0c5-e25f099204c4","Type":"ContainerStarted","Data":"124c26f1f489505ddde7c5fedd4dc0853250ddf7ad07873202149ec0a78af095"} Dec 13 07:16:31 crc kubenswrapper[4971]: I1213 07:16:31.769288 4971 scope.go:117] "RemoveContainer" containerID="e8beee43d15719dccb6892d6f5711674492425300efa098be89033bc65977d14" Dec 13 07:16:31 crc kubenswrapper[4971]: E1213 07:16:31.770184 4971 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-82xjz_openshift-machine-config-operator(e0c4af12-edda-41d7-9b44-f87396174ef2)\"" pod="openshift-machine-config-operator/machine-config-daemon-82xjz" podUID="e0c4af12-edda-41d7-9b44-f87396174ef2" Dec 13 07:16:32 crc kubenswrapper[4971]: I1213 07:16:32.676811 4971 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-74d67684c-cb4m9"] Dec 13 07:16:32 crc kubenswrapper[4971]: I1213 07:16:32.678659 4971 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-74d67684c-cb4m9" Dec 13 07:16:32 crc kubenswrapper[4971]: I1213 07:16:32.681141 4971 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-edpm-ipam" Dec 13 07:16:32 crc kubenswrapper[4971]: I1213 07:16:32.692035 4971 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-74d67684c-cb4m9"] Dec 13 07:16:32 crc kubenswrapper[4971]: I1213 07:16:32.846165 4971 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/62db389f-20fa-444c-af08-3941ad42ae32-ovsdbserver-sb\") pod \"dnsmasq-dns-74d67684c-cb4m9\" (UID: \"62db389f-20fa-444c-af08-3941ad42ae32\") " pod="openstack/dnsmasq-dns-74d67684c-cb4m9" Dec 13 07:16:32 crc kubenswrapper[4971]: I1213 07:16:32.846489 4971 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/62db389f-20fa-444c-af08-3941ad42ae32-ovsdbserver-nb\") pod \"dnsmasq-dns-74d67684c-cb4m9\" (UID: \"62db389f-20fa-444c-af08-3941ad42ae32\") " pod="openstack/dnsmasq-dns-74d67684c-cb4m9" Dec 13 07:16:32 crc kubenswrapper[4971]: I1213 07:16:32.846560 4971 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-qmdgk\" (UniqueName: \"kubernetes.io/projected/62db389f-20fa-444c-af08-3941ad42ae32-kube-api-access-qmdgk\") pod \"dnsmasq-dns-74d67684c-cb4m9\" (UID: \"62db389f-20fa-444c-af08-3941ad42ae32\") " pod="openstack/dnsmasq-dns-74d67684c-cb4m9" Dec 13 07:16:32 crc kubenswrapper[4971]: I1213 07:16:32.846822 4971 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/62db389f-20fa-444c-af08-3941ad42ae32-dns-svc\") pod \"dnsmasq-dns-74d67684c-cb4m9\" (UID: \"62db389f-20fa-444c-af08-3941ad42ae32\") " pod="openstack/dnsmasq-dns-74d67684c-cb4m9" Dec 13 07:16:32 crc kubenswrapper[4971]: I1213 07:16:32.846850 4971 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/configmap/62db389f-20fa-444c-af08-3941ad42ae32-openstack-edpm-ipam\") pod \"dnsmasq-dns-74d67684c-cb4m9\" (UID: \"62db389f-20fa-444c-af08-3941ad42ae32\") " pod="openstack/dnsmasq-dns-74d67684c-cb4m9" Dec 13 07:16:32 crc kubenswrapper[4971]: I1213 07:16:32.846925 4971 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/62db389f-20fa-444c-af08-3941ad42ae32-config\") pod \"dnsmasq-dns-74d67684c-cb4m9\" (UID: \"62db389f-20fa-444c-af08-3941ad42ae32\") " pod="openstack/dnsmasq-dns-74d67684c-cb4m9" Dec 13 07:16:32 crc kubenswrapper[4971]: I1213 07:16:32.948881 4971 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/62db389f-20fa-444c-af08-3941ad42ae32-ovsdbserver-nb\") pod \"dnsmasq-dns-74d67684c-cb4m9\" (UID: \"62db389f-20fa-444c-af08-3941ad42ae32\") " pod="openstack/dnsmasq-dns-74d67684c-cb4m9" Dec 13 07:16:32 crc kubenswrapper[4971]: I1213 07:16:32.948952 4971 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-qmdgk\" (UniqueName: \"kubernetes.io/projected/62db389f-20fa-444c-af08-3941ad42ae32-kube-api-access-qmdgk\") pod \"dnsmasq-dns-74d67684c-cb4m9\" (UID: \"62db389f-20fa-444c-af08-3941ad42ae32\") " pod="openstack/dnsmasq-dns-74d67684c-cb4m9" Dec 13 07:16:32 crc kubenswrapper[4971]: I1213 07:16:32.949046 4971 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/62db389f-20fa-444c-af08-3941ad42ae32-dns-svc\") pod \"dnsmasq-dns-74d67684c-cb4m9\" (UID: \"62db389f-20fa-444c-af08-3941ad42ae32\") " pod="openstack/dnsmasq-dns-74d67684c-cb4m9" Dec 13 07:16:32 crc kubenswrapper[4971]: I1213 07:16:32.949066 4971 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/configmap/62db389f-20fa-444c-af08-3941ad42ae32-openstack-edpm-ipam\") pod \"dnsmasq-dns-74d67684c-cb4m9\" (UID: \"62db389f-20fa-444c-af08-3941ad42ae32\") " pod="openstack/dnsmasq-dns-74d67684c-cb4m9" Dec 13 07:16:32 crc kubenswrapper[4971]: I1213 07:16:32.949096 4971 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/62db389f-20fa-444c-af08-3941ad42ae32-config\") pod \"dnsmasq-dns-74d67684c-cb4m9\" (UID: \"62db389f-20fa-444c-af08-3941ad42ae32\") " pod="openstack/dnsmasq-dns-74d67684c-cb4m9" Dec 13 07:16:32 crc kubenswrapper[4971]: I1213 07:16:32.949132 4971 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/62db389f-20fa-444c-af08-3941ad42ae32-ovsdbserver-sb\") pod \"dnsmasq-dns-74d67684c-cb4m9\" (UID: \"62db389f-20fa-444c-af08-3941ad42ae32\") " pod="openstack/dnsmasq-dns-74d67684c-cb4m9" Dec 13 07:16:32 crc kubenswrapper[4971]: I1213 07:16:32.949785 4971 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/62db389f-20fa-444c-af08-3941ad42ae32-ovsdbserver-nb\") pod \"dnsmasq-dns-74d67684c-cb4m9\" (UID: \"62db389f-20fa-444c-af08-3941ad42ae32\") " pod="openstack/dnsmasq-dns-74d67684c-cb4m9" Dec 13 07:16:32 crc kubenswrapper[4971]: I1213 07:16:32.949882 4971 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/configmap/62db389f-20fa-444c-af08-3941ad42ae32-openstack-edpm-ipam\") pod \"dnsmasq-dns-74d67684c-cb4m9\" (UID: \"62db389f-20fa-444c-af08-3941ad42ae32\") " pod="openstack/dnsmasq-dns-74d67684c-cb4m9" Dec 13 07:16:32 crc kubenswrapper[4971]: I1213 07:16:32.950140 4971 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/62db389f-20fa-444c-af08-3941ad42ae32-ovsdbserver-sb\") pod \"dnsmasq-dns-74d67684c-cb4m9\" (UID: \"62db389f-20fa-444c-af08-3941ad42ae32\") " pod="openstack/dnsmasq-dns-74d67684c-cb4m9" Dec 13 07:16:32 crc kubenswrapper[4971]: I1213 07:16:32.950400 4971 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/62db389f-20fa-444c-af08-3941ad42ae32-config\") pod \"dnsmasq-dns-74d67684c-cb4m9\" (UID: \"62db389f-20fa-444c-af08-3941ad42ae32\") " pod="openstack/dnsmasq-dns-74d67684c-cb4m9" Dec 13 07:16:32 crc kubenswrapper[4971]: I1213 07:16:32.950406 4971 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/62db389f-20fa-444c-af08-3941ad42ae32-dns-svc\") pod \"dnsmasq-dns-74d67684c-cb4m9\" (UID: \"62db389f-20fa-444c-af08-3941ad42ae32\") " pod="openstack/dnsmasq-dns-74d67684c-cb4m9" Dec 13 07:16:32 crc kubenswrapper[4971]: I1213 07:16:32.969287 4971 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-qmdgk\" (UniqueName: \"kubernetes.io/projected/62db389f-20fa-444c-af08-3941ad42ae32-kube-api-access-qmdgk\") pod \"dnsmasq-dns-74d67684c-cb4m9\" (UID: \"62db389f-20fa-444c-af08-3941ad42ae32\") " pod="openstack/dnsmasq-dns-74d67684c-cb4m9" Dec 13 07:16:32 crc kubenswrapper[4971]: I1213 07:16:32.999295 4971 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-74d67684c-cb4m9" Dec 13 07:16:33 crc kubenswrapper[4971]: I1213 07:16:33.451188 4971 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-74d67684c-cb4m9"] Dec 13 07:16:33 crc kubenswrapper[4971]: I1213 07:16:33.527616 4971 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-74d67684c-cb4m9" event={"ID":"62db389f-20fa-444c-af08-3941ad42ae32","Type":"ContainerStarted","Data":"9d55ff49d2681ca36b30e18756a3482706624680084d2a7297d1fd419065e7ee"} Dec 13 07:16:34 crc kubenswrapper[4971]: I1213 07:16:34.542133 4971 generic.go:334] "Generic (PLEG): container finished" podID="62db389f-20fa-444c-af08-3941ad42ae32" containerID="0d7551b57ee7908771a5ae1149f919bcf0096d26110da2b6b4e458fb2c2c0848" exitCode=0 Dec 13 07:16:34 crc kubenswrapper[4971]: I1213 07:16:34.542279 4971 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-74d67684c-cb4m9" event={"ID":"62db389f-20fa-444c-af08-3941ad42ae32","Type":"ContainerDied","Data":"0d7551b57ee7908771a5ae1149f919bcf0096d26110da2b6b4e458fb2c2c0848"} Dec 13 07:16:35 crc kubenswrapper[4971]: I1213 07:16:35.556466 4971 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-74d67684c-cb4m9" event={"ID":"62db389f-20fa-444c-af08-3941ad42ae32","Type":"ContainerStarted","Data":"f6ccf72e6b95053161756a47f6fa73f9875b865403fb3f99f7595ac221bfcfcb"} Dec 13 07:16:35 crc kubenswrapper[4971]: I1213 07:16:35.556819 4971 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/dnsmasq-dns-74d67684c-cb4m9" Dec 13 07:16:35 crc kubenswrapper[4971]: I1213 07:16:35.581802 4971 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/dnsmasq-dns-74d67684c-cb4m9" podStartSLOduration=3.581782916 podStartE2EDuration="3.581782916s" podCreationTimestamp="2025-12-13 07:16:32 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-13 07:16:35.576571917 +0000 UTC m=+1652.180981505" watchObservedRunningTime="2025-12-13 07:16:35.581782916 +0000 UTC m=+1652.186192364" Dec 13 07:16:43 crc kubenswrapper[4971]: I1213 07:16:43.001721 4971 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/dnsmasq-dns-74d67684c-cb4m9" Dec 13 07:16:43 crc kubenswrapper[4971]: I1213 07:16:43.060774 4971 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-78c596d7cf-9jdzl"] Dec 13 07:16:43 crc kubenswrapper[4971]: I1213 07:16:43.061017 4971 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/dnsmasq-dns-78c596d7cf-9jdzl" podUID="aadeb784-643a-4e78-909c-4a167bfe11eb" containerName="dnsmasq-dns" containerID="cri-o://5de93625f93b3f1510893ddd803bdb0ed482aa9dedb3ae4499ae9c0394002778" gracePeriod=10 Dec 13 07:16:43 crc kubenswrapper[4971]: I1213 07:16:43.256239 4971 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-55d9fc99f5-565xd"] Dec 13 07:16:43 crc kubenswrapper[4971]: I1213 07:16:43.259249 4971 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-55d9fc99f5-565xd" Dec 13 07:16:43 crc kubenswrapper[4971]: I1213 07:16:43.268836 4971 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-55d9fc99f5-565xd"] Dec 13 07:16:43 crc kubenswrapper[4971]: I1213 07:16:43.387572 4971 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/27e4748a-9fc8-4526-9b48-0d3e30f24268-config\") pod \"dnsmasq-dns-55d9fc99f5-565xd\" (UID: \"27e4748a-9fc8-4526-9b48-0d3e30f24268\") " pod="openstack/dnsmasq-dns-55d9fc99f5-565xd" Dec 13 07:16:43 crc kubenswrapper[4971]: I1213 07:16:43.388229 4971 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/configmap/27e4748a-9fc8-4526-9b48-0d3e30f24268-openstack-edpm-ipam\") pod \"dnsmasq-dns-55d9fc99f5-565xd\" (UID: \"27e4748a-9fc8-4526-9b48-0d3e30f24268\") " pod="openstack/dnsmasq-dns-55d9fc99f5-565xd" Dec 13 07:16:43 crc kubenswrapper[4971]: I1213 07:16:43.388495 4971 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/27e4748a-9fc8-4526-9b48-0d3e30f24268-dns-svc\") pod \"dnsmasq-dns-55d9fc99f5-565xd\" (UID: \"27e4748a-9fc8-4526-9b48-0d3e30f24268\") " pod="openstack/dnsmasq-dns-55d9fc99f5-565xd" Dec 13 07:16:43 crc kubenswrapper[4971]: I1213 07:16:43.388593 4971 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/27e4748a-9fc8-4526-9b48-0d3e30f24268-ovsdbserver-nb\") pod \"dnsmasq-dns-55d9fc99f5-565xd\" (UID: \"27e4748a-9fc8-4526-9b48-0d3e30f24268\") " pod="openstack/dnsmasq-dns-55d9fc99f5-565xd" Dec 13 07:16:43 crc kubenswrapper[4971]: I1213 07:16:43.388734 4971 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/27e4748a-9fc8-4526-9b48-0d3e30f24268-ovsdbserver-sb\") pod \"dnsmasq-dns-55d9fc99f5-565xd\" (UID: \"27e4748a-9fc8-4526-9b48-0d3e30f24268\") " pod="openstack/dnsmasq-dns-55d9fc99f5-565xd" Dec 13 07:16:43 crc kubenswrapper[4971]: I1213 07:16:43.388770 4971 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-9grd2\" (UniqueName: \"kubernetes.io/projected/27e4748a-9fc8-4526-9b48-0d3e30f24268-kube-api-access-9grd2\") pod \"dnsmasq-dns-55d9fc99f5-565xd\" (UID: \"27e4748a-9fc8-4526-9b48-0d3e30f24268\") " pod="openstack/dnsmasq-dns-55d9fc99f5-565xd" Dec 13 07:16:43 crc kubenswrapper[4971]: I1213 07:16:43.490742 4971 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/27e4748a-9fc8-4526-9b48-0d3e30f24268-dns-svc\") pod \"dnsmasq-dns-55d9fc99f5-565xd\" (UID: \"27e4748a-9fc8-4526-9b48-0d3e30f24268\") " pod="openstack/dnsmasq-dns-55d9fc99f5-565xd" Dec 13 07:16:43 crc kubenswrapper[4971]: I1213 07:16:43.490790 4971 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/27e4748a-9fc8-4526-9b48-0d3e30f24268-ovsdbserver-nb\") pod \"dnsmasq-dns-55d9fc99f5-565xd\" (UID: \"27e4748a-9fc8-4526-9b48-0d3e30f24268\") " pod="openstack/dnsmasq-dns-55d9fc99f5-565xd" Dec 13 07:16:43 crc kubenswrapper[4971]: I1213 07:16:43.490831 4971 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/27e4748a-9fc8-4526-9b48-0d3e30f24268-ovsdbserver-sb\") pod \"dnsmasq-dns-55d9fc99f5-565xd\" (UID: \"27e4748a-9fc8-4526-9b48-0d3e30f24268\") " pod="openstack/dnsmasq-dns-55d9fc99f5-565xd" Dec 13 07:16:43 crc kubenswrapper[4971]: I1213 07:16:43.490852 4971 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-9grd2\" (UniqueName: \"kubernetes.io/projected/27e4748a-9fc8-4526-9b48-0d3e30f24268-kube-api-access-9grd2\") pod \"dnsmasq-dns-55d9fc99f5-565xd\" (UID: \"27e4748a-9fc8-4526-9b48-0d3e30f24268\") " pod="openstack/dnsmasq-dns-55d9fc99f5-565xd" Dec 13 07:16:43 crc kubenswrapper[4971]: I1213 07:16:43.490884 4971 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/27e4748a-9fc8-4526-9b48-0d3e30f24268-config\") pod \"dnsmasq-dns-55d9fc99f5-565xd\" (UID: \"27e4748a-9fc8-4526-9b48-0d3e30f24268\") " pod="openstack/dnsmasq-dns-55d9fc99f5-565xd" Dec 13 07:16:43 crc kubenswrapper[4971]: I1213 07:16:43.490946 4971 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/configmap/27e4748a-9fc8-4526-9b48-0d3e30f24268-openstack-edpm-ipam\") pod \"dnsmasq-dns-55d9fc99f5-565xd\" (UID: \"27e4748a-9fc8-4526-9b48-0d3e30f24268\") " pod="openstack/dnsmasq-dns-55d9fc99f5-565xd" Dec 13 07:16:43 crc kubenswrapper[4971]: I1213 07:16:43.491644 4971 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/27e4748a-9fc8-4526-9b48-0d3e30f24268-dns-svc\") pod \"dnsmasq-dns-55d9fc99f5-565xd\" (UID: \"27e4748a-9fc8-4526-9b48-0d3e30f24268\") " pod="openstack/dnsmasq-dns-55d9fc99f5-565xd" Dec 13 07:16:43 crc kubenswrapper[4971]: I1213 07:16:43.491834 4971 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/configmap/27e4748a-9fc8-4526-9b48-0d3e30f24268-openstack-edpm-ipam\") pod \"dnsmasq-dns-55d9fc99f5-565xd\" (UID: \"27e4748a-9fc8-4526-9b48-0d3e30f24268\") " pod="openstack/dnsmasq-dns-55d9fc99f5-565xd" Dec 13 07:16:43 crc kubenswrapper[4971]: I1213 07:16:43.492244 4971 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/27e4748a-9fc8-4526-9b48-0d3e30f24268-ovsdbserver-sb\") pod \"dnsmasq-dns-55d9fc99f5-565xd\" (UID: \"27e4748a-9fc8-4526-9b48-0d3e30f24268\") " pod="openstack/dnsmasq-dns-55d9fc99f5-565xd" Dec 13 07:16:43 crc kubenswrapper[4971]: I1213 07:16:43.492444 4971 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/27e4748a-9fc8-4526-9b48-0d3e30f24268-ovsdbserver-nb\") pod \"dnsmasq-dns-55d9fc99f5-565xd\" (UID: \"27e4748a-9fc8-4526-9b48-0d3e30f24268\") " pod="openstack/dnsmasq-dns-55d9fc99f5-565xd" Dec 13 07:16:43 crc kubenswrapper[4971]: I1213 07:16:43.492812 4971 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/27e4748a-9fc8-4526-9b48-0d3e30f24268-config\") pod \"dnsmasq-dns-55d9fc99f5-565xd\" (UID: \"27e4748a-9fc8-4526-9b48-0d3e30f24268\") " pod="openstack/dnsmasq-dns-55d9fc99f5-565xd" Dec 13 07:16:43 crc kubenswrapper[4971]: I1213 07:16:43.524455 4971 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-9grd2\" (UniqueName: \"kubernetes.io/projected/27e4748a-9fc8-4526-9b48-0d3e30f24268-kube-api-access-9grd2\") pod \"dnsmasq-dns-55d9fc99f5-565xd\" (UID: \"27e4748a-9fc8-4526-9b48-0d3e30f24268\") " pod="openstack/dnsmasq-dns-55d9fc99f5-565xd" Dec 13 07:16:43 crc kubenswrapper[4971]: I1213 07:16:43.577864 4971 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-55d9fc99f5-565xd" Dec 13 07:16:43 crc kubenswrapper[4971]: I1213 07:16:43.617363 4971 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-78c596d7cf-9jdzl" Dec 13 07:16:43 crc kubenswrapper[4971]: I1213 07:16:43.672588 4971 generic.go:334] "Generic (PLEG): container finished" podID="aadeb784-643a-4e78-909c-4a167bfe11eb" containerID="5de93625f93b3f1510893ddd803bdb0ed482aa9dedb3ae4499ae9c0394002778" exitCode=0 Dec 13 07:16:43 crc kubenswrapper[4971]: I1213 07:16:43.672636 4971 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-78c596d7cf-9jdzl" event={"ID":"aadeb784-643a-4e78-909c-4a167bfe11eb","Type":"ContainerDied","Data":"5de93625f93b3f1510893ddd803bdb0ed482aa9dedb3ae4499ae9c0394002778"} Dec 13 07:16:43 crc kubenswrapper[4971]: I1213 07:16:43.672666 4971 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-78c596d7cf-9jdzl" event={"ID":"aadeb784-643a-4e78-909c-4a167bfe11eb","Type":"ContainerDied","Data":"39b7203f122b7ff1e610263e4bb3c7a4f458ce9c2a599f2d6c594aff255c3a77"} Dec 13 07:16:43 crc kubenswrapper[4971]: I1213 07:16:43.672684 4971 scope.go:117] "RemoveContainer" containerID="5de93625f93b3f1510893ddd803bdb0ed482aa9dedb3ae4499ae9c0394002778" Dec 13 07:16:43 crc kubenswrapper[4971]: I1213 07:16:43.672785 4971 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-78c596d7cf-9jdzl" Dec 13 07:16:43 crc kubenswrapper[4971]: I1213 07:16:43.771749 4971 scope.go:117] "RemoveContainer" containerID="7326cf828fc3f8902b2798804319ee20d2885e7a77300034607853426b6231d1" Dec 13 07:16:43 crc kubenswrapper[4971]: I1213 07:16:43.797933 4971 scope.go:117] "RemoveContainer" containerID="5de93625f93b3f1510893ddd803bdb0ed482aa9dedb3ae4499ae9c0394002778" Dec 13 07:16:43 crc kubenswrapper[4971]: E1213 07:16:43.798672 4971 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"5de93625f93b3f1510893ddd803bdb0ed482aa9dedb3ae4499ae9c0394002778\": container with ID starting with 5de93625f93b3f1510893ddd803bdb0ed482aa9dedb3ae4499ae9c0394002778 not found: ID does not exist" containerID="5de93625f93b3f1510893ddd803bdb0ed482aa9dedb3ae4499ae9c0394002778" Dec 13 07:16:43 crc kubenswrapper[4971]: I1213 07:16:43.798762 4971 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"5de93625f93b3f1510893ddd803bdb0ed482aa9dedb3ae4499ae9c0394002778"} err="failed to get container status \"5de93625f93b3f1510893ddd803bdb0ed482aa9dedb3ae4499ae9c0394002778\": rpc error: code = NotFound desc = could not find container \"5de93625f93b3f1510893ddd803bdb0ed482aa9dedb3ae4499ae9c0394002778\": container with ID starting with 5de93625f93b3f1510893ddd803bdb0ed482aa9dedb3ae4499ae9c0394002778 not found: ID does not exist" Dec 13 07:16:43 crc kubenswrapper[4971]: I1213 07:16:43.798787 4971 scope.go:117] "RemoveContainer" containerID="7326cf828fc3f8902b2798804319ee20d2885e7a77300034607853426b6231d1" Dec 13 07:16:43 crc kubenswrapper[4971]: I1213 07:16:43.799461 4971 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-7jkrp\" (UniqueName: \"kubernetes.io/projected/aadeb784-643a-4e78-909c-4a167bfe11eb-kube-api-access-7jkrp\") pod \"aadeb784-643a-4e78-909c-4a167bfe11eb\" (UID: \"aadeb784-643a-4e78-909c-4a167bfe11eb\") " Dec 13 07:16:43 crc kubenswrapper[4971]: I1213 07:16:43.799613 4971 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/aadeb784-643a-4e78-909c-4a167bfe11eb-ovsdbserver-nb\") pod \"aadeb784-643a-4e78-909c-4a167bfe11eb\" (UID: \"aadeb784-643a-4e78-909c-4a167bfe11eb\") " Dec 13 07:16:43 crc kubenswrapper[4971]: I1213 07:16:43.799691 4971 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/aadeb784-643a-4e78-909c-4a167bfe11eb-dns-svc\") pod \"aadeb784-643a-4e78-909c-4a167bfe11eb\" (UID: \"aadeb784-643a-4e78-909c-4a167bfe11eb\") " Dec 13 07:16:43 crc kubenswrapper[4971]: I1213 07:16:43.799718 4971 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/aadeb784-643a-4e78-909c-4a167bfe11eb-config\") pod \"aadeb784-643a-4e78-909c-4a167bfe11eb\" (UID: \"aadeb784-643a-4e78-909c-4a167bfe11eb\") " Dec 13 07:16:43 crc kubenswrapper[4971]: I1213 07:16:43.799780 4971 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/aadeb784-643a-4e78-909c-4a167bfe11eb-ovsdbserver-sb\") pod \"aadeb784-643a-4e78-909c-4a167bfe11eb\" (UID: \"aadeb784-643a-4e78-909c-4a167bfe11eb\") " Dec 13 07:16:43 crc kubenswrapper[4971]: E1213 07:16:43.800600 4971 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"7326cf828fc3f8902b2798804319ee20d2885e7a77300034607853426b6231d1\": container with ID starting with 7326cf828fc3f8902b2798804319ee20d2885e7a77300034607853426b6231d1 not found: ID does not exist" containerID="7326cf828fc3f8902b2798804319ee20d2885e7a77300034607853426b6231d1" Dec 13 07:16:43 crc kubenswrapper[4971]: I1213 07:16:43.800625 4971 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"7326cf828fc3f8902b2798804319ee20d2885e7a77300034607853426b6231d1"} err="failed to get container status \"7326cf828fc3f8902b2798804319ee20d2885e7a77300034607853426b6231d1\": rpc error: code = NotFound desc = could not find container \"7326cf828fc3f8902b2798804319ee20d2885e7a77300034607853426b6231d1\": container with ID starting with 7326cf828fc3f8902b2798804319ee20d2885e7a77300034607853426b6231d1 not found: ID does not exist" Dec 13 07:16:43 crc kubenswrapper[4971]: I1213 07:16:43.804673 4971 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/aadeb784-643a-4e78-909c-4a167bfe11eb-kube-api-access-7jkrp" (OuterVolumeSpecName: "kube-api-access-7jkrp") pod "aadeb784-643a-4e78-909c-4a167bfe11eb" (UID: "aadeb784-643a-4e78-909c-4a167bfe11eb"). InnerVolumeSpecName "kube-api-access-7jkrp". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 13 07:16:43 crc kubenswrapper[4971]: I1213 07:16:43.858438 4971 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/aadeb784-643a-4e78-909c-4a167bfe11eb-ovsdbserver-nb" (OuterVolumeSpecName: "ovsdbserver-nb") pod "aadeb784-643a-4e78-909c-4a167bfe11eb" (UID: "aadeb784-643a-4e78-909c-4a167bfe11eb"). InnerVolumeSpecName "ovsdbserver-nb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 13 07:16:43 crc kubenswrapper[4971]: I1213 07:16:43.860889 4971 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/aadeb784-643a-4e78-909c-4a167bfe11eb-ovsdbserver-sb" (OuterVolumeSpecName: "ovsdbserver-sb") pod "aadeb784-643a-4e78-909c-4a167bfe11eb" (UID: "aadeb784-643a-4e78-909c-4a167bfe11eb"). InnerVolumeSpecName "ovsdbserver-sb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 13 07:16:43 crc kubenswrapper[4971]: I1213 07:16:43.874223 4971 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/aadeb784-643a-4e78-909c-4a167bfe11eb-config" (OuterVolumeSpecName: "config") pod "aadeb784-643a-4e78-909c-4a167bfe11eb" (UID: "aadeb784-643a-4e78-909c-4a167bfe11eb"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 13 07:16:43 crc kubenswrapper[4971]: I1213 07:16:43.883155 4971 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/aadeb784-643a-4e78-909c-4a167bfe11eb-dns-svc" (OuterVolumeSpecName: "dns-svc") pod "aadeb784-643a-4e78-909c-4a167bfe11eb" (UID: "aadeb784-643a-4e78-909c-4a167bfe11eb"). InnerVolumeSpecName "dns-svc". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 13 07:16:43 crc kubenswrapper[4971]: I1213 07:16:43.902275 4971 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/aadeb784-643a-4e78-909c-4a167bfe11eb-ovsdbserver-nb\") on node \"crc\" DevicePath \"\"" Dec 13 07:16:43 crc kubenswrapper[4971]: I1213 07:16:43.902314 4971 reconciler_common.go:293] "Volume detached for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/aadeb784-643a-4e78-909c-4a167bfe11eb-dns-svc\") on node \"crc\" DevicePath \"\"" Dec 13 07:16:43 crc kubenswrapper[4971]: I1213 07:16:43.902323 4971 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/aadeb784-643a-4e78-909c-4a167bfe11eb-config\") on node \"crc\" DevicePath \"\"" Dec 13 07:16:43 crc kubenswrapper[4971]: I1213 07:16:43.902332 4971 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/aadeb784-643a-4e78-909c-4a167bfe11eb-ovsdbserver-sb\") on node \"crc\" DevicePath \"\"" Dec 13 07:16:43 crc kubenswrapper[4971]: I1213 07:16:43.902340 4971 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-7jkrp\" (UniqueName: \"kubernetes.io/projected/aadeb784-643a-4e78-909c-4a167bfe11eb-kube-api-access-7jkrp\") on node \"crc\" DevicePath \"\"" Dec 13 07:16:44 crc kubenswrapper[4971]: I1213 07:16:44.009490 4971 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-78c596d7cf-9jdzl"] Dec 13 07:16:44 crc kubenswrapper[4971]: I1213 07:16:44.019744 4971 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-78c596d7cf-9jdzl"] Dec 13 07:16:44 crc kubenswrapper[4971]: I1213 07:16:44.126600 4971 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-55d9fc99f5-565xd"] Dec 13 07:16:44 crc kubenswrapper[4971]: I1213 07:16:44.683778 4971 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-55d9fc99f5-565xd" event={"ID":"27e4748a-9fc8-4526-9b48-0d3e30f24268","Type":"ContainerStarted","Data":"ed75a47ddb443493474354190272325d4def38a026694f37cf823cbc035877cf"} Dec 13 07:16:44 crc kubenswrapper[4971]: I1213 07:16:44.684074 4971 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-55d9fc99f5-565xd" event={"ID":"27e4748a-9fc8-4526-9b48-0d3e30f24268","Type":"ContainerStarted","Data":"bec210e8168458ec8df26ec71e1cd92391945df2c7b244bddb888e5a106c9d61"} Dec 13 07:16:45 crc kubenswrapper[4971]: I1213 07:16:45.693551 4971 generic.go:334] "Generic (PLEG): container finished" podID="27e4748a-9fc8-4526-9b48-0d3e30f24268" containerID="ed75a47ddb443493474354190272325d4def38a026694f37cf823cbc035877cf" exitCode=0 Dec 13 07:16:45 crc kubenswrapper[4971]: I1213 07:16:45.693787 4971 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-55d9fc99f5-565xd" event={"ID":"27e4748a-9fc8-4526-9b48-0d3e30f24268","Type":"ContainerDied","Data":"ed75a47ddb443493474354190272325d4def38a026694f37cf823cbc035877cf"} Dec 13 07:16:45 crc kubenswrapper[4971]: I1213 07:16:45.768669 4971 scope.go:117] "RemoveContainer" containerID="e8beee43d15719dccb6892d6f5711674492425300efa098be89033bc65977d14" Dec 13 07:16:45 crc kubenswrapper[4971]: E1213 07:16:45.769155 4971 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-82xjz_openshift-machine-config-operator(e0c4af12-edda-41d7-9b44-f87396174ef2)\"" pod="openshift-machine-config-operator/machine-config-daemon-82xjz" podUID="e0c4af12-edda-41d7-9b44-f87396174ef2" Dec 13 07:16:45 crc kubenswrapper[4971]: I1213 07:16:45.782798 4971 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="aadeb784-643a-4e78-909c-4a167bfe11eb" path="/var/lib/kubelet/pods/aadeb784-643a-4e78-909c-4a167bfe11eb/volumes" Dec 13 07:16:46 crc kubenswrapper[4971]: I1213 07:16:46.712278 4971 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-55d9fc99f5-565xd" event={"ID":"27e4748a-9fc8-4526-9b48-0d3e30f24268","Type":"ContainerStarted","Data":"80f9dad47026016e812f4cd0fc75f8b495a90d7f6dd16030d03e794885c387fa"} Dec 13 07:16:46 crc kubenswrapper[4971]: I1213 07:16:46.713034 4971 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/dnsmasq-dns-55d9fc99f5-565xd" Dec 13 07:16:46 crc kubenswrapper[4971]: I1213 07:16:46.747811 4971 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/dnsmasq-dns-55d9fc99f5-565xd" podStartSLOduration=3.7477930109999997 podStartE2EDuration="3.747793011s" podCreationTimestamp="2025-12-13 07:16:43 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-13 07:16:46.744043548 +0000 UTC m=+1663.348453066" watchObservedRunningTime="2025-12-13 07:16:46.747793011 +0000 UTC m=+1663.352202469" Dec 13 07:16:53 crc kubenswrapper[4971]: I1213 07:16:53.580500 4971 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/dnsmasq-dns-55d9fc99f5-565xd" Dec 13 07:16:53 crc kubenswrapper[4971]: I1213 07:16:53.760244 4971 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-74d67684c-cb4m9"] Dec 13 07:16:53 crc kubenswrapper[4971]: I1213 07:16:53.760478 4971 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/dnsmasq-dns-74d67684c-cb4m9" podUID="62db389f-20fa-444c-af08-3941ad42ae32" containerName="dnsmasq-dns" containerID="cri-o://f6ccf72e6b95053161756a47f6fa73f9875b865403fb3f99f7595ac221bfcfcb" gracePeriod=10 Dec 13 07:16:54 crc kubenswrapper[4971]: I1213 07:16:54.301015 4971 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-74d67684c-cb4m9" Dec 13 07:16:54 crc kubenswrapper[4971]: I1213 07:16:54.494961 4971 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/62db389f-20fa-444c-af08-3941ad42ae32-config\") pod \"62db389f-20fa-444c-af08-3941ad42ae32\" (UID: \"62db389f-20fa-444c-af08-3941ad42ae32\") " Dec 13 07:16:54 crc kubenswrapper[4971]: I1213 07:16:54.495095 4971 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/62db389f-20fa-444c-af08-3941ad42ae32-ovsdbserver-sb\") pod \"62db389f-20fa-444c-af08-3941ad42ae32\" (UID: \"62db389f-20fa-444c-af08-3941ad42ae32\") " Dec 13 07:16:54 crc kubenswrapper[4971]: I1213 07:16:54.495146 4971 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-qmdgk\" (UniqueName: \"kubernetes.io/projected/62db389f-20fa-444c-af08-3941ad42ae32-kube-api-access-qmdgk\") pod \"62db389f-20fa-444c-af08-3941ad42ae32\" (UID: \"62db389f-20fa-444c-af08-3941ad42ae32\") " Dec 13 07:16:54 crc kubenswrapper[4971]: I1213 07:16:54.495303 4971 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/62db389f-20fa-444c-af08-3941ad42ae32-ovsdbserver-nb\") pod \"62db389f-20fa-444c-af08-3941ad42ae32\" (UID: \"62db389f-20fa-444c-af08-3941ad42ae32\") " Dec 13 07:16:54 crc kubenswrapper[4971]: I1213 07:16:54.495360 4971 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/configmap/62db389f-20fa-444c-af08-3941ad42ae32-openstack-edpm-ipam\") pod \"62db389f-20fa-444c-af08-3941ad42ae32\" (UID: \"62db389f-20fa-444c-af08-3941ad42ae32\") " Dec 13 07:16:54 crc kubenswrapper[4971]: I1213 07:16:54.495403 4971 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/62db389f-20fa-444c-af08-3941ad42ae32-dns-svc\") pod \"62db389f-20fa-444c-af08-3941ad42ae32\" (UID: \"62db389f-20fa-444c-af08-3941ad42ae32\") " Dec 13 07:16:54 crc kubenswrapper[4971]: I1213 07:16:54.501134 4971 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/62db389f-20fa-444c-af08-3941ad42ae32-kube-api-access-qmdgk" (OuterVolumeSpecName: "kube-api-access-qmdgk") pod "62db389f-20fa-444c-af08-3941ad42ae32" (UID: "62db389f-20fa-444c-af08-3941ad42ae32"). InnerVolumeSpecName "kube-api-access-qmdgk". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 13 07:16:54 crc kubenswrapper[4971]: I1213 07:16:54.539558 4971 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/62db389f-20fa-444c-af08-3941ad42ae32-ovsdbserver-sb" (OuterVolumeSpecName: "ovsdbserver-sb") pod "62db389f-20fa-444c-af08-3941ad42ae32" (UID: "62db389f-20fa-444c-af08-3941ad42ae32"). InnerVolumeSpecName "ovsdbserver-sb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 13 07:16:54 crc kubenswrapper[4971]: I1213 07:16:54.542969 4971 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/62db389f-20fa-444c-af08-3941ad42ae32-dns-svc" (OuterVolumeSpecName: "dns-svc") pod "62db389f-20fa-444c-af08-3941ad42ae32" (UID: "62db389f-20fa-444c-af08-3941ad42ae32"). InnerVolumeSpecName "dns-svc". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 13 07:16:54 crc kubenswrapper[4971]: I1213 07:16:54.544858 4971 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/62db389f-20fa-444c-af08-3941ad42ae32-ovsdbserver-nb" (OuterVolumeSpecName: "ovsdbserver-nb") pod "62db389f-20fa-444c-af08-3941ad42ae32" (UID: "62db389f-20fa-444c-af08-3941ad42ae32"). InnerVolumeSpecName "ovsdbserver-nb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 13 07:16:54 crc kubenswrapper[4971]: I1213 07:16:54.550254 4971 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/62db389f-20fa-444c-af08-3941ad42ae32-config" (OuterVolumeSpecName: "config") pod "62db389f-20fa-444c-af08-3941ad42ae32" (UID: "62db389f-20fa-444c-af08-3941ad42ae32"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 13 07:16:54 crc kubenswrapper[4971]: I1213 07:16:54.553372 4971 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/62db389f-20fa-444c-af08-3941ad42ae32-openstack-edpm-ipam" (OuterVolumeSpecName: "openstack-edpm-ipam") pod "62db389f-20fa-444c-af08-3941ad42ae32" (UID: "62db389f-20fa-444c-af08-3941ad42ae32"). InnerVolumeSpecName "openstack-edpm-ipam". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 13 07:16:54 crc kubenswrapper[4971]: I1213 07:16:54.598081 4971 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/62db389f-20fa-444c-af08-3941ad42ae32-config\") on node \"crc\" DevicePath \"\"" Dec 13 07:16:54 crc kubenswrapper[4971]: I1213 07:16:54.598124 4971 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/62db389f-20fa-444c-af08-3941ad42ae32-ovsdbserver-sb\") on node \"crc\" DevicePath \"\"" Dec 13 07:16:54 crc kubenswrapper[4971]: I1213 07:16:54.598141 4971 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-qmdgk\" (UniqueName: \"kubernetes.io/projected/62db389f-20fa-444c-af08-3941ad42ae32-kube-api-access-qmdgk\") on node \"crc\" DevicePath \"\"" Dec 13 07:16:54 crc kubenswrapper[4971]: I1213 07:16:54.598154 4971 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/62db389f-20fa-444c-af08-3941ad42ae32-ovsdbserver-nb\") on node \"crc\" DevicePath \"\"" Dec 13 07:16:54 crc kubenswrapper[4971]: I1213 07:16:54.598166 4971 reconciler_common.go:293] "Volume detached for volume \"openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/configmap/62db389f-20fa-444c-af08-3941ad42ae32-openstack-edpm-ipam\") on node \"crc\" DevicePath \"\"" Dec 13 07:16:54 crc kubenswrapper[4971]: I1213 07:16:54.598177 4971 reconciler_common.go:293] "Volume detached for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/62db389f-20fa-444c-af08-3941ad42ae32-dns-svc\") on node \"crc\" DevicePath \"\"" Dec 13 07:16:54 crc kubenswrapper[4971]: I1213 07:16:54.797540 4971 generic.go:334] "Generic (PLEG): container finished" podID="62db389f-20fa-444c-af08-3941ad42ae32" containerID="f6ccf72e6b95053161756a47f6fa73f9875b865403fb3f99f7595ac221bfcfcb" exitCode=0 Dec 13 07:16:54 crc kubenswrapper[4971]: I1213 07:16:54.797590 4971 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-74d67684c-cb4m9" Dec 13 07:16:54 crc kubenswrapper[4971]: I1213 07:16:54.797608 4971 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-74d67684c-cb4m9" event={"ID":"62db389f-20fa-444c-af08-3941ad42ae32","Type":"ContainerDied","Data":"f6ccf72e6b95053161756a47f6fa73f9875b865403fb3f99f7595ac221bfcfcb"} Dec 13 07:16:54 crc kubenswrapper[4971]: I1213 07:16:54.797978 4971 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-74d67684c-cb4m9" event={"ID":"62db389f-20fa-444c-af08-3941ad42ae32","Type":"ContainerDied","Data":"9d55ff49d2681ca36b30e18756a3482706624680084d2a7297d1fd419065e7ee"} Dec 13 07:16:54 crc kubenswrapper[4971]: I1213 07:16:54.798000 4971 scope.go:117] "RemoveContainer" containerID="f6ccf72e6b95053161756a47f6fa73f9875b865403fb3f99f7595ac221bfcfcb" Dec 13 07:16:54 crc kubenswrapper[4971]: I1213 07:16:54.828794 4971 scope.go:117] "RemoveContainer" containerID="0d7551b57ee7908771a5ae1149f919bcf0096d26110da2b6b4e458fb2c2c0848" Dec 13 07:16:54 crc kubenswrapper[4971]: I1213 07:16:54.839276 4971 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-74d67684c-cb4m9"] Dec 13 07:16:54 crc kubenswrapper[4971]: I1213 07:16:54.846787 4971 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-74d67684c-cb4m9"] Dec 13 07:16:54 crc kubenswrapper[4971]: I1213 07:16:54.859396 4971 scope.go:117] "RemoveContainer" containerID="f6ccf72e6b95053161756a47f6fa73f9875b865403fb3f99f7595ac221bfcfcb" Dec 13 07:16:54 crc kubenswrapper[4971]: E1213 07:16:54.859882 4971 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"f6ccf72e6b95053161756a47f6fa73f9875b865403fb3f99f7595ac221bfcfcb\": container with ID starting with f6ccf72e6b95053161756a47f6fa73f9875b865403fb3f99f7595ac221bfcfcb not found: ID does not exist" containerID="f6ccf72e6b95053161756a47f6fa73f9875b865403fb3f99f7595ac221bfcfcb" Dec 13 07:16:54 crc kubenswrapper[4971]: I1213 07:16:54.859912 4971 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"f6ccf72e6b95053161756a47f6fa73f9875b865403fb3f99f7595ac221bfcfcb"} err="failed to get container status \"f6ccf72e6b95053161756a47f6fa73f9875b865403fb3f99f7595ac221bfcfcb\": rpc error: code = NotFound desc = could not find container \"f6ccf72e6b95053161756a47f6fa73f9875b865403fb3f99f7595ac221bfcfcb\": container with ID starting with f6ccf72e6b95053161756a47f6fa73f9875b865403fb3f99f7595ac221bfcfcb not found: ID does not exist" Dec 13 07:16:54 crc kubenswrapper[4971]: I1213 07:16:54.859933 4971 scope.go:117] "RemoveContainer" containerID="0d7551b57ee7908771a5ae1149f919bcf0096d26110da2b6b4e458fb2c2c0848" Dec 13 07:16:54 crc kubenswrapper[4971]: E1213 07:16:54.860172 4971 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"0d7551b57ee7908771a5ae1149f919bcf0096d26110da2b6b4e458fb2c2c0848\": container with ID starting with 0d7551b57ee7908771a5ae1149f919bcf0096d26110da2b6b4e458fb2c2c0848 not found: ID does not exist" containerID="0d7551b57ee7908771a5ae1149f919bcf0096d26110da2b6b4e458fb2c2c0848" Dec 13 07:16:54 crc kubenswrapper[4971]: I1213 07:16:54.860190 4971 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"0d7551b57ee7908771a5ae1149f919bcf0096d26110da2b6b4e458fb2c2c0848"} err="failed to get container status \"0d7551b57ee7908771a5ae1149f919bcf0096d26110da2b6b4e458fb2c2c0848\": rpc error: code = NotFound desc = could not find container \"0d7551b57ee7908771a5ae1149f919bcf0096d26110da2b6b4e458fb2c2c0848\": container with ID starting with 0d7551b57ee7908771a5ae1149f919bcf0096d26110da2b6b4e458fb2c2c0848 not found: ID does not exist" Dec 13 07:16:55 crc kubenswrapper[4971]: I1213 07:16:55.784167 4971 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="62db389f-20fa-444c-af08-3941ad42ae32" path="/var/lib/kubelet/pods/62db389f-20fa-444c-af08-3941ad42ae32/volumes" Dec 13 07:16:59 crc kubenswrapper[4971]: I1213 07:16:59.389347 4971 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-rk2vb"] Dec 13 07:16:59 crc kubenswrapper[4971]: E1213 07:16:59.390222 4971 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="62db389f-20fa-444c-af08-3941ad42ae32" containerName="dnsmasq-dns" Dec 13 07:16:59 crc kubenswrapper[4971]: I1213 07:16:59.390236 4971 state_mem.go:107] "Deleted CPUSet assignment" podUID="62db389f-20fa-444c-af08-3941ad42ae32" containerName="dnsmasq-dns" Dec 13 07:16:59 crc kubenswrapper[4971]: E1213 07:16:59.390254 4971 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="62db389f-20fa-444c-af08-3941ad42ae32" containerName="init" Dec 13 07:16:59 crc kubenswrapper[4971]: I1213 07:16:59.390261 4971 state_mem.go:107] "Deleted CPUSet assignment" podUID="62db389f-20fa-444c-af08-3941ad42ae32" containerName="init" Dec 13 07:16:59 crc kubenswrapper[4971]: E1213 07:16:59.390299 4971 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="aadeb784-643a-4e78-909c-4a167bfe11eb" containerName="dnsmasq-dns" Dec 13 07:16:59 crc kubenswrapper[4971]: I1213 07:16:59.390306 4971 state_mem.go:107] "Deleted CPUSet assignment" podUID="aadeb784-643a-4e78-909c-4a167bfe11eb" containerName="dnsmasq-dns" Dec 13 07:16:59 crc kubenswrapper[4971]: E1213 07:16:59.390323 4971 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="aadeb784-643a-4e78-909c-4a167bfe11eb" containerName="init" Dec 13 07:16:59 crc kubenswrapper[4971]: I1213 07:16:59.390329 4971 state_mem.go:107] "Deleted CPUSet assignment" podUID="aadeb784-643a-4e78-909c-4a167bfe11eb" containerName="init" Dec 13 07:16:59 crc kubenswrapper[4971]: I1213 07:16:59.390559 4971 memory_manager.go:354] "RemoveStaleState removing state" podUID="aadeb784-643a-4e78-909c-4a167bfe11eb" containerName="dnsmasq-dns" Dec 13 07:16:59 crc kubenswrapper[4971]: I1213 07:16:59.390580 4971 memory_manager.go:354] "RemoveStaleState removing state" podUID="62db389f-20fa-444c-af08-3941ad42ae32" containerName="dnsmasq-dns" Dec 13 07:16:59 crc kubenswrapper[4971]: I1213 07:16:59.391150 4971 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-rk2vb" Dec 13 07:16:59 crc kubenswrapper[4971]: I1213 07:16:59.394035 4971 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplane-ansible-ssh-private-key-secret" Dec 13 07:16:59 crc kubenswrapper[4971]: I1213 07:16:59.394731 4971 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplanenodeset-openstack-edpm-ipam" Dec 13 07:16:59 crc kubenswrapper[4971]: I1213 07:16:59.394848 4971 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-aee-default-env" Dec 13 07:16:59 crc kubenswrapper[4971]: I1213 07:16:59.396026 4971 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-edpm-ipam-dockercfg-zb2j5" Dec 13 07:16:59 crc kubenswrapper[4971]: I1213 07:16:59.404366 4971 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-rk2vb"] Dec 13 07:16:59 crc kubenswrapper[4971]: I1213 07:16:59.453557 4971 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-qjlgq\" (UniqueName: \"kubernetes.io/projected/f00ef9b1-1293-4273-90d1-7a7419ebf996-kube-api-access-qjlgq\") pod \"repo-setup-edpm-deployment-openstack-edpm-ipam-rk2vb\" (UID: \"f00ef9b1-1293-4273-90d1-7a7419ebf996\") " pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-rk2vb" Dec 13 07:16:59 crc kubenswrapper[4971]: I1213 07:16:59.453630 4971 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"repo-setup-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/f00ef9b1-1293-4273-90d1-7a7419ebf996-repo-setup-combined-ca-bundle\") pod \"repo-setup-edpm-deployment-openstack-edpm-ipam-rk2vb\" (UID: \"f00ef9b1-1293-4273-90d1-7a7419ebf996\") " pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-rk2vb" Dec 13 07:16:59 crc kubenswrapper[4971]: I1213 07:16:59.453753 4971 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/f00ef9b1-1293-4273-90d1-7a7419ebf996-inventory\") pod \"repo-setup-edpm-deployment-openstack-edpm-ipam-rk2vb\" (UID: \"f00ef9b1-1293-4273-90d1-7a7419ebf996\") " pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-rk2vb" Dec 13 07:16:59 crc kubenswrapper[4971]: I1213 07:16:59.453807 4971 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/f00ef9b1-1293-4273-90d1-7a7419ebf996-ssh-key\") pod \"repo-setup-edpm-deployment-openstack-edpm-ipam-rk2vb\" (UID: \"f00ef9b1-1293-4273-90d1-7a7419ebf996\") " pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-rk2vb" Dec 13 07:16:59 crc kubenswrapper[4971]: I1213 07:16:59.554588 4971 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-qjlgq\" (UniqueName: \"kubernetes.io/projected/f00ef9b1-1293-4273-90d1-7a7419ebf996-kube-api-access-qjlgq\") pod \"repo-setup-edpm-deployment-openstack-edpm-ipam-rk2vb\" (UID: \"f00ef9b1-1293-4273-90d1-7a7419ebf996\") " pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-rk2vb" Dec 13 07:16:59 crc kubenswrapper[4971]: I1213 07:16:59.554947 4971 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"repo-setup-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/f00ef9b1-1293-4273-90d1-7a7419ebf996-repo-setup-combined-ca-bundle\") pod \"repo-setup-edpm-deployment-openstack-edpm-ipam-rk2vb\" (UID: \"f00ef9b1-1293-4273-90d1-7a7419ebf996\") " pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-rk2vb" Dec 13 07:16:59 crc kubenswrapper[4971]: I1213 07:16:59.555056 4971 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/f00ef9b1-1293-4273-90d1-7a7419ebf996-inventory\") pod \"repo-setup-edpm-deployment-openstack-edpm-ipam-rk2vb\" (UID: \"f00ef9b1-1293-4273-90d1-7a7419ebf996\") " pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-rk2vb" Dec 13 07:16:59 crc kubenswrapper[4971]: I1213 07:16:59.555102 4971 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/f00ef9b1-1293-4273-90d1-7a7419ebf996-ssh-key\") pod \"repo-setup-edpm-deployment-openstack-edpm-ipam-rk2vb\" (UID: \"f00ef9b1-1293-4273-90d1-7a7419ebf996\") " pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-rk2vb" Dec 13 07:16:59 crc kubenswrapper[4971]: I1213 07:16:59.560181 4971 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/f00ef9b1-1293-4273-90d1-7a7419ebf996-ssh-key\") pod \"repo-setup-edpm-deployment-openstack-edpm-ipam-rk2vb\" (UID: \"f00ef9b1-1293-4273-90d1-7a7419ebf996\") " pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-rk2vb" Dec 13 07:16:59 crc kubenswrapper[4971]: I1213 07:16:59.560404 4971 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"repo-setup-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/f00ef9b1-1293-4273-90d1-7a7419ebf996-repo-setup-combined-ca-bundle\") pod \"repo-setup-edpm-deployment-openstack-edpm-ipam-rk2vb\" (UID: \"f00ef9b1-1293-4273-90d1-7a7419ebf996\") " pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-rk2vb" Dec 13 07:16:59 crc kubenswrapper[4971]: I1213 07:16:59.564289 4971 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/f00ef9b1-1293-4273-90d1-7a7419ebf996-inventory\") pod \"repo-setup-edpm-deployment-openstack-edpm-ipam-rk2vb\" (UID: \"f00ef9b1-1293-4273-90d1-7a7419ebf996\") " pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-rk2vb" Dec 13 07:16:59 crc kubenswrapper[4971]: I1213 07:16:59.574184 4971 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-qjlgq\" (UniqueName: \"kubernetes.io/projected/f00ef9b1-1293-4273-90d1-7a7419ebf996-kube-api-access-qjlgq\") pod \"repo-setup-edpm-deployment-openstack-edpm-ipam-rk2vb\" (UID: \"f00ef9b1-1293-4273-90d1-7a7419ebf996\") " pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-rk2vb" Dec 13 07:16:59 crc kubenswrapper[4971]: I1213 07:16:59.715876 4971 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-rk2vb" Dec 13 07:17:00 crc kubenswrapper[4971]: W1213 07:17:00.318858 4971 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podf00ef9b1_1293_4273_90d1_7a7419ebf996.slice/crio-6042c24b2485dd4820aa9a44e686d6a7dddeb613c3a5fbfa49a5d34467634902 WatchSource:0}: Error finding container 6042c24b2485dd4820aa9a44e686d6a7dddeb613c3a5fbfa49a5d34467634902: Status 404 returned error can't find the container with id 6042c24b2485dd4820aa9a44e686d6a7dddeb613c3a5fbfa49a5d34467634902 Dec 13 07:17:00 crc kubenswrapper[4971]: I1213 07:17:00.321946 4971 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-rk2vb"] Dec 13 07:17:00 crc kubenswrapper[4971]: I1213 07:17:00.768639 4971 scope.go:117] "RemoveContainer" containerID="e8beee43d15719dccb6892d6f5711674492425300efa098be89033bc65977d14" Dec 13 07:17:00 crc kubenswrapper[4971]: E1213 07:17:00.769285 4971 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-82xjz_openshift-machine-config-operator(e0c4af12-edda-41d7-9b44-f87396174ef2)\"" pod="openshift-machine-config-operator/machine-config-daemon-82xjz" podUID="e0c4af12-edda-41d7-9b44-f87396174ef2" Dec 13 07:17:00 crc kubenswrapper[4971]: I1213 07:17:00.876063 4971 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-rk2vb" event={"ID":"f00ef9b1-1293-4273-90d1-7a7419ebf996","Type":"ContainerStarted","Data":"6042c24b2485dd4820aa9a44e686d6a7dddeb613c3a5fbfa49a5d34467634902"} Dec 13 07:17:02 crc kubenswrapper[4971]: I1213 07:17:02.896488 4971 generic.go:334] "Generic (PLEG): container finished" podID="42eb5011-6651-43cb-a562-c89cc6df2c53" containerID="61c8ef8fd4dd958d475d5f013ee5bfe47c0dda61b86af7a1955684dfb85be60b" exitCode=0 Dec 13 07:17:02 crc kubenswrapper[4971]: I1213 07:17:02.896613 4971 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-server-0" event={"ID":"42eb5011-6651-43cb-a562-c89cc6df2c53","Type":"ContainerDied","Data":"61c8ef8fd4dd958d475d5f013ee5bfe47c0dda61b86af7a1955684dfb85be60b"} Dec 13 07:17:03 crc kubenswrapper[4971]: I1213 07:17:03.921858 4971 generic.go:334] "Generic (PLEG): container finished" podID="299317ec-a652-40f2-b0c5-e25f099204c4" containerID="124c26f1f489505ddde7c5fedd4dc0853250ddf7ad07873202149ec0a78af095" exitCode=0 Dec 13 07:17:03 crc kubenswrapper[4971]: I1213 07:17:03.922460 4971 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-cell1-server-0" event={"ID":"299317ec-a652-40f2-b0c5-e25f099204c4","Type":"ContainerDied","Data":"124c26f1f489505ddde7c5fedd4dc0853250ddf7ad07873202149ec0a78af095"} Dec 13 07:17:04 crc kubenswrapper[4971]: I1213 07:17:04.937073 4971 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-cell1-server-0" event={"ID":"299317ec-a652-40f2-b0c5-e25f099204c4","Type":"ContainerStarted","Data":"89fab04e1ba30673ebeb75daee58bd4b9324982177d76d6ef659d05f1f0fd48f"} Dec 13 07:17:04 crc kubenswrapper[4971]: I1213 07:17:04.937653 4971 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/rabbitmq-cell1-server-0" Dec 13 07:17:04 crc kubenswrapper[4971]: I1213 07:17:04.941271 4971 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-server-0" event={"ID":"42eb5011-6651-43cb-a562-c89cc6df2c53","Type":"ContainerStarted","Data":"c16b262c11df26849f81317fd8ab64f048b4d1b36886afe55ec576e20e047856"} Dec 13 07:17:04 crc kubenswrapper[4971]: I1213 07:17:04.941512 4971 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/rabbitmq-server-0" Dec 13 07:17:04 crc kubenswrapper[4971]: I1213 07:17:04.974789 4971 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/rabbitmq-cell1-server-0" podStartSLOduration=37.974772764 podStartE2EDuration="37.974772764s" podCreationTimestamp="2025-12-13 07:16:27 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-13 07:17:04.966774496 +0000 UTC m=+1681.571183944" watchObservedRunningTime="2025-12-13 07:17:04.974772764 +0000 UTC m=+1681.579182212" Dec 13 07:17:04 crc kubenswrapper[4971]: I1213 07:17:04.992202 4971 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/rabbitmq-server-0" podStartSLOduration=38.992188892 podStartE2EDuration="38.992188892s" podCreationTimestamp="2025-12-13 07:16:26 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-13 07:17:04.987075887 +0000 UTC m=+1681.591485335" watchObservedRunningTime="2025-12-13 07:17:04.992188892 +0000 UTC m=+1681.596598340" Dec 13 07:17:13 crc kubenswrapper[4971]: I1213 07:17:13.127952 4971 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-aee-default-env" Dec 13 07:17:14 crc kubenswrapper[4971]: I1213 07:17:14.091793 4971 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-rk2vb" event={"ID":"f00ef9b1-1293-4273-90d1-7a7419ebf996","Type":"ContainerStarted","Data":"b1f374774ca878889bdc02f923abb290b5fdaf86d3cfbee33de205252d676d59"} Dec 13 07:17:14 crc kubenswrapper[4971]: I1213 07:17:14.108554 4971 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-rk2vb" podStartSLOduration=2.30568814 podStartE2EDuration="15.108539116s" podCreationTimestamp="2025-12-13 07:16:59 +0000 UTC" firstStartedPulling="2025-12-13 07:17:00.32146948 +0000 UTC m=+1676.925878928" lastFinishedPulling="2025-12-13 07:17:13.124320446 +0000 UTC m=+1689.728729904" observedRunningTime="2025-12-13 07:17:14.107134682 +0000 UTC m=+1690.711544130" watchObservedRunningTime="2025-12-13 07:17:14.108539116 +0000 UTC m=+1690.712948564" Dec 13 07:17:14 crc kubenswrapper[4971]: I1213 07:17:14.768823 4971 scope.go:117] "RemoveContainer" containerID="e8beee43d15719dccb6892d6f5711674492425300efa098be89033bc65977d14" Dec 13 07:17:14 crc kubenswrapper[4971]: E1213 07:17:14.770189 4971 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-82xjz_openshift-machine-config-operator(e0c4af12-edda-41d7-9b44-f87396174ef2)\"" pod="openshift-machine-config-operator/machine-config-daemon-82xjz" podUID="e0c4af12-edda-41d7-9b44-f87396174ef2" Dec 13 07:17:16 crc kubenswrapper[4971]: I1213 07:17:16.865757 4971 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/rabbitmq-server-0" Dec 13 07:17:18 crc kubenswrapper[4971]: I1213 07:17:18.129708 4971 generic.go:334] "Generic (PLEG): container finished" podID="f00ef9b1-1293-4273-90d1-7a7419ebf996" containerID="b1f374774ca878889bdc02f923abb290b5fdaf86d3cfbee33de205252d676d59" exitCode=2 Dec 13 07:17:18 crc kubenswrapper[4971]: I1213 07:17:18.129751 4971 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-rk2vb" event={"ID":"f00ef9b1-1293-4273-90d1-7a7419ebf996","Type":"ContainerDied","Data":"b1f374774ca878889bdc02f923abb290b5fdaf86d3cfbee33de205252d676d59"} Dec 13 07:17:18 crc kubenswrapper[4971]: I1213 07:17:18.181712 4971 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/rabbitmq-cell1-server-0" Dec 13 07:17:19 crc kubenswrapper[4971]: I1213 07:17:19.965004 4971 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-rk2vb" Dec 13 07:17:20 crc kubenswrapper[4971]: I1213 07:17:20.125350 4971 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"repo-setup-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/f00ef9b1-1293-4273-90d1-7a7419ebf996-repo-setup-combined-ca-bundle\") pod \"f00ef9b1-1293-4273-90d1-7a7419ebf996\" (UID: \"f00ef9b1-1293-4273-90d1-7a7419ebf996\") " Dec 13 07:17:20 crc kubenswrapper[4971]: I1213 07:17:20.125682 4971 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-qjlgq\" (UniqueName: \"kubernetes.io/projected/f00ef9b1-1293-4273-90d1-7a7419ebf996-kube-api-access-qjlgq\") pod \"f00ef9b1-1293-4273-90d1-7a7419ebf996\" (UID: \"f00ef9b1-1293-4273-90d1-7a7419ebf996\") " Dec 13 07:17:20 crc kubenswrapper[4971]: I1213 07:17:20.125753 4971 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/f00ef9b1-1293-4273-90d1-7a7419ebf996-ssh-key\") pod \"f00ef9b1-1293-4273-90d1-7a7419ebf996\" (UID: \"f00ef9b1-1293-4273-90d1-7a7419ebf996\") " Dec 13 07:17:20 crc kubenswrapper[4971]: I1213 07:17:20.125801 4971 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/f00ef9b1-1293-4273-90d1-7a7419ebf996-inventory\") pod \"f00ef9b1-1293-4273-90d1-7a7419ebf996\" (UID: \"f00ef9b1-1293-4273-90d1-7a7419ebf996\") " Dec 13 07:17:20 crc kubenswrapper[4971]: I1213 07:17:20.130771 4971 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/f00ef9b1-1293-4273-90d1-7a7419ebf996-kube-api-access-qjlgq" (OuterVolumeSpecName: "kube-api-access-qjlgq") pod "f00ef9b1-1293-4273-90d1-7a7419ebf996" (UID: "f00ef9b1-1293-4273-90d1-7a7419ebf996"). InnerVolumeSpecName "kube-api-access-qjlgq". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 13 07:17:20 crc kubenswrapper[4971]: I1213 07:17:20.131702 4971 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/f00ef9b1-1293-4273-90d1-7a7419ebf996-repo-setup-combined-ca-bundle" (OuterVolumeSpecName: "repo-setup-combined-ca-bundle") pod "f00ef9b1-1293-4273-90d1-7a7419ebf996" (UID: "f00ef9b1-1293-4273-90d1-7a7419ebf996"). InnerVolumeSpecName "repo-setup-combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 13 07:17:20 crc kubenswrapper[4971]: I1213 07:17:20.159908 4971 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/f00ef9b1-1293-4273-90d1-7a7419ebf996-inventory" (OuterVolumeSpecName: "inventory") pod "f00ef9b1-1293-4273-90d1-7a7419ebf996" (UID: "f00ef9b1-1293-4273-90d1-7a7419ebf996"). InnerVolumeSpecName "inventory". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 13 07:17:20 crc kubenswrapper[4971]: I1213 07:17:20.165585 4971 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-rk2vb" event={"ID":"f00ef9b1-1293-4273-90d1-7a7419ebf996","Type":"ContainerDied","Data":"6042c24b2485dd4820aa9a44e686d6a7dddeb613c3a5fbfa49a5d34467634902"} Dec 13 07:17:20 crc kubenswrapper[4971]: I1213 07:17:20.165633 4971 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="6042c24b2485dd4820aa9a44e686d6a7dddeb613c3a5fbfa49a5d34467634902" Dec 13 07:17:20 crc kubenswrapper[4971]: I1213 07:17:20.165703 4971 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-rk2vb" Dec 13 07:17:20 crc kubenswrapper[4971]: I1213 07:17:20.172203 4971 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/f00ef9b1-1293-4273-90d1-7a7419ebf996-ssh-key" (OuterVolumeSpecName: "ssh-key") pod "f00ef9b1-1293-4273-90d1-7a7419ebf996" (UID: "f00ef9b1-1293-4273-90d1-7a7419ebf996"). InnerVolumeSpecName "ssh-key". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 13 07:17:20 crc kubenswrapper[4971]: I1213 07:17:20.228105 4971 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-qjlgq\" (UniqueName: \"kubernetes.io/projected/f00ef9b1-1293-4273-90d1-7a7419ebf996-kube-api-access-qjlgq\") on node \"crc\" DevicePath \"\"" Dec 13 07:17:20 crc kubenswrapper[4971]: I1213 07:17:20.228136 4971 reconciler_common.go:293] "Volume detached for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/f00ef9b1-1293-4273-90d1-7a7419ebf996-ssh-key\") on node \"crc\" DevicePath \"\"" Dec 13 07:17:20 crc kubenswrapper[4971]: I1213 07:17:20.228146 4971 reconciler_common.go:293] "Volume detached for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/f00ef9b1-1293-4273-90d1-7a7419ebf996-inventory\") on node \"crc\" DevicePath \"\"" Dec 13 07:17:20 crc kubenswrapper[4971]: I1213 07:17:20.228156 4971 reconciler_common.go:293] "Volume detached for volume \"repo-setup-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/f00ef9b1-1293-4273-90d1-7a7419ebf996-repo-setup-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 13 07:17:26 crc kubenswrapper[4971]: I1213 07:17:26.769213 4971 scope.go:117] "RemoveContainer" containerID="e8beee43d15719dccb6892d6f5711674492425300efa098be89033bc65977d14" Dec 13 07:17:26 crc kubenswrapper[4971]: E1213 07:17:26.770264 4971 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-82xjz_openshift-machine-config-operator(e0c4af12-edda-41d7-9b44-f87396174ef2)\"" pod="openshift-machine-config-operator/machine-config-daemon-82xjz" podUID="e0c4af12-edda-41d7-9b44-f87396174ef2" Dec 13 07:17:27 crc kubenswrapper[4971]: I1213 07:17:27.022551 4971 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-q8m6f"] Dec 13 07:17:27 crc kubenswrapper[4971]: E1213 07:17:27.022962 4971 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f00ef9b1-1293-4273-90d1-7a7419ebf996" containerName="repo-setup-edpm-deployment-openstack-edpm-ipam" Dec 13 07:17:27 crc kubenswrapper[4971]: I1213 07:17:27.022984 4971 state_mem.go:107] "Deleted CPUSet assignment" podUID="f00ef9b1-1293-4273-90d1-7a7419ebf996" containerName="repo-setup-edpm-deployment-openstack-edpm-ipam" Dec 13 07:17:27 crc kubenswrapper[4971]: I1213 07:17:27.023156 4971 memory_manager.go:354] "RemoveStaleState removing state" podUID="f00ef9b1-1293-4273-90d1-7a7419ebf996" containerName="repo-setup-edpm-deployment-openstack-edpm-ipam" Dec 13 07:17:27 crc kubenswrapper[4971]: I1213 07:17:27.023773 4971 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-q8m6f" Dec 13 07:17:27 crc kubenswrapper[4971]: I1213 07:17:27.026661 4971 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplane-ansible-ssh-private-key-secret" Dec 13 07:17:27 crc kubenswrapper[4971]: I1213 07:17:27.027445 4971 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-aee-default-env" Dec 13 07:17:27 crc kubenswrapper[4971]: I1213 07:17:27.027680 4971 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplanenodeset-openstack-edpm-ipam" Dec 13 07:17:27 crc kubenswrapper[4971]: I1213 07:17:27.027898 4971 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-edpm-ipam-dockercfg-zb2j5" Dec 13 07:17:27 crc kubenswrapper[4971]: I1213 07:17:27.039031 4971 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-q8m6f"] Dec 13 07:17:27 crc kubenswrapper[4971]: I1213 07:17:27.157993 4971 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-8hdgh\" (UniqueName: \"kubernetes.io/projected/895771e3-67cb-4847-ab4b-3b7edfe79102-kube-api-access-8hdgh\") pod \"repo-setup-edpm-deployment-openstack-edpm-ipam-q8m6f\" (UID: \"895771e3-67cb-4847-ab4b-3b7edfe79102\") " pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-q8m6f" Dec 13 07:17:27 crc kubenswrapper[4971]: I1213 07:17:27.158127 4971 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/895771e3-67cb-4847-ab4b-3b7edfe79102-ssh-key\") pod \"repo-setup-edpm-deployment-openstack-edpm-ipam-q8m6f\" (UID: \"895771e3-67cb-4847-ab4b-3b7edfe79102\") " pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-q8m6f" Dec 13 07:17:27 crc kubenswrapper[4971]: I1213 07:17:27.158170 4971 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"repo-setup-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/895771e3-67cb-4847-ab4b-3b7edfe79102-repo-setup-combined-ca-bundle\") pod \"repo-setup-edpm-deployment-openstack-edpm-ipam-q8m6f\" (UID: \"895771e3-67cb-4847-ab4b-3b7edfe79102\") " pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-q8m6f" Dec 13 07:17:27 crc kubenswrapper[4971]: I1213 07:17:27.158295 4971 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/895771e3-67cb-4847-ab4b-3b7edfe79102-inventory\") pod \"repo-setup-edpm-deployment-openstack-edpm-ipam-q8m6f\" (UID: \"895771e3-67cb-4847-ab4b-3b7edfe79102\") " pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-q8m6f" Dec 13 07:17:27 crc kubenswrapper[4971]: I1213 07:17:27.260543 4971 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/895771e3-67cb-4847-ab4b-3b7edfe79102-ssh-key\") pod \"repo-setup-edpm-deployment-openstack-edpm-ipam-q8m6f\" (UID: \"895771e3-67cb-4847-ab4b-3b7edfe79102\") " pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-q8m6f" Dec 13 07:17:27 crc kubenswrapper[4971]: I1213 07:17:27.260602 4971 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"repo-setup-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/895771e3-67cb-4847-ab4b-3b7edfe79102-repo-setup-combined-ca-bundle\") pod \"repo-setup-edpm-deployment-openstack-edpm-ipam-q8m6f\" (UID: \"895771e3-67cb-4847-ab4b-3b7edfe79102\") " pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-q8m6f" Dec 13 07:17:27 crc kubenswrapper[4971]: I1213 07:17:27.260699 4971 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/895771e3-67cb-4847-ab4b-3b7edfe79102-inventory\") pod \"repo-setup-edpm-deployment-openstack-edpm-ipam-q8m6f\" (UID: \"895771e3-67cb-4847-ab4b-3b7edfe79102\") " pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-q8m6f" Dec 13 07:17:27 crc kubenswrapper[4971]: I1213 07:17:27.260852 4971 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-8hdgh\" (UniqueName: \"kubernetes.io/projected/895771e3-67cb-4847-ab4b-3b7edfe79102-kube-api-access-8hdgh\") pod \"repo-setup-edpm-deployment-openstack-edpm-ipam-q8m6f\" (UID: \"895771e3-67cb-4847-ab4b-3b7edfe79102\") " pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-q8m6f" Dec 13 07:17:27 crc kubenswrapper[4971]: I1213 07:17:27.268065 4971 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"repo-setup-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/895771e3-67cb-4847-ab4b-3b7edfe79102-repo-setup-combined-ca-bundle\") pod \"repo-setup-edpm-deployment-openstack-edpm-ipam-q8m6f\" (UID: \"895771e3-67cb-4847-ab4b-3b7edfe79102\") " pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-q8m6f" Dec 13 07:17:27 crc kubenswrapper[4971]: I1213 07:17:27.268418 4971 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/895771e3-67cb-4847-ab4b-3b7edfe79102-inventory\") pod \"repo-setup-edpm-deployment-openstack-edpm-ipam-q8m6f\" (UID: \"895771e3-67cb-4847-ab4b-3b7edfe79102\") " pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-q8m6f" Dec 13 07:17:27 crc kubenswrapper[4971]: I1213 07:17:27.270231 4971 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/895771e3-67cb-4847-ab4b-3b7edfe79102-ssh-key\") pod \"repo-setup-edpm-deployment-openstack-edpm-ipam-q8m6f\" (UID: \"895771e3-67cb-4847-ab4b-3b7edfe79102\") " pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-q8m6f" Dec 13 07:17:27 crc kubenswrapper[4971]: I1213 07:17:27.282046 4971 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-8hdgh\" (UniqueName: \"kubernetes.io/projected/895771e3-67cb-4847-ab4b-3b7edfe79102-kube-api-access-8hdgh\") pod \"repo-setup-edpm-deployment-openstack-edpm-ipam-q8m6f\" (UID: \"895771e3-67cb-4847-ab4b-3b7edfe79102\") " pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-q8m6f" Dec 13 07:17:27 crc kubenswrapper[4971]: I1213 07:17:27.350745 4971 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-q8m6f" Dec 13 07:17:27 crc kubenswrapper[4971]: I1213 07:17:27.871608 4971 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-q8m6f"] Dec 13 07:17:28 crc kubenswrapper[4971]: I1213 07:17:28.242907 4971 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-q8m6f" event={"ID":"895771e3-67cb-4847-ab4b-3b7edfe79102","Type":"ContainerStarted","Data":"46c0fe6fb0da705578eddedecd487ebad764b834531e3bc7e4c1d8b97eab4804"} Dec 13 07:17:29 crc kubenswrapper[4971]: I1213 07:17:29.256732 4971 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-q8m6f" event={"ID":"895771e3-67cb-4847-ab4b-3b7edfe79102","Type":"ContainerStarted","Data":"0ef36e730b8b14be062f8a9cd0480202fc65350a0964daedd0d8ca68547dd758"} Dec 13 07:17:29 crc kubenswrapper[4971]: I1213 07:17:29.282282 4971 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-q8m6f" podStartSLOduration=1.828503091 podStartE2EDuration="2.282259276s" podCreationTimestamp="2025-12-13 07:17:27 +0000 UTC" firstStartedPulling="2025-12-13 07:17:27.88096467 +0000 UTC m=+1704.485374118" lastFinishedPulling="2025-12-13 07:17:28.334720855 +0000 UTC m=+1704.939130303" observedRunningTime="2025-12-13 07:17:29.275273363 +0000 UTC m=+1705.879682821" watchObservedRunningTime="2025-12-13 07:17:29.282259276 +0000 UTC m=+1705.886668734" Dec 13 07:17:39 crc kubenswrapper[4971]: I1213 07:17:39.768670 4971 scope.go:117] "RemoveContainer" containerID="e8beee43d15719dccb6892d6f5711674492425300efa098be89033bc65977d14" Dec 13 07:17:39 crc kubenswrapper[4971]: E1213 07:17:39.769674 4971 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-82xjz_openshift-machine-config-operator(e0c4af12-edda-41d7-9b44-f87396174ef2)\"" pod="openshift-machine-config-operator/machine-config-daemon-82xjz" podUID="e0c4af12-edda-41d7-9b44-f87396174ef2" Dec 13 07:17:44 crc kubenswrapper[4971]: I1213 07:17:44.435948 4971 generic.go:334] "Generic (PLEG): container finished" podID="895771e3-67cb-4847-ab4b-3b7edfe79102" containerID="0ef36e730b8b14be062f8a9cd0480202fc65350a0964daedd0d8ca68547dd758" exitCode=2 Dec 13 07:17:44 crc kubenswrapper[4971]: I1213 07:17:44.436072 4971 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-q8m6f" event={"ID":"895771e3-67cb-4847-ab4b-3b7edfe79102","Type":"ContainerDied","Data":"0ef36e730b8b14be062f8a9cd0480202fc65350a0964daedd0d8ca68547dd758"} Dec 13 07:17:45 crc kubenswrapper[4971]: I1213 07:17:45.922672 4971 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-q8m6f" Dec 13 07:17:46 crc kubenswrapper[4971]: I1213 07:17:46.037255 4971 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-8hdgh\" (UniqueName: \"kubernetes.io/projected/895771e3-67cb-4847-ab4b-3b7edfe79102-kube-api-access-8hdgh\") pod \"895771e3-67cb-4847-ab4b-3b7edfe79102\" (UID: \"895771e3-67cb-4847-ab4b-3b7edfe79102\") " Dec 13 07:17:46 crc kubenswrapper[4971]: I1213 07:17:46.037413 4971 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"repo-setup-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/895771e3-67cb-4847-ab4b-3b7edfe79102-repo-setup-combined-ca-bundle\") pod \"895771e3-67cb-4847-ab4b-3b7edfe79102\" (UID: \"895771e3-67cb-4847-ab4b-3b7edfe79102\") " Dec 13 07:17:46 crc kubenswrapper[4971]: I1213 07:17:46.037447 4971 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/895771e3-67cb-4847-ab4b-3b7edfe79102-ssh-key\") pod \"895771e3-67cb-4847-ab4b-3b7edfe79102\" (UID: \"895771e3-67cb-4847-ab4b-3b7edfe79102\") " Dec 13 07:17:46 crc kubenswrapper[4971]: I1213 07:17:46.037479 4971 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/895771e3-67cb-4847-ab4b-3b7edfe79102-inventory\") pod \"895771e3-67cb-4847-ab4b-3b7edfe79102\" (UID: \"895771e3-67cb-4847-ab4b-3b7edfe79102\") " Dec 13 07:17:46 crc kubenswrapper[4971]: I1213 07:17:46.043390 4971 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/895771e3-67cb-4847-ab4b-3b7edfe79102-kube-api-access-8hdgh" (OuterVolumeSpecName: "kube-api-access-8hdgh") pod "895771e3-67cb-4847-ab4b-3b7edfe79102" (UID: "895771e3-67cb-4847-ab4b-3b7edfe79102"). InnerVolumeSpecName "kube-api-access-8hdgh". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 13 07:17:46 crc kubenswrapper[4971]: I1213 07:17:46.044447 4971 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/895771e3-67cb-4847-ab4b-3b7edfe79102-repo-setup-combined-ca-bundle" (OuterVolumeSpecName: "repo-setup-combined-ca-bundle") pod "895771e3-67cb-4847-ab4b-3b7edfe79102" (UID: "895771e3-67cb-4847-ab4b-3b7edfe79102"). InnerVolumeSpecName "repo-setup-combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 13 07:17:46 crc kubenswrapper[4971]: I1213 07:17:46.072604 4971 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/895771e3-67cb-4847-ab4b-3b7edfe79102-inventory" (OuterVolumeSpecName: "inventory") pod "895771e3-67cb-4847-ab4b-3b7edfe79102" (UID: "895771e3-67cb-4847-ab4b-3b7edfe79102"). InnerVolumeSpecName "inventory". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 13 07:17:46 crc kubenswrapper[4971]: I1213 07:17:46.076961 4971 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/895771e3-67cb-4847-ab4b-3b7edfe79102-ssh-key" (OuterVolumeSpecName: "ssh-key") pod "895771e3-67cb-4847-ab4b-3b7edfe79102" (UID: "895771e3-67cb-4847-ab4b-3b7edfe79102"). InnerVolumeSpecName "ssh-key". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 13 07:17:46 crc kubenswrapper[4971]: I1213 07:17:46.140166 4971 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-8hdgh\" (UniqueName: \"kubernetes.io/projected/895771e3-67cb-4847-ab4b-3b7edfe79102-kube-api-access-8hdgh\") on node \"crc\" DevicePath \"\"" Dec 13 07:17:46 crc kubenswrapper[4971]: I1213 07:17:46.140202 4971 reconciler_common.go:293] "Volume detached for volume \"repo-setup-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/895771e3-67cb-4847-ab4b-3b7edfe79102-repo-setup-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 13 07:17:46 crc kubenswrapper[4971]: I1213 07:17:46.140217 4971 reconciler_common.go:293] "Volume detached for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/895771e3-67cb-4847-ab4b-3b7edfe79102-ssh-key\") on node \"crc\" DevicePath \"\"" Dec 13 07:17:46 crc kubenswrapper[4971]: I1213 07:17:46.140230 4971 reconciler_common.go:293] "Volume detached for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/895771e3-67cb-4847-ab4b-3b7edfe79102-inventory\") on node \"crc\" DevicePath \"\"" Dec 13 07:17:46 crc kubenswrapper[4971]: I1213 07:17:46.462125 4971 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-q8m6f" event={"ID":"895771e3-67cb-4847-ab4b-3b7edfe79102","Type":"ContainerDied","Data":"46c0fe6fb0da705578eddedecd487ebad764b834531e3bc7e4c1d8b97eab4804"} Dec 13 07:17:46 crc kubenswrapper[4971]: I1213 07:17:46.462198 4971 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-q8m6f" Dec 13 07:17:46 crc kubenswrapper[4971]: I1213 07:17:46.462204 4971 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="46c0fe6fb0da705578eddedecd487ebad764b834531e3bc7e4c1d8b97eab4804" Dec 13 07:17:51 crc kubenswrapper[4971]: I1213 07:17:51.768019 4971 scope.go:117] "RemoveContainer" containerID="e8beee43d15719dccb6892d6f5711674492425300efa098be89033bc65977d14" Dec 13 07:17:51 crc kubenswrapper[4971]: E1213 07:17:51.770081 4971 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-82xjz_openshift-machine-config-operator(e0c4af12-edda-41d7-9b44-f87396174ef2)\"" pod="openshift-machine-config-operator/machine-config-daemon-82xjz" podUID="e0c4af12-edda-41d7-9b44-f87396174ef2" Dec 13 07:18:03 crc kubenswrapper[4971]: I1213 07:18:03.067608 4971 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-fgt8j"] Dec 13 07:18:03 crc kubenswrapper[4971]: E1213 07:18:03.069043 4971 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="895771e3-67cb-4847-ab4b-3b7edfe79102" containerName="repo-setup-edpm-deployment-openstack-edpm-ipam" Dec 13 07:18:03 crc kubenswrapper[4971]: I1213 07:18:03.069076 4971 state_mem.go:107] "Deleted CPUSet assignment" podUID="895771e3-67cb-4847-ab4b-3b7edfe79102" containerName="repo-setup-edpm-deployment-openstack-edpm-ipam" Dec 13 07:18:03 crc kubenswrapper[4971]: I1213 07:18:03.069418 4971 memory_manager.go:354] "RemoveStaleState removing state" podUID="895771e3-67cb-4847-ab4b-3b7edfe79102" containerName="repo-setup-edpm-deployment-openstack-edpm-ipam" Dec 13 07:18:03 crc kubenswrapper[4971]: I1213 07:18:03.070608 4971 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-fgt8j" Dec 13 07:18:03 crc kubenswrapper[4971]: I1213 07:18:03.073163 4971 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-edpm-ipam-dockercfg-zb2j5" Dec 13 07:18:03 crc kubenswrapper[4971]: I1213 07:18:03.073864 4971 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-aee-default-env" Dec 13 07:18:03 crc kubenswrapper[4971]: I1213 07:18:03.074067 4971 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplane-ansible-ssh-private-key-secret" Dec 13 07:18:03 crc kubenswrapper[4971]: I1213 07:18:03.076712 4971 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplanenodeset-openstack-edpm-ipam" Dec 13 07:18:03 crc kubenswrapper[4971]: I1213 07:18:03.083340 4971 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-fgt8j"] Dec 13 07:18:03 crc kubenswrapper[4971]: I1213 07:18:03.174740 4971 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/90351fb3-f8c9-4db4-b577-8820f50ccf15-inventory\") pod \"repo-setup-edpm-deployment-openstack-edpm-ipam-fgt8j\" (UID: \"90351fb3-f8c9-4db4-b577-8820f50ccf15\") " pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-fgt8j" Dec 13 07:18:03 crc kubenswrapper[4971]: I1213 07:18:03.174780 4971 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"repo-setup-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/90351fb3-f8c9-4db4-b577-8820f50ccf15-repo-setup-combined-ca-bundle\") pod \"repo-setup-edpm-deployment-openstack-edpm-ipam-fgt8j\" (UID: \"90351fb3-f8c9-4db4-b577-8820f50ccf15\") " pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-fgt8j" Dec 13 07:18:03 crc kubenswrapper[4971]: I1213 07:18:03.175025 4971 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-z84fz\" (UniqueName: \"kubernetes.io/projected/90351fb3-f8c9-4db4-b577-8820f50ccf15-kube-api-access-z84fz\") pod \"repo-setup-edpm-deployment-openstack-edpm-ipam-fgt8j\" (UID: \"90351fb3-f8c9-4db4-b577-8820f50ccf15\") " pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-fgt8j" Dec 13 07:18:03 crc kubenswrapper[4971]: I1213 07:18:03.175179 4971 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/90351fb3-f8c9-4db4-b577-8820f50ccf15-ssh-key\") pod \"repo-setup-edpm-deployment-openstack-edpm-ipam-fgt8j\" (UID: \"90351fb3-f8c9-4db4-b577-8820f50ccf15\") " pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-fgt8j" Dec 13 07:18:03 crc kubenswrapper[4971]: I1213 07:18:03.277583 4971 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/90351fb3-f8c9-4db4-b577-8820f50ccf15-inventory\") pod \"repo-setup-edpm-deployment-openstack-edpm-ipam-fgt8j\" (UID: \"90351fb3-f8c9-4db4-b577-8820f50ccf15\") " pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-fgt8j" Dec 13 07:18:03 crc kubenswrapper[4971]: I1213 07:18:03.277706 4971 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"repo-setup-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/90351fb3-f8c9-4db4-b577-8820f50ccf15-repo-setup-combined-ca-bundle\") pod \"repo-setup-edpm-deployment-openstack-edpm-ipam-fgt8j\" (UID: \"90351fb3-f8c9-4db4-b577-8820f50ccf15\") " pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-fgt8j" Dec 13 07:18:03 crc kubenswrapper[4971]: I1213 07:18:03.277902 4971 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-z84fz\" (UniqueName: \"kubernetes.io/projected/90351fb3-f8c9-4db4-b577-8820f50ccf15-kube-api-access-z84fz\") pod \"repo-setup-edpm-deployment-openstack-edpm-ipam-fgt8j\" (UID: \"90351fb3-f8c9-4db4-b577-8820f50ccf15\") " pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-fgt8j" Dec 13 07:18:03 crc kubenswrapper[4971]: I1213 07:18:03.278045 4971 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/90351fb3-f8c9-4db4-b577-8820f50ccf15-ssh-key\") pod \"repo-setup-edpm-deployment-openstack-edpm-ipam-fgt8j\" (UID: \"90351fb3-f8c9-4db4-b577-8820f50ccf15\") " pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-fgt8j" Dec 13 07:18:03 crc kubenswrapper[4971]: I1213 07:18:03.284168 4971 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"repo-setup-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/90351fb3-f8c9-4db4-b577-8820f50ccf15-repo-setup-combined-ca-bundle\") pod \"repo-setup-edpm-deployment-openstack-edpm-ipam-fgt8j\" (UID: \"90351fb3-f8c9-4db4-b577-8820f50ccf15\") " pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-fgt8j" Dec 13 07:18:03 crc kubenswrapper[4971]: I1213 07:18:03.286729 4971 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/90351fb3-f8c9-4db4-b577-8820f50ccf15-inventory\") pod \"repo-setup-edpm-deployment-openstack-edpm-ipam-fgt8j\" (UID: \"90351fb3-f8c9-4db4-b577-8820f50ccf15\") " pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-fgt8j" Dec 13 07:18:03 crc kubenswrapper[4971]: I1213 07:18:03.287101 4971 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/90351fb3-f8c9-4db4-b577-8820f50ccf15-ssh-key\") pod \"repo-setup-edpm-deployment-openstack-edpm-ipam-fgt8j\" (UID: \"90351fb3-f8c9-4db4-b577-8820f50ccf15\") " pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-fgt8j" Dec 13 07:18:03 crc kubenswrapper[4971]: I1213 07:18:03.307604 4971 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-z84fz\" (UniqueName: \"kubernetes.io/projected/90351fb3-f8c9-4db4-b577-8820f50ccf15-kube-api-access-z84fz\") pod \"repo-setup-edpm-deployment-openstack-edpm-ipam-fgt8j\" (UID: \"90351fb3-f8c9-4db4-b577-8820f50ccf15\") " pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-fgt8j" Dec 13 07:18:03 crc kubenswrapper[4971]: I1213 07:18:03.399862 4971 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-fgt8j" Dec 13 07:18:03 crc kubenswrapper[4971]: I1213 07:18:03.956784 4971 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-fgt8j"] Dec 13 07:18:04 crc kubenswrapper[4971]: I1213 07:18:04.652965 4971 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-fgt8j" event={"ID":"90351fb3-f8c9-4db4-b577-8820f50ccf15","Type":"ContainerStarted","Data":"c0c03eb1bae10ab0acaf5f9f83c2abfa0e7fbe8507e2671682675bd60fba9418"} Dec 13 07:18:04 crc kubenswrapper[4971]: I1213 07:18:04.768716 4971 scope.go:117] "RemoveContainer" containerID="e8beee43d15719dccb6892d6f5711674492425300efa098be89033bc65977d14" Dec 13 07:18:04 crc kubenswrapper[4971]: E1213 07:18:04.769245 4971 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-82xjz_openshift-machine-config-operator(e0c4af12-edda-41d7-9b44-f87396174ef2)\"" pod="openshift-machine-config-operator/machine-config-daemon-82xjz" podUID="e0c4af12-edda-41d7-9b44-f87396174ef2" Dec 13 07:18:04 crc kubenswrapper[4971]: I1213 07:18:04.833324 4971 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-aee-default-env" Dec 13 07:18:05 crc kubenswrapper[4971]: I1213 07:18:05.663405 4971 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-fgt8j" event={"ID":"90351fb3-f8c9-4db4-b577-8820f50ccf15","Type":"ContainerStarted","Data":"54824cd9cbea8808582e7a1609afdfdf04d051d1509664c0e14aa36943354de6"} Dec 13 07:18:05 crc kubenswrapper[4971]: I1213 07:18:05.682843 4971 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-fgt8j" podStartSLOduration=1.81892886 podStartE2EDuration="2.682826371s" podCreationTimestamp="2025-12-13 07:18:03 +0000 UTC" firstStartedPulling="2025-12-13 07:18:03.965430928 +0000 UTC m=+1740.569840376" lastFinishedPulling="2025-12-13 07:18:04.829328419 +0000 UTC m=+1741.433737887" observedRunningTime="2025-12-13 07:18:05.682085083 +0000 UTC m=+1742.286494571" watchObservedRunningTime="2025-12-13 07:18:05.682826371 +0000 UTC m=+1742.287235819" Dec 13 07:18:09 crc kubenswrapper[4971]: I1213 07:18:09.702856 4971 generic.go:334] "Generic (PLEG): container finished" podID="90351fb3-f8c9-4db4-b577-8820f50ccf15" containerID="54824cd9cbea8808582e7a1609afdfdf04d051d1509664c0e14aa36943354de6" exitCode=2 Dec 13 07:18:09 crc kubenswrapper[4971]: I1213 07:18:09.702928 4971 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-fgt8j" event={"ID":"90351fb3-f8c9-4db4-b577-8820f50ccf15","Type":"ContainerDied","Data":"54824cd9cbea8808582e7a1609afdfdf04d051d1509664c0e14aa36943354de6"} Dec 13 07:18:11 crc kubenswrapper[4971]: I1213 07:18:11.094380 4971 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-fgt8j" Dec 13 07:18:11 crc kubenswrapper[4971]: I1213 07:18:11.190303 4971 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/90351fb3-f8c9-4db4-b577-8820f50ccf15-inventory\") pod \"90351fb3-f8c9-4db4-b577-8820f50ccf15\" (UID: \"90351fb3-f8c9-4db4-b577-8820f50ccf15\") " Dec 13 07:18:11 crc kubenswrapper[4971]: I1213 07:18:11.190357 4971 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/90351fb3-f8c9-4db4-b577-8820f50ccf15-ssh-key\") pod \"90351fb3-f8c9-4db4-b577-8820f50ccf15\" (UID: \"90351fb3-f8c9-4db4-b577-8820f50ccf15\") " Dec 13 07:18:11 crc kubenswrapper[4971]: I1213 07:18:11.190438 4971 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-z84fz\" (UniqueName: \"kubernetes.io/projected/90351fb3-f8c9-4db4-b577-8820f50ccf15-kube-api-access-z84fz\") pod \"90351fb3-f8c9-4db4-b577-8820f50ccf15\" (UID: \"90351fb3-f8c9-4db4-b577-8820f50ccf15\") " Dec 13 07:18:11 crc kubenswrapper[4971]: I1213 07:18:11.190464 4971 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"repo-setup-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/90351fb3-f8c9-4db4-b577-8820f50ccf15-repo-setup-combined-ca-bundle\") pod \"90351fb3-f8c9-4db4-b577-8820f50ccf15\" (UID: \"90351fb3-f8c9-4db4-b577-8820f50ccf15\") " Dec 13 07:18:11 crc kubenswrapper[4971]: I1213 07:18:11.195724 4971 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/90351fb3-f8c9-4db4-b577-8820f50ccf15-repo-setup-combined-ca-bundle" (OuterVolumeSpecName: "repo-setup-combined-ca-bundle") pod "90351fb3-f8c9-4db4-b577-8820f50ccf15" (UID: "90351fb3-f8c9-4db4-b577-8820f50ccf15"). InnerVolumeSpecName "repo-setup-combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 13 07:18:11 crc kubenswrapper[4971]: I1213 07:18:11.195826 4971 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/90351fb3-f8c9-4db4-b577-8820f50ccf15-kube-api-access-z84fz" (OuterVolumeSpecName: "kube-api-access-z84fz") pod "90351fb3-f8c9-4db4-b577-8820f50ccf15" (UID: "90351fb3-f8c9-4db4-b577-8820f50ccf15"). InnerVolumeSpecName "kube-api-access-z84fz". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 13 07:18:11 crc kubenswrapper[4971]: I1213 07:18:11.223747 4971 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/90351fb3-f8c9-4db4-b577-8820f50ccf15-inventory" (OuterVolumeSpecName: "inventory") pod "90351fb3-f8c9-4db4-b577-8820f50ccf15" (UID: "90351fb3-f8c9-4db4-b577-8820f50ccf15"). InnerVolumeSpecName "inventory". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 13 07:18:11 crc kubenswrapper[4971]: I1213 07:18:11.247057 4971 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/90351fb3-f8c9-4db4-b577-8820f50ccf15-ssh-key" (OuterVolumeSpecName: "ssh-key") pod "90351fb3-f8c9-4db4-b577-8820f50ccf15" (UID: "90351fb3-f8c9-4db4-b577-8820f50ccf15"). InnerVolumeSpecName "ssh-key". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 13 07:18:11 crc kubenswrapper[4971]: I1213 07:18:11.291973 4971 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-z84fz\" (UniqueName: \"kubernetes.io/projected/90351fb3-f8c9-4db4-b577-8820f50ccf15-kube-api-access-z84fz\") on node \"crc\" DevicePath \"\"" Dec 13 07:18:11 crc kubenswrapper[4971]: I1213 07:18:11.292010 4971 reconciler_common.go:293] "Volume detached for volume \"repo-setup-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/90351fb3-f8c9-4db4-b577-8820f50ccf15-repo-setup-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 13 07:18:11 crc kubenswrapper[4971]: I1213 07:18:11.292031 4971 reconciler_common.go:293] "Volume detached for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/90351fb3-f8c9-4db4-b577-8820f50ccf15-inventory\") on node \"crc\" DevicePath \"\"" Dec 13 07:18:11 crc kubenswrapper[4971]: I1213 07:18:11.292040 4971 reconciler_common.go:293] "Volume detached for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/90351fb3-f8c9-4db4-b577-8820f50ccf15-ssh-key\") on node \"crc\" DevicePath \"\"" Dec 13 07:18:11 crc kubenswrapper[4971]: I1213 07:18:11.727676 4971 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-fgt8j" event={"ID":"90351fb3-f8c9-4db4-b577-8820f50ccf15","Type":"ContainerDied","Data":"c0c03eb1bae10ab0acaf5f9f83c2abfa0e7fbe8507e2671682675bd60fba9418"} Dec 13 07:18:11 crc kubenswrapper[4971]: I1213 07:18:11.727825 4971 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="c0c03eb1bae10ab0acaf5f9f83c2abfa0e7fbe8507e2671682675bd60fba9418" Dec 13 07:18:11 crc kubenswrapper[4971]: I1213 07:18:11.727926 4971 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-fgt8j" Dec 13 07:18:13 crc kubenswrapper[4971]: I1213 07:18:13.134734 4971 scope.go:117] "RemoveContainer" containerID="32818695048dd3362ff1e26ded731fa65a905cf3d8e201439acefbf1750b86da" Dec 13 07:18:13 crc kubenswrapper[4971]: I1213 07:18:13.190707 4971 scope.go:117] "RemoveContainer" containerID="804e6a307f26bd01f1e6f6fd6603421e30380f05416bb5c66264ede91de4fc73" Dec 13 07:18:13 crc kubenswrapper[4971]: I1213 07:18:13.222692 4971 scope.go:117] "RemoveContainer" containerID="ce0d05172c6bf7f2de3ae358f7e695533010db00a93a51ea0dcaacfe0ce9399f" Dec 13 07:18:16 crc kubenswrapper[4971]: I1213 07:18:16.800997 4971 scope.go:117] "RemoveContainer" containerID="e8beee43d15719dccb6892d6f5711674492425300efa098be89033bc65977d14" Dec 13 07:18:16 crc kubenswrapper[4971]: E1213 07:18:16.804219 4971 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-82xjz_openshift-machine-config-operator(e0c4af12-edda-41d7-9b44-f87396174ef2)\"" pod="openshift-machine-config-operator/machine-config-daemon-82xjz" podUID="e0c4af12-edda-41d7-9b44-f87396174ef2" Dec 13 07:18:30 crc kubenswrapper[4971]: I1213 07:18:30.769182 4971 scope.go:117] "RemoveContainer" containerID="e8beee43d15719dccb6892d6f5711674492425300efa098be89033bc65977d14" Dec 13 07:18:30 crc kubenswrapper[4971]: E1213 07:18:30.769981 4971 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-82xjz_openshift-machine-config-operator(e0c4af12-edda-41d7-9b44-f87396174ef2)\"" pod="openshift-machine-config-operator/machine-config-daemon-82xjz" podUID="e0c4af12-edda-41d7-9b44-f87396174ef2" Dec 13 07:18:44 crc kubenswrapper[4971]: I1213 07:18:44.768703 4971 scope.go:117] "RemoveContainer" containerID="e8beee43d15719dccb6892d6f5711674492425300efa098be89033bc65977d14" Dec 13 07:18:44 crc kubenswrapper[4971]: E1213 07:18:44.769677 4971 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-82xjz_openshift-machine-config-operator(e0c4af12-edda-41d7-9b44-f87396174ef2)\"" pod="openshift-machine-config-operator/machine-config-daemon-82xjz" podUID="e0c4af12-edda-41d7-9b44-f87396174ef2" Dec 13 07:18:49 crc kubenswrapper[4971]: I1213 07:18:49.056314 4971 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-8vg77"] Dec 13 07:18:49 crc kubenswrapper[4971]: E1213 07:18:49.057732 4971 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="90351fb3-f8c9-4db4-b577-8820f50ccf15" containerName="repo-setup-edpm-deployment-openstack-edpm-ipam" Dec 13 07:18:49 crc kubenswrapper[4971]: I1213 07:18:49.057762 4971 state_mem.go:107] "Deleted CPUSet assignment" podUID="90351fb3-f8c9-4db4-b577-8820f50ccf15" containerName="repo-setup-edpm-deployment-openstack-edpm-ipam" Dec 13 07:18:49 crc kubenswrapper[4971]: I1213 07:18:49.058182 4971 memory_manager.go:354] "RemoveStaleState removing state" podUID="90351fb3-f8c9-4db4-b577-8820f50ccf15" containerName="repo-setup-edpm-deployment-openstack-edpm-ipam" Dec 13 07:18:49 crc kubenswrapper[4971]: I1213 07:18:49.058905 4971 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-8vg77" Dec 13 07:18:49 crc kubenswrapper[4971]: I1213 07:18:49.063010 4971 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-edpm-ipam-dockercfg-zb2j5" Dec 13 07:18:49 crc kubenswrapper[4971]: I1213 07:18:49.063153 4971 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplanenodeset-openstack-edpm-ipam" Dec 13 07:18:49 crc kubenswrapper[4971]: I1213 07:18:49.063328 4971 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplane-ansible-ssh-private-key-secret" Dec 13 07:18:49 crc kubenswrapper[4971]: I1213 07:18:49.063591 4971 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-aee-default-env" Dec 13 07:18:49 crc kubenswrapper[4971]: I1213 07:18:49.073178 4971 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-8vg77"] Dec 13 07:18:49 crc kubenswrapper[4971]: I1213 07:18:49.159510 4971 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-jwt9k\" (UniqueName: \"kubernetes.io/projected/68fb082e-8ed6-4ecc-88a9-d349440dd1b3-kube-api-access-jwt9k\") pod \"repo-setup-edpm-deployment-openstack-edpm-ipam-8vg77\" (UID: \"68fb082e-8ed6-4ecc-88a9-d349440dd1b3\") " pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-8vg77" Dec 13 07:18:49 crc kubenswrapper[4971]: I1213 07:18:49.159681 4971 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"repo-setup-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/68fb082e-8ed6-4ecc-88a9-d349440dd1b3-repo-setup-combined-ca-bundle\") pod \"repo-setup-edpm-deployment-openstack-edpm-ipam-8vg77\" (UID: \"68fb082e-8ed6-4ecc-88a9-d349440dd1b3\") " pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-8vg77" Dec 13 07:18:49 crc kubenswrapper[4971]: I1213 07:18:49.159894 4971 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/68fb082e-8ed6-4ecc-88a9-d349440dd1b3-inventory\") pod \"repo-setup-edpm-deployment-openstack-edpm-ipam-8vg77\" (UID: \"68fb082e-8ed6-4ecc-88a9-d349440dd1b3\") " pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-8vg77" Dec 13 07:18:49 crc kubenswrapper[4971]: I1213 07:18:49.160008 4971 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/68fb082e-8ed6-4ecc-88a9-d349440dd1b3-ssh-key\") pod \"repo-setup-edpm-deployment-openstack-edpm-ipam-8vg77\" (UID: \"68fb082e-8ed6-4ecc-88a9-d349440dd1b3\") " pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-8vg77" Dec 13 07:18:49 crc kubenswrapper[4971]: I1213 07:18:49.261840 4971 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-jwt9k\" (UniqueName: \"kubernetes.io/projected/68fb082e-8ed6-4ecc-88a9-d349440dd1b3-kube-api-access-jwt9k\") pod \"repo-setup-edpm-deployment-openstack-edpm-ipam-8vg77\" (UID: \"68fb082e-8ed6-4ecc-88a9-d349440dd1b3\") " pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-8vg77" Dec 13 07:18:49 crc kubenswrapper[4971]: I1213 07:18:49.261935 4971 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"repo-setup-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/68fb082e-8ed6-4ecc-88a9-d349440dd1b3-repo-setup-combined-ca-bundle\") pod \"repo-setup-edpm-deployment-openstack-edpm-ipam-8vg77\" (UID: \"68fb082e-8ed6-4ecc-88a9-d349440dd1b3\") " pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-8vg77" Dec 13 07:18:49 crc kubenswrapper[4971]: I1213 07:18:49.262002 4971 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/68fb082e-8ed6-4ecc-88a9-d349440dd1b3-inventory\") pod \"repo-setup-edpm-deployment-openstack-edpm-ipam-8vg77\" (UID: \"68fb082e-8ed6-4ecc-88a9-d349440dd1b3\") " pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-8vg77" Dec 13 07:18:49 crc kubenswrapper[4971]: I1213 07:18:49.262045 4971 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/68fb082e-8ed6-4ecc-88a9-d349440dd1b3-ssh-key\") pod \"repo-setup-edpm-deployment-openstack-edpm-ipam-8vg77\" (UID: \"68fb082e-8ed6-4ecc-88a9-d349440dd1b3\") " pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-8vg77" Dec 13 07:18:49 crc kubenswrapper[4971]: I1213 07:18:49.269415 4971 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/68fb082e-8ed6-4ecc-88a9-d349440dd1b3-inventory\") pod \"repo-setup-edpm-deployment-openstack-edpm-ipam-8vg77\" (UID: \"68fb082e-8ed6-4ecc-88a9-d349440dd1b3\") " pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-8vg77" Dec 13 07:18:49 crc kubenswrapper[4971]: I1213 07:18:49.270221 4971 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/68fb082e-8ed6-4ecc-88a9-d349440dd1b3-ssh-key\") pod \"repo-setup-edpm-deployment-openstack-edpm-ipam-8vg77\" (UID: \"68fb082e-8ed6-4ecc-88a9-d349440dd1b3\") " pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-8vg77" Dec 13 07:18:49 crc kubenswrapper[4971]: I1213 07:18:49.270439 4971 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"repo-setup-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/68fb082e-8ed6-4ecc-88a9-d349440dd1b3-repo-setup-combined-ca-bundle\") pod \"repo-setup-edpm-deployment-openstack-edpm-ipam-8vg77\" (UID: \"68fb082e-8ed6-4ecc-88a9-d349440dd1b3\") " pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-8vg77" Dec 13 07:18:49 crc kubenswrapper[4971]: I1213 07:18:49.283202 4971 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-jwt9k\" (UniqueName: \"kubernetes.io/projected/68fb082e-8ed6-4ecc-88a9-d349440dd1b3-kube-api-access-jwt9k\") pod \"repo-setup-edpm-deployment-openstack-edpm-ipam-8vg77\" (UID: \"68fb082e-8ed6-4ecc-88a9-d349440dd1b3\") " pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-8vg77" Dec 13 07:18:49 crc kubenswrapper[4971]: I1213 07:18:49.402440 4971 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-8vg77" Dec 13 07:18:50 crc kubenswrapper[4971]: I1213 07:18:50.584297 4971 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-8vg77"] Dec 13 07:18:50 crc kubenswrapper[4971]: W1213 07:18:50.593002 4971 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod68fb082e_8ed6_4ecc_88a9_d349440dd1b3.slice/crio-52bbdf2dc31edaa9092046534c87c92d27490ce6d5d57607abf8bb2a5c6fd331 WatchSource:0}: Error finding container 52bbdf2dc31edaa9092046534c87c92d27490ce6d5d57607abf8bb2a5c6fd331: Status 404 returned error can't find the container with id 52bbdf2dc31edaa9092046534c87c92d27490ce6d5d57607abf8bb2a5c6fd331 Dec 13 07:18:50 crc kubenswrapper[4971]: I1213 07:18:50.597188 4971 provider.go:102] Refreshing cache for provider: *credentialprovider.defaultDockerConfigProvider Dec 13 07:18:51 crc kubenswrapper[4971]: I1213 07:18:51.321112 4971 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-8vg77" event={"ID":"68fb082e-8ed6-4ecc-88a9-d349440dd1b3","Type":"ContainerStarted","Data":"64e238fe45a615cf21cd269809f116b1a84e3b58edd6ed2187c07198bfd90fe6"} Dec 13 07:18:51 crc kubenswrapper[4971]: I1213 07:18:51.321637 4971 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-8vg77" event={"ID":"68fb082e-8ed6-4ecc-88a9-d349440dd1b3","Type":"ContainerStarted","Data":"52bbdf2dc31edaa9092046534c87c92d27490ce6d5d57607abf8bb2a5c6fd331"} Dec 13 07:18:51 crc kubenswrapper[4971]: I1213 07:18:51.350130 4971 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-8vg77" podStartSLOduration=1.911301052 podStartE2EDuration="2.350103637s" podCreationTimestamp="2025-12-13 07:18:49 +0000 UTC" firstStartedPulling="2025-12-13 07:18:50.59697462 +0000 UTC m=+1787.201384068" lastFinishedPulling="2025-12-13 07:18:51.035777165 +0000 UTC m=+1787.640186653" observedRunningTime="2025-12-13 07:18:51.337826183 +0000 UTC m=+1787.942235631" watchObservedRunningTime="2025-12-13 07:18:51.350103637 +0000 UTC m=+1787.954513115" Dec 13 07:18:55 crc kubenswrapper[4971]: I1213 07:18:55.359028 4971 generic.go:334] "Generic (PLEG): container finished" podID="68fb082e-8ed6-4ecc-88a9-d349440dd1b3" containerID="64e238fe45a615cf21cd269809f116b1a84e3b58edd6ed2187c07198bfd90fe6" exitCode=2 Dec 13 07:18:55 crc kubenswrapper[4971]: I1213 07:18:55.359149 4971 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-8vg77" event={"ID":"68fb082e-8ed6-4ecc-88a9-d349440dd1b3","Type":"ContainerDied","Data":"64e238fe45a615cf21cd269809f116b1a84e3b58edd6ed2187c07198bfd90fe6"} Dec 13 07:18:56 crc kubenswrapper[4971]: I1213 07:18:56.741075 4971 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-8vg77" Dec 13 07:18:56 crc kubenswrapper[4971]: I1213 07:18:56.767220 4971 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"repo-setup-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/68fb082e-8ed6-4ecc-88a9-d349440dd1b3-repo-setup-combined-ca-bundle\") pod \"68fb082e-8ed6-4ecc-88a9-d349440dd1b3\" (UID: \"68fb082e-8ed6-4ecc-88a9-d349440dd1b3\") " Dec 13 07:18:56 crc kubenswrapper[4971]: I1213 07:18:56.767352 4971 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-jwt9k\" (UniqueName: \"kubernetes.io/projected/68fb082e-8ed6-4ecc-88a9-d349440dd1b3-kube-api-access-jwt9k\") pod \"68fb082e-8ed6-4ecc-88a9-d349440dd1b3\" (UID: \"68fb082e-8ed6-4ecc-88a9-d349440dd1b3\") " Dec 13 07:18:56 crc kubenswrapper[4971]: I1213 07:18:56.767405 4971 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/68fb082e-8ed6-4ecc-88a9-d349440dd1b3-inventory\") pod \"68fb082e-8ed6-4ecc-88a9-d349440dd1b3\" (UID: \"68fb082e-8ed6-4ecc-88a9-d349440dd1b3\") " Dec 13 07:18:56 crc kubenswrapper[4971]: I1213 07:18:56.767554 4971 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/68fb082e-8ed6-4ecc-88a9-d349440dd1b3-ssh-key\") pod \"68fb082e-8ed6-4ecc-88a9-d349440dd1b3\" (UID: \"68fb082e-8ed6-4ecc-88a9-d349440dd1b3\") " Dec 13 07:18:56 crc kubenswrapper[4971]: I1213 07:18:56.767725 4971 scope.go:117] "RemoveContainer" containerID="e8beee43d15719dccb6892d6f5711674492425300efa098be89033bc65977d14" Dec 13 07:18:56 crc kubenswrapper[4971]: E1213 07:18:56.768260 4971 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-82xjz_openshift-machine-config-operator(e0c4af12-edda-41d7-9b44-f87396174ef2)\"" pod="openshift-machine-config-operator/machine-config-daemon-82xjz" podUID="e0c4af12-edda-41d7-9b44-f87396174ef2" Dec 13 07:18:56 crc kubenswrapper[4971]: I1213 07:18:56.782856 4971 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/68fb082e-8ed6-4ecc-88a9-d349440dd1b3-repo-setup-combined-ca-bundle" (OuterVolumeSpecName: "repo-setup-combined-ca-bundle") pod "68fb082e-8ed6-4ecc-88a9-d349440dd1b3" (UID: "68fb082e-8ed6-4ecc-88a9-d349440dd1b3"). InnerVolumeSpecName "repo-setup-combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 13 07:18:56 crc kubenswrapper[4971]: I1213 07:18:56.792014 4971 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/68fb082e-8ed6-4ecc-88a9-d349440dd1b3-kube-api-access-jwt9k" (OuterVolumeSpecName: "kube-api-access-jwt9k") pod "68fb082e-8ed6-4ecc-88a9-d349440dd1b3" (UID: "68fb082e-8ed6-4ecc-88a9-d349440dd1b3"). InnerVolumeSpecName "kube-api-access-jwt9k". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 13 07:18:56 crc kubenswrapper[4971]: I1213 07:18:56.817382 4971 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/68fb082e-8ed6-4ecc-88a9-d349440dd1b3-ssh-key" (OuterVolumeSpecName: "ssh-key") pod "68fb082e-8ed6-4ecc-88a9-d349440dd1b3" (UID: "68fb082e-8ed6-4ecc-88a9-d349440dd1b3"). InnerVolumeSpecName "ssh-key". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 13 07:18:56 crc kubenswrapper[4971]: I1213 07:18:56.820009 4971 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/68fb082e-8ed6-4ecc-88a9-d349440dd1b3-inventory" (OuterVolumeSpecName: "inventory") pod "68fb082e-8ed6-4ecc-88a9-d349440dd1b3" (UID: "68fb082e-8ed6-4ecc-88a9-d349440dd1b3"). InnerVolumeSpecName "inventory". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 13 07:18:56 crc kubenswrapper[4971]: I1213 07:18:56.870441 4971 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-jwt9k\" (UniqueName: \"kubernetes.io/projected/68fb082e-8ed6-4ecc-88a9-d349440dd1b3-kube-api-access-jwt9k\") on node \"crc\" DevicePath \"\"" Dec 13 07:18:56 crc kubenswrapper[4971]: I1213 07:18:56.870471 4971 reconciler_common.go:293] "Volume detached for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/68fb082e-8ed6-4ecc-88a9-d349440dd1b3-inventory\") on node \"crc\" DevicePath \"\"" Dec 13 07:18:56 crc kubenswrapper[4971]: I1213 07:18:56.870481 4971 reconciler_common.go:293] "Volume detached for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/68fb082e-8ed6-4ecc-88a9-d349440dd1b3-ssh-key\") on node \"crc\" DevicePath \"\"" Dec 13 07:18:56 crc kubenswrapper[4971]: I1213 07:18:56.870491 4971 reconciler_common.go:293] "Volume detached for volume \"repo-setup-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/68fb082e-8ed6-4ecc-88a9-d349440dd1b3-repo-setup-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 13 07:18:57 crc kubenswrapper[4971]: I1213 07:18:57.381737 4971 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-8vg77" event={"ID":"68fb082e-8ed6-4ecc-88a9-d349440dd1b3","Type":"ContainerDied","Data":"52bbdf2dc31edaa9092046534c87c92d27490ce6d5d57607abf8bb2a5c6fd331"} Dec 13 07:18:57 crc kubenswrapper[4971]: I1213 07:18:57.381788 4971 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="52bbdf2dc31edaa9092046534c87c92d27490ce6d5d57607abf8bb2a5c6fd331" Dec 13 07:18:57 crc kubenswrapper[4971]: I1213 07:18:57.381875 4971 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-8vg77" Dec 13 07:19:08 crc kubenswrapper[4971]: I1213 07:19:08.769463 4971 scope.go:117] "RemoveContainer" containerID="e8beee43d15719dccb6892d6f5711674492425300efa098be89033bc65977d14" Dec 13 07:19:08 crc kubenswrapper[4971]: E1213 07:19:08.770506 4971 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-82xjz_openshift-machine-config-operator(e0c4af12-edda-41d7-9b44-f87396174ef2)\"" pod="openshift-machine-config-operator/machine-config-daemon-82xjz" podUID="e0c4af12-edda-41d7-9b44-f87396174ef2" Dec 13 07:19:13 crc kubenswrapper[4971]: I1213 07:19:13.321227 4971 scope.go:117] "RemoveContainer" containerID="5f0a1e401c9c7ec2cb4fb7ef1c38076d6e666c4f5c4d211000b63a3fb564c157" Dec 13 07:19:13 crc kubenswrapper[4971]: I1213 07:19:13.363867 4971 scope.go:117] "RemoveContainer" containerID="aebcdf008ab057d1a88376abfdad8b3a2c538ea0af5a431a1b81134323b14d9c" Dec 13 07:19:19 crc kubenswrapper[4971]: I1213 07:19:19.769297 4971 scope.go:117] "RemoveContainer" containerID="e8beee43d15719dccb6892d6f5711674492425300efa098be89033bc65977d14" Dec 13 07:19:19 crc kubenswrapper[4971]: E1213 07:19:19.770060 4971 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-82xjz_openshift-machine-config-operator(e0c4af12-edda-41d7-9b44-f87396174ef2)\"" pod="openshift-machine-config-operator/machine-config-daemon-82xjz" podUID="e0c4af12-edda-41d7-9b44-f87396174ef2" Dec 13 07:19:30 crc kubenswrapper[4971]: I1213 07:19:30.772055 4971 scope.go:117] "RemoveContainer" containerID="e8beee43d15719dccb6892d6f5711674492425300efa098be89033bc65977d14" Dec 13 07:19:30 crc kubenswrapper[4971]: E1213 07:19:30.773116 4971 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-82xjz_openshift-machine-config-operator(e0c4af12-edda-41d7-9b44-f87396174ef2)\"" pod="openshift-machine-config-operator/machine-config-daemon-82xjz" podUID="e0c4af12-edda-41d7-9b44-f87396174ef2" Dec 13 07:19:43 crc kubenswrapper[4971]: I1213 07:19:43.774257 4971 scope.go:117] "RemoveContainer" containerID="e8beee43d15719dccb6892d6f5711674492425300efa098be89033bc65977d14" Dec 13 07:19:43 crc kubenswrapper[4971]: E1213 07:19:43.775799 4971 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-82xjz_openshift-machine-config-operator(e0c4af12-edda-41d7-9b44-f87396174ef2)\"" pod="openshift-machine-config-operator/machine-config-daemon-82xjz" podUID="e0c4af12-edda-41d7-9b44-f87396174ef2" Dec 13 07:19:57 crc kubenswrapper[4971]: I1213 07:19:57.770036 4971 scope.go:117] "RemoveContainer" containerID="e8beee43d15719dccb6892d6f5711674492425300efa098be89033bc65977d14" Dec 13 07:19:58 crc kubenswrapper[4971]: I1213 07:19:58.171147 4971 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-82xjz" event={"ID":"e0c4af12-edda-41d7-9b44-f87396174ef2","Type":"ContainerStarted","Data":"adc3c82aa161f8f1a6007b84b7b30285aa46a34a77bebf9c44580ee9879d3298"} Dec 13 07:20:15 crc kubenswrapper[4971]: I1213 07:20:15.059433 4971 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-8jf6x"] Dec 13 07:20:15 crc kubenswrapper[4971]: E1213 07:20:15.060874 4971 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="68fb082e-8ed6-4ecc-88a9-d349440dd1b3" containerName="repo-setup-edpm-deployment-openstack-edpm-ipam" Dec 13 07:20:15 crc kubenswrapper[4971]: I1213 07:20:15.060901 4971 state_mem.go:107] "Deleted CPUSet assignment" podUID="68fb082e-8ed6-4ecc-88a9-d349440dd1b3" containerName="repo-setup-edpm-deployment-openstack-edpm-ipam" Dec 13 07:20:15 crc kubenswrapper[4971]: I1213 07:20:15.061395 4971 memory_manager.go:354] "RemoveStaleState removing state" podUID="68fb082e-8ed6-4ecc-88a9-d349440dd1b3" containerName="repo-setup-edpm-deployment-openstack-edpm-ipam" Dec 13 07:20:15 crc kubenswrapper[4971]: I1213 07:20:15.062695 4971 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-8jf6x" Dec 13 07:20:15 crc kubenswrapper[4971]: I1213 07:20:15.065665 4971 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplanenodeset-openstack-edpm-ipam" Dec 13 07:20:15 crc kubenswrapper[4971]: I1213 07:20:15.067757 4971 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-edpm-ipam-dockercfg-zb2j5" Dec 13 07:20:15 crc kubenswrapper[4971]: I1213 07:20:15.067865 4971 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplane-ansible-ssh-private-key-secret" Dec 13 07:20:15 crc kubenswrapper[4971]: I1213 07:20:15.068108 4971 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-aee-default-env" Dec 13 07:20:15 crc kubenswrapper[4971]: I1213 07:20:15.079555 4971 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-8jf6x"] Dec 13 07:20:15 crc kubenswrapper[4971]: I1213 07:20:15.139870 4971 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"repo-setup-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/fc0d806f-23ce-4920-8932-920a739c700c-repo-setup-combined-ca-bundle\") pod \"repo-setup-edpm-deployment-openstack-edpm-ipam-8jf6x\" (UID: \"fc0d806f-23ce-4920-8932-920a739c700c\") " pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-8jf6x" Dec 13 07:20:15 crc kubenswrapper[4971]: I1213 07:20:15.140007 4971 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-xsqxk\" (UniqueName: \"kubernetes.io/projected/fc0d806f-23ce-4920-8932-920a739c700c-kube-api-access-xsqxk\") pod \"repo-setup-edpm-deployment-openstack-edpm-ipam-8jf6x\" (UID: \"fc0d806f-23ce-4920-8932-920a739c700c\") " pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-8jf6x" Dec 13 07:20:15 crc kubenswrapper[4971]: I1213 07:20:15.140064 4971 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/fc0d806f-23ce-4920-8932-920a739c700c-inventory\") pod \"repo-setup-edpm-deployment-openstack-edpm-ipam-8jf6x\" (UID: \"fc0d806f-23ce-4920-8932-920a739c700c\") " pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-8jf6x" Dec 13 07:20:15 crc kubenswrapper[4971]: I1213 07:20:15.140302 4971 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/fc0d806f-23ce-4920-8932-920a739c700c-ssh-key\") pod \"repo-setup-edpm-deployment-openstack-edpm-ipam-8jf6x\" (UID: \"fc0d806f-23ce-4920-8932-920a739c700c\") " pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-8jf6x" Dec 13 07:20:15 crc kubenswrapper[4971]: I1213 07:20:15.241544 4971 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-xsqxk\" (UniqueName: \"kubernetes.io/projected/fc0d806f-23ce-4920-8932-920a739c700c-kube-api-access-xsqxk\") pod \"repo-setup-edpm-deployment-openstack-edpm-ipam-8jf6x\" (UID: \"fc0d806f-23ce-4920-8932-920a739c700c\") " pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-8jf6x" Dec 13 07:20:15 crc kubenswrapper[4971]: I1213 07:20:15.241605 4971 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/fc0d806f-23ce-4920-8932-920a739c700c-inventory\") pod \"repo-setup-edpm-deployment-openstack-edpm-ipam-8jf6x\" (UID: \"fc0d806f-23ce-4920-8932-920a739c700c\") " pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-8jf6x" Dec 13 07:20:15 crc kubenswrapper[4971]: I1213 07:20:15.241705 4971 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/fc0d806f-23ce-4920-8932-920a739c700c-ssh-key\") pod \"repo-setup-edpm-deployment-openstack-edpm-ipam-8jf6x\" (UID: \"fc0d806f-23ce-4920-8932-920a739c700c\") " pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-8jf6x" Dec 13 07:20:15 crc kubenswrapper[4971]: I1213 07:20:15.241775 4971 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"repo-setup-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/fc0d806f-23ce-4920-8932-920a739c700c-repo-setup-combined-ca-bundle\") pod \"repo-setup-edpm-deployment-openstack-edpm-ipam-8jf6x\" (UID: \"fc0d806f-23ce-4920-8932-920a739c700c\") " pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-8jf6x" Dec 13 07:20:15 crc kubenswrapper[4971]: I1213 07:20:15.249119 4971 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"repo-setup-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/fc0d806f-23ce-4920-8932-920a739c700c-repo-setup-combined-ca-bundle\") pod \"repo-setup-edpm-deployment-openstack-edpm-ipam-8jf6x\" (UID: \"fc0d806f-23ce-4920-8932-920a739c700c\") " pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-8jf6x" Dec 13 07:20:15 crc kubenswrapper[4971]: I1213 07:20:15.249189 4971 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/fc0d806f-23ce-4920-8932-920a739c700c-inventory\") pod \"repo-setup-edpm-deployment-openstack-edpm-ipam-8jf6x\" (UID: \"fc0d806f-23ce-4920-8932-920a739c700c\") " pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-8jf6x" Dec 13 07:20:15 crc kubenswrapper[4971]: I1213 07:20:15.249500 4971 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/fc0d806f-23ce-4920-8932-920a739c700c-ssh-key\") pod \"repo-setup-edpm-deployment-openstack-edpm-ipam-8jf6x\" (UID: \"fc0d806f-23ce-4920-8932-920a739c700c\") " pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-8jf6x" Dec 13 07:20:15 crc kubenswrapper[4971]: I1213 07:20:15.259153 4971 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-xsqxk\" (UniqueName: \"kubernetes.io/projected/fc0d806f-23ce-4920-8932-920a739c700c-kube-api-access-xsqxk\") pod \"repo-setup-edpm-deployment-openstack-edpm-ipam-8jf6x\" (UID: \"fc0d806f-23ce-4920-8932-920a739c700c\") " pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-8jf6x" Dec 13 07:20:15 crc kubenswrapper[4971]: I1213 07:20:15.401891 4971 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-8jf6x" Dec 13 07:20:16 crc kubenswrapper[4971]: I1213 07:20:16.024160 4971 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-8jf6x"] Dec 13 07:20:16 crc kubenswrapper[4971]: W1213 07:20:16.027119 4971 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podfc0d806f_23ce_4920_8932_920a739c700c.slice/crio-ab1ac3a33126f099fe2811d1edd1d42e403a55b58d4c3206c281a864ea65349d WatchSource:0}: Error finding container ab1ac3a33126f099fe2811d1edd1d42e403a55b58d4c3206c281a864ea65349d: Status 404 returned error can't find the container with id ab1ac3a33126f099fe2811d1edd1d42e403a55b58d4c3206c281a864ea65349d Dec 13 07:20:16 crc kubenswrapper[4971]: I1213 07:20:16.368101 4971 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-8jf6x" event={"ID":"fc0d806f-23ce-4920-8932-920a739c700c","Type":"ContainerStarted","Data":"ab1ac3a33126f099fe2811d1edd1d42e403a55b58d4c3206c281a864ea65349d"} Dec 13 07:20:17 crc kubenswrapper[4971]: I1213 07:20:17.377539 4971 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-8jf6x" event={"ID":"fc0d806f-23ce-4920-8932-920a739c700c","Type":"ContainerStarted","Data":"c6717e7f9fd2a04c93e249f4661a0232169d358b869932582c249bc440616523"} Dec 13 07:20:29 crc kubenswrapper[4971]: I1213 07:20:29.524225 4971 generic.go:334] "Generic (PLEG): container finished" podID="fc0d806f-23ce-4920-8932-920a739c700c" containerID="c6717e7f9fd2a04c93e249f4661a0232169d358b869932582c249bc440616523" exitCode=0 Dec 13 07:20:29 crc kubenswrapper[4971]: I1213 07:20:29.524291 4971 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-8jf6x" event={"ID":"fc0d806f-23ce-4920-8932-920a739c700c","Type":"ContainerDied","Data":"c6717e7f9fd2a04c93e249f4661a0232169d358b869932582c249bc440616523"} Dec 13 07:20:31 crc kubenswrapper[4971]: I1213 07:20:31.004133 4971 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-8jf6x" Dec 13 07:20:31 crc kubenswrapper[4971]: I1213 07:20:31.101844 4971 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/fc0d806f-23ce-4920-8932-920a739c700c-ssh-key\") pod \"fc0d806f-23ce-4920-8932-920a739c700c\" (UID: \"fc0d806f-23ce-4920-8932-920a739c700c\") " Dec 13 07:20:31 crc kubenswrapper[4971]: I1213 07:20:31.101927 4971 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"repo-setup-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/fc0d806f-23ce-4920-8932-920a739c700c-repo-setup-combined-ca-bundle\") pod \"fc0d806f-23ce-4920-8932-920a739c700c\" (UID: \"fc0d806f-23ce-4920-8932-920a739c700c\") " Dec 13 07:20:31 crc kubenswrapper[4971]: I1213 07:20:31.102089 4971 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-xsqxk\" (UniqueName: \"kubernetes.io/projected/fc0d806f-23ce-4920-8932-920a739c700c-kube-api-access-xsqxk\") pod \"fc0d806f-23ce-4920-8932-920a739c700c\" (UID: \"fc0d806f-23ce-4920-8932-920a739c700c\") " Dec 13 07:20:31 crc kubenswrapper[4971]: I1213 07:20:31.102157 4971 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/fc0d806f-23ce-4920-8932-920a739c700c-inventory\") pod \"fc0d806f-23ce-4920-8932-920a739c700c\" (UID: \"fc0d806f-23ce-4920-8932-920a739c700c\") " Dec 13 07:20:31 crc kubenswrapper[4971]: I1213 07:20:31.108265 4971 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/fc0d806f-23ce-4920-8932-920a739c700c-repo-setup-combined-ca-bundle" (OuterVolumeSpecName: "repo-setup-combined-ca-bundle") pod "fc0d806f-23ce-4920-8932-920a739c700c" (UID: "fc0d806f-23ce-4920-8932-920a739c700c"). InnerVolumeSpecName "repo-setup-combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 13 07:20:31 crc kubenswrapper[4971]: I1213 07:20:31.119097 4971 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/fc0d806f-23ce-4920-8932-920a739c700c-kube-api-access-xsqxk" (OuterVolumeSpecName: "kube-api-access-xsqxk") pod "fc0d806f-23ce-4920-8932-920a739c700c" (UID: "fc0d806f-23ce-4920-8932-920a739c700c"). InnerVolumeSpecName "kube-api-access-xsqxk". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 13 07:20:31 crc kubenswrapper[4971]: I1213 07:20:31.145336 4971 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/fc0d806f-23ce-4920-8932-920a739c700c-ssh-key" (OuterVolumeSpecName: "ssh-key") pod "fc0d806f-23ce-4920-8932-920a739c700c" (UID: "fc0d806f-23ce-4920-8932-920a739c700c"). InnerVolumeSpecName "ssh-key". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 13 07:20:31 crc kubenswrapper[4971]: I1213 07:20:31.153220 4971 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/fc0d806f-23ce-4920-8932-920a739c700c-inventory" (OuterVolumeSpecName: "inventory") pod "fc0d806f-23ce-4920-8932-920a739c700c" (UID: "fc0d806f-23ce-4920-8932-920a739c700c"). InnerVolumeSpecName "inventory". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 13 07:20:31 crc kubenswrapper[4971]: I1213 07:20:31.204808 4971 reconciler_common.go:293] "Volume detached for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/fc0d806f-23ce-4920-8932-920a739c700c-ssh-key\") on node \"crc\" DevicePath \"\"" Dec 13 07:20:31 crc kubenswrapper[4971]: I1213 07:20:31.204861 4971 reconciler_common.go:293] "Volume detached for volume \"repo-setup-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/fc0d806f-23ce-4920-8932-920a739c700c-repo-setup-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 13 07:20:31 crc kubenswrapper[4971]: I1213 07:20:31.204879 4971 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-xsqxk\" (UniqueName: \"kubernetes.io/projected/fc0d806f-23ce-4920-8932-920a739c700c-kube-api-access-xsqxk\") on node \"crc\" DevicePath \"\"" Dec 13 07:20:31 crc kubenswrapper[4971]: I1213 07:20:31.204893 4971 reconciler_common.go:293] "Volume detached for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/fc0d806f-23ce-4920-8932-920a739c700c-inventory\") on node \"crc\" DevicePath \"\"" Dec 13 07:20:31 crc kubenswrapper[4971]: I1213 07:20:31.551084 4971 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-8jf6x" event={"ID":"fc0d806f-23ce-4920-8932-920a739c700c","Type":"ContainerDied","Data":"ab1ac3a33126f099fe2811d1edd1d42e403a55b58d4c3206c281a864ea65349d"} Dec 13 07:20:31 crc kubenswrapper[4971]: I1213 07:20:31.551160 4971 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="ab1ac3a33126f099fe2811d1edd1d42e403a55b58d4c3206c281a864ea65349d" Dec 13 07:20:31 crc kubenswrapper[4971]: I1213 07:20:31.551176 4971 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-8jf6x" Dec 13 07:20:31 crc kubenswrapper[4971]: I1213 07:20:31.677702 4971 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-7nvdr"] Dec 13 07:20:31 crc kubenswrapper[4971]: E1213 07:20:31.678424 4971 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="fc0d806f-23ce-4920-8932-920a739c700c" containerName="repo-setup-edpm-deployment-openstack-edpm-ipam" Dec 13 07:20:31 crc kubenswrapper[4971]: I1213 07:20:31.678455 4971 state_mem.go:107] "Deleted CPUSet assignment" podUID="fc0d806f-23ce-4920-8932-920a739c700c" containerName="repo-setup-edpm-deployment-openstack-edpm-ipam" Dec 13 07:20:31 crc kubenswrapper[4971]: I1213 07:20:31.678806 4971 memory_manager.go:354] "RemoveStaleState removing state" podUID="fc0d806f-23ce-4920-8932-920a739c700c" containerName="repo-setup-edpm-deployment-openstack-edpm-ipam" Dec 13 07:20:31 crc kubenswrapper[4971]: I1213 07:20:31.679852 4971 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-7nvdr" Dec 13 07:20:31 crc kubenswrapper[4971]: I1213 07:20:31.683864 4971 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-aee-default-env" Dec 13 07:20:31 crc kubenswrapper[4971]: I1213 07:20:31.684847 4971 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-edpm-ipam-dockercfg-zb2j5" Dec 13 07:20:31 crc kubenswrapper[4971]: I1213 07:20:31.684851 4971 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplanenodeset-openstack-edpm-ipam" Dec 13 07:20:31 crc kubenswrapper[4971]: I1213 07:20:31.686311 4971 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplane-ansible-ssh-private-key-secret" Dec 13 07:20:31 crc kubenswrapper[4971]: I1213 07:20:31.695162 4971 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-7nvdr"] Dec 13 07:20:31 crc kubenswrapper[4971]: I1213 07:20:31.715639 4971 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"bootstrap-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/80f5a7cb-cc56-4d8e-ba65-0198d682bcc5-bootstrap-combined-ca-bundle\") pod \"bootstrap-edpm-deployment-openstack-edpm-ipam-7nvdr\" (UID: \"80f5a7cb-cc56-4d8e-ba65-0198d682bcc5\") " pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-7nvdr" Dec 13 07:20:31 crc kubenswrapper[4971]: I1213 07:20:31.716029 4971 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/80f5a7cb-cc56-4d8e-ba65-0198d682bcc5-ssh-key\") pod \"bootstrap-edpm-deployment-openstack-edpm-ipam-7nvdr\" (UID: \"80f5a7cb-cc56-4d8e-ba65-0198d682bcc5\") " pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-7nvdr" Dec 13 07:20:31 crc kubenswrapper[4971]: I1213 07:20:31.716312 4971 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/80f5a7cb-cc56-4d8e-ba65-0198d682bcc5-inventory\") pod \"bootstrap-edpm-deployment-openstack-edpm-ipam-7nvdr\" (UID: \"80f5a7cb-cc56-4d8e-ba65-0198d682bcc5\") " pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-7nvdr" Dec 13 07:20:31 crc kubenswrapper[4971]: I1213 07:20:31.716509 4971 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-gntbn\" (UniqueName: \"kubernetes.io/projected/80f5a7cb-cc56-4d8e-ba65-0198d682bcc5-kube-api-access-gntbn\") pod \"bootstrap-edpm-deployment-openstack-edpm-ipam-7nvdr\" (UID: \"80f5a7cb-cc56-4d8e-ba65-0198d682bcc5\") " pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-7nvdr" Dec 13 07:20:31 crc kubenswrapper[4971]: I1213 07:20:31.818343 4971 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"bootstrap-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/80f5a7cb-cc56-4d8e-ba65-0198d682bcc5-bootstrap-combined-ca-bundle\") pod \"bootstrap-edpm-deployment-openstack-edpm-ipam-7nvdr\" (UID: \"80f5a7cb-cc56-4d8e-ba65-0198d682bcc5\") " pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-7nvdr" Dec 13 07:20:31 crc kubenswrapper[4971]: I1213 07:20:31.818981 4971 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/80f5a7cb-cc56-4d8e-ba65-0198d682bcc5-ssh-key\") pod \"bootstrap-edpm-deployment-openstack-edpm-ipam-7nvdr\" (UID: \"80f5a7cb-cc56-4d8e-ba65-0198d682bcc5\") " pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-7nvdr" Dec 13 07:20:31 crc kubenswrapper[4971]: I1213 07:20:31.819324 4971 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/80f5a7cb-cc56-4d8e-ba65-0198d682bcc5-inventory\") pod \"bootstrap-edpm-deployment-openstack-edpm-ipam-7nvdr\" (UID: \"80f5a7cb-cc56-4d8e-ba65-0198d682bcc5\") " pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-7nvdr" Dec 13 07:20:31 crc kubenswrapper[4971]: I1213 07:20:31.819585 4971 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-gntbn\" (UniqueName: \"kubernetes.io/projected/80f5a7cb-cc56-4d8e-ba65-0198d682bcc5-kube-api-access-gntbn\") pod \"bootstrap-edpm-deployment-openstack-edpm-ipam-7nvdr\" (UID: \"80f5a7cb-cc56-4d8e-ba65-0198d682bcc5\") " pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-7nvdr" Dec 13 07:20:31 crc kubenswrapper[4971]: I1213 07:20:31.824384 4971 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/80f5a7cb-cc56-4d8e-ba65-0198d682bcc5-ssh-key\") pod \"bootstrap-edpm-deployment-openstack-edpm-ipam-7nvdr\" (UID: \"80f5a7cb-cc56-4d8e-ba65-0198d682bcc5\") " pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-7nvdr" Dec 13 07:20:31 crc kubenswrapper[4971]: I1213 07:20:31.825784 4971 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"bootstrap-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/80f5a7cb-cc56-4d8e-ba65-0198d682bcc5-bootstrap-combined-ca-bundle\") pod \"bootstrap-edpm-deployment-openstack-edpm-ipam-7nvdr\" (UID: \"80f5a7cb-cc56-4d8e-ba65-0198d682bcc5\") " pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-7nvdr" Dec 13 07:20:31 crc kubenswrapper[4971]: I1213 07:20:31.826068 4971 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/80f5a7cb-cc56-4d8e-ba65-0198d682bcc5-inventory\") pod \"bootstrap-edpm-deployment-openstack-edpm-ipam-7nvdr\" (UID: \"80f5a7cb-cc56-4d8e-ba65-0198d682bcc5\") " pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-7nvdr" Dec 13 07:20:31 crc kubenswrapper[4971]: I1213 07:20:31.850298 4971 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-gntbn\" (UniqueName: \"kubernetes.io/projected/80f5a7cb-cc56-4d8e-ba65-0198d682bcc5-kube-api-access-gntbn\") pod \"bootstrap-edpm-deployment-openstack-edpm-ipam-7nvdr\" (UID: \"80f5a7cb-cc56-4d8e-ba65-0198d682bcc5\") " pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-7nvdr" Dec 13 07:20:32 crc kubenswrapper[4971]: I1213 07:20:32.011745 4971 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-7nvdr" Dec 13 07:20:32 crc kubenswrapper[4971]: I1213 07:20:32.583479 4971 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-7nvdr"] Dec 13 07:20:33 crc kubenswrapper[4971]: I1213 07:20:33.578939 4971 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-7nvdr" event={"ID":"80f5a7cb-cc56-4d8e-ba65-0198d682bcc5","Type":"ContainerStarted","Data":"66b156cf0ae30ae6d8916884aac86cb1ee337155f7ef722bd822b01428f81924"} Dec 13 07:20:33 crc kubenswrapper[4971]: I1213 07:20:33.579430 4971 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-7nvdr" event={"ID":"80f5a7cb-cc56-4d8e-ba65-0198d682bcc5","Type":"ContainerStarted","Data":"6834ce97e2c8534adaa3f8861449d6c40d5d35a4171c46b13ab0b2a625ebc948"} Dec 13 07:20:33 crc kubenswrapper[4971]: I1213 07:20:33.599157 4971 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-7nvdr" podStartSLOduration=1.964688959 podStartE2EDuration="2.599127076s" podCreationTimestamp="2025-12-13 07:20:31 +0000 UTC" firstStartedPulling="2025-12-13 07:20:32.593294764 +0000 UTC m=+1889.197704212" lastFinishedPulling="2025-12-13 07:20:33.227732841 +0000 UTC m=+1889.832142329" observedRunningTime="2025-12-13 07:20:33.597094725 +0000 UTC m=+1890.201504213" watchObservedRunningTime="2025-12-13 07:20:33.599127076 +0000 UTC m=+1890.203536564" Dec 13 07:21:00 crc kubenswrapper[4971]: I1213 07:21:00.059366 4971 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/glance-db-create-9wmnw"] Dec 13 07:21:00 crc kubenswrapper[4971]: I1213 07:21:00.090074 4971 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/glance-db-create-9wmnw"] Dec 13 07:21:01 crc kubenswrapper[4971]: I1213 07:21:01.783813 4971 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="ba68c497-330b-4e22-8ecd-d84ee54815d0" path="/var/lib/kubelet/pods/ba68c497-330b-4e22-8ecd-d84ee54815d0/volumes" Dec 13 07:21:02 crc kubenswrapper[4971]: I1213 07:21:02.032433 4971 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/glance-a8e5-account-create-update-7xfzr"] Dec 13 07:21:02 crc kubenswrapper[4971]: I1213 07:21:02.042132 4971 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/glance-a8e5-account-create-update-7xfzr"] Dec 13 07:21:03 crc kubenswrapper[4971]: I1213 07:21:03.780610 4971 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="00df1bf7-20e4-485c-b1c2-6f00df8379ff" path="/var/lib/kubelet/pods/00df1bf7-20e4-485c-b1c2-6f00df8379ff/volumes" Dec 13 07:21:09 crc kubenswrapper[4971]: I1213 07:21:09.114715 4971 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/placement-e878-account-create-update-lnhvj"] Dec 13 07:21:09 crc kubenswrapper[4971]: I1213 07:21:09.125770 4971 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/keystone-2ba7-account-create-update-v5s9z"] Dec 13 07:21:09 crc kubenswrapper[4971]: I1213 07:21:09.137317 4971 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/keystone-db-create-gzpfj"] Dec 13 07:21:09 crc kubenswrapper[4971]: I1213 07:21:09.145790 4971 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/placement-db-create-jwvzj"] Dec 13 07:21:09 crc kubenswrapper[4971]: I1213 07:21:09.153946 4971 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/placement-e878-account-create-update-lnhvj"] Dec 13 07:21:09 crc kubenswrapper[4971]: I1213 07:21:09.162552 4971 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/keystone-2ba7-account-create-update-v5s9z"] Dec 13 07:21:09 crc kubenswrapper[4971]: I1213 07:21:09.170023 4971 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/keystone-db-create-gzpfj"] Dec 13 07:21:09 crc kubenswrapper[4971]: I1213 07:21:09.177157 4971 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/placement-db-create-jwvzj"] Dec 13 07:21:09 crc kubenswrapper[4971]: I1213 07:21:09.797726 4971 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="292d6752-a759-4b19-9759-f0101a1b01e6" path="/var/lib/kubelet/pods/292d6752-a759-4b19-9759-f0101a1b01e6/volumes" Dec 13 07:21:09 crc kubenswrapper[4971]: I1213 07:21:09.799958 4971 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="b2645870-215e-4acb-9f8c-cfa2bb0a5cf6" path="/var/lib/kubelet/pods/b2645870-215e-4acb-9f8c-cfa2bb0a5cf6/volumes" Dec 13 07:21:09 crc kubenswrapper[4971]: I1213 07:21:09.801441 4971 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="d0ed4d40-1309-4641-b91c-f7dc117d6f4a" path="/var/lib/kubelet/pods/d0ed4d40-1309-4641-b91c-f7dc117d6f4a/volumes" Dec 13 07:21:09 crc kubenswrapper[4971]: I1213 07:21:09.803552 4971 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="e07e072c-e921-4f2d-a967-1966c35ff49b" path="/var/lib/kubelet/pods/e07e072c-e921-4f2d-a967-1966c35ff49b/volumes" Dec 13 07:21:13 crc kubenswrapper[4971]: I1213 07:21:13.475140 4971 scope.go:117] "RemoveContainer" containerID="ae179455b245343d250e77a8751d8b15e9f429baa9c54c6b6b56737c04ca7cdf" Dec 13 07:21:13 crc kubenswrapper[4971]: I1213 07:21:13.510157 4971 scope.go:117] "RemoveContainer" containerID="693a818b1ae2abfe18c3d6a10996082200f8677d5087a58ec65f4b2d7dac24ed" Dec 13 07:21:13 crc kubenswrapper[4971]: I1213 07:21:13.541981 4971 scope.go:117] "RemoveContainer" containerID="5901f03e4f1457dd53da63ac9fcae002c832ee5d16b994772d7e84ecbd94c434" Dec 13 07:21:13 crc kubenswrapper[4971]: I1213 07:21:13.606013 4971 scope.go:117] "RemoveContainer" containerID="f48369c49c58c454eabe532f1f904b7c0c3a079ce470fb0caf73489fc5a83168" Dec 13 07:21:13 crc kubenswrapper[4971]: I1213 07:21:13.648206 4971 scope.go:117] "RemoveContainer" containerID="3c10d92d48ac251e515ced70913bfe1054c399803e7e124888cbabe71d4abe76" Dec 13 07:21:13 crc kubenswrapper[4971]: I1213 07:21:13.687373 4971 scope.go:117] "RemoveContainer" containerID="9954aa1f2e47bd1129572705d83bf81bf56d842edd10741b7415b648d7ac42b9" Dec 13 07:21:13 crc kubenswrapper[4971]: I1213 07:21:13.736136 4971 scope.go:117] "RemoveContainer" containerID="7d89c39b751696f0853323248c94bdc2f2bd46d2ecf99447674a6e1e32eff850" Dec 13 07:21:13 crc kubenswrapper[4971]: I1213 07:21:13.774803 4971 scope.go:117] "RemoveContainer" containerID="71c48a05c5bd070dcf1c7058c0af96d1e9e2a917513a2d26b09d1778b89a2045" Dec 13 07:21:13 crc kubenswrapper[4971]: I1213 07:21:13.820996 4971 scope.go:117] "RemoveContainer" containerID="0478cc6d9e47009edd4f98ea96f3a639adea516ff1dd78509c7bfda6bffe3874" Dec 13 07:21:13 crc kubenswrapper[4971]: I1213 07:21:13.865754 4971 scope.go:117] "RemoveContainer" containerID="a26e3edab88870893204ebdf9829573548f908e202436604df866f2e829e670c" Dec 13 07:21:44 crc kubenswrapper[4971]: I1213 07:21:44.044422 4971 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/glance-db-sync-s24b9"] Dec 13 07:21:44 crc kubenswrapper[4971]: I1213 07:21:44.054320 4971 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/glance-db-sync-s24b9"] Dec 13 07:21:45 crc kubenswrapper[4971]: I1213 07:21:45.789739 4971 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="6d483ba4-aee0-4c3b-a416-f3ee7ba19feb" path="/var/lib/kubelet/pods/6d483ba4-aee0-4c3b-a416-f3ee7ba19feb/volumes" Dec 13 07:21:50 crc kubenswrapper[4971]: I1213 07:21:50.039764 4971 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/cinder-db-create-8sbpv"] Dec 13 07:21:50 crc kubenswrapper[4971]: I1213 07:21:50.053938 4971 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/barbican-5251-account-create-update-dr8tv"] Dec 13 07:21:50 crc kubenswrapper[4971]: I1213 07:21:50.082619 4971 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/barbican-5251-account-create-update-dr8tv"] Dec 13 07:21:50 crc kubenswrapper[4971]: I1213 07:21:50.096955 4971 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/cinder-db-create-8sbpv"] Dec 13 07:21:51 crc kubenswrapper[4971]: I1213 07:21:51.026775 4971 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/barbican-db-create-5czp4"] Dec 13 07:21:51 crc kubenswrapper[4971]: I1213 07:21:51.041705 4971 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/cinder-51d6-account-create-update-hm8kr"] Dec 13 07:21:51 crc kubenswrapper[4971]: I1213 07:21:51.047457 4971 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/neutron-db-create-s42n4"] Dec 13 07:21:51 crc kubenswrapper[4971]: I1213 07:21:51.053990 4971 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/barbican-db-create-5czp4"] Dec 13 07:21:51 crc kubenswrapper[4971]: I1213 07:21:51.060534 4971 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/neutron-db-create-s42n4"] Dec 13 07:21:51 crc kubenswrapper[4971]: I1213 07:21:51.067336 4971 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/cinder-51d6-account-create-update-hm8kr"] Dec 13 07:21:51 crc kubenswrapper[4971]: I1213 07:21:51.783919 4971 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="20b5092f-e5a2-467f-8dbc-e083156b0e20" path="/var/lib/kubelet/pods/20b5092f-e5a2-467f-8dbc-e083156b0e20/volumes" Dec 13 07:21:51 crc kubenswrapper[4971]: I1213 07:21:51.785334 4971 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="48a5104a-2e2e-4828-bb54-bf40553d7ee0" path="/var/lib/kubelet/pods/48a5104a-2e2e-4828-bb54-bf40553d7ee0/volumes" Dec 13 07:21:51 crc kubenswrapper[4971]: I1213 07:21:51.786381 4971 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="4c7ef6d3-1b68-4fb6-9784-0ee0bb154a8d" path="/var/lib/kubelet/pods/4c7ef6d3-1b68-4fb6-9784-0ee0bb154a8d/volumes" Dec 13 07:21:51 crc kubenswrapper[4971]: I1213 07:21:51.787338 4971 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="ad253e3a-226e-464a-9772-46f4b0081314" path="/var/lib/kubelet/pods/ad253e3a-226e-464a-9772-46f4b0081314/volumes" Dec 13 07:21:51 crc kubenswrapper[4971]: I1213 07:21:51.789398 4971 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="f136ffc5-3a49-4807-8025-10356e78ee46" path="/var/lib/kubelet/pods/f136ffc5-3a49-4807-8025-10356e78ee46/volumes" Dec 13 07:21:52 crc kubenswrapper[4971]: I1213 07:21:52.041224 4971 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/neutron-2db2-account-create-update-t6l5w"] Dec 13 07:21:52 crc kubenswrapper[4971]: I1213 07:21:52.056947 4971 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/neutron-2db2-account-create-update-t6l5w"] Dec 13 07:21:53 crc kubenswrapper[4971]: I1213 07:21:53.784587 4971 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="349ffc46-502d-4c10-9d0c-69bcfdb46fe7" path="/var/lib/kubelet/pods/349ffc46-502d-4c10-9d0c-69bcfdb46fe7/volumes" Dec 13 07:21:57 crc kubenswrapper[4971]: E1213 07:21:57.918081 4971 cadvisor_stats_provider.go:516] "Partial failure issuing cadvisor.ContainerInfoV2" err="partial failures: [\"/system.slice/rpm-ostreed.service\": RecentStats: unable to find data in memory cache]" Dec 13 07:22:03 crc kubenswrapper[4971]: I1213 07:22:03.040439 4971 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/keystone-db-sync-b2cmm"] Dec 13 07:22:03 crc kubenswrapper[4971]: I1213 07:22:03.053244 4971 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/keystone-db-sync-b2cmm"] Dec 13 07:22:03 crc kubenswrapper[4971]: I1213 07:22:03.783481 4971 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="d6b11c79-8c13-42d7-8dce-c434dfab1912" path="/var/lib/kubelet/pods/d6b11c79-8c13-42d7-8dce-c434dfab1912/volumes" Dec 13 07:22:14 crc kubenswrapper[4971]: I1213 07:22:14.091668 4971 scope.go:117] "RemoveContainer" containerID="3869966112be8a7f6f89d5ea0ceb4a3a1dea2af5d13e82953998769f10e53115" Dec 13 07:22:14 crc kubenswrapper[4971]: I1213 07:22:14.129797 4971 scope.go:117] "RemoveContainer" containerID="3c9811af77b10cff3571a757300c8a4590fe34ec9a412c81599a4241a91ba697" Dec 13 07:22:14 crc kubenswrapper[4971]: I1213 07:22:14.159159 4971 scope.go:117] "RemoveContainer" containerID="74efeeab58249e168c5ec8b783a7a55a261ee15a5da946cae38e1d8ca17d36bc" Dec 13 07:22:14 crc kubenswrapper[4971]: I1213 07:22:14.220080 4971 scope.go:117] "RemoveContainer" containerID="092c2442d88a2d6edbfffc8f033cfd7a1abeb2f3a58b906d2f18d96ae471ca6b" Dec 13 07:22:14 crc kubenswrapper[4971]: I1213 07:22:14.246948 4971 scope.go:117] "RemoveContainer" containerID="9850b81303e2106176d4957d891cc8a2a2779060307683a31e2a544ab57a390b" Dec 13 07:22:14 crc kubenswrapper[4971]: I1213 07:22:14.284147 4971 scope.go:117] "RemoveContainer" containerID="e620d182d52f7c29766f7d0a12dbd3f37b8c5f4a578281d86ab99036ece89ae9" Dec 13 07:22:14 crc kubenswrapper[4971]: I1213 07:22:14.319700 4971 scope.go:117] "RemoveContainer" containerID="3f7c5ba1c6814836f2d82610a709511d47a2e71f2dcf2e913651c6519d7acf4a" Dec 13 07:22:14 crc kubenswrapper[4971]: I1213 07:22:14.356788 4971 scope.go:117] "RemoveContainer" containerID="360bfe834cf511f784e167d260a057730d73c2f25a0284f0622fe6be6d7b48d2" Dec 13 07:22:14 crc kubenswrapper[4971]: I1213 07:22:14.400372 4971 scope.go:117] "RemoveContainer" containerID="465494d76081e5489304d7756f8b20158debd5991405b16dba4332284efc2ab6" Dec 13 07:22:14 crc kubenswrapper[4971]: I1213 07:22:14.424056 4971 scope.go:117] "RemoveContainer" containerID="b783904e1f84f1d7bad0d373f6e98e2e2ac2c1af86ccbccb0a67b46357c25dca" Dec 13 07:22:16 crc kubenswrapper[4971]: I1213 07:22:16.153753 4971 patch_prober.go:28] interesting pod/machine-config-daemon-82xjz container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 13 07:22:16 crc kubenswrapper[4971]: I1213 07:22:16.154084 4971 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-82xjz" podUID="e0c4af12-edda-41d7-9b44-f87396174ef2" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 13 07:22:38 crc kubenswrapper[4971]: I1213 07:22:38.088380 4971 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/neutron-db-sync-8z2nm"] Dec 13 07:22:38 crc kubenswrapper[4971]: I1213 07:22:38.097017 4971 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/neutron-db-sync-8z2nm"] Dec 13 07:22:39 crc kubenswrapper[4971]: I1213 07:22:39.780369 4971 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="0813350d-430f-4564-a09a-062a3239ba2f" path="/var/lib/kubelet/pods/0813350d-430f-4564-a09a-062a3239ba2f/volumes" Dec 13 07:22:46 crc kubenswrapper[4971]: I1213 07:22:46.041472 4971 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/placement-db-sync-954vn"] Dec 13 07:22:46 crc kubenswrapper[4971]: I1213 07:22:46.051349 4971 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/placement-db-sync-954vn"] Dec 13 07:22:46 crc kubenswrapper[4971]: I1213 07:22:46.153925 4971 patch_prober.go:28] interesting pod/machine-config-daemon-82xjz container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 13 07:22:46 crc kubenswrapper[4971]: I1213 07:22:46.153993 4971 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-82xjz" podUID="e0c4af12-edda-41d7-9b44-f87396174ef2" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 13 07:22:47 crc kubenswrapper[4971]: I1213 07:22:47.777883 4971 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="166612b7-15b4-4a04-b7bb-62603363de80" path="/var/lib/kubelet/pods/166612b7-15b4-4a04-b7bb-62603363de80/volumes" Dec 13 07:22:48 crc kubenswrapper[4971]: I1213 07:22:48.033304 4971 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/keystone-bootstrap-dn9h9"] Dec 13 07:22:48 crc kubenswrapper[4971]: I1213 07:22:48.042719 4971 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/keystone-bootstrap-dn9h9"] Dec 13 07:22:49 crc kubenswrapper[4971]: I1213 07:22:49.781271 4971 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="ee3afc2f-79c6-444c-b6a3-2d4c475c3aac" path="/var/lib/kubelet/pods/ee3afc2f-79c6-444c-b6a3-2d4c475c3aac/volumes" Dec 13 07:23:04 crc kubenswrapper[4971]: I1213 07:23:04.041813 4971 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/barbican-db-sync-dvchm"] Dec 13 07:23:04 crc kubenswrapper[4971]: I1213 07:23:04.048920 4971 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/cinder-db-sync-26k47"] Dec 13 07:23:04 crc kubenswrapper[4971]: I1213 07:23:04.057398 4971 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/barbican-db-sync-dvchm"] Dec 13 07:23:04 crc kubenswrapper[4971]: I1213 07:23:04.064789 4971 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/cinder-db-sync-26k47"] Dec 13 07:23:05 crc kubenswrapper[4971]: I1213 07:23:05.791368 4971 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="c1d819d0-339f-4e7e-9180-63d2f7c0042e" path="/var/lib/kubelet/pods/c1d819d0-339f-4e7e-9180-63d2f7c0042e/volumes" Dec 13 07:23:05 crc kubenswrapper[4971]: I1213 07:23:05.792508 4971 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="d7cf1b55-0dae-4a17-857a-6336fe71ff24" path="/var/lib/kubelet/pods/d7cf1b55-0dae-4a17-857a-6336fe71ff24/volumes" Dec 13 07:23:06 crc kubenswrapper[4971]: I1213 07:23:06.925642 4971 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/community-operators-mjjv2"] Dec 13 07:23:06 crc kubenswrapper[4971]: I1213 07:23:06.929114 4971 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-mjjv2" Dec 13 07:23:06 crc kubenswrapper[4971]: I1213 07:23:06.935524 4971 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-mjjv2"] Dec 13 07:23:07 crc kubenswrapper[4971]: I1213 07:23:07.128334 4971 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/71083227-c9dc-420e-b242-2adf03b636c5-catalog-content\") pod \"community-operators-mjjv2\" (UID: \"71083227-c9dc-420e-b242-2adf03b636c5\") " pod="openshift-marketplace/community-operators-mjjv2" Dec 13 07:23:07 crc kubenswrapper[4971]: I1213 07:23:07.128401 4971 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-bh64p\" (UniqueName: \"kubernetes.io/projected/71083227-c9dc-420e-b242-2adf03b636c5-kube-api-access-bh64p\") pod \"community-operators-mjjv2\" (UID: \"71083227-c9dc-420e-b242-2adf03b636c5\") " pod="openshift-marketplace/community-operators-mjjv2" Dec 13 07:23:07 crc kubenswrapper[4971]: I1213 07:23:07.128693 4971 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/71083227-c9dc-420e-b242-2adf03b636c5-utilities\") pod \"community-operators-mjjv2\" (UID: \"71083227-c9dc-420e-b242-2adf03b636c5\") " pod="openshift-marketplace/community-operators-mjjv2" Dec 13 07:23:07 crc kubenswrapper[4971]: I1213 07:23:07.234264 4971 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/71083227-c9dc-420e-b242-2adf03b636c5-catalog-content\") pod \"community-operators-mjjv2\" (UID: \"71083227-c9dc-420e-b242-2adf03b636c5\") " pod="openshift-marketplace/community-operators-mjjv2" Dec 13 07:23:07 crc kubenswrapper[4971]: I1213 07:23:07.234414 4971 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-bh64p\" (UniqueName: \"kubernetes.io/projected/71083227-c9dc-420e-b242-2adf03b636c5-kube-api-access-bh64p\") pod \"community-operators-mjjv2\" (UID: \"71083227-c9dc-420e-b242-2adf03b636c5\") " pod="openshift-marketplace/community-operators-mjjv2" Dec 13 07:23:07 crc kubenswrapper[4971]: I1213 07:23:07.234556 4971 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/71083227-c9dc-420e-b242-2adf03b636c5-utilities\") pod \"community-operators-mjjv2\" (UID: \"71083227-c9dc-420e-b242-2adf03b636c5\") " pod="openshift-marketplace/community-operators-mjjv2" Dec 13 07:23:07 crc kubenswrapper[4971]: I1213 07:23:07.235645 4971 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/71083227-c9dc-420e-b242-2adf03b636c5-utilities\") pod \"community-operators-mjjv2\" (UID: \"71083227-c9dc-420e-b242-2adf03b636c5\") " pod="openshift-marketplace/community-operators-mjjv2" Dec 13 07:23:07 crc kubenswrapper[4971]: I1213 07:23:07.268811 4971 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/71083227-c9dc-420e-b242-2adf03b636c5-catalog-content\") pod \"community-operators-mjjv2\" (UID: \"71083227-c9dc-420e-b242-2adf03b636c5\") " pod="openshift-marketplace/community-operators-mjjv2" Dec 13 07:23:07 crc kubenswrapper[4971]: I1213 07:23:07.298431 4971 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-bh64p\" (UniqueName: \"kubernetes.io/projected/71083227-c9dc-420e-b242-2adf03b636c5-kube-api-access-bh64p\") pod \"community-operators-mjjv2\" (UID: \"71083227-c9dc-420e-b242-2adf03b636c5\") " pod="openshift-marketplace/community-operators-mjjv2" Dec 13 07:23:07 crc kubenswrapper[4971]: I1213 07:23:07.568303 4971 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-mjjv2" Dec 13 07:23:08 crc kubenswrapper[4971]: I1213 07:23:08.088432 4971 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-mjjv2"] Dec 13 07:23:10 crc kubenswrapper[4971]: I1213 07:23:10.143058 4971 generic.go:334] "Generic (PLEG): container finished" podID="71083227-c9dc-420e-b242-2adf03b636c5" containerID="89ea9768c86688de6fc20f29eee7c384b2e190ff5862cfeae9266a11011546e0" exitCode=0 Dec 13 07:23:10 crc kubenswrapper[4971]: I1213 07:23:10.143141 4971 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-mjjv2" event={"ID":"71083227-c9dc-420e-b242-2adf03b636c5","Type":"ContainerDied","Data":"89ea9768c86688de6fc20f29eee7c384b2e190ff5862cfeae9266a11011546e0"} Dec 13 07:23:10 crc kubenswrapper[4971]: I1213 07:23:10.143561 4971 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-mjjv2" event={"ID":"71083227-c9dc-420e-b242-2adf03b636c5","Type":"ContainerStarted","Data":"654d05fe120905b43a46eda225c765af18f4a1c8af3f3b41dda916ae247dd7af"} Dec 13 07:23:12 crc kubenswrapper[4971]: I1213 07:23:12.166778 4971 generic.go:334] "Generic (PLEG): container finished" podID="71083227-c9dc-420e-b242-2adf03b636c5" containerID="eacfac38c2440049c59b71905e0f0ab0c2bbd2cbea97333cf36ee4c35c878610" exitCode=0 Dec 13 07:23:12 crc kubenswrapper[4971]: I1213 07:23:12.166834 4971 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-mjjv2" event={"ID":"71083227-c9dc-420e-b242-2adf03b636c5","Type":"ContainerDied","Data":"eacfac38c2440049c59b71905e0f0ab0c2bbd2cbea97333cf36ee4c35c878610"} Dec 13 07:23:13 crc kubenswrapper[4971]: I1213 07:23:13.181019 4971 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-mjjv2" event={"ID":"71083227-c9dc-420e-b242-2adf03b636c5","Type":"ContainerStarted","Data":"19097e73c91810ffe8a4c776ace05bdc8956f8f58da395d477b28d9805191b19"} Dec 13 07:23:13 crc kubenswrapper[4971]: I1213 07:23:13.197065 4971 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/community-operators-mjjv2" podStartSLOduration=4.678682732 podStartE2EDuration="7.197044971s" podCreationTimestamp="2025-12-13 07:23:06 +0000 UTC" firstStartedPulling="2025-12-13 07:23:10.14550398 +0000 UTC m=+2046.749913468" lastFinishedPulling="2025-12-13 07:23:12.663866249 +0000 UTC m=+2049.268275707" observedRunningTime="2025-12-13 07:23:13.196958719 +0000 UTC m=+2049.801368167" watchObservedRunningTime="2025-12-13 07:23:13.197044971 +0000 UTC m=+2049.801454429" Dec 13 07:23:14 crc kubenswrapper[4971]: I1213 07:23:14.576432 4971 scope.go:117] "RemoveContainer" containerID="8c01f00505b9ed9c18e1f1f83e29164af0bbff0a698dc2db5d4e5ba5e0a656f9" Dec 13 07:23:15 crc kubenswrapper[4971]: I1213 07:23:15.081230 4971 scope.go:117] "RemoveContainer" containerID="79444a316a054f8a31c3874025055b3501be54e3988c3b3a02890b8643e62d72" Dec 13 07:23:15 crc kubenswrapper[4971]: I1213 07:23:15.152951 4971 scope.go:117] "RemoveContainer" containerID="0b05501843e82c1e08bcb4071dfc72aa660772984720ae7dcfa4a26088cbd9be" Dec 13 07:23:15 crc kubenswrapper[4971]: I1213 07:23:15.191139 4971 scope.go:117] "RemoveContainer" containerID="9d3fdcdb191f5a67a4fa2a8f1e92401ee6d44779f8a7c3fead7ebd96791a001b" Dec 13 07:23:15 crc kubenswrapper[4971]: I1213 07:23:15.237028 4971 scope.go:117] "RemoveContainer" containerID="3397bd79f11b055f37e31637c81e296356926fc8aaa643a9a2f444c08a08c661" Dec 13 07:23:16 crc kubenswrapper[4971]: I1213 07:23:16.153472 4971 patch_prober.go:28] interesting pod/machine-config-daemon-82xjz container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 13 07:23:16 crc kubenswrapper[4971]: I1213 07:23:16.153839 4971 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-82xjz" podUID="e0c4af12-edda-41d7-9b44-f87396174ef2" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 13 07:23:16 crc kubenswrapper[4971]: I1213 07:23:16.153883 4971 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-82xjz" Dec 13 07:23:16 crc kubenswrapper[4971]: I1213 07:23:16.154614 4971 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"adc3c82aa161f8f1a6007b84b7b30285aa46a34a77bebf9c44580ee9879d3298"} pod="openshift-machine-config-operator/machine-config-daemon-82xjz" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Dec 13 07:23:16 crc kubenswrapper[4971]: I1213 07:23:16.154673 4971 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-82xjz" podUID="e0c4af12-edda-41d7-9b44-f87396174ef2" containerName="machine-config-daemon" containerID="cri-o://adc3c82aa161f8f1a6007b84b7b30285aa46a34a77bebf9c44580ee9879d3298" gracePeriod=600 Dec 13 07:23:17 crc kubenswrapper[4971]: I1213 07:23:17.257631 4971 generic.go:334] "Generic (PLEG): container finished" podID="e0c4af12-edda-41d7-9b44-f87396174ef2" containerID="adc3c82aa161f8f1a6007b84b7b30285aa46a34a77bebf9c44580ee9879d3298" exitCode=0 Dec 13 07:23:17 crc kubenswrapper[4971]: I1213 07:23:17.257711 4971 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-82xjz" event={"ID":"e0c4af12-edda-41d7-9b44-f87396174ef2","Type":"ContainerDied","Data":"adc3c82aa161f8f1a6007b84b7b30285aa46a34a77bebf9c44580ee9879d3298"} Dec 13 07:23:17 crc kubenswrapper[4971]: I1213 07:23:17.259439 4971 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-82xjz" event={"ID":"e0c4af12-edda-41d7-9b44-f87396174ef2","Type":"ContainerStarted","Data":"ec2e4b1312b5d5d0545354a836fef4af376ca3b34327ab1b9e0b72c102976079"} Dec 13 07:23:17 crc kubenswrapper[4971]: I1213 07:23:17.259525 4971 scope.go:117] "RemoveContainer" containerID="e8beee43d15719dccb6892d6f5711674492425300efa098be89033bc65977d14" Dec 13 07:23:17 crc kubenswrapper[4971]: I1213 07:23:17.569569 4971 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/community-operators-mjjv2" Dec 13 07:23:17 crc kubenswrapper[4971]: I1213 07:23:17.569626 4971 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/community-operators-mjjv2" Dec 13 07:23:17 crc kubenswrapper[4971]: I1213 07:23:17.623846 4971 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/community-operators-mjjv2" Dec 13 07:23:18 crc kubenswrapper[4971]: I1213 07:23:18.313957 4971 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/community-operators-mjjv2" Dec 13 07:23:18 crc kubenswrapper[4971]: I1213 07:23:18.359462 4971 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-mjjv2"] Dec 13 07:23:20 crc kubenswrapper[4971]: I1213 07:23:20.284782 4971 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/community-operators-mjjv2" podUID="71083227-c9dc-420e-b242-2adf03b636c5" containerName="registry-server" containerID="cri-o://19097e73c91810ffe8a4c776ace05bdc8956f8f58da395d477b28d9805191b19" gracePeriod=2 Dec 13 07:23:20 crc kubenswrapper[4971]: I1213 07:23:20.764500 4971 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-mjjv2" Dec 13 07:23:20 crc kubenswrapper[4971]: I1213 07:23:20.934414 4971 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/71083227-c9dc-420e-b242-2adf03b636c5-utilities\") pod \"71083227-c9dc-420e-b242-2adf03b636c5\" (UID: \"71083227-c9dc-420e-b242-2adf03b636c5\") " Dec 13 07:23:20 crc kubenswrapper[4971]: I1213 07:23:20.934578 4971 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-bh64p\" (UniqueName: \"kubernetes.io/projected/71083227-c9dc-420e-b242-2adf03b636c5-kube-api-access-bh64p\") pod \"71083227-c9dc-420e-b242-2adf03b636c5\" (UID: \"71083227-c9dc-420e-b242-2adf03b636c5\") " Dec 13 07:23:20 crc kubenswrapper[4971]: I1213 07:23:20.934605 4971 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/71083227-c9dc-420e-b242-2adf03b636c5-catalog-content\") pod \"71083227-c9dc-420e-b242-2adf03b636c5\" (UID: \"71083227-c9dc-420e-b242-2adf03b636c5\") " Dec 13 07:23:20 crc kubenswrapper[4971]: I1213 07:23:20.935219 4971 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/71083227-c9dc-420e-b242-2adf03b636c5-utilities" (OuterVolumeSpecName: "utilities") pod "71083227-c9dc-420e-b242-2adf03b636c5" (UID: "71083227-c9dc-420e-b242-2adf03b636c5"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 13 07:23:20 crc kubenswrapper[4971]: I1213 07:23:20.942823 4971 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/71083227-c9dc-420e-b242-2adf03b636c5-kube-api-access-bh64p" (OuterVolumeSpecName: "kube-api-access-bh64p") pod "71083227-c9dc-420e-b242-2adf03b636c5" (UID: "71083227-c9dc-420e-b242-2adf03b636c5"). InnerVolumeSpecName "kube-api-access-bh64p". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 13 07:23:21 crc kubenswrapper[4971]: I1213 07:23:21.009596 4971 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/71083227-c9dc-420e-b242-2adf03b636c5-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "71083227-c9dc-420e-b242-2adf03b636c5" (UID: "71083227-c9dc-420e-b242-2adf03b636c5"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 13 07:23:21 crc kubenswrapper[4971]: I1213 07:23:21.036315 4971 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/71083227-c9dc-420e-b242-2adf03b636c5-utilities\") on node \"crc\" DevicePath \"\"" Dec 13 07:23:21 crc kubenswrapper[4971]: I1213 07:23:21.036354 4971 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-bh64p\" (UniqueName: \"kubernetes.io/projected/71083227-c9dc-420e-b242-2adf03b636c5-kube-api-access-bh64p\") on node \"crc\" DevicePath \"\"" Dec 13 07:23:21 crc kubenswrapper[4971]: I1213 07:23:21.036365 4971 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/71083227-c9dc-420e-b242-2adf03b636c5-catalog-content\") on node \"crc\" DevicePath \"\"" Dec 13 07:23:21 crc kubenswrapper[4971]: I1213 07:23:21.296707 4971 generic.go:334] "Generic (PLEG): container finished" podID="71083227-c9dc-420e-b242-2adf03b636c5" containerID="19097e73c91810ffe8a4c776ace05bdc8956f8f58da395d477b28d9805191b19" exitCode=0 Dec 13 07:23:21 crc kubenswrapper[4971]: I1213 07:23:21.296761 4971 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-mjjv2" event={"ID":"71083227-c9dc-420e-b242-2adf03b636c5","Type":"ContainerDied","Data":"19097e73c91810ffe8a4c776ace05bdc8956f8f58da395d477b28d9805191b19"} Dec 13 07:23:21 crc kubenswrapper[4971]: I1213 07:23:21.296789 4971 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-mjjv2" event={"ID":"71083227-c9dc-420e-b242-2adf03b636c5","Type":"ContainerDied","Data":"654d05fe120905b43a46eda225c765af18f4a1c8af3f3b41dda916ae247dd7af"} Dec 13 07:23:21 crc kubenswrapper[4971]: I1213 07:23:21.296807 4971 scope.go:117] "RemoveContainer" containerID="19097e73c91810ffe8a4c776ace05bdc8956f8f58da395d477b28d9805191b19" Dec 13 07:23:21 crc kubenswrapper[4971]: I1213 07:23:21.296811 4971 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-mjjv2" Dec 13 07:23:21 crc kubenswrapper[4971]: I1213 07:23:21.321447 4971 scope.go:117] "RemoveContainer" containerID="eacfac38c2440049c59b71905e0f0ab0c2bbd2cbea97333cf36ee4c35c878610" Dec 13 07:23:21 crc kubenswrapper[4971]: I1213 07:23:21.347683 4971 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-mjjv2"] Dec 13 07:23:21 crc kubenswrapper[4971]: I1213 07:23:21.359972 4971 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/community-operators-mjjv2"] Dec 13 07:23:21 crc kubenswrapper[4971]: I1213 07:23:21.363820 4971 scope.go:117] "RemoveContainer" containerID="89ea9768c86688de6fc20f29eee7c384b2e190ff5862cfeae9266a11011546e0" Dec 13 07:23:21 crc kubenswrapper[4971]: I1213 07:23:21.389270 4971 scope.go:117] "RemoveContainer" containerID="19097e73c91810ffe8a4c776ace05bdc8956f8f58da395d477b28d9805191b19" Dec 13 07:23:21 crc kubenswrapper[4971]: E1213 07:23:21.389727 4971 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"19097e73c91810ffe8a4c776ace05bdc8956f8f58da395d477b28d9805191b19\": container with ID starting with 19097e73c91810ffe8a4c776ace05bdc8956f8f58da395d477b28d9805191b19 not found: ID does not exist" containerID="19097e73c91810ffe8a4c776ace05bdc8956f8f58da395d477b28d9805191b19" Dec 13 07:23:21 crc kubenswrapper[4971]: I1213 07:23:21.389773 4971 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"19097e73c91810ffe8a4c776ace05bdc8956f8f58da395d477b28d9805191b19"} err="failed to get container status \"19097e73c91810ffe8a4c776ace05bdc8956f8f58da395d477b28d9805191b19\": rpc error: code = NotFound desc = could not find container \"19097e73c91810ffe8a4c776ace05bdc8956f8f58da395d477b28d9805191b19\": container with ID starting with 19097e73c91810ffe8a4c776ace05bdc8956f8f58da395d477b28d9805191b19 not found: ID does not exist" Dec 13 07:23:21 crc kubenswrapper[4971]: I1213 07:23:21.389801 4971 scope.go:117] "RemoveContainer" containerID="eacfac38c2440049c59b71905e0f0ab0c2bbd2cbea97333cf36ee4c35c878610" Dec 13 07:23:21 crc kubenswrapper[4971]: E1213 07:23:21.390220 4971 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"eacfac38c2440049c59b71905e0f0ab0c2bbd2cbea97333cf36ee4c35c878610\": container with ID starting with eacfac38c2440049c59b71905e0f0ab0c2bbd2cbea97333cf36ee4c35c878610 not found: ID does not exist" containerID="eacfac38c2440049c59b71905e0f0ab0c2bbd2cbea97333cf36ee4c35c878610" Dec 13 07:23:21 crc kubenswrapper[4971]: I1213 07:23:21.390249 4971 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"eacfac38c2440049c59b71905e0f0ab0c2bbd2cbea97333cf36ee4c35c878610"} err="failed to get container status \"eacfac38c2440049c59b71905e0f0ab0c2bbd2cbea97333cf36ee4c35c878610\": rpc error: code = NotFound desc = could not find container \"eacfac38c2440049c59b71905e0f0ab0c2bbd2cbea97333cf36ee4c35c878610\": container with ID starting with eacfac38c2440049c59b71905e0f0ab0c2bbd2cbea97333cf36ee4c35c878610 not found: ID does not exist" Dec 13 07:23:21 crc kubenswrapper[4971]: I1213 07:23:21.390272 4971 scope.go:117] "RemoveContainer" containerID="89ea9768c86688de6fc20f29eee7c384b2e190ff5862cfeae9266a11011546e0" Dec 13 07:23:21 crc kubenswrapper[4971]: E1213 07:23:21.390986 4971 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"89ea9768c86688de6fc20f29eee7c384b2e190ff5862cfeae9266a11011546e0\": container with ID starting with 89ea9768c86688de6fc20f29eee7c384b2e190ff5862cfeae9266a11011546e0 not found: ID does not exist" containerID="89ea9768c86688de6fc20f29eee7c384b2e190ff5862cfeae9266a11011546e0" Dec 13 07:23:21 crc kubenswrapper[4971]: I1213 07:23:21.391013 4971 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"89ea9768c86688de6fc20f29eee7c384b2e190ff5862cfeae9266a11011546e0"} err="failed to get container status \"89ea9768c86688de6fc20f29eee7c384b2e190ff5862cfeae9266a11011546e0\": rpc error: code = NotFound desc = could not find container \"89ea9768c86688de6fc20f29eee7c384b2e190ff5862cfeae9266a11011546e0\": container with ID starting with 89ea9768c86688de6fc20f29eee7c384b2e190ff5862cfeae9266a11011546e0 not found: ID does not exist" Dec 13 07:23:21 crc kubenswrapper[4971]: I1213 07:23:21.779884 4971 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="71083227-c9dc-420e-b242-2adf03b636c5" path="/var/lib/kubelet/pods/71083227-c9dc-420e-b242-2adf03b636c5/volumes" Dec 13 07:23:38 crc kubenswrapper[4971]: I1213 07:23:38.051460 4971 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-cell1-db-create-4j49j"] Dec 13 07:23:38 crc kubenswrapper[4971]: I1213 07:23:38.062848 4971 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-api-048e-account-create-update-kggpx"] Dec 13 07:23:38 crc kubenswrapper[4971]: I1213 07:23:38.074200 4971 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-cell1-db-create-4j49j"] Dec 13 07:23:38 crc kubenswrapper[4971]: I1213 07:23:38.085721 4971 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-api-048e-account-create-update-kggpx"] Dec 13 07:23:38 crc kubenswrapper[4971]: I1213 07:23:38.939685 4971 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-operators-sbkvv"] Dec 13 07:23:38 crc kubenswrapper[4971]: E1213 07:23:38.940422 4971 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="71083227-c9dc-420e-b242-2adf03b636c5" containerName="extract-content" Dec 13 07:23:38 crc kubenswrapper[4971]: I1213 07:23:38.941582 4971 state_mem.go:107] "Deleted CPUSet assignment" podUID="71083227-c9dc-420e-b242-2adf03b636c5" containerName="extract-content" Dec 13 07:23:38 crc kubenswrapper[4971]: E1213 07:23:38.941687 4971 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="71083227-c9dc-420e-b242-2adf03b636c5" containerName="extract-utilities" Dec 13 07:23:38 crc kubenswrapper[4971]: I1213 07:23:38.941698 4971 state_mem.go:107] "Deleted CPUSet assignment" podUID="71083227-c9dc-420e-b242-2adf03b636c5" containerName="extract-utilities" Dec 13 07:23:38 crc kubenswrapper[4971]: E1213 07:23:38.941711 4971 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="71083227-c9dc-420e-b242-2adf03b636c5" containerName="registry-server" Dec 13 07:23:38 crc kubenswrapper[4971]: I1213 07:23:38.941718 4971 state_mem.go:107] "Deleted CPUSet assignment" podUID="71083227-c9dc-420e-b242-2adf03b636c5" containerName="registry-server" Dec 13 07:23:38 crc kubenswrapper[4971]: I1213 07:23:38.942071 4971 memory_manager.go:354] "RemoveStaleState removing state" podUID="71083227-c9dc-420e-b242-2adf03b636c5" containerName="registry-server" Dec 13 07:23:38 crc kubenswrapper[4971]: I1213 07:23:38.943746 4971 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-sbkvv" Dec 13 07:23:38 crc kubenswrapper[4971]: I1213 07:23:38.957496 4971 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-sbkvv"] Dec 13 07:23:39 crc kubenswrapper[4971]: I1213 07:23:39.042194 4971 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/fbe986fc-eacf-42a4-98f6-288d943470dc-utilities\") pod \"redhat-operators-sbkvv\" (UID: \"fbe986fc-eacf-42a4-98f6-288d943470dc\") " pod="openshift-marketplace/redhat-operators-sbkvv" Dec 13 07:23:39 crc kubenswrapper[4971]: I1213 07:23:39.042273 4971 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-z688t\" (UniqueName: \"kubernetes.io/projected/fbe986fc-eacf-42a4-98f6-288d943470dc-kube-api-access-z688t\") pod \"redhat-operators-sbkvv\" (UID: \"fbe986fc-eacf-42a4-98f6-288d943470dc\") " pod="openshift-marketplace/redhat-operators-sbkvv" Dec 13 07:23:39 crc kubenswrapper[4971]: I1213 07:23:39.042345 4971 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/fbe986fc-eacf-42a4-98f6-288d943470dc-catalog-content\") pod \"redhat-operators-sbkvv\" (UID: \"fbe986fc-eacf-42a4-98f6-288d943470dc\") " pod="openshift-marketplace/redhat-operators-sbkvv" Dec 13 07:23:39 crc kubenswrapper[4971]: I1213 07:23:39.043043 4971 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-cell0-db-create-jsdsv"] Dec 13 07:23:39 crc kubenswrapper[4971]: I1213 07:23:39.065743 4971 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-cell0-db-create-jsdsv"] Dec 13 07:23:39 crc kubenswrapper[4971]: I1213 07:23:39.075499 4971 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-cell0-ed10-account-create-update-tth4x"] Dec 13 07:23:39 crc kubenswrapper[4971]: I1213 07:23:39.084592 4971 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-api-db-create-ndmqx"] Dec 13 07:23:39 crc kubenswrapper[4971]: I1213 07:23:39.091973 4971 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-cell1-61ad-account-create-update-4ckp9"] Dec 13 07:23:39 crc kubenswrapper[4971]: I1213 07:23:39.099580 4971 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-cell0-ed10-account-create-update-tth4x"] Dec 13 07:23:39 crc kubenswrapper[4971]: I1213 07:23:39.106583 4971 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-api-db-create-ndmqx"] Dec 13 07:23:39 crc kubenswrapper[4971]: I1213 07:23:39.114342 4971 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-cell1-61ad-account-create-update-4ckp9"] Dec 13 07:23:39 crc kubenswrapper[4971]: I1213 07:23:39.143913 4971 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/fbe986fc-eacf-42a4-98f6-288d943470dc-catalog-content\") pod \"redhat-operators-sbkvv\" (UID: \"fbe986fc-eacf-42a4-98f6-288d943470dc\") " pod="openshift-marketplace/redhat-operators-sbkvv" Dec 13 07:23:39 crc kubenswrapper[4971]: I1213 07:23:39.144021 4971 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/fbe986fc-eacf-42a4-98f6-288d943470dc-utilities\") pod \"redhat-operators-sbkvv\" (UID: \"fbe986fc-eacf-42a4-98f6-288d943470dc\") " pod="openshift-marketplace/redhat-operators-sbkvv" Dec 13 07:23:39 crc kubenswrapper[4971]: I1213 07:23:39.144062 4971 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-z688t\" (UniqueName: \"kubernetes.io/projected/fbe986fc-eacf-42a4-98f6-288d943470dc-kube-api-access-z688t\") pod \"redhat-operators-sbkvv\" (UID: \"fbe986fc-eacf-42a4-98f6-288d943470dc\") " pod="openshift-marketplace/redhat-operators-sbkvv" Dec 13 07:23:39 crc kubenswrapper[4971]: I1213 07:23:39.144616 4971 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/fbe986fc-eacf-42a4-98f6-288d943470dc-catalog-content\") pod \"redhat-operators-sbkvv\" (UID: \"fbe986fc-eacf-42a4-98f6-288d943470dc\") " pod="openshift-marketplace/redhat-operators-sbkvv" Dec 13 07:23:39 crc kubenswrapper[4971]: I1213 07:23:39.144624 4971 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/fbe986fc-eacf-42a4-98f6-288d943470dc-utilities\") pod \"redhat-operators-sbkvv\" (UID: \"fbe986fc-eacf-42a4-98f6-288d943470dc\") " pod="openshift-marketplace/redhat-operators-sbkvv" Dec 13 07:23:39 crc kubenswrapper[4971]: I1213 07:23:39.172618 4971 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-z688t\" (UniqueName: \"kubernetes.io/projected/fbe986fc-eacf-42a4-98f6-288d943470dc-kube-api-access-z688t\") pod \"redhat-operators-sbkvv\" (UID: \"fbe986fc-eacf-42a4-98f6-288d943470dc\") " pod="openshift-marketplace/redhat-operators-sbkvv" Dec 13 07:23:39 crc kubenswrapper[4971]: I1213 07:23:39.269318 4971 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-sbkvv" Dec 13 07:23:39 crc kubenswrapper[4971]: I1213 07:23:39.778323 4971 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="0ed1405a-8326-4fe9-8fb4-1d97b9e8befb" path="/var/lib/kubelet/pods/0ed1405a-8326-4fe9-8fb4-1d97b9e8befb/volumes" Dec 13 07:23:39 crc kubenswrapper[4971]: I1213 07:23:39.779298 4971 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="5ba87867-bb90-4587-9012-6db9d2901b10" path="/var/lib/kubelet/pods/5ba87867-bb90-4587-9012-6db9d2901b10/volumes" Dec 13 07:23:39 crc kubenswrapper[4971]: I1213 07:23:39.779919 4971 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="5ffa477d-1f7b-445f-9a97-65a7889e44d4" path="/var/lib/kubelet/pods/5ffa477d-1f7b-445f-9a97-65a7889e44d4/volumes" Dec 13 07:23:39 crc kubenswrapper[4971]: I1213 07:23:39.780573 4971 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="8b60a7ee-8d5e-4bf2-8d64-161aeced2357" path="/var/lib/kubelet/pods/8b60a7ee-8d5e-4bf2-8d64-161aeced2357/volumes" Dec 13 07:23:39 crc kubenswrapper[4971]: I1213 07:23:39.781722 4971 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="e0e9f238-3cb4-43ff-bac3-f1c283d0915f" path="/var/lib/kubelet/pods/e0e9f238-3cb4-43ff-bac3-f1c283d0915f/volumes" Dec 13 07:23:39 crc kubenswrapper[4971]: I1213 07:23:39.782339 4971 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="edac9227-16d2-48b0-ac2b-3f71c67d6cf8" path="/var/lib/kubelet/pods/edac9227-16d2-48b0-ac2b-3f71c67d6cf8/volumes" Dec 13 07:23:39 crc kubenswrapper[4971]: I1213 07:23:39.817209 4971 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-sbkvv"] Dec 13 07:23:39 crc kubenswrapper[4971]: W1213 07:23:39.841823 4971 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podfbe986fc_eacf_42a4_98f6_288d943470dc.slice/crio-24d381ef9b49217881442c4dc31812e0cee8fbbbffc9bbf4320e59fe18573b7c WatchSource:0}: Error finding container 24d381ef9b49217881442c4dc31812e0cee8fbbbffc9bbf4320e59fe18573b7c: Status 404 returned error can't find the container with id 24d381ef9b49217881442c4dc31812e0cee8fbbbffc9bbf4320e59fe18573b7c Dec 13 07:23:40 crc kubenswrapper[4971]: I1213 07:23:40.491024 4971 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-sbkvv" event={"ID":"fbe986fc-eacf-42a4-98f6-288d943470dc","Type":"ContainerStarted","Data":"24d381ef9b49217881442c4dc31812e0cee8fbbbffc9bbf4320e59fe18573b7c"} Dec 13 07:23:41 crc kubenswrapper[4971]: I1213 07:23:41.309361 4971 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/certified-operators-587w4"] Dec 13 07:23:41 crc kubenswrapper[4971]: I1213 07:23:41.311958 4971 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-587w4" Dec 13 07:23:41 crc kubenswrapper[4971]: I1213 07:23:41.324770 4971 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-587w4"] Dec 13 07:23:41 crc kubenswrapper[4971]: I1213 07:23:41.392404 4971 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/3e083b0a-4cc8-4a7e-83bf-956166f10069-utilities\") pod \"certified-operators-587w4\" (UID: \"3e083b0a-4cc8-4a7e-83bf-956166f10069\") " pod="openshift-marketplace/certified-operators-587w4" Dec 13 07:23:41 crc kubenswrapper[4971]: I1213 07:23:41.392580 4971 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/3e083b0a-4cc8-4a7e-83bf-956166f10069-catalog-content\") pod \"certified-operators-587w4\" (UID: \"3e083b0a-4cc8-4a7e-83bf-956166f10069\") " pod="openshift-marketplace/certified-operators-587w4" Dec 13 07:23:41 crc kubenswrapper[4971]: I1213 07:23:41.392626 4971 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-lhfh8\" (UniqueName: \"kubernetes.io/projected/3e083b0a-4cc8-4a7e-83bf-956166f10069-kube-api-access-lhfh8\") pod \"certified-operators-587w4\" (UID: \"3e083b0a-4cc8-4a7e-83bf-956166f10069\") " pod="openshift-marketplace/certified-operators-587w4" Dec 13 07:23:41 crc kubenswrapper[4971]: I1213 07:23:41.494244 4971 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/3e083b0a-4cc8-4a7e-83bf-956166f10069-catalog-content\") pod \"certified-operators-587w4\" (UID: \"3e083b0a-4cc8-4a7e-83bf-956166f10069\") " pod="openshift-marketplace/certified-operators-587w4" Dec 13 07:23:41 crc kubenswrapper[4971]: I1213 07:23:41.494352 4971 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-lhfh8\" (UniqueName: \"kubernetes.io/projected/3e083b0a-4cc8-4a7e-83bf-956166f10069-kube-api-access-lhfh8\") pod \"certified-operators-587w4\" (UID: \"3e083b0a-4cc8-4a7e-83bf-956166f10069\") " pod="openshift-marketplace/certified-operators-587w4" Dec 13 07:23:41 crc kubenswrapper[4971]: I1213 07:23:41.494477 4971 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/3e083b0a-4cc8-4a7e-83bf-956166f10069-utilities\") pod \"certified-operators-587w4\" (UID: \"3e083b0a-4cc8-4a7e-83bf-956166f10069\") " pod="openshift-marketplace/certified-operators-587w4" Dec 13 07:23:41 crc kubenswrapper[4971]: I1213 07:23:41.495368 4971 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/3e083b0a-4cc8-4a7e-83bf-956166f10069-utilities\") pod \"certified-operators-587w4\" (UID: \"3e083b0a-4cc8-4a7e-83bf-956166f10069\") " pod="openshift-marketplace/certified-operators-587w4" Dec 13 07:23:41 crc kubenswrapper[4971]: I1213 07:23:41.495679 4971 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/3e083b0a-4cc8-4a7e-83bf-956166f10069-catalog-content\") pod \"certified-operators-587w4\" (UID: \"3e083b0a-4cc8-4a7e-83bf-956166f10069\") " pod="openshift-marketplace/certified-operators-587w4" Dec 13 07:23:41 crc kubenswrapper[4971]: I1213 07:23:41.503310 4971 generic.go:334] "Generic (PLEG): container finished" podID="fbe986fc-eacf-42a4-98f6-288d943470dc" containerID="b95450ada02de13d4d8f2422676cf14a569d4f84c2be200fa6be72fcea571096" exitCode=0 Dec 13 07:23:41 crc kubenswrapper[4971]: I1213 07:23:41.503353 4971 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-sbkvv" event={"ID":"fbe986fc-eacf-42a4-98f6-288d943470dc","Type":"ContainerDied","Data":"b95450ada02de13d4d8f2422676cf14a569d4f84c2be200fa6be72fcea571096"} Dec 13 07:23:41 crc kubenswrapper[4971]: I1213 07:23:41.529111 4971 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-lhfh8\" (UniqueName: \"kubernetes.io/projected/3e083b0a-4cc8-4a7e-83bf-956166f10069-kube-api-access-lhfh8\") pod \"certified-operators-587w4\" (UID: \"3e083b0a-4cc8-4a7e-83bf-956166f10069\") " pod="openshift-marketplace/certified-operators-587w4" Dec 13 07:23:41 crc kubenswrapper[4971]: I1213 07:23:41.643133 4971 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-587w4" Dec 13 07:23:42 crc kubenswrapper[4971]: W1213 07:23:42.157461 4971 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod3e083b0a_4cc8_4a7e_83bf_956166f10069.slice/crio-828302ef4afbae64996e0d7090e90ba4e76dfc5c1e319a88440111bde7ea52d7 WatchSource:0}: Error finding container 828302ef4afbae64996e0d7090e90ba4e76dfc5c1e319a88440111bde7ea52d7: Status 404 returned error can't find the container with id 828302ef4afbae64996e0d7090e90ba4e76dfc5c1e319a88440111bde7ea52d7 Dec 13 07:23:42 crc kubenswrapper[4971]: I1213 07:23:42.159506 4971 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-587w4"] Dec 13 07:23:42 crc kubenswrapper[4971]: I1213 07:23:42.512793 4971 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-587w4" event={"ID":"3e083b0a-4cc8-4a7e-83bf-956166f10069","Type":"ContainerStarted","Data":"828302ef4afbae64996e0d7090e90ba4e76dfc5c1e319a88440111bde7ea52d7"} Dec 13 07:23:43 crc kubenswrapper[4971]: I1213 07:23:43.524342 4971 generic.go:334] "Generic (PLEG): container finished" podID="3e083b0a-4cc8-4a7e-83bf-956166f10069" containerID="bac558e7b4872b3dec8fb88e42d3a2ee993882f4a4d572dee1e4f4ba767591cb" exitCode=0 Dec 13 07:23:43 crc kubenswrapper[4971]: I1213 07:23:43.524577 4971 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-587w4" event={"ID":"3e083b0a-4cc8-4a7e-83bf-956166f10069","Type":"ContainerDied","Data":"bac558e7b4872b3dec8fb88e42d3a2ee993882f4a4d572dee1e4f4ba767591cb"} Dec 13 07:23:49 crc kubenswrapper[4971]: I1213 07:23:49.587381 4971 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-sbkvv" event={"ID":"fbe986fc-eacf-42a4-98f6-288d943470dc","Type":"ContainerStarted","Data":"aa99e3a9c2358daf739038ac0fa7be8391fe24da9e5dd84f2f83edd9f63c9901"} Dec 13 07:23:52 crc kubenswrapper[4971]: I1213 07:23:51.619990 4971 generic.go:334] "Generic (PLEG): container finished" podID="fbe986fc-eacf-42a4-98f6-288d943470dc" containerID="aa99e3a9c2358daf739038ac0fa7be8391fe24da9e5dd84f2f83edd9f63c9901" exitCode=0 Dec 13 07:23:52 crc kubenswrapper[4971]: I1213 07:23:51.620098 4971 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-sbkvv" event={"ID":"fbe986fc-eacf-42a4-98f6-288d943470dc","Type":"ContainerDied","Data":"aa99e3a9c2358daf739038ac0fa7be8391fe24da9e5dd84f2f83edd9f63c9901"} Dec 13 07:23:52 crc kubenswrapper[4971]: I1213 07:23:51.624282 4971 provider.go:102] Refreshing cache for provider: *credentialprovider.defaultDockerConfigProvider Dec 13 07:23:54 crc kubenswrapper[4971]: I1213 07:23:54.664129 4971 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-587w4" event={"ID":"3e083b0a-4cc8-4a7e-83bf-956166f10069","Type":"ContainerStarted","Data":"e8f8bda0e7d32492e4fdf48d04e708279803e50d1ade1d6b080adf6af134dd0e"} Dec 13 07:23:55 crc kubenswrapper[4971]: I1213 07:23:55.676298 4971 generic.go:334] "Generic (PLEG): container finished" podID="3e083b0a-4cc8-4a7e-83bf-956166f10069" containerID="e8f8bda0e7d32492e4fdf48d04e708279803e50d1ade1d6b080adf6af134dd0e" exitCode=0 Dec 13 07:23:55 crc kubenswrapper[4971]: I1213 07:23:55.676329 4971 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-587w4" event={"ID":"3e083b0a-4cc8-4a7e-83bf-956166f10069","Type":"ContainerDied","Data":"e8f8bda0e7d32492e4fdf48d04e708279803e50d1ade1d6b080adf6af134dd0e"} Dec 13 07:24:00 crc kubenswrapper[4971]: I1213 07:24:00.735934 4971 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-sbkvv" event={"ID":"fbe986fc-eacf-42a4-98f6-288d943470dc","Type":"ContainerStarted","Data":"51d90418c1a42b289e90ac8160fe1831853dd7604a28213a37f478e1aecccee5"} Dec 13 07:24:02 crc kubenswrapper[4971]: I1213 07:24:02.770924 4971 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-587w4" event={"ID":"3e083b0a-4cc8-4a7e-83bf-956166f10069","Type":"ContainerStarted","Data":"5f7cd20e93b0677908dfeefe7527764d6bad5619e32d02cbe2cd660ad8713972"} Dec 13 07:24:02 crc kubenswrapper[4971]: I1213 07:24:02.774361 4971 generic.go:334] "Generic (PLEG): container finished" podID="80f5a7cb-cc56-4d8e-ba65-0198d682bcc5" containerID="66b156cf0ae30ae6d8916884aac86cb1ee337155f7ef722bd822b01428f81924" exitCode=0 Dec 13 07:24:02 crc kubenswrapper[4971]: I1213 07:24:02.774538 4971 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-7nvdr" event={"ID":"80f5a7cb-cc56-4d8e-ba65-0198d682bcc5","Type":"ContainerDied","Data":"66b156cf0ae30ae6d8916884aac86cb1ee337155f7ef722bd822b01428f81924"} Dec 13 07:24:02 crc kubenswrapper[4971]: I1213 07:24:02.792381 4971 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/certified-operators-587w4" podStartSLOduration=4.567287566 podStartE2EDuration="21.792363137s" podCreationTimestamp="2025-12-13 07:23:41 +0000 UTC" firstStartedPulling="2025-12-13 07:23:45.163201708 +0000 UTC m=+2081.767611166" lastFinishedPulling="2025-12-13 07:24:02.388277249 +0000 UTC m=+2098.992686737" observedRunningTime="2025-12-13 07:24:02.789246989 +0000 UTC m=+2099.393656437" watchObservedRunningTime="2025-12-13 07:24:02.792363137 +0000 UTC m=+2099.396772585" Dec 13 07:24:02 crc kubenswrapper[4971]: I1213 07:24:02.812776 4971 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-operators-sbkvv" podStartSLOduration=12.332858241 podStartE2EDuration="24.812756669s" podCreationTimestamp="2025-12-13 07:23:38 +0000 UTC" firstStartedPulling="2025-12-13 07:23:41.505842589 +0000 UTC m=+2078.110252067" lastFinishedPulling="2025-12-13 07:23:53.985741047 +0000 UTC m=+2090.590150495" observedRunningTime="2025-12-13 07:24:02.806361078 +0000 UTC m=+2099.410770576" watchObservedRunningTime="2025-12-13 07:24:02.812756669 +0000 UTC m=+2099.417166137" Dec 13 07:24:04 crc kubenswrapper[4971]: I1213 07:24:04.202705 4971 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-7nvdr" Dec 13 07:24:04 crc kubenswrapper[4971]: I1213 07:24:04.389233 4971 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"bootstrap-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/80f5a7cb-cc56-4d8e-ba65-0198d682bcc5-bootstrap-combined-ca-bundle\") pod \"80f5a7cb-cc56-4d8e-ba65-0198d682bcc5\" (UID: \"80f5a7cb-cc56-4d8e-ba65-0198d682bcc5\") " Dec 13 07:24:04 crc kubenswrapper[4971]: I1213 07:24:04.389364 4971 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/80f5a7cb-cc56-4d8e-ba65-0198d682bcc5-ssh-key\") pod \"80f5a7cb-cc56-4d8e-ba65-0198d682bcc5\" (UID: \"80f5a7cb-cc56-4d8e-ba65-0198d682bcc5\") " Dec 13 07:24:04 crc kubenswrapper[4971]: I1213 07:24:04.389435 4971 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-gntbn\" (UniqueName: \"kubernetes.io/projected/80f5a7cb-cc56-4d8e-ba65-0198d682bcc5-kube-api-access-gntbn\") pod \"80f5a7cb-cc56-4d8e-ba65-0198d682bcc5\" (UID: \"80f5a7cb-cc56-4d8e-ba65-0198d682bcc5\") " Dec 13 07:24:04 crc kubenswrapper[4971]: I1213 07:24:04.389495 4971 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/80f5a7cb-cc56-4d8e-ba65-0198d682bcc5-inventory\") pod \"80f5a7cb-cc56-4d8e-ba65-0198d682bcc5\" (UID: \"80f5a7cb-cc56-4d8e-ba65-0198d682bcc5\") " Dec 13 07:24:04 crc kubenswrapper[4971]: I1213 07:24:04.404166 4971 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/80f5a7cb-cc56-4d8e-ba65-0198d682bcc5-kube-api-access-gntbn" (OuterVolumeSpecName: "kube-api-access-gntbn") pod "80f5a7cb-cc56-4d8e-ba65-0198d682bcc5" (UID: "80f5a7cb-cc56-4d8e-ba65-0198d682bcc5"). InnerVolumeSpecName "kube-api-access-gntbn". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 13 07:24:04 crc kubenswrapper[4971]: I1213 07:24:04.405778 4971 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/80f5a7cb-cc56-4d8e-ba65-0198d682bcc5-bootstrap-combined-ca-bundle" (OuterVolumeSpecName: "bootstrap-combined-ca-bundle") pod "80f5a7cb-cc56-4d8e-ba65-0198d682bcc5" (UID: "80f5a7cb-cc56-4d8e-ba65-0198d682bcc5"). InnerVolumeSpecName "bootstrap-combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 13 07:24:04 crc kubenswrapper[4971]: I1213 07:24:04.420054 4971 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/80f5a7cb-cc56-4d8e-ba65-0198d682bcc5-ssh-key" (OuterVolumeSpecName: "ssh-key") pod "80f5a7cb-cc56-4d8e-ba65-0198d682bcc5" (UID: "80f5a7cb-cc56-4d8e-ba65-0198d682bcc5"). InnerVolumeSpecName "ssh-key". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 13 07:24:04 crc kubenswrapper[4971]: I1213 07:24:04.426490 4971 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/80f5a7cb-cc56-4d8e-ba65-0198d682bcc5-inventory" (OuterVolumeSpecName: "inventory") pod "80f5a7cb-cc56-4d8e-ba65-0198d682bcc5" (UID: "80f5a7cb-cc56-4d8e-ba65-0198d682bcc5"). InnerVolumeSpecName "inventory". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 13 07:24:04 crc kubenswrapper[4971]: I1213 07:24:04.491584 4971 reconciler_common.go:293] "Volume detached for volume \"bootstrap-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/80f5a7cb-cc56-4d8e-ba65-0198d682bcc5-bootstrap-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 13 07:24:04 crc kubenswrapper[4971]: I1213 07:24:04.491623 4971 reconciler_common.go:293] "Volume detached for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/80f5a7cb-cc56-4d8e-ba65-0198d682bcc5-ssh-key\") on node \"crc\" DevicePath \"\"" Dec 13 07:24:04 crc kubenswrapper[4971]: I1213 07:24:04.491638 4971 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-gntbn\" (UniqueName: \"kubernetes.io/projected/80f5a7cb-cc56-4d8e-ba65-0198d682bcc5-kube-api-access-gntbn\") on node \"crc\" DevicePath \"\"" Dec 13 07:24:04 crc kubenswrapper[4971]: I1213 07:24:04.491650 4971 reconciler_common.go:293] "Volume detached for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/80f5a7cb-cc56-4d8e-ba65-0198d682bcc5-inventory\") on node \"crc\" DevicePath \"\"" Dec 13 07:24:04 crc kubenswrapper[4971]: I1213 07:24:04.793791 4971 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-7nvdr" event={"ID":"80f5a7cb-cc56-4d8e-ba65-0198d682bcc5","Type":"ContainerDied","Data":"6834ce97e2c8534adaa3f8861449d6c40d5d35a4171c46b13ab0b2a625ebc948"} Dec 13 07:24:04 crc kubenswrapper[4971]: I1213 07:24:04.793834 4971 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="6834ce97e2c8534adaa3f8861449d6c40d5d35a4171c46b13ab0b2a625ebc948" Dec 13 07:24:04 crc kubenswrapper[4971]: I1213 07:24:04.794669 4971 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-7nvdr" Dec 13 07:24:04 crc kubenswrapper[4971]: I1213 07:24:04.940957 4971 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/configure-network-edpm-deployment-openstack-edpm-ipam-qj5wr"] Dec 13 07:24:04 crc kubenswrapper[4971]: E1213 07:24:04.941360 4971 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="80f5a7cb-cc56-4d8e-ba65-0198d682bcc5" containerName="bootstrap-edpm-deployment-openstack-edpm-ipam" Dec 13 07:24:04 crc kubenswrapper[4971]: I1213 07:24:04.941382 4971 state_mem.go:107] "Deleted CPUSet assignment" podUID="80f5a7cb-cc56-4d8e-ba65-0198d682bcc5" containerName="bootstrap-edpm-deployment-openstack-edpm-ipam" Dec 13 07:24:04 crc kubenswrapper[4971]: I1213 07:24:04.941623 4971 memory_manager.go:354] "RemoveStaleState removing state" podUID="80f5a7cb-cc56-4d8e-ba65-0198d682bcc5" containerName="bootstrap-edpm-deployment-openstack-edpm-ipam" Dec 13 07:24:04 crc kubenswrapper[4971]: I1213 07:24:04.942289 4971 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/configure-network-edpm-deployment-openstack-edpm-ipam-qj5wr" Dec 13 07:24:04 crc kubenswrapper[4971]: I1213 07:24:04.945633 4971 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-edpm-ipam-dockercfg-zb2j5" Dec 13 07:24:04 crc kubenswrapper[4971]: I1213 07:24:04.945670 4971 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplanenodeset-openstack-edpm-ipam" Dec 13 07:24:04 crc kubenswrapper[4971]: I1213 07:24:04.945698 4971 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplane-ansible-ssh-private-key-secret" Dec 13 07:24:04 crc kubenswrapper[4971]: I1213 07:24:04.945845 4971 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-aee-default-env" Dec 13 07:24:04 crc kubenswrapper[4971]: I1213 07:24:04.971793 4971 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/configure-network-edpm-deployment-openstack-edpm-ipam-qj5wr"] Dec 13 07:24:05 crc kubenswrapper[4971]: I1213 07:24:05.104609 4971 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/edef63aa-a18d-4591-a477-924e5cef56d8-inventory\") pod \"configure-network-edpm-deployment-openstack-edpm-ipam-qj5wr\" (UID: \"edef63aa-a18d-4591-a477-924e5cef56d8\") " pod="openstack/configure-network-edpm-deployment-openstack-edpm-ipam-qj5wr" Dec 13 07:24:05 crc kubenswrapper[4971]: I1213 07:24:05.104883 4971 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-7vm5c\" (UniqueName: \"kubernetes.io/projected/edef63aa-a18d-4591-a477-924e5cef56d8-kube-api-access-7vm5c\") pod \"configure-network-edpm-deployment-openstack-edpm-ipam-qj5wr\" (UID: \"edef63aa-a18d-4591-a477-924e5cef56d8\") " pod="openstack/configure-network-edpm-deployment-openstack-edpm-ipam-qj5wr" Dec 13 07:24:05 crc kubenswrapper[4971]: I1213 07:24:05.105465 4971 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/edef63aa-a18d-4591-a477-924e5cef56d8-ssh-key\") pod \"configure-network-edpm-deployment-openstack-edpm-ipam-qj5wr\" (UID: \"edef63aa-a18d-4591-a477-924e5cef56d8\") " pod="openstack/configure-network-edpm-deployment-openstack-edpm-ipam-qj5wr" Dec 13 07:24:05 crc kubenswrapper[4971]: I1213 07:24:05.207858 4971 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-7vm5c\" (UniqueName: \"kubernetes.io/projected/edef63aa-a18d-4591-a477-924e5cef56d8-kube-api-access-7vm5c\") pod \"configure-network-edpm-deployment-openstack-edpm-ipam-qj5wr\" (UID: \"edef63aa-a18d-4591-a477-924e5cef56d8\") " pod="openstack/configure-network-edpm-deployment-openstack-edpm-ipam-qj5wr" Dec 13 07:24:05 crc kubenswrapper[4971]: I1213 07:24:05.207973 4971 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/edef63aa-a18d-4591-a477-924e5cef56d8-ssh-key\") pod \"configure-network-edpm-deployment-openstack-edpm-ipam-qj5wr\" (UID: \"edef63aa-a18d-4591-a477-924e5cef56d8\") " pod="openstack/configure-network-edpm-deployment-openstack-edpm-ipam-qj5wr" Dec 13 07:24:05 crc kubenswrapper[4971]: I1213 07:24:05.208099 4971 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/edef63aa-a18d-4591-a477-924e5cef56d8-inventory\") pod \"configure-network-edpm-deployment-openstack-edpm-ipam-qj5wr\" (UID: \"edef63aa-a18d-4591-a477-924e5cef56d8\") " pod="openstack/configure-network-edpm-deployment-openstack-edpm-ipam-qj5wr" Dec 13 07:24:05 crc kubenswrapper[4971]: I1213 07:24:05.214902 4971 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/edef63aa-a18d-4591-a477-924e5cef56d8-inventory\") pod \"configure-network-edpm-deployment-openstack-edpm-ipam-qj5wr\" (UID: \"edef63aa-a18d-4591-a477-924e5cef56d8\") " pod="openstack/configure-network-edpm-deployment-openstack-edpm-ipam-qj5wr" Dec 13 07:24:05 crc kubenswrapper[4971]: I1213 07:24:05.215449 4971 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/edef63aa-a18d-4591-a477-924e5cef56d8-ssh-key\") pod \"configure-network-edpm-deployment-openstack-edpm-ipam-qj5wr\" (UID: \"edef63aa-a18d-4591-a477-924e5cef56d8\") " pod="openstack/configure-network-edpm-deployment-openstack-edpm-ipam-qj5wr" Dec 13 07:24:05 crc kubenswrapper[4971]: I1213 07:24:05.241959 4971 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-7vm5c\" (UniqueName: \"kubernetes.io/projected/edef63aa-a18d-4591-a477-924e5cef56d8-kube-api-access-7vm5c\") pod \"configure-network-edpm-deployment-openstack-edpm-ipam-qj5wr\" (UID: \"edef63aa-a18d-4591-a477-924e5cef56d8\") " pod="openstack/configure-network-edpm-deployment-openstack-edpm-ipam-qj5wr" Dec 13 07:24:05 crc kubenswrapper[4971]: I1213 07:24:05.303255 4971 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/configure-network-edpm-deployment-openstack-edpm-ipam-qj5wr" Dec 13 07:24:05 crc kubenswrapper[4971]: I1213 07:24:05.674757 4971 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/configure-network-edpm-deployment-openstack-edpm-ipam-qj5wr"] Dec 13 07:24:05 crc kubenswrapper[4971]: W1213 07:24:05.681005 4971 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podedef63aa_a18d_4591_a477_924e5cef56d8.slice/crio-33cc89611a94a06c47c3084f0fc6e0e327ccba0d46d7d6375ff44c3035736cb6 WatchSource:0}: Error finding container 33cc89611a94a06c47c3084f0fc6e0e327ccba0d46d7d6375ff44c3035736cb6: Status 404 returned error can't find the container with id 33cc89611a94a06c47c3084f0fc6e0e327ccba0d46d7d6375ff44c3035736cb6 Dec 13 07:24:05 crc kubenswrapper[4971]: I1213 07:24:05.805957 4971 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/configure-network-edpm-deployment-openstack-edpm-ipam-qj5wr" event={"ID":"edef63aa-a18d-4591-a477-924e5cef56d8","Type":"ContainerStarted","Data":"33cc89611a94a06c47c3084f0fc6e0e327ccba0d46d7d6375ff44c3035736cb6"} Dec 13 07:24:09 crc kubenswrapper[4971]: I1213 07:24:09.269678 4971 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-operators-sbkvv" Dec 13 07:24:09 crc kubenswrapper[4971]: I1213 07:24:09.270170 4971 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-operators-sbkvv" Dec 13 07:24:09 crc kubenswrapper[4971]: I1213 07:24:09.344149 4971 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-operators-sbkvv" Dec 13 07:24:09 crc kubenswrapper[4971]: I1213 07:24:09.954910 4971 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-operators-sbkvv" Dec 13 07:24:10 crc kubenswrapper[4971]: I1213 07:24:10.141985 4971 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-sbkvv"] Dec 13 07:24:11 crc kubenswrapper[4971]: I1213 07:24:11.643374 4971 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/certified-operators-587w4" Dec 13 07:24:11 crc kubenswrapper[4971]: I1213 07:24:11.643695 4971 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/certified-operators-587w4" Dec 13 07:24:11 crc kubenswrapper[4971]: I1213 07:24:11.721704 4971 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/certified-operators-587w4" Dec 13 07:24:11 crc kubenswrapper[4971]: I1213 07:24:11.876214 4971 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-operators-sbkvv" podUID="fbe986fc-eacf-42a4-98f6-288d943470dc" containerName="registry-server" containerID="cri-o://51d90418c1a42b289e90ac8160fe1831853dd7604a28213a37f478e1aecccee5" gracePeriod=2 Dec 13 07:24:11 crc kubenswrapper[4971]: I1213 07:24:11.967981 4971 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/certified-operators-587w4" Dec 13 07:24:12 crc kubenswrapper[4971]: I1213 07:24:12.551400 4971 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-587w4"] Dec 13 07:24:13 crc kubenswrapper[4971]: I1213 07:24:13.930179 4971 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/certified-operators-587w4" podUID="3e083b0a-4cc8-4a7e-83bf-956166f10069" containerName="registry-server" containerID="cri-o://5f7cd20e93b0677908dfeefe7527764d6bad5619e32d02cbe2cd660ad8713972" gracePeriod=2 Dec 13 07:24:15 crc kubenswrapper[4971]: I1213 07:24:15.386183 4971 scope.go:117] "RemoveContainer" containerID="44cc99feda7a93b0ad3cddca4023a75dbf6b44a85ac3a407d60f674701f50cab" Dec 13 07:24:16 crc kubenswrapper[4971]: I1213 07:24:16.740336 4971 prober.go:107] "Probe failed" probeType="Liveness" pod="openstack/ceilometer-0" podUID="7cfb4dbf-cc55-48b7-b50c-65d6d4ae64ba" containerName="ceilometer-central-agent" probeResult="failure" output="command timed out" Dec 13 07:24:17 crc kubenswrapper[4971]: I1213 07:24:17.758032 4971 scope.go:117] "RemoveContainer" containerID="3dc5bacbb497492ceb200a8f4079f1fe1919ec2f11414de66931136305d0fa67" Dec 13 07:24:17 crc kubenswrapper[4971]: I1213 07:24:17.825355 4971 scope.go:117] "RemoveContainer" containerID="9509f0f8e726726fa1500e026105ca7cecd8caa43de3dcf3443517afda368ef9" Dec 13 07:24:17 crc kubenswrapper[4971]: I1213 07:24:17.916610 4971 scope.go:117] "RemoveContainer" containerID="0035d5bc04fb0f44fabb4f7127f141f9fb4c42242812ceedc881646754b08540" Dec 13 07:24:17 crc kubenswrapper[4971]: I1213 07:24:17.948295 4971 scope.go:117] "RemoveContainer" containerID="ab8949e28c39670644fadb5e1c3465a991ef49397ef4b090a05d46af40d1b3a9" Dec 13 07:24:17 crc kubenswrapper[4971]: I1213 07:24:17.987961 4971 scope.go:117] "RemoveContainer" containerID="cb21d277a6ad833347a698fee9c98b7876b7a07a2de9d2e3630e0994cda10e94" Dec 13 07:24:18 crc kubenswrapper[4971]: I1213 07:24:18.456355 4971 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-operators-sbkvv_fbe986fc-eacf-42a4-98f6-288d943470dc/registry-server/0.log" Dec 13 07:24:18 crc kubenswrapper[4971]: I1213 07:24:18.457502 4971 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-sbkvv" Dec 13 07:24:18 crc kubenswrapper[4971]: I1213 07:24:18.500772 4971 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-operators-sbkvv_fbe986fc-eacf-42a4-98f6-288d943470dc/registry-server/0.log" Dec 13 07:24:18 crc kubenswrapper[4971]: I1213 07:24:18.505284 4971 generic.go:334] "Generic (PLEG): container finished" podID="fbe986fc-eacf-42a4-98f6-288d943470dc" containerID="51d90418c1a42b289e90ac8160fe1831853dd7604a28213a37f478e1aecccee5" exitCode=137 Dec 13 07:24:18 crc kubenswrapper[4971]: I1213 07:24:18.505363 4971 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-sbkvv" event={"ID":"fbe986fc-eacf-42a4-98f6-288d943470dc","Type":"ContainerDied","Data":"51d90418c1a42b289e90ac8160fe1831853dd7604a28213a37f478e1aecccee5"} Dec 13 07:24:18 crc kubenswrapper[4971]: I1213 07:24:18.505404 4971 scope.go:117] "RemoveContainer" containerID="51d90418c1a42b289e90ac8160fe1831853dd7604a28213a37f478e1aecccee5" Dec 13 07:24:18 crc kubenswrapper[4971]: I1213 07:24:18.525017 4971 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/fbe986fc-eacf-42a4-98f6-288d943470dc-utilities\") pod \"fbe986fc-eacf-42a4-98f6-288d943470dc\" (UID: \"fbe986fc-eacf-42a4-98f6-288d943470dc\") " Dec 13 07:24:18 crc kubenswrapper[4971]: I1213 07:24:18.525219 4971 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/fbe986fc-eacf-42a4-98f6-288d943470dc-catalog-content\") pod \"fbe986fc-eacf-42a4-98f6-288d943470dc\" (UID: \"fbe986fc-eacf-42a4-98f6-288d943470dc\") " Dec 13 07:24:18 crc kubenswrapper[4971]: I1213 07:24:18.525241 4971 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-z688t\" (UniqueName: \"kubernetes.io/projected/fbe986fc-eacf-42a4-98f6-288d943470dc-kube-api-access-z688t\") pod \"fbe986fc-eacf-42a4-98f6-288d943470dc\" (UID: \"fbe986fc-eacf-42a4-98f6-288d943470dc\") " Dec 13 07:24:18 crc kubenswrapper[4971]: I1213 07:24:18.525810 4971 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/fbe986fc-eacf-42a4-98f6-288d943470dc-utilities" (OuterVolumeSpecName: "utilities") pod "fbe986fc-eacf-42a4-98f6-288d943470dc" (UID: "fbe986fc-eacf-42a4-98f6-288d943470dc"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 13 07:24:18 crc kubenswrapper[4971]: I1213 07:24:18.531823 4971 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/fbe986fc-eacf-42a4-98f6-288d943470dc-kube-api-access-z688t" (OuterVolumeSpecName: "kube-api-access-z688t") pod "fbe986fc-eacf-42a4-98f6-288d943470dc" (UID: "fbe986fc-eacf-42a4-98f6-288d943470dc"). InnerVolumeSpecName "kube-api-access-z688t". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 13 07:24:18 crc kubenswrapper[4971]: I1213 07:24:18.627569 4971 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/fbe986fc-eacf-42a4-98f6-288d943470dc-utilities\") on node \"crc\" DevicePath \"\"" Dec 13 07:24:18 crc kubenswrapper[4971]: I1213 07:24:18.627602 4971 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-z688t\" (UniqueName: \"kubernetes.io/projected/fbe986fc-eacf-42a4-98f6-288d943470dc-kube-api-access-z688t\") on node \"crc\" DevicePath \"\"" Dec 13 07:24:18 crc kubenswrapper[4971]: I1213 07:24:18.638976 4971 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/fbe986fc-eacf-42a4-98f6-288d943470dc-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "fbe986fc-eacf-42a4-98f6-288d943470dc" (UID: "fbe986fc-eacf-42a4-98f6-288d943470dc"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 13 07:24:18 crc kubenswrapper[4971]: I1213 07:24:18.641047 4971 scope.go:117] "RemoveContainer" containerID="aa99e3a9c2358daf739038ac0fa7be8391fe24da9e5dd84f2f83edd9f63c9901" Dec 13 07:24:18 crc kubenswrapper[4971]: I1213 07:24:18.687488 4971 scope.go:117] "RemoveContainer" containerID="b95450ada02de13d4d8f2422676cf14a569d4f84c2be200fa6be72fcea571096" Dec 13 07:24:18 crc kubenswrapper[4971]: I1213 07:24:18.729245 4971 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/fbe986fc-eacf-42a4-98f6-288d943470dc-catalog-content\") on node \"crc\" DevicePath \"\"" Dec 13 07:24:19 crc kubenswrapper[4971]: I1213 07:24:19.241022 4971 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_certified-operators-587w4_3e083b0a-4cc8-4a7e-83bf-956166f10069/registry-server/0.log" Dec 13 07:24:19 crc kubenswrapper[4971]: I1213 07:24:19.242502 4971 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-587w4" Dec 13 07:24:19 crc kubenswrapper[4971]: I1213 07:24:19.339693 4971 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-lhfh8\" (UniqueName: \"kubernetes.io/projected/3e083b0a-4cc8-4a7e-83bf-956166f10069-kube-api-access-lhfh8\") pod \"3e083b0a-4cc8-4a7e-83bf-956166f10069\" (UID: \"3e083b0a-4cc8-4a7e-83bf-956166f10069\") " Dec 13 07:24:19 crc kubenswrapper[4971]: I1213 07:24:19.339805 4971 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/3e083b0a-4cc8-4a7e-83bf-956166f10069-catalog-content\") pod \"3e083b0a-4cc8-4a7e-83bf-956166f10069\" (UID: \"3e083b0a-4cc8-4a7e-83bf-956166f10069\") " Dec 13 07:24:19 crc kubenswrapper[4971]: I1213 07:24:19.339908 4971 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/3e083b0a-4cc8-4a7e-83bf-956166f10069-utilities\") pod \"3e083b0a-4cc8-4a7e-83bf-956166f10069\" (UID: \"3e083b0a-4cc8-4a7e-83bf-956166f10069\") " Dec 13 07:24:19 crc kubenswrapper[4971]: I1213 07:24:19.340768 4971 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/3e083b0a-4cc8-4a7e-83bf-956166f10069-utilities" (OuterVolumeSpecName: "utilities") pod "3e083b0a-4cc8-4a7e-83bf-956166f10069" (UID: "3e083b0a-4cc8-4a7e-83bf-956166f10069"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 13 07:24:19 crc kubenswrapper[4971]: I1213 07:24:19.344131 4971 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/3e083b0a-4cc8-4a7e-83bf-956166f10069-kube-api-access-lhfh8" (OuterVolumeSpecName: "kube-api-access-lhfh8") pod "3e083b0a-4cc8-4a7e-83bf-956166f10069" (UID: "3e083b0a-4cc8-4a7e-83bf-956166f10069"). InnerVolumeSpecName "kube-api-access-lhfh8". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 13 07:24:19 crc kubenswrapper[4971]: I1213 07:24:19.392305 4971 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/3e083b0a-4cc8-4a7e-83bf-956166f10069-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "3e083b0a-4cc8-4a7e-83bf-956166f10069" (UID: "3e083b0a-4cc8-4a7e-83bf-956166f10069"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 13 07:24:19 crc kubenswrapper[4971]: I1213 07:24:19.441577 4971 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-lhfh8\" (UniqueName: \"kubernetes.io/projected/3e083b0a-4cc8-4a7e-83bf-956166f10069-kube-api-access-lhfh8\") on node \"crc\" DevicePath \"\"" Dec 13 07:24:19 crc kubenswrapper[4971]: I1213 07:24:19.441610 4971 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/3e083b0a-4cc8-4a7e-83bf-956166f10069-catalog-content\") on node \"crc\" DevicePath \"\"" Dec 13 07:24:19 crc kubenswrapper[4971]: I1213 07:24:19.441619 4971 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/3e083b0a-4cc8-4a7e-83bf-956166f10069-utilities\") on node \"crc\" DevicePath \"\"" Dec 13 07:24:19 crc kubenswrapper[4971]: I1213 07:24:19.526162 4971 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-sbkvv" event={"ID":"fbe986fc-eacf-42a4-98f6-288d943470dc","Type":"ContainerDied","Data":"24d381ef9b49217881442c4dc31812e0cee8fbbbffc9bbf4320e59fe18573b7c"} Dec 13 07:24:19 crc kubenswrapper[4971]: I1213 07:24:19.526267 4971 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-sbkvv" Dec 13 07:24:19 crc kubenswrapper[4971]: I1213 07:24:19.528840 4971 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/configure-network-edpm-deployment-openstack-edpm-ipam-qj5wr" event={"ID":"edef63aa-a18d-4591-a477-924e5cef56d8","Type":"ContainerStarted","Data":"1b8738b60e78e46c6c3201e6dc3104d9218dd08393b34b5a75578292a38c23c0"} Dec 13 07:24:19 crc kubenswrapper[4971]: I1213 07:24:19.543458 4971 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_certified-operators-587w4_3e083b0a-4cc8-4a7e-83bf-956166f10069/registry-server/0.log" Dec 13 07:24:19 crc kubenswrapper[4971]: I1213 07:24:19.548847 4971 generic.go:334] "Generic (PLEG): container finished" podID="3e083b0a-4cc8-4a7e-83bf-956166f10069" containerID="5f7cd20e93b0677908dfeefe7527764d6bad5619e32d02cbe2cd660ad8713972" exitCode=137 Dec 13 07:24:19 crc kubenswrapper[4971]: I1213 07:24:19.548916 4971 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-587w4" event={"ID":"3e083b0a-4cc8-4a7e-83bf-956166f10069","Type":"ContainerDied","Data":"5f7cd20e93b0677908dfeefe7527764d6bad5619e32d02cbe2cd660ad8713972"} Dec 13 07:24:19 crc kubenswrapper[4971]: I1213 07:24:19.548946 4971 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-587w4" event={"ID":"3e083b0a-4cc8-4a7e-83bf-956166f10069","Type":"ContainerDied","Data":"828302ef4afbae64996e0d7090e90ba4e76dfc5c1e319a88440111bde7ea52d7"} Dec 13 07:24:19 crc kubenswrapper[4971]: I1213 07:24:19.548977 4971 scope.go:117] "RemoveContainer" containerID="5f7cd20e93b0677908dfeefe7527764d6bad5619e32d02cbe2cd660ad8713972" Dec 13 07:24:19 crc kubenswrapper[4971]: I1213 07:24:19.549167 4971 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-587w4" Dec 13 07:24:19 crc kubenswrapper[4971]: I1213 07:24:19.561212 4971 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/configure-network-edpm-deployment-openstack-edpm-ipam-qj5wr" podStartSLOduration=2.165328687 podStartE2EDuration="15.56119254s" podCreationTimestamp="2025-12-13 07:24:04 +0000 UTC" firstStartedPulling="2025-12-13 07:24:05.684379539 +0000 UTC m=+2102.288788987" lastFinishedPulling="2025-12-13 07:24:19.080243392 +0000 UTC m=+2115.684652840" observedRunningTime="2025-12-13 07:24:19.553485877 +0000 UTC m=+2116.157895335" watchObservedRunningTime="2025-12-13 07:24:19.56119254 +0000 UTC m=+2116.165601988" Dec 13 07:24:19 crc kubenswrapper[4971]: I1213 07:24:19.581160 4971 scope.go:117] "RemoveContainer" containerID="e8f8bda0e7d32492e4fdf48d04e708279803e50d1ade1d6b080adf6af134dd0e" Dec 13 07:24:19 crc kubenswrapper[4971]: I1213 07:24:19.602240 4971 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-sbkvv"] Dec 13 07:24:19 crc kubenswrapper[4971]: I1213 07:24:19.602298 4971 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-operators-sbkvv"] Dec 13 07:24:19 crc kubenswrapper[4971]: I1213 07:24:19.614182 4971 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-587w4"] Dec 13 07:24:19 crc kubenswrapper[4971]: I1213 07:24:19.614398 4971 scope.go:117] "RemoveContainer" containerID="bac558e7b4872b3dec8fb88e42d3a2ee993882f4a4d572dee1e4f4ba767591cb" Dec 13 07:24:19 crc kubenswrapper[4971]: I1213 07:24:19.622659 4971 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/certified-operators-587w4"] Dec 13 07:24:19 crc kubenswrapper[4971]: I1213 07:24:19.643027 4971 scope.go:117] "RemoveContainer" containerID="5f7cd20e93b0677908dfeefe7527764d6bad5619e32d02cbe2cd660ad8713972" Dec 13 07:24:19 crc kubenswrapper[4971]: E1213 07:24:19.643554 4971 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"5f7cd20e93b0677908dfeefe7527764d6bad5619e32d02cbe2cd660ad8713972\": container with ID starting with 5f7cd20e93b0677908dfeefe7527764d6bad5619e32d02cbe2cd660ad8713972 not found: ID does not exist" containerID="5f7cd20e93b0677908dfeefe7527764d6bad5619e32d02cbe2cd660ad8713972" Dec 13 07:24:19 crc kubenswrapper[4971]: I1213 07:24:19.643596 4971 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"5f7cd20e93b0677908dfeefe7527764d6bad5619e32d02cbe2cd660ad8713972"} err="failed to get container status \"5f7cd20e93b0677908dfeefe7527764d6bad5619e32d02cbe2cd660ad8713972\": rpc error: code = NotFound desc = could not find container \"5f7cd20e93b0677908dfeefe7527764d6bad5619e32d02cbe2cd660ad8713972\": container with ID starting with 5f7cd20e93b0677908dfeefe7527764d6bad5619e32d02cbe2cd660ad8713972 not found: ID does not exist" Dec 13 07:24:19 crc kubenswrapper[4971]: I1213 07:24:19.643624 4971 scope.go:117] "RemoveContainer" containerID="e8f8bda0e7d32492e4fdf48d04e708279803e50d1ade1d6b080adf6af134dd0e" Dec 13 07:24:19 crc kubenswrapper[4971]: E1213 07:24:19.643951 4971 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"e8f8bda0e7d32492e4fdf48d04e708279803e50d1ade1d6b080adf6af134dd0e\": container with ID starting with e8f8bda0e7d32492e4fdf48d04e708279803e50d1ade1d6b080adf6af134dd0e not found: ID does not exist" containerID="e8f8bda0e7d32492e4fdf48d04e708279803e50d1ade1d6b080adf6af134dd0e" Dec 13 07:24:19 crc kubenswrapper[4971]: I1213 07:24:19.644037 4971 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"e8f8bda0e7d32492e4fdf48d04e708279803e50d1ade1d6b080adf6af134dd0e"} err="failed to get container status \"e8f8bda0e7d32492e4fdf48d04e708279803e50d1ade1d6b080adf6af134dd0e\": rpc error: code = NotFound desc = could not find container \"e8f8bda0e7d32492e4fdf48d04e708279803e50d1ade1d6b080adf6af134dd0e\": container with ID starting with e8f8bda0e7d32492e4fdf48d04e708279803e50d1ade1d6b080adf6af134dd0e not found: ID does not exist" Dec 13 07:24:19 crc kubenswrapper[4971]: I1213 07:24:19.644068 4971 scope.go:117] "RemoveContainer" containerID="bac558e7b4872b3dec8fb88e42d3a2ee993882f4a4d572dee1e4f4ba767591cb" Dec 13 07:24:19 crc kubenswrapper[4971]: E1213 07:24:19.644449 4971 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"bac558e7b4872b3dec8fb88e42d3a2ee993882f4a4d572dee1e4f4ba767591cb\": container with ID starting with bac558e7b4872b3dec8fb88e42d3a2ee993882f4a4d572dee1e4f4ba767591cb not found: ID does not exist" containerID="bac558e7b4872b3dec8fb88e42d3a2ee993882f4a4d572dee1e4f4ba767591cb" Dec 13 07:24:19 crc kubenswrapper[4971]: I1213 07:24:19.644476 4971 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"bac558e7b4872b3dec8fb88e42d3a2ee993882f4a4d572dee1e4f4ba767591cb"} err="failed to get container status \"bac558e7b4872b3dec8fb88e42d3a2ee993882f4a4d572dee1e4f4ba767591cb\": rpc error: code = NotFound desc = could not find container \"bac558e7b4872b3dec8fb88e42d3a2ee993882f4a4d572dee1e4f4ba767591cb\": container with ID starting with bac558e7b4872b3dec8fb88e42d3a2ee993882f4a4d572dee1e4f4ba767591cb not found: ID does not exist" Dec 13 07:24:19 crc kubenswrapper[4971]: I1213 07:24:19.778400 4971 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="3e083b0a-4cc8-4a7e-83bf-956166f10069" path="/var/lib/kubelet/pods/3e083b0a-4cc8-4a7e-83bf-956166f10069/volumes" Dec 13 07:24:19 crc kubenswrapper[4971]: I1213 07:24:19.779038 4971 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="fbe986fc-eacf-42a4-98f6-288d943470dc" path="/var/lib/kubelet/pods/fbe986fc-eacf-42a4-98f6-288d943470dc/volumes" Dec 13 07:24:27 crc kubenswrapper[4971]: I1213 07:24:27.038634 4971 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-cell0-conductor-db-sync-g9t27"] Dec 13 07:24:27 crc kubenswrapper[4971]: I1213 07:24:27.047849 4971 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-cell0-conductor-db-sync-g9t27"] Dec 13 07:24:27 crc kubenswrapper[4971]: I1213 07:24:27.778449 4971 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="31dddcf3-380b-4003-a766-88d4f5bc00de" path="/var/lib/kubelet/pods/31dddcf3-380b-4003-a766-88d4f5bc00de/volumes" Dec 13 07:24:50 crc kubenswrapper[4971]: I1213 07:24:50.776318 4971 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-cell0-cell-mapping-rgjbg"] Dec 13 07:24:50 crc kubenswrapper[4971]: I1213 07:24:50.796279 4971 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-cell0-cell-mapping-rgjbg"] Dec 13 07:24:51 crc kubenswrapper[4971]: I1213 07:24:51.785962 4971 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="74fcba46-f364-4c8a-a4e3-ace14451d1f7" path="/var/lib/kubelet/pods/74fcba46-f364-4c8a-a4e3-ace14451d1f7/volumes" Dec 13 07:24:54 crc kubenswrapper[4971]: I1213 07:24:54.037962 4971 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-cell1-conductor-db-sync-kvw2t"] Dec 13 07:24:54 crc kubenswrapper[4971]: I1213 07:24:54.046390 4971 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-cell1-conductor-db-sync-kvw2t"] Dec 13 07:24:55 crc kubenswrapper[4971]: I1213 07:24:55.784770 4971 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="6c6829bb-25ee-4102-8335-b747bdbe6121" path="/var/lib/kubelet/pods/6c6829bb-25ee-4102-8335-b747bdbe6121/volumes" Dec 13 07:24:57 crc kubenswrapper[4971]: I1213 07:24:57.129630 4971 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-marketplace-b4xjx"] Dec 13 07:24:57 crc kubenswrapper[4971]: E1213 07:24:57.130232 4971 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="fbe986fc-eacf-42a4-98f6-288d943470dc" containerName="extract-content" Dec 13 07:24:57 crc kubenswrapper[4971]: I1213 07:24:57.130253 4971 state_mem.go:107] "Deleted CPUSet assignment" podUID="fbe986fc-eacf-42a4-98f6-288d943470dc" containerName="extract-content" Dec 13 07:24:57 crc kubenswrapper[4971]: E1213 07:24:57.130299 4971 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="3e083b0a-4cc8-4a7e-83bf-956166f10069" containerName="extract-content" Dec 13 07:24:57 crc kubenswrapper[4971]: I1213 07:24:57.130313 4971 state_mem.go:107] "Deleted CPUSet assignment" podUID="3e083b0a-4cc8-4a7e-83bf-956166f10069" containerName="extract-content" Dec 13 07:24:57 crc kubenswrapper[4971]: E1213 07:24:57.130331 4971 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="fbe986fc-eacf-42a4-98f6-288d943470dc" containerName="extract-utilities" Dec 13 07:24:57 crc kubenswrapper[4971]: I1213 07:24:57.130345 4971 state_mem.go:107] "Deleted CPUSet assignment" podUID="fbe986fc-eacf-42a4-98f6-288d943470dc" containerName="extract-utilities" Dec 13 07:24:57 crc kubenswrapper[4971]: E1213 07:24:57.130366 4971 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="3e083b0a-4cc8-4a7e-83bf-956166f10069" containerName="registry-server" Dec 13 07:24:57 crc kubenswrapper[4971]: I1213 07:24:57.130377 4971 state_mem.go:107] "Deleted CPUSet assignment" podUID="3e083b0a-4cc8-4a7e-83bf-956166f10069" containerName="registry-server" Dec 13 07:24:57 crc kubenswrapper[4971]: E1213 07:24:57.130400 4971 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="fbe986fc-eacf-42a4-98f6-288d943470dc" containerName="registry-server" Dec 13 07:24:57 crc kubenswrapper[4971]: I1213 07:24:57.130412 4971 state_mem.go:107] "Deleted CPUSet assignment" podUID="fbe986fc-eacf-42a4-98f6-288d943470dc" containerName="registry-server" Dec 13 07:24:57 crc kubenswrapper[4971]: E1213 07:24:57.130432 4971 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="3e083b0a-4cc8-4a7e-83bf-956166f10069" containerName="extract-utilities" Dec 13 07:24:57 crc kubenswrapper[4971]: I1213 07:24:57.130445 4971 state_mem.go:107] "Deleted CPUSet assignment" podUID="3e083b0a-4cc8-4a7e-83bf-956166f10069" containerName="extract-utilities" Dec 13 07:24:57 crc kubenswrapper[4971]: I1213 07:24:57.130792 4971 memory_manager.go:354] "RemoveStaleState removing state" podUID="3e083b0a-4cc8-4a7e-83bf-956166f10069" containerName="registry-server" Dec 13 07:24:57 crc kubenswrapper[4971]: I1213 07:24:57.130837 4971 memory_manager.go:354] "RemoveStaleState removing state" podUID="fbe986fc-eacf-42a4-98f6-288d943470dc" containerName="registry-server" Dec 13 07:24:57 crc kubenswrapper[4971]: I1213 07:24:57.132991 4971 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-b4xjx" Dec 13 07:24:57 crc kubenswrapper[4971]: I1213 07:24:57.139644 4971 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-b4xjx"] Dec 13 07:24:57 crc kubenswrapper[4971]: I1213 07:24:57.140000 4971 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/088781cb-6d76-4036-b6fe-44a74dc0b88e-catalog-content\") pod \"redhat-marketplace-b4xjx\" (UID: \"088781cb-6d76-4036-b6fe-44a74dc0b88e\") " pod="openshift-marketplace/redhat-marketplace-b4xjx" Dec 13 07:24:57 crc kubenswrapper[4971]: I1213 07:24:57.140049 4971 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/088781cb-6d76-4036-b6fe-44a74dc0b88e-utilities\") pod \"redhat-marketplace-b4xjx\" (UID: \"088781cb-6d76-4036-b6fe-44a74dc0b88e\") " pod="openshift-marketplace/redhat-marketplace-b4xjx" Dec 13 07:24:57 crc kubenswrapper[4971]: I1213 07:24:57.140207 4971 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-296dw\" (UniqueName: \"kubernetes.io/projected/088781cb-6d76-4036-b6fe-44a74dc0b88e-kube-api-access-296dw\") pod \"redhat-marketplace-b4xjx\" (UID: \"088781cb-6d76-4036-b6fe-44a74dc0b88e\") " pod="openshift-marketplace/redhat-marketplace-b4xjx" Dec 13 07:24:57 crc kubenswrapper[4971]: I1213 07:24:57.242071 4971 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/088781cb-6d76-4036-b6fe-44a74dc0b88e-catalog-content\") pod \"redhat-marketplace-b4xjx\" (UID: \"088781cb-6d76-4036-b6fe-44a74dc0b88e\") " pod="openshift-marketplace/redhat-marketplace-b4xjx" Dec 13 07:24:57 crc kubenswrapper[4971]: I1213 07:24:57.242127 4971 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/088781cb-6d76-4036-b6fe-44a74dc0b88e-utilities\") pod \"redhat-marketplace-b4xjx\" (UID: \"088781cb-6d76-4036-b6fe-44a74dc0b88e\") " pod="openshift-marketplace/redhat-marketplace-b4xjx" Dec 13 07:24:57 crc kubenswrapper[4971]: I1213 07:24:57.242215 4971 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-296dw\" (UniqueName: \"kubernetes.io/projected/088781cb-6d76-4036-b6fe-44a74dc0b88e-kube-api-access-296dw\") pod \"redhat-marketplace-b4xjx\" (UID: \"088781cb-6d76-4036-b6fe-44a74dc0b88e\") " pod="openshift-marketplace/redhat-marketplace-b4xjx" Dec 13 07:24:57 crc kubenswrapper[4971]: I1213 07:24:57.242660 4971 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/088781cb-6d76-4036-b6fe-44a74dc0b88e-catalog-content\") pod \"redhat-marketplace-b4xjx\" (UID: \"088781cb-6d76-4036-b6fe-44a74dc0b88e\") " pod="openshift-marketplace/redhat-marketplace-b4xjx" Dec 13 07:24:57 crc kubenswrapper[4971]: I1213 07:24:57.243031 4971 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/088781cb-6d76-4036-b6fe-44a74dc0b88e-utilities\") pod \"redhat-marketplace-b4xjx\" (UID: \"088781cb-6d76-4036-b6fe-44a74dc0b88e\") " pod="openshift-marketplace/redhat-marketplace-b4xjx" Dec 13 07:24:57 crc kubenswrapper[4971]: I1213 07:24:57.268787 4971 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-296dw\" (UniqueName: \"kubernetes.io/projected/088781cb-6d76-4036-b6fe-44a74dc0b88e-kube-api-access-296dw\") pod \"redhat-marketplace-b4xjx\" (UID: \"088781cb-6d76-4036-b6fe-44a74dc0b88e\") " pod="openshift-marketplace/redhat-marketplace-b4xjx" Dec 13 07:24:57 crc kubenswrapper[4971]: I1213 07:24:57.461144 4971 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-b4xjx" Dec 13 07:24:57 crc kubenswrapper[4971]: I1213 07:24:57.966287 4971 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-b4xjx"] Dec 13 07:24:58 crc kubenswrapper[4971]: I1213 07:24:58.211225 4971 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-b4xjx" event={"ID":"088781cb-6d76-4036-b6fe-44a74dc0b88e","Type":"ContainerStarted","Data":"487d08fab5a97d8f55c9955018b903150c8d4068716d1406e27bcb6dc3fc75f8"} Dec 13 07:24:58 crc kubenswrapper[4971]: I1213 07:24:58.211263 4971 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-b4xjx" event={"ID":"088781cb-6d76-4036-b6fe-44a74dc0b88e","Type":"ContainerStarted","Data":"55f1e156a2960c17d36213a4df571764dd7a1701ff6e213c172869331272d603"} Dec 13 07:24:59 crc kubenswrapper[4971]: I1213 07:24:59.222446 4971 generic.go:334] "Generic (PLEG): container finished" podID="088781cb-6d76-4036-b6fe-44a74dc0b88e" containerID="487d08fab5a97d8f55c9955018b903150c8d4068716d1406e27bcb6dc3fc75f8" exitCode=0 Dec 13 07:24:59 crc kubenswrapper[4971]: I1213 07:24:59.222760 4971 generic.go:334] "Generic (PLEG): container finished" podID="088781cb-6d76-4036-b6fe-44a74dc0b88e" containerID="71e378a86e610e89278282d167c2ec0c6e8a9c70cbb2290358c0fc757b2aadd5" exitCode=0 Dec 13 07:24:59 crc kubenswrapper[4971]: I1213 07:24:59.222556 4971 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-b4xjx" event={"ID":"088781cb-6d76-4036-b6fe-44a74dc0b88e","Type":"ContainerDied","Data":"487d08fab5a97d8f55c9955018b903150c8d4068716d1406e27bcb6dc3fc75f8"} Dec 13 07:24:59 crc kubenswrapper[4971]: I1213 07:24:59.222801 4971 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-b4xjx" event={"ID":"088781cb-6d76-4036-b6fe-44a74dc0b88e","Type":"ContainerDied","Data":"71e378a86e610e89278282d167c2ec0c6e8a9c70cbb2290358c0fc757b2aadd5"} Dec 13 07:25:00 crc kubenswrapper[4971]: I1213 07:25:00.235111 4971 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-b4xjx" event={"ID":"088781cb-6d76-4036-b6fe-44a74dc0b88e","Type":"ContainerStarted","Data":"fbbb0df4c2f227da6b3301afc069e02413cab194de1f64ad81ef2b572fb4aa0c"} Dec 13 07:25:00 crc kubenswrapper[4971]: I1213 07:25:00.262347 4971 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-marketplace-b4xjx" podStartSLOduration=1.693771329 podStartE2EDuration="3.262327585s" podCreationTimestamp="2025-12-13 07:24:57 +0000 UTC" firstStartedPulling="2025-12-13 07:24:58.213060775 +0000 UTC m=+2154.817470233" lastFinishedPulling="2025-12-13 07:24:59.781617001 +0000 UTC m=+2156.386026489" observedRunningTime="2025-12-13 07:25:00.25736825 +0000 UTC m=+2156.861777768" watchObservedRunningTime="2025-12-13 07:25:00.262327585 +0000 UTC m=+2156.866737043" Dec 13 07:25:07 crc kubenswrapper[4971]: I1213 07:25:07.461687 4971 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-marketplace-b4xjx" Dec 13 07:25:07 crc kubenswrapper[4971]: I1213 07:25:07.462292 4971 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-marketplace-b4xjx" Dec 13 07:25:07 crc kubenswrapper[4971]: I1213 07:25:07.536618 4971 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-marketplace-b4xjx" Dec 13 07:25:08 crc kubenswrapper[4971]: I1213 07:25:08.382451 4971 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-marketplace-b4xjx" Dec 13 07:25:08 crc kubenswrapper[4971]: I1213 07:25:08.461285 4971 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-b4xjx"] Dec 13 07:25:10 crc kubenswrapper[4971]: I1213 07:25:10.339917 4971 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-marketplace-b4xjx" podUID="088781cb-6d76-4036-b6fe-44a74dc0b88e" containerName="registry-server" containerID="cri-o://fbbb0df4c2f227da6b3301afc069e02413cab194de1f64ad81ef2b572fb4aa0c" gracePeriod=2 Dec 13 07:25:11 crc kubenswrapper[4971]: I1213 07:25:11.267363 4971 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-b4xjx" Dec 13 07:25:11 crc kubenswrapper[4971]: I1213 07:25:11.352121 4971 generic.go:334] "Generic (PLEG): container finished" podID="088781cb-6d76-4036-b6fe-44a74dc0b88e" containerID="fbbb0df4c2f227da6b3301afc069e02413cab194de1f64ad81ef2b572fb4aa0c" exitCode=0 Dec 13 07:25:11 crc kubenswrapper[4971]: I1213 07:25:11.352159 4971 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-b4xjx" event={"ID":"088781cb-6d76-4036-b6fe-44a74dc0b88e","Type":"ContainerDied","Data":"fbbb0df4c2f227da6b3301afc069e02413cab194de1f64ad81ef2b572fb4aa0c"} Dec 13 07:25:11 crc kubenswrapper[4971]: I1213 07:25:11.352184 4971 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-b4xjx" event={"ID":"088781cb-6d76-4036-b6fe-44a74dc0b88e","Type":"ContainerDied","Data":"55f1e156a2960c17d36213a4df571764dd7a1701ff6e213c172869331272d603"} Dec 13 07:25:11 crc kubenswrapper[4971]: I1213 07:25:11.352202 4971 scope.go:117] "RemoveContainer" containerID="fbbb0df4c2f227da6b3301afc069e02413cab194de1f64ad81ef2b572fb4aa0c" Dec 13 07:25:11 crc kubenswrapper[4971]: I1213 07:25:11.353107 4971 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-b4xjx" Dec 13 07:25:11 crc kubenswrapper[4971]: I1213 07:25:11.376854 4971 scope.go:117] "RemoveContainer" containerID="71e378a86e610e89278282d167c2ec0c6e8a9c70cbb2290358c0fc757b2aadd5" Dec 13 07:25:11 crc kubenswrapper[4971]: I1213 07:25:11.407319 4971 scope.go:117] "RemoveContainer" containerID="487d08fab5a97d8f55c9955018b903150c8d4068716d1406e27bcb6dc3fc75f8" Dec 13 07:25:11 crc kubenswrapper[4971]: I1213 07:25:11.431315 4971 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-296dw\" (UniqueName: \"kubernetes.io/projected/088781cb-6d76-4036-b6fe-44a74dc0b88e-kube-api-access-296dw\") pod \"088781cb-6d76-4036-b6fe-44a74dc0b88e\" (UID: \"088781cb-6d76-4036-b6fe-44a74dc0b88e\") " Dec 13 07:25:11 crc kubenswrapper[4971]: I1213 07:25:11.431559 4971 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/088781cb-6d76-4036-b6fe-44a74dc0b88e-catalog-content\") pod \"088781cb-6d76-4036-b6fe-44a74dc0b88e\" (UID: \"088781cb-6d76-4036-b6fe-44a74dc0b88e\") " Dec 13 07:25:11 crc kubenswrapper[4971]: I1213 07:25:11.431646 4971 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/088781cb-6d76-4036-b6fe-44a74dc0b88e-utilities\") pod \"088781cb-6d76-4036-b6fe-44a74dc0b88e\" (UID: \"088781cb-6d76-4036-b6fe-44a74dc0b88e\") " Dec 13 07:25:11 crc kubenswrapper[4971]: I1213 07:25:11.432825 4971 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/088781cb-6d76-4036-b6fe-44a74dc0b88e-utilities" (OuterVolumeSpecName: "utilities") pod "088781cb-6d76-4036-b6fe-44a74dc0b88e" (UID: "088781cb-6d76-4036-b6fe-44a74dc0b88e"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 13 07:25:11 crc kubenswrapper[4971]: I1213 07:25:11.445008 4971 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/088781cb-6d76-4036-b6fe-44a74dc0b88e-kube-api-access-296dw" (OuterVolumeSpecName: "kube-api-access-296dw") pod "088781cb-6d76-4036-b6fe-44a74dc0b88e" (UID: "088781cb-6d76-4036-b6fe-44a74dc0b88e"). InnerVolumeSpecName "kube-api-access-296dw". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 13 07:25:11 crc kubenswrapper[4971]: I1213 07:25:11.455496 4971 scope.go:117] "RemoveContainer" containerID="fbbb0df4c2f227da6b3301afc069e02413cab194de1f64ad81ef2b572fb4aa0c" Dec 13 07:25:11 crc kubenswrapper[4971]: E1213 07:25:11.456016 4971 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"fbbb0df4c2f227da6b3301afc069e02413cab194de1f64ad81ef2b572fb4aa0c\": container with ID starting with fbbb0df4c2f227da6b3301afc069e02413cab194de1f64ad81ef2b572fb4aa0c not found: ID does not exist" containerID="fbbb0df4c2f227da6b3301afc069e02413cab194de1f64ad81ef2b572fb4aa0c" Dec 13 07:25:11 crc kubenswrapper[4971]: I1213 07:25:11.456066 4971 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"fbbb0df4c2f227da6b3301afc069e02413cab194de1f64ad81ef2b572fb4aa0c"} err="failed to get container status \"fbbb0df4c2f227da6b3301afc069e02413cab194de1f64ad81ef2b572fb4aa0c\": rpc error: code = NotFound desc = could not find container \"fbbb0df4c2f227da6b3301afc069e02413cab194de1f64ad81ef2b572fb4aa0c\": container with ID starting with fbbb0df4c2f227da6b3301afc069e02413cab194de1f64ad81ef2b572fb4aa0c not found: ID does not exist" Dec 13 07:25:11 crc kubenswrapper[4971]: I1213 07:25:11.456095 4971 scope.go:117] "RemoveContainer" containerID="71e378a86e610e89278282d167c2ec0c6e8a9c70cbb2290358c0fc757b2aadd5" Dec 13 07:25:11 crc kubenswrapper[4971]: E1213 07:25:11.456488 4971 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"71e378a86e610e89278282d167c2ec0c6e8a9c70cbb2290358c0fc757b2aadd5\": container with ID starting with 71e378a86e610e89278282d167c2ec0c6e8a9c70cbb2290358c0fc757b2aadd5 not found: ID does not exist" containerID="71e378a86e610e89278282d167c2ec0c6e8a9c70cbb2290358c0fc757b2aadd5" Dec 13 07:25:11 crc kubenswrapper[4971]: I1213 07:25:11.456533 4971 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"71e378a86e610e89278282d167c2ec0c6e8a9c70cbb2290358c0fc757b2aadd5"} err="failed to get container status \"71e378a86e610e89278282d167c2ec0c6e8a9c70cbb2290358c0fc757b2aadd5\": rpc error: code = NotFound desc = could not find container \"71e378a86e610e89278282d167c2ec0c6e8a9c70cbb2290358c0fc757b2aadd5\": container with ID starting with 71e378a86e610e89278282d167c2ec0c6e8a9c70cbb2290358c0fc757b2aadd5 not found: ID does not exist" Dec 13 07:25:11 crc kubenswrapper[4971]: I1213 07:25:11.456556 4971 scope.go:117] "RemoveContainer" containerID="487d08fab5a97d8f55c9955018b903150c8d4068716d1406e27bcb6dc3fc75f8" Dec 13 07:25:11 crc kubenswrapper[4971]: E1213 07:25:11.456847 4971 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"487d08fab5a97d8f55c9955018b903150c8d4068716d1406e27bcb6dc3fc75f8\": container with ID starting with 487d08fab5a97d8f55c9955018b903150c8d4068716d1406e27bcb6dc3fc75f8 not found: ID does not exist" containerID="487d08fab5a97d8f55c9955018b903150c8d4068716d1406e27bcb6dc3fc75f8" Dec 13 07:25:11 crc kubenswrapper[4971]: I1213 07:25:11.456869 4971 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"487d08fab5a97d8f55c9955018b903150c8d4068716d1406e27bcb6dc3fc75f8"} err="failed to get container status \"487d08fab5a97d8f55c9955018b903150c8d4068716d1406e27bcb6dc3fc75f8\": rpc error: code = NotFound desc = could not find container \"487d08fab5a97d8f55c9955018b903150c8d4068716d1406e27bcb6dc3fc75f8\": container with ID starting with 487d08fab5a97d8f55c9955018b903150c8d4068716d1406e27bcb6dc3fc75f8 not found: ID does not exist" Dec 13 07:25:11 crc kubenswrapper[4971]: I1213 07:25:11.463107 4971 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/088781cb-6d76-4036-b6fe-44a74dc0b88e-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "088781cb-6d76-4036-b6fe-44a74dc0b88e" (UID: "088781cb-6d76-4036-b6fe-44a74dc0b88e"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 13 07:25:11 crc kubenswrapper[4971]: I1213 07:25:11.535602 4971 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/088781cb-6d76-4036-b6fe-44a74dc0b88e-utilities\") on node \"crc\" DevicePath \"\"" Dec 13 07:25:11 crc kubenswrapper[4971]: I1213 07:25:11.535697 4971 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-296dw\" (UniqueName: \"kubernetes.io/projected/088781cb-6d76-4036-b6fe-44a74dc0b88e-kube-api-access-296dw\") on node \"crc\" DevicePath \"\"" Dec 13 07:25:11 crc kubenswrapper[4971]: I1213 07:25:11.535758 4971 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/088781cb-6d76-4036-b6fe-44a74dc0b88e-catalog-content\") on node \"crc\" DevicePath \"\"" Dec 13 07:25:11 crc kubenswrapper[4971]: I1213 07:25:11.704375 4971 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-b4xjx"] Dec 13 07:25:11 crc kubenswrapper[4971]: I1213 07:25:11.727811 4971 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-marketplace-b4xjx"] Dec 13 07:25:11 crc kubenswrapper[4971]: I1213 07:25:11.781981 4971 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="088781cb-6d76-4036-b6fe-44a74dc0b88e" path="/var/lib/kubelet/pods/088781cb-6d76-4036-b6fe-44a74dc0b88e/volumes" Dec 13 07:25:16 crc kubenswrapper[4971]: I1213 07:25:16.154134 4971 patch_prober.go:28] interesting pod/machine-config-daemon-82xjz container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 13 07:25:16 crc kubenswrapper[4971]: I1213 07:25:16.154755 4971 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-82xjz" podUID="e0c4af12-edda-41d7-9b44-f87396174ef2" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 13 07:25:18 crc kubenswrapper[4971]: I1213 07:25:18.148815 4971 scope.go:117] "RemoveContainer" containerID="76f58f854276b22f5878baf9dc07e632f05e6335bb6779edc28b4315b134c334" Dec 13 07:25:18 crc kubenswrapper[4971]: I1213 07:25:18.201673 4971 scope.go:117] "RemoveContainer" containerID="cc91778c66ca205f06be892439fa7bd0075527b3d7db3226d2d4b560aac675ab" Dec 13 07:25:18 crc kubenswrapper[4971]: I1213 07:25:18.261667 4971 scope.go:117] "RemoveContainer" containerID="153628e66d4982d16e87f162d44c475a8042e5c3ad6fd999fa22c399d1af2b9f" Dec 13 07:25:36 crc kubenswrapper[4971]: I1213 07:25:36.056233 4971 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-cell1-cell-mapping-kqkzw"] Dec 13 07:25:36 crc kubenswrapper[4971]: I1213 07:25:36.064104 4971 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-cell1-cell-mapping-kqkzw"] Dec 13 07:25:37 crc kubenswrapper[4971]: I1213 07:25:37.781251 4971 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="20d00522-0b07-416b-90b4-0b588f99c41f" path="/var/lib/kubelet/pods/20d00522-0b07-416b-90b4-0b588f99c41f/volumes" Dec 13 07:25:46 crc kubenswrapper[4971]: I1213 07:25:46.153619 4971 patch_prober.go:28] interesting pod/machine-config-daemon-82xjz container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 13 07:25:46 crc kubenswrapper[4971]: I1213 07:25:46.154798 4971 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-82xjz" podUID="e0c4af12-edda-41d7-9b44-f87396174ef2" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 13 07:26:00 crc kubenswrapper[4971]: I1213 07:26:00.869087 4971 generic.go:334] "Generic (PLEG): container finished" podID="edef63aa-a18d-4591-a477-924e5cef56d8" containerID="1b8738b60e78e46c6c3201e6dc3104d9218dd08393b34b5a75578292a38c23c0" exitCode=0 Dec 13 07:26:00 crc kubenswrapper[4971]: I1213 07:26:00.869207 4971 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/configure-network-edpm-deployment-openstack-edpm-ipam-qj5wr" event={"ID":"edef63aa-a18d-4591-a477-924e5cef56d8","Type":"ContainerDied","Data":"1b8738b60e78e46c6c3201e6dc3104d9218dd08393b34b5a75578292a38c23c0"} Dec 13 07:26:02 crc kubenswrapper[4971]: I1213 07:26:02.521292 4971 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/configure-network-edpm-deployment-openstack-edpm-ipam-qj5wr" Dec 13 07:26:02 crc kubenswrapper[4971]: I1213 07:26:02.744205 4971 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/edef63aa-a18d-4591-a477-924e5cef56d8-inventory\") pod \"edef63aa-a18d-4591-a477-924e5cef56d8\" (UID: \"edef63aa-a18d-4591-a477-924e5cef56d8\") " Dec 13 07:26:02 crc kubenswrapper[4971]: I1213 07:26:02.744271 4971 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-7vm5c\" (UniqueName: \"kubernetes.io/projected/edef63aa-a18d-4591-a477-924e5cef56d8-kube-api-access-7vm5c\") pod \"edef63aa-a18d-4591-a477-924e5cef56d8\" (UID: \"edef63aa-a18d-4591-a477-924e5cef56d8\") " Dec 13 07:26:02 crc kubenswrapper[4971]: I1213 07:26:02.744336 4971 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/edef63aa-a18d-4591-a477-924e5cef56d8-ssh-key\") pod \"edef63aa-a18d-4591-a477-924e5cef56d8\" (UID: \"edef63aa-a18d-4591-a477-924e5cef56d8\") " Dec 13 07:26:02 crc kubenswrapper[4971]: I1213 07:26:02.749324 4971 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/edef63aa-a18d-4591-a477-924e5cef56d8-kube-api-access-7vm5c" (OuterVolumeSpecName: "kube-api-access-7vm5c") pod "edef63aa-a18d-4591-a477-924e5cef56d8" (UID: "edef63aa-a18d-4591-a477-924e5cef56d8"). InnerVolumeSpecName "kube-api-access-7vm5c". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 13 07:26:02 crc kubenswrapper[4971]: I1213 07:26:02.770647 4971 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/edef63aa-a18d-4591-a477-924e5cef56d8-inventory" (OuterVolumeSpecName: "inventory") pod "edef63aa-a18d-4591-a477-924e5cef56d8" (UID: "edef63aa-a18d-4591-a477-924e5cef56d8"). InnerVolumeSpecName "inventory". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 13 07:26:02 crc kubenswrapper[4971]: I1213 07:26:02.773233 4971 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/edef63aa-a18d-4591-a477-924e5cef56d8-ssh-key" (OuterVolumeSpecName: "ssh-key") pod "edef63aa-a18d-4591-a477-924e5cef56d8" (UID: "edef63aa-a18d-4591-a477-924e5cef56d8"). InnerVolumeSpecName "ssh-key". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 13 07:26:02 crc kubenswrapper[4971]: I1213 07:26:02.846320 4971 reconciler_common.go:293] "Volume detached for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/edef63aa-a18d-4591-a477-924e5cef56d8-inventory\") on node \"crc\" DevicePath \"\"" Dec 13 07:26:02 crc kubenswrapper[4971]: I1213 07:26:02.846357 4971 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-7vm5c\" (UniqueName: \"kubernetes.io/projected/edef63aa-a18d-4591-a477-924e5cef56d8-kube-api-access-7vm5c\") on node \"crc\" DevicePath \"\"" Dec 13 07:26:02 crc kubenswrapper[4971]: I1213 07:26:02.846368 4971 reconciler_common.go:293] "Volume detached for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/edef63aa-a18d-4591-a477-924e5cef56d8-ssh-key\") on node \"crc\" DevicePath \"\"" Dec 13 07:26:02 crc kubenswrapper[4971]: I1213 07:26:02.888332 4971 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/configure-network-edpm-deployment-openstack-edpm-ipam-qj5wr" event={"ID":"edef63aa-a18d-4591-a477-924e5cef56d8","Type":"ContainerDied","Data":"33cc89611a94a06c47c3084f0fc6e0e327ccba0d46d7d6375ff44c3035736cb6"} Dec 13 07:26:02 crc kubenswrapper[4971]: I1213 07:26:02.888406 4971 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="33cc89611a94a06c47c3084f0fc6e0e327ccba0d46d7d6375ff44c3035736cb6" Dec 13 07:26:02 crc kubenswrapper[4971]: I1213 07:26:02.888560 4971 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/configure-network-edpm-deployment-openstack-edpm-ipam-qj5wr" Dec 13 07:26:02 crc kubenswrapper[4971]: I1213 07:26:02.984986 4971 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/validate-network-edpm-deployment-openstack-edpm-ipam-wx2k2"] Dec 13 07:26:02 crc kubenswrapper[4971]: E1213 07:26:02.985512 4971 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="edef63aa-a18d-4591-a477-924e5cef56d8" containerName="configure-network-edpm-deployment-openstack-edpm-ipam" Dec 13 07:26:02 crc kubenswrapper[4971]: I1213 07:26:02.985551 4971 state_mem.go:107] "Deleted CPUSet assignment" podUID="edef63aa-a18d-4591-a477-924e5cef56d8" containerName="configure-network-edpm-deployment-openstack-edpm-ipam" Dec 13 07:26:02 crc kubenswrapper[4971]: E1213 07:26:02.985574 4971 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="088781cb-6d76-4036-b6fe-44a74dc0b88e" containerName="extract-content" Dec 13 07:26:02 crc kubenswrapper[4971]: I1213 07:26:02.985581 4971 state_mem.go:107] "Deleted CPUSet assignment" podUID="088781cb-6d76-4036-b6fe-44a74dc0b88e" containerName="extract-content" Dec 13 07:26:02 crc kubenswrapper[4971]: E1213 07:26:02.985598 4971 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="088781cb-6d76-4036-b6fe-44a74dc0b88e" containerName="registry-server" Dec 13 07:26:02 crc kubenswrapper[4971]: I1213 07:26:02.985610 4971 state_mem.go:107] "Deleted CPUSet assignment" podUID="088781cb-6d76-4036-b6fe-44a74dc0b88e" containerName="registry-server" Dec 13 07:26:02 crc kubenswrapper[4971]: E1213 07:26:02.985625 4971 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="088781cb-6d76-4036-b6fe-44a74dc0b88e" containerName="extract-utilities" Dec 13 07:26:02 crc kubenswrapper[4971]: I1213 07:26:02.985634 4971 state_mem.go:107] "Deleted CPUSet assignment" podUID="088781cb-6d76-4036-b6fe-44a74dc0b88e" containerName="extract-utilities" Dec 13 07:26:02 crc kubenswrapper[4971]: I1213 07:26:02.985853 4971 memory_manager.go:354] "RemoveStaleState removing state" podUID="edef63aa-a18d-4591-a477-924e5cef56d8" containerName="configure-network-edpm-deployment-openstack-edpm-ipam" Dec 13 07:26:02 crc kubenswrapper[4971]: I1213 07:26:02.985883 4971 memory_manager.go:354] "RemoveStaleState removing state" podUID="088781cb-6d76-4036-b6fe-44a74dc0b88e" containerName="registry-server" Dec 13 07:26:02 crc kubenswrapper[4971]: I1213 07:26:02.986629 4971 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/validate-network-edpm-deployment-openstack-edpm-ipam-wx2k2" Dec 13 07:26:02 crc kubenswrapper[4971]: I1213 07:26:02.989008 4971 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplanenodeset-openstack-edpm-ipam" Dec 13 07:26:02 crc kubenswrapper[4971]: I1213 07:26:02.989095 4971 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplane-ansible-ssh-private-key-secret" Dec 13 07:26:02 crc kubenswrapper[4971]: I1213 07:26:02.989095 4971 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-aee-default-env" Dec 13 07:26:02 crc kubenswrapper[4971]: I1213 07:26:02.989249 4971 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-edpm-ipam-dockercfg-zb2j5" Dec 13 07:26:03 crc kubenswrapper[4971]: I1213 07:26:03.002873 4971 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/validate-network-edpm-deployment-openstack-edpm-ipam-wx2k2"] Dec 13 07:26:03 crc kubenswrapper[4971]: I1213 07:26:03.155000 4971 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/f468bc03-51b5-4b38-a2b5-b9061002f4d7-inventory\") pod \"validate-network-edpm-deployment-openstack-edpm-ipam-wx2k2\" (UID: \"f468bc03-51b5-4b38-a2b5-b9061002f4d7\") " pod="openstack/validate-network-edpm-deployment-openstack-edpm-ipam-wx2k2" Dec 13 07:26:03 crc kubenswrapper[4971]: I1213 07:26:03.155110 4971 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-2nrm7\" (UniqueName: \"kubernetes.io/projected/f468bc03-51b5-4b38-a2b5-b9061002f4d7-kube-api-access-2nrm7\") pod \"validate-network-edpm-deployment-openstack-edpm-ipam-wx2k2\" (UID: \"f468bc03-51b5-4b38-a2b5-b9061002f4d7\") " pod="openstack/validate-network-edpm-deployment-openstack-edpm-ipam-wx2k2" Dec 13 07:26:03 crc kubenswrapper[4971]: I1213 07:26:03.155221 4971 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/f468bc03-51b5-4b38-a2b5-b9061002f4d7-ssh-key\") pod \"validate-network-edpm-deployment-openstack-edpm-ipam-wx2k2\" (UID: \"f468bc03-51b5-4b38-a2b5-b9061002f4d7\") " pod="openstack/validate-network-edpm-deployment-openstack-edpm-ipam-wx2k2" Dec 13 07:26:03 crc kubenswrapper[4971]: I1213 07:26:03.256647 4971 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/f468bc03-51b5-4b38-a2b5-b9061002f4d7-inventory\") pod \"validate-network-edpm-deployment-openstack-edpm-ipam-wx2k2\" (UID: \"f468bc03-51b5-4b38-a2b5-b9061002f4d7\") " pod="openstack/validate-network-edpm-deployment-openstack-edpm-ipam-wx2k2" Dec 13 07:26:03 crc kubenswrapper[4971]: I1213 07:26:03.256769 4971 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-2nrm7\" (UniqueName: \"kubernetes.io/projected/f468bc03-51b5-4b38-a2b5-b9061002f4d7-kube-api-access-2nrm7\") pod \"validate-network-edpm-deployment-openstack-edpm-ipam-wx2k2\" (UID: \"f468bc03-51b5-4b38-a2b5-b9061002f4d7\") " pod="openstack/validate-network-edpm-deployment-openstack-edpm-ipam-wx2k2" Dec 13 07:26:03 crc kubenswrapper[4971]: I1213 07:26:03.256846 4971 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/f468bc03-51b5-4b38-a2b5-b9061002f4d7-ssh-key\") pod \"validate-network-edpm-deployment-openstack-edpm-ipam-wx2k2\" (UID: \"f468bc03-51b5-4b38-a2b5-b9061002f4d7\") " pod="openstack/validate-network-edpm-deployment-openstack-edpm-ipam-wx2k2" Dec 13 07:26:03 crc kubenswrapper[4971]: I1213 07:26:03.262260 4971 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/f468bc03-51b5-4b38-a2b5-b9061002f4d7-ssh-key\") pod \"validate-network-edpm-deployment-openstack-edpm-ipam-wx2k2\" (UID: \"f468bc03-51b5-4b38-a2b5-b9061002f4d7\") " pod="openstack/validate-network-edpm-deployment-openstack-edpm-ipam-wx2k2" Dec 13 07:26:03 crc kubenswrapper[4971]: I1213 07:26:03.262982 4971 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/f468bc03-51b5-4b38-a2b5-b9061002f4d7-inventory\") pod \"validate-network-edpm-deployment-openstack-edpm-ipam-wx2k2\" (UID: \"f468bc03-51b5-4b38-a2b5-b9061002f4d7\") " pod="openstack/validate-network-edpm-deployment-openstack-edpm-ipam-wx2k2" Dec 13 07:26:03 crc kubenswrapper[4971]: I1213 07:26:03.288260 4971 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-2nrm7\" (UniqueName: \"kubernetes.io/projected/f468bc03-51b5-4b38-a2b5-b9061002f4d7-kube-api-access-2nrm7\") pod \"validate-network-edpm-deployment-openstack-edpm-ipam-wx2k2\" (UID: \"f468bc03-51b5-4b38-a2b5-b9061002f4d7\") " pod="openstack/validate-network-edpm-deployment-openstack-edpm-ipam-wx2k2" Dec 13 07:26:03 crc kubenswrapper[4971]: I1213 07:26:03.312876 4971 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/validate-network-edpm-deployment-openstack-edpm-ipam-wx2k2" Dec 13 07:26:04 crc kubenswrapper[4971]: I1213 07:26:04.093901 4971 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/validate-network-edpm-deployment-openstack-edpm-ipam-wx2k2"] Dec 13 07:26:04 crc kubenswrapper[4971]: I1213 07:26:04.715085 4971 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-aee-default-env" Dec 13 07:26:04 crc kubenswrapper[4971]: I1213 07:26:04.908411 4971 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/validate-network-edpm-deployment-openstack-edpm-ipam-wx2k2" event={"ID":"f468bc03-51b5-4b38-a2b5-b9061002f4d7","Type":"ContainerStarted","Data":"167c95463f0741e8137a0c32131ce3f7dce8bff750f2c89ddac28a72c0f36572"} Dec 13 07:26:05 crc kubenswrapper[4971]: I1213 07:26:05.918444 4971 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/validate-network-edpm-deployment-openstack-edpm-ipam-wx2k2" event={"ID":"f468bc03-51b5-4b38-a2b5-b9061002f4d7","Type":"ContainerStarted","Data":"456322a0d13d44bab8798554c48b5c8b6cc23cb2481bfbbc2a4d19fa60a84d7e"} Dec 13 07:26:05 crc kubenswrapper[4971]: I1213 07:26:05.941705 4971 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/validate-network-edpm-deployment-openstack-edpm-ipam-wx2k2" podStartSLOduration=3.33274765 podStartE2EDuration="3.941688177s" podCreationTimestamp="2025-12-13 07:26:02 +0000 UTC" firstStartedPulling="2025-12-13 07:26:04.102634794 +0000 UTC m=+2220.707044242" lastFinishedPulling="2025-12-13 07:26:04.711575321 +0000 UTC m=+2221.315984769" observedRunningTime="2025-12-13 07:26:05.938001056 +0000 UTC m=+2222.542410504" watchObservedRunningTime="2025-12-13 07:26:05.941688177 +0000 UTC m=+2222.546097625" Dec 13 07:26:11 crc kubenswrapper[4971]: I1213 07:26:11.973169 4971 generic.go:334] "Generic (PLEG): container finished" podID="f468bc03-51b5-4b38-a2b5-b9061002f4d7" containerID="456322a0d13d44bab8798554c48b5c8b6cc23cb2481bfbbc2a4d19fa60a84d7e" exitCode=0 Dec 13 07:26:11 crc kubenswrapper[4971]: I1213 07:26:11.973545 4971 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/validate-network-edpm-deployment-openstack-edpm-ipam-wx2k2" event={"ID":"f468bc03-51b5-4b38-a2b5-b9061002f4d7","Type":"ContainerDied","Data":"456322a0d13d44bab8798554c48b5c8b6cc23cb2481bfbbc2a4d19fa60a84d7e"} Dec 13 07:26:13 crc kubenswrapper[4971]: I1213 07:26:13.457210 4971 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/validate-network-edpm-deployment-openstack-edpm-ipam-wx2k2" Dec 13 07:26:13 crc kubenswrapper[4971]: I1213 07:26:13.463016 4971 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/f468bc03-51b5-4b38-a2b5-b9061002f4d7-inventory\") pod \"f468bc03-51b5-4b38-a2b5-b9061002f4d7\" (UID: \"f468bc03-51b5-4b38-a2b5-b9061002f4d7\") " Dec 13 07:26:13 crc kubenswrapper[4971]: I1213 07:26:13.508976 4971 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/f468bc03-51b5-4b38-a2b5-b9061002f4d7-inventory" (OuterVolumeSpecName: "inventory") pod "f468bc03-51b5-4b38-a2b5-b9061002f4d7" (UID: "f468bc03-51b5-4b38-a2b5-b9061002f4d7"). InnerVolumeSpecName "inventory". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 13 07:26:13 crc kubenswrapper[4971]: I1213 07:26:13.565624 4971 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-2nrm7\" (UniqueName: \"kubernetes.io/projected/f468bc03-51b5-4b38-a2b5-b9061002f4d7-kube-api-access-2nrm7\") pod \"f468bc03-51b5-4b38-a2b5-b9061002f4d7\" (UID: \"f468bc03-51b5-4b38-a2b5-b9061002f4d7\") " Dec 13 07:26:13 crc kubenswrapper[4971]: I1213 07:26:13.566616 4971 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/f468bc03-51b5-4b38-a2b5-b9061002f4d7-ssh-key\") pod \"f468bc03-51b5-4b38-a2b5-b9061002f4d7\" (UID: \"f468bc03-51b5-4b38-a2b5-b9061002f4d7\") " Dec 13 07:26:13 crc kubenswrapper[4971]: I1213 07:26:13.567376 4971 reconciler_common.go:293] "Volume detached for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/f468bc03-51b5-4b38-a2b5-b9061002f4d7-inventory\") on node \"crc\" DevicePath \"\"" Dec 13 07:26:13 crc kubenswrapper[4971]: I1213 07:26:13.569295 4971 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/f468bc03-51b5-4b38-a2b5-b9061002f4d7-kube-api-access-2nrm7" (OuterVolumeSpecName: "kube-api-access-2nrm7") pod "f468bc03-51b5-4b38-a2b5-b9061002f4d7" (UID: "f468bc03-51b5-4b38-a2b5-b9061002f4d7"). InnerVolumeSpecName "kube-api-access-2nrm7". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 13 07:26:13 crc kubenswrapper[4971]: I1213 07:26:13.598405 4971 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/f468bc03-51b5-4b38-a2b5-b9061002f4d7-ssh-key" (OuterVolumeSpecName: "ssh-key") pod "f468bc03-51b5-4b38-a2b5-b9061002f4d7" (UID: "f468bc03-51b5-4b38-a2b5-b9061002f4d7"). InnerVolumeSpecName "ssh-key". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 13 07:26:13 crc kubenswrapper[4971]: I1213 07:26:13.669779 4971 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-2nrm7\" (UniqueName: \"kubernetes.io/projected/f468bc03-51b5-4b38-a2b5-b9061002f4d7-kube-api-access-2nrm7\") on node \"crc\" DevicePath \"\"" Dec 13 07:26:13 crc kubenswrapper[4971]: I1213 07:26:13.669828 4971 reconciler_common.go:293] "Volume detached for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/f468bc03-51b5-4b38-a2b5-b9061002f4d7-ssh-key\") on node \"crc\" DevicePath \"\"" Dec 13 07:26:13 crc kubenswrapper[4971]: I1213 07:26:13.997316 4971 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/validate-network-edpm-deployment-openstack-edpm-ipam-wx2k2" event={"ID":"f468bc03-51b5-4b38-a2b5-b9061002f4d7","Type":"ContainerDied","Data":"167c95463f0741e8137a0c32131ce3f7dce8bff750f2c89ddac28a72c0f36572"} Dec 13 07:26:13 crc kubenswrapper[4971]: I1213 07:26:13.997383 4971 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="167c95463f0741e8137a0c32131ce3f7dce8bff750f2c89ddac28a72c0f36572" Dec 13 07:26:13 crc kubenswrapper[4971]: I1213 07:26:13.997413 4971 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/validate-network-edpm-deployment-openstack-edpm-ipam-wx2k2" Dec 13 07:26:14 crc kubenswrapper[4971]: I1213 07:26:14.078489 4971 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/install-os-edpm-deployment-openstack-edpm-ipam-n2hlm"] Dec 13 07:26:14 crc kubenswrapper[4971]: E1213 07:26:14.078860 4971 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f468bc03-51b5-4b38-a2b5-b9061002f4d7" containerName="validate-network-edpm-deployment-openstack-edpm-ipam" Dec 13 07:26:14 crc kubenswrapper[4971]: I1213 07:26:14.078874 4971 state_mem.go:107] "Deleted CPUSet assignment" podUID="f468bc03-51b5-4b38-a2b5-b9061002f4d7" containerName="validate-network-edpm-deployment-openstack-edpm-ipam" Dec 13 07:26:14 crc kubenswrapper[4971]: I1213 07:26:14.079088 4971 memory_manager.go:354] "RemoveStaleState removing state" podUID="f468bc03-51b5-4b38-a2b5-b9061002f4d7" containerName="validate-network-edpm-deployment-openstack-edpm-ipam" Dec 13 07:26:14 crc kubenswrapper[4971]: I1213 07:26:14.079695 4971 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/install-os-edpm-deployment-openstack-edpm-ipam-n2hlm" Dec 13 07:26:14 crc kubenswrapper[4971]: I1213 07:26:14.083280 4971 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplane-ansible-ssh-private-key-secret" Dec 13 07:26:14 crc kubenswrapper[4971]: I1213 07:26:14.083493 4971 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-aee-default-env" Dec 13 07:26:14 crc kubenswrapper[4971]: I1213 07:26:14.089840 4971 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/install-os-edpm-deployment-openstack-edpm-ipam-n2hlm"] Dec 13 07:26:14 crc kubenswrapper[4971]: I1213 07:26:14.094216 4971 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-edpm-ipam-dockercfg-zb2j5" Dec 13 07:26:14 crc kubenswrapper[4971]: I1213 07:26:14.094424 4971 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplanenodeset-openstack-edpm-ipam" Dec 13 07:26:14 crc kubenswrapper[4971]: I1213 07:26:14.179650 4971 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-v8bh9\" (UniqueName: \"kubernetes.io/projected/a4963e97-7f24-46ad-bd42-8b6c2b2d0274-kube-api-access-v8bh9\") pod \"install-os-edpm-deployment-openstack-edpm-ipam-n2hlm\" (UID: \"a4963e97-7f24-46ad-bd42-8b6c2b2d0274\") " pod="openstack/install-os-edpm-deployment-openstack-edpm-ipam-n2hlm" Dec 13 07:26:14 crc kubenswrapper[4971]: I1213 07:26:14.180012 4971 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/a4963e97-7f24-46ad-bd42-8b6c2b2d0274-inventory\") pod \"install-os-edpm-deployment-openstack-edpm-ipam-n2hlm\" (UID: \"a4963e97-7f24-46ad-bd42-8b6c2b2d0274\") " pod="openstack/install-os-edpm-deployment-openstack-edpm-ipam-n2hlm" Dec 13 07:26:14 crc kubenswrapper[4971]: I1213 07:26:14.180193 4971 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/a4963e97-7f24-46ad-bd42-8b6c2b2d0274-ssh-key\") pod \"install-os-edpm-deployment-openstack-edpm-ipam-n2hlm\" (UID: \"a4963e97-7f24-46ad-bd42-8b6c2b2d0274\") " pod="openstack/install-os-edpm-deployment-openstack-edpm-ipam-n2hlm" Dec 13 07:26:14 crc kubenswrapper[4971]: I1213 07:26:14.294253 4971 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/a4963e97-7f24-46ad-bd42-8b6c2b2d0274-inventory\") pod \"install-os-edpm-deployment-openstack-edpm-ipam-n2hlm\" (UID: \"a4963e97-7f24-46ad-bd42-8b6c2b2d0274\") " pod="openstack/install-os-edpm-deployment-openstack-edpm-ipam-n2hlm" Dec 13 07:26:14 crc kubenswrapper[4971]: I1213 07:26:14.294405 4971 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/a4963e97-7f24-46ad-bd42-8b6c2b2d0274-ssh-key\") pod \"install-os-edpm-deployment-openstack-edpm-ipam-n2hlm\" (UID: \"a4963e97-7f24-46ad-bd42-8b6c2b2d0274\") " pod="openstack/install-os-edpm-deployment-openstack-edpm-ipam-n2hlm" Dec 13 07:26:14 crc kubenswrapper[4971]: I1213 07:26:14.294747 4971 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-v8bh9\" (UniqueName: \"kubernetes.io/projected/a4963e97-7f24-46ad-bd42-8b6c2b2d0274-kube-api-access-v8bh9\") pod \"install-os-edpm-deployment-openstack-edpm-ipam-n2hlm\" (UID: \"a4963e97-7f24-46ad-bd42-8b6c2b2d0274\") " pod="openstack/install-os-edpm-deployment-openstack-edpm-ipam-n2hlm" Dec 13 07:26:14 crc kubenswrapper[4971]: I1213 07:26:14.298079 4971 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/a4963e97-7f24-46ad-bd42-8b6c2b2d0274-ssh-key\") pod \"install-os-edpm-deployment-openstack-edpm-ipam-n2hlm\" (UID: \"a4963e97-7f24-46ad-bd42-8b6c2b2d0274\") " pod="openstack/install-os-edpm-deployment-openstack-edpm-ipam-n2hlm" Dec 13 07:26:14 crc kubenswrapper[4971]: I1213 07:26:14.299260 4971 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/a4963e97-7f24-46ad-bd42-8b6c2b2d0274-inventory\") pod \"install-os-edpm-deployment-openstack-edpm-ipam-n2hlm\" (UID: \"a4963e97-7f24-46ad-bd42-8b6c2b2d0274\") " pod="openstack/install-os-edpm-deployment-openstack-edpm-ipam-n2hlm" Dec 13 07:26:14 crc kubenswrapper[4971]: I1213 07:26:14.324789 4971 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-v8bh9\" (UniqueName: \"kubernetes.io/projected/a4963e97-7f24-46ad-bd42-8b6c2b2d0274-kube-api-access-v8bh9\") pod \"install-os-edpm-deployment-openstack-edpm-ipam-n2hlm\" (UID: \"a4963e97-7f24-46ad-bd42-8b6c2b2d0274\") " pod="openstack/install-os-edpm-deployment-openstack-edpm-ipam-n2hlm" Dec 13 07:26:14 crc kubenswrapper[4971]: I1213 07:26:14.397146 4971 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/install-os-edpm-deployment-openstack-edpm-ipam-n2hlm" Dec 13 07:26:15 crc kubenswrapper[4971]: I1213 07:26:15.014631 4971 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/install-os-edpm-deployment-openstack-edpm-ipam-n2hlm"] Dec 13 07:26:15 crc kubenswrapper[4971]: W1213 07:26:15.022382 4971 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-poda4963e97_7f24_46ad_bd42_8b6c2b2d0274.slice/crio-c96c0ec7ce7d831395f05a16ff2bfc1aac8de3cd226c73849035218cbcfcc2b0 WatchSource:0}: Error finding container c96c0ec7ce7d831395f05a16ff2bfc1aac8de3cd226c73849035218cbcfcc2b0: Status 404 returned error can't find the container with id c96c0ec7ce7d831395f05a16ff2bfc1aac8de3cd226c73849035218cbcfcc2b0 Dec 13 07:26:16 crc kubenswrapper[4971]: I1213 07:26:16.017604 4971 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/install-os-edpm-deployment-openstack-edpm-ipam-n2hlm" event={"ID":"a4963e97-7f24-46ad-bd42-8b6c2b2d0274","Type":"ContainerStarted","Data":"58181d74ba49de7a80cf6dac9d470068037e36d6ae703590820a64b82fd344a6"} Dec 13 07:26:16 crc kubenswrapper[4971]: I1213 07:26:16.017991 4971 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/install-os-edpm-deployment-openstack-edpm-ipam-n2hlm" event={"ID":"a4963e97-7f24-46ad-bd42-8b6c2b2d0274","Type":"ContainerStarted","Data":"c96c0ec7ce7d831395f05a16ff2bfc1aac8de3cd226c73849035218cbcfcc2b0"} Dec 13 07:26:16 crc kubenswrapper[4971]: I1213 07:26:16.039595 4971 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/install-os-edpm-deployment-openstack-edpm-ipam-n2hlm" podStartSLOduration=1.36205036 podStartE2EDuration="2.039576286s" podCreationTimestamp="2025-12-13 07:26:14 +0000 UTC" firstStartedPulling="2025-12-13 07:26:15.024855561 +0000 UTC m=+2231.629264999" lastFinishedPulling="2025-12-13 07:26:15.702381477 +0000 UTC m=+2232.306790925" observedRunningTime="2025-12-13 07:26:16.033699821 +0000 UTC m=+2232.638109279" watchObservedRunningTime="2025-12-13 07:26:16.039576286 +0000 UTC m=+2232.643985734" Dec 13 07:26:16 crc kubenswrapper[4971]: I1213 07:26:16.153324 4971 patch_prober.go:28] interesting pod/machine-config-daemon-82xjz container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 13 07:26:16 crc kubenswrapper[4971]: I1213 07:26:16.153378 4971 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-82xjz" podUID="e0c4af12-edda-41d7-9b44-f87396174ef2" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 13 07:26:16 crc kubenswrapper[4971]: I1213 07:26:16.153414 4971 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-82xjz" Dec 13 07:26:16 crc kubenswrapper[4971]: I1213 07:26:16.154107 4971 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"ec2e4b1312b5d5d0545354a836fef4af376ca3b34327ab1b9e0b72c102976079"} pod="openshift-machine-config-operator/machine-config-daemon-82xjz" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Dec 13 07:26:16 crc kubenswrapper[4971]: I1213 07:26:16.154166 4971 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-82xjz" podUID="e0c4af12-edda-41d7-9b44-f87396174ef2" containerName="machine-config-daemon" containerID="cri-o://ec2e4b1312b5d5d0545354a836fef4af376ca3b34327ab1b9e0b72c102976079" gracePeriod=600 Dec 13 07:26:16 crc kubenswrapper[4971]: E1213 07:26:16.274138 4971 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-82xjz_openshift-machine-config-operator(e0c4af12-edda-41d7-9b44-f87396174ef2)\"" pod="openshift-machine-config-operator/machine-config-daemon-82xjz" podUID="e0c4af12-edda-41d7-9b44-f87396174ef2" Dec 13 07:26:17 crc kubenswrapper[4971]: I1213 07:26:17.033483 4971 generic.go:334] "Generic (PLEG): container finished" podID="e0c4af12-edda-41d7-9b44-f87396174ef2" containerID="ec2e4b1312b5d5d0545354a836fef4af376ca3b34327ab1b9e0b72c102976079" exitCode=0 Dec 13 07:26:17 crc kubenswrapper[4971]: I1213 07:26:17.033569 4971 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-82xjz" event={"ID":"e0c4af12-edda-41d7-9b44-f87396174ef2","Type":"ContainerDied","Data":"ec2e4b1312b5d5d0545354a836fef4af376ca3b34327ab1b9e0b72c102976079"} Dec 13 07:26:17 crc kubenswrapper[4971]: I1213 07:26:17.033905 4971 scope.go:117] "RemoveContainer" containerID="adc3c82aa161f8f1a6007b84b7b30285aa46a34a77bebf9c44580ee9879d3298" Dec 13 07:26:17 crc kubenswrapper[4971]: I1213 07:26:17.034728 4971 scope.go:117] "RemoveContainer" containerID="ec2e4b1312b5d5d0545354a836fef4af376ca3b34327ab1b9e0b72c102976079" Dec 13 07:26:17 crc kubenswrapper[4971]: E1213 07:26:17.035014 4971 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-82xjz_openshift-machine-config-operator(e0c4af12-edda-41d7-9b44-f87396174ef2)\"" pod="openshift-machine-config-operator/machine-config-daemon-82xjz" podUID="e0c4af12-edda-41d7-9b44-f87396174ef2" Dec 13 07:26:18 crc kubenswrapper[4971]: I1213 07:26:18.548764 4971 scope.go:117] "RemoveContainer" containerID="c993f4cf7451d6815046085857a794dcbdad51b484925bf88562a78935bb1b7f" Dec 13 07:26:29 crc kubenswrapper[4971]: I1213 07:26:29.768870 4971 scope.go:117] "RemoveContainer" containerID="ec2e4b1312b5d5d0545354a836fef4af376ca3b34327ab1b9e0b72c102976079" Dec 13 07:26:29 crc kubenswrapper[4971]: E1213 07:26:29.770204 4971 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-82xjz_openshift-machine-config-operator(e0c4af12-edda-41d7-9b44-f87396174ef2)\"" pod="openshift-machine-config-operator/machine-config-daemon-82xjz" podUID="e0c4af12-edda-41d7-9b44-f87396174ef2" Dec 13 07:26:41 crc kubenswrapper[4971]: I1213 07:26:41.769078 4971 scope.go:117] "RemoveContainer" containerID="ec2e4b1312b5d5d0545354a836fef4af376ca3b34327ab1b9e0b72c102976079" Dec 13 07:26:41 crc kubenswrapper[4971]: E1213 07:26:41.769988 4971 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-82xjz_openshift-machine-config-operator(e0c4af12-edda-41d7-9b44-f87396174ef2)\"" pod="openshift-machine-config-operator/machine-config-daemon-82xjz" podUID="e0c4af12-edda-41d7-9b44-f87396174ef2" Dec 13 07:26:52 crc kubenswrapper[4971]: I1213 07:26:52.768767 4971 scope.go:117] "RemoveContainer" containerID="ec2e4b1312b5d5d0545354a836fef4af376ca3b34327ab1b9e0b72c102976079" Dec 13 07:26:52 crc kubenswrapper[4971]: E1213 07:26:52.769601 4971 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-82xjz_openshift-machine-config-operator(e0c4af12-edda-41d7-9b44-f87396174ef2)\"" pod="openshift-machine-config-operator/machine-config-daemon-82xjz" podUID="e0c4af12-edda-41d7-9b44-f87396174ef2" Dec 13 07:27:01 crc kubenswrapper[4971]: I1213 07:27:01.481587 4971 generic.go:334] "Generic (PLEG): container finished" podID="a4963e97-7f24-46ad-bd42-8b6c2b2d0274" containerID="58181d74ba49de7a80cf6dac9d470068037e36d6ae703590820a64b82fd344a6" exitCode=0 Dec 13 07:27:01 crc kubenswrapper[4971]: I1213 07:27:01.481636 4971 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/install-os-edpm-deployment-openstack-edpm-ipam-n2hlm" event={"ID":"a4963e97-7f24-46ad-bd42-8b6c2b2d0274","Type":"ContainerDied","Data":"58181d74ba49de7a80cf6dac9d470068037e36d6ae703590820a64b82fd344a6"} Dec 13 07:27:02 crc kubenswrapper[4971]: I1213 07:27:02.915247 4971 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/install-os-edpm-deployment-openstack-edpm-ipam-n2hlm" Dec 13 07:27:02 crc kubenswrapper[4971]: I1213 07:27:02.975062 4971 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/a4963e97-7f24-46ad-bd42-8b6c2b2d0274-inventory\") pod \"a4963e97-7f24-46ad-bd42-8b6c2b2d0274\" (UID: \"a4963e97-7f24-46ad-bd42-8b6c2b2d0274\") " Dec 13 07:27:02 crc kubenswrapper[4971]: I1213 07:27:02.975353 4971 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/a4963e97-7f24-46ad-bd42-8b6c2b2d0274-ssh-key\") pod \"a4963e97-7f24-46ad-bd42-8b6c2b2d0274\" (UID: \"a4963e97-7f24-46ad-bd42-8b6c2b2d0274\") " Dec 13 07:27:02 crc kubenswrapper[4971]: I1213 07:27:02.975469 4971 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-v8bh9\" (UniqueName: \"kubernetes.io/projected/a4963e97-7f24-46ad-bd42-8b6c2b2d0274-kube-api-access-v8bh9\") pod \"a4963e97-7f24-46ad-bd42-8b6c2b2d0274\" (UID: \"a4963e97-7f24-46ad-bd42-8b6c2b2d0274\") " Dec 13 07:27:02 crc kubenswrapper[4971]: I1213 07:27:02.999078 4971 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/a4963e97-7f24-46ad-bd42-8b6c2b2d0274-kube-api-access-v8bh9" (OuterVolumeSpecName: "kube-api-access-v8bh9") pod "a4963e97-7f24-46ad-bd42-8b6c2b2d0274" (UID: "a4963e97-7f24-46ad-bd42-8b6c2b2d0274"). InnerVolumeSpecName "kube-api-access-v8bh9". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 13 07:27:03 crc kubenswrapper[4971]: I1213 07:27:03.022790 4971 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/a4963e97-7f24-46ad-bd42-8b6c2b2d0274-inventory" (OuterVolumeSpecName: "inventory") pod "a4963e97-7f24-46ad-bd42-8b6c2b2d0274" (UID: "a4963e97-7f24-46ad-bd42-8b6c2b2d0274"). InnerVolumeSpecName "inventory". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 13 07:27:03 crc kubenswrapper[4971]: I1213 07:27:03.024237 4971 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/a4963e97-7f24-46ad-bd42-8b6c2b2d0274-ssh-key" (OuterVolumeSpecName: "ssh-key") pod "a4963e97-7f24-46ad-bd42-8b6c2b2d0274" (UID: "a4963e97-7f24-46ad-bd42-8b6c2b2d0274"). InnerVolumeSpecName "ssh-key". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 13 07:27:03 crc kubenswrapper[4971]: I1213 07:27:03.085808 4971 reconciler_common.go:293] "Volume detached for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/a4963e97-7f24-46ad-bd42-8b6c2b2d0274-inventory\") on node \"crc\" DevicePath \"\"" Dec 13 07:27:03 crc kubenswrapper[4971]: I1213 07:27:03.085843 4971 reconciler_common.go:293] "Volume detached for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/a4963e97-7f24-46ad-bd42-8b6c2b2d0274-ssh-key\") on node \"crc\" DevicePath \"\"" Dec 13 07:27:03 crc kubenswrapper[4971]: I1213 07:27:03.085855 4971 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-v8bh9\" (UniqueName: \"kubernetes.io/projected/a4963e97-7f24-46ad-bd42-8b6c2b2d0274-kube-api-access-v8bh9\") on node \"crc\" DevicePath \"\"" Dec 13 07:27:03 crc kubenswrapper[4971]: I1213 07:27:03.500798 4971 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/install-os-edpm-deployment-openstack-edpm-ipam-n2hlm" event={"ID":"a4963e97-7f24-46ad-bd42-8b6c2b2d0274","Type":"ContainerDied","Data":"c96c0ec7ce7d831395f05a16ff2bfc1aac8de3cd226c73849035218cbcfcc2b0"} Dec 13 07:27:03 crc kubenswrapper[4971]: I1213 07:27:03.501158 4971 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="c96c0ec7ce7d831395f05a16ff2bfc1aac8de3cd226c73849035218cbcfcc2b0" Dec 13 07:27:03 crc kubenswrapper[4971]: I1213 07:27:03.500896 4971 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/install-os-edpm-deployment-openstack-edpm-ipam-n2hlm" Dec 13 07:27:03 crc kubenswrapper[4971]: I1213 07:27:03.640415 4971 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ceph-hci-pre-edpm-deployment-openstack-edpm-ipam-llgqk"] Dec 13 07:27:03 crc kubenswrapper[4971]: E1213 07:27:03.640987 4971 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="a4963e97-7f24-46ad-bd42-8b6c2b2d0274" containerName="install-os-edpm-deployment-openstack-edpm-ipam" Dec 13 07:27:03 crc kubenswrapper[4971]: I1213 07:27:03.641004 4971 state_mem.go:107] "Deleted CPUSet assignment" podUID="a4963e97-7f24-46ad-bd42-8b6c2b2d0274" containerName="install-os-edpm-deployment-openstack-edpm-ipam" Dec 13 07:27:03 crc kubenswrapper[4971]: I1213 07:27:03.641158 4971 memory_manager.go:354] "RemoveStaleState removing state" podUID="a4963e97-7f24-46ad-bd42-8b6c2b2d0274" containerName="install-os-edpm-deployment-openstack-edpm-ipam" Dec 13 07:27:03 crc kubenswrapper[4971]: I1213 07:27:03.641789 4971 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceph-hci-pre-edpm-deployment-openstack-edpm-ipam-llgqk" Dec 13 07:27:03 crc kubenswrapper[4971]: I1213 07:27:03.644358 4971 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplane-ansible-ssh-private-key-secret" Dec 13 07:27:03 crc kubenswrapper[4971]: I1213 07:27:03.644545 4971 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-aee-default-env" Dec 13 07:27:03 crc kubenswrapper[4971]: I1213 07:27:03.644839 4971 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplanenodeset-openstack-edpm-ipam" Dec 13 07:27:03 crc kubenswrapper[4971]: I1213 07:27:03.644983 4971 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-edpm-ipam-dockercfg-zb2j5" Dec 13 07:27:03 crc kubenswrapper[4971]: I1213 07:27:03.656674 4971 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ceph-hci-pre-edpm-deployment-openstack-edpm-ipam-llgqk"] Dec 13 07:27:03 crc kubenswrapper[4971]: I1213 07:27:03.695932 4971 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/c437fb6f-411b-4af7-bf7c-3aad6306e4f2-inventory\") pod \"ceph-hci-pre-edpm-deployment-openstack-edpm-ipam-llgqk\" (UID: \"c437fb6f-411b-4af7-bf7c-3aad6306e4f2\") " pod="openstack/ceph-hci-pre-edpm-deployment-openstack-edpm-ipam-llgqk" Dec 13 07:27:03 crc kubenswrapper[4971]: I1213 07:27:03.696258 4971 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-szpcf\" (UniqueName: \"kubernetes.io/projected/c437fb6f-411b-4af7-bf7c-3aad6306e4f2-kube-api-access-szpcf\") pod \"ceph-hci-pre-edpm-deployment-openstack-edpm-ipam-llgqk\" (UID: \"c437fb6f-411b-4af7-bf7c-3aad6306e4f2\") " pod="openstack/ceph-hci-pre-edpm-deployment-openstack-edpm-ipam-llgqk" Dec 13 07:27:03 crc kubenswrapper[4971]: I1213 07:27:03.696384 4971 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/c437fb6f-411b-4af7-bf7c-3aad6306e4f2-ssh-key\") pod \"ceph-hci-pre-edpm-deployment-openstack-edpm-ipam-llgqk\" (UID: \"c437fb6f-411b-4af7-bf7c-3aad6306e4f2\") " pod="openstack/ceph-hci-pre-edpm-deployment-openstack-edpm-ipam-llgqk" Dec 13 07:27:03 crc kubenswrapper[4971]: I1213 07:27:03.797634 4971 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-szpcf\" (UniqueName: \"kubernetes.io/projected/c437fb6f-411b-4af7-bf7c-3aad6306e4f2-kube-api-access-szpcf\") pod \"ceph-hci-pre-edpm-deployment-openstack-edpm-ipam-llgqk\" (UID: \"c437fb6f-411b-4af7-bf7c-3aad6306e4f2\") " pod="openstack/ceph-hci-pre-edpm-deployment-openstack-edpm-ipam-llgqk" Dec 13 07:27:03 crc kubenswrapper[4971]: I1213 07:27:03.797716 4971 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/c437fb6f-411b-4af7-bf7c-3aad6306e4f2-ssh-key\") pod \"ceph-hci-pre-edpm-deployment-openstack-edpm-ipam-llgqk\" (UID: \"c437fb6f-411b-4af7-bf7c-3aad6306e4f2\") " pod="openstack/ceph-hci-pre-edpm-deployment-openstack-edpm-ipam-llgqk" Dec 13 07:27:03 crc kubenswrapper[4971]: I1213 07:27:03.797831 4971 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/c437fb6f-411b-4af7-bf7c-3aad6306e4f2-inventory\") pod \"ceph-hci-pre-edpm-deployment-openstack-edpm-ipam-llgqk\" (UID: \"c437fb6f-411b-4af7-bf7c-3aad6306e4f2\") " pod="openstack/ceph-hci-pre-edpm-deployment-openstack-edpm-ipam-llgqk" Dec 13 07:27:03 crc kubenswrapper[4971]: I1213 07:27:03.804642 4971 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/c437fb6f-411b-4af7-bf7c-3aad6306e4f2-ssh-key\") pod \"ceph-hci-pre-edpm-deployment-openstack-edpm-ipam-llgqk\" (UID: \"c437fb6f-411b-4af7-bf7c-3aad6306e4f2\") " pod="openstack/ceph-hci-pre-edpm-deployment-openstack-edpm-ipam-llgqk" Dec 13 07:27:03 crc kubenswrapper[4971]: I1213 07:27:03.804687 4971 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/c437fb6f-411b-4af7-bf7c-3aad6306e4f2-inventory\") pod \"ceph-hci-pre-edpm-deployment-openstack-edpm-ipam-llgqk\" (UID: \"c437fb6f-411b-4af7-bf7c-3aad6306e4f2\") " pod="openstack/ceph-hci-pre-edpm-deployment-openstack-edpm-ipam-llgqk" Dec 13 07:27:03 crc kubenswrapper[4971]: I1213 07:27:03.813166 4971 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-szpcf\" (UniqueName: \"kubernetes.io/projected/c437fb6f-411b-4af7-bf7c-3aad6306e4f2-kube-api-access-szpcf\") pod \"ceph-hci-pre-edpm-deployment-openstack-edpm-ipam-llgqk\" (UID: \"c437fb6f-411b-4af7-bf7c-3aad6306e4f2\") " pod="openstack/ceph-hci-pre-edpm-deployment-openstack-edpm-ipam-llgqk" Dec 13 07:27:03 crc kubenswrapper[4971]: I1213 07:27:03.961963 4971 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-edpm-ipam-dockercfg-zb2j5" Dec 13 07:27:03 crc kubenswrapper[4971]: I1213 07:27:03.970243 4971 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceph-hci-pre-edpm-deployment-openstack-edpm-ipam-llgqk" Dec 13 07:27:04 crc kubenswrapper[4971]: I1213 07:27:04.624877 4971 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ceph-hci-pre-edpm-deployment-openstack-edpm-ipam-llgqk"] Dec 13 07:27:04 crc kubenswrapper[4971]: I1213 07:27:04.769672 4971 scope.go:117] "RemoveContainer" containerID="ec2e4b1312b5d5d0545354a836fef4af376ca3b34327ab1b9e0b72c102976079" Dec 13 07:27:04 crc kubenswrapper[4971]: E1213 07:27:04.770145 4971 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-82xjz_openshift-machine-config-operator(e0c4af12-edda-41d7-9b44-f87396174ef2)\"" pod="openshift-machine-config-operator/machine-config-daemon-82xjz" podUID="e0c4af12-edda-41d7-9b44-f87396174ef2" Dec 13 07:27:05 crc kubenswrapper[4971]: I1213 07:27:05.530220 4971 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceph-hci-pre-edpm-deployment-openstack-edpm-ipam-llgqk" event={"ID":"c437fb6f-411b-4af7-bf7c-3aad6306e4f2","Type":"ContainerStarted","Data":"fcdab57d739fa73388e766ab9160f98e76d7e3e4e634d391f8753e62ee5e191e"} Dec 13 07:27:08 crc kubenswrapper[4971]: I1213 07:27:08.589024 4971 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-aee-default-env" Dec 13 07:27:09 crc kubenswrapper[4971]: I1213 07:27:09.567170 4971 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceph-hci-pre-edpm-deployment-openstack-edpm-ipam-llgqk" event={"ID":"c437fb6f-411b-4af7-bf7c-3aad6306e4f2","Type":"ContainerStarted","Data":"b64b8d40637b7c78fb325e0b9ef34c5a9d5b3db7ad47fdd2f1fb42c0e8af00ad"} Dec 13 07:27:09 crc kubenswrapper[4971]: I1213 07:27:09.596709 4971 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/ceph-hci-pre-edpm-deployment-openstack-edpm-ipam-llgqk" podStartSLOduration=2.6595619409999998 podStartE2EDuration="6.596690181s" podCreationTimestamp="2025-12-13 07:27:03 +0000 UTC" firstStartedPulling="2025-12-13 07:27:04.635053695 +0000 UTC m=+2281.239463153" lastFinishedPulling="2025-12-13 07:27:08.572181905 +0000 UTC m=+2285.176591393" observedRunningTime="2025-12-13 07:27:09.587061474 +0000 UTC m=+2286.191470962" watchObservedRunningTime="2025-12-13 07:27:09.596690181 +0000 UTC m=+2286.201099629" Dec 13 07:27:17 crc kubenswrapper[4971]: I1213 07:27:17.655613 4971 generic.go:334] "Generic (PLEG): container finished" podID="c437fb6f-411b-4af7-bf7c-3aad6306e4f2" containerID="b64b8d40637b7c78fb325e0b9ef34c5a9d5b3db7ad47fdd2f1fb42c0e8af00ad" exitCode=0 Dec 13 07:27:17 crc kubenswrapper[4971]: I1213 07:27:17.656108 4971 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceph-hci-pre-edpm-deployment-openstack-edpm-ipam-llgqk" event={"ID":"c437fb6f-411b-4af7-bf7c-3aad6306e4f2","Type":"ContainerDied","Data":"b64b8d40637b7c78fb325e0b9ef34c5a9d5b3db7ad47fdd2f1fb42c0e8af00ad"} Dec 13 07:27:18 crc kubenswrapper[4971]: I1213 07:27:18.769401 4971 scope.go:117] "RemoveContainer" containerID="ec2e4b1312b5d5d0545354a836fef4af376ca3b34327ab1b9e0b72c102976079" Dec 13 07:27:18 crc kubenswrapper[4971]: E1213 07:27:18.770177 4971 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-82xjz_openshift-machine-config-operator(e0c4af12-edda-41d7-9b44-f87396174ef2)\"" pod="openshift-machine-config-operator/machine-config-daemon-82xjz" podUID="e0c4af12-edda-41d7-9b44-f87396174ef2" Dec 13 07:27:19 crc kubenswrapper[4971]: I1213 07:27:19.229064 4971 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ceph-hci-pre-edpm-deployment-openstack-edpm-ipam-llgqk" Dec 13 07:27:19 crc kubenswrapper[4971]: I1213 07:27:19.252501 4971 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/c437fb6f-411b-4af7-bf7c-3aad6306e4f2-ssh-key\") pod \"c437fb6f-411b-4af7-bf7c-3aad6306e4f2\" (UID: \"c437fb6f-411b-4af7-bf7c-3aad6306e4f2\") " Dec 13 07:27:19 crc kubenswrapper[4971]: I1213 07:27:19.252781 4971 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/c437fb6f-411b-4af7-bf7c-3aad6306e4f2-inventory\") pod \"c437fb6f-411b-4af7-bf7c-3aad6306e4f2\" (UID: \"c437fb6f-411b-4af7-bf7c-3aad6306e4f2\") " Dec 13 07:27:19 crc kubenswrapper[4971]: I1213 07:27:19.252832 4971 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-szpcf\" (UniqueName: \"kubernetes.io/projected/c437fb6f-411b-4af7-bf7c-3aad6306e4f2-kube-api-access-szpcf\") pod \"c437fb6f-411b-4af7-bf7c-3aad6306e4f2\" (UID: \"c437fb6f-411b-4af7-bf7c-3aad6306e4f2\") " Dec 13 07:27:19 crc kubenswrapper[4971]: I1213 07:27:19.272120 4971 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/c437fb6f-411b-4af7-bf7c-3aad6306e4f2-kube-api-access-szpcf" (OuterVolumeSpecName: "kube-api-access-szpcf") pod "c437fb6f-411b-4af7-bf7c-3aad6306e4f2" (UID: "c437fb6f-411b-4af7-bf7c-3aad6306e4f2"). InnerVolumeSpecName "kube-api-access-szpcf". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 13 07:27:19 crc kubenswrapper[4971]: I1213 07:27:19.288958 4971 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/c437fb6f-411b-4af7-bf7c-3aad6306e4f2-inventory" (OuterVolumeSpecName: "inventory") pod "c437fb6f-411b-4af7-bf7c-3aad6306e4f2" (UID: "c437fb6f-411b-4af7-bf7c-3aad6306e4f2"). InnerVolumeSpecName "inventory". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 13 07:27:19 crc kubenswrapper[4971]: I1213 07:27:19.295438 4971 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/c437fb6f-411b-4af7-bf7c-3aad6306e4f2-ssh-key" (OuterVolumeSpecName: "ssh-key") pod "c437fb6f-411b-4af7-bf7c-3aad6306e4f2" (UID: "c437fb6f-411b-4af7-bf7c-3aad6306e4f2"). InnerVolumeSpecName "ssh-key". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 13 07:27:19 crc kubenswrapper[4971]: I1213 07:27:19.355156 4971 reconciler_common.go:293] "Volume detached for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/c437fb6f-411b-4af7-bf7c-3aad6306e4f2-inventory\") on node \"crc\" DevicePath \"\"" Dec 13 07:27:19 crc kubenswrapper[4971]: I1213 07:27:19.355207 4971 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-szpcf\" (UniqueName: \"kubernetes.io/projected/c437fb6f-411b-4af7-bf7c-3aad6306e4f2-kube-api-access-szpcf\") on node \"crc\" DevicePath \"\"" Dec 13 07:27:19 crc kubenswrapper[4971]: I1213 07:27:19.355219 4971 reconciler_common.go:293] "Volume detached for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/c437fb6f-411b-4af7-bf7c-3aad6306e4f2-ssh-key\") on node \"crc\" DevicePath \"\"" Dec 13 07:27:19 crc kubenswrapper[4971]: I1213 07:27:19.678708 4971 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceph-hci-pre-edpm-deployment-openstack-edpm-ipam-llgqk" event={"ID":"c437fb6f-411b-4af7-bf7c-3aad6306e4f2","Type":"ContainerDied","Data":"fcdab57d739fa73388e766ab9160f98e76d7e3e4e634d391f8753e62ee5e191e"} Dec 13 07:27:19 crc kubenswrapper[4971]: I1213 07:27:19.678757 4971 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="fcdab57d739fa73388e766ab9160f98e76d7e3e4e634d391f8753e62ee5e191e" Dec 13 07:27:19 crc kubenswrapper[4971]: I1213 07:27:19.678822 4971 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ceph-hci-pre-edpm-deployment-openstack-edpm-ipam-llgqk" Dec 13 07:27:19 crc kubenswrapper[4971]: I1213 07:27:19.739871 4971 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/configure-os-edpm-deployment-openstack-edpm-ipam-zrl24"] Dec 13 07:27:19 crc kubenswrapper[4971]: E1213 07:27:19.740213 4971 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="c437fb6f-411b-4af7-bf7c-3aad6306e4f2" containerName="ceph-hci-pre-edpm-deployment-openstack-edpm-ipam" Dec 13 07:27:19 crc kubenswrapper[4971]: I1213 07:27:19.740234 4971 state_mem.go:107] "Deleted CPUSet assignment" podUID="c437fb6f-411b-4af7-bf7c-3aad6306e4f2" containerName="ceph-hci-pre-edpm-deployment-openstack-edpm-ipam" Dec 13 07:27:19 crc kubenswrapper[4971]: I1213 07:27:19.740409 4971 memory_manager.go:354] "RemoveStaleState removing state" podUID="c437fb6f-411b-4af7-bf7c-3aad6306e4f2" containerName="ceph-hci-pre-edpm-deployment-openstack-edpm-ipam" Dec 13 07:27:19 crc kubenswrapper[4971]: I1213 07:27:19.740999 4971 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-zrl24" Dec 13 07:27:19 crc kubenswrapper[4971]: I1213 07:27:19.744654 4971 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplanenodeset-openstack-edpm-ipam" Dec 13 07:27:19 crc kubenswrapper[4971]: I1213 07:27:19.744897 4971 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-edpm-ipam-dockercfg-zb2j5" Dec 13 07:27:19 crc kubenswrapper[4971]: I1213 07:27:19.745049 4971 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-aee-default-env" Dec 13 07:27:19 crc kubenswrapper[4971]: I1213 07:27:19.745201 4971 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplane-ansible-ssh-private-key-secret" Dec 13 07:27:19 crc kubenswrapper[4971]: I1213 07:27:19.751482 4971 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/configure-os-edpm-deployment-openstack-edpm-ipam-zrl24"] Dec 13 07:27:19 crc kubenswrapper[4971]: I1213 07:27:19.766738 4971 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/912b00fe-da52-454a-924c-e49620ecd7fd-ssh-key\") pod \"configure-os-edpm-deployment-openstack-edpm-ipam-zrl24\" (UID: \"912b00fe-da52-454a-924c-e49620ecd7fd\") " pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-zrl24" Dec 13 07:27:19 crc kubenswrapper[4971]: I1213 07:27:19.767170 4971 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/912b00fe-da52-454a-924c-e49620ecd7fd-inventory\") pod \"configure-os-edpm-deployment-openstack-edpm-ipam-zrl24\" (UID: \"912b00fe-da52-454a-924c-e49620ecd7fd\") " pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-zrl24" Dec 13 07:27:19 crc kubenswrapper[4971]: I1213 07:27:19.868430 4971 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-vmdlk\" (UniqueName: \"kubernetes.io/projected/912b00fe-da52-454a-924c-e49620ecd7fd-kube-api-access-vmdlk\") pod \"configure-os-edpm-deployment-openstack-edpm-ipam-zrl24\" (UID: \"912b00fe-da52-454a-924c-e49620ecd7fd\") " pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-zrl24" Dec 13 07:27:19 crc kubenswrapper[4971]: I1213 07:27:19.868562 4971 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/912b00fe-da52-454a-924c-e49620ecd7fd-ssh-key\") pod \"configure-os-edpm-deployment-openstack-edpm-ipam-zrl24\" (UID: \"912b00fe-da52-454a-924c-e49620ecd7fd\") " pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-zrl24" Dec 13 07:27:19 crc kubenswrapper[4971]: I1213 07:27:19.868746 4971 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/912b00fe-da52-454a-924c-e49620ecd7fd-inventory\") pod \"configure-os-edpm-deployment-openstack-edpm-ipam-zrl24\" (UID: \"912b00fe-da52-454a-924c-e49620ecd7fd\") " pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-zrl24" Dec 13 07:27:19 crc kubenswrapper[4971]: I1213 07:27:19.877708 4971 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/912b00fe-da52-454a-924c-e49620ecd7fd-inventory\") pod \"configure-os-edpm-deployment-openstack-edpm-ipam-zrl24\" (UID: \"912b00fe-da52-454a-924c-e49620ecd7fd\") " pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-zrl24" Dec 13 07:27:19 crc kubenswrapper[4971]: I1213 07:27:19.877853 4971 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/912b00fe-da52-454a-924c-e49620ecd7fd-ssh-key\") pod \"configure-os-edpm-deployment-openstack-edpm-ipam-zrl24\" (UID: \"912b00fe-da52-454a-924c-e49620ecd7fd\") " pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-zrl24" Dec 13 07:27:19 crc kubenswrapper[4971]: I1213 07:27:19.971007 4971 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-vmdlk\" (UniqueName: \"kubernetes.io/projected/912b00fe-da52-454a-924c-e49620ecd7fd-kube-api-access-vmdlk\") pod \"configure-os-edpm-deployment-openstack-edpm-ipam-zrl24\" (UID: \"912b00fe-da52-454a-924c-e49620ecd7fd\") " pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-zrl24" Dec 13 07:27:19 crc kubenswrapper[4971]: I1213 07:27:19.985893 4971 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-vmdlk\" (UniqueName: \"kubernetes.io/projected/912b00fe-da52-454a-924c-e49620ecd7fd-kube-api-access-vmdlk\") pod \"configure-os-edpm-deployment-openstack-edpm-ipam-zrl24\" (UID: \"912b00fe-da52-454a-924c-e49620ecd7fd\") " pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-zrl24" Dec 13 07:27:20 crc kubenswrapper[4971]: I1213 07:27:20.056564 4971 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-zrl24" Dec 13 07:27:20 crc kubenswrapper[4971]: I1213 07:27:20.495762 4971 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/configure-os-edpm-deployment-openstack-edpm-ipam-zrl24"] Dec 13 07:27:20 crc kubenswrapper[4971]: I1213 07:27:20.691277 4971 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-zrl24" event={"ID":"912b00fe-da52-454a-924c-e49620ecd7fd","Type":"ContainerStarted","Data":"d548c820eb2311ad4829cf2007c699ca9818dba2b11cb4f7e57526ee76c006cb"} Dec 13 07:27:32 crc kubenswrapper[4971]: I1213 07:27:32.768436 4971 scope.go:117] "RemoveContainer" containerID="ec2e4b1312b5d5d0545354a836fef4af376ca3b34327ab1b9e0b72c102976079" Dec 13 07:27:32 crc kubenswrapper[4971]: E1213 07:27:32.769463 4971 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-82xjz_openshift-machine-config-operator(e0c4af12-edda-41d7-9b44-f87396174ef2)\"" pod="openshift-machine-config-operator/machine-config-daemon-82xjz" podUID="e0c4af12-edda-41d7-9b44-f87396174ef2" Dec 13 07:27:32 crc kubenswrapper[4971]: I1213 07:27:32.940371 4971 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-zrl24" event={"ID":"912b00fe-da52-454a-924c-e49620ecd7fd","Type":"ContainerStarted","Data":"0b61890aecd8dd55185aba04582cb72fabb37f4ccbe4d8a42f4ca0117703e827"} Dec 13 07:27:33 crc kubenswrapper[4971]: I1213 07:27:33.973819 4971 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-zrl24" podStartSLOduration=4.76985392 podStartE2EDuration="14.973801189s" podCreationTimestamp="2025-12-13 07:27:19 +0000 UTC" firstStartedPulling="2025-12-13 07:27:20.510890891 +0000 UTC m=+2297.115300379" lastFinishedPulling="2025-12-13 07:27:30.71483816 +0000 UTC m=+2307.319247648" observedRunningTime="2025-12-13 07:27:33.964801968 +0000 UTC m=+2310.569211426" watchObservedRunningTime="2025-12-13 07:27:33.973801189 +0000 UTC m=+2310.578210637" Dec 13 07:27:44 crc kubenswrapper[4971]: I1213 07:27:44.768067 4971 scope.go:117] "RemoveContainer" containerID="ec2e4b1312b5d5d0545354a836fef4af376ca3b34327ab1b9e0b72c102976079" Dec 13 07:27:44 crc kubenswrapper[4971]: E1213 07:27:44.768671 4971 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-82xjz_openshift-machine-config-operator(e0c4af12-edda-41d7-9b44-f87396174ef2)\"" pod="openshift-machine-config-operator/machine-config-daemon-82xjz" podUID="e0c4af12-edda-41d7-9b44-f87396174ef2" Dec 13 07:27:58 crc kubenswrapper[4971]: I1213 07:27:58.768840 4971 scope.go:117] "RemoveContainer" containerID="ec2e4b1312b5d5d0545354a836fef4af376ca3b34327ab1b9e0b72c102976079" Dec 13 07:27:58 crc kubenswrapper[4971]: E1213 07:27:58.769595 4971 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-82xjz_openshift-machine-config-operator(e0c4af12-edda-41d7-9b44-f87396174ef2)\"" pod="openshift-machine-config-operator/machine-config-daemon-82xjz" podUID="e0c4af12-edda-41d7-9b44-f87396174ef2" Dec 13 07:28:12 crc kubenswrapper[4971]: I1213 07:28:12.768193 4971 scope.go:117] "RemoveContainer" containerID="ec2e4b1312b5d5d0545354a836fef4af376ca3b34327ab1b9e0b72c102976079" Dec 13 07:28:12 crc kubenswrapper[4971]: E1213 07:28:12.771139 4971 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-82xjz_openshift-machine-config-operator(e0c4af12-edda-41d7-9b44-f87396174ef2)\"" pod="openshift-machine-config-operator/machine-config-daemon-82xjz" podUID="e0c4af12-edda-41d7-9b44-f87396174ef2" Dec 13 07:28:23 crc kubenswrapper[4971]: I1213 07:28:23.778585 4971 scope.go:117] "RemoveContainer" containerID="ec2e4b1312b5d5d0545354a836fef4af376ca3b34327ab1b9e0b72c102976079" Dec 13 07:28:23 crc kubenswrapper[4971]: E1213 07:28:23.779420 4971 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-82xjz_openshift-machine-config-operator(e0c4af12-edda-41d7-9b44-f87396174ef2)\"" pod="openshift-machine-config-operator/machine-config-daemon-82xjz" podUID="e0c4af12-edda-41d7-9b44-f87396174ef2" Dec 13 07:28:33 crc kubenswrapper[4971]: I1213 07:28:33.518621 4971 generic.go:334] "Generic (PLEG): container finished" podID="912b00fe-da52-454a-924c-e49620ecd7fd" containerID="0b61890aecd8dd55185aba04582cb72fabb37f4ccbe4d8a42f4ca0117703e827" exitCode=0 Dec 13 07:28:33 crc kubenswrapper[4971]: I1213 07:28:33.518686 4971 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-zrl24" event={"ID":"912b00fe-da52-454a-924c-e49620ecd7fd","Type":"ContainerDied","Data":"0b61890aecd8dd55185aba04582cb72fabb37f4ccbe4d8a42f4ca0117703e827"} Dec 13 07:28:34 crc kubenswrapper[4971]: I1213 07:28:34.928234 4971 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-zrl24" Dec 13 07:28:35 crc kubenswrapper[4971]: I1213 07:28:35.125761 4971 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/912b00fe-da52-454a-924c-e49620ecd7fd-inventory\") pod \"912b00fe-da52-454a-924c-e49620ecd7fd\" (UID: \"912b00fe-da52-454a-924c-e49620ecd7fd\") " Dec 13 07:28:35 crc kubenswrapper[4971]: I1213 07:28:35.125863 4971 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-vmdlk\" (UniqueName: \"kubernetes.io/projected/912b00fe-da52-454a-924c-e49620ecd7fd-kube-api-access-vmdlk\") pod \"912b00fe-da52-454a-924c-e49620ecd7fd\" (UID: \"912b00fe-da52-454a-924c-e49620ecd7fd\") " Dec 13 07:28:35 crc kubenswrapper[4971]: I1213 07:28:35.126021 4971 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/912b00fe-da52-454a-924c-e49620ecd7fd-ssh-key\") pod \"912b00fe-da52-454a-924c-e49620ecd7fd\" (UID: \"912b00fe-da52-454a-924c-e49620ecd7fd\") " Dec 13 07:28:35 crc kubenswrapper[4971]: I1213 07:28:35.132154 4971 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/912b00fe-da52-454a-924c-e49620ecd7fd-kube-api-access-vmdlk" (OuterVolumeSpecName: "kube-api-access-vmdlk") pod "912b00fe-da52-454a-924c-e49620ecd7fd" (UID: "912b00fe-da52-454a-924c-e49620ecd7fd"). InnerVolumeSpecName "kube-api-access-vmdlk". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 13 07:28:35 crc kubenswrapper[4971]: I1213 07:28:35.156285 4971 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/912b00fe-da52-454a-924c-e49620ecd7fd-inventory" (OuterVolumeSpecName: "inventory") pod "912b00fe-da52-454a-924c-e49620ecd7fd" (UID: "912b00fe-da52-454a-924c-e49620ecd7fd"). InnerVolumeSpecName "inventory". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 13 07:28:35 crc kubenswrapper[4971]: I1213 07:28:35.164187 4971 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/912b00fe-da52-454a-924c-e49620ecd7fd-ssh-key" (OuterVolumeSpecName: "ssh-key") pod "912b00fe-da52-454a-924c-e49620ecd7fd" (UID: "912b00fe-da52-454a-924c-e49620ecd7fd"). InnerVolumeSpecName "ssh-key". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 13 07:28:35 crc kubenswrapper[4971]: I1213 07:28:35.227566 4971 reconciler_common.go:293] "Volume detached for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/912b00fe-da52-454a-924c-e49620ecd7fd-ssh-key\") on node \"crc\" DevicePath \"\"" Dec 13 07:28:35 crc kubenswrapper[4971]: I1213 07:28:35.227590 4971 reconciler_common.go:293] "Volume detached for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/912b00fe-da52-454a-924c-e49620ecd7fd-inventory\") on node \"crc\" DevicePath \"\"" Dec 13 07:28:35 crc kubenswrapper[4971]: I1213 07:28:35.227603 4971 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-vmdlk\" (UniqueName: \"kubernetes.io/projected/912b00fe-da52-454a-924c-e49620ecd7fd-kube-api-access-vmdlk\") on node \"crc\" DevicePath \"\"" Dec 13 07:28:35 crc kubenswrapper[4971]: I1213 07:28:35.582204 4971 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-zrl24" event={"ID":"912b00fe-da52-454a-924c-e49620ecd7fd","Type":"ContainerDied","Data":"d548c820eb2311ad4829cf2007c699ca9818dba2b11cb4f7e57526ee76c006cb"} Dec 13 07:28:35 crc kubenswrapper[4971]: I1213 07:28:35.582250 4971 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-zrl24" Dec 13 07:28:35 crc kubenswrapper[4971]: I1213 07:28:35.582255 4971 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="d548c820eb2311ad4829cf2007c699ca9818dba2b11cb4f7e57526ee76c006cb" Dec 13 07:28:35 crc kubenswrapper[4971]: I1213 07:28:35.665332 4971 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ssh-known-hosts-edpm-deployment-2stjx"] Dec 13 07:28:35 crc kubenswrapper[4971]: E1213 07:28:35.665735 4971 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="912b00fe-da52-454a-924c-e49620ecd7fd" containerName="configure-os-edpm-deployment-openstack-edpm-ipam" Dec 13 07:28:35 crc kubenswrapper[4971]: I1213 07:28:35.665761 4971 state_mem.go:107] "Deleted CPUSet assignment" podUID="912b00fe-da52-454a-924c-e49620ecd7fd" containerName="configure-os-edpm-deployment-openstack-edpm-ipam" Dec 13 07:28:35 crc kubenswrapper[4971]: I1213 07:28:35.665981 4971 memory_manager.go:354] "RemoveStaleState removing state" podUID="912b00fe-da52-454a-924c-e49620ecd7fd" containerName="configure-os-edpm-deployment-openstack-edpm-ipam" Dec 13 07:28:35 crc kubenswrapper[4971]: I1213 07:28:35.667695 4971 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ssh-known-hosts-edpm-deployment-2stjx" Dec 13 07:28:35 crc kubenswrapper[4971]: I1213 07:28:35.670344 4971 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-edpm-ipam-dockercfg-zb2j5" Dec 13 07:28:35 crc kubenswrapper[4971]: I1213 07:28:35.670533 4971 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-aee-default-env" Dec 13 07:28:35 crc kubenswrapper[4971]: I1213 07:28:35.670772 4971 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplanenodeset-openstack-edpm-ipam" Dec 13 07:28:35 crc kubenswrapper[4971]: I1213 07:28:35.671075 4971 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplane-ansible-ssh-private-key-secret" Dec 13 07:28:35 crc kubenswrapper[4971]: I1213 07:28:35.680487 4971 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ssh-known-hosts-edpm-deployment-2stjx"] Dec 13 07:28:35 crc kubenswrapper[4971]: I1213 07:28:35.838863 4971 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-l88mx\" (UniqueName: \"kubernetes.io/projected/be382cbb-023b-4d16-9e01-b2eb44663da2-kube-api-access-l88mx\") pod \"ssh-known-hosts-edpm-deployment-2stjx\" (UID: \"be382cbb-023b-4d16-9e01-b2eb44663da2\") " pod="openstack/ssh-known-hosts-edpm-deployment-2stjx" Dec 13 07:28:35 crc kubenswrapper[4971]: I1213 07:28:35.838932 4971 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key-openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/secret/be382cbb-023b-4d16-9e01-b2eb44663da2-ssh-key-openstack-edpm-ipam\") pod \"ssh-known-hosts-edpm-deployment-2stjx\" (UID: \"be382cbb-023b-4d16-9e01-b2eb44663da2\") " pod="openstack/ssh-known-hosts-edpm-deployment-2stjx" Dec 13 07:28:35 crc kubenswrapper[4971]: I1213 07:28:35.839168 4971 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"inventory-0\" (UniqueName: \"kubernetes.io/secret/be382cbb-023b-4d16-9e01-b2eb44663da2-inventory-0\") pod \"ssh-known-hosts-edpm-deployment-2stjx\" (UID: \"be382cbb-023b-4d16-9e01-b2eb44663da2\") " pod="openstack/ssh-known-hosts-edpm-deployment-2stjx" Dec 13 07:28:35 crc kubenswrapper[4971]: I1213 07:28:35.941102 4971 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-l88mx\" (UniqueName: \"kubernetes.io/projected/be382cbb-023b-4d16-9e01-b2eb44663da2-kube-api-access-l88mx\") pod \"ssh-known-hosts-edpm-deployment-2stjx\" (UID: \"be382cbb-023b-4d16-9e01-b2eb44663da2\") " pod="openstack/ssh-known-hosts-edpm-deployment-2stjx" Dec 13 07:28:35 crc kubenswrapper[4971]: I1213 07:28:35.941219 4971 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key-openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/secret/be382cbb-023b-4d16-9e01-b2eb44663da2-ssh-key-openstack-edpm-ipam\") pod \"ssh-known-hosts-edpm-deployment-2stjx\" (UID: \"be382cbb-023b-4d16-9e01-b2eb44663da2\") " pod="openstack/ssh-known-hosts-edpm-deployment-2stjx" Dec 13 07:28:35 crc kubenswrapper[4971]: I1213 07:28:35.941352 4971 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"inventory-0\" (UniqueName: \"kubernetes.io/secret/be382cbb-023b-4d16-9e01-b2eb44663da2-inventory-0\") pod \"ssh-known-hosts-edpm-deployment-2stjx\" (UID: \"be382cbb-023b-4d16-9e01-b2eb44663da2\") " pod="openstack/ssh-known-hosts-edpm-deployment-2stjx" Dec 13 07:28:35 crc kubenswrapper[4971]: I1213 07:28:35.947867 4971 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"inventory-0\" (UniqueName: \"kubernetes.io/secret/be382cbb-023b-4d16-9e01-b2eb44663da2-inventory-0\") pod \"ssh-known-hosts-edpm-deployment-2stjx\" (UID: \"be382cbb-023b-4d16-9e01-b2eb44663da2\") " pod="openstack/ssh-known-hosts-edpm-deployment-2stjx" Dec 13 07:28:35 crc kubenswrapper[4971]: I1213 07:28:35.952397 4971 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key-openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/secret/be382cbb-023b-4d16-9e01-b2eb44663da2-ssh-key-openstack-edpm-ipam\") pod \"ssh-known-hosts-edpm-deployment-2stjx\" (UID: \"be382cbb-023b-4d16-9e01-b2eb44663da2\") " pod="openstack/ssh-known-hosts-edpm-deployment-2stjx" Dec 13 07:28:35 crc kubenswrapper[4971]: I1213 07:28:35.960691 4971 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-l88mx\" (UniqueName: \"kubernetes.io/projected/be382cbb-023b-4d16-9e01-b2eb44663da2-kube-api-access-l88mx\") pod \"ssh-known-hosts-edpm-deployment-2stjx\" (UID: \"be382cbb-023b-4d16-9e01-b2eb44663da2\") " pod="openstack/ssh-known-hosts-edpm-deployment-2stjx" Dec 13 07:28:35 crc kubenswrapper[4971]: I1213 07:28:35.994070 4971 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ssh-known-hosts-edpm-deployment-2stjx" Dec 13 07:28:36 crc kubenswrapper[4971]: I1213 07:28:36.727373 4971 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ssh-known-hosts-edpm-deployment-2stjx"] Dec 13 07:28:37 crc kubenswrapper[4971]: I1213 07:28:37.602310 4971 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ssh-known-hosts-edpm-deployment-2stjx" event={"ID":"be382cbb-023b-4d16-9e01-b2eb44663da2","Type":"ContainerStarted","Data":"9926e510c6ad6805e77b89affc083a7f8b2bd984e0c63a1919c2f412e158d433"} Dec 13 07:28:37 crc kubenswrapper[4971]: I1213 07:28:37.768946 4971 scope.go:117] "RemoveContainer" containerID="ec2e4b1312b5d5d0545354a836fef4af376ca3b34327ab1b9e0b72c102976079" Dec 13 07:28:37 crc kubenswrapper[4971]: E1213 07:28:37.769629 4971 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-82xjz_openshift-machine-config-operator(e0c4af12-edda-41d7-9b44-f87396174ef2)\"" pod="openshift-machine-config-operator/machine-config-daemon-82xjz" podUID="e0c4af12-edda-41d7-9b44-f87396174ef2" Dec 13 07:28:38 crc kubenswrapper[4971]: I1213 07:28:38.612509 4971 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ssh-known-hosts-edpm-deployment-2stjx" event={"ID":"be382cbb-023b-4d16-9e01-b2eb44663da2","Type":"ContainerStarted","Data":"8e528e869a31af117915d49484380c02654b1d3b2432c0d6edd12bc017ba2b67"} Dec 13 07:28:38 crc kubenswrapper[4971]: I1213 07:28:38.637403 4971 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/ssh-known-hosts-edpm-deployment-2stjx" podStartSLOduration=2.93602995 podStartE2EDuration="3.637382391s" podCreationTimestamp="2025-12-13 07:28:35 +0000 UTC" firstStartedPulling="2025-12-13 07:28:36.727437974 +0000 UTC m=+2373.331847432" lastFinishedPulling="2025-12-13 07:28:37.428790405 +0000 UTC m=+2374.033199873" observedRunningTime="2025-12-13 07:28:38.634815878 +0000 UTC m=+2375.239225336" watchObservedRunningTime="2025-12-13 07:28:38.637382391 +0000 UTC m=+2375.241791849" Dec 13 07:28:46 crc kubenswrapper[4971]: I1213 07:28:46.688419 4971 generic.go:334] "Generic (PLEG): container finished" podID="be382cbb-023b-4d16-9e01-b2eb44663da2" containerID="8e528e869a31af117915d49484380c02654b1d3b2432c0d6edd12bc017ba2b67" exitCode=0 Dec 13 07:28:46 crc kubenswrapper[4971]: I1213 07:28:46.688646 4971 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ssh-known-hosts-edpm-deployment-2stjx" event={"ID":"be382cbb-023b-4d16-9e01-b2eb44663da2","Type":"ContainerDied","Data":"8e528e869a31af117915d49484380c02654b1d3b2432c0d6edd12bc017ba2b67"} Dec 13 07:28:48 crc kubenswrapper[4971]: I1213 07:28:48.148592 4971 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ssh-known-hosts-edpm-deployment-2stjx" Dec 13 07:28:48 crc kubenswrapper[4971]: I1213 07:28:48.280291 4971 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key-openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/secret/be382cbb-023b-4d16-9e01-b2eb44663da2-ssh-key-openstack-edpm-ipam\") pod \"be382cbb-023b-4d16-9e01-b2eb44663da2\" (UID: \"be382cbb-023b-4d16-9e01-b2eb44663da2\") " Dec 13 07:28:48 crc kubenswrapper[4971]: I1213 07:28:48.280455 4971 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-l88mx\" (UniqueName: \"kubernetes.io/projected/be382cbb-023b-4d16-9e01-b2eb44663da2-kube-api-access-l88mx\") pod \"be382cbb-023b-4d16-9e01-b2eb44663da2\" (UID: \"be382cbb-023b-4d16-9e01-b2eb44663da2\") " Dec 13 07:28:48 crc kubenswrapper[4971]: I1213 07:28:48.280553 4971 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"inventory-0\" (UniqueName: \"kubernetes.io/secret/be382cbb-023b-4d16-9e01-b2eb44663da2-inventory-0\") pod \"be382cbb-023b-4d16-9e01-b2eb44663da2\" (UID: \"be382cbb-023b-4d16-9e01-b2eb44663da2\") " Dec 13 07:28:48 crc kubenswrapper[4971]: I1213 07:28:48.288826 4971 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/be382cbb-023b-4d16-9e01-b2eb44663da2-kube-api-access-l88mx" (OuterVolumeSpecName: "kube-api-access-l88mx") pod "be382cbb-023b-4d16-9e01-b2eb44663da2" (UID: "be382cbb-023b-4d16-9e01-b2eb44663da2"). InnerVolumeSpecName "kube-api-access-l88mx". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 13 07:28:48 crc kubenswrapper[4971]: I1213 07:28:48.329094 4971 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/be382cbb-023b-4d16-9e01-b2eb44663da2-inventory-0" (OuterVolumeSpecName: "inventory-0") pod "be382cbb-023b-4d16-9e01-b2eb44663da2" (UID: "be382cbb-023b-4d16-9e01-b2eb44663da2"). InnerVolumeSpecName "inventory-0". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 13 07:28:48 crc kubenswrapper[4971]: I1213 07:28:48.329704 4971 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/be382cbb-023b-4d16-9e01-b2eb44663da2-ssh-key-openstack-edpm-ipam" (OuterVolumeSpecName: "ssh-key-openstack-edpm-ipam") pod "be382cbb-023b-4d16-9e01-b2eb44663da2" (UID: "be382cbb-023b-4d16-9e01-b2eb44663da2"). InnerVolumeSpecName "ssh-key-openstack-edpm-ipam". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 13 07:28:48 crc kubenswrapper[4971]: I1213 07:28:48.383184 4971 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-l88mx\" (UniqueName: \"kubernetes.io/projected/be382cbb-023b-4d16-9e01-b2eb44663da2-kube-api-access-l88mx\") on node \"crc\" DevicePath \"\"" Dec 13 07:28:48 crc kubenswrapper[4971]: I1213 07:28:48.383241 4971 reconciler_common.go:293] "Volume detached for volume \"inventory-0\" (UniqueName: \"kubernetes.io/secret/be382cbb-023b-4d16-9e01-b2eb44663da2-inventory-0\") on node \"crc\" DevicePath \"\"" Dec 13 07:28:48 crc kubenswrapper[4971]: I1213 07:28:48.383263 4971 reconciler_common.go:293] "Volume detached for volume \"ssh-key-openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/secret/be382cbb-023b-4d16-9e01-b2eb44663da2-ssh-key-openstack-edpm-ipam\") on node \"crc\" DevicePath \"\"" Dec 13 07:28:48 crc kubenswrapper[4971]: I1213 07:28:48.735382 4971 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ssh-known-hosts-edpm-deployment-2stjx" event={"ID":"be382cbb-023b-4d16-9e01-b2eb44663da2","Type":"ContainerDied","Data":"9926e510c6ad6805e77b89affc083a7f8b2bd984e0c63a1919c2f412e158d433"} Dec 13 07:28:48 crc kubenswrapper[4971]: I1213 07:28:48.735795 4971 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="9926e510c6ad6805e77b89affc083a7f8b2bd984e0c63a1919c2f412e158d433" Dec 13 07:28:48 crc kubenswrapper[4971]: I1213 07:28:48.735547 4971 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ssh-known-hosts-edpm-deployment-2stjx" Dec 13 07:28:48 crc kubenswrapper[4971]: I1213 07:28:48.937412 4971 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/run-os-edpm-deployment-openstack-edpm-ipam-dfs5x"] Dec 13 07:28:48 crc kubenswrapper[4971]: E1213 07:28:48.937992 4971 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="be382cbb-023b-4d16-9e01-b2eb44663da2" containerName="ssh-known-hosts-edpm-deployment" Dec 13 07:28:48 crc kubenswrapper[4971]: I1213 07:28:48.938017 4971 state_mem.go:107] "Deleted CPUSet assignment" podUID="be382cbb-023b-4d16-9e01-b2eb44663da2" containerName="ssh-known-hosts-edpm-deployment" Dec 13 07:28:48 crc kubenswrapper[4971]: I1213 07:28:48.938278 4971 memory_manager.go:354] "RemoveStaleState removing state" podUID="be382cbb-023b-4d16-9e01-b2eb44663da2" containerName="ssh-known-hosts-edpm-deployment" Dec 13 07:28:48 crc kubenswrapper[4971]: I1213 07:28:48.939140 4971 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/run-os-edpm-deployment-openstack-edpm-ipam-dfs5x" Dec 13 07:28:48 crc kubenswrapper[4971]: I1213 07:28:48.945985 4971 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-aee-default-env" Dec 13 07:28:48 crc kubenswrapper[4971]: I1213 07:28:48.946474 4971 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplanenodeset-openstack-edpm-ipam" Dec 13 07:28:48 crc kubenswrapper[4971]: I1213 07:28:48.946732 4971 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/run-os-edpm-deployment-openstack-edpm-ipam-dfs5x"] Dec 13 07:28:48 crc kubenswrapper[4971]: I1213 07:28:48.950415 4971 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-edpm-ipam-dockercfg-zb2j5" Dec 13 07:28:48 crc kubenswrapper[4971]: I1213 07:28:48.950536 4971 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplane-ansible-ssh-private-key-secret" Dec 13 07:28:49 crc kubenswrapper[4971]: I1213 07:28:49.040136 4971 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/e3950b9b-9046-4139-8cd6-8f099cef93e4-ssh-key\") pod \"run-os-edpm-deployment-openstack-edpm-ipam-dfs5x\" (UID: \"e3950b9b-9046-4139-8cd6-8f099cef93e4\") " pod="openstack/run-os-edpm-deployment-openstack-edpm-ipam-dfs5x" Dec 13 07:28:49 crc kubenswrapper[4971]: I1213 07:28:49.040243 4971 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/e3950b9b-9046-4139-8cd6-8f099cef93e4-inventory\") pod \"run-os-edpm-deployment-openstack-edpm-ipam-dfs5x\" (UID: \"e3950b9b-9046-4139-8cd6-8f099cef93e4\") " pod="openstack/run-os-edpm-deployment-openstack-edpm-ipam-dfs5x" Dec 13 07:28:49 crc kubenswrapper[4971]: I1213 07:28:49.040878 4971 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-stpcn\" (UniqueName: \"kubernetes.io/projected/e3950b9b-9046-4139-8cd6-8f099cef93e4-kube-api-access-stpcn\") pod \"run-os-edpm-deployment-openstack-edpm-ipam-dfs5x\" (UID: \"e3950b9b-9046-4139-8cd6-8f099cef93e4\") " pod="openstack/run-os-edpm-deployment-openstack-edpm-ipam-dfs5x" Dec 13 07:28:49 crc kubenswrapper[4971]: I1213 07:28:49.143318 4971 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-stpcn\" (UniqueName: \"kubernetes.io/projected/e3950b9b-9046-4139-8cd6-8f099cef93e4-kube-api-access-stpcn\") pod \"run-os-edpm-deployment-openstack-edpm-ipam-dfs5x\" (UID: \"e3950b9b-9046-4139-8cd6-8f099cef93e4\") " pod="openstack/run-os-edpm-deployment-openstack-edpm-ipam-dfs5x" Dec 13 07:28:49 crc kubenswrapper[4971]: I1213 07:28:49.143416 4971 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/e3950b9b-9046-4139-8cd6-8f099cef93e4-ssh-key\") pod \"run-os-edpm-deployment-openstack-edpm-ipam-dfs5x\" (UID: \"e3950b9b-9046-4139-8cd6-8f099cef93e4\") " pod="openstack/run-os-edpm-deployment-openstack-edpm-ipam-dfs5x" Dec 13 07:28:49 crc kubenswrapper[4971]: I1213 07:28:49.143493 4971 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/e3950b9b-9046-4139-8cd6-8f099cef93e4-inventory\") pod \"run-os-edpm-deployment-openstack-edpm-ipam-dfs5x\" (UID: \"e3950b9b-9046-4139-8cd6-8f099cef93e4\") " pod="openstack/run-os-edpm-deployment-openstack-edpm-ipam-dfs5x" Dec 13 07:28:49 crc kubenswrapper[4971]: I1213 07:28:49.151170 4971 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/e3950b9b-9046-4139-8cd6-8f099cef93e4-inventory\") pod \"run-os-edpm-deployment-openstack-edpm-ipam-dfs5x\" (UID: \"e3950b9b-9046-4139-8cd6-8f099cef93e4\") " pod="openstack/run-os-edpm-deployment-openstack-edpm-ipam-dfs5x" Dec 13 07:28:49 crc kubenswrapper[4971]: I1213 07:28:49.151346 4971 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/e3950b9b-9046-4139-8cd6-8f099cef93e4-ssh-key\") pod \"run-os-edpm-deployment-openstack-edpm-ipam-dfs5x\" (UID: \"e3950b9b-9046-4139-8cd6-8f099cef93e4\") " pod="openstack/run-os-edpm-deployment-openstack-edpm-ipam-dfs5x" Dec 13 07:28:49 crc kubenswrapper[4971]: I1213 07:28:49.175921 4971 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-stpcn\" (UniqueName: \"kubernetes.io/projected/e3950b9b-9046-4139-8cd6-8f099cef93e4-kube-api-access-stpcn\") pod \"run-os-edpm-deployment-openstack-edpm-ipam-dfs5x\" (UID: \"e3950b9b-9046-4139-8cd6-8f099cef93e4\") " pod="openstack/run-os-edpm-deployment-openstack-edpm-ipam-dfs5x" Dec 13 07:28:49 crc kubenswrapper[4971]: I1213 07:28:49.266821 4971 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/run-os-edpm-deployment-openstack-edpm-ipam-dfs5x" Dec 13 07:28:50 crc kubenswrapper[4971]: I1213 07:28:50.054152 4971 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/run-os-edpm-deployment-openstack-edpm-ipam-dfs5x"] Dec 13 07:28:51 crc kubenswrapper[4971]: I1213 07:28:50.762478 4971 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/run-os-edpm-deployment-openstack-edpm-ipam-dfs5x" event={"ID":"e3950b9b-9046-4139-8cd6-8f099cef93e4","Type":"ContainerStarted","Data":"5889853cfe99225199787e9b38f5f647fcf7f77e74b4887954e802226cccd161"} Dec 13 07:28:51 crc kubenswrapper[4971]: I1213 07:28:51.787004 4971 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/run-os-edpm-deployment-openstack-edpm-ipam-dfs5x" event={"ID":"e3950b9b-9046-4139-8cd6-8f099cef93e4","Type":"ContainerStarted","Data":"9822e0c105b43e179352e287724d20e8458c235bdae5f99e81482c895bc83d21"} Dec 13 07:28:51 crc kubenswrapper[4971]: I1213 07:28:51.814892 4971 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/run-os-edpm-deployment-openstack-edpm-ipam-dfs5x" podStartSLOduration=3.272993428 podStartE2EDuration="3.814868745s" podCreationTimestamp="2025-12-13 07:28:48 +0000 UTC" firstStartedPulling="2025-12-13 07:28:50.062485565 +0000 UTC m=+2386.666895013" lastFinishedPulling="2025-12-13 07:28:50.604360872 +0000 UTC m=+2387.208770330" observedRunningTime="2025-12-13 07:28:51.805119455 +0000 UTC m=+2388.409528943" watchObservedRunningTime="2025-12-13 07:28:51.814868745 +0000 UTC m=+2388.419278203" Dec 13 07:28:52 crc kubenswrapper[4971]: I1213 07:28:52.770453 4971 scope.go:117] "RemoveContainer" containerID="ec2e4b1312b5d5d0545354a836fef4af376ca3b34327ab1b9e0b72c102976079" Dec 13 07:28:52 crc kubenswrapper[4971]: E1213 07:28:52.770924 4971 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-82xjz_openshift-machine-config-operator(e0c4af12-edda-41d7-9b44-f87396174ef2)\"" pod="openshift-machine-config-operator/machine-config-daemon-82xjz" podUID="e0c4af12-edda-41d7-9b44-f87396174ef2" Dec 13 07:29:00 crc kubenswrapper[4971]: I1213 07:29:00.886272 4971 generic.go:334] "Generic (PLEG): container finished" podID="e3950b9b-9046-4139-8cd6-8f099cef93e4" containerID="9822e0c105b43e179352e287724d20e8458c235bdae5f99e81482c895bc83d21" exitCode=0 Dec 13 07:29:00 crc kubenswrapper[4971]: I1213 07:29:00.886358 4971 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/run-os-edpm-deployment-openstack-edpm-ipam-dfs5x" event={"ID":"e3950b9b-9046-4139-8cd6-8f099cef93e4","Type":"ContainerDied","Data":"9822e0c105b43e179352e287724d20e8458c235bdae5f99e81482c895bc83d21"} Dec 13 07:29:02 crc kubenswrapper[4971]: I1213 07:29:02.441261 4971 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/run-os-edpm-deployment-openstack-edpm-ipam-dfs5x" Dec 13 07:29:02 crc kubenswrapper[4971]: I1213 07:29:02.626804 4971 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/e3950b9b-9046-4139-8cd6-8f099cef93e4-ssh-key\") pod \"e3950b9b-9046-4139-8cd6-8f099cef93e4\" (UID: \"e3950b9b-9046-4139-8cd6-8f099cef93e4\") " Dec 13 07:29:02 crc kubenswrapper[4971]: I1213 07:29:02.627065 4971 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/e3950b9b-9046-4139-8cd6-8f099cef93e4-inventory\") pod \"e3950b9b-9046-4139-8cd6-8f099cef93e4\" (UID: \"e3950b9b-9046-4139-8cd6-8f099cef93e4\") " Dec 13 07:29:02 crc kubenswrapper[4971]: I1213 07:29:02.627133 4971 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-stpcn\" (UniqueName: \"kubernetes.io/projected/e3950b9b-9046-4139-8cd6-8f099cef93e4-kube-api-access-stpcn\") pod \"e3950b9b-9046-4139-8cd6-8f099cef93e4\" (UID: \"e3950b9b-9046-4139-8cd6-8f099cef93e4\") " Dec 13 07:29:02 crc kubenswrapper[4971]: I1213 07:29:02.637103 4971 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/e3950b9b-9046-4139-8cd6-8f099cef93e4-kube-api-access-stpcn" (OuterVolumeSpecName: "kube-api-access-stpcn") pod "e3950b9b-9046-4139-8cd6-8f099cef93e4" (UID: "e3950b9b-9046-4139-8cd6-8f099cef93e4"). InnerVolumeSpecName "kube-api-access-stpcn". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 13 07:29:02 crc kubenswrapper[4971]: I1213 07:29:02.663438 4971 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/e3950b9b-9046-4139-8cd6-8f099cef93e4-inventory" (OuterVolumeSpecName: "inventory") pod "e3950b9b-9046-4139-8cd6-8f099cef93e4" (UID: "e3950b9b-9046-4139-8cd6-8f099cef93e4"). InnerVolumeSpecName "inventory". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 13 07:29:02 crc kubenswrapper[4971]: I1213 07:29:02.666149 4971 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/e3950b9b-9046-4139-8cd6-8f099cef93e4-ssh-key" (OuterVolumeSpecName: "ssh-key") pod "e3950b9b-9046-4139-8cd6-8f099cef93e4" (UID: "e3950b9b-9046-4139-8cd6-8f099cef93e4"). InnerVolumeSpecName "ssh-key". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 13 07:29:02 crc kubenswrapper[4971]: I1213 07:29:02.729867 4971 reconciler_common.go:293] "Volume detached for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/e3950b9b-9046-4139-8cd6-8f099cef93e4-inventory\") on node \"crc\" DevicePath \"\"" Dec 13 07:29:02 crc kubenswrapper[4971]: I1213 07:29:02.729910 4971 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-stpcn\" (UniqueName: \"kubernetes.io/projected/e3950b9b-9046-4139-8cd6-8f099cef93e4-kube-api-access-stpcn\") on node \"crc\" DevicePath \"\"" Dec 13 07:29:02 crc kubenswrapper[4971]: I1213 07:29:02.729925 4971 reconciler_common.go:293] "Volume detached for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/e3950b9b-9046-4139-8cd6-8f099cef93e4-ssh-key\") on node \"crc\" DevicePath \"\"" Dec 13 07:29:02 crc kubenswrapper[4971]: I1213 07:29:02.909425 4971 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/run-os-edpm-deployment-openstack-edpm-ipam-dfs5x" event={"ID":"e3950b9b-9046-4139-8cd6-8f099cef93e4","Type":"ContainerDied","Data":"5889853cfe99225199787e9b38f5f647fcf7f77e74b4887954e802226cccd161"} Dec 13 07:29:02 crc kubenswrapper[4971]: I1213 07:29:02.909495 4971 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="5889853cfe99225199787e9b38f5f647fcf7f77e74b4887954e802226cccd161" Dec 13 07:29:02 crc kubenswrapper[4971]: I1213 07:29:02.909501 4971 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/run-os-edpm-deployment-openstack-edpm-ipam-dfs5x" Dec 13 07:29:03 crc kubenswrapper[4971]: I1213 07:29:03.003107 4971 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-hjgk2"] Dec 13 07:29:03 crc kubenswrapper[4971]: E1213 07:29:03.003777 4971 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="e3950b9b-9046-4139-8cd6-8f099cef93e4" containerName="run-os-edpm-deployment-openstack-edpm-ipam" Dec 13 07:29:03 crc kubenswrapper[4971]: I1213 07:29:03.003801 4971 state_mem.go:107] "Deleted CPUSet assignment" podUID="e3950b9b-9046-4139-8cd6-8f099cef93e4" containerName="run-os-edpm-deployment-openstack-edpm-ipam" Dec 13 07:29:03 crc kubenswrapper[4971]: I1213 07:29:03.004018 4971 memory_manager.go:354] "RemoveStaleState removing state" podUID="e3950b9b-9046-4139-8cd6-8f099cef93e4" containerName="run-os-edpm-deployment-openstack-edpm-ipam" Dec 13 07:29:03 crc kubenswrapper[4971]: I1213 07:29:03.004788 4971 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-hjgk2" Dec 13 07:29:03 crc kubenswrapper[4971]: I1213 07:29:03.007732 4971 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-aee-default-env" Dec 13 07:29:03 crc kubenswrapper[4971]: I1213 07:29:03.007790 4971 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-edpm-ipam-dockercfg-zb2j5" Dec 13 07:29:03 crc kubenswrapper[4971]: I1213 07:29:03.007954 4971 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplanenodeset-openstack-edpm-ipam" Dec 13 07:29:03 crc kubenswrapper[4971]: I1213 07:29:03.007974 4971 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplane-ansible-ssh-private-key-secret" Dec 13 07:29:03 crc kubenswrapper[4971]: I1213 07:29:03.017991 4971 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-hjgk2"] Dec 13 07:29:03 crc kubenswrapper[4971]: I1213 07:29:03.142303 4971 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/62eac784-f59b-4764-ad97-681df8129497-inventory\") pod \"reboot-os-edpm-deployment-openstack-edpm-ipam-hjgk2\" (UID: \"62eac784-f59b-4764-ad97-681df8129497\") " pod="openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-hjgk2" Dec 13 07:29:03 crc kubenswrapper[4971]: I1213 07:29:03.142837 4971 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/62eac784-f59b-4764-ad97-681df8129497-ssh-key\") pod \"reboot-os-edpm-deployment-openstack-edpm-ipam-hjgk2\" (UID: \"62eac784-f59b-4764-ad97-681df8129497\") " pod="openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-hjgk2" Dec 13 07:29:03 crc kubenswrapper[4971]: I1213 07:29:03.142924 4971 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-xz2n4\" (UniqueName: \"kubernetes.io/projected/62eac784-f59b-4764-ad97-681df8129497-kube-api-access-xz2n4\") pod \"reboot-os-edpm-deployment-openstack-edpm-ipam-hjgk2\" (UID: \"62eac784-f59b-4764-ad97-681df8129497\") " pod="openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-hjgk2" Dec 13 07:29:03 crc kubenswrapper[4971]: I1213 07:29:03.245383 4971 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/62eac784-f59b-4764-ad97-681df8129497-ssh-key\") pod \"reboot-os-edpm-deployment-openstack-edpm-ipam-hjgk2\" (UID: \"62eac784-f59b-4764-ad97-681df8129497\") " pod="openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-hjgk2" Dec 13 07:29:03 crc kubenswrapper[4971]: I1213 07:29:03.245449 4971 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-xz2n4\" (UniqueName: \"kubernetes.io/projected/62eac784-f59b-4764-ad97-681df8129497-kube-api-access-xz2n4\") pod \"reboot-os-edpm-deployment-openstack-edpm-ipam-hjgk2\" (UID: \"62eac784-f59b-4764-ad97-681df8129497\") " pod="openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-hjgk2" Dec 13 07:29:03 crc kubenswrapper[4971]: I1213 07:29:03.245555 4971 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/62eac784-f59b-4764-ad97-681df8129497-inventory\") pod \"reboot-os-edpm-deployment-openstack-edpm-ipam-hjgk2\" (UID: \"62eac784-f59b-4764-ad97-681df8129497\") " pod="openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-hjgk2" Dec 13 07:29:03 crc kubenswrapper[4971]: I1213 07:29:03.251705 4971 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/62eac784-f59b-4764-ad97-681df8129497-inventory\") pod \"reboot-os-edpm-deployment-openstack-edpm-ipam-hjgk2\" (UID: \"62eac784-f59b-4764-ad97-681df8129497\") " pod="openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-hjgk2" Dec 13 07:29:03 crc kubenswrapper[4971]: I1213 07:29:03.255342 4971 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/62eac784-f59b-4764-ad97-681df8129497-ssh-key\") pod \"reboot-os-edpm-deployment-openstack-edpm-ipam-hjgk2\" (UID: \"62eac784-f59b-4764-ad97-681df8129497\") " pod="openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-hjgk2" Dec 13 07:29:03 crc kubenswrapper[4971]: I1213 07:29:03.272829 4971 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-xz2n4\" (UniqueName: \"kubernetes.io/projected/62eac784-f59b-4764-ad97-681df8129497-kube-api-access-xz2n4\") pod \"reboot-os-edpm-deployment-openstack-edpm-ipam-hjgk2\" (UID: \"62eac784-f59b-4764-ad97-681df8129497\") " pod="openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-hjgk2" Dec 13 07:29:03 crc kubenswrapper[4971]: I1213 07:29:03.322026 4971 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-hjgk2" Dec 13 07:29:03 crc kubenswrapper[4971]: I1213 07:29:03.912936 4971 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-hjgk2"] Dec 13 07:29:03 crc kubenswrapper[4971]: W1213 07:29:03.935741 4971 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod62eac784_f59b_4764_ad97_681df8129497.slice/crio-2b930532ac336448c05d921a16d4293823a2e05394829d9592ecb8cd08ee11c2 WatchSource:0}: Error finding container 2b930532ac336448c05d921a16d4293823a2e05394829d9592ecb8cd08ee11c2: Status 404 returned error can't find the container with id 2b930532ac336448c05d921a16d4293823a2e05394829d9592ecb8cd08ee11c2 Dec 13 07:29:03 crc kubenswrapper[4971]: I1213 07:29:03.939951 4971 provider.go:102] Refreshing cache for provider: *credentialprovider.defaultDockerConfigProvider Dec 13 07:29:04 crc kubenswrapper[4971]: I1213 07:29:04.472665 4971 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-aee-default-env" Dec 13 07:29:04 crc kubenswrapper[4971]: I1213 07:29:04.933955 4971 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-hjgk2" event={"ID":"62eac784-f59b-4764-ad97-681df8129497","Type":"ContainerStarted","Data":"caa6c4125172e3f7d61499c91b31dfad33bb6803bce76d35a19b16d325ddb492"} Dec 13 07:29:04 crc kubenswrapper[4971]: I1213 07:29:04.934024 4971 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-hjgk2" event={"ID":"62eac784-f59b-4764-ad97-681df8129497","Type":"ContainerStarted","Data":"2b930532ac336448c05d921a16d4293823a2e05394829d9592ecb8cd08ee11c2"} Dec 13 07:29:04 crc kubenswrapper[4971]: I1213 07:29:04.956480 4971 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-hjgk2" podStartSLOduration=2.427006295 podStartE2EDuration="2.956463326s" podCreationTimestamp="2025-12-13 07:29:02 +0000 UTC" firstStartedPulling="2025-12-13 07:29:03.939478185 +0000 UTC m=+2400.543887643" lastFinishedPulling="2025-12-13 07:29:04.468935226 +0000 UTC m=+2401.073344674" observedRunningTime="2025-12-13 07:29:04.950234212 +0000 UTC m=+2401.554643660" watchObservedRunningTime="2025-12-13 07:29:04.956463326 +0000 UTC m=+2401.560872774" Dec 13 07:29:07 crc kubenswrapper[4971]: I1213 07:29:07.770276 4971 scope.go:117] "RemoveContainer" containerID="ec2e4b1312b5d5d0545354a836fef4af376ca3b34327ab1b9e0b72c102976079" Dec 13 07:29:07 crc kubenswrapper[4971]: E1213 07:29:07.771105 4971 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-82xjz_openshift-machine-config-operator(e0c4af12-edda-41d7-9b44-f87396174ef2)\"" pod="openshift-machine-config-operator/machine-config-daemon-82xjz" podUID="e0c4af12-edda-41d7-9b44-f87396174ef2" Dec 13 07:29:16 crc kubenswrapper[4971]: I1213 07:29:16.051998 4971 generic.go:334] "Generic (PLEG): container finished" podID="62eac784-f59b-4764-ad97-681df8129497" containerID="caa6c4125172e3f7d61499c91b31dfad33bb6803bce76d35a19b16d325ddb492" exitCode=0 Dec 13 07:29:16 crc kubenswrapper[4971]: I1213 07:29:16.052249 4971 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-hjgk2" event={"ID":"62eac784-f59b-4764-ad97-681df8129497","Type":"ContainerDied","Data":"caa6c4125172e3f7d61499c91b31dfad33bb6803bce76d35a19b16d325ddb492"} Dec 13 07:29:17 crc kubenswrapper[4971]: I1213 07:29:17.585152 4971 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-hjgk2" Dec 13 07:29:17 crc kubenswrapper[4971]: I1213 07:29:17.772320 4971 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/62eac784-f59b-4764-ad97-681df8129497-ssh-key\") pod \"62eac784-f59b-4764-ad97-681df8129497\" (UID: \"62eac784-f59b-4764-ad97-681df8129497\") " Dec 13 07:29:17 crc kubenswrapper[4971]: I1213 07:29:17.772427 4971 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/62eac784-f59b-4764-ad97-681df8129497-inventory\") pod \"62eac784-f59b-4764-ad97-681df8129497\" (UID: \"62eac784-f59b-4764-ad97-681df8129497\") " Dec 13 07:29:17 crc kubenswrapper[4971]: I1213 07:29:17.773348 4971 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-xz2n4\" (UniqueName: \"kubernetes.io/projected/62eac784-f59b-4764-ad97-681df8129497-kube-api-access-xz2n4\") pod \"62eac784-f59b-4764-ad97-681df8129497\" (UID: \"62eac784-f59b-4764-ad97-681df8129497\") " Dec 13 07:29:17 crc kubenswrapper[4971]: I1213 07:29:17.780749 4971 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/62eac784-f59b-4764-ad97-681df8129497-kube-api-access-xz2n4" (OuterVolumeSpecName: "kube-api-access-xz2n4") pod "62eac784-f59b-4764-ad97-681df8129497" (UID: "62eac784-f59b-4764-ad97-681df8129497"). InnerVolumeSpecName "kube-api-access-xz2n4". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 13 07:29:17 crc kubenswrapper[4971]: I1213 07:29:17.820236 4971 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/62eac784-f59b-4764-ad97-681df8129497-ssh-key" (OuterVolumeSpecName: "ssh-key") pod "62eac784-f59b-4764-ad97-681df8129497" (UID: "62eac784-f59b-4764-ad97-681df8129497"). InnerVolumeSpecName "ssh-key". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 13 07:29:17 crc kubenswrapper[4971]: I1213 07:29:17.839883 4971 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/62eac784-f59b-4764-ad97-681df8129497-inventory" (OuterVolumeSpecName: "inventory") pod "62eac784-f59b-4764-ad97-681df8129497" (UID: "62eac784-f59b-4764-ad97-681df8129497"). InnerVolumeSpecName "inventory". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 13 07:29:17 crc kubenswrapper[4971]: I1213 07:29:17.880770 4971 reconciler_common.go:293] "Volume detached for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/62eac784-f59b-4764-ad97-681df8129497-inventory\") on node \"crc\" DevicePath \"\"" Dec 13 07:29:17 crc kubenswrapper[4971]: I1213 07:29:17.880827 4971 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-xz2n4\" (UniqueName: \"kubernetes.io/projected/62eac784-f59b-4764-ad97-681df8129497-kube-api-access-xz2n4\") on node \"crc\" DevicePath \"\"" Dec 13 07:29:17 crc kubenswrapper[4971]: I1213 07:29:17.880842 4971 reconciler_common.go:293] "Volume detached for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/62eac784-f59b-4764-ad97-681df8129497-ssh-key\") on node \"crc\" DevicePath \"\"" Dec 13 07:29:18 crc kubenswrapper[4971]: I1213 07:29:18.077835 4971 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-hjgk2" event={"ID":"62eac784-f59b-4764-ad97-681df8129497","Type":"ContainerDied","Data":"2b930532ac336448c05d921a16d4293823a2e05394829d9592ecb8cd08ee11c2"} Dec 13 07:29:18 crc kubenswrapper[4971]: I1213 07:29:18.077896 4971 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="2b930532ac336448c05d921a16d4293823a2e05394829d9592ecb8cd08ee11c2" Dec 13 07:29:18 crc kubenswrapper[4971]: I1213 07:29:18.077930 4971 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-hjgk2" Dec 13 07:29:21 crc kubenswrapper[4971]: I1213 07:29:21.769726 4971 scope.go:117] "RemoveContainer" containerID="ec2e4b1312b5d5d0545354a836fef4af376ca3b34327ab1b9e0b72c102976079" Dec 13 07:29:21 crc kubenswrapper[4971]: E1213 07:29:21.770793 4971 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-82xjz_openshift-machine-config-operator(e0c4af12-edda-41d7-9b44-f87396174ef2)\"" pod="openshift-machine-config-operator/machine-config-daemon-82xjz" podUID="e0c4af12-edda-41d7-9b44-f87396174ef2" Dec 13 07:29:33 crc kubenswrapper[4971]: I1213 07:29:33.779579 4971 scope.go:117] "RemoveContainer" containerID="ec2e4b1312b5d5d0545354a836fef4af376ca3b34327ab1b9e0b72c102976079" Dec 13 07:29:33 crc kubenswrapper[4971]: E1213 07:29:33.782012 4971 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-82xjz_openshift-machine-config-operator(e0c4af12-edda-41d7-9b44-f87396174ef2)\"" pod="openshift-machine-config-operator/machine-config-daemon-82xjz" podUID="e0c4af12-edda-41d7-9b44-f87396174ef2" Dec 13 07:29:48 crc kubenswrapper[4971]: I1213 07:29:48.835150 4971 scope.go:117] "RemoveContainer" containerID="ec2e4b1312b5d5d0545354a836fef4af376ca3b34327ab1b9e0b72c102976079" Dec 13 07:29:48 crc kubenswrapper[4971]: E1213 07:29:48.836971 4971 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-82xjz_openshift-machine-config-operator(e0c4af12-edda-41d7-9b44-f87396174ef2)\"" pod="openshift-machine-config-operator/machine-config-daemon-82xjz" podUID="e0c4af12-edda-41d7-9b44-f87396174ef2" Dec 13 07:30:00 crc kubenswrapper[4971]: I1213 07:30:00.160794 4971 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29426850-rpprk"] Dec 13 07:30:00 crc kubenswrapper[4971]: E1213 07:30:00.161819 4971 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="62eac784-f59b-4764-ad97-681df8129497" containerName="reboot-os-edpm-deployment-openstack-edpm-ipam" Dec 13 07:30:00 crc kubenswrapper[4971]: I1213 07:30:00.161838 4971 state_mem.go:107] "Deleted CPUSet assignment" podUID="62eac784-f59b-4764-ad97-681df8129497" containerName="reboot-os-edpm-deployment-openstack-edpm-ipam" Dec 13 07:30:00 crc kubenswrapper[4971]: I1213 07:30:00.162066 4971 memory_manager.go:354] "RemoveStaleState removing state" podUID="62eac784-f59b-4764-ad97-681df8129497" containerName="reboot-os-edpm-deployment-openstack-edpm-ipam" Dec 13 07:30:00 crc kubenswrapper[4971]: I1213 07:30:00.162845 4971 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29426850-rpprk" Dec 13 07:30:00 crc kubenswrapper[4971]: I1213 07:30:00.165080 4971 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"collect-profiles-dockercfg-kzf4t" Dec 13 07:30:00 crc kubenswrapper[4971]: I1213 07:30:00.165971 4971 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-operator-lifecycle-manager"/"collect-profiles-config" Dec 13 07:30:00 crc kubenswrapper[4971]: I1213 07:30:00.174685 4971 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29426850-rpprk"] Dec 13 07:30:00 crc kubenswrapper[4971]: I1213 07:30:00.217061 4971 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-5lv8k\" (UniqueName: \"kubernetes.io/projected/0e2af7b4-e1cb-4276-bece-56851503f2b0-kube-api-access-5lv8k\") pod \"collect-profiles-29426850-rpprk\" (UID: \"0e2af7b4-e1cb-4276-bece-56851503f2b0\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29426850-rpprk" Dec 13 07:30:00 crc kubenswrapper[4971]: I1213 07:30:00.217292 4971 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/0e2af7b4-e1cb-4276-bece-56851503f2b0-config-volume\") pod \"collect-profiles-29426850-rpprk\" (UID: \"0e2af7b4-e1cb-4276-bece-56851503f2b0\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29426850-rpprk" Dec 13 07:30:00 crc kubenswrapper[4971]: I1213 07:30:00.217335 4971 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/0e2af7b4-e1cb-4276-bece-56851503f2b0-secret-volume\") pod \"collect-profiles-29426850-rpprk\" (UID: \"0e2af7b4-e1cb-4276-bece-56851503f2b0\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29426850-rpprk" Dec 13 07:30:00 crc kubenswrapper[4971]: I1213 07:30:00.319305 4971 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-5lv8k\" (UniqueName: \"kubernetes.io/projected/0e2af7b4-e1cb-4276-bece-56851503f2b0-kube-api-access-5lv8k\") pod \"collect-profiles-29426850-rpprk\" (UID: \"0e2af7b4-e1cb-4276-bece-56851503f2b0\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29426850-rpprk" Dec 13 07:30:00 crc kubenswrapper[4971]: I1213 07:30:00.319481 4971 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/0e2af7b4-e1cb-4276-bece-56851503f2b0-config-volume\") pod \"collect-profiles-29426850-rpprk\" (UID: \"0e2af7b4-e1cb-4276-bece-56851503f2b0\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29426850-rpprk" Dec 13 07:30:00 crc kubenswrapper[4971]: I1213 07:30:00.319598 4971 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/0e2af7b4-e1cb-4276-bece-56851503f2b0-secret-volume\") pod \"collect-profiles-29426850-rpprk\" (UID: \"0e2af7b4-e1cb-4276-bece-56851503f2b0\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29426850-rpprk" Dec 13 07:30:00 crc kubenswrapper[4971]: I1213 07:30:00.321710 4971 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/0e2af7b4-e1cb-4276-bece-56851503f2b0-config-volume\") pod \"collect-profiles-29426850-rpprk\" (UID: \"0e2af7b4-e1cb-4276-bece-56851503f2b0\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29426850-rpprk" Dec 13 07:30:00 crc kubenswrapper[4971]: I1213 07:30:00.330335 4971 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/0e2af7b4-e1cb-4276-bece-56851503f2b0-secret-volume\") pod \"collect-profiles-29426850-rpprk\" (UID: \"0e2af7b4-e1cb-4276-bece-56851503f2b0\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29426850-rpprk" Dec 13 07:30:00 crc kubenswrapper[4971]: I1213 07:30:00.342591 4971 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-5lv8k\" (UniqueName: \"kubernetes.io/projected/0e2af7b4-e1cb-4276-bece-56851503f2b0-kube-api-access-5lv8k\") pod \"collect-profiles-29426850-rpprk\" (UID: \"0e2af7b4-e1cb-4276-bece-56851503f2b0\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29426850-rpprk" Dec 13 07:30:00 crc kubenswrapper[4971]: I1213 07:30:00.548816 4971 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29426850-rpprk" Dec 13 07:30:01 crc kubenswrapper[4971]: I1213 07:30:01.048389 4971 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29426850-rpprk"] Dec 13 07:30:01 crc kubenswrapper[4971]: I1213 07:30:01.537480 4971 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29426850-rpprk" event={"ID":"0e2af7b4-e1cb-4276-bece-56851503f2b0","Type":"ContainerStarted","Data":"c6325e4a33d42e8703e692af3bb85b8d6f6cd00439d91984ba490882209c6b44"} Dec 13 07:30:01 crc kubenswrapper[4971]: I1213 07:30:01.537589 4971 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29426850-rpprk" event={"ID":"0e2af7b4-e1cb-4276-bece-56851503f2b0","Type":"ContainerStarted","Data":"c2fd4f110444789b7c3509ef25a965112f905a1e6e8f9b00b18f6adfea16c5b4"} Dec 13 07:30:01 crc kubenswrapper[4971]: I1213 07:30:01.558957 4971 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-operator-lifecycle-manager/collect-profiles-29426850-rpprk" podStartSLOduration=1.558939823 podStartE2EDuration="1.558939823s" podCreationTimestamp="2025-12-13 07:30:00 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-13 07:30:01.555689992 +0000 UTC m=+2458.160099440" watchObservedRunningTime="2025-12-13 07:30:01.558939823 +0000 UTC m=+2458.163349291" Dec 13 07:30:02 crc kubenswrapper[4971]: I1213 07:30:02.551072 4971 generic.go:334] "Generic (PLEG): container finished" podID="0e2af7b4-e1cb-4276-bece-56851503f2b0" containerID="c6325e4a33d42e8703e692af3bb85b8d6f6cd00439d91984ba490882209c6b44" exitCode=0 Dec 13 07:30:02 crc kubenswrapper[4971]: I1213 07:30:02.551155 4971 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29426850-rpprk" event={"ID":"0e2af7b4-e1cb-4276-bece-56851503f2b0","Type":"ContainerDied","Data":"c6325e4a33d42e8703e692af3bb85b8d6f6cd00439d91984ba490882209c6b44"} Dec 13 07:30:03 crc kubenswrapper[4971]: I1213 07:30:03.779083 4971 scope.go:117] "RemoveContainer" containerID="ec2e4b1312b5d5d0545354a836fef4af376ca3b34327ab1b9e0b72c102976079" Dec 13 07:30:03 crc kubenswrapper[4971]: E1213 07:30:03.779840 4971 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-82xjz_openshift-machine-config-operator(e0c4af12-edda-41d7-9b44-f87396174ef2)\"" pod="openshift-machine-config-operator/machine-config-daemon-82xjz" podUID="e0c4af12-edda-41d7-9b44-f87396174ef2" Dec 13 07:30:03 crc kubenswrapper[4971]: I1213 07:30:03.883672 4971 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29426850-rpprk" Dec 13 07:30:04 crc kubenswrapper[4971]: I1213 07:30:04.071768 4971 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-5lv8k\" (UniqueName: \"kubernetes.io/projected/0e2af7b4-e1cb-4276-bece-56851503f2b0-kube-api-access-5lv8k\") pod \"0e2af7b4-e1cb-4276-bece-56851503f2b0\" (UID: \"0e2af7b4-e1cb-4276-bece-56851503f2b0\") " Dec 13 07:30:04 crc kubenswrapper[4971]: I1213 07:30:04.071983 4971 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/0e2af7b4-e1cb-4276-bece-56851503f2b0-config-volume\") pod \"0e2af7b4-e1cb-4276-bece-56851503f2b0\" (UID: \"0e2af7b4-e1cb-4276-bece-56851503f2b0\") " Dec 13 07:30:04 crc kubenswrapper[4971]: I1213 07:30:04.072029 4971 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/0e2af7b4-e1cb-4276-bece-56851503f2b0-secret-volume\") pod \"0e2af7b4-e1cb-4276-bece-56851503f2b0\" (UID: \"0e2af7b4-e1cb-4276-bece-56851503f2b0\") " Dec 13 07:30:04 crc kubenswrapper[4971]: I1213 07:30:04.072858 4971 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/0e2af7b4-e1cb-4276-bece-56851503f2b0-config-volume" (OuterVolumeSpecName: "config-volume") pod "0e2af7b4-e1cb-4276-bece-56851503f2b0" (UID: "0e2af7b4-e1cb-4276-bece-56851503f2b0"). InnerVolumeSpecName "config-volume". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 13 07:30:04 crc kubenswrapper[4971]: I1213 07:30:04.081905 4971 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/0e2af7b4-e1cb-4276-bece-56851503f2b0-kube-api-access-5lv8k" (OuterVolumeSpecName: "kube-api-access-5lv8k") pod "0e2af7b4-e1cb-4276-bece-56851503f2b0" (UID: "0e2af7b4-e1cb-4276-bece-56851503f2b0"). InnerVolumeSpecName "kube-api-access-5lv8k". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 13 07:30:04 crc kubenswrapper[4971]: I1213 07:30:04.083671 4971 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/0e2af7b4-e1cb-4276-bece-56851503f2b0-secret-volume" (OuterVolumeSpecName: "secret-volume") pod "0e2af7b4-e1cb-4276-bece-56851503f2b0" (UID: "0e2af7b4-e1cb-4276-bece-56851503f2b0"). InnerVolumeSpecName "secret-volume". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 13 07:30:04 crc kubenswrapper[4971]: I1213 07:30:04.173896 4971 reconciler_common.go:293] "Volume detached for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/0e2af7b4-e1cb-4276-bece-56851503f2b0-config-volume\") on node \"crc\" DevicePath \"\"" Dec 13 07:30:04 crc kubenswrapper[4971]: I1213 07:30:04.173927 4971 reconciler_common.go:293] "Volume detached for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/0e2af7b4-e1cb-4276-bece-56851503f2b0-secret-volume\") on node \"crc\" DevicePath \"\"" Dec 13 07:30:04 crc kubenswrapper[4971]: I1213 07:30:04.173939 4971 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-5lv8k\" (UniqueName: \"kubernetes.io/projected/0e2af7b4-e1cb-4276-bece-56851503f2b0-kube-api-access-5lv8k\") on node \"crc\" DevicePath \"\"" Dec 13 07:30:04 crc kubenswrapper[4971]: I1213 07:30:04.572585 4971 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29426850-rpprk" event={"ID":"0e2af7b4-e1cb-4276-bece-56851503f2b0","Type":"ContainerDied","Data":"c2fd4f110444789b7c3509ef25a965112f905a1e6e8f9b00b18f6adfea16c5b4"} Dec 13 07:30:04 crc kubenswrapper[4971]: I1213 07:30:04.572619 4971 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29426850-rpprk" Dec 13 07:30:04 crc kubenswrapper[4971]: I1213 07:30:04.572632 4971 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="c2fd4f110444789b7c3509ef25a965112f905a1e6e8f9b00b18f6adfea16c5b4" Dec 13 07:30:04 crc kubenswrapper[4971]: I1213 07:30:04.680854 4971 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29426805-fxxrz"] Dec 13 07:30:04 crc kubenswrapper[4971]: I1213 07:30:04.691066 4971 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29426805-fxxrz"] Dec 13 07:30:07 crc kubenswrapper[4971]: I1213 07:30:07.814610 4971 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="c1575ff9-1c53-4732-bc48-252f9f270597" path="/var/lib/kubelet/pods/c1575ff9-1c53-4732-bc48-252f9f270597/volumes" Dec 13 07:30:07 crc kubenswrapper[4971]: E1213 07:30:07.816571 4971 kubelet.go:2526] "Housekeeping took longer than expected" err="housekeeping took too long" expected="1s" actual="2.049s" Dec 13 07:30:15 crc kubenswrapper[4971]: I1213 07:30:15.770755 4971 scope.go:117] "RemoveContainer" containerID="ec2e4b1312b5d5d0545354a836fef4af376ca3b34327ab1b9e0b72c102976079" Dec 13 07:30:15 crc kubenswrapper[4971]: E1213 07:30:15.771450 4971 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-82xjz_openshift-machine-config-operator(e0c4af12-edda-41d7-9b44-f87396174ef2)\"" pod="openshift-machine-config-operator/machine-config-daemon-82xjz" podUID="e0c4af12-edda-41d7-9b44-f87396174ef2" Dec 13 07:30:18 crc kubenswrapper[4971]: I1213 07:30:18.691772 4971 scope.go:117] "RemoveContainer" containerID="7279bfe22b885129c7c9a8d98bc482013cb0cdf21dfdc5a026eda208800095de" Dec 13 07:30:26 crc kubenswrapper[4971]: I1213 07:30:26.769014 4971 scope.go:117] "RemoveContainer" containerID="ec2e4b1312b5d5d0545354a836fef4af376ca3b34327ab1b9e0b72c102976079" Dec 13 07:30:26 crc kubenswrapper[4971]: E1213 07:30:26.770615 4971 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-82xjz_openshift-machine-config-operator(e0c4af12-edda-41d7-9b44-f87396174ef2)\"" pod="openshift-machine-config-operator/machine-config-daemon-82xjz" podUID="e0c4af12-edda-41d7-9b44-f87396174ef2" Dec 13 07:30:38 crc kubenswrapper[4971]: I1213 07:30:38.769681 4971 scope.go:117] "RemoveContainer" containerID="ec2e4b1312b5d5d0545354a836fef4af376ca3b34327ab1b9e0b72c102976079" Dec 13 07:30:38 crc kubenswrapper[4971]: E1213 07:30:38.771000 4971 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-82xjz_openshift-machine-config-operator(e0c4af12-edda-41d7-9b44-f87396174ef2)\"" pod="openshift-machine-config-operator/machine-config-daemon-82xjz" podUID="e0c4af12-edda-41d7-9b44-f87396174ef2" Dec 13 07:30:53 crc kubenswrapper[4971]: I1213 07:30:53.779849 4971 scope.go:117] "RemoveContainer" containerID="ec2e4b1312b5d5d0545354a836fef4af376ca3b34327ab1b9e0b72c102976079" Dec 13 07:30:53 crc kubenswrapper[4971]: E1213 07:30:53.780803 4971 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-82xjz_openshift-machine-config-operator(e0c4af12-edda-41d7-9b44-f87396174ef2)\"" pod="openshift-machine-config-operator/machine-config-daemon-82xjz" podUID="e0c4af12-edda-41d7-9b44-f87396174ef2" Dec 13 07:31:07 crc kubenswrapper[4971]: I1213 07:31:07.768421 4971 scope.go:117] "RemoveContainer" containerID="ec2e4b1312b5d5d0545354a836fef4af376ca3b34327ab1b9e0b72c102976079" Dec 13 07:31:07 crc kubenswrapper[4971]: E1213 07:31:07.769265 4971 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-82xjz_openshift-machine-config-operator(e0c4af12-edda-41d7-9b44-f87396174ef2)\"" pod="openshift-machine-config-operator/machine-config-daemon-82xjz" podUID="e0c4af12-edda-41d7-9b44-f87396174ef2" Dec 13 07:31:19 crc kubenswrapper[4971]: I1213 07:31:19.769029 4971 scope.go:117] "RemoveContainer" containerID="ec2e4b1312b5d5d0545354a836fef4af376ca3b34327ab1b9e0b72c102976079" Dec 13 07:31:20 crc kubenswrapper[4971]: I1213 07:31:20.381595 4971 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-82xjz" event={"ID":"e0c4af12-edda-41d7-9b44-f87396174ef2","Type":"ContainerStarted","Data":"22e9f7096ef3c500a655cea9dbd5645fc5a0ca55768d6f95bfcaf07af351d70a"} Dec 13 07:33:46 crc kubenswrapper[4971]: I1213 07:33:46.153316 4971 patch_prober.go:28] interesting pod/machine-config-daemon-82xjz container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 13 07:33:46 crc kubenswrapper[4971]: I1213 07:33:46.154038 4971 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-82xjz" podUID="e0c4af12-edda-41d7-9b44-f87396174ef2" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 13 07:34:06 crc kubenswrapper[4971]: I1213 07:34:06.739810 4971 prober.go:107] "Probe failed" probeType="Liveness" pod="openstack/ceilometer-0" podUID="7cfb4dbf-cc55-48b7-b50c-65d6d4ae64ba" containerName="ceilometer-central-agent" probeResult="failure" output="command timed out" Dec 13 07:34:06 crc kubenswrapper[4971]: I1213 07:34:06.740479 4971 prober.go:107] "Probe failed" probeType="Liveness" pod="openstack/ceilometer-0" podUID="7cfb4dbf-cc55-48b7-b50c-65d6d4ae64ba" containerName="ceilometer-notification-agent" probeResult="failure" output="command timed out" Dec 13 07:34:11 crc kubenswrapper[4971]: I1213 07:34:11.739222 4971 prober.go:107] "Probe failed" probeType="Liveness" pod="openstack/ceilometer-0" podUID="7cfb4dbf-cc55-48b7-b50c-65d6d4ae64ba" containerName="ceilometer-central-agent" probeResult="failure" output="command timed out" Dec 13 07:34:16 crc kubenswrapper[4971]: I1213 07:34:16.153852 4971 patch_prober.go:28] interesting pod/machine-config-daemon-82xjz container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 13 07:34:16 crc kubenswrapper[4971]: I1213 07:34:16.154707 4971 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-82xjz" podUID="e0c4af12-edda-41d7-9b44-f87396174ef2" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 13 07:34:16 crc kubenswrapper[4971]: I1213 07:34:16.739325 4971 prober.go:107] "Probe failed" probeType="Liveness" pod="openstack/ceilometer-0" podUID="7cfb4dbf-cc55-48b7-b50c-65d6d4ae64ba" containerName="ceilometer-central-agent" probeResult="failure" output="command timed out" Dec 13 07:34:16 crc kubenswrapper[4971]: I1213 07:34:16.739499 4971 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openstack/ceilometer-0" Dec 13 07:34:16 crc kubenswrapper[4971]: I1213 07:34:16.740755 4971 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="ceilometer-central-agent" containerStatusID={"Type":"cri-o","ID":"5d071b2ffdcfc4bb8c866abc518564c7b5dac13fb9b40ac923cc1c5a8d52504f"} pod="openstack/ceilometer-0" containerMessage="Container ceilometer-central-agent failed liveness probe, will be restarted" Dec 13 07:34:16 crc kubenswrapper[4971]: I1213 07:34:16.740934 4971 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="7cfb4dbf-cc55-48b7-b50c-65d6d4ae64ba" containerName="ceilometer-central-agent" containerID="cri-o://5d071b2ffdcfc4bb8c866abc518564c7b5dac13fb9b40ac923cc1c5a8d52504f" gracePeriod=30 Dec 13 07:34:20 crc kubenswrapper[4971]: I1213 07:34:20.401021 4971 provider.go:102] Refreshing cache for provider: *credentialprovider.defaultDockerConfigProvider Dec 13 07:34:20 crc kubenswrapper[4971]: I1213 07:34:20.516684 4971 generic.go:334] "Generic (PLEG): container finished" podID="7cfb4dbf-cc55-48b7-b50c-65d6d4ae64ba" containerID="5d071b2ffdcfc4bb8c866abc518564c7b5dac13fb9b40ac923cc1c5a8d52504f" exitCode=0 Dec 13 07:34:20 crc kubenswrapper[4971]: I1213 07:34:20.516723 4971 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"7cfb4dbf-cc55-48b7-b50c-65d6d4ae64ba","Type":"ContainerDied","Data":"5d071b2ffdcfc4bb8c866abc518564c7b5dac13fb9b40ac923cc1c5a8d52504f"} Dec 13 07:34:21 crc kubenswrapper[4971]: I1213 07:34:21.533286 4971 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"7cfb4dbf-cc55-48b7-b50c-65d6d4ae64ba","Type":"ContainerStarted","Data":"2270e1af9b8023fb36226ee428231c2e187e0f781fcb819a9e8e50ef5b0046f5"} Dec 13 07:34:46 crc kubenswrapper[4971]: I1213 07:34:46.153556 4971 patch_prober.go:28] interesting pod/machine-config-daemon-82xjz container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 13 07:34:46 crc kubenswrapper[4971]: I1213 07:34:46.154078 4971 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-82xjz" podUID="e0c4af12-edda-41d7-9b44-f87396174ef2" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 13 07:34:46 crc kubenswrapper[4971]: I1213 07:34:46.154120 4971 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-82xjz" Dec 13 07:34:46 crc kubenswrapper[4971]: I1213 07:34:46.155945 4971 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"22e9f7096ef3c500a655cea9dbd5645fc5a0ca55768d6f95bfcaf07af351d70a"} pod="openshift-machine-config-operator/machine-config-daemon-82xjz" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Dec 13 07:34:46 crc kubenswrapper[4971]: I1213 07:34:46.156048 4971 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-82xjz" podUID="e0c4af12-edda-41d7-9b44-f87396174ef2" containerName="machine-config-daemon" containerID="cri-o://22e9f7096ef3c500a655cea9dbd5645fc5a0ca55768d6f95bfcaf07af351d70a" gracePeriod=600 Dec 13 07:34:46 crc kubenswrapper[4971]: I1213 07:34:46.795231 4971 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-82xjz" event={"ID":"e0c4af12-edda-41d7-9b44-f87396174ef2","Type":"ContainerDied","Data":"22e9f7096ef3c500a655cea9dbd5645fc5a0ca55768d6f95bfcaf07af351d70a"} Dec 13 07:34:46 crc kubenswrapper[4971]: I1213 07:34:46.795491 4971 scope.go:117] "RemoveContainer" containerID="ec2e4b1312b5d5d0545354a836fef4af376ca3b34327ab1b9e0b72c102976079" Dec 13 07:34:46 crc kubenswrapper[4971]: I1213 07:34:46.795264 4971 generic.go:334] "Generic (PLEG): container finished" podID="e0c4af12-edda-41d7-9b44-f87396174ef2" containerID="22e9f7096ef3c500a655cea9dbd5645fc5a0ca55768d6f95bfcaf07af351d70a" exitCode=0 Dec 13 07:34:47 crc kubenswrapper[4971]: I1213 07:34:47.804091 4971 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-82xjz" event={"ID":"e0c4af12-edda-41d7-9b44-f87396174ef2","Type":"ContainerStarted","Data":"09d0904c76ce78aa0b074447b1e64cf63160f573da086f934c8cb712615b71d4"} Dec 13 07:34:52 crc kubenswrapper[4971]: I1213 07:34:52.690654 4971 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/certified-operators-zvt24"] Dec 13 07:34:52 crc kubenswrapper[4971]: E1213 07:34:52.691832 4971 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="0e2af7b4-e1cb-4276-bece-56851503f2b0" containerName="collect-profiles" Dec 13 07:34:52 crc kubenswrapper[4971]: I1213 07:34:52.691854 4971 state_mem.go:107] "Deleted CPUSet assignment" podUID="0e2af7b4-e1cb-4276-bece-56851503f2b0" containerName="collect-profiles" Dec 13 07:34:52 crc kubenswrapper[4971]: I1213 07:34:52.692158 4971 memory_manager.go:354] "RemoveStaleState removing state" podUID="0e2af7b4-e1cb-4276-bece-56851503f2b0" containerName="collect-profiles" Dec 13 07:34:52 crc kubenswrapper[4971]: I1213 07:34:52.695727 4971 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-zvt24" Dec 13 07:34:52 crc kubenswrapper[4971]: I1213 07:34:52.718770 4971 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-zvt24"] Dec 13 07:34:52 crc kubenswrapper[4971]: I1213 07:34:52.822600 4971 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-d6f9b\" (UniqueName: \"kubernetes.io/projected/9ca340e2-ba26-4388-821e-6c034f3cd062-kube-api-access-d6f9b\") pod \"certified-operators-zvt24\" (UID: \"9ca340e2-ba26-4388-821e-6c034f3cd062\") " pod="openshift-marketplace/certified-operators-zvt24" Dec 13 07:34:52 crc kubenswrapper[4971]: I1213 07:34:52.822749 4971 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/9ca340e2-ba26-4388-821e-6c034f3cd062-catalog-content\") pod \"certified-operators-zvt24\" (UID: \"9ca340e2-ba26-4388-821e-6c034f3cd062\") " pod="openshift-marketplace/certified-operators-zvt24" Dec 13 07:34:52 crc kubenswrapper[4971]: I1213 07:34:52.822833 4971 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/9ca340e2-ba26-4388-821e-6c034f3cd062-utilities\") pod \"certified-operators-zvt24\" (UID: \"9ca340e2-ba26-4388-821e-6c034f3cd062\") " pod="openshift-marketplace/certified-operators-zvt24" Dec 13 07:34:52 crc kubenswrapper[4971]: I1213 07:34:52.924539 4971 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/9ca340e2-ba26-4388-821e-6c034f3cd062-catalog-content\") pod \"certified-operators-zvt24\" (UID: \"9ca340e2-ba26-4388-821e-6c034f3cd062\") " pod="openshift-marketplace/certified-operators-zvt24" Dec 13 07:34:52 crc kubenswrapper[4971]: I1213 07:34:52.925003 4971 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/9ca340e2-ba26-4388-821e-6c034f3cd062-utilities\") pod \"certified-operators-zvt24\" (UID: \"9ca340e2-ba26-4388-821e-6c034f3cd062\") " pod="openshift-marketplace/certified-operators-zvt24" Dec 13 07:34:52 crc kubenswrapper[4971]: I1213 07:34:52.925553 4971 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/9ca340e2-ba26-4388-821e-6c034f3cd062-utilities\") pod \"certified-operators-zvt24\" (UID: \"9ca340e2-ba26-4388-821e-6c034f3cd062\") " pod="openshift-marketplace/certified-operators-zvt24" Dec 13 07:34:52 crc kubenswrapper[4971]: I1213 07:34:52.925666 4971 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/9ca340e2-ba26-4388-821e-6c034f3cd062-catalog-content\") pod \"certified-operators-zvt24\" (UID: \"9ca340e2-ba26-4388-821e-6c034f3cd062\") " pod="openshift-marketplace/certified-operators-zvt24" Dec 13 07:34:52 crc kubenswrapper[4971]: I1213 07:34:52.926304 4971 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-d6f9b\" (UniqueName: \"kubernetes.io/projected/9ca340e2-ba26-4388-821e-6c034f3cd062-kube-api-access-d6f9b\") pod \"certified-operators-zvt24\" (UID: \"9ca340e2-ba26-4388-821e-6c034f3cd062\") " pod="openshift-marketplace/certified-operators-zvt24" Dec 13 07:34:52 crc kubenswrapper[4971]: I1213 07:34:52.956656 4971 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-d6f9b\" (UniqueName: \"kubernetes.io/projected/9ca340e2-ba26-4388-821e-6c034f3cd062-kube-api-access-d6f9b\") pod \"certified-operators-zvt24\" (UID: \"9ca340e2-ba26-4388-821e-6c034f3cd062\") " pod="openshift-marketplace/certified-operators-zvt24" Dec 13 07:34:53 crc kubenswrapper[4971]: I1213 07:34:53.022050 4971 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-zvt24" Dec 13 07:34:53 crc kubenswrapper[4971]: I1213 07:34:53.841914 4971 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-zvt24"] Dec 13 07:34:53 crc kubenswrapper[4971]: I1213 07:34:53.871105 4971 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-zvt24" event={"ID":"9ca340e2-ba26-4388-821e-6c034f3cd062","Type":"ContainerStarted","Data":"acb82fc2305e4517d491f8c925e76cf96e6a22734d6f4ebdaa420020a5b2bff5"} Dec 13 07:34:55 crc kubenswrapper[4971]: I1213 07:34:55.900380 4971 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-zvt24" event={"ID":"9ca340e2-ba26-4388-821e-6c034f3cd062","Type":"ContainerStarted","Data":"90a1634e4a179a9f8427b648a0013179523f5192295db8b2c71f6fb7ec5ad698"} Dec 13 07:34:56 crc kubenswrapper[4971]: I1213 07:34:56.912760 4971 generic.go:334] "Generic (PLEG): container finished" podID="9ca340e2-ba26-4388-821e-6c034f3cd062" containerID="90a1634e4a179a9f8427b648a0013179523f5192295db8b2c71f6fb7ec5ad698" exitCode=0 Dec 13 07:34:56 crc kubenswrapper[4971]: I1213 07:34:56.912870 4971 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-zvt24" event={"ID":"9ca340e2-ba26-4388-821e-6c034f3cd062","Type":"ContainerDied","Data":"90a1634e4a179a9f8427b648a0013179523f5192295db8b2c71f6fb7ec5ad698"} Dec 13 07:35:06 crc kubenswrapper[4971]: I1213 07:35:06.734990 4971 prober.go:107] "Probe failed" probeType="Liveness" pod="openstack/ceilometer-0" podUID="7cfb4dbf-cc55-48b7-b50c-65d6d4ae64ba" containerName="ceilometer-notification-agent" probeResult="failure" output="command timed out" Dec 13 07:35:07 crc kubenswrapper[4971]: I1213 07:35:07.875959 4971 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/kube-state-metrics-0" podUID="8314d6ed-4778-4709-a494-0642e0ce0617" containerName="kube-state-metrics" probeResult="failure" output="Get \"https://10.217.0.167:8081/readyz\": net/http: request canceled (Client.Timeout exceeded while awaiting headers)" Dec 13 07:35:25 crc kubenswrapper[4971]: I1213 07:35:25.207915 4971 generic.go:334] "Generic (PLEG): container finished" podID="9ca340e2-ba26-4388-821e-6c034f3cd062" containerID="9e08517a9e2441af7fe8d60feea944535375b36edd16a346ab145f0d4b246951" exitCode=0 Dec 13 07:35:25 crc kubenswrapper[4971]: I1213 07:35:25.207962 4971 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-zvt24" event={"ID":"9ca340e2-ba26-4388-821e-6c034f3cd062","Type":"ContainerDied","Data":"9e08517a9e2441af7fe8d60feea944535375b36edd16a346ab145f0d4b246951"} Dec 13 07:35:27 crc kubenswrapper[4971]: I1213 07:35:27.229610 4971 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-zvt24" event={"ID":"9ca340e2-ba26-4388-821e-6c034f3cd062","Type":"ContainerStarted","Data":"1b84e6138e16f2e388d7b18bd562490aba92d8cfe2b0209b054ff62d812ef4fc"} Dec 13 07:35:28 crc kubenswrapper[4971]: I1213 07:35:28.256605 4971 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/certified-operators-zvt24" podStartSLOduration=8.46128692 podStartE2EDuration="36.256587247s" podCreationTimestamp="2025-12-13 07:34:52 +0000 UTC" firstStartedPulling="2025-12-13 07:34:57.93134929 +0000 UTC m=+2754.535758768" lastFinishedPulling="2025-12-13 07:35:25.726649647 +0000 UTC m=+2782.331059095" observedRunningTime="2025-12-13 07:35:28.254730322 +0000 UTC m=+2784.859139820" watchObservedRunningTime="2025-12-13 07:35:28.256587247 +0000 UTC m=+2784.860996705" Dec 13 07:35:33 crc kubenswrapper[4971]: I1213 07:35:33.022683 4971 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/certified-operators-zvt24" Dec 13 07:35:33 crc kubenswrapper[4971]: I1213 07:35:33.023294 4971 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/certified-operators-zvt24" Dec 13 07:35:33 crc kubenswrapper[4971]: I1213 07:35:33.066027 4971 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/certified-operators-zvt24" Dec 13 07:35:33 crc kubenswrapper[4971]: I1213 07:35:33.340441 4971 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/certified-operators-zvt24" Dec 13 07:35:33 crc kubenswrapper[4971]: I1213 07:35:33.403908 4971 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-zvt24"] Dec 13 07:35:35 crc kubenswrapper[4971]: I1213 07:35:35.303710 4971 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/certified-operators-zvt24" podUID="9ca340e2-ba26-4388-821e-6c034f3cd062" containerName="registry-server" containerID="cri-o://1b84e6138e16f2e388d7b18bd562490aba92d8cfe2b0209b054ff62d812ef4fc" gracePeriod=2 Dec 13 07:35:36 crc kubenswrapper[4971]: I1213 07:35:36.283231 4971 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-zvt24" Dec 13 07:35:36 crc kubenswrapper[4971]: I1213 07:35:36.326091 4971 generic.go:334] "Generic (PLEG): container finished" podID="9ca340e2-ba26-4388-821e-6c034f3cd062" containerID="1b84e6138e16f2e388d7b18bd562490aba92d8cfe2b0209b054ff62d812ef4fc" exitCode=0 Dec 13 07:35:36 crc kubenswrapper[4971]: I1213 07:35:36.326146 4971 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-zvt24" event={"ID":"9ca340e2-ba26-4388-821e-6c034f3cd062","Type":"ContainerDied","Data":"1b84e6138e16f2e388d7b18bd562490aba92d8cfe2b0209b054ff62d812ef4fc"} Dec 13 07:35:36 crc kubenswrapper[4971]: I1213 07:35:36.326185 4971 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-zvt24" event={"ID":"9ca340e2-ba26-4388-821e-6c034f3cd062","Type":"ContainerDied","Data":"acb82fc2305e4517d491f8c925e76cf96e6a22734d6f4ebdaa420020a5b2bff5"} Dec 13 07:35:36 crc kubenswrapper[4971]: I1213 07:35:36.326206 4971 scope.go:117] "RemoveContainer" containerID="1b84e6138e16f2e388d7b18bd562490aba92d8cfe2b0209b054ff62d812ef4fc" Dec 13 07:35:36 crc kubenswrapper[4971]: I1213 07:35:36.327108 4971 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-zvt24" Dec 13 07:35:36 crc kubenswrapper[4971]: I1213 07:35:36.343796 4971 scope.go:117] "RemoveContainer" containerID="9e08517a9e2441af7fe8d60feea944535375b36edd16a346ab145f0d4b246951" Dec 13 07:35:36 crc kubenswrapper[4971]: I1213 07:35:36.364880 4971 scope.go:117] "RemoveContainer" containerID="90a1634e4a179a9f8427b648a0013179523f5192295db8b2c71f6fb7ec5ad698" Dec 13 07:35:36 crc kubenswrapper[4971]: I1213 07:35:36.400388 4971 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/9ca340e2-ba26-4388-821e-6c034f3cd062-utilities\") pod \"9ca340e2-ba26-4388-821e-6c034f3cd062\" (UID: \"9ca340e2-ba26-4388-821e-6c034f3cd062\") " Dec 13 07:35:36 crc kubenswrapper[4971]: I1213 07:35:36.400713 4971 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/9ca340e2-ba26-4388-821e-6c034f3cd062-catalog-content\") pod \"9ca340e2-ba26-4388-821e-6c034f3cd062\" (UID: \"9ca340e2-ba26-4388-821e-6c034f3cd062\") " Dec 13 07:35:36 crc kubenswrapper[4971]: I1213 07:35:36.400861 4971 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-d6f9b\" (UniqueName: \"kubernetes.io/projected/9ca340e2-ba26-4388-821e-6c034f3cd062-kube-api-access-d6f9b\") pod \"9ca340e2-ba26-4388-821e-6c034f3cd062\" (UID: \"9ca340e2-ba26-4388-821e-6c034f3cd062\") " Dec 13 07:35:36 crc kubenswrapper[4971]: I1213 07:35:36.401487 4971 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/9ca340e2-ba26-4388-821e-6c034f3cd062-utilities" (OuterVolumeSpecName: "utilities") pod "9ca340e2-ba26-4388-821e-6c034f3cd062" (UID: "9ca340e2-ba26-4388-821e-6c034f3cd062"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 13 07:35:36 crc kubenswrapper[4971]: I1213 07:35:36.407761 4971 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/9ca340e2-ba26-4388-821e-6c034f3cd062-kube-api-access-d6f9b" (OuterVolumeSpecName: "kube-api-access-d6f9b") pod "9ca340e2-ba26-4388-821e-6c034f3cd062" (UID: "9ca340e2-ba26-4388-821e-6c034f3cd062"). InnerVolumeSpecName "kube-api-access-d6f9b". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 13 07:35:36 crc kubenswrapper[4971]: I1213 07:35:36.413506 4971 scope.go:117] "RemoveContainer" containerID="1b84e6138e16f2e388d7b18bd562490aba92d8cfe2b0209b054ff62d812ef4fc" Dec 13 07:35:36 crc kubenswrapper[4971]: E1213 07:35:36.414239 4971 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"1b84e6138e16f2e388d7b18bd562490aba92d8cfe2b0209b054ff62d812ef4fc\": container with ID starting with 1b84e6138e16f2e388d7b18bd562490aba92d8cfe2b0209b054ff62d812ef4fc not found: ID does not exist" containerID="1b84e6138e16f2e388d7b18bd562490aba92d8cfe2b0209b054ff62d812ef4fc" Dec 13 07:35:36 crc kubenswrapper[4971]: I1213 07:35:36.414366 4971 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"1b84e6138e16f2e388d7b18bd562490aba92d8cfe2b0209b054ff62d812ef4fc"} err="failed to get container status \"1b84e6138e16f2e388d7b18bd562490aba92d8cfe2b0209b054ff62d812ef4fc\": rpc error: code = NotFound desc = could not find container \"1b84e6138e16f2e388d7b18bd562490aba92d8cfe2b0209b054ff62d812ef4fc\": container with ID starting with 1b84e6138e16f2e388d7b18bd562490aba92d8cfe2b0209b054ff62d812ef4fc not found: ID does not exist" Dec 13 07:35:36 crc kubenswrapper[4971]: I1213 07:35:36.414471 4971 scope.go:117] "RemoveContainer" containerID="9e08517a9e2441af7fe8d60feea944535375b36edd16a346ab145f0d4b246951" Dec 13 07:35:36 crc kubenswrapper[4971]: E1213 07:35:36.418745 4971 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"9e08517a9e2441af7fe8d60feea944535375b36edd16a346ab145f0d4b246951\": container with ID starting with 9e08517a9e2441af7fe8d60feea944535375b36edd16a346ab145f0d4b246951 not found: ID does not exist" containerID="9e08517a9e2441af7fe8d60feea944535375b36edd16a346ab145f0d4b246951" Dec 13 07:35:36 crc kubenswrapper[4971]: I1213 07:35:36.418802 4971 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"9e08517a9e2441af7fe8d60feea944535375b36edd16a346ab145f0d4b246951"} err="failed to get container status \"9e08517a9e2441af7fe8d60feea944535375b36edd16a346ab145f0d4b246951\": rpc error: code = NotFound desc = could not find container \"9e08517a9e2441af7fe8d60feea944535375b36edd16a346ab145f0d4b246951\": container with ID starting with 9e08517a9e2441af7fe8d60feea944535375b36edd16a346ab145f0d4b246951 not found: ID does not exist" Dec 13 07:35:36 crc kubenswrapper[4971]: I1213 07:35:36.418835 4971 scope.go:117] "RemoveContainer" containerID="90a1634e4a179a9f8427b648a0013179523f5192295db8b2c71f6fb7ec5ad698" Dec 13 07:35:36 crc kubenswrapper[4971]: E1213 07:35:36.421025 4971 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"90a1634e4a179a9f8427b648a0013179523f5192295db8b2c71f6fb7ec5ad698\": container with ID starting with 90a1634e4a179a9f8427b648a0013179523f5192295db8b2c71f6fb7ec5ad698 not found: ID does not exist" containerID="90a1634e4a179a9f8427b648a0013179523f5192295db8b2c71f6fb7ec5ad698" Dec 13 07:35:36 crc kubenswrapper[4971]: I1213 07:35:36.421079 4971 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"90a1634e4a179a9f8427b648a0013179523f5192295db8b2c71f6fb7ec5ad698"} err="failed to get container status \"90a1634e4a179a9f8427b648a0013179523f5192295db8b2c71f6fb7ec5ad698\": rpc error: code = NotFound desc = could not find container \"90a1634e4a179a9f8427b648a0013179523f5192295db8b2c71f6fb7ec5ad698\": container with ID starting with 90a1634e4a179a9f8427b648a0013179523f5192295db8b2c71f6fb7ec5ad698 not found: ID does not exist" Dec 13 07:35:36 crc kubenswrapper[4971]: I1213 07:35:36.464061 4971 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/9ca340e2-ba26-4388-821e-6c034f3cd062-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "9ca340e2-ba26-4388-821e-6c034f3cd062" (UID: "9ca340e2-ba26-4388-821e-6c034f3cd062"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 13 07:35:36 crc kubenswrapper[4971]: I1213 07:35:36.502789 4971 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/9ca340e2-ba26-4388-821e-6c034f3cd062-utilities\") on node \"crc\" DevicePath \"\"" Dec 13 07:35:36 crc kubenswrapper[4971]: I1213 07:35:36.502829 4971 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/9ca340e2-ba26-4388-821e-6c034f3cd062-catalog-content\") on node \"crc\" DevicePath \"\"" Dec 13 07:35:36 crc kubenswrapper[4971]: I1213 07:35:36.502845 4971 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-d6f9b\" (UniqueName: \"kubernetes.io/projected/9ca340e2-ba26-4388-821e-6c034f3cd062-kube-api-access-d6f9b\") on node \"crc\" DevicePath \"\"" Dec 13 07:35:36 crc kubenswrapper[4971]: I1213 07:35:36.659430 4971 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-zvt24"] Dec 13 07:35:36 crc kubenswrapper[4971]: I1213 07:35:36.669613 4971 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/certified-operators-zvt24"] Dec 13 07:35:37 crc kubenswrapper[4971]: I1213 07:35:37.786670 4971 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="9ca340e2-ba26-4388-821e-6c034f3cd062" path="/var/lib/kubelet/pods/9ca340e2-ba26-4388-821e-6c034f3cd062/volumes" Dec 13 07:35:40 crc kubenswrapper[4971]: I1213 07:35:40.582911 4971 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-7nvdr"] Dec 13 07:35:40 crc kubenswrapper[4971]: I1213 07:35:40.597641 4971 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/configure-network-edpm-deployment-openstack-edpm-ipam-qj5wr"] Dec 13 07:35:40 crc kubenswrapper[4971]: I1213 07:35:40.610905 4971 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-8jf6x"] Dec 13 07:35:40 crc kubenswrapper[4971]: I1213 07:35:40.618598 4971 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-fgt8j"] Dec 13 07:35:40 crc kubenswrapper[4971]: I1213 07:35:40.627014 4971 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-q8m6f"] Dec 13 07:35:40 crc kubenswrapper[4971]: I1213 07:35:40.633235 4971 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ssh-known-hosts-edpm-deployment-2stjx"] Dec 13 07:35:40 crc kubenswrapper[4971]: I1213 07:35:40.639529 4971 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/validate-network-edpm-deployment-openstack-edpm-ipam-wx2k2"] Dec 13 07:35:40 crc kubenswrapper[4971]: I1213 07:35:40.644926 4971 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-7nvdr"] Dec 13 07:35:40 crc kubenswrapper[4971]: I1213 07:35:40.650255 4971 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-rk2vb"] Dec 13 07:35:40 crc kubenswrapper[4971]: I1213 07:35:40.655798 4971 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-8vg77"] Dec 13 07:35:40 crc kubenswrapper[4971]: I1213 07:35:40.661158 4971 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ceph-hci-pre-edpm-deployment-openstack-edpm-ipam-llgqk"] Dec 13 07:35:40 crc kubenswrapper[4971]: I1213 07:35:40.667715 4971 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/install-os-edpm-deployment-openstack-edpm-ipam-n2hlm"] Dec 13 07:35:40 crc kubenswrapper[4971]: I1213 07:35:40.673188 4971 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/configure-os-edpm-deployment-openstack-edpm-ipam-zrl24"] Dec 13 07:35:40 crc kubenswrapper[4971]: I1213 07:35:40.679009 4971 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/configure-network-edpm-deployment-openstack-edpm-ipam-qj5wr"] Dec 13 07:35:40 crc kubenswrapper[4971]: I1213 07:35:40.684213 4971 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/run-os-edpm-deployment-openstack-edpm-ipam-dfs5x"] Dec 13 07:35:40 crc kubenswrapper[4971]: I1213 07:35:40.689416 4971 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-q8m6f"] Dec 13 07:35:40 crc kubenswrapper[4971]: I1213 07:35:40.694434 4971 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-8jf6x"] Dec 13 07:35:40 crc kubenswrapper[4971]: I1213 07:35:40.704172 4971 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/ceph-hci-pre-edpm-deployment-openstack-edpm-ipam-llgqk"] Dec 13 07:35:40 crc kubenswrapper[4971]: I1213 07:35:40.709412 4971 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/validate-network-edpm-deployment-openstack-edpm-ipam-wx2k2"] Dec 13 07:35:40 crc kubenswrapper[4971]: I1213 07:35:40.714756 4971 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-fgt8j"] Dec 13 07:35:40 crc kubenswrapper[4971]: I1213 07:35:40.734188 4971 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/ssh-known-hosts-edpm-deployment-2stjx"] Dec 13 07:35:40 crc kubenswrapper[4971]: I1213 07:35:40.747555 4971 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/install-os-edpm-deployment-openstack-edpm-ipam-n2hlm"] Dec 13 07:35:40 crc kubenswrapper[4971]: I1213 07:35:40.759708 4971 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-rk2vb"] Dec 13 07:35:40 crc kubenswrapper[4971]: I1213 07:35:40.768883 4971 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/run-os-edpm-deployment-openstack-edpm-ipam-dfs5x"] Dec 13 07:35:40 crc kubenswrapper[4971]: I1213 07:35:40.779145 4971 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/configure-os-edpm-deployment-openstack-edpm-ipam-zrl24"] Dec 13 07:35:40 crc kubenswrapper[4971]: I1213 07:35:40.785222 4971 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-hjgk2"] Dec 13 07:35:40 crc kubenswrapper[4971]: I1213 07:35:40.793352 4971 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-hjgk2"] Dec 13 07:35:40 crc kubenswrapper[4971]: I1213 07:35:40.802361 4971 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-8vg77"] Dec 13 07:35:41 crc kubenswrapper[4971]: I1213 07:35:41.779259 4971 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="62eac784-f59b-4764-ad97-681df8129497" path="/var/lib/kubelet/pods/62eac784-f59b-4764-ad97-681df8129497/volumes" Dec 13 07:35:41 crc kubenswrapper[4971]: I1213 07:35:41.780449 4971 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="68fb082e-8ed6-4ecc-88a9-d349440dd1b3" path="/var/lib/kubelet/pods/68fb082e-8ed6-4ecc-88a9-d349440dd1b3/volumes" Dec 13 07:35:41 crc kubenswrapper[4971]: I1213 07:35:41.781219 4971 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="80f5a7cb-cc56-4d8e-ba65-0198d682bcc5" path="/var/lib/kubelet/pods/80f5a7cb-cc56-4d8e-ba65-0198d682bcc5/volumes" Dec 13 07:35:41 crc kubenswrapper[4971]: I1213 07:35:41.781970 4971 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="895771e3-67cb-4847-ab4b-3b7edfe79102" path="/var/lib/kubelet/pods/895771e3-67cb-4847-ab4b-3b7edfe79102/volumes" Dec 13 07:35:41 crc kubenswrapper[4971]: I1213 07:35:41.783294 4971 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="90351fb3-f8c9-4db4-b577-8820f50ccf15" path="/var/lib/kubelet/pods/90351fb3-f8c9-4db4-b577-8820f50ccf15/volumes" Dec 13 07:35:41 crc kubenswrapper[4971]: I1213 07:35:41.784060 4971 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="912b00fe-da52-454a-924c-e49620ecd7fd" path="/var/lib/kubelet/pods/912b00fe-da52-454a-924c-e49620ecd7fd/volumes" Dec 13 07:35:41 crc kubenswrapper[4971]: I1213 07:35:41.784822 4971 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="a4963e97-7f24-46ad-bd42-8b6c2b2d0274" path="/var/lib/kubelet/pods/a4963e97-7f24-46ad-bd42-8b6c2b2d0274/volumes" Dec 13 07:35:41 crc kubenswrapper[4971]: I1213 07:35:41.786307 4971 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="be382cbb-023b-4d16-9e01-b2eb44663da2" path="/var/lib/kubelet/pods/be382cbb-023b-4d16-9e01-b2eb44663da2/volumes" Dec 13 07:35:41 crc kubenswrapper[4971]: I1213 07:35:41.787025 4971 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="c437fb6f-411b-4af7-bf7c-3aad6306e4f2" path="/var/lib/kubelet/pods/c437fb6f-411b-4af7-bf7c-3aad6306e4f2/volumes" Dec 13 07:35:41 crc kubenswrapper[4971]: I1213 07:35:41.787920 4971 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="e3950b9b-9046-4139-8cd6-8f099cef93e4" path="/var/lib/kubelet/pods/e3950b9b-9046-4139-8cd6-8f099cef93e4/volumes" Dec 13 07:35:41 crc kubenswrapper[4971]: I1213 07:35:41.789212 4971 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="edef63aa-a18d-4591-a477-924e5cef56d8" path="/var/lib/kubelet/pods/edef63aa-a18d-4591-a477-924e5cef56d8/volumes" Dec 13 07:35:41 crc kubenswrapper[4971]: I1213 07:35:41.790021 4971 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="f00ef9b1-1293-4273-90d1-7a7419ebf996" path="/var/lib/kubelet/pods/f00ef9b1-1293-4273-90d1-7a7419ebf996/volumes" Dec 13 07:35:41 crc kubenswrapper[4971]: I1213 07:35:41.790804 4971 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="f468bc03-51b5-4b38-a2b5-b9061002f4d7" path="/var/lib/kubelet/pods/f468bc03-51b5-4b38-a2b5-b9061002f4d7/volumes" Dec 13 07:35:41 crc kubenswrapper[4971]: I1213 07:35:41.791464 4971 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="fc0d806f-23ce-4920-8932-920a739c700c" path="/var/lib/kubelet/pods/fc0d806f-23ce-4920-8932-920a739c700c/volumes" Dec 13 07:35:46 crc kubenswrapper[4971]: I1213 07:35:46.750895 4971 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-fpgpm"] Dec 13 07:35:46 crc kubenswrapper[4971]: E1213 07:35:46.751702 4971 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="9ca340e2-ba26-4388-821e-6c034f3cd062" containerName="registry-server" Dec 13 07:35:46 crc kubenswrapper[4971]: I1213 07:35:46.751714 4971 state_mem.go:107] "Deleted CPUSet assignment" podUID="9ca340e2-ba26-4388-821e-6c034f3cd062" containerName="registry-server" Dec 13 07:35:46 crc kubenswrapper[4971]: E1213 07:35:46.751725 4971 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="9ca340e2-ba26-4388-821e-6c034f3cd062" containerName="extract-utilities" Dec 13 07:35:46 crc kubenswrapper[4971]: I1213 07:35:46.751732 4971 state_mem.go:107] "Deleted CPUSet assignment" podUID="9ca340e2-ba26-4388-821e-6c034f3cd062" containerName="extract-utilities" Dec 13 07:35:46 crc kubenswrapper[4971]: E1213 07:35:46.751746 4971 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="9ca340e2-ba26-4388-821e-6c034f3cd062" containerName="extract-content" Dec 13 07:35:46 crc kubenswrapper[4971]: I1213 07:35:46.751754 4971 state_mem.go:107] "Deleted CPUSet assignment" podUID="9ca340e2-ba26-4388-821e-6c034f3cd062" containerName="extract-content" Dec 13 07:35:46 crc kubenswrapper[4971]: I1213 07:35:46.751924 4971 memory_manager.go:354] "RemoveStaleState removing state" podUID="9ca340e2-ba26-4388-821e-6c034f3cd062" containerName="registry-server" Dec 13 07:35:46 crc kubenswrapper[4971]: I1213 07:35:46.752449 4971 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-fpgpm" Dec 13 07:35:46 crc kubenswrapper[4971]: I1213 07:35:46.754888 4971 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplane-ansible-ssh-private-key-secret" Dec 13 07:35:46 crc kubenswrapper[4971]: I1213 07:35:46.760319 4971 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceph-conf-files" Dec 13 07:35:46 crc kubenswrapper[4971]: I1213 07:35:46.760601 4971 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-edpm-ipam-dockercfg-zb2j5" Dec 13 07:35:46 crc kubenswrapper[4971]: I1213 07:35:46.760758 4971 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplanenodeset-openstack-edpm-ipam" Dec 13 07:35:46 crc kubenswrapper[4971]: I1213 07:35:46.762676 4971 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-aee-default-env" Dec 13 07:35:46 crc kubenswrapper[4971]: I1213 07:35:46.785978 4971 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-fpgpm"] Dec 13 07:35:46 crc kubenswrapper[4971]: I1213 07:35:46.895087 4971 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-9zqc8\" (UniqueName: \"kubernetes.io/projected/53d7daa1-9679-44ba-9023-808ce5c250a6-kube-api-access-9zqc8\") pod \"repo-setup-edpm-deployment-openstack-edpm-ipam-fpgpm\" (UID: \"53d7daa1-9679-44ba-9023-808ce5c250a6\") " pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-fpgpm" Dec 13 07:35:46 crc kubenswrapper[4971]: I1213 07:35:46.895154 4971 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"repo-setup-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/53d7daa1-9679-44ba-9023-808ce5c250a6-repo-setup-combined-ca-bundle\") pod \"repo-setup-edpm-deployment-openstack-edpm-ipam-fpgpm\" (UID: \"53d7daa1-9679-44ba-9023-808ce5c250a6\") " pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-fpgpm" Dec 13 07:35:46 crc kubenswrapper[4971]: I1213 07:35:46.895326 4971 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/53d7daa1-9679-44ba-9023-808ce5c250a6-inventory\") pod \"repo-setup-edpm-deployment-openstack-edpm-ipam-fpgpm\" (UID: \"53d7daa1-9679-44ba-9023-808ce5c250a6\") " pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-fpgpm" Dec 13 07:35:46 crc kubenswrapper[4971]: I1213 07:35:46.895587 4971 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/53d7daa1-9679-44ba-9023-808ce5c250a6-ssh-key\") pod \"repo-setup-edpm-deployment-openstack-edpm-ipam-fpgpm\" (UID: \"53d7daa1-9679-44ba-9023-808ce5c250a6\") " pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-fpgpm" Dec 13 07:35:46 crc kubenswrapper[4971]: I1213 07:35:46.895817 4971 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/53d7daa1-9679-44ba-9023-808ce5c250a6-ceph\") pod \"repo-setup-edpm-deployment-openstack-edpm-ipam-fpgpm\" (UID: \"53d7daa1-9679-44ba-9023-808ce5c250a6\") " pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-fpgpm" Dec 13 07:35:46 crc kubenswrapper[4971]: I1213 07:35:46.997490 4971 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-9zqc8\" (UniqueName: \"kubernetes.io/projected/53d7daa1-9679-44ba-9023-808ce5c250a6-kube-api-access-9zqc8\") pod \"repo-setup-edpm-deployment-openstack-edpm-ipam-fpgpm\" (UID: \"53d7daa1-9679-44ba-9023-808ce5c250a6\") " pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-fpgpm" Dec 13 07:35:46 crc kubenswrapper[4971]: I1213 07:35:46.997558 4971 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"repo-setup-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/53d7daa1-9679-44ba-9023-808ce5c250a6-repo-setup-combined-ca-bundle\") pod \"repo-setup-edpm-deployment-openstack-edpm-ipam-fpgpm\" (UID: \"53d7daa1-9679-44ba-9023-808ce5c250a6\") " pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-fpgpm" Dec 13 07:35:46 crc kubenswrapper[4971]: I1213 07:35:46.997600 4971 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/53d7daa1-9679-44ba-9023-808ce5c250a6-inventory\") pod \"repo-setup-edpm-deployment-openstack-edpm-ipam-fpgpm\" (UID: \"53d7daa1-9679-44ba-9023-808ce5c250a6\") " pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-fpgpm" Dec 13 07:35:46 crc kubenswrapper[4971]: I1213 07:35:46.997644 4971 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/53d7daa1-9679-44ba-9023-808ce5c250a6-ssh-key\") pod \"repo-setup-edpm-deployment-openstack-edpm-ipam-fpgpm\" (UID: \"53d7daa1-9679-44ba-9023-808ce5c250a6\") " pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-fpgpm" Dec 13 07:35:46 crc kubenswrapper[4971]: I1213 07:35:46.997682 4971 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/53d7daa1-9679-44ba-9023-808ce5c250a6-ceph\") pod \"repo-setup-edpm-deployment-openstack-edpm-ipam-fpgpm\" (UID: \"53d7daa1-9679-44ba-9023-808ce5c250a6\") " pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-fpgpm" Dec 13 07:35:47 crc kubenswrapper[4971]: I1213 07:35:47.004339 4971 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/53d7daa1-9679-44ba-9023-808ce5c250a6-inventory\") pod \"repo-setup-edpm-deployment-openstack-edpm-ipam-fpgpm\" (UID: \"53d7daa1-9679-44ba-9023-808ce5c250a6\") " pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-fpgpm" Dec 13 07:35:47 crc kubenswrapper[4971]: I1213 07:35:47.004380 4971 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/53d7daa1-9679-44ba-9023-808ce5c250a6-ceph\") pod \"repo-setup-edpm-deployment-openstack-edpm-ipam-fpgpm\" (UID: \"53d7daa1-9679-44ba-9023-808ce5c250a6\") " pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-fpgpm" Dec 13 07:35:47 crc kubenswrapper[4971]: I1213 07:35:47.005791 4971 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"repo-setup-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/53d7daa1-9679-44ba-9023-808ce5c250a6-repo-setup-combined-ca-bundle\") pod \"repo-setup-edpm-deployment-openstack-edpm-ipam-fpgpm\" (UID: \"53d7daa1-9679-44ba-9023-808ce5c250a6\") " pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-fpgpm" Dec 13 07:35:47 crc kubenswrapper[4971]: I1213 07:35:47.006621 4971 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/53d7daa1-9679-44ba-9023-808ce5c250a6-ssh-key\") pod \"repo-setup-edpm-deployment-openstack-edpm-ipam-fpgpm\" (UID: \"53d7daa1-9679-44ba-9023-808ce5c250a6\") " pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-fpgpm" Dec 13 07:35:47 crc kubenswrapper[4971]: I1213 07:35:47.012978 4971 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-9zqc8\" (UniqueName: \"kubernetes.io/projected/53d7daa1-9679-44ba-9023-808ce5c250a6-kube-api-access-9zqc8\") pod \"repo-setup-edpm-deployment-openstack-edpm-ipam-fpgpm\" (UID: \"53d7daa1-9679-44ba-9023-808ce5c250a6\") " pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-fpgpm" Dec 13 07:35:47 crc kubenswrapper[4971]: I1213 07:35:47.087330 4971 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-fpgpm" Dec 13 07:35:47 crc kubenswrapper[4971]: I1213 07:35:47.592758 4971 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-fpgpm"] Dec 13 07:35:48 crc kubenswrapper[4971]: I1213 07:35:48.437129 4971 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-fpgpm" event={"ID":"53d7daa1-9679-44ba-9023-808ce5c250a6","Type":"ContainerStarted","Data":"08567213e9e080d44b32aeece48b01aaed5896da02e3626a51a3418002df7866"} Dec 13 07:35:49 crc kubenswrapper[4971]: I1213 07:35:49.451214 4971 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-fpgpm" event={"ID":"53d7daa1-9679-44ba-9023-808ce5c250a6","Type":"ContainerStarted","Data":"0e8eb58c175eeacdc0c96ced58fd751e4b04cffb0f1c6ba2efe9fa41ef2ab9cd"} Dec 13 07:35:49 crc kubenswrapper[4971]: I1213 07:35:49.487461 4971 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-fpgpm" podStartSLOduration=2.7055326109999998 podStartE2EDuration="3.48743028s" podCreationTimestamp="2025-12-13 07:35:46 +0000 UTC" firstStartedPulling="2025-12-13 07:35:47.600064146 +0000 UTC m=+2804.204473594" lastFinishedPulling="2025-12-13 07:35:48.381961815 +0000 UTC m=+2804.986371263" observedRunningTime="2025-12-13 07:35:49.466465266 +0000 UTC m=+2806.070874714" watchObservedRunningTime="2025-12-13 07:35:49.48743028 +0000 UTC m=+2806.091839758" Dec 13 07:36:02 crc kubenswrapper[4971]: I1213 07:36:02.576166 4971 generic.go:334] "Generic (PLEG): container finished" podID="53d7daa1-9679-44ba-9023-808ce5c250a6" containerID="0e8eb58c175eeacdc0c96ced58fd751e4b04cffb0f1c6ba2efe9fa41ef2ab9cd" exitCode=0 Dec 13 07:36:02 crc kubenswrapper[4971]: I1213 07:36:02.576268 4971 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-fpgpm" event={"ID":"53d7daa1-9679-44ba-9023-808ce5c250a6","Type":"ContainerDied","Data":"0e8eb58c175eeacdc0c96ced58fd751e4b04cffb0f1c6ba2efe9fa41ef2ab9cd"} Dec 13 07:36:04 crc kubenswrapper[4971]: I1213 07:36:04.046755 4971 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-fpgpm" Dec 13 07:36:04 crc kubenswrapper[4971]: I1213 07:36:04.244255 4971 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/53d7daa1-9679-44ba-9023-808ce5c250a6-ceph\") pod \"53d7daa1-9679-44ba-9023-808ce5c250a6\" (UID: \"53d7daa1-9679-44ba-9023-808ce5c250a6\") " Dec 13 07:36:04 crc kubenswrapper[4971]: I1213 07:36:04.244320 4971 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-9zqc8\" (UniqueName: \"kubernetes.io/projected/53d7daa1-9679-44ba-9023-808ce5c250a6-kube-api-access-9zqc8\") pod \"53d7daa1-9679-44ba-9023-808ce5c250a6\" (UID: \"53d7daa1-9679-44ba-9023-808ce5c250a6\") " Dec 13 07:36:04 crc kubenswrapper[4971]: I1213 07:36:04.244375 4971 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/53d7daa1-9679-44ba-9023-808ce5c250a6-ssh-key\") pod \"53d7daa1-9679-44ba-9023-808ce5c250a6\" (UID: \"53d7daa1-9679-44ba-9023-808ce5c250a6\") " Dec 13 07:36:04 crc kubenswrapper[4971]: I1213 07:36:04.244418 4971 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/53d7daa1-9679-44ba-9023-808ce5c250a6-inventory\") pod \"53d7daa1-9679-44ba-9023-808ce5c250a6\" (UID: \"53d7daa1-9679-44ba-9023-808ce5c250a6\") " Dec 13 07:36:04 crc kubenswrapper[4971]: I1213 07:36:04.244456 4971 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"repo-setup-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/53d7daa1-9679-44ba-9023-808ce5c250a6-repo-setup-combined-ca-bundle\") pod \"53d7daa1-9679-44ba-9023-808ce5c250a6\" (UID: \"53d7daa1-9679-44ba-9023-808ce5c250a6\") " Dec 13 07:36:04 crc kubenswrapper[4971]: I1213 07:36:04.250388 4971 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/53d7daa1-9679-44ba-9023-808ce5c250a6-repo-setup-combined-ca-bundle" (OuterVolumeSpecName: "repo-setup-combined-ca-bundle") pod "53d7daa1-9679-44ba-9023-808ce5c250a6" (UID: "53d7daa1-9679-44ba-9023-808ce5c250a6"). InnerVolumeSpecName "repo-setup-combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 13 07:36:04 crc kubenswrapper[4971]: I1213 07:36:04.251299 4971 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/53d7daa1-9679-44ba-9023-808ce5c250a6-kube-api-access-9zqc8" (OuterVolumeSpecName: "kube-api-access-9zqc8") pod "53d7daa1-9679-44ba-9023-808ce5c250a6" (UID: "53d7daa1-9679-44ba-9023-808ce5c250a6"). InnerVolumeSpecName "kube-api-access-9zqc8". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 13 07:36:04 crc kubenswrapper[4971]: I1213 07:36:04.254940 4971 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/53d7daa1-9679-44ba-9023-808ce5c250a6-ceph" (OuterVolumeSpecName: "ceph") pod "53d7daa1-9679-44ba-9023-808ce5c250a6" (UID: "53d7daa1-9679-44ba-9023-808ce5c250a6"). InnerVolumeSpecName "ceph". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 13 07:36:04 crc kubenswrapper[4971]: I1213 07:36:04.272565 4971 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/53d7daa1-9679-44ba-9023-808ce5c250a6-inventory" (OuterVolumeSpecName: "inventory") pod "53d7daa1-9679-44ba-9023-808ce5c250a6" (UID: "53d7daa1-9679-44ba-9023-808ce5c250a6"). InnerVolumeSpecName "inventory". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 13 07:36:04 crc kubenswrapper[4971]: I1213 07:36:04.275660 4971 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/53d7daa1-9679-44ba-9023-808ce5c250a6-ssh-key" (OuterVolumeSpecName: "ssh-key") pod "53d7daa1-9679-44ba-9023-808ce5c250a6" (UID: "53d7daa1-9679-44ba-9023-808ce5c250a6"). InnerVolumeSpecName "ssh-key". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 13 07:36:04 crc kubenswrapper[4971]: I1213 07:36:04.347739 4971 reconciler_common.go:293] "Volume detached for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/53d7daa1-9679-44ba-9023-808ce5c250a6-ceph\") on node \"crc\" DevicePath \"\"" Dec 13 07:36:04 crc kubenswrapper[4971]: I1213 07:36:04.347779 4971 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-9zqc8\" (UniqueName: \"kubernetes.io/projected/53d7daa1-9679-44ba-9023-808ce5c250a6-kube-api-access-9zqc8\") on node \"crc\" DevicePath \"\"" Dec 13 07:36:04 crc kubenswrapper[4971]: I1213 07:36:04.347794 4971 reconciler_common.go:293] "Volume detached for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/53d7daa1-9679-44ba-9023-808ce5c250a6-ssh-key\") on node \"crc\" DevicePath \"\"" Dec 13 07:36:04 crc kubenswrapper[4971]: I1213 07:36:04.347806 4971 reconciler_common.go:293] "Volume detached for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/53d7daa1-9679-44ba-9023-808ce5c250a6-inventory\") on node \"crc\" DevicePath \"\"" Dec 13 07:36:04 crc kubenswrapper[4971]: I1213 07:36:04.347818 4971 reconciler_common.go:293] "Volume detached for volume \"repo-setup-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/53d7daa1-9679-44ba-9023-808ce5c250a6-repo-setup-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 13 07:36:04 crc kubenswrapper[4971]: I1213 07:36:04.631483 4971 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-fpgpm" event={"ID":"53d7daa1-9679-44ba-9023-808ce5c250a6","Type":"ContainerDied","Data":"08567213e9e080d44b32aeece48b01aaed5896da02e3626a51a3418002df7866"} Dec 13 07:36:04 crc kubenswrapper[4971]: I1213 07:36:04.631806 4971 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="08567213e9e080d44b32aeece48b01aaed5896da02e3626a51a3418002df7866" Dec 13 07:36:04 crc kubenswrapper[4971]: I1213 07:36:04.631564 4971 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-fpgpm" Dec 13 07:36:04 crc kubenswrapper[4971]: I1213 07:36:04.742968 4971 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-jdcq5"] Dec 13 07:36:04 crc kubenswrapper[4971]: E1213 07:36:04.743582 4971 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="53d7daa1-9679-44ba-9023-808ce5c250a6" containerName="repo-setup-edpm-deployment-openstack-edpm-ipam" Dec 13 07:36:04 crc kubenswrapper[4971]: I1213 07:36:04.743613 4971 state_mem.go:107] "Deleted CPUSet assignment" podUID="53d7daa1-9679-44ba-9023-808ce5c250a6" containerName="repo-setup-edpm-deployment-openstack-edpm-ipam" Dec 13 07:36:04 crc kubenswrapper[4971]: I1213 07:36:04.743941 4971 memory_manager.go:354] "RemoveStaleState removing state" podUID="53d7daa1-9679-44ba-9023-808ce5c250a6" containerName="repo-setup-edpm-deployment-openstack-edpm-ipam" Dec 13 07:36:04 crc kubenswrapper[4971]: I1213 07:36:04.744878 4971 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-jdcq5" Dec 13 07:36:04 crc kubenswrapper[4971]: I1213 07:36:04.746782 4971 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplanenodeset-openstack-edpm-ipam" Dec 13 07:36:04 crc kubenswrapper[4971]: I1213 07:36:04.747206 4971 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceph-conf-files" Dec 13 07:36:04 crc kubenswrapper[4971]: I1213 07:36:04.747457 4971 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplane-ansible-ssh-private-key-secret" Dec 13 07:36:04 crc kubenswrapper[4971]: I1213 07:36:04.747769 4971 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-aee-default-env" Dec 13 07:36:04 crc kubenswrapper[4971]: I1213 07:36:04.748766 4971 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-edpm-ipam-dockercfg-zb2j5" Dec 13 07:36:04 crc kubenswrapper[4971]: I1213 07:36:04.758722 4971 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-jdcq5"] Dec 13 07:36:04 crc kubenswrapper[4971]: I1213 07:36:04.858107 4971 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/27dafb6f-1f04-448c-97d2-0287c277c26a-ceph\") pod \"bootstrap-edpm-deployment-openstack-edpm-ipam-jdcq5\" (UID: \"27dafb6f-1f04-448c-97d2-0287c277c26a\") " pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-jdcq5" Dec 13 07:36:04 crc kubenswrapper[4971]: I1213 07:36:04.858175 4971 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/27dafb6f-1f04-448c-97d2-0287c277c26a-inventory\") pod \"bootstrap-edpm-deployment-openstack-edpm-ipam-jdcq5\" (UID: \"27dafb6f-1f04-448c-97d2-0287c277c26a\") " pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-jdcq5" Dec 13 07:36:04 crc kubenswrapper[4971]: I1213 07:36:04.858202 4971 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/27dafb6f-1f04-448c-97d2-0287c277c26a-ssh-key\") pod \"bootstrap-edpm-deployment-openstack-edpm-ipam-jdcq5\" (UID: \"27dafb6f-1f04-448c-97d2-0287c277c26a\") " pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-jdcq5" Dec 13 07:36:04 crc kubenswrapper[4971]: I1213 07:36:04.858236 4971 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-vj54l\" (UniqueName: \"kubernetes.io/projected/27dafb6f-1f04-448c-97d2-0287c277c26a-kube-api-access-vj54l\") pod \"bootstrap-edpm-deployment-openstack-edpm-ipam-jdcq5\" (UID: \"27dafb6f-1f04-448c-97d2-0287c277c26a\") " pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-jdcq5" Dec 13 07:36:04 crc kubenswrapper[4971]: I1213 07:36:04.858335 4971 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"bootstrap-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/27dafb6f-1f04-448c-97d2-0287c277c26a-bootstrap-combined-ca-bundle\") pod \"bootstrap-edpm-deployment-openstack-edpm-ipam-jdcq5\" (UID: \"27dafb6f-1f04-448c-97d2-0287c277c26a\") " pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-jdcq5" Dec 13 07:36:04 crc kubenswrapper[4971]: I1213 07:36:04.960405 4971 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/27dafb6f-1f04-448c-97d2-0287c277c26a-ssh-key\") pod \"bootstrap-edpm-deployment-openstack-edpm-ipam-jdcq5\" (UID: \"27dafb6f-1f04-448c-97d2-0287c277c26a\") " pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-jdcq5" Dec 13 07:36:04 crc kubenswrapper[4971]: I1213 07:36:04.960475 4971 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-vj54l\" (UniqueName: \"kubernetes.io/projected/27dafb6f-1f04-448c-97d2-0287c277c26a-kube-api-access-vj54l\") pod \"bootstrap-edpm-deployment-openstack-edpm-ipam-jdcq5\" (UID: \"27dafb6f-1f04-448c-97d2-0287c277c26a\") " pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-jdcq5" Dec 13 07:36:04 crc kubenswrapper[4971]: I1213 07:36:04.960577 4971 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"bootstrap-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/27dafb6f-1f04-448c-97d2-0287c277c26a-bootstrap-combined-ca-bundle\") pod \"bootstrap-edpm-deployment-openstack-edpm-ipam-jdcq5\" (UID: \"27dafb6f-1f04-448c-97d2-0287c277c26a\") " pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-jdcq5" Dec 13 07:36:04 crc kubenswrapper[4971]: I1213 07:36:04.960660 4971 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/27dafb6f-1f04-448c-97d2-0287c277c26a-ceph\") pod \"bootstrap-edpm-deployment-openstack-edpm-ipam-jdcq5\" (UID: \"27dafb6f-1f04-448c-97d2-0287c277c26a\") " pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-jdcq5" Dec 13 07:36:04 crc kubenswrapper[4971]: I1213 07:36:04.960691 4971 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/27dafb6f-1f04-448c-97d2-0287c277c26a-inventory\") pod \"bootstrap-edpm-deployment-openstack-edpm-ipam-jdcq5\" (UID: \"27dafb6f-1f04-448c-97d2-0287c277c26a\") " pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-jdcq5" Dec 13 07:36:04 crc kubenswrapper[4971]: I1213 07:36:04.965264 4971 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/27dafb6f-1f04-448c-97d2-0287c277c26a-ceph\") pod \"bootstrap-edpm-deployment-openstack-edpm-ipam-jdcq5\" (UID: \"27dafb6f-1f04-448c-97d2-0287c277c26a\") " pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-jdcq5" Dec 13 07:36:04 crc kubenswrapper[4971]: I1213 07:36:04.966756 4971 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/27dafb6f-1f04-448c-97d2-0287c277c26a-inventory\") pod \"bootstrap-edpm-deployment-openstack-edpm-ipam-jdcq5\" (UID: \"27dafb6f-1f04-448c-97d2-0287c277c26a\") " pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-jdcq5" Dec 13 07:36:04 crc kubenswrapper[4971]: I1213 07:36:04.968151 4971 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"bootstrap-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/27dafb6f-1f04-448c-97d2-0287c277c26a-bootstrap-combined-ca-bundle\") pod \"bootstrap-edpm-deployment-openstack-edpm-ipam-jdcq5\" (UID: \"27dafb6f-1f04-448c-97d2-0287c277c26a\") " pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-jdcq5" Dec 13 07:36:04 crc kubenswrapper[4971]: I1213 07:36:04.974070 4971 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/27dafb6f-1f04-448c-97d2-0287c277c26a-ssh-key\") pod \"bootstrap-edpm-deployment-openstack-edpm-ipam-jdcq5\" (UID: \"27dafb6f-1f04-448c-97d2-0287c277c26a\") " pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-jdcq5" Dec 13 07:36:04 crc kubenswrapper[4971]: I1213 07:36:04.996307 4971 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-vj54l\" (UniqueName: \"kubernetes.io/projected/27dafb6f-1f04-448c-97d2-0287c277c26a-kube-api-access-vj54l\") pod \"bootstrap-edpm-deployment-openstack-edpm-ipam-jdcq5\" (UID: \"27dafb6f-1f04-448c-97d2-0287c277c26a\") " pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-jdcq5" Dec 13 07:36:05 crc kubenswrapper[4971]: I1213 07:36:05.077244 4971 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-jdcq5" Dec 13 07:36:05 crc kubenswrapper[4971]: I1213 07:36:05.888293 4971 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-jdcq5"] Dec 13 07:36:06 crc kubenswrapper[4971]: I1213 07:36:06.654094 4971 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-jdcq5" event={"ID":"27dafb6f-1f04-448c-97d2-0287c277c26a","Type":"ContainerStarted","Data":"e4c766b216d62d39ce3d5d0c6123cc0fe4642a24fc10cac357f86656ca7d0fd0"} Dec 13 07:36:10 crc kubenswrapper[4971]: I1213 07:36:10.693816 4971 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-jdcq5" event={"ID":"27dafb6f-1f04-448c-97d2-0287c277c26a","Type":"ContainerStarted","Data":"7c07425b85175e6b917d6be6f1e8b599614963feb58702c98acee16dfb00d65d"} Dec 13 07:36:10 crc kubenswrapper[4971]: I1213 07:36:10.784616 4971 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-jdcq5" podStartSLOduration=2.988919352 podStartE2EDuration="6.784595912s" podCreationTimestamp="2025-12-13 07:36:04 +0000 UTC" firstStartedPulling="2025-12-13 07:36:05.887593106 +0000 UTC m=+2822.492002594" lastFinishedPulling="2025-12-13 07:36:09.683269696 +0000 UTC m=+2826.287679154" observedRunningTime="2025-12-13 07:36:10.778243109 +0000 UTC m=+2827.382652587" watchObservedRunningTime="2025-12-13 07:36:10.784595912 +0000 UTC m=+2827.389005370" Dec 13 07:36:18 crc kubenswrapper[4971]: I1213 07:36:18.879151 4971 scope.go:117] "RemoveContainer" containerID="54824cd9cbea8808582e7a1609afdfdf04d051d1509664c0e14aa36943354de6" Dec 13 07:36:18 crc kubenswrapper[4971]: I1213 07:36:18.912496 4971 scope.go:117] "RemoveContainer" containerID="0b61890aecd8dd55185aba04582cb72fabb37f4ccbe4d8a42f4ca0117703e827" Dec 13 07:36:18 crc kubenswrapper[4971]: I1213 07:36:18.986425 4971 scope.go:117] "RemoveContainer" containerID="8e528e869a31af117915d49484380c02654b1d3b2432c0d6edd12bc017ba2b67" Dec 13 07:36:19 crc kubenswrapper[4971]: I1213 07:36:19.020856 4971 scope.go:117] "RemoveContainer" containerID="b64b8d40637b7c78fb325e0b9ef34c5a9d5b3db7ad47fdd2f1fb42c0e8af00ad" Dec 13 07:36:19 crc kubenswrapper[4971]: I1213 07:36:19.061580 4971 scope.go:117] "RemoveContainer" containerID="456322a0d13d44bab8798554c48b5c8b6cc23cb2481bfbbc2a4d19fa60a84d7e" Dec 13 07:36:19 crc kubenswrapper[4971]: I1213 07:36:19.097988 4971 scope.go:117] "RemoveContainer" containerID="0ef36e730b8b14be062f8a9cd0480202fc65350a0964daedd0d8ca68547dd758" Dec 13 07:36:19 crc kubenswrapper[4971]: I1213 07:36:19.131055 4971 scope.go:117] "RemoveContainer" containerID="b1f374774ca878889bdc02f923abb290b5fdaf86d3cfbee33de205252d676d59" Dec 13 07:36:19 crc kubenswrapper[4971]: I1213 07:36:19.168766 4971 scope.go:117] "RemoveContainer" containerID="58181d74ba49de7a80cf6dac9d470068037e36d6ae703590820a64b82fd344a6" Dec 13 07:36:19 crc kubenswrapper[4971]: I1213 07:36:19.229348 4971 scope.go:117] "RemoveContainer" containerID="1b8738b60e78e46c6c3201e6dc3104d9218dd08393b34b5a75578292a38c23c0" Dec 13 07:36:19 crc kubenswrapper[4971]: I1213 07:36:19.279449 4971 scope.go:117] "RemoveContainer" containerID="66b156cf0ae30ae6d8916884aac86cb1ee337155f7ef722bd822b01428f81924" Dec 13 07:36:19 crc kubenswrapper[4971]: I1213 07:36:19.350507 4971 scope.go:117] "RemoveContainer" containerID="64e238fe45a615cf21cd269809f116b1a84e3b58edd6ed2187c07198bfd90fe6" Dec 13 07:36:19 crc kubenswrapper[4971]: I1213 07:36:19.381276 4971 scope.go:117] "RemoveContainer" containerID="caa6c4125172e3f7d61499c91b31dfad33bb6803bce76d35a19b16d325ddb492" Dec 13 07:36:19 crc kubenswrapper[4971]: I1213 07:36:19.430943 4971 scope.go:117] "RemoveContainer" containerID="9822e0c105b43e179352e287724d20e8458c235bdae5f99e81482c895bc83d21" Dec 13 07:36:19 crc kubenswrapper[4971]: I1213 07:36:19.458400 4971 scope.go:117] "RemoveContainer" containerID="c6717e7f9fd2a04c93e249f4661a0232169d358b869932582c249bc440616523" Dec 13 07:36:36 crc kubenswrapper[4971]: I1213 07:36:36.740240 4971 prober.go:107] "Probe failed" probeType="Liveness" pod="openstack/ceilometer-0" podUID="7cfb4dbf-cc55-48b7-b50c-65d6d4ae64ba" containerName="ceilometer-notification-agent" probeResult="failure" output="command timed out" Dec 13 07:37:16 crc kubenswrapper[4971]: I1213 07:37:16.153416 4971 patch_prober.go:28] interesting pod/machine-config-daemon-82xjz container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 13 07:37:16 crc kubenswrapper[4971]: I1213 07:37:16.154060 4971 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-82xjz" podUID="e0c4af12-edda-41d7-9b44-f87396174ef2" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 13 07:37:46 crc kubenswrapper[4971]: I1213 07:37:46.153922 4971 patch_prober.go:28] interesting pod/machine-config-daemon-82xjz container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 13 07:37:46 crc kubenswrapper[4971]: I1213 07:37:46.154660 4971 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-82xjz" podUID="e0c4af12-edda-41d7-9b44-f87396174ef2" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 13 07:37:51 crc kubenswrapper[4971]: I1213 07:37:51.796989 4971 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-marketplace-gczff"] Dec 13 07:37:51 crc kubenswrapper[4971]: I1213 07:37:51.802997 4971 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-gczff" Dec 13 07:37:51 crc kubenswrapper[4971]: I1213 07:37:51.806176 4971 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-gczff"] Dec 13 07:37:52 crc kubenswrapper[4971]: I1213 07:37:52.046135 4971 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-4zdbf\" (UniqueName: \"kubernetes.io/projected/b1cb4a29-db37-4323-9ebb-1c5e67d1668e-kube-api-access-4zdbf\") pod \"redhat-marketplace-gczff\" (UID: \"b1cb4a29-db37-4323-9ebb-1c5e67d1668e\") " pod="openshift-marketplace/redhat-marketplace-gczff" Dec 13 07:37:52 crc kubenswrapper[4971]: I1213 07:37:52.046471 4971 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/b1cb4a29-db37-4323-9ebb-1c5e67d1668e-catalog-content\") pod \"redhat-marketplace-gczff\" (UID: \"b1cb4a29-db37-4323-9ebb-1c5e67d1668e\") " pod="openshift-marketplace/redhat-marketplace-gczff" Dec 13 07:37:52 crc kubenswrapper[4971]: I1213 07:37:52.046542 4971 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/b1cb4a29-db37-4323-9ebb-1c5e67d1668e-utilities\") pod \"redhat-marketplace-gczff\" (UID: \"b1cb4a29-db37-4323-9ebb-1c5e67d1668e\") " pod="openshift-marketplace/redhat-marketplace-gczff" Dec 13 07:37:52 crc kubenswrapper[4971]: I1213 07:37:52.081880 4971 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-operators-7cbwp"] Dec 13 07:37:52 crc kubenswrapper[4971]: I1213 07:37:52.084219 4971 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-7cbwp" Dec 13 07:37:52 crc kubenswrapper[4971]: I1213 07:37:52.093323 4971 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-7cbwp"] Dec 13 07:37:52 crc kubenswrapper[4971]: I1213 07:37:52.148545 4971 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/b1cb4a29-db37-4323-9ebb-1c5e67d1668e-utilities\") pod \"redhat-marketplace-gczff\" (UID: \"b1cb4a29-db37-4323-9ebb-1c5e67d1668e\") " pod="openshift-marketplace/redhat-marketplace-gczff" Dec 13 07:37:52 crc kubenswrapper[4971]: I1213 07:37:52.148619 4971 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-h7kdw\" (UniqueName: \"kubernetes.io/projected/a35ae77d-f799-4da5-9f79-c6e746b947b8-kube-api-access-h7kdw\") pod \"redhat-operators-7cbwp\" (UID: \"a35ae77d-f799-4da5-9f79-c6e746b947b8\") " pod="openshift-marketplace/redhat-operators-7cbwp" Dec 13 07:37:52 crc kubenswrapper[4971]: I1213 07:37:52.148747 4971 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/a35ae77d-f799-4da5-9f79-c6e746b947b8-catalog-content\") pod \"redhat-operators-7cbwp\" (UID: \"a35ae77d-f799-4da5-9f79-c6e746b947b8\") " pod="openshift-marketplace/redhat-operators-7cbwp" Dec 13 07:37:52 crc kubenswrapper[4971]: I1213 07:37:52.148801 4971 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-4zdbf\" (UniqueName: \"kubernetes.io/projected/b1cb4a29-db37-4323-9ebb-1c5e67d1668e-kube-api-access-4zdbf\") pod \"redhat-marketplace-gczff\" (UID: \"b1cb4a29-db37-4323-9ebb-1c5e67d1668e\") " pod="openshift-marketplace/redhat-marketplace-gczff" Dec 13 07:37:52 crc kubenswrapper[4971]: I1213 07:37:52.148980 4971 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/b1cb4a29-db37-4323-9ebb-1c5e67d1668e-catalog-content\") pod \"redhat-marketplace-gczff\" (UID: \"b1cb4a29-db37-4323-9ebb-1c5e67d1668e\") " pod="openshift-marketplace/redhat-marketplace-gczff" Dec 13 07:37:52 crc kubenswrapper[4971]: I1213 07:37:52.149149 4971 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/b1cb4a29-db37-4323-9ebb-1c5e67d1668e-utilities\") pod \"redhat-marketplace-gczff\" (UID: \"b1cb4a29-db37-4323-9ebb-1c5e67d1668e\") " pod="openshift-marketplace/redhat-marketplace-gczff" Dec 13 07:37:52 crc kubenswrapper[4971]: I1213 07:37:52.149172 4971 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/a35ae77d-f799-4da5-9f79-c6e746b947b8-utilities\") pod \"redhat-operators-7cbwp\" (UID: \"a35ae77d-f799-4da5-9f79-c6e746b947b8\") " pod="openshift-marketplace/redhat-operators-7cbwp" Dec 13 07:37:52 crc kubenswrapper[4971]: I1213 07:37:52.149355 4971 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/b1cb4a29-db37-4323-9ebb-1c5e67d1668e-catalog-content\") pod \"redhat-marketplace-gczff\" (UID: \"b1cb4a29-db37-4323-9ebb-1c5e67d1668e\") " pod="openshift-marketplace/redhat-marketplace-gczff" Dec 13 07:37:52 crc kubenswrapper[4971]: I1213 07:37:52.169148 4971 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-4zdbf\" (UniqueName: \"kubernetes.io/projected/b1cb4a29-db37-4323-9ebb-1c5e67d1668e-kube-api-access-4zdbf\") pod \"redhat-marketplace-gczff\" (UID: \"b1cb4a29-db37-4323-9ebb-1c5e67d1668e\") " pod="openshift-marketplace/redhat-marketplace-gczff" Dec 13 07:37:52 crc kubenswrapper[4971]: I1213 07:37:52.250606 4971 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/a35ae77d-f799-4da5-9f79-c6e746b947b8-utilities\") pod \"redhat-operators-7cbwp\" (UID: \"a35ae77d-f799-4da5-9f79-c6e746b947b8\") " pod="openshift-marketplace/redhat-operators-7cbwp" Dec 13 07:37:52 crc kubenswrapper[4971]: I1213 07:37:52.250777 4971 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-h7kdw\" (UniqueName: \"kubernetes.io/projected/a35ae77d-f799-4da5-9f79-c6e746b947b8-kube-api-access-h7kdw\") pod \"redhat-operators-7cbwp\" (UID: \"a35ae77d-f799-4da5-9f79-c6e746b947b8\") " pod="openshift-marketplace/redhat-operators-7cbwp" Dec 13 07:37:52 crc kubenswrapper[4971]: I1213 07:37:52.250904 4971 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/a35ae77d-f799-4da5-9f79-c6e746b947b8-catalog-content\") pod \"redhat-operators-7cbwp\" (UID: \"a35ae77d-f799-4da5-9f79-c6e746b947b8\") " pod="openshift-marketplace/redhat-operators-7cbwp" Dec 13 07:37:52 crc kubenswrapper[4971]: I1213 07:37:52.251237 4971 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/a35ae77d-f799-4da5-9f79-c6e746b947b8-utilities\") pod \"redhat-operators-7cbwp\" (UID: \"a35ae77d-f799-4da5-9f79-c6e746b947b8\") " pod="openshift-marketplace/redhat-operators-7cbwp" Dec 13 07:37:52 crc kubenswrapper[4971]: I1213 07:37:52.251541 4971 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/a35ae77d-f799-4da5-9f79-c6e746b947b8-catalog-content\") pod \"redhat-operators-7cbwp\" (UID: \"a35ae77d-f799-4da5-9f79-c6e746b947b8\") " pod="openshift-marketplace/redhat-operators-7cbwp" Dec 13 07:37:52 crc kubenswrapper[4971]: I1213 07:37:52.271960 4971 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-h7kdw\" (UniqueName: \"kubernetes.io/projected/a35ae77d-f799-4da5-9f79-c6e746b947b8-kube-api-access-h7kdw\") pod \"redhat-operators-7cbwp\" (UID: \"a35ae77d-f799-4da5-9f79-c6e746b947b8\") " pod="openshift-marketplace/redhat-operators-7cbwp" Dec 13 07:37:52 crc kubenswrapper[4971]: I1213 07:37:52.410253 4971 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-7cbwp" Dec 13 07:37:52 crc kubenswrapper[4971]: I1213 07:37:52.434756 4971 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-gczff" Dec 13 07:37:52 crc kubenswrapper[4971]: I1213 07:37:52.981542 4971 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-7cbwp"] Dec 13 07:37:53 crc kubenswrapper[4971]: I1213 07:37:53.026405 4971 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-gczff"] Dec 13 07:37:53 crc kubenswrapper[4971]: W1213 07:37:53.041124 4971 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podb1cb4a29_db37_4323_9ebb_1c5e67d1668e.slice/crio-6fc43ca17b8ed1c2b21d8076d3fe281c3e7b9260485e430eed3591c6880b2a55 WatchSource:0}: Error finding container 6fc43ca17b8ed1c2b21d8076d3fe281c3e7b9260485e430eed3591c6880b2a55: Status 404 returned error can't find the container with id 6fc43ca17b8ed1c2b21d8076d3fe281c3e7b9260485e430eed3591c6880b2a55 Dec 13 07:37:53 crc kubenswrapper[4971]: I1213 07:37:53.895791 4971 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-7cbwp" event={"ID":"a35ae77d-f799-4da5-9f79-c6e746b947b8","Type":"ContainerStarted","Data":"49644044b57c792b8cd49dbd312aee4fb37d64a5670cc2a45873b6d750a74c64"} Dec 13 07:37:53 crc kubenswrapper[4971]: I1213 07:37:53.897751 4971 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-gczff" event={"ID":"b1cb4a29-db37-4323-9ebb-1c5e67d1668e","Type":"ContainerStarted","Data":"6fc43ca17b8ed1c2b21d8076d3fe281c3e7b9260485e430eed3591c6880b2a55"} Dec 13 07:37:54 crc kubenswrapper[4971]: I1213 07:37:54.259192 4971 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/community-operators-kzw5r"] Dec 13 07:37:54 crc kubenswrapper[4971]: I1213 07:37:54.261351 4971 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-kzw5r" Dec 13 07:37:54 crc kubenswrapper[4971]: I1213 07:37:54.286549 4971 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-kzw5r"] Dec 13 07:37:54 crc kubenswrapper[4971]: I1213 07:37:54.350995 4971 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-vl5lt\" (UniqueName: \"kubernetes.io/projected/b2475f4e-bfe7-4678-8779-2b56c189b3b0-kube-api-access-vl5lt\") pod \"community-operators-kzw5r\" (UID: \"b2475f4e-bfe7-4678-8779-2b56c189b3b0\") " pod="openshift-marketplace/community-operators-kzw5r" Dec 13 07:37:54 crc kubenswrapper[4971]: I1213 07:37:54.351067 4971 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/b2475f4e-bfe7-4678-8779-2b56c189b3b0-utilities\") pod \"community-operators-kzw5r\" (UID: \"b2475f4e-bfe7-4678-8779-2b56c189b3b0\") " pod="openshift-marketplace/community-operators-kzw5r" Dec 13 07:37:54 crc kubenswrapper[4971]: I1213 07:37:54.351137 4971 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/b2475f4e-bfe7-4678-8779-2b56c189b3b0-catalog-content\") pod \"community-operators-kzw5r\" (UID: \"b2475f4e-bfe7-4678-8779-2b56c189b3b0\") " pod="openshift-marketplace/community-operators-kzw5r" Dec 13 07:37:54 crc kubenswrapper[4971]: I1213 07:37:54.452574 4971 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-vl5lt\" (UniqueName: \"kubernetes.io/projected/b2475f4e-bfe7-4678-8779-2b56c189b3b0-kube-api-access-vl5lt\") pod \"community-operators-kzw5r\" (UID: \"b2475f4e-bfe7-4678-8779-2b56c189b3b0\") " pod="openshift-marketplace/community-operators-kzw5r" Dec 13 07:37:54 crc kubenswrapper[4971]: I1213 07:37:54.452684 4971 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/b2475f4e-bfe7-4678-8779-2b56c189b3b0-utilities\") pod \"community-operators-kzw5r\" (UID: \"b2475f4e-bfe7-4678-8779-2b56c189b3b0\") " pod="openshift-marketplace/community-operators-kzw5r" Dec 13 07:37:54 crc kubenswrapper[4971]: I1213 07:37:54.453326 4971 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/b2475f4e-bfe7-4678-8779-2b56c189b3b0-utilities\") pod \"community-operators-kzw5r\" (UID: \"b2475f4e-bfe7-4678-8779-2b56c189b3b0\") " pod="openshift-marketplace/community-operators-kzw5r" Dec 13 07:37:54 crc kubenswrapper[4971]: I1213 07:37:54.453477 4971 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/b2475f4e-bfe7-4678-8779-2b56c189b3b0-catalog-content\") pod \"community-operators-kzw5r\" (UID: \"b2475f4e-bfe7-4678-8779-2b56c189b3b0\") " pod="openshift-marketplace/community-operators-kzw5r" Dec 13 07:37:54 crc kubenswrapper[4971]: I1213 07:37:54.453889 4971 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/b2475f4e-bfe7-4678-8779-2b56c189b3b0-catalog-content\") pod \"community-operators-kzw5r\" (UID: \"b2475f4e-bfe7-4678-8779-2b56c189b3b0\") " pod="openshift-marketplace/community-operators-kzw5r" Dec 13 07:37:54 crc kubenswrapper[4971]: I1213 07:37:54.489636 4971 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-vl5lt\" (UniqueName: \"kubernetes.io/projected/b2475f4e-bfe7-4678-8779-2b56c189b3b0-kube-api-access-vl5lt\") pod \"community-operators-kzw5r\" (UID: \"b2475f4e-bfe7-4678-8779-2b56c189b3b0\") " pod="openshift-marketplace/community-operators-kzw5r" Dec 13 07:37:54 crc kubenswrapper[4971]: I1213 07:37:54.587054 4971 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-kzw5r" Dec 13 07:37:54 crc kubenswrapper[4971]: I1213 07:37:54.927436 4971 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-7cbwp" event={"ID":"a35ae77d-f799-4da5-9f79-c6e746b947b8","Type":"ContainerStarted","Data":"47b2b2b58223551c494b2ffbe0403cb6c1a27b9ac87464342f7cd8393866e53e"} Dec 13 07:37:54 crc kubenswrapper[4971]: I1213 07:37:54.933957 4971 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-gczff" event={"ID":"b1cb4a29-db37-4323-9ebb-1c5e67d1668e","Type":"ContainerStarted","Data":"2271396c7c130beac1da746bac38527b915b89bea9fb90c02c1e36fea202a98b"} Dec 13 07:37:55 crc kubenswrapper[4971]: I1213 07:37:55.223645 4971 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-kzw5r"] Dec 13 07:37:55 crc kubenswrapper[4971]: W1213 07:37:55.281018 4971 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podb2475f4e_bfe7_4678_8779_2b56c189b3b0.slice/crio-f9fe3ec558c101650e816df1e11c455cb6314aae4b0853a285decf65b1bef543 WatchSource:0}: Error finding container f9fe3ec558c101650e816df1e11c455cb6314aae4b0853a285decf65b1bef543: Status 404 returned error can't find the container with id f9fe3ec558c101650e816df1e11c455cb6314aae4b0853a285decf65b1bef543 Dec 13 07:37:55 crc kubenswrapper[4971]: I1213 07:37:55.945336 4971 generic.go:334] "Generic (PLEG): container finished" podID="b2475f4e-bfe7-4678-8779-2b56c189b3b0" containerID="2ad39db5591c90628f77debcb15b5c9919f98d2bc5fde6f43ce1b35ff420379d" exitCode=0 Dec 13 07:37:55 crc kubenswrapper[4971]: I1213 07:37:55.945391 4971 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-kzw5r" event={"ID":"b2475f4e-bfe7-4678-8779-2b56c189b3b0","Type":"ContainerDied","Data":"2ad39db5591c90628f77debcb15b5c9919f98d2bc5fde6f43ce1b35ff420379d"} Dec 13 07:37:55 crc kubenswrapper[4971]: I1213 07:37:55.945740 4971 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-kzw5r" event={"ID":"b2475f4e-bfe7-4678-8779-2b56c189b3b0","Type":"ContainerStarted","Data":"f9fe3ec558c101650e816df1e11c455cb6314aae4b0853a285decf65b1bef543"} Dec 13 07:37:55 crc kubenswrapper[4971]: I1213 07:37:55.950024 4971 generic.go:334] "Generic (PLEG): container finished" podID="a35ae77d-f799-4da5-9f79-c6e746b947b8" containerID="47b2b2b58223551c494b2ffbe0403cb6c1a27b9ac87464342f7cd8393866e53e" exitCode=0 Dec 13 07:37:55 crc kubenswrapper[4971]: I1213 07:37:55.950108 4971 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-7cbwp" event={"ID":"a35ae77d-f799-4da5-9f79-c6e746b947b8","Type":"ContainerDied","Data":"47b2b2b58223551c494b2ffbe0403cb6c1a27b9ac87464342f7cd8393866e53e"} Dec 13 07:37:55 crc kubenswrapper[4971]: I1213 07:37:55.955939 4971 generic.go:334] "Generic (PLEG): container finished" podID="b1cb4a29-db37-4323-9ebb-1c5e67d1668e" containerID="2271396c7c130beac1da746bac38527b915b89bea9fb90c02c1e36fea202a98b" exitCode=0 Dec 13 07:37:55 crc kubenswrapper[4971]: I1213 07:37:55.955980 4971 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-gczff" event={"ID":"b1cb4a29-db37-4323-9ebb-1c5e67d1668e","Type":"ContainerDied","Data":"2271396c7c130beac1da746bac38527b915b89bea9fb90c02c1e36fea202a98b"} Dec 13 07:38:10 crc kubenswrapper[4971]: I1213 07:38:10.218650 4971 generic.go:334] "Generic (PLEG): container finished" podID="27dafb6f-1f04-448c-97d2-0287c277c26a" containerID="7c07425b85175e6b917d6be6f1e8b599614963feb58702c98acee16dfb00d65d" exitCode=0 Dec 13 07:38:10 crc kubenswrapper[4971]: I1213 07:38:10.218781 4971 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-jdcq5" event={"ID":"27dafb6f-1f04-448c-97d2-0287c277c26a","Type":"ContainerDied","Data":"7c07425b85175e6b917d6be6f1e8b599614963feb58702c98acee16dfb00d65d"} Dec 13 07:38:11 crc kubenswrapper[4971]: I1213 07:38:11.744026 4971 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-jdcq5" Dec 13 07:38:11 crc kubenswrapper[4971]: I1213 07:38:11.846184 4971 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/27dafb6f-1f04-448c-97d2-0287c277c26a-ssh-key\") pod \"27dafb6f-1f04-448c-97d2-0287c277c26a\" (UID: \"27dafb6f-1f04-448c-97d2-0287c277c26a\") " Dec 13 07:38:11 crc kubenswrapper[4971]: I1213 07:38:11.846345 4971 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"bootstrap-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/27dafb6f-1f04-448c-97d2-0287c277c26a-bootstrap-combined-ca-bundle\") pod \"27dafb6f-1f04-448c-97d2-0287c277c26a\" (UID: \"27dafb6f-1f04-448c-97d2-0287c277c26a\") " Dec 13 07:38:11 crc kubenswrapper[4971]: I1213 07:38:11.846409 4971 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-vj54l\" (UniqueName: \"kubernetes.io/projected/27dafb6f-1f04-448c-97d2-0287c277c26a-kube-api-access-vj54l\") pod \"27dafb6f-1f04-448c-97d2-0287c277c26a\" (UID: \"27dafb6f-1f04-448c-97d2-0287c277c26a\") " Dec 13 07:38:11 crc kubenswrapper[4971]: I1213 07:38:11.846540 4971 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/27dafb6f-1f04-448c-97d2-0287c277c26a-inventory\") pod \"27dafb6f-1f04-448c-97d2-0287c277c26a\" (UID: \"27dafb6f-1f04-448c-97d2-0287c277c26a\") " Dec 13 07:38:11 crc kubenswrapper[4971]: I1213 07:38:11.846573 4971 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/27dafb6f-1f04-448c-97d2-0287c277c26a-ceph\") pod \"27dafb6f-1f04-448c-97d2-0287c277c26a\" (UID: \"27dafb6f-1f04-448c-97d2-0287c277c26a\") " Dec 13 07:38:11 crc kubenswrapper[4971]: I1213 07:38:11.853705 4971 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/27dafb6f-1f04-448c-97d2-0287c277c26a-bootstrap-combined-ca-bundle" (OuterVolumeSpecName: "bootstrap-combined-ca-bundle") pod "27dafb6f-1f04-448c-97d2-0287c277c26a" (UID: "27dafb6f-1f04-448c-97d2-0287c277c26a"). InnerVolumeSpecName "bootstrap-combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 13 07:38:11 crc kubenswrapper[4971]: I1213 07:38:11.863814 4971 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/27dafb6f-1f04-448c-97d2-0287c277c26a-ceph" (OuterVolumeSpecName: "ceph") pod "27dafb6f-1f04-448c-97d2-0287c277c26a" (UID: "27dafb6f-1f04-448c-97d2-0287c277c26a"). InnerVolumeSpecName "ceph". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 13 07:38:11 crc kubenswrapper[4971]: I1213 07:38:11.863823 4971 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/27dafb6f-1f04-448c-97d2-0287c277c26a-kube-api-access-vj54l" (OuterVolumeSpecName: "kube-api-access-vj54l") pod "27dafb6f-1f04-448c-97d2-0287c277c26a" (UID: "27dafb6f-1f04-448c-97d2-0287c277c26a"). InnerVolumeSpecName "kube-api-access-vj54l". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 13 07:38:11 crc kubenswrapper[4971]: I1213 07:38:11.874565 4971 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/27dafb6f-1f04-448c-97d2-0287c277c26a-ssh-key" (OuterVolumeSpecName: "ssh-key") pod "27dafb6f-1f04-448c-97d2-0287c277c26a" (UID: "27dafb6f-1f04-448c-97d2-0287c277c26a"). InnerVolumeSpecName "ssh-key". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 13 07:38:11 crc kubenswrapper[4971]: I1213 07:38:11.890036 4971 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/27dafb6f-1f04-448c-97d2-0287c277c26a-inventory" (OuterVolumeSpecName: "inventory") pod "27dafb6f-1f04-448c-97d2-0287c277c26a" (UID: "27dafb6f-1f04-448c-97d2-0287c277c26a"). InnerVolumeSpecName "inventory". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 13 07:38:11 crc kubenswrapper[4971]: I1213 07:38:11.948208 4971 reconciler_common.go:293] "Volume detached for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/27dafb6f-1f04-448c-97d2-0287c277c26a-ssh-key\") on node \"crc\" DevicePath \"\"" Dec 13 07:38:11 crc kubenswrapper[4971]: I1213 07:38:11.948377 4971 reconciler_common.go:293] "Volume detached for volume \"bootstrap-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/27dafb6f-1f04-448c-97d2-0287c277c26a-bootstrap-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 13 07:38:11 crc kubenswrapper[4971]: I1213 07:38:11.948444 4971 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-vj54l\" (UniqueName: \"kubernetes.io/projected/27dafb6f-1f04-448c-97d2-0287c277c26a-kube-api-access-vj54l\") on node \"crc\" DevicePath \"\"" Dec 13 07:38:11 crc kubenswrapper[4971]: I1213 07:38:11.948498 4971 reconciler_common.go:293] "Volume detached for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/27dafb6f-1f04-448c-97d2-0287c277c26a-inventory\") on node \"crc\" DevicePath \"\"" Dec 13 07:38:11 crc kubenswrapper[4971]: I1213 07:38:11.948575 4971 reconciler_common.go:293] "Volume detached for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/27dafb6f-1f04-448c-97d2-0287c277c26a-ceph\") on node \"crc\" DevicePath \"\"" Dec 13 07:38:12 crc kubenswrapper[4971]: I1213 07:38:12.245366 4971 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-jdcq5" event={"ID":"27dafb6f-1f04-448c-97d2-0287c277c26a","Type":"ContainerDied","Data":"e4c766b216d62d39ce3d5d0c6123cc0fe4642a24fc10cac357f86656ca7d0fd0"} Dec 13 07:38:12 crc kubenswrapper[4971]: I1213 07:38:12.245475 4971 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="e4c766b216d62d39ce3d5d0c6123cc0fe4642a24fc10cac357f86656ca7d0fd0" Dec 13 07:38:12 crc kubenswrapper[4971]: I1213 07:38:12.245841 4971 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-jdcq5" Dec 13 07:38:12 crc kubenswrapper[4971]: I1213 07:38:12.401371 4971 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/configure-network-edpm-deployment-openstack-edpm-ipam-95c87"] Dec 13 07:38:12 crc kubenswrapper[4971]: E1213 07:38:12.401913 4971 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="27dafb6f-1f04-448c-97d2-0287c277c26a" containerName="bootstrap-edpm-deployment-openstack-edpm-ipam" Dec 13 07:38:12 crc kubenswrapper[4971]: I1213 07:38:12.401944 4971 state_mem.go:107] "Deleted CPUSet assignment" podUID="27dafb6f-1f04-448c-97d2-0287c277c26a" containerName="bootstrap-edpm-deployment-openstack-edpm-ipam" Dec 13 07:38:12 crc kubenswrapper[4971]: I1213 07:38:12.402213 4971 memory_manager.go:354] "RemoveStaleState removing state" podUID="27dafb6f-1f04-448c-97d2-0287c277c26a" containerName="bootstrap-edpm-deployment-openstack-edpm-ipam" Dec 13 07:38:12 crc kubenswrapper[4971]: I1213 07:38:12.403036 4971 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/configure-network-edpm-deployment-openstack-edpm-ipam-95c87" Dec 13 07:38:12 crc kubenswrapper[4971]: I1213 07:38:12.410829 4971 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplanenodeset-openstack-edpm-ipam" Dec 13 07:38:12 crc kubenswrapper[4971]: I1213 07:38:12.411036 4971 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-edpm-ipam-dockercfg-zb2j5" Dec 13 07:38:12 crc kubenswrapper[4971]: I1213 07:38:12.411236 4971 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceph-conf-files" Dec 13 07:38:12 crc kubenswrapper[4971]: I1213 07:38:12.411373 4971 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-aee-default-env" Dec 13 07:38:12 crc kubenswrapper[4971]: I1213 07:38:12.411602 4971 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplane-ansible-ssh-private-key-secret" Dec 13 07:38:12 crc kubenswrapper[4971]: I1213 07:38:12.413078 4971 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/configure-network-edpm-deployment-openstack-edpm-ipam-95c87"] Dec 13 07:38:12 crc kubenswrapper[4971]: I1213 07:38:12.562664 4971 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-rr4ff\" (UniqueName: \"kubernetes.io/projected/ef08e99c-0bd8-4f27-ae28-2cfdeb4ec045-kube-api-access-rr4ff\") pod \"configure-network-edpm-deployment-openstack-edpm-ipam-95c87\" (UID: \"ef08e99c-0bd8-4f27-ae28-2cfdeb4ec045\") " pod="openstack/configure-network-edpm-deployment-openstack-edpm-ipam-95c87" Dec 13 07:38:12 crc kubenswrapper[4971]: I1213 07:38:12.562750 4971 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/ef08e99c-0bd8-4f27-ae28-2cfdeb4ec045-inventory\") pod \"configure-network-edpm-deployment-openstack-edpm-ipam-95c87\" (UID: \"ef08e99c-0bd8-4f27-ae28-2cfdeb4ec045\") " pod="openstack/configure-network-edpm-deployment-openstack-edpm-ipam-95c87" Dec 13 07:38:12 crc kubenswrapper[4971]: I1213 07:38:12.563118 4971 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/ef08e99c-0bd8-4f27-ae28-2cfdeb4ec045-ceph\") pod \"configure-network-edpm-deployment-openstack-edpm-ipam-95c87\" (UID: \"ef08e99c-0bd8-4f27-ae28-2cfdeb4ec045\") " pod="openstack/configure-network-edpm-deployment-openstack-edpm-ipam-95c87" Dec 13 07:38:12 crc kubenswrapper[4971]: I1213 07:38:12.563203 4971 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/ef08e99c-0bd8-4f27-ae28-2cfdeb4ec045-ssh-key\") pod \"configure-network-edpm-deployment-openstack-edpm-ipam-95c87\" (UID: \"ef08e99c-0bd8-4f27-ae28-2cfdeb4ec045\") " pod="openstack/configure-network-edpm-deployment-openstack-edpm-ipam-95c87" Dec 13 07:38:12 crc kubenswrapper[4971]: I1213 07:38:12.665416 4971 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/ef08e99c-0bd8-4f27-ae28-2cfdeb4ec045-ceph\") pod \"configure-network-edpm-deployment-openstack-edpm-ipam-95c87\" (UID: \"ef08e99c-0bd8-4f27-ae28-2cfdeb4ec045\") " pod="openstack/configure-network-edpm-deployment-openstack-edpm-ipam-95c87" Dec 13 07:38:12 crc kubenswrapper[4971]: I1213 07:38:12.665561 4971 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/ef08e99c-0bd8-4f27-ae28-2cfdeb4ec045-ssh-key\") pod \"configure-network-edpm-deployment-openstack-edpm-ipam-95c87\" (UID: \"ef08e99c-0bd8-4f27-ae28-2cfdeb4ec045\") " pod="openstack/configure-network-edpm-deployment-openstack-edpm-ipam-95c87" Dec 13 07:38:12 crc kubenswrapper[4971]: I1213 07:38:12.665750 4971 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-rr4ff\" (UniqueName: \"kubernetes.io/projected/ef08e99c-0bd8-4f27-ae28-2cfdeb4ec045-kube-api-access-rr4ff\") pod \"configure-network-edpm-deployment-openstack-edpm-ipam-95c87\" (UID: \"ef08e99c-0bd8-4f27-ae28-2cfdeb4ec045\") " pod="openstack/configure-network-edpm-deployment-openstack-edpm-ipam-95c87" Dec 13 07:38:12 crc kubenswrapper[4971]: I1213 07:38:12.665791 4971 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/ef08e99c-0bd8-4f27-ae28-2cfdeb4ec045-inventory\") pod \"configure-network-edpm-deployment-openstack-edpm-ipam-95c87\" (UID: \"ef08e99c-0bd8-4f27-ae28-2cfdeb4ec045\") " pod="openstack/configure-network-edpm-deployment-openstack-edpm-ipam-95c87" Dec 13 07:38:12 crc kubenswrapper[4971]: I1213 07:38:12.671118 4971 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/ef08e99c-0bd8-4f27-ae28-2cfdeb4ec045-ssh-key\") pod \"configure-network-edpm-deployment-openstack-edpm-ipam-95c87\" (UID: \"ef08e99c-0bd8-4f27-ae28-2cfdeb4ec045\") " pod="openstack/configure-network-edpm-deployment-openstack-edpm-ipam-95c87" Dec 13 07:38:12 crc kubenswrapper[4971]: I1213 07:38:12.671786 4971 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/ef08e99c-0bd8-4f27-ae28-2cfdeb4ec045-ceph\") pod \"configure-network-edpm-deployment-openstack-edpm-ipam-95c87\" (UID: \"ef08e99c-0bd8-4f27-ae28-2cfdeb4ec045\") " pod="openstack/configure-network-edpm-deployment-openstack-edpm-ipam-95c87" Dec 13 07:38:12 crc kubenswrapper[4971]: I1213 07:38:12.672934 4971 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/ef08e99c-0bd8-4f27-ae28-2cfdeb4ec045-inventory\") pod \"configure-network-edpm-deployment-openstack-edpm-ipam-95c87\" (UID: \"ef08e99c-0bd8-4f27-ae28-2cfdeb4ec045\") " pod="openstack/configure-network-edpm-deployment-openstack-edpm-ipam-95c87" Dec 13 07:38:12 crc kubenswrapper[4971]: I1213 07:38:12.701459 4971 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-rr4ff\" (UniqueName: \"kubernetes.io/projected/ef08e99c-0bd8-4f27-ae28-2cfdeb4ec045-kube-api-access-rr4ff\") pod \"configure-network-edpm-deployment-openstack-edpm-ipam-95c87\" (UID: \"ef08e99c-0bd8-4f27-ae28-2cfdeb4ec045\") " pod="openstack/configure-network-edpm-deployment-openstack-edpm-ipam-95c87" Dec 13 07:38:12 crc kubenswrapper[4971]: I1213 07:38:12.736040 4971 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/configure-network-edpm-deployment-openstack-edpm-ipam-95c87" Dec 13 07:38:16 crc kubenswrapper[4971]: I1213 07:38:16.154210 4971 patch_prober.go:28] interesting pod/machine-config-daemon-82xjz container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 13 07:38:16 crc kubenswrapper[4971]: I1213 07:38:16.154828 4971 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-82xjz" podUID="e0c4af12-edda-41d7-9b44-f87396174ef2" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 13 07:38:16 crc kubenswrapper[4971]: I1213 07:38:16.154873 4971 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-82xjz" Dec 13 07:38:16 crc kubenswrapper[4971]: I1213 07:38:16.155659 4971 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"09d0904c76ce78aa0b074447b1e64cf63160f573da086f934c8cb712615b71d4"} pod="openshift-machine-config-operator/machine-config-daemon-82xjz" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Dec 13 07:38:16 crc kubenswrapper[4971]: I1213 07:38:16.155708 4971 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-82xjz" podUID="e0c4af12-edda-41d7-9b44-f87396174ef2" containerName="machine-config-daemon" containerID="cri-o://09d0904c76ce78aa0b074447b1e64cf63160f573da086f934c8cb712615b71d4" gracePeriod=600 Dec 13 07:38:16 crc kubenswrapper[4971]: I1213 07:38:16.545991 4971 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/configure-network-edpm-deployment-openstack-edpm-ipam-95c87"] Dec 13 07:38:17 crc kubenswrapper[4971]: I1213 07:38:17.296740 4971 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-7cbwp" event={"ID":"a35ae77d-f799-4da5-9f79-c6e746b947b8","Type":"ContainerStarted","Data":"13041742c489627311b990c47b0b1fb64e78198e38710e78517b779991176e80"} Dec 13 07:38:17 crc kubenswrapper[4971]: I1213 07:38:17.299603 4971 generic.go:334] "Generic (PLEG): container finished" podID="b1cb4a29-db37-4323-9ebb-1c5e67d1668e" containerID="bc1edf3b6a1cfd239bc684daa88f85a6503e61d05084a620396ef36b1187130a" exitCode=0 Dec 13 07:38:17 crc kubenswrapper[4971]: I1213 07:38:17.299702 4971 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-gczff" event={"ID":"b1cb4a29-db37-4323-9ebb-1c5e67d1668e","Type":"ContainerDied","Data":"bc1edf3b6a1cfd239bc684daa88f85a6503e61d05084a620396ef36b1187130a"} Dec 13 07:38:17 crc kubenswrapper[4971]: I1213 07:38:17.301431 4971 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/configure-network-edpm-deployment-openstack-edpm-ipam-95c87" event={"ID":"ef08e99c-0bd8-4f27-ae28-2cfdeb4ec045","Type":"ContainerStarted","Data":"e15bc3336be6c26a6455e4ca601bcfafe6f0ff74d5f73888e456d091b81c726c"} Dec 13 07:38:17 crc kubenswrapper[4971]: I1213 07:38:17.305150 4971 generic.go:334] "Generic (PLEG): container finished" podID="e0c4af12-edda-41d7-9b44-f87396174ef2" containerID="09d0904c76ce78aa0b074447b1e64cf63160f573da086f934c8cb712615b71d4" exitCode=0 Dec 13 07:38:17 crc kubenswrapper[4971]: I1213 07:38:17.305186 4971 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-82xjz" event={"ID":"e0c4af12-edda-41d7-9b44-f87396174ef2","Type":"ContainerDied","Data":"09d0904c76ce78aa0b074447b1e64cf63160f573da086f934c8cb712615b71d4"} Dec 13 07:38:17 crc kubenswrapper[4971]: I1213 07:38:17.305214 4971 scope.go:117] "RemoveContainer" containerID="22e9f7096ef3c500a655cea9dbd5645fc5a0ca55768d6f95bfcaf07af351d70a" Dec 13 07:38:20 crc kubenswrapper[4971]: I1213 07:38:20.348349 4971 generic.go:334] "Generic (PLEG): container finished" podID="a35ae77d-f799-4da5-9f79-c6e746b947b8" containerID="13041742c489627311b990c47b0b1fb64e78198e38710e78517b779991176e80" exitCode=0 Dec 13 07:38:20 crc kubenswrapper[4971]: I1213 07:38:20.349894 4971 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-7cbwp" event={"ID":"a35ae77d-f799-4da5-9f79-c6e746b947b8","Type":"ContainerDied","Data":"13041742c489627311b990c47b0b1fb64e78198e38710e78517b779991176e80"} Dec 13 07:38:20 crc kubenswrapper[4971]: E1213 07:38:20.845126 4971 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-82xjz_openshift-machine-config-operator(e0c4af12-edda-41d7-9b44-f87396174ef2)\"" pod="openshift-machine-config-operator/machine-config-daemon-82xjz" podUID="e0c4af12-edda-41d7-9b44-f87396174ef2" Dec 13 07:38:21 crc kubenswrapper[4971]: I1213 07:38:21.355745 4971 scope.go:117] "RemoveContainer" containerID="09d0904c76ce78aa0b074447b1e64cf63160f573da086f934c8cb712615b71d4" Dec 13 07:38:21 crc kubenswrapper[4971]: E1213 07:38:21.356010 4971 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-82xjz_openshift-machine-config-operator(e0c4af12-edda-41d7-9b44-f87396174ef2)\"" pod="openshift-machine-config-operator/machine-config-daemon-82xjz" podUID="e0c4af12-edda-41d7-9b44-f87396174ef2" Dec 13 07:38:33 crc kubenswrapper[4971]: I1213 07:38:33.801847 4971 scope.go:117] "RemoveContainer" containerID="09d0904c76ce78aa0b074447b1e64cf63160f573da086f934c8cb712615b71d4" Dec 13 07:38:33 crc kubenswrapper[4971]: E1213 07:38:33.804338 4971 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-82xjz_openshift-machine-config-operator(e0c4af12-edda-41d7-9b44-f87396174ef2)\"" pod="openshift-machine-config-operator/machine-config-daemon-82xjz" podUID="e0c4af12-edda-41d7-9b44-f87396174ef2" Dec 13 07:38:38 crc kubenswrapper[4971]: I1213 07:38:38.521737 4971 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-7cbwp" event={"ID":"a35ae77d-f799-4da5-9f79-c6e746b947b8","Type":"ContainerStarted","Data":"59b83a72b81de97df9b965c104c26152370c0667b798646695c2cab0f5b22ad2"} Dec 13 07:38:38 crc kubenswrapper[4971]: I1213 07:38:38.524750 4971 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-gczff" event={"ID":"b1cb4a29-db37-4323-9ebb-1c5e67d1668e","Type":"ContainerStarted","Data":"aa5c5f2fb90eac90d93fff5686813d010b84ffd5c4258ca436fba098d2a533fc"} Dec 13 07:38:38 crc kubenswrapper[4971]: I1213 07:38:38.527786 4971 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-kzw5r" event={"ID":"b2475f4e-bfe7-4678-8779-2b56c189b3b0","Type":"ContainerStarted","Data":"11f1862e923bece9b65315bff525b8527dadd56db1be11e6f2b2034f0ba96eb8"} Dec 13 07:38:38 crc kubenswrapper[4971]: I1213 07:38:38.535899 4971 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/configure-network-edpm-deployment-openstack-edpm-ipam-95c87" event={"ID":"ef08e99c-0bd8-4f27-ae28-2cfdeb4ec045","Type":"ContainerStarted","Data":"4278d8e4b0edce7206d8f770cca876e18985c8ab127d9b7cc25e851d326f368c"} Dec 13 07:38:38 crc kubenswrapper[4971]: I1213 07:38:38.552715 4971 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-operators-7cbwp" podStartSLOduration=5.013035601 podStartE2EDuration="46.552690885s" podCreationTimestamp="2025-12-13 07:37:52 +0000 UTC" firstStartedPulling="2025-12-13 07:37:55.952857583 +0000 UTC m=+2932.557267071" lastFinishedPulling="2025-12-13 07:38:37.492512907 +0000 UTC m=+2974.096922355" observedRunningTime="2025-12-13 07:38:38.550635795 +0000 UTC m=+2975.155045253" watchObservedRunningTime="2025-12-13 07:38:38.552690885 +0000 UTC m=+2975.157100373" Dec 13 07:38:38 crc kubenswrapper[4971]: I1213 07:38:38.602192 4971 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/configure-network-edpm-deployment-openstack-edpm-ipam-95c87" podStartSLOduration=5.6628468640000005 podStartE2EDuration="26.602172692s" podCreationTimestamp="2025-12-13 07:38:12 +0000 UTC" firstStartedPulling="2025-12-13 07:38:16.554346437 +0000 UTC m=+2953.158755885" lastFinishedPulling="2025-12-13 07:38:37.493672255 +0000 UTC m=+2974.098081713" observedRunningTime="2025-12-13 07:38:38.593480414 +0000 UTC m=+2975.197889872" watchObservedRunningTime="2025-12-13 07:38:38.602172692 +0000 UTC m=+2975.206582150" Dec 13 07:38:38 crc kubenswrapper[4971]: I1213 07:38:38.625373 4971 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-marketplace-gczff" podStartSLOduration=6.031249829 podStartE2EDuration="47.625344599s" podCreationTimestamp="2025-12-13 07:37:51 +0000 UTC" firstStartedPulling="2025-12-13 07:37:55.957811382 +0000 UTC m=+2932.562220830" lastFinishedPulling="2025-12-13 07:38:37.551906142 +0000 UTC m=+2974.156315600" observedRunningTime="2025-12-13 07:38:38.621189929 +0000 UTC m=+2975.225599377" watchObservedRunningTime="2025-12-13 07:38:38.625344599 +0000 UTC m=+2975.229754057" Dec 13 07:38:39 crc kubenswrapper[4971]: I1213 07:38:39.547130 4971 generic.go:334] "Generic (PLEG): container finished" podID="b2475f4e-bfe7-4678-8779-2b56c189b3b0" containerID="11f1862e923bece9b65315bff525b8527dadd56db1be11e6f2b2034f0ba96eb8" exitCode=0 Dec 13 07:38:39 crc kubenswrapper[4971]: I1213 07:38:39.547192 4971 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-kzw5r" event={"ID":"b2475f4e-bfe7-4678-8779-2b56c189b3b0","Type":"ContainerDied","Data":"11f1862e923bece9b65315bff525b8527dadd56db1be11e6f2b2034f0ba96eb8"} Dec 13 07:38:42 crc kubenswrapper[4971]: I1213 07:38:42.410859 4971 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-operators-7cbwp" Dec 13 07:38:42 crc kubenswrapper[4971]: I1213 07:38:42.411473 4971 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-operators-7cbwp" Dec 13 07:38:42 crc kubenswrapper[4971]: I1213 07:38:42.435549 4971 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-marketplace-gczff" Dec 13 07:38:42 crc kubenswrapper[4971]: I1213 07:38:42.435920 4971 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-marketplace-gczff" Dec 13 07:38:42 crc kubenswrapper[4971]: I1213 07:38:42.487657 4971 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-marketplace-gczff" Dec 13 07:38:42 crc kubenswrapper[4971]: I1213 07:38:42.625296 4971 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-marketplace-gczff" Dec 13 07:38:42 crc kubenswrapper[4971]: I1213 07:38:42.774759 4971 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-gczff"] Dec 13 07:38:43 crc kubenswrapper[4971]: I1213 07:38:43.460879 4971 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-marketplace/redhat-operators-7cbwp" podUID="a35ae77d-f799-4da5-9f79-c6e746b947b8" containerName="registry-server" probeResult="failure" output=< Dec 13 07:38:43 crc kubenswrapper[4971]: timeout: failed to connect service ":50051" within 1s Dec 13 07:38:43 crc kubenswrapper[4971]: > Dec 13 07:38:44 crc kubenswrapper[4971]: I1213 07:38:44.587591 4971 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-marketplace-gczff" podUID="b1cb4a29-db37-4323-9ebb-1c5e67d1668e" containerName="registry-server" containerID="cri-o://aa5c5f2fb90eac90d93fff5686813d010b84ffd5c4258ca436fba098d2a533fc" gracePeriod=2 Dec 13 07:38:46 crc kubenswrapper[4971]: I1213 07:38:46.613158 4971 generic.go:334] "Generic (PLEG): container finished" podID="b1cb4a29-db37-4323-9ebb-1c5e67d1668e" containerID="aa5c5f2fb90eac90d93fff5686813d010b84ffd5c4258ca436fba098d2a533fc" exitCode=0 Dec 13 07:38:46 crc kubenswrapper[4971]: I1213 07:38:46.613236 4971 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-gczff" event={"ID":"b1cb4a29-db37-4323-9ebb-1c5e67d1668e","Type":"ContainerDied","Data":"aa5c5f2fb90eac90d93fff5686813d010b84ffd5c4258ca436fba098d2a533fc"} Dec 13 07:38:47 crc kubenswrapper[4971]: I1213 07:38:47.008548 4971 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-gczff" Dec 13 07:38:47 crc kubenswrapper[4971]: I1213 07:38:47.026637 4971 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/b1cb4a29-db37-4323-9ebb-1c5e67d1668e-utilities\") pod \"b1cb4a29-db37-4323-9ebb-1c5e67d1668e\" (UID: \"b1cb4a29-db37-4323-9ebb-1c5e67d1668e\") " Dec 13 07:38:47 crc kubenswrapper[4971]: I1213 07:38:47.028415 4971 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/b1cb4a29-db37-4323-9ebb-1c5e67d1668e-catalog-content\") pod \"b1cb4a29-db37-4323-9ebb-1c5e67d1668e\" (UID: \"b1cb4a29-db37-4323-9ebb-1c5e67d1668e\") " Dec 13 07:38:47 crc kubenswrapper[4971]: I1213 07:38:47.028449 4971 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-4zdbf\" (UniqueName: \"kubernetes.io/projected/b1cb4a29-db37-4323-9ebb-1c5e67d1668e-kube-api-access-4zdbf\") pod \"b1cb4a29-db37-4323-9ebb-1c5e67d1668e\" (UID: \"b1cb4a29-db37-4323-9ebb-1c5e67d1668e\") " Dec 13 07:38:47 crc kubenswrapper[4971]: I1213 07:38:47.030187 4971 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/b1cb4a29-db37-4323-9ebb-1c5e67d1668e-utilities" (OuterVolumeSpecName: "utilities") pod "b1cb4a29-db37-4323-9ebb-1c5e67d1668e" (UID: "b1cb4a29-db37-4323-9ebb-1c5e67d1668e"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 13 07:38:47 crc kubenswrapper[4971]: I1213 07:38:47.047039 4971 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/b1cb4a29-db37-4323-9ebb-1c5e67d1668e-kube-api-access-4zdbf" (OuterVolumeSpecName: "kube-api-access-4zdbf") pod "b1cb4a29-db37-4323-9ebb-1c5e67d1668e" (UID: "b1cb4a29-db37-4323-9ebb-1c5e67d1668e"). InnerVolumeSpecName "kube-api-access-4zdbf". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 13 07:38:47 crc kubenswrapper[4971]: I1213 07:38:47.067418 4971 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/b1cb4a29-db37-4323-9ebb-1c5e67d1668e-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "b1cb4a29-db37-4323-9ebb-1c5e67d1668e" (UID: "b1cb4a29-db37-4323-9ebb-1c5e67d1668e"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 13 07:38:47 crc kubenswrapper[4971]: I1213 07:38:47.130050 4971 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/b1cb4a29-db37-4323-9ebb-1c5e67d1668e-utilities\") on node \"crc\" DevicePath \"\"" Dec 13 07:38:47 crc kubenswrapper[4971]: I1213 07:38:47.130084 4971 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/b1cb4a29-db37-4323-9ebb-1c5e67d1668e-catalog-content\") on node \"crc\" DevicePath \"\"" Dec 13 07:38:47 crc kubenswrapper[4971]: I1213 07:38:47.130096 4971 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-4zdbf\" (UniqueName: \"kubernetes.io/projected/b1cb4a29-db37-4323-9ebb-1c5e67d1668e-kube-api-access-4zdbf\") on node \"crc\" DevicePath \"\"" Dec 13 07:38:47 crc kubenswrapper[4971]: I1213 07:38:47.623024 4971 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-kzw5r" event={"ID":"b2475f4e-bfe7-4678-8779-2b56c189b3b0","Type":"ContainerStarted","Data":"790af74352159d524f4828c358893b82ca6f68004bcf699dc115133cdcc23bc1"} Dec 13 07:38:47 crc kubenswrapper[4971]: I1213 07:38:47.626308 4971 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-gczff" event={"ID":"b1cb4a29-db37-4323-9ebb-1c5e67d1668e","Type":"ContainerDied","Data":"6fc43ca17b8ed1c2b21d8076d3fe281c3e7b9260485e430eed3591c6880b2a55"} Dec 13 07:38:47 crc kubenswrapper[4971]: I1213 07:38:47.626350 4971 scope.go:117] "RemoveContainer" containerID="aa5c5f2fb90eac90d93fff5686813d010b84ffd5c4258ca436fba098d2a533fc" Dec 13 07:38:47 crc kubenswrapper[4971]: I1213 07:38:47.626400 4971 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-gczff" Dec 13 07:38:47 crc kubenswrapper[4971]: I1213 07:38:47.649460 4971 scope.go:117] "RemoveContainer" containerID="bc1edf3b6a1cfd239bc684daa88f85a6503e61d05084a620396ef36b1187130a" Dec 13 07:38:47 crc kubenswrapper[4971]: I1213 07:38:47.650476 4971 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/community-operators-kzw5r" podStartSLOduration=3.252470899 podStartE2EDuration="53.650458586s" podCreationTimestamp="2025-12-13 07:37:54 +0000 UTC" firstStartedPulling="2025-12-13 07:37:55.947661519 +0000 UTC m=+2932.552070967" lastFinishedPulling="2025-12-13 07:38:46.345649206 +0000 UTC m=+2982.950058654" observedRunningTime="2025-12-13 07:38:47.647438023 +0000 UTC m=+2984.251847461" watchObservedRunningTime="2025-12-13 07:38:47.650458586 +0000 UTC m=+2984.254868034" Dec 13 07:38:47 crc kubenswrapper[4971]: I1213 07:38:47.670150 4971 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-gczff"] Dec 13 07:38:47 crc kubenswrapper[4971]: I1213 07:38:47.680007 4971 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-marketplace-gczff"] Dec 13 07:38:47 crc kubenswrapper[4971]: I1213 07:38:47.691945 4971 scope.go:117] "RemoveContainer" containerID="2271396c7c130beac1da746bac38527b915b89bea9fb90c02c1e36fea202a98b" Dec 13 07:38:47 crc kubenswrapper[4971]: I1213 07:38:47.782757 4971 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="b1cb4a29-db37-4323-9ebb-1c5e67d1668e" path="/var/lib/kubelet/pods/b1cb4a29-db37-4323-9ebb-1c5e67d1668e/volumes" Dec 13 07:38:48 crc kubenswrapper[4971]: I1213 07:38:48.768362 4971 scope.go:117] "RemoveContainer" containerID="09d0904c76ce78aa0b074447b1e64cf63160f573da086f934c8cb712615b71d4" Dec 13 07:38:48 crc kubenswrapper[4971]: E1213 07:38:48.768901 4971 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-82xjz_openshift-machine-config-operator(e0c4af12-edda-41d7-9b44-f87396174ef2)\"" pod="openshift-machine-config-operator/machine-config-daemon-82xjz" podUID="e0c4af12-edda-41d7-9b44-f87396174ef2" Dec 13 07:38:52 crc kubenswrapper[4971]: I1213 07:38:52.468706 4971 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-operators-7cbwp" Dec 13 07:38:52 crc kubenswrapper[4971]: I1213 07:38:52.525695 4971 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-operators-7cbwp" Dec 13 07:38:53 crc kubenswrapper[4971]: I1213 07:38:53.007258 4971 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-7cbwp"] Dec 13 07:38:53 crc kubenswrapper[4971]: I1213 07:38:53.695719 4971 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-operators-7cbwp" podUID="a35ae77d-f799-4da5-9f79-c6e746b947b8" containerName="registry-server" containerID="cri-o://59b83a72b81de97df9b965c104c26152370c0667b798646695c2cab0f5b22ad2" gracePeriod=2 Dec 13 07:38:54 crc kubenswrapper[4971]: I1213 07:38:54.588060 4971 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/community-operators-kzw5r" Dec 13 07:38:54 crc kubenswrapper[4971]: I1213 07:38:54.589263 4971 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/community-operators-kzw5r" Dec 13 07:38:54 crc kubenswrapper[4971]: I1213 07:38:54.636503 4971 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/community-operators-kzw5r" Dec 13 07:38:54 crc kubenswrapper[4971]: I1213 07:38:54.658946 4971 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-7cbwp" Dec 13 07:38:54 crc kubenswrapper[4971]: I1213 07:38:54.705898 4971 generic.go:334] "Generic (PLEG): container finished" podID="a35ae77d-f799-4da5-9f79-c6e746b947b8" containerID="59b83a72b81de97df9b965c104c26152370c0667b798646695c2cab0f5b22ad2" exitCode=0 Dec 13 07:38:54 crc kubenswrapper[4971]: I1213 07:38:54.705964 4971 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-7cbwp" Dec 13 07:38:54 crc kubenswrapper[4971]: I1213 07:38:54.705997 4971 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-7cbwp" event={"ID":"a35ae77d-f799-4da5-9f79-c6e746b947b8","Type":"ContainerDied","Data":"59b83a72b81de97df9b965c104c26152370c0667b798646695c2cab0f5b22ad2"} Dec 13 07:38:54 crc kubenswrapper[4971]: I1213 07:38:54.706051 4971 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-7cbwp" event={"ID":"a35ae77d-f799-4da5-9f79-c6e746b947b8","Type":"ContainerDied","Data":"49644044b57c792b8cd49dbd312aee4fb37d64a5670cc2a45873b6d750a74c64"} Dec 13 07:38:54 crc kubenswrapper[4971]: I1213 07:38:54.706068 4971 scope.go:117] "RemoveContainer" containerID="59b83a72b81de97df9b965c104c26152370c0667b798646695c2cab0f5b22ad2" Dec 13 07:38:54 crc kubenswrapper[4971]: I1213 07:38:54.725291 4971 scope.go:117] "RemoveContainer" containerID="13041742c489627311b990c47b0b1fb64e78198e38710e78517b779991176e80" Dec 13 07:38:54 crc kubenswrapper[4971]: I1213 07:38:54.747119 4971 scope.go:117] "RemoveContainer" containerID="47b2b2b58223551c494b2ffbe0403cb6c1a27b9ac87464342f7cd8393866e53e" Dec 13 07:38:54 crc kubenswrapper[4971]: I1213 07:38:54.749749 4971 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/community-operators-kzw5r" Dec 13 07:38:54 crc kubenswrapper[4971]: I1213 07:38:54.782312 4971 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-h7kdw\" (UniqueName: \"kubernetes.io/projected/a35ae77d-f799-4da5-9f79-c6e746b947b8-kube-api-access-h7kdw\") pod \"a35ae77d-f799-4da5-9f79-c6e746b947b8\" (UID: \"a35ae77d-f799-4da5-9f79-c6e746b947b8\") " Dec 13 07:38:54 crc kubenswrapper[4971]: I1213 07:38:54.782420 4971 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/a35ae77d-f799-4da5-9f79-c6e746b947b8-utilities\") pod \"a35ae77d-f799-4da5-9f79-c6e746b947b8\" (UID: \"a35ae77d-f799-4da5-9f79-c6e746b947b8\") " Dec 13 07:38:54 crc kubenswrapper[4971]: I1213 07:38:54.782472 4971 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/a35ae77d-f799-4da5-9f79-c6e746b947b8-catalog-content\") pod \"a35ae77d-f799-4da5-9f79-c6e746b947b8\" (UID: \"a35ae77d-f799-4da5-9f79-c6e746b947b8\") " Dec 13 07:38:54 crc kubenswrapper[4971]: I1213 07:38:54.783953 4971 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/a35ae77d-f799-4da5-9f79-c6e746b947b8-utilities" (OuterVolumeSpecName: "utilities") pod "a35ae77d-f799-4da5-9f79-c6e746b947b8" (UID: "a35ae77d-f799-4da5-9f79-c6e746b947b8"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 13 07:38:54 crc kubenswrapper[4971]: I1213 07:38:54.785918 4971 scope.go:117] "RemoveContainer" containerID="59b83a72b81de97df9b965c104c26152370c0667b798646695c2cab0f5b22ad2" Dec 13 07:38:54 crc kubenswrapper[4971]: E1213 07:38:54.788737 4971 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"59b83a72b81de97df9b965c104c26152370c0667b798646695c2cab0f5b22ad2\": container with ID starting with 59b83a72b81de97df9b965c104c26152370c0667b798646695c2cab0f5b22ad2 not found: ID does not exist" containerID="59b83a72b81de97df9b965c104c26152370c0667b798646695c2cab0f5b22ad2" Dec 13 07:38:54 crc kubenswrapper[4971]: I1213 07:38:54.788799 4971 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"59b83a72b81de97df9b965c104c26152370c0667b798646695c2cab0f5b22ad2"} err="failed to get container status \"59b83a72b81de97df9b965c104c26152370c0667b798646695c2cab0f5b22ad2\": rpc error: code = NotFound desc = could not find container \"59b83a72b81de97df9b965c104c26152370c0667b798646695c2cab0f5b22ad2\": container with ID starting with 59b83a72b81de97df9b965c104c26152370c0667b798646695c2cab0f5b22ad2 not found: ID does not exist" Dec 13 07:38:54 crc kubenswrapper[4971]: I1213 07:38:54.788831 4971 scope.go:117] "RemoveContainer" containerID="13041742c489627311b990c47b0b1fb64e78198e38710e78517b779991176e80" Dec 13 07:38:54 crc kubenswrapper[4971]: E1213 07:38:54.789602 4971 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"13041742c489627311b990c47b0b1fb64e78198e38710e78517b779991176e80\": container with ID starting with 13041742c489627311b990c47b0b1fb64e78198e38710e78517b779991176e80 not found: ID does not exist" containerID="13041742c489627311b990c47b0b1fb64e78198e38710e78517b779991176e80" Dec 13 07:38:54 crc kubenswrapper[4971]: I1213 07:38:54.789646 4971 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"13041742c489627311b990c47b0b1fb64e78198e38710e78517b779991176e80"} err="failed to get container status \"13041742c489627311b990c47b0b1fb64e78198e38710e78517b779991176e80\": rpc error: code = NotFound desc = could not find container \"13041742c489627311b990c47b0b1fb64e78198e38710e78517b779991176e80\": container with ID starting with 13041742c489627311b990c47b0b1fb64e78198e38710e78517b779991176e80 not found: ID does not exist" Dec 13 07:38:54 crc kubenswrapper[4971]: I1213 07:38:54.789677 4971 scope.go:117] "RemoveContainer" containerID="47b2b2b58223551c494b2ffbe0403cb6c1a27b9ac87464342f7cd8393866e53e" Dec 13 07:38:54 crc kubenswrapper[4971]: I1213 07:38:54.790007 4971 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/a35ae77d-f799-4da5-9f79-c6e746b947b8-kube-api-access-h7kdw" (OuterVolumeSpecName: "kube-api-access-h7kdw") pod "a35ae77d-f799-4da5-9f79-c6e746b947b8" (UID: "a35ae77d-f799-4da5-9f79-c6e746b947b8"). InnerVolumeSpecName "kube-api-access-h7kdw". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 13 07:38:54 crc kubenswrapper[4971]: E1213 07:38:54.790436 4971 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"47b2b2b58223551c494b2ffbe0403cb6c1a27b9ac87464342f7cd8393866e53e\": container with ID starting with 47b2b2b58223551c494b2ffbe0403cb6c1a27b9ac87464342f7cd8393866e53e not found: ID does not exist" containerID="47b2b2b58223551c494b2ffbe0403cb6c1a27b9ac87464342f7cd8393866e53e" Dec 13 07:38:54 crc kubenswrapper[4971]: I1213 07:38:54.790474 4971 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"47b2b2b58223551c494b2ffbe0403cb6c1a27b9ac87464342f7cd8393866e53e"} err="failed to get container status \"47b2b2b58223551c494b2ffbe0403cb6c1a27b9ac87464342f7cd8393866e53e\": rpc error: code = NotFound desc = could not find container \"47b2b2b58223551c494b2ffbe0403cb6c1a27b9ac87464342f7cd8393866e53e\": container with ID starting with 47b2b2b58223551c494b2ffbe0403cb6c1a27b9ac87464342f7cd8393866e53e not found: ID does not exist" Dec 13 07:38:54 crc kubenswrapper[4971]: I1213 07:38:54.885424 4971 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-h7kdw\" (UniqueName: \"kubernetes.io/projected/a35ae77d-f799-4da5-9f79-c6e746b947b8-kube-api-access-h7kdw\") on node \"crc\" DevicePath \"\"" Dec 13 07:38:54 crc kubenswrapper[4971]: I1213 07:38:54.885461 4971 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/a35ae77d-f799-4da5-9f79-c6e746b947b8-utilities\") on node \"crc\" DevicePath \"\"" Dec 13 07:38:54 crc kubenswrapper[4971]: I1213 07:38:54.906049 4971 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/a35ae77d-f799-4da5-9f79-c6e746b947b8-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "a35ae77d-f799-4da5-9f79-c6e746b947b8" (UID: "a35ae77d-f799-4da5-9f79-c6e746b947b8"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 13 07:38:54 crc kubenswrapper[4971]: I1213 07:38:54.988350 4971 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/a35ae77d-f799-4da5-9f79-c6e746b947b8-catalog-content\") on node \"crc\" DevicePath \"\"" Dec 13 07:38:55 crc kubenswrapper[4971]: I1213 07:38:55.039377 4971 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-7cbwp"] Dec 13 07:38:55 crc kubenswrapper[4971]: I1213 07:38:55.045753 4971 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-operators-7cbwp"] Dec 13 07:38:55 crc kubenswrapper[4971]: I1213 07:38:55.779038 4971 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="a35ae77d-f799-4da5-9f79-c6e746b947b8" path="/var/lib/kubelet/pods/a35ae77d-f799-4da5-9f79-c6e746b947b8/volumes" Dec 13 07:38:56 crc kubenswrapper[4971]: I1213 07:38:56.028497 4971 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-kzw5r"] Dec 13 07:38:56 crc kubenswrapper[4971]: I1213 07:38:56.405089 4971 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-z62gm"] Dec 13 07:38:56 crc kubenswrapper[4971]: I1213 07:38:56.405603 4971 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/community-operators-z62gm" podUID="af866491-3f91-4f43-809b-47b473e2e53d" containerName="registry-server" containerID="cri-o://16ad47ec9ccf05452ca25d481805aa6c176618de22b4b31f0c2a40625bead6dc" gracePeriod=2 Dec 13 07:38:56 crc kubenswrapper[4971]: I1213 07:38:56.731757 4971 generic.go:334] "Generic (PLEG): container finished" podID="af866491-3f91-4f43-809b-47b473e2e53d" containerID="16ad47ec9ccf05452ca25d481805aa6c176618de22b4b31f0c2a40625bead6dc" exitCode=0 Dec 13 07:38:56 crc kubenswrapper[4971]: I1213 07:38:56.732006 4971 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-z62gm" event={"ID":"af866491-3f91-4f43-809b-47b473e2e53d","Type":"ContainerDied","Data":"16ad47ec9ccf05452ca25d481805aa6c176618de22b4b31f0c2a40625bead6dc"} Dec 13 07:38:56 crc kubenswrapper[4971]: I1213 07:38:56.893146 4971 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-z62gm" Dec 13 07:38:57 crc kubenswrapper[4971]: I1213 07:38:57.023654 4971 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-dz96m\" (UniqueName: \"kubernetes.io/projected/af866491-3f91-4f43-809b-47b473e2e53d-kube-api-access-dz96m\") pod \"af866491-3f91-4f43-809b-47b473e2e53d\" (UID: \"af866491-3f91-4f43-809b-47b473e2e53d\") " Dec 13 07:38:57 crc kubenswrapper[4971]: I1213 07:38:57.024076 4971 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/af866491-3f91-4f43-809b-47b473e2e53d-catalog-content\") pod \"af866491-3f91-4f43-809b-47b473e2e53d\" (UID: \"af866491-3f91-4f43-809b-47b473e2e53d\") " Dec 13 07:38:57 crc kubenswrapper[4971]: I1213 07:38:57.024136 4971 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/af866491-3f91-4f43-809b-47b473e2e53d-utilities\") pod \"af866491-3f91-4f43-809b-47b473e2e53d\" (UID: \"af866491-3f91-4f43-809b-47b473e2e53d\") " Dec 13 07:38:57 crc kubenswrapper[4971]: I1213 07:38:57.024745 4971 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/af866491-3f91-4f43-809b-47b473e2e53d-utilities" (OuterVolumeSpecName: "utilities") pod "af866491-3f91-4f43-809b-47b473e2e53d" (UID: "af866491-3f91-4f43-809b-47b473e2e53d"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 13 07:38:57 crc kubenswrapper[4971]: I1213 07:38:57.029260 4971 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/af866491-3f91-4f43-809b-47b473e2e53d-kube-api-access-dz96m" (OuterVolumeSpecName: "kube-api-access-dz96m") pod "af866491-3f91-4f43-809b-47b473e2e53d" (UID: "af866491-3f91-4f43-809b-47b473e2e53d"). InnerVolumeSpecName "kube-api-access-dz96m". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 13 07:38:57 crc kubenswrapper[4971]: I1213 07:38:57.076351 4971 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/af866491-3f91-4f43-809b-47b473e2e53d-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "af866491-3f91-4f43-809b-47b473e2e53d" (UID: "af866491-3f91-4f43-809b-47b473e2e53d"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 13 07:38:57 crc kubenswrapper[4971]: I1213 07:38:57.125982 4971 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/af866491-3f91-4f43-809b-47b473e2e53d-catalog-content\") on node \"crc\" DevicePath \"\"" Dec 13 07:38:57 crc kubenswrapper[4971]: I1213 07:38:57.126017 4971 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/af866491-3f91-4f43-809b-47b473e2e53d-utilities\") on node \"crc\" DevicePath \"\"" Dec 13 07:38:57 crc kubenswrapper[4971]: I1213 07:38:57.126029 4971 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-dz96m\" (UniqueName: \"kubernetes.io/projected/af866491-3f91-4f43-809b-47b473e2e53d-kube-api-access-dz96m\") on node \"crc\" DevicePath \"\"" Dec 13 07:38:57 crc kubenswrapper[4971]: I1213 07:38:57.743590 4971 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-z62gm" event={"ID":"af866491-3f91-4f43-809b-47b473e2e53d","Type":"ContainerDied","Data":"f445f08f77bd2f2654a517a3a35f5a0a57a0a0afb83ad082a26a7401778a6f38"} Dec 13 07:38:57 crc kubenswrapper[4971]: I1213 07:38:57.743634 4971 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-z62gm" Dec 13 07:38:57 crc kubenswrapper[4971]: I1213 07:38:57.743659 4971 scope.go:117] "RemoveContainer" containerID="16ad47ec9ccf05452ca25d481805aa6c176618de22b4b31f0c2a40625bead6dc" Dec 13 07:38:57 crc kubenswrapper[4971]: I1213 07:38:57.767004 4971 scope.go:117] "RemoveContainer" containerID="6e0143cb7ec4ee22e7f2a12722efb6a8345395ac858586f4608d91ec233f7f63" Dec 13 07:38:57 crc kubenswrapper[4971]: I1213 07:38:57.778913 4971 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-z62gm"] Dec 13 07:38:57 crc kubenswrapper[4971]: I1213 07:38:57.787291 4971 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/community-operators-z62gm"] Dec 13 07:38:57 crc kubenswrapper[4971]: I1213 07:38:57.806950 4971 scope.go:117] "RemoveContainer" containerID="9dadbcb68ebd5832cdbbf8fba0f228defdeccac4a5b7715e71465d39c164d555" Dec 13 07:38:59 crc kubenswrapper[4971]: I1213 07:38:59.801379 4971 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="af866491-3f91-4f43-809b-47b473e2e53d" path="/var/lib/kubelet/pods/af866491-3f91-4f43-809b-47b473e2e53d/volumes" Dec 13 07:39:01 crc kubenswrapper[4971]: I1213 07:39:01.768847 4971 scope.go:117] "RemoveContainer" containerID="09d0904c76ce78aa0b074447b1e64cf63160f573da086f934c8cb712615b71d4" Dec 13 07:39:01 crc kubenswrapper[4971]: E1213 07:39:01.769536 4971 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-82xjz_openshift-machine-config-operator(e0c4af12-edda-41d7-9b44-f87396174ef2)\"" pod="openshift-machine-config-operator/machine-config-daemon-82xjz" podUID="e0c4af12-edda-41d7-9b44-f87396174ef2" Dec 13 07:39:08 crc kubenswrapper[4971]: I1213 07:39:08.847096 4971 generic.go:334] "Generic (PLEG): container finished" podID="ef08e99c-0bd8-4f27-ae28-2cfdeb4ec045" containerID="4278d8e4b0edce7206d8f770cca876e18985c8ab127d9b7cc25e851d326f368c" exitCode=0 Dec 13 07:39:08 crc kubenswrapper[4971]: I1213 07:39:08.847207 4971 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/configure-network-edpm-deployment-openstack-edpm-ipam-95c87" event={"ID":"ef08e99c-0bd8-4f27-ae28-2cfdeb4ec045","Type":"ContainerDied","Data":"4278d8e4b0edce7206d8f770cca876e18985c8ab127d9b7cc25e851d326f368c"} Dec 13 07:39:10 crc kubenswrapper[4971]: I1213 07:39:10.339672 4971 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/configure-network-edpm-deployment-openstack-edpm-ipam-95c87" Dec 13 07:39:10 crc kubenswrapper[4971]: I1213 07:39:10.419212 4971 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/ef08e99c-0bd8-4f27-ae28-2cfdeb4ec045-ceph\") pod \"ef08e99c-0bd8-4f27-ae28-2cfdeb4ec045\" (UID: \"ef08e99c-0bd8-4f27-ae28-2cfdeb4ec045\") " Dec 13 07:39:10 crc kubenswrapper[4971]: I1213 07:39:10.419655 4971 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/ef08e99c-0bd8-4f27-ae28-2cfdeb4ec045-ssh-key\") pod \"ef08e99c-0bd8-4f27-ae28-2cfdeb4ec045\" (UID: \"ef08e99c-0bd8-4f27-ae28-2cfdeb4ec045\") " Dec 13 07:39:10 crc kubenswrapper[4971]: I1213 07:39:10.419806 4971 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-rr4ff\" (UniqueName: \"kubernetes.io/projected/ef08e99c-0bd8-4f27-ae28-2cfdeb4ec045-kube-api-access-rr4ff\") pod \"ef08e99c-0bd8-4f27-ae28-2cfdeb4ec045\" (UID: \"ef08e99c-0bd8-4f27-ae28-2cfdeb4ec045\") " Dec 13 07:39:10 crc kubenswrapper[4971]: I1213 07:39:10.419864 4971 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/ef08e99c-0bd8-4f27-ae28-2cfdeb4ec045-inventory\") pod \"ef08e99c-0bd8-4f27-ae28-2cfdeb4ec045\" (UID: \"ef08e99c-0bd8-4f27-ae28-2cfdeb4ec045\") " Dec 13 07:39:10 crc kubenswrapper[4971]: I1213 07:39:10.427466 4971 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/ef08e99c-0bd8-4f27-ae28-2cfdeb4ec045-kube-api-access-rr4ff" (OuterVolumeSpecName: "kube-api-access-rr4ff") pod "ef08e99c-0bd8-4f27-ae28-2cfdeb4ec045" (UID: "ef08e99c-0bd8-4f27-ae28-2cfdeb4ec045"). InnerVolumeSpecName "kube-api-access-rr4ff". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 13 07:39:10 crc kubenswrapper[4971]: I1213 07:39:10.429863 4971 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/ef08e99c-0bd8-4f27-ae28-2cfdeb4ec045-ceph" (OuterVolumeSpecName: "ceph") pod "ef08e99c-0bd8-4f27-ae28-2cfdeb4ec045" (UID: "ef08e99c-0bd8-4f27-ae28-2cfdeb4ec045"). InnerVolumeSpecName "ceph". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 13 07:39:10 crc kubenswrapper[4971]: I1213 07:39:10.444475 4971 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/ef08e99c-0bd8-4f27-ae28-2cfdeb4ec045-inventory" (OuterVolumeSpecName: "inventory") pod "ef08e99c-0bd8-4f27-ae28-2cfdeb4ec045" (UID: "ef08e99c-0bd8-4f27-ae28-2cfdeb4ec045"). InnerVolumeSpecName "inventory". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 13 07:39:10 crc kubenswrapper[4971]: I1213 07:39:10.445587 4971 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/ef08e99c-0bd8-4f27-ae28-2cfdeb4ec045-ssh-key" (OuterVolumeSpecName: "ssh-key") pod "ef08e99c-0bd8-4f27-ae28-2cfdeb4ec045" (UID: "ef08e99c-0bd8-4f27-ae28-2cfdeb4ec045"). InnerVolumeSpecName "ssh-key". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 13 07:39:10 crc kubenswrapper[4971]: I1213 07:39:10.521583 4971 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-rr4ff\" (UniqueName: \"kubernetes.io/projected/ef08e99c-0bd8-4f27-ae28-2cfdeb4ec045-kube-api-access-rr4ff\") on node \"crc\" DevicePath \"\"" Dec 13 07:39:10 crc kubenswrapper[4971]: I1213 07:39:10.521620 4971 reconciler_common.go:293] "Volume detached for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/ef08e99c-0bd8-4f27-ae28-2cfdeb4ec045-inventory\") on node \"crc\" DevicePath \"\"" Dec 13 07:39:10 crc kubenswrapper[4971]: I1213 07:39:10.521629 4971 reconciler_common.go:293] "Volume detached for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/ef08e99c-0bd8-4f27-ae28-2cfdeb4ec045-ceph\") on node \"crc\" DevicePath \"\"" Dec 13 07:39:10 crc kubenswrapper[4971]: I1213 07:39:10.521638 4971 reconciler_common.go:293] "Volume detached for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/ef08e99c-0bd8-4f27-ae28-2cfdeb4ec045-ssh-key\") on node \"crc\" DevicePath \"\"" Dec 13 07:39:10 crc kubenswrapper[4971]: I1213 07:39:10.870405 4971 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/configure-network-edpm-deployment-openstack-edpm-ipam-95c87" event={"ID":"ef08e99c-0bd8-4f27-ae28-2cfdeb4ec045","Type":"ContainerDied","Data":"e15bc3336be6c26a6455e4ca601bcfafe6f0ff74d5f73888e456d091b81c726c"} Dec 13 07:39:10 crc kubenswrapper[4971]: I1213 07:39:10.870473 4971 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="e15bc3336be6c26a6455e4ca601bcfafe6f0ff74d5f73888e456d091b81c726c" Dec 13 07:39:10 crc kubenswrapper[4971]: I1213 07:39:10.870484 4971 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/configure-network-edpm-deployment-openstack-edpm-ipam-95c87" Dec 13 07:39:10 crc kubenswrapper[4971]: I1213 07:39:10.969147 4971 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/validate-network-edpm-deployment-openstack-edpm-ipam-4lhml"] Dec 13 07:39:10 crc kubenswrapper[4971]: E1213 07:39:10.969627 4971 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="ef08e99c-0bd8-4f27-ae28-2cfdeb4ec045" containerName="configure-network-edpm-deployment-openstack-edpm-ipam" Dec 13 07:39:10 crc kubenswrapper[4971]: I1213 07:39:10.969652 4971 state_mem.go:107] "Deleted CPUSet assignment" podUID="ef08e99c-0bd8-4f27-ae28-2cfdeb4ec045" containerName="configure-network-edpm-deployment-openstack-edpm-ipam" Dec 13 07:39:10 crc kubenswrapper[4971]: E1213 07:39:10.969670 4971 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="b1cb4a29-db37-4323-9ebb-1c5e67d1668e" containerName="extract-content" Dec 13 07:39:10 crc kubenswrapper[4971]: I1213 07:39:10.969679 4971 state_mem.go:107] "Deleted CPUSet assignment" podUID="b1cb4a29-db37-4323-9ebb-1c5e67d1668e" containerName="extract-content" Dec 13 07:39:10 crc kubenswrapper[4971]: E1213 07:39:10.969708 4971 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="af866491-3f91-4f43-809b-47b473e2e53d" containerName="registry-server" Dec 13 07:39:10 crc kubenswrapper[4971]: I1213 07:39:10.969716 4971 state_mem.go:107] "Deleted CPUSet assignment" podUID="af866491-3f91-4f43-809b-47b473e2e53d" containerName="registry-server" Dec 13 07:39:10 crc kubenswrapper[4971]: E1213 07:39:10.969729 4971 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="af866491-3f91-4f43-809b-47b473e2e53d" containerName="extract-utilities" Dec 13 07:39:10 crc kubenswrapper[4971]: I1213 07:39:10.969735 4971 state_mem.go:107] "Deleted CPUSet assignment" podUID="af866491-3f91-4f43-809b-47b473e2e53d" containerName="extract-utilities" Dec 13 07:39:10 crc kubenswrapper[4971]: E1213 07:39:10.969749 4971 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="af866491-3f91-4f43-809b-47b473e2e53d" containerName="extract-content" Dec 13 07:39:10 crc kubenswrapper[4971]: I1213 07:39:10.969756 4971 state_mem.go:107] "Deleted CPUSet assignment" podUID="af866491-3f91-4f43-809b-47b473e2e53d" containerName="extract-content" Dec 13 07:39:10 crc kubenswrapper[4971]: E1213 07:39:10.969765 4971 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="a35ae77d-f799-4da5-9f79-c6e746b947b8" containerName="extract-content" Dec 13 07:39:10 crc kubenswrapper[4971]: I1213 07:39:10.969771 4971 state_mem.go:107] "Deleted CPUSet assignment" podUID="a35ae77d-f799-4da5-9f79-c6e746b947b8" containerName="extract-content" Dec 13 07:39:10 crc kubenswrapper[4971]: E1213 07:39:10.969782 4971 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="a35ae77d-f799-4da5-9f79-c6e746b947b8" containerName="registry-server" Dec 13 07:39:10 crc kubenswrapper[4971]: I1213 07:39:10.969974 4971 state_mem.go:107] "Deleted CPUSet assignment" podUID="a35ae77d-f799-4da5-9f79-c6e746b947b8" containerName="registry-server" Dec 13 07:39:10 crc kubenswrapper[4971]: E1213 07:39:10.969988 4971 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="b1cb4a29-db37-4323-9ebb-1c5e67d1668e" containerName="registry-server" Dec 13 07:39:10 crc kubenswrapper[4971]: I1213 07:39:10.969996 4971 state_mem.go:107] "Deleted CPUSet assignment" podUID="b1cb4a29-db37-4323-9ebb-1c5e67d1668e" containerName="registry-server" Dec 13 07:39:10 crc kubenswrapper[4971]: E1213 07:39:10.970009 4971 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="a35ae77d-f799-4da5-9f79-c6e746b947b8" containerName="extract-utilities" Dec 13 07:39:10 crc kubenswrapper[4971]: I1213 07:39:10.970017 4971 state_mem.go:107] "Deleted CPUSet assignment" podUID="a35ae77d-f799-4da5-9f79-c6e746b947b8" containerName="extract-utilities" Dec 13 07:39:10 crc kubenswrapper[4971]: E1213 07:39:10.970027 4971 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="b1cb4a29-db37-4323-9ebb-1c5e67d1668e" containerName="extract-utilities" Dec 13 07:39:10 crc kubenswrapper[4971]: I1213 07:39:10.970034 4971 state_mem.go:107] "Deleted CPUSet assignment" podUID="b1cb4a29-db37-4323-9ebb-1c5e67d1668e" containerName="extract-utilities" Dec 13 07:39:10 crc kubenswrapper[4971]: I1213 07:39:10.970229 4971 memory_manager.go:354] "RemoveStaleState removing state" podUID="b1cb4a29-db37-4323-9ebb-1c5e67d1668e" containerName="registry-server" Dec 13 07:39:10 crc kubenswrapper[4971]: I1213 07:39:10.970249 4971 memory_manager.go:354] "RemoveStaleState removing state" podUID="ef08e99c-0bd8-4f27-ae28-2cfdeb4ec045" containerName="configure-network-edpm-deployment-openstack-edpm-ipam" Dec 13 07:39:10 crc kubenswrapper[4971]: I1213 07:39:10.970275 4971 memory_manager.go:354] "RemoveStaleState removing state" podUID="a35ae77d-f799-4da5-9f79-c6e746b947b8" containerName="registry-server" Dec 13 07:39:10 crc kubenswrapper[4971]: I1213 07:39:10.970291 4971 memory_manager.go:354] "RemoveStaleState removing state" podUID="af866491-3f91-4f43-809b-47b473e2e53d" containerName="registry-server" Dec 13 07:39:10 crc kubenswrapper[4971]: I1213 07:39:10.971081 4971 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/validate-network-edpm-deployment-openstack-edpm-ipam-4lhml" Dec 13 07:39:10 crc kubenswrapper[4971]: I1213 07:39:10.975732 4971 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplane-ansible-ssh-private-key-secret" Dec 13 07:39:10 crc kubenswrapper[4971]: I1213 07:39:10.976279 4971 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-edpm-ipam-dockercfg-zb2j5" Dec 13 07:39:10 crc kubenswrapper[4971]: I1213 07:39:10.976890 4971 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-aee-default-env" Dec 13 07:39:10 crc kubenswrapper[4971]: I1213 07:39:10.977074 4971 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceph-conf-files" Dec 13 07:39:10 crc kubenswrapper[4971]: I1213 07:39:10.979274 4971 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplanenodeset-openstack-edpm-ipam" Dec 13 07:39:10 crc kubenswrapper[4971]: I1213 07:39:10.984877 4971 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/validate-network-edpm-deployment-openstack-edpm-ipam-4lhml"] Dec 13 07:39:11 crc kubenswrapper[4971]: I1213 07:39:11.133339 4971 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/2185de1e-5198-4f0f-8918-1e90aadb7fb7-ssh-key\") pod \"validate-network-edpm-deployment-openstack-edpm-ipam-4lhml\" (UID: \"2185de1e-5198-4f0f-8918-1e90aadb7fb7\") " pod="openstack/validate-network-edpm-deployment-openstack-edpm-ipam-4lhml" Dec 13 07:39:11 crc kubenswrapper[4971]: I1213 07:39:11.133411 4971 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/2185de1e-5198-4f0f-8918-1e90aadb7fb7-inventory\") pod \"validate-network-edpm-deployment-openstack-edpm-ipam-4lhml\" (UID: \"2185de1e-5198-4f0f-8918-1e90aadb7fb7\") " pod="openstack/validate-network-edpm-deployment-openstack-edpm-ipam-4lhml" Dec 13 07:39:11 crc kubenswrapper[4971]: I1213 07:39:11.133683 4971 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-r899w\" (UniqueName: \"kubernetes.io/projected/2185de1e-5198-4f0f-8918-1e90aadb7fb7-kube-api-access-r899w\") pod \"validate-network-edpm-deployment-openstack-edpm-ipam-4lhml\" (UID: \"2185de1e-5198-4f0f-8918-1e90aadb7fb7\") " pod="openstack/validate-network-edpm-deployment-openstack-edpm-ipam-4lhml" Dec 13 07:39:11 crc kubenswrapper[4971]: I1213 07:39:11.133767 4971 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/2185de1e-5198-4f0f-8918-1e90aadb7fb7-ceph\") pod \"validate-network-edpm-deployment-openstack-edpm-ipam-4lhml\" (UID: \"2185de1e-5198-4f0f-8918-1e90aadb7fb7\") " pod="openstack/validate-network-edpm-deployment-openstack-edpm-ipam-4lhml" Dec 13 07:39:11 crc kubenswrapper[4971]: I1213 07:39:11.301344 4971 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-r899w\" (UniqueName: \"kubernetes.io/projected/2185de1e-5198-4f0f-8918-1e90aadb7fb7-kube-api-access-r899w\") pod \"validate-network-edpm-deployment-openstack-edpm-ipam-4lhml\" (UID: \"2185de1e-5198-4f0f-8918-1e90aadb7fb7\") " pod="openstack/validate-network-edpm-deployment-openstack-edpm-ipam-4lhml" Dec 13 07:39:11 crc kubenswrapper[4971]: I1213 07:39:11.301428 4971 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/2185de1e-5198-4f0f-8918-1e90aadb7fb7-ceph\") pod \"validate-network-edpm-deployment-openstack-edpm-ipam-4lhml\" (UID: \"2185de1e-5198-4f0f-8918-1e90aadb7fb7\") " pod="openstack/validate-network-edpm-deployment-openstack-edpm-ipam-4lhml" Dec 13 07:39:11 crc kubenswrapper[4971]: I1213 07:39:11.301573 4971 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/2185de1e-5198-4f0f-8918-1e90aadb7fb7-ssh-key\") pod \"validate-network-edpm-deployment-openstack-edpm-ipam-4lhml\" (UID: \"2185de1e-5198-4f0f-8918-1e90aadb7fb7\") " pod="openstack/validate-network-edpm-deployment-openstack-edpm-ipam-4lhml" Dec 13 07:39:11 crc kubenswrapper[4971]: I1213 07:39:11.301622 4971 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/2185de1e-5198-4f0f-8918-1e90aadb7fb7-inventory\") pod \"validate-network-edpm-deployment-openstack-edpm-ipam-4lhml\" (UID: \"2185de1e-5198-4f0f-8918-1e90aadb7fb7\") " pod="openstack/validate-network-edpm-deployment-openstack-edpm-ipam-4lhml" Dec 13 07:39:11 crc kubenswrapper[4971]: I1213 07:39:11.332277 4971 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/2185de1e-5198-4f0f-8918-1e90aadb7fb7-ssh-key\") pod \"validate-network-edpm-deployment-openstack-edpm-ipam-4lhml\" (UID: \"2185de1e-5198-4f0f-8918-1e90aadb7fb7\") " pod="openstack/validate-network-edpm-deployment-openstack-edpm-ipam-4lhml" Dec 13 07:39:11 crc kubenswrapper[4971]: I1213 07:39:11.333201 4971 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/2185de1e-5198-4f0f-8918-1e90aadb7fb7-inventory\") pod \"validate-network-edpm-deployment-openstack-edpm-ipam-4lhml\" (UID: \"2185de1e-5198-4f0f-8918-1e90aadb7fb7\") " pod="openstack/validate-network-edpm-deployment-openstack-edpm-ipam-4lhml" Dec 13 07:39:11 crc kubenswrapper[4971]: I1213 07:39:11.338376 4971 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/2185de1e-5198-4f0f-8918-1e90aadb7fb7-ceph\") pod \"validate-network-edpm-deployment-openstack-edpm-ipam-4lhml\" (UID: \"2185de1e-5198-4f0f-8918-1e90aadb7fb7\") " pod="openstack/validate-network-edpm-deployment-openstack-edpm-ipam-4lhml" Dec 13 07:39:11 crc kubenswrapper[4971]: I1213 07:39:11.343244 4971 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-r899w\" (UniqueName: \"kubernetes.io/projected/2185de1e-5198-4f0f-8918-1e90aadb7fb7-kube-api-access-r899w\") pod \"validate-network-edpm-deployment-openstack-edpm-ipam-4lhml\" (UID: \"2185de1e-5198-4f0f-8918-1e90aadb7fb7\") " pod="openstack/validate-network-edpm-deployment-openstack-edpm-ipam-4lhml" Dec 13 07:39:11 crc kubenswrapper[4971]: I1213 07:39:11.598012 4971 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/validate-network-edpm-deployment-openstack-edpm-ipam-4lhml" Dec 13 07:39:12 crc kubenswrapper[4971]: I1213 07:39:12.311802 4971 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/validate-network-edpm-deployment-openstack-edpm-ipam-4lhml"] Dec 13 07:39:12 crc kubenswrapper[4971]: I1213 07:39:12.894667 4971 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/validate-network-edpm-deployment-openstack-edpm-ipam-4lhml" event={"ID":"2185de1e-5198-4f0f-8918-1e90aadb7fb7","Type":"ContainerStarted","Data":"a188d8a459fdd3c1cb43a6a649704f840396c4186accc263fc8552ee142cc4b2"} Dec 13 07:39:14 crc kubenswrapper[4971]: I1213 07:39:14.769003 4971 scope.go:117] "RemoveContainer" containerID="09d0904c76ce78aa0b074447b1e64cf63160f573da086f934c8cb712615b71d4" Dec 13 07:39:14 crc kubenswrapper[4971]: E1213 07:39:14.769734 4971 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-82xjz_openshift-machine-config-operator(e0c4af12-edda-41d7-9b44-f87396174ef2)\"" pod="openshift-machine-config-operator/machine-config-daemon-82xjz" podUID="e0c4af12-edda-41d7-9b44-f87396174ef2" Dec 13 07:39:17 crc kubenswrapper[4971]: I1213 07:39:17.956336 4971 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/validate-network-edpm-deployment-openstack-edpm-ipam-4lhml" event={"ID":"2185de1e-5198-4f0f-8918-1e90aadb7fb7","Type":"ContainerStarted","Data":"08785fa5a36ada22e4c767ec58a4dcd6765a2b3fa81f8dbc816bfbdd2125aa30"} Dec 13 07:39:17 crc kubenswrapper[4971]: I1213 07:39:17.971116 4971 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/validate-network-edpm-deployment-openstack-edpm-ipam-4lhml" podStartSLOduration=3.057919853 podStartE2EDuration="7.971099618s" podCreationTimestamp="2025-12-13 07:39:10 +0000 UTC" firstStartedPulling="2025-12-13 07:39:12.316719042 +0000 UTC m=+3008.921128490" lastFinishedPulling="2025-12-13 07:39:17.229898807 +0000 UTC m=+3013.834308255" observedRunningTime="2025-12-13 07:39:17.969105781 +0000 UTC m=+3014.573515229" watchObservedRunningTime="2025-12-13 07:39:17.971099618 +0000 UTC m=+3014.575509066" Dec 13 07:39:24 crc kubenswrapper[4971]: I1213 07:39:24.070401 4971 generic.go:334] "Generic (PLEG): container finished" podID="2185de1e-5198-4f0f-8918-1e90aadb7fb7" containerID="08785fa5a36ada22e4c767ec58a4dcd6765a2b3fa81f8dbc816bfbdd2125aa30" exitCode=0 Dec 13 07:39:24 crc kubenswrapper[4971]: I1213 07:39:24.070569 4971 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/validate-network-edpm-deployment-openstack-edpm-ipam-4lhml" event={"ID":"2185de1e-5198-4f0f-8918-1e90aadb7fb7","Type":"ContainerDied","Data":"08785fa5a36ada22e4c767ec58a4dcd6765a2b3fa81f8dbc816bfbdd2125aa30"} Dec 13 07:39:25 crc kubenswrapper[4971]: I1213 07:39:25.572023 4971 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/validate-network-edpm-deployment-openstack-edpm-ipam-4lhml" Dec 13 07:39:25 crc kubenswrapper[4971]: I1213 07:39:25.817118 4971 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-r899w\" (UniqueName: \"kubernetes.io/projected/2185de1e-5198-4f0f-8918-1e90aadb7fb7-kube-api-access-r899w\") pod \"2185de1e-5198-4f0f-8918-1e90aadb7fb7\" (UID: \"2185de1e-5198-4f0f-8918-1e90aadb7fb7\") " Dec 13 07:39:25 crc kubenswrapper[4971]: I1213 07:39:25.817193 4971 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/2185de1e-5198-4f0f-8918-1e90aadb7fb7-ceph\") pod \"2185de1e-5198-4f0f-8918-1e90aadb7fb7\" (UID: \"2185de1e-5198-4f0f-8918-1e90aadb7fb7\") " Dec 13 07:39:25 crc kubenswrapper[4971]: I1213 07:39:25.817288 4971 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/2185de1e-5198-4f0f-8918-1e90aadb7fb7-inventory\") pod \"2185de1e-5198-4f0f-8918-1e90aadb7fb7\" (UID: \"2185de1e-5198-4f0f-8918-1e90aadb7fb7\") " Dec 13 07:39:25 crc kubenswrapper[4971]: I1213 07:39:25.817506 4971 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/2185de1e-5198-4f0f-8918-1e90aadb7fb7-ssh-key\") pod \"2185de1e-5198-4f0f-8918-1e90aadb7fb7\" (UID: \"2185de1e-5198-4f0f-8918-1e90aadb7fb7\") " Dec 13 07:39:25 crc kubenswrapper[4971]: I1213 07:39:25.830178 4971 scope.go:117] "RemoveContainer" containerID="09d0904c76ce78aa0b074447b1e64cf63160f573da086f934c8cb712615b71d4" Dec 13 07:39:25 crc kubenswrapper[4971]: E1213 07:39:25.830608 4971 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-82xjz_openshift-machine-config-operator(e0c4af12-edda-41d7-9b44-f87396174ef2)\"" pod="openshift-machine-config-operator/machine-config-daemon-82xjz" podUID="e0c4af12-edda-41d7-9b44-f87396174ef2" Dec 13 07:39:25 crc kubenswrapper[4971]: I1213 07:39:25.833625 4971 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/2185de1e-5198-4f0f-8918-1e90aadb7fb7-ceph" (OuterVolumeSpecName: "ceph") pod "2185de1e-5198-4f0f-8918-1e90aadb7fb7" (UID: "2185de1e-5198-4f0f-8918-1e90aadb7fb7"). InnerVolumeSpecName "ceph". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 13 07:39:25 crc kubenswrapper[4971]: I1213 07:39:25.837215 4971 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/2185de1e-5198-4f0f-8918-1e90aadb7fb7-kube-api-access-r899w" (OuterVolumeSpecName: "kube-api-access-r899w") pod "2185de1e-5198-4f0f-8918-1e90aadb7fb7" (UID: "2185de1e-5198-4f0f-8918-1e90aadb7fb7"). InnerVolumeSpecName "kube-api-access-r899w". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 13 07:39:25 crc kubenswrapper[4971]: I1213 07:39:25.862804 4971 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/2185de1e-5198-4f0f-8918-1e90aadb7fb7-inventory" (OuterVolumeSpecName: "inventory") pod "2185de1e-5198-4f0f-8918-1e90aadb7fb7" (UID: "2185de1e-5198-4f0f-8918-1e90aadb7fb7"). InnerVolumeSpecName "inventory". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 13 07:39:25 crc kubenswrapper[4971]: I1213 07:39:25.870380 4971 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/2185de1e-5198-4f0f-8918-1e90aadb7fb7-ssh-key" (OuterVolumeSpecName: "ssh-key") pod "2185de1e-5198-4f0f-8918-1e90aadb7fb7" (UID: "2185de1e-5198-4f0f-8918-1e90aadb7fb7"). InnerVolumeSpecName "ssh-key". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 13 07:39:25 crc kubenswrapper[4971]: I1213 07:39:25.921202 4971 reconciler_common.go:293] "Volume detached for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/2185de1e-5198-4f0f-8918-1e90aadb7fb7-ssh-key\") on node \"crc\" DevicePath \"\"" Dec 13 07:39:25 crc kubenswrapper[4971]: I1213 07:39:25.921231 4971 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-r899w\" (UniqueName: \"kubernetes.io/projected/2185de1e-5198-4f0f-8918-1e90aadb7fb7-kube-api-access-r899w\") on node \"crc\" DevicePath \"\"" Dec 13 07:39:25 crc kubenswrapper[4971]: I1213 07:39:25.921244 4971 reconciler_common.go:293] "Volume detached for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/2185de1e-5198-4f0f-8918-1e90aadb7fb7-ceph\") on node \"crc\" DevicePath \"\"" Dec 13 07:39:25 crc kubenswrapper[4971]: I1213 07:39:25.921252 4971 reconciler_common.go:293] "Volume detached for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/2185de1e-5198-4f0f-8918-1e90aadb7fb7-inventory\") on node \"crc\" DevicePath \"\"" Dec 13 07:39:26 crc kubenswrapper[4971]: I1213 07:39:26.175613 4971 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/validate-network-edpm-deployment-openstack-edpm-ipam-4lhml" event={"ID":"2185de1e-5198-4f0f-8918-1e90aadb7fb7","Type":"ContainerDied","Data":"a188d8a459fdd3c1cb43a6a649704f840396c4186accc263fc8552ee142cc4b2"} Dec 13 07:39:26 crc kubenswrapper[4971]: I1213 07:39:26.176005 4971 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="a188d8a459fdd3c1cb43a6a649704f840396c4186accc263fc8552ee142cc4b2" Dec 13 07:39:26 crc kubenswrapper[4971]: I1213 07:39:26.175707 4971 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/validate-network-edpm-deployment-openstack-edpm-ipam-4lhml" Dec 13 07:39:26 crc kubenswrapper[4971]: I1213 07:39:26.282301 4971 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/install-os-edpm-deployment-openstack-edpm-ipam-ffdjh"] Dec 13 07:39:26 crc kubenswrapper[4971]: E1213 07:39:26.282779 4971 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="2185de1e-5198-4f0f-8918-1e90aadb7fb7" containerName="validate-network-edpm-deployment-openstack-edpm-ipam" Dec 13 07:39:26 crc kubenswrapper[4971]: I1213 07:39:26.282801 4971 state_mem.go:107] "Deleted CPUSet assignment" podUID="2185de1e-5198-4f0f-8918-1e90aadb7fb7" containerName="validate-network-edpm-deployment-openstack-edpm-ipam" Dec 13 07:39:26 crc kubenswrapper[4971]: I1213 07:39:26.282974 4971 memory_manager.go:354] "RemoveStaleState removing state" podUID="2185de1e-5198-4f0f-8918-1e90aadb7fb7" containerName="validate-network-edpm-deployment-openstack-edpm-ipam" Dec 13 07:39:26 crc kubenswrapper[4971]: I1213 07:39:26.283624 4971 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/install-os-edpm-deployment-openstack-edpm-ipam-ffdjh" Dec 13 07:39:26 crc kubenswrapper[4971]: I1213 07:39:26.290115 4971 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceph-conf-files" Dec 13 07:39:26 crc kubenswrapper[4971]: I1213 07:39:26.290329 4971 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-edpm-ipam-dockercfg-zb2j5" Dec 13 07:39:26 crc kubenswrapper[4971]: I1213 07:39:26.290358 4971 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplanenodeset-openstack-edpm-ipam" Dec 13 07:39:26 crc kubenswrapper[4971]: I1213 07:39:26.290445 4971 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-aee-default-env" Dec 13 07:39:26 crc kubenswrapper[4971]: I1213 07:39:26.291186 4971 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplane-ansible-ssh-private-key-secret" Dec 13 07:39:26 crc kubenswrapper[4971]: I1213 07:39:26.291563 4971 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/install-os-edpm-deployment-openstack-edpm-ipam-ffdjh"] Dec 13 07:39:26 crc kubenswrapper[4971]: I1213 07:39:26.433143 4971 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-lwcqm\" (UniqueName: \"kubernetes.io/projected/b9a8435d-184f-4451-82cc-3b88e2e19050-kube-api-access-lwcqm\") pod \"install-os-edpm-deployment-openstack-edpm-ipam-ffdjh\" (UID: \"b9a8435d-184f-4451-82cc-3b88e2e19050\") " pod="openstack/install-os-edpm-deployment-openstack-edpm-ipam-ffdjh" Dec 13 07:39:26 crc kubenswrapper[4971]: I1213 07:39:26.433918 4971 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/b9a8435d-184f-4451-82cc-3b88e2e19050-inventory\") pod \"install-os-edpm-deployment-openstack-edpm-ipam-ffdjh\" (UID: \"b9a8435d-184f-4451-82cc-3b88e2e19050\") " pod="openstack/install-os-edpm-deployment-openstack-edpm-ipam-ffdjh" Dec 13 07:39:26 crc kubenswrapper[4971]: I1213 07:39:26.435499 4971 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/b9a8435d-184f-4451-82cc-3b88e2e19050-ssh-key\") pod \"install-os-edpm-deployment-openstack-edpm-ipam-ffdjh\" (UID: \"b9a8435d-184f-4451-82cc-3b88e2e19050\") " pod="openstack/install-os-edpm-deployment-openstack-edpm-ipam-ffdjh" Dec 13 07:39:26 crc kubenswrapper[4971]: I1213 07:39:26.435632 4971 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/b9a8435d-184f-4451-82cc-3b88e2e19050-ceph\") pod \"install-os-edpm-deployment-openstack-edpm-ipam-ffdjh\" (UID: \"b9a8435d-184f-4451-82cc-3b88e2e19050\") " pod="openstack/install-os-edpm-deployment-openstack-edpm-ipam-ffdjh" Dec 13 07:39:26 crc kubenswrapper[4971]: I1213 07:39:26.540838 4971 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-lwcqm\" (UniqueName: \"kubernetes.io/projected/b9a8435d-184f-4451-82cc-3b88e2e19050-kube-api-access-lwcqm\") pod \"install-os-edpm-deployment-openstack-edpm-ipam-ffdjh\" (UID: \"b9a8435d-184f-4451-82cc-3b88e2e19050\") " pod="openstack/install-os-edpm-deployment-openstack-edpm-ipam-ffdjh" Dec 13 07:39:26 crc kubenswrapper[4971]: I1213 07:39:26.540970 4971 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/b9a8435d-184f-4451-82cc-3b88e2e19050-inventory\") pod \"install-os-edpm-deployment-openstack-edpm-ipam-ffdjh\" (UID: \"b9a8435d-184f-4451-82cc-3b88e2e19050\") " pod="openstack/install-os-edpm-deployment-openstack-edpm-ipam-ffdjh" Dec 13 07:39:26 crc kubenswrapper[4971]: I1213 07:39:26.541016 4971 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/b9a8435d-184f-4451-82cc-3b88e2e19050-ssh-key\") pod \"install-os-edpm-deployment-openstack-edpm-ipam-ffdjh\" (UID: \"b9a8435d-184f-4451-82cc-3b88e2e19050\") " pod="openstack/install-os-edpm-deployment-openstack-edpm-ipam-ffdjh" Dec 13 07:39:26 crc kubenswrapper[4971]: I1213 07:39:26.541033 4971 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/b9a8435d-184f-4451-82cc-3b88e2e19050-ceph\") pod \"install-os-edpm-deployment-openstack-edpm-ipam-ffdjh\" (UID: \"b9a8435d-184f-4451-82cc-3b88e2e19050\") " pod="openstack/install-os-edpm-deployment-openstack-edpm-ipam-ffdjh" Dec 13 07:39:26 crc kubenswrapper[4971]: I1213 07:39:26.545319 4971 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/b9a8435d-184f-4451-82cc-3b88e2e19050-inventory\") pod \"install-os-edpm-deployment-openstack-edpm-ipam-ffdjh\" (UID: \"b9a8435d-184f-4451-82cc-3b88e2e19050\") " pod="openstack/install-os-edpm-deployment-openstack-edpm-ipam-ffdjh" Dec 13 07:39:26 crc kubenswrapper[4971]: I1213 07:39:26.545409 4971 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/b9a8435d-184f-4451-82cc-3b88e2e19050-ssh-key\") pod \"install-os-edpm-deployment-openstack-edpm-ipam-ffdjh\" (UID: \"b9a8435d-184f-4451-82cc-3b88e2e19050\") " pod="openstack/install-os-edpm-deployment-openstack-edpm-ipam-ffdjh" Dec 13 07:39:26 crc kubenswrapper[4971]: I1213 07:39:26.546457 4971 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/b9a8435d-184f-4451-82cc-3b88e2e19050-ceph\") pod \"install-os-edpm-deployment-openstack-edpm-ipam-ffdjh\" (UID: \"b9a8435d-184f-4451-82cc-3b88e2e19050\") " pod="openstack/install-os-edpm-deployment-openstack-edpm-ipam-ffdjh" Dec 13 07:39:26 crc kubenswrapper[4971]: I1213 07:39:26.562866 4971 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-lwcqm\" (UniqueName: \"kubernetes.io/projected/b9a8435d-184f-4451-82cc-3b88e2e19050-kube-api-access-lwcqm\") pod \"install-os-edpm-deployment-openstack-edpm-ipam-ffdjh\" (UID: \"b9a8435d-184f-4451-82cc-3b88e2e19050\") " pod="openstack/install-os-edpm-deployment-openstack-edpm-ipam-ffdjh" Dec 13 07:39:26 crc kubenswrapper[4971]: I1213 07:39:26.692273 4971 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/install-os-edpm-deployment-openstack-edpm-ipam-ffdjh" Dec 13 07:39:27 crc kubenswrapper[4971]: I1213 07:39:27.324256 4971 provider.go:102] Refreshing cache for provider: *credentialprovider.defaultDockerConfigProvider Dec 13 07:39:27 crc kubenswrapper[4971]: I1213 07:39:27.326067 4971 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/install-os-edpm-deployment-openstack-edpm-ipam-ffdjh"] Dec 13 07:39:28 crc kubenswrapper[4971]: I1213 07:39:28.194394 4971 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/install-os-edpm-deployment-openstack-edpm-ipam-ffdjh" event={"ID":"b9a8435d-184f-4451-82cc-3b88e2e19050","Type":"ContainerStarted","Data":"ef8d352a2632aa7791676ea05f1d628cb84e234c0e385f66cb3fd4fd2274e2c6"} Dec 13 07:39:29 crc kubenswrapper[4971]: I1213 07:39:29.204815 4971 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/install-os-edpm-deployment-openstack-edpm-ipam-ffdjh" event={"ID":"b9a8435d-184f-4451-82cc-3b88e2e19050","Type":"ContainerStarted","Data":"a8574d99b6c465309e27c4f9809190af275732c737e25295b9b5328f67bc849b"} Dec 13 07:39:29 crc kubenswrapper[4971]: I1213 07:39:29.220435 4971 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/install-os-edpm-deployment-openstack-edpm-ipam-ffdjh" podStartSLOduration=1.89882438 podStartE2EDuration="3.220416014s" podCreationTimestamp="2025-12-13 07:39:26 +0000 UTC" firstStartedPulling="2025-12-13 07:39:27.324042584 +0000 UTC m=+3023.928452032" lastFinishedPulling="2025-12-13 07:39:28.645634198 +0000 UTC m=+3025.250043666" observedRunningTime="2025-12-13 07:39:29.217277469 +0000 UTC m=+3025.821686907" watchObservedRunningTime="2025-12-13 07:39:29.220416014 +0000 UTC m=+3025.824825472" Dec 13 07:39:39 crc kubenswrapper[4971]: I1213 07:39:39.768310 4971 scope.go:117] "RemoveContainer" containerID="09d0904c76ce78aa0b074447b1e64cf63160f573da086f934c8cb712615b71d4" Dec 13 07:39:39 crc kubenswrapper[4971]: E1213 07:39:39.769110 4971 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-82xjz_openshift-machine-config-operator(e0c4af12-edda-41d7-9b44-f87396174ef2)\"" pod="openshift-machine-config-operator/machine-config-daemon-82xjz" podUID="e0c4af12-edda-41d7-9b44-f87396174ef2" Dec 13 07:39:54 crc kubenswrapper[4971]: I1213 07:39:54.769274 4971 scope.go:117] "RemoveContainer" containerID="09d0904c76ce78aa0b074447b1e64cf63160f573da086f934c8cb712615b71d4" Dec 13 07:39:54 crc kubenswrapper[4971]: E1213 07:39:54.770136 4971 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-82xjz_openshift-machine-config-operator(e0c4af12-edda-41d7-9b44-f87396174ef2)\"" pod="openshift-machine-config-operator/machine-config-daemon-82xjz" podUID="e0c4af12-edda-41d7-9b44-f87396174ef2" Dec 13 07:40:07 crc kubenswrapper[4971]: I1213 07:40:07.768891 4971 scope.go:117] "RemoveContainer" containerID="09d0904c76ce78aa0b074447b1e64cf63160f573da086f934c8cb712615b71d4" Dec 13 07:40:07 crc kubenswrapper[4971]: E1213 07:40:07.771025 4971 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-82xjz_openshift-machine-config-operator(e0c4af12-edda-41d7-9b44-f87396174ef2)\"" pod="openshift-machine-config-operator/machine-config-daemon-82xjz" podUID="e0c4af12-edda-41d7-9b44-f87396174ef2" Dec 13 07:40:11 crc kubenswrapper[4971]: I1213 07:40:11.554365 4971 generic.go:334] "Generic (PLEG): container finished" podID="b9a8435d-184f-4451-82cc-3b88e2e19050" containerID="a8574d99b6c465309e27c4f9809190af275732c737e25295b9b5328f67bc849b" exitCode=0 Dec 13 07:40:11 crc kubenswrapper[4971]: I1213 07:40:11.554510 4971 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/install-os-edpm-deployment-openstack-edpm-ipam-ffdjh" event={"ID":"b9a8435d-184f-4451-82cc-3b88e2e19050","Type":"ContainerDied","Data":"a8574d99b6c465309e27c4f9809190af275732c737e25295b9b5328f67bc849b"} Dec 13 07:40:13 crc kubenswrapper[4971]: I1213 07:40:13.120444 4971 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/install-os-edpm-deployment-openstack-edpm-ipam-ffdjh" Dec 13 07:40:13 crc kubenswrapper[4971]: I1213 07:40:13.203703 4971 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-lwcqm\" (UniqueName: \"kubernetes.io/projected/b9a8435d-184f-4451-82cc-3b88e2e19050-kube-api-access-lwcqm\") pod \"b9a8435d-184f-4451-82cc-3b88e2e19050\" (UID: \"b9a8435d-184f-4451-82cc-3b88e2e19050\") " Dec 13 07:40:13 crc kubenswrapper[4971]: I1213 07:40:13.203816 4971 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/b9a8435d-184f-4451-82cc-3b88e2e19050-ceph\") pod \"b9a8435d-184f-4451-82cc-3b88e2e19050\" (UID: \"b9a8435d-184f-4451-82cc-3b88e2e19050\") " Dec 13 07:40:13 crc kubenswrapper[4971]: I1213 07:40:13.203911 4971 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/b9a8435d-184f-4451-82cc-3b88e2e19050-ssh-key\") pod \"b9a8435d-184f-4451-82cc-3b88e2e19050\" (UID: \"b9a8435d-184f-4451-82cc-3b88e2e19050\") " Dec 13 07:40:13 crc kubenswrapper[4971]: I1213 07:40:13.203967 4971 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/b9a8435d-184f-4451-82cc-3b88e2e19050-inventory\") pod \"b9a8435d-184f-4451-82cc-3b88e2e19050\" (UID: \"b9a8435d-184f-4451-82cc-3b88e2e19050\") " Dec 13 07:40:13 crc kubenswrapper[4971]: I1213 07:40:13.210825 4971 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/b9a8435d-184f-4451-82cc-3b88e2e19050-ceph" (OuterVolumeSpecName: "ceph") pod "b9a8435d-184f-4451-82cc-3b88e2e19050" (UID: "b9a8435d-184f-4451-82cc-3b88e2e19050"). InnerVolumeSpecName "ceph". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 13 07:40:13 crc kubenswrapper[4971]: I1213 07:40:13.218812 4971 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/b9a8435d-184f-4451-82cc-3b88e2e19050-kube-api-access-lwcqm" (OuterVolumeSpecName: "kube-api-access-lwcqm") pod "b9a8435d-184f-4451-82cc-3b88e2e19050" (UID: "b9a8435d-184f-4451-82cc-3b88e2e19050"). InnerVolumeSpecName "kube-api-access-lwcqm". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 13 07:40:13 crc kubenswrapper[4971]: I1213 07:40:13.238547 4971 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/b9a8435d-184f-4451-82cc-3b88e2e19050-ssh-key" (OuterVolumeSpecName: "ssh-key") pod "b9a8435d-184f-4451-82cc-3b88e2e19050" (UID: "b9a8435d-184f-4451-82cc-3b88e2e19050"). InnerVolumeSpecName "ssh-key". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 13 07:40:13 crc kubenswrapper[4971]: I1213 07:40:13.240193 4971 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/b9a8435d-184f-4451-82cc-3b88e2e19050-inventory" (OuterVolumeSpecName: "inventory") pod "b9a8435d-184f-4451-82cc-3b88e2e19050" (UID: "b9a8435d-184f-4451-82cc-3b88e2e19050"). InnerVolumeSpecName "inventory". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 13 07:40:13 crc kubenswrapper[4971]: I1213 07:40:13.305981 4971 reconciler_common.go:293] "Volume detached for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/b9a8435d-184f-4451-82cc-3b88e2e19050-ssh-key\") on node \"crc\" DevicePath \"\"" Dec 13 07:40:13 crc kubenswrapper[4971]: I1213 07:40:13.306022 4971 reconciler_common.go:293] "Volume detached for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/b9a8435d-184f-4451-82cc-3b88e2e19050-inventory\") on node \"crc\" DevicePath \"\"" Dec 13 07:40:13 crc kubenswrapper[4971]: I1213 07:40:13.306040 4971 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-lwcqm\" (UniqueName: \"kubernetes.io/projected/b9a8435d-184f-4451-82cc-3b88e2e19050-kube-api-access-lwcqm\") on node \"crc\" DevicePath \"\"" Dec 13 07:40:13 crc kubenswrapper[4971]: I1213 07:40:13.306054 4971 reconciler_common.go:293] "Volume detached for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/b9a8435d-184f-4451-82cc-3b88e2e19050-ceph\") on node \"crc\" DevicePath \"\"" Dec 13 07:40:13 crc kubenswrapper[4971]: I1213 07:40:13.578719 4971 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/install-os-edpm-deployment-openstack-edpm-ipam-ffdjh" event={"ID":"b9a8435d-184f-4451-82cc-3b88e2e19050","Type":"ContainerDied","Data":"ef8d352a2632aa7791676ea05f1d628cb84e234c0e385f66cb3fd4fd2274e2c6"} Dec 13 07:40:13 crc kubenswrapper[4971]: I1213 07:40:13.579064 4971 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="ef8d352a2632aa7791676ea05f1d628cb84e234c0e385f66cb3fd4fd2274e2c6" Dec 13 07:40:13 crc kubenswrapper[4971]: I1213 07:40:13.578784 4971 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/install-os-edpm-deployment-openstack-edpm-ipam-ffdjh" Dec 13 07:40:13 crc kubenswrapper[4971]: I1213 07:40:13.694458 4971 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ceph-hci-pre-edpm-deployment-openstack-edpm-ipam-4fh5s"] Dec 13 07:40:13 crc kubenswrapper[4971]: E1213 07:40:13.694918 4971 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="b9a8435d-184f-4451-82cc-3b88e2e19050" containerName="install-os-edpm-deployment-openstack-edpm-ipam" Dec 13 07:40:13 crc kubenswrapper[4971]: I1213 07:40:13.694949 4971 state_mem.go:107] "Deleted CPUSet assignment" podUID="b9a8435d-184f-4451-82cc-3b88e2e19050" containerName="install-os-edpm-deployment-openstack-edpm-ipam" Dec 13 07:40:13 crc kubenswrapper[4971]: I1213 07:40:13.695212 4971 memory_manager.go:354] "RemoveStaleState removing state" podUID="b9a8435d-184f-4451-82cc-3b88e2e19050" containerName="install-os-edpm-deployment-openstack-edpm-ipam" Dec 13 07:40:13 crc kubenswrapper[4971]: I1213 07:40:13.696064 4971 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceph-hci-pre-edpm-deployment-openstack-edpm-ipam-4fh5s" Dec 13 07:40:13 crc kubenswrapper[4971]: I1213 07:40:13.697857 4971 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-edpm-ipam-dockercfg-zb2j5" Dec 13 07:40:13 crc kubenswrapper[4971]: I1213 07:40:13.698221 4971 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplane-ansible-ssh-private-key-secret" Dec 13 07:40:13 crc kubenswrapper[4971]: I1213 07:40:13.698456 4971 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-aee-default-env" Dec 13 07:40:13 crc kubenswrapper[4971]: I1213 07:40:13.698609 4971 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceph-conf-files" Dec 13 07:40:13 crc kubenswrapper[4971]: I1213 07:40:13.698729 4971 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplanenodeset-openstack-edpm-ipam" Dec 13 07:40:13 crc kubenswrapper[4971]: I1213 07:40:13.709469 4971 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ceph-hci-pre-edpm-deployment-openstack-edpm-ipam-4fh5s"] Dec 13 07:40:13 crc kubenswrapper[4971]: I1213 07:40:13.815011 4971 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-2fpbb\" (UniqueName: \"kubernetes.io/projected/bbbc7fc6-2323-4765-9160-a1c085de09d8-kube-api-access-2fpbb\") pod \"ceph-hci-pre-edpm-deployment-openstack-edpm-ipam-4fh5s\" (UID: \"bbbc7fc6-2323-4765-9160-a1c085de09d8\") " pod="openstack/ceph-hci-pre-edpm-deployment-openstack-edpm-ipam-4fh5s" Dec 13 07:40:13 crc kubenswrapper[4971]: I1213 07:40:13.815321 4971 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/bbbc7fc6-2323-4765-9160-a1c085de09d8-ceph\") pod \"ceph-hci-pre-edpm-deployment-openstack-edpm-ipam-4fh5s\" (UID: \"bbbc7fc6-2323-4765-9160-a1c085de09d8\") " pod="openstack/ceph-hci-pre-edpm-deployment-openstack-edpm-ipam-4fh5s" Dec 13 07:40:13 crc kubenswrapper[4971]: I1213 07:40:13.815377 4971 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/bbbc7fc6-2323-4765-9160-a1c085de09d8-ssh-key\") pod \"ceph-hci-pre-edpm-deployment-openstack-edpm-ipam-4fh5s\" (UID: \"bbbc7fc6-2323-4765-9160-a1c085de09d8\") " pod="openstack/ceph-hci-pre-edpm-deployment-openstack-edpm-ipam-4fh5s" Dec 13 07:40:13 crc kubenswrapper[4971]: I1213 07:40:13.815551 4971 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/bbbc7fc6-2323-4765-9160-a1c085de09d8-inventory\") pod \"ceph-hci-pre-edpm-deployment-openstack-edpm-ipam-4fh5s\" (UID: \"bbbc7fc6-2323-4765-9160-a1c085de09d8\") " pod="openstack/ceph-hci-pre-edpm-deployment-openstack-edpm-ipam-4fh5s" Dec 13 07:40:13 crc kubenswrapper[4971]: I1213 07:40:13.917898 4971 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-2fpbb\" (UniqueName: \"kubernetes.io/projected/bbbc7fc6-2323-4765-9160-a1c085de09d8-kube-api-access-2fpbb\") pod \"ceph-hci-pre-edpm-deployment-openstack-edpm-ipam-4fh5s\" (UID: \"bbbc7fc6-2323-4765-9160-a1c085de09d8\") " pod="openstack/ceph-hci-pre-edpm-deployment-openstack-edpm-ipam-4fh5s" Dec 13 07:40:13 crc kubenswrapper[4971]: I1213 07:40:13.918112 4971 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/bbbc7fc6-2323-4765-9160-a1c085de09d8-ceph\") pod \"ceph-hci-pre-edpm-deployment-openstack-edpm-ipam-4fh5s\" (UID: \"bbbc7fc6-2323-4765-9160-a1c085de09d8\") " pod="openstack/ceph-hci-pre-edpm-deployment-openstack-edpm-ipam-4fh5s" Dec 13 07:40:13 crc kubenswrapper[4971]: I1213 07:40:13.918163 4971 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/bbbc7fc6-2323-4765-9160-a1c085de09d8-ssh-key\") pod \"ceph-hci-pre-edpm-deployment-openstack-edpm-ipam-4fh5s\" (UID: \"bbbc7fc6-2323-4765-9160-a1c085de09d8\") " pod="openstack/ceph-hci-pre-edpm-deployment-openstack-edpm-ipam-4fh5s" Dec 13 07:40:13 crc kubenswrapper[4971]: I1213 07:40:13.918262 4971 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/bbbc7fc6-2323-4765-9160-a1c085de09d8-inventory\") pod \"ceph-hci-pre-edpm-deployment-openstack-edpm-ipam-4fh5s\" (UID: \"bbbc7fc6-2323-4765-9160-a1c085de09d8\") " pod="openstack/ceph-hci-pre-edpm-deployment-openstack-edpm-ipam-4fh5s" Dec 13 07:40:13 crc kubenswrapper[4971]: I1213 07:40:13.925129 4971 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/bbbc7fc6-2323-4765-9160-a1c085de09d8-ceph\") pod \"ceph-hci-pre-edpm-deployment-openstack-edpm-ipam-4fh5s\" (UID: \"bbbc7fc6-2323-4765-9160-a1c085de09d8\") " pod="openstack/ceph-hci-pre-edpm-deployment-openstack-edpm-ipam-4fh5s" Dec 13 07:40:13 crc kubenswrapper[4971]: I1213 07:40:13.926674 4971 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/bbbc7fc6-2323-4765-9160-a1c085de09d8-ssh-key\") pod \"ceph-hci-pre-edpm-deployment-openstack-edpm-ipam-4fh5s\" (UID: \"bbbc7fc6-2323-4765-9160-a1c085de09d8\") " pod="openstack/ceph-hci-pre-edpm-deployment-openstack-edpm-ipam-4fh5s" Dec 13 07:40:13 crc kubenswrapper[4971]: I1213 07:40:13.934508 4971 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/bbbc7fc6-2323-4765-9160-a1c085de09d8-inventory\") pod \"ceph-hci-pre-edpm-deployment-openstack-edpm-ipam-4fh5s\" (UID: \"bbbc7fc6-2323-4765-9160-a1c085de09d8\") " pod="openstack/ceph-hci-pre-edpm-deployment-openstack-edpm-ipam-4fh5s" Dec 13 07:40:13 crc kubenswrapper[4971]: I1213 07:40:13.937112 4971 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-2fpbb\" (UniqueName: \"kubernetes.io/projected/bbbc7fc6-2323-4765-9160-a1c085de09d8-kube-api-access-2fpbb\") pod \"ceph-hci-pre-edpm-deployment-openstack-edpm-ipam-4fh5s\" (UID: \"bbbc7fc6-2323-4765-9160-a1c085de09d8\") " pod="openstack/ceph-hci-pre-edpm-deployment-openstack-edpm-ipam-4fh5s" Dec 13 07:40:14 crc kubenswrapper[4971]: I1213 07:40:14.014418 4971 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceph-hci-pre-edpm-deployment-openstack-edpm-ipam-4fh5s" Dec 13 07:40:14 crc kubenswrapper[4971]: I1213 07:40:14.597462 4971 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ceph-hci-pre-edpm-deployment-openstack-edpm-ipam-4fh5s"] Dec 13 07:40:14 crc kubenswrapper[4971]: W1213 07:40:14.608717 4971 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podbbbc7fc6_2323_4765_9160_a1c085de09d8.slice/crio-df4e7470f1f8199ce398550c5a35f1a88e7812fd0a74e323000426adb40f66bc WatchSource:0}: Error finding container df4e7470f1f8199ce398550c5a35f1a88e7812fd0a74e323000426adb40f66bc: Status 404 returned error can't find the container with id df4e7470f1f8199ce398550c5a35f1a88e7812fd0a74e323000426adb40f66bc Dec 13 07:40:15 crc kubenswrapper[4971]: I1213 07:40:15.595055 4971 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceph-hci-pre-edpm-deployment-openstack-edpm-ipam-4fh5s" event={"ID":"bbbc7fc6-2323-4765-9160-a1c085de09d8","Type":"ContainerStarted","Data":"df4e7470f1f8199ce398550c5a35f1a88e7812fd0a74e323000426adb40f66bc"} Dec 13 07:40:17 crc kubenswrapper[4971]: I1213 07:40:17.608873 4971 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceph-hci-pre-edpm-deployment-openstack-edpm-ipam-4fh5s" event={"ID":"bbbc7fc6-2323-4765-9160-a1c085de09d8","Type":"ContainerStarted","Data":"5c133a2beb28b323047345c8af9937d617594894485c1454e26761cc22830b9d"} Dec 13 07:40:17 crc kubenswrapper[4971]: I1213 07:40:17.633577 4971 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/ceph-hci-pre-edpm-deployment-openstack-edpm-ipam-4fh5s" podStartSLOduration=2.99416223 podStartE2EDuration="4.633556482s" podCreationTimestamp="2025-12-13 07:40:13 +0000 UTC" firstStartedPulling="2025-12-13 07:40:14.612068995 +0000 UTC m=+3071.216478443" lastFinishedPulling="2025-12-13 07:40:16.251463247 +0000 UTC m=+3072.855872695" observedRunningTime="2025-12-13 07:40:17.625642322 +0000 UTC m=+3074.230051790" watchObservedRunningTime="2025-12-13 07:40:17.633556482 +0000 UTC m=+3074.237965950" Dec 13 07:40:18 crc kubenswrapper[4971]: I1213 07:40:18.768820 4971 scope.go:117] "RemoveContainer" containerID="09d0904c76ce78aa0b074447b1e64cf63160f573da086f934c8cb712615b71d4" Dec 13 07:40:18 crc kubenswrapper[4971]: E1213 07:40:18.769309 4971 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-82xjz_openshift-machine-config-operator(e0c4af12-edda-41d7-9b44-f87396174ef2)\"" pod="openshift-machine-config-operator/machine-config-daemon-82xjz" podUID="e0c4af12-edda-41d7-9b44-f87396174ef2" Dec 13 07:40:24 crc kubenswrapper[4971]: I1213 07:40:24.682101 4971 generic.go:334] "Generic (PLEG): container finished" podID="bbbc7fc6-2323-4765-9160-a1c085de09d8" containerID="5c133a2beb28b323047345c8af9937d617594894485c1454e26761cc22830b9d" exitCode=0 Dec 13 07:40:24 crc kubenswrapper[4971]: I1213 07:40:24.682198 4971 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceph-hci-pre-edpm-deployment-openstack-edpm-ipam-4fh5s" event={"ID":"bbbc7fc6-2323-4765-9160-a1c085de09d8","Type":"ContainerDied","Data":"5c133a2beb28b323047345c8af9937d617594894485c1454e26761cc22830b9d"} Dec 13 07:40:26 crc kubenswrapper[4971]: I1213 07:40:26.076978 4971 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ceph-hci-pre-edpm-deployment-openstack-edpm-ipam-4fh5s" Dec 13 07:40:26 crc kubenswrapper[4971]: I1213 07:40:26.185659 4971 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-2fpbb\" (UniqueName: \"kubernetes.io/projected/bbbc7fc6-2323-4765-9160-a1c085de09d8-kube-api-access-2fpbb\") pod \"bbbc7fc6-2323-4765-9160-a1c085de09d8\" (UID: \"bbbc7fc6-2323-4765-9160-a1c085de09d8\") " Dec 13 07:40:26 crc kubenswrapper[4971]: I1213 07:40:26.185846 4971 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/bbbc7fc6-2323-4765-9160-a1c085de09d8-inventory\") pod \"bbbc7fc6-2323-4765-9160-a1c085de09d8\" (UID: \"bbbc7fc6-2323-4765-9160-a1c085de09d8\") " Dec 13 07:40:26 crc kubenswrapper[4971]: I1213 07:40:26.185883 4971 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/bbbc7fc6-2323-4765-9160-a1c085de09d8-ceph\") pod \"bbbc7fc6-2323-4765-9160-a1c085de09d8\" (UID: \"bbbc7fc6-2323-4765-9160-a1c085de09d8\") " Dec 13 07:40:26 crc kubenswrapper[4971]: I1213 07:40:26.185952 4971 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/bbbc7fc6-2323-4765-9160-a1c085de09d8-ssh-key\") pod \"bbbc7fc6-2323-4765-9160-a1c085de09d8\" (UID: \"bbbc7fc6-2323-4765-9160-a1c085de09d8\") " Dec 13 07:40:26 crc kubenswrapper[4971]: I1213 07:40:26.262175 4971 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/bbbc7fc6-2323-4765-9160-a1c085de09d8-ceph" (OuterVolumeSpecName: "ceph") pod "bbbc7fc6-2323-4765-9160-a1c085de09d8" (UID: "bbbc7fc6-2323-4765-9160-a1c085de09d8"). InnerVolumeSpecName "ceph". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 13 07:40:26 crc kubenswrapper[4971]: I1213 07:40:26.268728 4971 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/bbbc7fc6-2323-4765-9160-a1c085de09d8-inventory" (OuterVolumeSpecName: "inventory") pod "bbbc7fc6-2323-4765-9160-a1c085de09d8" (UID: "bbbc7fc6-2323-4765-9160-a1c085de09d8"). InnerVolumeSpecName "inventory". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 13 07:40:26 crc kubenswrapper[4971]: I1213 07:40:26.272819 4971 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/bbbc7fc6-2323-4765-9160-a1c085de09d8-kube-api-access-2fpbb" (OuterVolumeSpecName: "kube-api-access-2fpbb") pod "bbbc7fc6-2323-4765-9160-a1c085de09d8" (UID: "bbbc7fc6-2323-4765-9160-a1c085de09d8"). InnerVolumeSpecName "kube-api-access-2fpbb". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 13 07:40:26 crc kubenswrapper[4971]: I1213 07:40:26.290171 4971 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-2fpbb\" (UniqueName: \"kubernetes.io/projected/bbbc7fc6-2323-4765-9160-a1c085de09d8-kube-api-access-2fpbb\") on node \"crc\" DevicePath \"\"" Dec 13 07:40:26 crc kubenswrapper[4971]: I1213 07:40:26.290218 4971 reconciler_common.go:293] "Volume detached for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/bbbc7fc6-2323-4765-9160-a1c085de09d8-inventory\") on node \"crc\" DevicePath \"\"" Dec 13 07:40:26 crc kubenswrapper[4971]: I1213 07:40:26.290229 4971 reconciler_common.go:293] "Volume detached for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/bbbc7fc6-2323-4765-9160-a1c085de09d8-ceph\") on node \"crc\" DevicePath \"\"" Dec 13 07:40:26 crc kubenswrapper[4971]: I1213 07:40:26.294907 4971 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/bbbc7fc6-2323-4765-9160-a1c085de09d8-ssh-key" (OuterVolumeSpecName: "ssh-key") pod "bbbc7fc6-2323-4765-9160-a1c085de09d8" (UID: "bbbc7fc6-2323-4765-9160-a1c085de09d8"). InnerVolumeSpecName "ssh-key". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 13 07:40:26 crc kubenswrapper[4971]: I1213 07:40:26.392405 4971 reconciler_common.go:293] "Volume detached for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/bbbc7fc6-2323-4765-9160-a1c085de09d8-ssh-key\") on node \"crc\" DevicePath \"\"" Dec 13 07:40:26 crc kubenswrapper[4971]: I1213 07:40:26.704008 4971 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceph-hci-pre-edpm-deployment-openstack-edpm-ipam-4fh5s" event={"ID":"bbbc7fc6-2323-4765-9160-a1c085de09d8","Type":"ContainerDied","Data":"df4e7470f1f8199ce398550c5a35f1a88e7812fd0a74e323000426adb40f66bc"} Dec 13 07:40:26 crc kubenswrapper[4971]: I1213 07:40:26.704060 4971 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="df4e7470f1f8199ce398550c5a35f1a88e7812fd0a74e323000426adb40f66bc" Dec 13 07:40:26 crc kubenswrapper[4971]: I1213 07:40:26.704130 4971 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ceph-hci-pre-edpm-deployment-openstack-edpm-ipam-4fh5s" Dec 13 07:40:26 crc kubenswrapper[4971]: I1213 07:40:26.862602 4971 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/configure-os-edpm-deployment-openstack-edpm-ipam-7qg86"] Dec 13 07:40:26 crc kubenswrapper[4971]: E1213 07:40:26.863037 4971 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="bbbc7fc6-2323-4765-9160-a1c085de09d8" containerName="ceph-hci-pre-edpm-deployment-openstack-edpm-ipam" Dec 13 07:40:26 crc kubenswrapper[4971]: I1213 07:40:26.863055 4971 state_mem.go:107] "Deleted CPUSet assignment" podUID="bbbc7fc6-2323-4765-9160-a1c085de09d8" containerName="ceph-hci-pre-edpm-deployment-openstack-edpm-ipam" Dec 13 07:40:26 crc kubenswrapper[4971]: I1213 07:40:26.863238 4971 memory_manager.go:354] "RemoveStaleState removing state" podUID="bbbc7fc6-2323-4765-9160-a1c085de09d8" containerName="ceph-hci-pre-edpm-deployment-openstack-edpm-ipam" Dec 13 07:40:26 crc kubenswrapper[4971]: I1213 07:40:26.863832 4971 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-7qg86" Dec 13 07:40:26 crc kubenswrapper[4971]: I1213 07:40:26.878316 4971 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplane-ansible-ssh-private-key-secret" Dec 13 07:40:26 crc kubenswrapper[4971]: I1213 07:40:26.878597 4971 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceph-conf-files" Dec 13 07:40:26 crc kubenswrapper[4971]: I1213 07:40:26.878816 4971 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-aee-default-env" Dec 13 07:40:26 crc kubenswrapper[4971]: I1213 07:40:26.879023 4971 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-edpm-ipam-dockercfg-zb2j5" Dec 13 07:40:26 crc kubenswrapper[4971]: I1213 07:40:26.879164 4971 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplanenodeset-openstack-edpm-ipam" Dec 13 07:40:26 crc kubenswrapper[4971]: I1213 07:40:26.886208 4971 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/configure-os-edpm-deployment-openstack-edpm-ipam-7qg86"] Dec 13 07:40:27 crc kubenswrapper[4971]: I1213 07:40:27.009843 4971 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-w2jfs\" (UniqueName: \"kubernetes.io/projected/b03f6448-21d3-417f-a1a9-6e7c733550ec-kube-api-access-w2jfs\") pod \"configure-os-edpm-deployment-openstack-edpm-ipam-7qg86\" (UID: \"b03f6448-21d3-417f-a1a9-6e7c733550ec\") " pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-7qg86" Dec 13 07:40:27 crc kubenswrapper[4971]: I1213 07:40:27.009911 4971 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/b03f6448-21d3-417f-a1a9-6e7c733550ec-ssh-key\") pod \"configure-os-edpm-deployment-openstack-edpm-ipam-7qg86\" (UID: \"b03f6448-21d3-417f-a1a9-6e7c733550ec\") " pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-7qg86" Dec 13 07:40:27 crc kubenswrapper[4971]: I1213 07:40:27.010050 4971 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/b03f6448-21d3-417f-a1a9-6e7c733550ec-inventory\") pod \"configure-os-edpm-deployment-openstack-edpm-ipam-7qg86\" (UID: \"b03f6448-21d3-417f-a1a9-6e7c733550ec\") " pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-7qg86" Dec 13 07:40:27 crc kubenswrapper[4971]: I1213 07:40:27.010369 4971 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/b03f6448-21d3-417f-a1a9-6e7c733550ec-ceph\") pod \"configure-os-edpm-deployment-openstack-edpm-ipam-7qg86\" (UID: \"b03f6448-21d3-417f-a1a9-6e7c733550ec\") " pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-7qg86" Dec 13 07:40:27 crc kubenswrapper[4971]: I1213 07:40:27.112216 4971 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/b03f6448-21d3-417f-a1a9-6e7c733550ec-ceph\") pod \"configure-os-edpm-deployment-openstack-edpm-ipam-7qg86\" (UID: \"b03f6448-21d3-417f-a1a9-6e7c733550ec\") " pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-7qg86" Dec 13 07:40:27 crc kubenswrapper[4971]: I1213 07:40:27.112283 4971 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-w2jfs\" (UniqueName: \"kubernetes.io/projected/b03f6448-21d3-417f-a1a9-6e7c733550ec-kube-api-access-w2jfs\") pod \"configure-os-edpm-deployment-openstack-edpm-ipam-7qg86\" (UID: \"b03f6448-21d3-417f-a1a9-6e7c733550ec\") " pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-7qg86" Dec 13 07:40:27 crc kubenswrapper[4971]: I1213 07:40:27.112323 4971 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/b03f6448-21d3-417f-a1a9-6e7c733550ec-ssh-key\") pod \"configure-os-edpm-deployment-openstack-edpm-ipam-7qg86\" (UID: \"b03f6448-21d3-417f-a1a9-6e7c733550ec\") " pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-7qg86" Dec 13 07:40:27 crc kubenswrapper[4971]: I1213 07:40:27.112359 4971 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/b03f6448-21d3-417f-a1a9-6e7c733550ec-inventory\") pod \"configure-os-edpm-deployment-openstack-edpm-ipam-7qg86\" (UID: \"b03f6448-21d3-417f-a1a9-6e7c733550ec\") " pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-7qg86" Dec 13 07:40:27 crc kubenswrapper[4971]: I1213 07:40:27.119001 4971 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/b03f6448-21d3-417f-a1a9-6e7c733550ec-ceph\") pod \"configure-os-edpm-deployment-openstack-edpm-ipam-7qg86\" (UID: \"b03f6448-21d3-417f-a1a9-6e7c733550ec\") " pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-7qg86" Dec 13 07:40:27 crc kubenswrapper[4971]: I1213 07:40:27.119922 4971 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/b03f6448-21d3-417f-a1a9-6e7c733550ec-inventory\") pod \"configure-os-edpm-deployment-openstack-edpm-ipam-7qg86\" (UID: \"b03f6448-21d3-417f-a1a9-6e7c733550ec\") " pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-7qg86" Dec 13 07:40:27 crc kubenswrapper[4971]: I1213 07:40:27.125969 4971 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/b03f6448-21d3-417f-a1a9-6e7c733550ec-ssh-key\") pod \"configure-os-edpm-deployment-openstack-edpm-ipam-7qg86\" (UID: \"b03f6448-21d3-417f-a1a9-6e7c733550ec\") " pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-7qg86" Dec 13 07:40:27 crc kubenswrapper[4971]: I1213 07:40:27.134680 4971 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-w2jfs\" (UniqueName: \"kubernetes.io/projected/b03f6448-21d3-417f-a1a9-6e7c733550ec-kube-api-access-w2jfs\") pod \"configure-os-edpm-deployment-openstack-edpm-ipam-7qg86\" (UID: \"b03f6448-21d3-417f-a1a9-6e7c733550ec\") " pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-7qg86" Dec 13 07:40:27 crc kubenswrapper[4971]: I1213 07:40:27.182775 4971 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-7qg86" Dec 13 07:40:27 crc kubenswrapper[4971]: I1213 07:40:27.802126 4971 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/configure-os-edpm-deployment-openstack-edpm-ipam-7qg86"] Dec 13 07:40:27 crc kubenswrapper[4971]: W1213 07:40:27.807685 4971 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podb03f6448_21d3_417f_a1a9_6e7c733550ec.slice/crio-abffdfc69b129bfccfa9ab67ef64d13d4b1c972fca8806a723f544b69002ae96 WatchSource:0}: Error finding container abffdfc69b129bfccfa9ab67ef64d13d4b1c972fca8806a723f544b69002ae96: Status 404 returned error can't find the container with id abffdfc69b129bfccfa9ab67ef64d13d4b1c972fca8806a723f544b69002ae96 Dec 13 07:40:28 crc kubenswrapper[4971]: I1213 07:40:28.722336 4971 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-7qg86" event={"ID":"b03f6448-21d3-417f-a1a9-6e7c733550ec","Type":"ContainerStarted","Data":"abffdfc69b129bfccfa9ab67ef64d13d4b1c972fca8806a723f544b69002ae96"} Dec 13 07:40:30 crc kubenswrapper[4971]: I1213 07:40:30.738739 4971 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-7qg86" event={"ID":"b03f6448-21d3-417f-a1a9-6e7c733550ec","Type":"ContainerStarted","Data":"fae89683ea6cfaaa107d7dbadbc23c229598502b238c7837a8ee6451c38d906d"} Dec 13 07:40:30 crc kubenswrapper[4971]: I1213 07:40:30.769504 4971 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-7qg86" podStartSLOduration=3.234550672 podStartE2EDuration="4.769487426s" podCreationTimestamp="2025-12-13 07:40:26 +0000 UTC" firstStartedPulling="2025-12-13 07:40:27.80972362 +0000 UTC m=+3084.414133078" lastFinishedPulling="2025-12-13 07:40:29.344660344 +0000 UTC m=+3085.949069832" observedRunningTime="2025-12-13 07:40:30.764429024 +0000 UTC m=+3087.368838482" watchObservedRunningTime="2025-12-13 07:40:30.769487426 +0000 UTC m=+3087.373896874" Dec 13 07:40:32 crc kubenswrapper[4971]: I1213 07:40:32.786039 4971 scope.go:117] "RemoveContainer" containerID="09d0904c76ce78aa0b074447b1e64cf63160f573da086f934c8cb712615b71d4" Dec 13 07:40:32 crc kubenswrapper[4971]: E1213 07:40:32.786672 4971 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-82xjz_openshift-machine-config-operator(e0c4af12-edda-41d7-9b44-f87396174ef2)\"" pod="openshift-machine-config-operator/machine-config-daemon-82xjz" podUID="e0c4af12-edda-41d7-9b44-f87396174ef2" Dec 13 07:40:46 crc kubenswrapper[4971]: I1213 07:40:46.742023 4971 prober.go:107] "Probe failed" probeType="Liveness" pod="openstack/ceilometer-0" podUID="7cfb4dbf-cc55-48b7-b50c-65d6d4ae64ba" containerName="ceilometer-central-agent" probeResult="failure" output="command timed out" Dec 13 07:40:47 crc kubenswrapper[4971]: I1213 07:40:47.769372 4971 scope.go:117] "RemoveContainer" containerID="09d0904c76ce78aa0b074447b1e64cf63160f573da086f934c8cb712615b71d4" Dec 13 07:40:47 crc kubenswrapper[4971]: E1213 07:40:47.769843 4971 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-82xjz_openshift-machine-config-operator(e0c4af12-edda-41d7-9b44-f87396174ef2)\"" pod="openshift-machine-config-operator/machine-config-daemon-82xjz" podUID="e0c4af12-edda-41d7-9b44-f87396174ef2" Dec 13 07:40:51 crc kubenswrapper[4971]: I1213 07:40:51.740331 4971 prober.go:107] "Probe failed" probeType="Liveness" pod="openstack/ceilometer-0" podUID="7cfb4dbf-cc55-48b7-b50c-65d6d4ae64ba" containerName="ceilometer-central-agent" probeResult="failure" output="command timed out" Dec 13 07:40:53 crc kubenswrapper[4971]: I1213 07:40:53.870037 4971 prober.go:107] "Probe failed" probeType="Liveness" pod="openstack/ceilometer-0" podUID="7cfb4dbf-cc55-48b7-b50c-65d6d4ae64ba" containerName="ceilometer-central-agent" probeResult="failure" output=< Dec 13 07:40:53 crc kubenswrapper[4971]: Unkown error: Expecting value: line 1 column 1 (char 0) Dec 13 07:40:53 crc kubenswrapper[4971]: > Dec 13 07:40:53 crc kubenswrapper[4971]: I1213 07:40:53.870130 4971 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openstack/ceilometer-0" Dec 13 07:40:53 crc kubenswrapper[4971]: I1213 07:40:53.871012 4971 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="ceilometer-central-agent" containerStatusID={"Type":"cri-o","ID":"2270e1af9b8023fb36226ee428231c2e187e0f781fcb819a9e8e50ef5b0046f5"} pod="openstack/ceilometer-0" containerMessage="Container ceilometer-central-agent failed liveness probe, will be restarted" Dec 13 07:40:53 crc kubenswrapper[4971]: I1213 07:40:53.871103 4971 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="7cfb4dbf-cc55-48b7-b50c-65d6d4ae64ba" containerName="ceilometer-central-agent" containerID="cri-o://2270e1af9b8023fb36226ee428231c2e187e0f781fcb819a9e8e50ef5b0046f5" gracePeriod=30 Dec 13 07:40:58 crc kubenswrapper[4971]: I1213 07:40:58.066841 4971 generic.go:334] "Generic (PLEG): container finished" podID="7cfb4dbf-cc55-48b7-b50c-65d6d4ae64ba" containerID="2270e1af9b8023fb36226ee428231c2e187e0f781fcb819a9e8e50ef5b0046f5" exitCode=0 Dec 13 07:40:58 crc kubenswrapper[4971]: I1213 07:40:58.066971 4971 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"7cfb4dbf-cc55-48b7-b50c-65d6d4ae64ba","Type":"ContainerDied","Data":"2270e1af9b8023fb36226ee428231c2e187e0f781fcb819a9e8e50ef5b0046f5"} Dec 13 07:40:58 crc kubenswrapper[4971]: I1213 07:40:58.067811 4971 scope.go:117] "RemoveContainer" containerID="5d071b2ffdcfc4bb8c866abc518564c7b5dac13fb9b40ac923cc1c5a8d52504f" Dec 13 07:41:01 crc kubenswrapper[4971]: I1213 07:41:01.768558 4971 scope.go:117] "RemoveContainer" containerID="09d0904c76ce78aa0b074447b1e64cf63160f573da086f934c8cb712615b71d4" Dec 13 07:41:01 crc kubenswrapper[4971]: E1213 07:41:01.769961 4971 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-82xjz_openshift-machine-config-operator(e0c4af12-edda-41d7-9b44-f87396174ef2)\"" pod="openshift-machine-config-operator/machine-config-daemon-82xjz" podUID="e0c4af12-edda-41d7-9b44-f87396174ef2" Dec 13 07:41:08 crc kubenswrapper[4971]: I1213 07:41:08.195542 4971 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"7cfb4dbf-cc55-48b7-b50c-65d6d4ae64ba","Type":"ContainerStarted","Data":"7472627612cd5a2ad24c6810633935f98b010f9d67b67ec6bddf4a6475b06a08"} Dec 13 07:41:14 crc kubenswrapper[4971]: I1213 07:41:14.769237 4971 scope.go:117] "RemoveContainer" containerID="09d0904c76ce78aa0b074447b1e64cf63160f573da086f934c8cb712615b71d4" Dec 13 07:41:14 crc kubenswrapper[4971]: E1213 07:41:14.769942 4971 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-82xjz_openshift-machine-config-operator(e0c4af12-edda-41d7-9b44-f87396174ef2)\"" pod="openshift-machine-config-operator/machine-config-daemon-82xjz" podUID="e0c4af12-edda-41d7-9b44-f87396174ef2" Dec 13 07:41:23 crc kubenswrapper[4971]: I1213 07:41:23.335103 4971 generic.go:334] "Generic (PLEG): container finished" podID="b03f6448-21d3-417f-a1a9-6e7c733550ec" containerID="fae89683ea6cfaaa107d7dbadbc23c229598502b238c7837a8ee6451c38d906d" exitCode=0 Dec 13 07:41:23 crc kubenswrapper[4971]: I1213 07:41:23.335207 4971 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-7qg86" event={"ID":"b03f6448-21d3-417f-a1a9-6e7c733550ec","Type":"ContainerDied","Data":"fae89683ea6cfaaa107d7dbadbc23c229598502b238c7837a8ee6451c38d906d"} Dec 13 07:41:24 crc kubenswrapper[4971]: I1213 07:41:24.778836 4971 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-7qg86" Dec 13 07:41:24 crc kubenswrapper[4971]: I1213 07:41:24.828303 4971 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/b03f6448-21d3-417f-a1a9-6e7c733550ec-inventory\") pod \"b03f6448-21d3-417f-a1a9-6e7c733550ec\" (UID: \"b03f6448-21d3-417f-a1a9-6e7c733550ec\") " Dec 13 07:41:24 crc kubenswrapper[4971]: I1213 07:41:24.828491 4971 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/b03f6448-21d3-417f-a1a9-6e7c733550ec-ssh-key\") pod \"b03f6448-21d3-417f-a1a9-6e7c733550ec\" (UID: \"b03f6448-21d3-417f-a1a9-6e7c733550ec\") " Dec 13 07:41:24 crc kubenswrapper[4971]: I1213 07:41:24.828591 4971 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-w2jfs\" (UniqueName: \"kubernetes.io/projected/b03f6448-21d3-417f-a1a9-6e7c733550ec-kube-api-access-w2jfs\") pod \"b03f6448-21d3-417f-a1a9-6e7c733550ec\" (UID: \"b03f6448-21d3-417f-a1a9-6e7c733550ec\") " Dec 13 07:41:24 crc kubenswrapper[4971]: I1213 07:41:24.828642 4971 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/b03f6448-21d3-417f-a1a9-6e7c733550ec-ceph\") pod \"b03f6448-21d3-417f-a1a9-6e7c733550ec\" (UID: \"b03f6448-21d3-417f-a1a9-6e7c733550ec\") " Dec 13 07:41:24 crc kubenswrapper[4971]: I1213 07:41:24.837701 4971 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/b03f6448-21d3-417f-a1a9-6e7c733550ec-ceph" (OuterVolumeSpecName: "ceph") pod "b03f6448-21d3-417f-a1a9-6e7c733550ec" (UID: "b03f6448-21d3-417f-a1a9-6e7c733550ec"). InnerVolumeSpecName "ceph". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 13 07:41:24 crc kubenswrapper[4971]: I1213 07:41:24.837835 4971 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/b03f6448-21d3-417f-a1a9-6e7c733550ec-kube-api-access-w2jfs" (OuterVolumeSpecName: "kube-api-access-w2jfs") pod "b03f6448-21d3-417f-a1a9-6e7c733550ec" (UID: "b03f6448-21d3-417f-a1a9-6e7c733550ec"). InnerVolumeSpecName "kube-api-access-w2jfs". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 13 07:41:24 crc kubenswrapper[4971]: I1213 07:41:24.861226 4971 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/b03f6448-21d3-417f-a1a9-6e7c733550ec-inventory" (OuterVolumeSpecName: "inventory") pod "b03f6448-21d3-417f-a1a9-6e7c733550ec" (UID: "b03f6448-21d3-417f-a1a9-6e7c733550ec"). InnerVolumeSpecName "inventory". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 13 07:41:24 crc kubenswrapper[4971]: I1213 07:41:24.869938 4971 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/b03f6448-21d3-417f-a1a9-6e7c733550ec-ssh-key" (OuterVolumeSpecName: "ssh-key") pod "b03f6448-21d3-417f-a1a9-6e7c733550ec" (UID: "b03f6448-21d3-417f-a1a9-6e7c733550ec"). InnerVolumeSpecName "ssh-key". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 13 07:41:24 crc kubenswrapper[4971]: I1213 07:41:24.930604 4971 reconciler_common.go:293] "Volume detached for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/b03f6448-21d3-417f-a1a9-6e7c733550ec-ssh-key\") on node \"crc\" DevicePath \"\"" Dec 13 07:41:24 crc kubenswrapper[4971]: I1213 07:41:24.930639 4971 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-w2jfs\" (UniqueName: \"kubernetes.io/projected/b03f6448-21d3-417f-a1a9-6e7c733550ec-kube-api-access-w2jfs\") on node \"crc\" DevicePath \"\"" Dec 13 07:41:24 crc kubenswrapper[4971]: I1213 07:41:24.930650 4971 reconciler_common.go:293] "Volume detached for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/b03f6448-21d3-417f-a1a9-6e7c733550ec-ceph\") on node \"crc\" DevicePath \"\"" Dec 13 07:41:24 crc kubenswrapper[4971]: I1213 07:41:24.930658 4971 reconciler_common.go:293] "Volume detached for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/b03f6448-21d3-417f-a1a9-6e7c733550ec-inventory\") on node \"crc\" DevicePath \"\"" Dec 13 07:41:25 crc kubenswrapper[4971]: I1213 07:41:25.354662 4971 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-7qg86" event={"ID":"b03f6448-21d3-417f-a1a9-6e7c733550ec","Type":"ContainerDied","Data":"abffdfc69b129bfccfa9ab67ef64d13d4b1c972fca8806a723f544b69002ae96"} Dec 13 07:41:25 crc kubenswrapper[4971]: I1213 07:41:25.354724 4971 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="abffdfc69b129bfccfa9ab67ef64d13d4b1c972fca8806a723f544b69002ae96" Dec 13 07:41:25 crc kubenswrapper[4971]: I1213 07:41:25.354763 4971 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-7qg86" Dec 13 07:41:25 crc kubenswrapper[4971]: I1213 07:41:25.449480 4971 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ssh-known-hosts-edpm-deployment-b4wgp"] Dec 13 07:41:25 crc kubenswrapper[4971]: E1213 07:41:25.449975 4971 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="b03f6448-21d3-417f-a1a9-6e7c733550ec" containerName="configure-os-edpm-deployment-openstack-edpm-ipam" Dec 13 07:41:25 crc kubenswrapper[4971]: I1213 07:41:25.450001 4971 state_mem.go:107] "Deleted CPUSet assignment" podUID="b03f6448-21d3-417f-a1a9-6e7c733550ec" containerName="configure-os-edpm-deployment-openstack-edpm-ipam" Dec 13 07:41:25 crc kubenswrapper[4971]: I1213 07:41:25.450243 4971 memory_manager.go:354] "RemoveStaleState removing state" podUID="b03f6448-21d3-417f-a1a9-6e7c733550ec" containerName="configure-os-edpm-deployment-openstack-edpm-ipam" Dec 13 07:41:25 crc kubenswrapper[4971]: I1213 07:41:25.462474 4971 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ssh-known-hosts-edpm-deployment-b4wgp"] Dec 13 07:41:25 crc kubenswrapper[4971]: I1213 07:41:25.462637 4971 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ssh-known-hosts-edpm-deployment-b4wgp" Dec 13 07:41:25 crc kubenswrapper[4971]: I1213 07:41:25.465225 4971 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplane-ansible-ssh-private-key-secret" Dec 13 07:41:25 crc kubenswrapper[4971]: I1213 07:41:25.465749 4971 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplanenodeset-openstack-edpm-ipam" Dec 13 07:41:25 crc kubenswrapper[4971]: I1213 07:41:25.466038 4971 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-edpm-ipam-dockercfg-zb2j5" Dec 13 07:41:25 crc kubenswrapper[4971]: I1213 07:41:25.466368 4971 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceph-conf-files" Dec 13 07:41:25 crc kubenswrapper[4971]: I1213 07:41:25.466603 4971 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-aee-default-env" Dec 13 07:41:25 crc kubenswrapper[4971]: I1213 07:41:25.543607 4971 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key-openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/secret/82daff50-eacf-45ef-b206-ffe0e1d2fe89-ssh-key-openstack-edpm-ipam\") pod \"ssh-known-hosts-edpm-deployment-b4wgp\" (UID: \"82daff50-eacf-45ef-b206-ffe0e1d2fe89\") " pod="openstack/ssh-known-hosts-edpm-deployment-b4wgp" Dec 13 07:41:25 crc kubenswrapper[4971]: I1213 07:41:25.544143 4971 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-7cp2x\" (UniqueName: \"kubernetes.io/projected/82daff50-eacf-45ef-b206-ffe0e1d2fe89-kube-api-access-7cp2x\") pod \"ssh-known-hosts-edpm-deployment-b4wgp\" (UID: \"82daff50-eacf-45ef-b206-ffe0e1d2fe89\") " pod="openstack/ssh-known-hosts-edpm-deployment-b4wgp" Dec 13 07:41:25 crc kubenswrapper[4971]: I1213 07:41:25.544410 4971 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/82daff50-eacf-45ef-b206-ffe0e1d2fe89-ceph\") pod \"ssh-known-hosts-edpm-deployment-b4wgp\" (UID: \"82daff50-eacf-45ef-b206-ffe0e1d2fe89\") " pod="openstack/ssh-known-hosts-edpm-deployment-b4wgp" Dec 13 07:41:25 crc kubenswrapper[4971]: I1213 07:41:25.544810 4971 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"inventory-0\" (UniqueName: \"kubernetes.io/secret/82daff50-eacf-45ef-b206-ffe0e1d2fe89-inventory-0\") pod \"ssh-known-hosts-edpm-deployment-b4wgp\" (UID: \"82daff50-eacf-45ef-b206-ffe0e1d2fe89\") " pod="openstack/ssh-known-hosts-edpm-deployment-b4wgp" Dec 13 07:41:25 crc kubenswrapper[4971]: I1213 07:41:25.646665 4971 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"inventory-0\" (UniqueName: \"kubernetes.io/secret/82daff50-eacf-45ef-b206-ffe0e1d2fe89-inventory-0\") pod \"ssh-known-hosts-edpm-deployment-b4wgp\" (UID: \"82daff50-eacf-45ef-b206-ffe0e1d2fe89\") " pod="openstack/ssh-known-hosts-edpm-deployment-b4wgp" Dec 13 07:41:25 crc kubenswrapper[4971]: I1213 07:41:25.646726 4971 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key-openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/secret/82daff50-eacf-45ef-b206-ffe0e1d2fe89-ssh-key-openstack-edpm-ipam\") pod \"ssh-known-hosts-edpm-deployment-b4wgp\" (UID: \"82daff50-eacf-45ef-b206-ffe0e1d2fe89\") " pod="openstack/ssh-known-hosts-edpm-deployment-b4wgp" Dec 13 07:41:25 crc kubenswrapper[4971]: I1213 07:41:25.646792 4971 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-7cp2x\" (UniqueName: \"kubernetes.io/projected/82daff50-eacf-45ef-b206-ffe0e1d2fe89-kube-api-access-7cp2x\") pod \"ssh-known-hosts-edpm-deployment-b4wgp\" (UID: \"82daff50-eacf-45ef-b206-ffe0e1d2fe89\") " pod="openstack/ssh-known-hosts-edpm-deployment-b4wgp" Dec 13 07:41:25 crc kubenswrapper[4971]: I1213 07:41:25.646817 4971 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/82daff50-eacf-45ef-b206-ffe0e1d2fe89-ceph\") pod \"ssh-known-hosts-edpm-deployment-b4wgp\" (UID: \"82daff50-eacf-45ef-b206-ffe0e1d2fe89\") " pod="openstack/ssh-known-hosts-edpm-deployment-b4wgp" Dec 13 07:41:25 crc kubenswrapper[4971]: I1213 07:41:25.651460 4971 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"inventory-0\" (UniqueName: \"kubernetes.io/secret/82daff50-eacf-45ef-b206-ffe0e1d2fe89-inventory-0\") pod \"ssh-known-hosts-edpm-deployment-b4wgp\" (UID: \"82daff50-eacf-45ef-b206-ffe0e1d2fe89\") " pod="openstack/ssh-known-hosts-edpm-deployment-b4wgp" Dec 13 07:41:25 crc kubenswrapper[4971]: I1213 07:41:25.657112 4971 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/82daff50-eacf-45ef-b206-ffe0e1d2fe89-ceph\") pod \"ssh-known-hosts-edpm-deployment-b4wgp\" (UID: \"82daff50-eacf-45ef-b206-ffe0e1d2fe89\") " pod="openstack/ssh-known-hosts-edpm-deployment-b4wgp" Dec 13 07:41:25 crc kubenswrapper[4971]: I1213 07:41:25.661186 4971 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key-openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/secret/82daff50-eacf-45ef-b206-ffe0e1d2fe89-ssh-key-openstack-edpm-ipam\") pod \"ssh-known-hosts-edpm-deployment-b4wgp\" (UID: \"82daff50-eacf-45ef-b206-ffe0e1d2fe89\") " pod="openstack/ssh-known-hosts-edpm-deployment-b4wgp" Dec 13 07:41:25 crc kubenswrapper[4971]: I1213 07:41:25.665316 4971 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-7cp2x\" (UniqueName: \"kubernetes.io/projected/82daff50-eacf-45ef-b206-ffe0e1d2fe89-kube-api-access-7cp2x\") pod \"ssh-known-hosts-edpm-deployment-b4wgp\" (UID: \"82daff50-eacf-45ef-b206-ffe0e1d2fe89\") " pod="openstack/ssh-known-hosts-edpm-deployment-b4wgp" Dec 13 07:41:25 crc kubenswrapper[4971]: I1213 07:41:25.786901 4971 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ssh-known-hosts-edpm-deployment-b4wgp" Dec 13 07:41:26 crc kubenswrapper[4971]: I1213 07:41:26.274079 4971 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ssh-known-hosts-edpm-deployment-b4wgp"] Dec 13 07:41:26 crc kubenswrapper[4971]: I1213 07:41:26.363609 4971 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ssh-known-hosts-edpm-deployment-b4wgp" event={"ID":"82daff50-eacf-45ef-b206-ffe0e1d2fe89","Type":"ContainerStarted","Data":"ba9e889704a2692f493a78e8601cda3d5606b0faef8ba3b931d81497c785fef4"} Dec 13 07:41:28 crc kubenswrapper[4971]: I1213 07:41:28.769346 4971 scope.go:117] "RemoveContainer" containerID="09d0904c76ce78aa0b074447b1e64cf63160f573da086f934c8cb712615b71d4" Dec 13 07:41:28 crc kubenswrapper[4971]: E1213 07:41:28.770297 4971 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-82xjz_openshift-machine-config-operator(e0c4af12-edda-41d7-9b44-f87396174ef2)\"" pod="openshift-machine-config-operator/machine-config-daemon-82xjz" podUID="e0c4af12-edda-41d7-9b44-f87396174ef2" Dec 13 07:41:33 crc kubenswrapper[4971]: I1213 07:41:33.430916 4971 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ssh-known-hosts-edpm-deployment-b4wgp" event={"ID":"82daff50-eacf-45ef-b206-ffe0e1d2fe89","Type":"ContainerStarted","Data":"89a0f112ec24cb027c74c7a93d2e2c6b24566875cbd8b8e5db4bc050e022c529"} Dec 13 07:41:33 crc kubenswrapper[4971]: I1213 07:41:33.450129 4971 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/ssh-known-hosts-edpm-deployment-b4wgp" podStartSLOduration=3.455631035 podStartE2EDuration="8.450110202s" podCreationTimestamp="2025-12-13 07:41:25 +0000 UTC" firstStartedPulling="2025-12-13 07:41:26.278396263 +0000 UTC m=+3142.882805731" lastFinishedPulling="2025-12-13 07:41:31.27287545 +0000 UTC m=+3147.877284898" observedRunningTime="2025-12-13 07:41:33.445617634 +0000 UTC m=+3150.050027082" watchObservedRunningTime="2025-12-13 07:41:33.450110202 +0000 UTC m=+3150.054519650" Dec 13 07:41:40 crc kubenswrapper[4971]: I1213 07:41:40.769095 4971 scope.go:117] "RemoveContainer" containerID="09d0904c76ce78aa0b074447b1e64cf63160f573da086f934c8cb712615b71d4" Dec 13 07:41:40 crc kubenswrapper[4971]: E1213 07:41:40.770803 4971 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-82xjz_openshift-machine-config-operator(e0c4af12-edda-41d7-9b44-f87396174ef2)\"" pod="openshift-machine-config-operator/machine-config-daemon-82xjz" podUID="e0c4af12-edda-41d7-9b44-f87396174ef2" Dec 13 07:41:44 crc kubenswrapper[4971]: I1213 07:41:44.534290 4971 generic.go:334] "Generic (PLEG): container finished" podID="82daff50-eacf-45ef-b206-ffe0e1d2fe89" containerID="89a0f112ec24cb027c74c7a93d2e2c6b24566875cbd8b8e5db4bc050e022c529" exitCode=0 Dec 13 07:41:44 crc kubenswrapper[4971]: I1213 07:41:44.534551 4971 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ssh-known-hosts-edpm-deployment-b4wgp" event={"ID":"82daff50-eacf-45ef-b206-ffe0e1d2fe89","Type":"ContainerDied","Data":"89a0f112ec24cb027c74c7a93d2e2c6b24566875cbd8b8e5db4bc050e022c529"} Dec 13 07:41:45 crc kubenswrapper[4971]: I1213 07:41:45.958734 4971 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ssh-known-hosts-edpm-deployment-b4wgp" Dec 13 07:41:46 crc kubenswrapper[4971]: I1213 07:41:46.050573 4971 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/82daff50-eacf-45ef-b206-ffe0e1d2fe89-ceph\") pod \"82daff50-eacf-45ef-b206-ffe0e1d2fe89\" (UID: \"82daff50-eacf-45ef-b206-ffe0e1d2fe89\") " Dec 13 07:41:46 crc kubenswrapper[4971]: I1213 07:41:46.050648 4971 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-7cp2x\" (UniqueName: \"kubernetes.io/projected/82daff50-eacf-45ef-b206-ffe0e1d2fe89-kube-api-access-7cp2x\") pod \"82daff50-eacf-45ef-b206-ffe0e1d2fe89\" (UID: \"82daff50-eacf-45ef-b206-ffe0e1d2fe89\") " Dec 13 07:41:46 crc kubenswrapper[4971]: I1213 07:41:46.050691 4971 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"inventory-0\" (UniqueName: \"kubernetes.io/secret/82daff50-eacf-45ef-b206-ffe0e1d2fe89-inventory-0\") pod \"82daff50-eacf-45ef-b206-ffe0e1d2fe89\" (UID: \"82daff50-eacf-45ef-b206-ffe0e1d2fe89\") " Dec 13 07:41:46 crc kubenswrapper[4971]: I1213 07:41:46.050765 4971 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key-openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/secret/82daff50-eacf-45ef-b206-ffe0e1d2fe89-ssh-key-openstack-edpm-ipam\") pod \"82daff50-eacf-45ef-b206-ffe0e1d2fe89\" (UID: \"82daff50-eacf-45ef-b206-ffe0e1d2fe89\") " Dec 13 07:41:46 crc kubenswrapper[4971]: I1213 07:41:46.056833 4971 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/82daff50-eacf-45ef-b206-ffe0e1d2fe89-kube-api-access-7cp2x" (OuterVolumeSpecName: "kube-api-access-7cp2x") pod "82daff50-eacf-45ef-b206-ffe0e1d2fe89" (UID: "82daff50-eacf-45ef-b206-ffe0e1d2fe89"). InnerVolumeSpecName "kube-api-access-7cp2x". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 13 07:41:46 crc kubenswrapper[4971]: I1213 07:41:46.058109 4971 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/82daff50-eacf-45ef-b206-ffe0e1d2fe89-ceph" (OuterVolumeSpecName: "ceph") pod "82daff50-eacf-45ef-b206-ffe0e1d2fe89" (UID: "82daff50-eacf-45ef-b206-ffe0e1d2fe89"). InnerVolumeSpecName "ceph". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 13 07:41:46 crc kubenswrapper[4971]: I1213 07:41:46.088790 4971 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/82daff50-eacf-45ef-b206-ffe0e1d2fe89-ssh-key-openstack-edpm-ipam" (OuterVolumeSpecName: "ssh-key-openstack-edpm-ipam") pod "82daff50-eacf-45ef-b206-ffe0e1d2fe89" (UID: "82daff50-eacf-45ef-b206-ffe0e1d2fe89"). InnerVolumeSpecName "ssh-key-openstack-edpm-ipam". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 13 07:41:46 crc kubenswrapper[4971]: I1213 07:41:46.093014 4971 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/82daff50-eacf-45ef-b206-ffe0e1d2fe89-inventory-0" (OuterVolumeSpecName: "inventory-0") pod "82daff50-eacf-45ef-b206-ffe0e1d2fe89" (UID: "82daff50-eacf-45ef-b206-ffe0e1d2fe89"). InnerVolumeSpecName "inventory-0". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 13 07:41:46 crc kubenswrapper[4971]: I1213 07:41:46.154592 4971 reconciler_common.go:293] "Volume detached for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/82daff50-eacf-45ef-b206-ffe0e1d2fe89-ceph\") on node \"crc\" DevicePath \"\"" Dec 13 07:41:46 crc kubenswrapper[4971]: I1213 07:41:46.154630 4971 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-7cp2x\" (UniqueName: \"kubernetes.io/projected/82daff50-eacf-45ef-b206-ffe0e1d2fe89-kube-api-access-7cp2x\") on node \"crc\" DevicePath \"\"" Dec 13 07:41:46 crc kubenswrapper[4971]: I1213 07:41:46.154645 4971 reconciler_common.go:293] "Volume detached for volume \"inventory-0\" (UniqueName: \"kubernetes.io/secret/82daff50-eacf-45ef-b206-ffe0e1d2fe89-inventory-0\") on node \"crc\" DevicePath \"\"" Dec 13 07:41:46 crc kubenswrapper[4971]: I1213 07:41:46.154659 4971 reconciler_common.go:293] "Volume detached for volume \"ssh-key-openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/secret/82daff50-eacf-45ef-b206-ffe0e1d2fe89-ssh-key-openstack-edpm-ipam\") on node \"crc\" DevicePath \"\"" Dec 13 07:41:46 crc kubenswrapper[4971]: I1213 07:41:46.558979 4971 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ssh-known-hosts-edpm-deployment-b4wgp" event={"ID":"82daff50-eacf-45ef-b206-ffe0e1d2fe89","Type":"ContainerDied","Data":"ba9e889704a2692f493a78e8601cda3d5606b0faef8ba3b931d81497c785fef4"} Dec 13 07:41:46 crc kubenswrapper[4971]: I1213 07:41:46.559042 4971 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="ba9e889704a2692f493a78e8601cda3d5606b0faef8ba3b931d81497c785fef4" Dec 13 07:41:46 crc kubenswrapper[4971]: I1213 07:41:46.559120 4971 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ssh-known-hosts-edpm-deployment-b4wgp" Dec 13 07:41:46 crc kubenswrapper[4971]: I1213 07:41:46.680007 4971 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/run-os-edpm-deployment-openstack-edpm-ipam-tn7h4"] Dec 13 07:41:46 crc kubenswrapper[4971]: E1213 07:41:46.681795 4971 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="82daff50-eacf-45ef-b206-ffe0e1d2fe89" containerName="ssh-known-hosts-edpm-deployment" Dec 13 07:41:46 crc kubenswrapper[4971]: I1213 07:41:46.681836 4971 state_mem.go:107] "Deleted CPUSet assignment" podUID="82daff50-eacf-45ef-b206-ffe0e1d2fe89" containerName="ssh-known-hosts-edpm-deployment" Dec 13 07:41:46 crc kubenswrapper[4971]: I1213 07:41:46.682158 4971 memory_manager.go:354] "RemoveStaleState removing state" podUID="82daff50-eacf-45ef-b206-ffe0e1d2fe89" containerName="ssh-known-hosts-edpm-deployment" Dec 13 07:41:46 crc kubenswrapper[4971]: I1213 07:41:46.683595 4971 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/run-os-edpm-deployment-openstack-edpm-ipam-tn7h4" Dec 13 07:41:46 crc kubenswrapper[4971]: I1213 07:41:46.686031 4971 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceph-conf-files" Dec 13 07:41:46 crc kubenswrapper[4971]: I1213 07:41:46.686077 4971 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-aee-default-env" Dec 13 07:41:46 crc kubenswrapper[4971]: I1213 07:41:46.686574 4971 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-edpm-ipam-dockercfg-zb2j5" Dec 13 07:41:46 crc kubenswrapper[4971]: I1213 07:41:46.686843 4971 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplanenodeset-openstack-edpm-ipam" Dec 13 07:41:46 crc kubenswrapper[4971]: I1213 07:41:46.687067 4971 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplane-ansible-ssh-private-key-secret" Dec 13 07:41:46 crc kubenswrapper[4971]: I1213 07:41:46.697594 4971 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/run-os-edpm-deployment-openstack-edpm-ipam-tn7h4"] Dec 13 07:41:46 crc kubenswrapper[4971]: E1213 07:41:46.811049 4971 cadvisor_stats_provider.go:516] "Partial failure issuing cadvisor.ContainerInfoV2" err="partial failures: [\"/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod82daff50_eacf_45ef_b206_ffe0e1d2fe89.slice/crio-ba9e889704a2692f493a78e8601cda3d5606b0faef8ba3b931d81497c785fef4\": RecentStats: unable to find data in memory cache], [\"/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod82daff50_eacf_45ef_b206_ffe0e1d2fe89.slice\": RecentStats: unable to find data in memory cache]" Dec 13 07:41:46 crc kubenswrapper[4971]: I1213 07:41:46.870994 4971 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/88c03dc6-31a0-4c8c-93b6-61a9c895dffe-ssh-key\") pod \"run-os-edpm-deployment-openstack-edpm-ipam-tn7h4\" (UID: \"88c03dc6-31a0-4c8c-93b6-61a9c895dffe\") " pod="openstack/run-os-edpm-deployment-openstack-edpm-ipam-tn7h4" Dec 13 07:41:46 crc kubenswrapper[4971]: I1213 07:41:46.871116 4971 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/88c03dc6-31a0-4c8c-93b6-61a9c895dffe-inventory\") pod \"run-os-edpm-deployment-openstack-edpm-ipam-tn7h4\" (UID: \"88c03dc6-31a0-4c8c-93b6-61a9c895dffe\") " pod="openstack/run-os-edpm-deployment-openstack-edpm-ipam-tn7h4" Dec 13 07:41:46 crc kubenswrapper[4971]: I1213 07:41:46.871143 4971 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/88c03dc6-31a0-4c8c-93b6-61a9c895dffe-ceph\") pod \"run-os-edpm-deployment-openstack-edpm-ipam-tn7h4\" (UID: \"88c03dc6-31a0-4c8c-93b6-61a9c895dffe\") " pod="openstack/run-os-edpm-deployment-openstack-edpm-ipam-tn7h4" Dec 13 07:41:46 crc kubenswrapper[4971]: I1213 07:41:46.871255 4971 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-zmml2\" (UniqueName: \"kubernetes.io/projected/88c03dc6-31a0-4c8c-93b6-61a9c895dffe-kube-api-access-zmml2\") pod \"run-os-edpm-deployment-openstack-edpm-ipam-tn7h4\" (UID: \"88c03dc6-31a0-4c8c-93b6-61a9c895dffe\") " pod="openstack/run-os-edpm-deployment-openstack-edpm-ipam-tn7h4" Dec 13 07:41:46 crc kubenswrapper[4971]: I1213 07:41:46.972308 4971 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/88c03dc6-31a0-4c8c-93b6-61a9c895dffe-ssh-key\") pod \"run-os-edpm-deployment-openstack-edpm-ipam-tn7h4\" (UID: \"88c03dc6-31a0-4c8c-93b6-61a9c895dffe\") " pod="openstack/run-os-edpm-deployment-openstack-edpm-ipam-tn7h4" Dec 13 07:41:46 crc kubenswrapper[4971]: I1213 07:41:46.972361 4971 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/88c03dc6-31a0-4c8c-93b6-61a9c895dffe-inventory\") pod \"run-os-edpm-deployment-openstack-edpm-ipam-tn7h4\" (UID: \"88c03dc6-31a0-4c8c-93b6-61a9c895dffe\") " pod="openstack/run-os-edpm-deployment-openstack-edpm-ipam-tn7h4" Dec 13 07:41:46 crc kubenswrapper[4971]: I1213 07:41:46.972381 4971 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/88c03dc6-31a0-4c8c-93b6-61a9c895dffe-ceph\") pod \"run-os-edpm-deployment-openstack-edpm-ipam-tn7h4\" (UID: \"88c03dc6-31a0-4c8c-93b6-61a9c895dffe\") " pod="openstack/run-os-edpm-deployment-openstack-edpm-ipam-tn7h4" Dec 13 07:41:46 crc kubenswrapper[4971]: I1213 07:41:46.972442 4971 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-zmml2\" (UniqueName: \"kubernetes.io/projected/88c03dc6-31a0-4c8c-93b6-61a9c895dffe-kube-api-access-zmml2\") pod \"run-os-edpm-deployment-openstack-edpm-ipam-tn7h4\" (UID: \"88c03dc6-31a0-4c8c-93b6-61a9c895dffe\") " pod="openstack/run-os-edpm-deployment-openstack-edpm-ipam-tn7h4" Dec 13 07:41:46 crc kubenswrapper[4971]: I1213 07:41:46.984902 4971 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/88c03dc6-31a0-4c8c-93b6-61a9c895dffe-inventory\") pod \"run-os-edpm-deployment-openstack-edpm-ipam-tn7h4\" (UID: \"88c03dc6-31a0-4c8c-93b6-61a9c895dffe\") " pod="openstack/run-os-edpm-deployment-openstack-edpm-ipam-tn7h4" Dec 13 07:41:46 crc kubenswrapper[4971]: I1213 07:41:46.989368 4971 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/88c03dc6-31a0-4c8c-93b6-61a9c895dffe-ssh-key\") pod \"run-os-edpm-deployment-openstack-edpm-ipam-tn7h4\" (UID: \"88c03dc6-31a0-4c8c-93b6-61a9c895dffe\") " pod="openstack/run-os-edpm-deployment-openstack-edpm-ipam-tn7h4" Dec 13 07:41:46 crc kubenswrapper[4971]: I1213 07:41:46.990131 4971 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/88c03dc6-31a0-4c8c-93b6-61a9c895dffe-ceph\") pod \"run-os-edpm-deployment-openstack-edpm-ipam-tn7h4\" (UID: \"88c03dc6-31a0-4c8c-93b6-61a9c895dffe\") " pod="openstack/run-os-edpm-deployment-openstack-edpm-ipam-tn7h4" Dec 13 07:41:46 crc kubenswrapper[4971]: I1213 07:41:46.992785 4971 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-zmml2\" (UniqueName: \"kubernetes.io/projected/88c03dc6-31a0-4c8c-93b6-61a9c895dffe-kube-api-access-zmml2\") pod \"run-os-edpm-deployment-openstack-edpm-ipam-tn7h4\" (UID: \"88c03dc6-31a0-4c8c-93b6-61a9c895dffe\") " pod="openstack/run-os-edpm-deployment-openstack-edpm-ipam-tn7h4" Dec 13 07:41:47 crc kubenswrapper[4971]: I1213 07:41:47.007656 4971 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/run-os-edpm-deployment-openstack-edpm-ipam-tn7h4" Dec 13 07:41:47 crc kubenswrapper[4971]: I1213 07:41:47.371904 4971 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/run-os-edpm-deployment-openstack-edpm-ipam-tn7h4"] Dec 13 07:41:47 crc kubenswrapper[4971]: W1213 07:41:47.372901 4971 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod88c03dc6_31a0_4c8c_93b6_61a9c895dffe.slice/crio-25d7181f13934b49eaf31f347a1b21b668fba5e678557d04efc4ed41500c539a WatchSource:0}: Error finding container 25d7181f13934b49eaf31f347a1b21b668fba5e678557d04efc4ed41500c539a: Status 404 returned error can't find the container with id 25d7181f13934b49eaf31f347a1b21b668fba5e678557d04efc4ed41500c539a Dec 13 07:41:47 crc kubenswrapper[4971]: I1213 07:41:47.567394 4971 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/run-os-edpm-deployment-openstack-edpm-ipam-tn7h4" event={"ID":"88c03dc6-31a0-4c8c-93b6-61a9c895dffe","Type":"ContainerStarted","Data":"25d7181f13934b49eaf31f347a1b21b668fba5e678557d04efc4ed41500c539a"} Dec 13 07:41:48 crc kubenswrapper[4971]: I1213 07:41:48.577860 4971 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/run-os-edpm-deployment-openstack-edpm-ipam-tn7h4" event={"ID":"88c03dc6-31a0-4c8c-93b6-61a9c895dffe","Type":"ContainerStarted","Data":"98de38d5b38a816a5918866b64e3efa749a71316b2060606ffe47c4d34b72ce4"} Dec 13 07:41:48 crc kubenswrapper[4971]: I1213 07:41:48.601425 4971 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/run-os-edpm-deployment-openstack-edpm-ipam-tn7h4" podStartSLOduration=1.654597883 podStartE2EDuration="2.60140398s" podCreationTimestamp="2025-12-13 07:41:46 +0000 UTC" firstStartedPulling="2025-12-13 07:41:47.375874063 +0000 UTC m=+3163.980283521" lastFinishedPulling="2025-12-13 07:41:48.32268016 +0000 UTC m=+3164.927089618" observedRunningTime="2025-12-13 07:41:48.595141091 +0000 UTC m=+3165.199550539" watchObservedRunningTime="2025-12-13 07:41:48.60140398 +0000 UTC m=+3165.205813438" Dec 13 07:41:51 crc kubenswrapper[4971]: I1213 07:41:51.768433 4971 scope.go:117] "RemoveContainer" containerID="09d0904c76ce78aa0b074447b1e64cf63160f573da086f934c8cb712615b71d4" Dec 13 07:41:51 crc kubenswrapper[4971]: E1213 07:41:51.769272 4971 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-82xjz_openshift-machine-config-operator(e0c4af12-edda-41d7-9b44-f87396174ef2)\"" pod="openshift-machine-config-operator/machine-config-daemon-82xjz" podUID="e0c4af12-edda-41d7-9b44-f87396174ef2" Dec 13 07:42:05 crc kubenswrapper[4971]: I1213 07:42:05.769210 4971 scope.go:117] "RemoveContainer" containerID="09d0904c76ce78aa0b074447b1e64cf63160f573da086f934c8cb712615b71d4" Dec 13 07:42:05 crc kubenswrapper[4971]: E1213 07:42:05.770052 4971 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-82xjz_openshift-machine-config-operator(e0c4af12-edda-41d7-9b44-f87396174ef2)\"" pod="openshift-machine-config-operator/machine-config-daemon-82xjz" podUID="e0c4af12-edda-41d7-9b44-f87396174ef2" Dec 13 07:42:08 crc kubenswrapper[4971]: I1213 07:42:08.760500 4971 generic.go:334] "Generic (PLEG): container finished" podID="88c03dc6-31a0-4c8c-93b6-61a9c895dffe" containerID="98de38d5b38a816a5918866b64e3efa749a71316b2060606ffe47c4d34b72ce4" exitCode=0 Dec 13 07:42:08 crc kubenswrapper[4971]: I1213 07:42:08.760875 4971 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/run-os-edpm-deployment-openstack-edpm-ipam-tn7h4" event={"ID":"88c03dc6-31a0-4c8c-93b6-61a9c895dffe","Type":"ContainerDied","Data":"98de38d5b38a816a5918866b64e3efa749a71316b2060606ffe47c4d34b72ce4"} Dec 13 07:42:10 crc kubenswrapper[4971]: I1213 07:42:10.359836 4971 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/run-os-edpm-deployment-openstack-edpm-ipam-tn7h4" Dec 13 07:42:10 crc kubenswrapper[4971]: I1213 07:42:10.509532 4971 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-zmml2\" (UniqueName: \"kubernetes.io/projected/88c03dc6-31a0-4c8c-93b6-61a9c895dffe-kube-api-access-zmml2\") pod \"88c03dc6-31a0-4c8c-93b6-61a9c895dffe\" (UID: \"88c03dc6-31a0-4c8c-93b6-61a9c895dffe\") " Dec 13 07:42:10 crc kubenswrapper[4971]: I1213 07:42:10.509617 4971 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/88c03dc6-31a0-4c8c-93b6-61a9c895dffe-ssh-key\") pod \"88c03dc6-31a0-4c8c-93b6-61a9c895dffe\" (UID: \"88c03dc6-31a0-4c8c-93b6-61a9c895dffe\") " Dec 13 07:42:10 crc kubenswrapper[4971]: I1213 07:42:10.509645 4971 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/88c03dc6-31a0-4c8c-93b6-61a9c895dffe-inventory\") pod \"88c03dc6-31a0-4c8c-93b6-61a9c895dffe\" (UID: \"88c03dc6-31a0-4c8c-93b6-61a9c895dffe\") " Dec 13 07:42:10 crc kubenswrapper[4971]: I1213 07:42:10.536719 4971 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/88c03dc6-31a0-4c8c-93b6-61a9c895dffe-kube-api-access-zmml2" (OuterVolumeSpecName: "kube-api-access-zmml2") pod "88c03dc6-31a0-4c8c-93b6-61a9c895dffe" (UID: "88c03dc6-31a0-4c8c-93b6-61a9c895dffe"). InnerVolumeSpecName "kube-api-access-zmml2". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 13 07:42:10 crc kubenswrapper[4971]: I1213 07:42:10.559560 4971 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/88c03dc6-31a0-4c8c-93b6-61a9c895dffe-inventory" (OuterVolumeSpecName: "inventory") pod "88c03dc6-31a0-4c8c-93b6-61a9c895dffe" (UID: "88c03dc6-31a0-4c8c-93b6-61a9c895dffe"). InnerVolumeSpecName "inventory". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 13 07:42:10 crc kubenswrapper[4971]: I1213 07:42:10.607593 4971 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/88c03dc6-31a0-4c8c-93b6-61a9c895dffe-ssh-key" (OuterVolumeSpecName: "ssh-key") pod "88c03dc6-31a0-4c8c-93b6-61a9c895dffe" (UID: "88c03dc6-31a0-4c8c-93b6-61a9c895dffe"). InnerVolumeSpecName "ssh-key". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 13 07:42:10 crc kubenswrapper[4971]: I1213 07:42:10.610879 4971 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/88c03dc6-31a0-4c8c-93b6-61a9c895dffe-ceph\") pod \"88c03dc6-31a0-4c8c-93b6-61a9c895dffe\" (UID: \"88c03dc6-31a0-4c8c-93b6-61a9c895dffe\") " Dec 13 07:42:10 crc kubenswrapper[4971]: I1213 07:42:10.611211 4971 reconciler_common.go:293] "Volume detached for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/88c03dc6-31a0-4c8c-93b6-61a9c895dffe-ssh-key\") on node \"crc\" DevicePath \"\"" Dec 13 07:42:10 crc kubenswrapper[4971]: I1213 07:42:10.611230 4971 reconciler_common.go:293] "Volume detached for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/88c03dc6-31a0-4c8c-93b6-61a9c895dffe-inventory\") on node \"crc\" DevicePath \"\"" Dec 13 07:42:10 crc kubenswrapper[4971]: I1213 07:42:10.611240 4971 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-zmml2\" (UniqueName: \"kubernetes.io/projected/88c03dc6-31a0-4c8c-93b6-61a9c895dffe-kube-api-access-zmml2\") on node \"crc\" DevicePath \"\"" Dec 13 07:42:10 crc kubenswrapper[4971]: I1213 07:42:10.613611 4971 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/88c03dc6-31a0-4c8c-93b6-61a9c895dffe-ceph" (OuterVolumeSpecName: "ceph") pod "88c03dc6-31a0-4c8c-93b6-61a9c895dffe" (UID: "88c03dc6-31a0-4c8c-93b6-61a9c895dffe"). InnerVolumeSpecName "ceph". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 13 07:42:10 crc kubenswrapper[4971]: I1213 07:42:10.712559 4971 reconciler_common.go:293] "Volume detached for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/88c03dc6-31a0-4c8c-93b6-61a9c895dffe-ceph\") on node \"crc\" DevicePath \"\"" Dec 13 07:42:10 crc kubenswrapper[4971]: I1213 07:42:10.782347 4971 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/run-os-edpm-deployment-openstack-edpm-ipam-tn7h4" event={"ID":"88c03dc6-31a0-4c8c-93b6-61a9c895dffe","Type":"ContainerDied","Data":"25d7181f13934b49eaf31f347a1b21b668fba5e678557d04efc4ed41500c539a"} Dec 13 07:42:10 crc kubenswrapper[4971]: I1213 07:42:10.782384 4971 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="25d7181f13934b49eaf31f347a1b21b668fba5e678557d04efc4ed41500c539a" Dec 13 07:42:10 crc kubenswrapper[4971]: I1213 07:42:10.782430 4971 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/run-os-edpm-deployment-openstack-edpm-ipam-tn7h4" Dec 13 07:42:11 crc kubenswrapper[4971]: I1213 07:42:11.014296 4971 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-mmm7v"] Dec 13 07:42:11 crc kubenswrapper[4971]: E1213 07:42:11.014762 4971 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="88c03dc6-31a0-4c8c-93b6-61a9c895dffe" containerName="run-os-edpm-deployment-openstack-edpm-ipam" Dec 13 07:42:11 crc kubenswrapper[4971]: I1213 07:42:11.014784 4971 state_mem.go:107] "Deleted CPUSet assignment" podUID="88c03dc6-31a0-4c8c-93b6-61a9c895dffe" containerName="run-os-edpm-deployment-openstack-edpm-ipam" Dec 13 07:42:11 crc kubenswrapper[4971]: I1213 07:42:11.015020 4971 memory_manager.go:354] "RemoveStaleState removing state" podUID="88c03dc6-31a0-4c8c-93b6-61a9c895dffe" containerName="run-os-edpm-deployment-openstack-edpm-ipam" Dec 13 07:42:11 crc kubenswrapper[4971]: I1213 07:42:11.015724 4971 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-mmm7v" Dec 13 07:42:11 crc kubenswrapper[4971]: I1213 07:42:11.018085 4971 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceph-conf-files" Dec 13 07:42:11 crc kubenswrapper[4971]: I1213 07:42:11.018691 4971 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplane-ansible-ssh-private-key-secret" Dec 13 07:42:11 crc kubenswrapper[4971]: I1213 07:42:11.018773 4971 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplanenodeset-openstack-edpm-ipam" Dec 13 07:42:11 crc kubenswrapper[4971]: I1213 07:42:11.019170 4971 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-aee-default-env" Dec 13 07:42:11 crc kubenswrapper[4971]: I1213 07:42:11.019883 4971 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-2nmwh\" (UniqueName: \"kubernetes.io/projected/94cbb455-7651-4d43-ad24-495457496a2d-kube-api-access-2nmwh\") pod \"reboot-os-edpm-deployment-openstack-edpm-ipam-mmm7v\" (UID: \"94cbb455-7651-4d43-ad24-495457496a2d\") " pod="openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-mmm7v" Dec 13 07:42:11 crc kubenswrapper[4971]: I1213 07:42:11.019957 4971 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/94cbb455-7651-4d43-ad24-495457496a2d-ssh-key\") pod \"reboot-os-edpm-deployment-openstack-edpm-ipam-mmm7v\" (UID: \"94cbb455-7651-4d43-ad24-495457496a2d\") " pod="openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-mmm7v" Dec 13 07:42:11 crc kubenswrapper[4971]: I1213 07:42:11.019987 4971 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/94cbb455-7651-4d43-ad24-495457496a2d-inventory\") pod \"reboot-os-edpm-deployment-openstack-edpm-ipam-mmm7v\" (UID: \"94cbb455-7651-4d43-ad24-495457496a2d\") " pod="openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-mmm7v" Dec 13 07:42:11 crc kubenswrapper[4971]: I1213 07:42:11.020030 4971 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/94cbb455-7651-4d43-ad24-495457496a2d-ceph\") pod \"reboot-os-edpm-deployment-openstack-edpm-ipam-mmm7v\" (UID: \"94cbb455-7651-4d43-ad24-495457496a2d\") " pod="openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-mmm7v" Dec 13 07:42:11 crc kubenswrapper[4971]: I1213 07:42:11.021951 4971 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-edpm-ipam-dockercfg-zb2j5" Dec 13 07:42:11 crc kubenswrapper[4971]: I1213 07:42:11.124169 4971 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-mmm7v"] Dec 13 07:42:11 crc kubenswrapper[4971]: I1213 07:42:11.127129 4971 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/94cbb455-7651-4d43-ad24-495457496a2d-ssh-key\") pod \"reboot-os-edpm-deployment-openstack-edpm-ipam-mmm7v\" (UID: \"94cbb455-7651-4d43-ad24-495457496a2d\") " pod="openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-mmm7v" Dec 13 07:42:11 crc kubenswrapper[4971]: I1213 07:42:11.127187 4971 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/94cbb455-7651-4d43-ad24-495457496a2d-inventory\") pod \"reboot-os-edpm-deployment-openstack-edpm-ipam-mmm7v\" (UID: \"94cbb455-7651-4d43-ad24-495457496a2d\") " pod="openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-mmm7v" Dec 13 07:42:11 crc kubenswrapper[4971]: I1213 07:42:11.127230 4971 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/94cbb455-7651-4d43-ad24-495457496a2d-ceph\") pod \"reboot-os-edpm-deployment-openstack-edpm-ipam-mmm7v\" (UID: \"94cbb455-7651-4d43-ad24-495457496a2d\") " pod="openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-mmm7v" Dec 13 07:42:11 crc kubenswrapper[4971]: I1213 07:42:11.127376 4971 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-2nmwh\" (UniqueName: \"kubernetes.io/projected/94cbb455-7651-4d43-ad24-495457496a2d-kube-api-access-2nmwh\") pod \"reboot-os-edpm-deployment-openstack-edpm-ipam-mmm7v\" (UID: \"94cbb455-7651-4d43-ad24-495457496a2d\") " pod="openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-mmm7v" Dec 13 07:42:11 crc kubenswrapper[4971]: I1213 07:42:11.132228 4971 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/94cbb455-7651-4d43-ad24-495457496a2d-ssh-key\") pod \"reboot-os-edpm-deployment-openstack-edpm-ipam-mmm7v\" (UID: \"94cbb455-7651-4d43-ad24-495457496a2d\") " pod="openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-mmm7v" Dec 13 07:42:11 crc kubenswrapper[4971]: I1213 07:42:11.139266 4971 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/94cbb455-7651-4d43-ad24-495457496a2d-inventory\") pod \"reboot-os-edpm-deployment-openstack-edpm-ipam-mmm7v\" (UID: \"94cbb455-7651-4d43-ad24-495457496a2d\") " pod="openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-mmm7v" Dec 13 07:42:11 crc kubenswrapper[4971]: I1213 07:42:11.140885 4971 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/94cbb455-7651-4d43-ad24-495457496a2d-ceph\") pod \"reboot-os-edpm-deployment-openstack-edpm-ipam-mmm7v\" (UID: \"94cbb455-7651-4d43-ad24-495457496a2d\") " pod="openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-mmm7v" Dec 13 07:42:11 crc kubenswrapper[4971]: I1213 07:42:11.154240 4971 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-2nmwh\" (UniqueName: \"kubernetes.io/projected/94cbb455-7651-4d43-ad24-495457496a2d-kube-api-access-2nmwh\") pod \"reboot-os-edpm-deployment-openstack-edpm-ipam-mmm7v\" (UID: \"94cbb455-7651-4d43-ad24-495457496a2d\") " pod="openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-mmm7v" Dec 13 07:42:11 crc kubenswrapper[4971]: I1213 07:42:11.440258 4971 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-mmm7v" Dec 13 07:42:12 crc kubenswrapper[4971]: I1213 07:42:12.151928 4971 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-mmm7v"] Dec 13 07:42:12 crc kubenswrapper[4971]: I1213 07:42:12.802603 4971 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-mmm7v" event={"ID":"94cbb455-7651-4d43-ad24-495457496a2d","Type":"ContainerStarted","Data":"9954d38a1098cd82cf3b5cf3c0a82412973e68ce725670504824eae875c2a3dc"} Dec 13 07:42:16 crc kubenswrapper[4971]: I1213 07:42:16.768194 4971 scope.go:117] "RemoveContainer" containerID="09d0904c76ce78aa0b074447b1e64cf63160f573da086f934c8cb712615b71d4" Dec 13 07:42:16 crc kubenswrapper[4971]: E1213 07:42:16.769166 4971 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-82xjz_openshift-machine-config-operator(e0c4af12-edda-41d7-9b44-f87396174ef2)\"" pod="openshift-machine-config-operator/machine-config-daemon-82xjz" podUID="e0c4af12-edda-41d7-9b44-f87396174ef2" Dec 13 07:42:23 crc kubenswrapper[4971]: I1213 07:42:23.946932 4971 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-mmm7v" event={"ID":"94cbb455-7651-4d43-ad24-495457496a2d","Type":"ContainerStarted","Data":"c900ba399fc79d9e2b76a7c47a31893069e03b2a3dca86723c690a45f9392d52"} Dec 13 07:42:24 crc kubenswrapper[4971]: I1213 07:42:24.980308 4971 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-mmm7v" podStartSLOduration=3.852019216 podStartE2EDuration="14.980290981s" podCreationTimestamp="2025-12-13 07:42:10 +0000 UTC" firstStartedPulling="2025-12-13 07:42:12.145053174 +0000 UTC m=+3188.749462642" lastFinishedPulling="2025-12-13 07:42:23.273324919 +0000 UTC m=+3199.877734407" observedRunningTime="2025-12-13 07:42:24.968138398 +0000 UTC m=+3201.572547846" watchObservedRunningTime="2025-12-13 07:42:24.980290981 +0000 UTC m=+3201.584700429" Dec 13 07:42:30 crc kubenswrapper[4971]: I1213 07:42:30.769002 4971 scope.go:117] "RemoveContainer" containerID="09d0904c76ce78aa0b074447b1e64cf63160f573da086f934c8cb712615b71d4" Dec 13 07:42:30 crc kubenswrapper[4971]: E1213 07:42:30.770085 4971 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-82xjz_openshift-machine-config-operator(e0c4af12-edda-41d7-9b44-f87396174ef2)\"" pod="openshift-machine-config-operator/machine-config-daemon-82xjz" podUID="e0c4af12-edda-41d7-9b44-f87396174ef2" Dec 13 07:42:39 crc kubenswrapper[4971]: I1213 07:42:39.133828 4971 generic.go:334] "Generic (PLEG): container finished" podID="94cbb455-7651-4d43-ad24-495457496a2d" containerID="c900ba399fc79d9e2b76a7c47a31893069e03b2a3dca86723c690a45f9392d52" exitCode=0 Dec 13 07:42:39 crc kubenswrapper[4971]: I1213 07:42:39.133898 4971 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-mmm7v" event={"ID":"94cbb455-7651-4d43-ad24-495457496a2d","Type":"ContainerDied","Data":"c900ba399fc79d9e2b76a7c47a31893069e03b2a3dca86723c690a45f9392d52"} Dec 13 07:42:40 crc kubenswrapper[4971]: I1213 07:42:40.585425 4971 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-mmm7v" Dec 13 07:42:40 crc kubenswrapper[4971]: I1213 07:42:40.765191 4971 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/94cbb455-7651-4d43-ad24-495457496a2d-ssh-key\") pod \"94cbb455-7651-4d43-ad24-495457496a2d\" (UID: \"94cbb455-7651-4d43-ad24-495457496a2d\") " Dec 13 07:42:40 crc kubenswrapper[4971]: I1213 07:42:40.765234 4971 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-2nmwh\" (UniqueName: \"kubernetes.io/projected/94cbb455-7651-4d43-ad24-495457496a2d-kube-api-access-2nmwh\") pod \"94cbb455-7651-4d43-ad24-495457496a2d\" (UID: \"94cbb455-7651-4d43-ad24-495457496a2d\") " Dec 13 07:42:40 crc kubenswrapper[4971]: I1213 07:42:40.765294 4971 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/94cbb455-7651-4d43-ad24-495457496a2d-inventory\") pod \"94cbb455-7651-4d43-ad24-495457496a2d\" (UID: \"94cbb455-7651-4d43-ad24-495457496a2d\") " Dec 13 07:42:40 crc kubenswrapper[4971]: I1213 07:42:40.765393 4971 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/94cbb455-7651-4d43-ad24-495457496a2d-ceph\") pod \"94cbb455-7651-4d43-ad24-495457496a2d\" (UID: \"94cbb455-7651-4d43-ad24-495457496a2d\") " Dec 13 07:42:40 crc kubenswrapper[4971]: I1213 07:42:40.771543 4971 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/94cbb455-7651-4d43-ad24-495457496a2d-ceph" (OuterVolumeSpecName: "ceph") pod "94cbb455-7651-4d43-ad24-495457496a2d" (UID: "94cbb455-7651-4d43-ad24-495457496a2d"). InnerVolumeSpecName "ceph". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 13 07:42:40 crc kubenswrapper[4971]: I1213 07:42:40.773680 4971 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/94cbb455-7651-4d43-ad24-495457496a2d-kube-api-access-2nmwh" (OuterVolumeSpecName: "kube-api-access-2nmwh") pod "94cbb455-7651-4d43-ad24-495457496a2d" (UID: "94cbb455-7651-4d43-ad24-495457496a2d"). InnerVolumeSpecName "kube-api-access-2nmwh". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 13 07:42:40 crc kubenswrapper[4971]: I1213 07:42:40.796824 4971 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/94cbb455-7651-4d43-ad24-495457496a2d-inventory" (OuterVolumeSpecName: "inventory") pod "94cbb455-7651-4d43-ad24-495457496a2d" (UID: "94cbb455-7651-4d43-ad24-495457496a2d"). InnerVolumeSpecName "inventory". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 13 07:42:40 crc kubenswrapper[4971]: I1213 07:42:40.813725 4971 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/94cbb455-7651-4d43-ad24-495457496a2d-ssh-key" (OuterVolumeSpecName: "ssh-key") pod "94cbb455-7651-4d43-ad24-495457496a2d" (UID: "94cbb455-7651-4d43-ad24-495457496a2d"). InnerVolumeSpecName "ssh-key". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 13 07:42:40 crc kubenswrapper[4971]: I1213 07:42:40.868123 4971 reconciler_common.go:293] "Volume detached for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/94cbb455-7651-4d43-ad24-495457496a2d-ssh-key\") on node \"crc\" DevicePath \"\"" Dec 13 07:42:40 crc kubenswrapper[4971]: I1213 07:42:40.868676 4971 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-2nmwh\" (UniqueName: \"kubernetes.io/projected/94cbb455-7651-4d43-ad24-495457496a2d-kube-api-access-2nmwh\") on node \"crc\" DevicePath \"\"" Dec 13 07:42:40 crc kubenswrapper[4971]: I1213 07:42:40.869043 4971 reconciler_common.go:293] "Volume detached for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/94cbb455-7651-4d43-ad24-495457496a2d-inventory\") on node \"crc\" DevicePath \"\"" Dec 13 07:42:40 crc kubenswrapper[4971]: I1213 07:42:40.869087 4971 reconciler_common.go:293] "Volume detached for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/94cbb455-7651-4d43-ad24-495457496a2d-ceph\") on node \"crc\" DevicePath \"\"" Dec 13 07:42:41 crc kubenswrapper[4971]: I1213 07:42:41.158572 4971 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-mmm7v" event={"ID":"94cbb455-7651-4d43-ad24-495457496a2d","Type":"ContainerDied","Data":"9954d38a1098cd82cf3b5cf3c0a82412973e68ce725670504824eae875c2a3dc"} Dec 13 07:42:41 crc kubenswrapper[4971]: I1213 07:42:41.158629 4971 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="9954d38a1098cd82cf3b5cf3c0a82412973e68ce725670504824eae875c2a3dc" Dec 13 07:42:41 crc kubenswrapper[4971]: I1213 07:42:41.158727 4971 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-mmm7v" Dec 13 07:42:41 crc kubenswrapper[4971]: I1213 07:42:41.286196 4971 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/install-certs-edpm-deployment-openstack-edpm-ipam-lkhcs"] Dec 13 07:42:41 crc kubenswrapper[4971]: E1213 07:42:41.286778 4971 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="94cbb455-7651-4d43-ad24-495457496a2d" containerName="reboot-os-edpm-deployment-openstack-edpm-ipam" Dec 13 07:42:41 crc kubenswrapper[4971]: I1213 07:42:41.286808 4971 state_mem.go:107] "Deleted CPUSet assignment" podUID="94cbb455-7651-4d43-ad24-495457496a2d" containerName="reboot-os-edpm-deployment-openstack-edpm-ipam" Dec 13 07:42:41 crc kubenswrapper[4971]: I1213 07:42:41.287094 4971 memory_manager.go:354] "RemoveStaleState removing state" podUID="94cbb455-7651-4d43-ad24-495457496a2d" containerName="reboot-os-edpm-deployment-openstack-edpm-ipam" Dec 13 07:42:41 crc kubenswrapper[4971]: I1213 07:42:41.287983 4971 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-lkhcs" Dec 13 07:42:41 crc kubenswrapper[4971]: I1213 07:42:41.296252 4971 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-edpm-ipam-ovn-default-certs-0" Dec 13 07:42:41 crc kubenswrapper[4971]: I1213 07:42:41.296971 4971 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceph-conf-files" Dec 13 07:42:41 crc kubenswrapper[4971]: I1213 07:42:41.297485 4971 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplanenodeset-openstack-edpm-ipam" Dec 13 07:42:41 crc kubenswrapper[4971]: I1213 07:42:41.297700 4971 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-edpm-ipam-libvirt-default-certs-0" Dec 13 07:42:41 crc kubenswrapper[4971]: I1213 07:42:41.298476 4971 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-edpm-ipam-neutron-metadata-default-certs-0" Dec 13 07:42:41 crc kubenswrapper[4971]: I1213 07:42:41.298636 4971 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-aee-default-env" Dec 13 07:42:41 crc kubenswrapper[4971]: I1213 07:42:41.299120 4971 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/install-certs-edpm-deployment-openstack-edpm-ipam-lkhcs"] Dec 13 07:42:41 crc kubenswrapper[4971]: I1213 07:42:41.299584 4971 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplane-ansible-ssh-private-key-secret" Dec 13 07:42:41 crc kubenswrapper[4971]: I1213 07:42:41.301278 4971 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-edpm-ipam-dockercfg-zb2j5" Dec 13 07:42:41 crc kubenswrapper[4971]: I1213 07:42:41.479569 4971 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"openstack-edpm-ipam-ovn-default-certs-0\" (UniqueName: \"kubernetes.io/projected/2f26ed2c-695e-4425-85cc-6e03e59b39db-openstack-edpm-ipam-ovn-default-certs-0\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-lkhcs\" (UID: \"2f26ed2c-695e-4425-85cc-6e03e59b39db\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-lkhcs" Dec 13 07:42:41 crc kubenswrapper[4971]: I1213 07:42:41.479615 4971 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-r9blj\" (UniqueName: \"kubernetes.io/projected/2f26ed2c-695e-4425-85cc-6e03e59b39db-kube-api-access-r9blj\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-lkhcs\" (UID: \"2f26ed2c-695e-4425-85cc-6e03e59b39db\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-lkhcs" Dec 13 07:42:41 crc kubenswrapper[4971]: I1213 07:42:41.479651 4971 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"libvirt-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/2f26ed2c-695e-4425-85cc-6e03e59b39db-libvirt-combined-ca-bundle\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-lkhcs\" (UID: \"2f26ed2c-695e-4425-85cc-6e03e59b39db\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-lkhcs" Dec 13 07:42:41 crc kubenswrapper[4971]: I1213 07:42:41.479679 4971 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/2f26ed2c-695e-4425-85cc-6e03e59b39db-inventory\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-lkhcs\" (UID: \"2f26ed2c-695e-4425-85cc-6e03e59b39db\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-lkhcs" Dec 13 07:42:41 crc kubenswrapper[4971]: I1213 07:42:41.479699 4971 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"nova-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/2f26ed2c-695e-4425-85cc-6e03e59b39db-nova-combined-ca-bundle\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-lkhcs\" (UID: \"2f26ed2c-695e-4425-85cc-6e03e59b39db\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-lkhcs" Dec 13 07:42:41 crc kubenswrapper[4971]: I1213 07:42:41.479852 4971 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"repo-setup-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/2f26ed2c-695e-4425-85cc-6e03e59b39db-repo-setup-combined-ca-bundle\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-lkhcs\" (UID: \"2f26ed2c-695e-4425-85cc-6e03e59b39db\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-lkhcs" Dec 13 07:42:41 crc kubenswrapper[4971]: I1213 07:42:41.479933 4971 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovn-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/2f26ed2c-695e-4425-85cc-6e03e59b39db-ovn-combined-ca-bundle\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-lkhcs\" (UID: \"2f26ed2c-695e-4425-85cc-6e03e59b39db\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-lkhcs" Dec 13 07:42:41 crc kubenswrapper[4971]: I1213 07:42:41.479982 4971 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"openstack-edpm-ipam-libvirt-default-certs-0\" (UniqueName: \"kubernetes.io/projected/2f26ed2c-695e-4425-85cc-6e03e59b39db-openstack-edpm-ipam-libvirt-default-certs-0\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-lkhcs\" (UID: \"2f26ed2c-695e-4425-85cc-6e03e59b39db\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-lkhcs" Dec 13 07:42:41 crc kubenswrapper[4971]: I1213 07:42:41.480003 4971 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"openstack-edpm-ipam-neutron-metadata-default-certs-0\" (UniqueName: \"kubernetes.io/projected/2f26ed2c-695e-4425-85cc-6e03e59b39db-openstack-edpm-ipam-neutron-metadata-default-certs-0\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-lkhcs\" (UID: \"2f26ed2c-695e-4425-85cc-6e03e59b39db\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-lkhcs" Dec 13 07:42:41 crc kubenswrapper[4971]: I1213 07:42:41.480028 4971 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"bootstrap-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/2f26ed2c-695e-4425-85cc-6e03e59b39db-bootstrap-combined-ca-bundle\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-lkhcs\" (UID: \"2f26ed2c-695e-4425-85cc-6e03e59b39db\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-lkhcs" Dec 13 07:42:41 crc kubenswrapper[4971]: I1213 07:42:41.480102 4971 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/2f26ed2c-695e-4425-85cc-6e03e59b39db-ssh-key\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-lkhcs\" (UID: \"2f26ed2c-695e-4425-85cc-6e03e59b39db\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-lkhcs" Dec 13 07:42:41 crc kubenswrapper[4971]: I1213 07:42:41.480176 4971 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/2f26ed2c-695e-4425-85cc-6e03e59b39db-ceph\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-lkhcs\" (UID: \"2f26ed2c-695e-4425-85cc-6e03e59b39db\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-lkhcs" Dec 13 07:42:41 crc kubenswrapper[4971]: I1213 07:42:41.480231 4971 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"neutron-metadata-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/2f26ed2c-695e-4425-85cc-6e03e59b39db-neutron-metadata-combined-ca-bundle\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-lkhcs\" (UID: \"2f26ed2c-695e-4425-85cc-6e03e59b39db\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-lkhcs" Dec 13 07:42:41 crc kubenswrapper[4971]: I1213 07:42:41.582736 4971 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"openstack-edpm-ipam-ovn-default-certs-0\" (UniqueName: \"kubernetes.io/projected/2f26ed2c-695e-4425-85cc-6e03e59b39db-openstack-edpm-ipam-ovn-default-certs-0\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-lkhcs\" (UID: \"2f26ed2c-695e-4425-85cc-6e03e59b39db\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-lkhcs" Dec 13 07:42:41 crc kubenswrapper[4971]: I1213 07:42:41.582845 4971 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-r9blj\" (UniqueName: \"kubernetes.io/projected/2f26ed2c-695e-4425-85cc-6e03e59b39db-kube-api-access-r9blj\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-lkhcs\" (UID: \"2f26ed2c-695e-4425-85cc-6e03e59b39db\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-lkhcs" Dec 13 07:42:41 crc kubenswrapper[4971]: I1213 07:42:41.582933 4971 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"libvirt-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/2f26ed2c-695e-4425-85cc-6e03e59b39db-libvirt-combined-ca-bundle\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-lkhcs\" (UID: \"2f26ed2c-695e-4425-85cc-6e03e59b39db\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-lkhcs" Dec 13 07:42:41 crc kubenswrapper[4971]: I1213 07:42:41.583005 4971 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/2f26ed2c-695e-4425-85cc-6e03e59b39db-inventory\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-lkhcs\" (UID: \"2f26ed2c-695e-4425-85cc-6e03e59b39db\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-lkhcs" Dec 13 07:42:41 crc kubenswrapper[4971]: I1213 07:42:41.583057 4971 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nova-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/2f26ed2c-695e-4425-85cc-6e03e59b39db-nova-combined-ca-bundle\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-lkhcs\" (UID: \"2f26ed2c-695e-4425-85cc-6e03e59b39db\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-lkhcs" Dec 13 07:42:41 crc kubenswrapper[4971]: I1213 07:42:41.583131 4971 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"repo-setup-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/2f26ed2c-695e-4425-85cc-6e03e59b39db-repo-setup-combined-ca-bundle\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-lkhcs\" (UID: \"2f26ed2c-695e-4425-85cc-6e03e59b39db\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-lkhcs" Dec 13 07:42:41 crc kubenswrapper[4971]: I1213 07:42:41.583209 4971 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovn-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/2f26ed2c-695e-4425-85cc-6e03e59b39db-ovn-combined-ca-bundle\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-lkhcs\" (UID: \"2f26ed2c-695e-4425-85cc-6e03e59b39db\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-lkhcs" Dec 13 07:42:41 crc kubenswrapper[4971]: I1213 07:42:41.583283 4971 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"openstack-edpm-ipam-libvirt-default-certs-0\" (UniqueName: \"kubernetes.io/projected/2f26ed2c-695e-4425-85cc-6e03e59b39db-openstack-edpm-ipam-libvirt-default-certs-0\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-lkhcs\" (UID: \"2f26ed2c-695e-4425-85cc-6e03e59b39db\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-lkhcs" Dec 13 07:42:41 crc kubenswrapper[4971]: I1213 07:42:41.583334 4971 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"openstack-edpm-ipam-neutron-metadata-default-certs-0\" (UniqueName: \"kubernetes.io/projected/2f26ed2c-695e-4425-85cc-6e03e59b39db-openstack-edpm-ipam-neutron-metadata-default-certs-0\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-lkhcs\" (UID: \"2f26ed2c-695e-4425-85cc-6e03e59b39db\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-lkhcs" Dec 13 07:42:41 crc kubenswrapper[4971]: I1213 07:42:41.583388 4971 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"bootstrap-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/2f26ed2c-695e-4425-85cc-6e03e59b39db-bootstrap-combined-ca-bundle\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-lkhcs\" (UID: \"2f26ed2c-695e-4425-85cc-6e03e59b39db\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-lkhcs" Dec 13 07:42:41 crc kubenswrapper[4971]: I1213 07:42:41.583505 4971 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/2f26ed2c-695e-4425-85cc-6e03e59b39db-ssh-key\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-lkhcs\" (UID: \"2f26ed2c-695e-4425-85cc-6e03e59b39db\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-lkhcs" Dec 13 07:42:41 crc kubenswrapper[4971]: I1213 07:42:41.583643 4971 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/2f26ed2c-695e-4425-85cc-6e03e59b39db-ceph\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-lkhcs\" (UID: \"2f26ed2c-695e-4425-85cc-6e03e59b39db\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-lkhcs" Dec 13 07:42:41 crc kubenswrapper[4971]: I1213 07:42:41.583724 4971 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"neutron-metadata-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/2f26ed2c-695e-4425-85cc-6e03e59b39db-neutron-metadata-combined-ca-bundle\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-lkhcs\" (UID: \"2f26ed2c-695e-4425-85cc-6e03e59b39db\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-lkhcs" Dec 13 07:42:41 crc kubenswrapper[4971]: I1213 07:42:41.589037 4971 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/2f26ed2c-695e-4425-85cc-6e03e59b39db-inventory\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-lkhcs\" (UID: \"2f26ed2c-695e-4425-85cc-6e03e59b39db\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-lkhcs" Dec 13 07:42:41 crc kubenswrapper[4971]: I1213 07:42:41.589222 4971 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"nova-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/2f26ed2c-695e-4425-85cc-6e03e59b39db-nova-combined-ca-bundle\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-lkhcs\" (UID: \"2f26ed2c-695e-4425-85cc-6e03e59b39db\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-lkhcs" Dec 13 07:42:41 crc kubenswrapper[4971]: I1213 07:42:41.589882 4971 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovn-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/2f26ed2c-695e-4425-85cc-6e03e59b39db-ovn-combined-ca-bundle\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-lkhcs\" (UID: \"2f26ed2c-695e-4425-85cc-6e03e59b39db\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-lkhcs" Dec 13 07:42:41 crc kubenswrapper[4971]: I1213 07:42:41.590143 4971 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/2f26ed2c-695e-4425-85cc-6e03e59b39db-ssh-key\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-lkhcs\" (UID: \"2f26ed2c-695e-4425-85cc-6e03e59b39db\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-lkhcs" Dec 13 07:42:41 crc kubenswrapper[4971]: I1213 07:42:41.590672 4971 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"neutron-metadata-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/2f26ed2c-695e-4425-85cc-6e03e59b39db-neutron-metadata-combined-ca-bundle\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-lkhcs\" (UID: \"2f26ed2c-695e-4425-85cc-6e03e59b39db\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-lkhcs" Dec 13 07:42:41 crc kubenswrapper[4971]: I1213 07:42:41.590727 4971 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"libvirt-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/2f26ed2c-695e-4425-85cc-6e03e59b39db-libvirt-combined-ca-bundle\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-lkhcs\" (UID: \"2f26ed2c-695e-4425-85cc-6e03e59b39db\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-lkhcs" Dec 13 07:42:41 crc kubenswrapper[4971]: I1213 07:42:41.591500 4971 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/2f26ed2c-695e-4425-85cc-6e03e59b39db-ceph\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-lkhcs\" (UID: \"2f26ed2c-695e-4425-85cc-6e03e59b39db\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-lkhcs" Dec 13 07:42:41 crc kubenswrapper[4971]: I1213 07:42:41.592060 4971 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"bootstrap-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/2f26ed2c-695e-4425-85cc-6e03e59b39db-bootstrap-combined-ca-bundle\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-lkhcs\" (UID: \"2f26ed2c-695e-4425-85cc-6e03e59b39db\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-lkhcs" Dec 13 07:42:41 crc kubenswrapper[4971]: I1213 07:42:41.592804 4971 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"repo-setup-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/2f26ed2c-695e-4425-85cc-6e03e59b39db-repo-setup-combined-ca-bundle\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-lkhcs\" (UID: \"2f26ed2c-695e-4425-85cc-6e03e59b39db\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-lkhcs" Dec 13 07:42:41 crc kubenswrapper[4971]: I1213 07:42:41.594373 4971 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"openstack-edpm-ipam-ovn-default-certs-0\" (UniqueName: \"kubernetes.io/projected/2f26ed2c-695e-4425-85cc-6e03e59b39db-openstack-edpm-ipam-ovn-default-certs-0\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-lkhcs\" (UID: \"2f26ed2c-695e-4425-85cc-6e03e59b39db\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-lkhcs" Dec 13 07:42:41 crc kubenswrapper[4971]: I1213 07:42:41.596163 4971 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"openstack-edpm-ipam-libvirt-default-certs-0\" (UniqueName: \"kubernetes.io/projected/2f26ed2c-695e-4425-85cc-6e03e59b39db-openstack-edpm-ipam-libvirt-default-certs-0\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-lkhcs\" (UID: \"2f26ed2c-695e-4425-85cc-6e03e59b39db\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-lkhcs" Dec 13 07:42:41 crc kubenswrapper[4971]: I1213 07:42:41.596984 4971 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"openstack-edpm-ipam-neutron-metadata-default-certs-0\" (UniqueName: \"kubernetes.io/projected/2f26ed2c-695e-4425-85cc-6e03e59b39db-openstack-edpm-ipam-neutron-metadata-default-certs-0\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-lkhcs\" (UID: \"2f26ed2c-695e-4425-85cc-6e03e59b39db\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-lkhcs" Dec 13 07:42:41 crc kubenswrapper[4971]: I1213 07:42:41.598493 4971 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-r9blj\" (UniqueName: \"kubernetes.io/projected/2f26ed2c-695e-4425-85cc-6e03e59b39db-kube-api-access-r9blj\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-lkhcs\" (UID: \"2f26ed2c-695e-4425-85cc-6e03e59b39db\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-lkhcs" Dec 13 07:42:41 crc kubenswrapper[4971]: I1213 07:42:41.659684 4971 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-lkhcs" Dec 13 07:42:42 crc kubenswrapper[4971]: I1213 07:42:42.228774 4971 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/install-certs-edpm-deployment-openstack-edpm-ipam-lkhcs"] Dec 13 07:42:43 crc kubenswrapper[4971]: I1213 07:42:43.182717 4971 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-lkhcs" event={"ID":"2f26ed2c-695e-4425-85cc-6e03e59b39db","Type":"ContainerStarted","Data":"bc347832235712f07d8accbf39bb77ba6b808fea1d49885891d73d8d2d087ef5"} Dec 13 07:42:44 crc kubenswrapper[4971]: I1213 07:42:44.769910 4971 scope.go:117] "RemoveContainer" containerID="09d0904c76ce78aa0b074447b1e64cf63160f573da086f934c8cb712615b71d4" Dec 13 07:42:44 crc kubenswrapper[4971]: E1213 07:42:44.770833 4971 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-82xjz_openshift-machine-config-operator(e0c4af12-edda-41d7-9b44-f87396174ef2)\"" pod="openshift-machine-config-operator/machine-config-daemon-82xjz" podUID="e0c4af12-edda-41d7-9b44-f87396174ef2" Dec 13 07:42:49 crc kubenswrapper[4971]: I1213 07:42:49.261855 4971 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-lkhcs" event={"ID":"2f26ed2c-695e-4425-85cc-6e03e59b39db","Type":"ContainerStarted","Data":"b704fafcb835517ce26e6502687682cb0dada4231f45f8a9ebbeb901d2ae3b9c"} Dec 13 07:42:50 crc kubenswrapper[4971]: I1213 07:42:50.298900 4971 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-lkhcs" podStartSLOduration=3.5810887510000002 podStartE2EDuration="9.298853907s" podCreationTimestamp="2025-12-13 07:42:41 +0000 UTC" firstStartedPulling="2025-12-13 07:42:42.234931757 +0000 UTC m=+3218.839341235" lastFinishedPulling="2025-12-13 07:42:47.952696913 +0000 UTC m=+3224.557106391" observedRunningTime="2025-12-13 07:42:50.295861024 +0000 UTC m=+3226.900270502" watchObservedRunningTime="2025-12-13 07:42:50.298853907 +0000 UTC m=+3226.903263375" Dec 13 07:42:57 crc kubenswrapper[4971]: I1213 07:42:57.769198 4971 scope.go:117] "RemoveContainer" containerID="09d0904c76ce78aa0b074447b1e64cf63160f573da086f934c8cb712615b71d4" Dec 13 07:42:57 crc kubenswrapper[4971]: E1213 07:42:57.770302 4971 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-82xjz_openshift-machine-config-operator(e0c4af12-edda-41d7-9b44-f87396174ef2)\"" pod="openshift-machine-config-operator/machine-config-daemon-82xjz" podUID="e0c4af12-edda-41d7-9b44-f87396174ef2" Dec 13 07:43:10 crc kubenswrapper[4971]: I1213 07:43:10.768953 4971 scope.go:117] "RemoveContainer" containerID="09d0904c76ce78aa0b074447b1e64cf63160f573da086f934c8cb712615b71d4" Dec 13 07:43:10 crc kubenswrapper[4971]: E1213 07:43:10.769733 4971 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-82xjz_openshift-machine-config-operator(e0c4af12-edda-41d7-9b44-f87396174ef2)\"" pod="openshift-machine-config-operator/machine-config-daemon-82xjz" podUID="e0c4af12-edda-41d7-9b44-f87396174ef2" Dec 13 07:43:23 crc kubenswrapper[4971]: I1213 07:43:23.571283 4971 generic.go:334] "Generic (PLEG): container finished" podID="2f26ed2c-695e-4425-85cc-6e03e59b39db" containerID="b704fafcb835517ce26e6502687682cb0dada4231f45f8a9ebbeb901d2ae3b9c" exitCode=0 Dec 13 07:43:23 crc kubenswrapper[4971]: I1213 07:43:23.571380 4971 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-lkhcs" event={"ID":"2f26ed2c-695e-4425-85cc-6e03e59b39db","Type":"ContainerDied","Data":"b704fafcb835517ce26e6502687682cb0dada4231f45f8a9ebbeb901d2ae3b9c"} Dec 13 07:43:24 crc kubenswrapper[4971]: I1213 07:43:24.966430 4971 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-lkhcs" Dec 13 07:43:25 crc kubenswrapper[4971]: I1213 07:43:25.118746 4971 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"openstack-edpm-ipam-neutron-metadata-default-certs-0\" (UniqueName: \"kubernetes.io/projected/2f26ed2c-695e-4425-85cc-6e03e59b39db-openstack-edpm-ipam-neutron-metadata-default-certs-0\") pod \"2f26ed2c-695e-4425-85cc-6e03e59b39db\" (UID: \"2f26ed2c-695e-4425-85cc-6e03e59b39db\") " Dec 13 07:43:25 crc kubenswrapper[4971]: I1213 07:43:25.119816 4971 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/2f26ed2c-695e-4425-85cc-6e03e59b39db-ceph\") pod \"2f26ed2c-695e-4425-85cc-6e03e59b39db\" (UID: \"2f26ed2c-695e-4425-85cc-6e03e59b39db\") " Dec 13 07:43:25 crc kubenswrapper[4971]: I1213 07:43:25.119879 4971 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"openstack-edpm-ipam-ovn-default-certs-0\" (UniqueName: \"kubernetes.io/projected/2f26ed2c-695e-4425-85cc-6e03e59b39db-openstack-edpm-ipam-ovn-default-certs-0\") pod \"2f26ed2c-695e-4425-85cc-6e03e59b39db\" (UID: \"2f26ed2c-695e-4425-85cc-6e03e59b39db\") " Dec 13 07:43:25 crc kubenswrapper[4971]: I1213 07:43:25.119905 4971 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"repo-setup-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/2f26ed2c-695e-4425-85cc-6e03e59b39db-repo-setup-combined-ca-bundle\") pod \"2f26ed2c-695e-4425-85cc-6e03e59b39db\" (UID: \"2f26ed2c-695e-4425-85cc-6e03e59b39db\") " Dec 13 07:43:25 crc kubenswrapper[4971]: I1213 07:43:25.119930 4971 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"libvirt-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/2f26ed2c-695e-4425-85cc-6e03e59b39db-libvirt-combined-ca-bundle\") pod \"2f26ed2c-695e-4425-85cc-6e03e59b39db\" (UID: \"2f26ed2c-695e-4425-85cc-6e03e59b39db\") " Dec 13 07:43:25 crc kubenswrapper[4971]: I1213 07:43:25.119965 4971 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"bootstrap-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/2f26ed2c-695e-4425-85cc-6e03e59b39db-bootstrap-combined-ca-bundle\") pod \"2f26ed2c-695e-4425-85cc-6e03e59b39db\" (UID: \"2f26ed2c-695e-4425-85cc-6e03e59b39db\") " Dec 13 07:43:25 crc kubenswrapper[4971]: I1213 07:43:25.120018 4971 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"openstack-edpm-ipam-libvirt-default-certs-0\" (UniqueName: \"kubernetes.io/projected/2f26ed2c-695e-4425-85cc-6e03e59b39db-openstack-edpm-ipam-libvirt-default-certs-0\") pod \"2f26ed2c-695e-4425-85cc-6e03e59b39db\" (UID: \"2f26ed2c-695e-4425-85cc-6e03e59b39db\") " Dec 13 07:43:25 crc kubenswrapper[4971]: I1213 07:43:25.120067 4971 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/2f26ed2c-695e-4425-85cc-6e03e59b39db-ssh-key\") pod \"2f26ed2c-695e-4425-85cc-6e03e59b39db\" (UID: \"2f26ed2c-695e-4425-85cc-6e03e59b39db\") " Dec 13 07:43:25 crc kubenswrapper[4971]: I1213 07:43:25.120104 4971 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"nova-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/2f26ed2c-695e-4425-85cc-6e03e59b39db-nova-combined-ca-bundle\") pod \"2f26ed2c-695e-4425-85cc-6e03e59b39db\" (UID: \"2f26ed2c-695e-4425-85cc-6e03e59b39db\") " Dec 13 07:43:25 crc kubenswrapper[4971]: I1213 07:43:25.120128 4971 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/2f26ed2c-695e-4425-85cc-6e03e59b39db-inventory\") pod \"2f26ed2c-695e-4425-85cc-6e03e59b39db\" (UID: \"2f26ed2c-695e-4425-85cc-6e03e59b39db\") " Dec 13 07:43:25 crc kubenswrapper[4971]: I1213 07:43:25.120164 4971 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"neutron-metadata-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/2f26ed2c-695e-4425-85cc-6e03e59b39db-neutron-metadata-combined-ca-bundle\") pod \"2f26ed2c-695e-4425-85cc-6e03e59b39db\" (UID: \"2f26ed2c-695e-4425-85cc-6e03e59b39db\") " Dec 13 07:43:25 crc kubenswrapper[4971]: I1213 07:43:25.120188 4971 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovn-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/2f26ed2c-695e-4425-85cc-6e03e59b39db-ovn-combined-ca-bundle\") pod \"2f26ed2c-695e-4425-85cc-6e03e59b39db\" (UID: \"2f26ed2c-695e-4425-85cc-6e03e59b39db\") " Dec 13 07:43:25 crc kubenswrapper[4971]: I1213 07:43:25.120210 4971 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-r9blj\" (UniqueName: \"kubernetes.io/projected/2f26ed2c-695e-4425-85cc-6e03e59b39db-kube-api-access-r9blj\") pod \"2f26ed2c-695e-4425-85cc-6e03e59b39db\" (UID: \"2f26ed2c-695e-4425-85cc-6e03e59b39db\") " Dec 13 07:43:25 crc kubenswrapper[4971]: I1213 07:43:25.125296 4971 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/2f26ed2c-695e-4425-85cc-6e03e59b39db-openstack-edpm-ipam-libvirt-default-certs-0" (OuterVolumeSpecName: "openstack-edpm-ipam-libvirt-default-certs-0") pod "2f26ed2c-695e-4425-85cc-6e03e59b39db" (UID: "2f26ed2c-695e-4425-85cc-6e03e59b39db"). InnerVolumeSpecName "openstack-edpm-ipam-libvirt-default-certs-0". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 13 07:43:25 crc kubenswrapper[4971]: I1213 07:43:25.126618 4971 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/2f26ed2c-695e-4425-85cc-6e03e59b39db-kube-api-access-r9blj" (OuterVolumeSpecName: "kube-api-access-r9blj") pod "2f26ed2c-695e-4425-85cc-6e03e59b39db" (UID: "2f26ed2c-695e-4425-85cc-6e03e59b39db"). InnerVolumeSpecName "kube-api-access-r9blj". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 13 07:43:25 crc kubenswrapper[4971]: I1213 07:43:25.126778 4971 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/2f26ed2c-695e-4425-85cc-6e03e59b39db-bootstrap-combined-ca-bundle" (OuterVolumeSpecName: "bootstrap-combined-ca-bundle") pod "2f26ed2c-695e-4425-85cc-6e03e59b39db" (UID: "2f26ed2c-695e-4425-85cc-6e03e59b39db"). InnerVolumeSpecName "bootstrap-combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 13 07:43:25 crc kubenswrapper[4971]: I1213 07:43:25.127087 4971 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/2f26ed2c-695e-4425-85cc-6e03e59b39db-openstack-edpm-ipam-neutron-metadata-default-certs-0" (OuterVolumeSpecName: "openstack-edpm-ipam-neutron-metadata-default-certs-0") pod "2f26ed2c-695e-4425-85cc-6e03e59b39db" (UID: "2f26ed2c-695e-4425-85cc-6e03e59b39db"). InnerVolumeSpecName "openstack-edpm-ipam-neutron-metadata-default-certs-0". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 13 07:43:25 crc kubenswrapper[4971]: I1213 07:43:25.128651 4971 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/2f26ed2c-695e-4425-85cc-6e03e59b39db-repo-setup-combined-ca-bundle" (OuterVolumeSpecName: "repo-setup-combined-ca-bundle") pod "2f26ed2c-695e-4425-85cc-6e03e59b39db" (UID: "2f26ed2c-695e-4425-85cc-6e03e59b39db"). InnerVolumeSpecName "repo-setup-combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 13 07:43:25 crc kubenswrapper[4971]: I1213 07:43:25.131045 4971 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/2f26ed2c-695e-4425-85cc-6e03e59b39db-libvirt-combined-ca-bundle" (OuterVolumeSpecName: "libvirt-combined-ca-bundle") pod "2f26ed2c-695e-4425-85cc-6e03e59b39db" (UID: "2f26ed2c-695e-4425-85cc-6e03e59b39db"). InnerVolumeSpecName "libvirt-combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 13 07:43:25 crc kubenswrapper[4971]: I1213 07:43:25.134910 4971 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/2f26ed2c-695e-4425-85cc-6e03e59b39db-nova-combined-ca-bundle" (OuterVolumeSpecName: "nova-combined-ca-bundle") pod "2f26ed2c-695e-4425-85cc-6e03e59b39db" (UID: "2f26ed2c-695e-4425-85cc-6e03e59b39db"). InnerVolumeSpecName "nova-combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 13 07:43:25 crc kubenswrapper[4971]: I1213 07:43:25.135101 4971 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/2f26ed2c-695e-4425-85cc-6e03e59b39db-openstack-edpm-ipam-ovn-default-certs-0" (OuterVolumeSpecName: "openstack-edpm-ipam-ovn-default-certs-0") pod "2f26ed2c-695e-4425-85cc-6e03e59b39db" (UID: "2f26ed2c-695e-4425-85cc-6e03e59b39db"). InnerVolumeSpecName "openstack-edpm-ipam-ovn-default-certs-0". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 13 07:43:25 crc kubenswrapper[4971]: I1213 07:43:25.135796 4971 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/2f26ed2c-695e-4425-85cc-6e03e59b39db-ovn-combined-ca-bundle" (OuterVolumeSpecName: "ovn-combined-ca-bundle") pod "2f26ed2c-695e-4425-85cc-6e03e59b39db" (UID: "2f26ed2c-695e-4425-85cc-6e03e59b39db"). InnerVolumeSpecName "ovn-combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 13 07:43:25 crc kubenswrapper[4971]: I1213 07:43:25.139338 4971 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/2f26ed2c-695e-4425-85cc-6e03e59b39db-ceph" (OuterVolumeSpecName: "ceph") pod "2f26ed2c-695e-4425-85cc-6e03e59b39db" (UID: "2f26ed2c-695e-4425-85cc-6e03e59b39db"). InnerVolumeSpecName "ceph". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 13 07:43:25 crc kubenswrapper[4971]: I1213 07:43:25.148845 4971 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/2f26ed2c-695e-4425-85cc-6e03e59b39db-neutron-metadata-combined-ca-bundle" (OuterVolumeSpecName: "neutron-metadata-combined-ca-bundle") pod "2f26ed2c-695e-4425-85cc-6e03e59b39db" (UID: "2f26ed2c-695e-4425-85cc-6e03e59b39db"). InnerVolumeSpecName "neutron-metadata-combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 13 07:43:25 crc kubenswrapper[4971]: I1213 07:43:25.153442 4971 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/2f26ed2c-695e-4425-85cc-6e03e59b39db-ssh-key" (OuterVolumeSpecName: "ssh-key") pod "2f26ed2c-695e-4425-85cc-6e03e59b39db" (UID: "2f26ed2c-695e-4425-85cc-6e03e59b39db"). InnerVolumeSpecName "ssh-key". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 13 07:43:25 crc kubenswrapper[4971]: I1213 07:43:25.167902 4971 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/2f26ed2c-695e-4425-85cc-6e03e59b39db-inventory" (OuterVolumeSpecName: "inventory") pod "2f26ed2c-695e-4425-85cc-6e03e59b39db" (UID: "2f26ed2c-695e-4425-85cc-6e03e59b39db"). InnerVolumeSpecName "inventory". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 13 07:43:25 crc kubenswrapper[4971]: I1213 07:43:25.221888 4971 reconciler_common.go:293] "Volume detached for volume \"bootstrap-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/2f26ed2c-695e-4425-85cc-6e03e59b39db-bootstrap-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 13 07:43:25 crc kubenswrapper[4971]: I1213 07:43:25.221923 4971 reconciler_common.go:293] "Volume detached for volume \"openstack-edpm-ipam-libvirt-default-certs-0\" (UniqueName: \"kubernetes.io/projected/2f26ed2c-695e-4425-85cc-6e03e59b39db-openstack-edpm-ipam-libvirt-default-certs-0\") on node \"crc\" DevicePath \"\"" Dec 13 07:43:25 crc kubenswrapper[4971]: I1213 07:43:25.221938 4971 reconciler_common.go:293] "Volume detached for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/2f26ed2c-695e-4425-85cc-6e03e59b39db-ssh-key\") on node \"crc\" DevicePath \"\"" Dec 13 07:43:25 crc kubenswrapper[4971]: I1213 07:43:25.221949 4971 reconciler_common.go:293] "Volume detached for volume \"nova-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/2f26ed2c-695e-4425-85cc-6e03e59b39db-nova-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 13 07:43:25 crc kubenswrapper[4971]: I1213 07:43:25.221960 4971 reconciler_common.go:293] "Volume detached for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/2f26ed2c-695e-4425-85cc-6e03e59b39db-inventory\") on node \"crc\" DevicePath \"\"" Dec 13 07:43:25 crc kubenswrapper[4971]: I1213 07:43:25.221969 4971 reconciler_common.go:293] "Volume detached for volume \"neutron-metadata-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/2f26ed2c-695e-4425-85cc-6e03e59b39db-neutron-metadata-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 13 07:43:25 crc kubenswrapper[4971]: I1213 07:43:25.221981 4971 reconciler_common.go:293] "Volume detached for volume \"ovn-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/2f26ed2c-695e-4425-85cc-6e03e59b39db-ovn-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 13 07:43:25 crc kubenswrapper[4971]: I1213 07:43:25.221992 4971 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-r9blj\" (UniqueName: \"kubernetes.io/projected/2f26ed2c-695e-4425-85cc-6e03e59b39db-kube-api-access-r9blj\") on node \"crc\" DevicePath \"\"" Dec 13 07:43:25 crc kubenswrapper[4971]: I1213 07:43:25.222002 4971 reconciler_common.go:293] "Volume detached for volume \"openstack-edpm-ipam-neutron-metadata-default-certs-0\" (UniqueName: \"kubernetes.io/projected/2f26ed2c-695e-4425-85cc-6e03e59b39db-openstack-edpm-ipam-neutron-metadata-default-certs-0\") on node \"crc\" DevicePath \"\"" Dec 13 07:43:25 crc kubenswrapper[4971]: I1213 07:43:25.222014 4971 reconciler_common.go:293] "Volume detached for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/2f26ed2c-695e-4425-85cc-6e03e59b39db-ceph\") on node \"crc\" DevicePath \"\"" Dec 13 07:43:25 crc kubenswrapper[4971]: I1213 07:43:25.222025 4971 reconciler_common.go:293] "Volume detached for volume \"openstack-edpm-ipam-ovn-default-certs-0\" (UniqueName: \"kubernetes.io/projected/2f26ed2c-695e-4425-85cc-6e03e59b39db-openstack-edpm-ipam-ovn-default-certs-0\") on node \"crc\" DevicePath \"\"" Dec 13 07:43:25 crc kubenswrapper[4971]: I1213 07:43:25.222037 4971 reconciler_common.go:293] "Volume detached for volume \"repo-setup-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/2f26ed2c-695e-4425-85cc-6e03e59b39db-repo-setup-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 13 07:43:25 crc kubenswrapper[4971]: I1213 07:43:25.222047 4971 reconciler_common.go:293] "Volume detached for volume \"libvirt-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/2f26ed2c-695e-4425-85cc-6e03e59b39db-libvirt-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 13 07:43:25 crc kubenswrapper[4971]: I1213 07:43:25.593666 4971 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-lkhcs" event={"ID":"2f26ed2c-695e-4425-85cc-6e03e59b39db","Type":"ContainerDied","Data":"bc347832235712f07d8accbf39bb77ba6b808fea1d49885891d73d8d2d087ef5"} Dec 13 07:43:25 crc kubenswrapper[4971]: I1213 07:43:25.593713 4971 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-lkhcs" Dec 13 07:43:25 crc kubenswrapper[4971]: I1213 07:43:25.593722 4971 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="bc347832235712f07d8accbf39bb77ba6b808fea1d49885891d73d8d2d087ef5" Dec 13 07:43:25 crc kubenswrapper[4971]: I1213 07:43:25.707794 4971 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ceph-client-edpm-deployment-openstack-edpm-ipam-7grkn"] Dec 13 07:43:25 crc kubenswrapper[4971]: E1213 07:43:25.708251 4971 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="2f26ed2c-695e-4425-85cc-6e03e59b39db" containerName="install-certs-edpm-deployment-openstack-edpm-ipam" Dec 13 07:43:25 crc kubenswrapper[4971]: I1213 07:43:25.708269 4971 state_mem.go:107] "Deleted CPUSet assignment" podUID="2f26ed2c-695e-4425-85cc-6e03e59b39db" containerName="install-certs-edpm-deployment-openstack-edpm-ipam" Dec 13 07:43:25 crc kubenswrapper[4971]: I1213 07:43:25.708448 4971 memory_manager.go:354] "RemoveStaleState removing state" podUID="2f26ed2c-695e-4425-85cc-6e03e59b39db" containerName="install-certs-edpm-deployment-openstack-edpm-ipam" Dec 13 07:43:25 crc kubenswrapper[4971]: I1213 07:43:25.709093 4971 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceph-client-edpm-deployment-openstack-edpm-ipam-7grkn" Dec 13 07:43:25 crc kubenswrapper[4971]: I1213 07:43:25.711881 4971 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplane-ansible-ssh-private-key-secret" Dec 13 07:43:25 crc kubenswrapper[4971]: I1213 07:43:25.712474 4971 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplanenodeset-openstack-edpm-ipam" Dec 13 07:43:25 crc kubenswrapper[4971]: I1213 07:43:25.712895 4971 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-aee-default-env" Dec 13 07:43:25 crc kubenswrapper[4971]: I1213 07:43:25.713386 4971 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-edpm-ipam-dockercfg-zb2j5" Dec 13 07:43:25 crc kubenswrapper[4971]: I1213 07:43:25.713871 4971 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceph-conf-files" Dec 13 07:43:25 crc kubenswrapper[4971]: I1213 07:43:25.716356 4971 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ceph-client-edpm-deployment-openstack-edpm-ipam-7grkn"] Dec 13 07:43:25 crc kubenswrapper[4971]: I1213 07:43:25.771060 4971 scope.go:117] "RemoveContainer" containerID="09d0904c76ce78aa0b074447b1e64cf63160f573da086f934c8cb712615b71d4" Dec 13 07:43:25 crc kubenswrapper[4971]: I1213 07:43:25.845308 4971 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-85prg\" (UniqueName: \"kubernetes.io/projected/42cff6d1-138e-4a16-a570-cdcbcbe6ab49-kube-api-access-85prg\") pod \"ceph-client-edpm-deployment-openstack-edpm-ipam-7grkn\" (UID: \"42cff6d1-138e-4a16-a570-cdcbcbe6ab49\") " pod="openstack/ceph-client-edpm-deployment-openstack-edpm-ipam-7grkn" Dec 13 07:43:25 crc kubenswrapper[4971]: I1213 07:43:25.845701 4971 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/42cff6d1-138e-4a16-a570-cdcbcbe6ab49-ceph\") pod \"ceph-client-edpm-deployment-openstack-edpm-ipam-7grkn\" (UID: \"42cff6d1-138e-4a16-a570-cdcbcbe6ab49\") " pod="openstack/ceph-client-edpm-deployment-openstack-edpm-ipam-7grkn" Dec 13 07:43:25 crc kubenswrapper[4971]: I1213 07:43:25.846009 4971 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/42cff6d1-138e-4a16-a570-cdcbcbe6ab49-inventory\") pod \"ceph-client-edpm-deployment-openstack-edpm-ipam-7grkn\" (UID: \"42cff6d1-138e-4a16-a570-cdcbcbe6ab49\") " pod="openstack/ceph-client-edpm-deployment-openstack-edpm-ipam-7grkn" Dec 13 07:43:25 crc kubenswrapper[4971]: I1213 07:43:25.846089 4971 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/42cff6d1-138e-4a16-a570-cdcbcbe6ab49-ssh-key\") pod \"ceph-client-edpm-deployment-openstack-edpm-ipam-7grkn\" (UID: \"42cff6d1-138e-4a16-a570-cdcbcbe6ab49\") " pod="openstack/ceph-client-edpm-deployment-openstack-edpm-ipam-7grkn" Dec 13 07:43:25 crc kubenswrapper[4971]: I1213 07:43:25.948314 4971 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-85prg\" (UniqueName: \"kubernetes.io/projected/42cff6d1-138e-4a16-a570-cdcbcbe6ab49-kube-api-access-85prg\") pod \"ceph-client-edpm-deployment-openstack-edpm-ipam-7grkn\" (UID: \"42cff6d1-138e-4a16-a570-cdcbcbe6ab49\") " pod="openstack/ceph-client-edpm-deployment-openstack-edpm-ipam-7grkn" Dec 13 07:43:25 crc kubenswrapper[4971]: I1213 07:43:25.948393 4971 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/42cff6d1-138e-4a16-a570-cdcbcbe6ab49-ceph\") pod \"ceph-client-edpm-deployment-openstack-edpm-ipam-7grkn\" (UID: \"42cff6d1-138e-4a16-a570-cdcbcbe6ab49\") " pod="openstack/ceph-client-edpm-deployment-openstack-edpm-ipam-7grkn" Dec 13 07:43:25 crc kubenswrapper[4971]: I1213 07:43:25.948543 4971 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/42cff6d1-138e-4a16-a570-cdcbcbe6ab49-inventory\") pod \"ceph-client-edpm-deployment-openstack-edpm-ipam-7grkn\" (UID: \"42cff6d1-138e-4a16-a570-cdcbcbe6ab49\") " pod="openstack/ceph-client-edpm-deployment-openstack-edpm-ipam-7grkn" Dec 13 07:43:25 crc kubenswrapper[4971]: I1213 07:43:25.948589 4971 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/42cff6d1-138e-4a16-a570-cdcbcbe6ab49-ssh-key\") pod \"ceph-client-edpm-deployment-openstack-edpm-ipam-7grkn\" (UID: \"42cff6d1-138e-4a16-a570-cdcbcbe6ab49\") " pod="openstack/ceph-client-edpm-deployment-openstack-edpm-ipam-7grkn" Dec 13 07:43:25 crc kubenswrapper[4971]: I1213 07:43:25.967207 4971 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/42cff6d1-138e-4a16-a570-cdcbcbe6ab49-ceph\") pod \"ceph-client-edpm-deployment-openstack-edpm-ipam-7grkn\" (UID: \"42cff6d1-138e-4a16-a570-cdcbcbe6ab49\") " pod="openstack/ceph-client-edpm-deployment-openstack-edpm-ipam-7grkn" Dec 13 07:43:25 crc kubenswrapper[4971]: I1213 07:43:25.967868 4971 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/42cff6d1-138e-4a16-a570-cdcbcbe6ab49-inventory\") pod \"ceph-client-edpm-deployment-openstack-edpm-ipam-7grkn\" (UID: \"42cff6d1-138e-4a16-a570-cdcbcbe6ab49\") " pod="openstack/ceph-client-edpm-deployment-openstack-edpm-ipam-7grkn" Dec 13 07:43:25 crc kubenswrapper[4971]: I1213 07:43:25.969387 4971 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/42cff6d1-138e-4a16-a570-cdcbcbe6ab49-ssh-key\") pod \"ceph-client-edpm-deployment-openstack-edpm-ipam-7grkn\" (UID: \"42cff6d1-138e-4a16-a570-cdcbcbe6ab49\") " pod="openstack/ceph-client-edpm-deployment-openstack-edpm-ipam-7grkn" Dec 13 07:43:25 crc kubenswrapper[4971]: I1213 07:43:25.985315 4971 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-85prg\" (UniqueName: \"kubernetes.io/projected/42cff6d1-138e-4a16-a570-cdcbcbe6ab49-kube-api-access-85prg\") pod \"ceph-client-edpm-deployment-openstack-edpm-ipam-7grkn\" (UID: \"42cff6d1-138e-4a16-a570-cdcbcbe6ab49\") " pod="openstack/ceph-client-edpm-deployment-openstack-edpm-ipam-7grkn" Dec 13 07:43:26 crc kubenswrapper[4971]: I1213 07:43:26.074033 4971 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceph-client-edpm-deployment-openstack-edpm-ipam-7grkn" Dec 13 07:43:26 crc kubenswrapper[4971]: I1213 07:43:26.614565 4971 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ceph-client-edpm-deployment-openstack-edpm-ipam-7grkn"] Dec 13 07:43:26 crc kubenswrapper[4971]: W1213 07:43:26.619257 4971 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod42cff6d1_138e_4a16_a570_cdcbcbe6ab49.slice/crio-25e0089d6072d819e70cb60b533e2a4cda9c5bf9c8bb62ce7a54172b039a1c09 WatchSource:0}: Error finding container 25e0089d6072d819e70cb60b533e2a4cda9c5bf9c8bb62ce7a54172b039a1c09: Status 404 returned error can't find the container with id 25e0089d6072d819e70cb60b533e2a4cda9c5bf9c8bb62ce7a54172b039a1c09 Dec 13 07:43:27 crc kubenswrapper[4971]: I1213 07:43:27.609596 4971 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-82xjz" event={"ID":"e0c4af12-edda-41d7-9b44-f87396174ef2","Type":"ContainerStarted","Data":"95874f24e419c2abe48edb2569aa8a50c62f04d8d5d3a5215adeee275f78b4d8"} Dec 13 07:43:27 crc kubenswrapper[4971]: I1213 07:43:27.611186 4971 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceph-client-edpm-deployment-openstack-edpm-ipam-7grkn" event={"ID":"42cff6d1-138e-4a16-a570-cdcbcbe6ab49","Type":"ContainerStarted","Data":"25e0089d6072d819e70cb60b533e2a4cda9c5bf9c8bb62ce7a54172b039a1c09"} Dec 13 07:43:29 crc kubenswrapper[4971]: I1213 07:43:29.627232 4971 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceph-client-edpm-deployment-openstack-edpm-ipam-7grkn" event={"ID":"42cff6d1-138e-4a16-a570-cdcbcbe6ab49","Type":"ContainerStarted","Data":"415b9c4a9d837b9ec35bead37f37480388cc0382ace776bfd12d38e6261b25b6"} Dec 13 07:43:29 crc kubenswrapper[4971]: I1213 07:43:29.646799 4971 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/ceph-client-edpm-deployment-openstack-edpm-ipam-7grkn" podStartSLOduration=2.50663812 podStartE2EDuration="4.646780335s" podCreationTimestamp="2025-12-13 07:43:25 +0000 UTC" firstStartedPulling="2025-12-13 07:43:26.622943483 +0000 UTC m=+3263.227352931" lastFinishedPulling="2025-12-13 07:43:28.763085708 +0000 UTC m=+3265.367495146" observedRunningTime="2025-12-13 07:43:29.641813846 +0000 UTC m=+3266.246223294" watchObservedRunningTime="2025-12-13 07:43:29.646780335 +0000 UTC m=+3266.251189783" Dec 13 07:43:35 crc kubenswrapper[4971]: I1213 07:43:35.681469 4971 generic.go:334] "Generic (PLEG): container finished" podID="42cff6d1-138e-4a16-a570-cdcbcbe6ab49" containerID="415b9c4a9d837b9ec35bead37f37480388cc0382ace776bfd12d38e6261b25b6" exitCode=0 Dec 13 07:43:35 crc kubenswrapper[4971]: I1213 07:43:35.681577 4971 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceph-client-edpm-deployment-openstack-edpm-ipam-7grkn" event={"ID":"42cff6d1-138e-4a16-a570-cdcbcbe6ab49","Type":"ContainerDied","Data":"415b9c4a9d837b9ec35bead37f37480388cc0382ace776bfd12d38e6261b25b6"} Dec 13 07:43:37 crc kubenswrapper[4971]: I1213 07:43:37.082674 4971 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ceph-client-edpm-deployment-openstack-edpm-ipam-7grkn" Dec 13 07:43:37 crc kubenswrapper[4971]: I1213 07:43:37.241231 4971 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/42cff6d1-138e-4a16-a570-cdcbcbe6ab49-ssh-key\") pod \"42cff6d1-138e-4a16-a570-cdcbcbe6ab49\" (UID: \"42cff6d1-138e-4a16-a570-cdcbcbe6ab49\") " Dec 13 07:43:37 crc kubenswrapper[4971]: I1213 07:43:37.241305 4971 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/42cff6d1-138e-4a16-a570-cdcbcbe6ab49-ceph\") pod \"42cff6d1-138e-4a16-a570-cdcbcbe6ab49\" (UID: \"42cff6d1-138e-4a16-a570-cdcbcbe6ab49\") " Dec 13 07:43:37 crc kubenswrapper[4971]: I1213 07:43:37.241362 4971 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/42cff6d1-138e-4a16-a570-cdcbcbe6ab49-inventory\") pod \"42cff6d1-138e-4a16-a570-cdcbcbe6ab49\" (UID: \"42cff6d1-138e-4a16-a570-cdcbcbe6ab49\") " Dec 13 07:43:37 crc kubenswrapper[4971]: I1213 07:43:37.241432 4971 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-85prg\" (UniqueName: \"kubernetes.io/projected/42cff6d1-138e-4a16-a570-cdcbcbe6ab49-kube-api-access-85prg\") pod \"42cff6d1-138e-4a16-a570-cdcbcbe6ab49\" (UID: \"42cff6d1-138e-4a16-a570-cdcbcbe6ab49\") " Dec 13 07:43:37 crc kubenswrapper[4971]: I1213 07:43:37.247744 4971 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/42cff6d1-138e-4a16-a570-cdcbcbe6ab49-ceph" (OuterVolumeSpecName: "ceph") pod "42cff6d1-138e-4a16-a570-cdcbcbe6ab49" (UID: "42cff6d1-138e-4a16-a570-cdcbcbe6ab49"). InnerVolumeSpecName "ceph". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 13 07:43:37 crc kubenswrapper[4971]: I1213 07:43:37.249829 4971 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/42cff6d1-138e-4a16-a570-cdcbcbe6ab49-kube-api-access-85prg" (OuterVolumeSpecName: "kube-api-access-85prg") pod "42cff6d1-138e-4a16-a570-cdcbcbe6ab49" (UID: "42cff6d1-138e-4a16-a570-cdcbcbe6ab49"). InnerVolumeSpecName "kube-api-access-85prg". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 13 07:43:37 crc kubenswrapper[4971]: I1213 07:43:37.267740 4971 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/42cff6d1-138e-4a16-a570-cdcbcbe6ab49-ssh-key" (OuterVolumeSpecName: "ssh-key") pod "42cff6d1-138e-4a16-a570-cdcbcbe6ab49" (UID: "42cff6d1-138e-4a16-a570-cdcbcbe6ab49"). InnerVolumeSpecName "ssh-key". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 13 07:43:37 crc kubenswrapper[4971]: I1213 07:43:37.272398 4971 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/42cff6d1-138e-4a16-a570-cdcbcbe6ab49-inventory" (OuterVolumeSpecName: "inventory") pod "42cff6d1-138e-4a16-a570-cdcbcbe6ab49" (UID: "42cff6d1-138e-4a16-a570-cdcbcbe6ab49"). InnerVolumeSpecName "inventory". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 13 07:43:37 crc kubenswrapper[4971]: I1213 07:43:37.343409 4971 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-85prg\" (UniqueName: \"kubernetes.io/projected/42cff6d1-138e-4a16-a570-cdcbcbe6ab49-kube-api-access-85prg\") on node \"crc\" DevicePath \"\"" Dec 13 07:43:37 crc kubenswrapper[4971]: I1213 07:43:37.343454 4971 reconciler_common.go:293] "Volume detached for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/42cff6d1-138e-4a16-a570-cdcbcbe6ab49-ssh-key\") on node \"crc\" DevicePath \"\"" Dec 13 07:43:37 crc kubenswrapper[4971]: I1213 07:43:37.343529 4971 reconciler_common.go:293] "Volume detached for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/42cff6d1-138e-4a16-a570-cdcbcbe6ab49-ceph\") on node \"crc\" DevicePath \"\"" Dec 13 07:43:37 crc kubenswrapper[4971]: I1213 07:43:37.343543 4971 reconciler_common.go:293] "Volume detached for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/42cff6d1-138e-4a16-a570-cdcbcbe6ab49-inventory\") on node \"crc\" DevicePath \"\"" Dec 13 07:43:37 crc kubenswrapper[4971]: I1213 07:43:37.699191 4971 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceph-client-edpm-deployment-openstack-edpm-ipam-7grkn" event={"ID":"42cff6d1-138e-4a16-a570-cdcbcbe6ab49","Type":"ContainerDied","Data":"25e0089d6072d819e70cb60b533e2a4cda9c5bf9c8bb62ce7a54172b039a1c09"} Dec 13 07:43:37 crc kubenswrapper[4971]: I1213 07:43:37.699456 4971 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="25e0089d6072d819e70cb60b533e2a4cda9c5bf9c8bb62ce7a54172b039a1c09" Dec 13 07:43:37 crc kubenswrapper[4971]: I1213 07:43:37.699249 4971 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ceph-client-edpm-deployment-openstack-edpm-ipam-7grkn" Dec 13 07:43:37 crc kubenswrapper[4971]: I1213 07:43:37.792256 4971 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ovn-edpm-deployment-openstack-edpm-ipam-lswgl"] Dec 13 07:43:37 crc kubenswrapper[4971]: E1213 07:43:37.793089 4971 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="42cff6d1-138e-4a16-a570-cdcbcbe6ab49" containerName="ceph-client-edpm-deployment-openstack-edpm-ipam" Dec 13 07:43:37 crc kubenswrapper[4971]: I1213 07:43:37.793110 4971 state_mem.go:107] "Deleted CPUSet assignment" podUID="42cff6d1-138e-4a16-a570-cdcbcbe6ab49" containerName="ceph-client-edpm-deployment-openstack-edpm-ipam" Dec 13 07:43:37 crc kubenswrapper[4971]: I1213 07:43:37.793583 4971 memory_manager.go:354] "RemoveStaleState removing state" podUID="42cff6d1-138e-4a16-a570-cdcbcbe6ab49" containerName="ceph-client-edpm-deployment-openstack-edpm-ipam" Dec 13 07:43:37 crc kubenswrapper[4971]: I1213 07:43:37.794642 4971 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-lswgl" Dec 13 07:43:37 crc kubenswrapper[4971]: I1213 07:43:37.797213 4971 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"ovncontroller-config" Dec 13 07:43:37 crc kubenswrapper[4971]: I1213 07:43:37.798800 4971 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplane-ansible-ssh-private-key-secret" Dec 13 07:43:37 crc kubenswrapper[4971]: I1213 07:43:37.798995 4971 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-edpm-ipam-dockercfg-zb2j5" Dec 13 07:43:37 crc kubenswrapper[4971]: I1213 07:43:37.799211 4971 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceph-conf-files" Dec 13 07:43:37 crc kubenswrapper[4971]: I1213 07:43:37.799579 4971 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-aee-default-env" Dec 13 07:43:37 crc kubenswrapper[4971]: I1213 07:43:37.799672 4971 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplanenodeset-openstack-edpm-ipam" Dec 13 07:43:37 crc kubenswrapper[4971]: I1213 07:43:37.815505 4971 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovn-edpm-deployment-openstack-edpm-ipam-lswgl"] Dec 13 07:43:37 crc kubenswrapper[4971]: I1213 07:43:37.867634 4971 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/9f1e6cde-d76e-45f1-a6a6-8546b16e389b-ssh-key\") pod \"ovn-edpm-deployment-openstack-edpm-ipam-lswgl\" (UID: \"9f1e6cde-d76e-45f1-a6a6-8546b16e389b\") " pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-lswgl" Dec 13 07:43:37 crc kubenswrapper[4971]: I1213 07:43:37.867699 4971 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovn-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/9f1e6cde-d76e-45f1-a6a6-8546b16e389b-ovn-combined-ca-bundle\") pod \"ovn-edpm-deployment-openstack-edpm-ipam-lswgl\" (UID: \"9f1e6cde-d76e-45f1-a6a6-8546b16e389b\") " pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-lswgl" Dec 13 07:43:37 crc kubenswrapper[4971]: I1213 07:43:37.867719 4971 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovncontroller-config-0\" (UniqueName: \"kubernetes.io/configmap/9f1e6cde-d76e-45f1-a6a6-8546b16e389b-ovncontroller-config-0\") pod \"ovn-edpm-deployment-openstack-edpm-ipam-lswgl\" (UID: \"9f1e6cde-d76e-45f1-a6a6-8546b16e389b\") " pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-lswgl" Dec 13 07:43:37 crc kubenswrapper[4971]: I1213 07:43:37.867769 4971 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/9f1e6cde-d76e-45f1-a6a6-8546b16e389b-ceph\") pod \"ovn-edpm-deployment-openstack-edpm-ipam-lswgl\" (UID: \"9f1e6cde-d76e-45f1-a6a6-8546b16e389b\") " pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-lswgl" Dec 13 07:43:37 crc kubenswrapper[4971]: I1213 07:43:37.867887 4971 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/9f1e6cde-d76e-45f1-a6a6-8546b16e389b-inventory\") pod \"ovn-edpm-deployment-openstack-edpm-ipam-lswgl\" (UID: \"9f1e6cde-d76e-45f1-a6a6-8546b16e389b\") " pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-lswgl" Dec 13 07:43:37 crc kubenswrapper[4971]: I1213 07:43:37.867914 4971 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-rbntd\" (UniqueName: \"kubernetes.io/projected/9f1e6cde-d76e-45f1-a6a6-8546b16e389b-kube-api-access-rbntd\") pod \"ovn-edpm-deployment-openstack-edpm-ipam-lswgl\" (UID: \"9f1e6cde-d76e-45f1-a6a6-8546b16e389b\") " pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-lswgl" Dec 13 07:43:37 crc kubenswrapper[4971]: I1213 07:43:37.969191 4971 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/9f1e6cde-d76e-45f1-a6a6-8546b16e389b-inventory\") pod \"ovn-edpm-deployment-openstack-edpm-ipam-lswgl\" (UID: \"9f1e6cde-d76e-45f1-a6a6-8546b16e389b\") " pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-lswgl" Dec 13 07:43:37 crc kubenswrapper[4971]: I1213 07:43:37.969258 4971 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-rbntd\" (UniqueName: \"kubernetes.io/projected/9f1e6cde-d76e-45f1-a6a6-8546b16e389b-kube-api-access-rbntd\") pod \"ovn-edpm-deployment-openstack-edpm-ipam-lswgl\" (UID: \"9f1e6cde-d76e-45f1-a6a6-8546b16e389b\") " pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-lswgl" Dec 13 07:43:37 crc kubenswrapper[4971]: I1213 07:43:37.969332 4971 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/9f1e6cde-d76e-45f1-a6a6-8546b16e389b-ssh-key\") pod \"ovn-edpm-deployment-openstack-edpm-ipam-lswgl\" (UID: \"9f1e6cde-d76e-45f1-a6a6-8546b16e389b\") " pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-lswgl" Dec 13 07:43:37 crc kubenswrapper[4971]: I1213 07:43:37.969358 4971 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovn-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/9f1e6cde-d76e-45f1-a6a6-8546b16e389b-ovn-combined-ca-bundle\") pod \"ovn-edpm-deployment-openstack-edpm-ipam-lswgl\" (UID: \"9f1e6cde-d76e-45f1-a6a6-8546b16e389b\") " pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-lswgl" Dec 13 07:43:37 crc kubenswrapper[4971]: I1213 07:43:37.969377 4971 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovncontroller-config-0\" (UniqueName: \"kubernetes.io/configmap/9f1e6cde-d76e-45f1-a6a6-8546b16e389b-ovncontroller-config-0\") pod \"ovn-edpm-deployment-openstack-edpm-ipam-lswgl\" (UID: \"9f1e6cde-d76e-45f1-a6a6-8546b16e389b\") " pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-lswgl" Dec 13 07:43:37 crc kubenswrapper[4971]: I1213 07:43:37.969413 4971 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/9f1e6cde-d76e-45f1-a6a6-8546b16e389b-ceph\") pod \"ovn-edpm-deployment-openstack-edpm-ipam-lswgl\" (UID: \"9f1e6cde-d76e-45f1-a6a6-8546b16e389b\") " pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-lswgl" Dec 13 07:43:37 crc kubenswrapper[4971]: I1213 07:43:37.970868 4971 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovncontroller-config-0\" (UniqueName: \"kubernetes.io/configmap/9f1e6cde-d76e-45f1-a6a6-8546b16e389b-ovncontroller-config-0\") pod \"ovn-edpm-deployment-openstack-edpm-ipam-lswgl\" (UID: \"9f1e6cde-d76e-45f1-a6a6-8546b16e389b\") " pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-lswgl" Dec 13 07:43:37 crc kubenswrapper[4971]: I1213 07:43:37.973356 4971 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/9f1e6cde-d76e-45f1-a6a6-8546b16e389b-inventory\") pod \"ovn-edpm-deployment-openstack-edpm-ipam-lswgl\" (UID: \"9f1e6cde-d76e-45f1-a6a6-8546b16e389b\") " pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-lswgl" Dec 13 07:43:37 crc kubenswrapper[4971]: I1213 07:43:37.973984 4971 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/9f1e6cde-d76e-45f1-a6a6-8546b16e389b-ssh-key\") pod \"ovn-edpm-deployment-openstack-edpm-ipam-lswgl\" (UID: \"9f1e6cde-d76e-45f1-a6a6-8546b16e389b\") " pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-lswgl" Dec 13 07:43:37 crc kubenswrapper[4971]: I1213 07:43:37.974168 4971 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovn-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/9f1e6cde-d76e-45f1-a6a6-8546b16e389b-ovn-combined-ca-bundle\") pod \"ovn-edpm-deployment-openstack-edpm-ipam-lswgl\" (UID: \"9f1e6cde-d76e-45f1-a6a6-8546b16e389b\") " pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-lswgl" Dec 13 07:43:37 crc kubenswrapper[4971]: I1213 07:43:37.977121 4971 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/9f1e6cde-d76e-45f1-a6a6-8546b16e389b-ceph\") pod \"ovn-edpm-deployment-openstack-edpm-ipam-lswgl\" (UID: \"9f1e6cde-d76e-45f1-a6a6-8546b16e389b\") " pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-lswgl" Dec 13 07:43:37 crc kubenswrapper[4971]: I1213 07:43:37.987630 4971 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-rbntd\" (UniqueName: \"kubernetes.io/projected/9f1e6cde-d76e-45f1-a6a6-8546b16e389b-kube-api-access-rbntd\") pod \"ovn-edpm-deployment-openstack-edpm-ipam-lswgl\" (UID: \"9f1e6cde-d76e-45f1-a6a6-8546b16e389b\") " pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-lswgl" Dec 13 07:43:38 crc kubenswrapper[4971]: I1213 07:43:38.119599 4971 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-lswgl" Dec 13 07:43:38 crc kubenswrapper[4971]: I1213 07:43:38.647791 4971 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovn-edpm-deployment-openstack-edpm-ipam-lswgl"] Dec 13 07:43:38 crc kubenswrapper[4971]: I1213 07:43:38.708055 4971 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-lswgl" event={"ID":"9f1e6cde-d76e-45f1-a6a6-8546b16e389b","Type":"ContainerStarted","Data":"f192445f47d86d3382d70bec772d818c31373aff1881d16cd57dc21a4c54ef3f"} Dec 13 07:43:39 crc kubenswrapper[4971]: I1213 07:43:39.716966 4971 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-lswgl" event={"ID":"9f1e6cde-d76e-45f1-a6a6-8546b16e389b","Type":"ContainerStarted","Data":"ff79f971b7696f671fb81c5beebb8e0b73d05e1efb44b948a6f6a65e9ca3cccb"} Dec 13 07:43:39 crc kubenswrapper[4971]: I1213 07:43:39.756190 4971 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-lswgl" podStartSLOduration=2.008816636 podStartE2EDuration="2.756167102s" podCreationTimestamp="2025-12-13 07:43:37 +0000 UTC" firstStartedPulling="2025-12-13 07:43:38.657634054 +0000 UTC m=+3275.262043502" lastFinishedPulling="2025-12-13 07:43:39.40498452 +0000 UTC m=+3276.009393968" observedRunningTime="2025-12-13 07:43:39.747329838 +0000 UTC m=+3276.351739286" watchObservedRunningTime="2025-12-13 07:43:39.756167102 +0000 UTC m=+3276.360576550" Dec 13 07:45:00 crc kubenswrapper[4971]: I1213 07:45:00.158662 4971 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29426865-mqtkj"] Dec 13 07:45:00 crc kubenswrapper[4971]: I1213 07:45:00.160651 4971 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29426865-mqtkj" Dec 13 07:45:00 crc kubenswrapper[4971]: I1213 07:45:00.164811 4971 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-operator-lifecycle-manager"/"collect-profiles-config" Dec 13 07:45:00 crc kubenswrapper[4971]: I1213 07:45:00.164847 4971 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"collect-profiles-dockercfg-kzf4t" Dec 13 07:45:00 crc kubenswrapper[4971]: I1213 07:45:00.181120 4971 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29426865-mqtkj"] Dec 13 07:45:00 crc kubenswrapper[4971]: I1213 07:45:00.259991 4971 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/099e03d5-1210-4976-ae03-e984b101d681-config-volume\") pod \"collect-profiles-29426865-mqtkj\" (UID: \"099e03d5-1210-4976-ae03-e984b101d681\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29426865-mqtkj" Dec 13 07:45:00 crc kubenswrapper[4971]: I1213 07:45:00.260052 4971 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-rkjfx\" (UniqueName: \"kubernetes.io/projected/099e03d5-1210-4976-ae03-e984b101d681-kube-api-access-rkjfx\") pod \"collect-profiles-29426865-mqtkj\" (UID: \"099e03d5-1210-4976-ae03-e984b101d681\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29426865-mqtkj" Dec 13 07:45:00 crc kubenswrapper[4971]: I1213 07:45:00.260194 4971 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/099e03d5-1210-4976-ae03-e984b101d681-secret-volume\") pod \"collect-profiles-29426865-mqtkj\" (UID: \"099e03d5-1210-4976-ae03-e984b101d681\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29426865-mqtkj" Dec 13 07:45:00 crc kubenswrapper[4971]: I1213 07:45:00.362408 4971 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/099e03d5-1210-4976-ae03-e984b101d681-secret-volume\") pod \"collect-profiles-29426865-mqtkj\" (UID: \"099e03d5-1210-4976-ae03-e984b101d681\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29426865-mqtkj" Dec 13 07:45:00 crc kubenswrapper[4971]: I1213 07:45:00.362511 4971 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/099e03d5-1210-4976-ae03-e984b101d681-config-volume\") pod \"collect-profiles-29426865-mqtkj\" (UID: \"099e03d5-1210-4976-ae03-e984b101d681\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29426865-mqtkj" Dec 13 07:45:00 crc kubenswrapper[4971]: I1213 07:45:00.362556 4971 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-rkjfx\" (UniqueName: \"kubernetes.io/projected/099e03d5-1210-4976-ae03-e984b101d681-kube-api-access-rkjfx\") pod \"collect-profiles-29426865-mqtkj\" (UID: \"099e03d5-1210-4976-ae03-e984b101d681\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29426865-mqtkj" Dec 13 07:45:00 crc kubenswrapper[4971]: I1213 07:45:00.364401 4971 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/099e03d5-1210-4976-ae03-e984b101d681-config-volume\") pod \"collect-profiles-29426865-mqtkj\" (UID: \"099e03d5-1210-4976-ae03-e984b101d681\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29426865-mqtkj" Dec 13 07:45:00 crc kubenswrapper[4971]: I1213 07:45:00.368896 4971 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/099e03d5-1210-4976-ae03-e984b101d681-secret-volume\") pod \"collect-profiles-29426865-mqtkj\" (UID: \"099e03d5-1210-4976-ae03-e984b101d681\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29426865-mqtkj" Dec 13 07:45:00 crc kubenswrapper[4971]: I1213 07:45:00.379705 4971 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-rkjfx\" (UniqueName: \"kubernetes.io/projected/099e03d5-1210-4976-ae03-e984b101d681-kube-api-access-rkjfx\") pod \"collect-profiles-29426865-mqtkj\" (UID: \"099e03d5-1210-4976-ae03-e984b101d681\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29426865-mqtkj" Dec 13 07:45:00 crc kubenswrapper[4971]: I1213 07:45:00.483820 4971 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29426865-mqtkj" Dec 13 07:45:01 crc kubenswrapper[4971]: I1213 07:45:01.007306 4971 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29426865-mqtkj"] Dec 13 07:45:01 crc kubenswrapper[4971]: I1213 07:45:01.554053 4971 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29426865-mqtkj" event={"ID":"099e03d5-1210-4976-ae03-e984b101d681","Type":"ContainerStarted","Data":"d5e45b0e5f6f46fa05fbb1f93081db75b1488ecfccc055daa354973e965247b4"} Dec 13 07:45:02 crc kubenswrapper[4971]: I1213 07:45:02.565476 4971 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29426865-mqtkj" event={"ID":"099e03d5-1210-4976-ae03-e984b101d681","Type":"ContainerStarted","Data":"6c76911f291cffb765fc22afc90bcd5afabdd2c106785593da6f0d3f650d878f"} Dec 13 07:45:02 crc kubenswrapper[4971]: I1213 07:45:02.585815 4971 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-operator-lifecycle-manager/collect-profiles-29426865-mqtkj" podStartSLOduration=2.58579597 podStartE2EDuration="2.58579597s" podCreationTimestamp="2025-12-13 07:45:00 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-13 07:45:02.580564587 +0000 UTC m=+3359.184974035" watchObservedRunningTime="2025-12-13 07:45:02.58579597 +0000 UTC m=+3359.190205408" Dec 13 07:45:03 crc kubenswrapper[4971]: I1213 07:45:03.577947 4971 generic.go:334] "Generic (PLEG): container finished" podID="099e03d5-1210-4976-ae03-e984b101d681" containerID="6c76911f291cffb765fc22afc90bcd5afabdd2c106785593da6f0d3f650d878f" exitCode=0 Dec 13 07:45:03 crc kubenswrapper[4971]: I1213 07:45:03.578001 4971 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29426865-mqtkj" event={"ID":"099e03d5-1210-4976-ae03-e984b101d681","Type":"ContainerDied","Data":"6c76911f291cffb765fc22afc90bcd5afabdd2c106785593da6f0d3f650d878f"} Dec 13 07:45:04 crc kubenswrapper[4971]: I1213 07:45:04.956095 4971 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29426865-mqtkj" Dec 13 07:45:05 crc kubenswrapper[4971]: I1213 07:45:05.080848 4971 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-rkjfx\" (UniqueName: \"kubernetes.io/projected/099e03d5-1210-4976-ae03-e984b101d681-kube-api-access-rkjfx\") pod \"099e03d5-1210-4976-ae03-e984b101d681\" (UID: \"099e03d5-1210-4976-ae03-e984b101d681\") " Dec 13 07:45:05 crc kubenswrapper[4971]: I1213 07:45:05.081540 4971 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/099e03d5-1210-4976-ae03-e984b101d681-secret-volume\") pod \"099e03d5-1210-4976-ae03-e984b101d681\" (UID: \"099e03d5-1210-4976-ae03-e984b101d681\") " Dec 13 07:45:05 crc kubenswrapper[4971]: I1213 07:45:05.081686 4971 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/099e03d5-1210-4976-ae03-e984b101d681-config-volume\") pod \"099e03d5-1210-4976-ae03-e984b101d681\" (UID: \"099e03d5-1210-4976-ae03-e984b101d681\") " Dec 13 07:45:05 crc kubenswrapper[4971]: I1213 07:45:05.084265 4971 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/099e03d5-1210-4976-ae03-e984b101d681-config-volume" (OuterVolumeSpecName: "config-volume") pod "099e03d5-1210-4976-ae03-e984b101d681" (UID: "099e03d5-1210-4976-ae03-e984b101d681"). InnerVolumeSpecName "config-volume". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 13 07:45:05 crc kubenswrapper[4971]: I1213 07:45:05.089294 4971 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/099e03d5-1210-4976-ae03-e984b101d681-secret-volume" (OuterVolumeSpecName: "secret-volume") pod "099e03d5-1210-4976-ae03-e984b101d681" (UID: "099e03d5-1210-4976-ae03-e984b101d681"). InnerVolumeSpecName "secret-volume". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 13 07:45:05 crc kubenswrapper[4971]: I1213 07:45:05.089306 4971 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/099e03d5-1210-4976-ae03-e984b101d681-kube-api-access-rkjfx" (OuterVolumeSpecName: "kube-api-access-rkjfx") pod "099e03d5-1210-4976-ae03-e984b101d681" (UID: "099e03d5-1210-4976-ae03-e984b101d681"). InnerVolumeSpecName "kube-api-access-rkjfx". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 13 07:45:05 crc kubenswrapper[4971]: I1213 07:45:05.185236 4971 reconciler_common.go:293] "Volume detached for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/099e03d5-1210-4976-ae03-e984b101d681-secret-volume\") on node \"crc\" DevicePath \"\"" Dec 13 07:45:05 crc kubenswrapper[4971]: I1213 07:45:05.185269 4971 reconciler_common.go:293] "Volume detached for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/099e03d5-1210-4976-ae03-e984b101d681-config-volume\") on node \"crc\" DevicePath \"\"" Dec 13 07:45:05 crc kubenswrapper[4971]: I1213 07:45:05.185279 4971 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-rkjfx\" (UniqueName: \"kubernetes.io/projected/099e03d5-1210-4976-ae03-e984b101d681-kube-api-access-rkjfx\") on node \"crc\" DevicePath \"\"" Dec 13 07:45:05 crc kubenswrapper[4971]: I1213 07:45:05.596756 4971 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29426865-mqtkj" event={"ID":"099e03d5-1210-4976-ae03-e984b101d681","Type":"ContainerDied","Data":"d5e45b0e5f6f46fa05fbb1f93081db75b1488ecfccc055daa354973e965247b4"} Dec 13 07:45:05 crc kubenswrapper[4971]: I1213 07:45:05.596793 4971 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="d5e45b0e5f6f46fa05fbb1f93081db75b1488ecfccc055daa354973e965247b4" Dec 13 07:45:05 crc kubenswrapper[4971]: I1213 07:45:05.596800 4971 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29426865-mqtkj" Dec 13 07:45:05 crc kubenswrapper[4971]: I1213 07:45:05.652446 4971 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29426820-pbkhw"] Dec 13 07:45:05 crc kubenswrapper[4971]: I1213 07:45:05.659602 4971 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29426820-pbkhw"] Dec 13 07:45:05 crc kubenswrapper[4971]: I1213 07:45:05.783953 4971 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="48585415-a489-4ea0-9d68-981bbebd10cb" path="/var/lib/kubelet/pods/48585415-a489-4ea0-9d68-981bbebd10cb/volumes" Dec 13 07:45:12 crc kubenswrapper[4971]: I1213 07:45:12.764876 4971 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/certified-operators-tfdxq"] Dec 13 07:45:12 crc kubenswrapper[4971]: E1213 07:45:12.765874 4971 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="099e03d5-1210-4976-ae03-e984b101d681" containerName="collect-profiles" Dec 13 07:45:12 crc kubenswrapper[4971]: I1213 07:45:12.765890 4971 state_mem.go:107] "Deleted CPUSet assignment" podUID="099e03d5-1210-4976-ae03-e984b101d681" containerName="collect-profiles" Dec 13 07:45:12 crc kubenswrapper[4971]: I1213 07:45:12.766062 4971 memory_manager.go:354] "RemoveStaleState removing state" podUID="099e03d5-1210-4976-ae03-e984b101d681" containerName="collect-profiles" Dec 13 07:45:12 crc kubenswrapper[4971]: I1213 07:45:12.768476 4971 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-tfdxq" Dec 13 07:45:12 crc kubenswrapper[4971]: I1213 07:45:12.769307 4971 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/fa8714e3-d1ca-412a-9008-3f116341671a-catalog-content\") pod \"certified-operators-tfdxq\" (UID: \"fa8714e3-d1ca-412a-9008-3f116341671a\") " pod="openshift-marketplace/certified-operators-tfdxq" Dec 13 07:45:12 crc kubenswrapper[4971]: I1213 07:45:12.769601 4971 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/fa8714e3-d1ca-412a-9008-3f116341671a-utilities\") pod \"certified-operators-tfdxq\" (UID: \"fa8714e3-d1ca-412a-9008-3f116341671a\") " pod="openshift-marketplace/certified-operators-tfdxq" Dec 13 07:45:12 crc kubenswrapper[4971]: I1213 07:45:12.770091 4971 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-m6c8c\" (UniqueName: \"kubernetes.io/projected/fa8714e3-d1ca-412a-9008-3f116341671a-kube-api-access-m6c8c\") pod \"certified-operators-tfdxq\" (UID: \"fa8714e3-d1ca-412a-9008-3f116341671a\") " pod="openshift-marketplace/certified-operators-tfdxq" Dec 13 07:45:12 crc kubenswrapper[4971]: I1213 07:45:12.781304 4971 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-tfdxq"] Dec 13 07:45:12 crc kubenswrapper[4971]: I1213 07:45:12.871210 4971 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-m6c8c\" (UniqueName: \"kubernetes.io/projected/fa8714e3-d1ca-412a-9008-3f116341671a-kube-api-access-m6c8c\") pod \"certified-operators-tfdxq\" (UID: \"fa8714e3-d1ca-412a-9008-3f116341671a\") " pod="openshift-marketplace/certified-operators-tfdxq" Dec 13 07:45:12 crc kubenswrapper[4971]: I1213 07:45:12.871341 4971 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/fa8714e3-d1ca-412a-9008-3f116341671a-catalog-content\") pod \"certified-operators-tfdxq\" (UID: \"fa8714e3-d1ca-412a-9008-3f116341671a\") " pod="openshift-marketplace/certified-operators-tfdxq" Dec 13 07:45:12 crc kubenswrapper[4971]: I1213 07:45:12.871438 4971 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/fa8714e3-d1ca-412a-9008-3f116341671a-utilities\") pod \"certified-operators-tfdxq\" (UID: \"fa8714e3-d1ca-412a-9008-3f116341671a\") " pod="openshift-marketplace/certified-operators-tfdxq" Dec 13 07:45:12 crc kubenswrapper[4971]: I1213 07:45:12.872491 4971 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/fa8714e3-d1ca-412a-9008-3f116341671a-utilities\") pod \"certified-operators-tfdxq\" (UID: \"fa8714e3-d1ca-412a-9008-3f116341671a\") " pod="openshift-marketplace/certified-operators-tfdxq" Dec 13 07:45:12 crc kubenswrapper[4971]: I1213 07:45:12.873330 4971 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/fa8714e3-d1ca-412a-9008-3f116341671a-catalog-content\") pod \"certified-operators-tfdxq\" (UID: \"fa8714e3-d1ca-412a-9008-3f116341671a\") " pod="openshift-marketplace/certified-operators-tfdxq" Dec 13 07:45:12 crc kubenswrapper[4971]: I1213 07:45:12.899807 4971 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-m6c8c\" (UniqueName: \"kubernetes.io/projected/fa8714e3-d1ca-412a-9008-3f116341671a-kube-api-access-m6c8c\") pod \"certified-operators-tfdxq\" (UID: \"fa8714e3-d1ca-412a-9008-3f116341671a\") " pod="openshift-marketplace/certified-operators-tfdxq" Dec 13 07:45:13 crc kubenswrapper[4971]: I1213 07:45:13.104407 4971 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-tfdxq" Dec 13 07:45:13 crc kubenswrapper[4971]: I1213 07:45:13.586013 4971 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-tfdxq"] Dec 13 07:45:13 crc kubenswrapper[4971]: I1213 07:45:13.691769 4971 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-tfdxq" event={"ID":"fa8714e3-d1ca-412a-9008-3f116341671a","Type":"ContainerStarted","Data":"4a9c774b3d386377dd6bf47ad2977272bdec87518e130bdde523075e58cc7cc4"} Dec 13 07:45:16 crc kubenswrapper[4971]: I1213 07:45:16.721574 4971 generic.go:334] "Generic (PLEG): container finished" podID="fa8714e3-d1ca-412a-9008-3f116341671a" containerID="3837336ebfa471e9824e194ccca736bb21334907a318e55692ccfed429e58d6f" exitCode=0 Dec 13 07:45:16 crc kubenswrapper[4971]: I1213 07:45:16.721647 4971 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-tfdxq" event={"ID":"fa8714e3-d1ca-412a-9008-3f116341671a","Type":"ContainerDied","Data":"3837336ebfa471e9824e194ccca736bb21334907a318e55692ccfed429e58d6f"} Dec 13 07:45:16 crc kubenswrapper[4971]: I1213 07:45:16.724357 4971 provider.go:102] Refreshing cache for provider: *credentialprovider.defaultDockerConfigProvider Dec 13 07:45:20 crc kubenswrapper[4971]: I1213 07:45:20.068971 4971 scope.go:117] "RemoveContainer" containerID="2c7b6bd72e36013d046f5737a2251f46017870ee010e28c0aa3441cf21916bf0" Dec 13 07:45:20 crc kubenswrapper[4971]: I1213 07:45:20.761336 4971 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-tfdxq" event={"ID":"fa8714e3-d1ca-412a-9008-3f116341671a","Type":"ContainerStarted","Data":"cb77e650071dfbe6e95e3877602233de0265b6c0055877cf6f75a5854939c4e2"} Dec 13 07:45:21 crc kubenswrapper[4971]: I1213 07:45:21.772811 4971 generic.go:334] "Generic (PLEG): container finished" podID="fa8714e3-d1ca-412a-9008-3f116341671a" containerID="cb77e650071dfbe6e95e3877602233de0265b6c0055877cf6f75a5854939c4e2" exitCode=0 Dec 13 07:45:21 crc kubenswrapper[4971]: I1213 07:45:21.782418 4971 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-tfdxq" event={"ID":"fa8714e3-d1ca-412a-9008-3f116341671a","Type":"ContainerDied","Data":"cb77e650071dfbe6e95e3877602233de0265b6c0055877cf6f75a5854939c4e2"} Dec 13 07:45:25 crc kubenswrapper[4971]: I1213 07:45:25.821667 4971 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-tfdxq" event={"ID":"fa8714e3-d1ca-412a-9008-3f116341671a","Type":"ContainerStarted","Data":"fb633bebbf9c57e666268a311bd32732e6372f95e5634dce2cb405f3b9227b04"} Dec 13 07:45:26 crc kubenswrapper[4971]: I1213 07:45:26.877655 4971 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/certified-operators-tfdxq" podStartSLOduration=6.379210349 podStartE2EDuration="14.877633591s" podCreationTimestamp="2025-12-13 07:45:12 +0000 UTC" firstStartedPulling="2025-12-13 07:45:16.724033393 +0000 UTC m=+3373.328442841" lastFinishedPulling="2025-12-13 07:45:25.222456635 +0000 UTC m=+3381.826866083" observedRunningTime="2025-12-13 07:45:26.85200617 +0000 UTC m=+3383.456415638" watchObservedRunningTime="2025-12-13 07:45:26.877633591 +0000 UTC m=+3383.482043039" Dec 13 07:45:33 crc kubenswrapper[4971]: I1213 07:45:33.105189 4971 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/certified-operators-tfdxq" Dec 13 07:45:33 crc kubenswrapper[4971]: I1213 07:45:33.105732 4971 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/certified-operators-tfdxq" Dec 13 07:45:33 crc kubenswrapper[4971]: I1213 07:45:33.164358 4971 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/certified-operators-tfdxq" Dec 13 07:45:33 crc kubenswrapper[4971]: I1213 07:45:33.938647 4971 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/certified-operators-tfdxq" Dec 13 07:45:33 crc kubenswrapper[4971]: I1213 07:45:33.978910 4971 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-tfdxq"] Dec 13 07:45:35 crc kubenswrapper[4971]: I1213 07:45:35.938327 4971 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/certified-operators-tfdxq" podUID="fa8714e3-d1ca-412a-9008-3f116341671a" containerName="registry-server" containerID="cri-o://fb633bebbf9c57e666268a311bd32732e6372f95e5634dce2cb405f3b9227b04" gracePeriod=2 Dec 13 07:45:38 crc kubenswrapper[4971]: I1213 07:45:38.982406 4971 generic.go:334] "Generic (PLEG): container finished" podID="fa8714e3-d1ca-412a-9008-3f116341671a" containerID="fb633bebbf9c57e666268a311bd32732e6372f95e5634dce2cb405f3b9227b04" exitCode=0 Dec 13 07:45:38 crc kubenswrapper[4971]: I1213 07:45:38.982487 4971 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-tfdxq" event={"ID":"fa8714e3-d1ca-412a-9008-3f116341671a","Type":"ContainerDied","Data":"fb633bebbf9c57e666268a311bd32732e6372f95e5634dce2cb405f3b9227b04"} Dec 13 07:45:39 crc kubenswrapper[4971]: I1213 07:45:39.105998 4971 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-tfdxq" Dec 13 07:45:39 crc kubenswrapper[4971]: I1213 07:45:39.248146 4971 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/fa8714e3-d1ca-412a-9008-3f116341671a-utilities\") pod \"fa8714e3-d1ca-412a-9008-3f116341671a\" (UID: \"fa8714e3-d1ca-412a-9008-3f116341671a\") " Dec 13 07:45:39 crc kubenswrapper[4971]: I1213 07:45:39.248272 4971 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-m6c8c\" (UniqueName: \"kubernetes.io/projected/fa8714e3-d1ca-412a-9008-3f116341671a-kube-api-access-m6c8c\") pod \"fa8714e3-d1ca-412a-9008-3f116341671a\" (UID: \"fa8714e3-d1ca-412a-9008-3f116341671a\") " Dec 13 07:45:39 crc kubenswrapper[4971]: I1213 07:45:39.248358 4971 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/fa8714e3-d1ca-412a-9008-3f116341671a-catalog-content\") pod \"fa8714e3-d1ca-412a-9008-3f116341671a\" (UID: \"fa8714e3-d1ca-412a-9008-3f116341671a\") " Dec 13 07:45:39 crc kubenswrapper[4971]: I1213 07:45:39.250066 4971 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/fa8714e3-d1ca-412a-9008-3f116341671a-utilities" (OuterVolumeSpecName: "utilities") pod "fa8714e3-d1ca-412a-9008-3f116341671a" (UID: "fa8714e3-d1ca-412a-9008-3f116341671a"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 13 07:45:39 crc kubenswrapper[4971]: I1213 07:45:39.254949 4971 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/fa8714e3-d1ca-412a-9008-3f116341671a-kube-api-access-m6c8c" (OuterVolumeSpecName: "kube-api-access-m6c8c") pod "fa8714e3-d1ca-412a-9008-3f116341671a" (UID: "fa8714e3-d1ca-412a-9008-3f116341671a"). InnerVolumeSpecName "kube-api-access-m6c8c". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 13 07:45:39 crc kubenswrapper[4971]: I1213 07:45:39.301446 4971 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/fa8714e3-d1ca-412a-9008-3f116341671a-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "fa8714e3-d1ca-412a-9008-3f116341671a" (UID: "fa8714e3-d1ca-412a-9008-3f116341671a"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 13 07:45:39 crc kubenswrapper[4971]: I1213 07:45:39.350157 4971 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/fa8714e3-d1ca-412a-9008-3f116341671a-utilities\") on node \"crc\" DevicePath \"\"" Dec 13 07:45:39 crc kubenswrapper[4971]: I1213 07:45:39.350190 4971 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-m6c8c\" (UniqueName: \"kubernetes.io/projected/fa8714e3-d1ca-412a-9008-3f116341671a-kube-api-access-m6c8c\") on node \"crc\" DevicePath \"\"" Dec 13 07:45:39 crc kubenswrapper[4971]: I1213 07:45:39.350200 4971 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/fa8714e3-d1ca-412a-9008-3f116341671a-catalog-content\") on node \"crc\" DevicePath \"\"" Dec 13 07:45:40 crc kubenswrapper[4971]: I1213 07:45:40.004986 4971 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-tfdxq" event={"ID":"fa8714e3-d1ca-412a-9008-3f116341671a","Type":"ContainerDied","Data":"4a9c774b3d386377dd6bf47ad2977272bdec87518e130bdde523075e58cc7cc4"} Dec 13 07:45:40 crc kubenswrapper[4971]: I1213 07:45:40.005590 4971 scope.go:117] "RemoveContainer" containerID="fb633bebbf9c57e666268a311bd32732e6372f95e5634dce2cb405f3b9227b04" Dec 13 07:45:40 crc kubenswrapper[4971]: I1213 07:45:40.005443 4971 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-tfdxq" Dec 13 07:45:40 crc kubenswrapper[4971]: I1213 07:45:40.036205 4971 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-tfdxq"] Dec 13 07:45:40 crc kubenswrapper[4971]: I1213 07:45:40.036670 4971 scope.go:117] "RemoveContainer" containerID="cb77e650071dfbe6e95e3877602233de0265b6c0055877cf6f75a5854939c4e2" Dec 13 07:45:40 crc kubenswrapper[4971]: I1213 07:45:40.045061 4971 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/certified-operators-tfdxq"] Dec 13 07:45:40 crc kubenswrapper[4971]: I1213 07:45:40.060976 4971 scope.go:117] "RemoveContainer" containerID="3837336ebfa471e9824e194ccca736bb21334907a318e55692ccfed429e58d6f" Dec 13 07:45:41 crc kubenswrapper[4971]: I1213 07:45:41.788073 4971 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="fa8714e3-d1ca-412a-9008-3f116341671a" path="/var/lib/kubelet/pods/fa8714e3-d1ca-412a-9008-3f116341671a/volumes" Dec 13 07:45:46 crc kubenswrapper[4971]: I1213 07:45:46.154624 4971 patch_prober.go:28] interesting pod/machine-config-daemon-82xjz container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 13 07:45:46 crc kubenswrapper[4971]: I1213 07:45:46.155170 4971 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-82xjz" podUID="e0c4af12-edda-41d7-9b44-f87396174ef2" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 13 07:45:47 crc kubenswrapper[4971]: I1213 07:45:47.081187 4971 generic.go:334] "Generic (PLEG): container finished" podID="9f1e6cde-d76e-45f1-a6a6-8546b16e389b" containerID="ff79f971b7696f671fb81c5beebb8e0b73d05e1efb44b948a6f6a65e9ca3cccb" exitCode=0 Dec 13 07:45:47 crc kubenswrapper[4971]: I1213 07:45:47.081264 4971 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-lswgl" event={"ID":"9f1e6cde-d76e-45f1-a6a6-8546b16e389b","Type":"ContainerDied","Data":"ff79f971b7696f671fb81c5beebb8e0b73d05e1efb44b948a6f6a65e9ca3cccb"} Dec 13 07:45:48 crc kubenswrapper[4971]: I1213 07:45:48.524834 4971 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-lswgl" Dec 13 07:45:48 crc kubenswrapper[4971]: I1213 07:45:48.636726 4971 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/9f1e6cde-d76e-45f1-a6a6-8546b16e389b-ceph\") pod \"9f1e6cde-d76e-45f1-a6a6-8546b16e389b\" (UID: \"9f1e6cde-d76e-45f1-a6a6-8546b16e389b\") " Dec 13 07:45:48 crc kubenswrapper[4971]: I1213 07:45:48.637259 4971 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovn-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/9f1e6cde-d76e-45f1-a6a6-8546b16e389b-ovn-combined-ca-bundle\") pod \"9f1e6cde-d76e-45f1-a6a6-8546b16e389b\" (UID: \"9f1e6cde-d76e-45f1-a6a6-8546b16e389b\") " Dec 13 07:45:48 crc kubenswrapper[4971]: I1213 07:45:48.637563 4971 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/9f1e6cde-d76e-45f1-a6a6-8546b16e389b-ssh-key\") pod \"9f1e6cde-d76e-45f1-a6a6-8546b16e389b\" (UID: \"9f1e6cde-d76e-45f1-a6a6-8546b16e389b\") " Dec 13 07:45:48 crc kubenswrapper[4971]: I1213 07:45:48.637755 4971 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-rbntd\" (UniqueName: \"kubernetes.io/projected/9f1e6cde-d76e-45f1-a6a6-8546b16e389b-kube-api-access-rbntd\") pod \"9f1e6cde-d76e-45f1-a6a6-8546b16e389b\" (UID: \"9f1e6cde-d76e-45f1-a6a6-8546b16e389b\") " Dec 13 07:45:48 crc kubenswrapper[4971]: I1213 07:45:48.637974 4971 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/9f1e6cde-d76e-45f1-a6a6-8546b16e389b-inventory\") pod \"9f1e6cde-d76e-45f1-a6a6-8546b16e389b\" (UID: \"9f1e6cde-d76e-45f1-a6a6-8546b16e389b\") " Dec 13 07:45:48 crc kubenswrapper[4971]: I1213 07:45:48.638133 4971 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovncontroller-config-0\" (UniqueName: \"kubernetes.io/configmap/9f1e6cde-d76e-45f1-a6a6-8546b16e389b-ovncontroller-config-0\") pod \"9f1e6cde-d76e-45f1-a6a6-8546b16e389b\" (UID: \"9f1e6cde-d76e-45f1-a6a6-8546b16e389b\") " Dec 13 07:45:48 crc kubenswrapper[4971]: I1213 07:45:48.642996 4971 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/9f1e6cde-d76e-45f1-a6a6-8546b16e389b-ovn-combined-ca-bundle" (OuterVolumeSpecName: "ovn-combined-ca-bundle") pod "9f1e6cde-d76e-45f1-a6a6-8546b16e389b" (UID: "9f1e6cde-d76e-45f1-a6a6-8546b16e389b"). InnerVolumeSpecName "ovn-combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 13 07:45:48 crc kubenswrapper[4971]: I1213 07:45:48.643422 4971 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/9f1e6cde-d76e-45f1-a6a6-8546b16e389b-kube-api-access-rbntd" (OuterVolumeSpecName: "kube-api-access-rbntd") pod "9f1e6cde-d76e-45f1-a6a6-8546b16e389b" (UID: "9f1e6cde-d76e-45f1-a6a6-8546b16e389b"). InnerVolumeSpecName "kube-api-access-rbntd". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 13 07:45:48 crc kubenswrapper[4971]: I1213 07:45:48.646682 4971 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/9f1e6cde-d76e-45f1-a6a6-8546b16e389b-ceph" (OuterVolumeSpecName: "ceph") pod "9f1e6cde-d76e-45f1-a6a6-8546b16e389b" (UID: "9f1e6cde-d76e-45f1-a6a6-8546b16e389b"). InnerVolumeSpecName "ceph". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 13 07:45:48 crc kubenswrapper[4971]: I1213 07:45:48.666278 4971 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/9f1e6cde-d76e-45f1-a6a6-8546b16e389b-ovncontroller-config-0" (OuterVolumeSpecName: "ovncontroller-config-0") pod "9f1e6cde-d76e-45f1-a6a6-8546b16e389b" (UID: "9f1e6cde-d76e-45f1-a6a6-8546b16e389b"). InnerVolumeSpecName "ovncontroller-config-0". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 13 07:45:48 crc kubenswrapper[4971]: I1213 07:45:48.668655 4971 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/9f1e6cde-d76e-45f1-a6a6-8546b16e389b-inventory" (OuterVolumeSpecName: "inventory") pod "9f1e6cde-d76e-45f1-a6a6-8546b16e389b" (UID: "9f1e6cde-d76e-45f1-a6a6-8546b16e389b"). InnerVolumeSpecName "inventory". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 13 07:45:48 crc kubenswrapper[4971]: I1213 07:45:48.670077 4971 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/9f1e6cde-d76e-45f1-a6a6-8546b16e389b-ssh-key" (OuterVolumeSpecName: "ssh-key") pod "9f1e6cde-d76e-45f1-a6a6-8546b16e389b" (UID: "9f1e6cde-d76e-45f1-a6a6-8546b16e389b"). InnerVolumeSpecName "ssh-key". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 13 07:45:48 crc kubenswrapper[4971]: I1213 07:45:48.741359 4971 reconciler_common.go:293] "Volume detached for volume \"ovn-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/9f1e6cde-d76e-45f1-a6a6-8546b16e389b-ovn-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 13 07:45:48 crc kubenswrapper[4971]: I1213 07:45:48.741412 4971 reconciler_common.go:293] "Volume detached for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/9f1e6cde-d76e-45f1-a6a6-8546b16e389b-ssh-key\") on node \"crc\" DevicePath \"\"" Dec 13 07:45:48 crc kubenswrapper[4971]: I1213 07:45:48.741431 4971 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-rbntd\" (UniqueName: \"kubernetes.io/projected/9f1e6cde-d76e-45f1-a6a6-8546b16e389b-kube-api-access-rbntd\") on node \"crc\" DevicePath \"\"" Dec 13 07:45:48 crc kubenswrapper[4971]: I1213 07:45:48.741449 4971 reconciler_common.go:293] "Volume detached for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/9f1e6cde-d76e-45f1-a6a6-8546b16e389b-inventory\") on node \"crc\" DevicePath \"\"" Dec 13 07:45:48 crc kubenswrapper[4971]: I1213 07:45:48.741469 4971 reconciler_common.go:293] "Volume detached for volume \"ovncontroller-config-0\" (UniqueName: \"kubernetes.io/configmap/9f1e6cde-d76e-45f1-a6a6-8546b16e389b-ovncontroller-config-0\") on node \"crc\" DevicePath \"\"" Dec 13 07:45:48 crc kubenswrapper[4971]: I1213 07:45:48.741492 4971 reconciler_common.go:293] "Volume detached for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/9f1e6cde-d76e-45f1-a6a6-8546b16e389b-ceph\") on node \"crc\" DevicePath \"\"" Dec 13 07:45:49 crc kubenswrapper[4971]: I1213 07:45:49.103196 4971 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-lswgl" event={"ID":"9f1e6cde-d76e-45f1-a6a6-8546b16e389b","Type":"ContainerDied","Data":"f192445f47d86d3382d70bec772d818c31373aff1881d16cd57dc21a4c54ef3f"} Dec 13 07:45:49 crc kubenswrapper[4971]: I1213 07:45:49.103716 4971 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="f192445f47d86d3382d70bec772d818c31373aff1881d16cd57dc21a4c54ef3f" Dec 13 07:45:49 crc kubenswrapper[4971]: I1213 07:45:49.103265 4971 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-lswgl" Dec 13 07:45:49 crc kubenswrapper[4971]: I1213 07:45:49.210237 4971 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-tpz7w"] Dec 13 07:45:49 crc kubenswrapper[4971]: E1213 07:45:49.211192 4971 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="fa8714e3-d1ca-412a-9008-3f116341671a" containerName="extract-content" Dec 13 07:45:49 crc kubenswrapper[4971]: I1213 07:45:49.211213 4971 state_mem.go:107] "Deleted CPUSet assignment" podUID="fa8714e3-d1ca-412a-9008-3f116341671a" containerName="extract-content" Dec 13 07:45:49 crc kubenswrapper[4971]: E1213 07:45:49.211227 4971 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="fa8714e3-d1ca-412a-9008-3f116341671a" containerName="registry-server" Dec 13 07:45:49 crc kubenswrapper[4971]: I1213 07:45:49.211234 4971 state_mem.go:107] "Deleted CPUSet assignment" podUID="fa8714e3-d1ca-412a-9008-3f116341671a" containerName="registry-server" Dec 13 07:45:49 crc kubenswrapper[4971]: E1213 07:45:49.211258 4971 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="fa8714e3-d1ca-412a-9008-3f116341671a" containerName="extract-utilities" Dec 13 07:45:49 crc kubenswrapper[4971]: I1213 07:45:49.211266 4971 state_mem.go:107] "Deleted CPUSet assignment" podUID="fa8714e3-d1ca-412a-9008-3f116341671a" containerName="extract-utilities" Dec 13 07:45:49 crc kubenswrapper[4971]: E1213 07:45:49.211277 4971 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="9f1e6cde-d76e-45f1-a6a6-8546b16e389b" containerName="ovn-edpm-deployment-openstack-edpm-ipam" Dec 13 07:45:49 crc kubenswrapper[4971]: I1213 07:45:49.211283 4971 state_mem.go:107] "Deleted CPUSet assignment" podUID="9f1e6cde-d76e-45f1-a6a6-8546b16e389b" containerName="ovn-edpm-deployment-openstack-edpm-ipam" Dec 13 07:45:49 crc kubenswrapper[4971]: I1213 07:45:49.211497 4971 memory_manager.go:354] "RemoveStaleState removing state" podUID="fa8714e3-d1ca-412a-9008-3f116341671a" containerName="registry-server" Dec 13 07:45:49 crc kubenswrapper[4971]: I1213 07:45:49.211544 4971 memory_manager.go:354] "RemoveStaleState removing state" podUID="9f1e6cde-d76e-45f1-a6a6-8546b16e389b" containerName="ovn-edpm-deployment-openstack-edpm-ipam" Dec 13 07:45:49 crc kubenswrapper[4971]: I1213 07:45:49.212308 4971 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-tpz7w" Dec 13 07:45:49 crc kubenswrapper[4971]: I1213 07:45:49.215971 4971 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-aee-default-env" Dec 13 07:45:49 crc kubenswrapper[4971]: I1213 07:45:49.216175 4971 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplanenodeset-openstack-edpm-ipam" Dec 13 07:45:49 crc kubenswrapper[4971]: I1213 07:45:49.216249 4971 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-metadata-neutron-config" Dec 13 07:45:49 crc kubenswrapper[4971]: I1213 07:45:49.216397 4971 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-edpm-ipam-dockercfg-zb2j5" Dec 13 07:45:49 crc kubenswrapper[4971]: I1213 07:45:49.216417 4971 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceph-conf-files" Dec 13 07:45:49 crc kubenswrapper[4971]: I1213 07:45:49.216577 4971 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"neutron-ovn-metadata-agent-neutron-config" Dec 13 07:45:49 crc kubenswrapper[4971]: I1213 07:45:49.217983 4971 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplane-ansible-ssh-private-key-secret" Dec 13 07:45:49 crc kubenswrapper[4971]: I1213 07:45:49.226788 4971 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-tpz7w"] Dec 13 07:45:49 crc kubenswrapper[4971]: I1213 07:45:49.355587 4971 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"neutron-ovn-metadata-agent-neutron-config-0\" (UniqueName: \"kubernetes.io/secret/3a9bc76d-7313-4304-ad56-d6c075837f50-neutron-ovn-metadata-agent-neutron-config-0\") pod \"neutron-metadata-edpm-deployment-openstack-edpm-ipam-tpz7w\" (UID: \"3a9bc76d-7313-4304-ad56-d6c075837f50\") " pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-tpz7w" Dec 13 07:45:49 crc kubenswrapper[4971]: I1213 07:45:49.355653 4971 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/3a9bc76d-7313-4304-ad56-d6c075837f50-ssh-key\") pod \"neutron-metadata-edpm-deployment-openstack-edpm-ipam-tpz7w\" (UID: \"3a9bc76d-7313-4304-ad56-d6c075837f50\") " pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-tpz7w" Dec 13 07:45:49 crc kubenswrapper[4971]: I1213 07:45:49.355696 4971 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/3a9bc76d-7313-4304-ad56-d6c075837f50-inventory\") pod \"neutron-metadata-edpm-deployment-openstack-edpm-ipam-tpz7w\" (UID: \"3a9bc76d-7313-4304-ad56-d6c075837f50\") " pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-tpz7w" Dec 13 07:45:49 crc kubenswrapper[4971]: I1213 07:45:49.355725 4971 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"neutron-metadata-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/3a9bc76d-7313-4304-ad56-d6c075837f50-neutron-metadata-combined-ca-bundle\") pod \"neutron-metadata-edpm-deployment-openstack-edpm-ipam-tpz7w\" (UID: \"3a9bc76d-7313-4304-ad56-d6c075837f50\") " pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-tpz7w" Dec 13 07:45:49 crc kubenswrapper[4971]: I1213 07:45:49.355778 4971 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"nova-metadata-neutron-config-0\" (UniqueName: \"kubernetes.io/secret/3a9bc76d-7313-4304-ad56-d6c075837f50-nova-metadata-neutron-config-0\") pod \"neutron-metadata-edpm-deployment-openstack-edpm-ipam-tpz7w\" (UID: \"3a9bc76d-7313-4304-ad56-d6c075837f50\") " pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-tpz7w" Dec 13 07:45:49 crc kubenswrapper[4971]: I1213 07:45:49.355811 4971 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/3a9bc76d-7313-4304-ad56-d6c075837f50-ceph\") pod \"neutron-metadata-edpm-deployment-openstack-edpm-ipam-tpz7w\" (UID: \"3a9bc76d-7313-4304-ad56-d6c075837f50\") " pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-tpz7w" Dec 13 07:45:49 crc kubenswrapper[4971]: I1213 07:45:49.355843 4971 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-pvdfh\" (UniqueName: \"kubernetes.io/projected/3a9bc76d-7313-4304-ad56-d6c075837f50-kube-api-access-pvdfh\") pod \"neutron-metadata-edpm-deployment-openstack-edpm-ipam-tpz7w\" (UID: \"3a9bc76d-7313-4304-ad56-d6c075837f50\") " pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-tpz7w" Dec 13 07:45:49 crc kubenswrapper[4971]: I1213 07:45:49.457359 4971 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/3a9bc76d-7313-4304-ad56-d6c075837f50-ssh-key\") pod \"neutron-metadata-edpm-deployment-openstack-edpm-ipam-tpz7w\" (UID: \"3a9bc76d-7313-4304-ad56-d6c075837f50\") " pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-tpz7w" Dec 13 07:45:49 crc kubenswrapper[4971]: I1213 07:45:49.457447 4971 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/3a9bc76d-7313-4304-ad56-d6c075837f50-inventory\") pod \"neutron-metadata-edpm-deployment-openstack-edpm-ipam-tpz7w\" (UID: \"3a9bc76d-7313-4304-ad56-d6c075837f50\") " pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-tpz7w" Dec 13 07:45:49 crc kubenswrapper[4971]: I1213 07:45:49.457492 4971 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"neutron-metadata-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/3a9bc76d-7313-4304-ad56-d6c075837f50-neutron-metadata-combined-ca-bundle\") pod \"neutron-metadata-edpm-deployment-openstack-edpm-ipam-tpz7w\" (UID: \"3a9bc76d-7313-4304-ad56-d6c075837f50\") " pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-tpz7w" Dec 13 07:45:49 crc kubenswrapper[4971]: I1213 07:45:49.457586 4971 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nova-metadata-neutron-config-0\" (UniqueName: \"kubernetes.io/secret/3a9bc76d-7313-4304-ad56-d6c075837f50-nova-metadata-neutron-config-0\") pod \"neutron-metadata-edpm-deployment-openstack-edpm-ipam-tpz7w\" (UID: \"3a9bc76d-7313-4304-ad56-d6c075837f50\") " pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-tpz7w" Dec 13 07:45:49 crc kubenswrapper[4971]: I1213 07:45:49.457609 4971 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/3a9bc76d-7313-4304-ad56-d6c075837f50-ceph\") pod \"neutron-metadata-edpm-deployment-openstack-edpm-ipam-tpz7w\" (UID: \"3a9bc76d-7313-4304-ad56-d6c075837f50\") " pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-tpz7w" Dec 13 07:45:49 crc kubenswrapper[4971]: I1213 07:45:49.457654 4971 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-pvdfh\" (UniqueName: \"kubernetes.io/projected/3a9bc76d-7313-4304-ad56-d6c075837f50-kube-api-access-pvdfh\") pod \"neutron-metadata-edpm-deployment-openstack-edpm-ipam-tpz7w\" (UID: \"3a9bc76d-7313-4304-ad56-d6c075837f50\") " pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-tpz7w" Dec 13 07:45:49 crc kubenswrapper[4971]: I1213 07:45:49.457708 4971 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"neutron-ovn-metadata-agent-neutron-config-0\" (UniqueName: \"kubernetes.io/secret/3a9bc76d-7313-4304-ad56-d6c075837f50-neutron-ovn-metadata-agent-neutron-config-0\") pod \"neutron-metadata-edpm-deployment-openstack-edpm-ipam-tpz7w\" (UID: \"3a9bc76d-7313-4304-ad56-d6c075837f50\") " pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-tpz7w" Dec 13 07:45:49 crc kubenswrapper[4971]: I1213 07:45:49.463352 4971 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/3a9bc76d-7313-4304-ad56-d6c075837f50-inventory\") pod \"neutron-metadata-edpm-deployment-openstack-edpm-ipam-tpz7w\" (UID: \"3a9bc76d-7313-4304-ad56-d6c075837f50\") " pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-tpz7w" Dec 13 07:45:49 crc kubenswrapper[4971]: I1213 07:45:49.463600 4971 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"neutron-ovn-metadata-agent-neutron-config-0\" (UniqueName: \"kubernetes.io/secret/3a9bc76d-7313-4304-ad56-d6c075837f50-neutron-ovn-metadata-agent-neutron-config-0\") pod \"neutron-metadata-edpm-deployment-openstack-edpm-ipam-tpz7w\" (UID: \"3a9bc76d-7313-4304-ad56-d6c075837f50\") " pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-tpz7w" Dec 13 07:45:49 crc kubenswrapper[4971]: I1213 07:45:49.463970 4971 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"nova-metadata-neutron-config-0\" (UniqueName: \"kubernetes.io/secret/3a9bc76d-7313-4304-ad56-d6c075837f50-nova-metadata-neutron-config-0\") pod \"neutron-metadata-edpm-deployment-openstack-edpm-ipam-tpz7w\" (UID: \"3a9bc76d-7313-4304-ad56-d6c075837f50\") " pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-tpz7w" Dec 13 07:45:49 crc kubenswrapper[4971]: I1213 07:45:49.464905 4971 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/3a9bc76d-7313-4304-ad56-d6c075837f50-ssh-key\") pod \"neutron-metadata-edpm-deployment-openstack-edpm-ipam-tpz7w\" (UID: \"3a9bc76d-7313-4304-ad56-d6c075837f50\") " pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-tpz7w" Dec 13 07:45:49 crc kubenswrapper[4971]: I1213 07:45:49.465642 4971 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"neutron-metadata-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/3a9bc76d-7313-4304-ad56-d6c075837f50-neutron-metadata-combined-ca-bundle\") pod \"neutron-metadata-edpm-deployment-openstack-edpm-ipam-tpz7w\" (UID: \"3a9bc76d-7313-4304-ad56-d6c075837f50\") " pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-tpz7w" Dec 13 07:45:49 crc kubenswrapper[4971]: I1213 07:45:49.466813 4971 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/3a9bc76d-7313-4304-ad56-d6c075837f50-ceph\") pod \"neutron-metadata-edpm-deployment-openstack-edpm-ipam-tpz7w\" (UID: \"3a9bc76d-7313-4304-ad56-d6c075837f50\") " pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-tpz7w" Dec 13 07:45:49 crc kubenswrapper[4971]: I1213 07:45:49.482411 4971 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-pvdfh\" (UniqueName: \"kubernetes.io/projected/3a9bc76d-7313-4304-ad56-d6c075837f50-kube-api-access-pvdfh\") pod \"neutron-metadata-edpm-deployment-openstack-edpm-ipam-tpz7w\" (UID: \"3a9bc76d-7313-4304-ad56-d6c075837f50\") " pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-tpz7w" Dec 13 07:45:49 crc kubenswrapper[4971]: I1213 07:45:49.538502 4971 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-tpz7w" Dec 13 07:45:50 crc kubenswrapper[4971]: I1213 07:45:50.509907 4971 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-tpz7w"] Dec 13 07:45:51 crc kubenswrapper[4971]: I1213 07:45:51.257346 4971 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-tpz7w" event={"ID":"3a9bc76d-7313-4304-ad56-d6c075837f50","Type":"ContainerStarted","Data":"38a52126afa3241e2fd7ddc4cb0f4bcf858247a7084f211020a53e41c9cd3e8d"} Dec 13 07:45:52 crc kubenswrapper[4971]: I1213 07:45:52.273323 4971 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-tpz7w" event={"ID":"3a9bc76d-7313-4304-ad56-d6c075837f50","Type":"ContainerStarted","Data":"ddf5ebab40f7fe925c6975b9dd123bbc339822f41090991739c0cd9e4faf0cf8"} Dec 13 07:45:52 crc kubenswrapper[4971]: I1213 07:45:52.298386 4971 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-tpz7w" podStartSLOduration=2.144765501 podStartE2EDuration="3.298354908s" podCreationTimestamp="2025-12-13 07:45:49 +0000 UTC" firstStartedPulling="2025-12-13 07:45:50.521118958 +0000 UTC m=+3407.125528406" lastFinishedPulling="2025-12-13 07:45:51.674708325 +0000 UTC m=+3408.279117813" observedRunningTime="2025-12-13 07:45:52.28647751 +0000 UTC m=+3408.890886988" watchObservedRunningTime="2025-12-13 07:45:52.298354908 +0000 UTC m=+3408.902764396" Dec 13 07:46:16 crc kubenswrapper[4971]: I1213 07:46:16.153407 4971 patch_prober.go:28] interesting pod/machine-config-daemon-82xjz container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 13 07:46:16 crc kubenswrapper[4971]: I1213 07:46:16.154153 4971 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-82xjz" podUID="e0c4af12-edda-41d7-9b44-f87396174ef2" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 13 07:46:46 crc kubenswrapper[4971]: I1213 07:46:46.154060 4971 patch_prober.go:28] interesting pod/machine-config-daemon-82xjz container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 13 07:46:46 crc kubenswrapper[4971]: I1213 07:46:46.154534 4971 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-82xjz" podUID="e0c4af12-edda-41d7-9b44-f87396174ef2" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 13 07:46:46 crc kubenswrapper[4971]: I1213 07:46:46.154576 4971 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-82xjz" Dec 13 07:46:46 crc kubenswrapper[4971]: I1213 07:46:46.155262 4971 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"95874f24e419c2abe48edb2569aa8a50c62f04d8d5d3a5215adeee275f78b4d8"} pod="openshift-machine-config-operator/machine-config-daemon-82xjz" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Dec 13 07:46:46 crc kubenswrapper[4971]: I1213 07:46:46.155313 4971 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-82xjz" podUID="e0c4af12-edda-41d7-9b44-f87396174ef2" containerName="machine-config-daemon" containerID="cri-o://95874f24e419c2abe48edb2569aa8a50c62f04d8d5d3a5215adeee275f78b4d8" gracePeriod=600 Dec 13 07:46:46 crc kubenswrapper[4971]: I1213 07:46:46.786808 4971 generic.go:334] "Generic (PLEG): container finished" podID="e0c4af12-edda-41d7-9b44-f87396174ef2" containerID="95874f24e419c2abe48edb2569aa8a50c62f04d8d5d3a5215adeee275f78b4d8" exitCode=0 Dec 13 07:46:46 crc kubenswrapper[4971]: I1213 07:46:46.787008 4971 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-82xjz" event={"ID":"e0c4af12-edda-41d7-9b44-f87396174ef2","Type":"ContainerDied","Data":"95874f24e419c2abe48edb2569aa8a50c62f04d8d5d3a5215adeee275f78b4d8"} Dec 13 07:46:46 crc kubenswrapper[4971]: I1213 07:46:46.787155 4971 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-82xjz" event={"ID":"e0c4af12-edda-41d7-9b44-f87396174ef2","Type":"ContainerStarted","Data":"e975a5822755c28b3c504aa81fe38e85a7a58425e2a7f942f681aff10344af88"} Dec 13 07:46:46 crc kubenswrapper[4971]: I1213 07:46:46.787186 4971 scope.go:117] "RemoveContainer" containerID="09d0904c76ce78aa0b074447b1e64cf63160f573da086f934c8cb712615b71d4" Dec 13 07:47:54 crc kubenswrapper[4971]: I1213 07:47:54.674316 4971 generic.go:334] "Generic (PLEG): container finished" podID="3a9bc76d-7313-4304-ad56-d6c075837f50" containerID="ddf5ebab40f7fe925c6975b9dd123bbc339822f41090991739c0cd9e4faf0cf8" exitCode=0 Dec 13 07:47:54 crc kubenswrapper[4971]: I1213 07:47:54.674396 4971 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-tpz7w" event={"ID":"3a9bc76d-7313-4304-ad56-d6c075837f50","Type":"ContainerDied","Data":"ddf5ebab40f7fe925c6975b9dd123bbc339822f41090991739c0cd9e4faf0cf8"} Dec 13 07:47:56 crc kubenswrapper[4971]: I1213 07:47:56.099867 4971 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-tpz7w" Dec 13 07:47:56 crc kubenswrapper[4971]: I1213 07:47:56.253726 4971 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"nova-metadata-neutron-config-0\" (UniqueName: \"kubernetes.io/secret/3a9bc76d-7313-4304-ad56-d6c075837f50-nova-metadata-neutron-config-0\") pod \"3a9bc76d-7313-4304-ad56-d6c075837f50\" (UID: \"3a9bc76d-7313-4304-ad56-d6c075837f50\") " Dec 13 07:47:56 crc kubenswrapper[4971]: I1213 07:47:56.253838 4971 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"neutron-metadata-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/3a9bc76d-7313-4304-ad56-d6c075837f50-neutron-metadata-combined-ca-bundle\") pod \"3a9bc76d-7313-4304-ad56-d6c075837f50\" (UID: \"3a9bc76d-7313-4304-ad56-d6c075837f50\") " Dec 13 07:47:56 crc kubenswrapper[4971]: I1213 07:47:56.253925 4971 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/3a9bc76d-7313-4304-ad56-d6c075837f50-ssh-key\") pod \"3a9bc76d-7313-4304-ad56-d6c075837f50\" (UID: \"3a9bc76d-7313-4304-ad56-d6c075837f50\") " Dec 13 07:47:56 crc kubenswrapper[4971]: I1213 07:47:56.253984 4971 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/3a9bc76d-7313-4304-ad56-d6c075837f50-ceph\") pod \"3a9bc76d-7313-4304-ad56-d6c075837f50\" (UID: \"3a9bc76d-7313-4304-ad56-d6c075837f50\") " Dec 13 07:47:56 crc kubenswrapper[4971]: I1213 07:47:56.254013 4971 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"neutron-ovn-metadata-agent-neutron-config-0\" (UniqueName: \"kubernetes.io/secret/3a9bc76d-7313-4304-ad56-d6c075837f50-neutron-ovn-metadata-agent-neutron-config-0\") pod \"3a9bc76d-7313-4304-ad56-d6c075837f50\" (UID: \"3a9bc76d-7313-4304-ad56-d6c075837f50\") " Dec 13 07:47:56 crc kubenswrapper[4971]: I1213 07:47:56.254101 4971 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/3a9bc76d-7313-4304-ad56-d6c075837f50-inventory\") pod \"3a9bc76d-7313-4304-ad56-d6c075837f50\" (UID: \"3a9bc76d-7313-4304-ad56-d6c075837f50\") " Dec 13 07:47:56 crc kubenswrapper[4971]: I1213 07:47:56.254220 4971 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-pvdfh\" (UniqueName: \"kubernetes.io/projected/3a9bc76d-7313-4304-ad56-d6c075837f50-kube-api-access-pvdfh\") pod \"3a9bc76d-7313-4304-ad56-d6c075837f50\" (UID: \"3a9bc76d-7313-4304-ad56-d6c075837f50\") " Dec 13 07:47:56 crc kubenswrapper[4971]: I1213 07:47:56.263430 4971 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/3a9bc76d-7313-4304-ad56-d6c075837f50-neutron-metadata-combined-ca-bundle" (OuterVolumeSpecName: "neutron-metadata-combined-ca-bundle") pod "3a9bc76d-7313-4304-ad56-d6c075837f50" (UID: "3a9bc76d-7313-4304-ad56-d6c075837f50"). InnerVolumeSpecName "neutron-metadata-combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 13 07:47:56 crc kubenswrapper[4971]: I1213 07:47:56.263501 4971 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/3a9bc76d-7313-4304-ad56-d6c075837f50-ceph" (OuterVolumeSpecName: "ceph") pod "3a9bc76d-7313-4304-ad56-d6c075837f50" (UID: "3a9bc76d-7313-4304-ad56-d6c075837f50"). InnerVolumeSpecName "ceph". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 13 07:47:56 crc kubenswrapper[4971]: I1213 07:47:56.263634 4971 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/3a9bc76d-7313-4304-ad56-d6c075837f50-kube-api-access-pvdfh" (OuterVolumeSpecName: "kube-api-access-pvdfh") pod "3a9bc76d-7313-4304-ad56-d6c075837f50" (UID: "3a9bc76d-7313-4304-ad56-d6c075837f50"). InnerVolumeSpecName "kube-api-access-pvdfh". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 13 07:47:56 crc kubenswrapper[4971]: I1213 07:47:56.285559 4971 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/3a9bc76d-7313-4304-ad56-d6c075837f50-nova-metadata-neutron-config-0" (OuterVolumeSpecName: "nova-metadata-neutron-config-0") pod "3a9bc76d-7313-4304-ad56-d6c075837f50" (UID: "3a9bc76d-7313-4304-ad56-d6c075837f50"). InnerVolumeSpecName "nova-metadata-neutron-config-0". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 13 07:47:56 crc kubenswrapper[4971]: I1213 07:47:56.285658 4971 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/3a9bc76d-7313-4304-ad56-d6c075837f50-inventory" (OuterVolumeSpecName: "inventory") pod "3a9bc76d-7313-4304-ad56-d6c075837f50" (UID: "3a9bc76d-7313-4304-ad56-d6c075837f50"). InnerVolumeSpecName "inventory". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 13 07:47:56 crc kubenswrapper[4971]: I1213 07:47:56.286188 4971 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/3a9bc76d-7313-4304-ad56-d6c075837f50-neutron-ovn-metadata-agent-neutron-config-0" (OuterVolumeSpecName: "neutron-ovn-metadata-agent-neutron-config-0") pod "3a9bc76d-7313-4304-ad56-d6c075837f50" (UID: "3a9bc76d-7313-4304-ad56-d6c075837f50"). InnerVolumeSpecName "neutron-ovn-metadata-agent-neutron-config-0". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 13 07:47:56 crc kubenswrapper[4971]: I1213 07:47:56.288116 4971 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/3a9bc76d-7313-4304-ad56-d6c075837f50-ssh-key" (OuterVolumeSpecName: "ssh-key") pod "3a9bc76d-7313-4304-ad56-d6c075837f50" (UID: "3a9bc76d-7313-4304-ad56-d6c075837f50"). InnerVolumeSpecName "ssh-key". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 13 07:47:56 crc kubenswrapper[4971]: I1213 07:47:56.356149 4971 reconciler_common.go:293] "Volume detached for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/3a9bc76d-7313-4304-ad56-d6c075837f50-inventory\") on node \"crc\" DevicePath \"\"" Dec 13 07:47:56 crc kubenswrapper[4971]: I1213 07:47:56.356188 4971 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-pvdfh\" (UniqueName: \"kubernetes.io/projected/3a9bc76d-7313-4304-ad56-d6c075837f50-kube-api-access-pvdfh\") on node \"crc\" DevicePath \"\"" Dec 13 07:47:56 crc kubenswrapper[4971]: I1213 07:47:56.356201 4971 reconciler_common.go:293] "Volume detached for volume \"nova-metadata-neutron-config-0\" (UniqueName: \"kubernetes.io/secret/3a9bc76d-7313-4304-ad56-d6c075837f50-nova-metadata-neutron-config-0\") on node \"crc\" DevicePath \"\"" Dec 13 07:47:56 crc kubenswrapper[4971]: I1213 07:47:56.356211 4971 reconciler_common.go:293] "Volume detached for volume \"neutron-metadata-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/3a9bc76d-7313-4304-ad56-d6c075837f50-neutron-metadata-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 13 07:47:56 crc kubenswrapper[4971]: I1213 07:47:56.356222 4971 reconciler_common.go:293] "Volume detached for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/3a9bc76d-7313-4304-ad56-d6c075837f50-ssh-key\") on node \"crc\" DevicePath \"\"" Dec 13 07:47:56 crc kubenswrapper[4971]: I1213 07:47:56.356280 4971 reconciler_common.go:293] "Volume detached for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/3a9bc76d-7313-4304-ad56-d6c075837f50-ceph\") on node \"crc\" DevicePath \"\"" Dec 13 07:47:56 crc kubenswrapper[4971]: I1213 07:47:56.356292 4971 reconciler_common.go:293] "Volume detached for volume \"neutron-ovn-metadata-agent-neutron-config-0\" (UniqueName: \"kubernetes.io/secret/3a9bc76d-7313-4304-ad56-d6c075837f50-neutron-ovn-metadata-agent-neutron-config-0\") on node \"crc\" DevicePath \"\"" Dec 13 07:47:56 crc kubenswrapper[4971]: I1213 07:47:56.692541 4971 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-tpz7w" event={"ID":"3a9bc76d-7313-4304-ad56-d6c075837f50","Type":"ContainerDied","Data":"38a52126afa3241e2fd7ddc4cb0f4bcf858247a7084f211020a53e41c9cd3e8d"} Dec 13 07:47:56 crc kubenswrapper[4971]: I1213 07:47:56.693079 4971 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="38a52126afa3241e2fd7ddc4cb0f4bcf858247a7084f211020a53e41c9cd3e8d" Dec 13 07:47:56 crc kubenswrapper[4971]: I1213 07:47:56.692851 4971 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-tpz7w" Dec 13 07:47:56 crc kubenswrapper[4971]: I1213 07:47:56.781329 4971 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/libvirt-edpm-deployment-openstack-edpm-ipam-47wpm"] Dec 13 07:47:56 crc kubenswrapper[4971]: E1213 07:47:56.781718 4971 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="3a9bc76d-7313-4304-ad56-d6c075837f50" containerName="neutron-metadata-edpm-deployment-openstack-edpm-ipam" Dec 13 07:47:56 crc kubenswrapper[4971]: I1213 07:47:56.781738 4971 state_mem.go:107] "Deleted CPUSet assignment" podUID="3a9bc76d-7313-4304-ad56-d6c075837f50" containerName="neutron-metadata-edpm-deployment-openstack-edpm-ipam" Dec 13 07:47:56 crc kubenswrapper[4971]: I1213 07:47:56.781906 4971 memory_manager.go:354] "RemoveStaleState removing state" podUID="3a9bc76d-7313-4304-ad56-d6c075837f50" containerName="neutron-metadata-edpm-deployment-openstack-edpm-ipam" Dec 13 07:47:56 crc kubenswrapper[4971]: I1213 07:47:56.782505 4971 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/libvirt-edpm-deployment-openstack-edpm-ipam-47wpm" Dec 13 07:47:56 crc kubenswrapper[4971]: I1213 07:47:56.785283 4971 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplanenodeset-openstack-edpm-ipam" Dec 13 07:47:56 crc kubenswrapper[4971]: I1213 07:47:56.785316 4971 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplane-ansible-ssh-private-key-secret" Dec 13 07:47:56 crc kubenswrapper[4971]: I1213 07:47:56.785450 4971 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-aee-default-env" Dec 13 07:47:56 crc kubenswrapper[4971]: I1213 07:47:56.785615 4971 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-edpm-ipam-dockercfg-zb2j5" Dec 13 07:47:56 crc kubenswrapper[4971]: I1213 07:47:56.785675 4971 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"libvirt-secret" Dec 13 07:47:56 crc kubenswrapper[4971]: I1213 07:47:56.787919 4971 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceph-conf-files" Dec 13 07:47:56 crc kubenswrapper[4971]: I1213 07:47:56.811711 4971 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/libvirt-edpm-deployment-openstack-edpm-ipam-47wpm"] Dec 13 07:47:56 crc kubenswrapper[4971]: I1213 07:47:56.865343 4971 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/ce803e3c-be8a-4ee3-ac0b-42b035c043a5-ssh-key\") pod \"libvirt-edpm-deployment-openstack-edpm-ipam-47wpm\" (UID: \"ce803e3c-be8a-4ee3-ac0b-42b035c043a5\") " pod="openstack/libvirt-edpm-deployment-openstack-edpm-ipam-47wpm" Dec 13 07:47:56 crc kubenswrapper[4971]: I1213 07:47:56.865406 4971 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"libvirt-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/ce803e3c-be8a-4ee3-ac0b-42b035c043a5-libvirt-combined-ca-bundle\") pod \"libvirt-edpm-deployment-openstack-edpm-ipam-47wpm\" (UID: \"ce803e3c-be8a-4ee3-ac0b-42b035c043a5\") " pod="openstack/libvirt-edpm-deployment-openstack-edpm-ipam-47wpm" Dec 13 07:47:56 crc kubenswrapper[4971]: I1213 07:47:56.865588 4971 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/ce803e3c-be8a-4ee3-ac0b-42b035c043a5-ceph\") pod \"libvirt-edpm-deployment-openstack-edpm-ipam-47wpm\" (UID: \"ce803e3c-be8a-4ee3-ac0b-42b035c043a5\") " pod="openstack/libvirt-edpm-deployment-openstack-edpm-ipam-47wpm" Dec 13 07:47:56 crc kubenswrapper[4971]: I1213 07:47:56.865822 4971 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/ce803e3c-be8a-4ee3-ac0b-42b035c043a5-inventory\") pod \"libvirt-edpm-deployment-openstack-edpm-ipam-47wpm\" (UID: \"ce803e3c-be8a-4ee3-ac0b-42b035c043a5\") " pod="openstack/libvirt-edpm-deployment-openstack-edpm-ipam-47wpm" Dec 13 07:47:56 crc kubenswrapper[4971]: I1213 07:47:56.865889 4971 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-mpl6j\" (UniqueName: \"kubernetes.io/projected/ce803e3c-be8a-4ee3-ac0b-42b035c043a5-kube-api-access-mpl6j\") pod \"libvirt-edpm-deployment-openstack-edpm-ipam-47wpm\" (UID: \"ce803e3c-be8a-4ee3-ac0b-42b035c043a5\") " pod="openstack/libvirt-edpm-deployment-openstack-edpm-ipam-47wpm" Dec 13 07:47:56 crc kubenswrapper[4971]: I1213 07:47:56.865969 4971 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"libvirt-secret-0\" (UniqueName: \"kubernetes.io/secret/ce803e3c-be8a-4ee3-ac0b-42b035c043a5-libvirt-secret-0\") pod \"libvirt-edpm-deployment-openstack-edpm-ipam-47wpm\" (UID: \"ce803e3c-be8a-4ee3-ac0b-42b035c043a5\") " pod="openstack/libvirt-edpm-deployment-openstack-edpm-ipam-47wpm" Dec 13 07:47:56 crc kubenswrapper[4971]: I1213 07:47:56.967021 4971 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/ce803e3c-be8a-4ee3-ac0b-42b035c043a5-ceph\") pod \"libvirt-edpm-deployment-openstack-edpm-ipam-47wpm\" (UID: \"ce803e3c-be8a-4ee3-ac0b-42b035c043a5\") " pod="openstack/libvirt-edpm-deployment-openstack-edpm-ipam-47wpm" Dec 13 07:47:56 crc kubenswrapper[4971]: I1213 07:47:56.967183 4971 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/ce803e3c-be8a-4ee3-ac0b-42b035c043a5-inventory\") pod \"libvirt-edpm-deployment-openstack-edpm-ipam-47wpm\" (UID: \"ce803e3c-be8a-4ee3-ac0b-42b035c043a5\") " pod="openstack/libvirt-edpm-deployment-openstack-edpm-ipam-47wpm" Dec 13 07:47:56 crc kubenswrapper[4971]: I1213 07:47:56.967219 4971 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-mpl6j\" (UniqueName: \"kubernetes.io/projected/ce803e3c-be8a-4ee3-ac0b-42b035c043a5-kube-api-access-mpl6j\") pod \"libvirt-edpm-deployment-openstack-edpm-ipam-47wpm\" (UID: \"ce803e3c-be8a-4ee3-ac0b-42b035c043a5\") " pod="openstack/libvirt-edpm-deployment-openstack-edpm-ipam-47wpm" Dec 13 07:47:56 crc kubenswrapper[4971]: I1213 07:47:56.967284 4971 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"libvirt-secret-0\" (UniqueName: \"kubernetes.io/secret/ce803e3c-be8a-4ee3-ac0b-42b035c043a5-libvirt-secret-0\") pod \"libvirt-edpm-deployment-openstack-edpm-ipam-47wpm\" (UID: \"ce803e3c-be8a-4ee3-ac0b-42b035c043a5\") " pod="openstack/libvirt-edpm-deployment-openstack-edpm-ipam-47wpm" Dec 13 07:47:56 crc kubenswrapper[4971]: I1213 07:47:56.967360 4971 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/ce803e3c-be8a-4ee3-ac0b-42b035c043a5-ssh-key\") pod \"libvirt-edpm-deployment-openstack-edpm-ipam-47wpm\" (UID: \"ce803e3c-be8a-4ee3-ac0b-42b035c043a5\") " pod="openstack/libvirt-edpm-deployment-openstack-edpm-ipam-47wpm" Dec 13 07:47:56 crc kubenswrapper[4971]: I1213 07:47:56.967403 4971 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"libvirt-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/ce803e3c-be8a-4ee3-ac0b-42b035c043a5-libvirt-combined-ca-bundle\") pod \"libvirt-edpm-deployment-openstack-edpm-ipam-47wpm\" (UID: \"ce803e3c-be8a-4ee3-ac0b-42b035c043a5\") " pod="openstack/libvirt-edpm-deployment-openstack-edpm-ipam-47wpm" Dec 13 07:47:56 crc kubenswrapper[4971]: I1213 07:47:56.971871 4971 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/ce803e3c-be8a-4ee3-ac0b-42b035c043a5-inventory\") pod \"libvirt-edpm-deployment-openstack-edpm-ipam-47wpm\" (UID: \"ce803e3c-be8a-4ee3-ac0b-42b035c043a5\") " pod="openstack/libvirt-edpm-deployment-openstack-edpm-ipam-47wpm" Dec 13 07:47:56 crc kubenswrapper[4971]: I1213 07:47:56.972391 4971 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/ce803e3c-be8a-4ee3-ac0b-42b035c043a5-ceph\") pod \"libvirt-edpm-deployment-openstack-edpm-ipam-47wpm\" (UID: \"ce803e3c-be8a-4ee3-ac0b-42b035c043a5\") " pod="openstack/libvirt-edpm-deployment-openstack-edpm-ipam-47wpm" Dec 13 07:47:56 crc kubenswrapper[4971]: I1213 07:47:56.974174 4971 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/ce803e3c-be8a-4ee3-ac0b-42b035c043a5-ssh-key\") pod \"libvirt-edpm-deployment-openstack-edpm-ipam-47wpm\" (UID: \"ce803e3c-be8a-4ee3-ac0b-42b035c043a5\") " pod="openstack/libvirt-edpm-deployment-openstack-edpm-ipam-47wpm" Dec 13 07:47:56 crc kubenswrapper[4971]: I1213 07:47:56.974420 4971 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"libvirt-secret-0\" (UniqueName: \"kubernetes.io/secret/ce803e3c-be8a-4ee3-ac0b-42b035c043a5-libvirt-secret-0\") pod \"libvirt-edpm-deployment-openstack-edpm-ipam-47wpm\" (UID: \"ce803e3c-be8a-4ee3-ac0b-42b035c043a5\") " pod="openstack/libvirt-edpm-deployment-openstack-edpm-ipam-47wpm" Dec 13 07:47:56 crc kubenswrapper[4971]: I1213 07:47:56.975793 4971 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"libvirt-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/ce803e3c-be8a-4ee3-ac0b-42b035c043a5-libvirt-combined-ca-bundle\") pod \"libvirt-edpm-deployment-openstack-edpm-ipam-47wpm\" (UID: \"ce803e3c-be8a-4ee3-ac0b-42b035c043a5\") " pod="openstack/libvirt-edpm-deployment-openstack-edpm-ipam-47wpm" Dec 13 07:47:56 crc kubenswrapper[4971]: I1213 07:47:56.986383 4971 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-mpl6j\" (UniqueName: \"kubernetes.io/projected/ce803e3c-be8a-4ee3-ac0b-42b035c043a5-kube-api-access-mpl6j\") pod \"libvirt-edpm-deployment-openstack-edpm-ipam-47wpm\" (UID: \"ce803e3c-be8a-4ee3-ac0b-42b035c043a5\") " pod="openstack/libvirt-edpm-deployment-openstack-edpm-ipam-47wpm" Dec 13 07:47:57 crc kubenswrapper[4971]: I1213 07:47:57.110215 4971 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/libvirt-edpm-deployment-openstack-edpm-ipam-47wpm" Dec 13 07:47:57 crc kubenswrapper[4971]: I1213 07:47:57.627203 4971 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/libvirt-edpm-deployment-openstack-edpm-ipam-47wpm"] Dec 13 07:47:57 crc kubenswrapper[4971]: I1213 07:47:57.703218 4971 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/libvirt-edpm-deployment-openstack-edpm-ipam-47wpm" event={"ID":"ce803e3c-be8a-4ee3-ac0b-42b035c043a5","Type":"ContainerStarted","Data":"1aafacb341ba7b8e7758aa19be358053c9cc26dcc6a38caf8b4f2569407c0613"} Dec 13 07:47:58 crc kubenswrapper[4971]: I1213 07:47:58.714931 4971 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/libvirt-edpm-deployment-openstack-edpm-ipam-47wpm" event={"ID":"ce803e3c-be8a-4ee3-ac0b-42b035c043a5","Type":"ContainerStarted","Data":"3dd6fb5bc5f369c42726e16bbfdbf2ee181d92464f6ddc5409e5863fb73c10b2"} Dec 13 07:47:58 crc kubenswrapper[4971]: I1213 07:47:58.739839 4971 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/libvirt-edpm-deployment-openstack-edpm-ipam-47wpm" podStartSLOduration=2.182502003 podStartE2EDuration="2.739814326s" podCreationTimestamp="2025-12-13 07:47:56 +0000 UTC" firstStartedPulling="2025-12-13 07:47:57.634972488 +0000 UTC m=+3534.239381946" lastFinishedPulling="2025-12-13 07:47:58.192284781 +0000 UTC m=+3534.796694269" observedRunningTime="2025-12-13 07:47:58.735624518 +0000 UTC m=+3535.340034006" watchObservedRunningTime="2025-12-13 07:47:58.739814326 +0000 UTC m=+3535.344223794" Dec 13 07:48:26 crc kubenswrapper[4971]: I1213 07:48:26.556231 4971 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/community-operators-6nsgx"] Dec 13 07:48:26 crc kubenswrapper[4971]: I1213 07:48:26.560998 4971 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-6nsgx" Dec 13 07:48:26 crc kubenswrapper[4971]: I1213 07:48:26.571297 4971 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-6nsgx"] Dec 13 07:48:26 crc kubenswrapper[4971]: I1213 07:48:26.644456 4971 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/b807ae25-7d1d-4335-946a-45ffb0160d36-catalog-content\") pod \"community-operators-6nsgx\" (UID: \"b807ae25-7d1d-4335-946a-45ffb0160d36\") " pod="openshift-marketplace/community-operators-6nsgx" Dec 13 07:48:26 crc kubenswrapper[4971]: I1213 07:48:26.644724 4971 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/b807ae25-7d1d-4335-946a-45ffb0160d36-utilities\") pod \"community-operators-6nsgx\" (UID: \"b807ae25-7d1d-4335-946a-45ffb0160d36\") " pod="openshift-marketplace/community-operators-6nsgx" Dec 13 07:48:26 crc kubenswrapper[4971]: I1213 07:48:26.644783 4971 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-m5dgm\" (UniqueName: \"kubernetes.io/projected/b807ae25-7d1d-4335-946a-45ffb0160d36-kube-api-access-m5dgm\") pod \"community-operators-6nsgx\" (UID: \"b807ae25-7d1d-4335-946a-45ffb0160d36\") " pod="openshift-marketplace/community-operators-6nsgx" Dec 13 07:48:26 crc kubenswrapper[4971]: I1213 07:48:26.746793 4971 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/b807ae25-7d1d-4335-946a-45ffb0160d36-utilities\") pod \"community-operators-6nsgx\" (UID: \"b807ae25-7d1d-4335-946a-45ffb0160d36\") " pod="openshift-marketplace/community-operators-6nsgx" Dec 13 07:48:26 crc kubenswrapper[4971]: I1213 07:48:26.746850 4971 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-m5dgm\" (UniqueName: \"kubernetes.io/projected/b807ae25-7d1d-4335-946a-45ffb0160d36-kube-api-access-m5dgm\") pod \"community-operators-6nsgx\" (UID: \"b807ae25-7d1d-4335-946a-45ffb0160d36\") " pod="openshift-marketplace/community-operators-6nsgx" Dec 13 07:48:26 crc kubenswrapper[4971]: I1213 07:48:26.746927 4971 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/b807ae25-7d1d-4335-946a-45ffb0160d36-catalog-content\") pod \"community-operators-6nsgx\" (UID: \"b807ae25-7d1d-4335-946a-45ffb0160d36\") " pod="openshift-marketplace/community-operators-6nsgx" Dec 13 07:48:26 crc kubenswrapper[4971]: I1213 07:48:26.747298 4971 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/b807ae25-7d1d-4335-946a-45ffb0160d36-utilities\") pod \"community-operators-6nsgx\" (UID: \"b807ae25-7d1d-4335-946a-45ffb0160d36\") " pod="openshift-marketplace/community-operators-6nsgx" Dec 13 07:48:26 crc kubenswrapper[4971]: I1213 07:48:26.747388 4971 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/b807ae25-7d1d-4335-946a-45ffb0160d36-catalog-content\") pod \"community-operators-6nsgx\" (UID: \"b807ae25-7d1d-4335-946a-45ffb0160d36\") " pod="openshift-marketplace/community-operators-6nsgx" Dec 13 07:48:26 crc kubenswrapper[4971]: I1213 07:48:26.773431 4971 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-m5dgm\" (UniqueName: \"kubernetes.io/projected/b807ae25-7d1d-4335-946a-45ffb0160d36-kube-api-access-m5dgm\") pod \"community-operators-6nsgx\" (UID: \"b807ae25-7d1d-4335-946a-45ffb0160d36\") " pod="openshift-marketplace/community-operators-6nsgx" Dec 13 07:48:26 crc kubenswrapper[4971]: I1213 07:48:26.898658 4971 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-6nsgx" Dec 13 07:48:27 crc kubenswrapper[4971]: I1213 07:48:27.467354 4971 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-6nsgx"] Dec 13 07:48:28 crc kubenswrapper[4971]: I1213 07:48:28.014439 4971 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-6nsgx" event={"ID":"b807ae25-7d1d-4335-946a-45ffb0160d36","Type":"ContainerStarted","Data":"747393bd09404d668ac9f8feb01e0d5f24f87f4b6793f12d9c7c33cad40538d5"} Dec 13 07:48:29 crc kubenswrapper[4971]: I1213 07:48:29.024003 4971 generic.go:334] "Generic (PLEG): container finished" podID="b807ae25-7d1d-4335-946a-45ffb0160d36" containerID="b637e61b5b3c1aefd1d1b602e972b2219e1d724cee342090f605ea27b4722be3" exitCode=0 Dec 13 07:48:29 crc kubenswrapper[4971]: I1213 07:48:29.024278 4971 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-6nsgx" event={"ID":"b807ae25-7d1d-4335-946a-45ffb0160d36","Type":"ContainerDied","Data":"b637e61b5b3c1aefd1d1b602e972b2219e1d724cee342090f605ea27b4722be3"} Dec 13 07:48:31 crc kubenswrapper[4971]: I1213 07:48:31.046641 4971 generic.go:334] "Generic (PLEG): container finished" podID="b807ae25-7d1d-4335-946a-45ffb0160d36" containerID="c1b860aae170e2efbd9b88faa1fe108392ee0fa7c75c88aba4be6d0bfb88f568" exitCode=0 Dec 13 07:48:31 crc kubenswrapper[4971]: I1213 07:48:31.046708 4971 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-6nsgx" event={"ID":"b807ae25-7d1d-4335-946a-45ffb0160d36","Type":"ContainerDied","Data":"c1b860aae170e2efbd9b88faa1fe108392ee0fa7c75c88aba4be6d0bfb88f568"} Dec 13 07:48:33 crc kubenswrapper[4971]: I1213 07:48:33.066089 4971 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-6nsgx" event={"ID":"b807ae25-7d1d-4335-946a-45ffb0160d36","Type":"ContainerStarted","Data":"2d778bfafde3c5c8513578df5229341a4a3a53a549f8965d7f39597ed0a7622a"} Dec 13 07:48:33 crc kubenswrapper[4971]: I1213 07:48:33.095256 4971 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/community-operators-6nsgx" podStartSLOduration=4.109245861 podStartE2EDuration="7.095235607s" podCreationTimestamp="2025-12-13 07:48:26 +0000 UTC" firstStartedPulling="2025-12-13 07:48:29.02593248 +0000 UTC m=+3565.630341938" lastFinishedPulling="2025-12-13 07:48:32.011922236 +0000 UTC m=+3568.616331684" observedRunningTime="2025-12-13 07:48:33.08698701 +0000 UTC m=+3569.691396458" watchObservedRunningTime="2025-12-13 07:48:33.095235607 +0000 UTC m=+3569.699645055" Dec 13 07:48:36 crc kubenswrapper[4971]: I1213 07:48:36.899751 4971 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/community-operators-6nsgx" Dec 13 07:48:36 crc kubenswrapper[4971]: I1213 07:48:36.900509 4971 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/community-operators-6nsgx" Dec 13 07:48:36 crc kubenswrapper[4971]: I1213 07:48:36.973751 4971 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/community-operators-6nsgx" Dec 13 07:48:37 crc kubenswrapper[4971]: I1213 07:48:37.155580 4971 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/community-operators-6nsgx" Dec 13 07:48:37 crc kubenswrapper[4971]: I1213 07:48:37.243403 4971 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-6nsgx"] Dec 13 07:48:39 crc kubenswrapper[4971]: I1213 07:48:39.117754 4971 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/community-operators-6nsgx" podUID="b807ae25-7d1d-4335-946a-45ffb0160d36" containerName="registry-server" containerID="cri-o://2d778bfafde3c5c8513578df5229341a4a3a53a549f8965d7f39597ed0a7622a" gracePeriod=2 Dec 13 07:48:39 crc kubenswrapper[4971]: I1213 07:48:39.586057 4971 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-6nsgx" Dec 13 07:48:39 crc kubenswrapper[4971]: I1213 07:48:39.717051 4971 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-m5dgm\" (UniqueName: \"kubernetes.io/projected/b807ae25-7d1d-4335-946a-45ffb0160d36-kube-api-access-m5dgm\") pod \"b807ae25-7d1d-4335-946a-45ffb0160d36\" (UID: \"b807ae25-7d1d-4335-946a-45ffb0160d36\") " Dec 13 07:48:39 crc kubenswrapper[4971]: I1213 07:48:39.717197 4971 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/b807ae25-7d1d-4335-946a-45ffb0160d36-catalog-content\") pod \"b807ae25-7d1d-4335-946a-45ffb0160d36\" (UID: \"b807ae25-7d1d-4335-946a-45ffb0160d36\") " Dec 13 07:48:39 crc kubenswrapper[4971]: I1213 07:48:39.717353 4971 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/b807ae25-7d1d-4335-946a-45ffb0160d36-utilities\") pod \"b807ae25-7d1d-4335-946a-45ffb0160d36\" (UID: \"b807ae25-7d1d-4335-946a-45ffb0160d36\") " Dec 13 07:48:39 crc kubenswrapper[4971]: I1213 07:48:39.718874 4971 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/b807ae25-7d1d-4335-946a-45ffb0160d36-utilities" (OuterVolumeSpecName: "utilities") pod "b807ae25-7d1d-4335-946a-45ffb0160d36" (UID: "b807ae25-7d1d-4335-946a-45ffb0160d36"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 13 07:48:39 crc kubenswrapper[4971]: I1213 07:48:39.730072 4971 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/b807ae25-7d1d-4335-946a-45ffb0160d36-kube-api-access-m5dgm" (OuterVolumeSpecName: "kube-api-access-m5dgm") pod "b807ae25-7d1d-4335-946a-45ffb0160d36" (UID: "b807ae25-7d1d-4335-946a-45ffb0160d36"). InnerVolumeSpecName "kube-api-access-m5dgm". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 13 07:48:39 crc kubenswrapper[4971]: I1213 07:48:39.779345 4971 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/b807ae25-7d1d-4335-946a-45ffb0160d36-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "b807ae25-7d1d-4335-946a-45ffb0160d36" (UID: "b807ae25-7d1d-4335-946a-45ffb0160d36"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 13 07:48:39 crc kubenswrapper[4971]: I1213 07:48:39.819897 4971 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/b807ae25-7d1d-4335-946a-45ffb0160d36-utilities\") on node \"crc\" DevicePath \"\"" Dec 13 07:48:39 crc kubenswrapper[4971]: I1213 07:48:39.819946 4971 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-m5dgm\" (UniqueName: \"kubernetes.io/projected/b807ae25-7d1d-4335-946a-45ffb0160d36-kube-api-access-m5dgm\") on node \"crc\" DevicePath \"\"" Dec 13 07:48:39 crc kubenswrapper[4971]: I1213 07:48:39.819963 4971 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/b807ae25-7d1d-4335-946a-45ffb0160d36-catalog-content\") on node \"crc\" DevicePath \"\"" Dec 13 07:48:40 crc kubenswrapper[4971]: I1213 07:48:40.128921 4971 generic.go:334] "Generic (PLEG): container finished" podID="b807ae25-7d1d-4335-946a-45ffb0160d36" containerID="2d778bfafde3c5c8513578df5229341a4a3a53a549f8965d7f39597ed0a7622a" exitCode=0 Dec 13 07:48:40 crc kubenswrapper[4971]: I1213 07:48:40.129017 4971 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-6nsgx" Dec 13 07:48:40 crc kubenswrapper[4971]: I1213 07:48:40.129022 4971 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-6nsgx" event={"ID":"b807ae25-7d1d-4335-946a-45ffb0160d36","Type":"ContainerDied","Data":"2d778bfafde3c5c8513578df5229341a4a3a53a549f8965d7f39597ed0a7622a"} Dec 13 07:48:40 crc kubenswrapper[4971]: I1213 07:48:40.129405 4971 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-6nsgx" event={"ID":"b807ae25-7d1d-4335-946a-45ffb0160d36","Type":"ContainerDied","Data":"747393bd09404d668ac9f8feb01e0d5f24f87f4b6793f12d9c7c33cad40538d5"} Dec 13 07:48:40 crc kubenswrapper[4971]: I1213 07:48:40.129429 4971 scope.go:117] "RemoveContainer" containerID="2d778bfafde3c5c8513578df5229341a4a3a53a549f8965d7f39597ed0a7622a" Dec 13 07:48:40 crc kubenswrapper[4971]: I1213 07:48:40.164581 4971 scope.go:117] "RemoveContainer" containerID="c1b860aae170e2efbd9b88faa1fe108392ee0fa7c75c88aba4be6d0bfb88f568" Dec 13 07:48:40 crc kubenswrapper[4971]: I1213 07:48:40.164947 4971 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-6nsgx"] Dec 13 07:48:40 crc kubenswrapper[4971]: I1213 07:48:40.174969 4971 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/community-operators-6nsgx"] Dec 13 07:48:40 crc kubenswrapper[4971]: I1213 07:48:40.185801 4971 scope.go:117] "RemoveContainer" containerID="b637e61b5b3c1aefd1d1b602e972b2219e1d724cee342090f605ea27b4722be3" Dec 13 07:48:40 crc kubenswrapper[4971]: I1213 07:48:40.236148 4971 scope.go:117] "RemoveContainer" containerID="2d778bfafde3c5c8513578df5229341a4a3a53a549f8965d7f39597ed0a7622a" Dec 13 07:48:40 crc kubenswrapper[4971]: E1213 07:48:40.236648 4971 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"2d778bfafde3c5c8513578df5229341a4a3a53a549f8965d7f39597ed0a7622a\": container with ID starting with 2d778bfafde3c5c8513578df5229341a4a3a53a549f8965d7f39597ed0a7622a not found: ID does not exist" containerID="2d778bfafde3c5c8513578df5229341a4a3a53a549f8965d7f39597ed0a7622a" Dec 13 07:48:40 crc kubenswrapper[4971]: I1213 07:48:40.236688 4971 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"2d778bfafde3c5c8513578df5229341a4a3a53a549f8965d7f39597ed0a7622a"} err="failed to get container status \"2d778bfafde3c5c8513578df5229341a4a3a53a549f8965d7f39597ed0a7622a\": rpc error: code = NotFound desc = could not find container \"2d778bfafde3c5c8513578df5229341a4a3a53a549f8965d7f39597ed0a7622a\": container with ID starting with 2d778bfafde3c5c8513578df5229341a4a3a53a549f8965d7f39597ed0a7622a not found: ID does not exist" Dec 13 07:48:40 crc kubenswrapper[4971]: I1213 07:48:40.236748 4971 scope.go:117] "RemoveContainer" containerID="c1b860aae170e2efbd9b88faa1fe108392ee0fa7c75c88aba4be6d0bfb88f568" Dec 13 07:48:40 crc kubenswrapper[4971]: E1213 07:48:40.237147 4971 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"c1b860aae170e2efbd9b88faa1fe108392ee0fa7c75c88aba4be6d0bfb88f568\": container with ID starting with c1b860aae170e2efbd9b88faa1fe108392ee0fa7c75c88aba4be6d0bfb88f568 not found: ID does not exist" containerID="c1b860aae170e2efbd9b88faa1fe108392ee0fa7c75c88aba4be6d0bfb88f568" Dec 13 07:48:40 crc kubenswrapper[4971]: I1213 07:48:40.237166 4971 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"c1b860aae170e2efbd9b88faa1fe108392ee0fa7c75c88aba4be6d0bfb88f568"} err="failed to get container status \"c1b860aae170e2efbd9b88faa1fe108392ee0fa7c75c88aba4be6d0bfb88f568\": rpc error: code = NotFound desc = could not find container \"c1b860aae170e2efbd9b88faa1fe108392ee0fa7c75c88aba4be6d0bfb88f568\": container with ID starting with c1b860aae170e2efbd9b88faa1fe108392ee0fa7c75c88aba4be6d0bfb88f568 not found: ID does not exist" Dec 13 07:48:40 crc kubenswrapper[4971]: I1213 07:48:40.237183 4971 scope.go:117] "RemoveContainer" containerID="b637e61b5b3c1aefd1d1b602e972b2219e1d724cee342090f605ea27b4722be3" Dec 13 07:48:40 crc kubenswrapper[4971]: E1213 07:48:40.237482 4971 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"b637e61b5b3c1aefd1d1b602e972b2219e1d724cee342090f605ea27b4722be3\": container with ID starting with b637e61b5b3c1aefd1d1b602e972b2219e1d724cee342090f605ea27b4722be3 not found: ID does not exist" containerID="b637e61b5b3c1aefd1d1b602e972b2219e1d724cee342090f605ea27b4722be3" Dec 13 07:48:40 crc kubenswrapper[4971]: I1213 07:48:40.237505 4971 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"b637e61b5b3c1aefd1d1b602e972b2219e1d724cee342090f605ea27b4722be3"} err="failed to get container status \"b637e61b5b3c1aefd1d1b602e972b2219e1d724cee342090f605ea27b4722be3\": rpc error: code = NotFound desc = could not find container \"b637e61b5b3c1aefd1d1b602e972b2219e1d724cee342090f605ea27b4722be3\": container with ID starting with b637e61b5b3c1aefd1d1b602e972b2219e1d724cee342090f605ea27b4722be3 not found: ID does not exist" Dec 13 07:48:41 crc kubenswrapper[4971]: I1213 07:48:41.778816 4971 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="b807ae25-7d1d-4335-946a-45ffb0160d36" path="/var/lib/kubelet/pods/b807ae25-7d1d-4335-946a-45ffb0160d36/volumes" Dec 13 07:48:46 crc kubenswrapper[4971]: I1213 07:48:46.153426 4971 patch_prober.go:28] interesting pod/machine-config-daemon-82xjz container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 13 07:48:46 crc kubenswrapper[4971]: I1213 07:48:46.154007 4971 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-82xjz" podUID="e0c4af12-edda-41d7-9b44-f87396174ef2" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 13 07:48:53 crc kubenswrapper[4971]: I1213 07:48:53.765709 4971 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-marketplace-nf2hb"] Dec 13 07:48:53 crc kubenswrapper[4971]: E1213 07:48:53.766574 4971 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="b807ae25-7d1d-4335-946a-45ffb0160d36" containerName="extract-utilities" Dec 13 07:48:53 crc kubenswrapper[4971]: I1213 07:48:53.766588 4971 state_mem.go:107] "Deleted CPUSet assignment" podUID="b807ae25-7d1d-4335-946a-45ffb0160d36" containerName="extract-utilities" Dec 13 07:48:53 crc kubenswrapper[4971]: E1213 07:48:53.766602 4971 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="b807ae25-7d1d-4335-946a-45ffb0160d36" containerName="registry-server" Dec 13 07:48:53 crc kubenswrapper[4971]: I1213 07:48:53.766608 4971 state_mem.go:107] "Deleted CPUSet assignment" podUID="b807ae25-7d1d-4335-946a-45ffb0160d36" containerName="registry-server" Dec 13 07:48:53 crc kubenswrapper[4971]: E1213 07:48:53.766623 4971 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="b807ae25-7d1d-4335-946a-45ffb0160d36" containerName="extract-content" Dec 13 07:48:53 crc kubenswrapper[4971]: I1213 07:48:53.766631 4971 state_mem.go:107] "Deleted CPUSet assignment" podUID="b807ae25-7d1d-4335-946a-45ffb0160d36" containerName="extract-content" Dec 13 07:48:53 crc kubenswrapper[4971]: I1213 07:48:53.766854 4971 memory_manager.go:354] "RemoveStaleState removing state" podUID="b807ae25-7d1d-4335-946a-45ffb0160d36" containerName="registry-server" Dec 13 07:48:53 crc kubenswrapper[4971]: I1213 07:48:53.768312 4971 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-nf2hb" Dec 13 07:48:53 crc kubenswrapper[4971]: I1213 07:48:53.788900 4971 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-nf2hb"] Dec 13 07:48:53 crc kubenswrapper[4971]: I1213 07:48:53.872341 4971 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-bb47k\" (UniqueName: \"kubernetes.io/projected/6de28224-6009-4541-97d8-3d01a36c7a01-kube-api-access-bb47k\") pod \"redhat-marketplace-nf2hb\" (UID: \"6de28224-6009-4541-97d8-3d01a36c7a01\") " pod="openshift-marketplace/redhat-marketplace-nf2hb" Dec 13 07:48:53 crc kubenswrapper[4971]: I1213 07:48:53.872478 4971 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/6de28224-6009-4541-97d8-3d01a36c7a01-utilities\") pod \"redhat-marketplace-nf2hb\" (UID: \"6de28224-6009-4541-97d8-3d01a36c7a01\") " pod="openshift-marketplace/redhat-marketplace-nf2hb" Dec 13 07:48:53 crc kubenswrapper[4971]: I1213 07:48:53.872549 4971 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/6de28224-6009-4541-97d8-3d01a36c7a01-catalog-content\") pod \"redhat-marketplace-nf2hb\" (UID: \"6de28224-6009-4541-97d8-3d01a36c7a01\") " pod="openshift-marketplace/redhat-marketplace-nf2hb" Dec 13 07:48:53 crc kubenswrapper[4971]: I1213 07:48:53.960461 4971 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-operators-wwm29"] Dec 13 07:48:53 crc kubenswrapper[4971]: I1213 07:48:53.962834 4971 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-wwm29" Dec 13 07:48:53 crc kubenswrapper[4971]: I1213 07:48:53.973892 4971 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/6de28224-6009-4541-97d8-3d01a36c7a01-utilities\") pod \"redhat-marketplace-nf2hb\" (UID: \"6de28224-6009-4541-97d8-3d01a36c7a01\") " pod="openshift-marketplace/redhat-marketplace-nf2hb" Dec 13 07:48:53 crc kubenswrapper[4971]: I1213 07:48:53.973958 4971 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/6de28224-6009-4541-97d8-3d01a36c7a01-catalog-content\") pod \"redhat-marketplace-nf2hb\" (UID: \"6de28224-6009-4541-97d8-3d01a36c7a01\") " pod="openshift-marketplace/redhat-marketplace-nf2hb" Dec 13 07:48:53 crc kubenswrapper[4971]: I1213 07:48:53.974069 4971 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-bb47k\" (UniqueName: \"kubernetes.io/projected/6de28224-6009-4541-97d8-3d01a36c7a01-kube-api-access-bb47k\") pod \"redhat-marketplace-nf2hb\" (UID: \"6de28224-6009-4541-97d8-3d01a36c7a01\") " pod="openshift-marketplace/redhat-marketplace-nf2hb" Dec 13 07:48:53 crc kubenswrapper[4971]: I1213 07:48:53.974503 4971 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/6de28224-6009-4541-97d8-3d01a36c7a01-utilities\") pod \"redhat-marketplace-nf2hb\" (UID: \"6de28224-6009-4541-97d8-3d01a36c7a01\") " pod="openshift-marketplace/redhat-marketplace-nf2hb" Dec 13 07:48:53 crc kubenswrapper[4971]: I1213 07:48:53.974637 4971 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/6de28224-6009-4541-97d8-3d01a36c7a01-catalog-content\") pod \"redhat-marketplace-nf2hb\" (UID: \"6de28224-6009-4541-97d8-3d01a36c7a01\") " pod="openshift-marketplace/redhat-marketplace-nf2hb" Dec 13 07:48:53 crc kubenswrapper[4971]: I1213 07:48:53.978936 4971 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-wwm29"] Dec 13 07:48:54 crc kubenswrapper[4971]: I1213 07:48:54.014886 4971 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-bb47k\" (UniqueName: \"kubernetes.io/projected/6de28224-6009-4541-97d8-3d01a36c7a01-kube-api-access-bb47k\") pod \"redhat-marketplace-nf2hb\" (UID: \"6de28224-6009-4541-97d8-3d01a36c7a01\") " pod="openshift-marketplace/redhat-marketplace-nf2hb" Dec 13 07:48:54 crc kubenswrapper[4971]: I1213 07:48:54.075939 4971 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/a45b0b6c-0728-4808-b2c1-f4194222e636-utilities\") pod \"redhat-operators-wwm29\" (UID: \"a45b0b6c-0728-4808-b2c1-f4194222e636\") " pod="openshift-marketplace/redhat-operators-wwm29" Dec 13 07:48:54 crc kubenswrapper[4971]: I1213 07:48:54.076802 4971 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-9ns5k\" (UniqueName: \"kubernetes.io/projected/a45b0b6c-0728-4808-b2c1-f4194222e636-kube-api-access-9ns5k\") pod \"redhat-operators-wwm29\" (UID: \"a45b0b6c-0728-4808-b2c1-f4194222e636\") " pod="openshift-marketplace/redhat-operators-wwm29" Dec 13 07:48:54 crc kubenswrapper[4971]: I1213 07:48:54.077043 4971 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/a45b0b6c-0728-4808-b2c1-f4194222e636-catalog-content\") pod \"redhat-operators-wwm29\" (UID: \"a45b0b6c-0728-4808-b2c1-f4194222e636\") " pod="openshift-marketplace/redhat-operators-wwm29" Dec 13 07:48:54 crc kubenswrapper[4971]: I1213 07:48:54.108980 4971 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-nf2hb" Dec 13 07:48:54 crc kubenswrapper[4971]: I1213 07:48:54.257937 4971 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/a45b0b6c-0728-4808-b2c1-f4194222e636-utilities\") pod \"redhat-operators-wwm29\" (UID: \"a45b0b6c-0728-4808-b2c1-f4194222e636\") " pod="openshift-marketplace/redhat-operators-wwm29" Dec 13 07:48:54 crc kubenswrapper[4971]: I1213 07:48:54.259766 4971 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/a45b0b6c-0728-4808-b2c1-f4194222e636-utilities\") pod \"redhat-operators-wwm29\" (UID: \"a45b0b6c-0728-4808-b2c1-f4194222e636\") " pod="openshift-marketplace/redhat-operators-wwm29" Dec 13 07:48:54 crc kubenswrapper[4971]: I1213 07:48:54.268120 4971 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-9ns5k\" (UniqueName: \"kubernetes.io/projected/a45b0b6c-0728-4808-b2c1-f4194222e636-kube-api-access-9ns5k\") pod \"redhat-operators-wwm29\" (UID: \"a45b0b6c-0728-4808-b2c1-f4194222e636\") " pod="openshift-marketplace/redhat-operators-wwm29" Dec 13 07:48:54 crc kubenswrapper[4971]: I1213 07:48:54.268291 4971 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/a45b0b6c-0728-4808-b2c1-f4194222e636-catalog-content\") pod \"redhat-operators-wwm29\" (UID: \"a45b0b6c-0728-4808-b2c1-f4194222e636\") " pod="openshift-marketplace/redhat-operators-wwm29" Dec 13 07:48:54 crc kubenswrapper[4971]: I1213 07:48:54.269147 4971 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/a45b0b6c-0728-4808-b2c1-f4194222e636-catalog-content\") pod \"redhat-operators-wwm29\" (UID: \"a45b0b6c-0728-4808-b2c1-f4194222e636\") " pod="openshift-marketplace/redhat-operators-wwm29" Dec 13 07:48:54 crc kubenswrapper[4971]: I1213 07:48:54.288708 4971 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-9ns5k\" (UniqueName: \"kubernetes.io/projected/a45b0b6c-0728-4808-b2c1-f4194222e636-kube-api-access-9ns5k\") pod \"redhat-operators-wwm29\" (UID: \"a45b0b6c-0728-4808-b2c1-f4194222e636\") " pod="openshift-marketplace/redhat-operators-wwm29" Dec 13 07:48:54 crc kubenswrapper[4971]: I1213 07:48:54.578620 4971 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-wwm29" Dec 13 07:48:54 crc kubenswrapper[4971]: I1213 07:48:54.783326 4971 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-nf2hb"] Dec 13 07:48:55 crc kubenswrapper[4971]: I1213 07:48:55.050562 4971 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-wwm29"] Dec 13 07:48:55 crc kubenswrapper[4971]: I1213 07:48:55.278453 4971 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-wwm29" event={"ID":"a45b0b6c-0728-4808-b2c1-f4194222e636","Type":"ContainerStarted","Data":"d8befbfe74416a928b159aa11868df889307368c95ea12f80ec45c5086b1e381"} Dec 13 07:48:55 crc kubenswrapper[4971]: I1213 07:48:55.279845 4971 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-nf2hb" event={"ID":"6de28224-6009-4541-97d8-3d01a36c7a01","Type":"ContainerStarted","Data":"dd3b8f3fd42af823ff407cf0595b093d476141a932f61b2d287c8fc987757c16"} Dec 13 07:48:56 crc kubenswrapper[4971]: I1213 07:48:56.292669 4971 generic.go:334] "Generic (PLEG): container finished" podID="a45b0b6c-0728-4808-b2c1-f4194222e636" containerID="52fe63bd9b6433d9895fc4a9b49f8988877b01994174d1c5754b5e0422be45af" exitCode=0 Dec 13 07:48:56 crc kubenswrapper[4971]: I1213 07:48:56.292790 4971 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-wwm29" event={"ID":"a45b0b6c-0728-4808-b2c1-f4194222e636","Type":"ContainerDied","Data":"52fe63bd9b6433d9895fc4a9b49f8988877b01994174d1c5754b5e0422be45af"} Dec 13 07:48:56 crc kubenswrapper[4971]: I1213 07:48:56.296392 4971 generic.go:334] "Generic (PLEG): container finished" podID="6de28224-6009-4541-97d8-3d01a36c7a01" containerID="21bc0c3e39d5f093358b67e5488614f59a788c163393139801a03710afb502c7" exitCode=0 Dec 13 07:48:56 crc kubenswrapper[4971]: I1213 07:48:56.296428 4971 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-nf2hb" event={"ID":"6de28224-6009-4541-97d8-3d01a36c7a01","Type":"ContainerDied","Data":"21bc0c3e39d5f093358b67e5488614f59a788c163393139801a03710afb502c7"} Dec 13 07:48:57 crc kubenswrapper[4971]: I1213 07:48:57.305727 4971 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-nf2hb" event={"ID":"6de28224-6009-4541-97d8-3d01a36c7a01","Type":"ContainerStarted","Data":"c0a90784d6fe9adc4f37e8468b69677d0a454a35c15418dd04d6f99393c35a10"} Dec 13 07:48:57 crc kubenswrapper[4971]: I1213 07:48:57.307914 4971 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-wwm29" event={"ID":"a45b0b6c-0728-4808-b2c1-f4194222e636","Type":"ContainerStarted","Data":"0f54aada2974ffff33374268d9cfcbe1163fa444673593da831b17d404399c70"} Dec 13 07:48:58 crc kubenswrapper[4971]: I1213 07:48:58.354728 4971 generic.go:334] "Generic (PLEG): container finished" podID="6de28224-6009-4541-97d8-3d01a36c7a01" containerID="c0a90784d6fe9adc4f37e8468b69677d0a454a35c15418dd04d6f99393c35a10" exitCode=0 Dec 13 07:48:58 crc kubenswrapper[4971]: I1213 07:48:58.354780 4971 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-nf2hb" event={"ID":"6de28224-6009-4541-97d8-3d01a36c7a01","Type":"ContainerDied","Data":"c0a90784d6fe9adc4f37e8468b69677d0a454a35c15418dd04d6f99393c35a10"} Dec 13 07:49:00 crc kubenswrapper[4971]: I1213 07:49:00.381051 4971 generic.go:334] "Generic (PLEG): container finished" podID="a45b0b6c-0728-4808-b2c1-f4194222e636" containerID="0f54aada2974ffff33374268d9cfcbe1163fa444673593da831b17d404399c70" exitCode=0 Dec 13 07:49:00 crc kubenswrapper[4971]: I1213 07:49:00.381127 4971 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-wwm29" event={"ID":"a45b0b6c-0728-4808-b2c1-f4194222e636","Type":"ContainerDied","Data":"0f54aada2974ffff33374268d9cfcbe1163fa444673593da831b17d404399c70"} Dec 13 07:49:03 crc kubenswrapper[4971]: I1213 07:49:03.420980 4971 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-nf2hb" event={"ID":"6de28224-6009-4541-97d8-3d01a36c7a01","Type":"ContainerStarted","Data":"c631d470cef727c5ef474cdf3f9891925e6e38064b9bb8356e2af2a79c3874fd"} Dec 13 07:49:03 crc kubenswrapper[4971]: I1213 07:49:03.424244 4971 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-wwm29" event={"ID":"a45b0b6c-0728-4808-b2c1-f4194222e636","Type":"ContainerStarted","Data":"1f2d6ed3d776e76e7c46f53a2f75ab4475f7eff02f50ed180ee228a659251d93"} Dec 13 07:49:03 crc kubenswrapper[4971]: I1213 07:49:03.439499 4971 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-marketplace-nf2hb" podStartSLOduration=4.314336357 podStartE2EDuration="10.439476736s" podCreationTimestamp="2025-12-13 07:48:53 +0000 UTC" firstStartedPulling="2025-12-13 07:48:56.298038152 +0000 UTC m=+3592.902447620" lastFinishedPulling="2025-12-13 07:49:02.423178531 +0000 UTC m=+3599.027587999" observedRunningTime="2025-12-13 07:49:03.435933101 +0000 UTC m=+3600.040342559" watchObservedRunningTime="2025-12-13 07:49:03.439476736 +0000 UTC m=+3600.043886194" Dec 13 07:49:03 crc kubenswrapper[4971]: I1213 07:49:03.464028 4971 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-operators-wwm29" podStartSLOduration=4.323161997 podStartE2EDuration="10.46400725s" podCreationTimestamp="2025-12-13 07:48:53 +0000 UTC" firstStartedPulling="2025-12-13 07:48:56.294740064 +0000 UTC m=+3592.899149512" lastFinishedPulling="2025-12-13 07:49:02.435585307 +0000 UTC m=+3599.039994765" observedRunningTime="2025-12-13 07:49:03.455585929 +0000 UTC m=+3600.059995387" watchObservedRunningTime="2025-12-13 07:49:03.46400725 +0000 UTC m=+3600.068416698" Dec 13 07:49:04 crc kubenswrapper[4971]: I1213 07:49:04.109901 4971 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-marketplace-nf2hb" Dec 13 07:49:04 crc kubenswrapper[4971]: I1213 07:49:04.109941 4971 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-marketplace-nf2hb" Dec 13 07:49:04 crc kubenswrapper[4971]: I1213 07:49:04.158486 4971 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-marketplace-nf2hb" Dec 13 07:49:04 crc kubenswrapper[4971]: I1213 07:49:04.578802 4971 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-operators-wwm29" Dec 13 07:49:04 crc kubenswrapper[4971]: I1213 07:49:04.578859 4971 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-operators-wwm29" Dec 13 07:49:05 crc kubenswrapper[4971]: I1213 07:49:05.620881 4971 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-marketplace/redhat-operators-wwm29" podUID="a45b0b6c-0728-4808-b2c1-f4194222e636" containerName="registry-server" probeResult="failure" output=< Dec 13 07:49:05 crc kubenswrapper[4971]: timeout: failed to connect service ":50051" within 1s Dec 13 07:49:05 crc kubenswrapper[4971]: > Dec 13 07:49:14 crc kubenswrapper[4971]: I1213 07:49:14.173316 4971 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-marketplace-nf2hb" Dec 13 07:49:14 crc kubenswrapper[4971]: I1213 07:49:14.221133 4971 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-nf2hb"] Dec 13 07:49:14 crc kubenswrapper[4971]: I1213 07:49:14.523383 4971 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-marketplace-nf2hb" podUID="6de28224-6009-4541-97d8-3d01a36c7a01" containerName="registry-server" containerID="cri-o://c631d470cef727c5ef474cdf3f9891925e6e38064b9bb8356e2af2a79c3874fd" gracePeriod=2 Dec 13 07:49:14 crc kubenswrapper[4971]: I1213 07:49:14.626632 4971 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-operators-wwm29" Dec 13 07:49:14 crc kubenswrapper[4971]: I1213 07:49:14.691657 4971 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-operators-wwm29" Dec 13 07:49:14 crc kubenswrapper[4971]: I1213 07:49:14.991613 4971 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-nf2hb" Dec 13 07:49:15 crc kubenswrapper[4971]: I1213 07:49:15.121351 4971 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/6de28224-6009-4541-97d8-3d01a36c7a01-utilities\") pod \"6de28224-6009-4541-97d8-3d01a36c7a01\" (UID: \"6de28224-6009-4541-97d8-3d01a36c7a01\") " Dec 13 07:49:15 crc kubenswrapper[4971]: I1213 07:49:15.121422 4971 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/6de28224-6009-4541-97d8-3d01a36c7a01-catalog-content\") pod \"6de28224-6009-4541-97d8-3d01a36c7a01\" (UID: \"6de28224-6009-4541-97d8-3d01a36c7a01\") " Dec 13 07:49:15 crc kubenswrapper[4971]: I1213 07:49:15.122323 4971 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/6de28224-6009-4541-97d8-3d01a36c7a01-utilities" (OuterVolumeSpecName: "utilities") pod "6de28224-6009-4541-97d8-3d01a36c7a01" (UID: "6de28224-6009-4541-97d8-3d01a36c7a01"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 13 07:49:15 crc kubenswrapper[4971]: I1213 07:49:15.122659 4971 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-bb47k\" (UniqueName: \"kubernetes.io/projected/6de28224-6009-4541-97d8-3d01a36c7a01-kube-api-access-bb47k\") pod \"6de28224-6009-4541-97d8-3d01a36c7a01\" (UID: \"6de28224-6009-4541-97d8-3d01a36c7a01\") " Dec 13 07:49:15 crc kubenswrapper[4971]: I1213 07:49:15.123442 4971 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/6de28224-6009-4541-97d8-3d01a36c7a01-utilities\") on node \"crc\" DevicePath \"\"" Dec 13 07:49:15 crc kubenswrapper[4971]: I1213 07:49:15.130259 4971 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/6de28224-6009-4541-97d8-3d01a36c7a01-kube-api-access-bb47k" (OuterVolumeSpecName: "kube-api-access-bb47k") pod "6de28224-6009-4541-97d8-3d01a36c7a01" (UID: "6de28224-6009-4541-97d8-3d01a36c7a01"). InnerVolumeSpecName "kube-api-access-bb47k". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 13 07:49:15 crc kubenswrapper[4971]: I1213 07:49:15.146020 4971 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/6de28224-6009-4541-97d8-3d01a36c7a01-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "6de28224-6009-4541-97d8-3d01a36c7a01" (UID: "6de28224-6009-4541-97d8-3d01a36c7a01"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 13 07:49:15 crc kubenswrapper[4971]: I1213 07:49:15.226064 4971 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/6de28224-6009-4541-97d8-3d01a36c7a01-catalog-content\") on node \"crc\" DevicePath \"\"" Dec 13 07:49:15 crc kubenswrapper[4971]: I1213 07:49:15.226101 4971 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-bb47k\" (UniqueName: \"kubernetes.io/projected/6de28224-6009-4541-97d8-3d01a36c7a01-kube-api-access-bb47k\") on node \"crc\" DevicePath \"\"" Dec 13 07:49:15 crc kubenswrapper[4971]: I1213 07:49:15.537050 4971 generic.go:334] "Generic (PLEG): container finished" podID="6de28224-6009-4541-97d8-3d01a36c7a01" containerID="c631d470cef727c5ef474cdf3f9891925e6e38064b9bb8356e2af2a79c3874fd" exitCode=0 Dec 13 07:49:15 crc kubenswrapper[4971]: I1213 07:49:15.537960 4971 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-nf2hb" event={"ID":"6de28224-6009-4541-97d8-3d01a36c7a01","Type":"ContainerDied","Data":"c631d470cef727c5ef474cdf3f9891925e6e38064b9bb8356e2af2a79c3874fd"} Dec 13 07:49:15 crc kubenswrapper[4971]: I1213 07:49:15.537999 4971 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-nf2hb" event={"ID":"6de28224-6009-4541-97d8-3d01a36c7a01","Type":"ContainerDied","Data":"dd3b8f3fd42af823ff407cf0595b093d476141a932f61b2d287c8fc987757c16"} Dec 13 07:49:15 crc kubenswrapper[4971]: I1213 07:49:15.538021 4971 scope.go:117] "RemoveContainer" containerID="c631d470cef727c5ef474cdf3f9891925e6e38064b9bb8356e2af2a79c3874fd" Dec 13 07:49:15 crc kubenswrapper[4971]: I1213 07:49:15.538178 4971 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-nf2hb" Dec 13 07:49:15 crc kubenswrapper[4971]: I1213 07:49:15.576907 4971 scope.go:117] "RemoveContainer" containerID="c0a90784d6fe9adc4f37e8468b69677d0a454a35c15418dd04d6f99393c35a10" Dec 13 07:49:15 crc kubenswrapper[4971]: I1213 07:49:15.603963 4971 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-nf2hb"] Dec 13 07:49:15 crc kubenswrapper[4971]: I1213 07:49:15.615116 4971 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-marketplace-nf2hb"] Dec 13 07:49:15 crc kubenswrapper[4971]: I1213 07:49:15.627050 4971 scope.go:117] "RemoveContainer" containerID="21bc0c3e39d5f093358b67e5488614f59a788c163393139801a03710afb502c7" Dec 13 07:49:15 crc kubenswrapper[4971]: I1213 07:49:15.683276 4971 scope.go:117] "RemoveContainer" containerID="c631d470cef727c5ef474cdf3f9891925e6e38064b9bb8356e2af2a79c3874fd" Dec 13 07:49:15 crc kubenswrapper[4971]: E1213 07:49:15.683796 4971 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"c631d470cef727c5ef474cdf3f9891925e6e38064b9bb8356e2af2a79c3874fd\": container with ID starting with c631d470cef727c5ef474cdf3f9891925e6e38064b9bb8356e2af2a79c3874fd not found: ID does not exist" containerID="c631d470cef727c5ef474cdf3f9891925e6e38064b9bb8356e2af2a79c3874fd" Dec 13 07:49:15 crc kubenswrapper[4971]: I1213 07:49:15.683830 4971 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"c631d470cef727c5ef474cdf3f9891925e6e38064b9bb8356e2af2a79c3874fd"} err="failed to get container status \"c631d470cef727c5ef474cdf3f9891925e6e38064b9bb8356e2af2a79c3874fd\": rpc error: code = NotFound desc = could not find container \"c631d470cef727c5ef474cdf3f9891925e6e38064b9bb8356e2af2a79c3874fd\": container with ID starting with c631d470cef727c5ef474cdf3f9891925e6e38064b9bb8356e2af2a79c3874fd not found: ID does not exist" Dec 13 07:49:15 crc kubenswrapper[4971]: I1213 07:49:15.683866 4971 scope.go:117] "RemoveContainer" containerID="c0a90784d6fe9adc4f37e8468b69677d0a454a35c15418dd04d6f99393c35a10" Dec 13 07:49:15 crc kubenswrapper[4971]: E1213 07:49:15.684130 4971 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"c0a90784d6fe9adc4f37e8468b69677d0a454a35c15418dd04d6f99393c35a10\": container with ID starting with c0a90784d6fe9adc4f37e8468b69677d0a454a35c15418dd04d6f99393c35a10 not found: ID does not exist" containerID="c0a90784d6fe9adc4f37e8468b69677d0a454a35c15418dd04d6f99393c35a10" Dec 13 07:49:15 crc kubenswrapper[4971]: I1213 07:49:15.684169 4971 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"c0a90784d6fe9adc4f37e8468b69677d0a454a35c15418dd04d6f99393c35a10"} err="failed to get container status \"c0a90784d6fe9adc4f37e8468b69677d0a454a35c15418dd04d6f99393c35a10\": rpc error: code = NotFound desc = could not find container \"c0a90784d6fe9adc4f37e8468b69677d0a454a35c15418dd04d6f99393c35a10\": container with ID starting with c0a90784d6fe9adc4f37e8468b69677d0a454a35c15418dd04d6f99393c35a10 not found: ID does not exist" Dec 13 07:49:15 crc kubenswrapper[4971]: I1213 07:49:15.684184 4971 scope.go:117] "RemoveContainer" containerID="21bc0c3e39d5f093358b67e5488614f59a788c163393139801a03710afb502c7" Dec 13 07:49:15 crc kubenswrapper[4971]: E1213 07:49:15.684607 4971 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"21bc0c3e39d5f093358b67e5488614f59a788c163393139801a03710afb502c7\": container with ID starting with 21bc0c3e39d5f093358b67e5488614f59a788c163393139801a03710afb502c7 not found: ID does not exist" containerID="21bc0c3e39d5f093358b67e5488614f59a788c163393139801a03710afb502c7" Dec 13 07:49:15 crc kubenswrapper[4971]: I1213 07:49:15.684682 4971 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"21bc0c3e39d5f093358b67e5488614f59a788c163393139801a03710afb502c7"} err="failed to get container status \"21bc0c3e39d5f093358b67e5488614f59a788c163393139801a03710afb502c7\": rpc error: code = NotFound desc = could not find container \"21bc0c3e39d5f093358b67e5488614f59a788c163393139801a03710afb502c7\": container with ID starting with 21bc0c3e39d5f093358b67e5488614f59a788c163393139801a03710afb502c7 not found: ID does not exist" Dec 13 07:49:15 crc kubenswrapper[4971]: I1213 07:49:15.780216 4971 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="6de28224-6009-4541-97d8-3d01a36c7a01" path="/var/lib/kubelet/pods/6de28224-6009-4541-97d8-3d01a36c7a01/volumes" Dec 13 07:49:16 crc kubenswrapper[4971]: I1213 07:49:16.154154 4971 patch_prober.go:28] interesting pod/machine-config-daemon-82xjz container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 13 07:49:16 crc kubenswrapper[4971]: I1213 07:49:16.154612 4971 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-82xjz" podUID="e0c4af12-edda-41d7-9b44-f87396174ef2" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 13 07:49:16 crc kubenswrapper[4971]: I1213 07:49:16.405904 4971 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-wwm29"] Dec 13 07:49:16 crc kubenswrapper[4971]: I1213 07:49:16.550432 4971 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-operators-wwm29" podUID="a45b0b6c-0728-4808-b2c1-f4194222e636" containerName="registry-server" containerID="cri-o://1f2d6ed3d776e76e7c46f53a2f75ab4475f7eff02f50ed180ee228a659251d93" gracePeriod=2 Dec 13 07:49:16 crc kubenswrapper[4971]: I1213 07:49:16.983258 4971 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-wwm29" Dec 13 07:49:17 crc kubenswrapper[4971]: I1213 07:49:17.064991 4971 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/a45b0b6c-0728-4808-b2c1-f4194222e636-catalog-content\") pod \"a45b0b6c-0728-4808-b2c1-f4194222e636\" (UID: \"a45b0b6c-0728-4808-b2c1-f4194222e636\") " Dec 13 07:49:17 crc kubenswrapper[4971]: I1213 07:49:17.065276 4971 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-9ns5k\" (UniqueName: \"kubernetes.io/projected/a45b0b6c-0728-4808-b2c1-f4194222e636-kube-api-access-9ns5k\") pod \"a45b0b6c-0728-4808-b2c1-f4194222e636\" (UID: \"a45b0b6c-0728-4808-b2c1-f4194222e636\") " Dec 13 07:49:17 crc kubenswrapper[4971]: I1213 07:49:17.065443 4971 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/a45b0b6c-0728-4808-b2c1-f4194222e636-utilities\") pod \"a45b0b6c-0728-4808-b2c1-f4194222e636\" (UID: \"a45b0b6c-0728-4808-b2c1-f4194222e636\") " Dec 13 07:49:17 crc kubenswrapper[4971]: I1213 07:49:17.066279 4971 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/a45b0b6c-0728-4808-b2c1-f4194222e636-utilities" (OuterVolumeSpecName: "utilities") pod "a45b0b6c-0728-4808-b2c1-f4194222e636" (UID: "a45b0b6c-0728-4808-b2c1-f4194222e636"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 13 07:49:17 crc kubenswrapper[4971]: I1213 07:49:17.071283 4971 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/a45b0b6c-0728-4808-b2c1-f4194222e636-kube-api-access-9ns5k" (OuterVolumeSpecName: "kube-api-access-9ns5k") pod "a45b0b6c-0728-4808-b2c1-f4194222e636" (UID: "a45b0b6c-0728-4808-b2c1-f4194222e636"). InnerVolumeSpecName "kube-api-access-9ns5k". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 13 07:49:17 crc kubenswrapper[4971]: I1213 07:49:17.168394 4971 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-9ns5k\" (UniqueName: \"kubernetes.io/projected/a45b0b6c-0728-4808-b2c1-f4194222e636-kube-api-access-9ns5k\") on node \"crc\" DevicePath \"\"" Dec 13 07:49:17 crc kubenswrapper[4971]: I1213 07:49:17.168707 4971 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/a45b0b6c-0728-4808-b2c1-f4194222e636-utilities\") on node \"crc\" DevicePath \"\"" Dec 13 07:49:17 crc kubenswrapper[4971]: I1213 07:49:17.182589 4971 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/a45b0b6c-0728-4808-b2c1-f4194222e636-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "a45b0b6c-0728-4808-b2c1-f4194222e636" (UID: "a45b0b6c-0728-4808-b2c1-f4194222e636"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 13 07:49:17 crc kubenswrapper[4971]: I1213 07:49:17.269904 4971 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/a45b0b6c-0728-4808-b2c1-f4194222e636-catalog-content\") on node \"crc\" DevicePath \"\"" Dec 13 07:49:17 crc kubenswrapper[4971]: I1213 07:49:17.562589 4971 generic.go:334] "Generic (PLEG): container finished" podID="a45b0b6c-0728-4808-b2c1-f4194222e636" containerID="1f2d6ed3d776e76e7c46f53a2f75ab4475f7eff02f50ed180ee228a659251d93" exitCode=0 Dec 13 07:49:17 crc kubenswrapper[4971]: I1213 07:49:17.562630 4971 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-wwm29" event={"ID":"a45b0b6c-0728-4808-b2c1-f4194222e636","Type":"ContainerDied","Data":"1f2d6ed3d776e76e7c46f53a2f75ab4475f7eff02f50ed180ee228a659251d93"} Dec 13 07:49:17 crc kubenswrapper[4971]: I1213 07:49:17.562658 4971 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-wwm29" event={"ID":"a45b0b6c-0728-4808-b2c1-f4194222e636","Type":"ContainerDied","Data":"d8befbfe74416a928b159aa11868df889307368c95ea12f80ec45c5086b1e381"} Dec 13 07:49:17 crc kubenswrapper[4971]: I1213 07:49:17.562678 4971 scope.go:117] "RemoveContainer" containerID="1f2d6ed3d776e76e7c46f53a2f75ab4475f7eff02f50ed180ee228a659251d93" Dec 13 07:49:17 crc kubenswrapper[4971]: I1213 07:49:17.562749 4971 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-wwm29" Dec 13 07:49:17 crc kubenswrapper[4971]: I1213 07:49:17.600272 4971 scope.go:117] "RemoveContainer" containerID="0f54aada2974ffff33374268d9cfcbe1163fa444673593da831b17d404399c70" Dec 13 07:49:17 crc kubenswrapper[4971]: I1213 07:49:17.621441 4971 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-wwm29"] Dec 13 07:49:17 crc kubenswrapper[4971]: I1213 07:49:17.628612 4971 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-operators-wwm29"] Dec 13 07:49:17 crc kubenswrapper[4971]: I1213 07:49:17.647874 4971 scope.go:117] "RemoveContainer" containerID="52fe63bd9b6433d9895fc4a9b49f8988877b01994174d1c5754b5e0422be45af" Dec 13 07:49:17 crc kubenswrapper[4971]: I1213 07:49:17.691181 4971 scope.go:117] "RemoveContainer" containerID="1f2d6ed3d776e76e7c46f53a2f75ab4475f7eff02f50ed180ee228a659251d93" Dec 13 07:49:17 crc kubenswrapper[4971]: E1213 07:49:17.691855 4971 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"1f2d6ed3d776e76e7c46f53a2f75ab4475f7eff02f50ed180ee228a659251d93\": container with ID starting with 1f2d6ed3d776e76e7c46f53a2f75ab4475f7eff02f50ed180ee228a659251d93 not found: ID does not exist" containerID="1f2d6ed3d776e76e7c46f53a2f75ab4475f7eff02f50ed180ee228a659251d93" Dec 13 07:49:17 crc kubenswrapper[4971]: I1213 07:49:17.691914 4971 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"1f2d6ed3d776e76e7c46f53a2f75ab4475f7eff02f50ed180ee228a659251d93"} err="failed to get container status \"1f2d6ed3d776e76e7c46f53a2f75ab4475f7eff02f50ed180ee228a659251d93\": rpc error: code = NotFound desc = could not find container \"1f2d6ed3d776e76e7c46f53a2f75ab4475f7eff02f50ed180ee228a659251d93\": container with ID starting with 1f2d6ed3d776e76e7c46f53a2f75ab4475f7eff02f50ed180ee228a659251d93 not found: ID does not exist" Dec 13 07:49:17 crc kubenswrapper[4971]: I1213 07:49:17.691979 4971 scope.go:117] "RemoveContainer" containerID="0f54aada2974ffff33374268d9cfcbe1163fa444673593da831b17d404399c70" Dec 13 07:49:17 crc kubenswrapper[4971]: E1213 07:49:17.692425 4971 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"0f54aada2974ffff33374268d9cfcbe1163fa444673593da831b17d404399c70\": container with ID starting with 0f54aada2974ffff33374268d9cfcbe1163fa444673593da831b17d404399c70 not found: ID does not exist" containerID="0f54aada2974ffff33374268d9cfcbe1163fa444673593da831b17d404399c70" Dec 13 07:49:17 crc kubenswrapper[4971]: I1213 07:49:17.692478 4971 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"0f54aada2974ffff33374268d9cfcbe1163fa444673593da831b17d404399c70"} err="failed to get container status \"0f54aada2974ffff33374268d9cfcbe1163fa444673593da831b17d404399c70\": rpc error: code = NotFound desc = could not find container \"0f54aada2974ffff33374268d9cfcbe1163fa444673593da831b17d404399c70\": container with ID starting with 0f54aada2974ffff33374268d9cfcbe1163fa444673593da831b17d404399c70 not found: ID does not exist" Dec 13 07:49:17 crc kubenswrapper[4971]: I1213 07:49:17.692543 4971 scope.go:117] "RemoveContainer" containerID="52fe63bd9b6433d9895fc4a9b49f8988877b01994174d1c5754b5e0422be45af" Dec 13 07:49:17 crc kubenswrapper[4971]: E1213 07:49:17.693057 4971 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"52fe63bd9b6433d9895fc4a9b49f8988877b01994174d1c5754b5e0422be45af\": container with ID starting with 52fe63bd9b6433d9895fc4a9b49f8988877b01994174d1c5754b5e0422be45af not found: ID does not exist" containerID="52fe63bd9b6433d9895fc4a9b49f8988877b01994174d1c5754b5e0422be45af" Dec 13 07:49:17 crc kubenswrapper[4971]: I1213 07:49:17.693085 4971 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"52fe63bd9b6433d9895fc4a9b49f8988877b01994174d1c5754b5e0422be45af"} err="failed to get container status \"52fe63bd9b6433d9895fc4a9b49f8988877b01994174d1c5754b5e0422be45af\": rpc error: code = NotFound desc = could not find container \"52fe63bd9b6433d9895fc4a9b49f8988877b01994174d1c5754b5e0422be45af\": container with ID starting with 52fe63bd9b6433d9895fc4a9b49f8988877b01994174d1c5754b5e0422be45af not found: ID does not exist" Dec 13 07:49:17 crc kubenswrapper[4971]: I1213 07:49:17.787427 4971 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="a45b0b6c-0728-4808-b2c1-f4194222e636" path="/var/lib/kubelet/pods/a45b0b6c-0728-4808-b2c1-f4194222e636/volumes" Dec 13 07:49:46 crc kubenswrapper[4971]: I1213 07:49:46.153609 4971 patch_prober.go:28] interesting pod/machine-config-daemon-82xjz container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 13 07:49:46 crc kubenswrapper[4971]: I1213 07:49:46.154036 4971 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-82xjz" podUID="e0c4af12-edda-41d7-9b44-f87396174ef2" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 13 07:49:46 crc kubenswrapper[4971]: I1213 07:49:46.154072 4971 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-82xjz" Dec 13 07:49:46 crc kubenswrapper[4971]: I1213 07:49:46.154731 4971 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"e975a5822755c28b3c504aa81fe38e85a7a58425e2a7f942f681aff10344af88"} pod="openshift-machine-config-operator/machine-config-daemon-82xjz" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Dec 13 07:49:46 crc kubenswrapper[4971]: I1213 07:49:46.154772 4971 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-82xjz" podUID="e0c4af12-edda-41d7-9b44-f87396174ef2" containerName="machine-config-daemon" containerID="cri-o://e975a5822755c28b3c504aa81fe38e85a7a58425e2a7f942f681aff10344af88" gracePeriod=600 Dec 13 07:49:46 crc kubenswrapper[4971]: E1213 07:49:46.290482 4971 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-82xjz_openshift-machine-config-operator(e0c4af12-edda-41d7-9b44-f87396174ef2)\"" pod="openshift-machine-config-operator/machine-config-daemon-82xjz" podUID="e0c4af12-edda-41d7-9b44-f87396174ef2" Dec 13 07:49:46 crc kubenswrapper[4971]: I1213 07:49:46.855241 4971 generic.go:334] "Generic (PLEG): container finished" podID="e0c4af12-edda-41d7-9b44-f87396174ef2" containerID="e975a5822755c28b3c504aa81fe38e85a7a58425e2a7f942f681aff10344af88" exitCode=0 Dec 13 07:49:46 crc kubenswrapper[4971]: I1213 07:49:46.855284 4971 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-82xjz" event={"ID":"e0c4af12-edda-41d7-9b44-f87396174ef2","Type":"ContainerDied","Data":"e975a5822755c28b3c504aa81fe38e85a7a58425e2a7f942f681aff10344af88"} Dec 13 07:49:46 crc kubenswrapper[4971]: I1213 07:49:46.855326 4971 scope.go:117] "RemoveContainer" containerID="95874f24e419c2abe48edb2569aa8a50c62f04d8d5d3a5215adeee275f78b4d8" Dec 13 07:49:46 crc kubenswrapper[4971]: I1213 07:49:46.856099 4971 scope.go:117] "RemoveContainer" containerID="e975a5822755c28b3c504aa81fe38e85a7a58425e2a7f942f681aff10344af88" Dec 13 07:49:46 crc kubenswrapper[4971]: E1213 07:49:46.856449 4971 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-82xjz_openshift-machine-config-operator(e0c4af12-edda-41d7-9b44-f87396174ef2)\"" pod="openshift-machine-config-operator/machine-config-daemon-82xjz" podUID="e0c4af12-edda-41d7-9b44-f87396174ef2" Dec 13 07:49:59 crc kubenswrapper[4971]: I1213 07:49:59.769363 4971 scope.go:117] "RemoveContainer" containerID="e975a5822755c28b3c504aa81fe38e85a7a58425e2a7f942f681aff10344af88" Dec 13 07:49:59 crc kubenswrapper[4971]: E1213 07:49:59.770395 4971 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-82xjz_openshift-machine-config-operator(e0c4af12-edda-41d7-9b44-f87396174ef2)\"" pod="openshift-machine-config-operator/machine-config-daemon-82xjz" podUID="e0c4af12-edda-41d7-9b44-f87396174ef2" Dec 13 07:50:12 crc kubenswrapper[4971]: I1213 07:50:12.768983 4971 scope.go:117] "RemoveContainer" containerID="e975a5822755c28b3c504aa81fe38e85a7a58425e2a7f942f681aff10344af88" Dec 13 07:50:12 crc kubenswrapper[4971]: E1213 07:50:12.769810 4971 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-82xjz_openshift-machine-config-operator(e0c4af12-edda-41d7-9b44-f87396174ef2)\"" pod="openshift-machine-config-operator/machine-config-daemon-82xjz" podUID="e0c4af12-edda-41d7-9b44-f87396174ef2" Dec 13 07:50:27 crc kubenswrapper[4971]: I1213 07:50:27.768380 4971 scope.go:117] "RemoveContainer" containerID="e975a5822755c28b3c504aa81fe38e85a7a58425e2a7f942f681aff10344af88" Dec 13 07:50:27 crc kubenswrapper[4971]: E1213 07:50:27.769334 4971 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-82xjz_openshift-machine-config-operator(e0c4af12-edda-41d7-9b44-f87396174ef2)\"" pod="openshift-machine-config-operator/machine-config-daemon-82xjz" podUID="e0c4af12-edda-41d7-9b44-f87396174ef2" Dec 13 07:50:40 crc kubenswrapper[4971]: I1213 07:50:40.769413 4971 scope.go:117] "RemoveContainer" containerID="e975a5822755c28b3c504aa81fe38e85a7a58425e2a7f942f681aff10344af88" Dec 13 07:50:40 crc kubenswrapper[4971]: E1213 07:50:40.770421 4971 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-82xjz_openshift-machine-config-operator(e0c4af12-edda-41d7-9b44-f87396174ef2)\"" pod="openshift-machine-config-operator/machine-config-daemon-82xjz" podUID="e0c4af12-edda-41d7-9b44-f87396174ef2" Dec 13 07:50:51 crc kubenswrapper[4971]: I1213 07:50:51.776549 4971 scope.go:117] "RemoveContainer" containerID="e975a5822755c28b3c504aa81fe38e85a7a58425e2a7f942f681aff10344af88" Dec 13 07:50:51 crc kubenswrapper[4971]: E1213 07:50:51.777325 4971 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-82xjz_openshift-machine-config-operator(e0c4af12-edda-41d7-9b44-f87396174ef2)\"" pod="openshift-machine-config-operator/machine-config-daemon-82xjz" podUID="e0c4af12-edda-41d7-9b44-f87396174ef2" Dec 13 07:51:05 crc kubenswrapper[4971]: I1213 07:51:05.768874 4971 scope.go:117] "RemoveContainer" containerID="e975a5822755c28b3c504aa81fe38e85a7a58425e2a7f942f681aff10344af88" Dec 13 07:51:05 crc kubenswrapper[4971]: E1213 07:51:05.769733 4971 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-82xjz_openshift-machine-config-operator(e0c4af12-edda-41d7-9b44-f87396174ef2)\"" pod="openshift-machine-config-operator/machine-config-daemon-82xjz" podUID="e0c4af12-edda-41d7-9b44-f87396174ef2" Dec 13 07:51:19 crc kubenswrapper[4971]: I1213 07:51:19.768742 4971 scope.go:117] "RemoveContainer" containerID="e975a5822755c28b3c504aa81fe38e85a7a58425e2a7f942f681aff10344af88" Dec 13 07:51:19 crc kubenswrapper[4971]: E1213 07:51:19.769488 4971 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-82xjz_openshift-machine-config-operator(e0c4af12-edda-41d7-9b44-f87396174ef2)\"" pod="openshift-machine-config-operator/machine-config-daemon-82xjz" podUID="e0c4af12-edda-41d7-9b44-f87396174ef2" Dec 13 07:51:30 crc kubenswrapper[4971]: I1213 07:51:30.768593 4971 scope.go:117] "RemoveContainer" containerID="e975a5822755c28b3c504aa81fe38e85a7a58425e2a7f942f681aff10344af88" Dec 13 07:51:30 crc kubenswrapper[4971]: E1213 07:51:30.769341 4971 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-82xjz_openshift-machine-config-operator(e0c4af12-edda-41d7-9b44-f87396174ef2)\"" pod="openshift-machine-config-operator/machine-config-daemon-82xjz" podUID="e0c4af12-edda-41d7-9b44-f87396174ef2" Dec 13 07:51:42 crc kubenswrapper[4971]: I1213 07:51:42.768332 4971 scope.go:117] "RemoveContainer" containerID="e975a5822755c28b3c504aa81fe38e85a7a58425e2a7f942f681aff10344af88" Dec 13 07:51:42 crc kubenswrapper[4971]: E1213 07:51:42.769380 4971 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-82xjz_openshift-machine-config-operator(e0c4af12-edda-41d7-9b44-f87396174ef2)\"" pod="openshift-machine-config-operator/machine-config-daemon-82xjz" podUID="e0c4af12-edda-41d7-9b44-f87396174ef2" Dec 13 07:51:57 crc kubenswrapper[4971]: I1213 07:51:57.768921 4971 scope.go:117] "RemoveContainer" containerID="e975a5822755c28b3c504aa81fe38e85a7a58425e2a7f942f681aff10344af88" Dec 13 07:51:57 crc kubenswrapper[4971]: E1213 07:51:57.769751 4971 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-82xjz_openshift-machine-config-operator(e0c4af12-edda-41d7-9b44-f87396174ef2)\"" pod="openshift-machine-config-operator/machine-config-daemon-82xjz" podUID="e0c4af12-edda-41d7-9b44-f87396174ef2" Dec 13 07:52:09 crc kubenswrapper[4971]: I1213 07:52:09.768973 4971 scope.go:117] "RemoveContainer" containerID="e975a5822755c28b3c504aa81fe38e85a7a58425e2a7f942f681aff10344af88" Dec 13 07:52:09 crc kubenswrapper[4971]: E1213 07:52:09.769868 4971 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-82xjz_openshift-machine-config-operator(e0c4af12-edda-41d7-9b44-f87396174ef2)\"" pod="openshift-machine-config-operator/machine-config-daemon-82xjz" podUID="e0c4af12-edda-41d7-9b44-f87396174ef2" Dec 13 07:52:20 crc kubenswrapper[4971]: I1213 07:52:20.768792 4971 scope.go:117] "RemoveContainer" containerID="e975a5822755c28b3c504aa81fe38e85a7a58425e2a7f942f681aff10344af88" Dec 13 07:52:20 crc kubenswrapper[4971]: E1213 07:52:20.770630 4971 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-82xjz_openshift-machine-config-operator(e0c4af12-edda-41d7-9b44-f87396174ef2)\"" pod="openshift-machine-config-operator/machine-config-daemon-82xjz" podUID="e0c4af12-edda-41d7-9b44-f87396174ef2" Dec 13 07:52:33 crc kubenswrapper[4971]: I1213 07:52:33.775006 4971 scope.go:117] "RemoveContainer" containerID="e975a5822755c28b3c504aa81fe38e85a7a58425e2a7f942f681aff10344af88" Dec 13 07:52:33 crc kubenswrapper[4971]: E1213 07:52:33.775797 4971 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-82xjz_openshift-machine-config-operator(e0c4af12-edda-41d7-9b44-f87396174ef2)\"" pod="openshift-machine-config-operator/machine-config-daemon-82xjz" podUID="e0c4af12-edda-41d7-9b44-f87396174ef2" Dec 13 07:52:47 crc kubenswrapper[4971]: I1213 07:52:47.769298 4971 scope.go:117] "RemoveContainer" containerID="e975a5822755c28b3c504aa81fe38e85a7a58425e2a7f942f681aff10344af88" Dec 13 07:52:47 crc kubenswrapper[4971]: E1213 07:52:47.770100 4971 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-82xjz_openshift-machine-config-operator(e0c4af12-edda-41d7-9b44-f87396174ef2)\"" pod="openshift-machine-config-operator/machine-config-daemon-82xjz" podUID="e0c4af12-edda-41d7-9b44-f87396174ef2" Dec 13 07:52:58 crc kubenswrapper[4971]: I1213 07:52:58.770374 4971 scope.go:117] "RemoveContainer" containerID="e975a5822755c28b3c504aa81fe38e85a7a58425e2a7f942f681aff10344af88" Dec 13 07:52:58 crc kubenswrapper[4971]: E1213 07:52:58.771804 4971 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-82xjz_openshift-machine-config-operator(e0c4af12-edda-41d7-9b44-f87396174ef2)\"" pod="openshift-machine-config-operator/machine-config-daemon-82xjz" podUID="e0c4af12-edda-41d7-9b44-f87396174ef2" Dec 13 07:53:13 crc kubenswrapper[4971]: I1213 07:53:13.776812 4971 scope.go:117] "RemoveContainer" containerID="e975a5822755c28b3c504aa81fe38e85a7a58425e2a7f942f681aff10344af88" Dec 13 07:53:13 crc kubenswrapper[4971]: E1213 07:53:13.777730 4971 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-82xjz_openshift-machine-config-operator(e0c4af12-edda-41d7-9b44-f87396174ef2)\"" pod="openshift-machine-config-operator/machine-config-daemon-82xjz" podUID="e0c4af12-edda-41d7-9b44-f87396174ef2" Dec 13 07:53:27 crc kubenswrapper[4971]: I1213 07:53:27.768918 4971 scope.go:117] "RemoveContainer" containerID="e975a5822755c28b3c504aa81fe38e85a7a58425e2a7f942f681aff10344af88" Dec 13 07:53:27 crc kubenswrapper[4971]: E1213 07:53:27.769582 4971 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-82xjz_openshift-machine-config-operator(e0c4af12-edda-41d7-9b44-f87396174ef2)\"" pod="openshift-machine-config-operator/machine-config-daemon-82xjz" podUID="e0c4af12-edda-41d7-9b44-f87396174ef2" Dec 13 07:53:38 crc kubenswrapper[4971]: I1213 07:53:38.768804 4971 scope.go:117] "RemoveContainer" containerID="e975a5822755c28b3c504aa81fe38e85a7a58425e2a7f942f681aff10344af88" Dec 13 07:53:38 crc kubenswrapper[4971]: E1213 07:53:38.769652 4971 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-82xjz_openshift-machine-config-operator(e0c4af12-edda-41d7-9b44-f87396174ef2)\"" pod="openshift-machine-config-operator/machine-config-daemon-82xjz" podUID="e0c4af12-edda-41d7-9b44-f87396174ef2" Dec 13 07:53:51 crc kubenswrapper[4971]: I1213 07:53:51.769193 4971 scope.go:117] "RemoveContainer" containerID="e975a5822755c28b3c504aa81fe38e85a7a58425e2a7f942f681aff10344af88" Dec 13 07:53:51 crc kubenswrapper[4971]: E1213 07:53:51.770112 4971 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-82xjz_openshift-machine-config-operator(e0c4af12-edda-41d7-9b44-f87396174ef2)\"" pod="openshift-machine-config-operator/machine-config-daemon-82xjz" podUID="e0c4af12-edda-41d7-9b44-f87396174ef2" Dec 13 07:54:03 crc kubenswrapper[4971]: I1213 07:54:03.776282 4971 scope.go:117] "RemoveContainer" containerID="e975a5822755c28b3c504aa81fe38e85a7a58425e2a7f942f681aff10344af88" Dec 13 07:54:03 crc kubenswrapper[4971]: E1213 07:54:03.777296 4971 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-82xjz_openshift-machine-config-operator(e0c4af12-edda-41d7-9b44-f87396174ef2)\"" pod="openshift-machine-config-operator/machine-config-daemon-82xjz" podUID="e0c4af12-edda-41d7-9b44-f87396174ef2" Dec 13 07:54:16 crc kubenswrapper[4971]: I1213 07:54:16.771869 4971 scope.go:117] "RemoveContainer" containerID="e975a5822755c28b3c504aa81fe38e85a7a58425e2a7f942f681aff10344af88" Dec 13 07:54:16 crc kubenswrapper[4971]: E1213 07:54:16.772764 4971 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-82xjz_openshift-machine-config-operator(e0c4af12-edda-41d7-9b44-f87396174ef2)\"" pod="openshift-machine-config-operator/machine-config-daemon-82xjz" podUID="e0c4af12-edda-41d7-9b44-f87396174ef2" Dec 13 07:54:27 crc kubenswrapper[4971]: I1213 07:54:27.768860 4971 scope.go:117] "RemoveContainer" containerID="e975a5822755c28b3c504aa81fe38e85a7a58425e2a7f942f681aff10344af88" Dec 13 07:54:27 crc kubenswrapper[4971]: E1213 07:54:27.769671 4971 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-82xjz_openshift-machine-config-operator(e0c4af12-edda-41d7-9b44-f87396174ef2)\"" pod="openshift-machine-config-operator/machine-config-daemon-82xjz" podUID="e0c4af12-edda-41d7-9b44-f87396174ef2" Dec 13 07:54:41 crc kubenswrapper[4971]: I1213 07:54:41.769108 4971 scope.go:117] "RemoveContainer" containerID="e975a5822755c28b3c504aa81fe38e85a7a58425e2a7f942f681aff10344af88" Dec 13 07:54:41 crc kubenswrapper[4971]: E1213 07:54:41.769977 4971 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-82xjz_openshift-machine-config-operator(e0c4af12-edda-41d7-9b44-f87396174ef2)\"" pod="openshift-machine-config-operator/machine-config-daemon-82xjz" podUID="e0c4af12-edda-41d7-9b44-f87396174ef2" Dec 13 07:54:45 crc kubenswrapper[4971]: I1213 07:54:45.096710 4971 generic.go:334] "Generic (PLEG): container finished" podID="ce803e3c-be8a-4ee3-ac0b-42b035c043a5" containerID="3dd6fb5bc5f369c42726e16bbfdbf2ee181d92464f6ddc5409e5863fb73c10b2" exitCode=0 Dec 13 07:54:45 crc kubenswrapper[4971]: I1213 07:54:45.096807 4971 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/libvirt-edpm-deployment-openstack-edpm-ipam-47wpm" event={"ID":"ce803e3c-be8a-4ee3-ac0b-42b035c043a5","Type":"ContainerDied","Data":"3dd6fb5bc5f369c42726e16bbfdbf2ee181d92464f6ddc5409e5863fb73c10b2"} Dec 13 07:54:46 crc kubenswrapper[4971]: I1213 07:54:46.576954 4971 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/libvirt-edpm-deployment-openstack-edpm-ipam-47wpm" Dec 13 07:54:46 crc kubenswrapper[4971]: I1213 07:54:46.740970 4971 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"libvirt-secret-0\" (UniqueName: \"kubernetes.io/secret/ce803e3c-be8a-4ee3-ac0b-42b035c043a5-libvirt-secret-0\") pod \"ce803e3c-be8a-4ee3-ac0b-42b035c043a5\" (UID: \"ce803e3c-be8a-4ee3-ac0b-42b035c043a5\") " Dec 13 07:54:46 crc kubenswrapper[4971]: I1213 07:54:46.741323 4971 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/ce803e3c-be8a-4ee3-ac0b-42b035c043a5-inventory\") pod \"ce803e3c-be8a-4ee3-ac0b-42b035c043a5\" (UID: \"ce803e3c-be8a-4ee3-ac0b-42b035c043a5\") " Dec 13 07:54:46 crc kubenswrapper[4971]: I1213 07:54:46.741384 4971 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-mpl6j\" (UniqueName: \"kubernetes.io/projected/ce803e3c-be8a-4ee3-ac0b-42b035c043a5-kube-api-access-mpl6j\") pod \"ce803e3c-be8a-4ee3-ac0b-42b035c043a5\" (UID: \"ce803e3c-be8a-4ee3-ac0b-42b035c043a5\") " Dec 13 07:54:46 crc kubenswrapper[4971]: I1213 07:54:46.741451 4971 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/ce803e3c-be8a-4ee3-ac0b-42b035c043a5-ssh-key\") pod \"ce803e3c-be8a-4ee3-ac0b-42b035c043a5\" (UID: \"ce803e3c-be8a-4ee3-ac0b-42b035c043a5\") " Dec 13 07:54:46 crc kubenswrapper[4971]: I1213 07:54:46.741481 4971 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/ce803e3c-be8a-4ee3-ac0b-42b035c043a5-ceph\") pod \"ce803e3c-be8a-4ee3-ac0b-42b035c043a5\" (UID: \"ce803e3c-be8a-4ee3-ac0b-42b035c043a5\") " Dec 13 07:54:46 crc kubenswrapper[4971]: I1213 07:54:46.741504 4971 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"libvirt-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/ce803e3c-be8a-4ee3-ac0b-42b035c043a5-libvirt-combined-ca-bundle\") pod \"ce803e3c-be8a-4ee3-ac0b-42b035c043a5\" (UID: \"ce803e3c-be8a-4ee3-ac0b-42b035c043a5\") " Dec 13 07:54:46 crc kubenswrapper[4971]: I1213 07:54:46.748156 4971 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/ce803e3c-be8a-4ee3-ac0b-42b035c043a5-ceph" (OuterVolumeSpecName: "ceph") pod "ce803e3c-be8a-4ee3-ac0b-42b035c043a5" (UID: "ce803e3c-be8a-4ee3-ac0b-42b035c043a5"). InnerVolumeSpecName "ceph". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 13 07:54:46 crc kubenswrapper[4971]: I1213 07:54:46.749032 4971 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/ce803e3c-be8a-4ee3-ac0b-42b035c043a5-kube-api-access-mpl6j" (OuterVolumeSpecName: "kube-api-access-mpl6j") pod "ce803e3c-be8a-4ee3-ac0b-42b035c043a5" (UID: "ce803e3c-be8a-4ee3-ac0b-42b035c043a5"). InnerVolumeSpecName "kube-api-access-mpl6j". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 13 07:54:46 crc kubenswrapper[4971]: I1213 07:54:46.749914 4971 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/ce803e3c-be8a-4ee3-ac0b-42b035c043a5-libvirt-combined-ca-bundle" (OuterVolumeSpecName: "libvirt-combined-ca-bundle") pod "ce803e3c-be8a-4ee3-ac0b-42b035c043a5" (UID: "ce803e3c-be8a-4ee3-ac0b-42b035c043a5"). InnerVolumeSpecName "libvirt-combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 13 07:54:46 crc kubenswrapper[4971]: I1213 07:54:46.769599 4971 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/ce803e3c-be8a-4ee3-ac0b-42b035c043a5-libvirt-secret-0" (OuterVolumeSpecName: "libvirt-secret-0") pod "ce803e3c-be8a-4ee3-ac0b-42b035c043a5" (UID: "ce803e3c-be8a-4ee3-ac0b-42b035c043a5"). InnerVolumeSpecName "libvirt-secret-0". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 13 07:54:46 crc kubenswrapper[4971]: I1213 07:54:46.774891 4971 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/ce803e3c-be8a-4ee3-ac0b-42b035c043a5-ssh-key" (OuterVolumeSpecName: "ssh-key") pod "ce803e3c-be8a-4ee3-ac0b-42b035c043a5" (UID: "ce803e3c-be8a-4ee3-ac0b-42b035c043a5"). InnerVolumeSpecName "ssh-key". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 13 07:54:46 crc kubenswrapper[4971]: I1213 07:54:46.778683 4971 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/ce803e3c-be8a-4ee3-ac0b-42b035c043a5-inventory" (OuterVolumeSpecName: "inventory") pod "ce803e3c-be8a-4ee3-ac0b-42b035c043a5" (UID: "ce803e3c-be8a-4ee3-ac0b-42b035c043a5"). InnerVolumeSpecName "inventory". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 13 07:54:46 crc kubenswrapper[4971]: I1213 07:54:46.843619 4971 reconciler_common.go:293] "Volume detached for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/ce803e3c-be8a-4ee3-ac0b-42b035c043a5-ceph\") on node \"crc\" DevicePath \"\"" Dec 13 07:54:46 crc kubenswrapper[4971]: I1213 07:54:46.843653 4971 reconciler_common.go:293] "Volume detached for volume \"libvirt-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/ce803e3c-be8a-4ee3-ac0b-42b035c043a5-libvirt-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 13 07:54:46 crc kubenswrapper[4971]: I1213 07:54:46.843672 4971 reconciler_common.go:293] "Volume detached for volume \"libvirt-secret-0\" (UniqueName: \"kubernetes.io/secret/ce803e3c-be8a-4ee3-ac0b-42b035c043a5-libvirt-secret-0\") on node \"crc\" DevicePath \"\"" Dec 13 07:54:46 crc kubenswrapper[4971]: I1213 07:54:46.843685 4971 reconciler_common.go:293] "Volume detached for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/ce803e3c-be8a-4ee3-ac0b-42b035c043a5-inventory\") on node \"crc\" DevicePath \"\"" Dec 13 07:54:46 crc kubenswrapper[4971]: I1213 07:54:46.843708 4971 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-mpl6j\" (UniqueName: \"kubernetes.io/projected/ce803e3c-be8a-4ee3-ac0b-42b035c043a5-kube-api-access-mpl6j\") on node \"crc\" DevicePath \"\"" Dec 13 07:54:46 crc kubenswrapper[4971]: I1213 07:54:46.843718 4971 reconciler_common.go:293] "Volume detached for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/ce803e3c-be8a-4ee3-ac0b-42b035c043a5-ssh-key\") on node \"crc\" DevicePath \"\"" Dec 13 07:54:47 crc kubenswrapper[4971]: I1213 07:54:47.114405 4971 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/libvirt-edpm-deployment-openstack-edpm-ipam-47wpm" event={"ID":"ce803e3c-be8a-4ee3-ac0b-42b035c043a5","Type":"ContainerDied","Data":"1aafacb341ba7b8e7758aa19be358053c9cc26dcc6a38caf8b4f2569407c0613"} Dec 13 07:54:47 crc kubenswrapper[4971]: I1213 07:54:47.114464 4971 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="1aafacb341ba7b8e7758aa19be358053c9cc26dcc6a38caf8b4f2569407c0613" Dec 13 07:54:47 crc kubenswrapper[4971]: I1213 07:54:47.114548 4971 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/libvirt-edpm-deployment-openstack-edpm-ipam-47wpm" Dec 13 07:54:47 crc kubenswrapper[4971]: I1213 07:54:47.274876 4971 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-custom-ceph-edpm-deployment-openstack-edpm-ipam-4lrjf"] Dec 13 07:54:47 crc kubenswrapper[4971]: E1213 07:54:47.275468 4971 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="ce803e3c-be8a-4ee3-ac0b-42b035c043a5" containerName="libvirt-edpm-deployment-openstack-edpm-ipam" Dec 13 07:54:47 crc kubenswrapper[4971]: I1213 07:54:47.275491 4971 state_mem.go:107] "Deleted CPUSet assignment" podUID="ce803e3c-be8a-4ee3-ac0b-42b035c043a5" containerName="libvirt-edpm-deployment-openstack-edpm-ipam" Dec 13 07:54:47 crc kubenswrapper[4971]: E1213 07:54:47.275527 4971 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="6de28224-6009-4541-97d8-3d01a36c7a01" containerName="registry-server" Dec 13 07:54:47 crc kubenswrapper[4971]: I1213 07:54:47.275541 4971 state_mem.go:107] "Deleted CPUSet assignment" podUID="6de28224-6009-4541-97d8-3d01a36c7a01" containerName="registry-server" Dec 13 07:54:47 crc kubenswrapper[4971]: E1213 07:54:47.275572 4971 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="6de28224-6009-4541-97d8-3d01a36c7a01" containerName="extract-content" Dec 13 07:54:47 crc kubenswrapper[4971]: I1213 07:54:47.275580 4971 state_mem.go:107] "Deleted CPUSet assignment" podUID="6de28224-6009-4541-97d8-3d01a36c7a01" containerName="extract-content" Dec 13 07:54:47 crc kubenswrapper[4971]: E1213 07:54:47.275594 4971 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="a45b0b6c-0728-4808-b2c1-f4194222e636" containerName="registry-server" Dec 13 07:54:47 crc kubenswrapper[4971]: I1213 07:54:47.275603 4971 state_mem.go:107] "Deleted CPUSet assignment" podUID="a45b0b6c-0728-4808-b2c1-f4194222e636" containerName="registry-server" Dec 13 07:54:47 crc kubenswrapper[4971]: E1213 07:54:47.275633 4971 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="6de28224-6009-4541-97d8-3d01a36c7a01" containerName="extract-utilities" Dec 13 07:54:47 crc kubenswrapper[4971]: I1213 07:54:47.275641 4971 state_mem.go:107] "Deleted CPUSet assignment" podUID="6de28224-6009-4541-97d8-3d01a36c7a01" containerName="extract-utilities" Dec 13 07:54:47 crc kubenswrapper[4971]: E1213 07:54:47.275654 4971 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="a45b0b6c-0728-4808-b2c1-f4194222e636" containerName="extract-utilities" Dec 13 07:54:47 crc kubenswrapper[4971]: I1213 07:54:47.275661 4971 state_mem.go:107] "Deleted CPUSet assignment" podUID="a45b0b6c-0728-4808-b2c1-f4194222e636" containerName="extract-utilities" Dec 13 07:54:47 crc kubenswrapper[4971]: E1213 07:54:47.275703 4971 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="a45b0b6c-0728-4808-b2c1-f4194222e636" containerName="extract-content" Dec 13 07:54:47 crc kubenswrapper[4971]: I1213 07:54:47.275712 4971 state_mem.go:107] "Deleted CPUSet assignment" podUID="a45b0b6c-0728-4808-b2c1-f4194222e636" containerName="extract-content" Dec 13 07:54:47 crc kubenswrapper[4971]: I1213 07:54:47.275971 4971 memory_manager.go:354] "RemoveStaleState removing state" podUID="a45b0b6c-0728-4808-b2c1-f4194222e636" containerName="registry-server" Dec 13 07:54:47 crc kubenswrapper[4971]: I1213 07:54:47.275994 4971 memory_manager.go:354] "RemoveStaleState removing state" podUID="6de28224-6009-4541-97d8-3d01a36c7a01" containerName="registry-server" Dec 13 07:54:47 crc kubenswrapper[4971]: I1213 07:54:47.276009 4971 memory_manager.go:354] "RemoveStaleState removing state" podUID="ce803e3c-be8a-4ee3-ac0b-42b035c043a5" containerName="libvirt-edpm-deployment-openstack-edpm-ipam" Dec 13 07:54:47 crc kubenswrapper[4971]: I1213 07:54:47.276920 4971 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-custom-ceph-edpm-deployment-openstack-edpm-ipam-4lrjf" Dec 13 07:54:47 crc kubenswrapper[4971]: I1213 07:54:47.281487 4971 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-custom-ceph-edpm-deployment-openstack-edpm-ipam-4lrjf"] Dec 13 07:54:47 crc kubenswrapper[4971]: I1213 07:54:47.285475 4971 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"ceph-nova" Dec 13 07:54:47 crc kubenswrapper[4971]: I1213 07:54:47.285684 4971 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-aee-default-env" Dec 13 07:54:47 crc kubenswrapper[4971]: I1213 07:54:47.285813 4971 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"nova-extra-config" Dec 13 07:54:47 crc kubenswrapper[4971]: I1213 07:54:47.285998 4971 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-migration-ssh-key" Dec 13 07:54:47 crc kubenswrapper[4971]: I1213 07:54:47.286597 4971 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplane-ansible-ssh-private-key-secret" Dec 13 07:54:47 crc kubenswrapper[4971]: I1213 07:54:47.293527 4971 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceph-conf-files" Dec 13 07:54:47 crc kubenswrapper[4971]: I1213 07:54:47.293806 4971 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-cell1-compute-config" Dec 13 07:54:47 crc kubenswrapper[4971]: I1213 07:54:47.294102 4971 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplanenodeset-openstack-edpm-ipam" Dec 13 07:54:47 crc kubenswrapper[4971]: I1213 07:54:47.294369 4971 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-edpm-ipam-dockercfg-zb2j5" Dec 13 07:54:47 crc kubenswrapper[4971]: I1213 07:54:47.456285 4971 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/e3e414b3-58e3-4da8-9542-7b6d2f229b8d-ssh-key\") pod \"nova-custom-ceph-edpm-deployment-openstack-edpm-ipam-4lrjf\" (UID: \"e3e414b3-58e3-4da8-9542-7b6d2f229b8d\") " pod="openstack/nova-custom-ceph-edpm-deployment-openstack-edpm-ipam-4lrjf" Dec 13 07:54:47 crc kubenswrapper[4971]: I1213 07:54:47.456340 4971 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/e3e414b3-58e3-4da8-9542-7b6d2f229b8d-inventory\") pod \"nova-custom-ceph-edpm-deployment-openstack-edpm-ipam-4lrjf\" (UID: \"e3e414b3-58e3-4da8-9542-7b6d2f229b8d\") " pod="openstack/nova-custom-ceph-edpm-deployment-openstack-edpm-ipam-4lrjf" Dec 13 07:54:47 crc kubenswrapper[4971]: I1213 07:54:47.456407 4971 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-bdzlh\" (UniqueName: \"kubernetes.io/projected/e3e414b3-58e3-4da8-9542-7b6d2f229b8d-kube-api-access-bdzlh\") pod \"nova-custom-ceph-edpm-deployment-openstack-edpm-ipam-4lrjf\" (UID: \"e3e414b3-58e3-4da8-9542-7b6d2f229b8d\") " pod="openstack/nova-custom-ceph-edpm-deployment-openstack-edpm-ipam-4lrjf" Dec 13 07:54:47 crc kubenswrapper[4971]: I1213 07:54:47.456427 4971 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"nova-migration-ssh-key-0\" (UniqueName: \"kubernetes.io/secret/e3e414b3-58e3-4da8-9542-7b6d2f229b8d-nova-migration-ssh-key-0\") pod \"nova-custom-ceph-edpm-deployment-openstack-edpm-ipam-4lrjf\" (UID: \"e3e414b3-58e3-4da8-9542-7b6d2f229b8d\") " pod="openstack/nova-custom-ceph-edpm-deployment-openstack-edpm-ipam-4lrjf" Dec 13 07:54:47 crc kubenswrapper[4971]: I1213 07:54:47.456454 4971 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"nova-custom-ceph-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/e3e414b3-58e3-4da8-9542-7b6d2f229b8d-nova-custom-ceph-combined-ca-bundle\") pod \"nova-custom-ceph-edpm-deployment-openstack-edpm-ipam-4lrjf\" (UID: \"e3e414b3-58e3-4da8-9542-7b6d2f229b8d\") " pod="openstack/nova-custom-ceph-edpm-deployment-openstack-edpm-ipam-4lrjf" Dec 13 07:54:47 crc kubenswrapper[4971]: I1213 07:54:47.456478 4971 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"nova-cell1-compute-config-1\" (UniqueName: \"kubernetes.io/secret/e3e414b3-58e3-4da8-9542-7b6d2f229b8d-nova-cell1-compute-config-1\") pod \"nova-custom-ceph-edpm-deployment-openstack-edpm-ipam-4lrjf\" (UID: \"e3e414b3-58e3-4da8-9542-7b6d2f229b8d\") " pod="openstack/nova-custom-ceph-edpm-deployment-openstack-edpm-ipam-4lrjf" Dec 13 07:54:47 crc kubenswrapper[4971]: I1213 07:54:47.456543 4971 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"nova-cell1-compute-config-0\" (UniqueName: \"kubernetes.io/secret/e3e414b3-58e3-4da8-9542-7b6d2f229b8d-nova-cell1-compute-config-0\") pod \"nova-custom-ceph-edpm-deployment-openstack-edpm-ipam-4lrjf\" (UID: \"e3e414b3-58e3-4da8-9542-7b6d2f229b8d\") " pod="openstack/nova-custom-ceph-edpm-deployment-openstack-edpm-ipam-4lrjf" Dec 13 07:54:47 crc kubenswrapper[4971]: I1213 07:54:47.456596 4971 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/e3e414b3-58e3-4da8-9542-7b6d2f229b8d-ceph\") pod \"nova-custom-ceph-edpm-deployment-openstack-edpm-ipam-4lrjf\" (UID: \"e3e414b3-58e3-4da8-9542-7b6d2f229b8d\") " pod="openstack/nova-custom-ceph-edpm-deployment-openstack-edpm-ipam-4lrjf" Dec 13 07:54:47 crc kubenswrapper[4971]: I1213 07:54:47.456639 4971 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ceph-nova-0\" (UniqueName: \"kubernetes.io/configmap/e3e414b3-58e3-4da8-9542-7b6d2f229b8d-ceph-nova-0\") pod \"nova-custom-ceph-edpm-deployment-openstack-edpm-ipam-4lrjf\" (UID: \"e3e414b3-58e3-4da8-9542-7b6d2f229b8d\") " pod="openstack/nova-custom-ceph-edpm-deployment-openstack-edpm-ipam-4lrjf" Dec 13 07:54:47 crc kubenswrapper[4971]: I1213 07:54:47.456658 4971 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"nova-migration-ssh-key-1\" (UniqueName: \"kubernetes.io/secret/e3e414b3-58e3-4da8-9542-7b6d2f229b8d-nova-migration-ssh-key-1\") pod \"nova-custom-ceph-edpm-deployment-openstack-edpm-ipam-4lrjf\" (UID: \"e3e414b3-58e3-4da8-9542-7b6d2f229b8d\") " pod="openstack/nova-custom-ceph-edpm-deployment-openstack-edpm-ipam-4lrjf" Dec 13 07:54:47 crc kubenswrapper[4971]: I1213 07:54:47.456696 4971 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"nova-extra-config-0\" (UniqueName: \"kubernetes.io/configmap/e3e414b3-58e3-4da8-9542-7b6d2f229b8d-nova-extra-config-0\") pod \"nova-custom-ceph-edpm-deployment-openstack-edpm-ipam-4lrjf\" (UID: \"e3e414b3-58e3-4da8-9542-7b6d2f229b8d\") " pod="openstack/nova-custom-ceph-edpm-deployment-openstack-edpm-ipam-4lrjf" Dec 13 07:54:47 crc kubenswrapper[4971]: I1213 07:54:47.558765 4971 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-bdzlh\" (UniqueName: \"kubernetes.io/projected/e3e414b3-58e3-4da8-9542-7b6d2f229b8d-kube-api-access-bdzlh\") pod \"nova-custom-ceph-edpm-deployment-openstack-edpm-ipam-4lrjf\" (UID: \"e3e414b3-58e3-4da8-9542-7b6d2f229b8d\") " pod="openstack/nova-custom-ceph-edpm-deployment-openstack-edpm-ipam-4lrjf" Dec 13 07:54:47 crc kubenswrapper[4971]: I1213 07:54:47.558815 4971 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nova-custom-ceph-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/e3e414b3-58e3-4da8-9542-7b6d2f229b8d-nova-custom-ceph-combined-ca-bundle\") pod \"nova-custom-ceph-edpm-deployment-openstack-edpm-ipam-4lrjf\" (UID: \"e3e414b3-58e3-4da8-9542-7b6d2f229b8d\") " pod="openstack/nova-custom-ceph-edpm-deployment-openstack-edpm-ipam-4lrjf" Dec 13 07:54:47 crc kubenswrapper[4971]: I1213 07:54:47.558838 4971 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nova-migration-ssh-key-0\" (UniqueName: \"kubernetes.io/secret/e3e414b3-58e3-4da8-9542-7b6d2f229b8d-nova-migration-ssh-key-0\") pod \"nova-custom-ceph-edpm-deployment-openstack-edpm-ipam-4lrjf\" (UID: \"e3e414b3-58e3-4da8-9542-7b6d2f229b8d\") " pod="openstack/nova-custom-ceph-edpm-deployment-openstack-edpm-ipam-4lrjf" Dec 13 07:54:47 crc kubenswrapper[4971]: I1213 07:54:47.558868 4971 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nova-cell1-compute-config-1\" (UniqueName: \"kubernetes.io/secret/e3e414b3-58e3-4da8-9542-7b6d2f229b8d-nova-cell1-compute-config-1\") pod \"nova-custom-ceph-edpm-deployment-openstack-edpm-ipam-4lrjf\" (UID: \"e3e414b3-58e3-4da8-9542-7b6d2f229b8d\") " pod="openstack/nova-custom-ceph-edpm-deployment-openstack-edpm-ipam-4lrjf" Dec 13 07:54:47 crc kubenswrapper[4971]: I1213 07:54:47.558917 4971 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nova-cell1-compute-config-0\" (UniqueName: \"kubernetes.io/secret/e3e414b3-58e3-4da8-9542-7b6d2f229b8d-nova-cell1-compute-config-0\") pod \"nova-custom-ceph-edpm-deployment-openstack-edpm-ipam-4lrjf\" (UID: \"e3e414b3-58e3-4da8-9542-7b6d2f229b8d\") " pod="openstack/nova-custom-ceph-edpm-deployment-openstack-edpm-ipam-4lrjf" Dec 13 07:54:47 crc kubenswrapper[4971]: I1213 07:54:47.558964 4971 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/e3e414b3-58e3-4da8-9542-7b6d2f229b8d-ceph\") pod \"nova-custom-ceph-edpm-deployment-openstack-edpm-ipam-4lrjf\" (UID: \"e3e414b3-58e3-4da8-9542-7b6d2f229b8d\") " pod="openstack/nova-custom-ceph-edpm-deployment-openstack-edpm-ipam-4lrjf" Dec 13 07:54:47 crc kubenswrapper[4971]: I1213 07:54:47.559031 4971 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ceph-nova-0\" (UniqueName: \"kubernetes.io/configmap/e3e414b3-58e3-4da8-9542-7b6d2f229b8d-ceph-nova-0\") pod \"nova-custom-ceph-edpm-deployment-openstack-edpm-ipam-4lrjf\" (UID: \"e3e414b3-58e3-4da8-9542-7b6d2f229b8d\") " pod="openstack/nova-custom-ceph-edpm-deployment-openstack-edpm-ipam-4lrjf" Dec 13 07:54:47 crc kubenswrapper[4971]: I1213 07:54:47.559055 4971 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nova-migration-ssh-key-1\" (UniqueName: \"kubernetes.io/secret/e3e414b3-58e3-4da8-9542-7b6d2f229b8d-nova-migration-ssh-key-1\") pod \"nova-custom-ceph-edpm-deployment-openstack-edpm-ipam-4lrjf\" (UID: \"e3e414b3-58e3-4da8-9542-7b6d2f229b8d\") " pod="openstack/nova-custom-ceph-edpm-deployment-openstack-edpm-ipam-4lrjf" Dec 13 07:54:47 crc kubenswrapper[4971]: I1213 07:54:47.559103 4971 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nova-extra-config-0\" (UniqueName: \"kubernetes.io/configmap/e3e414b3-58e3-4da8-9542-7b6d2f229b8d-nova-extra-config-0\") pod \"nova-custom-ceph-edpm-deployment-openstack-edpm-ipam-4lrjf\" (UID: \"e3e414b3-58e3-4da8-9542-7b6d2f229b8d\") " pod="openstack/nova-custom-ceph-edpm-deployment-openstack-edpm-ipam-4lrjf" Dec 13 07:54:47 crc kubenswrapper[4971]: I1213 07:54:47.559140 4971 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/e3e414b3-58e3-4da8-9542-7b6d2f229b8d-ssh-key\") pod \"nova-custom-ceph-edpm-deployment-openstack-edpm-ipam-4lrjf\" (UID: \"e3e414b3-58e3-4da8-9542-7b6d2f229b8d\") " pod="openstack/nova-custom-ceph-edpm-deployment-openstack-edpm-ipam-4lrjf" Dec 13 07:54:47 crc kubenswrapper[4971]: I1213 07:54:47.559162 4971 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/e3e414b3-58e3-4da8-9542-7b6d2f229b8d-inventory\") pod \"nova-custom-ceph-edpm-deployment-openstack-edpm-ipam-4lrjf\" (UID: \"e3e414b3-58e3-4da8-9542-7b6d2f229b8d\") " pod="openstack/nova-custom-ceph-edpm-deployment-openstack-edpm-ipam-4lrjf" Dec 13 07:54:47 crc kubenswrapper[4971]: I1213 07:54:47.560701 4971 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ceph-nova-0\" (UniqueName: \"kubernetes.io/configmap/e3e414b3-58e3-4da8-9542-7b6d2f229b8d-ceph-nova-0\") pod \"nova-custom-ceph-edpm-deployment-openstack-edpm-ipam-4lrjf\" (UID: \"e3e414b3-58e3-4da8-9542-7b6d2f229b8d\") " pod="openstack/nova-custom-ceph-edpm-deployment-openstack-edpm-ipam-4lrjf" Dec 13 07:54:47 crc kubenswrapper[4971]: I1213 07:54:47.560712 4971 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"nova-extra-config-0\" (UniqueName: \"kubernetes.io/configmap/e3e414b3-58e3-4da8-9542-7b6d2f229b8d-nova-extra-config-0\") pod \"nova-custom-ceph-edpm-deployment-openstack-edpm-ipam-4lrjf\" (UID: \"e3e414b3-58e3-4da8-9542-7b6d2f229b8d\") " pod="openstack/nova-custom-ceph-edpm-deployment-openstack-edpm-ipam-4lrjf" Dec 13 07:54:47 crc kubenswrapper[4971]: I1213 07:54:47.564059 4971 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/e3e414b3-58e3-4da8-9542-7b6d2f229b8d-inventory\") pod \"nova-custom-ceph-edpm-deployment-openstack-edpm-ipam-4lrjf\" (UID: \"e3e414b3-58e3-4da8-9542-7b6d2f229b8d\") " pod="openstack/nova-custom-ceph-edpm-deployment-openstack-edpm-ipam-4lrjf" Dec 13 07:54:47 crc kubenswrapper[4971]: I1213 07:54:47.564102 4971 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"nova-cell1-compute-config-0\" (UniqueName: \"kubernetes.io/secret/e3e414b3-58e3-4da8-9542-7b6d2f229b8d-nova-cell1-compute-config-0\") pod \"nova-custom-ceph-edpm-deployment-openstack-edpm-ipam-4lrjf\" (UID: \"e3e414b3-58e3-4da8-9542-7b6d2f229b8d\") " pod="openstack/nova-custom-ceph-edpm-deployment-openstack-edpm-ipam-4lrjf" Dec 13 07:54:47 crc kubenswrapper[4971]: I1213 07:54:47.564262 4971 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"nova-migration-ssh-key-1\" (UniqueName: \"kubernetes.io/secret/e3e414b3-58e3-4da8-9542-7b6d2f229b8d-nova-migration-ssh-key-1\") pod \"nova-custom-ceph-edpm-deployment-openstack-edpm-ipam-4lrjf\" (UID: \"e3e414b3-58e3-4da8-9542-7b6d2f229b8d\") " pod="openstack/nova-custom-ceph-edpm-deployment-openstack-edpm-ipam-4lrjf" Dec 13 07:54:47 crc kubenswrapper[4971]: I1213 07:54:47.564761 4971 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"nova-migration-ssh-key-0\" (UniqueName: \"kubernetes.io/secret/e3e414b3-58e3-4da8-9542-7b6d2f229b8d-nova-migration-ssh-key-0\") pod \"nova-custom-ceph-edpm-deployment-openstack-edpm-ipam-4lrjf\" (UID: \"e3e414b3-58e3-4da8-9542-7b6d2f229b8d\") " pod="openstack/nova-custom-ceph-edpm-deployment-openstack-edpm-ipam-4lrjf" Dec 13 07:54:47 crc kubenswrapper[4971]: I1213 07:54:47.566904 4971 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"nova-cell1-compute-config-1\" (UniqueName: \"kubernetes.io/secret/e3e414b3-58e3-4da8-9542-7b6d2f229b8d-nova-cell1-compute-config-1\") pod \"nova-custom-ceph-edpm-deployment-openstack-edpm-ipam-4lrjf\" (UID: \"e3e414b3-58e3-4da8-9542-7b6d2f229b8d\") " pod="openstack/nova-custom-ceph-edpm-deployment-openstack-edpm-ipam-4lrjf" Dec 13 07:54:47 crc kubenswrapper[4971]: I1213 07:54:47.573185 4971 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/e3e414b3-58e3-4da8-9542-7b6d2f229b8d-ceph\") pod \"nova-custom-ceph-edpm-deployment-openstack-edpm-ipam-4lrjf\" (UID: \"e3e414b3-58e3-4da8-9542-7b6d2f229b8d\") " pod="openstack/nova-custom-ceph-edpm-deployment-openstack-edpm-ipam-4lrjf" Dec 13 07:54:47 crc kubenswrapper[4971]: I1213 07:54:47.574367 4971 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"nova-custom-ceph-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/e3e414b3-58e3-4da8-9542-7b6d2f229b8d-nova-custom-ceph-combined-ca-bundle\") pod \"nova-custom-ceph-edpm-deployment-openstack-edpm-ipam-4lrjf\" (UID: \"e3e414b3-58e3-4da8-9542-7b6d2f229b8d\") " pod="openstack/nova-custom-ceph-edpm-deployment-openstack-edpm-ipam-4lrjf" Dec 13 07:54:47 crc kubenswrapper[4971]: I1213 07:54:47.575052 4971 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/e3e414b3-58e3-4da8-9542-7b6d2f229b8d-ssh-key\") pod \"nova-custom-ceph-edpm-deployment-openstack-edpm-ipam-4lrjf\" (UID: \"e3e414b3-58e3-4da8-9542-7b6d2f229b8d\") " pod="openstack/nova-custom-ceph-edpm-deployment-openstack-edpm-ipam-4lrjf" Dec 13 07:54:47 crc kubenswrapper[4971]: I1213 07:54:47.582641 4971 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-bdzlh\" (UniqueName: \"kubernetes.io/projected/e3e414b3-58e3-4da8-9542-7b6d2f229b8d-kube-api-access-bdzlh\") pod \"nova-custom-ceph-edpm-deployment-openstack-edpm-ipam-4lrjf\" (UID: \"e3e414b3-58e3-4da8-9542-7b6d2f229b8d\") " pod="openstack/nova-custom-ceph-edpm-deployment-openstack-edpm-ipam-4lrjf" Dec 13 07:54:47 crc kubenswrapper[4971]: I1213 07:54:47.607894 4971 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-custom-ceph-edpm-deployment-openstack-edpm-ipam-4lrjf" Dec 13 07:54:48 crc kubenswrapper[4971]: I1213 07:54:48.114832 4971 provider.go:102] Refreshing cache for provider: *credentialprovider.defaultDockerConfigProvider Dec 13 07:54:48 crc kubenswrapper[4971]: I1213 07:54:48.121083 4971 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-custom-ceph-edpm-deployment-openstack-edpm-ipam-4lrjf"] Dec 13 07:54:48 crc kubenswrapper[4971]: I1213 07:54:48.127010 4971 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-custom-ceph-edpm-deployment-openstack-edpm-ipam-4lrjf" event={"ID":"e3e414b3-58e3-4da8-9542-7b6d2f229b8d","Type":"ContainerStarted","Data":"2c015b66146cc082fe12ada47babed1bf43fdf32e2911b4265bb80326cd7e9a8"} Dec 13 07:54:51 crc kubenswrapper[4971]: I1213 07:54:51.160683 4971 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-custom-ceph-edpm-deployment-openstack-edpm-ipam-4lrjf" event={"ID":"e3e414b3-58e3-4da8-9542-7b6d2f229b8d","Type":"ContainerStarted","Data":"d498c90947d65120478c2dd0f02643a6a679451d58bb633a23f7329a6e28662b"} Dec 13 07:54:51 crc kubenswrapper[4971]: I1213 07:54:51.214609 4971 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-custom-ceph-edpm-deployment-openstack-edpm-ipam-4lrjf" podStartSLOduration=3.49673694 podStartE2EDuration="4.214587405s" podCreationTimestamp="2025-12-13 07:54:47 +0000 UTC" firstStartedPulling="2025-12-13 07:54:48.114585921 +0000 UTC m=+3944.718995369" lastFinishedPulling="2025-12-13 07:54:48.832436346 +0000 UTC m=+3945.436845834" observedRunningTime="2025-12-13 07:54:51.205271819 +0000 UTC m=+3947.809681297" watchObservedRunningTime="2025-12-13 07:54:51.214587405 +0000 UTC m=+3947.818996863" Dec 13 07:54:55 crc kubenswrapper[4971]: I1213 07:54:55.768954 4971 scope.go:117] "RemoveContainer" containerID="e975a5822755c28b3c504aa81fe38e85a7a58425e2a7f942f681aff10344af88" Dec 13 07:54:56 crc kubenswrapper[4971]: I1213 07:54:56.206320 4971 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-82xjz" event={"ID":"e0c4af12-edda-41d7-9b44-f87396174ef2","Type":"ContainerStarted","Data":"d6378ab93da056e5dac3c55cbcf1f29105621255967ad878505c69af98017008"} Dec 13 07:55:51 crc kubenswrapper[4971]: I1213 07:55:51.957111 4971 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/certified-operators-9kcds"] Dec 13 07:55:51 crc kubenswrapper[4971]: I1213 07:55:51.960482 4971 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-9kcds" Dec 13 07:55:51 crc kubenswrapper[4971]: I1213 07:55:51.966428 4971 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-9kcds"] Dec 13 07:55:52 crc kubenswrapper[4971]: I1213 07:55:52.086368 4971 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-csqw2\" (UniqueName: \"kubernetes.io/projected/1c3b2935-e742-4689-a926-1579059acf6a-kube-api-access-csqw2\") pod \"certified-operators-9kcds\" (UID: \"1c3b2935-e742-4689-a926-1579059acf6a\") " pod="openshift-marketplace/certified-operators-9kcds" Dec 13 07:55:52 crc kubenswrapper[4971]: I1213 07:55:52.086691 4971 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/1c3b2935-e742-4689-a926-1579059acf6a-catalog-content\") pod \"certified-operators-9kcds\" (UID: \"1c3b2935-e742-4689-a926-1579059acf6a\") " pod="openshift-marketplace/certified-operators-9kcds" Dec 13 07:55:52 crc kubenswrapper[4971]: I1213 07:55:52.086928 4971 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/1c3b2935-e742-4689-a926-1579059acf6a-utilities\") pod \"certified-operators-9kcds\" (UID: \"1c3b2935-e742-4689-a926-1579059acf6a\") " pod="openshift-marketplace/certified-operators-9kcds" Dec 13 07:55:52 crc kubenswrapper[4971]: I1213 07:55:52.189364 4971 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/1c3b2935-e742-4689-a926-1579059acf6a-catalog-content\") pod \"certified-operators-9kcds\" (UID: \"1c3b2935-e742-4689-a926-1579059acf6a\") " pod="openshift-marketplace/certified-operators-9kcds" Dec 13 07:55:52 crc kubenswrapper[4971]: I1213 07:55:52.189490 4971 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/1c3b2935-e742-4689-a926-1579059acf6a-utilities\") pod \"certified-operators-9kcds\" (UID: \"1c3b2935-e742-4689-a926-1579059acf6a\") " pod="openshift-marketplace/certified-operators-9kcds" Dec 13 07:55:52 crc kubenswrapper[4971]: I1213 07:55:52.189580 4971 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-csqw2\" (UniqueName: \"kubernetes.io/projected/1c3b2935-e742-4689-a926-1579059acf6a-kube-api-access-csqw2\") pod \"certified-operators-9kcds\" (UID: \"1c3b2935-e742-4689-a926-1579059acf6a\") " pod="openshift-marketplace/certified-operators-9kcds" Dec 13 07:55:52 crc kubenswrapper[4971]: I1213 07:55:52.189992 4971 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/1c3b2935-e742-4689-a926-1579059acf6a-catalog-content\") pod \"certified-operators-9kcds\" (UID: \"1c3b2935-e742-4689-a926-1579059acf6a\") " pod="openshift-marketplace/certified-operators-9kcds" Dec 13 07:55:52 crc kubenswrapper[4971]: I1213 07:55:52.189993 4971 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/1c3b2935-e742-4689-a926-1579059acf6a-utilities\") pod \"certified-operators-9kcds\" (UID: \"1c3b2935-e742-4689-a926-1579059acf6a\") " pod="openshift-marketplace/certified-operators-9kcds" Dec 13 07:55:52 crc kubenswrapper[4971]: I1213 07:55:52.211328 4971 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-csqw2\" (UniqueName: \"kubernetes.io/projected/1c3b2935-e742-4689-a926-1579059acf6a-kube-api-access-csqw2\") pod \"certified-operators-9kcds\" (UID: \"1c3b2935-e742-4689-a926-1579059acf6a\") " pod="openshift-marketplace/certified-operators-9kcds" Dec 13 07:55:52 crc kubenswrapper[4971]: I1213 07:55:52.302307 4971 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-9kcds" Dec 13 07:55:53 crc kubenswrapper[4971]: I1213 07:55:53.334415 4971 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-9kcds"] Dec 13 07:55:53 crc kubenswrapper[4971]: I1213 07:55:53.708556 4971 generic.go:334] "Generic (PLEG): container finished" podID="1c3b2935-e742-4689-a926-1579059acf6a" containerID="ec4da79e98adfd8da6b3419f75f781e432449a898de30ee47113932574b59070" exitCode=0 Dec 13 07:55:53 crc kubenswrapper[4971]: I1213 07:55:53.708744 4971 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-9kcds" event={"ID":"1c3b2935-e742-4689-a926-1579059acf6a","Type":"ContainerDied","Data":"ec4da79e98adfd8da6b3419f75f781e432449a898de30ee47113932574b59070"} Dec 13 07:55:53 crc kubenswrapper[4971]: I1213 07:55:53.708925 4971 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-9kcds" event={"ID":"1c3b2935-e742-4689-a926-1579059acf6a","Type":"ContainerStarted","Data":"2b3a0a4ec3f58f979faa00f081fd011e6f2eb7937a57f1d0322f3c200551d466"} Dec 13 07:55:55 crc kubenswrapper[4971]: I1213 07:55:55.734607 4971 generic.go:334] "Generic (PLEG): container finished" podID="1c3b2935-e742-4689-a926-1579059acf6a" containerID="16f32af8b3e0032c00bffb84954c7efa2ca4aadc32e309b78799120251f1990e" exitCode=0 Dec 13 07:55:55 crc kubenswrapper[4971]: I1213 07:55:55.734679 4971 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-9kcds" event={"ID":"1c3b2935-e742-4689-a926-1579059acf6a","Type":"ContainerDied","Data":"16f32af8b3e0032c00bffb84954c7efa2ca4aadc32e309b78799120251f1990e"} Dec 13 07:55:56 crc kubenswrapper[4971]: I1213 07:55:56.746087 4971 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-9kcds" event={"ID":"1c3b2935-e742-4689-a926-1579059acf6a","Type":"ContainerStarted","Data":"3b2e2e7d999a735ff571a49c08fcaa9753703f98e679add488b20a23230823f5"} Dec 13 07:55:56 crc kubenswrapper[4971]: I1213 07:55:56.765188 4971 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/certified-operators-9kcds" podStartSLOduration=3.263425491 podStartE2EDuration="5.765165974s" podCreationTimestamp="2025-12-13 07:55:51 +0000 UTC" firstStartedPulling="2025-12-13 07:55:53.712348672 +0000 UTC m=+4010.316758120" lastFinishedPulling="2025-12-13 07:55:56.214089155 +0000 UTC m=+4012.818498603" observedRunningTime="2025-12-13 07:55:56.765123163 +0000 UTC m=+4013.369532611" watchObservedRunningTime="2025-12-13 07:55:56.765165974 +0000 UTC m=+4013.369575432" Dec 13 07:56:02 crc kubenswrapper[4971]: I1213 07:56:02.302456 4971 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/certified-operators-9kcds" Dec 13 07:56:02 crc kubenswrapper[4971]: I1213 07:56:02.302956 4971 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/certified-operators-9kcds" Dec 13 07:56:02 crc kubenswrapper[4971]: I1213 07:56:02.360741 4971 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/certified-operators-9kcds" Dec 13 07:56:02 crc kubenswrapper[4971]: I1213 07:56:02.863764 4971 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/certified-operators-9kcds" Dec 13 07:56:02 crc kubenswrapper[4971]: I1213 07:56:02.914748 4971 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-9kcds"] Dec 13 07:56:04 crc kubenswrapper[4971]: I1213 07:56:04.835808 4971 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/certified-operators-9kcds" podUID="1c3b2935-e742-4689-a926-1579059acf6a" containerName="registry-server" containerID="cri-o://3b2e2e7d999a735ff571a49c08fcaa9753703f98e679add488b20a23230823f5" gracePeriod=2 Dec 13 07:56:05 crc kubenswrapper[4971]: I1213 07:56:05.848101 4971 generic.go:334] "Generic (PLEG): container finished" podID="1c3b2935-e742-4689-a926-1579059acf6a" containerID="3b2e2e7d999a735ff571a49c08fcaa9753703f98e679add488b20a23230823f5" exitCode=0 Dec 13 07:56:05 crc kubenswrapper[4971]: I1213 07:56:05.848395 4971 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-9kcds" event={"ID":"1c3b2935-e742-4689-a926-1579059acf6a","Type":"ContainerDied","Data":"3b2e2e7d999a735ff571a49c08fcaa9753703f98e679add488b20a23230823f5"} Dec 13 07:56:06 crc kubenswrapper[4971]: I1213 07:56:06.397085 4971 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-9kcds" Dec 13 07:56:06 crc kubenswrapper[4971]: I1213 07:56:06.456530 4971 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/1c3b2935-e742-4689-a926-1579059acf6a-catalog-content\") pod \"1c3b2935-e742-4689-a926-1579059acf6a\" (UID: \"1c3b2935-e742-4689-a926-1579059acf6a\") " Dec 13 07:56:06 crc kubenswrapper[4971]: I1213 07:56:06.456724 4971 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/1c3b2935-e742-4689-a926-1579059acf6a-utilities\") pod \"1c3b2935-e742-4689-a926-1579059acf6a\" (UID: \"1c3b2935-e742-4689-a926-1579059acf6a\") " Dec 13 07:56:06 crc kubenswrapper[4971]: I1213 07:56:06.456755 4971 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-csqw2\" (UniqueName: \"kubernetes.io/projected/1c3b2935-e742-4689-a926-1579059acf6a-kube-api-access-csqw2\") pod \"1c3b2935-e742-4689-a926-1579059acf6a\" (UID: \"1c3b2935-e742-4689-a926-1579059acf6a\") " Dec 13 07:56:06 crc kubenswrapper[4971]: I1213 07:56:06.457661 4971 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/1c3b2935-e742-4689-a926-1579059acf6a-utilities" (OuterVolumeSpecName: "utilities") pod "1c3b2935-e742-4689-a926-1579059acf6a" (UID: "1c3b2935-e742-4689-a926-1579059acf6a"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 13 07:56:06 crc kubenswrapper[4971]: I1213 07:56:06.462018 4971 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/1c3b2935-e742-4689-a926-1579059acf6a-kube-api-access-csqw2" (OuterVolumeSpecName: "kube-api-access-csqw2") pod "1c3b2935-e742-4689-a926-1579059acf6a" (UID: "1c3b2935-e742-4689-a926-1579059acf6a"). InnerVolumeSpecName "kube-api-access-csqw2". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 13 07:56:06 crc kubenswrapper[4971]: I1213 07:56:06.521831 4971 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/1c3b2935-e742-4689-a926-1579059acf6a-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "1c3b2935-e742-4689-a926-1579059acf6a" (UID: "1c3b2935-e742-4689-a926-1579059acf6a"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 13 07:56:06 crc kubenswrapper[4971]: I1213 07:56:06.558680 4971 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/1c3b2935-e742-4689-a926-1579059acf6a-utilities\") on node \"crc\" DevicePath \"\"" Dec 13 07:56:06 crc kubenswrapper[4971]: I1213 07:56:06.558719 4971 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-csqw2\" (UniqueName: \"kubernetes.io/projected/1c3b2935-e742-4689-a926-1579059acf6a-kube-api-access-csqw2\") on node \"crc\" DevicePath \"\"" Dec 13 07:56:06 crc kubenswrapper[4971]: I1213 07:56:06.558733 4971 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/1c3b2935-e742-4689-a926-1579059acf6a-catalog-content\") on node \"crc\" DevicePath \"\"" Dec 13 07:56:06 crc kubenswrapper[4971]: I1213 07:56:06.859496 4971 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-9kcds" event={"ID":"1c3b2935-e742-4689-a926-1579059acf6a","Type":"ContainerDied","Data":"2b3a0a4ec3f58f979faa00f081fd011e6f2eb7937a57f1d0322f3c200551d466"} Dec 13 07:56:06 crc kubenswrapper[4971]: I1213 07:56:06.859567 4971 scope.go:117] "RemoveContainer" containerID="3b2e2e7d999a735ff571a49c08fcaa9753703f98e679add488b20a23230823f5" Dec 13 07:56:06 crc kubenswrapper[4971]: I1213 07:56:06.859570 4971 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-9kcds" Dec 13 07:56:06 crc kubenswrapper[4971]: I1213 07:56:06.889419 4971 scope.go:117] "RemoveContainer" containerID="16f32af8b3e0032c00bffb84954c7efa2ca4aadc32e309b78799120251f1990e" Dec 13 07:56:06 crc kubenswrapper[4971]: I1213 07:56:06.892020 4971 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-9kcds"] Dec 13 07:56:06 crc kubenswrapper[4971]: I1213 07:56:06.900375 4971 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/certified-operators-9kcds"] Dec 13 07:56:06 crc kubenswrapper[4971]: I1213 07:56:06.931112 4971 scope.go:117] "RemoveContainer" containerID="ec4da79e98adfd8da6b3419f75f781e432449a898de30ee47113932574b59070" Dec 13 07:56:07 crc kubenswrapper[4971]: I1213 07:56:07.789658 4971 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="1c3b2935-e742-4689-a926-1579059acf6a" path="/var/lib/kubelet/pods/1c3b2935-e742-4689-a926-1579059acf6a/volumes" Dec 13 07:57:16 crc kubenswrapper[4971]: I1213 07:57:16.153756 4971 patch_prober.go:28] interesting pod/machine-config-daemon-82xjz container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 13 07:57:16 crc kubenswrapper[4971]: I1213 07:57:16.154363 4971 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-82xjz" podUID="e0c4af12-edda-41d7-9b44-f87396174ef2" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 13 07:57:46 crc kubenswrapper[4971]: I1213 07:57:46.153254 4971 patch_prober.go:28] interesting pod/machine-config-daemon-82xjz container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 13 07:57:46 crc kubenswrapper[4971]: I1213 07:57:46.153732 4971 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-82xjz" podUID="e0c4af12-edda-41d7-9b44-f87396174ef2" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 13 07:58:16 crc kubenswrapper[4971]: I1213 07:58:16.153992 4971 patch_prober.go:28] interesting pod/machine-config-daemon-82xjz container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 13 07:58:16 crc kubenswrapper[4971]: I1213 07:58:16.155960 4971 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-82xjz" podUID="e0c4af12-edda-41d7-9b44-f87396174ef2" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 13 07:58:16 crc kubenswrapper[4971]: I1213 07:58:16.156123 4971 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-82xjz" Dec 13 07:58:16 crc kubenswrapper[4971]: I1213 07:58:16.157018 4971 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"d6378ab93da056e5dac3c55cbcf1f29105621255967ad878505c69af98017008"} pod="openshift-machine-config-operator/machine-config-daemon-82xjz" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Dec 13 07:58:16 crc kubenswrapper[4971]: I1213 07:58:16.157180 4971 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-82xjz" podUID="e0c4af12-edda-41d7-9b44-f87396174ef2" containerName="machine-config-daemon" containerID="cri-o://d6378ab93da056e5dac3c55cbcf1f29105621255967ad878505c69af98017008" gracePeriod=600 Dec 13 07:58:20 crc kubenswrapper[4971]: I1213 07:58:20.448089 4971 generic.go:334] "Generic (PLEG): container finished" podID="e0c4af12-edda-41d7-9b44-f87396174ef2" containerID="d6378ab93da056e5dac3c55cbcf1f29105621255967ad878505c69af98017008" exitCode=0 Dec 13 07:58:20 crc kubenswrapper[4971]: I1213 07:58:20.448302 4971 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-82xjz" event={"ID":"e0c4af12-edda-41d7-9b44-f87396174ef2","Type":"ContainerDied","Data":"d6378ab93da056e5dac3c55cbcf1f29105621255967ad878505c69af98017008"} Dec 13 07:58:20 crc kubenswrapper[4971]: I1213 07:58:20.448779 4971 scope.go:117] "RemoveContainer" containerID="e975a5822755c28b3c504aa81fe38e85a7a58425e2a7f942f681aff10344af88" Dec 13 07:58:21 crc kubenswrapper[4971]: I1213 07:58:21.487850 4971 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-82xjz" event={"ID":"e0c4af12-edda-41d7-9b44-f87396174ef2","Type":"ContainerStarted","Data":"e9e9ecdca81954997172b8f30f0cebf348daa8452034fa9732b767f423fdc0de"} Dec 13 07:58:56 crc kubenswrapper[4971]: I1213 07:58:56.503752 4971 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-marketplace-tb9ft"] Dec 13 07:58:56 crc kubenswrapper[4971]: E1213 07:58:56.504661 4971 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="1c3b2935-e742-4689-a926-1579059acf6a" containerName="registry-server" Dec 13 07:58:56 crc kubenswrapper[4971]: I1213 07:58:56.504675 4971 state_mem.go:107] "Deleted CPUSet assignment" podUID="1c3b2935-e742-4689-a926-1579059acf6a" containerName="registry-server" Dec 13 07:58:56 crc kubenswrapper[4971]: E1213 07:58:56.504707 4971 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="1c3b2935-e742-4689-a926-1579059acf6a" containerName="extract-content" Dec 13 07:58:56 crc kubenswrapper[4971]: I1213 07:58:56.504713 4971 state_mem.go:107] "Deleted CPUSet assignment" podUID="1c3b2935-e742-4689-a926-1579059acf6a" containerName="extract-content" Dec 13 07:58:56 crc kubenswrapper[4971]: E1213 07:58:56.504726 4971 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="1c3b2935-e742-4689-a926-1579059acf6a" containerName="extract-utilities" Dec 13 07:58:56 crc kubenswrapper[4971]: I1213 07:58:56.504734 4971 state_mem.go:107] "Deleted CPUSet assignment" podUID="1c3b2935-e742-4689-a926-1579059acf6a" containerName="extract-utilities" Dec 13 07:58:56 crc kubenswrapper[4971]: I1213 07:58:56.504921 4971 memory_manager.go:354] "RemoveStaleState removing state" podUID="1c3b2935-e742-4689-a926-1579059acf6a" containerName="registry-server" Dec 13 07:58:56 crc kubenswrapper[4971]: I1213 07:58:56.506401 4971 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-tb9ft" Dec 13 07:58:56 crc kubenswrapper[4971]: I1213 07:58:56.514039 4971 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-tb9ft"] Dec 13 07:58:56 crc kubenswrapper[4971]: I1213 07:58:56.658503 4971 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-xhtc9\" (UniqueName: \"kubernetes.io/projected/792209dd-f370-404b-934c-a9a7846463a6-kube-api-access-xhtc9\") pod \"redhat-marketplace-tb9ft\" (UID: \"792209dd-f370-404b-934c-a9a7846463a6\") " pod="openshift-marketplace/redhat-marketplace-tb9ft" Dec 13 07:58:56 crc kubenswrapper[4971]: I1213 07:58:56.658788 4971 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/792209dd-f370-404b-934c-a9a7846463a6-catalog-content\") pod \"redhat-marketplace-tb9ft\" (UID: \"792209dd-f370-404b-934c-a9a7846463a6\") " pod="openshift-marketplace/redhat-marketplace-tb9ft" Dec 13 07:58:56 crc kubenswrapper[4971]: I1213 07:58:56.658926 4971 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/792209dd-f370-404b-934c-a9a7846463a6-utilities\") pod \"redhat-marketplace-tb9ft\" (UID: \"792209dd-f370-404b-934c-a9a7846463a6\") " pod="openshift-marketplace/redhat-marketplace-tb9ft" Dec 13 07:58:56 crc kubenswrapper[4971]: I1213 07:58:56.760660 4971 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-xhtc9\" (UniqueName: \"kubernetes.io/projected/792209dd-f370-404b-934c-a9a7846463a6-kube-api-access-xhtc9\") pod \"redhat-marketplace-tb9ft\" (UID: \"792209dd-f370-404b-934c-a9a7846463a6\") " pod="openshift-marketplace/redhat-marketplace-tb9ft" Dec 13 07:58:56 crc kubenswrapper[4971]: I1213 07:58:56.760800 4971 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/792209dd-f370-404b-934c-a9a7846463a6-catalog-content\") pod \"redhat-marketplace-tb9ft\" (UID: \"792209dd-f370-404b-934c-a9a7846463a6\") " pod="openshift-marketplace/redhat-marketplace-tb9ft" Dec 13 07:58:56 crc kubenswrapper[4971]: I1213 07:58:56.760858 4971 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/792209dd-f370-404b-934c-a9a7846463a6-utilities\") pod \"redhat-marketplace-tb9ft\" (UID: \"792209dd-f370-404b-934c-a9a7846463a6\") " pod="openshift-marketplace/redhat-marketplace-tb9ft" Dec 13 07:58:56 crc kubenswrapper[4971]: I1213 07:58:56.761581 4971 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/792209dd-f370-404b-934c-a9a7846463a6-utilities\") pod \"redhat-marketplace-tb9ft\" (UID: \"792209dd-f370-404b-934c-a9a7846463a6\") " pod="openshift-marketplace/redhat-marketplace-tb9ft" Dec 13 07:58:56 crc kubenswrapper[4971]: I1213 07:58:56.761679 4971 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/792209dd-f370-404b-934c-a9a7846463a6-catalog-content\") pod \"redhat-marketplace-tb9ft\" (UID: \"792209dd-f370-404b-934c-a9a7846463a6\") " pod="openshift-marketplace/redhat-marketplace-tb9ft" Dec 13 07:58:56 crc kubenswrapper[4971]: I1213 07:58:56.786445 4971 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-xhtc9\" (UniqueName: \"kubernetes.io/projected/792209dd-f370-404b-934c-a9a7846463a6-kube-api-access-xhtc9\") pod \"redhat-marketplace-tb9ft\" (UID: \"792209dd-f370-404b-934c-a9a7846463a6\") " pod="openshift-marketplace/redhat-marketplace-tb9ft" Dec 13 07:58:56 crc kubenswrapper[4971]: I1213 07:58:56.826171 4971 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-tb9ft" Dec 13 07:58:57 crc kubenswrapper[4971]: I1213 07:58:57.328260 4971 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-tb9ft"] Dec 13 07:58:57 crc kubenswrapper[4971]: I1213 07:58:57.978630 4971 generic.go:334] "Generic (PLEG): container finished" podID="792209dd-f370-404b-934c-a9a7846463a6" containerID="afe306a349220ad251bd7c62bd1f804ee2cb3ee8bda101ff9863f664cd39648a" exitCode=0 Dec 13 07:58:57 crc kubenswrapper[4971]: I1213 07:58:57.978977 4971 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-tb9ft" event={"ID":"792209dd-f370-404b-934c-a9a7846463a6","Type":"ContainerDied","Data":"afe306a349220ad251bd7c62bd1f804ee2cb3ee8bda101ff9863f664cd39648a"} Dec 13 07:58:57 crc kubenswrapper[4971]: I1213 07:58:57.979013 4971 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-tb9ft" event={"ID":"792209dd-f370-404b-934c-a9a7846463a6","Type":"ContainerStarted","Data":"633b7376dd1451a265ec177fb77d34df1b628cdee37c7a90a22d4d5122b13b97"} Dec 13 07:58:58 crc kubenswrapper[4971]: I1213 07:58:58.988929 4971 generic.go:334] "Generic (PLEG): container finished" podID="792209dd-f370-404b-934c-a9a7846463a6" containerID="fdee2bc0ff932d1163d6ff67498498054e1811dd984054d2ac14db20870e3b12" exitCode=0 Dec 13 07:58:58 crc kubenswrapper[4971]: I1213 07:58:58.988977 4971 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-tb9ft" event={"ID":"792209dd-f370-404b-934c-a9a7846463a6","Type":"ContainerDied","Data":"fdee2bc0ff932d1163d6ff67498498054e1811dd984054d2ac14db20870e3b12"} Dec 13 07:59:01 crc kubenswrapper[4971]: I1213 07:59:01.009589 4971 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-tb9ft" event={"ID":"792209dd-f370-404b-934c-a9a7846463a6","Type":"ContainerStarted","Data":"ce8b677bcae11f4920b9b02ae95b2453726f317b63824efd51fe43e0daa969ee"} Dec 13 07:59:02 crc kubenswrapper[4971]: I1213 07:59:02.036723 4971 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-marketplace-tb9ft" podStartSLOduration=4.554211733 podStartE2EDuration="6.036706066s" podCreationTimestamp="2025-12-13 07:58:56 +0000 UTC" firstStartedPulling="2025-12-13 07:58:57.981045912 +0000 UTC m=+4194.585455360" lastFinishedPulling="2025-12-13 07:58:59.463540245 +0000 UTC m=+4196.067949693" observedRunningTime="2025-12-13 07:59:02.034559705 +0000 UTC m=+4198.638969163" watchObservedRunningTime="2025-12-13 07:59:02.036706066 +0000 UTC m=+4198.641115514" Dec 13 07:59:06 crc kubenswrapper[4971]: I1213 07:59:06.826938 4971 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-marketplace-tb9ft" Dec 13 07:59:06 crc kubenswrapper[4971]: I1213 07:59:06.827472 4971 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-marketplace-tb9ft" Dec 13 07:59:06 crc kubenswrapper[4971]: I1213 07:59:06.871954 4971 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-marketplace-tb9ft" Dec 13 07:59:07 crc kubenswrapper[4971]: I1213 07:59:07.109610 4971 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-marketplace-tb9ft" Dec 13 07:59:07 crc kubenswrapper[4971]: I1213 07:59:07.154955 4971 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-tb9ft"] Dec 13 07:59:09 crc kubenswrapper[4971]: I1213 07:59:09.089437 4971 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-marketplace-tb9ft" podUID="792209dd-f370-404b-934c-a9a7846463a6" containerName="registry-server" containerID="cri-o://ce8b677bcae11f4920b9b02ae95b2453726f317b63824efd51fe43e0daa969ee" gracePeriod=2 Dec 13 07:59:09 crc kubenswrapper[4971]: I1213 07:59:09.574424 4971 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-tb9ft" Dec 13 07:59:09 crc kubenswrapper[4971]: I1213 07:59:09.605026 4971 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/792209dd-f370-404b-934c-a9a7846463a6-utilities\") pod \"792209dd-f370-404b-934c-a9a7846463a6\" (UID: \"792209dd-f370-404b-934c-a9a7846463a6\") " Dec 13 07:59:09 crc kubenswrapper[4971]: I1213 07:59:09.605200 4971 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/792209dd-f370-404b-934c-a9a7846463a6-catalog-content\") pod \"792209dd-f370-404b-934c-a9a7846463a6\" (UID: \"792209dd-f370-404b-934c-a9a7846463a6\") " Dec 13 07:59:09 crc kubenswrapper[4971]: I1213 07:59:09.605250 4971 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-xhtc9\" (UniqueName: \"kubernetes.io/projected/792209dd-f370-404b-934c-a9a7846463a6-kube-api-access-xhtc9\") pod \"792209dd-f370-404b-934c-a9a7846463a6\" (UID: \"792209dd-f370-404b-934c-a9a7846463a6\") " Dec 13 07:59:09 crc kubenswrapper[4971]: I1213 07:59:09.612071 4971 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/792209dd-f370-404b-934c-a9a7846463a6-utilities" (OuterVolumeSpecName: "utilities") pod "792209dd-f370-404b-934c-a9a7846463a6" (UID: "792209dd-f370-404b-934c-a9a7846463a6"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 13 07:59:09 crc kubenswrapper[4971]: I1213 07:59:09.613282 4971 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/792209dd-f370-404b-934c-a9a7846463a6-kube-api-access-xhtc9" (OuterVolumeSpecName: "kube-api-access-xhtc9") pod "792209dd-f370-404b-934c-a9a7846463a6" (UID: "792209dd-f370-404b-934c-a9a7846463a6"). InnerVolumeSpecName "kube-api-access-xhtc9". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 13 07:59:09 crc kubenswrapper[4971]: I1213 07:59:09.633506 4971 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/792209dd-f370-404b-934c-a9a7846463a6-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "792209dd-f370-404b-934c-a9a7846463a6" (UID: "792209dd-f370-404b-934c-a9a7846463a6"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 13 07:59:09 crc kubenswrapper[4971]: I1213 07:59:09.706896 4971 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/792209dd-f370-404b-934c-a9a7846463a6-catalog-content\") on node \"crc\" DevicePath \"\"" Dec 13 07:59:09 crc kubenswrapper[4971]: I1213 07:59:09.706927 4971 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-xhtc9\" (UniqueName: \"kubernetes.io/projected/792209dd-f370-404b-934c-a9a7846463a6-kube-api-access-xhtc9\") on node \"crc\" DevicePath \"\"" Dec 13 07:59:09 crc kubenswrapper[4971]: I1213 07:59:09.706939 4971 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/792209dd-f370-404b-934c-a9a7846463a6-utilities\") on node \"crc\" DevicePath \"\"" Dec 13 07:59:10 crc kubenswrapper[4971]: I1213 07:59:10.102637 4971 generic.go:334] "Generic (PLEG): container finished" podID="792209dd-f370-404b-934c-a9a7846463a6" containerID="ce8b677bcae11f4920b9b02ae95b2453726f317b63824efd51fe43e0daa969ee" exitCode=0 Dec 13 07:59:10 crc kubenswrapper[4971]: I1213 07:59:10.102706 4971 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-tb9ft" event={"ID":"792209dd-f370-404b-934c-a9a7846463a6","Type":"ContainerDied","Data":"ce8b677bcae11f4920b9b02ae95b2453726f317b63824efd51fe43e0daa969ee"} Dec 13 07:59:10 crc kubenswrapper[4971]: I1213 07:59:10.102766 4971 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-tb9ft" event={"ID":"792209dd-f370-404b-934c-a9a7846463a6","Type":"ContainerDied","Data":"633b7376dd1451a265ec177fb77d34df1b628cdee37c7a90a22d4d5122b13b97"} Dec 13 07:59:10 crc kubenswrapper[4971]: I1213 07:59:10.102807 4971 scope.go:117] "RemoveContainer" containerID="ce8b677bcae11f4920b9b02ae95b2453726f317b63824efd51fe43e0daa969ee" Dec 13 07:59:10 crc kubenswrapper[4971]: I1213 07:59:10.102898 4971 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-tb9ft" Dec 13 07:59:10 crc kubenswrapper[4971]: I1213 07:59:10.152039 4971 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-tb9ft"] Dec 13 07:59:10 crc kubenswrapper[4971]: I1213 07:59:10.152899 4971 scope.go:117] "RemoveContainer" containerID="fdee2bc0ff932d1163d6ff67498498054e1811dd984054d2ac14db20870e3b12" Dec 13 07:59:10 crc kubenswrapper[4971]: I1213 07:59:10.171324 4971 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-marketplace-tb9ft"] Dec 13 07:59:10 crc kubenswrapper[4971]: I1213 07:59:10.182100 4971 scope.go:117] "RemoveContainer" containerID="afe306a349220ad251bd7c62bd1f804ee2cb3ee8bda101ff9863f664cd39648a" Dec 13 07:59:10 crc kubenswrapper[4971]: I1213 07:59:10.219227 4971 scope.go:117] "RemoveContainer" containerID="ce8b677bcae11f4920b9b02ae95b2453726f317b63824efd51fe43e0daa969ee" Dec 13 07:59:10 crc kubenswrapper[4971]: E1213 07:59:10.220773 4971 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"ce8b677bcae11f4920b9b02ae95b2453726f317b63824efd51fe43e0daa969ee\": container with ID starting with ce8b677bcae11f4920b9b02ae95b2453726f317b63824efd51fe43e0daa969ee not found: ID does not exist" containerID="ce8b677bcae11f4920b9b02ae95b2453726f317b63824efd51fe43e0daa969ee" Dec 13 07:59:10 crc kubenswrapper[4971]: I1213 07:59:10.220835 4971 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"ce8b677bcae11f4920b9b02ae95b2453726f317b63824efd51fe43e0daa969ee"} err="failed to get container status \"ce8b677bcae11f4920b9b02ae95b2453726f317b63824efd51fe43e0daa969ee\": rpc error: code = NotFound desc = could not find container \"ce8b677bcae11f4920b9b02ae95b2453726f317b63824efd51fe43e0daa969ee\": container with ID starting with ce8b677bcae11f4920b9b02ae95b2453726f317b63824efd51fe43e0daa969ee not found: ID does not exist" Dec 13 07:59:10 crc kubenswrapper[4971]: I1213 07:59:10.220873 4971 scope.go:117] "RemoveContainer" containerID="fdee2bc0ff932d1163d6ff67498498054e1811dd984054d2ac14db20870e3b12" Dec 13 07:59:10 crc kubenswrapper[4971]: E1213 07:59:10.221235 4971 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"fdee2bc0ff932d1163d6ff67498498054e1811dd984054d2ac14db20870e3b12\": container with ID starting with fdee2bc0ff932d1163d6ff67498498054e1811dd984054d2ac14db20870e3b12 not found: ID does not exist" containerID="fdee2bc0ff932d1163d6ff67498498054e1811dd984054d2ac14db20870e3b12" Dec 13 07:59:10 crc kubenswrapper[4971]: I1213 07:59:10.221296 4971 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"fdee2bc0ff932d1163d6ff67498498054e1811dd984054d2ac14db20870e3b12"} err="failed to get container status \"fdee2bc0ff932d1163d6ff67498498054e1811dd984054d2ac14db20870e3b12\": rpc error: code = NotFound desc = could not find container \"fdee2bc0ff932d1163d6ff67498498054e1811dd984054d2ac14db20870e3b12\": container with ID starting with fdee2bc0ff932d1163d6ff67498498054e1811dd984054d2ac14db20870e3b12 not found: ID does not exist" Dec 13 07:59:10 crc kubenswrapper[4971]: I1213 07:59:10.221329 4971 scope.go:117] "RemoveContainer" containerID="afe306a349220ad251bd7c62bd1f804ee2cb3ee8bda101ff9863f664cd39648a" Dec 13 07:59:10 crc kubenswrapper[4971]: E1213 07:59:10.221741 4971 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"afe306a349220ad251bd7c62bd1f804ee2cb3ee8bda101ff9863f664cd39648a\": container with ID starting with afe306a349220ad251bd7c62bd1f804ee2cb3ee8bda101ff9863f664cd39648a not found: ID does not exist" containerID="afe306a349220ad251bd7c62bd1f804ee2cb3ee8bda101ff9863f664cd39648a" Dec 13 07:59:10 crc kubenswrapper[4971]: I1213 07:59:10.221782 4971 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"afe306a349220ad251bd7c62bd1f804ee2cb3ee8bda101ff9863f664cd39648a"} err="failed to get container status \"afe306a349220ad251bd7c62bd1f804ee2cb3ee8bda101ff9863f664cd39648a\": rpc error: code = NotFound desc = could not find container \"afe306a349220ad251bd7c62bd1f804ee2cb3ee8bda101ff9863f664cd39648a\": container with ID starting with afe306a349220ad251bd7c62bd1f804ee2cb3ee8bda101ff9863f664cd39648a not found: ID does not exist" Dec 13 07:59:11 crc kubenswrapper[4971]: I1213 07:59:11.788461 4971 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="792209dd-f370-404b-934c-a9a7846463a6" path="/var/lib/kubelet/pods/792209dd-f370-404b-934c-a9a7846463a6/volumes" Dec 13 07:59:12 crc kubenswrapper[4971]: I1213 07:59:12.528963 4971 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/community-operators-gqklt"] Dec 13 07:59:12 crc kubenswrapper[4971]: E1213 07:59:12.529744 4971 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="792209dd-f370-404b-934c-a9a7846463a6" containerName="extract-content" Dec 13 07:59:12 crc kubenswrapper[4971]: I1213 07:59:12.529762 4971 state_mem.go:107] "Deleted CPUSet assignment" podUID="792209dd-f370-404b-934c-a9a7846463a6" containerName="extract-content" Dec 13 07:59:12 crc kubenswrapper[4971]: E1213 07:59:12.529778 4971 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="792209dd-f370-404b-934c-a9a7846463a6" containerName="extract-utilities" Dec 13 07:59:12 crc kubenswrapper[4971]: I1213 07:59:12.529786 4971 state_mem.go:107] "Deleted CPUSet assignment" podUID="792209dd-f370-404b-934c-a9a7846463a6" containerName="extract-utilities" Dec 13 07:59:12 crc kubenswrapper[4971]: E1213 07:59:12.529811 4971 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="792209dd-f370-404b-934c-a9a7846463a6" containerName="registry-server" Dec 13 07:59:12 crc kubenswrapper[4971]: I1213 07:59:12.529818 4971 state_mem.go:107] "Deleted CPUSet assignment" podUID="792209dd-f370-404b-934c-a9a7846463a6" containerName="registry-server" Dec 13 07:59:12 crc kubenswrapper[4971]: I1213 07:59:12.530023 4971 memory_manager.go:354] "RemoveStaleState removing state" podUID="792209dd-f370-404b-934c-a9a7846463a6" containerName="registry-server" Dec 13 07:59:12 crc kubenswrapper[4971]: I1213 07:59:12.531338 4971 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-gqklt" Dec 13 07:59:12 crc kubenswrapper[4971]: I1213 07:59:12.552467 4971 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-gqklt"] Dec 13 07:59:12 crc kubenswrapper[4971]: I1213 07:59:12.567718 4971 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-pgdb9\" (UniqueName: \"kubernetes.io/projected/2436be9d-5009-43b6-8f67-76d1ec28c83f-kube-api-access-pgdb9\") pod \"community-operators-gqklt\" (UID: \"2436be9d-5009-43b6-8f67-76d1ec28c83f\") " pod="openshift-marketplace/community-operators-gqklt" Dec 13 07:59:12 crc kubenswrapper[4971]: I1213 07:59:12.567799 4971 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/2436be9d-5009-43b6-8f67-76d1ec28c83f-utilities\") pod \"community-operators-gqklt\" (UID: \"2436be9d-5009-43b6-8f67-76d1ec28c83f\") " pod="openshift-marketplace/community-operators-gqklt" Dec 13 07:59:12 crc kubenswrapper[4971]: I1213 07:59:12.568249 4971 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/2436be9d-5009-43b6-8f67-76d1ec28c83f-catalog-content\") pod \"community-operators-gqklt\" (UID: \"2436be9d-5009-43b6-8f67-76d1ec28c83f\") " pod="openshift-marketplace/community-operators-gqklt" Dec 13 07:59:12 crc kubenswrapper[4971]: I1213 07:59:12.670389 4971 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-pgdb9\" (UniqueName: \"kubernetes.io/projected/2436be9d-5009-43b6-8f67-76d1ec28c83f-kube-api-access-pgdb9\") pod \"community-operators-gqklt\" (UID: \"2436be9d-5009-43b6-8f67-76d1ec28c83f\") " pod="openshift-marketplace/community-operators-gqklt" Dec 13 07:59:12 crc kubenswrapper[4971]: I1213 07:59:12.670437 4971 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/2436be9d-5009-43b6-8f67-76d1ec28c83f-utilities\") pod \"community-operators-gqklt\" (UID: \"2436be9d-5009-43b6-8f67-76d1ec28c83f\") " pod="openshift-marketplace/community-operators-gqklt" Dec 13 07:59:12 crc kubenswrapper[4971]: I1213 07:59:12.670505 4971 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/2436be9d-5009-43b6-8f67-76d1ec28c83f-catalog-content\") pod \"community-operators-gqklt\" (UID: \"2436be9d-5009-43b6-8f67-76d1ec28c83f\") " pod="openshift-marketplace/community-operators-gqklt" Dec 13 07:59:12 crc kubenswrapper[4971]: I1213 07:59:12.671062 4971 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/2436be9d-5009-43b6-8f67-76d1ec28c83f-catalog-content\") pod \"community-operators-gqklt\" (UID: \"2436be9d-5009-43b6-8f67-76d1ec28c83f\") " pod="openshift-marketplace/community-operators-gqklt" Dec 13 07:59:12 crc kubenswrapper[4971]: I1213 07:59:12.671192 4971 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/2436be9d-5009-43b6-8f67-76d1ec28c83f-utilities\") pod \"community-operators-gqklt\" (UID: \"2436be9d-5009-43b6-8f67-76d1ec28c83f\") " pod="openshift-marketplace/community-operators-gqklt" Dec 13 07:59:12 crc kubenswrapper[4971]: I1213 07:59:12.951709 4971 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-pgdb9\" (UniqueName: \"kubernetes.io/projected/2436be9d-5009-43b6-8f67-76d1ec28c83f-kube-api-access-pgdb9\") pod \"community-operators-gqklt\" (UID: \"2436be9d-5009-43b6-8f67-76d1ec28c83f\") " pod="openshift-marketplace/community-operators-gqklt" Dec 13 07:59:13 crc kubenswrapper[4971]: I1213 07:59:13.163824 4971 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-gqklt" Dec 13 07:59:13 crc kubenswrapper[4971]: I1213 07:59:13.630603 4971 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-gqklt"] Dec 13 07:59:13 crc kubenswrapper[4971]: W1213 07:59:13.637597 4971 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod2436be9d_5009_43b6_8f67_76d1ec28c83f.slice/crio-709d1194aec0a13f758ed0640e4a4e909a6f1e7d87a09efe5010317f70963224 WatchSource:0}: Error finding container 709d1194aec0a13f758ed0640e4a4e909a6f1e7d87a09efe5010317f70963224: Status 404 returned error can't find the container with id 709d1194aec0a13f758ed0640e4a4e909a6f1e7d87a09efe5010317f70963224 Dec 13 07:59:14 crc kubenswrapper[4971]: I1213 07:59:14.138422 4971 generic.go:334] "Generic (PLEG): container finished" podID="2436be9d-5009-43b6-8f67-76d1ec28c83f" containerID="f5146d46fde89df3c9d490af08b96b43e114368fbc0385a89dc92bf89018ccd4" exitCode=0 Dec 13 07:59:14 crc kubenswrapper[4971]: I1213 07:59:14.138482 4971 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-gqklt" event={"ID":"2436be9d-5009-43b6-8f67-76d1ec28c83f","Type":"ContainerDied","Data":"f5146d46fde89df3c9d490af08b96b43e114368fbc0385a89dc92bf89018ccd4"} Dec 13 07:59:14 crc kubenswrapper[4971]: I1213 07:59:14.139776 4971 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-gqklt" event={"ID":"2436be9d-5009-43b6-8f67-76d1ec28c83f","Type":"ContainerStarted","Data":"709d1194aec0a13f758ed0640e4a4e909a6f1e7d87a09efe5010317f70963224"} Dec 13 07:59:16 crc kubenswrapper[4971]: I1213 07:59:16.164695 4971 generic.go:334] "Generic (PLEG): container finished" podID="2436be9d-5009-43b6-8f67-76d1ec28c83f" containerID="0f67fcfbdf5cf7ec936b642339b9eaa172523dddb03d068cb06321d9e88e842d" exitCode=0 Dec 13 07:59:16 crc kubenswrapper[4971]: I1213 07:59:16.165297 4971 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-gqklt" event={"ID":"2436be9d-5009-43b6-8f67-76d1ec28c83f","Type":"ContainerDied","Data":"0f67fcfbdf5cf7ec936b642339b9eaa172523dddb03d068cb06321d9e88e842d"} Dec 13 07:59:17 crc kubenswrapper[4971]: I1213 07:59:17.111136 4971 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-operators-284t8"] Dec 13 07:59:17 crc kubenswrapper[4971]: I1213 07:59:17.113082 4971 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-284t8" Dec 13 07:59:17 crc kubenswrapper[4971]: I1213 07:59:17.141956 4971 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-284t8"] Dec 13 07:59:17 crc kubenswrapper[4971]: I1213 07:59:17.157545 4971 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-q8892\" (UniqueName: \"kubernetes.io/projected/83b8c4fa-b85d-4cdb-ae08-71461f8dc452-kube-api-access-q8892\") pod \"redhat-operators-284t8\" (UID: \"83b8c4fa-b85d-4cdb-ae08-71461f8dc452\") " pod="openshift-marketplace/redhat-operators-284t8" Dec 13 07:59:17 crc kubenswrapper[4971]: I1213 07:59:17.157618 4971 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/83b8c4fa-b85d-4cdb-ae08-71461f8dc452-catalog-content\") pod \"redhat-operators-284t8\" (UID: \"83b8c4fa-b85d-4cdb-ae08-71461f8dc452\") " pod="openshift-marketplace/redhat-operators-284t8" Dec 13 07:59:17 crc kubenswrapper[4971]: I1213 07:59:17.157683 4971 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/83b8c4fa-b85d-4cdb-ae08-71461f8dc452-utilities\") pod \"redhat-operators-284t8\" (UID: \"83b8c4fa-b85d-4cdb-ae08-71461f8dc452\") " pod="openshift-marketplace/redhat-operators-284t8" Dec 13 07:59:17 crc kubenswrapper[4971]: I1213 07:59:17.183803 4971 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-gqklt" event={"ID":"2436be9d-5009-43b6-8f67-76d1ec28c83f","Type":"ContainerStarted","Data":"d697b14083a212818e76f90258f69d8b24d28ab2c46c69d29ad04a40cbcc395c"} Dec 13 07:59:17 crc kubenswrapper[4971]: I1213 07:59:17.212697 4971 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/community-operators-gqklt" podStartSLOduration=2.689306626 podStartE2EDuration="5.212681503s" podCreationTimestamp="2025-12-13 07:59:12 +0000 UTC" firstStartedPulling="2025-12-13 07:59:14.140571314 +0000 UTC m=+4210.744980762" lastFinishedPulling="2025-12-13 07:59:16.663946191 +0000 UTC m=+4213.268355639" observedRunningTime="2025-12-13 07:59:17.208596064 +0000 UTC m=+4213.813005562" watchObservedRunningTime="2025-12-13 07:59:17.212681503 +0000 UTC m=+4213.817090951" Dec 13 07:59:17 crc kubenswrapper[4971]: I1213 07:59:17.260120 4971 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-q8892\" (UniqueName: \"kubernetes.io/projected/83b8c4fa-b85d-4cdb-ae08-71461f8dc452-kube-api-access-q8892\") pod \"redhat-operators-284t8\" (UID: \"83b8c4fa-b85d-4cdb-ae08-71461f8dc452\") " pod="openshift-marketplace/redhat-operators-284t8" Dec 13 07:59:17 crc kubenswrapper[4971]: I1213 07:59:17.260196 4971 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/83b8c4fa-b85d-4cdb-ae08-71461f8dc452-catalog-content\") pod \"redhat-operators-284t8\" (UID: \"83b8c4fa-b85d-4cdb-ae08-71461f8dc452\") " pod="openshift-marketplace/redhat-operators-284t8" Dec 13 07:59:17 crc kubenswrapper[4971]: I1213 07:59:17.260256 4971 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/83b8c4fa-b85d-4cdb-ae08-71461f8dc452-utilities\") pod \"redhat-operators-284t8\" (UID: \"83b8c4fa-b85d-4cdb-ae08-71461f8dc452\") " pod="openshift-marketplace/redhat-operators-284t8" Dec 13 07:59:17 crc kubenswrapper[4971]: I1213 07:59:17.261281 4971 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/83b8c4fa-b85d-4cdb-ae08-71461f8dc452-utilities\") pod \"redhat-operators-284t8\" (UID: \"83b8c4fa-b85d-4cdb-ae08-71461f8dc452\") " pod="openshift-marketplace/redhat-operators-284t8" Dec 13 07:59:17 crc kubenswrapper[4971]: I1213 07:59:17.261560 4971 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/83b8c4fa-b85d-4cdb-ae08-71461f8dc452-catalog-content\") pod \"redhat-operators-284t8\" (UID: \"83b8c4fa-b85d-4cdb-ae08-71461f8dc452\") " pod="openshift-marketplace/redhat-operators-284t8" Dec 13 07:59:17 crc kubenswrapper[4971]: I1213 07:59:17.290427 4971 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-q8892\" (UniqueName: \"kubernetes.io/projected/83b8c4fa-b85d-4cdb-ae08-71461f8dc452-kube-api-access-q8892\") pod \"redhat-operators-284t8\" (UID: \"83b8c4fa-b85d-4cdb-ae08-71461f8dc452\") " pod="openshift-marketplace/redhat-operators-284t8" Dec 13 07:59:17 crc kubenswrapper[4971]: I1213 07:59:17.432497 4971 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-284t8" Dec 13 07:59:17 crc kubenswrapper[4971]: I1213 07:59:17.906031 4971 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-284t8"] Dec 13 07:59:18 crc kubenswrapper[4971]: I1213 07:59:18.207779 4971 generic.go:334] "Generic (PLEG): container finished" podID="83b8c4fa-b85d-4cdb-ae08-71461f8dc452" containerID="5bf93847e204b1e7bd63e4ec1873420df1968c262541390c8b0570715a252761" exitCode=0 Dec 13 07:59:18 crc kubenswrapper[4971]: I1213 07:59:18.208658 4971 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-284t8" event={"ID":"83b8c4fa-b85d-4cdb-ae08-71461f8dc452","Type":"ContainerDied","Data":"5bf93847e204b1e7bd63e4ec1873420df1968c262541390c8b0570715a252761"} Dec 13 07:59:18 crc kubenswrapper[4971]: I1213 07:59:18.208683 4971 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-284t8" event={"ID":"83b8c4fa-b85d-4cdb-ae08-71461f8dc452","Type":"ContainerStarted","Data":"054d8ea76715026ce89268541f37c9c667c09791c0adc8782ecd51587144d29b"} Dec 13 07:59:19 crc kubenswrapper[4971]: I1213 07:59:19.237556 4971 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-284t8" event={"ID":"83b8c4fa-b85d-4cdb-ae08-71461f8dc452","Type":"ContainerStarted","Data":"07d758c7b52844138aa133128b536957b1e022a2389d88268ddde52893c33df8"} Dec 13 07:59:21 crc kubenswrapper[4971]: I1213 07:59:21.261164 4971 generic.go:334] "Generic (PLEG): container finished" podID="83b8c4fa-b85d-4cdb-ae08-71461f8dc452" containerID="07d758c7b52844138aa133128b536957b1e022a2389d88268ddde52893c33df8" exitCode=0 Dec 13 07:59:21 crc kubenswrapper[4971]: I1213 07:59:21.261224 4971 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-284t8" event={"ID":"83b8c4fa-b85d-4cdb-ae08-71461f8dc452","Type":"ContainerDied","Data":"07d758c7b52844138aa133128b536957b1e022a2389d88268ddde52893c33df8"} Dec 13 07:59:23 crc kubenswrapper[4971]: I1213 07:59:23.164845 4971 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/community-operators-gqklt" Dec 13 07:59:23 crc kubenswrapper[4971]: I1213 07:59:23.165541 4971 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/community-operators-gqklt" Dec 13 07:59:23 crc kubenswrapper[4971]: I1213 07:59:23.221397 4971 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/community-operators-gqklt" Dec 13 07:59:23 crc kubenswrapper[4971]: I1213 07:59:23.283148 4971 generic.go:334] "Generic (PLEG): container finished" podID="e3e414b3-58e3-4da8-9542-7b6d2f229b8d" containerID="d498c90947d65120478c2dd0f02643a6a679451d58bb633a23f7329a6e28662b" exitCode=0 Dec 13 07:59:23 crc kubenswrapper[4971]: I1213 07:59:23.283254 4971 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-custom-ceph-edpm-deployment-openstack-edpm-ipam-4lrjf" event={"ID":"e3e414b3-58e3-4da8-9542-7b6d2f229b8d","Type":"ContainerDied","Data":"d498c90947d65120478c2dd0f02643a6a679451d58bb633a23f7329a6e28662b"} Dec 13 07:59:23 crc kubenswrapper[4971]: I1213 07:59:23.373633 4971 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/community-operators-gqklt" Dec 13 07:59:23 crc kubenswrapper[4971]: I1213 07:59:23.910272 4971 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-gqklt"] Dec 13 07:59:24 crc kubenswrapper[4971]: I1213 07:59:24.298487 4971 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-284t8" event={"ID":"83b8c4fa-b85d-4cdb-ae08-71461f8dc452","Type":"ContainerStarted","Data":"b9fe34225f31127367f411a45e14ea2e4cfda0298892c91dbf4db6971bf79192"} Dec 13 07:59:24 crc kubenswrapper[4971]: I1213 07:59:24.337315 4971 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-operators-284t8" podStartSLOduration=3.125554837 podStartE2EDuration="7.33728304s" podCreationTimestamp="2025-12-13 07:59:17 +0000 UTC" firstStartedPulling="2025-12-13 07:59:18.209561141 +0000 UTC m=+4214.813970589" lastFinishedPulling="2025-12-13 07:59:22.421289334 +0000 UTC m=+4219.025698792" observedRunningTime="2025-12-13 07:59:24.320800831 +0000 UTC m=+4220.925210349" watchObservedRunningTime="2025-12-13 07:59:24.33728304 +0000 UTC m=+4220.941692528" Dec 13 07:59:25 crc kubenswrapper[4971]: I1213 07:59:25.146444 4971 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-custom-ceph-edpm-deployment-openstack-edpm-ipam-4lrjf" Dec 13 07:59:25 crc kubenswrapper[4971]: I1213 07:59:25.227429 4971 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"nova-custom-ceph-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/e3e414b3-58e3-4da8-9542-7b6d2f229b8d-nova-custom-ceph-combined-ca-bundle\") pod \"e3e414b3-58e3-4da8-9542-7b6d2f229b8d\" (UID: \"e3e414b3-58e3-4da8-9542-7b6d2f229b8d\") " Dec 13 07:59:25 crc kubenswrapper[4971]: I1213 07:59:25.227846 4971 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"nova-cell1-compute-config-1\" (UniqueName: \"kubernetes.io/secret/e3e414b3-58e3-4da8-9542-7b6d2f229b8d-nova-cell1-compute-config-1\") pod \"e3e414b3-58e3-4da8-9542-7b6d2f229b8d\" (UID: \"e3e414b3-58e3-4da8-9542-7b6d2f229b8d\") " Dec 13 07:59:25 crc kubenswrapper[4971]: I1213 07:59:25.227966 4971 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-bdzlh\" (UniqueName: \"kubernetes.io/projected/e3e414b3-58e3-4da8-9542-7b6d2f229b8d-kube-api-access-bdzlh\") pod \"e3e414b3-58e3-4da8-9542-7b6d2f229b8d\" (UID: \"e3e414b3-58e3-4da8-9542-7b6d2f229b8d\") " Dec 13 07:59:25 crc kubenswrapper[4971]: I1213 07:59:25.228010 4971 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"nova-extra-config-0\" (UniqueName: \"kubernetes.io/configmap/e3e414b3-58e3-4da8-9542-7b6d2f229b8d-nova-extra-config-0\") pod \"e3e414b3-58e3-4da8-9542-7b6d2f229b8d\" (UID: \"e3e414b3-58e3-4da8-9542-7b6d2f229b8d\") " Dec 13 07:59:25 crc kubenswrapper[4971]: I1213 07:59:25.228090 4971 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/e3e414b3-58e3-4da8-9542-7b6d2f229b8d-ceph\") pod \"e3e414b3-58e3-4da8-9542-7b6d2f229b8d\" (UID: \"e3e414b3-58e3-4da8-9542-7b6d2f229b8d\") " Dec 13 07:59:25 crc kubenswrapper[4971]: I1213 07:59:25.228113 4971 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"nova-migration-ssh-key-1\" (UniqueName: \"kubernetes.io/secret/e3e414b3-58e3-4da8-9542-7b6d2f229b8d-nova-migration-ssh-key-1\") pod \"e3e414b3-58e3-4da8-9542-7b6d2f229b8d\" (UID: \"e3e414b3-58e3-4da8-9542-7b6d2f229b8d\") " Dec 13 07:59:25 crc kubenswrapper[4971]: I1213 07:59:25.228212 4971 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"nova-migration-ssh-key-0\" (UniqueName: \"kubernetes.io/secret/e3e414b3-58e3-4da8-9542-7b6d2f229b8d-nova-migration-ssh-key-0\") pod \"e3e414b3-58e3-4da8-9542-7b6d2f229b8d\" (UID: \"e3e414b3-58e3-4da8-9542-7b6d2f229b8d\") " Dec 13 07:59:25 crc kubenswrapper[4971]: I1213 07:59:25.228245 4971 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/e3e414b3-58e3-4da8-9542-7b6d2f229b8d-ssh-key\") pod \"e3e414b3-58e3-4da8-9542-7b6d2f229b8d\" (UID: \"e3e414b3-58e3-4da8-9542-7b6d2f229b8d\") " Dec 13 07:59:25 crc kubenswrapper[4971]: I1213 07:59:25.228322 4971 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"nova-cell1-compute-config-0\" (UniqueName: \"kubernetes.io/secret/e3e414b3-58e3-4da8-9542-7b6d2f229b8d-nova-cell1-compute-config-0\") pod \"e3e414b3-58e3-4da8-9542-7b6d2f229b8d\" (UID: \"e3e414b3-58e3-4da8-9542-7b6d2f229b8d\") " Dec 13 07:59:25 crc kubenswrapper[4971]: I1213 07:59:25.228361 4971 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/e3e414b3-58e3-4da8-9542-7b6d2f229b8d-inventory\") pod \"e3e414b3-58e3-4da8-9542-7b6d2f229b8d\" (UID: \"e3e414b3-58e3-4da8-9542-7b6d2f229b8d\") " Dec 13 07:59:25 crc kubenswrapper[4971]: I1213 07:59:25.228386 4971 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ceph-nova-0\" (UniqueName: \"kubernetes.io/configmap/e3e414b3-58e3-4da8-9542-7b6d2f229b8d-ceph-nova-0\") pod \"e3e414b3-58e3-4da8-9542-7b6d2f229b8d\" (UID: \"e3e414b3-58e3-4da8-9542-7b6d2f229b8d\") " Dec 13 07:59:25 crc kubenswrapper[4971]: I1213 07:59:25.234808 4971 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/e3e414b3-58e3-4da8-9542-7b6d2f229b8d-kube-api-access-bdzlh" (OuterVolumeSpecName: "kube-api-access-bdzlh") pod "e3e414b3-58e3-4da8-9542-7b6d2f229b8d" (UID: "e3e414b3-58e3-4da8-9542-7b6d2f229b8d"). InnerVolumeSpecName "kube-api-access-bdzlh". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 13 07:59:25 crc kubenswrapper[4971]: I1213 07:59:25.236239 4971 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/e3e414b3-58e3-4da8-9542-7b6d2f229b8d-nova-custom-ceph-combined-ca-bundle" (OuterVolumeSpecName: "nova-custom-ceph-combined-ca-bundle") pod "e3e414b3-58e3-4da8-9542-7b6d2f229b8d" (UID: "e3e414b3-58e3-4da8-9542-7b6d2f229b8d"). InnerVolumeSpecName "nova-custom-ceph-combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 13 07:59:25 crc kubenswrapper[4971]: I1213 07:59:25.236716 4971 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/e3e414b3-58e3-4da8-9542-7b6d2f229b8d-ceph" (OuterVolumeSpecName: "ceph") pod "e3e414b3-58e3-4da8-9542-7b6d2f229b8d" (UID: "e3e414b3-58e3-4da8-9542-7b6d2f229b8d"). InnerVolumeSpecName "ceph". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 13 07:59:25 crc kubenswrapper[4971]: I1213 07:59:25.254337 4971 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/e3e414b3-58e3-4da8-9542-7b6d2f229b8d-ceph-nova-0" (OuterVolumeSpecName: "ceph-nova-0") pod "e3e414b3-58e3-4da8-9542-7b6d2f229b8d" (UID: "e3e414b3-58e3-4da8-9542-7b6d2f229b8d"). InnerVolumeSpecName "ceph-nova-0". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 13 07:59:25 crc kubenswrapper[4971]: I1213 07:59:25.262620 4971 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/e3e414b3-58e3-4da8-9542-7b6d2f229b8d-nova-cell1-compute-config-0" (OuterVolumeSpecName: "nova-cell1-compute-config-0") pod "e3e414b3-58e3-4da8-9542-7b6d2f229b8d" (UID: "e3e414b3-58e3-4da8-9542-7b6d2f229b8d"). InnerVolumeSpecName "nova-cell1-compute-config-0". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 13 07:59:25 crc kubenswrapper[4971]: I1213 07:59:25.270207 4971 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/e3e414b3-58e3-4da8-9542-7b6d2f229b8d-nova-extra-config-0" (OuterVolumeSpecName: "nova-extra-config-0") pod "e3e414b3-58e3-4da8-9542-7b6d2f229b8d" (UID: "e3e414b3-58e3-4da8-9542-7b6d2f229b8d"). InnerVolumeSpecName "nova-extra-config-0". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 13 07:59:25 crc kubenswrapper[4971]: I1213 07:59:25.279150 4971 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/e3e414b3-58e3-4da8-9542-7b6d2f229b8d-ssh-key" (OuterVolumeSpecName: "ssh-key") pod "e3e414b3-58e3-4da8-9542-7b6d2f229b8d" (UID: "e3e414b3-58e3-4da8-9542-7b6d2f229b8d"). InnerVolumeSpecName "ssh-key". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 13 07:59:25 crc kubenswrapper[4971]: I1213 07:59:25.281104 4971 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/e3e414b3-58e3-4da8-9542-7b6d2f229b8d-nova-cell1-compute-config-1" (OuterVolumeSpecName: "nova-cell1-compute-config-1") pod "e3e414b3-58e3-4da8-9542-7b6d2f229b8d" (UID: "e3e414b3-58e3-4da8-9542-7b6d2f229b8d"). InnerVolumeSpecName "nova-cell1-compute-config-1". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 13 07:59:25 crc kubenswrapper[4971]: I1213 07:59:25.281623 4971 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/e3e414b3-58e3-4da8-9542-7b6d2f229b8d-nova-migration-ssh-key-0" (OuterVolumeSpecName: "nova-migration-ssh-key-0") pod "e3e414b3-58e3-4da8-9542-7b6d2f229b8d" (UID: "e3e414b3-58e3-4da8-9542-7b6d2f229b8d"). InnerVolumeSpecName "nova-migration-ssh-key-0". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 13 07:59:25 crc kubenswrapper[4971]: I1213 07:59:25.288398 4971 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/e3e414b3-58e3-4da8-9542-7b6d2f229b8d-nova-migration-ssh-key-1" (OuterVolumeSpecName: "nova-migration-ssh-key-1") pod "e3e414b3-58e3-4da8-9542-7b6d2f229b8d" (UID: "e3e414b3-58e3-4da8-9542-7b6d2f229b8d"). InnerVolumeSpecName "nova-migration-ssh-key-1". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 13 07:59:25 crc kubenswrapper[4971]: I1213 07:59:25.289846 4971 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/e3e414b3-58e3-4da8-9542-7b6d2f229b8d-inventory" (OuterVolumeSpecName: "inventory") pod "e3e414b3-58e3-4da8-9542-7b6d2f229b8d" (UID: "e3e414b3-58e3-4da8-9542-7b6d2f229b8d"). InnerVolumeSpecName "inventory". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 13 07:59:25 crc kubenswrapper[4971]: I1213 07:59:25.311891 4971 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/community-operators-gqklt" podUID="2436be9d-5009-43b6-8f67-76d1ec28c83f" containerName="registry-server" containerID="cri-o://d697b14083a212818e76f90258f69d8b24d28ab2c46c69d29ad04a40cbcc395c" gracePeriod=2 Dec 13 07:59:25 crc kubenswrapper[4971]: I1213 07:59:25.312024 4971 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-custom-ceph-edpm-deployment-openstack-edpm-ipam-4lrjf" Dec 13 07:59:25 crc kubenswrapper[4971]: I1213 07:59:25.312835 4971 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-custom-ceph-edpm-deployment-openstack-edpm-ipam-4lrjf" event={"ID":"e3e414b3-58e3-4da8-9542-7b6d2f229b8d","Type":"ContainerDied","Data":"2c015b66146cc082fe12ada47babed1bf43fdf32e2911b4265bb80326cd7e9a8"} Dec 13 07:59:25 crc kubenswrapper[4971]: I1213 07:59:25.312917 4971 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="2c015b66146cc082fe12ada47babed1bf43fdf32e2911b4265bb80326cd7e9a8" Dec 13 07:59:25 crc kubenswrapper[4971]: I1213 07:59:25.331167 4971 reconciler_common.go:293] "Volume detached for volume \"nova-cell1-compute-config-0\" (UniqueName: \"kubernetes.io/secret/e3e414b3-58e3-4da8-9542-7b6d2f229b8d-nova-cell1-compute-config-0\") on node \"crc\" DevicePath \"\"" Dec 13 07:59:25 crc kubenswrapper[4971]: I1213 07:59:25.331202 4971 reconciler_common.go:293] "Volume detached for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/e3e414b3-58e3-4da8-9542-7b6d2f229b8d-inventory\") on node \"crc\" DevicePath \"\"" Dec 13 07:59:25 crc kubenswrapper[4971]: I1213 07:59:25.331214 4971 reconciler_common.go:293] "Volume detached for volume \"ceph-nova-0\" (UniqueName: \"kubernetes.io/configmap/e3e414b3-58e3-4da8-9542-7b6d2f229b8d-ceph-nova-0\") on node \"crc\" DevicePath \"\"" Dec 13 07:59:25 crc kubenswrapper[4971]: I1213 07:59:25.331229 4971 reconciler_common.go:293] "Volume detached for volume \"nova-custom-ceph-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/e3e414b3-58e3-4da8-9542-7b6d2f229b8d-nova-custom-ceph-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 13 07:59:25 crc kubenswrapper[4971]: I1213 07:59:25.331245 4971 reconciler_common.go:293] "Volume detached for volume \"nova-cell1-compute-config-1\" (UniqueName: \"kubernetes.io/secret/e3e414b3-58e3-4da8-9542-7b6d2f229b8d-nova-cell1-compute-config-1\") on node \"crc\" DevicePath \"\"" Dec 13 07:59:25 crc kubenswrapper[4971]: I1213 07:59:25.331258 4971 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-bdzlh\" (UniqueName: \"kubernetes.io/projected/e3e414b3-58e3-4da8-9542-7b6d2f229b8d-kube-api-access-bdzlh\") on node \"crc\" DevicePath \"\"" Dec 13 07:59:25 crc kubenswrapper[4971]: I1213 07:59:25.331268 4971 reconciler_common.go:293] "Volume detached for volume \"nova-extra-config-0\" (UniqueName: \"kubernetes.io/configmap/e3e414b3-58e3-4da8-9542-7b6d2f229b8d-nova-extra-config-0\") on node \"crc\" DevicePath \"\"" Dec 13 07:59:25 crc kubenswrapper[4971]: I1213 07:59:25.331278 4971 reconciler_common.go:293] "Volume detached for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/e3e414b3-58e3-4da8-9542-7b6d2f229b8d-ceph\") on node \"crc\" DevicePath \"\"" Dec 13 07:59:25 crc kubenswrapper[4971]: I1213 07:59:25.331289 4971 reconciler_common.go:293] "Volume detached for volume \"nova-migration-ssh-key-1\" (UniqueName: \"kubernetes.io/secret/e3e414b3-58e3-4da8-9542-7b6d2f229b8d-nova-migration-ssh-key-1\") on node \"crc\" DevicePath \"\"" Dec 13 07:59:25 crc kubenswrapper[4971]: I1213 07:59:25.331299 4971 reconciler_common.go:293] "Volume detached for volume \"nova-migration-ssh-key-0\" (UniqueName: \"kubernetes.io/secret/e3e414b3-58e3-4da8-9542-7b6d2f229b8d-nova-migration-ssh-key-0\") on node \"crc\" DevicePath \"\"" Dec 13 07:59:25 crc kubenswrapper[4971]: I1213 07:59:25.331309 4971 reconciler_common.go:293] "Volume detached for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/e3e414b3-58e3-4da8-9542-7b6d2f229b8d-ssh-key\") on node \"crc\" DevicePath \"\"" Dec 13 07:59:26 crc kubenswrapper[4971]: I1213 07:59:26.322703 4971 generic.go:334] "Generic (PLEG): container finished" podID="2436be9d-5009-43b6-8f67-76d1ec28c83f" containerID="d697b14083a212818e76f90258f69d8b24d28ab2c46c69d29ad04a40cbcc395c" exitCode=0 Dec 13 07:59:26 crc kubenswrapper[4971]: I1213 07:59:26.323061 4971 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-gqklt" event={"ID":"2436be9d-5009-43b6-8f67-76d1ec28c83f","Type":"ContainerDied","Data":"d697b14083a212818e76f90258f69d8b24d28ab2c46c69d29ad04a40cbcc395c"} Dec 13 07:59:27 crc kubenswrapper[4971]: I1213 07:59:27.092165 4971 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-gqklt" Dec 13 07:59:27 crc kubenswrapper[4971]: I1213 07:59:27.164827 4971 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/2436be9d-5009-43b6-8f67-76d1ec28c83f-utilities\") pod \"2436be9d-5009-43b6-8f67-76d1ec28c83f\" (UID: \"2436be9d-5009-43b6-8f67-76d1ec28c83f\") " Dec 13 07:59:27 crc kubenswrapper[4971]: I1213 07:59:27.165087 4971 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-pgdb9\" (UniqueName: \"kubernetes.io/projected/2436be9d-5009-43b6-8f67-76d1ec28c83f-kube-api-access-pgdb9\") pod \"2436be9d-5009-43b6-8f67-76d1ec28c83f\" (UID: \"2436be9d-5009-43b6-8f67-76d1ec28c83f\") " Dec 13 07:59:27 crc kubenswrapper[4971]: I1213 07:59:27.165119 4971 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/2436be9d-5009-43b6-8f67-76d1ec28c83f-catalog-content\") pod \"2436be9d-5009-43b6-8f67-76d1ec28c83f\" (UID: \"2436be9d-5009-43b6-8f67-76d1ec28c83f\") " Dec 13 07:59:27 crc kubenswrapper[4971]: I1213 07:59:27.165612 4971 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/2436be9d-5009-43b6-8f67-76d1ec28c83f-utilities" (OuterVolumeSpecName: "utilities") pod "2436be9d-5009-43b6-8f67-76d1ec28c83f" (UID: "2436be9d-5009-43b6-8f67-76d1ec28c83f"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 13 07:59:27 crc kubenswrapper[4971]: I1213 07:59:27.179131 4971 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/2436be9d-5009-43b6-8f67-76d1ec28c83f-kube-api-access-pgdb9" (OuterVolumeSpecName: "kube-api-access-pgdb9") pod "2436be9d-5009-43b6-8f67-76d1ec28c83f" (UID: "2436be9d-5009-43b6-8f67-76d1ec28c83f"). InnerVolumeSpecName "kube-api-access-pgdb9". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 13 07:59:27 crc kubenswrapper[4971]: I1213 07:59:27.222448 4971 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/2436be9d-5009-43b6-8f67-76d1ec28c83f-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "2436be9d-5009-43b6-8f67-76d1ec28c83f" (UID: "2436be9d-5009-43b6-8f67-76d1ec28c83f"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 13 07:59:27 crc kubenswrapper[4971]: I1213 07:59:27.278709 4971 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-pgdb9\" (UniqueName: \"kubernetes.io/projected/2436be9d-5009-43b6-8f67-76d1ec28c83f-kube-api-access-pgdb9\") on node \"crc\" DevicePath \"\"" Dec 13 07:59:27 crc kubenswrapper[4971]: I1213 07:59:27.278743 4971 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/2436be9d-5009-43b6-8f67-76d1ec28c83f-catalog-content\") on node \"crc\" DevicePath \"\"" Dec 13 07:59:27 crc kubenswrapper[4971]: I1213 07:59:27.278752 4971 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/2436be9d-5009-43b6-8f67-76d1ec28c83f-utilities\") on node \"crc\" DevicePath \"\"" Dec 13 07:59:27 crc kubenswrapper[4971]: I1213 07:59:27.368016 4971 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-gqklt" event={"ID":"2436be9d-5009-43b6-8f67-76d1ec28c83f","Type":"ContainerDied","Data":"709d1194aec0a13f758ed0640e4a4e909a6f1e7d87a09efe5010317f70963224"} Dec 13 07:59:27 crc kubenswrapper[4971]: I1213 07:59:27.368084 4971 scope.go:117] "RemoveContainer" containerID="d697b14083a212818e76f90258f69d8b24d28ab2c46c69d29ad04a40cbcc395c" Dec 13 07:59:27 crc kubenswrapper[4971]: I1213 07:59:27.368244 4971 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-gqklt" Dec 13 07:59:27 crc kubenswrapper[4971]: I1213 07:59:27.415157 4971 scope.go:117] "RemoveContainer" containerID="0f67fcfbdf5cf7ec936b642339b9eaa172523dddb03d068cb06321d9e88e842d" Dec 13 07:59:27 crc kubenswrapper[4971]: I1213 07:59:27.444646 4971 scope.go:117] "RemoveContainer" containerID="f5146d46fde89df3c9d490af08b96b43e114368fbc0385a89dc92bf89018ccd4" Dec 13 07:59:27 crc kubenswrapper[4971]: I1213 07:59:27.444790 4971 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-operators-284t8" Dec 13 07:59:27 crc kubenswrapper[4971]: I1213 07:59:27.444830 4971 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-operators-284t8" Dec 13 07:59:27 crc kubenswrapper[4971]: I1213 07:59:27.460841 4971 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-gqklt"] Dec 13 07:59:27 crc kubenswrapper[4971]: I1213 07:59:27.487355 4971 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/community-operators-gqklt"] Dec 13 07:59:27 crc kubenswrapper[4971]: I1213 07:59:27.777275 4971 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="2436be9d-5009-43b6-8f67-76d1ec28c83f" path="/var/lib/kubelet/pods/2436be9d-5009-43b6-8f67-76d1ec28c83f/volumes" Dec 13 07:59:28 crc kubenswrapper[4971]: I1213 07:59:28.512055 4971 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-marketplace/redhat-operators-284t8" podUID="83b8c4fa-b85d-4cdb-ae08-71461f8dc452" containerName="registry-server" probeResult="failure" output=< Dec 13 07:59:28 crc kubenswrapper[4971]: timeout: failed to connect service ":50051" within 1s Dec 13 07:59:28 crc kubenswrapper[4971]: > Dec 13 07:59:37 crc kubenswrapper[4971]: I1213 07:59:37.499646 4971 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-operators-284t8" Dec 13 07:59:37 crc kubenswrapper[4971]: I1213 07:59:37.583765 4971 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-operators-284t8" Dec 13 07:59:37 crc kubenswrapper[4971]: I1213 07:59:37.778034 4971 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-284t8"] Dec 13 07:59:39 crc kubenswrapper[4971]: I1213 07:59:39.486068 4971 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-operators-284t8" podUID="83b8c4fa-b85d-4cdb-ae08-71461f8dc452" containerName="registry-server" containerID="cri-o://b9fe34225f31127367f411a45e14ea2e4cfda0298892c91dbf4db6971bf79192" gracePeriod=2 Dec 13 07:59:40 crc kubenswrapper[4971]: I1213 07:59:40.496815 4971 generic.go:334] "Generic (PLEG): container finished" podID="83b8c4fa-b85d-4cdb-ae08-71461f8dc452" containerID="b9fe34225f31127367f411a45e14ea2e4cfda0298892c91dbf4db6971bf79192" exitCode=0 Dec 13 07:59:40 crc kubenswrapper[4971]: I1213 07:59:40.497142 4971 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-284t8" event={"ID":"83b8c4fa-b85d-4cdb-ae08-71461f8dc452","Type":"ContainerDied","Data":"b9fe34225f31127367f411a45e14ea2e4cfda0298892c91dbf4db6971bf79192"} Dec 13 07:59:40 crc kubenswrapper[4971]: I1213 07:59:40.497192 4971 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-284t8" event={"ID":"83b8c4fa-b85d-4cdb-ae08-71461f8dc452","Type":"ContainerDied","Data":"054d8ea76715026ce89268541f37c9c667c09791c0adc8782ecd51587144d29b"} Dec 13 07:59:40 crc kubenswrapper[4971]: I1213 07:59:40.497204 4971 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="054d8ea76715026ce89268541f37c9c667c09791c0adc8782ecd51587144d29b" Dec 13 07:59:40 crc kubenswrapper[4971]: I1213 07:59:40.558209 4971 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-284t8" Dec 13 07:59:40 crc kubenswrapper[4971]: I1213 07:59:40.650575 4971 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-q8892\" (UniqueName: \"kubernetes.io/projected/83b8c4fa-b85d-4cdb-ae08-71461f8dc452-kube-api-access-q8892\") pod \"83b8c4fa-b85d-4cdb-ae08-71461f8dc452\" (UID: \"83b8c4fa-b85d-4cdb-ae08-71461f8dc452\") " Dec 13 07:59:40 crc kubenswrapper[4971]: I1213 07:59:40.650672 4971 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/83b8c4fa-b85d-4cdb-ae08-71461f8dc452-utilities\") pod \"83b8c4fa-b85d-4cdb-ae08-71461f8dc452\" (UID: \"83b8c4fa-b85d-4cdb-ae08-71461f8dc452\") " Dec 13 07:59:40 crc kubenswrapper[4971]: I1213 07:59:40.650690 4971 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/83b8c4fa-b85d-4cdb-ae08-71461f8dc452-catalog-content\") pod \"83b8c4fa-b85d-4cdb-ae08-71461f8dc452\" (UID: \"83b8c4fa-b85d-4cdb-ae08-71461f8dc452\") " Dec 13 07:59:40 crc kubenswrapper[4971]: I1213 07:59:40.654225 4971 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/83b8c4fa-b85d-4cdb-ae08-71461f8dc452-utilities" (OuterVolumeSpecName: "utilities") pod "83b8c4fa-b85d-4cdb-ae08-71461f8dc452" (UID: "83b8c4fa-b85d-4cdb-ae08-71461f8dc452"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 13 07:59:40 crc kubenswrapper[4971]: I1213 07:59:40.658763 4971 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/83b8c4fa-b85d-4cdb-ae08-71461f8dc452-kube-api-access-q8892" (OuterVolumeSpecName: "kube-api-access-q8892") pod "83b8c4fa-b85d-4cdb-ae08-71461f8dc452" (UID: "83b8c4fa-b85d-4cdb-ae08-71461f8dc452"). InnerVolumeSpecName "kube-api-access-q8892". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 13 07:59:40 crc kubenswrapper[4971]: I1213 07:59:40.752584 4971 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-q8892\" (UniqueName: \"kubernetes.io/projected/83b8c4fa-b85d-4cdb-ae08-71461f8dc452-kube-api-access-q8892\") on node \"crc\" DevicePath \"\"" Dec 13 07:59:40 crc kubenswrapper[4971]: I1213 07:59:40.752622 4971 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/83b8c4fa-b85d-4cdb-ae08-71461f8dc452-utilities\") on node \"crc\" DevicePath \"\"" Dec 13 07:59:40 crc kubenswrapper[4971]: I1213 07:59:40.780915 4971 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/83b8c4fa-b85d-4cdb-ae08-71461f8dc452-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "83b8c4fa-b85d-4cdb-ae08-71461f8dc452" (UID: "83b8c4fa-b85d-4cdb-ae08-71461f8dc452"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 13 07:59:40 crc kubenswrapper[4971]: I1213 07:59:40.854633 4971 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/83b8c4fa-b85d-4cdb-ae08-71461f8dc452-catalog-content\") on node \"crc\" DevicePath \"\"" Dec 13 07:59:41 crc kubenswrapper[4971]: I1213 07:59:41.504065 4971 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-284t8" Dec 13 07:59:41 crc kubenswrapper[4971]: I1213 07:59:41.568561 4971 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-284t8"] Dec 13 07:59:41 crc kubenswrapper[4971]: I1213 07:59:41.600079 4971 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-operators-284t8"] Dec 13 07:59:41 crc kubenswrapper[4971]: E1213 07:59:41.706969 4971 cadvisor_stats_provider.go:516] "Partial failure issuing cadvisor.ContainerInfoV2" err="partial failures: [\"/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod83b8c4fa_b85d_4cdb_ae08_71461f8dc452.slice/crio-054d8ea76715026ce89268541f37c9c667c09791c0adc8782ecd51587144d29b\": RecentStats: unable to find data in memory cache]" Dec 13 07:59:41 crc kubenswrapper[4971]: I1213 07:59:41.780026 4971 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="83b8c4fa-b85d-4cdb-ae08-71461f8dc452" path="/var/lib/kubelet/pods/83b8c4fa-b85d-4cdb-ae08-71461f8dc452/volumes" Dec 13 07:59:42 crc kubenswrapper[4971]: I1213 07:59:42.288196 4971 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/cinder-volume-volume1-0"] Dec 13 07:59:42 crc kubenswrapper[4971]: E1213 07:59:42.288919 4971 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="83b8c4fa-b85d-4cdb-ae08-71461f8dc452" containerName="extract-utilities" Dec 13 07:59:42 crc kubenswrapper[4971]: I1213 07:59:42.288937 4971 state_mem.go:107] "Deleted CPUSet assignment" podUID="83b8c4fa-b85d-4cdb-ae08-71461f8dc452" containerName="extract-utilities" Dec 13 07:59:42 crc kubenswrapper[4971]: E1213 07:59:42.288950 4971 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="2436be9d-5009-43b6-8f67-76d1ec28c83f" containerName="extract-utilities" Dec 13 07:59:42 crc kubenswrapper[4971]: I1213 07:59:42.288957 4971 state_mem.go:107] "Deleted CPUSet assignment" podUID="2436be9d-5009-43b6-8f67-76d1ec28c83f" containerName="extract-utilities" Dec 13 07:59:42 crc kubenswrapper[4971]: E1213 07:59:42.288970 4971 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="2436be9d-5009-43b6-8f67-76d1ec28c83f" containerName="registry-server" Dec 13 07:59:42 crc kubenswrapper[4971]: I1213 07:59:42.288976 4971 state_mem.go:107] "Deleted CPUSet assignment" podUID="2436be9d-5009-43b6-8f67-76d1ec28c83f" containerName="registry-server" Dec 13 07:59:42 crc kubenswrapper[4971]: E1213 07:59:42.288991 4971 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="e3e414b3-58e3-4da8-9542-7b6d2f229b8d" containerName="nova-custom-ceph-edpm-deployment-openstack-edpm-ipam" Dec 13 07:59:42 crc kubenswrapper[4971]: I1213 07:59:42.288998 4971 state_mem.go:107] "Deleted CPUSet assignment" podUID="e3e414b3-58e3-4da8-9542-7b6d2f229b8d" containerName="nova-custom-ceph-edpm-deployment-openstack-edpm-ipam" Dec 13 07:59:42 crc kubenswrapper[4971]: E1213 07:59:42.289006 4971 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="83b8c4fa-b85d-4cdb-ae08-71461f8dc452" containerName="registry-server" Dec 13 07:59:42 crc kubenswrapper[4971]: I1213 07:59:42.289012 4971 state_mem.go:107] "Deleted CPUSet assignment" podUID="83b8c4fa-b85d-4cdb-ae08-71461f8dc452" containerName="registry-server" Dec 13 07:59:42 crc kubenswrapper[4971]: E1213 07:59:42.289022 4971 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="83b8c4fa-b85d-4cdb-ae08-71461f8dc452" containerName="extract-content" Dec 13 07:59:42 crc kubenswrapper[4971]: I1213 07:59:42.289027 4971 state_mem.go:107] "Deleted CPUSet assignment" podUID="83b8c4fa-b85d-4cdb-ae08-71461f8dc452" containerName="extract-content" Dec 13 07:59:42 crc kubenswrapper[4971]: E1213 07:59:42.289035 4971 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="2436be9d-5009-43b6-8f67-76d1ec28c83f" containerName="extract-content" Dec 13 07:59:42 crc kubenswrapper[4971]: I1213 07:59:42.289040 4971 state_mem.go:107] "Deleted CPUSet assignment" podUID="2436be9d-5009-43b6-8f67-76d1ec28c83f" containerName="extract-content" Dec 13 07:59:42 crc kubenswrapper[4971]: I1213 07:59:42.289196 4971 memory_manager.go:354] "RemoveStaleState removing state" podUID="2436be9d-5009-43b6-8f67-76d1ec28c83f" containerName="registry-server" Dec 13 07:59:42 crc kubenswrapper[4971]: I1213 07:59:42.289214 4971 memory_manager.go:354] "RemoveStaleState removing state" podUID="e3e414b3-58e3-4da8-9542-7b6d2f229b8d" containerName="nova-custom-ceph-edpm-deployment-openstack-edpm-ipam" Dec 13 07:59:42 crc kubenswrapper[4971]: I1213 07:59:42.289224 4971 memory_manager.go:354] "RemoveStaleState removing state" podUID="83b8c4fa-b85d-4cdb-ae08-71461f8dc452" containerName="registry-server" Dec 13 07:59:42 crc kubenswrapper[4971]: I1213 07:59:42.290541 4971 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-volume-volume1-0" Dec 13 07:59:42 crc kubenswrapper[4971]: I1213 07:59:42.292578 4971 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cinder-volume-volume1-config-data" Dec 13 07:59:42 crc kubenswrapper[4971]: I1213 07:59:42.305697 4971 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceph-conf-files" Dec 13 07:59:42 crc kubenswrapper[4971]: I1213 07:59:42.306384 4971 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/cinder-backup-0"] Dec 13 07:59:42 crc kubenswrapper[4971]: I1213 07:59:42.308209 4971 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-backup-0" Dec 13 07:59:42 crc kubenswrapper[4971]: I1213 07:59:42.309946 4971 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cinder-backup-config-data" Dec 13 07:59:42 crc kubenswrapper[4971]: I1213 07:59:42.318058 4971 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cinder-volume-volume1-0"] Dec 13 07:59:42 crc kubenswrapper[4971]: I1213 07:59:42.337918 4971 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cinder-backup-0"] Dec 13 07:59:42 crc kubenswrapper[4971]: I1213 07:59:42.382213 4971 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-kgdpc\" (UniqueName: \"kubernetes.io/projected/7bbf960d-93e9-49d6-8545-28c34e36229e-kube-api-access-kgdpc\") pod \"cinder-volume-volume1-0\" (UID: \"7bbf960d-93e9-49d6-8545-28c34e36229e\") " pod="openstack/cinder-volume-volume1-0" Dec 13 07:59:42 crc kubenswrapper[4971]: I1213 07:59:42.383264 4971 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/be3cc24a-8820-4884-8f26-e2885d933527-scripts\") pod \"cinder-backup-0\" (UID: \"be3cc24a-8820-4884-8f26-e2885d933527\") " pod="openstack/cinder-backup-0" Dec 13 07:59:42 crc kubenswrapper[4971]: I1213 07:59:42.384157 4971 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-nvme\" (UniqueName: \"kubernetes.io/host-path/be3cc24a-8820-4884-8f26-e2885d933527-etc-nvme\") pod \"cinder-backup-0\" (UID: \"be3cc24a-8820-4884-8f26-e2885d933527\") " pod="openstack/cinder-backup-0" Dec 13 07:59:42 crc kubenswrapper[4971]: I1213 07:59:42.384250 4971 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"lib-modules\" (UniqueName: \"kubernetes.io/host-path/7bbf960d-93e9-49d6-8545-28c34e36229e-lib-modules\") pod \"cinder-volume-volume1-0\" (UID: \"7bbf960d-93e9-49d6-8545-28c34e36229e\") " pod="openstack/cinder-volume-volume1-0" Dec 13 07:59:42 crc kubenswrapper[4971]: I1213 07:59:42.384331 4971 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-kphkb\" (UniqueName: \"kubernetes.io/projected/be3cc24a-8820-4884-8f26-e2885d933527-kube-api-access-kphkb\") pod \"cinder-backup-0\" (UID: \"be3cc24a-8820-4884-8f26-e2885d933527\") " pod="openstack/cinder-backup-0" Dec 13 07:59:42 crc kubenswrapper[4971]: I1213 07:59:42.384399 4971 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/be3cc24a-8820-4884-8f26-e2885d933527-config-data-custom\") pod \"cinder-backup-0\" (UID: \"be3cc24a-8820-4884-8f26-e2885d933527\") " pod="openstack/cinder-backup-0" Dec 13 07:59:42 crc kubenswrapper[4971]: I1213 07:59:42.384487 4971 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dev\" (UniqueName: \"kubernetes.io/host-path/be3cc24a-8820-4884-8f26-e2885d933527-dev\") pod \"cinder-backup-0\" (UID: \"be3cc24a-8820-4884-8f26-e2885d933527\") " pod="openstack/cinder-backup-0" Dec 13 07:59:42 crc kubenswrapper[4971]: I1213 07:59:42.384585 4971 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-lib-cinder\" (UniqueName: \"kubernetes.io/host-path/be3cc24a-8820-4884-8f26-e2885d933527-var-lib-cinder\") pod \"cinder-backup-0\" (UID: \"be3cc24a-8820-4884-8f26-e2885d933527\") " pod="openstack/cinder-backup-0" Dec 13 07:59:42 crc kubenswrapper[4971]: I1213 07:59:42.384836 4971 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-nvme\" (UniqueName: \"kubernetes.io/host-path/7bbf960d-93e9-49d6-8545-28c34e36229e-etc-nvme\") pod \"cinder-volume-volume1-0\" (UID: \"7bbf960d-93e9-49d6-8545-28c34e36229e\") " pod="openstack/cinder-volume-volume1-0" Dec 13 07:59:42 crc kubenswrapper[4971]: I1213 07:59:42.384964 4971 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-iscsi\" (UniqueName: \"kubernetes.io/host-path/be3cc24a-8820-4884-8f26-e2885d933527-etc-iscsi\") pod \"cinder-backup-0\" (UID: \"be3cc24a-8820-4884-8f26-e2885d933527\") " pod="openstack/cinder-backup-0" Dec 13 07:59:42 crc kubenswrapper[4971]: I1213 07:59:42.385060 4971 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ceph\" (UniqueName: \"kubernetes.io/projected/be3cc24a-8820-4884-8f26-e2885d933527-ceph\") pod \"cinder-backup-0\" (UID: \"be3cc24a-8820-4884-8f26-e2885d933527\") " pod="openstack/cinder-backup-0" Dec 13 07:59:42 crc kubenswrapper[4971]: I1213 07:59:42.385139 4971 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"sys\" (UniqueName: \"kubernetes.io/host-path/be3cc24a-8820-4884-8f26-e2885d933527-sys\") pod \"cinder-backup-0\" (UID: \"be3cc24a-8820-4884-8f26-e2885d933527\") " pod="openstack/cinder-backup-0" Dec 13 07:59:42 crc kubenswrapper[4971]: I1213 07:59:42.385216 4971 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/be3cc24a-8820-4884-8f26-e2885d933527-etc-machine-id\") pod \"cinder-backup-0\" (UID: \"be3cc24a-8820-4884-8f26-e2885d933527\") " pod="openstack/cinder-backup-0" Dec 13 07:59:42 crc kubenswrapper[4971]: I1213 07:59:42.385286 4971 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/7bbf960d-93e9-49d6-8545-28c34e36229e-combined-ca-bundle\") pod \"cinder-volume-volume1-0\" (UID: \"7bbf960d-93e9-49d6-8545-28c34e36229e\") " pod="openstack/cinder-volume-volume1-0" Dec 13 07:59:42 crc kubenswrapper[4971]: I1213 07:59:42.385352 4971 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ceph\" (UniqueName: \"kubernetes.io/projected/7bbf960d-93e9-49d6-8545-28c34e36229e-ceph\") pod \"cinder-volume-volume1-0\" (UID: \"7bbf960d-93e9-49d6-8545-28c34e36229e\") " pod="openstack/cinder-volume-volume1-0" Dec 13 07:59:42 crc kubenswrapper[4971]: I1213 07:59:42.385419 4971 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dev\" (UniqueName: \"kubernetes.io/host-path/7bbf960d-93e9-49d6-8545-28c34e36229e-dev\") pod \"cinder-volume-volume1-0\" (UID: \"7bbf960d-93e9-49d6-8545-28c34e36229e\") " pod="openstack/cinder-volume-volume1-0" Dec 13 07:59:42 crc kubenswrapper[4971]: I1213 07:59:42.385483 4971 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-locks-cinder\" (UniqueName: \"kubernetes.io/host-path/7bbf960d-93e9-49d6-8545-28c34e36229e-var-locks-cinder\") pod \"cinder-volume-volume1-0\" (UID: \"7bbf960d-93e9-49d6-8545-28c34e36229e\") " pod="openstack/cinder-volume-volume1-0" Dec 13 07:59:42 crc kubenswrapper[4971]: I1213 07:59:42.385575 4971 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-iscsi\" (UniqueName: \"kubernetes.io/host-path/7bbf960d-93e9-49d6-8545-28c34e36229e-etc-iscsi\") pod \"cinder-volume-volume1-0\" (UID: \"7bbf960d-93e9-49d6-8545-28c34e36229e\") " pod="openstack/cinder-volume-volume1-0" Dec 13 07:59:42 crc kubenswrapper[4971]: I1213 07:59:42.385694 4971 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/be3cc24a-8820-4884-8f26-e2885d933527-config-data\") pod \"cinder-backup-0\" (UID: \"be3cc24a-8820-4884-8f26-e2885d933527\") " pod="openstack/cinder-backup-0" Dec 13 07:59:42 crc kubenswrapper[4971]: I1213 07:59:42.385785 4971 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/7bbf960d-93e9-49d6-8545-28c34e36229e-scripts\") pod \"cinder-volume-volume1-0\" (UID: \"7bbf960d-93e9-49d6-8545-28c34e36229e\") " pod="openstack/cinder-volume-volume1-0" Dec 13 07:59:42 crc kubenswrapper[4971]: I1213 07:59:42.385874 4971 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"lib-modules\" (UniqueName: \"kubernetes.io/host-path/be3cc24a-8820-4884-8f26-e2885d933527-lib-modules\") pod \"cinder-backup-0\" (UID: \"be3cc24a-8820-4884-8f26-e2885d933527\") " pod="openstack/cinder-backup-0" Dec 13 07:59:42 crc kubenswrapper[4971]: I1213 07:59:42.385971 4971 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run\" (UniqueName: \"kubernetes.io/host-path/be3cc24a-8820-4884-8f26-e2885d933527-run\") pod \"cinder-backup-0\" (UID: \"be3cc24a-8820-4884-8f26-e2885d933527\") " pod="openstack/cinder-backup-0" Dec 13 07:59:42 crc kubenswrapper[4971]: I1213 07:59:42.386055 4971 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-locks-brick\" (UniqueName: \"kubernetes.io/host-path/be3cc24a-8820-4884-8f26-e2885d933527-var-locks-brick\") pod \"cinder-backup-0\" (UID: \"be3cc24a-8820-4884-8f26-e2885d933527\") " pod="openstack/cinder-backup-0" Dec 13 07:59:42 crc kubenswrapper[4971]: I1213 07:59:42.386127 4971 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/7bbf960d-93e9-49d6-8545-28c34e36229e-etc-machine-id\") pod \"cinder-volume-volume1-0\" (UID: \"7bbf960d-93e9-49d6-8545-28c34e36229e\") " pod="openstack/cinder-volume-volume1-0" Dec 13 07:59:42 crc kubenswrapper[4971]: I1213 07:59:42.386203 4971 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/7bbf960d-93e9-49d6-8545-28c34e36229e-config-data\") pod \"cinder-volume-volume1-0\" (UID: \"7bbf960d-93e9-49d6-8545-28c34e36229e\") " pod="openstack/cinder-volume-volume1-0" Dec 13 07:59:42 crc kubenswrapper[4971]: I1213 07:59:42.386269 4971 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/be3cc24a-8820-4884-8f26-e2885d933527-combined-ca-bundle\") pod \"cinder-backup-0\" (UID: \"be3cc24a-8820-4884-8f26-e2885d933527\") " pod="openstack/cinder-backup-0" Dec 13 07:59:42 crc kubenswrapper[4971]: I1213 07:59:42.386352 4971 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-lib-cinder\" (UniqueName: \"kubernetes.io/host-path/7bbf960d-93e9-49d6-8545-28c34e36229e-var-lib-cinder\") pod \"cinder-volume-volume1-0\" (UID: \"7bbf960d-93e9-49d6-8545-28c34e36229e\") " pod="openstack/cinder-volume-volume1-0" Dec 13 07:59:42 crc kubenswrapper[4971]: I1213 07:59:42.386424 4971 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-locks-cinder\" (UniqueName: \"kubernetes.io/host-path/be3cc24a-8820-4884-8f26-e2885d933527-var-locks-cinder\") pod \"cinder-backup-0\" (UID: \"be3cc24a-8820-4884-8f26-e2885d933527\") " pod="openstack/cinder-backup-0" Dec 13 07:59:42 crc kubenswrapper[4971]: I1213 07:59:42.386538 4971 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run\" (UniqueName: \"kubernetes.io/host-path/7bbf960d-93e9-49d6-8545-28c34e36229e-run\") pod \"cinder-volume-volume1-0\" (UID: \"7bbf960d-93e9-49d6-8545-28c34e36229e\") " pod="openstack/cinder-volume-volume1-0" Dec 13 07:59:42 crc kubenswrapper[4971]: I1213 07:59:42.386622 4971 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-locks-brick\" (UniqueName: \"kubernetes.io/host-path/7bbf960d-93e9-49d6-8545-28c34e36229e-var-locks-brick\") pod \"cinder-volume-volume1-0\" (UID: \"7bbf960d-93e9-49d6-8545-28c34e36229e\") " pod="openstack/cinder-volume-volume1-0" Dec 13 07:59:42 crc kubenswrapper[4971]: I1213 07:59:42.386688 4971 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/7bbf960d-93e9-49d6-8545-28c34e36229e-config-data-custom\") pod \"cinder-volume-volume1-0\" (UID: \"7bbf960d-93e9-49d6-8545-28c34e36229e\") " pod="openstack/cinder-volume-volume1-0" Dec 13 07:59:42 crc kubenswrapper[4971]: I1213 07:59:42.386757 4971 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"sys\" (UniqueName: \"kubernetes.io/host-path/7bbf960d-93e9-49d6-8545-28c34e36229e-sys\") pod \"cinder-volume-volume1-0\" (UID: \"7bbf960d-93e9-49d6-8545-28c34e36229e\") " pod="openstack/cinder-volume-volume1-0" Dec 13 07:59:42 crc kubenswrapper[4971]: I1213 07:59:42.488508 4971 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/be3cc24a-8820-4884-8f26-e2885d933527-config-data\") pod \"cinder-backup-0\" (UID: \"be3cc24a-8820-4884-8f26-e2885d933527\") " pod="openstack/cinder-backup-0" Dec 13 07:59:42 crc kubenswrapper[4971]: I1213 07:59:42.488584 4971 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/7bbf960d-93e9-49d6-8545-28c34e36229e-scripts\") pod \"cinder-volume-volume1-0\" (UID: \"7bbf960d-93e9-49d6-8545-28c34e36229e\") " pod="openstack/cinder-volume-volume1-0" Dec 13 07:59:42 crc kubenswrapper[4971]: I1213 07:59:42.488619 4971 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"lib-modules\" (UniqueName: \"kubernetes.io/host-path/be3cc24a-8820-4884-8f26-e2885d933527-lib-modules\") pod \"cinder-backup-0\" (UID: \"be3cc24a-8820-4884-8f26-e2885d933527\") " pod="openstack/cinder-backup-0" Dec 13 07:59:42 crc kubenswrapper[4971]: I1213 07:59:42.488638 4971 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run\" (UniqueName: \"kubernetes.io/host-path/be3cc24a-8820-4884-8f26-e2885d933527-run\") pod \"cinder-backup-0\" (UID: \"be3cc24a-8820-4884-8f26-e2885d933527\") " pod="openstack/cinder-backup-0" Dec 13 07:59:42 crc kubenswrapper[4971]: I1213 07:59:42.488656 4971 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-locks-brick\" (UniqueName: \"kubernetes.io/host-path/be3cc24a-8820-4884-8f26-e2885d933527-var-locks-brick\") pod \"cinder-backup-0\" (UID: \"be3cc24a-8820-4884-8f26-e2885d933527\") " pod="openstack/cinder-backup-0" Dec 13 07:59:42 crc kubenswrapper[4971]: I1213 07:59:42.488675 4971 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/7bbf960d-93e9-49d6-8545-28c34e36229e-etc-machine-id\") pod \"cinder-volume-volume1-0\" (UID: \"7bbf960d-93e9-49d6-8545-28c34e36229e\") " pod="openstack/cinder-volume-volume1-0" Dec 13 07:59:42 crc kubenswrapper[4971]: I1213 07:59:42.488694 4971 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/7bbf960d-93e9-49d6-8545-28c34e36229e-config-data\") pod \"cinder-volume-volume1-0\" (UID: \"7bbf960d-93e9-49d6-8545-28c34e36229e\") " pod="openstack/cinder-volume-volume1-0" Dec 13 07:59:42 crc kubenswrapper[4971]: I1213 07:59:42.488715 4971 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/be3cc24a-8820-4884-8f26-e2885d933527-combined-ca-bundle\") pod \"cinder-backup-0\" (UID: \"be3cc24a-8820-4884-8f26-e2885d933527\") " pod="openstack/cinder-backup-0" Dec 13 07:59:42 crc kubenswrapper[4971]: I1213 07:59:42.488737 4971 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-lib-cinder\" (UniqueName: \"kubernetes.io/host-path/7bbf960d-93e9-49d6-8545-28c34e36229e-var-lib-cinder\") pod \"cinder-volume-volume1-0\" (UID: \"7bbf960d-93e9-49d6-8545-28c34e36229e\") " pod="openstack/cinder-volume-volume1-0" Dec 13 07:59:42 crc kubenswrapper[4971]: I1213 07:59:42.488754 4971 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-locks-cinder\" (UniqueName: \"kubernetes.io/host-path/be3cc24a-8820-4884-8f26-e2885d933527-var-locks-cinder\") pod \"cinder-backup-0\" (UID: \"be3cc24a-8820-4884-8f26-e2885d933527\") " pod="openstack/cinder-backup-0" Dec 13 07:59:42 crc kubenswrapper[4971]: I1213 07:59:42.488785 4971 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run\" (UniqueName: \"kubernetes.io/host-path/7bbf960d-93e9-49d6-8545-28c34e36229e-run\") pod \"cinder-volume-volume1-0\" (UID: \"7bbf960d-93e9-49d6-8545-28c34e36229e\") " pod="openstack/cinder-volume-volume1-0" Dec 13 07:59:42 crc kubenswrapper[4971]: I1213 07:59:42.488802 4971 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-locks-brick\" (UniqueName: \"kubernetes.io/host-path/7bbf960d-93e9-49d6-8545-28c34e36229e-var-locks-brick\") pod \"cinder-volume-volume1-0\" (UID: \"7bbf960d-93e9-49d6-8545-28c34e36229e\") " pod="openstack/cinder-volume-volume1-0" Dec 13 07:59:42 crc kubenswrapper[4971]: I1213 07:59:42.488820 4971 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/7bbf960d-93e9-49d6-8545-28c34e36229e-config-data-custom\") pod \"cinder-volume-volume1-0\" (UID: \"7bbf960d-93e9-49d6-8545-28c34e36229e\") " pod="openstack/cinder-volume-volume1-0" Dec 13 07:59:42 crc kubenswrapper[4971]: I1213 07:59:42.488835 4971 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"sys\" (UniqueName: \"kubernetes.io/host-path/7bbf960d-93e9-49d6-8545-28c34e36229e-sys\") pod \"cinder-volume-volume1-0\" (UID: \"7bbf960d-93e9-49d6-8545-28c34e36229e\") " pod="openstack/cinder-volume-volume1-0" Dec 13 07:59:42 crc kubenswrapper[4971]: I1213 07:59:42.488852 4971 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-kgdpc\" (UniqueName: \"kubernetes.io/projected/7bbf960d-93e9-49d6-8545-28c34e36229e-kube-api-access-kgdpc\") pod \"cinder-volume-volume1-0\" (UID: \"7bbf960d-93e9-49d6-8545-28c34e36229e\") " pod="openstack/cinder-volume-volume1-0" Dec 13 07:59:42 crc kubenswrapper[4971]: I1213 07:59:42.488869 4971 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/be3cc24a-8820-4884-8f26-e2885d933527-scripts\") pod \"cinder-backup-0\" (UID: \"be3cc24a-8820-4884-8f26-e2885d933527\") " pod="openstack/cinder-backup-0" Dec 13 07:59:42 crc kubenswrapper[4971]: I1213 07:59:42.488883 4971 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-nvme\" (UniqueName: \"kubernetes.io/host-path/be3cc24a-8820-4884-8f26-e2885d933527-etc-nvme\") pod \"cinder-backup-0\" (UID: \"be3cc24a-8820-4884-8f26-e2885d933527\") " pod="openstack/cinder-backup-0" Dec 13 07:59:42 crc kubenswrapper[4971]: I1213 07:59:42.488908 4971 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"lib-modules\" (UniqueName: \"kubernetes.io/host-path/7bbf960d-93e9-49d6-8545-28c34e36229e-lib-modules\") pod \"cinder-volume-volume1-0\" (UID: \"7bbf960d-93e9-49d6-8545-28c34e36229e\") " pod="openstack/cinder-volume-volume1-0" Dec 13 07:59:42 crc kubenswrapper[4971]: I1213 07:59:42.488932 4971 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/be3cc24a-8820-4884-8f26-e2885d933527-config-data-custom\") pod \"cinder-backup-0\" (UID: \"be3cc24a-8820-4884-8f26-e2885d933527\") " pod="openstack/cinder-backup-0" Dec 13 07:59:42 crc kubenswrapper[4971]: I1213 07:59:42.488949 4971 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-kphkb\" (UniqueName: \"kubernetes.io/projected/be3cc24a-8820-4884-8f26-e2885d933527-kube-api-access-kphkb\") pod \"cinder-backup-0\" (UID: \"be3cc24a-8820-4884-8f26-e2885d933527\") " pod="openstack/cinder-backup-0" Dec 13 07:59:42 crc kubenswrapper[4971]: I1213 07:59:42.488974 4971 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dev\" (UniqueName: \"kubernetes.io/host-path/be3cc24a-8820-4884-8f26-e2885d933527-dev\") pod \"cinder-backup-0\" (UID: \"be3cc24a-8820-4884-8f26-e2885d933527\") " pod="openstack/cinder-backup-0" Dec 13 07:59:42 crc kubenswrapper[4971]: I1213 07:59:42.488995 4971 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-lib-cinder\" (UniqueName: \"kubernetes.io/host-path/be3cc24a-8820-4884-8f26-e2885d933527-var-lib-cinder\") pod \"cinder-backup-0\" (UID: \"be3cc24a-8820-4884-8f26-e2885d933527\") " pod="openstack/cinder-backup-0" Dec 13 07:59:42 crc kubenswrapper[4971]: I1213 07:59:42.489010 4971 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-nvme\" (UniqueName: \"kubernetes.io/host-path/7bbf960d-93e9-49d6-8545-28c34e36229e-etc-nvme\") pod \"cinder-volume-volume1-0\" (UID: \"7bbf960d-93e9-49d6-8545-28c34e36229e\") " pod="openstack/cinder-volume-volume1-0" Dec 13 07:59:42 crc kubenswrapper[4971]: I1213 07:59:42.489041 4971 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-iscsi\" (UniqueName: \"kubernetes.io/host-path/be3cc24a-8820-4884-8f26-e2885d933527-etc-iscsi\") pod \"cinder-backup-0\" (UID: \"be3cc24a-8820-4884-8f26-e2885d933527\") " pod="openstack/cinder-backup-0" Dec 13 07:59:42 crc kubenswrapper[4971]: I1213 07:59:42.489055 4971 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ceph\" (UniqueName: \"kubernetes.io/projected/be3cc24a-8820-4884-8f26-e2885d933527-ceph\") pod \"cinder-backup-0\" (UID: \"be3cc24a-8820-4884-8f26-e2885d933527\") " pod="openstack/cinder-backup-0" Dec 13 07:59:42 crc kubenswrapper[4971]: I1213 07:59:42.489075 4971 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"sys\" (UniqueName: \"kubernetes.io/host-path/be3cc24a-8820-4884-8f26-e2885d933527-sys\") pod \"cinder-backup-0\" (UID: \"be3cc24a-8820-4884-8f26-e2885d933527\") " pod="openstack/cinder-backup-0" Dec 13 07:59:42 crc kubenswrapper[4971]: I1213 07:59:42.489094 4971 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/be3cc24a-8820-4884-8f26-e2885d933527-etc-machine-id\") pod \"cinder-backup-0\" (UID: \"be3cc24a-8820-4884-8f26-e2885d933527\") " pod="openstack/cinder-backup-0" Dec 13 07:59:42 crc kubenswrapper[4971]: I1213 07:59:42.489110 4971 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/7bbf960d-93e9-49d6-8545-28c34e36229e-combined-ca-bundle\") pod \"cinder-volume-volume1-0\" (UID: \"7bbf960d-93e9-49d6-8545-28c34e36229e\") " pod="openstack/cinder-volume-volume1-0" Dec 13 07:59:42 crc kubenswrapper[4971]: I1213 07:59:42.489127 4971 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ceph\" (UniqueName: \"kubernetes.io/projected/7bbf960d-93e9-49d6-8545-28c34e36229e-ceph\") pod \"cinder-volume-volume1-0\" (UID: \"7bbf960d-93e9-49d6-8545-28c34e36229e\") " pod="openstack/cinder-volume-volume1-0" Dec 13 07:59:42 crc kubenswrapper[4971]: I1213 07:59:42.489140 4971 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dev\" (UniqueName: \"kubernetes.io/host-path/7bbf960d-93e9-49d6-8545-28c34e36229e-dev\") pod \"cinder-volume-volume1-0\" (UID: \"7bbf960d-93e9-49d6-8545-28c34e36229e\") " pod="openstack/cinder-volume-volume1-0" Dec 13 07:59:42 crc kubenswrapper[4971]: I1213 07:59:42.489154 4971 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-locks-cinder\" (UniqueName: \"kubernetes.io/host-path/7bbf960d-93e9-49d6-8545-28c34e36229e-var-locks-cinder\") pod \"cinder-volume-volume1-0\" (UID: \"7bbf960d-93e9-49d6-8545-28c34e36229e\") " pod="openstack/cinder-volume-volume1-0" Dec 13 07:59:42 crc kubenswrapper[4971]: I1213 07:59:42.489173 4971 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-iscsi\" (UniqueName: \"kubernetes.io/host-path/7bbf960d-93e9-49d6-8545-28c34e36229e-etc-iscsi\") pod \"cinder-volume-volume1-0\" (UID: \"7bbf960d-93e9-49d6-8545-28c34e36229e\") " pod="openstack/cinder-volume-volume1-0" Dec 13 07:59:42 crc kubenswrapper[4971]: I1213 07:59:42.489181 4971 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-locks-cinder\" (UniqueName: \"kubernetes.io/host-path/be3cc24a-8820-4884-8f26-e2885d933527-var-locks-cinder\") pod \"cinder-backup-0\" (UID: \"be3cc24a-8820-4884-8f26-e2885d933527\") " pod="openstack/cinder-backup-0" Dec 13 07:59:42 crc kubenswrapper[4971]: I1213 07:59:42.489247 4971 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-iscsi\" (UniqueName: \"kubernetes.io/host-path/7bbf960d-93e9-49d6-8545-28c34e36229e-etc-iscsi\") pod \"cinder-volume-volume1-0\" (UID: \"7bbf960d-93e9-49d6-8545-28c34e36229e\") " pod="openstack/cinder-volume-volume1-0" Dec 13 07:59:42 crc kubenswrapper[4971]: I1213 07:59:42.489284 4971 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run\" (UniqueName: \"kubernetes.io/host-path/7bbf960d-93e9-49d6-8545-28c34e36229e-run\") pod \"cinder-volume-volume1-0\" (UID: \"7bbf960d-93e9-49d6-8545-28c34e36229e\") " pod="openstack/cinder-volume-volume1-0" Dec 13 07:59:42 crc kubenswrapper[4971]: I1213 07:59:42.489412 4971 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-locks-brick\" (UniqueName: \"kubernetes.io/host-path/7bbf960d-93e9-49d6-8545-28c34e36229e-var-locks-brick\") pod \"cinder-volume-volume1-0\" (UID: \"7bbf960d-93e9-49d6-8545-28c34e36229e\") " pod="openstack/cinder-volume-volume1-0" Dec 13 07:59:42 crc kubenswrapper[4971]: I1213 07:59:42.489748 4971 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/7bbf960d-93e9-49d6-8545-28c34e36229e-etc-machine-id\") pod \"cinder-volume-volume1-0\" (UID: \"7bbf960d-93e9-49d6-8545-28c34e36229e\") " pod="openstack/cinder-volume-volume1-0" Dec 13 07:59:42 crc kubenswrapper[4971]: I1213 07:59:42.489781 4971 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-lib-cinder\" (UniqueName: \"kubernetes.io/host-path/be3cc24a-8820-4884-8f26-e2885d933527-var-lib-cinder\") pod \"cinder-backup-0\" (UID: \"be3cc24a-8820-4884-8f26-e2885d933527\") " pod="openstack/cinder-backup-0" Dec 13 07:59:42 crc kubenswrapper[4971]: I1213 07:59:42.489795 4971 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-nvme\" (UniqueName: \"kubernetes.io/host-path/be3cc24a-8820-4884-8f26-e2885d933527-etc-nvme\") pod \"cinder-backup-0\" (UID: \"be3cc24a-8820-4884-8f26-e2885d933527\") " pod="openstack/cinder-backup-0" Dec 13 07:59:42 crc kubenswrapper[4971]: I1213 07:59:42.489810 4971 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"lib-modules\" (UniqueName: \"kubernetes.io/host-path/be3cc24a-8820-4884-8f26-e2885d933527-lib-modules\") pod \"cinder-backup-0\" (UID: \"be3cc24a-8820-4884-8f26-e2885d933527\") " pod="openstack/cinder-backup-0" Dec 13 07:59:42 crc kubenswrapper[4971]: I1213 07:59:42.489829 4971 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"lib-modules\" (UniqueName: \"kubernetes.io/host-path/7bbf960d-93e9-49d6-8545-28c34e36229e-lib-modules\") pod \"cinder-volume-volume1-0\" (UID: \"7bbf960d-93e9-49d6-8545-28c34e36229e\") " pod="openstack/cinder-volume-volume1-0" Dec 13 07:59:42 crc kubenswrapper[4971]: I1213 07:59:42.489892 4971 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-locks-brick\" (UniqueName: \"kubernetes.io/host-path/be3cc24a-8820-4884-8f26-e2885d933527-var-locks-brick\") pod \"cinder-backup-0\" (UID: \"be3cc24a-8820-4884-8f26-e2885d933527\") " pod="openstack/cinder-backup-0" Dec 13 07:59:42 crc kubenswrapper[4971]: I1213 07:59:42.489892 4971 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-lib-cinder\" (UniqueName: \"kubernetes.io/host-path/7bbf960d-93e9-49d6-8545-28c34e36229e-var-lib-cinder\") pod \"cinder-volume-volume1-0\" (UID: \"7bbf960d-93e9-49d6-8545-28c34e36229e\") " pod="openstack/cinder-volume-volume1-0" Dec 13 07:59:42 crc kubenswrapper[4971]: I1213 07:59:42.490016 4971 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/be3cc24a-8820-4884-8f26-e2885d933527-etc-machine-id\") pod \"cinder-backup-0\" (UID: \"be3cc24a-8820-4884-8f26-e2885d933527\") " pod="openstack/cinder-backup-0" Dec 13 07:59:42 crc kubenswrapper[4971]: I1213 07:59:42.490040 4971 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dev\" (UniqueName: \"kubernetes.io/host-path/be3cc24a-8820-4884-8f26-e2885d933527-dev\") pod \"cinder-backup-0\" (UID: \"be3cc24a-8820-4884-8f26-e2885d933527\") " pod="openstack/cinder-backup-0" Dec 13 07:59:42 crc kubenswrapper[4971]: I1213 07:59:42.490057 4971 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-nvme\" (UniqueName: \"kubernetes.io/host-path/7bbf960d-93e9-49d6-8545-28c34e36229e-etc-nvme\") pod \"cinder-volume-volume1-0\" (UID: \"7bbf960d-93e9-49d6-8545-28c34e36229e\") " pod="openstack/cinder-volume-volume1-0" Dec 13 07:59:42 crc kubenswrapper[4971]: I1213 07:59:42.490084 4971 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-iscsi\" (UniqueName: \"kubernetes.io/host-path/be3cc24a-8820-4884-8f26-e2885d933527-etc-iscsi\") pod \"cinder-backup-0\" (UID: \"be3cc24a-8820-4884-8f26-e2885d933527\") " pod="openstack/cinder-backup-0" Dec 13 07:59:42 crc kubenswrapper[4971]: I1213 07:59:42.490136 4971 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dev\" (UniqueName: \"kubernetes.io/host-path/7bbf960d-93e9-49d6-8545-28c34e36229e-dev\") pod \"cinder-volume-volume1-0\" (UID: \"7bbf960d-93e9-49d6-8545-28c34e36229e\") " pod="openstack/cinder-volume-volume1-0" Dec 13 07:59:42 crc kubenswrapper[4971]: I1213 07:59:42.490422 4971 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"sys\" (UniqueName: \"kubernetes.io/host-path/7bbf960d-93e9-49d6-8545-28c34e36229e-sys\") pod \"cinder-volume-volume1-0\" (UID: \"7bbf960d-93e9-49d6-8545-28c34e36229e\") " pod="openstack/cinder-volume-volume1-0" Dec 13 07:59:42 crc kubenswrapper[4971]: I1213 07:59:42.490458 4971 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run\" (UniqueName: \"kubernetes.io/host-path/be3cc24a-8820-4884-8f26-e2885d933527-run\") pod \"cinder-backup-0\" (UID: \"be3cc24a-8820-4884-8f26-e2885d933527\") " pod="openstack/cinder-backup-0" Dec 13 07:59:42 crc kubenswrapper[4971]: I1213 07:59:42.490466 4971 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-locks-cinder\" (UniqueName: \"kubernetes.io/host-path/7bbf960d-93e9-49d6-8545-28c34e36229e-var-locks-cinder\") pod \"cinder-volume-volume1-0\" (UID: \"7bbf960d-93e9-49d6-8545-28c34e36229e\") " pod="openstack/cinder-volume-volume1-0" Dec 13 07:59:42 crc kubenswrapper[4971]: I1213 07:59:42.490499 4971 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"sys\" (UniqueName: \"kubernetes.io/host-path/be3cc24a-8820-4884-8f26-e2885d933527-sys\") pod \"cinder-backup-0\" (UID: \"be3cc24a-8820-4884-8f26-e2885d933527\") " pod="openstack/cinder-backup-0" Dec 13 07:59:42 crc kubenswrapper[4971]: I1213 07:59:42.494932 4971 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/be3cc24a-8820-4884-8f26-e2885d933527-combined-ca-bundle\") pod \"cinder-backup-0\" (UID: \"be3cc24a-8820-4884-8f26-e2885d933527\") " pod="openstack/cinder-backup-0" Dec 13 07:59:42 crc kubenswrapper[4971]: I1213 07:59:42.495642 4971 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/be3cc24a-8820-4884-8f26-e2885d933527-scripts\") pod \"cinder-backup-0\" (UID: \"be3cc24a-8820-4884-8f26-e2885d933527\") " pod="openstack/cinder-backup-0" Dec 13 07:59:42 crc kubenswrapper[4971]: I1213 07:59:42.496852 4971 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/7bbf960d-93e9-49d6-8545-28c34e36229e-config-data\") pod \"cinder-volume-volume1-0\" (UID: \"7bbf960d-93e9-49d6-8545-28c34e36229e\") " pod="openstack/cinder-volume-volume1-0" Dec 13 07:59:42 crc kubenswrapper[4971]: I1213 07:59:42.499078 4971 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/7bbf960d-93e9-49d6-8545-28c34e36229e-combined-ca-bundle\") pod \"cinder-volume-volume1-0\" (UID: \"7bbf960d-93e9-49d6-8545-28c34e36229e\") " pod="openstack/cinder-volume-volume1-0" Dec 13 07:59:42 crc kubenswrapper[4971]: I1213 07:59:42.501161 4971 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/7bbf960d-93e9-49d6-8545-28c34e36229e-config-data-custom\") pod \"cinder-volume-volume1-0\" (UID: \"7bbf960d-93e9-49d6-8545-28c34e36229e\") " pod="openstack/cinder-volume-volume1-0" Dec 13 07:59:42 crc kubenswrapper[4971]: I1213 07:59:42.501443 4971 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/be3cc24a-8820-4884-8f26-e2885d933527-config-data\") pod \"cinder-backup-0\" (UID: \"be3cc24a-8820-4884-8f26-e2885d933527\") " pod="openstack/cinder-backup-0" Dec 13 07:59:42 crc kubenswrapper[4971]: I1213 07:59:42.502265 4971 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/7bbf960d-93e9-49d6-8545-28c34e36229e-scripts\") pod \"cinder-volume-volume1-0\" (UID: \"7bbf960d-93e9-49d6-8545-28c34e36229e\") " pod="openstack/cinder-volume-volume1-0" Dec 13 07:59:42 crc kubenswrapper[4971]: I1213 07:59:42.502372 4971 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/be3cc24a-8820-4884-8f26-e2885d933527-config-data-custom\") pod \"cinder-backup-0\" (UID: \"be3cc24a-8820-4884-8f26-e2885d933527\") " pod="openstack/cinder-backup-0" Dec 13 07:59:42 crc kubenswrapper[4971]: I1213 07:59:42.502843 4971 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ceph\" (UniqueName: \"kubernetes.io/projected/be3cc24a-8820-4884-8f26-e2885d933527-ceph\") pod \"cinder-backup-0\" (UID: \"be3cc24a-8820-4884-8f26-e2885d933527\") " pod="openstack/cinder-backup-0" Dec 13 07:59:42 crc kubenswrapper[4971]: I1213 07:59:42.506124 4971 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ceph\" (UniqueName: \"kubernetes.io/projected/7bbf960d-93e9-49d6-8545-28c34e36229e-ceph\") pod \"cinder-volume-volume1-0\" (UID: \"7bbf960d-93e9-49d6-8545-28c34e36229e\") " pod="openstack/cinder-volume-volume1-0" Dec 13 07:59:42 crc kubenswrapper[4971]: I1213 07:59:42.513441 4971 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-kphkb\" (UniqueName: \"kubernetes.io/projected/be3cc24a-8820-4884-8f26-e2885d933527-kube-api-access-kphkb\") pod \"cinder-backup-0\" (UID: \"be3cc24a-8820-4884-8f26-e2885d933527\") " pod="openstack/cinder-backup-0" Dec 13 07:59:42 crc kubenswrapper[4971]: I1213 07:59:42.529619 4971 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-kgdpc\" (UniqueName: \"kubernetes.io/projected/7bbf960d-93e9-49d6-8545-28c34e36229e-kube-api-access-kgdpc\") pod \"cinder-volume-volume1-0\" (UID: \"7bbf960d-93e9-49d6-8545-28c34e36229e\") " pod="openstack/cinder-volume-volume1-0" Dec 13 07:59:42 crc kubenswrapper[4971]: I1213 07:59:42.610372 4971 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-volume-volume1-0" Dec 13 07:59:42 crc kubenswrapper[4971]: I1213 07:59:42.624345 4971 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-backup-0" Dec 13 07:59:43 crc kubenswrapper[4971]: I1213 07:59:43.102185 4971 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/manila-db-create-wmjhn"] Dec 13 07:59:43 crc kubenswrapper[4971]: I1213 07:59:43.103556 4971 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/manila-db-create-wmjhn" Dec 13 07:59:43 crc kubenswrapper[4971]: I1213 07:59:43.113814 4971 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/manila-db-create-wmjhn"] Dec 13 07:59:43 crc kubenswrapper[4971]: I1213 07:59:43.188590 4971 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/glance-default-external-api-0"] Dec 13 07:59:43 crc kubenswrapper[4971]: I1213 07:59:43.190845 4971 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-external-api-0" Dec 13 07:59:43 crc kubenswrapper[4971]: I1213 07:59:43.192909 4971 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"glance-default-external-config-data" Dec 13 07:59:43 crc kubenswrapper[4971]: I1213 07:59:43.193191 4971 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"glance-glance-dockercfg-k864m" Dec 13 07:59:43 crc kubenswrapper[4971]: I1213 07:59:43.193420 4971 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"glance-scripts" Dec 13 07:59:43 crc kubenswrapper[4971]: I1213 07:59:43.193573 4971 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-glance-default-public-svc" Dec 13 07:59:43 crc kubenswrapper[4971]: I1213 07:59:43.207576 4971 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-default-external-api-0"] Dec 13 07:59:43 crc kubenswrapper[4971]: I1213 07:59:43.212624 4971 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/44d3eebd-3532-454e-b687-dd752f549afe-operator-scripts\") pod \"manila-db-create-wmjhn\" (UID: \"44d3eebd-3532-454e-b687-dd752f549afe\") " pod="openstack/manila-db-create-wmjhn" Dec 13 07:59:43 crc kubenswrapper[4971]: I1213 07:59:43.212677 4971 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-rcdq2\" (UniqueName: \"kubernetes.io/projected/44d3eebd-3532-454e-b687-dd752f549afe-kube-api-access-rcdq2\") pod \"manila-db-create-wmjhn\" (UID: \"44d3eebd-3532-454e-b687-dd752f549afe\") " pod="openstack/manila-db-create-wmjhn" Dec 13 07:59:43 crc kubenswrapper[4971]: I1213 07:59:43.220580 4971 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/glance-default-internal-api-0"] Dec 13 07:59:43 crc kubenswrapper[4971]: I1213 07:59:43.222213 4971 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-internal-api-0" Dec 13 07:59:43 crc kubenswrapper[4971]: I1213 07:59:43.226916 4971 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"glance-default-internal-config-data" Dec 13 07:59:43 crc kubenswrapper[4971]: I1213 07:59:43.227414 4971 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-glance-default-internal-svc" Dec 13 07:59:43 crc kubenswrapper[4971]: I1213 07:59:43.247649 4971 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-default-internal-api-0"] Dec 13 07:59:43 crc kubenswrapper[4971]: I1213 07:59:43.315056 4971 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/3158ebe6-4c8c-41d1-bf95-6b4492a7e0aa-config-data\") pod \"glance-default-internal-api-0\" (UID: \"3158ebe6-4c8c-41d1-bf95-6b4492a7e0aa\") " pod="openstack/glance-default-internal-api-0" Dec 13 07:59:43 crc kubenswrapper[4971]: I1213 07:59:43.315110 4971 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-7gslt\" (UniqueName: \"kubernetes.io/projected/9fbf6500-bf43-4a4c-820c-14a60a27afb9-kube-api-access-7gslt\") pod \"glance-default-external-api-0\" (UID: \"9fbf6500-bf43-4a4c-820c-14a60a27afb9\") " pod="openstack/glance-default-external-api-0" Dec 13 07:59:43 crc kubenswrapper[4971]: I1213 07:59:43.315149 4971 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/3158ebe6-4c8c-41d1-bf95-6b4492a7e0aa-combined-ca-bundle\") pod \"glance-default-internal-api-0\" (UID: \"3158ebe6-4c8c-41d1-bf95-6b4492a7e0aa\") " pod="openstack/glance-default-internal-api-0" Dec 13 07:59:43 crc kubenswrapper[4971]: I1213 07:59:43.315181 4971 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/9fbf6500-bf43-4a4c-820c-14a60a27afb9-logs\") pod \"glance-default-external-api-0\" (UID: \"9fbf6500-bf43-4a4c-820c-14a60a27afb9\") " pod="openstack/glance-default-external-api-0" Dec 13 07:59:43 crc kubenswrapper[4971]: I1213 07:59:43.315281 4971 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/9fbf6500-bf43-4a4c-820c-14a60a27afb9-public-tls-certs\") pod \"glance-default-external-api-0\" (UID: \"9fbf6500-bf43-4a4c-820c-14a60a27afb9\") " pod="openstack/glance-default-external-api-0" Dec 13 07:59:43 crc kubenswrapper[4971]: I1213 07:59:43.315318 4971 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/9fbf6500-bf43-4a4c-820c-14a60a27afb9-httpd-run\") pod \"glance-default-external-api-0\" (UID: \"9fbf6500-bf43-4a4c-820c-14a60a27afb9\") " pod="openstack/glance-default-external-api-0" Dec 13 07:59:43 crc kubenswrapper[4971]: I1213 07:59:43.315354 4971 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ceph\" (UniqueName: \"kubernetes.io/projected/3158ebe6-4c8c-41d1-bf95-6b4492a7e0aa-ceph\") pod \"glance-default-internal-api-0\" (UID: \"3158ebe6-4c8c-41d1-bf95-6b4492a7e0aa\") " pod="openstack/glance-default-internal-api-0" Dec 13 07:59:43 crc kubenswrapper[4971]: I1213 07:59:43.315383 4971 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/44d3eebd-3532-454e-b687-dd752f549afe-operator-scripts\") pod \"manila-db-create-wmjhn\" (UID: \"44d3eebd-3532-454e-b687-dd752f549afe\") " pod="openstack/manila-db-create-wmjhn" Dec 13 07:59:43 crc kubenswrapper[4971]: I1213 07:59:43.315422 4971 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-rcdq2\" (UniqueName: \"kubernetes.io/projected/44d3eebd-3532-454e-b687-dd752f549afe-kube-api-access-rcdq2\") pod \"manila-db-create-wmjhn\" (UID: \"44d3eebd-3532-454e-b687-dd752f549afe\") " pod="openstack/manila-db-create-wmjhn" Dec 13 07:59:43 crc kubenswrapper[4971]: I1213 07:59:43.315453 4971 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/3158ebe6-4c8c-41d1-bf95-6b4492a7e0aa-internal-tls-certs\") pod \"glance-default-internal-api-0\" (UID: \"3158ebe6-4c8c-41d1-bf95-6b4492a7e0aa\") " pod="openstack/glance-default-internal-api-0" Dec 13 07:59:43 crc kubenswrapper[4971]: I1213 07:59:43.315477 4971 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/9fbf6500-bf43-4a4c-820c-14a60a27afb9-config-data\") pod \"glance-default-external-api-0\" (UID: \"9fbf6500-bf43-4a4c-820c-14a60a27afb9\") " pod="openstack/glance-default-external-api-0" Dec 13 07:59:43 crc kubenswrapper[4971]: I1213 07:59:43.315531 4971 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/3158ebe6-4c8c-41d1-bf95-6b4492a7e0aa-scripts\") pod \"glance-default-internal-api-0\" (UID: \"3158ebe6-4c8c-41d1-bf95-6b4492a7e0aa\") " pod="openstack/glance-default-internal-api-0" Dec 13 07:59:43 crc kubenswrapper[4971]: I1213 07:59:43.315573 4971 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-vmjq8\" (UniqueName: \"kubernetes.io/projected/3158ebe6-4c8c-41d1-bf95-6b4492a7e0aa-kube-api-access-vmjq8\") pod \"glance-default-internal-api-0\" (UID: \"3158ebe6-4c8c-41d1-bf95-6b4492a7e0aa\") " pod="openstack/glance-default-internal-api-0" Dec 13 07:59:43 crc kubenswrapper[4971]: I1213 07:59:43.315604 4971 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/9fbf6500-bf43-4a4c-820c-14a60a27afb9-combined-ca-bundle\") pod \"glance-default-external-api-0\" (UID: \"9fbf6500-bf43-4a4c-820c-14a60a27afb9\") " pod="openstack/glance-default-external-api-0" Dec 13 07:59:43 crc kubenswrapper[4971]: I1213 07:59:43.315628 4971 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"local-storage11-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage11-crc\") pod \"glance-default-external-api-0\" (UID: \"9fbf6500-bf43-4a4c-820c-14a60a27afb9\") " pod="openstack/glance-default-external-api-0" Dec 13 07:59:43 crc kubenswrapper[4971]: I1213 07:59:43.315759 4971 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ceph\" (UniqueName: \"kubernetes.io/projected/9fbf6500-bf43-4a4c-820c-14a60a27afb9-ceph\") pod \"glance-default-external-api-0\" (UID: \"9fbf6500-bf43-4a4c-820c-14a60a27afb9\") " pod="openstack/glance-default-external-api-0" Dec 13 07:59:43 crc kubenswrapper[4971]: I1213 07:59:43.315857 4971 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"local-storage01-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage01-crc\") pod \"glance-default-internal-api-0\" (UID: \"3158ebe6-4c8c-41d1-bf95-6b4492a7e0aa\") " pod="openstack/glance-default-internal-api-0" Dec 13 07:59:43 crc kubenswrapper[4971]: I1213 07:59:43.315892 4971 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/9fbf6500-bf43-4a4c-820c-14a60a27afb9-scripts\") pod \"glance-default-external-api-0\" (UID: \"9fbf6500-bf43-4a4c-820c-14a60a27afb9\") " pod="openstack/glance-default-external-api-0" Dec 13 07:59:43 crc kubenswrapper[4971]: I1213 07:59:43.316005 4971 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/3158ebe6-4c8c-41d1-bf95-6b4492a7e0aa-httpd-run\") pod \"glance-default-internal-api-0\" (UID: \"3158ebe6-4c8c-41d1-bf95-6b4492a7e0aa\") " pod="openstack/glance-default-internal-api-0" Dec 13 07:59:43 crc kubenswrapper[4971]: I1213 07:59:43.316070 4971 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/3158ebe6-4c8c-41d1-bf95-6b4492a7e0aa-logs\") pod \"glance-default-internal-api-0\" (UID: \"3158ebe6-4c8c-41d1-bf95-6b4492a7e0aa\") " pod="openstack/glance-default-internal-api-0" Dec 13 07:59:43 crc kubenswrapper[4971]: I1213 07:59:43.316126 4971 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/44d3eebd-3532-454e-b687-dd752f549afe-operator-scripts\") pod \"manila-db-create-wmjhn\" (UID: \"44d3eebd-3532-454e-b687-dd752f549afe\") " pod="openstack/manila-db-create-wmjhn" Dec 13 07:59:43 crc kubenswrapper[4971]: I1213 07:59:43.328253 4971 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/manila-8949-account-create-update-6pkvl"] Dec 13 07:59:43 crc kubenswrapper[4971]: I1213 07:59:43.329370 4971 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/manila-8949-account-create-update-6pkvl" Dec 13 07:59:43 crc kubenswrapper[4971]: I1213 07:59:43.336727 4971 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"manila-db-secret" Dec 13 07:59:43 crc kubenswrapper[4971]: I1213 07:59:43.337212 4971 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-rcdq2\" (UniqueName: \"kubernetes.io/projected/44d3eebd-3532-454e-b687-dd752f549afe-kube-api-access-rcdq2\") pod \"manila-db-create-wmjhn\" (UID: \"44d3eebd-3532-454e-b687-dd752f549afe\") " pod="openstack/manila-db-create-wmjhn" Dec 13 07:59:43 crc kubenswrapper[4971]: I1213 07:59:43.343352 4971 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cinder-volume-volume1-0"] Dec 13 07:59:43 crc kubenswrapper[4971]: I1213 07:59:43.354844 4971 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/manila-8949-account-create-update-6pkvl"] Dec 13 07:59:43 crc kubenswrapper[4971]: W1213 07:59:43.402710 4971 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podbe3cc24a_8820_4884_8f26_e2885d933527.slice/crio-dcf26701dd7d956bb6c52af205b2331096dbdc4067d0986ebfd0e18980181462 WatchSource:0}: Error finding container dcf26701dd7d956bb6c52af205b2331096dbdc4067d0986ebfd0e18980181462: Status 404 returned error can't find the container with id dcf26701dd7d956bb6c52af205b2331096dbdc4067d0986ebfd0e18980181462 Dec 13 07:59:43 crc kubenswrapper[4971]: I1213 07:59:43.404295 4971 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cinder-backup-0"] Dec 13 07:59:43 crc kubenswrapper[4971]: I1213 07:59:43.417539 4971 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/3158ebe6-4c8c-41d1-bf95-6b4492a7e0aa-scripts\") pod \"glance-default-internal-api-0\" (UID: \"3158ebe6-4c8c-41d1-bf95-6b4492a7e0aa\") " pod="openstack/glance-default-internal-api-0" Dec 13 07:59:43 crc kubenswrapper[4971]: I1213 07:59:43.417597 4971 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-7vx49\" (UniqueName: \"kubernetes.io/projected/ed7c6200-6e0f-4865-81a8-7eb5c39a996f-kube-api-access-7vx49\") pod \"manila-8949-account-create-update-6pkvl\" (UID: \"ed7c6200-6e0f-4865-81a8-7eb5c39a996f\") " pod="openstack/manila-8949-account-create-update-6pkvl" Dec 13 07:59:43 crc kubenswrapper[4971]: I1213 07:59:43.417628 4971 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-vmjq8\" (UniqueName: \"kubernetes.io/projected/3158ebe6-4c8c-41d1-bf95-6b4492a7e0aa-kube-api-access-vmjq8\") pod \"glance-default-internal-api-0\" (UID: \"3158ebe6-4c8c-41d1-bf95-6b4492a7e0aa\") " pod="openstack/glance-default-internal-api-0" Dec 13 07:59:43 crc kubenswrapper[4971]: I1213 07:59:43.417648 4971 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/9fbf6500-bf43-4a4c-820c-14a60a27afb9-combined-ca-bundle\") pod \"glance-default-external-api-0\" (UID: \"9fbf6500-bf43-4a4c-820c-14a60a27afb9\") " pod="openstack/glance-default-external-api-0" Dec 13 07:59:43 crc kubenswrapper[4971]: I1213 07:59:43.417667 4971 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"local-storage11-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage11-crc\") pod \"glance-default-external-api-0\" (UID: \"9fbf6500-bf43-4a4c-820c-14a60a27afb9\") " pod="openstack/glance-default-external-api-0" Dec 13 07:59:43 crc kubenswrapper[4971]: I1213 07:59:43.417699 4971 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ceph\" (UniqueName: \"kubernetes.io/projected/9fbf6500-bf43-4a4c-820c-14a60a27afb9-ceph\") pod \"glance-default-external-api-0\" (UID: \"9fbf6500-bf43-4a4c-820c-14a60a27afb9\") " pod="openstack/glance-default-external-api-0" Dec 13 07:59:43 crc kubenswrapper[4971]: I1213 07:59:43.417904 4971 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"local-storage01-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage01-crc\") pod \"glance-default-internal-api-0\" (UID: \"3158ebe6-4c8c-41d1-bf95-6b4492a7e0aa\") " pod="openstack/glance-default-internal-api-0" Dec 13 07:59:43 crc kubenswrapper[4971]: I1213 07:59:43.417922 4971 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/9fbf6500-bf43-4a4c-820c-14a60a27afb9-scripts\") pod \"glance-default-external-api-0\" (UID: \"9fbf6500-bf43-4a4c-820c-14a60a27afb9\") " pod="openstack/glance-default-external-api-0" Dec 13 07:59:43 crc kubenswrapper[4971]: I1213 07:59:43.417963 4971 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/3158ebe6-4c8c-41d1-bf95-6b4492a7e0aa-httpd-run\") pod \"glance-default-internal-api-0\" (UID: \"3158ebe6-4c8c-41d1-bf95-6b4492a7e0aa\") " pod="openstack/glance-default-internal-api-0" Dec 13 07:59:43 crc kubenswrapper[4971]: I1213 07:59:43.417981 4971 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/3158ebe6-4c8c-41d1-bf95-6b4492a7e0aa-logs\") pod \"glance-default-internal-api-0\" (UID: \"3158ebe6-4c8c-41d1-bf95-6b4492a7e0aa\") " pod="openstack/glance-default-internal-api-0" Dec 13 07:59:43 crc kubenswrapper[4971]: I1213 07:59:43.418013 4971 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/3158ebe6-4c8c-41d1-bf95-6b4492a7e0aa-config-data\") pod \"glance-default-internal-api-0\" (UID: \"3158ebe6-4c8c-41d1-bf95-6b4492a7e0aa\") " pod="openstack/glance-default-internal-api-0" Dec 13 07:59:43 crc kubenswrapper[4971]: I1213 07:59:43.418032 4971 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-7gslt\" (UniqueName: \"kubernetes.io/projected/9fbf6500-bf43-4a4c-820c-14a60a27afb9-kube-api-access-7gslt\") pod \"glance-default-external-api-0\" (UID: \"9fbf6500-bf43-4a4c-820c-14a60a27afb9\") " pod="openstack/glance-default-external-api-0" Dec 13 07:59:43 crc kubenswrapper[4971]: I1213 07:59:43.418050 4971 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/3158ebe6-4c8c-41d1-bf95-6b4492a7e0aa-combined-ca-bundle\") pod \"glance-default-internal-api-0\" (UID: \"3158ebe6-4c8c-41d1-bf95-6b4492a7e0aa\") " pod="openstack/glance-default-internal-api-0" Dec 13 07:59:43 crc kubenswrapper[4971]: I1213 07:59:43.418067 4971 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/9fbf6500-bf43-4a4c-820c-14a60a27afb9-logs\") pod \"glance-default-external-api-0\" (UID: \"9fbf6500-bf43-4a4c-820c-14a60a27afb9\") " pod="openstack/glance-default-external-api-0" Dec 13 07:59:43 crc kubenswrapper[4971]: I1213 07:59:43.418104 4971 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/9fbf6500-bf43-4a4c-820c-14a60a27afb9-public-tls-certs\") pod \"glance-default-external-api-0\" (UID: \"9fbf6500-bf43-4a4c-820c-14a60a27afb9\") " pod="openstack/glance-default-external-api-0" Dec 13 07:59:43 crc kubenswrapper[4971]: I1213 07:59:43.418123 4971 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/ed7c6200-6e0f-4865-81a8-7eb5c39a996f-operator-scripts\") pod \"manila-8949-account-create-update-6pkvl\" (UID: \"ed7c6200-6e0f-4865-81a8-7eb5c39a996f\") " pod="openstack/manila-8949-account-create-update-6pkvl" Dec 13 07:59:43 crc kubenswrapper[4971]: I1213 07:59:43.418148 4971 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/9fbf6500-bf43-4a4c-820c-14a60a27afb9-httpd-run\") pod \"glance-default-external-api-0\" (UID: \"9fbf6500-bf43-4a4c-820c-14a60a27afb9\") " pod="openstack/glance-default-external-api-0" Dec 13 07:59:43 crc kubenswrapper[4971]: I1213 07:59:43.418174 4971 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ceph\" (UniqueName: \"kubernetes.io/projected/3158ebe6-4c8c-41d1-bf95-6b4492a7e0aa-ceph\") pod \"glance-default-internal-api-0\" (UID: \"3158ebe6-4c8c-41d1-bf95-6b4492a7e0aa\") " pod="openstack/glance-default-internal-api-0" Dec 13 07:59:43 crc kubenswrapper[4971]: I1213 07:59:43.418193 4971 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/3158ebe6-4c8c-41d1-bf95-6b4492a7e0aa-internal-tls-certs\") pod \"glance-default-internal-api-0\" (UID: \"3158ebe6-4c8c-41d1-bf95-6b4492a7e0aa\") " pod="openstack/glance-default-internal-api-0" Dec 13 07:59:43 crc kubenswrapper[4971]: I1213 07:59:43.418209 4971 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/9fbf6500-bf43-4a4c-820c-14a60a27afb9-config-data\") pod \"glance-default-external-api-0\" (UID: \"9fbf6500-bf43-4a4c-820c-14a60a27afb9\") " pod="openstack/glance-default-external-api-0" Dec 13 07:59:43 crc kubenswrapper[4971]: I1213 07:59:43.418481 4971 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"local-storage11-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage11-crc\") pod \"glance-default-external-api-0\" (UID: \"9fbf6500-bf43-4a4c-820c-14a60a27afb9\") device mount path \"/mnt/openstack/pv11\"" pod="openstack/glance-default-external-api-0" Dec 13 07:59:43 crc kubenswrapper[4971]: I1213 07:59:43.418567 4971 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"local-storage01-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage01-crc\") pod \"glance-default-internal-api-0\" (UID: \"3158ebe6-4c8c-41d1-bf95-6b4492a7e0aa\") device mount path \"/mnt/openstack/pv01\"" pod="openstack/glance-default-internal-api-0" Dec 13 07:59:43 crc kubenswrapper[4971]: I1213 07:59:43.419380 4971 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/3158ebe6-4c8c-41d1-bf95-6b4492a7e0aa-logs\") pod \"glance-default-internal-api-0\" (UID: \"3158ebe6-4c8c-41d1-bf95-6b4492a7e0aa\") " pod="openstack/glance-default-internal-api-0" Dec 13 07:59:43 crc kubenswrapper[4971]: I1213 07:59:43.419415 4971 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/3158ebe6-4c8c-41d1-bf95-6b4492a7e0aa-httpd-run\") pod \"glance-default-internal-api-0\" (UID: \"3158ebe6-4c8c-41d1-bf95-6b4492a7e0aa\") " pod="openstack/glance-default-internal-api-0" Dec 13 07:59:43 crc kubenswrapper[4971]: I1213 07:59:43.419852 4971 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/9fbf6500-bf43-4a4c-820c-14a60a27afb9-httpd-run\") pod \"glance-default-external-api-0\" (UID: \"9fbf6500-bf43-4a4c-820c-14a60a27afb9\") " pod="openstack/glance-default-external-api-0" Dec 13 07:59:43 crc kubenswrapper[4971]: I1213 07:59:43.420657 4971 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/9fbf6500-bf43-4a4c-820c-14a60a27afb9-logs\") pod \"glance-default-external-api-0\" (UID: \"9fbf6500-bf43-4a4c-820c-14a60a27afb9\") " pod="openstack/glance-default-external-api-0" Dec 13 07:59:43 crc kubenswrapper[4971]: I1213 07:59:43.422806 4971 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/9fbf6500-bf43-4a4c-820c-14a60a27afb9-scripts\") pod \"glance-default-external-api-0\" (UID: \"9fbf6500-bf43-4a4c-820c-14a60a27afb9\") " pod="openstack/glance-default-external-api-0" Dec 13 07:59:43 crc kubenswrapper[4971]: I1213 07:59:43.422889 4971 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ceph\" (UniqueName: \"kubernetes.io/projected/9fbf6500-bf43-4a4c-820c-14a60a27afb9-ceph\") pod \"glance-default-external-api-0\" (UID: \"9fbf6500-bf43-4a4c-820c-14a60a27afb9\") " pod="openstack/glance-default-external-api-0" Dec 13 07:59:43 crc kubenswrapper[4971]: I1213 07:59:43.423244 4971 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/9fbf6500-bf43-4a4c-820c-14a60a27afb9-combined-ca-bundle\") pod \"glance-default-external-api-0\" (UID: \"9fbf6500-bf43-4a4c-820c-14a60a27afb9\") " pod="openstack/glance-default-external-api-0" Dec 13 07:59:43 crc kubenswrapper[4971]: I1213 07:59:43.423730 4971 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/3158ebe6-4c8c-41d1-bf95-6b4492a7e0aa-scripts\") pod \"glance-default-internal-api-0\" (UID: \"3158ebe6-4c8c-41d1-bf95-6b4492a7e0aa\") " pod="openstack/glance-default-internal-api-0" Dec 13 07:59:43 crc kubenswrapper[4971]: I1213 07:59:43.423923 4971 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/9fbf6500-bf43-4a4c-820c-14a60a27afb9-config-data\") pod \"glance-default-external-api-0\" (UID: \"9fbf6500-bf43-4a4c-820c-14a60a27afb9\") " pod="openstack/glance-default-external-api-0" Dec 13 07:59:43 crc kubenswrapper[4971]: I1213 07:59:43.426450 4971 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/3158ebe6-4c8c-41d1-bf95-6b4492a7e0aa-config-data\") pod \"glance-default-internal-api-0\" (UID: \"3158ebe6-4c8c-41d1-bf95-6b4492a7e0aa\") " pod="openstack/glance-default-internal-api-0" Dec 13 07:59:43 crc kubenswrapper[4971]: I1213 07:59:43.426465 4971 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ceph\" (UniqueName: \"kubernetes.io/projected/3158ebe6-4c8c-41d1-bf95-6b4492a7e0aa-ceph\") pod \"glance-default-internal-api-0\" (UID: \"3158ebe6-4c8c-41d1-bf95-6b4492a7e0aa\") " pod="openstack/glance-default-internal-api-0" Dec 13 07:59:43 crc kubenswrapper[4971]: I1213 07:59:43.428026 4971 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/9fbf6500-bf43-4a4c-820c-14a60a27afb9-public-tls-certs\") pod \"glance-default-external-api-0\" (UID: \"9fbf6500-bf43-4a4c-820c-14a60a27afb9\") " pod="openstack/glance-default-external-api-0" Dec 13 07:59:43 crc kubenswrapper[4971]: I1213 07:59:43.428391 4971 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/3158ebe6-4c8c-41d1-bf95-6b4492a7e0aa-combined-ca-bundle\") pod \"glance-default-internal-api-0\" (UID: \"3158ebe6-4c8c-41d1-bf95-6b4492a7e0aa\") " pod="openstack/glance-default-internal-api-0" Dec 13 07:59:43 crc kubenswrapper[4971]: I1213 07:59:43.429785 4971 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/3158ebe6-4c8c-41d1-bf95-6b4492a7e0aa-internal-tls-certs\") pod \"glance-default-internal-api-0\" (UID: \"3158ebe6-4c8c-41d1-bf95-6b4492a7e0aa\") " pod="openstack/glance-default-internal-api-0" Dec 13 07:59:43 crc kubenswrapper[4971]: I1213 07:59:43.435075 4971 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/manila-db-create-wmjhn" Dec 13 07:59:43 crc kubenswrapper[4971]: I1213 07:59:43.435965 4971 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-vmjq8\" (UniqueName: \"kubernetes.io/projected/3158ebe6-4c8c-41d1-bf95-6b4492a7e0aa-kube-api-access-vmjq8\") pod \"glance-default-internal-api-0\" (UID: \"3158ebe6-4c8c-41d1-bf95-6b4492a7e0aa\") " pod="openstack/glance-default-internal-api-0" Dec 13 07:59:43 crc kubenswrapper[4971]: I1213 07:59:43.436951 4971 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-7gslt\" (UniqueName: \"kubernetes.io/projected/9fbf6500-bf43-4a4c-820c-14a60a27afb9-kube-api-access-7gslt\") pod \"glance-default-external-api-0\" (UID: \"9fbf6500-bf43-4a4c-820c-14a60a27afb9\") " pod="openstack/glance-default-external-api-0" Dec 13 07:59:43 crc kubenswrapper[4971]: I1213 07:59:43.463798 4971 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"local-storage01-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage01-crc\") pod \"glance-default-internal-api-0\" (UID: \"3158ebe6-4c8c-41d1-bf95-6b4492a7e0aa\") " pod="openstack/glance-default-internal-api-0" Dec 13 07:59:43 crc kubenswrapper[4971]: I1213 07:59:43.472984 4971 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"local-storage11-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage11-crc\") pod \"glance-default-external-api-0\" (UID: \"9fbf6500-bf43-4a4c-820c-14a60a27afb9\") " pod="openstack/glance-default-external-api-0" Dec 13 07:59:43 crc kubenswrapper[4971]: I1213 07:59:43.504053 4971 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-external-api-0" Dec 13 07:59:43 crc kubenswrapper[4971]: I1213 07:59:43.519798 4971 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/ed7c6200-6e0f-4865-81a8-7eb5c39a996f-operator-scripts\") pod \"manila-8949-account-create-update-6pkvl\" (UID: \"ed7c6200-6e0f-4865-81a8-7eb5c39a996f\") " pod="openstack/manila-8949-account-create-update-6pkvl" Dec 13 07:59:43 crc kubenswrapper[4971]: I1213 07:59:43.520243 4971 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-7vx49\" (UniqueName: \"kubernetes.io/projected/ed7c6200-6e0f-4865-81a8-7eb5c39a996f-kube-api-access-7vx49\") pod \"manila-8949-account-create-update-6pkvl\" (UID: \"ed7c6200-6e0f-4865-81a8-7eb5c39a996f\") " pod="openstack/manila-8949-account-create-update-6pkvl" Dec 13 07:59:43 crc kubenswrapper[4971]: I1213 07:59:43.521407 4971 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/ed7c6200-6e0f-4865-81a8-7eb5c39a996f-operator-scripts\") pod \"manila-8949-account-create-update-6pkvl\" (UID: \"ed7c6200-6e0f-4865-81a8-7eb5c39a996f\") " pod="openstack/manila-8949-account-create-update-6pkvl" Dec 13 07:59:43 crc kubenswrapper[4971]: I1213 07:59:43.523707 4971 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-backup-0" event={"ID":"be3cc24a-8820-4884-8f26-e2885d933527","Type":"ContainerStarted","Data":"dcf26701dd7d956bb6c52af205b2331096dbdc4067d0986ebfd0e18980181462"} Dec 13 07:59:43 crc kubenswrapper[4971]: I1213 07:59:43.526070 4971 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-volume-volume1-0" event={"ID":"7bbf960d-93e9-49d6-8545-28c34e36229e","Type":"ContainerStarted","Data":"e1d84bf0ee0474bf27595a8d0b46a880f6799458081836b24da8ab268e6777a0"} Dec 13 07:59:43 crc kubenswrapper[4971]: I1213 07:59:43.545741 4971 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-internal-api-0" Dec 13 07:59:43 crc kubenswrapper[4971]: I1213 07:59:43.562501 4971 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-7vx49\" (UniqueName: \"kubernetes.io/projected/ed7c6200-6e0f-4865-81a8-7eb5c39a996f-kube-api-access-7vx49\") pod \"manila-8949-account-create-update-6pkvl\" (UID: \"ed7c6200-6e0f-4865-81a8-7eb5c39a996f\") " pod="openstack/manila-8949-account-create-update-6pkvl" Dec 13 07:59:43 crc kubenswrapper[4971]: I1213 07:59:43.657309 4971 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/manila-8949-account-create-update-6pkvl" Dec 13 07:59:44 crc kubenswrapper[4971]: I1213 07:59:44.028782 4971 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/manila-db-create-wmjhn"] Dec 13 07:59:44 crc kubenswrapper[4971]: I1213 07:59:44.274442 4971 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/manila-8949-account-create-update-6pkvl"] Dec 13 07:59:44 crc kubenswrapper[4971]: I1213 07:59:44.314847 4971 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-default-external-api-0"] Dec 13 07:59:44 crc kubenswrapper[4971]: I1213 07:59:44.357271 4971 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-default-internal-api-0"] Dec 13 07:59:44 crc kubenswrapper[4971]: W1213 07:59:44.445271 4971 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod3158ebe6_4c8c_41d1_bf95_6b4492a7e0aa.slice/crio-b4b4d864b28846e724abf36e679f5f83b9994a119985b2cc0ba8dc4917bcdcfc WatchSource:0}: Error finding container b4b4d864b28846e724abf36e679f5f83b9994a119985b2cc0ba8dc4917bcdcfc: Status 404 returned error can't find the container with id b4b4d864b28846e724abf36e679f5f83b9994a119985b2cc0ba8dc4917bcdcfc Dec 13 07:59:44 crc kubenswrapper[4971]: W1213 07:59:44.448495 4971 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod9fbf6500_bf43_4a4c_820c_14a60a27afb9.slice/crio-92a3b7d2f6d5b3c53cc716e6165d598205ba1334bbe1061fb1294209dd046ca4 WatchSource:0}: Error finding container 92a3b7d2f6d5b3c53cc716e6165d598205ba1334bbe1061fb1294209dd046ca4: Status 404 returned error can't find the container with id 92a3b7d2f6d5b3c53cc716e6165d598205ba1334bbe1061fb1294209dd046ca4 Dec 13 07:59:44 crc kubenswrapper[4971]: I1213 07:59:44.546960 4971 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-external-api-0" event={"ID":"9fbf6500-bf43-4a4c-820c-14a60a27afb9","Type":"ContainerStarted","Data":"92a3b7d2f6d5b3c53cc716e6165d598205ba1334bbe1061fb1294209dd046ca4"} Dec 13 07:59:44 crc kubenswrapper[4971]: I1213 07:59:44.552855 4971 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-internal-api-0" event={"ID":"3158ebe6-4c8c-41d1-bf95-6b4492a7e0aa","Type":"ContainerStarted","Data":"b4b4d864b28846e724abf36e679f5f83b9994a119985b2cc0ba8dc4917bcdcfc"} Dec 13 07:59:44 crc kubenswrapper[4971]: I1213 07:59:44.557926 4971 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/manila-8949-account-create-update-6pkvl" event={"ID":"ed7c6200-6e0f-4865-81a8-7eb5c39a996f","Type":"ContainerStarted","Data":"46f060657151ec00e17cc20b1ba13e0002a0703fb8b34c800aa94d572eff2a62"} Dec 13 07:59:44 crc kubenswrapper[4971]: I1213 07:59:44.568839 4971 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/manila-db-create-wmjhn" event={"ID":"44d3eebd-3532-454e-b687-dd752f549afe","Type":"ContainerStarted","Data":"f3d1cf893616810f172fc00110ecfa59d1af17a2991bd2c6dbe7580dee57f25d"} Dec 13 07:59:44 crc kubenswrapper[4971]: I1213 07:59:44.568893 4971 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/manila-db-create-wmjhn" event={"ID":"44d3eebd-3532-454e-b687-dd752f549afe","Type":"ContainerStarted","Data":"c99570c25013447fd2c3691880dd7a66cd9430c863e485014cb497e4ab449428"} Dec 13 07:59:44 crc kubenswrapper[4971]: I1213 07:59:44.607625 4971 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/manila-db-create-wmjhn" podStartSLOduration=1.607605293 podStartE2EDuration="1.607605293s" podCreationTimestamp="2025-12-13 07:59:43 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-13 07:59:44.589716869 +0000 UTC m=+4241.194126317" watchObservedRunningTime="2025-12-13 07:59:44.607605293 +0000 UTC m=+4241.212014731" Dec 13 07:59:45 crc kubenswrapper[4971]: I1213 07:59:45.590512 4971 generic.go:334] "Generic (PLEG): container finished" podID="44d3eebd-3532-454e-b687-dd752f549afe" containerID="f3d1cf893616810f172fc00110ecfa59d1af17a2991bd2c6dbe7580dee57f25d" exitCode=0 Dec 13 07:59:45 crc kubenswrapper[4971]: I1213 07:59:45.591054 4971 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/manila-db-create-wmjhn" event={"ID":"44d3eebd-3532-454e-b687-dd752f549afe","Type":"ContainerDied","Data":"f3d1cf893616810f172fc00110ecfa59d1af17a2991bd2c6dbe7580dee57f25d"} Dec 13 07:59:45 crc kubenswrapper[4971]: I1213 07:59:45.596583 4971 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-volume-volume1-0" event={"ID":"7bbf960d-93e9-49d6-8545-28c34e36229e","Type":"ContainerStarted","Data":"b2b3e725743988906a61815f43f9745696f3014ddf79c996c88cfd2842b3dbc1"} Dec 13 07:59:45 crc kubenswrapper[4971]: I1213 07:59:45.596638 4971 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-volume-volume1-0" event={"ID":"7bbf960d-93e9-49d6-8545-28c34e36229e","Type":"ContainerStarted","Data":"fa55af632a406cf49bb585d71bd44b6514fc7dbb5610e48b83596c717cacecbc"} Dec 13 07:59:45 crc kubenswrapper[4971]: I1213 07:59:45.599283 4971 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-external-api-0" event={"ID":"9fbf6500-bf43-4a4c-820c-14a60a27afb9","Type":"ContainerStarted","Data":"64acdb4bf763df4fc63a722008b8e2ba5a6c2da09e68f82f6baf52b7a1385ab5"} Dec 13 07:59:45 crc kubenswrapper[4971]: I1213 07:59:45.601680 4971 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-internal-api-0" event={"ID":"3158ebe6-4c8c-41d1-bf95-6b4492a7e0aa","Type":"ContainerStarted","Data":"496c16830725d21d64292b517300f72d2a85cc46c00967b4ff0d1c22a79a137f"} Dec 13 07:59:45 crc kubenswrapper[4971]: I1213 07:59:45.603668 4971 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-backup-0" event={"ID":"be3cc24a-8820-4884-8f26-e2885d933527","Type":"ContainerStarted","Data":"c56593752393f04996e485a0ff05b2c1974e5e26f4fc0c4b0c31c65abaeb8acb"} Dec 13 07:59:45 crc kubenswrapper[4971]: I1213 07:59:45.603697 4971 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-backup-0" event={"ID":"be3cc24a-8820-4884-8f26-e2885d933527","Type":"ContainerStarted","Data":"6f7eabb6ffa53a36c69546e916da1eda0cd0ab409cabc5679d2b5daf2ca5ba58"} Dec 13 07:59:45 crc kubenswrapper[4971]: I1213 07:59:45.611083 4971 generic.go:334] "Generic (PLEG): container finished" podID="ed7c6200-6e0f-4865-81a8-7eb5c39a996f" containerID="5ee1aa5ebf8687e00fb713d9535ae93929f98ad14f5e297b30bffdc3870be18e" exitCode=0 Dec 13 07:59:45 crc kubenswrapper[4971]: I1213 07:59:45.611145 4971 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/manila-8949-account-create-update-6pkvl" event={"ID":"ed7c6200-6e0f-4865-81a8-7eb5c39a996f","Type":"ContainerDied","Data":"5ee1aa5ebf8687e00fb713d9535ae93929f98ad14f5e297b30bffdc3870be18e"} Dec 13 07:59:45 crc kubenswrapper[4971]: I1213 07:59:45.638886 4971 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/cinder-backup-0" podStartSLOduration=2.362592942 podStartE2EDuration="3.638865624s" podCreationTimestamp="2025-12-13 07:59:42 +0000 UTC" firstStartedPulling="2025-12-13 07:59:43.404787889 +0000 UTC m=+4240.009197337" lastFinishedPulling="2025-12-13 07:59:44.681060571 +0000 UTC m=+4241.285470019" observedRunningTime="2025-12-13 07:59:45.628265857 +0000 UTC m=+4242.232675325" watchObservedRunningTime="2025-12-13 07:59:45.638865624 +0000 UTC m=+4242.243275072" Dec 13 07:59:45 crc kubenswrapper[4971]: I1213 07:59:45.681121 4971 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/cinder-volume-volume1-0" podStartSLOduration=2.379467471 podStartE2EDuration="3.681093216s" podCreationTimestamp="2025-12-13 07:59:42 +0000 UTC" firstStartedPulling="2025-12-13 07:59:43.328867932 +0000 UTC m=+4239.933277380" lastFinishedPulling="2025-12-13 07:59:44.630493677 +0000 UTC m=+4241.234903125" observedRunningTime="2025-12-13 07:59:45.660566439 +0000 UTC m=+4242.264975887" watchObservedRunningTime="2025-12-13 07:59:45.681093216 +0000 UTC m=+4242.285502664" Dec 13 07:59:46 crc kubenswrapper[4971]: I1213 07:59:46.637538 4971 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-external-api-0" event={"ID":"9fbf6500-bf43-4a4c-820c-14a60a27afb9","Type":"ContainerStarted","Data":"36064aa630a16af6699d8871aa8e46f5dee8c3c1657687aa33ef71cd4f03d051"} Dec 13 07:59:46 crc kubenswrapper[4971]: I1213 07:59:46.641834 4971 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-internal-api-0" event={"ID":"3158ebe6-4c8c-41d1-bf95-6b4492a7e0aa","Type":"ContainerStarted","Data":"7878e7872efb08684bd3c01a02cf433fc3dede331383f9285e78644179d387d8"} Dec 13 07:59:46 crc kubenswrapper[4971]: I1213 07:59:46.665476 4971 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/glance-default-external-api-0" podStartSLOduration=4.665456242 podStartE2EDuration="4.665456242s" podCreationTimestamp="2025-12-13 07:59:42 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-13 07:59:46.664646272 +0000 UTC m=+4243.269055710" watchObservedRunningTime="2025-12-13 07:59:46.665456242 +0000 UTC m=+4243.269865690" Dec 13 07:59:46 crc kubenswrapper[4971]: I1213 07:59:46.702478 4971 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/glance-default-internal-api-0" podStartSLOduration=4.702458577 podStartE2EDuration="4.702458577s" podCreationTimestamp="2025-12-13 07:59:42 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-13 07:59:46.691906882 +0000 UTC m=+4243.296316340" watchObservedRunningTime="2025-12-13 07:59:46.702458577 +0000 UTC m=+4243.306868045" Dec 13 07:59:47 crc kubenswrapper[4971]: I1213 07:59:47.074791 4971 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/manila-db-create-wmjhn" Dec 13 07:59:47 crc kubenswrapper[4971]: I1213 07:59:47.117230 4971 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/manila-8949-account-create-update-6pkvl" Dec 13 07:59:47 crc kubenswrapper[4971]: I1213 07:59:47.212012 4971 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-rcdq2\" (UniqueName: \"kubernetes.io/projected/44d3eebd-3532-454e-b687-dd752f549afe-kube-api-access-rcdq2\") pod \"44d3eebd-3532-454e-b687-dd752f549afe\" (UID: \"44d3eebd-3532-454e-b687-dd752f549afe\") " Dec 13 07:59:47 crc kubenswrapper[4971]: I1213 07:59:47.212140 4971 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-7vx49\" (UniqueName: \"kubernetes.io/projected/ed7c6200-6e0f-4865-81a8-7eb5c39a996f-kube-api-access-7vx49\") pod \"ed7c6200-6e0f-4865-81a8-7eb5c39a996f\" (UID: \"ed7c6200-6e0f-4865-81a8-7eb5c39a996f\") " Dec 13 07:59:47 crc kubenswrapper[4971]: I1213 07:59:47.213227 4971 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/ed7c6200-6e0f-4865-81a8-7eb5c39a996f-operator-scripts\") pod \"ed7c6200-6e0f-4865-81a8-7eb5c39a996f\" (UID: \"ed7c6200-6e0f-4865-81a8-7eb5c39a996f\") " Dec 13 07:59:47 crc kubenswrapper[4971]: I1213 07:59:47.213284 4971 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/44d3eebd-3532-454e-b687-dd752f549afe-operator-scripts\") pod \"44d3eebd-3532-454e-b687-dd752f549afe\" (UID: \"44d3eebd-3532-454e-b687-dd752f549afe\") " Dec 13 07:59:47 crc kubenswrapper[4971]: I1213 07:59:47.213817 4971 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/ed7c6200-6e0f-4865-81a8-7eb5c39a996f-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "ed7c6200-6e0f-4865-81a8-7eb5c39a996f" (UID: "ed7c6200-6e0f-4865-81a8-7eb5c39a996f"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 13 07:59:47 crc kubenswrapper[4971]: I1213 07:59:47.214145 4971 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/44d3eebd-3532-454e-b687-dd752f549afe-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "44d3eebd-3532-454e-b687-dd752f549afe" (UID: "44d3eebd-3532-454e-b687-dd752f549afe"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 13 07:59:47 crc kubenswrapper[4971]: I1213 07:59:47.214810 4971 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/ed7c6200-6e0f-4865-81a8-7eb5c39a996f-operator-scripts\") on node \"crc\" DevicePath \"\"" Dec 13 07:59:47 crc kubenswrapper[4971]: I1213 07:59:47.214836 4971 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/44d3eebd-3532-454e-b687-dd752f549afe-operator-scripts\") on node \"crc\" DevicePath \"\"" Dec 13 07:59:47 crc kubenswrapper[4971]: I1213 07:59:47.223008 4971 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/44d3eebd-3532-454e-b687-dd752f549afe-kube-api-access-rcdq2" (OuterVolumeSpecName: "kube-api-access-rcdq2") pod "44d3eebd-3532-454e-b687-dd752f549afe" (UID: "44d3eebd-3532-454e-b687-dd752f549afe"). InnerVolumeSpecName "kube-api-access-rcdq2". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 13 07:59:47 crc kubenswrapper[4971]: I1213 07:59:47.223275 4971 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/ed7c6200-6e0f-4865-81a8-7eb5c39a996f-kube-api-access-7vx49" (OuterVolumeSpecName: "kube-api-access-7vx49") pod "ed7c6200-6e0f-4865-81a8-7eb5c39a996f" (UID: "ed7c6200-6e0f-4865-81a8-7eb5c39a996f"). InnerVolumeSpecName "kube-api-access-7vx49". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 13 07:59:47 crc kubenswrapper[4971]: I1213 07:59:47.317970 4971 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-7vx49\" (UniqueName: \"kubernetes.io/projected/ed7c6200-6e0f-4865-81a8-7eb5c39a996f-kube-api-access-7vx49\") on node \"crc\" DevicePath \"\"" Dec 13 07:59:47 crc kubenswrapper[4971]: I1213 07:59:47.318177 4971 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-rcdq2\" (UniqueName: \"kubernetes.io/projected/44d3eebd-3532-454e-b687-dd752f549afe-kube-api-access-rcdq2\") on node \"crc\" DevicePath \"\"" Dec 13 07:59:47 crc kubenswrapper[4971]: I1213 07:59:47.611174 4971 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/cinder-volume-volume1-0" Dec 13 07:59:47 crc kubenswrapper[4971]: I1213 07:59:47.624827 4971 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/cinder-backup-0" Dec 13 07:59:47 crc kubenswrapper[4971]: I1213 07:59:47.652683 4971 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/manila-db-create-wmjhn" Dec 13 07:59:47 crc kubenswrapper[4971]: I1213 07:59:47.652677 4971 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/manila-db-create-wmjhn" event={"ID":"44d3eebd-3532-454e-b687-dd752f549afe","Type":"ContainerDied","Data":"c99570c25013447fd2c3691880dd7a66cd9430c863e485014cb497e4ab449428"} Dec 13 07:59:47 crc kubenswrapper[4971]: I1213 07:59:47.653034 4971 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="c99570c25013447fd2c3691880dd7a66cd9430c863e485014cb497e4ab449428" Dec 13 07:59:47 crc kubenswrapper[4971]: I1213 07:59:47.655748 4971 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/manila-8949-account-create-update-6pkvl" event={"ID":"ed7c6200-6e0f-4865-81a8-7eb5c39a996f","Type":"ContainerDied","Data":"46f060657151ec00e17cc20b1ba13e0002a0703fb8b34c800aa94d572eff2a62"} Dec 13 07:59:47 crc kubenswrapper[4971]: I1213 07:59:47.655783 4971 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="46f060657151ec00e17cc20b1ba13e0002a0703fb8b34c800aa94d572eff2a62" Dec 13 07:59:47 crc kubenswrapper[4971]: I1213 07:59:47.655816 4971 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/manila-8949-account-create-update-6pkvl" Dec 13 07:59:48 crc kubenswrapper[4971]: I1213 07:59:48.638478 4971 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/manila-db-sync-g9bws"] Dec 13 07:59:48 crc kubenswrapper[4971]: E1213 07:59:48.638918 4971 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="44d3eebd-3532-454e-b687-dd752f549afe" containerName="mariadb-database-create" Dec 13 07:59:48 crc kubenswrapper[4971]: I1213 07:59:48.638936 4971 state_mem.go:107] "Deleted CPUSet assignment" podUID="44d3eebd-3532-454e-b687-dd752f549afe" containerName="mariadb-database-create" Dec 13 07:59:48 crc kubenswrapper[4971]: E1213 07:59:48.638962 4971 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="ed7c6200-6e0f-4865-81a8-7eb5c39a996f" containerName="mariadb-account-create-update" Dec 13 07:59:48 crc kubenswrapper[4971]: I1213 07:59:48.638969 4971 state_mem.go:107] "Deleted CPUSet assignment" podUID="ed7c6200-6e0f-4865-81a8-7eb5c39a996f" containerName="mariadb-account-create-update" Dec 13 07:59:48 crc kubenswrapper[4971]: I1213 07:59:48.639164 4971 memory_manager.go:354] "RemoveStaleState removing state" podUID="44d3eebd-3532-454e-b687-dd752f549afe" containerName="mariadb-database-create" Dec 13 07:59:48 crc kubenswrapper[4971]: I1213 07:59:48.639181 4971 memory_manager.go:354] "RemoveStaleState removing state" podUID="ed7c6200-6e0f-4865-81a8-7eb5c39a996f" containerName="mariadb-account-create-update" Dec 13 07:59:48 crc kubenswrapper[4971]: I1213 07:59:48.639794 4971 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/manila-db-sync-g9bws" Dec 13 07:59:48 crc kubenswrapper[4971]: I1213 07:59:48.645096 4971 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"manila-manila-dockercfg-mn89t" Dec 13 07:59:48 crc kubenswrapper[4971]: I1213 07:59:48.645413 4971 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"manila-config-data" Dec 13 07:59:48 crc kubenswrapper[4971]: I1213 07:59:48.653261 4971 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/manila-db-sync-g9bws"] Dec 13 07:59:48 crc kubenswrapper[4971]: I1213 07:59:48.748555 4971 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"job-config-data\" (UniqueName: \"kubernetes.io/secret/bb2c5402-103a-4456-960f-7b0019111f1e-job-config-data\") pod \"manila-db-sync-g9bws\" (UID: \"bb2c5402-103a-4456-960f-7b0019111f1e\") " pod="openstack/manila-db-sync-g9bws" Dec 13 07:59:48 crc kubenswrapper[4971]: I1213 07:59:48.748985 4971 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/bb2c5402-103a-4456-960f-7b0019111f1e-config-data\") pod \"manila-db-sync-g9bws\" (UID: \"bb2c5402-103a-4456-960f-7b0019111f1e\") " pod="openstack/manila-db-sync-g9bws" Dec 13 07:59:48 crc kubenswrapper[4971]: I1213 07:59:48.749034 4971 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-2jfvl\" (UniqueName: \"kubernetes.io/projected/bb2c5402-103a-4456-960f-7b0019111f1e-kube-api-access-2jfvl\") pod \"manila-db-sync-g9bws\" (UID: \"bb2c5402-103a-4456-960f-7b0019111f1e\") " pod="openstack/manila-db-sync-g9bws" Dec 13 07:59:48 crc kubenswrapper[4971]: I1213 07:59:48.749130 4971 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/bb2c5402-103a-4456-960f-7b0019111f1e-combined-ca-bundle\") pod \"manila-db-sync-g9bws\" (UID: \"bb2c5402-103a-4456-960f-7b0019111f1e\") " pod="openstack/manila-db-sync-g9bws" Dec 13 07:59:48 crc kubenswrapper[4971]: I1213 07:59:48.851345 4971 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"job-config-data\" (UniqueName: \"kubernetes.io/secret/bb2c5402-103a-4456-960f-7b0019111f1e-job-config-data\") pod \"manila-db-sync-g9bws\" (UID: \"bb2c5402-103a-4456-960f-7b0019111f1e\") " pod="openstack/manila-db-sync-g9bws" Dec 13 07:59:48 crc kubenswrapper[4971]: I1213 07:59:48.851836 4971 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/bb2c5402-103a-4456-960f-7b0019111f1e-config-data\") pod \"manila-db-sync-g9bws\" (UID: \"bb2c5402-103a-4456-960f-7b0019111f1e\") " pod="openstack/manila-db-sync-g9bws" Dec 13 07:59:48 crc kubenswrapper[4971]: I1213 07:59:48.852616 4971 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-2jfvl\" (UniqueName: \"kubernetes.io/projected/bb2c5402-103a-4456-960f-7b0019111f1e-kube-api-access-2jfvl\") pod \"manila-db-sync-g9bws\" (UID: \"bb2c5402-103a-4456-960f-7b0019111f1e\") " pod="openstack/manila-db-sync-g9bws" Dec 13 07:59:48 crc kubenswrapper[4971]: I1213 07:59:48.852753 4971 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/bb2c5402-103a-4456-960f-7b0019111f1e-combined-ca-bundle\") pod \"manila-db-sync-g9bws\" (UID: \"bb2c5402-103a-4456-960f-7b0019111f1e\") " pod="openstack/manila-db-sync-g9bws" Dec 13 07:59:48 crc kubenswrapper[4971]: I1213 07:59:48.860219 4971 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/bb2c5402-103a-4456-960f-7b0019111f1e-combined-ca-bundle\") pod \"manila-db-sync-g9bws\" (UID: \"bb2c5402-103a-4456-960f-7b0019111f1e\") " pod="openstack/manila-db-sync-g9bws" Dec 13 07:59:48 crc kubenswrapper[4971]: I1213 07:59:48.861007 4971 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/bb2c5402-103a-4456-960f-7b0019111f1e-config-data\") pod \"manila-db-sync-g9bws\" (UID: \"bb2c5402-103a-4456-960f-7b0019111f1e\") " pod="openstack/manila-db-sync-g9bws" Dec 13 07:59:48 crc kubenswrapper[4971]: I1213 07:59:48.863248 4971 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"job-config-data\" (UniqueName: \"kubernetes.io/secret/bb2c5402-103a-4456-960f-7b0019111f1e-job-config-data\") pod \"manila-db-sync-g9bws\" (UID: \"bb2c5402-103a-4456-960f-7b0019111f1e\") " pod="openstack/manila-db-sync-g9bws" Dec 13 07:59:48 crc kubenswrapper[4971]: I1213 07:59:48.891794 4971 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-2jfvl\" (UniqueName: \"kubernetes.io/projected/bb2c5402-103a-4456-960f-7b0019111f1e-kube-api-access-2jfvl\") pod \"manila-db-sync-g9bws\" (UID: \"bb2c5402-103a-4456-960f-7b0019111f1e\") " pod="openstack/manila-db-sync-g9bws" Dec 13 07:59:48 crc kubenswrapper[4971]: I1213 07:59:48.966473 4971 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/manila-db-sync-g9bws" Dec 13 07:59:49 crc kubenswrapper[4971]: I1213 07:59:49.563891 4971 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/manila-db-sync-g9bws"] Dec 13 07:59:49 crc kubenswrapper[4971]: I1213 07:59:49.578242 4971 provider.go:102] Refreshing cache for provider: *credentialprovider.defaultDockerConfigProvider Dec 13 07:59:49 crc kubenswrapper[4971]: I1213 07:59:49.680145 4971 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/manila-db-sync-g9bws" event={"ID":"bb2c5402-103a-4456-960f-7b0019111f1e","Type":"ContainerStarted","Data":"a60235975f97ac9fc75fa5840f227359703bcbf5f22dc23a1ff768fd97ed5868"} Dec 13 07:59:52 crc kubenswrapper[4971]: I1213 07:59:52.861207 4971 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/cinder-backup-0" Dec 13 07:59:52 crc kubenswrapper[4971]: I1213 07:59:52.901832 4971 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/cinder-volume-volume1-0" Dec 13 07:59:53 crc kubenswrapper[4971]: I1213 07:59:53.505042 4971 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/glance-default-external-api-0" Dec 13 07:59:53 crc kubenswrapper[4971]: I1213 07:59:53.505095 4971 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/glance-default-external-api-0" Dec 13 07:59:53 crc kubenswrapper[4971]: I1213 07:59:53.546728 4971 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/glance-default-internal-api-0" Dec 13 07:59:53 crc kubenswrapper[4971]: I1213 07:59:53.547425 4971 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/glance-default-internal-api-0" Dec 13 07:59:53 crc kubenswrapper[4971]: I1213 07:59:53.557336 4971 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/glance-default-external-api-0" Dec 13 07:59:53 crc kubenswrapper[4971]: I1213 07:59:53.567843 4971 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/glance-default-external-api-0" Dec 13 07:59:53 crc kubenswrapper[4971]: I1213 07:59:53.583735 4971 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/glance-default-internal-api-0" Dec 13 07:59:53 crc kubenswrapper[4971]: I1213 07:59:53.612605 4971 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/glance-default-internal-api-0" Dec 13 07:59:53 crc kubenswrapper[4971]: I1213 07:59:53.729341 4971 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/glance-default-internal-api-0" Dec 13 07:59:53 crc kubenswrapper[4971]: I1213 07:59:53.729394 4971 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/glance-default-external-api-0" Dec 13 07:59:53 crc kubenswrapper[4971]: I1213 07:59:53.729415 4971 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/glance-default-external-api-0" Dec 13 07:59:53 crc kubenswrapper[4971]: I1213 07:59:53.729431 4971 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/glance-default-internal-api-0" Dec 13 07:59:56 crc kubenswrapper[4971]: I1213 07:59:56.515696 4971 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/glance-default-external-api-0" Dec 13 07:59:56 crc kubenswrapper[4971]: I1213 07:59:56.516203 4971 prober_manager.go:312] "Failed to trigger a manual run" probe="Readiness" Dec 13 07:59:56 crc kubenswrapper[4971]: I1213 07:59:56.518109 4971 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/glance-default-internal-api-0" Dec 13 07:59:56 crc kubenswrapper[4971]: I1213 07:59:56.518215 4971 prober_manager.go:312] "Failed to trigger a manual run" probe="Readiness" Dec 13 07:59:56 crc kubenswrapper[4971]: I1213 07:59:56.519209 4971 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/glance-default-internal-api-0" Dec 13 07:59:56 crc kubenswrapper[4971]: I1213 07:59:56.531301 4971 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/glance-default-external-api-0" Dec 13 07:59:56 crc kubenswrapper[4971]: I1213 07:59:56.761764 4971 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/manila-db-sync-g9bws" event={"ID":"bb2c5402-103a-4456-960f-7b0019111f1e","Type":"ContainerStarted","Data":"0339a9bf168c17c0589ea918f3a679eadb4868cae95155960974063bdd0efa2c"} Dec 13 07:59:56 crc kubenswrapper[4971]: I1213 07:59:56.782868 4971 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/manila-db-sync-g9bws" podStartSLOduration=3.076496338 podStartE2EDuration="8.782849958s" podCreationTimestamp="2025-12-13 07:59:48 +0000 UTC" firstStartedPulling="2025-12-13 07:59:49.578015779 +0000 UTC m=+4246.182425227" lastFinishedPulling="2025-12-13 07:59:55.284369369 +0000 UTC m=+4251.888778847" observedRunningTime="2025-12-13 07:59:56.774625069 +0000 UTC m=+4253.379034527" watchObservedRunningTime="2025-12-13 07:59:56.782849958 +0000 UTC m=+4253.387259406" Dec 13 08:00:00 crc kubenswrapper[4971]: I1213 08:00:00.171018 4971 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29426880-2x7zw"] Dec 13 08:00:00 crc kubenswrapper[4971]: I1213 08:00:00.172928 4971 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29426880-2x7zw" Dec 13 08:00:00 crc kubenswrapper[4971]: I1213 08:00:00.182020 4971 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-operator-lifecycle-manager"/"collect-profiles-config" Dec 13 08:00:00 crc kubenswrapper[4971]: I1213 08:00:00.182308 4971 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"collect-profiles-dockercfg-kzf4t" Dec 13 08:00:00 crc kubenswrapper[4971]: I1213 08:00:00.182844 4971 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29426880-2x7zw"] Dec 13 08:00:00 crc kubenswrapper[4971]: I1213 08:00:00.189331 4971 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/b9898935-5232-4bdb-846d-27084d2a8be6-config-volume\") pod \"collect-profiles-29426880-2x7zw\" (UID: \"b9898935-5232-4bdb-846d-27084d2a8be6\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29426880-2x7zw" Dec 13 08:00:00 crc kubenswrapper[4971]: I1213 08:00:00.189412 4971 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/b9898935-5232-4bdb-846d-27084d2a8be6-secret-volume\") pod \"collect-profiles-29426880-2x7zw\" (UID: \"b9898935-5232-4bdb-846d-27084d2a8be6\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29426880-2x7zw" Dec 13 08:00:00 crc kubenswrapper[4971]: I1213 08:00:00.189544 4971 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-gwm5k\" (UniqueName: \"kubernetes.io/projected/b9898935-5232-4bdb-846d-27084d2a8be6-kube-api-access-gwm5k\") pod \"collect-profiles-29426880-2x7zw\" (UID: \"b9898935-5232-4bdb-846d-27084d2a8be6\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29426880-2x7zw" Dec 13 08:00:00 crc kubenswrapper[4971]: I1213 08:00:00.291694 4971 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-gwm5k\" (UniqueName: \"kubernetes.io/projected/b9898935-5232-4bdb-846d-27084d2a8be6-kube-api-access-gwm5k\") pod \"collect-profiles-29426880-2x7zw\" (UID: \"b9898935-5232-4bdb-846d-27084d2a8be6\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29426880-2x7zw" Dec 13 08:00:00 crc kubenswrapper[4971]: I1213 08:00:00.291858 4971 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/b9898935-5232-4bdb-846d-27084d2a8be6-config-volume\") pod \"collect-profiles-29426880-2x7zw\" (UID: \"b9898935-5232-4bdb-846d-27084d2a8be6\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29426880-2x7zw" Dec 13 08:00:00 crc kubenswrapper[4971]: I1213 08:00:00.291893 4971 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/b9898935-5232-4bdb-846d-27084d2a8be6-secret-volume\") pod \"collect-profiles-29426880-2x7zw\" (UID: \"b9898935-5232-4bdb-846d-27084d2a8be6\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29426880-2x7zw" Dec 13 08:00:00 crc kubenswrapper[4971]: I1213 08:00:00.292772 4971 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/b9898935-5232-4bdb-846d-27084d2a8be6-config-volume\") pod \"collect-profiles-29426880-2x7zw\" (UID: \"b9898935-5232-4bdb-846d-27084d2a8be6\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29426880-2x7zw" Dec 13 08:00:00 crc kubenswrapper[4971]: I1213 08:00:00.297501 4971 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/b9898935-5232-4bdb-846d-27084d2a8be6-secret-volume\") pod \"collect-profiles-29426880-2x7zw\" (UID: \"b9898935-5232-4bdb-846d-27084d2a8be6\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29426880-2x7zw" Dec 13 08:00:00 crc kubenswrapper[4971]: I1213 08:00:00.310651 4971 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-gwm5k\" (UniqueName: \"kubernetes.io/projected/b9898935-5232-4bdb-846d-27084d2a8be6-kube-api-access-gwm5k\") pod \"collect-profiles-29426880-2x7zw\" (UID: \"b9898935-5232-4bdb-846d-27084d2a8be6\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29426880-2x7zw" Dec 13 08:00:00 crc kubenswrapper[4971]: I1213 08:00:00.567276 4971 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29426880-2x7zw" Dec 13 08:00:01 crc kubenswrapper[4971]: I1213 08:00:01.169447 4971 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29426880-2x7zw"] Dec 13 08:00:01 crc kubenswrapper[4971]: W1213 08:00:01.176206 4971 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podb9898935_5232_4bdb_846d_27084d2a8be6.slice/crio-7a4a2eb7676a601370250e5d1a183ef23fe7c1eecb46937c9b013da28932bd8f WatchSource:0}: Error finding container 7a4a2eb7676a601370250e5d1a183ef23fe7c1eecb46937c9b013da28932bd8f: Status 404 returned error can't find the container with id 7a4a2eb7676a601370250e5d1a183ef23fe7c1eecb46937c9b013da28932bd8f Dec 13 08:00:01 crc kubenswrapper[4971]: I1213 08:00:01.810108 4971 generic.go:334] "Generic (PLEG): container finished" podID="b9898935-5232-4bdb-846d-27084d2a8be6" containerID="802b1f844ed17148a197d5f8e8c87f118de71e681fde826e33cd782b93ef770a" exitCode=0 Dec 13 08:00:01 crc kubenswrapper[4971]: I1213 08:00:01.810203 4971 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29426880-2x7zw" event={"ID":"b9898935-5232-4bdb-846d-27084d2a8be6","Type":"ContainerDied","Data":"802b1f844ed17148a197d5f8e8c87f118de71e681fde826e33cd782b93ef770a"} Dec 13 08:00:01 crc kubenswrapper[4971]: I1213 08:00:01.810405 4971 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29426880-2x7zw" event={"ID":"b9898935-5232-4bdb-846d-27084d2a8be6","Type":"ContainerStarted","Data":"7a4a2eb7676a601370250e5d1a183ef23fe7c1eecb46937c9b013da28932bd8f"} Dec 13 08:00:03 crc kubenswrapper[4971]: I1213 08:00:03.206052 4971 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29426880-2x7zw" Dec 13 08:00:03 crc kubenswrapper[4971]: I1213 08:00:03.275121 4971 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/b9898935-5232-4bdb-846d-27084d2a8be6-secret-volume\") pod \"b9898935-5232-4bdb-846d-27084d2a8be6\" (UID: \"b9898935-5232-4bdb-846d-27084d2a8be6\") " Dec 13 08:00:03 crc kubenswrapper[4971]: I1213 08:00:03.275246 4971 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/b9898935-5232-4bdb-846d-27084d2a8be6-config-volume\") pod \"b9898935-5232-4bdb-846d-27084d2a8be6\" (UID: \"b9898935-5232-4bdb-846d-27084d2a8be6\") " Dec 13 08:00:03 crc kubenswrapper[4971]: I1213 08:00:03.275337 4971 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-gwm5k\" (UniqueName: \"kubernetes.io/projected/b9898935-5232-4bdb-846d-27084d2a8be6-kube-api-access-gwm5k\") pod \"b9898935-5232-4bdb-846d-27084d2a8be6\" (UID: \"b9898935-5232-4bdb-846d-27084d2a8be6\") " Dec 13 08:00:03 crc kubenswrapper[4971]: I1213 08:00:03.277917 4971 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/b9898935-5232-4bdb-846d-27084d2a8be6-config-volume" (OuterVolumeSpecName: "config-volume") pod "b9898935-5232-4bdb-846d-27084d2a8be6" (UID: "b9898935-5232-4bdb-846d-27084d2a8be6"). InnerVolumeSpecName "config-volume". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 13 08:00:03 crc kubenswrapper[4971]: I1213 08:00:03.282325 4971 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/b9898935-5232-4bdb-846d-27084d2a8be6-kube-api-access-gwm5k" (OuterVolumeSpecName: "kube-api-access-gwm5k") pod "b9898935-5232-4bdb-846d-27084d2a8be6" (UID: "b9898935-5232-4bdb-846d-27084d2a8be6"). InnerVolumeSpecName "kube-api-access-gwm5k". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 13 08:00:03 crc kubenswrapper[4971]: I1213 08:00:03.282827 4971 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/b9898935-5232-4bdb-846d-27084d2a8be6-secret-volume" (OuterVolumeSpecName: "secret-volume") pod "b9898935-5232-4bdb-846d-27084d2a8be6" (UID: "b9898935-5232-4bdb-846d-27084d2a8be6"). InnerVolumeSpecName "secret-volume". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 13 08:00:03 crc kubenswrapper[4971]: I1213 08:00:03.378071 4971 reconciler_common.go:293] "Volume detached for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/b9898935-5232-4bdb-846d-27084d2a8be6-secret-volume\") on node \"crc\" DevicePath \"\"" Dec 13 08:00:03 crc kubenswrapper[4971]: I1213 08:00:03.378107 4971 reconciler_common.go:293] "Volume detached for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/b9898935-5232-4bdb-846d-27084d2a8be6-config-volume\") on node \"crc\" DevicePath \"\"" Dec 13 08:00:03 crc kubenswrapper[4971]: I1213 08:00:03.378122 4971 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-gwm5k\" (UniqueName: \"kubernetes.io/projected/b9898935-5232-4bdb-846d-27084d2a8be6-kube-api-access-gwm5k\") on node \"crc\" DevicePath \"\"" Dec 13 08:00:03 crc kubenswrapper[4971]: I1213 08:00:03.848186 4971 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29426880-2x7zw" event={"ID":"b9898935-5232-4bdb-846d-27084d2a8be6","Type":"ContainerDied","Data":"7a4a2eb7676a601370250e5d1a183ef23fe7c1eecb46937c9b013da28932bd8f"} Dec 13 08:00:03 crc kubenswrapper[4971]: I1213 08:00:03.848594 4971 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="7a4a2eb7676a601370250e5d1a183ef23fe7c1eecb46937c9b013da28932bd8f" Dec 13 08:00:03 crc kubenswrapper[4971]: I1213 08:00:03.848750 4971 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29426880-2x7zw" Dec 13 08:00:04 crc kubenswrapper[4971]: I1213 08:00:04.281062 4971 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29426835-d7bs5"] Dec 13 08:00:04 crc kubenswrapper[4971]: I1213 08:00:04.288306 4971 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29426835-d7bs5"] Dec 13 08:00:05 crc kubenswrapper[4971]: I1213 08:00:05.851904 4971 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="2054169f-b228-4471-ba23-2a74cb0a126c" path="/var/lib/kubelet/pods/2054169f-b228-4471-ba23-2a74cb0a126c/volumes" Dec 13 08:00:07 crc kubenswrapper[4971]: I1213 08:00:07.890991 4971 generic.go:334] "Generic (PLEG): container finished" podID="bb2c5402-103a-4456-960f-7b0019111f1e" containerID="0339a9bf168c17c0589ea918f3a679eadb4868cae95155960974063bdd0efa2c" exitCode=0 Dec 13 08:00:07 crc kubenswrapper[4971]: I1213 08:00:07.891109 4971 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/manila-db-sync-g9bws" event={"ID":"bb2c5402-103a-4456-960f-7b0019111f1e","Type":"ContainerDied","Data":"0339a9bf168c17c0589ea918f3a679eadb4868cae95155960974063bdd0efa2c"} Dec 13 08:00:09 crc kubenswrapper[4971]: I1213 08:00:09.336010 4971 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/manila-db-sync-g9bws" Dec 13 08:00:09 crc kubenswrapper[4971]: I1213 08:00:09.521341 4971 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/bb2c5402-103a-4456-960f-7b0019111f1e-config-data\") pod \"bb2c5402-103a-4456-960f-7b0019111f1e\" (UID: \"bb2c5402-103a-4456-960f-7b0019111f1e\") " Dec 13 08:00:09 crc kubenswrapper[4971]: I1213 08:00:09.521398 4971 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-2jfvl\" (UniqueName: \"kubernetes.io/projected/bb2c5402-103a-4456-960f-7b0019111f1e-kube-api-access-2jfvl\") pod \"bb2c5402-103a-4456-960f-7b0019111f1e\" (UID: \"bb2c5402-103a-4456-960f-7b0019111f1e\") " Dec 13 08:00:09 crc kubenswrapper[4971]: I1213 08:00:09.521532 4971 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"job-config-data\" (UniqueName: \"kubernetes.io/secret/bb2c5402-103a-4456-960f-7b0019111f1e-job-config-data\") pod \"bb2c5402-103a-4456-960f-7b0019111f1e\" (UID: \"bb2c5402-103a-4456-960f-7b0019111f1e\") " Dec 13 08:00:09 crc kubenswrapper[4971]: I1213 08:00:09.521649 4971 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/bb2c5402-103a-4456-960f-7b0019111f1e-combined-ca-bundle\") pod \"bb2c5402-103a-4456-960f-7b0019111f1e\" (UID: \"bb2c5402-103a-4456-960f-7b0019111f1e\") " Dec 13 08:00:09 crc kubenswrapper[4971]: I1213 08:00:09.527960 4971 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/bb2c5402-103a-4456-960f-7b0019111f1e-job-config-data" (OuterVolumeSpecName: "job-config-data") pod "bb2c5402-103a-4456-960f-7b0019111f1e" (UID: "bb2c5402-103a-4456-960f-7b0019111f1e"). InnerVolumeSpecName "job-config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 13 08:00:09 crc kubenswrapper[4971]: I1213 08:00:09.531950 4971 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/bb2c5402-103a-4456-960f-7b0019111f1e-config-data" (OuterVolumeSpecName: "config-data") pod "bb2c5402-103a-4456-960f-7b0019111f1e" (UID: "bb2c5402-103a-4456-960f-7b0019111f1e"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 13 08:00:09 crc kubenswrapper[4971]: I1213 08:00:09.533577 4971 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/bb2c5402-103a-4456-960f-7b0019111f1e-kube-api-access-2jfvl" (OuterVolumeSpecName: "kube-api-access-2jfvl") pod "bb2c5402-103a-4456-960f-7b0019111f1e" (UID: "bb2c5402-103a-4456-960f-7b0019111f1e"). InnerVolumeSpecName "kube-api-access-2jfvl". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 13 08:00:09 crc kubenswrapper[4971]: I1213 08:00:09.572495 4971 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/bb2c5402-103a-4456-960f-7b0019111f1e-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "bb2c5402-103a-4456-960f-7b0019111f1e" (UID: "bb2c5402-103a-4456-960f-7b0019111f1e"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 13 08:00:09 crc kubenswrapper[4971]: I1213 08:00:09.624306 4971 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/bb2c5402-103a-4456-960f-7b0019111f1e-config-data\") on node \"crc\" DevicePath \"\"" Dec 13 08:00:09 crc kubenswrapper[4971]: I1213 08:00:09.624342 4971 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-2jfvl\" (UniqueName: \"kubernetes.io/projected/bb2c5402-103a-4456-960f-7b0019111f1e-kube-api-access-2jfvl\") on node \"crc\" DevicePath \"\"" Dec 13 08:00:09 crc kubenswrapper[4971]: I1213 08:00:09.624361 4971 reconciler_common.go:293] "Volume detached for volume \"job-config-data\" (UniqueName: \"kubernetes.io/secret/bb2c5402-103a-4456-960f-7b0019111f1e-job-config-data\") on node \"crc\" DevicePath \"\"" Dec 13 08:00:09 crc kubenswrapper[4971]: I1213 08:00:09.624374 4971 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/bb2c5402-103a-4456-960f-7b0019111f1e-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 13 08:00:09 crc kubenswrapper[4971]: I1213 08:00:09.914211 4971 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/manila-db-sync-g9bws" event={"ID":"bb2c5402-103a-4456-960f-7b0019111f1e","Type":"ContainerDied","Data":"a60235975f97ac9fc75fa5840f227359703bcbf5f22dc23a1ff768fd97ed5868"} Dec 13 08:00:09 crc kubenswrapper[4971]: I1213 08:00:09.914264 4971 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="a60235975f97ac9fc75fa5840f227359703bcbf5f22dc23a1ff768fd97ed5868" Dec 13 08:00:09 crc kubenswrapper[4971]: I1213 08:00:09.914260 4971 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/manila-db-sync-g9bws" Dec 13 08:00:10 crc kubenswrapper[4971]: I1213 08:00:10.407511 4971 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/manila-scheduler-0"] Dec 13 08:00:10 crc kubenswrapper[4971]: E1213 08:00:10.408501 4971 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="b9898935-5232-4bdb-846d-27084d2a8be6" containerName="collect-profiles" Dec 13 08:00:10 crc kubenswrapper[4971]: I1213 08:00:10.408548 4971 state_mem.go:107] "Deleted CPUSet assignment" podUID="b9898935-5232-4bdb-846d-27084d2a8be6" containerName="collect-profiles" Dec 13 08:00:10 crc kubenswrapper[4971]: E1213 08:00:10.408578 4971 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="bb2c5402-103a-4456-960f-7b0019111f1e" containerName="manila-db-sync" Dec 13 08:00:10 crc kubenswrapper[4971]: I1213 08:00:10.408591 4971 state_mem.go:107] "Deleted CPUSet assignment" podUID="bb2c5402-103a-4456-960f-7b0019111f1e" containerName="manila-db-sync" Dec 13 08:00:10 crc kubenswrapper[4971]: I1213 08:00:10.408965 4971 memory_manager.go:354] "RemoveStaleState removing state" podUID="b9898935-5232-4bdb-846d-27084d2a8be6" containerName="collect-profiles" Dec 13 08:00:10 crc kubenswrapper[4971]: I1213 08:00:10.409017 4971 memory_manager.go:354] "RemoveStaleState removing state" podUID="bb2c5402-103a-4456-960f-7b0019111f1e" containerName="manila-db-sync" Dec 13 08:00:10 crc kubenswrapper[4971]: I1213 08:00:10.410707 4971 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/manila-scheduler-0" Dec 13 08:00:10 crc kubenswrapper[4971]: I1213 08:00:10.418152 4971 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"manila-manila-dockercfg-mn89t" Dec 13 08:00:10 crc kubenswrapper[4971]: I1213 08:00:10.418344 4971 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"manila-scripts" Dec 13 08:00:10 crc kubenswrapper[4971]: I1213 08:00:10.418783 4971 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"manila-scheduler-config-data" Dec 13 08:00:10 crc kubenswrapper[4971]: I1213 08:00:10.419053 4971 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"manila-config-data" Dec 13 08:00:10 crc kubenswrapper[4971]: I1213 08:00:10.437868 4971 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/manila-scheduler-0"] Dec 13 08:00:10 crc kubenswrapper[4971]: I1213 08:00:10.470205 4971 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/manila-share-share1-0"] Dec 13 08:00:10 crc kubenswrapper[4971]: I1213 08:00:10.472476 4971 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/manila-share-share1-0" Dec 13 08:00:10 crc kubenswrapper[4971]: I1213 08:00:10.476948 4971 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"manila-share-share1-config-data" Dec 13 08:00:10 crc kubenswrapper[4971]: I1213 08:00:10.479501 4971 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/1d983511-44dd-4be9-9b43-2f496428a12e-combined-ca-bundle\") pod \"manila-share-share1-0\" (UID: \"1d983511-44dd-4be9-9b43-2f496428a12e\") " pod="openstack/manila-share-share1-0" Dec 13 08:00:10 crc kubenswrapper[4971]: I1213 08:00:10.479609 4971 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-cb2g9\" (UniqueName: \"kubernetes.io/projected/1d983511-44dd-4be9-9b43-2f496428a12e-kube-api-access-cb2g9\") pod \"manila-share-share1-0\" (UID: \"1d983511-44dd-4be9-9b43-2f496428a12e\") " pod="openstack/manila-share-share1-0" Dec 13 08:00:10 crc kubenswrapper[4971]: I1213 08:00:10.479884 4971 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-qgmtc\" (UniqueName: \"kubernetes.io/projected/c815d65d-6fc3-49c3-bda5-a2ea21b54438-kube-api-access-qgmtc\") pod \"manila-scheduler-0\" (UID: \"c815d65d-6fc3-49c3-bda5-a2ea21b54438\") " pod="openstack/manila-scheduler-0" Dec 13 08:00:10 crc kubenswrapper[4971]: I1213 08:00:10.479925 4971 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/1d983511-44dd-4be9-9b43-2f496428a12e-scripts\") pod \"manila-share-share1-0\" (UID: \"1d983511-44dd-4be9-9b43-2f496428a12e\") " pod="openstack/manila-share-share1-0" Dec 13 08:00:10 crc kubenswrapper[4971]: I1213 08:00:10.480209 4971 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-lib-manila\" (UniqueName: \"kubernetes.io/host-path/1d983511-44dd-4be9-9b43-2f496428a12e-var-lib-manila\") pod \"manila-share-share1-0\" (UID: \"1d983511-44dd-4be9-9b43-2f496428a12e\") " pod="openstack/manila-share-share1-0" Dec 13 08:00:10 crc kubenswrapper[4971]: I1213 08:00:10.480244 4971 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/1d983511-44dd-4be9-9b43-2f496428a12e-config-data\") pod \"manila-share-share1-0\" (UID: \"1d983511-44dd-4be9-9b43-2f496428a12e\") " pod="openstack/manila-share-share1-0" Dec 13 08:00:10 crc kubenswrapper[4971]: I1213 08:00:10.480280 4971 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/c815d65d-6fc3-49c3-bda5-a2ea21b54438-scripts\") pod \"manila-scheduler-0\" (UID: \"c815d65d-6fc3-49c3-bda5-a2ea21b54438\") " pod="openstack/manila-scheduler-0" Dec 13 08:00:10 crc kubenswrapper[4971]: I1213 08:00:10.480339 4971 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/c815d65d-6fc3-49c3-bda5-a2ea21b54438-combined-ca-bundle\") pod \"manila-scheduler-0\" (UID: \"c815d65d-6fc3-49c3-bda5-a2ea21b54438\") " pod="openstack/manila-scheduler-0" Dec 13 08:00:10 crc kubenswrapper[4971]: I1213 08:00:10.481890 4971 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/1d983511-44dd-4be9-9b43-2f496428a12e-config-data-custom\") pod \"manila-share-share1-0\" (UID: \"1d983511-44dd-4be9-9b43-2f496428a12e\") " pod="openstack/manila-share-share1-0" Dec 13 08:00:10 crc kubenswrapper[4971]: I1213 08:00:10.481963 4971 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/c815d65d-6fc3-49c3-bda5-a2ea21b54438-etc-machine-id\") pod \"manila-scheduler-0\" (UID: \"c815d65d-6fc3-49c3-bda5-a2ea21b54438\") " pod="openstack/manila-scheduler-0" Dec 13 08:00:10 crc kubenswrapper[4971]: I1213 08:00:10.481991 4971 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/c815d65d-6fc3-49c3-bda5-a2ea21b54438-config-data-custom\") pod \"manila-scheduler-0\" (UID: \"c815d65d-6fc3-49c3-bda5-a2ea21b54438\") " pod="openstack/manila-scheduler-0" Dec 13 08:00:10 crc kubenswrapper[4971]: I1213 08:00:10.482013 4971 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/1d983511-44dd-4be9-9b43-2f496428a12e-etc-machine-id\") pod \"manila-share-share1-0\" (UID: \"1d983511-44dd-4be9-9b43-2f496428a12e\") " pod="openstack/manila-share-share1-0" Dec 13 08:00:10 crc kubenswrapper[4971]: I1213 08:00:10.482051 4971 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/c815d65d-6fc3-49c3-bda5-a2ea21b54438-config-data\") pod \"manila-scheduler-0\" (UID: \"c815d65d-6fc3-49c3-bda5-a2ea21b54438\") " pod="openstack/manila-scheduler-0" Dec 13 08:00:10 crc kubenswrapper[4971]: I1213 08:00:10.482085 4971 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ceph\" (UniqueName: \"kubernetes.io/projected/1d983511-44dd-4be9-9b43-2f496428a12e-ceph\") pod \"manila-share-share1-0\" (UID: \"1d983511-44dd-4be9-9b43-2f496428a12e\") " pod="openstack/manila-share-share1-0" Dec 13 08:00:10 crc kubenswrapper[4971]: I1213 08:00:10.494239 4971 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/manila-share-share1-0"] Dec 13 08:00:10 crc kubenswrapper[4971]: I1213 08:00:10.586075 4971 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/1d983511-44dd-4be9-9b43-2f496428a12e-config-data-custom\") pod \"manila-share-share1-0\" (UID: \"1d983511-44dd-4be9-9b43-2f496428a12e\") " pod="openstack/manila-share-share1-0" Dec 13 08:00:10 crc kubenswrapper[4971]: I1213 08:00:10.586127 4971 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/c815d65d-6fc3-49c3-bda5-a2ea21b54438-etc-machine-id\") pod \"manila-scheduler-0\" (UID: \"c815d65d-6fc3-49c3-bda5-a2ea21b54438\") " pod="openstack/manila-scheduler-0" Dec 13 08:00:10 crc kubenswrapper[4971]: I1213 08:00:10.586149 4971 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/c815d65d-6fc3-49c3-bda5-a2ea21b54438-config-data-custom\") pod \"manila-scheduler-0\" (UID: \"c815d65d-6fc3-49c3-bda5-a2ea21b54438\") " pod="openstack/manila-scheduler-0" Dec 13 08:00:10 crc kubenswrapper[4971]: I1213 08:00:10.586172 4971 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/1d983511-44dd-4be9-9b43-2f496428a12e-etc-machine-id\") pod \"manila-share-share1-0\" (UID: \"1d983511-44dd-4be9-9b43-2f496428a12e\") " pod="openstack/manila-share-share1-0" Dec 13 08:00:10 crc kubenswrapper[4971]: I1213 08:00:10.586194 4971 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/c815d65d-6fc3-49c3-bda5-a2ea21b54438-config-data\") pod \"manila-scheduler-0\" (UID: \"c815d65d-6fc3-49c3-bda5-a2ea21b54438\") " pod="openstack/manila-scheduler-0" Dec 13 08:00:10 crc kubenswrapper[4971]: I1213 08:00:10.586214 4971 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ceph\" (UniqueName: \"kubernetes.io/projected/1d983511-44dd-4be9-9b43-2f496428a12e-ceph\") pod \"manila-share-share1-0\" (UID: \"1d983511-44dd-4be9-9b43-2f496428a12e\") " pod="openstack/manila-share-share1-0" Dec 13 08:00:10 crc kubenswrapper[4971]: I1213 08:00:10.586253 4971 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/1d983511-44dd-4be9-9b43-2f496428a12e-combined-ca-bundle\") pod \"manila-share-share1-0\" (UID: \"1d983511-44dd-4be9-9b43-2f496428a12e\") " pod="openstack/manila-share-share1-0" Dec 13 08:00:10 crc kubenswrapper[4971]: I1213 08:00:10.586279 4971 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cb2g9\" (UniqueName: \"kubernetes.io/projected/1d983511-44dd-4be9-9b43-2f496428a12e-kube-api-access-cb2g9\") pod \"manila-share-share1-0\" (UID: \"1d983511-44dd-4be9-9b43-2f496428a12e\") " pod="openstack/manila-share-share1-0" Dec 13 08:00:10 crc kubenswrapper[4971]: I1213 08:00:10.586321 4971 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-qgmtc\" (UniqueName: \"kubernetes.io/projected/c815d65d-6fc3-49c3-bda5-a2ea21b54438-kube-api-access-qgmtc\") pod \"manila-scheduler-0\" (UID: \"c815d65d-6fc3-49c3-bda5-a2ea21b54438\") " pod="openstack/manila-scheduler-0" Dec 13 08:00:10 crc kubenswrapper[4971]: I1213 08:00:10.586344 4971 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/1d983511-44dd-4be9-9b43-2f496428a12e-scripts\") pod \"manila-share-share1-0\" (UID: \"1d983511-44dd-4be9-9b43-2f496428a12e\") " pod="openstack/manila-share-share1-0" Dec 13 08:00:10 crc kubenswrapper[4971]: I1213 08:00:10.586363 4971 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-lib-manila\" (UniqueName: \"kubernetes.io/host-path/1d983511-44dd-4be9-9b43-2f496428a12e-var-lib-manila\") pod \"manila-share-share1-0\" (UID: \"1d983511-44dd-4be9-9b43-2f496428a12e\") " pod="openstack/manila-share-share1-0" Dec 13 08:00:10 crc kubenswrapper[4971]: I1213 08:00:10.586383 4971 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/1d983511-44dd-4be9-9b43-2f496428a12e-config-data\") pod \"manila-share-share1-0\" (UID: \"1d983511-44dd-4be9-9b43-2f496428a12e\") " pod="openstack/manila-share-share1-0" Dec 13 08:00:10 crc kubenswrapper[4971]: I1213 08:00:10.586406 4971 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/c815d65d-6fc3-49c3-bda5-a2ea21b54438-scripts\") pod \"manila-scheduler-0\" (UID: \"c815d65d-6fc3-49c3-bda5-a2ea21b54438\") " pod="openstack/manila-scheduler-0" Dec 13 08:00:10 crc kubenswrapper[4971]: I1213 08:00:10.586439 4971 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/c815d65d-6fc3-49c3-bda5-a2ea21b54438-combined-ca-bundle\") pod \"manila-scheduler-0\" (UID: \"c815d65d-6fc3-49c3-bda5-a2ea21b54438\") " pod="openstack/manila-scheduler-0" Dec 13 08:00:10 crc kubenswrapper[4971]: I1213 08:00:10.590583 4971 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/c815d65d-6fc3-49c3-bda5-a2ea21b54438-etc-machine-id\") pod \"manila-scheduler-0\" (UID: \"c815d65d-6fc3-49c3-bda5-a2ea21b54438\") " pod="openstack/manila-scheduler-0" Dec 13 08:00:10 crc kubenswrapper[4971]: I1213 08:00:10.590680 4971 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/1d983511-44dd-4be9-9b43-2f496428a12e-etc-machine-id\") pod \"manila-share-share1-0\" (UID: \"1d983511-44dd-4be9-9b43-2f496428a12e\") " pod="openstack/manila-share-share1-0" Dec 13 08:00:10 crc kubenswrapper[4971]: I1213 08:00:10.590789 4971 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-lib-manila\" (UniqueName: \"kubernetes.io/host-path/1d983511-44dd-4be9-9b43-2f496428a12e-var-lib-manila\") pod \"manila-share-share1-0\" (UID: \"1d983511-44dd-4be9-9b43-2f496428a12e\") " pod="openstack/manila-share-share1-0" Dec 13 08:00:10 crc kubenswrapper[4971]: I1213 08:00:10.591831 4971 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/c815d65d-6fc3-49c3-bda5-a2ea21b54438-config-data\") pod \"manila-scheduler-0\" (UID: \"c815d65d-6fc3-49c3-bda5-a2ea21b54438\") " pod="openstack/manila-scheduler-0" Dec 13 08:00:10 crc kubenswrapper[4971]: I1213 08:00:10.595965 4971 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/1d983511-44dd-4be9-9b43-2f496428a12e-combined-ca-bundle\") pod \"manila-share-share1-0\" (UID: \"1d983511-44dd-4be9-9b43-2f496428a12e\") " pod="openstack/manila-share-share1-0" Dec 13 08:00:10 crc kubenswrapper[4971]: I1213 08:00:10.596026 4971 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/c815d65d-6fc3-49c3-bda5-a2ea21b54438-config-data-custom\") pod \"manila-scheduler-0\" (UID: \"c815d65d-6fc3-49c3-bda5-a2ea21b54438\") " pod="openstack/manila-scheduler-0" Dec 13 08:00:10 crc kubenswrapper[4971]: I1213 08:00:10.598561 4971 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/1d983511-44dd-4be9-9b43-2f496428a12e-config-data\") pod \"manila-share-share1-0\" (UID: \"1d983511-44dd-4be9-9b43-2f496428a12e\") " pod="openstack/manila-share-share1-0" Dec 13 08:00:10 crc kubenswrapper[4971]: I1213 08:00:10.599047 4971 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/1d983511-44dd-4be9-9b43-2f496428a12e-scripts\") pod \"manila-share-share1-0\" (UID: \"1d983511-44dd-4be9-9b43-2f496428a12e\") " pod="openstack/manila-share-share1-0" Dec 13 08:00:10 crc kubenswrapper[4971]: I1213 08:00:10.601840 4971 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/1d983511-44dd-4be9-9b43-2f496428a12e-config-data-custom\") pod \"manila-share-share1-0\" (UID: \"1d983511-44dd-4be9-9b43-2f496428a12e\") " pod="openstack/manila-share-share1-0" Dec 13 08:00:10 crc kubenswrapper[4971]: I1213 08:00:10.603029 4971 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/c815d65d-6fc3-49c3-bda5-a2ea21b54438-combined-ca-bundle\") pod \"manila-scheduler-0\" (UID: \"c815d65d-6fc3-49c3-bda5-a2ea21b54438\") " pod="openstack/manila-scheduler-0" Dec 13 08:00:10 crc kubenswrapper[4971]: I1213 08:00:10.616788 4971 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/c815d65d-6fc3-49c3-bda5-a2ea21b54438-scripts\") pod \"manila-scheduler-0\" (UID: \"c815d65d-6fc3-49c3-bda5-a2ea21b54438\") " pod="openstack/manila-scheduler-0" Dec 13 08:00:10 crc kubenswrapper[4971]: I1213 08:00:10.620432 4971 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-qgmtc\" (UniqueName: \"kubernetes.io/projected/c815d65d-6fc3-49c3-bda5-a2ea21b54438-kube-api-access-qgmtc\") pod \"manila-scheduler-0\" (UID: \"c815d65d-6fc3-49c3-bda5-a2ea21b54438\") " pod="openstack/manila-scheduler-0" Dec 13 08:00:10 crc kubenswrapper[4971]: I1213 08:00:10.621694 4971 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ceph\" (UniqueName: \"kubernetes.io/projected/1d983511-44dd-4be9-9b43-2f496428a12e-ceph\") pod \"manila-share-share1-0\" (UID: \"1d983511-44dd-4be9-9b43-2f496428a12e\") " pod="openstack/manila-share-share1-0" Dec 13 08:00:10 crc kubenswrapper[4971]: I1213 08:00:10.622161 4971 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-cb2g9\" (UniqueName: \"kubernetes.io/projected/1d983511-44dd-4be9-9b43-2f496428a12e-kube-api-access-cb2g9\") pod \"manila-share-share1-0\" (UID: \"1d983511-44dd-4be9-9b43-2f496428a12e\") " pod="openstack/manila-share-share1-0" Dec 13 08:00:10 crc kubenswrapper[4971]: I1213 08:00:10.671013 4971 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-ccc856b4f-zh8np"] Dec 13 08:00:10 crc kubenswrapper[4971]: I1213 08:00:10.674031 4971 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-ccc856b4f-zh8np" Dec 13 08:00:10 crc kubenswrapper[4971]: I1213 08:00:10.687895 4971 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-ccc856b4f-zh8np"] Dec 13 08:00:10 crc kubenswrapper[4971]: I1213 08:00:10.688091 4971 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-8fl9c\" (UniqueName: \"kubernetes.io/projected/409ac970-9fc4-4b77-9548-1714f54c8d51-kube-api-access-8fl9c\") pod \"dnsmasq-dns-ccc856b4f-zh8np\" (UID: \"409ac970-9fc4-4b77-9548-1714f54c8d51\") " pod="openstack/dnsmasq-dns-ccc856b4f-zh8np" Dec 13 08:00:10 crc kubenswrapper[4971]: I1213 08:00:10.688134 4971 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/configmap/409ac970-9fc4-4b77-9548-1714f54c8d51-openstack-edpm-ipam\") pod \"dnsmasq-dns-ccc856b4f-zh8np\" (UID: \"409ac970-9fc4-4b77-9548-1714f54c8d51\") " pod="openstack/dnsmasq-dns-ccc856b4f-zh8np" Dec 13 08:00:10 crc kubenswrapper[4971]: I1213 08:00:10.688184 4971 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/409ac970-9fc4-4b77-9548-1714f54c8d51-ovsdbserver-sb\") pod \"dnsmasq-dns-ccc856b4f-zh8np\" (UID: \"409ac970-9fc4-4b77-9548-1714f54c8d51\") " pod="openstack/dnsmasq-dns-ccc856b4f-zh8np" Dec 13 08:00:10 crc kubenswrapper[4971]: I1213 08:00:10.688228 4971 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/409ac970-9fc4-4b77-9548-1714f54c8d51-ovsdbserver-nb\") pod \"dnsmasq-dns-ccc856b4f-zh8np\" (UID: \"409ac970-9fc4-4b77-9548-1714f54c8d51\") " pod="openstack/dnsmasq-dns-ccc856b4f-zh8np" Dec 13 08:00:10 crc kubenswrapper[4971]: I1213 08:00:10.688272 4971 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/409ac970-9fc4-4b77-9548-1714f54c8d51-dns-svc\") pod \"dnsmasq-dns-ccc856b4f-zh8np\" (UID: \"409ac970-9fc4-4b77-9548-1714f54c8d51\") " pod="openstack/dnsmasq-dns-ccc856b4f-zh8np" Dec 13 08:00:10 crc kubenswrapper[4971]: I1213 08:00:10.688322 4971 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/409ac970-9fc4-4b77-9548-1714f54c8d51-config\") pod \"dnsmasq-dns-ccc856b4f-zh8np\" (UID: \"409ac970-9fc4-4b77-9548-1714f54c8d51\") " pod="openstack/dnsmasq-dns-ccc856b4f-zh8np" Dec 13 08:00:10 crc kubenswrapper[4971]: I1213 08:00:10.745144 4971 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/manila-scheduler-0" Dec 13 08:00:10 crc kubenswrapper[4971]: I1213 08:00:10.786790 4971 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/manila-api-0"] Dec 13 08:00:10 crc kubenswrapper[4971]: I1213 08:00:10.788433 4971 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/manila-api-0" Dec 13 08:00:10 crc kubenswrapper[4971]: I1213 08:00:10.790735 4971 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-8fl9c\" (UniqueName: \"kubernetes.io/projected/409ac970-9fc4-4b77-9548-1714f54c8d51-kube-api-access-8fl9c\") pod \"dnsmasq-dns-ccc856b4f-zh8np\" (UID: \"409ac970-9fc4-4b77-9548-1714f54c8d51\") " pod="openstack/dnsmasq-dns-ccc856b4f-zh8np" Dec 13 08:00:10 crc kubenswrapper[4971]: I1213 08:00:10.790790 4971 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/configmap/409ac970-9fc4-4b77-9548-1714f54c8d51-openstack-edpm-ipam\") pod \"dnsmasq-dns-ccc856b4f-zh8np\" (UID: \"409ac970-9fc4-4b77-9548-1714f54c8d51\") " pod="openstack/dnsmasq-dns-ccc856b4f-zh8np" Dec 13 08:00:10 crc kubenswrapper[4971]: I1213 08:00:10.790834 4971 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/409ac970-9fc4-4b77-9548-1714f54c8d51-ovsdbserver-sb\") pod \"dnsmasq-dns-ccc856b4f-zh8np\" (UID: \"409ac970-9fc4-4b77-9548-1714f54c8d51\") " pod="openstack/dnsmasq-dns-ccc856b4f-zh8np" Dec 13 08:00:10 crc kubenswrapper[4971]: I1213 08:00:10.790877 4971 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/409ac970-9fc4-4b77-9548-1714f54c8d51-ovsdbserver-nb\") pod \"dnsmasq-dns-ccc856b4f-zh8np\" (UID: \"409ac970-9fc4-4b77-9548-1714f54c8d51\") " pod="openstack/dnsmasq-dns-ccc856b4f-zh8np" Dec 13 08:00:10 crc kubenswrapper[4971]: I1213 08:00:10.790940 4971 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/409ac970-9fc4-4b77-9548-1714f54c8d51-dns-svc\") pod \"dnsmasq-dns-ccc856b4f-zh8np\" (UID: \"409ac970-9fc4-4b77-9548-1714f54c8d51\") " pod="openstack/dnsmasq-dns-ccc856b4f-zh8np" Dec 13 08:00:10 crc kubenswrapper[4971]: I1213 08:00:10.790997 4971 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/409ac970-9fc4-4b77-9548-1714f54c8d51-config\") pod \"dnsmasq-dns-ccc856b4f-zh8np\" (UID: \"409ac970-9fc4-4b77-9548-1714f54c8d51\") " pod="openstack/dnsmasq-dns-ccc856b4f-zh8np" Dec 13 08:00:10 crc kubenswrapper[4971]: I1213 08:00:10.792314 4971 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/409ac970-9fc4-4b77-9548-1714f54c8d51-config\") pod \"dnsmasq-dns-ccc856b4f-zh8np\" (UID: \"409ac970-9fc4-4b77-9548-1714f54c8d51\") " pod="openstack/dnsmasq-dns-ccc856b4f-zh8np" Dec 13 08:00:10 crc kubenswrapper[4971]: I1213 08:00:10.792959 4971 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/409ac970-9fc4-4b77-9548-1714f54c8d51-ovsdbserver-sb\") pod \"dnsmasq-dns-ccc856b4f-zh8np\" (UID: \"409ac970-9fc4-4b77-9548-1714f54c8d51\") " pod="openstack/dnsmasq-dns-ccc856b4f-zh8np" Dec 13 08:00:10 crc kubenswrapper[4971]: I1213 08:00:10.793383 4971 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/configmap/409ac970-9fc4-4b77-9548-1714f54c8d51-openstack-edpm-ipam\") pod \"dnsmasq-dns-ccc856b4f-zh8np\" (UID: \"409ac970-9fc4-4b77-9548-1714f54c8d51\") " pod="openstack/dnsmasq-dns-ccc856b4f-zh8np" Dec 13 08:00:10 crc kubenswrapper[4971]: I1213 08:00:10.793853 4971 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"manila-api-config-data" Dec 13 08:00:10 crc kubenswrapper[4971]: I1213 08:00:10.794694 4971 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/409ac970-9fc4-4b77-9548-1714f54c8d51-dns-svc\") pod \"dnsmasq-dns-ccc856b4f-zh8np\" (UID: \"409ac970-9fc4-4b77-9548-1714f54c8d51\") " pod="openstack/dnsmasq-dns-ccc856b4f-zh8np" Dec 13 08:00:10 crc kubenswrapper[4971]: I1213 08:00:10.797879 4971 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/409ac970-9fc4-4b77-9548-1714f54c8d51-ovsdbserver-nb\") pod \"dnsmasq-dns-ccc856b4f-zh8np\" (UID: \"409ac970-9fc4-4b77-9548-1714f54c8d51\") " pod="openstack/dnsmasq-dns-ccc856b4f-zh8np" Dec 13 08:00:10 crc kubenswrapper[4971]: I1213 08:00:10.807543 4971 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/manila-api-0"] Dec 13 08:00:10 crc kubenswrapper[4971]: I1213 08:00:10.812832 4971 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/manila-share-share1-0" Dec 13 08:00:10 crc kubenswrapper[4971]: I1213 08:00:10.818745 4971 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-8fl9c\" (UniqueName: \"kubernetes.io/projected/409ac970-9fc4-4b77-9548-1714f54c8d51-kube-api-access-8fl9c\") pod \"dnsmasq-dns-ccc856b4f-zh8np\" (UID: \"409ac970-9fc4-4b77-9548-1714f54c8d51\") " pod="openstack/dnsmasq-dns-ccc856b4f-zh8np" Dec 13 08:00:10 crc kubenswrapper[4971]: I1213 08:00:10.893393 4971 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-44fds\" (UniqueName: \"kubernetes.io/projected/7812e279-0ade-4807-9d0c-d43185e20bd8-kube-api-access-44fds\") pod \"manila-api-0\" (UID: \"7812e279-0ade-4807-9d0c-d43185e20bd8\") " pod="openstack/manila-api-0" Dec 13 08:00:10 crc kubenswrapper[4971]: I1213 08:00:10.893509 4971 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/7812e279-0ade-4807-9d0c-d43185e20bd8-combined-ca-bundle\") pod \"manila-api-0\" (UID: \"7812e279-0ade-4807-9d0c-d43185e20bd8\") " pod="openstack/manila-api-0" Dec 13 08:00:10 crc kubenswrapper[4971]: I1213 08:00:10.893550 4971 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/7812e279-0ade-4807-9d0c-d43185e20bd8-config-data-custom\") pod \"manila-api-0\" (UID: \"7812e279-0ade-4807-9d0c-d43185e20bd8\") " pod="openstack/manila-api-0" Dec 13 08:00:10 crc kubenswrapper[4971]: I1213 08:00:10.893584 4971 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/7812e279-0ade-4807-9d0c-d43185e20bd8-config-data\") pod \"manila-api-0\" (UID: \"7812e279-0ade-4807-9d0c-d43185e20bd8\") " pod="openstack/manila-api-0" Dec 13 08:00:10 crc kubenswrapper[4971]: I1213 08:00:10.893603 4971 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/7812e279-0ade-4807-9d0c-d43185e20bd8-scripts\") pod \"manila-api-0\" (UID: \"7812e279-0ade-4807-9d0c-d43185e20bd8\") " pod="openstack/manila-api-0" Dec 13 08:00:10 crc kubenswrapper[4971]: I1213 08:00:10.893805 4971 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/7812e279-0ade-4807-9d0c-d43185e20bd8-etc-machine-id\") pod \"manila-api-0\" (UID: \"7812e279-0ade-4807-9d0c-d43185e20bd8\") " pod="openstack/manila-api-0" Dec 13 08:00:10 crc kubenswrapper[4971]: I1213 08:00:10.894072 4971 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/7812e279-0ade-4807-9d0c-d43185e20bd8-logs\") pod \"manila-api-0\" (UID: \"7812e279-0ade-4807-9d0c-d43185e20bd8\") " pod="openstack/manila-api-0" Dec 13 08:00:10 crc kubenswrapper[4971]: I1213 08:00:10.995831 4971 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/7812e279-0ade-4807-9d0c-d43185e20bd8-combined-ca-bundle\") pod \"manila-api-0\" (UID: \"7812e279-0ade-4807-9d0c-d43185e20bd8\") " pod="openstack/manila-api-0" Dec 13 08:00:10 crc kubenswrapper[4971]: I1213 08:00:10.995875 4971 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/7812e279-0ade-4807-9d0c-d43185e20bd8-config-data-custom\") pod \"manila-api-0\" (UID: \"7812e279-0ade-4807-9d0c-d43185e20bd8\") " pod="openstack/manila-api-0" Dec 13 08:00:10 crc kubenswrapper[4971]: I1213 08:00:10.995910 4971 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/7812e279-0ade-4807-9d0c-d43185e20bd8-config-data\") pod \"manila-api-0\" (UID: \"7812e279-0ade-4807-9d0c-d43185e20bd8\") " pod="openstack/manila-api-0" Dec 13 08:00:10 crc kubenswrapper[4971]: I1213 08:00:10.995928 4971 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/7812e279-0ade-4807-9d0c-d43185e20bd8-scripts\") pod \"manila-api-0\" (UID: \"7812e279-0ade-4807-9d0c-d43185e20bd8\") " pod="openstack/manila-api-0" Dec 13 08:00:10 crc kubenswrapper[4971]: I1213 08:00:10.995959 4971 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/7812e279-0ade-4807-9d0c-d43185e20bd8-etc-machine-id\") pod \"manila-api-0\" (UID: \"7812e279-0ade-4807-9d0c-d43185e20bd8\") " pod="openstack/manila-api-0" Dec 13 08:00:10 crc kubenswrapper[4971]: I1213 08:00:10.996017 4971 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/7812e279-0ade-4807-9d0c-d43185e20bd8-logs\") pod \"manila-api-0\" (UID: \"7812e279-0ade-4807-9d0c-d43185e20bd8\") " pod="openstack/manila-api-0" Dec 13 08:00:10 crc kubenswrapper[4971]: I1213 08:00:10.996101 4971 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-44fds\" (UniqueName: \"kubernetes.io/projected/7812e279-0ade-4807-9d0c-d43185e20bd8-kube-api-access-44fds\") pod \"manila-api-0\" (UID: \"7812e279-0ade-4807-9d0c-d43185e20bd8\") " pod="openstack/manila-api-0" Dec 13 08:00:10 crc kubenswrapper[4971]: I1213 08:00:10.996981 4971 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/7812e279-0ade-4807-9d0c-d43185e20bd8-etc-machine-id\") pod \"manila-api-0\" (UID: \"7812e279-0ade-4807-9d0c-d43185e20bd8\") " pod="openstack/manila-api-0" Dec 13 08:00:10 crc kubenswrapper[4971]: I1213 08:00:10.997391 4971 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/7812e279-0ade-4807-9d0c-d43185e20bd8-logs\") pod \"manila-api-0\" (UID: \"7812e279-0ade-4807-9d0c-d43185e20bd8\") " pod="openstack/manila-api-0" Dec 13 08:00:11 crc kubenswrapper[4971]: I1213 08:00:11.003655 4971 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/7812e279-0ade-4807-9d0c-d43185e20bd8-combined-ca-bundle\") pod \"manila-api-0\" (UID: \"7812e279-0ade-4807-9d0c-d43185e20bd8\") " pod="openstack/manila-api-0" Dec 13 08:00:11 crc kubenswrapper[4971]: I1213 08:00:11.005754 4971 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/7812e279-0ade-4807-9d0c-d43185e20bd8-config-data-custom\") pod \"manila-api-0\" (UID: \"7812e279-0ade-4807-9d0c-d43185e20bd8\") " pod="openstack/manila-api-0" Dec 13 08:00:11 crc kubenswrapper[4971]: I1213 08:00:11.006183 4971 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/7812e279-0ade-4807-9d0c-d43185e20bd8-scripts\") pod \"manila-api-0\" (UID: \"7812e279-0ade-4807-9d0c-d43185e20bd8\") " pod="openstack/manila-api-0" Dec 13 08:00:11 crc kubenswrapper[4971]: I1213 08:00:11.006304 4971 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/7812e279-0ade-4807-9d0c-d43185e20bd8-config-data\") pod \"manila-api-0\" (UID: \"7812e279-0ade-4807-9d0c-d43185e20bd8\") " pod="openstack/manila-api-0" Dec 13 08:00:11 crc kubenswrapper[4971]: I1213 08:00:11.137382 4971 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-44fds\" (UniqueName: \"kubernetes.io/projected/7812e279-0ade-4807-9d0c-d43185e20bd8-kube-api-access-44fds\") pod \"manila-api-0\" (UID: \"7812e279-0ade-4807-9d0c-d43185e20bd8\") " pod="openstack/manila-api-0" Dec 13 08:00:11 crc kubenswrapper[4971]: I1213 08:00:11.139003 4971 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-ccc856b4f-zh8np" Dec 13 08:00:11 crc kubenswrapper[4971]: I1213 08:00:11.243003 4971 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/manila-api-0" Dec 13 08:00:11 crc kubenswrapper[4971]: I1213 08:00:11.490417 4971 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/manila-scheduler-0"] Dec 13 08:00:11 crc kubenswrapper[4971]: I1213 08:00:11.751841 4971 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/manila-share-share1-0"] Dec 13 08:00:11 crc kubenswrapper[4971]: I1213 08:00:11.939112 4971 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/manila-share-share1-0" event={"ID":"1d983511-44dd-4be9-9b43-2f496428a12e","Type":"ContainerStarted","Data":"2277c634165d87b0973f14d4d41924b9594b3ab384ceba39a9141415b72f416a"} Dec 13 08:00:11 crc kubenswrapper[4971]: I1213 08:00:11.940087 4971 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/manila-scheduler-0" event={"ID":"c815d65d-6fc3-49c3-bda5-a2ea21b54438","Type":"ContainerStarted","Data":"e4d41c347bfa8e9b048bdc28d14e951a680bd1dc6808ae6d20d790a6e7afb521"} Dec 13 08:00:12 crc kubenswrapper[4971]: I1213 08:00:12.036646 4971 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-ccc856b4f-zh8np"] Dec 13 08:00:12 crc kubenswrapper[4971]: I1213 08:00:12.046876 4971 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/manila-api-0"] Dec 13 08:00:12 crc kubenswrapper[4971]: W1213 08:00:12.048654 4971 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod7812e279_0ade_4807_9d0c_d43185e20bd8.slice/crio-ab03bc9c668d61ba53a6227afe38206184ae84298bce512245d1e891deb6ce96 WatchSource:0}: Error finding container ab03bc9c668d61ba53a6227afe38206184ae84298bce512245d1e891deb6ce96: Status 404 returned error can't find the container with id ab03bc9c668d61ba53a6227afe38206184ae84298bce512245d1e891deb6ce96 Dec 13 08:00:12 crc kubenswrapper[4971]: W1213 08:00:12.054830 4971 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod409ac970_9fc4_4b77_9548_1714f54c8d51.slice/crio-533192ba6448be46fb0bd83cefe051680a29106cfd9276ab70f6e10764a3d8fa WatchSource:0}: Error finding container 533192ba6448be46fb0bd83cefe051680a29106cfd9276ab70f6e10764a3d8fa: Status 404 returned error can't find the container with id 533192ba6448be46fb0bd83cefe051680a29106cfd9276ab70f6e10764a3d8fa Dec 13 08:00:12 crc kubenswrapper[4971]: E1213 08:00:12.813971 4971 cadvisor_stats_provider.go:516] "Partial failure issuing cadvisor.ContainerInfoV2" err="partial failures: [\"/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod409ac970_9fc4_4b77_9548_1714f54c8d51.slice/crio-f79125d5f1a3c175292b0e583039325c7f17f13eba1e4848a4ab1d11487ee677.scope\": RecentStats: unable to find data in memory cache], [\"/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod409ac970_9fc4_4b77_9548_1714f54c8d51.slice/crio-conmon-f79125d5f1a3c175292b0e583039325c7f17f13eba1e4848a4ab1d11487ee677.scope\": RecentStats: unable to find data in memory cache]" Dec 13 08:00:12 crc kubenswrapper[4971]: I1213 08:00:12.952209 4971 generic.go:334] "Generic (PLEG): container finished" podID="409ac970-9fc4-4b77-9548-1714f54c8d51" containerID="f79125d5f1a3c175292b0e583039325c7f17f13eba1e4848a4ab1d11487ee677" exitCode=0 Dec 13 08:00:12 crc kubenswrapper[4971]: I1213 08:00:12.952261 4971 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-ccc856b4f-zh8np" event={"ID":"409ac970-9fc4-4b77-9548-1714f54c8d51","Type":"ContainerDied","Data":"f79125d5f1a3c175292b0e583039325c7f17f13eba1e4848a4ab1d11487ee677"} Dec 13 08:00:12 crc kubenswrapper[4971]: I1213 08:00:12.952308 4971 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-ccc856b4f-zh8np" event={"ID":"409ac970-9fc4-4b77-9548-1714f54c8d51","Type":"ContainerStarted","Data":"533192ba6448be46fb0bd83cefe051680a29106cfd9276ab70f6e10764a3d8fa"} Dec 13 08:00:12 crc kubenswrapper[4971]: I1213 08:00:12.957830 4971 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/manila-api-0" event={"ID":"7812e279-0ade-4807-9d0c-d43185e20bd8","Type":"ContainerStarted","Data":"20c6089c0a2d579f6703a936ef517f63451e4e25475c2e0b36ab650a0abd060a"} Dec 13 08:00:12 crc kubenswrapper[4971]: I1213 08:00:12.958307 4971 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/manila-api-0" event={"ID":"7812e279-0ade-4807-9d0c-d43185e20bd8","Type":"ContainerStarted","Data":"ab03bc9c668d61ba53a6227afe38206184ae84298bce512245d1e891deb6ce96"} Dec 13 08:00:13 crc kubenswrapper[4971]: I1213 08:00:13.987056 4971 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/manila-scheduler-0" event={"ID":"c815d65d-6fc3-49c3-bda5-a2ea21b54438","Type":"ContainerStarted","Data":"07f06b3979e8391497d4a21a6994af9fbae02b2bb2e5f2ff570057d31fb6ab8f"} Dec 13 08:00:13 crc kubenswrapper[4971]: I1213 08:00:13.994324 4971 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-ccc856b4f-zh8np" event={"ID":"409ac970-9fc4-4b77-9548-1714f54c8d51","Type":"ContainerStarted","Data":"43e5531fefc9e18591c9a2e8c3577db47533cb21206926f95d1d52da37542102"} Dec 13 08:00:13 crc kubenswrapper[4971]: I1213 08:00:13.994591 4971 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/dnsmasq-dns-ccc856b4f-zh8np" Dec 13 08:00:13 crc kubenswrapper[4971]: I1213 08:00:13.999216 4971 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/manila-api-0" event={"ID":"7812e279-0ade-4807-9d0c-d43185e20bd8","Type":"ContainerStarted","Data":"1e11a912c27d298ee4050c94e5294240a78d9b7cea657d1e457a83618e46d6ee"} Dec 13 08:00:13 crc kubenswrapper[4971]: I1213 08:00:13.999389 4971 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/manila-api-0" Dec 13 08:00:14 crc kubenswrapper[4971]: I1213 08:00:14.020072 4971 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/dnsmasq-dns-ccc856b4f-zh8np" podStartSLOduration=4.020055666 podStartE2EDuration="4.020055666s" podCreationTimestamp="2025-12-13 08:00:10 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-13 08:00:14.017178706 +0000 UTC m=+4270.621588154" watchObservedRunningTime="2025-12-13 08:00:14.020055666 +0000 UTC m=+4270.624465104" Dec 13 08:00:14 crc kubenswrapper[4971]: I1213 08:00:14.050938 4971 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/manila-api-0" podStartSLOduration=4.050916423 podStartE2EDuration="4.050916423s" podCreationTimestamp="2025-12-13 08:00:10 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-13 08:00:14.041982817 +0000 UTC m=+4270.646392265" watchObservedRunningTime="2025-12-13 08:00:14.050916423 +0000 UTC m=+4270.655325871" Dec 13 08:00:14 crc kubenswrapper[4971]: I1213 08:00:14.501937 4971 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/manila-api-0"] Dec 13 08:00:16 crc kubenswrapper[4971]: I1213 08:00:16.021433 4971 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/manila-api-0" podUID="7812e279-0ade-4807-9d0c-d43185e20bd8" containerName="manila-api-log" containerID="cri-o://20c6089c0a2d579f6703a936ef517f63451e4e25475c2e0b36ab650a0abd060a" gracePeriod=30 Dec 13 08:00:16 crc kubenswrapper[4971]: I1213 08:00:16.021860 4971 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/manila-api-0" podUID="7812e279-0ade-4807-9d0c-d43185e20bd8" containerName="manila-api" containerID="cri-o://1e11a912c27d298ee4050c94e5294240a78d9b7cea657d1e457a83618e46d6ee" gracePeriod=30 Dec 13 08:00:16 crc kubenswrapper[4971]: I1213 08:00:16.021943 4971 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/manila-scheduler-0" event={"ID":"c815d65d-6fc3-49c3-bda5-a2ea21b54438","Type":"ContainerStarted","Data":"303b59dbb886a69396d195934ac67cd67928e83e66852341c56d708cc4b419cf"} Dec 13 08:00:16 crc kubenswrapper[4971]: I1213 08:00:16.057026 4971 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/manila-scheduler-0" podStartSLOduration=4.61847049 podStartE2EDuration="6.057003729s" podCreationTimestamp="2025-12-13 08:00:10 +0000 UTC" firstStartedPulling="2025-12-13 08:00:11.518706562 +0000 UTC m=+4268.123116010" lastFinishedPulling="2025-12-13 08:00:12.957239811 +0000 UTC m=+4269.561649249" observedRunningTime="2025-12-13 08:00:16.044452236 +0000 UTC m=+4272.648861704" watchObservedRunningTime="2025-12-13 08:00:16.057003729 +0000 UTC m=+4272.661413187" Dec 13 08:00:17 crc kubenswrapper[4971]: I1213 08:00:17.043848 4971 generic.go:334] "Generic (PLEG): container finished" podID="7812e279-0ade-4807-9d0c-d43185e20bd8" containerID="1e11a912c27d298ee4050c94e5294240a78d9b7cea657d1e457a83618e46d6ee" exitCode=0 Dec 13 08:00:17 crc kubenswrapper[4971]: I1213 08:00:17.044141 4971 generic.go:334] "Generic (PLEG): container finished" podID="7812e279-0ade-4807-9d0c-d43185e20bd8" containerID="20c6089c0a2d579f6703a936ef517f63451e4e25475c2e0b36ab650a0abd060a" exitCode=143 Dec 13 08:00:17 crc kubenswrapper[4971]: I1213 08:00:17.045065 4971 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/manila-api-0" event={"ID":"7812e279-0ade-4807-9d0c-d43185e20bd8","Type":"ContainerDied","Data":"1e11a912c27d298ee4050c94e5294240a78d9b7cea657d1e457a83618e46d6ee"} Dec 13 08:00:17 crc kubenswrapper[4971]: I1213 08:00:17.045097 4971 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/manila-api-0" event={"ID":"7812e279-0ade-4807-9d0c-d43185e20bd8","Type":"ContainerDied","Data":"20c6089c0a2d579f6703a936ef517f63451e4e25475c2e0b36ab650a0abd060a"} Dec 13 08:00:17 crc kubenswrapper[4971]: I1213 08:00:17.465652 4971 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/manila-api-0" Dec 13 08:00:17 crc kubenswrapper[4971]: I1213 08:00:17.615699 4971 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/7812e279-0ade-4807-9d0c-d43185e20bd8-etc-machine-id\") pod \"7812e279-0ade-4807-9d0c-d43185e20bd8\" (UID: \"7812e279-0ade-4807-9d0c-d43185e20bd8\") " Dec 13 08:00:17 crc kubenswrapper[4971]: I1213 08:00:17.615776 4971 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/7812e279-0ade-4807-9d0c-d43185e20bd8-config-data-custom\") pod \"7812e279-0ade-4807-9d0c-d43185e20bd8\" (UID: \"7812e279-0ade-4807-9d0c-d43185e20bd8\") " Dec 13 08:00:17 crc kubenswrapper[4971]: I1213 08:00:17.616100 4971 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/7812e279-0ade-4807-9d0c-d43185e20bd8-scripts\") pod \"7812e279-0ade-4807-9d0c-d43185e20bd8\" (UID: \"7812e279-0ade-4807-9d0c-d43185e20bd8\") " Dec 13 08:00:17 crc kubenswrapper[4971]: I1213 08:00:17.616322 4971 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/7812e279-0ade-4807-9d0c-d43185e20bd8-logs\") pod \"7812e279-0ade-4807-9d0c-d43185e20bd8\" (UID: \"7812e279-0ade-4807-9d0c-d43185e20bd8\") " Dec 13 08:00:17 crc kubenswrapper[4971]: I1213 08:00:17.616449 4971 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/7812e279-0ade-4807-9d0c-d43185e20bd8-config-data\") pod \"7812e279-0ade-4807-9d0c-d43185e20bd8\" (UID: \"7812e279-0ade-4807-9d0c-d43185e20bd8\") " Dec 13 08:00:17 crc kubenswrapper[4971]: I1213 08:00:17.616489 4971 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/7812e279-0ade-4807-9d0c-d43185e20bd8-combined-ca-bundle\") pod \"7812e279-0ade-4807-9d0c-d43185e20bd8\" (UID: \"7812e279-0ade-4807-9d0c-d43185e20bd8\") " Dec 13 08:00:17 crc kubenswrapper[4971]: I1213 08:00:17.616554 4971 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-44fds\" (UniqueName: \"kubernetes.io/projected/7812e279-0ade-4807-9d0c-d43185e20bd8-kube-api-access-44fds\") pod \"7812e279-0ade-4807-9d0c-d43185e20bd8\" (UID: \"7812e279-0ade-4807-9d0c-d43185e20bd8\") " Dec 13 08:00:17 crc kubenswrapper[4971]: I1213 08:00:17.618076 4971 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/7812e279-0ade-4807-9d0c-d43185e20bd8-logs" (OuterVolumeSpecName: "logs") pod "7812e279-0ade-4807-9d0c-d43185e20bd8" (UID: "7812e279-0ade-4807-9d0c-d43185e20bd8"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 13 08:00:17 crc kubenswrapper[4971]: I1213 08:00:17.618130 4971 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/7812e279-0ade-4807-9d0c-d43185e20bd8-etc-machine-id" (OuterVolumeSpecName: "etc-machine-id") pod "7812e279-0ade-4807-9d0c-d43185e20bd8" (UID: "7812e279-0ade-4807-9d0c-d43185e20bd8"). InnerVolumeSpecName "etc-machine-id". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 13 08:00:17 crc kubenswrapper[4971]: I1213 08:00:17.623438 4971 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/7812e279-0ade-4807-9d0c-d43185e20bd8-config-data-custom" (OuterVolumeSpecName: "config-data-custom") pod "7812e279-0ade-4807-9d0c-d43185e20bd8" (UID: "7812e279-0ade-4807-9d0c-d43185e20bd8"). InnerVolumeSpecName "config-data-custom". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 13 08:00:17 crc kubenswrapper[4971]: I1213 08:00:17.623595 4971 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/7812e279-0ade-4807-9d0c-d43185e20bd8-kube-api-access-44fds" (OuterVolumeSpecName: "kube-api-access-44fds") pod "7812e279-0ade-4807-9d0c-d43185e20bd8" (UID: "7812e279-0ade-4807-9d0c-d43185e20bd8"). InnerVolumeSpecName "kube-api-access-44fds". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 13 08:00:17 crc kubenswrapper[4971]: I1213 08:00:17.637872 4971 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/7812e279-0ade-4807-9d0c-d43185e20bd8-scripts" (OuterVolumeSpecName: "scripts") pod "7812e279-0ade-4807-9d0c-d43185e20bd8" (UID: "7812e279-0ade-4807-9d0c-d43185e20bd8"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 13 08:00:17 crc kubenswrapper[4971]: I1213 08:00:17.661185 4971 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/7812e279-0ade-4807-9d0c-d43185e20bd8-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "7812e279-0ade-4807-9d0c-d43185e20bd8" (UID: "7812e279-0ade-4807-9d0c-d43185e20bd8"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 13 08:00:17 crc kubenswrapper[4971]: I1213 08:00:17.673638 4971 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/7812e279-0ade-4807-9d0c-d43185e20bd8-config-data" (OuterVolumeSpecName: "config-data") pod "7812e279-0ade-4807-9d0c-d43185e20bd8" (UID: "7812e279-0ade-4807-9d0c-d43185e20bd8"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 13 08:00:17 crc kubenswrapper[4971]: I1213 08:00:17.719360 4971 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/7812e279-0ade-4807-9d0c-d43185e20bd8-logs\") on node \"crc\" DevicePath \"\"" Dec 13 08:00:17 crc kubenswrapper[4971]: I1213 08:00:17.719406 4971 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/7812e279-0ade-4807-9d0c-d43185e20bd8-config-data\") on node \"crc\" DevicePath \"\"" Dec 13 08:00:17 crc kubenswrapper[4971]: I1213 08:00:17.719419 4971 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/7812e279-0ade-4807-9d0c-d43185e20bd8-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 13 08:00:17 crc kubenswrapper[4971]: I1213 08:00:17.719435 4971 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-44fds\" (UniqueName: \"kubernetes.io/projected/7812e279-0ade-4807-9d0c-d43185e20bd8-kube-api-access-44fds\") on node \"crc\" DevicePath \"\"" Dec 13 08:00:17 crc kubenswrapper[4971]: I1213 08:00:17.719450 4971 reconciler_common.go:293] "Volume detached for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/7812e279-0ade-4807-9d0c-d43185e20bd8-etc-machine-id\") on node \"crc\" DevicePath \"\"" Dec 13 08:00:17 crc kubenswrapper[4971]: I1213 08:00:17.719461 4971 reconciler_common.go:293] "Volume detached for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/7812e279-0ade-4807-9d0c-d43185e20bd8-config-data-custom\") on node \"crc\" DevicePath \"\"" Dec 13 08:00:17 crc kubenswrapper[4971]: I1213 08:00:17.719471 4971 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/7812e279-0ade-4807-9d0c-d43185e20bd8-scripts\") on node \"crc\" DevicePath \"\"" Dec 13 08:00:18 crc kubenswrapper[4971]: I1213 08:00:18.057503 4971 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/manila-api-0" event={"ID":"7812e279-0ade-4807-9d0c-d43185e20bd8","Type":"ContainerDied","Data":"ab03bc9c668d61ba53a6227afe38206184ae84298bce512245d1e891deb6ce96"} Dec 13 08:00:18 crc kubenswrapper[4971]: I1213 08:00:18.057567 4971 scope.go:117] "RemoveContainer" containerID="1e11a912c27d298ee4050c94e5294240a78d9b7cea657d1e457a83618e46d6ee" Dec 13 08:00:18 crc kubenswrapper[4971]: I1213 08:00:18.057621 4971 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/manila-api-0" Dec 13 08:00:18 crc kubenswrapper[4971]: I1213 08:00:18.086761 4971 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/manila-api-0"] Dec 13 08:00:18 crc kubenswrapper[4971]: I1213 08:00:18.095165 4971 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/manila-api-0"] Dec 13 08:00:18 crc kubenswrapper[4971]: I1213 08:00:18.110068 4971 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/manila-api-0"] Dec 13 08:00:18 crc kubenswrapper[4971]: E1213 08:00:18.110470 4971 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="7812e279-0ade-4807-9d0c-d43185e20bd8" containerName="manila-api" Dec 13 08:00:18 crc kubenswrapper[4971]: I1213 08:00:18.110488 4971 state_mem.go:107] "Deleted CPUSet assignment" podUID="7812e279-0ade-4807-9d0c-d43185e20bd8" containerName="manila-api" Dec 13 08:00:18 crc kubenswrapper[4971]: E1213 08:00:18.110508 4971 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="7812e279-0ade-4807-9d0c-d43185e20bd8" containerName="manila-api-log" Dec 13 08:00:18 crc kubenswrapper[4971]: I1213 08:00:18.110531 4971 state_mem.go:107] "Deleted CPUSet assignment" podUID="7812e279-0ade-4807-9d0c-d43185e20bd8" containerName="manila-api-log" Dec 13 08:00:18 crc kubenswrapper[4971]: I1213 08:00:18.110749 4971 memory_manager.go:354] "RemoveStaleState removing state" podUID="7812e279-0ade-4807-9d0c-d43185e20bd8" containerName="manila-api" Dec 13 08:00:18 crc kubenswrapper[4971]: I1213 08:00:18.110778 4971 memory_manager.go:354] "RemoveStaleState removing state" podUID="7812e279-0ade-4807-9d0c-d43185e20bd8" containerName="manila-api-log" Dec 13 08:00:18 crc kubenswrapper[4971]: I1213 08:00:18.111157 4971 scope.go:117] "RemoveContainer" containerID="20c6089c0a2d579f6703a936ef517f63451e4e25475c2e0b36ab650a0abd060a" Dec 13 08:00:18 crc kubenswrapper[4971]: I1213 08:00:18.111783 4971 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/manila-api-0" Dec 13 08:00:18 crc kubenswrapper[4971]: I1213 08:00:18.117149 4971 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-manila-internal-svc" Dec 13 08:00:18 crc kubenswrapper[4971]: I1213 08:00:18.117254 4971 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-manila-public-svc" Dec 13 08:00:18 crc kubenswrapper[4971]: I1213 08:00:18.117381 4971 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"manila-api-config-data" Dec 13 08:00:18 crc kubenswrapper[4971]: I1213 08:00:18.126816 4971 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/manila-api-0"] Dec 13 08:00:18 crc kubenswrapper[4971]: I1213 08:00:18.159096 4971 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/e8e297bb-43e7-45b9-a82b-a46c104519c1-config-data\") pod \"manila-api-0\" (UID: \"e8e297bb-43e7-45b9-a82b-a46c104519c1\") " pod="openstack/manila-api-0" Dec 13 08:00:18 crc kubenswrapper[4971]: I1213 08:00:18.159133 4971 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/e8e297bb-43e7-45b9-a82b-a46c104519c1-public-tls-certs\") pod \"manila-api-0\" (UID: \"e8e297bb-43e7-45b9-a82b-a46c104519c1\") " pod="openstack/manila-api-0" Dec 13 08:00:18 crc kubenswrapper[4971]: I1213 08:00:18.159169 4971 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/e8e297bb-43e7-45b9-a82b-a46c104519c1-combined-ca-bundle\") pod \"manila-api-0\" (UID: \"e8e297bb-43e7-45b9-a82b-a46c104519c1\") " pod="openstack/manila-api-0" Dec 13 08:00:18 crc kubenswrapper[4971]: I1213 08:00:18.159198 4971 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/e8e297bb-43e7-45b9-a82b-a46c104519c1-internal-tls-certs\") pod \"manila-api-0\" (UID: \"e8e297bb-43e7-45b9-a82b-a46c104519c1\") " pod="openstack/manila-api-0" Dec 13 08:00:18 crc kubenswrapper[4971]: I1213 08:00:18.159219 4971 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/e8e297bb-43e7-45b9-a82b-a46c104519c1-logs\") pod \"manila-api-0\" (UID: \"e8e297bb-43e7-45b9-a82b-a46c104519c1\") " pod="openstack/manila-api-0" Dec 13 08:00:18 crc kubenswrapper[4971]: I1213 08:00:18.159236 4971 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/e8e297bb-43e7-45b9-a82b-a46c104519c1-config-data-custom\") pod \"manila-api-0\" (UID: \"e8e297bb-43e7-45b9-a82b-a46c104519c1\") " pod="openstack/manila-api-0" Dec 13 08:00:18 crc kubenswrapper[4971]: I1213 08:00:18.159264 4971 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-kssdl\" (UniqueName: \"kubernetes.io/projected/e8e297bb-43e7-45b9-a82b-a46c104519c1-kube-api-access-kssdl\") pod \"manila-api-0\" (UID: \"e8e297bb-43e7-45b9-a82b-a46c104519c1\") " pod="openstack/manila-api-0" Dec 13 08:00:18 crc kubenswrapper[4971]: I1213 08:00:18.159332 4971 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/e8e297bb-43e7-45b9-a82b-a46c104519c1-etc-machine-id\") pod \"manila-api-0\" (UID: \"e8e297bb-43e7-45b9-a82b-a46c104519c1\") " pod="openstack/manila-api-0" Dec 13 08:00:18 crc kubenswrapper[4971]: I1213 08:00:18.159347 4971 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/e8e297bb-43e7-45b9-a82b-a46c104519c1-scripts\") pod \"manila-api-0\" (UID: \"e8e297bb-43e7-45b9-a82b-a46c104519c1\") " pod="openstack/manila-api-0" Dec 13 08:00:18 crc kubenswrapper[4971]: I1213 08:00:18.265070 4971 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/e8e297bb-43e7-45b9-a82b-a46c104519c1-logs\") pod \"manila-api-0\" (UID: \"e8e297bb-43e7-45b9-a82b-a46c104519c1\") " pod="openstack/manila-api-0" Dec 13 08:00:18 crc kubenswrapper[4971]: I1213 08:00:18.265152 4971 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/e8e297bb-43e7-45b9-a82b-a46c104519c1-config-data-custom\") pod \"manila-api-0\" (UID: \"e8e297bb-43e7-45b9-a82b-a46c104519c1\") " pod="openstack/manila-api-0" Dec 13 08:00:18 crc kubenswrapper[4971]: I1213 08:00:18.265212 4971 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-kssdl\" (UniqueName: \"kubernetes.io/projected/e8e297bb-43e7-45b9-a82b-a46c104519c1-kube-api-access-kssdl\") pod \"manila-api-0\" (UID: \"e8e297bb-43e7-45b9-a82b-a46c104519c1\") " pod="openstack/manila-api-0" Dec 13 08:00:18 crc kubenswrapper[4971]: I1213 08:00:18.265352 4971 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/e8e297bb-43e7-45b9-a82b-a46c104519c1-etc-machine-id\") pod \"manila-api-0\" (UID: \"e8e297bb-43e7-45b9-a82b-a46c104519c1\") " pod="openstack/manila-api-0" Dec 13 08:00:18 crc kubenswrapper[4971]: I1213 08:00:18.265380 4971 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/e8e297bb-43e7-45b9-a82b-a46c104519c1-scripts\") pod \"manila-api-0\" (UID: \"e8e297bb-43e7-45b9-a82b-a46c104519c1\") " pod="openstack/manila-api-0" Dec 13 08:00:18 crc kubenswrapper[4971]: I1213 08:00:18.265494 4971 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/e8e297bb-43e7-45b9-a82b-a46c104519c1-config-data\") pod \"manila-api-0\" (UID: \"e8e297bb-43e7-45b9-a82b-a46c104519c1\") " pod="openstack/manila-api-0" Dec 13 08:00:18 crc kubenswrapper[4971]: I1213 08:00:18.265541 4971 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/e8e297bb-43e7-45b9-a82b-a46c104519c1-public-tls-certs\") pod \"manila-api-0\" (UID: \"e8e297bb-43e7-45b9-a82b-a46c104519c1\") " pod="openstack/manila-api-0" Dec 13 08:00:18 crc kubenswrapper[4971]: I1213 08:00:18.265599 4971 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/e8e297bb-43e7-45b9-a82b-a46c104519c1-combined-ca-bundle\") pod \"manila-api-0\" (UID: \"e8e297bb-43e7-45b9-a82b-a46c104519c1\") " pod="openstack/manila-api-0" Dec 13 08:00:18 crc kubenswrapper[4971]: I1213 08:00:18.265641 4971 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/e8e297bb-43e7-45b9-a82b-a46c104519c1-internal-tls-certs\") pod \"manila-api-0\" (UID: \"e8e297bb-43e7-45b9-a82b-a46c104519c1\") " pod="openstack/manila-api-0" Dec 13 08:00:18 crc kubenswrapper[4971]: I1213 08:00:18.265721 4971 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/e8e297bb-43e7-45b9-a82b-a46c104519c1-logs\") pod \"manila-api-0\" (UID: \"e8e297bb-43e7-45b9-a82b-a46c104519c1\") " pod="openstack/manila-api-0" Dec 13 08:00:18 crc kubenswrapper[4971]: I1213 08:00:18.266844 4971 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/e8e297bb-43e7-45b9-a82b-a46c104519c1-etc-machine-id\") pod \"manila-api-0\" (UID: \"e8e297bb-43e7-45b9-a82b-a46c104519c1\") " pod="openstack/manila-api-0" Dec 13 08:00:18 crc kubenswrapper[4971]: I1213 08:00:18.283483 4971 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/e8e297bb-43e7-45b9-a82b-a46c104519c1-combined-ca-bundle\") pod \"manila-api-0\" (UID: \"e8e297bb-43e7-45b9-a82b-a46c104519c1\") " pod="openstack/manila-api-0" Dec 13 08:00:18 crc kubenswrapper[4971]: I1213 08:00:18.285508 4971 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/e8e297bb-43e7-45b9-a82b-a46c104519c1-config-data\") pod \"manila-api-0\" (UID: \"e8e297bb-43e7-45b9-a82b-a46c104519c1\") " pod="openstack/manila-api-0" Dec 13 08:00:18 crc kubenswrapper[4971]: I1213 08:00:18.296285 4971 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/e8e297bb-43e7-45b9-a82b-a46c104519c1-config-data-custom\") pod \"manila-api-0\" (UID: \"e8e297bb-43e7-45b9-a82b-a46c104519c1\") " pod="openstack/manila-api-0" Dec 13 08:00:18 crc kubenswrapper[4971]: I1213 08:00:18.305137 4971 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/e8e297bb-43e7-45b9-a82b-a46c104519c1-public-tls-certs\") pod \"manila-api-0\" (UID: \"e8e297bb-43e7-45b9-a82b-a46c104519c1\") " pod="openstack/manila-api-0" Dec 13 08:00:18 crc kubenswrapper[4971]: I1213 08:00:18.326288 4971 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/e8e297bb-43e7-45b9-a82b-a46c104519c1-scripts\") pod \"manila-api-0\" (UID: \"e8e297bb-43e7-45b9-a82b-a46c104519c1\") " pod="openstack/manila-api-0" Dec 13 08:00:18 crc kubenswrapper[4971]: I1213 08:00:18.327045 4971 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/e8e297bb-43e7-45b9-a82b-a46c104519c1-internal-tls-certs\") pod \"manila-api-0\" (UID: \"e8e297bb-43e7-45b9-a82b-a46c104519c1\") " pod="openstack/manila-api-0" Dec 13 08:00:18 crc kubenswrapper[4971]: I1213 08:00:18.336768 4971 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-kssdl\" (UniqueName: \"kubernetes.io/projected/e8e297bb-43e7-45b9-a82b-a46c104519c1-kube-api-access-kssdl\") pod \"manila-api-0\" (UID: \"e8e297bb-43e7-45b9-a82b-a46c104519c1\") " pod="openstack/manila-api-0" Dec 13 08:00:18 crc kubenswrapper[4971]: I1213 08:00:18.441018 4971 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/manila-api-0" Dec 13 08:00:19 crc kubenswrapper[4971]: I1213 08:00:19.067803 4971 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/manila-api-0"] Dec 13 08:00:19 crc kubenswrapper[4971]: W1213 08:00:19.082738 4971 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pode8e297bb_43e7_45b9_a82b_a46c104519c1.slice/crio-31025d8cbd4aa3a3853461fb8e1c342316d460d60ecc7d71ba3f6e20c9059be1 WatchSource:0}: Error finding container 31025d8cbd4aa3a3853461fb8e1c342316d460d60ecc7d71ba3f6e20c9059be1: Status 404 returned error can't find the container with id 31025d8cbd4aa3a3853461fb8e1c342316d460d60ecc7d71ba3f6e20c9059be1 Dec 13 08:00:19 crc kubenswrapper[4971]: I1213 08:00:19.849662 4971 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="7812e279-0ade-4807-9d0c-d43185e20bd8" path="/var/lib/kubelet/pods/7812e279-0ade-4807-9d0c-d43185e20bd8/volumes" Dec 13 08:00:20 crc kubenswrapper[4971]: I1213 08:00:20.078012 4971 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/manila-api-0" event={"ID":"e8e297bb-43e7-45b9-a82b-a46c104519c1","Type":"ContainerStarted","Data":"0392320f63717586dbb3c91f0abfbdbe5b9eefc49593849637ed80d458a3b702"} Dec 13 08:00:20 crc kubenswrapper[4971]: I1213 08:00:20.078335 4971 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/manila-api-0" event={"ID":"e8e297bb-43e7-45b9-a82b-a46c104519c1","Type":"ContainerStarted","Data":"31025d8cbd4aa3a3853461fb8e1c342316d460d60ecc7d71ba3f6e20c9059be1"} Dec 13 08:00:20 crc kubenswrapper[4971]: I1213 08:00:20.505437 4971 scope.go:117] "RemoveContainer" containerID="6334e16788385b038665f26911754d614f73fb16854784b3497d817139e836da" Dec 13 08:00:20 crc kubenswrapper[4971]: I1213 08:00:20.746126 4971 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/manila-scheduler-0" Dec 13 08:00:21 crc kubenswrapper[4971]: I1213 08:00:21.141724 4971 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/dnsmasq-dns-ccc856b4f-zh8np" Dec 13 08:00:21 crc kubenswrapper[4971]: I1213 08:00:21.232153 4971 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-55d9fc99f5-565xd"] Dec 13 08:00:21 crc kubenswrapper[4971]: I1213 08:00:21.232389 4971 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/dnsmasq-dns-55d9fc99f5-565xd" podUID="27e4748a-9fc8-4526-9b48-0d3e30f24268" containerName="dnsmasq-dns" containerID="cri-o://80f9dad47026016e812f4cd0fc75f8b495a90d7f6dd16030d03e794885c387fa" gracePeriod=10 Dec 13 08:00:22 crc kubenswrapper[4971]: I1213 08:00:22.225384 4971 generic.go:334] "Generic (PLEG): container finished" podID="27e4748a-9fc8-4526-9b48-0d3e30f24268" containerID="80f9dad47026016e812f4cd0fc75f8b495a90d7f6dd16030d03e794885c387fa" exitCode=0 Dec 13 08:00:22 crc kubenswrapper[4971]: I1213 08:00:22.225439 4971 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-55d9fc99f5-565xd" event={"ID":"27e4748a-9fc8-4526-9b48-0d3e30f24268","Type":"ContainerDied","Data":"80f9dad47026016e812f4cd0fc75f8b495a90d7f6dd16030d03e794885c387fa"} Dec 13 08:00:23 crc kubenswrapper[4971]: I1213 08:00:23.578664 4971 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/dnsmasq-dns-55d9fc99f5-565xd" podUID="27e4748a-9fc8-4526-9b48-0d3e30f24268" containerName="dnsmasq-dns" probeResult="failure" output="dial tcp 10.217.0.197:5353: connect: connection refused" Dec 13 08:00:24 crc kubenswrapper[4971]: I1213 08:00:24.040487 4971 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-55d9fc99f5-565xd" Dec 13 08:00:24 crc kubenswrapper[4971]: I1213 08:00:24.127545 4971 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/27e4748a-9fc8-4526-9b48-0d3e30f24268-dns-svc\") pod \"27e4748a-9fc8-4526-9b48-0d3e30f24268\" (UID: \"27e4748a-9fc8-4526-9b48-0d3e30f24268\") " Dec 13 08:00:24 crc kubenswrapper[4971]: I1213 08:00:24.127657 4971 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/27e4748a-9fc8-4526-9b48-0d3e30f24268-ovsdbserver-nb\") pod \"27e4748a-9fc8-4526-9b48-0d3e30f24268\" (UID: \"27e4748a-9fc8-4526-9b48-0d3e30f24268\") " Dec 13 08:00:24 crc kubenswrapper[4971]: I1213 08:00:24.127717 4971 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/configmap/27e4748a-9fc8-4526-9b48-0d3e30f24268-openstack-edpm-ipam\") pod \"27e4748a-9fc8-4526-9b48-0d3e30f24268\" (UID: \"27e4748a-9fc8-4526-9b48-0d3e30f24268\") " Dec 13 08:00:24 crc kubenswrapper[4971]: I1213 08:00:24.127863 4971 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/27e4748a-9fc8-4526-9b48-0d3e30f24268-ovsdbserver-sb\") pod \"27e4748a-9fc8-4526-9b48-0d3e30f24268\" (UID: \"27e4748a-9fc8-4526-9b48-0d3e30f24268\") " Dec 13 08:00:24 crc kubenswrapper[4971]: I1213 08:00:24.127989 4971 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-9grd2\" (UniqueName: \"kubernetes.io/projected/27e4748a-9fc8-4526-9b48-0d3e30f24268-kube-api-access-9grd2\") pod \"27e4748a-9fc8-4526-9b48-0d3e30f24268\" (UID: \"27e4748a-9fc8-4526-9b48-0d3e30f24268\") " Dec 13 08:00:24 crc kubenswrapper[4971]: I1213 08:00:24.128016 4971 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/27e4748a-9fc8-4526-9b48-0d3e30f24268-config\") pod \"27e4748a-9fc8-4526-9b48-0d3e30f24268\" (UID: \"27e4748a-9fc8-4526-9b48-0d3e30f24268\") " Dec 13 08:00:24 crc kubenswrapper[4971]: I1213 08:00:24.133940 4971 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/27e4748a-9fc8-4526-9b48-0d3e30f24268-kube-api-access-9grd2" (OuterVolumeSpecName: "kube-api-access-9grd2") pod "27e4748a-9fc8-4526-9b48-0d3e30f24268" (UID: "27e4748a-9fc8-4526-9b48-0d3e30f24268"). InnerVolumeSpecName "kube-api-access-9grd2". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 13 08:00:24 crc kubenswrapper[4971]: I1213 08:00:24.221068 4971 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/27e4748a-9fc8-4526-9b48-0d3e30f24268-ovsdbserver-sb" (OuterVolumeSpecName: "ovsdbserver-sb") pod "27e4748a-9fc8-4526-9b48-0d3e30f24268" (UID: "27e4748a-9fc8-4526-9b48-0d3e30f24268"). InnerVolumeSpecName "ovsdbserver-sb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 13 08:00:24 crc kubenswrapper[4971]: I1213 08:00:24.226269 4971 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/27e4748a-9fc8-4526-9b48-0d3e30f24268-config" (OuterVolumeSpecName: "config") pod "27e4748a-9fc8-4526-9b48-0d3e30f24268" (UID: "27e4748a-9fc8-4526-9b48-0d3e30f24268"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 13 08:00:24 crc kubenswrapper[4971]: I1213 08:00:24.233735 4971 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/27e4748a-9fc8-4526-9b48-0d3e30f24268-ovsdbserver-sb\") on node \"crc\" DevicePath \"\"" Dec 13 08:00:24 crc kubenswrapper[4971]: I1213 08:00:24.233761 4971 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-9grd2\" (UniqueName: \"kubernetes.io/projected/27e4748a-9fc8-4526-9b48-0d3e30f24268-kube-api-access-9grd2\") on node \"crc\" DevicePath \"\"" Dec 13 08:00:24 crc kubenswrapper[4971]: I1213 08:00:24.233772 4971 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/27e4748a-9fc8-4526-9b48-0d3e30f24268-config\") on node \"crc\" DevicePath \"\"" Dec 13 08:00:24 crc kubenswrapper[4971]: I1213 08:00:24.238050 4971 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/27e4748a-9fc8-4526-9b48-0d3e30f24268-openstack-edpm-ipam" (OuterVolumeSpecName: "openstack-edpm-ipam") pod "27e4748a-9fc8-4526-9b48-0d3e30f24268" (UID: "27e4748a-9fc8-4526-9b48-0d3e30f24268"). InnerVolumeSpecName "openstack-edpm-ipam". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 13 08:00:24 crc kubenswrapper[4971]: I1213 08:00:24.249851 4971 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-55d9fc99f5-565xd" event={"ID":"27e4748a-9fc8-4526-9b48-0d3e30f24268","Type":"ContainerDied","Data":"bec210e8168458ec8df26ec71e1cd92391945df2c7b244bddb888e5a106c9d61"} Dec 13 08:00:24 crc kubenswrapper[4971]: I1213 08:00:24.249918 4971 scope.go:117] "RemoveContainer" containerID="80f9dad47026016e812f4cd0fc75f8b495a90d7f6dd16030d03e794885c387fa" Dec 13 08:00:24 crc kubenswrapper[4971]: I1213 08:00:24.250067 4971 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-55d9fc99f5-565xd" Dec 13 08:00:24 crc kubenswrapper[4971]: I1213 08:00:24.255749 4971 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/27e4748a-9fc8-4526-9b48-0d3e30f24268-dns-svc" (OuterVolumeSpecName: "dns-svc") pod "27e4748a-9fc8-4526-9b48-0d3e30f24268" (UID: "27e4748a-9fc8-4526-9b48-0d3e30f24268"). InnerVolumeSpecName "dns-svc". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 13 08:00:24 crc kubenswrapper[4971]: I1213 08:00:24.261734 4971 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/27e4748a-9fc8-4526-9b48-0d3e30f24268-ovsdbserver-nb" (OuterVolumeSpecName: "ovsdbserver-nb") pod "27e4748a-9fc8-4526-9b48-0d3e30f24268" (UID: "27e4748a-9fc8-4526-9b48-0d3e30f24268"). InnerVolumeSpecName "ovsdbserver-nb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 13 08:00:24 crc kubenswrapper[4971]: I1213 08:00:24.293916 4971 scope.go:117] "RemoveContainer" containerID="ed75a47ddb443493474354190272325d4def38a026694f37cf823cbc035877cf" Dec 13 08:00:24 crc kubenswrapper[4971]: I1213 08:00:24.336475 4971 reconciler_common.go:293] "Volume detached for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/27e4748a-9fc8-4526-9b48-0d3e30f24268-dns-svc\") on node \"crc\" DevicePath \"\"" Dec 13 08:00:24 crc kubenswrapper[4971]: I1213 08:00:24.336533 4971 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/27e4748a-9fc8-4526-9b48-0d3e30f24268-ovsdbserver-nb\") on node \"crc\" DevicePath \"\"" Dec 13 08:00:24 crc kubenswrapper[4971]: I1213 08:00:24.336549 4971 reconciler_common.go:293] "Volume detached for volume \"openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/configmap/27e4748a-9fc8-4526-9b48-0d3e30f24268-openstack-edpm-ipam\") on node \"crc\" DevicePath \"\"" Dec 13 08:00:24 crc kubenswrapper[4971]: I1213 08:00:24.602594 4971 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-55d9fc99f5-565xd"] Dec 13 08:00:24 crc kubenswrapper[4971]: I1213 08:00:24.631827 4971 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-55d9fc99f5-565xd"] Dec 13 08:00:24 crc kubenswrapper[4971]: I1213 08:00:24.962893 4971 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ceilometer-0"] Dec 13 08:00:24 crc kubenswrapper[4971]: I1213 08:00:24.963454 4971 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="7cfb4dbf-cc55-48b7-b50c-65d6d4ae64ba" containerName="ceilometer-notification-agent" containerID="cri-o://aa8dbe1b33fab32fe56308f8a63d03f9d6b4e85ef1f5a3bacd7f345b960f96d7" gracePeriod=30 Dec 13 08:00:24 crc kubenswrapper[4971]: I1213 08:00:24.963624 4971 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="7cfb4dbf-cc55-48b7-b50c-65d6d4ae64ba" containerName="ceilometer-central-agent" containerID="cri-o://7472627612cd5a2ad24c6810633935f98b010f9d67b67ec6bddf4a6475b06a08" gracePeriod=30 Dec 13 08:00:24 crc kubenswrapper[4971]: I1213 08:00:24.963669 4971 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="7cfb4dbf-cc55-48b7-b50c-65d6d4ae64ba" containerName="proxy-httpd" containerID="cri-o://9392bd2978cae265c91fac659c8e2f215ca3c33f3a347e415f189d6226655a93" gracePeriod=30 Dec 13 08:00:24 crc kubenswrapper[4971]: I1213 08:00:24.963702 4971 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="7cfb4dbf-cc55-48b7-b50c-65d6d4ae64ba" containerName="sg-core" containerID="cri-o://fcb14c60d29a4cc2cb8ee4e92bbaef4cab76cb5fd4e53060369519333e285056" gracePeriod=30 Dec 13 08:00:25 crc kubenswrapper[4971]: I1213 08:00:25.264580 4971 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/manila-api-0" event={"ID":"e8e297bb-43e7-45b9-a82b-a46c104519c1","Type":"ContainerStarted","Data":"3f67c7df408c6857e4564e4c7cc55f0ccb0e6735ed175a6024aa20871b99f906"} Dec 13 08:00:25 crc kubenswrapper[4971]: I1213 08:00:25.265882 4971 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/manila-api-0" Dec 13 08:00:25 crc kubenswrapper[4971]: I1213 08:00:25.277809 4971 generic.go:334] "Generic (PLEG): container finished" podID="7cfb4dbf-cc55-48b7-b50c-65d6d4ae64ba" containerID="9392bd2978cae265c91fac659c8e2f215ca3c33f3a347e415f189d6226655a93" exitCode=0 Dec 13 08:00:25 crc kubenswrapper[4971]: I1213 08:00:25.277841 4971 generic.go:334] "Generic (PLEG): container finished" podID="7cfb4dbf-cc55-48b7-b50c-65d6d4ae64ba" containerID="fcb14c60d29a4cc2cb8ee4e92bbaef4cab76cb5fd4e53060369519333e285056" exitCode=2 Dec 13 08:00:25 crc kubenswrapper[4971]: I1213 08:00:25.277881 4971 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"7cfb4dbf-cc55-48b7-b50c-65d6d4ae64ba","Type":"ContainerDied","Data":"9392bd2978cae265c91fac659c8e2f215ca3c33f3a347e415f189d6226655a93"} Dec 13 08:00:25 crc kubenswrapper[4971]: I1213 08:00:25.277954 4971 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"7cfb4dbf-cc55-48b7-b50c-65d6d4ae64ba","Type":"ContainerDied","Data":"fcb14c60d29a4cc2cb8ee4e92bbaef4cab76cb5fd4e53060369519333e285056"} Dec 13 08:00:25 crc kubenswrapper[4971]: I1213 08:00:25.280071 4971 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/manila-share-share1-0" event={"ID":"1d983511-44dd-4be9-9b43-2f496428a12e","Type":"ContainerStarted","Data":"0399449dfe2e04f246bc19adf9eca94b501aca1e6be2bd628433385f39d17b3d"} Dec 13 08:00:25 crc kubenswrapper[4971]: I1213 08:00:25.280117 4971 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/manila-share-share1-0" event={"ID":"1d983511-44dd-4be9-9b43-2f496428a12e","Type":"ContainerStarted","Data":"9fb4b92a9f988de4c1a56dac883f2fac0249dc508a477173963001bbbf61c1e9"} Dec 13 08:00:25 crc kubenswrapper[4971]: I1213 08:00:25.312253 4971 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/manila-api-0" podStartSLOduration=7.312232787 podStartE2EDuration="7.312232787s" podCreationTimestamp="2025-12-13 08:00:18 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-13 08:00:25.292485399 +0000 UTC m=+4281.896894857" watchObservedRunningTime="2025-12-13 08:00:25.312232787 +0000 UTC m=+4281.916642235" Dec 13 08:00:25 crc kubenswrapper[4971]: I1213 08:00:25.314442 4971 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/manila-share-share1-0" podStartSLOduration=3.255601522 podStartE2EDuration="15.314435281s" podCreationTimestamp="2025-12-13 08:00:10 +0000 UTC" firstStartedPulling="2025-12-13 08:00:11.75786501 +0000 UTC m=+4268.362274458" lastFinishedPulling="2025-12-13 08:00:23.816698769 +0000 UTC m=+4280.421108217" observedRunningTime="2025-12-13 08:00:25.308683302 +0000 UTC m=+4281.913092750" watchObservedRunningTime="2025-12-13 08:00:25.314435281 +0000 UTC m=+4281.918844729" Dec 13 08:00:25 crc kubenswrapper[4971]: I1213 08:00:25.784832 4971 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="27e4748a-9fc8-4526-9b48-0d3e30f24268" path="/var/lib/kubelet/pods/27e4748a-9fc8-4526-9b48-0d3e30f24268/volumes" Dec 13 08:00:26 crc kubenswrapper[4971]: I1213 08:00:26.295563 4971 generic.go:334] "Generic (PLEG): container finished" podID="7cfb4dbf-cc55-48b7-b50c-65d6d4ae64ba" containerID="7472627612cd5a2ad24c6810633935f98b010f9d67b67ec6bddf4a6475b06a08" exitCode=0 Dec 13 08:00:26 crc kubenswrapper[4971]: I1213 08:00:26.295628 4971 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"7cfb4dbf-cc55-48b7-b50c-65d6d4ae64ba","Type":"ContainerDied","Data":"7472627612cd5a2ad24c6810633935f98b010f9d67b67ec6bddf4a6475b06a08"} Dec 13 08:00:26 crc kubenswrapper[4971]: I1213 08:00:26.295884 4971 scope.go:117] "RemoveContainer" containerID="2270e1af9b8023fb36226ee428231c2e187e0f781fcb819a9e8e50ef5b0046f5" Dec 13 08:00:28 crc kubenswrapper[4971]: I1213 08:00:28.002598 4971 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Dec 13 08:00:28 crc kubenswrapper[4971]: I1213 08:00:28.124796 4971 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/7cfb4dbf-cc55-48b7-b50c-65d6d4ae64ba-run-httpd\") pod \"7cfb4dbf-cc55-48b7-b50c-65d6d4ae64ba\" (UID: \"7cfb4dbf-cc55-48b7-b50c-65d6d4ae64ba\") " Dec 13 08:00:28 crc kubenswrapper[4971]: I1213 08:00:28.124846 4971 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/7cfb4dbf-cc55-48b7-b50c-65d6d4ae64ba-combined-ca-bundle\") pod \"7cfb4dbf-cc55-48b7-b50c-65d6d4ae64ba\" (UID: \"7cfb4dbf-cc55-48b7-b50c-65d6d4ae64ba\") " Dec 13 08:00:28 crc kubenswrapper[4971]: I1213 08:00:28.124926 4971 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-2tqrx\" (UniqueName: \"kubernetes.io/projected/7cfb4dbf-cc55-48b7-b50c-65d6d4ae64ba-kube-api-access-2tqrx\") pod \"7cfb4dbf-cc55-48b7-b50c-65d6d4ae64ba\" (UID: \"7cfb4dbf-cc55-48b7-b50c-65d6d4ae64ba\") " Dec 13 08:00:28 crc kubenswrapper[4971]: I1213 08:00:28.125028 4971 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/7cfb4dbf-cc55-48b7-b50c-65d6d4ae64ba-sg-core-conf-yaml\") pod \"7cfb4dbf-cc55-48b7-b50c-65d6d4ae64ba\" (UID: \"7cfb4dbf-cc55-48b7-b50c-65d6d4ae64ba\") " Dec 13 08:00:28 crc kubenswrapper[4971]: I1213 08:00:28.125060 4971 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/7cfb4dbf-cc55-48b7-b50c-65d6d4ae64ba-scripts\") pod \"7cfb4dbf-cc55-48b7-b50c-65d6d4ae64ba\" (UID: \"7cfb4dbf-cc55-48b7-b50c-65d6d4ae64ba\") " Dec 13 08:00:28 crc kubenswrapper[4971]: I1213 08:00:28.125108 4971 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/7cfb4dbf-cc55-48b7-b50c-65d6d4ae64ba-config-data\") pod \"7cfb4dbf-cc55-48b7-b50c-65d6d4ae64ba\" (UID: \"7cfb4dbf-cc55-48b7-b50c-65d6d4ae64ba\") " Dec 13 08:00:28 crc kubenswrapper[4971]: I1213 08:00:28.125153 4971 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/7cfb4dbf-cc55-48b7-b50c-65d6d4ae64ba-log-httpd\") pod \"7cfb4dbf-cc55-48b7-b50c-65d6d4ae64ba\" (UID: \"7cfb4dbf-cc55-48b7-b50c-65d6d4ae64ba\") " Dec 13 08:00:28 crc kubenswrapper[4971]: I1213 08:00:28.125194 4971 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ceilometer-tls-certs\" (UniqueName: \"kubernetes.io/secret/7cfb4dbf-cc55-48b7-b50c-65d6d4ae64ba-ceilometer-tls-certs\") pod \"7cfb4dbf-cc55-48b7-b50c-65d6d4ae64ba\" (UID: \"7cfb4dbf-cc55-48b7-b50c-65d6d4ae64ba\") " Dec 13 08:00:28 crc kubenswrapper[4971]: I1213 08:00:28.125970 4971 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/7cfb4dbf-cc55-48b7-b50c-65d6d4ae64ba-run-httpd" (OuterVolumeSpecName: "run-httpd") pod "7cfb4dbf-cc55-48b7-b50c-65d6d4ae64ba" (UID: "7cfb4dbf-cc55-48b7-b50c-65d6d4ae64ba"). InnerVolumeSpecName "run-httpd". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 13 08:00:28 crc kubenswrapper[4971]: I1213 08:00:28.126653 4971 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/7cfb4dbf-cc55-48b7-b50c-65d6d4ae64ba-log-httpd" (OuterVolumeSpecName: "log-httpd") pod "7cfb4dbf-cc55-48b7-b50c-65d6d4ae64ba" (UID: "7cfb4dbf-cc55-48b7-b50c-65d6d4ae64ba"). InnerVolumeSpecName "log-httpd". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 13 08:00:28 crc kubenswrapper[4971]: I1213 08:00:28.131371 4971 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/7cfb4dbf-cc55-48b7-b50c-65d6d4ae64ba-kube-api-access-2tqrx" (OuterVolumeSpecName: "kube-api-access-2tqrx") pod "7cfb4dbf-cc55-48b7-b50c-65d6d4ae64ba" (UID: "7cfb4dbf-cc55-48b7-b50c-65d6d4ae64ba"). InnerVolumeSpecName "kube-api-access-2tqrx". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 13 08:00:28 crc kubenswrapper[4971]: I1213 08:00:28.131744 4971 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/7cfb4dbf-cc55-48b7-b50c-65d6d4ae64ba-scripts" (OuterVolumeSpecName: "scripts") pod "7cfb4dbf-cc55-48b7-b50c-65d6d4ae64ba" (UID: "7cfb4dbf-cc55-48b7-b50c-65d6d4ae64ba"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 13 08:00:28 crc kubenswrapper[4971]: I1213 08:00:28.163120 4971 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/7cfb4dbf-cc55-48b7-b50c-65d6d4ae64ba-sg-core-conf-yaml" (OuterVolumeSpecName: "sg-core-conf-yaml") pod "7cfb4dbf-cc55-48b7-b50c-65d6d4ae64ba" (UID: "7cfb4dbf-cc55-48b7-b50c-65d6d4ae64ba"). InnerVolumeSpecName "sg-core-conf-yaml". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 13 08:00:28 crc kubenswrapper[4971]: I1213 08:00:28.180509 4971 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/7cfb4dbf-cc55-48b7-b50c-65d6d4ae64ba-ceilometer-tls-certs" (OuterVolumeSpecName: "ceilometer-tls-certs") pod "7cfb4dbf-cc55-48b7-b50c-65d6d4ae64ba" (UID: "7cfb4dbf-cc55-48b7-b50c-65d6d4ae64ba"). InnerVolumeSpecName "ceilometer-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 13 08:00:28 crc kubenswrapper[4971]: I1213 08:00:28.220006 4971 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/7cfb4dbf-cc55-48b7-b50c-65d6d4ae64ba-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "7cfb4dbf-cc55-48b7-b50c-65d6d4ae64ba" (UID: "7cfb4dbf-cc55-48b7-b50c-65d6d4ae64ba"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 13 08:00:28 crc kubenswrapper[4971]: I1213 08:00:28.228478 4971 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/7cfb4dbf-cc55-48b7-b50c-65d6d4ae64ba-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 13 08:00:28 crc kubenswrapper[4971]: I1213 08:00:28.228572 4971 reconciler_common.go:293] "Volume detached for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/7cfb4dbf-cc55-48b7-b50c-65d6d4ae64ba-run-httpd\") on node \"crc\" DevicePath \"\"" Dec 13 08:00:28 crc kubenswrapper[4971]: I1213 08:00:28.228588 4971 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-2tqrx\" (UniqueName: \"kubernetes.io/projected/7cfb4dbf-cc55-48b7-b50c-65d6d4ae64ba-kube-api-access-2tqrx\") on node \"crc\" DevicePath \"\"" Dec 13 08:00:28 crc kubenswrapper[4971]: I1213 08:00:28.228601 4971 reconciler_common.go:293] "Volume detached for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/7cfb4dbf-cc55-48b7-b50c-65d6d4ae64ba-sg-core-conf-yaml\") on node \"crc\" DevicePath \"\"" Dec 13 08:00:28 crc kubenswrapper[4971]: I1213 08:00:28.228612 4971 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/7cfb4dbf-cc55-48b7-b50c-65d6d4ae64ba-scripts\") on node \"crc\" DevicePath \"\"" Dec 13 08:00:28 crc kubenswrapper[4971]: I1213 08:00:28.228654 4971 reconciler_common.go:293] "Volume detached for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/7cfb4dbf-cc55-48b7-b50c-65d6d4ae64ba-log-httpd\") on node \"crc\" DevicePath \"\"" Dec 13 08:00:28 crc kubenswrapper[4971]: I1213 08:00:28.228666 4971 reconciler_common.go:293] "Volume detached for volume \"ceilometer-tls-certs\" (UniqueName: \"kubernetes.io/secret/7cfb4dbf-cc55-48b7-b50c-65d6d4ae64ba-ceilometer-tls-certs\") on node \"crc\" DevicePath \"\"" Dec 13 08:00:28 crc kubenswrapper[4971]: I1213 08:00:28.239408 4971 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/7cfb4dbf-cc55-48b7-b50c-65d6d4ae64ba-config-data" (OuterVolumeSpecName: "config-data") pod "7cfb4dbf-cc55-48b7-b50c-65d6d4ae64ba" (UID: "7cfb4dbf-cc55-48b7-b50c-65d6d4ae64ba"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 13 08:00:28 crc kubenswrapper[4971]: I1213 08:00:28.321969 4971 generic.go:334] "Generic (PLEG): container finished" podID="7cfb4dbf-cc55-48b7-b50c-65d6d4ae64ba" containerID="aa8dbe1b33fab32fe56308f8a63d03f9d6b4e85ef1f5a3bacd7f345b960f96d7" exitCode=0 Dec 13 08:00:28 crc kubenswrapper[4971]: I1213 08:00:28.322008 4971 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"7cfb4dbf-cc55-48b7-b50c-65d6d4ae64ba","Type":"ContainerDied","Data":"aa8dbe1b33fab32fe56308f8a63d03f9d6b4e85ef1f5a3bacd7f345b960f96d7"} Dec 13 08:00:28 crc kubenswrapper[4971]: I1213 08:00:28.322032 4971 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"7cfb4dbf-cc55-48b7-b50c-65d6d4ae64ba","Type":"ContainerDied","Data":"7d58386baf345d60c820859eeb823a8fa1776e41a1099901668394ee34ba7a40"} Dec 13 08:00:28 crc kubenswrapper[4971]: I1213 08:00:28.322048 4971 scope.go:117] "RemoveContainer" containerID="7472627612cd5a2ad24c6810633935f98b010f9d67b67ec6bddf4a6475b06a08" Dec 13 08:00:28 crc kubenswrapper[4971]: I1213 08:00:28.322145 4971 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Dec 13 08:00:28 crc kubenswrapper[4971]: I1213 08:00:28.333658 4971 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/7cfb4dbf-cc55-48b7-b50c-65d6d4ae64ba-config-data\") on node \"crc\" DevicePath \"\"" Dec 13 08:00:28 crc kubenswrapper[4971]: I1213 08:00:28.355322 4971 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ceilometer-0"] Dec 13 08:00:28 crc kubenswrapper[4971]: I1213 08:00:28.362815 4971 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/ceilometer-0"] Dec 13 08:00:28 crc kubenswrapper[4971]: I1213 08:00:28.391142 4971 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ceilometer-0"] Dec 13 08:00:28 crc kubenswrapper[4971]: I1213 08:00:28.391304 4971 scope.go:117] "RemoveContainer" containerID="9392bd2978cae265c91fac659c8e2f215ca3c33f3a347e415f189d6226655a93" Dec 13 08:00:28 crc kubenswrapper[4971]: E1213 08:00:28.391614 4971 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="7cfb4dbf-cc55-48b7-b50c-65d6d4ae64ba" containerName="ceilometer-central-agent" Dec 13 08:00:28 crc kubenswrapper[4971]: I1213 08:00:28.391634 4971 state_mem.go:107] "Deleted CPUSet assignment" podUID="7cfb4dbf-cc55-48b7-b50c-65d6d4ae64ba" containerName="ceilometer-central-agent" Dec 13 08:00:28 crc kubenswrapper[4971]: E1213 08:00:28.391658 4971 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="7cfb4dbf-cc55-48b7-b50c-65d6d4ae64ba" containerName="sg-core" Dec 13 08:00:28 crc kubenswrapper[4971]: I1213 08:00:28.391666 4971 state_mem.go:107] "Deleted CPUSet assignment" podUID="7cfb4dbf-cc55-48b7-b50c-65d6d4ae64ba" containerName="sg-core" Dec 13 08:00:28 crc kubenswrapper[4971]: E1213 08:00:28.391682 4971 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="7cfb4dbf-cc55-48b7-b50c-65d6d4ae64ba" containerName="ceilometer-central-agent" Dec 13 08:00:28 crc kubenswrapper[4971]: I1213 08:00:28.391689 4971 state_mem.go:107] "Deleted CPUSet assignment" podUID="7cfb4dbf-cc55-48b7-b50c-65d6d4ae64ba" containerName="ceilometer-central-agent" Dec 13 08:00:28 crc kubenswrapper[4971]: E1213 08:00:28.391708 4971 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="27e4748a-9fc8-4526-9b48-0d3e30f24268" containerName="dnsmasq-dns" Dec 13 08:00:28 crc kubenswrapper[4971]: I1213 08:00:28.391715 4971 state_mem.go:107] "Deleted CPUSet assignment" podUID="27e4748a-9fc8-4526-9b48-0d3e30f24268" containerName="dnsmasq-dns" Dec 13 08:00:28 crc kubenswrapper[4971]: E1213 08:00:28.391732 4971 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="27e4748a-9fc8-4526-9b48-0d3e30f24268" containerName="init" Dec 13 08:00:28 crc kubenswrapper[4971]: I1213 08:00:28.391740 4971 state_mem.go:107] "Deleted CPUSet assignment" podUID="27e4748a-9fc8-4526-9b48-0d3e30f24268" containerName="init" Dec 13 08:00:28 crc kubenswrapper[4971]: E1213 08:00:28.391760 4971 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="7cfb4dbf-cc55-48b7-b50c-65d6d4ae64ba" containerName="ceilometer-notification-agent" Dec 13 08:00:28 crc kubenswrapper[4971]: I1213 08:00:28.391768 4971 state_mem.go:107] "Deleted CPUSet assignment" podUID="7cfb4dbf-cc55-48b7-b50c-65d6d4ae64ba" containerName="ceilometer-notification-agent" Dec 13 08:00:28 crc kubenswrapper[4971]: E1213 08:00:28.391778 4971 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="7cfb4dbf-cc55-48b7-b50c-65d6d4ae64ba" containerName="proxy-httpd" Dec 13 08:00:28 crc kubenswrapper[4971]: I1213 08:00:28.391784 4971 state_mem.go:107] "Deleted CPUSet assignment" podUID="7cfb4dbf-cc55-48b7-b50c-65d6d4ae64ba" containerName="proxy-httpd" Dec 13 08:00:28 crc kubenswrapper[4971]: I1213 08:00:28.391981 4971 memory_manager.go:354] "RemoveStaleState removing state" podUID="7cfb4dbf-cc55-48b7-b50c-65d6d4ae64ba" containerName="ceilometer-notification-agent" Dec 13 08:00:28 crc kubenswrapper[4971]: I1213 08:00:28.392011 4971 memory_manager.go:354] "RemoveStaleState removing state" podUID="7cfb4dbf-cc55-48b7-b50c-65d6d4ae64ba" containerName="sg-core" Dec 13 08:00:28 crc kubenswrapper[4971]: I1213 08:00:28.392043 4971 memory_manager.go:354] "RemoveStaleState removing state" podUID="7cfb4dbf-cc55-48b7-b50c-65d6d4ae64ba" containerName="ceilometer-central-agent" Dec 13 08:00:28 crc kubenswrapper[4971]: I1213 08:00:28.392057 4971 memory_manager.go:354] "RemoveStaleState removing state" podUID="7cfb4dbf-cc55-48b7-b50c-65d6d4ae64ba" containerName="ceilometer-central-agent" Dec 13 08:00:28 crc kubenswrapper[4971]: I1213 08:00:28.392066 4971 memory_manager.go:354] "RemoveStaleState removing state" podUID="7cfb4dbf-cc55-48b7-b50c-65d6d4ae64ba" containerName="ceilometer-central-agent" Dec 13 08:00:28 crc kubenswrapper[4971]: I1213 08:00:28.392081 4971 memory_manager.go:354] "RemoveStaleState removing state" podUID="7cfb4dbf-cc55-48b7-b50c-65d6d4ae64ba" containerName="proxy-httpd" Dec 13 08:00:28 crc kubenswrapper[4971]: I1213 08:00:28.392089 4971 memory_manager.go:354] "RemoveStaleState removing state" podUID="27e4748a-9fc8-4526-9b48-0d3e30f24268" containerName="dnsmasq-dns" Dec 13 08:00:28 crc kubenswrapper[4971]: E1213 08:00:28.392296 4971 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="7cfb4dbf-cc55-48b7-b50c-65d6d4ae64ba" containerName="ceilometer-central-agent" Dec 13 08:00:28 crc kubenswrapper[4971]: I1213 08:00:28.392304 4971 state_mem.go:107] "Deleted CPUSet assignment" podUID="7cfb4dbf-cc55-48b7-b50c-65d6d4ae64ba" containerName="ceilometer-central-agent" Dec 13 08:00:28 crc kubenswrapper[4971]: I1213 08:00:28.393872 4971 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Dec 13 08:00:28 crc kubenswrapper[4971]: I1213 08:00:28.398003 4971 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceilometer-scripts" Dec 13 08:00:28 crc kubenswrapper[4971]: I1213 08:00:28.398134 4971 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceilometer-config-data" Dec 13 08:00:28 crc kubenswrapper[4971]: I1213 08:00:28.398165 4971 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-ceilometer-internal-svc" Dec 13 08:00:28 crc kubenswrapper[4971]: I1213 08:00:28.400983 4971 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ceilometer-0"] Dec 13 08:00:28 crc kubenswrapper[4971]: I1213 08:00:28.418485 4971 scope.go:117] "RemoveContainer" containerID="fcb14c60d29a4cc2cb8ee4e92bbaef4cab76cb5fd4e53060369519333e285056" Dec 13 08:00:28 crc kubenswrapper[4971]: I1213 08:00:28.436959 4971 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/7ba95a4c-be84-460a-9dfc-b6e1474e1b81-scripts\") pod \"ceilometer-0\" (UID: \"7ba95a4c-be84-460a-9dfc-b6e1474e1b81\") " pod="openstack/ceilometer-0" Dec 13 08:00:28 crc kubenswrapper[4971]: I1213 08:00:28.437123 4971 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-hsl8d\" (UniqueName: \"kubernetes.io/projected/7ba95a4c-be84-460a-9dfc-b6e1474e1b81-kube-api-access-hsl8d\") pod \"ceilometer-0\" (UID: \"7ba95a4c-be84-460a-9dfc-b6e1474e1b81\") " pod="openstack/ceilometer-0" Dec 13 08:00:28 crc kubenswrapper[4971]: I1213 08:00:28.437266 4971 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/7ba95a4c-be84-460a-9dfc-b6e1474e1b81-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"7ba95a4c-be84-460a-9dfc-b6e1474e1b81\") " pod="openstack/ceilometer-0" Dec 13 08:00:28 crc kubenswrapper[4971]: I1213 08:00:28.437309 4971 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/7ba95a4c-be84-460a-9dfc-b6e1474e1b81-config-data\") pod \"ceilometer-0\" (UID: \"7ba95a4c-be84-460a-9dfc-b6e1474e1b81\") " pod="openstack/ceilometer-0" Dec 13 08:00:28 crc kubenswrapper[4971]: I1213 08:00:28.437349 4971 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/7ba95a4c-be84-460a-9dfc-b6e1474e1b81-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"7ba95a4c-be84-460a-9dfc-b6e1474e1b81\") " pod="openstack/ceilometer-0" Dec 13 08:00:28 crc kubenswrapper[4971]: I1213 08:00:28.437392 4971 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/7ba95a4c-be84-460a-9dfc-b6e1474e1b81-run-httpd\") pod \"ceilometer-0\" (UID: \"7ba95a4c-be84-460a-9dfc-b6e1474e1b81\") " pod="openstack/ceilometer-0" Dec 13 08:00:28 crc kubenswrapper[4971]: I1213 08:00:28.437531 4971 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/7ba95a4c-be84-460a-9dfc-b6e1474e1b81-log-httpd\") pod \"ceilometer-0\" (UID: \"7ba95a4c-be84-460a-9dfc-b6e1474e1b81\") " pod="openstack/ceilometer-0" Dec 13 08:00:28 crc kubenswrapper[4971]: I1213 08:00:28.437598 4971 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ceilometer-tls-certs\" (UniqueName: \"kubernetes.io/secret/7ba95a4c-be84-460a-9dfc-b6e1474e1b81-ceilometer-tls-certs\") pod \"ceilometer-0\" (UID: \"7ba95a4c-be84-460a-9dfc-b6e1474e1b81\") " pod="openstack/ceilometer-0" Dec 13 08:00:28 crc kubenswrapper[4971]: I1213 08:00:28.443484 4971 scope.go:117] "RemoveContainer" containerID="aa8dbe1b33fab32fe56308f8a63d03f9d6b4e85ef1f5a3bacd7f345b960f96d7" Dec 13 08:00:28 crc kubenswrapper[4971]: I1213 08:00:28.463550 4971 scope.go:117] "RemoveContainer" containerID="7472627612cd5a2ad24c6810633935f98b010f9d67b67ec6bddf4a6475b06a08" Dec 13 08:00:28 crc kubenswrapper[4971]: E1213 08:00:28.463987 4971 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"7472627612cd5a2ad24c6810633935f98b010f9d67b67ec6bddf4a6475b06a08\": container with ID starting with 7472627612cd5a2ad24c6810633935f98b010f9d67b67ec6bddf4a6475b06a08 not found: ID does not exist" containerID="7472627612cd5a2ad24c6810633935f98b010f9d67b67ec6bddf4a6475b06a08" Dec 13 08:00:28 crc kubenswrapper[4971]: I1213 08:00:28.464086 4971 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"7472627612cd5a2ad24c6810633935f98b010f9d67b67ec6bddf4a6475b06a08"} err="failed to get container status \"7472627612cd5a2ad24c6810633935f98b010f9d67b67ec6bddf4a6475b06a08\": rpc error: code = NotFound desc = could not find container \"7472627612cd5a2ad24c6810633935f98b010f9d67b67ec6bddf4a6475b06a08\": container with ID starting with 7472627612cd5a2ad24c6810633935f98b010f9d67b67ec6bddf4a6475b06a08 not found: ID does not exist" Dec 13 08:00:28 crc kubenswrapper[4971]: I1213 08:00:28.464113 4971 scope.go:117] "RemoveContainer" containerID="9392bd2978cae265c91fac659c8e2f215ca3c33f3a347e415f189d6226655a93" Dec 13 08:00:28 crc kubenswrapper[4971]: E1213 08:00:28.467138 4971 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"9392bd2978cae265c91fac659c8e2f215ca3c33f3a347e415f189d6226655a93\": container with ID starting with 9392bd2978cae265c91fac659c8e2f215ca3c33f3a347e415f189d6226655a93 not found: ID does not exist" containerID="9392bd2978cae265c91fac659c8e2f215ca3c33f3a347e415f189d6226655a93" Dec 13 08:00:28 crc kubenswrapper[4971]: I1213 08:00:28.467181 4971 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"9392bd2978cae265c91fac659c8e2f215ca3c33f3a347e415f189d6226655a93"} err="failed to get container status \"9392bd2978cae265c91fac659c8e2f215ca3c33f3a347e415f189d6226655a93\": rpc error: code = NotFound desc = could not find container \"9392bd2978cae265c91fac659c8e2f215ca3c33f3a347e415f189d6226655a93\": container with ID starting with 9392bd2978cae265c91fac659c8e2f215ca3c33f3a347e415f189d6226655a93 not found: ID does not exist" Dec 13 08:00:28 crc kubenswrapper[4971]: I1213 08:00:28.467296 4971 scope.go:117] "RemoveContainer" containerID="fcb14c60d29a4cc2cb8ee4e92bbaef4cab76cb5fd4e53060369519333e285056" Dec 13 08:00:28 crc kubenswrapper[4971]: E1213 08:00:28.467835 4971 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"fcb14c60d29a4cc2cb8ee4e92bbaef4cab76cb5fd4e53060369519333e285056\": container with ID starting with fcb14c60d29a4cc2cb8ee4e92bbaef4cab76cb5fd4e53060369519333e285056 not found: ID does not exist" containerID="fcb14c60d29a4cc2cb8ee4e92bbaef4cab76cb5fd4e53060369519333e285056" Dec 13 08:00:28 crc kubenswrapper[4971]: I1213 08:00:28.467878 4971 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"fcb14c60d29a4cc2cb8ee4e92bbaef4cab76cb5fd4e53060369519333e285056"} err="failed to get container status \"fcb14c60d29a4cc2cb8ee4e92bbaef4cab76cb5fd4e53060369519333e285056\": rpc error: code = NotFound desc = could not find container \"fcb14c60d29a4cc2cb8ee4e92bbaef4cab76cb5fd4e53060369519333e285056\": container with ID starting with fcb14c60d29a4cc2cb8ee4e92bbaef4cab76cb5fd4e53060369519333e285056 not found: ID does not exist" Dec 13 08:00:28 crc kubenswrapper[4971]: I1213 08:00:28.467911 4971 scope.go:117] "RemoveContainer" containerID="aa8dbe1b33fab32fe56308f8a63d03f9d6b4e85ef1f5a3bacd7f345b960f96d7" Dec 13 08:00:28 crc kubenswrapper[4971]: E1213 08:00:28.468334 4971 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"aa8dbe1b33fab32fe56308f8a63d03f9d6b4e85ef1f5a3bacd7f345b960f96d7\": container with ID starting with aa8dbe1b33fab32fe56308f8a63d03f9d6b4e85ef1f5a3bacd7f345b960f96d7 not found: ID does not exist" containerID="aa8dbe1b33fab32fe56308f8a63d03f9d6b4e85ef1f5a3bacd7f345b960f96d7" Dec 13 08:00:28 crc kubenswrapper[4971]: I1213 08:00:28.468370 4971 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"aa8dbe1b33fab32fe56308f8a63d03f9d6b4e85ef1f5a3bacd7f345b960f96d7"} err="failed to get container status \"aa8dbe1b33fab32fe56308f8a63d03f9d6b4e85ef1f5a3bacd7f345b960f96d7\": rpc error: code = NotFound desc = could not find container \"aa8dbe1b33fab32fe56308f8a63d03f9d6b4e85ef1f5a3bacd7f345b960f96d7\": container with ID starting with aa8dbe1b33fab32fe56308f8a63d03f9d6b4e85ef1f5a3bacd7f345b960f96d7 not found: ID does not exist" Dec 13 08:00:28 crc kubenswrapper[4971]: I1213 08:00:28.539405 4971 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/7ba95a4c-be84-460a-9dfc-b6e1474e1b81-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"7ba95a4c-be84-460a-9dfc-b6e1474e1b81\") " pod="openstack/ceilometer-0" Dec 13 08:00:28 crc kubenswrapper[4971]: I1213 08:00:28.539782 4971 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/7ba95a4c-be84-460a-9dfc-b6e1474e1b81-config-data\") pod \"ceilometer-0\" (UID: \"7ba95a4c-be84-460a-9dfc-b6e1474e1b81\") " pod="openstack/ceilometer-0" Dec 13 08:00:28 crc kubenswrapper[4971]: I1213 08:00:28.539811 4971 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/7ba95a4c-be84-460a-9dfc-b6e1474e1b81-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"7ba95a4c-be84-460a-9dfc-b6e1474e1b81\") " pod="openstack/ceilometer-0" Dec 13 08:00:28 crc kubenswrapper[4971]: I1213 08:00:28.539839 4971 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/7ba95a4c-be84-460a-9dfc-b6e1474e1b81-run-httpd\") pod \"ceilometer-0\" (UID: \"7ba95a4c-be84-460a-9dfc-b6e1474e1b81\") " pod="openstack/ceilometer-0" Dec 13 08:00:28 crc kubenswrapper[4971]: I1213 08:00:28.539874 4971 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/7ba95a4c-be84-460a-9dfc-b6e1474e1b81-log-httpd\") pod \"ceilometer-0\" (UID: \"7ba95a4c-be84-460a-9dfc-b6e1474e1b81\") " pod="openstack/ceilometer-0" Dec 13 08:00:28 crc kubenswrapper[4971]: I1213 08:00:28.539923 4971 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ceilometer-tls-certs\" (UniqueName: \"kubernetes.io/secret/7ba95a4c-be84-460a-9dfc-b6e1474e1b81-ceilometer-tls-certs\") pod \"ceilometer-0\" (UID: \"7ba95a4c-be84-460a-9dfc-b6e1474e1b81\") " pod="openstack/ceilometer-0" Dec 13 08:00:28 crc kubenswrapper[4971]: I1213 08:00:28.539999 4971 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/7ba95a4c-be84-460a-9dfc-b6e1474e1b81-scripts\") pod \"ceilometer-0\" (UID: \"7ba95a4c-be84-460a-9dfc-b6e1474e1b81\") " pod="openstack/ceilometer-0" Dec 13 08:00:28 crc kubenswrapper[4971]: I1213 08:00:28.540090 4971 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-hsl8d\" (UniqueName: \"kubernetes.io/projected/7ba95a4c-be84-460a-9dfc-b6e1474e1b81-kube-api-access-hsl8d\") pod \"ceilometer-0\" (UID: \"7ba95a4c-be84-460a-9dfc-b6e1474e1b81\") " pod="openstack/ceilometer-0" Dec 13 08:00:28 crc kubenswrapper[4971]: I1213 08:00:28.542430 4971 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/7ba95a4c-be84-460a-9dfc-b6e1474e1b81-run-httpd\") pod \"ceilometer-0\" (UID: \"7ba95a4c-be84-460a-9dfc-b6e1474e1b81\") " pod="openstack/ceilometer-0" Dec 13 08:00:28 crc kubenswrapper[4971]: I1213 08:00:28.544781 4971 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ceilometer-tls-certs\" (UniqueName: \"kubernetes.io/secret/7ba95a4c-be84-460a-9dfc-b6e1474e1b81-ceilometer-tls-certs\") pod \"ceilometer-0\" (UID: \"7ba95a4c-be84-460a-9dfc-b6e1474e1b81\") " pod="openstack/ceilometer-0" Dec 13 08:00:28 crc kubenswrapper[4971]: I1213 08:00:28.545130 4971 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/7ba95a4c-be84-460a-9dfc-b6e1474e1b81-log-httpd\") pod \"ceilometer-0\" (UID: \"7ba95a4c-be84-460a-9dfc-b6e1474e1b81\") " pod="openstack/ceilometer-0" Dec 13 08:00:28 crc kubenswrapper[4971]: I1213 08:00:28.545383 4971 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/7ba95a4c-be84-460a-9dfc-b6e1474e1b81-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"7ba95a4c-be84-460a-9dfc-b6e1474e1b81\") " pod="openstack/ceilometer-0" Dec 13 08:00:28 crc kubenswrapper[4971]: I1213 08:00:28.545404 4971 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/7ba95a4c-be84-460a-9dfc-b6e1474e1b81-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"7ba95a4c-be84-460a-9dfc-b6e1474e1b81\") " pod="openstack/ceilometer-0" Dec 13 08:00:28 crc kubenswrapper[4971]: I1213 08:00:28.545436 4971 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/7ba95a4c-be84-460a-9dfc-b6e1474e1b81-scripts\") pod \"ceilometer-0\" (UID: \"7ba95a4c-be84-460a-9dfc-b6e1474e1b81\") " pod="openstack/ceilometer-0" Dec 13 08:00:28 crc kubenswrapper[4971]: I1213 08:00:28.546171 4971 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/7ba95a4c-be84-460a-9dfc-b6e1474e1b81-config-data\") pod \"ceilometer-0\" (UID: \"7ba95a4c-be84-460a-9dfc-b6e1474e1b81\") " pod="openstack/ceilometer-0" Dec 13 08:00:28 crc kubenswrapper[4971]: I1213 08:00:28.559568 4971 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-hsl8d\" (UniqueName: \"kubernetes.io/projected/7ba95a4c-be84-460a-9dfc-b6e1474e1b81-kube-api-access-hsl8d\") pod \"ceilometer-0\" (UID: \"7ba95a4c-be84-460a-9dfc-b6e1474e1b81\") " pod="openstack/ceilometer-0" Dec 13 08:00:28 crc kubenswrapper[4971]: I1213 08:00:28.720559 4971 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Dec 13 08:00:29 crc kubenswrapper[4971]: I1213 08:00:29.166208 4971 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ceilometer-0"] Dec 13 08:00:29 crc kubenswrapper[4971]: W1213 08:00:29.171055 4971 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod7ba95a4c_be84_460a_9dfc_b6e1474e1b81.slice/crio-f8e32fb170f1a92aef8737c47ffcf139d15c8c82ec5c3778bc5609000b467bca WatchSource:0}: Error finding container f8e32fb170f1a92aef8737c47ffcf139d15c8c82ec5c3778bc5609000b467bca: Status 404 returned error can't find the container with id f8e32fb170f1a92aef8737c47ffcf139d15c8c82ec5c3778bc5609000b467bca Dec 13 08:00:29 crc kubenswrapper[4971]: I1213 08:00:29.331952 4971 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"7ba95a4c-be84-460a-9dfc-b6e1474e1b81","Type":"ContainerStarted","Data":"f8e32fb170f1a92aef8737c47ffcf139d15c8c82ec5c3778bc5609000b467bca"} Dec 13 08:00:29 crc kubenswrapper[4971]: I1213 08:00:29.778232 4971 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="7cfb4dbf-cc55-48b7-b50c-65d6d4ae64ba" path="/var/lib/kubelet/pods/7cfb4dbf-cc55-48b7-b50c-65d6d4ae64ba/volumes" Dec 13 08:00:30 crc kubenswrapper[4971]: I1213 08:00:30.814469 4971 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/manila-share-share1-0" Dec 13 08:00:31 crc kubenswrapper[4971]: I1213 08:00:31.352893 4971 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"7ba95a4c-be84-460a-9dfc-b6e1474e1b81","Type":"ContainerStarted","Data":"6eb2fbba839f492eee9bd57995ecfad27d644708f75ac10616fb47ef3d2ba7b8"} Dec 13 08:00:32 crc kubenswrapper[4971]: I1213 08:00:32.324777 4971 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/manila-scheduler-0" Dec 13 08:00:32 crc kubenswrapper[4971]: I1213 08:00:32.375619 4971 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"7ba95a4c-be84-460a-9dfc-b6e1474e1b81","Type":"ContainerStarted","Data":"840555182ea84d20ab3bac86a12020d2ada7e86a2e28d079b04c9571598febe3"} Dec 13 08:00:32 crc kubenswrapper[4971]: I1213 08:00:32.395007 4971 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/manila-scheduler-0"] Dec 13 08:00:32 crc kubenswrapper[4971]: I1213 08:00:32.395229 4971 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/manila-scheduler-0" podUID="c815d65d-6fc3-49c3-bda5-a2ea21b54438" containerName="manila-scheduler" containerID="cri-o://07f06b3979e8391497d4a21a6994af9fbae02b2bb2e5f2ff570057d31fb6ab8f" gracePeriod=30 Dec 13 08:00:32 crc kubenswrapper[4971]: I1213 08:00:32.395364 4971 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/manila-scheduler-0" podUID="c815d65d-6fc3-49c3-bda5-a2ea21b54438" containerName="probe" containerID="cri-o://303b59dbb886a69396d195934ac67cd67928e83e66852341c56d708cc4b419cf" gracePeriod=30 Dec 13 08:00:33 crc kubenswrapper[4971]: I1213 08:00:33.388251 4971 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"7ba95a4c-be84-460a-9dfc-b6e1474e1b81","Type":"ContainerStarted","Data":"7074485ec5c578d2bdc407cf00901c81ed2bfcb6c580e47523f01fbebf12da23"} Dec 13 08:00:33 crc kubenswrapper[4971]: I1213 08:00:33.390764 4971 generic.go:334] "Generic (PLEG): container finished" podID="c815d65d-6fc3-49c3-bda5-a2ea21b54438" containerID="303b59dbb886a69396d195934ac67cd67928e83e66852341c56d708cc4b419cf" exitCode=0 Dec 13 08:00:33 crc kubenswrapper[4971]: I1213 08:00:33.390815 4971 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/manila-scheduler-0" event={"ID":"c815d65d-6fc3-49c3-bda5-a2ea21b54438","Type":"ContainerDied","Data":"303b59dbb886a69396d195934ac67cd67928e83e66852341c56d708cc4b419cf"} Dec 13 08:00:35 crc kubenswrapper[4971]: I1213 08:00:35.414816 4971 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"7ba95a4c-be84-460a-9dfc-b6e1474e1b81","Type":"ContainerStarted","Data":"9b9fab5fe6a2a26430de9d0f55d4df057baa2da21e7005c00e0e54bdcdfda941"} Dec 13 08:00:35 crc kubenswrapper[4971]: I1213 08:00:35.417111 4971 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/ceilometer-0" Dec 13 08:00:35 crc kubenswrapper[4971]: I1213 08:00:35.420080 4971 generic.go:334] "Generic (PLEG): container finished" podID="c815d65d-6fc3-49c3-bda5-a2ea21b54438" containerID="07f06b3979e8391497d4a21a6994af9fbae02b2bb2e5f2ff570057d31fb6ab8f" exitCode=0 Dec 13 08:00:35 crc kubenswrapper[4971]: I1213 08:00:35.420124 4971 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/manila-scheduler-0" event={"ID":"c815d65d-6fc3-49c3-bda5-a2ea21b54438","Type":"ContainerDied","Data":"07f06b3979e8391497d4a21a6994af9fbae02b2bb2e5f2ff570057d31fb6ab8f"} Dec 13 08:00:35 crc kubenswrapper[4971]: I1213 08:00:35.448207 4971 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/ceilometer-0" podStartSLOduration=1.9129657359999999 podStartE2EDuration="7.448125632s" podCreationTimestamp="2025-12-13 08:00:28 +0000 UTC" firstStartedPulling="2025-12-13 08:00:29.173860496 +0000 UTC m=+4285.778269944" lastFinishedPulling="2025-12-13 08:00:34.709020382 +0000 UTC m=+4291.313429840" observedRunningTime="2025-12-13 08:00:35.445621491 +0000 UTC m=+4292.050030939" watchObservedRunningTime="2025-12-13 08:00:35.448125632 +0000 UTC m=+4292.052535130" Dec 13 08:00:35 crc kubenswrapper[4971]: I1213 08:00:35.949776 4971 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/manila-scheduler-0" Dec 13 08:00:36 crc kubenswrapper[4971]: I1213 08:00:36.108259 4971 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/c815d65d-6fc3-49c3-bda5-a2ea21b54438-combined-ca-bundle\") pod \"c815d65d-6fc3-49c3-bda5-a2ea21b54438\" (UID: \"c815d65d-6fc3-49c3-bda5-a2ea21b54438\") " Dec 13 08:00:36 crc kubenswrapper[4971]: I1213 08:00:36.108304 4971 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/c815d65d-6fc3-49c3-bda5-a2ea21b54438-config-data-custom\") pod \"c815d65d-6fc3-49c3-bda5-a2ea21b54438\" (UID: \"c815d65d-6fc3-49c3-bda5-a2ea21b54438\") " Dec 13 08:00:36 crc kubenswrapper[4971]: I1213 08:00:36.108369 4971 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/c815d65d-6fc3-49c3-bda5-a2ea21b54438-config-data\") pod \"c815d65d-6fc3-49c3-bda5-a2ea21b54438\" (UID: \"c815d65d-6fc3-49c3-bda5-a2ea21b54438\") " Dec 13 08:00:36 crc kubenswrapper[4971]: I1213 08:00:36.108389 4971 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/c815d65d-6fc3-49c3-bda5-a2ea21b54438-scripts\") pod \"c815d65d-6fc3-49c3-bda5-a2ea21b54438\" (UID: \"c815d65d-6fc3-49c3-bda5-a2ea21b54438\") " Dec 13 08:00:36 crc kubenswrapper[4971]: I1213 08:00:36.108484 4971 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/c815d65d-6fc3-49c3-bda5-a2ea21b54438-etc-machine-id\") pod \"c815d65d-6fc3-49c3-bda5-a2ea21b54438\" (UID: \"c815d65d-6fc3-49c3-bda5-a2ea21b54438\") " Dec 13 08:00:36 crc kubenswrapper[4971]: I1213 08:00:36.108683 4971 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-qgmtc\" (UniqueName: \"kubernetes.io/projected/c815d65d-6fc3-49c3-bda5-a2ea21b54438-kube-api-access-qgmtc\") pod \"c815d65d-6fc3-49c3-bda5-a2ea21b54438\" (UID: \"c815d65d-6fc3-49c3-bda5-a2ea21b54438\") " Dec 13 08:00:36 crc kubenswrapper[4971]: I1213 08:00:36.108827 4971 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/c815d65d-6fc3-49c3-bda5-a2ea21b54438-etc-machine-id" (OuterVolumeSpecName: "etc-machine-id") pod "c815d65d-6fc3-49c3-bda5-a2ea21b54438" (UID: "c815d65d-6fc3-49c3-bda5-a2ea21b54438"). InnerVolumeSpecName "etc-machine-id". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 13 08:00:36 crc kubenswrapper[4971]: I1213 08:00:36.109303 4971 reconciler_common.go:293] "Volume detached for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/c815d65d-6fc3-49c3-bda5-a2ea21b54438-etc-machine-id\") on node \"crc\" DevicePath \"\"" Dec 13 08:00:36 crc kubenswrapper[4971]: I1213 08:00:36.114694 4971 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/c815d65d-6fc3-49c3-bda5-a2ea21b54438-scripts" (OuterVolumeSpecName: "scripts") pod "c815d65d-6fc3-49c3-bda5-a2ea21b54438" (UID: "c815d65d-6fc3-49c3-bda5-a2ea21b54438"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 13 08:00:36 crc kubenswrapper[4971]: I1213 08:00:36.133013 4971 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/c815d65d-6fc3-49c3-bda5-a2ea21b54438-kube-api-access-qgmtc" (OuterVolumeSpecName: "kube-api-access-qgmtc") pod "c815d65d-6fc3-49c3-bda5-a2ea21b54438" (UID: "c815d65d-6fc3-49c3-bda5-a2ea21b54438"). InnerVolumeSpecName "kube-api-access-qgmtc". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 13 08:00:36 crc kubenswrapper[4971]: I1213 08:00:36.135116 4971 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/c815d65d-6fc3-49c3-bda5-a2ea21b54438-config-data-custom" (OuterVolumeSpecName: "config-data-custom") pod "c815d65d-6fc3-49c3-bda5-a2ea21b54438" (UID: "c815d65d-6fc3-49c3-bda5-a2ea21b54438"). InnerVolumeSpecName "config-data-custom". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 13 08:00:36 crc kubenswrapper[4971]: I1213 08:00:36.203303 4971 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/c815d65d-6fc3-49c3-bda5-a2ea21b54438-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "c815d65d-6fc3-49c3-bda5-a2ea21b54438" (UID: "c815d65d-6fc3-49c3-bda5-a2ea21b54438"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 13 08:00:36 crc kubenswrapper[4971]: I1213 08:00:36.211633 4971 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-qgmtc\" (UniqueName: \"kubernetes.io/projected/c815d65d-6fc3-49c3-bda5-a2ea21b54438-kube-api-access-qgmtc\") on node \"crc\" DevicePath \"\"" Dec 13 08:00:36 crc kubenswrapper[4971]: I1213 08:00:36.211665 4971 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/c815d65d-6fc3-49c3-bda5-a2ea21b54438-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 13 08:00:36 crc kubenswrapper[4971]: I1213 08:00:36.211674 4971 reconciler_common.go:293] "Volume detached for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/c815d65d-6fc3-49c3-bda5-a2ea21b54438-config-data-custom\") on node \"crc\" DevicePath \"\"" Dec 13 08:00:36 crc kubenswrapper[4971]: I1213 08:00:36.211683 4971 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/c815d65d-6fc3-49c3-bda5-a2ea21b54438-scripts\") on node \"crc\" DevicePath \"\"" Dec 13 08:00:36 crc kubenswrapper[4971]: I1213 08:00:36.222695 4971 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/c815d65d-6fc3-49c3-bda5-a2ea21b54438-config-data" (OuterVolumeSpecName: "config-data") pod "c815d65d-6fc3-49c3-bda5-a2ea21b54438" (UID: "c815d65d-6fc3-49c3-bda5-a2ea21b54438"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 13 08:00:36 crc kubenswrapper[4971]: I1213 08:00:36.313200 4971 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/c815d65d-6fc3-49c3-bda5-a2ea21b54438-config-data\") on node \"crc\" DevicePath \"\"" Dec 13 08:00:36 crc kubenswrapper[4971]: I1213 08:00:36.436403 4971 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/manila-scheduler-0" Dec 13 08:00:36 crc kubenswrapper[4971]: I1213 08:00:36.439273 4971 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/manila-scheduler-0" event={"ID":"c815d65d-6fc3-49c3-bda5-a2ea21b54438","Type":"ContainerDied","Data":"e4d41c347bfa8e9b048bdc28d14e951a680bd1dc6808ae6d20d790a6e7afb521"} Dec 13 08:00:36 crc kubenswrapper[4971]: I1213 08:00:36.439413 4971 scope.go:117] "RemoveContainer" containerID="303b59dbb886a69396d195934ac67cd67928e83e66852341c56d708cc4b419cf" Dec 13 08:00:36 crc kubenswrapper[4971]: I1213 08:00:36.483238 4971 scope.go:117] "RemoveContainer" containerID="07f06b3979e8391497d4a21a6994af9fbae02b2bb2e5f2ff570057d31fb6ab8f" Dec 13 08:00:36 crc kubenswrapper[4971]: I1213 08:00:36.503941 4971 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/manila-scheduler-0"] Dec 13 08:00:36 crc kubenswrapper[4971]: I1213 08:00:36.526996 4971 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/manila-scheduler-0"] Dec 13 08:00:36 crc kubenswrapper[4971]: I1213 08:00:36.543191 4971 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/manila-scheduler-0"] Dec 13 08:00:36 crc kubenswrapper[4971]: E1213 08:00:36.543631 4971 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="c815d65d-6fc3-49c3-bda5-a2ea21b54438" containerName="manila-scheduler" Dec 13 08:00:36 crc kubenswrapper[4971]: I1213 08:00:36.543646 4971 state_mem.go:107] "Deleted CPUSet assignment" podUID="c815d65d-6fc3-49c3-bda5-a2ea21b54438" containerName="manila-scheduler" Dec 13 08:00:36 crc kubenswrapper[4971]: E1213 08:00:36.543665 4971 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="c815d65d-6fc3-49c3-bda5-a2ea21b54438" containerName="probe" Dec 13 08:00:36 crc kubenswrapper[4971]: I1213 08:00:36.543670 4971 state_mem.go:107] "Deleted CPUSet assignment" podUID="c815d65d-6fc3-49c3-bda5-a2ea21b54438" containerName="probe" Dec 13 08:00:36 crc kubenswrapper[4971]: I1213 08:00:36.543856 4971 memory_manager.go:354] "RemoveStaleState removing state" podUID="c815d65d-6fc3-49c3-bda5-a2ea21b54438" containerName="probe" Dec 13 08:00:36 crc kubenswrapper[4971]: I1213 08:00:36.543869 4971 memory_manager.go:354] "RemoveStaleState removing state" podUID="c815d65d-6fc3-49c3-bda5-a2ea21b54438" containerName="manila-scheduler" Dec 13 08:00:36 crc kubenswrapper[4971]: I1213 08:00:36.544834 4971 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/manila-scheduler-0" Dec 13 08:00:36 crc kubenswrapper[4971]: I1213 08:00:36.548258 4971 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"manila-scheduler-config-data" Dec 13 08:00:36 crc kubenswrapper[4971]: I1213 08:00:36.552418 4971 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/manila-scheduler-0"] Dec 13 08:00:36 crc kubenswrapper[4971]: I1213 08:00:36.721197 4971 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/13b76412-5c09-4398-b34f-f07adad9d270-config-data\") pod \"manila-scheduler-0\" (UID: \"13b76412-5c09-4398-b34f-f07adad9d270\") " pod="openstack/manila-scheduler-0" Dec 13 08:00:36 crc kubenswrapper[4971]: I1213 08:00:36.721732 4971 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-4525b\" (UniqueName: \"kubernetes.io/projected/13b76412-5c09-4398-b34f-f07adad9d270-kube-api-access-4525b\") pod \"manila-scheduler-0\" (UID: \"13b76412-5c09-4398-b34f-f07adad9d270\") " pod="openstack/manila-scheduler-0" Dec 13 08:00:36 crc kubenswrapper[4971]: I1213 08:00:36.721834 4971 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/13b76412-5c09-4398-b34f-f07adad9d270-scripts\") pod \"manila-scheduler-0\" (UID: \"13b76412-5c09-4398-b34f-f07adad9d270\") " pod="openstack/manila-scheduler-0" Dec 13 08:00:36 crc kubenswrapper[4971]: I1213 08:00:36.721927 4971 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/13b76412-5c09-4398-b34f-f07adad9d270-etc-machine-id\") pod \"manila-scheduler-0\" (UID: \"13b76412-5c09-4398-b34f-f07adad9d270\") " pod="openstack/manila-scheduler-0" Dec 13 08:00:36 crc kubenswrapper[4971]: I1213 08:00:36.721968 4971 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/13b76412-5c09-4398-b34f-f07adad9d270-combined-ca-bundle\") pod \"manila-scheduler-0\" (UID: \"13b76412-5c09-4398-b34f-f07adad9d270\") " pod="openstack/manila-scheduler-0" Dec 13 08:00:36 crc kubenswrapper[4971]: I1213 08:00:36.722044 4971 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/13b76412-5c09-4398-b34f-f07adad9d270-config-data-custom\") pod \"manila-scheduler-0\" (UID: \"13b76412-5c09-4398-b34f-f07adad9d270\") " pod="openstack/manila-scheduler-0" Dec 13 08:00:36 crc kubenswrapper[4971]: I1213 08:00:36.823740 4971 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/13b76412-5c09-4398-b34f-f07adad9d270-scripts\") pod \"manila-scheduler-0\" (UID: \"13b76412-5c09-4398-b34f-f07adad9d270\") " pod="openstack/manila-scheduler-0" Dec 13 08:00:36 crc kubenswrapper[4971]: I1213 08:00:36.823836 4971 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/13b76412-5c09-4398-b34f-f07adad9d270-etc-machine-id\") pod \"manila-scheduler-0\" (UID: \"13b76412-5c09-4398-b34f-f07adad9d270\") " pod="openstack/manila-scheduler-0" Dec 13 08:00:36 crc kubenswrapper[4971]: I1213 08:00:36.823863 4971 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/13b76412-5c09-4398-b34f-f07adad9d270-combined-ca-bundle\") pod \"manila-scheduler-0\" (UID: \"13b76412-5c09-4398-b34f-f07adad9d270\") " pod="openstack/manila-scheduler-0" Dec 13 08:00:36 crc kubenswrapper[4971]: I1213 08:00:36.823911 4971 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/13b76412-5c09-4398-b34f-f07adad9d270-config-data-custom\") pod \"manila-scheduler-0\" (UID: \"13b76412-5c09-4398-b34f-f07adad9d270\") " pod="openstack/manila-scheduler-0" Dec 13 08:00:36 crc kubenswrapper[4971]: I1213 08:00:36.823959 4971 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/13b76412-5c09-4398-b34f-f07adad9d270-config-data\") pod \"manila-scheduler-0\" (UID: \"13b76412-5c09-4398-b34f-f07adad9d270\") " pod="openstack/manila-scheduler-0" Dec 13 08:00:36 crc kubenswrapper[4971]: I1213 08:00:36.824036 4971 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-4525b\" (UniqueName: \"kubernetes.io/projected/13b76412-5c09-4398-b34f-f07adad9d270-kube-api-access-4525b\") pod \"manila-scheduler-0\" (UID: \"13b76412-5c09-4398-b34f-f07adad9d270\") " pod="openstack/manila-scheduler-0" Dec 13 08:00:36 crc kubenswrapper[4971]: I1213 08:00:36.825736 4971 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/13b76412-5c09-4398-b34f-f07adad9d270-etc-machine-id\") pod \"manila-scheduler-0\" (UID: \"13b76412-5c09-4398-b34f-f07adad9d270\") " pod="openstack/manila-scheduler-0" Dec 13 08:00:36 crc kubenswrapper[4971]: I1213 08:00:36.829249 4971 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/13b76412-5c09-4398-b34f-f07adad9d270-config-data-custom\") pod \"manila-scheduler-0\" (UID: \"13b76412-5c09-4398-b34f-f07adad9d270\") " pod="openstack/manila-scheduler-0" Dec 13 08:00:36 crc kubenswrapper[4971]: I1213 08:00:36.829363 4971 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/13b76412-5c09-4398-b34f-f07adad9d270-scripts\") pod \"manila-scheduler-0\" (UID: \"13b76412-5c09-4398-b34f-f07adad9d270\") " pod="openstack/manila-scheduler-0" Dec 13 08:00:36 crc kubenswrapper[4971]: I1213 08:00:36.829840 4971 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/13b76412-5c09-4398-b34f-f07adad9d270-combined-ca-bundle\") pod \"manila-scheduler-0\" (UID: \"13b76412-5c09-4398-b34f-f07adad9d270\") " pod="openstack/manila-scheduler-0" Dec 13 08:00:36 crc kubenswrapper[4971]: I1213 08:00:36.831071 4971 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/13b76412-5c09-4398-b34f-f07adad9d270-config-data\") pod \"manila-scheduler-0\" (UID: \"13b76412-5c09-4398-b34f-f07adad9d270\") " pod="openstack/manila-scheduler-0" Dec 13 08:00:36 crc kubenswrapper[4971]: I1213 08:00:36.843868 4971 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-4525b\" (UniqueName: \"kubernetes.io/projected/13b76412-5c09-4398-b34f-f07adad9d270-kube-api-access-4525b\") pod \"manila-scheduler-0\" (UID: \"13b76412-5c09-4398-b34f-f07adad9d270\") " pod="openstack/manila-scheduler-0" Dec 13 08:00:36 crc kubenswrapper[4971]: I1213 08:00:36.865666 4971 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/manila-scheduler-0" Dec 13 08:00:37 crc kubenswrapper[4971]: I1213 08:00:37.319393 4971 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/manila-scheduler-0"] Dec 13 08:00:37 crc kubenswrapper[4971]: I1213 08:00:37.450918 4971 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/manila-scheduler-0" event={"ID":"13b76412-5c09-4398-b34f-f07adad9d270","Type":"ContainerStarted","Data":"9afd759dbe4d255ca8a20b0c01b22eaf242da72c995cdf10ea39a747dc9d7e83"} Dec 13 08:00:37 crc kubenswrapper[4971]: I1213 08:00:37.778631 4971 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="c815d65d-6fc3-49c3-bda5-a2ea21b54438" path="/var/lib/kubelet/pods/c815d65d-6fc3-49c3-bda5-a2ea21b54438/volumes" Dec 13 08:00:38 crc kubenswrapper[4971]: I1213 08:00:38.467493 4971 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/manila-scheduler-0" event={"ID":"13b76412-5c09-4398-b34f-f07adad9d270","Type":"ContainerStarted","Data":"0ef569962d1ecf2db921bf756c5c98033d1860bf76eaee9744cd39179f56e0e4"} Dec 13 08:00:38 crc kubenswrapper[4971]: I1213 08:00:38.467872 4971 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/manila-scheduler-0" event={"ID":"13b76412-5c09-4398-b34f-f07adad9d270","Type":"ContainerStarted","Data":"6651628d7d57225dd2ed3e9db7b58212034ce0f41dbc3a405f333079889b54b7"} Dec 13 08:00:38 crc kubenswrapper[4971]: I1213 08:00:38.501735 4971 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/manila-scheduler-0" podStartSLOduration=2.501713622 podStartE2EDuration="2.501713622s" podCreationTimestamp="2025-12-13 08:00:36 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-13 08:00:38.500020422 +0000 UTC m=+4295.104429880" watchObservedRunningTime="2025-12-13 08:00:38.501713622 +0000 UTC m=+4295.106123080" Dec 13 08:00:39 crc kubenswrapper[4971]: I1213 08:00:39.707413 4971 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/manila-api-0" Dec 13 08:00:42 crc kubenswrapper[4971]: I1213 08:00:42.347673 4971 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/manila-share-share1-0" Dec 13 08:00:42 crc kubenswrapper[4971]: I1213 08:00:42.436750 4971 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/manila-share-share1-0"] Dec 13 08:00:42 crc kubenswrapper[4971]: I1213 08:00:42.505068 4971 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/manila-share-share1-0" podUID="1d983511-44dd-4be9-9b43-2f496428a12e" containerName="manila-share" containerID="cri-o://9fb4b92a9f988de4c1a56dac883f2fac0249dc508a477173963001bbbf61c1e9" gracePeriod=30 Dec 13 08:00:42 crc kubenswrapper[4971]: I1213 08:00:42.505217 4971 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/manila-share-share1-0" podUID="1d983511-44dd-4be9-9b43-2f496428a12e" containerName="probe" containerID="cri-o://0399449dfe2e04f246bc19adf9eca94b501aca1e6be2bd628433385f39d17b3d" gracePeriod=30 Dec 13 08:00:43 crc kubenswrapper[4971]: I1213 08:00:43.519640 4971 generic.go:334] "Generic (PLEG): container finished" podID="1d983511-44dd-4be9-9b43-2f496428a12e" containerID="0399449dfe2e04f246bc19adf9eca94b501aca1e6be2bd628433385f39d17b3d" exitCode=0 Dec 13 08:00:43 crc kubenswrapper[4971]: I1213 08:00:43.519980 4971 generic.go:334] "Generic (PLEG): container finished" podID="1d983511-44dd-4be9-9b43-2f496428a12e" containerID="9fb4b92a9f988de4c1a56dac883f2fac0249dc508a477173963001bbbf61c1e9" exitCode=1 Dec 13 08:00:43 crc kubenswrapper[4971]: I1213 08:00:43.519684 4971 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/manila-share-share1-0" event={"ID":"1d983511-44dd-4be9-9b43-2f496428a12e","Type":"ContainerDied","Data":"0399449dfe2e04f246bc19adf9eca94b501aca1e6be2bd628433385f39d17b3d"} Dec 13 08:00:43 crc kubenswrapper[4971]: I1213 08:00:43.520026 4971 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/manila-share-share1-0" event={"ID":"1d983511-44dd-4be9-9b43-2f496428a12e","Type":"ContainerDied","Data":"9fb4b92a9f988de4c1a56dac883f2fac0249dc508a477173963001bbbf61c1e9"} Dec 13 08:00:43 crc kubenswrapper[4971]: E1213 08:00:43.642909 4971 cadvisor_stats_provider.go:516] "Partial failure issuing cadvisor.ContainerInfoV2" err="partial failures: [\"/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod1d983511_44dd_4be9_9b43_2f496428a12e.slice/crio-conmon-0399449dfe2e04f246bc19adf9eca94b501aca1e6be2bd628433385f39d17b3d.scope\": RecentStats: unable to find data in memory cache], [\"/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod1d983511_44dd_4be9_9b43_2f496428a12e.slice/crio-conmon-9fb4b92a9f988de4c1a56dac883f2fac0249dc508a477173963001bbbf61c1e9.scope\": RecentStats: unable to find data in memory cache]" Dec 13 08:00:44 crc kubenswrapper[4971]: I1213 08:00:44.335111 4971 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/manila-share-share1-0" Dec 13 08:00:44 crc kubenswrapper[4971]: I1213 08:00:44.497368 4971 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/1d983511-44dd-4be9-9b43-2f496428a12e-config-data-custom\") pod \"1d983511-44dd-4be9-9b43-2f496428a12e\" (UID: \"1d983511-44dd-4be9-9b43-2f496428a12e\") " Dec 13 08:00:44 crc kubenswrapper[4971]: I1213 08:00:44.497449 4971 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ceph\" (UniqueName: \"kubernetes.io/projected/1d983511-44dd-4be9-9b43-2f496428a12e-ceph\") pod \"1d983511-44dd-4be9-9b43-2f496428a12e\" (UID: \"1d983511-44dd-4be9-9b43-2f496428a12e\") " Dec 13 08:00:44 crc kubenswrapper[4971]: I1213 08:00:44.497482 4971 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/1d983511-44dd-4be9-9b43-2f496428a12e-combined-ca-bundle\") pod \"1d983511-44dd-4be9-9b43-2f496428a12e\" (UID: \"1d983511-44dd-4be9-9b43-2f496428a12e\") " Dec 13 08:00:44 crc kubenswrapper[4971]: I1213 08:00:44.497570 4971 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"var-lib-manila\" (UniqueName: \"kubernetes.io/host-path/1d983511-44dd-4be9-9b43-2f496428a12e-var-lib-manila\") pod \"1d983511-44dd-4be9-9b43-2f496428a12e\" (UID: \"1d983511-44dd-4be9-9b43-2f496428a12e\") " Dec 13 08:00:44 crc kubenswrapper[4971]: I1213 08:00:44.497604 4971 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/1d983511-44dd-4be9-9b43-2f496428a12e-etc-machine-id\") pod \"1d983511-44dd-4be9-9b43-2f496428a12e\" (UID: \"1d983511-44dd-4be9-9b43-2f496428a12e\") " Dec 13 08:00:44 crc kubenswrapper[4971]: I1213 08:00:44.497693 4971 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/1d983511-44dd-4be9-9b43-2f496428a12e-var-lib-manila" (OuterVolumeSpecName: "var-lib-manila") pod "1d983511-44dd-4be9-9b43-2f496428a12e" (UID: "1d983511-44dd-4be9-9b43-2f496428a12e"). InnerVolumeSpecName "var-lib-manila". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 13 08:00:44 crc kubenswrapper[4971]: I1213 08:00:44.497752 4971 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/1d983511-44dd-4be9-9b43-2f496428a12e-scripts\") pod \"1d983511-44dd-4be9-9b43-2f496428a12e\" (UID: \"1d983511-44dd-4be9-9b43-2f496428a12e\") " Dec 13 08:00:44 crc kubenswrapper[4971]: I1213 08:00:44.497770 4971 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/1d983511-44dd-4be9-9b43-2f496428a12e-etc-machine-id" (OuterVolumeSpecName: "etc-machine-id") pod "1d983511-44dd-4be9-9b43-2f496428a12e" (UID: "1d983511-44dd-4be9-9b43-2f496428a12e"). InnerVolumeSpecName "etc-machine-id". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 13 08:00:44 crc kubenswrapper[4971]: I1213 08:00:44.497801 4971 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/1d983511-44dd-4be9-9b43-2f496428a12e-config-data\") pod \"1d983511-44dd-4be9-9b43-2f496428a12e\" (UID: \"1d983511-44dd-4be9-9b43-2f496428a12e\") " Dec 13 08:00:44 crc kubenswrapper[4971]: I1213 08:00:44.497848 4971 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-cb2g9\" (UniqueName: \"kubernetes.io/projected/1d983511-44dd-4be9-9b43-2f496428a12e-kube-api-access-cb2g9\") pod \"1d983511-44dd-4be9-9b43-2f496428a12e\" (UID: \"1d983511-44dd-4be9-9b43-2f496428a12e\") " Dec 13 08:00:44 crc kubenswrapper[4971]: I1213 08:00:44.498358 4971 reconciler_common.go:293] "Volume detached for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/1d983511-44dd-4be9-9b43-2f496428a12e-etc-machine-id\") on node \"crc\" DevicePath \"\"" Dec 13 08:00:44 crc kubenswrapper[4971]: I1213 08:00:44.498381 4971 reconciler_common.go:293] "Volume detached for volume \"var-lib-manila\" (UniqueName: \"kubernetes.io/host-path/1d983511-44dd-4be9-9b43-2f496428a12e-var-lib-manila\") on node \"crc\" DevicePath \"\"" Dec 13 08:00:44 crc kubenswrapper[4971]: I1213 08:00:44.505049 4971 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/1d983511-44dd-4be9-9b43-2f496428a12e-config-data-custom" (OuterVolumeSpecName: "config-data-custom") pod "1d983511-44dd-4be9-9b43-2f496428a12e" (UID: "1d983511-44dd-4be9-9b43-2f496428a12e"). InnerVolumeSpecName "config-data-custom". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 13 08:00:44 crc kubenswrapper[4971]: I1213 08:00:44.505888 4971 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/1d983511-44dd-4be9-9b43-2f496428a12e-scripts" (OuterVolumeSpecName: "scripts") pod "1d983511-44dd-4be9-9b43-2f496428a12e" (UID: "1d983511-44dd-4be9-9b43-2f496428a12e"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 13 08:00:44 crc kubenswrapper[4971]: I1213 08:00:44.505923 4971 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/1d983511-44dd-4be9-9b43-2f496428a12e-ceph" (OuterVolumeSpecName: "ceph") pod "1d983511-44dd-4be9-9b43-2f496428a12e" (UID: "1d983511-44dd-4be9-9b43-2f496428a12e"). InnerVolumeSpecName "ceph". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 13 08:00:44 crc kubenswrapper[4971]: I1213 08:00:44.506003 4971 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/1d983511-44dd-4be9-9b43-2f496428a12e-kube-api-access-cb2g9" (OuterVolumeSpecName: "kube-api-access-cb2g9") pod "1d983511-44dd-4be9-9b43-2f496428a12e" (UID: "1d983511-44dd-4be9-9b43-2f496428a12e"). InnerVolumeSpecName "kube-api-access-cb2g9". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 13 08:00:44 crc kubenswrapper[4971]: I1213 08:00:44.532271 4971 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/manila-share-share1-0" event={"ID":"1d983511-44dd-4be9-9b43-2f496428a12e","Type":"ContainerDied","Data":"2277c634165d87b0973f14d4d41924b9594b3ab384ceba39a9141415b72f416a"} Dec 13 08:00:44 crc kubenswrapper[4971]: I1213 08:00:44.532329 4971 scope.go:117] "RemoveContainer" containerID="0399449dfe2e04f246bc19adf9eca94b501aca1e6be2bd628433385f39d17b3d" Dec 13 08:00:44 crc kubenswrapper[4971]: I1213 08:00:44.532470 4971 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/manila-share-share1-0" Dec 13 08:00:44 crc kubenswrapper[4971]: I1213 08:00:44.572970 4971 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/1d983511-44dd-4be9-9b43-2f496428a12e-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "1d983511-44dd-4be9-9b43-2f496428a12e" (UID: "1d983511-44dd-4be9-9b43-2f496428a12e"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 13 08:00:44 crc kubenswrapper[4971]: I1213 08:00:44.600064 4971 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/1d983511-44dd-4be9-9b43-2f496428a12e-scripts\") on node \"crc\" DevicePath \"\"" Dec 13 08:00:44 crc kubenswrapper[4971]: I1213 08:00:44.600105 4971 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-cb2g9\" (UniqueName: \"kubernetes.io/projected/1d983511-44dd-4be9-9b43-2f496428a12e-kube-api-access-cb2g9\") on node \"crc\" DevicePath \"\"" Dec 13 08:00:44 crc kubenswrapper[4971]: I1213 08:00:44.600120 4971 reconciler_common.go:293] "Volume detached for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/1d983511-44dd-4be9-9b43-2f496428a12e-config-data-custom\") on node \"crc\" DevicePath \"\"" Dec 13 08:00:44 crc kubenswrapper[4971]: I1213 08:00:44.600132 4971 reconciler_common.go:293] "Volume detached for volume \"ceph\" (UniqueName: \"kubernetes.io/projected/1d983511-44dd-4be9-9b43-2f496428a12e-ceph\") on node \"crc\" DevicePath \"\"" Dec 13 08:00:44 crc kubenswrapper[4971]: I1213 08:00:44.600144 4971 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/1d983511-44dd-4be9-9b43-2f496428a12e-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 13 08:00:44 crc kubenswrapper[4971]: I1213 08:00:44.624422 4971 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/1d983511-44dd-4be9-9b43-2f496428a12e-config-data" (OuterVolumeSpecName: "config-data") pod "1d983511-44dd-4be9-9b43-2f496428a12e" (UID: "1d983511-44dd-4be9-9b43-2f496428a12e"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 13 08:00:44 crc kubenswrapper[4971]: I1213 08:00:44.702656 4971 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/1d983511-44dd-4be9-9b43-2f496428a12e-config-data\") on node \"crc\" DevicePath \"\"" Dec 13 08:00:44 crc kubenswrapper[4971]: I1213 08:00:44.712154 4971 scope.go:117] "RemoveContainer" containerID="9fb4b92a9f988de4c1a56dac883f2fac0249dc508a477173963001bbbf61c1e9" Dec 13 08:00:44 crc kubenswrapper[4971]: I1213 08:00:44.872381 4971 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/manila-share-share1-0"] Dec 13 08:00:44 crc kubenswrapper[4971]: I1213 08:00:44.880889 4971 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/manila-share-share1-0"] Dec 13 08:00:44 crc kubenswrapper[4971]: I1213 08:00:44.899155 4971 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/manila-share-share1-0"] Dec 13 08:00:44 crc kubenswrapper[4971]: E1213 08:00:44.899523 4971 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="1d983511-44dd-4be9-9b43-2f496428a12e" containerName="probe" Dec 13 08:00:44 crc kubenswrapper[4971]: I1213 08:00:44.899644 4971 state_mem.go:107] "Deleted CPUSet assignment" podUID="1d983511-44dd-4be9-9b43-2f496428a12e" containerName="probe" Dec 13 08:00:44 crc kubenswrapper[4971]: E1213 08:00:44.899682 4971 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="1d983511-44dd-4be9-9b43-2f496428a12e" containerName="manila-share" Dec 13 08:00:44 crc kubenswrapper[4971]: I1213 08:00:44.899689 4971 state_mem.go:107] "Deleted CPUSet assignment" podUID="1d983511-44dd-4be9-9b43-2f496428a12e" containerName="manila-share" Dec 13 08:00:44 crc kubenswrapper[4971]: I1213 08:00:44.899893 4971 memory_manager.go:354] "RemoveStaleState removing state" podUID="1d983511-44dd-4be9-9b43-2f496428a12e" containerName="probe" Dec 13 08:00:44 crc kubenswrapper[4971]: I1213 08:00:44.899910 4971 memory_manager.go:354] "RemoveStaleState removing state" podUID="1d983511-44dd-4be9-9b43-2f496428a12e" containerName="manila-share" Dec 13 08:00:44 crc kubenswrapper[4971]: I1213 08:00:44.900952 4971 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/manila-share-share1-0" Dec 13 08:00:44 crc kubenswrapper[4971]: I1213 08:00:44.904440 4971 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"manila-share-share1-config-data" Dec 13 08:00:44 crc kubenswrapper[4971]: I1213 08:00:44.922487 4971 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/manila-share-share1-0"] Dec 13 08:00:45 crc kubenswrapper[4971]: I1213 08:00:45.008080 4971 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-lib-manila\" (UniqueName: \"kubernetes.io/host-path/7d1dae22-44f3-47a3-99fa-a281fcffcdd2-var-lib-manila\") pod \"manila-share-share1-0\" (UID: \"7d1dae22-44f3-47a3-99fa-a281fcffcdd2\") " pod="openstack/manila-share-share1-0" Dec 13 08:00:45 crc kubenswrapper[4971]: I1213 08:00:45.008222 4971 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/7d1dae22-44f3-47a3-99fa-a281fcffcdd2-scripts\") pod \"manila-share-share1-0\" (UID: \"7d1dae22-44f3-47a3-99fa-a281fcffcdd2\") " pod="openstack/manila-share-share1-0" Dec 13 08:00:45 crc kubenswrapper[4971]: I1213 08:00:45.008249 4971 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/7d1dae22-44f3-47a3-99fa-a281fcffcdd2-etc-machine-id\") pod \"manila-share-share1-0\" (UID: \"7d1dae22-44f3-47a3-99fa-a281fcffcdd2\") " pod="openstack/manila-share-share1-0" Dec 13 08:00:45 crc kubenswrapper[4971]: I1213 08:00:45.008270 4971 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/7d1dae22-44f3-47a3-99fa-a281fcffcdd2-combined-ca-bundle\") pod \"manila-share-share1-0\" (UID: \"7d1dae22-44f3-47a3-99fa-a281fcffcdd2\") " pod="openstack/manila-share-share1-0" Dec 13 08:00:45 crc kubenswrapper[4971]: I1213 08:00:45.008405 4971 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/7d1dae22-44f3-47a3-99fa-a281fcffcdd2-config-data-custom\") pod \"manila-share-share1-0\" (UID: \"7d1dae22-44f3-47a3-99fa-a281fcffcdd2\") " pod="openstack/manila-share-share1-0" Dec 13 08:00:45 crc kubenswrapper[4971]: I1213 08:00:45.008452 4971 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ceph\" (UniqueName: \"kubernetes.io/projected/7d1dae22-44f3-47a3-99fa-a281fcffcdd2-ceph\") pod \"manila-share-share1-0\" (UID: \"7d1dae22-44f3-47a3-99fa-a281fcffcdd2\") " pod="openstack/manila-share-share1-0" Dec 13 08:00:45 crc kubenswrapper[4971]: I1213 08:00:45.008486 4971 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-s2pdx\" (UniqueName: \"kubernetes.io/projected/7d1dae22-44f3-47a3-99fa-a281fcffcdd2-kube-api-access-s2pdx\") pod \"manila-share-share1-0\" (UID: \"7d1dae22-44f3-47a3-99fa-a281fcffcdd2\") " pod="openstack/manila-share-share1-0" Dec 13 08:00:45 crc kubenswrapper[4971]: I1213 08:00:45.008607 4971 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/7d1dae22-44f3-47a3-99fa-a281fcffcdd2-config-data\") pod \"manila-share-share1-0\" (UID: \"7d1dae22-44f3-47a3-99fa-a281fcffcdd2\") " pod="openstack/manila-share-share1-0" Dec 13 08:00:45 crc kubenswrapper[4971]: I1213 08:00:45.110145 4971 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/7d1dae22-44f3-47a3-99fa-a281fcffcdd2-config-data\") pod \"manila-share-share1-0\" (UID: \"7d1dae22-44f3-47a3-99fa-a281fcffcdd2\") " pod="openstack/manila-share-share1-0" Dec 13 08:00:45 crc kubenswrapper[4971]: I1213 08:00:45.110945 4971 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-lib-manila\" (UniqueName: \"kubernetes.io/host-path/7d1dae22-44f3-47a3-99fa-a281fcffcdd2-var-lib-manila\") pod \"manila-share-share1-0\" (UID: \"7d1dae22-44f3-47a3-99fa-a281fcffcdd2\") " pod="openstack/manila-share-share1-0" Dec 13 08:00:45 crc kubenswrapper[4971]: I1213 08:00:45.111043 4971 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/7d1dae22-44f3-47a3-99fa-a281fcffcdd2-scripts\") pod \"manila-share-share1-0\" (UID: \"7d1dae22-44f3-47a3-99fa-a281fcffcdd2\") " pod="openstack/manila-share-share1-0" Dec 13 08:00:45 crc kubenswrapper[4971]: I1213 08:00:45.111065 4971 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-lib-manila\" (UniqueName: \"kubernetes.io/host-path/7d1dae22-44f3-47a3-99fa-a281fcffcdd2-var-lib-manila\") pod \"manila-share-share1-0\" (UID: \"7d1dae22-44f3-47a3-99fa-a281fcffcdd2\") " pod="openstack/manila-share-share1-0" Dec 13 08:00:45 crc kubenswrapper[4971]: I1213 08:00:45.111073 4971 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/7d1dae22-44f3-47a3-99fa-a281fcffcdd2-etc-machine-id\") pod \"manila-share-share1-0\" (UID: \"7d1dae22-44f3-47a3-99fa-a281fcffcdd2\") " pod="openstack/manila-share-share1-0" Dec 13 08:00:45 crc kubenswrapper[4971]: I1213 08:00:45.111095 4971 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/7d1dae22-44f3-47a3-99fa-a281fcffcdd2-combined-ca-bundle\") pod \"manila-share-share1-0\" (UID: \"7d1dae22-44f3-47a3-99fa-a281fcffcdd2\") " pod="openstack/manila-share-share1-0" Dec 13 08:00:45 crc kubenswrapper[4971]: I1213 08:00:45.111188 4971 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/7d1dae22-44f3-47a3-99fa-a281fcffcdd2-config-data-custom\") pod \"manila-share-share1-0\" (UID: \"7d1dae22-44f3-47a3-99fa-a281fcffcdd2\") " pod="openstack/manila-share-share1-0" Dec 13 08:00:45 crc kubenswrapper[4971]: I1213 08:00:45.111218 4971 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ceph\" (UniqueName: \"kubernetes.io/projected/7d1dae22-44f3-47a3-99fa-a281fcffcdd2-ceph\") pod \"manila-share-share1-0\" (UID: \"7d1dae22-44f3-47a3-99fa-a281fcffcdd2\") " pod="openstack/manila-share-share1-0" Dec 13 08:00:45 crc kubenswrapper[4971]: I1213 08:00:45.111259 4971 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-s2pdx\" (UniqueName: \"kubernetes.io/projected/7d1dae22-44f3-47a3-99fa-a281fcffcdd2-kube-api-access-s2pdx\") pod \"manila-share-share1-0\" (UID: \"7d1dae22-44f3-47a3-99fa-a281fcffcdd2\") " pod="openstack/manila-share-share1-0" Dec 13 08:00:45 crc kubenswrapper[4971]: I1213 08:00:45.112066 4971 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/7d1dae22-44f3-47a3-99fa-a281fcffcdd2-etc-machine-id\") pod \"manila-share-share1-0\" (UID: \"7d1dae22-44f3-47a3-99fa-a281fcffcdd2\") " pod="openstack/manila-share-share1-0" Dec 13 08:00:45 crc kubenswrapper[4971]: I1213 08:00:45.335009 4971 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ceph\" (UniqueName: \"kubernetes.io/projected/7d1dae22-44f3-47a3-99fa-a281fcffcdd2-ceph\") pod \"manila-share-share1-0\" (UID: \"7d1dae22-44f3-47a3-99fa-a281fcffcdd2\") " pod="openstack/manila-share-share1-0" Dec 13 08:00:45 crc kubenswrapper[4971]: I1213 08:00:45.336006 4971 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/7d1dae22-44f3-47a3-99fa-a281fcffcdd2-scripts\") pod \"manila-share-share1-0\" (UID: \"7d1dae22-44f3-47a3-99fa-a281fcffcdd2\") " pod="openstack/manila-share-share1-0" Dec 13 08:00:45 crc kubenswrapper[4971]: I1213 08:00:45.337084 4971 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/7d1dae22-44f3-47a3-99fa-a281fcffcdd2-config-data\") pod \"manila-share-share1-0\" (UID: \"7d1dae22-44f3-47a3-99fa-a281fcffcdd2\") " pod="openstack/manila-share-share1-0" Dec 13 08:00:45 crc kubenswrapper[4971]: I1213 08:00:45.337099 4971 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/7d1dae22-44f3-47a3-99fa-a281fcffcdd2-config-data-custom\") pod \"manila-share-share1-0\" (UID: \"7d1dae22-44f3-47a3-99fa-a281fcffcdd2\") " pod="openstack/manila-share-share1-0" Dec 13 08:00:45 crc kubenswrapper[4971]: I1213 08:00:45.337139 4971 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-s2pdx\" (UniqueName: \"kubernetes.io/projected/7d1dae22-44f3-47a3-99fa-a281fcffcdd2-kube-api-access-s2pdx\") pod \"manila-share-share1-0\" (UID: \"7d1dae22-44f3-47a3-99fa-a281fcffcdd2\") " pod="openstack/manila-share-share1-0" Dec 13 08:00:45 crc kubenswrapper[4971]: I1213 08:00:45.337693 4971 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/7d1dae22-44f3-47a3-99fa-a281fcffcdd2-combined-ca-bundle\") pod \"manila-share-share1-0\" (UID: \"7d1dae22-44f3-47a3-99fa-a281fcffcdd2\") " pod="openstack/manila-share-share1-0" Dec 13 08:00:45 crc kubenswrapper[4971]: I1213 08:00:45.518803 4971 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/manila-share-share1-0" Dec 13 08:00:45 crc kubenswrapper[4971]: I1213 08:00:45.787866 4971 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="1d983511-44dd-4be9-9b43-2f496428a12e" path="/var/lib/kubelet/pods/1d983511-44dd-4be9-9b43-2f496428a12e/volumes" Dec 13 08:00:46 crc kubenswrapper[4971]: I1213 08:00:46.079197 4971 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/manila-share-share1-0"] Dec 13 08:00:46 crc kubenswrapper[4971]: I1213 08:00:46.153686 4971 patch_prober.go:28] interesting pod/machine-config-daemon-82xjz container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 13 08:00:46 crc kubenswrapper[4971]: I1213 08:00:46.154003 4971 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-82xjz" podUID="e0c4af12-edda-41d7-9b44-f87396174ef2" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 13 08:00:46 crc kubenswrapper[4971]: I1213 08:00:46.569094 4971 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/manila-share-share1-0" event={"ID":"7d1dae22-44f3-47a3-99fa-a281fcffcdd2","Type":"ContainerStarted","Data":"cb055ef363cc90e6d9ada3473af0dc0ba3a308de27b3a82307efc0e7c81615d1"} Dec 13 08:00:46 crc kubenswrapper[4971]: I1213 08:00:46.866193 4971 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/manila-scheduler-0" Dec 13 08:00:47 crc kubenswrapper[4971]: I1213 08:00:47.579260 4971 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/manila-share-share1-0" event={"ID":"7d1dae22-44f3-47a3-99fa-a281fcffcdd2","Type":"ContainerStarted","Data":"b1b1a0fbeae1d8fed2561d941185b273c36177a984c10da21c8cd8f59ca8fe01"} Dec 13 08:00:47 crc kubenswrapper[4971]: I1213 08:00:47.579645 4971 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/manila-share-share1-0" event={"ID":"7d1dae22-44f3-47a3-99fa-a281fcffcdd2","Type":"ContainerStarted","Data":"e222ad9983b57900ca6295b9075807fbbae502782a94f08f07b7da2db31b64c7"} Dec 13 08:00:47 crc kubenswrapper[4971]: I1213 08:00:47.603033 4971 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/manila-share-share1-0" podStartSLOduration=3.603011064 podStartE2EDuration="3.603011064s" podCreationTimestamp="2025-12-13 08:00:44 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-13 08:00:47.594107259 +0000 UTC m=+4304.198516707" watchObservedRunningTime="2025-12-13 08:00:47.603011064 +0000 UTC m=+4304.207420512" Dec 13 08:00:55 crc kubenswrapper[4971]: I1213 08:00:55.519390 4971 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/manila-share-share1-0" Dec 13 08:00:58 crc kubenswrapper[4971]: I1213 08:00:58.316315 4971 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/manila-scheduler-0" Dec 13 08:00:58 crc kubenswrapper[4971]: I1213 08:00:58.728295 4971 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/ceilometer-0" Dec 13 08:01:00 crc kubenswrapper[4971]: I1213 08:01:00.167196 4971 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/keystone-cron-29426881-xx4lc"] Dec 13 08:01:00 crc kubenswrapper[4971]: I1213 08:01:00.171943 4971 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-cron-29426881-xx4lc" Dec 13 08:01:00 crc kubenswrapper[4971]: I1213 08:01:00.180796 4971 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/keystone-cron-29426881-xx4lc"] Dec 13 08:01:00 crc kubenswrapper[4971]: I1213 08:01:00.334972 4971 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-7mmgt\" (UniqueName: \"kubernetes.io/projected/13b568d7-cd02-480b-aa55-ef6074c71e60-kube-api-access-7mmgt\") pod \"keystone-cron-29426881-xx4lc\" (UID: \"13b568d7-cd02-480b-aa55-ef6074c71e60\") " pod="openstack/keystone-cron-29426881-xx4lc" Dec 13 08:01:00 crc kubenswrapper[4971]: I1213 08:01:00.335324 4971 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/13b568d7-cd02-480b-aa55-ef6074c71e60-config-data\") pod \"keystone-cron-29426881-xx4lc\" (UID: \"13b568d7-cd02-480b-aa55-ef6074c71e60\") " pod="openstack/keystone-cron-29426881-xx4lc" Dec 13 08:01:00 crc kubenswrapper[4971]: I1213 08:01:00.335413 4971 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/13b568d7-cd02-480b-aa55-ef6074c71e60-combined-ca-bundle\") pod \"keystone-cron-29426881-xx4lc\" (UID: \"13b568d7-cd02-480b-aa55-ef6074c71e60\") " pod="openstack/keystone-cron-29426881-xx4lc" Dec 13 08:01:00 crc kubenswrapper[4971]: I1213 08:01:00.335588 4971 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/13b568d7-cd02-480b-aa55-ef6074c71e60-fernet-keys\") pod \"keystone-cron-29426881-xx4lc\" (UID: \"13b568d7-cd02-480b-aa55-ef6074c71e60\") " pod="openstack/keystone-cron-29426881-xx4lc" Dec 13 08:01:00 crc kubenswrapper[4971]: I1213 08:01:00.437305 4971 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-7mmgt\" (UniqueName: \"kubernetes.io/projected/13b568d7-cd02-480b-aa55-ef6074c71e60-kube-api-access-7mmgt\") pod \"keystone-cron-29426881-xx4lc\" (UID: \"13b568d7-cd02-480b-aa55-ef6074c71e60\") " pod="openstack/keystone-cron-29426881-xx4lc" Dec 13 08:01:00 crc kubenswrapper[4971]: I1213 08:01:00.437709 4971 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/13b568d7-cd02-480b-aa55-ef6074c71e60-config-data\") pod \"keystone-cron-29426881-xx4lc\" (UID: \"13b568d7-cd02-480b-aa55-ef6074c71e60\") " pod="openstack/keystone-cron-29426881-xx4lc" Dec 13 08:01:00 crc kubenswrapper[4971]: I1213 08:01:00.437817 4971 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/13b568d7-cd02-480b-aa55-ef6074c71e60-combined-ca-bundle\") pod \"keystone-cron-29426881-xx4lc\" (UID: \"13b568d7-cd02-480b-aa55-ef6074c71e60\") " pod="openstack/keystone-cron-29426881-xx4lc" Dec 13 08:01:00 crc kubenswrapper[4971]: I1213 08:01:00.438068 4971 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/13b568d7-cd02-480b-aa55-ef6074c71e60-fernet-keys\") pod \"keystone-cron-29426881-xx4lc\" (UID: \"13b568d7-cd02-480b-aa55-ef6074c71e60\") " pod="openstack/keystone-cron-29426881-xx4lc" Dec 13 08:01:00 crc kubenswrapper[4971]: I1213 08:01:00.445074 4971 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/13b568d7-cd02-480b-aa55-ef6074c71e60-combined-ca-bundle\") pod \"keystone-cron-29426881-xx4lc\" (UID: \"13b568d7-cd02-480b-aa55-ef6074c71e60\") " pod="openstack/keystone-cron-29426881-xx4lc" Dec 13 08:01:00 crc kubenswrapper[4971]: I1213 08:01:00.446722 4971 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/13b568d7-cd02-480b-aa55-ef6074c71e60-config-data\") pod \"keystone-cron-29426881-xx4lc\" (UID: \"13b568d7-cd02-480b-aa55-ef6074c71e60\") " pod="openstack/keystone-cron-29426881-xx4lc" Dec 13 08:01:00 crc kubenswrapper[4971]: I1213 08:01:00.457392 4971 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/13b568d7-cd02-480b-aa55-ef6074c71e60-fernet-keys\") pod \"keystone-cron-29426881-xx4lc\" (UID: \"13b568d7-cd02-480b-aa55-ef6074c71e60\") " pod="openstack/keystone-cron-29426881-xx4lc" Dec 13 08:01:00 crc kubenswrapper[4971]: I1213 08:01:00.461555 4971 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-7mmgt\" (UniqueName: \"kubernetes.io/projected/13b568d7-cd02-480b-aa55-ef6074c71e60-kube-api-access-7mmgt\") pod \"keystone-cron-29426881-xx4lc\" (UID: \"13b568d7-cd02-480b-aa55-ef6074c71e60\") " pod="openstack/keystone-cron-29426881-xx4lc" Dec 13 08:01:00 crc kubenswrapper[4971]: I1213 08:01:00.561780 4971 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-cron-29426881-xx4lc" Dec 13 08:01:01 crc kubenswrapper[4971]: W1213 08:01:01.046359 4971 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod13b568d7_cd02_480b_aa55_ef6074c71e60.slice/crio-0e15e625e6ad2bb791453c0db711c33cd5d04e6e5a646ed803ed50145b96ddfe WatchSource:0}: Error finding container 0e15e625e6ad2bb791453c0db711c33cd5d04e6e5a646ed803ed50145b96ddfe: Status 404 returned error can't find the container with id 0e15e625e6ad2bb791453c0db711c33cd5d04e6e5a646ed803ed50145b96ddfe Dec 13 08:01:01 crc kubenswrapper[4971]: I1213 08:01:01.047410 4971 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/keystone-cron-29426881-xx4lc"] Dec 13 08:01:01 crc kubenswrapper[4971]: I1213 08:01:01.733051 4971 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-cron-29426881-xx4lc" event={"ID":"13b568d7-cd02-480b-aa55-ef6074c71e60","Type":"ContainerStarted","Data":"0e15e625e6ad2bb791453c0db711c33cd5d04e6e5a646ed803ed50145b96ddfe"} Dec 13 08:01:02 crc kubenswrapper[4971]: I1213 08:01:02.741439 4971 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-cron-29426881-xx4lc" event={"ID":"13b568d7-cd02-480b-aa55-ef6074c71e60","Type":"ContainerStarted","Data":"2ec3300649df49bd3c7a6f77fa370852ba96c92891c8df782f0c04ef9e7b405e"} Dec 13 08:01:04 crc kubenswrapper[4971]: I1213 08:01:04.792944 4971 generic.go:334] "Generic (PLEG): container finished" podID="13b568d7-cd02-480b-aa55-ef6074c71e60" containerID="2ec3300649df49bd3c7a6f77fa370852ba96c92891c8df782f0c04ef9e7b405e" exitCode=0 Dec 13 08:01:04 crc kubenswrapper[4971]: I1213 08:01:04.793081 4971 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-cron-29426881-xx4lc" event={"ID":"13b568d7-cd02-480b-aa55-ef6074c71e60","Type":"ContainerDied","Data":"2ec3300649df49bd3c7a6f77fa370852ba96c92891c8df782f0c04ef9e7b405e"} Dec 13 08:01:06 crc kubenswrapper[4971]: I1213 08:01:06.215416 4971 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-cron-29426881-xx4lc" Dec 13 08:01:06 crc kubenswrapper[4971]: I1213 08:01:06.362780 4971 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/13b568d7-cd02-480b-aa55-ef6074c71e60-combined-ca-bundle\") pod \"13b568d7-cd02-480b-aa55-ef6074c71e60\" (UID: \"13b568d7-cd02-480b-aa55-ef6074c71e60\") " Dec 13 08:01:06 crc kubenswrapper[4971]: I1213 08:01:06.362872 4971 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-7mmgt\" (UniqueName: \"kubernetes.io/projected/13b568d7-cd02-480b-aa55-ef6074c71e60-kube-api-access-7mmgt\") pod \"13b568d7-cd02-480b-aa55-ef6074c71e60\" (UID: \"13b568d7-cd02-480b-aa55-ef6074c71e60\") " Dec 13 08:01:06 crc kubenswrapper[4971]: I1213 08:01:06.362947 4971 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/13b568d7-cd02-480b-aa55-ef6074c71e60-fernet-keys\") pod \"13b568d7-cd02-480b-aa55-ef6074c71e60\" (UID: \"13b568d7-cd02-480b-aa55-ef6074c71e60\") " Dec 13 08:01:06 crc kubenswrapper[4971]: I1213 08:01:06.363000 4971 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/13b568d7-cd02-480b-aa55-ef6074c71e60-config-data\") pod \"13b568d7-cd02-480b-aa55-ef6074c71e60\" (UID: \"13b568d7-cd02-480b-aa55-ef6074c71e60\") " Dec 13 08:01:06 crc kubenswrapper[4971]: I1213 08:01:06.369495 4971 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/13b568d7-cd02-480b-aa55-ef6074c71e60-kube-api-access-7mmgt" (OuterVolumeSpecName: "kube-api-access-7mmgt") pod "13b568d7-cd02-480b-aa55-ef6074c71e60" (UID: "13b568d7-cd02-480b-aa55-ef6074c71e60"). InnerVolumeSpecName "kube-api-access-7mmgt". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 13 08:01:06 crc kubenswrapper[4971]: I1213 08:01:06.385609 4971 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/13b568d7-cd02-480b-aa55-ef6074c71e60-fernet-keys" (OuterVolumeSpecName: "fernet-keys") pod "13b568d7-cd02-480b-aa55-ef6074c71e60" (UID: "13b568d7-cd02-480b-aa55-ef6074c71e60"). InnerVolumeSpecName "fernet-keys". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 13 08:01:06 crc kubenswrapper[4971]: I1213 08:01:06.393142 4971 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/13b568d7-cd02-480b-aa55-ef6074c71e60-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "13b568d7-cd02-480b-aa55-ef6074c71e60" (UID: "13b568d7-cd02-480b-aa55-ef6074c71e60"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 13 08:01:06 crc kubenswrapper[4971]: I1213 08:01:06.425707 4971 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/13b568d7-cd02-480b-aa55-ef6074c71e60-config-data" (OuterVolumeSpecName: "config-data") pod "13b568d7-cd02-480b-aa55-ef6074c71e60" (UID: "13b568d7-cd02-480b-aa55-ef6074c71e60"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 13 08:01:06 crc kubenswrapper[4971]: I1213 08:01:06.464859 4971 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/13b568d7-cd02-480b-aa55-ef6074c71e60-config-data\") on node \"crc\" DevicePath \"\"" Dec 13 08:01:06 crc kubenswrapper[4971]: I1213 08:01:06.464901 4971 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/13b568d7-cd02-480b-aa55-ef6074c71e60-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 13 08:01:06 crc kubenswrapper[4971]: I1213 08:01:06.464914 4971 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-7mmgt\" (UniqueName: \"kubernetes.io/projected/13b568d7-cd02-480b-aa55-ef6074c71e60-kube-api-access-7mmgt\") on node \"crc\" DevicePath \"\"" Dec 13 08:01:06 crc kubenswrapper[4971]: I1213 08:01:06.464925 4971 reconciler_common.go:293] "Volume detached for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/13b568d7-cd02-480b-aa55-ef6074c71e60-fernet-keys\") on node \"crc\" DevicePath \"\"" Dec 13 08:01:06 crc kubenswrapper[4971]: I1213 08:01:06.817171 4971 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-cron-29426881-xx4lc" event={"ID":"13b568d7-cd02-480b-aa55-ef6074c71e60","Type":"ContainerDied","Data":"0e15e625e6ad2bb791453c0db711c33cd5d04e6e5a646ed803ed50145b96ddfe"} Dec 13 08:01:06 crc kubenswrapper[4971]: I1213 08:01:06.817467 4971 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="0e15e625e6ad2bb791453c0db711c33cd5d04e6e5a646ed803ed50145b96ddfe" Dec 13 08:01:06 crc kubenswrapper[4971]: I1213 08:01:06.817222 4971 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-cron-29426881-xx4lc" Dec 13 08:01:07 crc kubenswrapper[4971]: I1213 08:01:07.016757 4971 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/manila-share-share1-0" Dec 13 08:01:16 crc kubenswrapper[4971]: I1213 08:01:16.153693 4971 patch_prober.go:28] interesting pod/machine-config-daemon-82xjz container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 13 08:01:16 crc kubenswrapper[4971]: I1213 08:01:16.154335 4971 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-82xjz" podUID="e0c4af12-edda-41d7-9b44-f87396174ef2" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 13 08:01:46 crc kubenswrapper[4971]: I1213 08:01:46.154247 4971 patch_prober.go:28] interesting pod/machine-config-daemon-82xjz container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 13 08:01:46 crc kubenswrapper[4971]: I1213 08:01:46.154973 4971 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-82xjz" podUID="e0c4af12-edda-41d7-9b44-f87396174ef2" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 13 08:01:46 crc kubenswrapper[4971]: I1213 08:01:46.155032 4971 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-82xjz" Dec 13 08:01:46 crc kubenswrapper[4971]: I1213 08:01:46.155940 4971 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"e9e9ecdca81954997172b8f30f0cebf348daa8452034fa9732b767f423fdc0de"} pod="openshift-machine-config-operator/machine-config-daemon-82xjz" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Dec 13 08:01:46 crc kubenswrapper[4971]: I1213 08:01:46.156013 4971 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-82xjz" podUID="e0c4af12-edda-41d7-9b44-f87396174ef2" containerName="machine-config-daemon" containerID="cri-o://e9e9ecdca81954997172b8f30f0cebf348daa8452034fa9732b767f423fdc0de" gracePeriod=600 Dec 13 08:01:46 crc kubenswrapper[4971]: E1213 08:01:46.285162 4971 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-82xjz_openshift-machine-config-operator(e0c4af12-edda-41d7-9b44-f87396174ef2)\"" pod="openshift-machine-config-operator/machine-config-daemon-82xjz" podUID="e0c4af12-edda-41d7-9b44-f87396174ef2" Dec 13 08:01:47 crc kubenswrapper[4971]: I1213 08:01:47.248416 4971 generic.go:334] "Generic (PLEG): container finished" podID="e0c4af12-edda-41d7-9b44-f87396174ef2" containerID="e9e9ecdca81954997172b8f30f0cebf348daa8452034fa9732b767f423fdc0de" exitCode=0 Dec 13 08:01:47 crc kubenswrapper[4971]: I1213 08:01:47.248541 4971 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-82xjz" event={"ID":"e0c4af12-edda-41d7-9b44-f87396174ef2","Type":"ContainerDied","Data":"e9e9ecdca81954997172b8f30f0cebf348daa8452034fa9732b767f423fdc0de"} Dec 13 08:01:47 crc kubenswrapper[4971]: I1213 08:01:47.248777 4971 scope.go:117] "RemoveContainer" containerID="d6378ab93da056e5dac3c55cbcf1f29105621255967ad878505c69af98017008" Dec 13 08:01:47 crc kubenswrapper[4971]: I1213 08:01:47.249614 4971 scope.go:117] "RemoveContainer" containerID="e9e9ecdca81954997172b8f30f0cebf348daa8452034fa9732b767f423fdc0de" Dec 13 08:01:47 crc kubenswrapper[4971]: E1213 08:01:47.249873 4971 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-82xjz_openshift-machine-config-operator(e0c4af12-edda-41d7-9b44-f87396174ef2)\"" pod="openshift-machine-config-operator/machine-config-daemon-82xjz" podUID="e0c4af12-edda-41d7-9b44-f87396174ef2" Dec 13 08:01:57 crc kubenswrapper[4971]: I1213 08:01:57.836629 4971 scope.go:117] "RemoveContainer" containerID="e9e9ecdca81954997172b8f30f0cebf348daa8452034fa9732b767f423fdc0de" Dec 13 08:01:57 crc kubenswrapper[4971]: E1213 08:01:57.837405 4971 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-82xjz_openshift-machine-config-operator(e0c4af12-edda-41d7-9b44-f87396174ef2)\"" pod="openshift-machine-config-operator/machine-config-daemon-82xjz" podUID="e0c4af12-edda-41d7-9b44-f87396174ef2" Dec 13 08:02:07 crc kubenswrapper[4971]: I1213 08:02:07.275596 4971 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/openstack-operator-controller-operator-5bf59778d8-rrxtx"] Dec 13 08:02:07 crc kubenswrapper[4971]: E1213 08:02:07.276509 4971 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="13b568d7-cd02-480b-aa55-ef6074c71e60" containerName="keystone-cron" Dec 13 08:02:07 crc kubenswrapper[4971]: I1213 08:02:07.276538 4971 state_mem.go:107] "Deleted CPUSet assignment" podUID="13b568d7-cd02-480b-aa55-ef6074c71e60" containerName="keystone-cron" Dec 13 08:02:07 crc kubenswrapper[4971]: I1213 08:02:07.276715 4971 memory_manager.go:354] "RemoveStaleState removing state" podUID="13b568d7-cd02-480b-aa55-ef6074c71e60" containerName="keystone-cron" Dec 13 08:02:07 crc kubenswrapper[4971]: I1213 08:02:07.277337 4971 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/openstack-operator-controller-operator-5bf59778d8-rrxtx" Dec 13 08:02:07 crc kubenswrapper[4971]: I1213 08:02:07.374137 4971 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/openstack-operator-controller-operator-5bf59778d8-rrxtx"] Dec 13 08:02:07 crc kubenswrapper[4971]: I1213 08:02:07.447396 4971 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-qc9qz\" (UniqueName: \"kubernetes.io/projected/d0ece558-05e2-4e08-84e0-a314f68a453b-kube-api-access-qc9qz\") pod \"openstack-operator-controller-operator-5bf59778d8-rrxtx\" (UID: \"d0ece558-05e2-4e08-84e0-a314f68a453b\") " pod="openstack-operators/openstack-operator-controller-operator-5bf59778d8-rrxtx" Dec 13 08:02:07 crc kubenswrapper[4971]: I1213 08:02:07.549763 4971 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-qc9qz\" (UniqueName: \"kubernetes.io/projected/d0ece558-05e2-4e08-84e0-a314f68a453b-kube-api-access-qc9qz\") pod \"openstack-operator-controller-operator-5bf59778d8-rrxtx\" (UID: \"d0ece558-05e2-4e08-84e0-a314f68a453b\") " pod="openstack-operators/openstack-operator-controller-operator-5bf59778d8-rrxtx" Dec 13 08:02:07 crc kubenswrapper[4971]: I1213 08:02:07.574384 4971 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-qc9qz\" (UniqueName: \"kubernetes.io/projected/d0ece558-05e2-4e08-84e0-a314f68a453b-kube-api-access-qc9qz\") pod \"openstack-operator-controller-operator-5bf59778d8-rrxtx\" (UID: \"d0ece558-05e2-4e08-84e0-a314f68a453b\") " pod="openstack-operators/openstack-operator-controller-operator-5bf59778d8-rrxtx" Dec 13 08:02:07 crc kubenswrapper[4971]: I1213 08:02:07.676363 4971 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/openstack-operator-controller-operator-5bf59778d8-rrxtx" Dec 13 08:02:08 crc kubenswrapper[4971]: I1213 08:02:08.207866 4971 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/openstack-operator-controller-operator-5bf59778d8-rrxtx"] Dec 13 08:02:08 crc kubenswrapper[4971]: I1213 08:02:08.474992 4971 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/openstack-operator-controller-operator-5bf59778d8-rrxtx" event={"ID":"d0ece558-05e2-4e08-84e0-a314f68a453b","Type":"ContainerStarted","Data":"b778a07af4f30598bef071d20c955243151bac7454a47ae10695992abb693e37"} Dec 13 08:02:08 crc kubenswrapper[4971]: I1213 08:02:08.475423 4971 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/openstack-operator-controller-operator-5bf59778d8-rrxtx" Dec 13 08:02:08 crc kubenswrapper[4971]: I1213 08:02:08.475439 4971 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/openstack-operator-controller-operator-5bf59778d8-rrxtx" event={"ID":"d0ece558-05e2-4e08-84e0-a314f68a453b","Type":"ContainerStarted","Data":"0d97fe9bbd4fd5f2fdbdd7dc81f3c8257aaeec0a25754d358012f2f6949d8e86"} Dec 13 08:02:08 crc kubenswrapper[4971]: I1213 08:02:08.502924 4971 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/openstack-operator-controller-operator-5bf59778d8-rrxtx" podStartSLOduration=1.502906629 podStartE2EDuration="1.502906629s" podCreationTimestamp="2025-12-13 08:02:07 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-13 08:02:08.500430199 +0000 UTC m=+4385.104839667" watchObservedRunningTime="2025-12-13 08:02:08.502906629 +0000 UTC m=+4385.107316077" Dec 13 08:02:11 crc kubenswrapper[4971]: I1213 08:02:11.768583 4971 scope.go:117] "RemoveContainer" containerID="e9e9ecdca81954997172b8f30f0cebf348daa8452034fa9732b767f423fdc0de" Dec 13 08:02:11 crc kubenswrapper[4971]: E1213 08:02:11.769314 4971 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-82xjz_openshift-machine-config-operator(e0c4af12-edda-41d7-9b44-f87396174ef2)\"" pod="openshift-machine-config-operator/machine-config-daemon-82xjz" podUID="e0c4af12-edda-41d7-9b44-f87396174ef2" Dec 13 08:02:17 crc kubenswrapper[4971]: I1213 08:02:17.680374 4971 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/openstack-operator-controller-operator-5bf59778d8-rrxtx" Dec 13 08:02:17 crc kubenswrapper[4971]: I1213 08:02:17.782214 4971 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack-operators/openstack-operator-controller-operator-57bbbf4567-7k465"] Dec 13 08:02:17 crc kubenswrapper[4971]: I1213 08:02:17.782395 4971 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack-operators/openstack-operator-controller-operator-57bbbf4567-7k465" podUID="8e098434-439a-4045-b1ec-4fd4facfc8b5" containerName="operator" containerID="cri-o://a26ade213185196b113d9fe376779358a02eb587caef310deb2018d7638d59e0" gracePeriod=10 Dec 13 08:02:18 crc kubenswrapper[4971]: I1213 08:02:18.427287 4971 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack-operators/openstack-operator-controller-operator-57bbbf4567-7k465" Dec 13 08:02:18 crc kubenswrapper[4971]: I1213 08:02:18.501704 4971 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-v2fx8\" (UniqueName: \"kubernetes.io/projected/8e098434-439a-4045-b1ec-4fd4facfc8b5-kube-api-access-v2fx8\") pod \"8e098434-439a-4045-b1ec-4fd4facfc8b5\" (UID: \"8e098434-439a-4045-b1ec-4fd4facfc8b5\") " Dec 13 08:02:18 crc kubenswrapper[4971]: I1213 08:02:18.567427 4971 generic.go:334] "Generic (PLEG): container finished" podID="8e098434-439a-4045-b1ec-4fd4facfc8b5" containerID="a26ade213185196b113d9fe376779358a02eb587caef310deb2018d7638d59e0" exitCode=0 Dec 13 08:02:18 crc kubenswrapper[4971]: I1213 08:02:18.567469 4971 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/openstack-operator-controller-operator-57bbbf4567-7k465" event={"ID":"8e098434-439a-4045-b1ec-4fd4facfc8b5","Type":"ContainerDied","Data":"a26ade213185196b113d9fe376779358a02eb587caef310deb2018d7638d59e0"} Dec 13 08:02:18 crc kubenswrapper[4971]: I1213 08:02:18.567496 4971 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/openstack-operator-controller-operator-57bbbf4567-7k465" event={"ID":"8e098434-439a-4045-b1ec-4fd4facfc8b5","Type":"ContainerDied","Data":"26785ff0eb01b019639095424621c2a90bdb26d995948b6adf9d588225ac9079"} Dec 13 08:02:18 crc kubenswrapper[4971]: I1213 08:02:18.567527 4971 scope.go:117] "RemoveContainer" containerID="a26ade213185196b113d9fe376779358a02eb587caef310deb2018d7638d59e0" Dec 13 08:02:18 crc kubenswrapper[4971]: I1213 08:02:18.567571 4971 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack-operators/openstack-operator-controller-operator-57bbbf4567-7k465" Dec 13 08:02:18 crc kubenswrapper[4971]: I1213 08:02:18.587897 4971 scope.go:117] "RemoveContainer" containerID="a26ade213185196b113d9fe376779358a02eb587caef310deb2018d7638d59e0" Dec 13 08:02:18 crc kubenswrapper[4971]: E1213 08:02:18.589317 4971 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"a26ade213185196b113d9fe376779358a02eb587caef310deb2018d7638d59e0\": container with ID starting with a26ade213185196b113d9fe376779358a02eb587caef310deb2018d7638d59e0 not found: ID does not exist" containerID="a26ade213185196b113d9fe376779358a02eb587caef310deb2018d7638d59e0" Dec 13 08:02:18 crc kubenswrapper[4971]: I1213 08:02:18.589424 4971 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"a26ade213185196b113d9fe376779358a02eb587caef310deb2018d7638d59e0"} err="failed to get container status \"a26ade213185196b113d9fe376779358a02eb587caef310deb2018d7638d59e0\": rpc error: code = NotFound desc = could not find container \"a26ade213185196b113d9fe376779358a02eb587caef310deb2018d7638d59e0\": container with ID starting with a26ade213185196b113d9fe376779358a02eb587caef310deb2018d7638d59e0 not found: ID does not exist" Dec 13 08:02:19 crc kubenswrapper[4971]: I1213 08:02:19.033036 4971 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/8e098434-439a-4045-b1ec-4fd4facfc8b5-kube-api-access-v2fx8" (OuterVolumeSpecName: "kube-api-access-v2fx8") pod "8e098434-439a-4045-b1ec-4fd4facfc8b5" (UID: "8e098434-439a-4045-b1ec-4fd4facfc8b5"). InnerVolumeSpecName "kube-api-access-v2fx8". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 13 08:02:19 crc kubenswrapper[4971]: I1213 08:02:19.129208 4971 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-v2fx8\" (UniqueName: \"kubernetes.io/projected/8e098434-439a-4045-b1ec-4fd4facfc8b5-kube-api-access-v2fx8\") on node \"crc\" DevicePath \"\"" Dec 13 08:02:19 crc kubenswrapper[4971]: I1213 08:02:19.207722 4971 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack-operators/openstack-operator-controller-operator-57bbbf4567-7k465"] Dec 13 08:02:19 crc kubenswrapper[4971]: I1213 08:02:19.216621 4971 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack-operators/openstack-operator-controller-operator-57bbbf4567-7k465"] Dec 13 08:02:19 crc kubenswrapper[4971]: I1213 08:02:19.782920 4971 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="8e098434-439a-4045-b1ec-4fd4facfc8b5" path="/var/lib/kubelet/pods/8e098434-439a-4045-b1ec-4fd4facfc8b5/volumes" Dec 13 08:02:25 crc kubenswrapper[4971]: I1213 08:02:25.768065 4971 scope.go:117] "RemoveContainer" containerID="e9e9ecdca81954997172b8f30f0cebf348daa8452034fa9732b767f423fdc0de" Dec 13 08:02:25 crc kubenswrapper[4971]: E1213 08:02:25.768951 4971 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-82xjz_openshift-machine-config-operator(e0c4af12-edda-41d7-9b44-f87396174ef2)\"" pod="openshift-machine-config-operator/machine-config-daemon-82xjz" podUID="e0c4af12-edda-41d7-9b44-f87396174ef2" Dec 13 08:02:36 crc kubenswrapper[4971]: I1213 08:02:36.769365 4971 scope.go:117] "RemoveContainer" containerID="e9e9ecdca81954997172b8f30f0cebf348daa8452034fa9732b767f423fdc0de" Dec 13 08:02:36 crc kubenswrapper[4971]: E1213 08:02:36.770146 4971 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-82xjz_openshift-machine-config-operator(e0c4af12-edda-41d7-9b44-f87396174ef2)\"" pod="openshift-machine-config-operator/machine-config-daemon-82xjz" podUID="e0c4af12-edda-41d7-9b44-f87396174ef2" Dec 13 08:02:50 crc kubenswrapper[4971]: I1213 08:02:50.768872 4971 scope.go:117] "RemoveContainer" containerID="e9e9ecdca81954997172b8f30f0cebf348daa8452034fa9732b767f423fdc0de" Dec 13 08:02:50 crc kubenswrapper[4971]: E1213 08:02:50.769565 4971 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-82xjz_openshift-machine-config-operator(e0c4af12-edda-41d7-9b44-f87396174ef2)\"" pod="openshift-machine-config-operator/machine-config-daemon-82xjz" podUID="e0c4af12-edda-41d7-9b44-f87396174ef2" Dec 13 08:02:54 crc kubenswrapper[4971]: I1213 08:02:54.335146 4971 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/test-operator-controller-manager-85c4449d86-s4v55"] Dec 13 08:02:54 crc kubenswrapper[4971]: E1213 08:02:54.335957 4971 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="8e098434-439a-4045-b1ec-4fd4facfc8b5" containerName="operator" Dec 13 08:02:54 crc kubenswrapper[4971]: I1213 08:02:54.335970 4971 state_mem.go:107] "Deleted CPUSet assignment" podUID="8e098434-439a-4045-b1ec-4fd4facfc8b5" containerName="operator" Dec 13 08:02:54 crc kubenswrapper[4971]: I1213 08:02:54.336186 4971 memory_manager.go:354] "RemoveStaleState removing state" podUID="8e098434-439a-4045-b1ec-4fd4facfc8b5" containerName="operator" Dec 13 08:02:54 crc kubenswrapper[4971]: I1213 08:02:54.336842 4971 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/test-operator-controller-manager-85c4449d86-s4v55" Dec 13 08:02:54 crc kubenswrapper[4971]: I1213 08:02:54.343802 4971 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/test-operator-controller-manager-85c4449d86-s4v55"] Dec 13 08:02:54 crc kubenswrapper[4971]: I1213 08:02:54.390607 4971 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-9t4zz\" (UniqueName: \"kubernetes.io/projected/ccfc4136-eeb7-4fd8-a4d2-d09dd4a55802-kube-api-access-9t4zz\") pod \"test-operator-controller-manager-85c4449d86-s4v55\" (UID: \"ccfc4136-eeb7-4fd8-a4d2-d09dd4a55802\") " pod="openstack-operators/test-operator-controller-manager-85c4449d86-s4v55" Dec 13 08:02:54 crc kubenswrapper[4971]: I1213 08:02:54.493955 4971 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-9t4zz\" (UniqueName: \"kubernetes.io/projected/ccfc4136-eeb7-4fd8-a4d2-d09dd4a55802-kube-api-access-9t4zz\") pod \"test-operator-controller-manager-85c4449d86-s4v55\" (UID: \"ccfc4136-eeb7-4fd8-a4d2-d09dd4a55802\") " pod="openstack-operators/test-operator-controller-manager-85c4449d86-s4v55" Dec 13 08:02:54 crc kubenswrapper[4971]: I1213 08:02:54.515843 4971 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-9t4zz\" (UniqueName: \"kubernetes.io/projected/ccfc4136-eeb7-4fd8-a4d2-d09dd4a55802-kube-api-access-9t4zz\") pod \"test-operator-controller-manager-85c4449d86-s4v55\" (UID: \"ccfc4136-eeb7-4fd8-a4d2-d09dd4a55802\") " pod="openstack-operators/test-operator-controller-manager-85c4449d86-s4v55" Dec 13 08:02:54 crc kubenswrapper[4971]: I1213 08:02:54.678168 4971 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/test-operator-controller-manager-85c4449d86-s4v55" Dec 13 08:02:55 crc kubenswrapper[4971]: I1213 08:02:55.167755 4971 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/test-operator-controller-manager-85c4449d86-s4v55"] Dec 13 08:02:55 crc kubenswrapper[4971]: I1213 08:02:55.931407 4971 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/test-operator-controller-manager-85c4449d86-s4v55" event={"ID":"ccfc4136-eeb7-4fd8-a4d2-d09dd4a55802","Type":"ContainerStarted","Data":"9b71f9335e993e3154b6912870d01594006e2055490a05be72bc9cef615690d3"} Dec 13 08:02:56 crc kubenswrapper[4971]: I1213 08:02:56.941646 4971 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/test-operator-controller-manager-85c4449d86-s4v55" event={"ID":"ccfc4136-eeb7-4fd8-a4d2-d09dd4a55802","Type":"ContainerStarted","Data":"f9a47a80a3a676846b7eec5cd76bc502533816e800fd10365042f1607d6b8238"} Dec 13 08:02:56 crc kubenswrapper[4971]: I1213 08:02:56.941973 4971 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/test-operator-controller-manager-85c4449d86-s4v55" Dec 13 08:02:56 crc kubenswrapper[4971]: I1213 08:02:56.972585 4971 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/test-operator-controller-manager-85c4449d86-s4v55" podStartSLOduration=1.430497461 podStartE2EDuration="2.972564949s" podCreationTimestamp="2025-12-13 08:02:54 +0000 UTC" firstStartedPulling="2025-12-13 08:02:55.173605538 +0000 UTC m=+4431.778014996" lastFinishedPulling="2025-12-13 08:02:56.715673036 +0000 UTC m=+4433.320082484" observedRunningTime="2025-12-13 08:02:56.955678288 +0000 UTC m=+4433.560087776" watchObservedRunningTime="2025-12-13 08:02:56.972564949 +0000 UTC m=+4433.576974397" Dec 13 08:03:04 crc kubenswrapper[4971]: I1213 08:03:04.681761 4971 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/test-operator-controller-manager-85c4449d86-s4v55" Dec 13 08:03:04 crc kubenswrapper[4971]: I1213 08:03:04.733050 4971 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack-operators/test-operator-controller-manager-756ccf86c7-r8p79"] Dec 13 08:03:04 crc kubenswrapper[4971]: I1213 08:03:04.733306 4971 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack-operators/test-operator-controller-manager-756ccf86c7-r8p79" podUID="ee0625f5-5da4-4dc4-8b2e-f77c9f4459ed" containerName="manager" containerID="cri-o://1da9e8e8e877521032e722c9cfa8f7135044e21263ebfa94f409fb3d06d6e672" gracePeriod=10 Dec 13 08:03:04 crc kubenswrapper[4971]: I1213 08:03:04.770461 4971 scope.go:117] "RemoveContainer" containerID="e9e9ecdca81954997172b8f30f0cebf348daa8452034fa9732b767f423fdc0de" Dec 13 08:03:04 crc kubenswrapper[4971]: E1213 08:03:04.770792 4971 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-82xjz_openshift-machine-config-operator(e0c4af12-edda-41d7-9b44-f87396174ef2)\"" pod="openshift-machine-config-operator/machine-config-daemon-82xjz" podUID="e0c4af12-edda-41d7-9b44-f87396174ef2" Dec 13 08:03:05 crc kubenswrapper[4971]: I1213 08:03:05.027183 4971 generic.go:334] "Generic (PLEG): container finished" podID="ee0625f5-5da4-4dc4-8b2e-f77c9f4459ed" containerID="1da9e8e8e877521032e722c9cfa8f7135044e21263ebfa94f409fb3d06d6e672" exitCode=0 Dec 13 08:03:05 crc kubenswrapper[4971]: I1213 08:03:05.027473 4971 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/test-operator-controller-manager-756ccf86c7-r8p79" event={"ID":"ee0625f5-5da4-4dc4-8b2e-f77c9f4459ed","Type":"ContainerDied","Data":"1da9e8e8e877521032e722c9cfa8f7135044e21263ebfa94f409fb3d06d6e672"} Dec 13 08:03:05 crc kubenswrapper[4971]: I1213 08:03:05.185599 4971 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack-operators/test-operator-controller-manager-756ccf86c7-r8p79" Dec 13 08:03:05 crc kubenswrapper[4971]: I1213 08:03:05.352610 4971 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-cl8ll\" (UniqueName: \"kubernetes.io/projected/ee0625f5-5da4-4dc4-8b2e-f77c9f4459ed-kube-api-access-cl8ll\") pod \"ee0625f5-5da4-4dc4-8b2e-f77c9f4459ed\" (UID: \"ee0625f5-5da4-4dc4-8b2e-f77c9f4459ed\") " Dec 13 08:03:05 crc kubenswrapper[4971]: I1213 08:03:05.954033 4971 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/ee0625f5-5da4-4dc4-8b2e-f77c9f4459ed-kube-api-access-cl8ll" (OuterVolumeSpecName: "kube-api-access-cl8ll") pod "ee0625f5-5da4-4dc4-8b2e-f77c9f4459ed" (UID: "ee0625f5-5da4-4dc4-8b2e-f77c9f4459ed"). InnerVolumeSpecName "kube-api-access-cl8ll". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 13 08:03:05 crc kubenswrapper[4971]: I1213 08:03:05.974722 4971 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-cl8ll\" (UniqueName: \"kubernetes.io/projected/ee0625f5-5da4-4dc4-8b2e-f77c9f4459ed-kube-api-access-cl8ll\") on node \"crc\" DevicePath \"\"" Dec 13 08:03:06 crc kubenswrapper[4971]: I1213 08:03:06.038889 4971 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/test-operator-controller-manager-756ccf86c7-r8p79" event={"ID":"ee0625f5-5da4-4dc4-8b2e-f77c9f4459ed","Type":"ContainerDied","Data":"a2a04b4fa316f8603115e52b0c5cbbdc67f569fb0b6fe26e0bc650d4a7867b4c"} Dec 13 08:03:06 crc kubenswrapper[4971]: I1213 08:03:06.038952 4971 scope.go:117] "RemoveContainer" containerID="1da9e8e8e877521032e722c9cfa8f7135044e21263ebfa94f409fb3d06d6e672" Dec 13 08:03:06 crc kubenswrapper[4971]: I1213 08:03:06.039102 4971 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack-operators/test-operator-controller-manager-756ccf86c7-r8p79" Dec 13 08:03:06 crc kubenswrapper[4971]: I1213 08:03:06.082887 4971 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack-operators/test-operator-controller-manager-756ccf86c7-r8p79"] Dec 13 08:03:06 crc kubenswrapper[4971]: I1213 08:03:06.093570 4971 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack-operators/test-operator-controller-manager-756ccf86c7-r8p79"] Dec 13 08:03:07 crc kubenswrapper[4971]: I1213 08:03:07.787117 4971 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="ee0625f5-5da4-4dc4-8b2e-f77c9f4459ed" path="/var/lib/kubelet/pods/ee0625f5-5da4-4dc4-8b2e-f77c9f4459ed/volumes" Dec 13 08:03:19 crc kubenswrapper[4971]: I1213 08:03:19.768666 4971 scope.go:117] "RemoveContainer" containerID="e9e9ecdca81954997172b8f30f0cebf348daa8452034fa9732b767f423fdc0de" Dec 13 08:03:19 crc kubenswrapper[4971]: E1213 08:03:19.769401 4971 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-82xjz_openshift-machine-config-operator(e0c4af12-edda-41d7-9b44-f87396174ef2)\"" pod="openshift-machine-config-operator/machine-config-daemon-82xjz" podUID="e0c4af12-edda-41d7-9b44-f87396174ef2" Dec 13 08:03:33 crc kubenswrapper[4971]: I1213 08:03:33.777072 4971 scope.go:117] "RemoveContainer" containerID="e9e9ecdca81954997172b8f30f0cebf348daa8452034fa9732b767f423fdc0de" Dec 13 08:03:33 crc kubenswrapper[4971]: E1213 08:03:33.778330 4971 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-82xjz_openshift-machine-config-operator(e0c4af12-edda-41d7-9b44-f87396174ef2)\"" pod="openshift-machine-config-operator/machine-config-daemon-82xjz" podUID="e0c4af12-edda-41d7-9b44-f87396174ef2" Dec 13 08:03:47 crc kubenswrapper[4971]: I1213 08:03:47.768822 4971 scope.go:117] "RemoveContainer" containerID="e9e9ecdca81954997172b8f30f0cebf348daa8452034fa9732b767f423fdc0de" Dec 13 08:03:47 crc kubenswrapper[4971]: E1213 08:03:47.769468 4971 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-82xjz_openshift-machine-config-operator(e0c4af12-edda-41d7-9b44-f87396174ef2)\"" pod="openshift-machine-config-operator/machine-config-daemon-82xjz" podUID="e0c4af12-edda-41d7-9b44-f87396174ef2" Dec 13 08:03:59 crc kubenswrapper[4971]: I1213 08:03:59.769257 4971 scope.go:117] "RemoveContainer" containerID="e9e9ecdca81954997172b8f30f0cebf348daa8452034fa9732b767f423fdc0de" Dec 13 08:03:59 crc kubenswrapper[4971]: E1213 08:03:59.770143 4971 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-82xjz_openshift-machine-config-operator(e0c4af12-edda-41d7-9b44-f87396174ef2)\"" pod="openshift-machine-config-operator/machine-config-daemon-82xjz" podUID="e0c4af12-edda-41d7-9b44-f87396174ef2" Dec 13 08:04:14 crc kubenswrapper[4971]: I1213 08:04:14.768508 4971 scope.go:117] "RemoveContainer" containerID="e9e9ecdca81954997172b8f30f0cebf348daa8452034fa9732b767f423fdc0de" Dec 13 08:04:14 crc kubenswrapper[4971]: E1213 08:04:14.769579 4971 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-82xjz_openshift-machine-config-operator(e0c4af12-edda-41d7-9b44-f87396174ef2)\"" pod="openshift-machine-config-operator/machine-config-daemon-82xjz" podUID="e0c4af12-edda-41d7-9b44-f87396174ef2" Dec 13 08:04:25 crc kubenswrapper[4971]: I1213 08:04:25.768925 4971 scope.go:117] "RemoveContainer" containerID="e9e9ecdca81954997172b8f30f0cebf348daa8452034fa9732b767f423fdc0de" Dec 13 08:04:25 crc kubenswrapper[4971]: E1213 08:04:25.769764 4971 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-82xjz_openshift-machine-config-operator(e0c4af12-edda-41d7-9b44-f87396174ef2)\"" pod="openshift-machine-config-operator/machine-config-daemon-82xjz" podUID="e0c4af12-edda-41d7-9b44-f87396174ef2" Dec 13 08:04:40 crc kubenswrapper[4971]: I1213 08:04:40.769512 4971 scope.go:117] "RemoveContainer" containerID="e9e9ecdca81954997172b8f30f0cebf348daa8452034fa9732b767f423fdc0de" Dec 13 08:04:40 crc kubenswrapper[4971]: E1213 08:04:40.770230 4971 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-82xjz_openshift-machine-config-operator(e0c4af12-edda-41d7-9b44-f87396174ef2)\"" pod="openshift-machine-config-operator/machine-config-daemon-82xjz" podUID="e0c4af12-edda-41d7-9b44-f87396174ef2" Dec 13 08:04:55 crc kubenswrapper[4971]: I1213 08:04:55.768895 4971 scope.go:117] "RemoveContainer" containerID="e9e9ecdca81954997172b8f30f0cebf348daa8452034fa9732b767f423fdc0de" Dec 13 08:04:55 crc kubenswrapper[4971]: E1213 08:04:55.769781 4971 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-82xjz_openshift-machine-config-operator(e0c4af12-edda-41d7-9b44-f87396174ef2)\"" pod="openshift-machine-config-operator/machine-config-daemon-82xjz" podUID="e0c4af12-edda-41d7-9b44-f87396174ef2" Dec 13 08:05:10 crc kubenswrapper[4971]: I1213 08:05:10.769015 4971 scope.go:117] "RemoveContainer" containerID="e9e9ecdca81954997172b8f30f0cebf348daa8452034fa9732b767f423fdc0de" Dec 13 08:05:10 crc kubenswrapper[4971]: E1213 08:05:10.769850 4971 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-82xjz_openshift-machine-config-operator(e0c4af12-edda-41d7-9b44-f87396174ef2)\"" pod="openshift-machine-config-operator/machine-config-daemon-82xjz" podUID="e0c4af12-edda-41d7-9b44-f87396174ef2" Dec 13 08:05:13 crc kubenswrapper[4971]: I1213 08:05:13.575551 4971 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/tempest-tests-tempest-s00-full"] Dec 13 08:05:13 crc kubenswrapper[4971]: E1213 08:05:13.577031 4971 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="ee0625f5-5da4-4dc4-8b2e-f77c9f4459ed" containerName="manager" Dec 13 08:05:13 crc kubenswrapper[4971]: I1213 08:05:13.577067 4971 state_mem.go:107] "Deleted CPUSet assignment" podUID="ee0625f5-5da4-4dc4-8b2e-f77c9f4459ed" containerName="manager" Dec 13 08:05:13 crc kubenswrapper[4971]: I1213 08:05:13.577510 4971 memory_manager.go:354] "RemoveStaleState removing state" podUID="ee0625f5-5da4-4dc4-8b2e-f77c9f4459ed" containerName="manager" Dec 13 08:05:13 crc kubenswrapper[4971]: I1213 08:05:13.578851 4971 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/tempest-tests-tempest-s00-full" Dec 13 08:05:13 crc kubenswrapper[4971]: I1213 08:05:13.583676 4971 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"test-operator-controller-priv-key" Dec 13 08:05:13 crc kubenswrapper[4971]: I1213 08:05:13.583736 4971 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"tempest-tests-tempest-custom-data-s0" Dec 13 08:05:13 crc kubenswrapper[4971]: I1213 08:05:13.584135 4971 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"default-dockercfg-jxsqw" Dec 13 08:05:13 crc kubenswrapper[4971]: I1213 08:05:13.584189 4971 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"tempest-tests-tempest-env-vars-s0" Dec 13 08:05:13 crc kubenswrapper[4971]: I1213 08:05:13.588559 4971 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/tempest-tests-tempest-s00-full"] Dec 13 08:05:13 crc kubenswrapper[4971]: I1213 08:05:13.672611 4971 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ca-certs\" (UniqueName: \"kubernetes.io/secret/879e5242-5aa8-44fb-b759-6babf9c6559d-ca-certs\") pod \"tempest-tests-tempest-s00-full\" (UID: \"879e5242-5aa8-44fb-b759-6babf9c6559d\") " pod="openstack/tempest-tests-tempest-s00-full" Dec 13 08:05:13 crc kubenswrapper[4971]: I1213 08:05:13.672734 4971 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"openstack-config\" (UniqueName: \"kubernetes.io/configmap/879e5242-5aa8-44fb-b759-6babf9c6559d-openstack-config\") pod \"tempest-tests-tempest-s00-full\" (UID: \"879e5242-5aa8-44fb-b759-6babf9c6559d\") " pod="openstack/tempest-tests-tempest-s00-full" Dec 13 08:05:13 crc kubenswrapper[4971]: I1213 08:05:13.672785 4971 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/879e5242-5aa8-44fb-b759-6babf9c6559d-config-data\") pod \"tempest-tests-tempest-s00-full\" (UID: \"879e5242-5aa8-44fb-b759-6babf9c6559d\") " pod="openstack/tempest-tests-tempest-s00-full" Dec 13 08:05:13 crc kubenswrapper[4971]: I1213 08:05:13.672922 4971 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"openstack-config-secret\" (UniqueName: \"kubernetes.io/secret/879e5242-5aa8-44fb-b759-6babf9c6559d-openstack-config-secret\") pod \"tempest-tests-tempest-s00-full\" (UID: \"879e5242-5aa8-44fb-b759-6babf9c6559d\") " pod="openstack/tempest-tests-tempest-s00-full" Dec 13 08:05:13 crc kubenswrapper[4971]: I1213 08:05:13.672999 4971 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"local-storage12-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage12-crc\") pod \"tempest-tests-tempest-s00-full\" (UID: \"879e5242-5aa8-44fb-b759-6babf9c6559d\") " pod="openstack/tempest-tests-tempest-s00-full" Dec 13 08:05:13 crc kubenswrapper[4971]: I1213 08:05:13.673049 4971 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"test-operator-ephemeral-workdir\" (UniqueName: \"kubernetes.io/empty-dir/879e5242-5aa8-44fb-b759-6babf9c6559d-test-operator-ephemeral-workdir\") pod \"tempest-tests-tempest-s00-full\" (UID: \"879e5242-5aa8-44fb-b759-6babf9c6559d\") " pod="openstack/tempest-tests-tempest-s00-full" Dec 13 08:05:13 crc kubenswrapper[4971]: I1213 08:05:13.673074 4971 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/879e5242-5aa8-44fb-b759-6babf9c6559d-ssh-key\") pod \"tempest-tests-tempest-s00-full\" (UID: \"879e5242-5aa8-44fb-b759-6babf9c6559d\") " pod="openstack/tempest-tests-tempest-s00-full" Dec 13 08:05:13 crc kubenswrapper[4971]: I1213 08:05:13.673096 4971 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-n5hn4\" (UniqueName: \"kubernetes.io/projected/879e5242-5aa8-44fb-b759-6babf9c6559d-kube-api-access-n5hn4\") pod \"tempest-tests-tempest-s00-full\" (UID: \"879e5242-5aa8-44fb-b759-6babf9c6559d\") " pod="openstack/tempest-tests-tempest-s00-full" Dec 13 08:05:13 crc kubenswrapper[4971]: I1213 08:05:13.673132 4971 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"test-operator-ephemeral-temporary\" (UniqueName: \"kubernetes.io/empty-dir/879e5242-5aa8-44fb-b759-6babf9c6559d-test-operator-ephemeral-temporary\") pod \"tempest-tests-tempest-s00-full\" (UID: \"879e5242-5aa8-44fb-b759-6babf9c6559d\") " pod="openstack/tempest-tests-tempest-s00-full" Dec 13 08:05:13 crc kubenswrapper[4971]: I1213 08:05:13.673158 4971 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/879e5242-5aa8-44fb-b759-6babf9c6559d-ceph\") pod \"tempest-tests-tempest-s00-full\" (UID: \"879e5242-5aa8-44fb-b759-6babf9c6559d\") " pod="openstack/tempest-tests-tempest-s00-full" Dec 13 08:05:13 crc kubenswrapper[4971]: I1213 08:05:13.775579 4971 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"openstack-config\" (UniqueName: \"kubernetes.io/configmap/879e5242-5aa8-44fb-b759-6babf9c6559d-openstack-config\") pod \"tempest-tests-tempest-s00-full\" (UID: \"879e5242-5aa8-44fb-b759-6babf9c6559d\") " pod="openstack/tempest-tests-tempest-s00-full" Dec 13 08:05:13 crc kubenswrapper[4971]: I1213 08:05:13.775664 4971 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/879e5242-5aa8-44fb-b759-6babf9c6559d-config-data\") pod \"tempest-tests-tempest-s00-full\" (UID: \"879e5242-5aa8-44fb-b759-6babf9c6559d\") " pod="openstack/tempest-tests-tempest-s00-full" Dec 13 08:05:13 crc kubenswrapper[4971]: I1213 08:05:13.775721 4971 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"openstack-config-secret\" (UniqueName: \"kubernetes.io/secret/879e5242-5aa8-44fb-b759-6babf9c6559d-openstack-config-secret\") pod \"tempest-tests-tempest-s00-full\" (UID: \"879e5242-5aa8-44fb-b759-6babf9c6559d\") " pod="openstack/tempest-tests-tempest-s00-full" Dec 13 08:05:13 crc kubenswrapper[4971]: I1213 08:05:13.775810 4971 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"local-storage12-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage12-crc\") pod \"tempest-tests-tempest-s00-full\" (UID: \"879e5242-5aa8-44fb-b759-6babf9c6559d\") " pod="openstack/tempest-tests-tempest-s00-full" Dec 13 08:05:13 crc kubenswrapper[4971]: I1213 08:05:13.775863 4971 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"test-operator-ephemeral-workdir\" (UniqueName: \"kubernetes.io/empty-dir/879e5242-5aa8-44fb-b759-6babf9c6559d-test-operator-ephemeral-workdir\") pod \"tempest-tests-tempest-s00-full\" (UID: \"879e5242-5aa8-44fb-b759-6babf9c6559d\") " pod="openstack/tempest-tests-tempest-s00-full" Dec 13 08:05:13 crc kubenswrapper[4971]: I1213 08:05:13.775888 4971 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/879e5242-5aa8-44fb-b759-6babf9c6559d-ssh-key\") pod \"tempest-tests-tempest-s00-full\" (UID: \"879e5242-5aa8-44fb-b759-6babf9c6559d\") " pod="openstack/tempest-tests-tempest-s00-full" Dec 13 08:05:13 crc kubenswrapper[4971]: I1213 08:05:13.775912 4971 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-n5hn4\" (UniqueName: \"kubernetes.io/projected/879e5242-5aa8-44fb-b759-6babf9c6559d-kube-api-access-n5hn4\") pod \"tempest-tests-tempest-s00-full\" (UID: \"879e5242-5aa8-44fb-b759-6babf9c6559d\") " pod="openstack/tempest-tests-tempest-s00-full" Dec 13 08:05:13 crc kubenswrapper[4971]: I1213 08:05:13.775945 4971 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"test-operator-ephemeral-temporary\" (UniqueName: \"kubernetes.io/empty-dir/879e5242-5aa8-44fb-b759-6babf9c6559d-test-operator-ephemeral-temporary\") pod \"tempest-tests-tempest-s00-full\" (UID: \"879e5242-5aa8-44fb-b759-6babf9c6559d\") " pod="openstack/tempest-tests-tempest-s00-full" Dec 13 08:05:13 crc kubenswrapper[4971]: I1213 08:05:13.775971 4971 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/879e5242-5aa8-44fb-b759-6babf9c6559d-ceph\") pod \"tempest-tests-tempest-s00-full\" (UID: \"879e5242-5aa8-44fb-b759-6babf9c6559d\") " pod="openstack/tempest-tests-tempest-s00-full" Dec 13 08:05:13 crc kubenswrapper[4971]: I1213 08:05:13.776021 4971 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ca-certs\" (UniqueName: \"kubernetes.io/secret/879e5242-5aa8-44fb-b759-6babf9c6559d-ca-certs\") pod \"tempest-tests-tempest-s00-full\" (UID: \"879e5242-5aa8-44fb-b759-6babf9c6559d\") " pod="openstack/tempest-tests-tempest-s00-full" Dec 13 08:05:13 crc kubenswrapper[4971]: I1213 08:05:13.777056 4971 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"openstack-config\" (UniqueName: \"kubernetes.io/configmap/879e5242-5aa8-44fb-b759-6babf9c6559d-openstack-config\") pod \"tempest-tests-tempest-s00-full\" (UID: \"879e5242-5aa8-44fb-b759-6babf9c6559d\") " pod="openstack/tempest-tests-tempest-s00-full" Dec 13 08:05:13 crc kubenswrapper[4971]: I1213 08:05:13.777416 4971 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"local-storage12-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage12-crc\") pod \"tempest-tests-tempest-s00-full\" (UID: \"879e5242-5aa8-44fb-b759-6babf9c6559d\") device mount path \"/mnt/openstack/pv12\"" pod="openstack/tempest-tests-tempest-s00-full" Dec 13 08:05:13 crc kubenswrapper[4971]: I1213 08:05:13.777925 4971 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/879e5242-5aa8-44fb-b759-6babf9c6559d-config-data\") pod \"tempest-tests-tempest-s00-full\" (UID: \"879e5242-5aa8-44fb-b759-6babf9c6559d\") " pod="openstack/tempest-tests-tempest-s00-full" Dec 13 08:05:13 crc kubenswrapper[4971]: I1213 08:05:13.780664 4971 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"test-operator-ephemeral-workdir\" (UniqueName: \"kubernetes.io/empty-dir/879e5242-5aa8-44fb-b759-6babf9c6559d-test-operator-ephemeral-workdir\") pod \"tempest-tests-tempest-s00-full\" (UID: \"879e5242-5aa8-44fb-b759-6babf9c6559d\") " pod="openstack/tempest-tests-tempest-s00-full" Dec 13 08:05:13 crc kubenswrapper[4971]: I1213 08:05:13.780824 4971 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"test-operator-ephemeral-temporary\" (UniqueName: \"kubernetes.io/empty-dir/879e5242-5aa8-44fb-b759-6babf9c6559d-test-operator-ephemeral-temporary\") pod \"tempest-tests-tempest-s00-full\" (UID: \"879e5242-5aa8-44fb-b759-6babf9c6559d\") " pod="openstack/tempest-tests-tempest-s00-full" Dec 13 08:05:13 crc kubenswrapper[4971]: I1213 08:05:13.785141 4971 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/879e5242-5aa8-44fb-b759-6babf9c6559d-ceph\") pod \"tempest-tests-tempest-s00-full\" (UID: \"879e5242-5aa8-44fb-b759-6babf9c6559d\") " pod="openstack/tempest-tests-tempest-s00-full" Dec 13 08:05:13 crc kubenswrapper[4971]: I1213 08:05:13.786276 4971 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ca-certs\" (UniqueName: \"kubernetes.io/secret/879e5242-5aa8-44fb-b759-6babf9c6559d-ca-certs\") pod \"tempest-tests-tempest-s00-full\" (UID: \"879e5242-5aa8-44fb-b759-6babf9c6559d\") " pod="openstack/tempest-tests-tempest-s00-full" Dec 13 08:05:13 crc kubenswrapper[4971]: I1213 08:05:13.787414 4971 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/879e5242-5aa8-44fb-b759-6babf9c6559d-ssh-key\") pod \"tempest-tests-tempest-s00-full\" (UID: \"879e5242-5aa8-44fb-b759-6babf9c6559d\") " pod="openstack/tempest-tests-tempest-s00-full" Dec 13 08:05:13 crc kubenswrapper[4971]: I1213 08:05:13.791189 4971 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"openstack-config-secret\" (UniqueName: \"kubernetes.io/secret/879e5242-5aa8-44fb-b759-6babf9c6559d-openstack-config-secret\") pod \"tempest-tests-tempest-s00-full\" (UID: \"879e5242-5aa8-44fb-b759-6babf9c6559d\") " pod="openstack/tempest-tests-tempest-s00-full" Dec 13 08:05:13 crc kubenswrapper[4971]: I1213 08:05:13.795352 4971 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-n5hn4\" (UniqueName: \"kubernetes.io/projected/879e5242-5aa8-44fb-b759-6babf9c6559d-kube-api-access-n5hn4\") pod \"tempest-tests-tempest-s00-full\" (UID: \"879e5242-5aa8-44fb-b759-6babf9c6559d\") " pod="openstack/tempest-tests-tempest-s00-full" Dec 13 08:05:13 crc kubenswrapper[4971]: I1213 08:05:13.832460 4971 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"local-storage12-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage12-crc\") pod \"tempest-tests-tempest-s00-full\" (UID: \"879e5242-5aa8-44fb-b759-6babf9c6559d\") " pod="openstack/tempest-tests-tempest-s00-full" Dec 13 08:05:13 crc kubenswrapper[4971]: I1213 08:05:13.913925 4971 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/tempest-tests-tempest-s00-full" Dec 13 08:05:14 crc kubenswrapper[4971]: I1213 08:05:14.436511 4971 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/tempest-tests-tempest-s00-full"] Dec 13 08:05:14 crc kubenswrapper[4971]: I1213 08:05:14.439636 4971 provider.go:102] Refreshing cache for provider: *credentialprovider.defaultDockerConfigProvider Dec 13 08:05:14 crc kubenswrapper[4971]: I1213 08:05:14.718008 4971 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/tempest-tests-tempest-s00-full" event={"ID":"879e5242-5aa8-44fb-b759-6babf9c6559d","Type":"ContainerStarted","Data":"9135ef510bc8b400a7dbe31c87cd94bee1f16e0554efd8a7aaf5693eb3e181ef"} Dec 13 08:05:24 crc kubenswrapper[4971]: I1213 08:05:24.114627 4971 scope.go:117] "RemoveContainer" containerID="5bf93847e204b1e7bd63e4ec1873420df1968c262541390c8b0570715a252761" Dec 13 08:05:24 crc kubenswrapper[4971]: I1213 08:05:24.135844 4971 scope.go:117] "RemoveContainer" containerID="b9fe34225f31127367f411a45e14ea2e4cfda0298892c91dbf4db6971bf79192" Dec 13 08:05:24 crc kubenswrapper[4971]: I1213 08:05:24.182034 4971 scope.go:117] "RemoveContainer" containerID="07d758c7b52844138aa133128b536957b1e022a2389d88268ddde52893c33df8" Dec 13 08:05:25 crc kubenswrapper[4971]: I1213 08:05:25.772229 4971 scope.go:117] "RemoveContainer" containerID="e9e9ecdca81954997172b8f30f0cebf348daa8452034fa9732b767f423fdc0de" Dec 13 08:05:25 crc kubenswrapper[4971]: E1213 08:05:25.772828 4971 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-82xjz_openshift-machine-config-operator(e0c4af12-edda-41d7-9b44-f87396174ef2)\"" pod="openshift-machine-config-operator/machine-config-daemon-82xjz" podUID="e0c4af12-edda-41d7-9b44-f87396174ef2" Dec 13 08:05:39 crc kubenswrapper[4971]: I1213 08:05:39.769314 4971 scope.go:117] "RemoveContainer" containerID="e9e9ecdca81954997172b8f30f0cebf348daa8452034fa9732b767f423fdc0de" Dec 13 08:05:39 crc kubenswrapper[4971]: E1213 08:05:39.770142 4971 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-82xjz_openshift-machine-config-operator(e0c4af12-edda-41d7-9b44-f87396174ef2)\"" pod="openshift-machine-config-operator/machine-config-daemon-82xjz" podUID="e0c4af12-edda-41d7-9b44-f87396174ef2" Dec 13 08:05:48 crc kubenswrapper[4971]: E1213 08:05:48.896721 4971 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying config: context canceled" image="quay.io/podified-antelope-centos9/openstack-tempest-all:current-podified" Dec 13 08:05:48 crc kubenswrapper[4971]: E1213 08:05:48.897669 4971 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:tempest-tests-tempest-tests-runner,Image:quay.io/podified-antelope-centos9/openstack-tempest-all:current-podified,Command:[],Args:[],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:config-data,ReadOnly:false,MountPath:/etc/test_operator,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:test-operator-ephemeral-workdir,ReadOnly:false,MountPath:/var/lib/tempest,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:test-operator-ephemeral-temporary,ReadOnly:false,MountPath:/tmp,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:test-operator-logs,ReadOnly:false,MountPath:/var/lib/tempest/external_files,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:openstack-config,ReadOnly:true,MountPath:/etc/openstack/clouds.yaml,SubPath:clouds.yaml,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:openstack-config,ReadOnly:true,MountPath:/var/lib/tempest/.config/openstack/clouds.yaml,SubPath:clouds.yaml,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:openstack-config-secret,ReadOnly:false,MountPath:/etc/openstack/secure.yaml,SubPath:secure.yaml,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:ca-certs,ReadOnly:true,MountPath:/etc/pki/ca-trust/extracted/pem/tls-ca-bundle.pem,SubPath:tls-ca-bundle.pem,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:ssh-key,ReadOnly:false,MountPath:/var/lib/tempest/id_ecdsa,SubPath:ssh_key,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:ceph,ReadOnly:true,MountPath:/etc/ceph,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-n5hn4,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*42480,RunAsNonRoot:*false,ReadOnlyRootFilesystem:*false,AllowPrivilegeEscalation:*true,RunAsGroup:*42480,ProcMount:nil,WindowsOptions:nil,SeccompProfile:&SeccompProfile{Type:RuntimeDefault,LocalhostProfile:nil,},AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{EnvFromSource{Prefix:,ConfigMapRef:&ConfigMapEnvSource{LocalObjectReference:LocalObjectReference{Name:tempest-tests-tempest-custom-data-s0,},Optional:nil,},SecretRef:nil,},EnvFromSource{Prefix:,ConfigMapRef:&ConfigMapEnvSource{LocalObjectReference:LocalObjectReference{Name:tempest-tests-tempest-env-vars-s0,},Optional:nil,},SecretRef:nil,},},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod tempest-tests-tempest-s00-full_openstack(879e5242-5aa8-44fb-b759-6babf9c6559d): ErrImagePull: rpc error: code = Canceled desc = copying config: context canceled" logger="UnhandledError" Dec 13 08:05:48 crc kubenswrapper[4971]: E1213 08:05:48.899153 4971 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"tempest-tests-tempest-tests-runner\" with ErrImagePull: \"rpc error: code = Canceled desc = copying config: context canceled\"" pod="openstack/tempest-tests-tempest-s00-full" podUID="879e5242-5aa8-44fb-b759-6babf9c6559d" Dec 13 08:05:49 crc kubenswrapper[4971]: E1213 08:05:49.119996 4971 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"tempest-tests-tempest-tests-runner\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/podified-antelope-centos9/openstack-tempest-all:current-podified\\\"\"" pod="openstack/tempest-tests-tempest-s00-full" podUID="879e5242-5aa8-44fb-b759-6babf9c6559d" Dec 13 08:05:53 crc kubenswrapper[4971]: I1213 08:05:53.786614 4971 scope.go:117] "RemoveContainer" containerID="e9e9ecdca81954997172b8f30f0cebf348daa8452034fa9732b767f423fdc0de" Dec 13 08:05:53 crc kubenswrapper[4971]: E1213 08:05:53.787534 4971 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-82xjz_openshift-machine-config-operator(e0c4af12-edda-41d7-9b44-f87396174ef2)\"" pod="openshift-machine-config-operator/machine-config-daemon-82xjz" podUID="e0c4af12-edda-41d7-9b44-f87396174ef2" Dec 13 08:06:03 crc kubenswrapper[4971]: I1213 08:06:03.264344 4971 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/tempest-tests-tempest-s00-full" event={"ID":"879e5242-5aa8-44fb-b759-6babf9c6559d","Type":"ContainerStarted","Data":"9a7931b4191657a8f36dd12bd8870834b1b27a528abc6bce6a58dea36f6fa1df"} Dec 13 08:06:03 crc kubenswrapper[4971]: I1213 08:06:03.386818 4971 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/tempest-tests-tempest-s00-full" podStartSLOduration=4.491351902 podStartE2EDuration="51.386790357s" podCreationTimestamp="2025-12-13 08:05:12 +0000 UTC" firstStartedPulling="2025-12-13 08:05:14.439418089 +0000 UTC m=+4571.043827537" lastFinishedPulling="2025-12-13 08:06:01.334856544 +0000 UTC m=+4617.939265992" observedRunningTime="2025-12-13 08:06:03.281408297 +0000 UTC m=+4619.885817745" watchObservedRunningTime="2025-12-13 08:06:03.386790357 +0000 UTC m=+4619.991199805" Dec 13 08:06:08 crc kubenswrapper[4971]: I1213 08:06:08.769680 4971 scope.go:117] "RemoveContainer" containerID="e9e9ecdca81954997172b8f30f0cebf348daa8452034fa9732b767f423fdc0de" Dec 13 08:06:08 crc kubenswrapper[4971]: E1213 08:06:08.770384 4971 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-82xjz_openshift-machine-config-operator(e0c4af12-edda-41d7-9b44-f87396174ef2)\"" pod="openshift-machine-config-operator/machine-config-daemon-82xjz" podUID="e0c4af12-edda-41d7-9b44-f87396174ef2" Dec 13 08:06:11 crc kubenswrapper[4971]: I1213 08:06:11.626852 4971 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/certified-operators-rg4bm"] Dec 13 08:06:11 crc kubenswrapper[4971]: I1213 08:06:11.631066 4971 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-rg4bm" Dec 13 08:06:11 crc kubenswrapper[4971]: I1213 08:06:11.648404 4971 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-rg4bm"] Dec 13 08:06:11 crc kubenswrapper[4971]: I1213 08:06:11.656366 4971 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/d4c962e9-e046-4e64-9843-4fbba2cc434c-utilities\") pod \"certified-operators-rg4bm\" (UID: \"d4c962e9-e046-4e64-9843-4fbba2cc434c\") " pod="openshift-marketplace/certified-operators-rg4bm" Dec 13 08:06:11 crc kubenswrapper[4971]: I1213 08:06:11.656477 4971 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-ksksl\" (UniqueName: \"kubernetes.io/projected/d4c962e9-e046-4e64-9843-4fbba2cc434c-kube-api-access-ksksl\") pod \"certified-operators-rg4bm\" (UID: \"d4c962e9-e046-4e64-9843-4fbba2cc434c\") " pod="openshift-marketplace/certified-operators-rg4bm" Dec 13 08:06:11 crc kubenswrapper[4971]: I1213 08:06:11.656544 4971 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/d4c962e9-e046-4e64-9843-4fbba2cc434c-catalog-content\") pod \"certified-operators-rg4bm\" (UID: \"d4c962e9-e046-4e64-9843-4fbba2cc434c\") " pod="openshift-marketplace/certified-operators-rg4bm" Dec 13 08:06:11 crc kubenswrapper[4971]: I1213 08:06:11.759700 4971 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-ksksl\" (UniqueName: \"kubernetes.io/projected/d4c962e9-e046-4e64-9843-4fbba2cc434c-kube-api-access-ksksl\") pod \"certified-operators-rg4bm\" (UID: \"d4c962e9-e046-4e64-9843-4fbba2cc434c\") " pod="openshift-marketplace/certified-operators-rg4bm" Dec 13 08:06:11 crc kubenswrapper[4971]: I1213 08:06:11.759789 4971 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/d4c962e9-e046-4e64-9843-4fbba2cc434c-catalog-content\") pod \"certified-operators-rg4bm\" (UID: \"d4c962e9-e046-4e64-9843-4fbba2cc434c\") " pod="openshift-marketplace/certified-operators-rg4bm" Dec 13 08:06:11 crc kubenswrapper[4971]: I1213 08:06:11.760012 4971 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/d4c962e9-e046-4e64-9843-4fbba2cc434c-utilities\") pod \"certified-operators-rg4bm\" (UID: \"d4c962e9-e046-4e64-9843-4fbba2cc434c\") " pod="openshift-marketplace/certified-operators-rg4bm" Dec 13 08:06:11 crc kubenswrapper[4971]: I1213 08:06:11.760809 4971 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/d4c962e9-e046-4e64-9843-4fbba2cc434c-utilities\") pod \"certified-operators-rg4bm\" (UID: \"d4c962e9-e046-4e64-9843-4fbba2cc434c\") " pod="openshift-marketplace/certified-operators-rg4bm" Dec 13 08:06:11 crc kubenswrapper[4971]: I1213 08:06:11.761174 4971 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/d4c962e9-e046-4e64-9843-4fbba2cc434c-catalog-content\") pod \"certified-operators-rg4bm\" (UID: \"d4c962e9-e046-4e64-9843-4fbba2cc434c\") " pod="openshift-marketplace/certified-operators-rg4bm" Dec 13 08:06:11 crc kubenswrapper[4971]: I1213 08:06:11.785941 4971 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-ksksl\" (UniqueName: \"kubernetes.io/projected/d4c962e9-e046-4e64-9843-4fbba2cc434c-kube-api-access-ksksl\") pod \"certified-operators-rg4bm\" (UID: \"d4c962e9-e046-4e64-9843-4fbba2cc434c\") " pod="openshift-marketplace/certified-operators-rg4bm" Dec 13 08:06:11 crc kubenswrapper[4971]: I1213 08:06:11.956450 4971 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-rg4bm" Dec 13 08:06:12 crc kubenswrapper[4971]: I1213 08:06:12.532529 4971 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-rg4bm"] Dec 13 08:06:13 crc kubenswrapper[4971]: I1213 08:06:13.405901 4971 generic.go:334] "Generic (PLEG): container finished" podID="d4c962e9-e046-4e64-9843-4fbba2cc434c" containerID="18431a7fe7f4fa10984142ecadfbffaf7df42138b81e73ad53ab30f916f29f35" exitCode=0 Dec 13 08:06:13 crc kubenswrapper[4971]: I1213 08:06:13.406061 4971 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-rg4bm" event={"ID":"d4c962e9-e046-4e64-9843-4fbba2cc434c","Type":"ContainerDied","Data":"18431a7fe7f4fa10984142ecadfbffaf7df42138b81e73ad53ab30f916f29f35"} Dec 13 08:06:13 crc kubenswrapper[4971]: I1213 08:06:13.406493 4971 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-rg4bm" event={"ID":"d4c962e9-e046-4e64-9843-4fbba2cc434c","Type":"ContainerStarted","Data":"c7eeac5dd4766572885413c48b139c75cded4fdf803a9a8962cdde3bc3262834"} Dec 13 08:06:14 crc kubenswrapper[4971]: I1213 08:06:14.417441 4971 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-rg4bm" event={"ID":"d4c962e9-e046-4e64-9843-4fbba2cc434c","Type":"ContainerStarted","Data":"fd65a132350d969bdb1e10fbc9c0ac567403a6d30357fc7e7ebd0ad047e7ffb0"} Dec 13 08:06:15 crc kubenswrapper[4971]: I1213 08:06:15.434367 4971 generic.go:334] "Generic (PLEG): container finished" podID="d4c962e9-e046-4e64-9843-4fbba2cc434c" containerID="fd65a132350d969bdb1e10fbc9c0ac567403a6d30357fc7e7ebd0ad047e7ffb0" exitCode=0 Dec 13 08:06:15 crc kubenswrapper[4971]: I1213 08:06:15.434465 4971 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-rg4bm" event={"ID":"d4c962e9-e046-4e64-9843-4fbba2cc434c","Type":"ContainerDied","Data":"fd65a132350d969bdb1e10fbc9c0ac567403a6d30357fc7e7ebd0ad047e7ffb0"} Dec 13 08:06:16 crc kubenswrapper[4971]: I1213 08:06:16.446777 4971 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-rg4bm" event={"ID":"d4c962e9-e046-4e64-9843-4fbba2cc434c","Type":"ContainerStarted","Data":"5b875af5f2a5d77169b3956d4995571e67a17dd9f171140d1a34c93221f648d9"} Dec 13 08:06:16 crc kubenswrapper[4971]: I1213 08:06:16.484713 4971 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/certified-operators-rg4bm" podStartSLOduration=2.955582979 podStartE2EDuration="5.484690165s" podCreationTimestamp="2025-12-13 08:06:11 +0000 UTC" firstStartedPulling="2025-12-13 08:06:13.408468091 +0000 UTC m=+4630.012877539" lastFinishedPulling="2025-12-13 08:06:15.937575277 +0000 UTC m=+4632.541984725" observedRunningTime="2025-12-13 08:06:16.471667454 +0000 UTC m=+4633.076076902" watchObservedRunningTime="2025-12-13 08:06:16.484690165 +0000 UTC m=+4633.089099623" Dec 13 08:06:21 crc kubenswrapper[4971]: I1213 08:06:21.957001 4971 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/certified-operators-rg4bm" Dec 13 08:06:21 crc kubenswrapper[4971]: I1213 08:06:21.957651 4971 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/certified-operators-rg4bm" Dec 13 08:06:22 crc kubenswrapper[4971]: I1213 08:06:22.027403 4971 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/certified-operators-rg4bm" Dec 13 08:06:22 crc kubenswrapper[4971]: I1213 08:06:22.573308 4971 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/certified-operators-rg4bm" Dec 13 08:06:22 crc kubenswrapper[4971]: I1213 08:06:22.769162 4971 scope.go:117] "RemoveContainer" containerID="e9e9ecdca81954997172b8f30f0cebf348daa8452034fa9732b767f423fdc0de" Dec 13 08:06:22 crc kubenswrapper[4971]: E1213 08:06:22.769492 4971 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-82xjz_openshift-machine-config-operator(e0c4af12-edda-41d7-9b44-f87396174ef2)\"" pod="openshift-machine-config-operator/machine-config-daemon-82xjz" podUID="e0c4af12-edda-41d7-9b44-f87396174ef2" Dec 13 08:06:22 crc kubenswrapper[4971]: I1213 08:06:22.800036 4971 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-rg4bm"] Dec 13 08:06:24 crc kubenswrapper[4971]: I1213 08:06:24.527488 4971 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/certified-operators-rg4bm" podUID="d4c962e9-e046-4e64-9843-4fbba2cc434c" containerName="registry-server" containerID="cri-o://5b875af5f2a5d77169b3956d4995571e67a17dd9f171140d1a34c93221f648d9" gracePeriod=2 Dec 13 08:06:25 crc kubenswrapper[4971]: I1213 08:06:25.554338 4971 generic.go:334] "Generic (PLEG): container finished" podID="d4c962e9-e046-4e64-9843-4fbba2cc434c" containerID="5b875af5f2a5d77169b3956d4995571e67a17dd9f171140d1a34c93221f648d9" exitCode=0 Dec 13 08:06:25 crc kubenswrapper[4971]: I1213 08:06:25.554449 4971 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-rg4bm" event={"ID":"d4c962e9-e046-4e64-9843-4fbba2cc434c","Type":"ContainerDied","Data":"5b875af5f2a5d77169b3956d4995571e67a17dd9f171140d1a34c93221f648d9"} Dec 13 08:06:26 crc kubenswrapper[4971]: I1213 08:06:26.581663 4971 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-rg4bm" event={"ID":"d4c962e9-e046-4e64-9843-4fbba2cc434c","Type":"ContainerDied","Data":"c7eeac5dd4766572885413c48b139c75cded4fdf803a9a8962cdde3bc3262834"} Dec 13 08:06:26 crc kubenswrapper[4971]: I1213 08:06:26.582033 4971 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="c7eeac5dd4766572885413c48b139c75cded4fdf803a9a8962cdde3bc3262834" Dec 13 08:06:26 crc kubenswrapper[4971]: I1213 08:06:26.670560 4971 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-rg4bm" Dec 13 08:06:26 crc kubenswrapper[4971]: I1213 08:06:26.786310 4971 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/d4c962e9-e046-4e64-9843-4fbba2cc434c-catalog-content\") pod \"d4c962e9-e046-4e64-9843-4fbba2cc434c\" (UID: \"d4c962e9-e046-4e64-9843-4fbba2cc434c\") " Dec 13 08:06:26 crc kubenswrapper[4971]: I1213 08:06:26.786982 4971 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/d4c962e9-e046-4e64-9843-4fbba2cc434c-utilities\") pod \"d4c962e9-e046-4e64-9843-4fbba2cc434c\" (UID: \"d4c962e9-e046-4e64-9843-4fbba2cc434c\") " Dec 13 08:06:26 crc kubenswrapper[4971]: I1213 08:06:26.787133 4971 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-ksksl\" (UniqueName: \"kubernetes.io/projected/d4c962e9-e046-4e64-9843-4fbba2cc434c-kube-api-access-ksksl\") pod \"d4c962e9-e046-4e64-9843-4fbba2cc434c\" (UID: \"d4c962e9-e046-4e64-9843-4fbba2cc434c\") " Dec 13 08:06:26 crc kubenswrapper[4971]: I1213 08:06:26.788129 4971 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/d4c962e9-e046-4e64-9843-4fbba2cc434c-utilities" (OuterVolumeSpecName: "utilities") pod "d4c962e9-e046-4e64-9843-4fbba2cc434c" (UID: "d4c962e9-e046-4e64-9843-4fbba2cc434c"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 13 08:06:26 crc kubenswrapper[4971]: I1213 08:06:26.796286 4971 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/d4c962e9-e046-4e64-9843-4fbba2cc434c-kube-api-access-ksksl" (OuterVolumeSpecName: "kube-api-access-ksksl") pod "d4c962e9-e046-4e64-9843-4fbba2cc434c" (UID: "d4c962e9-e046-4e64-9843-4fbba2cc434c"). InnerVolumeSpecName "kube-api-access-ksksl". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 13 08:06:26 crc kubenswrapper[4971]: I1213 08:06:26.855724 4971 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/d4c962e9-e046-4e64-9843-4fbba2cc434c-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "d4c962e9-e046-4e64-9843-4fbba2cc434c" (UID: "d4c962e9-e046-4e64-9843-4fbba2cc434c"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 13 08:06:26 crc kubenswrapper[4971]: I1213 08:06:26.889618 4971 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/d4c962e9-e046-4e64-9843-4fbba2cc434c-utilities\") on node \"crc\" DevicePath \"\"" Dec 13 08:06:26 crc kubenswrapper[4971]: I1213 08:06:26.889657 4971 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-ksksl\" (UniqueName: \"kubernetes.io/projected/d4c962e9-e046-4e64-9843-4fbba2cc434c-kube-api-access-ksksl\") on node \"crc\" DevicePath \"\"" Dec 13 08:06:26 crc kubenswrapper[4971]: I1213 08:06:26.889669 4971 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/d4c962e9-e046-4e64-9843-4fbba2cc434c-catalog-content\") on node \"crc\" DevicePath \"\"" Dec 13 08:06:27 crc kubenswrapper[4971]: I1213 08:06:27.590579 4971 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-rg4bm" Dec 13 08:06:27 crc kubenswrapper[4971]: I1213 08:06:27.636611 4971 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-rg4bm"] Dec 13 08:06:27 crc kubenswrapper[4971]: I1213 08:06:27.647162 4971 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/certified-operators-rg4bm"] Dec 13 08:06:27 crc kubenswrapper[4971]: I1213 08:06:27.787428 4971 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="d4c962e9-e046-4e64-9843-4fbba2cc434c" path="/var/lib/kubelet/pods/d4c962e9-e046-4e64-9843-4fbba2cc434c/volumes" Dec 13 08:06:37 crc kubenswrapper[4971]: I1213 08:06:37.769675 4971 scope.go:117] "RemoveContainer" containerID="e9e9ecdca81954997172b8f30f0cebf348daa8452034fa9732b767f423fdc0de" Dec 13 08:06:37 crc kubenswrapper[4971]: E1213 08:06:37.770587 4971 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-82xjz_openshift-machine-config-operator(e0c4af12-edda-41d7-9b44-f87396174ef2)\"" pod="openshift-machine-config-operator/machine-config-daemon-82xjz" podUID="e0c4af12-edda-41d7-9b44-f87396174ef2" Dec 13 08:06:48 crc kubenswrapper[4971]: I1213 08:06:48.769107 4971 scope.go:117] "RemoveContainer" containerID="e9e9ecdca81954997172b8f30f0cebf348daa8452034fa9732b767f423fdc0de" Dec 13 08:06:49 crc kubenswrapper[4971]: I1213 08:06:49.870645 4971 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-82xjz" event={"ID":"e0c4af12-edda-41d7-9b44-f87396174ef2","Type":"ContainerStarted","Data":"fd226c5cf0ac150e1a98fe2ca7739d374a376da0b0db1d061f54febf1a32844a"} Dec 13 08:08:28 crc kubenswrapper[4971]: I1213 08:08:28.735642 4971 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/openstack-galera-0" podUID="91831fd5-fc52-46ec-ac49-63859378116b" containerName="galera" probeResult="failure" output="command timed out" Dec 13 08:08:28 crc kubenswrapper[4971]: I1213 08:08:28.735920 4971 prober.go:107] "Probe failed" probeType="Liveness" pod="openstack/openstack-galera-0" podUID="91831fd5-fc52-46ec-ac49-63859378116b" containerName="galera" probeResult="failure" output="command timed out" Dec 13 08:09:16 crc kubenswrapper[4971]: I1213 08:09:16.153411 4971 patch_prober.go:28] interesting pod/machine-config-daemon-82xjz container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 13 08:09:16 crc kubenswrapper[4971]: I1213 08:09:16.154111 4971 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-82xjz" podUID="e0c4af12-edda-41d7-9b44-f87396174ef2" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 13 08:09:46 crc kubenswrapper[4971]: I1213 08:09:46.153900 4971 patch_prober.go:28] interesting pod/machine-config-daemon-82xjz container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 13 08:09:46 crc kubenswrapper[4971]: I1213 08:09:46.154735 4971 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-82xjz" podUID="e0c4af12-edda-41d7-9b44-f87396174ef2" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 13 08:09:47 crc kubenswrapper[4971]: I1213 08:09:47.083954 4971 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/manila-db-create-wmjhn"] Dec 13 08:09:47 crc kubenswrapper[4971]: I1213 08:09:47.099068 4971 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/manila-db-create-wmjhn"] Dec 13 08:09:47 crc kubenswrapper[4971]: I1213 08:09:47.787194 4971 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="44d3eebd-3532-454e-b687-dd752f549afe" path="/var/lib/kubelet/pods/44d3eebd-3532-454e-b687-dd752f549afe/volumes" Dec 13 08:09:48 crc kubenswrapper[4971]: I1213 08:09:48.043002 4971 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/manila-8949-account-create-update-6pkvl"] Dec 13 08:09:48 crc kubenswrapper[4971]: I1213 08:09:48.054782 4971 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/manila-8949-account-create-update-6pkvl"] Dec 13 08:09:49 crc kubenswrapper[4971]: I1213 08:09:49.783557 4971 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="ed7c6200-6e0f-4865-81a8-7eb5c39a996f" path="/var/lib/kubelet/pods/ed7c6200-6e0f-4865-81a8-7eb5c39a996f/volumes" Dec 13 08:10:09 crc kubenswrapper[4971]: I1213 08:10:09.047451 4971 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/manila-db-sync-g9bws"] Dec 13 08:10:09 crc kubenswrapper[4971]: I1213 08:10:09.060327 4971 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/manila-db-sync-g9bws"] Dec 13 08:10:09 crc kubenswrapper[4971]: I1213 08:10:09.784019 4971 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="bb2c5402-103a-4456-960f-7b0019111f1e" path="/var/lib/kubelet/pods/bb2c5402-103a-4456-960f-7b0019111f1e/volumes" Dec 13 08:10:16 crc kubenswrapper[4971]: I1213 08:10:16.154044 4971 patch_prober.go:28] interesting pod/machine-config-daemon-82xjz container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 13 08:10:16 crc kubenswrapper[4971]: I1213 08:10:16.155017 4971 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-82xjz" podUID="e0c4af12-edda-41d7-9b44-f87396174ef2" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 13 08:10:16 crc kubenswrapper[4971]: I1213 08:10:16.155107 4971 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-82xjz" Dec 13 08:10:16 crc kubenswrapper[4971]: I1213 08:10:16.156542 4971 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"fd226c5cf0ac150e1a98fe2ca7739d374a376da0b0db1d061f54febf1a32844a"} pod="openshift-machine-config-operator/machine-config-daemon-82xjz" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Dec 13 08:10:16 crc kubenswrapper[4971]: I1213 08:10:16.156619 4971 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-82xjz" podUID="e0c4af12-edda-41d7-9b44-f87396174ef2" containerName="machine-config-daemon" containerID="cri-o://fd226c5cf0ac150e1a98fe2ca7739d374a376da0b0db1d061f54febf1a32844a" gracePeriod=600 Dec 13 08:10:24 crc kubenswrapper[4971]: I1213 08:10:24.342076 4971 scope.go:117] "RemoveContainer" containerID="0339a9bf168c17c0589ea918f3a679eadb4868cae95155960974063bdd0efa2c" Dec 13 08:10:25 crc kubenswrapper[4971]: I1213 08:10:25.477255 4971 trace.go:236] Trace[326465958]: "Calculate volume metrics of registry-storage for pod openshift-image-registry/image-registry-66df7c8f76-xjwwz" (13-Dec-2025 08:10:21.037) (total time: 4439ms): Dec 13 08:10:25 crc kubenswrapper[4971]: Trace[326465958]: [4.439735766s] [4.439735766s] END Dec 13 08:10:25 crc kubenswrapper[4971]: I1213 08:10:25.690962 4971 scope.go:117] "RemoveContainer" containerID="f3d1cf893616810f172fc00110ecfa59d1af17a2991bd2c6dbe7580dee57f25d" Dec 13 08:10:25 crc kubenswrapper[4971]: I1213 08:10:25.805506 4971 scope.go:117] "RemoveContainer" containerID="5ee1aa5ebf8687e00fb713d9535ae93929f98ad14f5e297b30bffdc3870be18e" Dec 13 08:10:26 crc kubenswrapper[4971]: I1213 08:10:26.418378 4971 generic.go:334] "Generic (PLEG): container finished" podID="e0c4af12-edda-41d7-9b44-f87396174ef2" containerID="fd226c5cf0ac150e1a98fe2ca7739d374a376da0b0db1d061f54febf1a32844a" exitCode=0 Dec 13 08:10:26 crc kubenswrapper[4971]: I1213 08:10:26.419432 4971 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-82xjz" event={"ID":"e0c4af12-edda-41d7-9b44-f87396174ef2","Type":"ContainerDied","Data":"fd226c5cf0ac150e1a98fe2ca7739d374a376da0b0db1d061f54febf1a32844a"} Dec 13 08:10:26 crc kubenswrapper[4971]: I1213 08:10:26.421715 4971 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-82xjz" event={"ID":"e0c4af12-edda-41d7-9b44-f87396174ef2","Type":"ContainerStarted","Data":"8b2617257e0c4b9039d9f1e54400a1b30b3fb18612ede5bd959442a8eca72ecd"} Dec 13 08:10:26 crc kubenswrapper[4971]: I1213 08:10:26.421775 4971 scope.go:117] "RemoveContainer" containerID="e9e9ecdca81954997172b8f30f0cebf348daa8452034fa9732b767f423fdc0de" Dec 13 08:10:44 crc kubenswrapper[4971]: I1213 08:10:44.395622 4971 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-operators-r64xp"] Dec 13 08:10:44 crc kubenswrapper[4971]: E1213 08:10:44.397028 4971 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="d4c962e9-e046-4e64-9843-4fbba2cc434c" containerName="extract-content" Dec 13 08:10:44 crc kubenswrapper[4971]: I1213 08:10:44.397052 4971 state_mem.go:107] "Deleted CPUSet assignment" podUID="d4c962e9-e046-4e64-9843-4fbba2cc434c" containerName="extract-content" Dec 13 08:10:44 crc kubenswrapper[4971]: E1213 08:10:44.397119 4971 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="d4c962e9-e046-4e64-9843-4fbba2cc434c" containerName="extract-utilities" Dec 13 08:10:44 crc kubenswrapper[4971]: I1213 08:10:44.397131 4971 state_mem.go:107] "Deleted CPUSet assignment" podUID="d4c962e9-e046-4e64-9843-4fbba2cc434c" containerName="extract-utilities" Dec 13 08:10:44 crc kubenswrapper[4971]: E1213 08:10:44.397157 4971 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="d4c962e9-e046-4e64-9843-4fbba2cc434c" containerName="registry-server" Dec 13 08:10:44 crc kubenswrapper[4971]: I1213 08:10:44.397170 4971 state_mem.go:107] "Deleted CPUSet assignment" podUID="d4c962e9-e046-4e64-9843-4fbba2cc434c" containerName="registry-server" Dec 13 08:10:44 crc kubenswrapper[4971]: I1213 08:10:44.397478 4971 memory_manager.go:354] "RemoveStaleState removing state" podUID="d4c962e9-e046-4e64-9843-4fbba2cc434c" containerName="registry-server" Dec 13 08:10:44 crc kubenswrapper[4971]: I1213 08:10:44.399395 4971 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-r64xp" Dec 13 08:10:44 crc kubenswrapper[4971]: I1213 08:10:44.410934 4971 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-r64xp"] Dec 13 08:10:44 crc kubenswrapper[4971]: I1213 08:10:44.548488 4971 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/9a41250a-e182-4912-a556-eb30e7834ff2-utilities\") pod \"redhat-operators-r64xp\" (UID: \"9a41250a-e182-4912-a556-eb30e7834ff2\") " pod="openshift-marketplace/redhat-operators-r64xp" Dec 13 08:10:44 crc kubenswrapper[4971]: I1213 08:10:44.548876 4971 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-kpm4w\" (UniqueName: \"kubernetes.io/projected/9a41250a-e182-4912-a556-eb30e7834ff2-kube-api-access-kpm4w\") pod \"redhat-operators-r64xp\" (UID: \"9a41250a-e182-4912-a556-eb30e7834ff2\") " pod="openshift-marketplace/redhat-operators-r64xp" Dec 13 08:10:44 crc kubenswrapper[4971]: I1213 08:10:44.548938 4971 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/9a41250a-e182-4912-a556-eb30e7834ff2-catalog-content\") pod \"redhat-operators-r64xp\" (UID: \"9a41250a-e182-4912-a556-eb30e7834ff2\") " pod="openshift-marketplace/redhat-operators-r64xp" Dec 13 08:10:44 crc kubenswrapper[4971]: I1213 08:10:44.650367 4971 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/9a41250a-e182-4912-a556-eb30e7834ff2-utilities\") pod \"redhat-operators-r64xp\" (UID: \"9a41250a-e182-4912-a556-eb30e7834ff2\") " pod="openshift-marketplace/redhat-operators-r64xp" Dec 13 08:10:44 crc kubenswrapper[4971]: I1213 08:10:44.650801 4971 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-kpm4w\" (UniqueName: \"kubernetes.io/projected/9a41250a-e182-4912-a556-eb30e7834ff2-kube-api-access-kpm4w\") pod \"redhat-operators-r64xp\" (UID: \"9a41250a-e182-4912-a556-eb30e7834ff2\") " pod="openshift-marketplace/redhat-operators-r64xp" Dec 13 08:10:44 crc kubenswrapper[4971]: I1213 08:10:44.650847 4971 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/9a41250a-e182-4912-a556-eb30e7834ff2-catalog-content\") pod \"redhat-operators-r64xp\" (UID: \"9a41250a-e182-4912-a556-eb30e7834ff2\") " pod="openshift-marketplace/redhat-operators-r64xp" Dec 13 08:10:44 crc kubenswrapper[4971]: I1213 08:10:44.651070 4971 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/9a41250a-e182-4912-a556-eb30e7834ff2-utilities\") pod \"redhat-operators-r64xp\" (UID: \"9a41250a-e182-4912-a556-eb30e7834ff2\") " pod="openshift-marketplace/redhat-operators-r64xp" Dec 13 08:10:44 crc kubenswrapper[4971]: I1213 08:10:44.651407 4971 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/9a41250a-e182-4912-a556-eb30e7834ff2-catalog-content\") pod \"redhat-operators-r64xp\" (UID: \"9a41250a-e182-4912-a556-eb30e7834ff2\") " pod="openshift-marketplace/redhat-operators-r64xp" Dec 13 08:10:44 crc kubenswrapper[4971]: I1213 08:10:44.671509 4971 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-kpm4w\" (UniqueName: \"kubernetes.io/projected/9a41250a-e182-4912-a556-eb30e7834ff2-kube-api-access-kpm4w\") pod \"redhat-operators-r64xp\" (UID: \"9a41250a-e182-4912-a556-eb30e7834ff2\") " pod="openshift-marketplace/redhat-operators-r64xp" Dec 13 08:10:44 crc kubenswrapper[4971]: I1213 08:10:44.722685 4971 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-r64xp" Dec 13 08:10:45 crc kubenswrapper[4971]: I1213 08:10:45.181984 4971 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-r64xp"] Dec 13 08:10:45 crc kubenswrapper[4971]: I1213 08:10:45.612781 4971 generic.go:334] "Generic (PLEG): container finished" podID="9a41250a-e182-4912-a556-eb30e7834ff2" containerID="63f28c2a680fa10ca9877c592bf5df95f01a806fb24f2191a87437fddc1edd53" exitCode=0 Dec 13 08:10:45 crc kubenswrapper[4971]: I1213 08:10:45.613049 4971 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-r64xp" event={"ID":"9a41250a-e182-4912-a556-eb30e7834ff2","Type":"ContainerDied","Data":"63f28c2a680fa10ca9877c592bf5df95f01a806fb24f2191a87437fddc1edd53"} Dec 13 08:10:45 crc kubenswrapper[4971]: I1213 08:10:45.613077 4971 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-r64xp" event={"ID":"9a41250a-e182-4912-a556-eb30e7834ff2","Type":"ContainerStarted","Data":"1465dd9e09198d515a81aeedc715be3ecc57f15661281d23c7efe0287f30c51c"} Dec 13 08:10:45 crc kubenswrapper[4971]: I1213 08:10:45.616759 4971 provider.go:102] Refreshing cache for provider: *credentialprovider.defaultDockerConfigProvider Dec 13 08:10:47 crc kubenswrapper[4971]: I1213 08:10:47.639897 4971 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-r64xp" event={"ID":"9a41250a-e182-4912-a556-eb30e7834ff2","Type":"ContainerStarted","Data":"8593772b1eda0c0f276b2fa0125bee804375b0829443663e2c2a1ba35520b0c7"} Dec 13 08:10:48 crc kubenswrapper[4971]: I1213 08:10:48.652220 4971 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-r64xp" event={"ID":"9a41250a-e182-4912-a556-eb30e7834ff2","Type":"ContainerDied","Data":"8593772b1eda0c0f276b2fa0125bee804375b0829443663e2c2a1ba35520b0c7"} Dec 13 08:10:48 crc kubenswrapper[4971]: I1213 08:10:48.652075 4971 generic.go:334] "Generic (PLEG): container finished" podID="9a41250a-e182-4912-a556-eb30e7834ff2" containerID="8593772b1eda0c0f276b2fa0125bee804375b0829443663e2c2a1ba35520b0c7" exitCode=0 Dec 13 08:10:49 crc kubenswrapper[4971]: I1213 08:10:49.666056 4971 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-r64xp" event={"ID":"9a41250a-e182-4912-a556-eb30e7834ff2","Type":"ContainerStarted","Data":"0b7583a44959c56719a826fbdb78e0e8cf69568f2577b246809ad750a44b80db"} Dec 13 08:10:49 crc kubenswrapper[4971]: I1213 08:10:49.690334 4971 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-operators-r64xp" podStartSLOduration=2.093927077 podStartE2EDuration="5.690310286s" podCreationTimestamp="2025-12-13 08:10:44 +0000 UTC" firstStartedPulling="2025-12-13 08:10:45.616404642 +0000 UTC m=+4902.220814100" lastFinishedPulling="2025-12-13 08:10:49.212787861 +0000 UTC m=+4905.817197309" observedRunningTime="2025-12-13 08:10:49.682985417 +0000 UTC m=+4906.287394865" watchObservedRunningTime="2025-12-13 08:10:49.690310286 +0000 UTC m=+4906.294719734" Dec 13 08:10:54 crc kubenswrapper[4971]: I1213 08:10:54.723554 4971 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-operators-r64xp" Dec 13 08:10:54 crc kubenswrapper[4971]: I1213 08:10:54.724344 4971 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-operators-r64xp" Dec 13 08:10:55 crc kubenswrapper[4971]: I1213 08:10:55.768186 4971 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-marketplace/redhat-operators-r64xp" podUID="9a41250a-e182-4912-a556-eb30e7834ff2" containerName="registry-server" probeResult="failure" output=< Dec 13 08:10:55 crc kubenswrapper[4971]: timeout: failed to connect service ":50051" within 1s Dec 13 08:10:55 crc kubenswrapper[4971]: > Dec 13 08:11:04 crc kubenswrapper[4971]: I1213 08:11:04.776324 4971 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-operators-r64xp" Dec 13 08:11:04 crc kubenswrapper[4971]: I1213 08:11:04.866743 4971 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-operators-r64xp" Dec 13 08:11:05 crc kubenswrapper[4971]: I1213 08:11:05.024719 4971 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-r64xp"] Dec 13 08:11:05 crc kubenswrapper[4971]: I1213 08:11:05.824749 4971 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-operators-r64xp" podUID="9a41250a-e182-4912-a556-eb30e7834ff2" containerName="registry-server" containerID="cri-o://0b7583a44959c56719a826fbdb78e0e8cf69568f2577b246809ad750a44b80db" gracePeriod=2 Dec 13 08:11:06 crc kubenswrapper[4971]: I1213 08:11:06.421844 4971 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-r64xp" Dec 13 08:11:06 crc kubenswrapper[4971]: I1213 08:11:06.495243 4971 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/9a41250a-e182-4912-a556-eb30e7834ff2-catalog-content\") pod \"9a41250a-e182-4912-a556-eb30e7834ff2\" (UID: \"9a41250a-e182-4912-a556-eb30e7834ff2\") " Dec 13 08:11:06 crc kubenswrapper[4971]: I1213 08:11:06.495440 4971 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-kpm4w\" (UniqueName: \"kubernetes.io/projected/9a41250a-e182-4912-a556-eb30e7834ff2-kube-api-access-kpm4w\") pod \"9a41250a-e182-4912-a556-eb30e7834ff2\" (UID: \"9a41250a-e182-4912-a556-eb30e7834ff2\") " Dec 13 08:11:06 crc kubenswrapper[4971]: I1213 08:11:06.495554 4971 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/9a41250a-e182-4912-a556-eb30e7834ff2-utilities\") pod \"9a41250a-e182-4912-a556-eb30e7834ff2\" (UID: \"9a41250a-e182-4912-a556-eb30e7834ff2\") " Dec 13 08:11:06 crc kubenswrapper[4971]: I1213 08:11:06.497262 4971 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/9a41250a-e182-4912-a556-eb30e7834ff2-utilities" (OuterVolumeSpecName: "utilities") pod "9a41250a-e182-4912-a556-eb30e7834ff2" (UID: "9a41250a-e182-4912-a556-eb30e7834ff2"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 13 08:11:06 crc kubenswrapper[4971]: I1213 08:11:06.503176 4971 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/9a41250a-e182-4912-a556-eb30e7834ff2-kube-api-access-kpm4w" (OuterVolumeSpecName: "kube-api-access-kpm4w") pod "9a41250a-e182-4912-a556-eb30e7834ff2" (UID: "9a41250a-e182-4912-a556-eb30e7834ff2"). InnerVolumeSpecName "kube-api-access-kpm4w". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 13 08:11:06 crc kubenswrapper[4971]: I1213 08:11:06.598842 4971 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/9a41250a-e182-4912-a556-eb30e7834ff2-utilities\") on node \"crc\" DevicePath \"\"" Dec 13 08:11:06 crc kubenswrapper[4971]: I1213 08:11:06.598909 4971 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-kpm4w\" (UniqueName: \"kubernetes.io/projected/9a41250a-e182-4912-a556-eb30e7834ff2-kube-api-access-kpm4w\") on node \"crc\" DevicePath \"\"" Dec 13 08:11:06 crc kubenswrapper[4971]: I1213 08:11:06.636740 4971 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/9a41250a-e182-4912-a556-eb30e7834ff2-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "9a41250a-e182-4912-a556-eb30e7834ff2" (UID: "9a41250a-e182-4912-a556-eb30e7834ff2"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 13 08:11:06 crc kubenswrapper[4971]: I1213 08:11:06.701192 4971 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/9a41250a-e182-4912-a556-eb30e7834ff2-catalog-content\") on node \"crc\" DevicePath \"\"" Dec 13 08:11:06 crc kubenswrapper[4971]: I1213 08:11:06.835038 4971 generic.go:334] "Generic (PLEG): container finished" podID="9a41250a-e182-4912-a556-eb30e7834ff2" containerID="0b7583a44959c56719a826fbdb78e0e8cf69568f2577b246809ad750a44b80db" exitCode=0 Dec 13 08:11:06 crc kubenswrapper[4971]: I1213 08:11:06.835154 4971 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-r64xp" Dec 13 08:11:06 crc kubenswrapper[4971]: I1213 08:11:06.835162 4971 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-r64xp" event={"ID":"9a41250a-e182-4912-a556-eb30e7834ff2","Type":"ContainerDied","Data":"0b7583a44959c56719a826fbdb78e0e8cf69568f2577b246809ad750a44b80db"} Dec 13 08:11:06 crc kubenswrapper[4971]: I1213 08:11:06.835578 4971 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-r64xp" event={"ID":"9a41250a-e182-4912-a556-eb30e7834ff2","Type":"ContainerDied","Data":"1465dd9e09198d515a81aeedc715be3ecc57f15661281d23c7efe0287f30c51c"} Dec 13 08:11:06 crc kubenswrapper[4971]: I1213 08:11:06.835605 4971 scope.go:117] "RemoveContainer" containerID="0b7583a44959c56719a826fbdb78e0e8cf69568f2577b246809ad750a44b80db" Dec 13 08:11:06 crc kubenswrapper[4971]: I1213 08:11:06.867299 4971 scope.go:117] "RemoveContainer" containerID="8593772b1eda0c0f276b2fa0125bee804375b0829443663e2c2a1ba35520b0c7" Dec 13 08:11:06 crc kubenswrapper[4971]: I1213 08:11:06.893469 4971 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-r64xp"] Dec 13 08:11:06 crc kubenswrapper[4971]: I1213 08:11:06.904385 4971 scope.go:117] "RemoveContainer" containerID="63f28c2a680fa10ca9877c592bf5df95f01a806fb24f2191a87437fddc1edd53" Dec 13 08:11:06 crc kubenswrapper[4971]: I1213 08:11:06.910551 4971 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-operators-r64xp"] Dec 13 08:11:06 crc kubenswrapper[4971]: I1213 08:11:06.956709 4971 scope.go:117] "RemoveContainer" containerID="0b7583a44959c56719a826fbdb78e0e8cf69568f2577b246809ad750a44b80db" Dec 13 08:11:06 crc kubenswrapper[4971]: E1213 08:11:06.957937 4971 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"0b7583a44959c56719a826fbdb78e0e8cf69568f2577b246809ad750a44b80db\": container with ID starting with 0b7583a44959c56719a826fbdb78e0e8cf69568f2577b246809ad750a44b80db not found: ID does not exist" containerID="0b7583a44959c56719a826fbdb78e0e8cf69568f2577b246809ad750a44b80db" Dec 13 08:11:06 crc kubenswrapper[4971]: I1213 08:11:06.957982 4971 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"0b7583a44959c56719a826fbdb78e0e8cf69568f2577b246809ad750a44b80db"} err="failed to get container status \"0b7583a44959c56719a826fbdb78e0e8cf69568f2577b246809ad750a44b80db\": rpc error: code = NotFound desc = could not find container \"0b7583a44959c56719a826fbdb78e0e8cf69568f2577b246809ad750a44b80db\": container with ID starting with 0b7583a44959c56719a826fbdb78e0e8cf69568f2577b246809ad750a44b80db not found: ID does not exist" Dec 13 08:11:06 crc kubenswrapper[4971]: I1213 08:11:06.958009 4971 scope.go:117] "RemoveContainer" containerID="8593772b1eda0c0f276b2fa0125bee804375b0829443663e2c2a1ba35520b0c7" Dec 13 08:11:06 crc kubenswrapper[4971]: E1213 08:11:06.958963 4971 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"8593772b1eda0c0f276b2fa0125bee804375b0829443663e2c2a1ba35520b0c7\": container with ID starting with 8593772b1eda0c0f276b2fa0125bee804375b0829443663e2c2a1ba35520b0c7 not found: ID does not exist" containerID="8593772b1eda0c0f276b2fa0125bee804375b0829443663e2c2a1ba35520b0c7" Dec 13 08:11:06 crc kubenswrapper[4971]: I1213 08:11:06.959042 4971 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"8593772b1eda0c0f276b2fa0125bee804375b0829443663e2c2a1ba35520b0c7"} err="failed to get container status \"8593772b1eda0c0f276b2fa0125bee804375b0829443663e2c2a1ba35520b0c7\": rpc error: code = NotFound desc = could not find container \"8593772b1eda0c0f276b2fa0125bee804375b0829443663e2c2a1ba35520b0c7\": container with ID starting with 8593772b1eda0c0f276b2fa0125bee804375b0829443663e2c2a1ba35520b0c7 not found: ID does not exist" Dec 13 08:11:06 crc kubenswrapper[4971]: I1213 08:11:06.959085 4971 scope.go:117] "RemoveContainer" containerID="63f28c2a680fa10ca9877c592bf5df95f01a806fb24f2191a87437fddc1edd53" Dec 13 08:11:06 crc kubenswrapper[4971]: E1213 08:11:06.959750 4971 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"63f28c2a680fa10ca9877c592bf5df95f01a806fb24f2191a87437fddc1edd53\": container with ID starting with 63f28c2a680fa10ca9877c592bf5df95f01a806fb24f2191a87437fddc1edd53 not found: ID does not exist" containerID="63f28c2a680fa10ca9877c592bf5df95f01a806fb24f2191a87437fddc1edd53" Dec 13 08:11:06 crc kubenswrapper[4971]: I1213 08:11:06.959789 4971 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"63f28c2a680fa10ca9877c592bf5df95f01a806fb24f2191a87437fddc1edd53"} err="failed to get container status \"63f28c2a680fa10ca9877c592bf5df95f01a806fb24f2191a87437fddc1edd53\": rpc error: code = NotFound desc = could not find container \"63f28c2a680fa10ca9877c592bf5df95f01a806fb24f2191a87437fddc1edd53\": container with ID starting with 63f28c2a680fa10ca9877c592bf5df95f01a806fb24f2191a87437fddc1edd53 not found: ID does not exist" Dec 13 08:11:07 crc kubenswrapper[4971]: I1213 08:11:07.787316 4971 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="9a41250a-e182-4912-a556-eb30e7834ff2" path="/var/lib/kubelet/pods/9a41250a-e182-4912-a556-eb30e7834ff2/volumes" Dec 13 08:12:18 crc kubenswrapper[4971]: I1213 08:12:18.433558 4971 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-marketplace-hd6kl"] Dec 13 08:12:18 crc kubenswrapper[4971]: E1213 08:12:18.435409 4971 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="9a41250a-e182-4912-a556-eb30e7834ff2" containerName="extract-utilities" Dec 13 08:12:18 crc kubenswrapper[4971]: I1213 08:12:18.435436 4971 state_mem.go:107] "Deleted CPUSet assignment" podUID="9a41250a-e182-4912-a556-eb30e7834ff2" containerName="extract-utilities" Dec 13 08:12:18 crc kubenswrapper[4971]: E1213 08:12:18.435465 4971 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="9a41250a-e182-4912-a556-eb30e7834ff2" containerName="extract-content" Dec 13 08:12:18 crc kubenswrapper[4971]: I1213 08:12:18.435471 4971 state_mem.go:107] "Deleted CPUSet assignment" podUID="9a41250a-e182-4912-a556-eb30e7834ff2" containerName="extract-content" Dec 13 08:12:18 crc kubenswrapper[4971]: E1213 08:12:18.435487 4971 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="9a41250a-e182-4912-a556-eb30e7834ff2" containerName="registry-server" Dec 13 08:12:18 crc kubenswrapper[4971]: I1213 08:12:18.435494 4971 state_mem.go:107] "Deleted CPUSet assignment" podUID="9a41250a-e182-4912-a556-eb30e7834ff2" containerName="registry-server" Dec 13 08:12:18 crc kubenswrapper[4971]: I1213 08:12:18.435806 4971 memory_manager.go:354] "RemoveStaleState removing state" podUID="9a41250a-e182-4912-a556-eb30e7834ff2" containerName="registry-server" Dec 13 08:12:18 crc kubenswrapper[4971]: I1213 08:12:18.437681 4971 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-hd6kl" Dec 13 08:12:18 crc kubenswrapper[4971]: I1213 08:12:18.452569 4971 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-hd6kl"] Dec 13 08:12:18 crc kubenswrapper[4971]: I1213 08:12:18.558839 4971 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/169ceaeb-da57-40cb-a236-614d1822d091-utilities\") pod \"redhat-marketplace-hd6kl\" (UID: \"169ceaeb-da57-40cb-a236-614d1822d091\") " pod="openshift-marketplace/redhat-marketplace-hd6kl" Dec 13 08:12:18 crc kubenswrapper[4971]: I1213 08:12:18.559013 4971 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-9m2td\" (UniqueName: \"kubernetes.io/projected/169ceaeb-da57-40cb-a236-614d1822d091-kube-api-access-9m2td\") pod \"redhat-marketplace-hd6kl\" (UID: \"169ceaeb-da57-40cb-a236-614d1822d091\") " pod="openshift-marketplace/redhat-marketplace-hd6kl" Dec 13 08:12:18 crc kubenswrapper[4971]: I1213 08:12:18.559131 4971 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/169ceaeb-da57-40cb-a236-614d1822d091-catalog-content\") pod \"redhat-marketplace-hd6kl\" (UID: \"169ceaeb-da57-40cb-a236-614d1822d091\") " pod="openshift-marketplace/redhat-marketplace-hd6kl" Dec 13 08:12:18 crc kubenswrapper[4971]: I1213 08:12:18.661666 4971 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/169ceaeb-da57-40cb-a236-614d1822d091-utilities\") pod \"redhat-marketplace-hd6kl\" (UID: \"169ceaeb-da57-40cb-a236-614d1822d091\") " pod="openshift-marketplace/redhat-marketplace-hd6kl" Dec 13 08:12:18 crc kubenswrapper[4971]: I1213 08:12:18.661837 4971 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-9m2td\" (UniqueName: \"kubernetes.io/projected/169ceaeb-da57-40cb-a236-614d1822d091-kube-api-access-9m2td\") pod \"redhat-marketplace-hd6kl\" (UID: \"169ceaeb-da57-40cb-a236-614d1822d091\") " pod="openshift-marketplace/redhat-marketplace-hd6kl" Dec 13 08:12:18 crc kubenswrapper[4971]: I1213 08:12:18.661937 4971 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/169ceaeb-da57-40cb-a236-614d1822d091-catalog-content\") pod \"redhat-marketplace-hd6kl\" (UID: \"169ceaeb-da57-40cb-a236-614d1822d091\") " pod="openshift-marketplace/redhat-marketplace-hd6kl" Dec 13 08:12:18 crc kubenswrapper[4971]: I1213 08:12:18.662996 4971 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/169ceaeb-da57-40cb-a236-614d1822d091-catalog-content\") pod \"redhat-marketplace-hd6kl\" (UID: \"169ceaeb-da57-40cb-a236-614d1822d091\") " pod="openshift-marketplace/redhat-marketplace-hd6kl" Dec 13 08:12:18 crc kubenswrapper[4971]: I1213 08:12:18.663021 4971 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/169ceaeb-da57-40cb-a236-614d1822d091-utilities\") pod \"redhat-marketplace-hd6kl\" (UID: \"169ceaeb-da57-40cb-a236-614d1822d091\") " pod="openshift-marketplace/redhat-marketplace-hd6kl" Dec 13 08:12:18 crc kubenswrapper[4971]: I1213 08:12:18.696372 4971 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-9m2td\" (UniqueName: \"kubernetes.io/projected/169ceaeb-da57-40cb-a236-614d1822d091-kube-api-access-9m2td\") pod \"redhat-marketplace-hd6kl\" (UID: \"169ceaeb-da57-40cb-a236-614d1822d091\") " pod="openshift-marketplace/redhat-marketplace-hd6kl" Dec 13 08:12:18 crc kubenswrapper[4971]: I1213 08:12:18.759360 4971 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-hd6kl" Dec 13 08:12:19 crc kubenswrapper[4971]: I1213 08:12:19.714281 4971 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-hd6kl"] Dec 13 08:12:20 crc kubenswrapper[4971]: I1213 08:12:20.369955 4971 generic.go:334] "Generic (PLEG): container finished" podID="169ceaeb-da57-40cb-a236-614d1822d091" containerID="1e38924608732da3c704758bfcbefdb9eec966b54e05d0857d0f2bb1d3f0f875" exitCode=0 Dec 13 08:12:20 crc kubenswrapper[4971]: I1213 08:12:20.370159 4971 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-hd6kl" event={"ID":"169ceaeb-da57-40cb-a236-614d1822d091","Type":"ContainerDied","Data":"1e38924608732da3c704758bfcbefdb9eec966b54e05d0857d0f2bb1d3f0f875"} Dec 13 08:12:20 crc kubenswrapper[4971]: I1213 08:12:20.370284 4971 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-hd6kl" event={"ID":"169ceaeb-da57-40cb-a236-614d1822d091","Type":"ContainerStarted","Data":"d30b65cfc8a5e6f4a8451749917b664787265cdb3073b7beb82e70ebe53dc63e"} Dec 13 08:12:20 crc kubenswrapper[4971]: I1213 08:12:20.622353 4971 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/community-operators-hxvbg"] Dec 13 08:12:20 crc kubenswrapper[4971]: I1213 08:12:20.625425 4971 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-hxvbg" Dec 13 08:12:20 crc kubenswrapper[4971]: I1213 08:12:20.640927 4971 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-hxvbg"] Dec 13 08:12:20 crc kubenswrapper[4971]: I1213 08:12:20.759403 4971 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/58991aaf-d599-4ab4-9393-5097bb3644a0-catalog-content\") pod \"community-operators-hxvbg\" (UID: \"58991aaf-d599-4ab4-9393-5097bb3644a0\") " pod="openshift-marketplace/community-operators-hxvbg" Dec 13 08:12:20 crc kubenswrapper[4971]: I1213 08:12:20.760917 4971 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-7tv85\" (UniqueName: \"kubernetes.io/projected/58991aaf-d599-4ab4-9393-5097bb3644a0-kube-api-access-7tv85\") pod \"community-operators-hxvbg\" (UID: \"58991aaf-d599-4ab4-9393-5097bb3644a0\") " pod="openshift-marketplace/community-operators-hxvbg" Dec 13 08:12:20 crc kubenswrapper[4971]: I1213 08:12:20.761274 4971 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/58991aaf-d599-4ab4-9393-5097bb3644a0-utilities\") pod \"community-operators-hxvbg\" (UID: \"58991aaf-d599-4ab4-9393-5097bb3644a0\") " pod="openshift-marketplace/community-operators-hxvbg" Dec 13 08:12:20 crc kubenswrapper[4971]: I1213 08:12:20.863042 4971 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/58991aaf-d599-4ab4-9393-5097bb3644a0-utilities\") pod \"community-operators-hxvbg\" (UID: \"58991aaf-d599-4ab4-9393-5097bb3644a0\") " pod="openshift-marketplace/community-operators-hxvbg" Dec 13 08:12:20 crc kubenswrapper[4971]: I1213 08:12:20.863641 4971 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/58991aaf-d599-4ab4-9393-5097bb3644a0-catalog-content\") pod \"community-operators-hxvbg\" (UID: \"58991aaf-d599-4ab4-9393-5097bb3644a0\") " pod="openshift-marketplace/community-operators-hxvbg" Dec 13 08:12:20 crc kubenswrapper[4971]: I1213 08:12:20.863704 4971 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-7tv85\" (UniqueName: \"kubernetes.io/projected/58991aaf-d599-4ab4-9393-5097bb3644a0-kube-api-access-7tv85\") pod \"community-operators-hxvbg\" (UID: \"58991aaf-d599-4ab4-9393-5097bb3644a0\") " pod="openshift-marketplace/community-operators-hxvbg" Dec 13 08:12:20 crc kubenswrapper[4971]: I1213 08:12:20.864264 4971 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/58991aaf-d599-4ab4-9393-5097bb3644a0-catalog-content\") pod \"community-operators-hxvbg\" (UID: \"58991aaf-d599-4ab4-9393-5097bb3644a0\") " pod="openshift-marketplace/community-operators-hxvbg" Dec 13 08:12:20 crc kubenswrapper[4971]: I1213 08:12:20.864538 4971 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/58991aaf-d599-4ab4-9393-5097bb3644a0-utilities\") pod \"community-operators-hxvbg\" (UID: \"58991aaf-d599-4ab4-9393-5097bb3644a0\") " pod="openshift-marketplace/community-operators-hxvbg" Dec 13 08:12:20 crc kubenswrapper[4971]: I1213 08:12:20.887153 4971 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-7tv85\" (UniqueName: \"kubernetes.io/projected/58991aaf-d599-4ab4-9393-5097bb3644a0-kube-api-access-7tv85\") pod \"community-operators-hxvbg\" (UID: \"58991aaf-d599-4ab4-9393-5097bb3644a0\") " pod="openshift-marketplace/community-operators-hxvbg" Dec 13 08:12:20 crc kubenswrapper[4971]: I1213 08:12:20.949184 4971 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-hxvbg" Dec 13 08:12:21 crc kubenswrapper[4971]: I1213 08:12:21.825257 4971 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-hxvbg"] Dec 13 08:12:21 crc kubenswrapper[4971]: W1213 08:12:21.831950 4971 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod58991aaf_d599_4ab4_9393_5097bb3644a0.slice/crio-da61acde57f1f1e7cb574fc591604682167858b5a06b9fd1ffe0e3c210c96a13 WatchSource:0}: Error finding container da61acde57f1f1e7cb574fc591604682167858b5a06b9fd1ffe0e3c210c96a13: Status 404 returned error can't find the container with id da61acde57f1f1e7cb574fc591604682167858b5a06b9fd1ffe0e3c210c96a13 Dec 13 08:12:22 crc kubenswrapper[4971]: I1213 08:12:22.660221 4971 generic.go:334] "Generic (PLEG): container finished" podID="169ceaeb-da57-40cb-a236-614d1822d091" containerID="614694710d451bd004c1d1f3705ce13ca7cee92d694e9bd172cf9060a4833a64" exitCode=0 Dec 13 08:12:22 crc kubenswrapper[4971]: I1213 08:12:22.661114 4971 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-hd6kl" event={"ID":"169ceaeb-da57-40cb-a236-614d1822d091","Type":"ContainerDied","Data":"614694710d451bd004c1d1f3705ce13ca7cee92d694e9bd172cf9060a4833a64"} Dec 13 08:12:22 crc kubenswrapper[4971]: I1213 08:12:22.666226 4971 generic.go:334] "Generic (PLEG): container finished" podID="58991aaf-d599-4ab4-9393-5097bb3644a0" containerID="d0ff62d04c1143fd1729898fd9c5184d4bebeceae28f3742c02e2896fbd4f036" exitCode=0 Dec 13 08:12:22 crc kubenswrapper[4971]: I1213 08:12:22.666296 4971 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-hxvbg" event={"ID":"58991aaf-d599-4ab4-9393-5097bb3644a0","Type":"ContainerDied","Data":"d0ff62d04c1143fd1729898fd9c5184d4bebeceae28f3742c02e2896fbd4f036"} Dec 13 08:12:22 crc kubenswrapper[4971]: I1213 08:12:22.666341 4971 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-hxvbg" event={"ID":"58991aaf-d599-4ab4-9393-5097bb3644a0","Type":"ContainerStarted","Data":"da61acde57f1f1e7cb574fc591604682167858b5a06b9fd1ffe0e3c210c96a13"} Dec 13 08:12:24 crc kubenswrapper[4971]: I1213 08:12:24.708144 4971 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-hd6kl" event={"ID":"169ceaeb-da57-40cb-a236-614d1822d091","Type":"ContainerStarted","Data":"30ec9d5261e21df87de9cab662675caa946e936f0a0584e3d87ba9c36dbc2099"} Dec 13 08:12:24 crc kubenswrapper[4971]: I1213 08:12:24.733123 4971 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-marketplace-hd6kl" podStartSLOduration=3.854333922 podStartE2EDuration="6.733098354s" podCreationTimestamp="2025-12-13 08:12:18 +0000 UTC" firstStartedPulling="2025-12-13 08:12:20.374265621 +0000 UTC m=+4996.978675069" lastFinishedPulling="2025-12-13 08:12:23.253030053 +0000 UTC m=+4999.857439501" observedRunningTime="2025-12-13 08:12:24.728747167 +0000 UTC m=+5001.333156625" watchObservedRunningTime="2025-12-13 08:12:24.733098354 +0000 UTC m=+5001.337507822" Dec 13 08:12:25 crc kubenswrapper[4971]: I1213 08:12:25.720681 4971 generic.go:334] "Generic (PLEG): container finished" podID="58991aaf-d599-4ab4-9393-5097bb3644a0" containerID="1709cd7bab150650b631896ec4eafcf066bdd5c12b9ace2abec45695dbdd055a" exitCode=0 Dec 13 08:12:25 crc kubenswrapper[4971]: I1213 08:12:25.720808 4971 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-hxvbg" event={"ID":"58991aaf-d599-4ab4-9393-5097bb3644a0","Type":"ContainerDied","Data":"1709cd7bab150650b631896ec4eafcf066bdd5c12b9ace2abec45695dbdd055a"} Dec 13 08:12:26 crc kubenswrapper[4971]: I1213 08:12:26.043841 4971 scope.go:117] "RemoveContainer" containerID="5b875af5f2a5d77169b3956d4995571e67a17dd9f171140d1a34c93221f648d9" Dec 13 08:12:26 crc kubenswrapper[4971]: I1213 08:12:26.070821 4971 scope.go:117] "RemoveContainer" containerID="fd65a132350d969bdb1e10fbc9c0ac567403a6d30357fc7e7ebd0ad047e7ffb0" Dec 13 08:12:26 crc kubenswrapper[4971]: I1213 08:12:26.098363 4971 scope.go:117] "RemoveContainer" containerID="18431a7fe7f4fa10984142ecadfbffaf7df42138b81e73ad53ab30f916f29f35" Dec 13 08:12:28 crc kubenswrapper[4971]: I1213 08:12:28.140358 4971 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-hxvbg" event={"ID":"58991aaf-d599-4ab4-9393-5097bb3644a0","Type":"ContainerStarted","Data":"bba26f62eb88c48690167b97e7c0598d94154e91295a2d4ca9ca546bd738104a"} Dec 13 08:12:28 crc kubenswrapper[4971]: I1213 08:12:28.169120 4971 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/community-operators-hxvbg" podStartSLOduration=4.384498492 podStartE2EDuration="8.169086817s" podCreationTimestamp="2025-12-13 08:12:20 +0000 UTC" firstStartedPulling="2025-12-13 08:12:22.668341671 +0000 UTC m=+4999.272751119" lastFinishedPulling="2025-12-13 08:12:26.452929996 +0000 UTC m=+5003.057339444" observedRunningTime="2025-12-13 08:12:28.165637123 +0000 UTC m=+5004.770046581" watchObservedRunningTime="2025-12-13 08:12:28.169086817 +0000 UTC m=+5004.773496265" Dec 13 08:12:28 crc kubenswrapper[4971]: I1213 08:12:28.868568 4971 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-marketplace-hd6kl" Dec 13 08:12:28 crc kubenswrapper[4971]: I1213 08:12:28.874916 4971 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-marketplace-hd6kl" Dec 13 08:12:29 crc kubenswrapper[4971]: I1213 08:12:29.015130 4971 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-marketplace-hd6kl" Dec 13 08:12:29 crc kubenswrapper[4971]: I1213 08:12:29.543375 4971 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-marketplace-hd6kl" Dec 13 08:12:30 crc kubenswrapper[4971]: I1213 08:12:30.809903 4971 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-hd6kl"] Dec 13 08:12:30 crc kubenswrapper[4971]: I1213 08:12:30.949995 4971 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/community-operators-hxvbg" Dec 13 08:12:30 crc kubenswrapper[4971]: I1213 08:12:30.951156 4971 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/community-operators-hxvbg" Dec 13 08:12:31 crc kubenswrapper[4971]: I1213 08:12:31.015022 4971 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/community-operators-hxvbg" Dec 13 08:12:32 crc kubenswrapper[4971]: I1213 08:12:32.392108 4971 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-marketplace-hd6kl" podUID="169ceaeb-da57-40cb-a236-614d1822d091" containerName="registry-server" containerID="cri-o://30ec9d5261e21df87de9cab662675caa946e936f0a0584e3d87ba9c36dbc2099" gracePeriod=2 Dec 13 08:12:32 crc kubenswrapper[4971]: I1213 08:12:32.459320 4971 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/community-operators-hxvbg" Dec 13 08:12:32 crc kubenswrapper[4971]: I1213 08:12:32.950771 4971 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-hd6kl" Dec 13 08:12:33 crc kubenswrapper[4971]: I1213 08:12:33.133143 4971 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/169ceaeb-da57-40cb-a236-614d1822d091-utilities\") pod \"169ceaeb-da57-40cb-a236-614d1822d091\" (UID: \"169ceaeb-da57-40cb-a236-614d1822d091\") " Dec 13 08:12:33 crc kubenswrapper[4971]: I1213 08:12:33.133208 4971 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-9m2td\" (UniqueName: \"kubernetes.io/projected/169ceaeb-da57-40cb-a236-614d1822d091-kube-api-access-9m2td\") pod \"169ceaeb-da57-40cb-a236-614d1822d091\" (UID: \"169ceaeb-da57-40cb-a236-614d1822d091\") " Dec 13 08:12:33 crc kubenswrapper[4971]: I1213 08:12:33.133418 4971 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/169ceaeb-da57-40cb-a236-614d1822d091-catalog-content\") pod \"169ceaeb-da57-40cb-a236-614d1822d091\" (UID: \"169ceaeb-da57-40cb-a236-614d1822d091\") " Dec 13 08:12:33 crc kubenswrapper[4971]: I1213 08:12:33.143637 4971 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/169ceaeb-da57-40cb-a236-614d1822d091-kube-api-access-9m2td" (OuterVolumeSpecName: "kube-api-access-9m2td") pod "169ceaeb-da57-40cb-a236-614d1822d091" (UID: "169ceaeb-da57-40cb-a236-614d1822d091"). InnerVolumeSpecName "kube-api-access-9m2td". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 13 08:12:33 crc kubenswrapper[4971]: I1213 08:12:33.151807 4971 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/169ceaeb-da57-40cb-a236-614d1822d091-utilities" (OuterVolumeSpecName: "utilities") pod "169ceaeb-da57-40cb-a236-614d1822d091" (UID: "169ceaeb-da57-40cb-a236-614d1822d091"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 13 08:12:33 crc kubenswrapper[4971]: I1213 08:12:33.159647 4971 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/169ceaeb-da57-40cb-a236-614d1822d091-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "169ceaeb-da57-40cb-a236-614d1822d091" (UID: "169ceaeb-da57-40cb-a236-614d1822d091"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 13 08:12:33 crc kubenswrapper[4971]: I1213 08:12:33.235875 4971 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/169ceaeb-da57-40cb-a236-614d1822d091-catalog-content\") on node \"crc\" DevicePath \"\"" Dec 13 08:12:33 crc kubenswrapper[4971]: I1213 08:12:33.236238 4971 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/169ceaeb-da57-40cb-a236-614d1822d091-utilities\") on node \"crc\" DevicePath \"\"" Dec 13 08:12:33 crc kubenswrapper[4971]: I1213 08:12:33.236254 4971 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-9m2td\" (UniqueName: \"kubernetes.io/projected/169ceaeb-da57-40cb-a236-614d1822d091-kube-api-access-9m2td\") on node \"crc\" DevicePath \"\"" Dec 13 08:12:33 crc kubenswrapper[4971]: I1213 08:12:33.406257 4971 generic.go:334] "Generic (PLEG): container finished" podID="169ceaeb-da57-40cb-a236-614d1822d091" containerID="30ec9d5261e21df87de9cab662675caa946e936f0a0584e3d87ba9c36dbc2099" exitCode=0 Dec 13 08:12:33 crc kubenswrapper[4971]: I1213 08:12:33.406326 4971 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-hd6kl" event={"ID":"169ceaeb-da57-40cb-a236-614d1822d091","Type":"ContainerDied","Data":"30ec9d5261e21df87de9cab662675caa946e936f0a0584e3d87ba9c36dbc2099"} Dec 13 08:12:33 crc kubenswrapper[4971]: I1213 08:12:33.406389 4971 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-hd6kl" Dec 13 08:12:33 crc kubenswrapper[4971]: I1213 08:12:33.406416 4971 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-hd6kl" event={"ID":"169ceaeb-da57-40cb-a236-614d1822d091","Type":"ContainerDied","Data":"d30b65cfc8a5e6f4a8451749917b664787265cdb3073b7beb82e70ebe53dc63e"} Dec 13 08:12:33 crc kubenswrapper[4971]: I1213 08:12:33.406442 4971 scope.go:117] "RemoveContainer" containerID="30ec9d5261e21df87de9cab662675caa946e936f0a0584e3d87ba9c36dbc2099" Dec 13 08:12:33 crc kubenswrapper[4971]: I1213 08:12:33.427004 4971 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-hxvbg"] Dec 13 08:12:33 crc kubenswrapper[4971]: I1213 08:12:33.436315 4971 scope.go:117] "RemoveContainer" containerID="614694710d451bd004c1d1f3705ce13ca7cee92d694e9bd172cf9060a4833a64" Dec 13 08:12:33 crc kubenswrapper[4971]: I1213 08:12:33.515625 4971 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-hd6kl"] Dec 13 08:12:33 crc kubenswrapper[4971]: I1213 08:12:33.516474 4971 scope.go:117] "RemoveContainer" containerID="1e38924608732da3c704758bfcbefdb9eec966b54e05d0857d0f2bb1d3f0f875" Dec 13 08:12:33 crc kubenswrapper[4971]: I1213 08:12:33.557281 4971 scope.go:117] "RemoveContainer" containerID="30ec9d5261e21df87de9cab662675caa946e936f0a0584e3d87ba9c36dbc2099" Dec 13 08:12:33 crc kubenswrapper[4971]: I1213 08:12:33.558236 4971 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-marketplace-hd6kl"] Dec 13 08:12:33 crc kubenswrapper[4971]: E1213 08:12:33.559365 4971 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"30ec9d5261e21df87de9cab662675caa946e936f0a0584e3d87ba9c36dbc2099\": container with ID starting with 30ec9d5261e21df87de9cab662675caa946e936f0a0584e3d87ba9c36dbc2099 not found: ID does not exist" containerID="30ec9d5261e21df87de9cab662675caa946e936f0a0584e3d87ba9c36dbc2099" Dec 13 08:12:33 crc kubenswrapper[4971]: I1213 08:12:33.559424 4971 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"30ec9d5261e21df87de9cab662675caa946e936f0a0584e3d87ba9c36dbc2099"} err="failed to get container status \"30ec9d5261e21df87de9cab662675caa946e936f0a0584e3d87ba9c36dbc2099\": rpc error: code = NotFound desc = could not find container \"30ec9d5261e21df87de9cab662675caa946e936f0a0584e3d87ba9c36dbc2099\": container with ID starting with 30ec9d5261e21df87de9cab662675caa946e936f0a0584e3d87ba9c36dbc2099 not found: ID does not exist" Dec 13 08:12:33 crc kubenswrapper[4971]: I1213 08:12:33.559457 4971 scope.go:117] "RemoveContainer" containerID="614694710d451bd004c1d1f3705ce13ca7cee92d694e9bd172cf9060a4833a64" Dec 13 08:12:33 crc kubenswrapper[4971]: E1213 08:12:33.560070 4971 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"614694710d451bd004c1d1f3705ce13ca7cee92d694e9bd172cf9060a4833a64\": container with ID starting with 614694710d451bd004c1d1f3705ce13ca7cee92d694e9bd172cf9060a4833a64 not found: ID does not exist" containerID="614694710d451bd004c1d1f3705ce13ca7cee92d694e9bd172cf9060a4833a64" Dec 13 08:12:33 crc kubenswrapper[4971]: I1213 08:12:33.560127 4971 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"614694710d451bd004c1d1f3705ce13ca7cee92d694e9bd172cf9060a4833a64"} err="failed to get container status \"614694710d451bd004c1d1f3705ce13ca7cee92d694e9bd172cf9060a4833a64\": rpc error: code = NotFound desc = could not find container \"614694710d451bd004c1d1f3705ce13ca7cee92d694e9bd172cf9060a4833a64\": container with ID starting with 614694710d451bd004c1d1f3705ce13ca7cee92d694e9bd172cf9060a4833a64 not found: ID does not exist" Dec 13 08:12:33 crc kubenswrapper[4971]: I1213 08:12:33.560161 4971 scope.go:117] "RemoveContainer" containerID="1e38924608732da3c704758bfcbefdb9eec966b54e05d0857d0f2bb1d3f0f875" Dec 13 08:12:33 crc kubenswrapper[4971]: E1213 08:12:33.560534 4971 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"1e38924608732da3c704758bfcbefdb9eec966b54e05d0857d0f2bb1d3f0f875\": container with ID starting with 1e38924608732da3c704758bfcbefdb9eec966b54e05d0857d0f2bb1d3f0f875 not found: ID does not exist" containerID="1e38924608732da3c704758bfcbefdb9eec966b54e05d0857d0f2bb1d3f0f875" Dec 13 08:12:33 crc kubenswrapper[4971]: I1213 08:12:33.560563 4971 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"1e38924608732da3c704758bfcbefdb9eec966b54e05d0857d0f2bb1d3f0f875"} err="failed to get container status \"1e38924608732da3c704758bfcbefdb9eec966b54e05d0857d0f2bb1d3f0f875\": rpc error: code = NotFound desc = could not find container \"1e38924608732da3c704758bfcbefdb9eec966b54e05d0857d0f2bb1d3f0f875\": container with ID starting with 1e38924608732da3c704758bfcbefdb9eec966b54e05d0857d0f2bb1d3f0f875 not found: ID does not exist" Dec 13 08:12:33 crc kubenswrapper[4971]: I1213 08:12:33.785467 4971 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="169ceaeb-da57-40cb-a236-614d1822d091" path="/var/lib/kubelet/pods/169ceaeb-da57-40cb-a236-614d1822d091/volumes" Dec 13 08:12:34 crc kubenswrapper[4971]: I1213 08:12:34.423964 4971 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/community-operators-hxvbg" podUID="58991aaf-d599-4ab4-9393-5097bb3644a0" containerName="registry-server" containerID="cri-o://bba26f62eb88c48690167b97e7c0598d94154e91295a2d4ca9ca546bd738104a" gracePeriod=2 Dec 13 08:12:35 crc kubenswrapper[4971]: I1213 08:12:35.174410 4971 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-hxvbg" Dec 13 08:12:35 crc kubenswrapper[4971]: I1213 08:12:35.297657 4971 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/58991aaf-d599-4ab4-9393-5097bb3644a0-catalog-content\") pod \"58991aaf-d599-4ab4-9393-5097bb3644a0\" (UID: \"58991aaf-d599-4ab4-9393-5097bb3644a0\") " Dec 13 08:12:35 crc kubenswrapper[4971]: I1213 08:12:35.297775 4971 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-7tv85\" (UniqueName: \"kubernetes.io/projected/58991aaf-d599-4ab4-9393-5097bb3644a0-kube-api-access-7tv85\") pod \"58991aaf-d599-4ab4-9393-5097bb3644a0\" (UID: \"58991aaf-d599-4ab4-9393-5097bb3644a0\") " Dec 13 08:12:35 crc kubenswrapper[4971]: I1213 08:12:35.297939 4971 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/58991aaf-d599-4ab4-9393-5097bb3644a0-utilities\") pod \"58991aaf-d599-4ab4-9393-5097bb3644a0\" (UID: \"58991aaf-d599-4ab4-9393-5097bb3644a0\") " Dec 13 08:12:35 crc kubenswrapper[4971]: I1213 08:12:35.300469 4971 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/58991aaf-d599-4ab4-9393-5097bb3644a0-utilities" (OuterVolumeSpecName: "utilities") pod "58991aaf-d599-4ab4-9393-5097bb3644a0" (UID: "58991aaf-d599-4ab4-9393-5097bb3644a0"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 13 08:12:35 crc kubenswrapper[4971]: I1213 08:12:35.310542 4971 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/58991aaf-d599-4ab4-9393-5097bb3644a0-kube-api-access-7tv85" (OuterVolumeSpecName: "kube-api-access-7tv85") pod "58991aaf-d599-4ab4-9393-5097bb3644a0" (UID: "58991aaf-d599-4ab4-9393-5097bb3644a0"). InnerVolumeSpecName "kube-api-access-7tv85". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 13 08:12:35 crc kubenswrapper[4971]: I1213 08:12:35.381859 4971 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/58991aaf-d599-4ab4-9393-5097bb3644a0-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "58991aaf-d599-4ab4-9393-5097bb3644a0" (UID: "58991aaf-d599-4ab4-9393-5097bb3644a0"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 13 08:12:35 crc kubenswrapper[4971]: I1213 08:12:35.401323 4971 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/58991aaf-d599-4ab4-9393-5097bb3644a0-utilities\") on node \"crc\" DevicePath \"\"" Dec 13 08:12:35 crc kubenswrapper[4971]: I1213 08:12:35.401386 4971 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/58991aaf-d599-4ab4-9393-5097bb3644a0-catalog-content\") on node \"crc\" DevicePath \"\"" Dec 13 08:12:35 crc kubenswrapper[4971]: I1213 08:12:35.401404 4971 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-7tv85\" (UniqueName: \"kubernetes.io/projected/58991aaf-d599-4ab4-9393-5097bb3644a0-kube-api-access-7tv85\") on node \"crc\" DevicePath \"\"" Dec 13 08:12:35 crc kubenswrapper[4971]: I1213 08:12:35.439987 4971 generic.go:334] "Generic (PLEG): container finished" podID="58991aaf-d599-4ab4-9393-5097bb3644a0" containerID="bba26f62eb88c48690167b97e7c0598d94154e91295a2d4ca9ca546bd738104a" exitCode=0 Dec 13 08:12:35 crc kubenswrapper[4971]: I1213 08:12:35.440050 4971 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-hxvbg" event={"ID":"58991aaf-d599-4ab4-9393-5097bb3644a0","Type":"ContainerDied","Data":"bba26f62eb88c48690167b97e7c0598d94154e91295a2d4ca9ca546bd738104a"} Dec 13 08:12:35 crc kubenswrapper[4971]: I1213 08:12:35.440092 4971 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-hxvbg" event={"ID":"58991aaf-d599-4ab4-9393-5097bb3644a0","Type":"ContainerDied","Data":"da61acde57f1f1e7cb574fc591604682167858b5a06b9fd1ffe0e3c210c96a13"} Dec 13 08:12:35 crc kubenswrapper[4971]: I1213 08:12:35.440117 4971 scope.go:117] "RemoveContainer" containerID="bba26f62eb88c48690167b97e7c0598d94154e91295a2d4ca9ca546bd738104a" Dec 13 08:12:35 crc kubenswrapper[4971]: I1213 08:12:35.440358 4971 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-hxvbg" Dec 13 08:12:35 crc kubenswrapper[4971]: I1213 08:12:35.500957 4971 scope.go:117] "RemoveContainer" containerID="1709cd7bab150650b631896ec4eafcf066bdd5c12b9ace2abec45695dbdd055a" Dec 13 08:12:35 crc kubenswrapper[4971]: I1213 08:12:35.503750 4971 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-hxvbg"] Dec 13 08:12:35 crc kubenswrapper[4971]: I1213 08:12:35.519658 4971 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/community-operators-hxvbg"] Dec 13 08:12:35 crc kubenswrapper[4971]: I1213 08:12:35.531164 4971 scope.go:117] "RemoveContainer" containerID="d0ff62d04c1143fd1729898fd9c5184d4bebeceae28f3742c02e2896fbd4f036" Dec 13 08:12:35 crc kubenswrapper[4971]: I1213 08:12:35.577654 4971 scope.go:117] "RemoveContainer" containerID="bba26f62eb88c48690167b97e7c0598d94154e91295a2d4ca9ca546bd738104a" Dec 13 08:12:35 crc kubenswrapper[4971]: E1213 08:12:35.578198 4971 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"bba26f62eb88c48690167b97e7c0598d94154e91295a2d4ca9ca546bd738104a\": container with ID starting with bba26f62eb88c48690167b97e7c0598d94154e91295a2d4ca9ca546bd738104a not found: ID does not exist" containerID="bba26f62eb88c48690167b97e7c0598d94154e91295a2d4ca9ca546bd738104a" Dec 13 08:12:35 crc kubenswrapper[4971]: I1213 08:12:35.578256 4971 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"bba26f62eb88c48690167b97e7c0598d94154e91295a2d4ca9ca546bd738104a"} err="failed to get container status \"bba26f62eb88c48690167b97e7c0598d94154e91295a2d4ca9ca546bd738104a\": rpc error: code = NotFound desc = could not find container \"bba26f62eb88c48690167b97e7c0598d94154e91295a2d4ca9ca546bd738104a\": container with ID starting with bba26f62eb88c48690167b97e7c0598d94154e91295a2d4ca9ca546bd738104a not found: ID does not exist" Dec 13 08:12:35 crc kubenswrapper[4971]: I1213 08:12:35.578295 4971 scope.go:117] "RemoveContainer" containerID="1709cd7bab150650b631896ec4eafcf066bdd5c12b9ace2abec45695dbdd055a" Dec 13 08:12:35 crc kubenswrapper[4971]: E1213 08:12:35.578726 4971 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"1709cd7bab150650b631896ec4eafcf066bdd5c12b9ace2abec45695dbdd055a\": container with ID starting with 1709cd7bab150650b631896ec4eafcf066bdd5c12b9ace2abec45695dbdd055a not found: ID does not exist" containerID="1709cd7bab150650b631896ec4eafcf066bdd5c12b9ace2abec45695dbdd055a" Dec 13 08:12:35 crc kubenswrapper[4971]: I1213 08:12:35.578757 4971 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"1709cd7bab150650b631896ec4eafcf066bdd5c12b9ace2abec45695dbdd055a"} err="failed to get container status \"1709cd7bab150650b631896ec4eafcf066bdd5c12b9ace2abec45695dbdd055a\": rpc error: code = NotFound desc = could not find container \"1709cd7bab150650b631896ec4eafcf066bdd5c12b9ace2abec45695dbdd055a\": container with ID starting with 1709cd7bab150650b631896ec4eafcf066bdd5c12b9ace2abec45695dbdd055a not found: ID does not exist" Dec 13 08:12:35 crc kubenswrapper[4971]: I1213 08:12:35.578778 4971 scope.go:117] "RemoveContainer" containerID="d0ff62d04c1143fd1729898fd9c5184d4bebeceae28f3742c02e2896fbd4f036" Dec 13 08:12:35 crc kubenswrapper[4971]: E1213 08:12:35.580347 4971 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"d0ff62d04c1143fd1729898fd9c5184d4bebeceae28f3742c02e2896fbd4f036\": container with ID starting with d0ff62d04c1143fd1729898fd9c5184d4bebeceae28f3742c02e2896fbd4f036 not found: ID does not exist" containerID="d0ff62d04c1143fd1729898fd9c5184d4bebeceae28f3742c02e2896fbd4f036" Dec 13 08:12:35 crc kubenswrapper[4971]: I1213 08:12:35.580383 4971 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"d0ff62d04c1143fd1729898fd9c5184d4bebeceae28f3742c02e2896fbd4f036"} err="failed to get container status \"d0ff62d04c1143fd1729898fd9c5184d4bebeceae28f3742c02e2896fbd4f036\": rpc error: code = NotFound desc = could not find container \"d0ff62d04c1143fd1729898fd9c5184d4bebeceae28f3742c02e2896fbd4f036\": container with ID starting with d0ff62d04c1143fd1729898fd9c5184d4bebeceae28f3742c02e2896fbd4f036 not found: ID does not exist" Dec 13 08:12:35 crc kubenswrapper[4971]: I1213 08:12:35.784602 4971 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="58991aaf-d599-4ab4-9393-5097bb3644a0" path="/var/lib/kubelet/pods/58991aaf-d599-4ab4-9393-5097bb3644a0/volumes" Dec 13 08:12:46 crc kubenswrapper[4971]: I1213 08:12:46.153361 4971 patch_prober.go:28] interesting pod/machine-config-daemon-82xjz container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 13 08:12:46 crc kubenswrapper[4971]: I1213 08:12:46.154368 4971 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-82xjz" podUID="e0c4af12-edda-41d7-9b44-f87396174ef2" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 13 08:13:16 crc kubenswrapper[4971]: I1213 08:13:16.153509 4971 patch_prober.go:28] interesting pod/machine-config-daemon-82xjz container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 13 08:13:16 crc kubenswrapper[4971]: I1213 08:13:16.155250 4971 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-82xjz" podUID="e0c4af12-edda-41d7-9b44-f87396174ef2" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 13 08:13:46 crc kubenswrapper[4971]: I1213 08:13:46.154065 4971 patch_prober.go:28] interesting pod/machine-config-daemon-82xjz container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 13 08:13:46 crc kubenswrapper[4971]: I1213 08:13:46.155137 4971 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-82xjz" podUID="e0c4af12-edda-41d7-9b44-f87396174ef2" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 13 08:13:46 crc kubenswrapper[4971]: I1213 08:13:46.155228 4971 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-82xjz" Dec 13 08:13:46 crc kubenswrapper[4971]: I1213 08:13:46.156669 4971 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"8b2617257e0c4b9039d9f1e54400a1b30b3fb18612ede5bd959442a8eca72ecd"} pod="openshift-machine-config-operator/machine-config-daemon-82xjz" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Dec 13 08:13:46 crc kubenswrapper[4971]: I1213 08:13:46.156739 4971 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-82xjz" podUID="e0c4af12-edda-41d7-9b44-f87396174ef2" containerName="machine-config-daemon" containerID="cri-o://8b2617257e0c4b9039d9f1e54400a1b30b3fb18612ede5bd959442a8eca72ecd" gracePeriod=600 Dec 13 08:13:46 crc kubenswrapper[4971]: E1213 08:13:46.316589 4971 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-82xjz_openshift-machine-config-operator(e0c4af12-edda-41d7-9b44-f87396174ef2)\"" pod="openshift-machine-config-operator/machine-config-daemon-82xjz" podUID="e0c4af12-edda-41d7-9b44-f87396174ef2" Dec 13 08:13:46 crc kubenswrapper[4971]: I1213 08:13:46.379333 4971 generic.go:334] "Generic (PLEG): container finished" podID="e0c4af12-edda-41d7-9b44-f87396174ef2" containerID="8b2617257e0c4b9039d9f1e54400a1b30b3fb18612ede5bd959442a8eca72ecd" exitCode=0 Dec 13 08:13:46 crc kubenswrapper[4971]: I1213 08:13:46.379404 4971 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-82xjz" event={"ID":"e0c4af12-edda-41d7-9b44-f87396174ef2","Type":"ContainerDied","Data":"8b2617257e0c4b9039d9f1e54400a1b30b3fb18612ede5bd959442a8eca72ecd"} Dec 13 08:13:46 crc kubenswrapper[4971]: I1213 08:13:46.379460 4971 scope.go:117] "RemoveContainer" containerID="fd226c5cf0ac150e1a98fe2ca7739d374a376da0b0db1d061f54febf1a32844a" Dec 13 08:13:46 crc kubenswrapper[4971]: I1213 08:13:46.380807 4971 scope.go:117] "RemoveContainer" containerID="8b2617257e0c4b9039d9f1e54400a1b30b3fb18612ede5bd959442a8eca72ecd" Dec 13 08:13:46 crc kubenswrapper[4971]: E1213 08:13:46.381147 4971 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-82xjz_openshift-machine-config-operator(e0c4af12-edda-41d7-9b44-f87396174ef2)\"" pod="openshift-machine-config-operator/machine-config-daemon-82xjz" podUID="e0c4af12-edda-41d7-9b44-f87396174ef2" Dec 13 08:14:00 crc kubenswrapper[4971]: I1213 08:14:00.769394 4971 scope.go:117] "RemoveContainer" containerID="8b2617257e0c4b9039d9f1e54400a1b30b3fb18612ede5bd959442a8eca72ecd" Dec 13 08:14:00 crc kubenswrapper[4971]: E1213 08:14:00.770510 4971 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-82xjz_openshift-machine-config-operator(e0c4af12-edda-41d7-9b44-f87396174ef2)\"" pod="openshift-machine-config-operator/machine-config-daemon-82xjz" podUID="e0c4af12-edda-41d7-9b44-f87396174ef2" Dec 13 08:14:15 crc kubenswrapper[4971]: I1213 08:14:14.865417 4971 scope.go:117] "RemoveContainer" containerID="8b2617257e0c4b9039d9f1e54400a1b30b3fb18612ede5bd959442a8eca72ecd" Dec 13 08:14:15 crc kubenswrapper[4971]: E1213 08:14:14.866910 4971 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-82xjz_openshift-machine-config-operator(e0c4af12-edda-41d7-9b44-f87396174ef2)\"" pod="openshift-machine-config-operator/machine-config-daemon-82xjz" podUID="e0c4af12-edda-41d7-9b44-f87396174ef2" Dec 13 08:14:27 crc kubenswrapper[4971]: I1213 08:14:27.769172 4971 scope.go:117] "RemoveContainer" containerID="8b2617257e0c4b9039d9f1e54400a1b30b3fb18612ede5bd959442a8eca72ecd" Dec 13 08:14:27 crc kubenswrapper[4971]: E1213 08:14:27.773835 4971 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-82xjz_openshift-machine-config-operator(e0c4af12-edda-41d7-9b44-f87396174ef2)\"" pod="openshift-machine-config-operator/machine-config-daemon-82xjz" podUID="e0c4af12-edda-41d7-9b44-f87396174ef2" Dec 13 08:14:42 crc kubenswrapper[4971]: I1213 08:14:42.769489 4971 scope.go:117] "RemoveContainer" containerID="8b2617257e0c4b9039d9f1e54400a1b30b3fb18612ede5bd959442a8eca72ecd" Dec 13 08:14:42 crc kubenswrapper[4971]: E1213 08:14:42.771015 4971 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-82xjz_openshift-machine-config-operator(e0c4af12-edda-41d7-9b44-f87396174ef2)\"" pod="openshift-machine-config-operator/machine-config-daemon-82xjz" podUID="e0c4af12-edda-41d7-9b44-f87396174ef2" Dec 13 08:14:55 crc kubenswrapper[4971]: I1213 08:14:55.769600 4971 scope.go:117] "RemoveContainer" containerID="8b2617257e0c4b9039d9f1e54400a1b30b3fb18612ede5bd959442a8eca72ecd" Dec 13 08:14:55 crc kubenswrapper[4971]: E1213 08:14:55.771217 4971 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-82xjz_openshift-machine-config-operator(e0c4af12-edda-41d7-9b44-f87396174ef2)\"" pod="openshift-machine-config-operator/machine-config-daemon-82xjz" podUID="e0c4af12-edda-41d7-9b44-f87396174ef2" Dec 13 08:15:00 crc kubenswrapper[4971]: I1213 08:15:00.163917 4971 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29426895-2sc8g"] Dec 13 08:15:00 crc kubenswrapper[4971]: E1213 08:15:00.165910 4971 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="169ceaeb-da57-40cb-a236-614d1822d091" containerName="extract-content" Dec 13 08:15:00 crc kubenswrapper[4971]: I1213 08:15:00.165996 4971 state_mem.go:107] "Deleted CPUSet assignment" podUID="169ceaeb-da57-40cb-a236-614d1822d091" containerName="extract-content" Dec 13 08:15:00 crc kubenswrapper[4971]: E1213 08:15:00.166061 4971 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="169ceaeb-da57-40cb-a236-614d1822d091" containerName="registry-server" Dec 13 08:15:00 crc kubenswrapper[4971]: I1213 08:15:00.166116 4971 state_mem.go:107] "Deleted CPUSet assignment" podUID="169ceaeb-da57-40cb-a236-614d1822d091" containerName="registry-server" Dec 13 08:15:00 crc kubenswrapper[4971]: E1213 08:15:00.166186 4971 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="58991aaf-d599-4ab4-9393-5097bb3644a0" containerName="registry-server" Dec 13 08:15:00 crc kubenswrapper[4971]: I1213 08:15:00.166248 4971 state_mem.go:107] "Deleted CPUSet assignment" podUID="58991aaf-d599-4ab4-9393-5097bb3644a0" containerName="registry-server" Dec 13 08:15:00 crc kubenswrapper[4971]: E1213 08:15:00.166309 4971 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="169ceaeb-da57-40cb-a236-614d1822d091" containerName="extract-utilities" Dec 13 08:15:00 crc kubenswrapper[4971]: I1213 08:15:00.166381 4971 state_mem.go:107] "Deleted CPUSet assignment" podUID="169ceaeb-da57-40cb-a236-614d1822d091" containerName="extract-utilities" Dec 13 08:15:00 crc kubenswrapper[4971]: E1213 08:15:00.166445 4971 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="58991aaf-d599-4ab4-9393-5097bb3644a0" containerName="extract-utilities" Dec 13 08:15:00 crc kubenswrapper[4971]: I1213 08:15:00.166502 4971 state_mem.go:107] "Deleted CPUSet assignment" podUID="58991aaf-d599-4ab4-9393-5097bb3644a0" containerName="extract-utilities" Dec 13 08:15:00 crc kubenswrapper[4971]: E1213 08:15:00.166587 4971 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="58991aaf-d599-4ab4-9393-5097bb3644a0" containerName="extract-content" Dec 13 08:15:00 crc kubenswrapper[4971]: I1213 08:15:00.166645 4971 state_mem.go:107] "Deleted CPUSet assignment" podUID="58991aaf-d599-4ab4-9393-5097bb3644a0" containerName="extract-content" Dec 13 08:15:00 crc kubenswrapper[4971]: I1213 08:15:00.166897 4971 memory_manager.go:354] "RemoveStaleState removing state" podUID="58991aaf-d599-4ab4-9393-5097bb3644a0" containerName="registry-server" Dec 13 08:15:00 crc kubenswrapper[4971]: I1213 08:15:00.166967 4971 memory_manager.go:354] "RemoveStaleState removing state" podUID="169ceaeb-da57-40cb-a236-614d1822d091" containerName="registry-server" Dec 13 08:15:00 crc kubenswrapper[4971]: I1213 08:15:00.167730 4971 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29426895-2sc8g" Dec 13 08:15:00 crc kubenswrapper[4971]: I1213 08:15:00.174154 4971 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"collect-profiles-dockercfg-kzf4t" Dec 13 08:15:00 crc kubenswrapper[4971]: I1213 08:15:00.174369 4971 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-operator-lifecycle-manager"/"collect-profiles-config" Dec 13 08:15:00 crc kubenswrapper[4971]: I1213 08:15:00.180473 4971 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29426895-2sc8g"] Dec 13 08:15:00 crc kubenswrapper[4971]: I1213 08:15:00.212786 4971 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-9qhhl\" (UniqueName: \"kubernetes.io/projected/ab8b4298-8945-4f7f-aa70-1d08d1a7b91c-kube-api-access-9qhhl\") pod \"collect-profiles-29426895-2sc8g\" (UID: \"ab8b4298-8945-4f7f-aa70-1d08d1a7b91c\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29426895-2sc8g" Dec 13 08:15:00 crc kubenswrapper[4971]: I1213 08:15:00.213251 4971 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/ab8b4298-8945-4f7f-aa70-1d08d1a7b91c-secret-volume\") pod \"collect-profiles-29426895-2sc8g\" (UID: \"ab8b4298-8945-4f7f-aa70-1d08d1a7b91c\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29426895-2sc8g" Dec 13 08:15:00 crc kubenswrapper[4971]: I1213 08:15:00.213601 4971 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/ab8b4298-8945-4f7f-aa70-1d08d1a7b91c-config-volume\") pod \"collect-profiles-29426895-2sc8g\" (UID: \"ab8b4298-8945-4f7f-aa70-1d08d1a7b91c\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29426895-2sc8g" Dec 13 08:15:00 crc kubenswrapper[4971]: I1213 08:15:00.316415 4971 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/ab8b4298-8945-4f7f-aa70-1d08d1a7b91c-secret-volume\") pod \"collect-profiles-29426895-2sc8g\" (UID: \"ab8b4298-8945-4f7f-aa70-1d08d1a7b91c\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29426895-2sc8g" Dec 13 08:15:00 crc kubenswrapper[4971]: I1213 08:15:00.316500 4971 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/ab8b4298-8945-4f7f-aa70-1d08d1a7b91c-config-volume\") pod \"collect-profiles-29426895-2sc8g\" (UID: \"ab8b4298-8945-4f7f-aa70-1d08d1a7b91c\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29426895-2sc8g" Dec 13 08:15:00 crc kubenswrapper[4971]: I1213 08:15:00.316785 4971 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-9qhhl\" (UniqueName: \"kubernetes.io/projected/ab8b4298-8945-4f7f-aa70-1d08d1a7b91c-kube-api-access-9qhhl\") pod \"collect-profiles-29426895-2sc8g\" (UID: \"ab8b4298-8945-4f7f-aa70-1d08d1a7b91c\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29426895-2sc8g" Dec 13 08:15:00 crc kubenswrapper[4971]: I1213 08:15:00.319107 4971 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/ab8b4298-8945-4f7f-aa70-1d08d1a7b91c-config-volume\") pod \"collect-profiles-29426895-2sc8g\" (UID: \"ab8b4298-8945-4f7f-aa70-1d08d1a7b91c\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29426895-2sc8g" Dec 13 08:15:00 crc kubenswrapper[4971]: I1213 08:15:00.328586 4971 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/ab8b4298-8945-4f7f-aa70-1d08d1a7b91c-secret-volume\") pod \"collect-profiles-29426895-2sc8g\" (UID: \"ab8b4298-8945-4f7f-aa70-1d08d1a7b91c\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29426895-2sc8g" Dec 13 08:15:00 crc kubenswrapper[4971]: I1213 08:15:00.342563 4971 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-9qhhl\" (UniqueName: \"kubernetes.io/projected/ab8b4298-8945-4f7f-aa70-1d08d1a7b91c-kube-api-access-9qhhl\") pod \"collect-profiles-29426895-2sc8g\" (UID: \"ab8b4298-8945-4f7f-aa70-1d08d1a7b91c\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29426895-2sc8g" Dec 13 08:15:00 crc kubenswrapper[4971]: I1213 08:15:00.497161 4971 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29426895-2sc8g" Dec 13 08:15:01 crc kubenswrapper[4971]: I1213 08:15:01.051249 4971 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29426895-2sc8g"] Dec 13 08:15:01 crc kubenswrapper[4971]: I1213 08:15:01.851426 4971 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29426895-2sc8g" event={"ID":"ab8b4298-8945-4f7f-aa70-1d08d1a7b91c","Type":"ContainerStarted","Data":"eda1293f48d70f7f701282611259ef794a7a56c71da9779a2716decce6d7d261"} Dec 13 08:15:01 crc kubenswrapper[4971]: I1213 08:15:01.851976 4971 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29426895-2sc8g" event={"ID":"ab8b4298-8945-4f7f-aa70-1d08d1a7b91c","Type":"ContainerStarted","Data":"f366d2774b216256a1e0135603f667be560ac877c80193ecfa524c7d9ebf59ef"} Dec 13 08:15:01 crc kubenswrapper[4971]: I1213 08:15:01.885637 4971 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-operator-lifecycle-manager/collect-profiles-29426895-2sc8g" podStartSLOduration=1.885601044 podStartE2EDuration="1.885601044s" podCreationTimestamp="2025-12-13 08:15:00 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-13 08:15:01.877879685 +0000 UTC m=+5158.482289143" watchObservedRunningTime="2025-12-13 08:15:01.885601044 +0000 UTC m=+5158.490010512" Dec 13 08:15:02 crc kubenswrapper[4971]: I1213 08:15:02.867130 4971 generic.go:334] "Generic (PLEG): container finished" podID="ab8b4298-8945-4f7f-aa70-1d08d1a7b91c" containerID="eda1293f48d70f7f701282611259ef794a7a56c71da9779a2716decce6d7d261" exitCode=0 Dec 13 08:15:02 crc kubenswrapper[4971]: I1213 08:15:02.867198 4971 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29426895-2sc8g" event={"ID":"ab8b4298-8945-4f7f-aa70-1d08d1a7b91c","Type":"ContainerDied","Data":"eda1293f48d70f7f701282611259ef794a7a56c71da9779a2716decce6d7d261"} Dec 13 08:15:04 crc kubenswrapper[4971]: I1213 08:15:04.892405 4971 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29426895-2sc8g" event={"ID":"ab8b4298-8945-4f7f-aa70-1d08d1a7b91c","Type":"ContainerDied","Data":"f366d2774b216256a1e0135603f667be560ac877c80193ecfa524c7d9ebf59ef"} Dec 13 08:15:04 crc kubenswrapper[4971]: I1213 08:15:04.894691 4971 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="f366d2774b216256a1e0135603f667be560ac877c80193ecfa524c7d9ebf59ef" Dec 13 08:15:05 crc kubenswrapper[4971]: I1213 08:15:05.214359 4971 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29426895-2sc8g" Dec 13 08:15:05 crc kubenswrapper[4971]: I1213 08:15:05.267623 4971 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-9qhhl\" (UniqueName: \"kubernetes.io/projected/ab8b4298-8945-4f7f-aa70-1d08d1a7b91c-kube-api-access-9qhhl\") pod \"ab8b4298-8945-4f7f-aa70-1d08d1a7b91c\" (UID: \"ab8b4298-8945-4f7f-aa70-1d08d1a7b91c\") " Dec 13 08:15:05 crc kubenswrapper[4971]: I1213 08:15:05.267727 4971 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/ab8b4298-8945-4f7f-aa70-1d08d1a7b91c-config-volume\") pod \"ab8b4298-8945-4f7f-aa70-1d08d1a7b91c\" (UID: \"ab8b4298-8945-4f7f-aa70-1d08d1a7b91c\") " Dec 13 08:15:05 crc kubenswrapper[4971]: I1213 08:15:05.267925 4971 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/ab8b4298-8945-4f7f-aa70-1d08d1a7b91c-secret-volume\") pod \"ab8b4298-8945-4f7f-aa70-1d08d1a7b91c\" (UID: \"ab8b4298-8945-4f7f-aa70-1d08d1a7b91c\") " Dec 13 08:15:05 crc kubenswrapper[4971]: I1213 08:15:05.269069 4971 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/ab8b4298-8945-4f7f-aa70-1d08d1a7b91c-config-volume" (OuterVolumeSpecName: "config-volume") pod "ab8b4298-8945-4f7f-aa70-1d08d1a7b91c" (UID: "ab8b4298-8945-4f7f-aa70-1d08d1a7b91c"). InnerVolumeSpecName "config-volume". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 13 08:15:05 crc kubenswrapper[4971]: I1213 08:15:05.277231 4971 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/ab8b4298-8945-4f7f-aa70-1d08d1a7b91c-secret-volume" (OuterVolumeSpecName: "secret-volume") pod "ab8b4298-8945-4f7f-aa70-1d08d1a7b91c" (UID: "ab8b4298-8945-4f7f-aa70-1d08d1a7b91c"). InnerVolumeSpecName "secret-volume". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 13 08:15:05 crc kubenswrapper[4971]: I1213 08:15:05.277953 4971 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/ab8b4298-8945-4f7f-aa70-1d08d1a7b91c-kube-api-access-9qhhl" (OuterVolumeSpecName: "kube-api-access-9qhhl") pod "ab8b4298-8945-4f7f-aa70-1d08d1a7b91c" (UID: "ab8b4298-8945-4f7f-aa70-1d08d1a7b91c"). InnerVolumeSpecName "kube-api-access-9qhhl". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 13 08:15:05 crc kubenswrapper[4971]: I1213 08:15:05.371072 4971 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-9qhhl\" (UniqueName: \"kubernetes.io/projected/ab8b4298-8945-4f7f-aa70-1d08d1a7b91c-kube-api-access-9qhhl\") on node \"crc\" DevicePath \"\"" Dec 13 08:15:05 crc kubenswrapper[4971]: I1213 08:15:05.371129 4971 reconciler_common.go:293] "Volume detached for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/ab8b4298-8945-4f7f-aa70-1d08d1a7b91c-config-volume\") on node \"crc\" DevicePath \"\"" Dec 13 08:15:05 crc kubenswrapper[4971]: I1213 08:15:05.371145 4971 reconciler_common.go:293] "Volume detached for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/ab8b4298-8945-4f7f-aa70-1d08d1a7b91c-secret-volume\") on node \"crc\" DevicePath \"\"" Dec 13 08:15:05 crc kubenswrapper[4971]: I1213 08:15:05.902852 4971 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29426895-2sc8g" Dec 13 08:15:06 crc kubenswrapper[4971]: I1213 08:15:06.342347 4971 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29426850-rpprk"] Dec 13 08:15:06 crc kubenswrapper[4971]: I1213 08:15:06.357818 4971 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29426850-rpprk"] Dec 13 08:15:07 crc kubenswrapper[4971]: I1213 08:15:07.769874 4971 scope.go:117] "RemoveContainer" containerID="8b2617257e0c4b9039d9f1e54400a1b30b3fb18612ede5bd959442a8eca72ecd" Dec 13 08:15:07 crc kubenswrapper[4971]: E1213 08:15:07.770832 4971 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-82xjz_openshift-machine-config-operator(e0c4af12-edda-41d7-9b44-f87396174ef2)\"" pod="openshift-machine-config-operator/machine-config-daemon-82xjz" podUID="e0c4af12-edda-41d7-9b44-f87396174ef2" Dec 13 08:15:07 crc kubenswrapper[4971]: I1213 08:15:07.787902 4971 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="0e2af7b4-e1cb-4276-bece-56851503f2b0" path="/var/lib/kubelet/pods/0e2af7b4-e1cb-4276-bece-56851503f2b0/volumes" Dec 13 08:15:21 crc kubenswrapper[4971]: I1213 08:15:21.768942 4971 scope.go:117] "RemoveContainer" containerID="8b2617257e0c4b9039d9f1e54400a1b30b3fb18612ede5bd959442a8eca72ecd" Dec 13 08:15:21 crc kubenswrapper[4971]: E1213 08:15:21.769847 4971 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-82xjz_openshift-machine-config-operator(e0c4af12-edda-41d7-9b44-f87396174ef2)\"" pod="openshift-machine-config-operator/machine-config-daemon-82xjz" podUID="e0c4af12-edda-41d7-9b44-f87396174ef2" Dec 13 08:15:26 crc kubenswrapper[4971]: I1213 08:15:26.309929 4971 scope.go:117] "RemoveContainer" containerID="c6325e4a33d42e8703e692af3bb85b8d6f6cd00439d91984ba490882209c6b44" Dec 13 08:15:34 crc kubenswrapper[4971]: I1213 08:15:34.770478 4971 scope.go:117] "RemoveContainer" containerID="8b2617257e0c4b9039d9f1e54400a1b30b3fb18612ede5bd959442a8eca72ecd" Dec 13 08:15:34 crc kubenswrapper[4971]: E1213 08:15:34.772593 4971 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-82xjz_openshift-machine-config-operator(e0c4af12-edda-41d7-9b44-f87396174ef2)\"" pod="openshift-machine-config-operator/machine-config-daemon-82xjz" podUID="e0c4af12-edda-41d7-9b44-f87396174ef2" Dec 13 08:15:49 crc kubenswrapper[4971]: I1213 08:15:49.770222 4971 scope.go:117] "RemoveContainer" containerID="8b2617257e0c4b9039d9f1e54400a1b30b3fb18612ede5bd959442a8eca72ecd" Dec 13 08:15:49 crc kubenswrapper[4971]: E1213 08:15:49.771616 4971 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-82xjz_openshift-machine-config-operator(e0c4af12-edda-41d7-9b44-f87396174ef2)\"" pod="openshift-machine-config-operator/machine-config-daemon-82xjz" podUID="e0c4af12-edda-41d7-9b44-f87396174ef2" Dec 13 08:16:00 crc kubenswrapper[4971]: I1213 08:16:00.769953 4971 scope.go:117] "RemoveContainer" containerID="8b2617257e0c4b9039d9f1e54400a1b30b3fb18612ede5bd959442a8eca72ecd" Dec 13 08:16:00 crc kubenswrapper[4971]: E1213 08:16:00.770680 4971 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-82xjz_openshift-machine-config-operator(e0c4af12-edda-41d7-9b44-f87396174ef2)\"" pod="openshift-machine-config-operator/machine-config-daemon-82xjz" podUID="e0c4af12-edda-41d7-9b44-f87396174ef2" Dec 13 08:16:11 crc kubenswrapper[4971]: I1213 08:16:11.769064 4971 scope.go:117] "RemoveContainer" containerID="8b2617257e0c4b9039d9f1e54400a1b30b3fb18612ede5bd959442a8eca72ecd" Dec 13 08:16:11 crc kubenswrapper[4971]: E1213 08:16:11.769897 4971 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-82xjz_openshift-machine-config-operator(e0c4af12-edda-41d7-9b44-f87396174ef2)\"" pod="openshift-machine-config-operator/machine-config-daemon-82xjz" podUID="e0c4af12-edda-41d7-9b44-f87396174ef2" Dec 13 08:16:20 crc kubenswrapper[4971]: I1213 08:16:20.441602 4971 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/certified-operators-5njt8"] Dec 13 08:16:20 crc kubenswrapper[4971]: E1213 08:16:20.443389 4971 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="ab8b4298-8945-4f7f-aa70-1d08d1a7b91c" containerName="collect-profiles" Dec 13 08:16:20 crc kubenswrapper[4971]: I1213 08:16:20.443415 4971 state_mem.go:107] "Deleted CPUSet assignment" podUID="ab8b4298-8945-4f7f-aa70-1d08d1a7b91c" containerName="collect-profiles" Dec 13 08:16:20 crc kubenswrapper[4971]: I1213 08:16:20.443743 4971 memory_manager.go:354] "RemoveStaleState removing state" podUID="ab8b4298-8945-4f7f-aa70-1d08d1a7b91c" containerName="collect-profiles" Dec 13 08:16:20 crc kubenswrapper[4971]: I1213 08:16:20.446088 4971 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-5njt8" Dec 13 08:16:20 crc kubenswrapper[4971]: I1213 08:16:20.460071 4971 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-5njt8"] Dec 13 08:16:20 crc kubenswrapper[4971]: I1213 08:16:20.521631 4971 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/c460f748-ea66-4bb5-96e2-47ca2686b29b-catalog-content\") pod \"certified-operators-5njt8\" (UID: \"c460f748-ea66-4bb5-96e2-47ca2686b29b\") " pod="openshift-marketplace/certified-operators-5njt8" Dec 13 08:16:20 crc kubenswrapper[4971]: I1213 08:16:20.521763 4971 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/c460f748-ea66-4bb5-96e2-47ca2686b29b-utilities\") pod \"certified-operators-5njt8\" (UID: \"c460f748-ea66-4bb5-96e2-47ca2686b29b\") " pod="openshift-marketplace/certified-operators-5njt8" Dec 13 08:16:20 crc kubenswrapper[4971]: I1213 08:16:20.521918 4971 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-5rlkl\" (UniqueName: \"kubernetes.io/projected/c460f748-ea66-4bb5-96e2-47ca2686b29b-kube-api-access-5rlkl\") pod \"certified-operators-5njt8\" (UID: \"c460f748-ea66-4bb5-96e2-47ca2686b29b\") " pod="openshift-marketplace/certified-operators-5njt8" Dec 13 08:16:20 crc kubenswrapper[4971]: I1213 08:16:20.624556 4971 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/c460f748-ea66-4bb5-96e2-47ca2686b29b-catalog-content\") pod \"certified-operators-5njt8\" (UID: \"c460f748-ea66-4bb5-96e2-47ca2686b29b\") " pod="openshift-marketplace/certified-operators-5njt8" Dec 13 08:16:20 crc kubenswrapper[4971]: I1213 08:16:20.624666 4971 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/c460f748-ea66-4bb5-96e2-47ca2686b29b-utilities\") pod \"certified-operators-5njt8\" (UID: \"c460f748-ea66-4bb5-96e2-47ca2686b29b\") " pod="openshift-marketplace/certified-operators-5njt8" Dec 13 08:16:20 crc kubenswrapper[4971]: I1213 08:16:20.624834 4971 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-5rlkl\" (UniqueName: \"kubernetes.io/projected/c460f748-ea66-4bb5-96e2-47ca2686b29b-kube-api-access-5rlkl\") pod \"certified-operators-5njt8\" (UID: \"c460f748-ea66-4bb5-96e2-47ca2686b29b\") " pod="openshift-marketplace/certified-operators-5njt8" Dec 13 08:16:20 crc kubenswrapper[4971]: I1213 08:16:20.625896 4971 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/c460f748-ea66-4bb5-96e2-47ca2686b29b-catalog-content\") pod \"certified-operators-5njt8\" (UID: \"c460f748-ea66-4bb5-96e2-47ca2686b29b\") " pod="openshift-marketplace/certified-operators-5njt8" Dec 13 08:16:20 crc kubenswrapper[4971]: I1213 08:16:20.625946 4971 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/c460f748-ea66-4bb5-96e2-47ca2686b29b-utilities\") pod \"certified-operators-5njt8\" (UID: \"c460f748-ea66-4bb5-96e2-47ca2686b29b\") " pod="openshift-marketplace/certified-operators-5njt8" Dec 13 08:16:20 crc kubenswrapper[4971]: I1213 08:16:20.665227 4971 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-5rlkl\" (UniqueName: \"kubernetes.io/projected/c460f748-ea66-4bb5-96e2-47ca2686b29b-kube-api-access-5rlkl\") pod \"certified-operators-5njt8\" (UID: \"c460f748-ea66-4bb5-96e2-47ca2686b29b\") " pod="openshift-marketplace/certified-operators-5njt8" Dec 13 08:16:20 crc kubenswrapper[4971]: I1213 08:16:20.778212 4971 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-5njt8" Dec 13 08:16:21 crc kubenswrapper[4971]: I1213 08:16:21.434217 4971 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-5njt8"] Dec 13 08:16:21 crc kubenswrapper[4971]: I1213 08:16:21.783871 4971 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-5njt8" event={"ID":"c460f748-ea66-4bb5-96e2-47ca2686b29b","Type":"ContainerStarted","Data":"ba33c667047cfd5e1a58d12d0e42b1420e52a200976a10a63044f874e7ae5501"} Dec 13 08:16:22 crc kubenswrapper[4971]: I1213 08:16:22.795590 4971 generic.go:334] "Generic (PLEG): container finished" podID="c460f748-ea66-4bb5-96e2-47ca2686b29b" containerID="cc264c930e1394428d95f33de4e76cacf6c84f77124060e9087e48ca88dc7876" exitCode=0 Dec 13 08:16:22 crc kubenswrapper[4971]: I1213 08:16:22.795965 4971 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-5njt8" event={"ID":"c460f748-ea66-4bb5-96e2-47ca2686b29b","Type":"ContainerDied","Data":"cc264c930e1394428d95f33de4e76cacf6c84f77124060e9087e48ca88dc7876"} Dec 13 08:16:22 crc kubenswrapper[4971]: I1213 08:16:22.798564 4971 provider.go:102] Refreshing cache for provider: *credentialprovider.defaultDockerConfigProvider Dec 13 08:16:24 crc kubenswrapper[4971]: I1213 08:16:24.823470 4971 generic.go:334] "Generic (PLEG): container finished" podID="c460f748-ea66-4bb5-96e2-47ca2686b29b" containerID="c89686f6017da92ac9fb4e9be2b004cdd76fc000164a965e22431c1f5e5770f6" exitCode=0 Dec 13 08:16:24 crc kubenswrapper[4971]: I1213 08:16:24.823554 4971 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-5njt8" event={"ID":"c460f748-ea66-4bb5-96e2-47ca2686b29b","Type":"ContainerDied","Data":"c89686f6017da92ac9fb4e9be2b004cdd76fc000164a965e22431c1f5e5770f6"} Dec 13 08:16:25 crc kubenswrapper[4971]: I1213 08:16:25.768764 4971 scope.go:117] "RemoveContainer" containerID="8b2617257e0c4b9039d9f1e54400a1b30b3fb18612ede5bd959442a8eca72ecd" Dec 13 08:16:25 crc kubenswrapper[4971]: E1213 08:16:25.769568 4971 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-82xjz_openshift-machine-config-operator(e0c4af12-edda-41d7-9b44-f87396174ef2)\"" pod="openshift-machine-config-operator/machine-config-daemon-82xjz" podUID="e0c4af12-edda-41d7-9b44-f87396174ef2" Dec 13 08:16:27 crc kubenswrapper[4971]: I1213 08:16:27.937965 4971 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-5njt8" event={"ID":"c460f748-ea66-4bb5-96e2-47ca2686b29b","Type":"ContainerStarted","Data":"95cf1bac07483a22865da258c7dfcc735daf52a711cadb484141630d4c9ffada"} Dec 13 08:16:30 crc kubenswrapper[4971]: I1213 08:16:30.778576 4971 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/certified-operators-5njt8" Dec 13 08:16:30 crc kubenswrapper[4971]: I1213 08:16:30.779558 4971 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/certified-operators-5njt8" Dec 13 08:16:30 crc kubenswrapper[4971]: I1213 08:16:30.853387 4971 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/certified-operators-5njt8" Dec 13 08:16:30 crc kubenswrapper[4971]: I1213 08:16:30.894293 4971 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/certified-operators-5njt8" podStartSLOduration=6.585093514 podStartE2EDuration="10.894250652s" podCreationTimestamp="2025-12-13 08:16:20 +0000 UTC" firstStartedPulling="2025-12-13 08:16:22.798260231 +0000 UTC m=+5239.402669679" lastFinishedPulling="2025-12-13 08:16:27.107417369 +0000 UTC m=+5243.711826817" observedRunningTime="2025-12-13 08:16:27.96570518 +0000 UTC m=+5244.570114638" watchObservedRunningTime="2025-12-13 08:16:30.894250652 +0000 UTC m=+5247.498660110" Dec 13 08:16:36 crc kubenswrapper[4971]: I1213 08:16:36.770164 4971 scope.go:117] "RemoveContainer" containerID="8b2617257e0c4b9039d9f1e54400a1b30b3fb18612ede5bd959442a8eca72ecd" Dec 13 08:16:36 crc kubenswrapper[4971]: E1213 08:16:36.772186 4971 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-82xjz_openshift-machine-config-operator(e0c4af12-edda-41d7-9b44-f87396174ef2)\"" pod="openshift-machine-config-operator/machine-config-daemon-82xjz" podUID="e0c4af12-edda-41d7-9b44-f87396174ef2" Dec 13 08:16:41 crc kubenswrapper[4971]: I1213 08:16:41.185308 4971 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/certified-operators-5njt8" Dec 13 08:16:41 crc kubenswrapper[4971]: I1213 08:16:41.255862 4971 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-5njt8"] Dec 13 08:16:42 crc kubenswrapper[4971]: I1213 08:16:42.185155 4971 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/certified-operators-5njt8" podUID="c460f748-ea66-4bb5-96e2-47ca2686b29b" containerName="registry-server" containerID="cri-o://95cf1bac07483a22865da258c7dfcc735daf52a711cadb484141630d4c9ffada" gracePeriod=2 Dec 13 08:16:42 crc kubenswrapper[4971]: I1213 08:16:42.988294 4971 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-5njt8" Dec 13 08:16:43 crc kubenswrapper[4971]: I1213 08:16:43.075454 4971 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-5rlkl\" (UniqueName: \"kubernetes.io/projected/c460f748-ea66-4bb5-96e2-47ca2686b29b-kube-api-access-5rlkl\") pod \"c460f748-ea66-4bb5-96e2-47ca2686b29b\" (UID: \"c460f748-ea66-4bb5-96e2-47ca2686b29b\") " Dec 13 08:16:43 crc kubenswrapper[4971]: I1213 08:16:43.075562 4971 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/c460f748-ea66-4bb5-96e2-47ca2686b29b-catalog-content\") pod \"c460f748-ea66-4bb5-96e2-47ca2686b29b\" (UID: \"c460f748-ea66-4bb5-96e2-47ca2686b29b\") " Dec 13 08:16:43 crc kubenswrapper[4971]: I1213 08:16:43.075614 4971 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/c460f748-ea66-4bb5-96e2-47ca2686b29b-utilities\") pod \"c460f748-ea66-4bb5-96e2-47ca2686b29b\" (UID: \"c460f748-ea66-4bb5-96e2-47ca2686b29b\") " Dec 13 08:16:43 crc kubenswrapper[4971]: I1213 08:16:43.077350 4971 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/c460f748-ea66-4bb5-96e2-47ca2686b29b-utilities" (OuterVolumeSpecName: "utilities") pod "c460f748-ea66-4bb5-96e2-47ca2686b29b" (UID: "c460f748-ea66-4bb5-96e2-47ca2686b29b"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 13 08:16:43 crc kubenswrapper[4971]: I1213 08:16:43.088456 4971 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/c460f748-ea66-4bb5-96e2-47ca2686b29b-kube-api-access-5rlkl" (OuterVolumeSpecName: "kube-api-access-5rlkl") pod "c460f748-ea66-4bb5-96e2-47ca2686b29b" (UID: "c460f748-ea66-4bb5-96e2-47ca2686b29b"). InnerVolumeSpecName "kube-api-access-5rlkl". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 13 08:16:43 crc kubenswrapper[4971]: I1213 08:16:43.138915 4971 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/c460f748-ea66-4bb5-96e2-47ca2686b29b-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "c460f748-ea66-4bb5-96e2-47ca2686b29b" (UID: "c460f748-ea66-4bb5-96e2-47ca2686b29b"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 13 08:16:43 crc kubenswrapper[4971]: I1213 08:16:43.178737 4971 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-5rlkl\" (UniqueName: \"kubernetes.io/projected/c460f748-ea66-4bb5-96e2-47ca2686b29b-kube-api-access-5rlkl\") on node \"crc\" DevicePath \"\"" Dec 13 08:16:43 crc kubenswrapper[4971]: I1213 08:16:43.178780 4971 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/c460f748-ea66-4bb5-96e2-47ca2686b29b-catalog-content\") on node \"crc\" DevicePath \"\"" Dec 13 08:16:43 crc kubenswrapper[4971]: I1213 08:16:43.178791 4971 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/c460f748-ea66-4bb5-96e2-47ca2686b29b-utilities\") on node \"crc\" DevicePath \"\"" Dec 13 08:16:43 crc kubenswrapper[4971]: I1213 08:16:43.200411 4971 generic.go:334] "Generic (PLEG): container finished" podID="c460f748-ea66-4bb5-96e2-47ca2686b29b" containerID="95cf1bac07483a22865da258c7dfcc735daf52a711cadb484141630d4c9ffada" exitCode=0 Dec 13 08:16:43 crc kubenswrapper[4971]: I1213 08:16:43.200536 4971 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-5njt8" Dec 13 08:16:43 crc kubenswrapper[4971]: I1213 08:16:43.200542 4971 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-5njt8" event={"ID":"c460f748-ea66-4bb5-96e2-47ca2686b29b","Type":"ContainerDied","Data":"95cf1bac07483a22865da258c7dfcc735daf52a711cadb484141630d4c9ffada"} Dec 13 08:16:43 crc kubenswrapper[4971]: I1213 08:16:43.200619 4971 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-5njt8" event={"ID":"c460f748-ea66-4bb5-96e2-47ca2686b29b","Type":"ContainerDied","Data":"ba33c667047cfd5e1a58d12d0e42b1420e52a200976a10a63044f874e7ae5501"} Dec 13 08:16:43 crc kubenswrapper[4971]: I1213 08:16:43.200644 4971 scope.go:117] "RemoveContainer" containerID="95cf1bac07483a22865da258c7dfcc735daf52a711cadb484141630d4c9ffada" Dec 13 08:16:43 crc kubenswrapper[4971]: I1213 08:16:43.236884 4971 scope.go:117] "RemoveContainer" containerID="c89686f6017da92ac9fb4e9be2b004cdd76fc000164a965e22431c1f5e5770f6" Dec 13 08:16:43 crc kubenswrapper[4971]: I1213 08:16:43.240296 4971 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-5njt8"] Dec 13 08:16:43 crc kubenswrapper[4971]: I1213 08:16:43.249941 4971 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/certified-operators-5njt8"] Dec 13 08:16:43 crc kubenswrapper[4971]: I1213 08:16:43.265312 4971 scope.go:117] "RemoveContainer" containerID="cc264c930e1394428d95f33de4e76cacf6c84f77124060e9087e48ca88dc7876" Dec 13 08:16:43 crc kubenswrapper[4971]: I1213 08:16:43.337390 4971 scope.go:117] "RemoveContainer" containerID="95cf1bac07483a22865da258c7dfcc735daf52a711cadb484141630d4c9ffada" Dec 13 08:16:43 crc kubenswrapper[4971]: E1213 08:16:43.340934 4971 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"95cf1bac07483a22865da258c7dfcc735daf52a711cadb484141630d4c9ffada\": container with ID starting with 95cf1bac07483a22865da258c7dfcc735daf52a711cadb484141630d4c9ffada not found: ID does not exist" containerID="95cf1bac07483a22865da258c7dfcc735daf52a711cadb484141630d4c9ffada" Dec 13 08:16:43 crc kubenswrapper[4971]: I1213 08:16:43.341006 4971 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"95cf1bac07483a22865da258c7dfcc735daf52a711cadb484141630d4c9ffada"} err="failed to get container status \"95cf1bac07483a22865da258c7dfcc735daf52a711cadb484141630d4c9ffada\": rpc error: code = NotFound desc = could not find container \"95cf1bac07483a22865da258c7dfcc735daf52a711cadb484141630d4c9ffada\": container with ID starting with 95cf1bac07483a22865da258c7dfcc735daf52a711cadb484141630d4c9ffada not found: ID does not exist" Dec 13 08:16:43 crc kubenswrapper[4971]: I1213 08:16:43.341036 4971 scope.go:117] "RemoveContainer" containerID="c89686f6017da92ac9fb4e9be2b004cdd76fc000164a965e22431c1f5e5770f6" Dec 13 08:16:43 crc kubenswrapper[4971]: E1213 08:16:43.347914 4971 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"c89686f6017da92ac9fb4e9be2b004cdd76fc000164a965e22431c1f5e5770f6\": container with ID starting with c89686f6017da92ac9fb4e9be2b004cdd76fc000164a965e22431c1f5e5770f6 not found: ID does not exist" containerID="c89686f6017da92ac9fb4e9be2b004cdd76fc000164a965e22431c1f5e5770f6" Dec 13 08:16:43 crc kubenswrapper[4971]: I1213 08:16:43.348013 4971 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"c89686f6017da92ac9fb4e9be2b004cdd76fc000164a965e22431c1f5e5770f6"} err="failed to get container status \"c89686f6017da92ac9fb4e9be2b004cdd76fc000164a965e22431c1f5e5770f6\": rpc error: code = NotFound desc = could not find container \"c89686f6017da92ac9fb4e9be2b004cdd76fc000164a965e22431c1f5e5770f6\": container with ID starting with c89686f6017da92ac9fb4e9be2b004cdd76fc000164a965e22431c1f5e5770f6 not found: ID does not exist" Dec 13 08:16:43 crc kubenswrapper[4971]: I1213 08:16:43.348081 4971 scope.go:117] "RemoveContainer" containerID="cc264c930e1394428d95f33de4e76cacf6c84f77124060e9087e48ca88dc7876" Dec 13 08:16:43 crc kubenswrapper[4971]: E1213 08:16:43.349174 4971 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"cc264c930e1394428d95f33de4e76cacf6c84f77124060e9087e48ca88dc7876\": container with ID starting with cc264c930e1394428d95f33de4e76cacf6c84f77124060e9087e48ca88dc7876 not found: ID does not exist" containerID="cc264c930e1394428d95f33de4e76cacf6c84f77124060e9087e48ca88dc7876" Dec 13 08:16:43 crc kubenswrapper[4971]: I1213 08:16:43.349227 4971 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"cc264c930e1394428d95f33de4e76cacf6c84f77124060e9087e48ca88dc7876"} err="failed to get container status \"cc264c930e1394428d95f33de4e76cacf6c84f77124060e9087e48ca88dc7876\": rpc error: code = NotFound desc = could not find container \"cc264c930e1394428d95f33de4e76cacf6c84f77124060e9087e48ca88dc7876\": container with ID starting with cc264c930e1394428d95f33de4e76cacf6c84f77124060e9087e48ca88dc7876 not found: ID does not exist" Dec 13 08:16:43 crc kubenswrapper[4971]: I1213 08:16:43.881042 4971 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="c460f748-ea66-4bb5-96e2-47ca2686b29b" path="/var/lib/kubelet/pods/c460f748-ea66-4bb5-96e2-47ca2686b29b/volumes" Dec 13 08:16:47 crc kubenswrapper[4971]: I1213 08:16:47.775699 4971 scope.go:117] "RemoveContainer" containerID="8b2617257e0c4b9039d9f1e54400a1b30b3fb18612ede5bd959442a8eca72ecd" Dec 13 08:16:47 crc kubenswrapper[4971]: E1213 08:16:47.776770 4971 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-82xjz_openshift-machine-config-operator(e0c4af12-edda-41d7-9b44-f87396174ef2)\"" pod="openshift-machine-config-operator/machine-config-daemon-82xjz" podUID="e0c4af12-edda-41d7-9b44-f87396174ef2" Dec 13 08:17:00 crc kubenswrapper[4971]: I1213 08:17:00.768257 4971 scope.go:117] "RemoveContainer" containerID="8b2617257e0c4b9039d9f1e54400a1b30b3fb18612ede5bd959442a8eca72ecd" Dec 13 08:17:00 crc kubenswrapper[4971]: E1213 08:17:00.769763 4971 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-82xjz_openshift-machine-config-operator(e0c4af12-edda-41d7-9b44-f87396174ef2)\"" pod="openshift-machine-config-operator/machine-config-daemon-82xjz" podUID="e0c4af12-edda-41d7-9b44-f87396174ef2" Dec 13 08:17:11 crc kubenswrapper[4971]: I1213 08:17:11.769508 4971 scope.go:117] "RemoveContainer" containerID="8b2617257e0c4b9039d9f1e54400a1b30b3fb18612ede5bd959442a8eca72ecd" Dec 13 08:17:11 crc kubenswrapper[4971]: E1213 08:17:11.771903 4971 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-82xjz_openshift-machine-config-operator(e0c4af12-edda-41d7-9b44-f87396174ef2)\"" pod="openshift-machine-config-operator/machine-config-daemon-82xjz" podUID="e0c4af12-edda-41d7-9b44-f87396174ef2" Dec 13 08:17:22 crc kubenswrapper[4971]: I1213 08:17:22.769377 4971 scope.go:117] "RemoveContainer" containerID="8b2617257e0c4b9039d9f1e54400a1b30b3fb18612ede5bd959442a8eca72ecd" Dec 13 08:17:22 crc kubenswrapper[4971]: E1213 08:17:22.770921 4971 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-82xjz_openshift-machine-config-operator(e0c4af12-edda-41d7-9b44-f87396174ef2)\"" pod="openshift-machine-config-operator/machine-config-daemon-82xjz" podUID="e0c4af12-edda-41d7-9b44-f87396174ef2" Dec 13 08:17:37 crc kubenswrapper[4971]: I1213 08:17:37.769292 4971 scope.go:117] "RemoveContainer" containerID="8b2617257e0c4b9039d9f1e54400a1b30b3fb18612ede5bd959442a8eca72ecd" Dec 13 08:17:37 crc kubenswrapper[4971]: E1213 08:17:37.770548 4971 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-82xjz_openshift-machine-config-operator(e0c4af12-edda-41d7-9b44-f87396174ef2)\"" pod="openshift-machine-config-operator/machine-config-daemon-82xjz" podUID="e0c4af12-edda-41d7-9b44-f87396174ef2" Dec 13 08:17:49 crc kubenswrapper[4971]: I1213 08:17:49.768834 4971 scope.go:117] "RemoveContainer" containerID="8b2617257e0c4b9039d9f1e54400a1b30b3fb18612ede5bd959442a8eca72ecd" Dec 13 08:17:49 crc kubenswrapper[4971]: E1213 08:17:49.769733 4971 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-82xjz_openshift-machine-config-operator(e0c4af12-edda-41d7-9b44-f87396174ef2)\"" pod="openshift-machine-config-operator/machine-config-daemon-82xjz" podUID="e0c4af12-edda-41d7-9b44-f87396174ef2" Dec 13 08:18:04 crc kubenswrapper[4971]: I1213 08:18:04.769049 4971 scope.go:117] "RemoveContainer" containerID="8b2617257e0c4b9039d9f1e54400a1b30b3fb18612ede5bd959442a8eca72ecd" Dec 13 08:18:04 crc kubenswrapper[4971]: E1213 08:18:04.769900 4971 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-82xjz_openshift-machine-config-operator(e0c4af12-edda-41d7-9b44-f87396174ef2)\"" pod="openshift-machine-config-operator/machine-config-daemon-82xjz" podUID="e0c4af12-edda-41d7-9b44-f87396174ef2" Dec 13 08:18:16 crc kubenswrapper[4971]: I1213 08:18:16.769000 4971 scope.go:117] "RemoveContainer" containerID="8b2617257e0c4b9039d9f1e54400a1b30b3fb18612ede5bd959442a8eca72ecd" Dec 13 08:18:16 crc kubenswrapper[4971]: E1213 08:18:16.770310 4971 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-82xjz_openshift-machine-config-operator(e0c4af12-edda-41d7-9b44-f87396174ef2)\"" pod="openshift-machine-config-operator/machine-config-daemon-82xjz" podUID="e0c4af12-edda-41d7-9b44-f87396174ef2" Dec 13 08:18:31 crc kubenswrapper[4971]: I1213 08:18:31.769075 4971 scope.go:117] "RemoveContainer" containerID="8b2617257e0c4b9039d9f1e54400a1b30b3fb18612ede5bd959442a8eca72ecd" Dec 13 08:18:31 crc kubenswrapper[4971]: E1213 08:18:31.770500 4971 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-82xjz_openshift-machine-config-operator(e0c4af12-edda-41d7-9b44-f87396174ef2)\"" pod="openshift-machine-config-operator/machine-config-daemon-82xjz" podUID="e0c4af12-edda-41d7-9b44-f87396174ef2" Dec 13 08:18:46 crc kubenswrapper[4971]: I1213 08:18:46.768900 4971 scope.go:117] "RemoveContainer" containerID="8b2617257e0c4b9039d9f1e54400a1b30b3fb18612ede5bd959442a8eca72ecd" Dec 13 08:18:47 crc kubenswrapper[4971]: I1213 08:18:47.780567 4971 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-82xjz" event={"ID":"e0c4af12-edda-41d7-9b44-f87396174ef2","Type":"ContainerStarted","Data":"b2fd564c27978e268006043ad8ebcbcac704f06f77bb3bdbd7878948103f567f"} Dec 13 08:21:16 crc kubenswrapper[4971]: I1213 08:21:16.153285 4971 patch_prober.go:28] interesting pod/machine-config-daemon-82xjz container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 13 08:21:16 crc kubenswrapper[4971]: I1213 08:21:16.154124 4971 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-82xjz" podUID="e0c4af12-edda-41d7-9b44-f87396174ef2" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 13 08:21:46 crc kubenswrapper[4971]: I1213 08:21:46.153473 4971 patch_prober.go:28] interesting pod/machine-config-daemon-82xjz container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 13 08:21:46 crc kubenswrapper[4971]: I1213 08:21:46.155800 4971 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-82xjz" podUID="e0c4af12-edda-41d7-9b44-f87396174ef2" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 13 08:22:01 crc kubenswrapper[4971]: I1213 08:22:01.601563 4971 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-operators-5zh2c"] Dec 13 08:22:01 crc kubenswrapper[4971]: E1213 08:22:01.602980 4971 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="c460f748-ea66-4bb5-96e2-47ca2686b29b" containerName="extract-content" Dec 13 08:22:01 crc kubenswrapper[4971]: I1213 08:22:01.603001 4971 state_mem.go:107] "Deleted CPUSet assignment" podUID="c460f748-ea66-4bb5-96e2-47ca2686b29b" containerName="extract-content" Dec 13 08:22:01 crc kubenswrapper[4971]: E1213 08:22:01.603020 4971 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="c460f748-ea66-4bb5-96e2-47ca2686b29b" containerName="extract-utilities" Dec 13 08:22:01 crc kubenswrapper[4971]: I1213 08:22:01.603029 4971 state_mem.go:107] "Deleted CPUSet assignment" podUID="c460f748-ea66-4bb5-96e2-47ca2686b29b" containerName="extract-utilities" Dec 13 08:22:01 crc kubenswrapper[4971]: E1213 08:22:01.603044 4971 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="c460f748-ea66-4bb5-96e2-47ca2686b29b" containerName="registry-server" Dec 13 08:22:01 crc kubenswrapper[4971]: I1213 08:22:01.603054 4971 state_mem.go:107] "Deleted CPUSet assignment" podUID="c460f748-ea66-4bb5-96e2-47ca2686b29b" containerName="registry-server" Dec 13 08:22:01 crc kubenswrapper[4971]: I1213 08:22:01.603303 4971 memory_manager.go:354] "RemoveStaleState removing state" podUID="c460f748-ea66-4bb5-96e2-47ca2686b29b" containerName="registry-server" Dec 13 08:22:01 crc kubenswrapper[4971]: I1213 08:22:01.604661 4971 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-5zh2c" Dec 13 08:22:01 crc kubenswrapper[4971]: I1213 08:22:01.636801 4971 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-5zh2c"] Dec 13 08:22:01 crc kubenswrapper[4971]: I1213 08:22:01.663820 4971 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/7fabfced-a2af-4a4c-b3e1-03461dc4ce53-catalog-content\") pod \"redhat-operators-5zh2c\" (UID: \"7fabfced-a2af-4a4c-b3e1-03461dc4ce53\") " pod="openshift-marketplace/redhat-operators-5zh2c" Dec 13 08:22:01 crc kubenswrapper[4971]: I1213 08:22:01.664179 4971 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-k7gqx\" (UniqueName: \"kubernetes.io/projected/7fabfced-a2af-4a4c-b3e1-03461dc4ce53-kube-api-access-k7gqx\") pod \"redhat-operators-5zh2c\" (UID: \"7fabfced-a2af-4a4c-b3e1-03461dc4ce53\") " pod="openshift-marketplace/redhat-operators-5zh2c" Dec 13 08:22:01 crc kubenswrapper[4971]: I1213 08:22:01.664635 4971 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/7fabfced-a2af-4a4c-b3e1-03461dc4ce53-utilities\") pod \"redhat-operators-5zh2c\" (UID: \"7fabfced-a2af-4a4c-b3e1-03461dc4ce53\") " pod="openshift-marketplace/redhat-operators-5zh2c" Dec 13 08:22:01 crc kubenswrapper[4971]: I1213 08:22:01.766462 4971 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/7fabfced-a2af-4a4c-b3e1-03461dc4ce53-utilities\") pod \"redhat-operators-5zh2c\" (UID: \"7fabfced-a2af-4a4c-b3e1-03461dc4ce53\") " pod="openshift-marketplace/redhat-operators-5zh2c" Dec 13 08:22:01 crc kubenswrapper[4971]: I1213 08:22:01.766670 4971 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/7fabfced-a2af-4a4c-b3e1-03461dc4ce53-catalog-content\") pod \"redhat-operators-5zh2c\" (UID: \"7fabfced-a2af-4a4c-b3e1-03461dc4ce53\") " pod="openshift-marketplace/redhat-operators-5zh2c" Dec 13 08:22:01 crc kubenswrapper[4971]: I1213 08:22:01.766748 4971 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-k7gqx\" (UniqueName: \"kubernetes.io/projected/7fabfced-a2af-4a4c-b3e1-03461dc4ce53-kube-api-access-k7gqx\") pod \"redhat-operators-5zh2c\" (UID: \"7fabfced-a2af-4a4c-b3e1-03461dc4ce53\") " pod="openshift-marketplace/redhat-operators-5zh2c" Dec 13 08:22:01 crc kubenswrapper[4971]: I1213 08:22:01.767460 4971 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/7fabfced-a2af-4a4c-b3e1-03461dc4ce53-utilities\") pod \"redhat-operators-5zh2c\" (UID: \"7fabfced-a2af-4a4c-b3e1-03461dc4ce53\") " pod="openshift-marketplace/redhat-operators-5zh2c" Dec 13 08:22:01 crc kubenswrapper[4971]: I1213 08:22:01.767758 4971 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/7fabfced-a2af-4a4c-b3e1-03461dc4ce53-catalog-content\") pod \"redhat-operators-5zh2c\" (UID: \"7fabfced-a2af-4a4c-b3e1-03461dc4ce53\") " pod="openshift-marketplace/redhat-operators-5zh2c" Dec 13 08:22:01 crc kubenswrapper[4971]: I1213 08:22:01.793057 4971 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-k7gqx\" (UniqueName: \"kubernetes.io/projected/7fabfced-a2af-4a4c-b3e1-03461dc4ce53-kube-api-access-k7gqx\") pod \"redhat-operators-5zh2c\" (UID: \"7fabfced-a2af-4a4c-b3e1-03461dc4ce53\") " pod="openshift-marketplace/redhat-operators-5zh2c" Dec 13 08:22:01 crc kubenswrapper[4971]: I1213 08:22:01.932658 4971 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-5zh2c" Dec 13 08:22:02 crc kubenswrapper[4971]: I1213 08:22:02.432374 4971 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-5zh2c"] Dec 13 08:22:03 crc kubenswrapper[4971]: I1213 08:22:03.029588 4971 generic.go:334] "Generic (PLEG): container finished" podID="7fabfced-a2af-4a4c-b3e1-03461dc4ce53" containerID="2b96f3ee27d9c0fb5e3f165bd3edb2714b46416277d56215775cb2cb57be96af" exitCode=0 Dec 13 08:22:03 crc kubenswrapper[4971]: I1213 08:22:03.029689 4971 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-5zh2c" event={"ID":"7fabfced-a2af-4a4c-b3e1-03461dc4ce53","Type":"ContainerDied","Data":"2b96f3ee27d9c0fb5e3f165bd3edb2714b46416277d56215775cb2cb57be96af"} Dec 13 08:22:03 crc kubenswrapper[4971]: I1213 08:22:03.029827 4971 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-5zh2c" event={"ID":"7fabfced-a2af-4a4c-b3e1-03461dc4ce53","Type":"ContainerStarted","Data":"339e83640ceb75328c717a374d1bb53c87245d152d8dcaf9c47f5c73095cdce8"} Dec 13 08:22:03 crc kubenswrapper[4971]: I1213 08:22:03.031629 4971 provider.go:102] Refreshing cache for provider: *credentialprovider.defaultDockerConfigProvider Dec 13 08:22:04 crc kubenswrapper[4971]: I1213 08:22:04.041330 4971 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-5zh2c" event={"ID":"7fabfced-a2af-4a4c-b3e1-03461dc4ce53","Type":"ContainerStarted","Data":"0a36d72ca16ce5be7880c2cb8e190ad8986ebb8b78e7a11aebc205d6cf1ac6a3"} Dec 13 08:22:07 crc kubenswrapper[4971]: I1213 08:22:07.074985 4971 generic.go:334] "Generic (PLEG): container finished" podID="7fabfced-a2af-4a4c-b3e1-03461dc4ce53" containerID="0a36d72ca16ce5be7880c2cb8e190ad8986ebb8b78e7a11aebc205d6cf1ac6a3" exitCode=0 Dec 13 08:22:07 crc kubenswrapper[4971]: I1213 08:22:07.075087 4971 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-5zh2c" event={"ID":"7fabfced-a2af-4a4c-b3e1-03461dc4ce53","Type":"ContainerDied","Data":"0a36d72ca16ce5be7880c2cb8e190ad8986ebb8b78e7a11aebc205d6cf1ac6a3"} Dec 13 08:22:09 crc kubenswrapper[4971]: I1213 08:22:09.094690 4971 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-5zh2c" event={"ID":"7fabfced-a2af-4a4c-b3e1-03461dc4ce53","Type":"ContainerStarted","Data":"73f44405eef2fedf0a637d913c0b17ef3c0295bd223b91d6bfcf534da9900440"} Dec 13 08:22:09 crc kubenswrapper[4971]: I1213 08:22:09.128958 4971 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-operators-5zh2c" podStartSLOduration=2.663135599 podStartE2EDuration="8.128926726s" podCreationTimestamp="2025-12-13 08:22:01 +0000 UTC" firstStartedPulling="2025-12-13 08:22:03.031382921 +0000 UTC m=+5579.635792369" lastFinishedPulling="2025-12-13 08:22:08.497174048 +0000 UTC m=+5585.101583496" observedRunningTime="2025-12-13 08:22:09.115974629 +0000 UTC m=+5585.720384077" watchObservedRunningTime="2025-12-13 08:22:09.128926726 +0000 UTC m=+5585.733336194" Dec 13 08:22:11 crc kubenswrapper[4971]: I1213 08:22:11.932941 4971 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-operators-5zh2c" Dec 13 08:22:11 crc kubenswrapper[4971]: I1213 08:22:11.933618 4971 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-operators-5zh2c" Dec 13 08:22:12 crc kubenswrapper[4971]: I1213 08:22:12.997879 4971 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-marketplace/redhat-operators-5zh2c" podUID="7fabfced-a2af-4a4c-b3e1-03461dc4ce53" containerName="registry-server" probeResult="failure" output=< Dec 13 08:22:12 crc kubenswrapper[4971]: timeout: failed to connect service ":50051" within 1s Dec 13 08:22:12 crc kubenswrapper[4971]: > Dec 13 08:22:16 crc kubenswrapper[4971]: I1213 08:22:16.153320 4971 patch_prober.go:28] interesting pod/machine-config-daemon-82xjz container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 13 08:22:16 crc kubenswrapper[4971]: I1213 08:22:16.153750 4971 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-82xjz" podUID="e0c4af12-edda-41d7-9b44-f87396174ef2" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 13 08:22:16 crc kubenswrapper[4971]: I1213 08:22:16.153796 4971 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-82xjz" Dec 13 08:22:16 crc kubenswrapper[4971]: I1213 08:22:16.154641 4971 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"b2fd564c27978e268006043ad8ebcbcac704f06f77bb3bdbd7878948103f567f"} pod="openshift-machine-config-operator/machine-config-daemon-82xjz" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Dec 13 08:22:16 crc kubenswrapper[4971]: I1213 08:22:16.155000 4971 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-82xjz" podUID="e0c4af12-edda-41d7-9b44-f87396174ef2" containerName="machine-config-daemon" containerID="cri-o://b2fd564c27978e268006043ad8ebcbcac704f06f77bb3bdbd7878948103f567f" gracePeriod=600 Dec 13 08:22:17 crc kubenswrapper[4971]: I1213 08:22:17.166324 4971 generic.go:334] "Generic (PLEG): container finished" podID="e0c4af12-edda-41d7-9b44-f87396174ef2" containerID="b2fd564c27978e268006043ad8ebcbcac704f06f77bb3bdbd7878948103f567f" exitCode=0 Dec 13 08:22:17 crc kubenswrapper[4971]: I1213 08:22:17.166387 4971 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-82xjz" event={"ID":"e0c4af12-edda-41d7-9b44-f87396174ef2","Type":"ContainerDied","Data":"b2fd564c27978e268006043ad8ebcbcac704f06f77bb3bdbd7878948103f567f"} Dec 13 08:22:17 crc kubenswrapper[4971]: I1213 08:22:17.167008 4971 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-82xjz" event={"ID":"e0c4af12-edda-41d7-9b44-f87396174ef2","Type":"ContainerStarted","Data":"bc1c06c3bcebf9c00aee813c7819c279356be670c515e1ea302b50b8618bc001"} Dec 13 08:22:17 crc kubenswrapper[4971]: I1213 08:22:17.167035 4971 scope.go:117] "RemoveContainer" containerID="8b2617257e0c4b9039d9f1e54400a1b30b3fb18612ede5bd959442a8eca72ecd" Dec 13 08:22:21 crc kubenswrapper[4971]: I1213 08:22:21.981134 4971 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-operators-5zh2c" Dec 13 08:22:22 crc kubenswrapper[4971]: I1213 08:22:22.075944 4971 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-operators-5zh2c" Dec 13 08:22:22 crc kubenswrapper[4971]: I1213 08:22:22.223166 4971 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-5zh2c"] Dec 13 08:22:23 crc kubenswrapper[4971]: I1213 08:22:23.226838 4971 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-operators-5zh2c" podUID="7fabfced-a2af-4a4c-b3e1-03461dc4ce53" containerName="registry-server" containerID="cri-o://73f44405eef2fedf0a637d913c0b17ef3c0295bd223b91d6bfcf534da9900440" gracePeriod=2 Dec 13 08:22:24 crc kubenswrapper[4971]: I1213 08:22:24.243408 4971 generic.go:334] "Generic (PLEG): container finished" podID="7fabfced-a2af-4a4c-b3e1-03461dc4ce53" containerID="73f44405eef2fedf0a637d913c0b17ef3c0295bd223b91d6bfcf534da9900440" exitCode=0 Dec 13 08:22:24 crc kubenswrapper[4971]: I1213 08:22:24.243484 4971 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-5zh2c" event={"ID":"7fabfced-a2af-4a4c-b3e1-03461dc4ce53","Type":"ContainerDied","Data":"73f44405eef2fedf0a637d913c0b17ef3c0295bd223b91d6bfcf534da9900440"} Dec 13 08:22:24 crc kubenswrapper[4971]: I1213 08:22:24.996274 4971 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-5zh2c" Dec 13 08:22:25 crc kubenswrapper[4971]: I1213 08:22:25.046394 4971 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/7fabfced-a2af-4a4c-b3e1-03461dc4ce53-utilities\") pod \"7fabfced-a2af-4a4c-b3e1-03461dc4ce53\" (UID: \"7fabfced-a2af-4a4c-b3e1-03461dc4ce53\") " Dec 13 08:22:25 crc kubenswrapper[4971]: I1213 08:22:25.046448 4971 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-k7gqx\" (UniqueName: \"kubernetes.io/projected/7fabfced-a2af-4a4c-b3e1-03461dc4ce53-kube-api-access-k7gqx\") pod \"7fabfced-a2af-4a4c-b3e1-03461dc4ce53\" (UID: \"7fabfced-a2af-4a4c-b3e1-03461dc4ce53\") " Dec 13 08:22:25 crc kubenswrapper[4971]: I1213 08:22:25.046621 4971 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/7fabfced-a2af-4a4c-b3e1-03461dc4ce53-catalog-content\") pod \"7fabfced-a2af-4a4c-b3e1-03461dc4ce53\" (UID: \"7fabfced-a2af-4a4c-b3e1-03461dc4ce53\") " Dec 13 08:22:25 crc kubenswrapper[4971]: I1213 08:22:25.047748 4971 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/7fabfced-a2af-4a4c-b3e1-03461dc4ce53-utilities" (OuterVolumeSpecName: "utilities") pod "7fabfced-a2af-4a4c-b3e1-03461dc4ce53" (UID: "7fabfced-a2af-4a4c-b3e1-03461dc4ce53"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 13 08:22:25 crc kubenswrapper[4971]: I1213 08:22:25.054034 4971 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/7fabfced-a2af-4a4c-b3e1-03461dc4ce53-kube-api-access-k7gqx" (OuterVolumeSpecName: "kube-api-access-k7gqx") pod "7fabfced-a2af-4a4c-b3e1-03461dc4ce53" (UID: "7fabfced-a2af-4a4c-b3e1-03461dc4ce53"). InnerVolumeSpecName "kube-api-access-k7gqx". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 13 08:22:25 crc kubenswrapper[4971]: I1213 08:22:25.149317 4971 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/7fabfced-a2af-4a4c-b3e1-03461dc4ce53-utilities\") on node \"crc\" DevicePath \"\"" Dec 13 08:22:25 crc kubenswrapper[4971]: I1213 08:22:25.149368 4971 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-k7gqx\" (UniqueName: \"kubernetes.io/projected/7fabfced-a2af-4a4c-b3e1-03461dc4ce53-kube-api-access-k7gqx\") on node \"crc\" DevicePath \"\"" Dec 13 08:22:25 crc kubenswrapper[4971]: I1213 08:22:25.182032 4971 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/7fabfced-a2af-4a4c-b3e1-03461dc4ce53-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "7fabfced-a2af-4a4c-b3e1-03461dc4ce53" (UID: "7fabfced-a2af-4a4c-b3e1-03461dc4ce53"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 13 08:22:25 crc kubenswrapper[4971]: I1213 08:22:25.252212 4971 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/7fabfced-a2af-4a4c-b3e1-03461dc4ce53-catalog-content\") on node \"crc\" DevicePath \"\"" Dec 13 08:22:25 crc kubenswrapper[4971]: I1213 08:22:25.254318 4971 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-5zh2c" event={"ID":"7fabfced-a2af-4a4c-b3e1-03461dc4ce53","Type":"ContainerDied","Data":"339e83640ceb75328c717a374d1bb53c87245d152d8dcaf9c47f5c73095cdce8"} Dec 13 08:22:25 crc kubenswrapper[4971]: I1213 08:22:25.254372 4971 scope.go:117] "RemoveContainer" containerID="73f44405eef2fedf0a637d913c0b17ef3c0295bd223b91d6bfcf534da9900440" Dec 13 08:22:25 crc kubenswrapper[4971]: I1213 08:22:25.254404 4971 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-5zh2c" Dec 13 08:22:25 crc kubenswrapper[4971]: I1213 08:22:25.275926 4971 scope.go:117] "RemoveContainer" containerID="0a36d72ca16ce5be7880c2cb8e190ad8986ebb8b78e7a11aebc205d6cf1ac6a3" Dec 13 08:22:25 crc kubenswrapper[4971]: I1213 08:22:25.286148 4971 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-5zh2c"] Dec 13 08:22:25 crc kubenswrapper[4971]: I1213 08:22:25.294444 4971 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-operators-5zh2c"] Dec 13 08:22:25 crc kubenswrapper[4971]: I1213 08:22:25.313425 4971 scope.go:117] "RemoveContainer" containerID="2b96f3ee27d9c0fb5e3f165bd3edb2714b46416277d56215775cb2cb57be96af" Dec 13 08:22:25 crc kubenswrapper[4971]: I1213 08:22:25.780892 4971 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="7fabfced-a2af-4a4c-b3e1-03461dc4ce53" path="/var/lib/kubelet/pods/7fabfced-a2af-4a4c-b3e1-03461dc4ce53/volumes" Dec 13 08:23:03 crc kubenswrapper[4971]: I1213 08:23:03.622195 4971 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-marketplace-lrjvn"] Dec 13 08:23:03 crc kubenswrapper[4971]: E1213 08:23:03.623667 4971 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="7fabfced-a2af-4a4c-b3e1-03461dc4ce53" containerName="extract-content" Dec 13 08:23:03 crc kubenswrapper[4971]: I1213 08:23:03.623690 4971 state_mem.go:107] "Deleted CPUSet assignment" podUID="7fabfced-a2af-4a4c-b3e1-03461dc4ce53" containerName="extract-content" Dec 13 08:23:03 crc kubenswrapper[4971]: E1213 08:23:03.623715 4971 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="7fabfced-a2af-4a4c-b3e1-03461dc4ce53" containerName="registry-server" Dec 13 08:23:03 crc kubenswrapper[4971]: I1213 08:23:03.623726 4971 state_mem.go:107] "Deleted CPUSet assignment" podUID="7fabfced-a2af-4a4c-b3e1-03461dc4ce53" containerName="registry-server" Dec 13 08:23:03 crc kubenswrapper[4971]: E1213 08:23:03.623782 4971 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="7fabfced-a2af-4a4c-b3e1-03461dc4ce53" containerName="extract-utilities" Dec 13 08:23:03 crc kubenswrapper[4971]: I1213 08:23:03.623796 4971 state_mem.go:107] "Deleted CPUSet assignment" podUID="7fabfced-a2af-4a4c-b3e1-03461dc4ce53" containerName="extract-utilities" Dec 13 08:23:03 crc kubenswrapper[4971]: I1213 08:23:03.624140 4971 memory_manager.go:354] "RemoveStaleState removing state" podUID="7fabfced-a2af-4a4c-b3e1-03461dc4ce53" containerName="registry-server" Dec 13 08:23:03 crc kubenswrapper[4971]: I1213 08:23:03.626435 4971 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-lrjvn" Dec 13 08:23:03 crc kubenswrapper[4971]: I1213 08:23:03.643070 4971 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-lrjvn"] Dec 13 08:23:03 crc kubenswrapper[4971]: I1213 08:23:03.740717 4971 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/ec383dc7-923c-4d8e-b714-4dbd5b3b9dc7-utilities\") pod \"redhat-marketplace-lrjvn\" (UID: \"ec383dc7-923c-4d8e-b714-4dbd5b3b9dc7\") " pod="openshift-marketplace/redhat-marketplace-lrjvn" Dec 13 08:23:03 crc kubenswrapper[4971]: I1213 08:23:03.741049 4971 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/ec383dc7-923c-4d8e-b714-4dbd5b3b9dc7-catalog-content\") pod \"redhat-marketplace-lrjvn\" (UID: \"ec383dc7-923c-4d8e-b714-4dbd5b3b9dc7\") " pod="openshift-marketplace/redhat-marketplace-lrjvn" Dec 13 08:23:03 crc kubenswrapper[4971]: I1213 08:23:03.741076 4971 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-vsrnd\" (UniqueName: \"kubernetes.io/projected/ec383dc7-923c-4d8e-b714-4dbd5b3b9dc7-kube-api-access-vsrnd\") pod \"redhat-marketplace-lrjvn\" (UID: \"ec383dc7-923c-4d8e-b714-4dbd5b3b9dc7\") " pod="openshift-marketplace/redhat-marketplace-lrjvn" Dec 13 08:23:03 crc kubenswrapper[4971]: I1213 08:23:03.843387 4971 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/ec383dc7-923c-4d8e-b714-4dbd5b3b9dc7-utilities\") pod \"redhat-marketplace-lrjvn\" (UID: \"ec383dc7-923c-4d8e-b714-4dbd5b3b9dc7\") " pod="openshift-marketplace/redhat-marketplace-lrjvn" Dec 13 08:23:03 crc kubenswrapper[4971]: I1213 08:23:03.843445 4971 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/ec383dc7-923c-4d8e-b714-4dbd5b3b9dc7-catalog-content\") pod \"redhat-marketplace-lrjvn\" (UID: \"ec383dc7-923c-4d8e-b714-4dbd5b3b9dc7\") " pod="openshift-marketplace/redhat-marketplace-lrjvn" Dec 13 08:23:03 crc kubenswrapper[4971]: I1213 08:23:03.843468 4971 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-vsrnd\" (UniqueName: \"kubernetes.io/projected/ec383dc7-923c-4d8e-b714-4dbd5b3b9dc7-kube-api-access-vsrnd\") pod \"redhat-marketplace-lrjvn\" (UID: \"ec383dc7-923c-4d8e-b714-4dbd5b3b9dc7\") " pod="openshift-marketplace/redhat-marketplace-lrjvn" Dec 13 08:23:03 crc kubenswrapper[4971]: I1213 08:23:03.845506 4971 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/ec383dc7-923c-4d8e-b714-4dbd5b3b9dc7-catalog-content\") pod \"redhat-marketplace-lrjvn\" (UID: \"ec383dc7-923c-4d8e-b714-4dbd5b3b9dc7\") " pod="openshift-marketplace/redhat-marketplace-lrjvn" Dec 13 08:23:03 crc kubenswrapper[4971]: I1213 08:23:03.845818 4971 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/ec383dc7-923c-4d8e-b714-4dbd5b3b9dc7-utilities\") pod \"redhat-marketplace-lrjvn\" (UID: \"ec383dc7-923c-4d8e-b714-4dbd5b3b9dc7\") " pod="openshift-marketplace/redhat-marketplace-lrjvn" Dec 13 08:23:03 crc kubenswrapper[4971]: I1213 08:23:03.872510 4971 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-vsrnd\" (UniqueName: \"kubernetes.io/projected/ec383dc7-923c-4d8e-b714-4dbd5b3b9dc7-kube-api-access-vsrnd\") pod \"redhat-marketplace-lrjvn\" (UID: \"ec383dc7-923c-4d8e-b714-4dbd5b3b9dc7\") " pod="openshift-marketplace/redhat-marketplace-lrjvn" Dec 13 08:23:03 crc kubenswrapper[4971]: I1213 08:23:03.969118 4971 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-lrjvn" Dec 13 08:23:04 crc kubenswrapper[4971]: I1213 08:23:04.508243 4971 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-lrjvn"] Dec 13 08:23:04 crc kubenswrapper[4971]: I1213 08:23:04.658822 4971 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-lrjvn" event={"ID":"ec383dc7-923c-4d8e-b714-4dbd5b3b9dc7","Type":"ContainerStarted","Data":"20874a677ddcc39f39eecc407d66f388c5e5cdc23e83e5e07c7398cf133cc84d"} Dec 13 08:23:05 crc kubenswrapper[4971]: I1213 08:23:05.670163 4971 generic.go:334] "Generic (PLEG): container finished" podID="ec383dc7-923c-4d8e-b714-4dbd5b3b9dc7" containerID="988ff3a97994a4ba54bc92a2d4be5852394d5775bae1876e6962eb65ecbd09a6" exitCode=0 Dec 13 08:23:05 crc kubenswrapper[4971]: I1213 08:23:05.670214 4971 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-lrjvn" event={"ID":"ec383dc7-923c-4d8e-b714-4dbd5b3b9dc7","Type":"ContainerDied","Data":"988ff3a97994a4ba54bc92a2d4be5852394d5775bae1876e6962eb65ecbd09a6"} Dec 13 08:23:06 crc kubenswrapper[4971]: I1213 08:23:06.687445 4971 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-lrjvn" event={"ID":"ec383dc7-923c-4d8e-b714-4dbd5b3b9dc7","Type":"ContainerStarted","Data":"30b2bf4251a4a6a611d97033e66cf4f81d195b6fc3f2381ea914bec9b0df1802"} Dec 13 08:23:07 crc kubenswrapper[4971]: I1213 08:23:07.697862 4971 generic.go:334] "Generic (PLEG): container finished" podID="ec383dc7-923c-4d8e-b714-4dbd5b3b9dc7" containerID="30b2bf4251a4a6a611d97033e66cf4f81d195b6fc3f2381ea914bec9b0df1802" exitCode=0 Dec 13 08:23:07 crc kubenswrapper[4971]: I1213 08:23:07.698022 4971 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-lrjvn" event={"ID":"ec383dc7-923c-4d8e-b714-4dbd5b3b9dc7","Type":"ContainerDied","Data":"30b2bf4251a4a6a611d97033e66cf4f81d195b6fc3f2381ea914bec9b0df1802"} Dec 13 08:23:08 crc kubenswrapper[4971]: I1213 08:23:08.709020 4971 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-lrjvn" event={"ID":"ec383dc7-923c-4d8e-b714-4dbd5b3b9dc7","Type":"ContainerStarted","Data":"0fae16d9179581f26f7a5aa53ed10489157fc40f087f4d00609dbbabd4bbf745"} Dec 13 08:23:08 crc kubenswrapper[4971]: I1213 08:23:08.729401 4971 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-marketplace-lrjvn" podStartSLOduration=3.268262553 podStartE2EDuration="5.729383771s" podCreationTimestamp="2025-12-13 08:23:03 +0000 UTC" firstStartedPulling="2025-12-13 08:23:05.672992303 +0000 UTC m=+5642.277401741" lastFinishedPulling="2025-12-13 08:23:08.134113501 +0000 UTC m=+5644.738522959" observedRunningTime="2025-12-13 08:23:08.725510016 +0000 UTC m=+5645.329919464" watchObservedRunningTime="2025-12-13 08:23:08.729383771 +0000 UTC m=+5645.333793219" Dec 13 08:23:13 crc kubenswrapper[4971]: I1213 08:23:13.969920 4971 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-marketplace-lrjvn" Dec 13 08:23:13 crc kubenswrapper[4971]: I1213 08:23:13.970509 4971 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-marketplace-lrjvn" Dec 13 08:23:14 crc kubenswrapper[4971]: I1213 08:23:14.033194 4971 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-marketplace-lrjvn" Dec 13 08:23:14 crc kubenswrapper[4971]: I1213 08:23:14.831937 4971 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-marketplace-lrjvn" Dec 13 08:23:14 crc kubenswrapper[4971]: I1213 08:23:14.894087 4971 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-lrjvn"] Dec 13 08:23:16 crc kubenswrapper[4971]: I1213 08:23:16.797036 4971 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-marketplace-lrjvn" podUID="ec383dc7-923c-4d8e-b714-4dbd5b3b9dc7" containerName="registry-server" containerID="cri-o://0fae16d9179581f26f7a5aa53ed10489157fc40f087f4d00609dbbabd4bbf745" gracePeriod=2 Dec 13 08:23:17 crc kubenswrapper[4971]: I1213 08:23:17.808394 4971 generic.go:334] "Generic (PLEG): container finished" podID="ec383dc7-923c-4d8e-b714-4dbd5b3b9dc7" containerID="0fae16d9179581f26f7a5aa53ed10489157fc40f087f4d00609dbbabd4bbf745" exitCode=0 Dec 13 08:23:17 crc kubenswrapper[4971]: I1213 08:23:17.808652 4971 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-lrjvn" event={"ID":"ec383dc7-923c-4d8e-b714-4dbd5b3b9dc7","Type":"ContainerDied","Data":"0fae16d9179581f26f7a5aa53ed10489157fc40f087f4d00609dbbabd4bbf745"} Dec 13 08:23:18 crc kubenswrapper[4971]: I1213 08:23:18.143020 4971 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-lrjvn" Dec 13 08:23:18 crc kubenswrapper[4971]: I1213 08:23:18.204773 4971 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/ec383dc7-923c-4d8e-b714-4dbd5b3b9dc7-utilities\") pod \"ec383dc7-923c-4d8e-b714-4dbd5b3b9dc7\" (UID: \"ec383dc7-923c-4d8e-b714-4dbd5b3b9dc7\") " Dec 13 08:23:18 crc kubenswrapper[4971]: I1213 08:23:18.204899 4971 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-vsrnd\" (UniqueName: \"kubernetes.io/projected/ec383dc7-923c-4d8e-b714-4dbd5b3b9dc7-kube-api-access-vsrnd\") pod \"ec383dc7-923c-4d8e-b714-4dbd5b3b9dc7\" (UID: \"ec383dc7-923c-4d8e-b714-4dbd5b3b9dc7\") " Dec 13 08:23:18 crc kubenswrapper[4971]: I1213 08:23:18.204934 4971 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/ec383dc7-923c-4d8e-b714-4dbd5b3b9dc7-catalog-content\") pod \"ec383dc7-923c-4d8e-b714-4dbd5b3b9dc7\" (UID: \"ec383dc7-923c-4d8e-b714-4dbd5b3b9dc7\") " Dec 13 08:23:18 crc kubenswrapper[4971]: I1213 08:23:18.206631 4971 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/ec383dc7-923c-4d8e-b714-4dbd5b3b9dc7-utilities" (OuterVolumeSpecName: "utilities") pod "ec383dc7-923c-4d8e-b714-4dbd5b3b9dc7" (UID: "ec383dc7-923c-4d8e-b714-4dbd5b3b9dc7"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 13 08:23:18 crc kubenswrapper[4971]: I1213 08:23:18.215759 4971 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/ec383dc7-923c-4d8e-b714-4dbd5b3b9dc7-kube-api-access-vsrnd" (OuterVolumeSpecName: "kube-api-access-vsrnd") pod "ec383dc7-923c-4d8e-b714-4dbd5b3b9dc7" (UID: "ec383dc7-923c-4d8e-b714-4dbd5b3b9dc7"). InnerVolumeSpecName "kube-api-access-vsrnd". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 13 08:23:18 crc kubenswrapper[4971]: I1213 08:23:18.232822 4971 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/ec383dc7-923c-4d8e-b714-4dbd5b3b9dc7-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "ec383dc7-923c-4d8e-b714-4dbd5b3b9dc7" (UID: "ec383dc7-923c-4d8e-b714-4dbd5b3b9dc7"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 13 08:23:18 crc kubenswrapper[4971]: I1213 08:23:18.307300 4971 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/ec383dc7-923c-4d8e-b714-4dbd5b3b9dc7-utilities\") on node \"crc\" DevicePath \"\"" Dec 13 08:23:18 crc kubenswrapper[4971]: I1213 08:23:18.307343 4971 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-vsrnd\" (UniqueName: \"kubernetes.io/projected/ec383dc7-923c-4d8e-b714-4dbd5b3b9dc7-kube-api-access-vsrnd\") on node \"crc\" DevicePath \"\"" Dec 13 08:23:18 crc kubenswrapper[4971]: I1213 08:23:18.307357 4971 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/ec383dc7-923c-4d8e-b714-4dbd5b3b9dc7-catalog-content\") on node \"crc\" DevicePath \"\"" Dec 13 08:23:18 crc kubenswrapper[4971]: I1213 08:23:18.822589 4971 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-lrjvn" event={"ID":"ec383dc7-923c-4d8e-b714-4dbd5b3b9dc7","Type":"ContainerDied","Data":"20874a677ddcc39f39eecc407d66f388c5e5cdc23e83e5e07c7398cf133cc84d"} Dec 13 08:23:18 crc kubenswrapper[4971]: I1213 08:23:18.822868 4971 scope.go:117] "RemoveContainer" containerID="0fae16d9179581f26f7a5aa53ed10489157fc40f087f4d00609dbbabd4bbf745" Dec 13 08:23:18 crc kubenswrapper[4971]: I1213 08:23:18.823011 4971 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-lrjvn" Dec 13 08:23:19 crc kubenswrapper[4971]: I1213 08:23:19.042445 4971 scope.go:117] "RemoveContainer" containerID="30b2bf4251a4a6a611d97033e66cf4f81d195b6fc3f2381ea914bec9b0df1802" Dec 13 08:23:19 crc kubenswrapper[4971]: I1213 08:23:19.083123 4971 scope.go:117] "RemoveContainer" containerID="988ff3a97994a4ba54bc92a2d4be5852394d5775bae1876e6962eb65ecbd09a6" Dec 13 08:23:19 crc kubenswrapper[4971]: I1213 08:23:19.151589 4971 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-lrjvn"] Dec 13 08:23:19 crc kubenswrapper[4971]: I1213 08:23:19.161102 4971 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-marketplace-lrjvn"] Dec 13 08:23:19 crc kubenswrapper[4971]: I1213 08:23:19.781725 4971 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="ec383dc7-923c-4d8e-b714-4dbd5b3b9dc7" path="/var/lib/kubelet/pods/ec383dc7-923c-4d8e-b714-4dbd5b3b9dc7/volumes" Dec 13 08:24:16 crc kubenswrapper[4971]: I1213 08:24:16.153376 4971 patch_prober.go:28] interesting pod/machine-config-daemon-82xjz container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 13 08:24:16 crc kubenswrapper[4971]: I1213 08:24:16.153949 4971 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-82xjz" podUID="e0c4af12-edda-41d7-9b44-f87396174ef2" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 13 08:24:33 crc kubenswrapper[4971]: I1213 08:24:33.002979 4971 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/community-operators-gz5gh"] Dec 13 08:24:33 crc kubenswrapper[4971]: E1213 08:24:33.003994 4971 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="ec383dc7-923c-4d8e-b714-4dbd5b3b9dc7" containerName="extract-content" Dec 13 08:24:33 crc kubenswrapper[4971]: I1213 08:24:33.004013 4971 state_mem.go:107] "Deleted CPUSet assignment" podUID="ec383dc7-923c-4d8e-b714-4dbd5b3b9dc7" containerName="extract-content" Dec 13 08:24:33 crc kubenswrapper[4971]: E1213 08:24:33.004031 4971 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="ec383dc7-923c-4d8e-b714-4dbd5b3b9dc7" containerName="extract-utilities" Dec 13 08:24:33 crc kubenswrapper[4971]: I1213 08:24:33.004050 4971 state_mem.go:107] "Deleted CPUSet assignment" podUID="ec383dc7-923c-4d8e-b714-4dbd5b3b9dc7" containerName="extract-utilities" Dec 13 08:24:33 crc kubenswrapper[4971]: E1213 08:24:33.004063 4971 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="ec383dc7-923c-4d8e-b714-4dbd5b3b9dc7" containerName="registry-server" Dec 13 08:24:33 crc kubenswrapper[4971]: I1213 08:24:33.004070 4971 state_mem.go:107] "Deleted CPUSet assignment" podUID="ec383dc7-923c-4d8e-b714-4dbd5b3b9dc7" containerName="registry-server" Dec 13 08:24:33 crc kubenswrapper[4971]: I1213 08:24:33.004314 4971 memory_manager.go:354] "RemoveStaleState removing state" podUID="ec383dc7-923c-4d8e-b714-4dbd5b3b9dc7" containerName="registry-server" Dec 13 08:24:33 crc kubenswrapper[4971]: I1213 08:24:33.007414 4971 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-gz5gh" Dec 13 08:24:33 crc kubenswrapper[4971]: I1213 08:24:33.020975 4971 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-gz5gh"] Dec 13 08:24:33 crc kubenswrapper[4971]: I1213 08:24:33.041259 4971 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/f8228e77-b9fc-436f-a438-33e7374f00f1-utilities\") pod \"community-operators-gz5gh\" (UID: \"f8228e77-b9fc-436f-a438-33e7374f00f1\") " pod="openshift-marketplace/community-operators-gz5gh" Dec 13 08:24:33 crc kubenswrapper[4971]: I1213 08:24:33.041356 4971 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-9sk9n\" (UniqueName: \"kubernetes.io/projected/f8228e77-b9fc-436f-a438-33e7374f00f1-kube-api-access-9sk9n\") pod \"community-operators-gz5gh\" (UID: \"f8228e77-b9fc-436f-a438-33e7374f00f1\") " pod="openshift-marketplace/community-operators-gz5gh" Dec 13 08:24:33 crc kubenswrapper[4971]: I1213 08:24:33.041536 4971 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/f8228e77-b9fc-436f-a438-33e7374f00f1-catalog-content\") pod \"community-operators-gz5gh\" (UID: \"f8228e77-b9fc-436f-a438-33e7374f00f1\") " pod="openshift-marketplace/community-operators-gz5gh" Dec 13 08:24:33 crc kubenswrapper[4971]: I1213 08:24:33.143317 4971 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/f8228e77-b9fc-436f-a438-33e7374f00f1-utilities\") pod \"community-operators-gz5gh\" (UID: \"f8228e77-b9fc-436f-a438-33e7374f00f1\") " pod="openshift-marketplace/community-operators-gz5gh" Dec 13 08:24:33 crc kubenswrapper[4971]: I1213 08:24:33.143411 4971 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-9sk9n\" (UniqueName: \"kubernetes.io/projected/f8228e77-b9fc-436f-a438-33e7374f00f1-kube-api-access-9sk9n\") pod \"community-operators-gz5gh\" (UID: \"f8228e77-b9fc-436f-a438-33e7374f00f1\") " pod="openshift-marketplace/community-operators-gz5gh" Dec 13 08:24:33 crc kubenswrapper[4971]: I1213 08:24:33.143605 4971 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/f8228e77-b9fc-436f-a438-33e7374f00f1-catalog-content\") pod \"community-operators-gz5gh\" (UID: \"f8228e77-b9fc-436f-a438-33e7374f00f1\") " pod="openshift-marketplace/community-operators-gz5gh" Dec 13 08:24:33 crc kubenswrapper[4971]: I1213 08:24:33.143981 4971 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/f8228e77-b9fc-436f-a438-33e7374f00f1-utilities\") pod \"community-operators-gz5gh\" (UID: \"f8228e77-b9fc-436f-a438-33e7374f00f1\") " pod="openshift-marketplace/community-operators-gz5gh" Dec 13 08:24:33 crc kubenswrapper[4971]: I1213 08:24:33.144143 4971 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/f8228e77-b9fc-436f-a438-33e7374f00f1-catalog-content\") pod \"community-operators-gz5gh\" (UID: \"f8228e77-b9fc-436f-a438-33e7374f00f1\") " pod="openshift-marketplace/community-operators-gz5gh" Dec 13 08:24:33 crc kubenswrapper[4971]: I1213 08:24:33.165444 4971 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-9sk9n\" (UniqueName: \"kubernetes.io/projected/f8228e77-b9fc-436f-a438-33e7374f00f1-kube-api-access-9sk9n\") pod \"community-operators-gz5gh\" (UID: \"f8228e77-b9fc-436f-a438-33e7374f00f1\") " pod="openshift-marketplace/community-operators-gz5gh" Dec 13 08:24:33 crc kubenswrapper[4971]: I1213 08:24:33.329855 4971 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-gz5gh" Dec 13 08:24:33 crc kubenswrapper[4971]: I1213 08:24:33.881377 4971 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-gz5gh"] Dec 13 08:24:34 crc kubenswrapper[4971]: I1213 08:24:34.167666 4971 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-gz5gh" event={"ID":"f8228e77-b9fc-436f-a438-33e7374f00f1","Type":"ContainerStarted","Data":"4d330f5c528b0eaee333655d5421ba0c08b92b93be77bf31ea53801afb77470d"} Dec 13 08:24:35 crc kubenswrapper[4971]: I1213 08:24:35.181298 4971 generic.go:334] "Generic (PLEG): container finished" podID="f8228e77-b9fc-436f-a438-33e7374f00f1" containerID="4f0934526f9489d7822440b4918c9f817f6315166837825ff2023a05e2f99c7a" exitCode=0 Dec 13 08:24:35 crc kubenswrapper[4971]: I1213 08:24:35.181368 4971 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-gz5gh" event={"ID":"f8228e77-b9fc-436f-a438-33e7374f00f1","Type":"ContainerDied","Data":"4f0934526f9489d7822440b4918c9f817f6315166837825ff2023a05e2f99c7a"} Dec 13 08:24:37 crc kubenswrapper[4971]: I1213 08:24:37.197987 4971 generic.go:334] "Generic (PLEG): container finished" podID="f8228e77-b9fc-436f-a438-33e7374f00f1" containerID="f36215752f2e5b5605056c30310050a37abde42018c2a914675c7d8a58c322d4" exitCode=0 Dec 13 08:24:37 crc kubenswrapper[4971]: I1213 08:24:37.198069 4971 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-gz5gh" event={"ID":"f8228e77-b9fc-436f-a438-33e7374f00f1","Type":"ContainerDied","Data":"f36215752f2e5b5605056c30310050a37abde42018c2a914675c7d8a58c322d4"} Dec 13 08:24:39 crc kubenswrapper[4971]: I1213 08:24:39.218707 4971 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-gz5gh" event={"ID":"f8228e77-b9fc-436f-a438-33e7374f00f1","Type":"ContainerStarted","Data":"34e77684bac41a3b34e543697cad3109b67391a97f3101163c9eea15a2d89200"} Dec 13 08:24:39 crc kubenswrapper[4971]: I1213 08:24:39.246782 4971 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/community-operators-gz5gh" podStartSLOduration=4.525636551 podStartE2EDuration="7.246763424s" podCreationTimestamp="2025-12-13 08:24:32 +0000 UTC" firstStartedPulling="2025-12-13 08:24:35.184431082 +0000 UTC m=+5731.788840530" lastFinishedPulling="2025-12-13 08:24:37.905557955 +0000 UTC m=+5734.509967403" observedRunningTime="2025-12-13 08:24:39.238025289 +0000 UTC m=+5735.842434737" watchObservedRunningTime="2025-12-13 08:24:39.246763424 +0000 UTC m=+5735.851172872" Dec 13 08:24:43 crc kubenswrapper[4971]: I1213 08:24:43.330156 4971 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/community-operators-gz5gh" Dec 13 08:24:43 crc kubenswrapper[4971]: I1213 08:24:43.330566 4971 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/community-operators-gz5gh" Dec 13 08:24:43 crc kubenswrapper[4971]: I1213 08:24:43.382787 4971 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/community-operators-gz5gh" Dec 13 08:24:44 crc kubenswrapper[4971]: I1213 08:24:44.312822 4971 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/community-operators-gz5gh" Dec 13 08:24:44 crc kubenswrapper[4971]: I1213 08:24:44.365243 4971 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-gz5gh"] Dec 13 08:24:46 crc kubenswrapper[4971]: I1213 08:24:46.153815 4971 patch_prober.go:28] interesting pod/machine-config-daemon-82xjz container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 13 08:24:46 crc kubenswrapper[4971]: I1213 08:24:46.154208 4971 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-82xjz" podUID="e0c4af12-edda-41d7-9b44-f87396174ef2" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 13 08:24:46 crc kubenswrapper[4971]: I1213 08:24:46.281969 4971 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/community-operators-gz5gh" podUID="f8228e77-b9fc-436f-a438-33e7374f00f1" containerName="registry-server" containerID="cri-o://34e77684bac41a3b34e543697cad3109b67391a97f3101163c9eea15a2d89200" gracePeriod=2 Dec 13 08:24:47 crc kubenswrapper[4971]: I1213 08:24:47.044903 4971 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-gz5gh" Dec 13 08:24:47 crc kubenswrapper[4971]: I1213 08:24:47.104636 4971 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/f8228e77-b9fc-436f-a438-33e7374f00f1-utilities\") pod \"f8228e77-b9fc-436f-a438-33e7374f00f1\" (UID: \"f8228e77-b9fc-436f-a438-33e7374f00f1\") " Dec 13 08:24:47 crc kubenswrapper[4971]: I1213 08:24:47.104713 4971 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/f8228e77-b9fc-436f-a438-33e7374f00f1-catalog-content\") pod \"f8228e77-b9fc-436f-a438-33e7374f00f1\" (UID: \"f8228e77-b9fc-436f-a438-33e7374f00f1\") " Dec 13 08:24:47 crc kubenswrapper[4971]: I1213 08:24:47.104912 4971 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-9sk9n\" (UniqueName: \"kubernetes.io/projected/f8228e77-b9fc-436f-a438-33e7374f00f1-kube-api-access-9sk9n\") pod \"f8228e77-b9fc-436f-a438-33e7374f00f1\" (UID: \"f8228e77-b9fc-436f-a438-33e7374f00f1\") " Dec 13 08:24:47 crc kubenswrapper[4971]: I1213 08:24:47.105857 4971 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/f8228e77-b9fc-436f-a438-33e7374f00f1-utilities" (OuterVolumeSpecName: "utilities") pod "f8228e77-b9fc-436f-a438-33e7374f00f1" (UID: "f8228e77-b9fc-436f-a438-33e7374f00f1"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 13 08:24:47 crc kubenswrapper[4971]: I1213 08:24:47.114015 4971 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/f8228e77-b9fc-436f-a438-33e7374f00f1-kube-api-access-9sk9n" (OuterVolumeSpecName: "kube-api-access-9sk9n") pod "f8228e77-b9fc-436f-a438-33e7374f00f1" (UID: "f8228e77-b9fc-436f-a438-33e7374f00f1"). InnerVolumeSpecName "kube-api-access-9sk9n". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 13 08:24:47 crc kubenswrapper[4971]: I1213 08:24:47.197296 4971 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/f8228e77-b9fc-436f-a438-33e7374f00f1-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "f8228e77-b9fc-436f-a438-33e7374f00f1" (UID: "f8228e77-b9fc-436f-a438-33e7374f00f1"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 13 08:24:47 crc kubenswrapper[4971]: I1213 08:24:47.207083 4971 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/f8228e77-b9fc-436f-a438-33e7374f00f1-utilities\") on node \"crc\" DevicePath \"\"" Dec 13 08:24:47 crc kubenswrapper[4971]: I1213 08:24:47.207120 4971 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/f8228e77-b9fc-436f-a438-33e7374f00f1-catalog-content\") on node \"crc\" DevicePath \"\"" Dec 13 08:24:47 crc kubenswrapper[4971]: I1213 08:24:47.207134 4971 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-9sk9n\" (UniqueName: \"kubernetes.io/projected/f8228e77-b9fc-436f-a438-33e7374f00f1-kube-api-access-9sk9n\") on node \"crc\" DevicePath \"\"" Dec 13 08:24:47 crc kubenswrapper[4971]: I1213 08:24:47.297592 4971 generic.go:334] "Generic (PLEG): container finished" podID="f8228e77-b9fc-436f-a438-33e7374f00f1" containerID="34e77684bac41a3b34e543697cad3109b67391a97f3101163c9eea15a2d89200" exitCode=0 Dec 13 08:24:47 crc kubenswrapper[4971]: I1213 08:24:47.297644 4971 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-gz5gh" event={"ID":"f8228e77-b9fc-436f-a438-33e7374f00f1","Type":"ContainerDied","Data":"34e77684bac41a3b34e543697cad3109b67391a97f3101163c9eea15a2d89200"} Dec 13 08:24:47 crc kubenswrapper[4971]: I1213 08:24:47.297680 4971 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-gz5gh" event={"ID":"f8228e77-b9fc-436f-a438-33e7374f00f1","Type":"ContainerDied","Data":"4d330f5c528b0eaee333655d5421ba0c08b92b93be77bf31ea53801afb77470d"} Dec 13 08:24:47 crc kubenswrapper[4971]: I1213 08:24:47.297705 4971 scope.go:117] "RemoveContainer" containerID="34e77684bac41a3b34e543697cad3109b67391a97f3101163c9eea15a2d89200" Dec 13 08:24:47 crc kubenswrapper[4971]: I1213 08:24:47.297844 4971 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-gz5gh" Dec 13 08:24:47 crc kubenswrapper[4971]: I1213 08:24:47.338251 4971 scope.go:117] "RemoveContainer" containerID="f36215752f2e5b5605056c30310050a37abde42018c2a914675c7d8a58c322d4" Dec 13 08:24:47 crc kubenswrapper[4971]: I1213 08:24:47.347373 4971 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-gz5gh"] Dec 13 08:24:47 crc kubenswrapper[4971]: I1213 08:24:47.358139 4971 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/community-operators-gz5gh"] Dec 13 08:24:47 crc kubenswrapper[4971]: I1213 08:24:47.367741 4971 scope.go:117] "RemoveContainer" containerID="4f0934526f9489d7822440b4918c9f817f6315166837825ff2023a05e2f99c7a" Dec 13 08:24:47 crc kubenswrapper[4971]: I1213 08:24:47.405176 4971 scope.go:117] "RemoveContainer" containerID="34e77684bac41a3b34e543697cad3109b67391a97f3101163c9eea15a2d89200" Dec 13 08:24:47 crc kubenswrapper[4971]: E1213 08:24:47.406199 4971 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"34e77684bac41a3b34e543697cad3109b67391a97f3101163c9eea15a2d89200\": container with ID starting with 34e77684bac41a3b34e543697cad3109b67391a97f3101163c9eea15a2d89200 not found: ID does not exist" containerID="34e77684bac41a3b34e543697cad3109b67391a97f3101163c9eea15a2d89200" Dec 13 08:24:47 crc kubenswrapper[4971]: I1213 08:24:47.406257 4971 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"34e77684bac41a3b34e543697cad3109b67391a97f3101163c9eea15a2d89200"} err="failed to get container status \"34e77684bac41a3b34e543697cad3109b67391a97f3101163c9eea15a2d89200\": rpc error: code = NotFound desc = could not find container \"34e77684bac41a3b34e543697cad3109b67391a97f3101163c9eea15a2d89200\": container with ID starting with 34e77684bac41a3b34e543697cad3109b67391a97f3101163c9eea15a2d89200 not found: ID does not exist" Dec 13 08:24:47 crc kubenswrapper[4971]: I1213 08:24:47.406292 4971 scope.go:117] "RemoveContainer" containerID="f36215752f2e5b5605056c30310050a37abde42018c2a914675c7d8a58c322d4" Dec 13 08:24:47 crc kubenswrapper[4971]: E1213 08:24:47.406778 4971 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"f36215752f2e5b5605056c30310050a37abde42018c2a914675c7d8a58c322d4\": container with ID starting with f36215752f2e5b5605056c30310050a37abde42018c2a914675c7d8a58c322d4 not found: ID does not exist" containerID="f36215752f2e5b5605056c30310050a37abde42018c2a914675c7d8a58c322d4" Dec 13 08:24:47 crc kubenswrapper[4971]: I1213 08:24:47.406812 4971 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"f36215752f2e5b5605056c30310050a37abde42018c2a914675c7d8a58c322d4"} err="failed to get container status \"f36215752f2e5b5605056c30310050a37abde42018c2a914675c7d8a58c322d4\": rpc error: code = NotFound desc = could not find container \"f36215752f2e5b5605056c30310050a37abde42018c2a914675c7d8a58c322d4\": container with ID starting with f36215752f2e5b5605056c30310050a37abde42018c2a914675c7d8a58c322d4 not found: ID does not exist" Dec 13 08:24:47 crc kubenswrapper[4971]: I1213 08:24:47.406835 4971 scope.go:117] "RemoveContainer" containerID="4f0934526f9489d7822440b4918c9f817f6315166837825ff2023a05e2f99c7a" Dec 13 08:24:47 crc kubenswrapper[4971]: E1213 08:24:47.408240 4971 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"4f0934526f9489d7822440b4918c9f817f6315166837825ff2023a05e2f99c7a\": container with ID starting with 4f0934526f9489d7822440b4918c9f817f6315166837825ff2023a05e2f99c7a not found: ID does not exist" containerID="4f0934526f9489d7822440b4918c9f817f6315166837825ff2023a05e2f99c7a" Dec 13 08:24:47 crc kubenswrapper[4971]: I1213 08:24:47.408300 4971 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"4f0934526f9489d7822440b4918c9f817f6315166837825ff2023a05e2f99c7a"} err="failed to get container status \"4f0934526f9489d7822440b4918c9f817f6315166837825ff2023a05e2f99c7a\": rpc error: code = NotFound desc = could not find container \"4f0934526f9489d7822440b4918c9f817f6315166837825ff2023a05e2f99c7a\": container with ID starting with 4f0934526f9489d7822440b4918c9f817f6315166837825ff2023a05e2f99c7a not found: ID does not exist" Dec 13 08:24:47 crc kubenswrapper[4971]: I1213 08:24:47.783899 4971 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="f8228e77-b9fc-436f-a438-33e7374f00f1" path="/var/lib/kubelet/pods/f8228e77-b9fc-436f-a438-33e7374f00f1/volumes" Dec 13 08:25:16 crc kubenswrapper[4971]: I1213 08:25:16.153746 4971 patch_prober.go:28] interesting pod/machine-config-daemon-82xjz container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 13 08:25:16 crc kubenswrapper[4971]: I1213 08:25:16.155263 4971 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-82xjz" podUID="e0c4af12-edda-41d7-9b44-f87396174ef2" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 13 08:25:16 crc kubenswrapper[4971]: I1213 08:25:16.155311 4971 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-82xjz" Dec 13 08:25:16 crc kubenswrapper[4971]: I1213 08:25:16.156077 4971 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"bc1c06c3bcebf9c00aee813c7819c279356be670c515e1ea302b50b8618bc001"} pod="openshift-machine-config-operator/machine-config-daemon-82xjz" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Dec 13 08:25:16 crc kubenswrapper[4971]: I1213 08:25:16.156162 4971 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-82xjz" podUID="e0c4af12-edda-41d7-9b44-f87396174ef2" containerName="machine-config-daemon" containerID="cri-o://bc1c06c3bcebf9c00aee813c7819c279356be670c515e1ea302b50b8618bc001" gracePeriod=600 Dec 13 08:25:16 crc kubenswrapper[4971]: E1213 08:25:16.419310 4971 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-82xjz_openshift-machine-config-operator(e0c4af12-edda-41d7-9b44-f87396174ef2)\"" pod="openshift-machine-config-operator/machine-config-daemon-82xjz" podUID="e0c4af12-edda-41d7-9b44-f87396174ef2" Dec 13 08:25:16 crc kubenswrapper[4971]: I1213 08:25:16.564673 4971 generic.go:334] "Generic (PLEG): container finished" podID="e0c4af12-edda-41d7-9b44-f87396174ef2" containerID="bc1c06c3bcebf9c00aee813c7819c279356be670c515e1ea302b50b8618bc001" exitCode=0 Dec 13 08:25:16 crc kubenswrapper[4971]: I1213 08:25:16.564733 4971 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-82xjz" event={"ID":"e0c4af12-edda-41d7-9b44-f87396174ef2","Type":"ContainerDied","Data":"bc1c06c3bcebf9c00aee813c7819c279356be670c515e1ea302b50b8618bc001"} Dec 13 08:25:16 crc kubenswrapper[4971]: I1213 08:25:16.564792 4971 scope.go:117] "RemoveContainer" containerID="b2fd564c27978e268006043ad8ebcbcac704f06f77bb3bdbd7878948103f567f" Dec 13 08:25:16 crc kubenswrapper[4971]: I1213 08:25:16.565587 4971 scope.go:117] "RemoveContainer" containerID="bc1c06c3bcebf9c00aee813c7819c279356be670c515e1ea302b50b8618bc001" Dec 13 08:25:16 crc kubenswrapper[4971]: E1213 08:25:16.565976 4971 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-82xjz_openshift-machine-config-operator(e0c4af12-edda-41d7-9b44-f87396174ef2)\"" pod="openshift-machine-config-operator/machine-config-daemon-82xjz" podUID="e0c4af12-edda-41d7-9b44-f87396174ef2" Dec 13 08:25:27 crc kubenswrapper[4971]: I1213 08:25:27.769722 4971 scope.go:117] "RemoveContainer" containerID="bc1c06c3bcebf9c00aee813c7819c279356be670c515e1ea302b50b8618bc001" Dec 13 08:25:27 crc kubenswrapper[4971]: E1213 08:25:27.770935 4971 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-82xjz_openshift-machine-config-operator(e0c4af12-edda-41d7-9b44-f87396174ef2)\"" pod="openshift-machine-config-operator/machine-config-daemon-82xjz" podUID="e0c4af12-edda-41d7-9b44-f87396174ef2" Dec 13 08:25:42 crc kubenswrapper[4971]: I1213 08:25:42.768985 4971 scope.go:117] "RemoveContainer" containerID="bc1c06c3bcebf9c00aee813c7819c279356be670c515e1ea302b50b8618bc001" Dec 13 08:25:42 crc kubenswrapper[4971]: E1213 08:25:42.769702 4971 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-82xjz_openshift-machine-config-operator(e0c4af12-edda-41d7-9b44-f87396174ef2)\"" pod="openshift-machine-config-operator/machine-config-daemon-82xjz" podUID="e0c4af12-edda-41d7-9b44-f87396174ef2" Dec 13 08:25:54 crc kubenswrapper[4971]: I1213 08:25:54.770534 4971 scope.go:117] "RemoveContainer" containerID="bc1c06c3bcebf9c00aee813c7819c279356be670c515e1ea302b50b8618bc001" Dec 13 08:25:54 crc kubenswrapper[4971]: E1213 08:25:54.771336 4971 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-82xjz_openshift-machine-config-operator(e0c4af12-edda-41d7-9b44-f87396174ef2)\"" pod="openshift-machine-config-operator/machine-config-daemon-82xjz" podUID="e0c4af12-edda-41d7-9b44-f87396174ef2" Dec 13 08:26:07 crc kubenswrapper[4971]: I1213 08:26:07.769045 4971 scope.go:117] "RemoveContainer" containerID="bc1c06c3bcebf9c00aee813c7819c279356be670c515e1ea302b50b8618bc001" Dec 13 08:26:07 crc kubenswrapper[4971]: E1213 08:26:07.769909 4971 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-82xjz_openshift-machine-config-operator(e0c4af12-edda-41d7-9b44-f87396174ef2)\"" pod="openshift-machine-config-operator/machine-config-daemon-82xjz" podUID="e0c4af12-edda-41d7-9b44-f87396174ef2" Dec 13 08:26:22 crc kubenswrapper[4971]: I1213 08:26:22.769208 4971 scope.go:117] "RemoveContainer" containerID="bc1c06c3bcebf9c00aee813c7819c279356be670c515e1ea302b50b8618bc001" Dec 13 08:26:22 crc kubenswrapper[4971]: E1213 08:26:22.770707 4971 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-82xjz_openshift-machine-config-operator(e0c4af12-edda-41d7-9b44-f87396174ef2)\"" pod="openshift-machine-config-operator/machine-config-daemon-82xjz" podUID="e0c4af12-edda-41d7-9b44-f87396174ef2" Dec 13 08:26:37 crc kubenswrapper[4971]: I1213 08:26:37.768780 4971 scope.go:117] "RemoveContainer" containerID="bc1c06c3bcebf9c00aee813c7819c279356be670c515e1ea302b50b8618bc001" Dec 13 08:26:37 crc kubenswrapper[4971]: E1213 08:26:37.769935 4971 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-82xjz_openshift-machine-config-operator(e0c4af12-edda-41d7-9b44-f87396174ef2)\"" pod="openshift-machine-config-operator/machine-config-daemon-82xjz" podUID="e0c4af12-edda-41d7-9b44-f87396174ef2" Dec 13 08:26:52 crc kubenswrapper[4971]: I1213 08:26:52.770797 4971 scope.go:117] "RemoveContainer" containerID="bc1c06c3bcebf9c00aee813c7819c279356be670c515e1ea302b50b8618bc001" Dec 13 08:26:52 crc kubenswrapper[4971]: E1213 08:26:52.772449 4971 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-82xjz_openshift-machine-config-operator(e0c4af12-edda-41d7-9b44-f87396174ef2)\"" pod="openshift-machine-config-operator/machine-config-daemon-82xjz" podUID="e0c4af12-edda-41d7-9b44-f87396174ef2" Dec 13 08:27:05 crc kubenswrapper[4971]: I1213 08:27:05.769703 4971 scope.go:117] "RemoveContainer" containerID="bc1c06c3bcebf9c00aee813c7819c279356be670c515e1ea302b50b8618bc001" Dec 13 08:27:05 crc kubenswrapper[4971]: E1213 08:27:05.772987 4971 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-82xjz_openshift-machine-config-operator(e0c4af12-edda-41d7-9b44-f87396174ef2)\"" pod="openshift-machine-config-operator/machine-config-daemon-82xjz" podUID="e0c4af12-edda-41d7-9b44-f87396174ef2" Dec 13 08:27:17 crc kubenswrapper[4971]: I1213 08:27:17.768580 4971 scope.go:117] "RemoveContainer" containerID="bc1c06c3bcebf9c00aee813c7819c279356be670c515e1ea302b50b8618bc001" Dec 13 08:27:17 crc kubenswrapper[4971]: E1213 08:27:17.769347 4971 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-82xjz_openshift-machine-config-operator(e0c4af12-edda-41d7-9b44-f87396174ef2)\"" pod="openshift-machine-config-operator/machine-config-daemon-82xjz" podUID="e0c4af12-edda-41d7-9b44-f87396174ef2" Dec 13 08:27:28 crc kubenswrapper[4971]: I1213 08:27:28.770320 4971 scope.go:117] "RemoveContainer" containerID="bc1c06c3bcebf9c00aee813c7819c279356be670c515e1ea302b50b8618bc001" Dec 13 08:27:28 crc kubenswrapper[4971]: E1213 08:27:28.771758 4971 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-82xjz_openshift-machine-config-operator(e0c4af12-edda-41d7-9b44-f87396174ef2)\"" pod="openshift-machine-config-operator/machine-config-daemon-82xjz" podUID="e0c4af12-edda-41d7-9b44-f87396174ef2" Dec 13 08:27:40 crc kubenswrapper[4971]: I1213 08:27:40.769000 4971 scope.go:117] "RemoveContainer" containerID="bc1c06c3bcebf9c00aee813c7819c279356be670c515e1ea302b50b8618bc001" Dec 13 08:27:40 crc kubenswrapper[4971]: E1213 08:27:40.770221 4971 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-82xjz_openshift-machine-config-operator(e0c4af12-edda-41d7-9b44-f87396174ef2)\"" pod="openshift-machine-config-operator/machine-config-daemon-82xjz" podUID="e0c4af12-edda-41d7-9b44-f87396174ef2" Dec 13 08:27:53 crc kubenswrapper[4971]: I1213 08:27:53.777402 4971 scope.go:117] "RemoveContainer" containerID="bc1c06c3bcebf9c00aee813c7819c279356be670c515e1ea302b50b8618bc001" Dec 13 08:27:53 crc kubenswrapper[4971]: E1213 08:27:53.778774 4971 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-82xjz_openshift-machine-config-operator(e0c4af12-edda-41d7-9b44-f87396174ef2)\"" pod="openshift-machine-config-operator/machine-config-daemon-82xjz" podUID="e0c4af12-edda-41d7-9b44-f87396174ef2" Dec 13 08:27:54 crc kubenswrapper[4971]: I1213 08:27:54.086241 4971 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/certified-operators-9fjp9"] Dec 13 08:27:54 crc kubenswrapper[4971]: E1213 08:27:54.087068 4971 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f8228e77-b9fc-436f-a438-33e7374f00f1" containerName="extract-content" Dec 13 08:27:54 crc kubenswrapper[4971]: I1213 08:27:54.087095 4971 state_mem.go:107] "Deleted CPUSet assignment" podUID="f8228e77-b9fc-436f-a438-33e7374f00f1" containerName="extract-content" Dec 13 08:27:54 crc kubenswrapper[4971]: E1213 08:27:54.087108 4971 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f8228e77-b9fc-436f-a438-33e7374f00f1" containerName="registry-server" Dec 13 08:27:54 crc kubenswrapper[4971]: I1213 08:27:54.087115 4971 state_mem.go:107] "Deleted CPUSet assignment" podUID="f8228e77-b9fc-436f-a438-33e7374f00f1" containerName="registry-server" Dec 13 08:27:54 crc kubenswrapper[4971]: E1213 08:27:54.087129 4971 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f8228e77-b9fc-436f-a438-33e7374f00f1" containerName="extract-utilities" Dec 13 08:27:54 crc kubenswrapper[4971]: I1213 08:27:54.087137 4971 state_mem.go:107] "Deleted CPUSet assignment" podUID="f8228e77-b9fc-436f-a438-33e7374f00f1" containerName="extract-utilities" Dec 13 08:27:54 crc kubenswrapper[4971]: I1213 08:27:54.087420 4971 memory_manager.go:354] "RemoveStaleState removing state" podUID="f8228e77-b9fc-436f-a438-33e7374f00f1" containerName="registry-server" Dec 13 08:27:54 crc kubenswrapper[4971]: I1213 08:27:54.089434 4971 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-9fjp9" Dec 13 08:27:54 crc kubenswrapper[4971]: I1213 08:27:54.105429 4971 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-wl9v7\" (UniqueName: \"kubernetes.io/projected/9ede661a-912c-4701-8c62-aa1bac23c7c6-kube-api-access-wl9v7\") pod \"certified-operators-9fjp9\" (UID: \"9ede661a-912c-4701-8c62-aa1bac23c7c6\") " pod="openshift-marketplace/certified-operators-9fjp9" Dec 13 08:27:54 crc kubenswrapper[4971]: I1213 08:27:54.105680 4971 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/9ede661a-912c-4701-8c62-aa1bac23c7c6-utilities\") pod \"certified-operators-9fjp9\" (UID: \"9ede661a-912c-4701-8c62-aa1bac23c7c6\") " pod="openshift-marketplace/certified-operators-9fjp9" Dec 13 08:27:54 crc kubenswrapper[4971]: I1213 08:27:54.105711 4971 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/9ede661a-912c-4701-8c62-aa1bac23c7c6-catalog-content\") pod \"certified-operators-9fjp9\" (UID: \"9ede661a-912c-4701-8c62-aa1bac23c7c6\") " pod="openshift-marketplace/certified-operators-9fjp9" Dec 13 08:27:54 crc kubenswrapper[4971]: I1213 08:27:54.132850 4971 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-9fjp9"] Dec 13 08:27:54 crc kubenswrapper[4971]: I1213 08:27:54.208707 4971 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/9ede661a-912c-4701-8c62-aa1bac23c7c6-utilities\") pod \"certified-operators-9fjp9\" (UID: \"9ede661a-912c-4701-8c62-aa1bac23c7c6\") " pod="openshift-marketplace/certified-operators-9fjp9" Dec 13 08:27:54 crc kubenswrapper[4971]: I1213 08:27:54.208786 4971 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/9ede661a-912c-4701-8c62-aa1bac23c7c6-catalog-content\") pod \"certified-operators-9fjp9\" (UID: \"9ede661a-912c-4701-8c62-aa1bac23c7c6\") " pod="openshift-marketplace/certified-operators-9fjp9" Dec 13 08:27:54 crc kubenswrapper[4971]: I1213 08:27:54.208840 4971 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-wl9v7\" (UniqueName: \"kubernetes.io/projected/9ede661a-912c-4701-8c62-aa1bac23c7c6-kube-api-access-wl9v7\") pod \"certified-operators-9fjp9\" (UID: \"9ede661a-912c-4701-8c62-aa1bac23c7c6\") " pod="openshift-marketplace/certified-operators-9fjp9" Dec 13 08:27:54 crc kubenswrapper[4971]: I1213 08:27:54.210088 4971 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/9ede661a-912c-4701-8c62-aa1bac23c7c6-catalog-content\") pod \"certified-operators-9fjp9\" (UID: \"9ede661a-912c-4701-8c62-aa1bac23c7c6\") " pod="openshift-marketplace/certified-operators-9fjp9" Dec 13 08:27:54 crc kubenswrapper[4971]: I1213 08:27:54.219569 4971 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/9ede661a-912c-4701-8c62-aa1bac23c7c6-utilities\") pod \"certified-operators-9fjp9\" (UID: \"9ede661a-912c-4701-8c62-aa1bac23c7c6\") " pod="openshift-marketplace/certified-operators-9fjp9" Dec 13 08:27:54 crc kubenswrapper[4971]: I1213 08:27:54.263530 4971 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-wl9v7\" (UniqueName: \"kubernetes.io/projected/9ede661a-912c-4701-8c62-aa1bac23c7c6-kube-api-access-wl9v7\") pod \"certified-operators-9fjp9\" (UID: \"9ede661a-912c-4701-8c62-aa1bac23c7c6\") " pod="openshift-marketplace/certified-operators-9fjp9" Dec 13 08:27:54 crc kubenswrapper[4971]: I1213 08:27:54.503327 4971 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-9fjp9" Dec 13 08:27:55 crc kubenswrapper[4971]: I1213 08:27:55.170775 4971 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-9fjp9"] Dec 13 08:27:56 crc kubenswrapper[4971]: I1213 08:27:56.051450 4971 generic.go:334] "Generic (PLEG): container finished" podID="9ede661a-912c-4701-8c62-aa1bac23c7c6" containerID="4f8adea8465d5f321bddf483086b4ade07f29919777ae960e7f22d9bcfeb61b1" exitCode=0 Dec 13 08:27:56 crc kubenswrapper[4971]: I1213 08:27:56.051585 4971 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-9fjp9" event={"ID":"9ede661a-912c-4701-8c62-aa1bac23c7c6","Type":"ContainerDied","Data":"4f8adea8465d5f321bddf483086b4ade07f29919777ae960e7f22d9bcfeb61b1"} Dec 13 08:27:56 crc kubenswrapper[4971]: I1213 08:27:56.051954 4971 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-9fjp9" event={"ID":"9ede661a-912c-4701-8c62-aa1bac23c7c6","Type":"ContainerStarted","Data":"68806a11edfbf75b67c5c1959ea6e2705e850a78c985104927b4b14d29f698b1"} Dec 13 08:27:56 crc kubenswrapper[4971]: I1213 08:27:56.055172 4971 provider.go:102] Refreshing cache for provider: *credentialprovider.defaultDockerConfigProvider Dec 13 08:27:58 crc kubenswrapper[4971]: I1213 08:27:58.075312 4971 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-9fjp9" event={"ID":"9ede661a-912c-4701-8c62-aa1bac23c7c6","Type":"ContainerStarted","Data":"5cc2cb61655c86e1525924a51d3453184a83078dae78fdcdaaf3b02a96d2db9b"} Dec 13 08:28:00 crc kubenswrapper[4971]: I1213 08:28:00.215884 4971 generic.go:334] "Generic (PLEG): container finished" podID="9ede661a-912c-4701-8c62-aa1bac23c7c6" containerID="5cc2cb61655c86e1525924a51d3453184a83078dae78fdcdaaf3b02a96d2db9b" exitCode=0 Dec 13 08:28:00 crc kubenswrapper[4971]: I1213 08:28:00.215966 4971 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-9fjp9" event={"ID":"9ede661a-912c-4701-8c62-aa1bac23c7c6","Type":"ContainerDied","Data":"5cc2cb61655c86e1525924a51d3453184a83078dae78fdcdaaf3b02a96d2db9b"} Dec 13 08:28:01 crc kubenswrapper[4971]: I1213 08:28:01.231671 4971 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-9fjp9" event={"ID":"9ede661a-912c-4701-8c62-aa1bac23c7c6","Type":"ContainerStarted","Data":"b0a1b112a3d1dee5f1cd1f5b040a3805bb8fc0c0951e9fe13cf24f65f52bc07c"} Dec 13 08:28:01 crc kubenswrapper[4971]: I1213 08:28:01.265412 4971 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/certified-operators-9fjp9" podStartSLOduration=2.459963697 podStartE2EDuration="7.265388725s" podCreationTimestamp="2025-12-13 08:27:54 +0000 UTC" firstStartedPulling="2025-12-13 08:27:56.054915488 +0000 UTC m=+5932.659324936" lastFinishedPulling="2025-12-13 08:28:00.860319915 +0000 UTC m=+5937.464749964" observedRunningTime="2025-12-13 08:28:01.25748486 +0000 UTC m=+5937.861894318" watchObservedRunningTime="2025-12-13 08:28:01.265388725 +0000 UTC m=+5937.869798173" Dec 13 08:28:04 crc kubenswrapper[4971]: I1213 08:28:04.504273 4971 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/certified-operators-9fjp9" Dec 13 08:28:04 crc kubenswrapper[4971]: I1213 08:28:04.505117 4971 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/certified-operators-9fjp9" Dec 13 08:28:04 crc kubenswrapper[4971]: I1213 08:28:04.563114 4971 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/certified-operators-9fjp9" Dec 13 08:28:06 crc kubenswrapper[4971]: I1213 08:28:06.768779 4971 scope.go:117] "RemoveContainer" containerID="bc1c06c3bcebf9c00aee813c7819c279356be670c515e1ea302b50b8618bc001" Dec 13 08:28:06 crc kubenswrapper[4971]: E1213 08:28:06.770096 4971 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-82xjz_openshift-machine-config-operator(e0c4af12-edda-41d7-9b44-f87396174ef2)\"" pod="openshift-machine-config-operator/machine-config-daemon-82xjz" podUID="e0c4af12-edda-41d7-9b44-f87396174ef2" Dec 13 08:28:14 crc kubenswrapper[4971]: I1213 08:28:14.562902 4971 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/certified-operators-9fjp9" Dec 13 08:28:14 crc kubenswrapper[4971]: I1213 08:28:14.641085 4971 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-9fjp9"] Dec 13 08:28:15 crc kubenswrapper[4971]: I1213 08:28:15.468297 4971 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/certified-operators-9fjp9" podUID="9ede661a-912c-4701-8c62-aa1bac23c7c6" containerName="registry-server" containerID="cri-o://b0a1b112a3d1dee5f1cd1f5b040a3805bb8fc0c0951e9fe13cf24f65f52bc07c" gracePeriod=2 Dec 13 08:28:16 crc kubenswrapper[4971]: I1213 08:28:16.247322 4971 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-9fjp9" Dec 13 08:28:16 crc kubenswrapper[4971]: I1213 08:28:16.295040 4971 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/9ede661a-912c-4701-8c62-aa1bac23c7c6-utilities" (OuterVolumeSpecName: "utilities") pod "9ede661a-912c-4701-8c62-aa1bac23c7c6" (UID: "9ede661a-912c-4701-8c62-aa1bac23c7c6"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 13 08:28:16 crc kubenswrapper[4971]: I1213 08:28:16.293889 4971 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/9ede661a-912c-4701-8c62-aa1bac23c7c6-utilities\") pod \"9ede661a-912c-4701-8c62-aa1bac23c7c6\" (UID: \"9ede661a-912c-4701-8c62-aa1bac23c7c6\") " Dec 13 08:28:16 crc kubenswrapper[4971]: I1213 08:28:16.295552 4971 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/9ede661a-912c-4701-8c62-aa1bac23c7c6-catalog-content\") pod \"9ede661a-912c-4701-8c62-aa1bac23c7c6\" (UID: \"9ede661a-912c-4701-8c62-aa1bac23c7c6\") " Dec 13 08:28:16 crc kubenswrapper[4971]: I1213 08:28:16.312479 4971 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-wl9v7\" (UniqueName: \"kubernetes.io/projected/9ede661a-912c-4701-8c62-aa1bac23c7c6-kube-api-access-wl9v7\") pod \"9ede661a-912c-4701-8c62-aa1bac23c7c6\" (UID: \"9ede661a-912c-4701-8c62-aa1bac23c7c6\") " Dec 13 08:28:16 crc kubenswrapper[4971]: I1213 08:28:16.315817 4971 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/9ede661a-912c-4701-8c62-aa1bac23c7c6-utilities\") on node \"crc\" DevicePath \"\"" Dec 13 08:28:16 crc kubenswrapper[4971]: I1213 08:28:16.325904 4971 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/9ede661a-912c-4701-8c62-aa1bac23c7c6-kube-api-access-wl9v7" (OuterVolumeSpecName: "kube-api-access-wl9v7") pod "9ede661a-912c-4701-8c62-aa1bac23c7c6" (UID: "9ede661a-912c-4701-8c62-aa1bac23c7c6"). InnerVolumeSpecName "kube-api-access-wl9v7". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 13 08:28:16 crc kubenswrapper[4971]: I1213 08:28:16.368964 4971 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/9ede661a-912c-4701-8c62-aa1bac23c7c6-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "9ede661a-912c-4701-8c62-aa1bac23c7c6" (UID: "9ede661a-912c-4701-8c62-aa1bac23c7c6"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 13 08:28:16 crc kubenswrapper[4971]: I1213 08:28:16.420931 4971 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/9ede661a-912c-4701-8c62-aa1bac23c7c6-catalog-content\") on node \"crc\" DevicePath \"\"" Dec 13 08:28:16 crc kubenswrapper[4971]: I1213 08:28:16.420978 4971 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-wl9v7\" (UniqueName: \"kubernetes.io/projected/9ede661a-912c-4701-8c62-aa1bac23c7c6-kube-api-access-wl9v7\") on node \"crc\" DevicePath \"\"" Dec 13 08:28:16 crc kubenswrapper[4971]: I1213 08:28:16.481542 4971 generic.go:334] "Generic (PLEG): container finished" podID="9ede661a-912c-4701-8c62-aa1bac23c7c6" containerID="b0a1b112a3d1dee5f1cd1f5b040a3805bb8fc0c0951e9fe13cf24f65f52bc07c" exitCode=0 Dec 13 08:28:16 crc kubenswrapper[4971]: I1213 08:28:16.481653 4971 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-9fjp9" event={"ID":"9ede661a-912c-4701-8c62-aa1bac23c7c6","Type":"ContainerDied","Data":"b0a1b112a3d1dee5f1cd1f5b040a3805bb8fc0c0951e9fe13cf24f65f52bc07c"} Dec 13 08:28:16 crc kubenswrapper[4971]: I1213 08:28:16.482010 4971 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-9fjp9" event={"ID":"9ede661a-912c-4701-8c62-aa1bac23c7c6","Type":"ContainerDied","Data":"68806a11edfbf75b67c5c1959ea6e2705e850a78c985104927b4b14d29f698b1"} Dec 13 08:28:16 crc kubenswrapper[4971]: I1213 08:28:16.482075 4971 scope.go:117] "RemoveContainer" containerID="b0a1b112a3d1dee5f1cd1f5b040a3805bb8fc0c0951e9fe13cf24f65f52bc07c" Dec 13 08:28:16 crc kubenswrapper[4971]: I1213 08:28:16.481695 4971 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-9fjp9" Dec 13 08:28:16 crc kubenswrapper[4971]: I1213 08:28:16.519125 4971 scope.go:117] "RemoveContainer" containerID="5cc2cb61655c86e1525924a51d3453184a83078dae78fdcdaaf3b02a96d2db9b" Dec 13 08:28:16 crc kubenswrapper[4971]: I1213 08:28:16.523318 4971 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-9fjp9"] Dec 13 08:28:16 crc kubenswrapper[4971]: I1213 08:28:16.534039 4971 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/certified-operators-9fjp9"] Dec 13 08:28:16 crc kubenswrapper[4971]: I1213 08:28:16.549177 4971 scope.go:117] "RemoveContainer" containerID="4f8adea8465d5f321bddf483086b4ade07f29919777ae960e7f22d9bcfeb61b1" Dec 13 08:28:16 crc kubenswrapper[4971]: I1213 08:28:16.589129 4971 scope.go:117] "RemoveContainer" containerID="b0a1b112a3d1dee5f1cd1f5b040a3805bb8fc0c0951e9fe13cf24f65f52bc07c" Dec 13 08:28:16 crc kubenswrapper[4971]: E1213 08:28:16.590102 4971 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"b0a1b112a3d1dee5f1cd1f5b040a3805bb8fc0c0951e9fe13cf24f65f52bc07c\": container with ID starting with b0a1b112a3d1dee5f1cd1f5b040a3805bb8fc0c0951e9fe13cf24f65f52bc07c not found: ID does not exist" containerID="b0a1b112a3d1dee5f1cd1f5b040a3805bb8fc0c0951e9fe13cf24f65f52bc07c" Dec 13 08:28:16 crc kubenswrapper[4971]: I1213 08:28:16.590159 4971 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"b0a1b112a3d1dee5f1cd1f5b040a3805bb8fc0c0951e9fe13cf24f65f52bc07c"} err="failed to get container status \"b0a1b112a3d1dee5f1cd1f5b040a3805bb8fc0c0951e9fe13cf24f65f52bc07c\": rpc error: code = NotFound desc = could not find container \"b0a1b112a3d1dee5f1cd1f5b040a3805bb8fc0c0951e9fe13cf24f65f52bc07c\": container with ID starting with b0a1b112a3d1dee5f1cd1f5b040a3805bb8fc0c0951e9fe13cf24f65f52bc07c not found: ID does not exist" Dec 13 08:28:16 crc kubenswrapper[4971]: I1213 08:28:16.590194 4971 scope.go:117] "RemoveContainer" containerID="5cc2cb61655c86e1525924a51d3453184a83078dae78fdcdaaf3b02a96d2db9b" Dec 13 08:28:16 crc kubenswrapper[4971]: E1213 08:28:16.590973 4971 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"5cc2cb61655c86e1525924a51d3453184a83078dae78fdcdaaf3b02a96d2db9b\": container with ID starting with 5cc2cb61655c86e1525924a51d3453184a83078dae78fdcdaaf3b02a96d2db9b not found: ID does not exist" containerID="5cc2cb61655c86e1525924a51d3453184a83078dae78fdcdaaf3b02a96d2db9b" Dec 13 08:28:16 crc kubenswrapper[4971]: I1213 08:28:16.591094 4971 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"5cc2cb61655c86e1525924a51d3453184a83078dae78fdcdaaf3b02a96d2db9b"} err="failed to get container status \"5cc2cb61655c86e1525924a51d3453184a83078dae78fdcdaaf3b02a96d2db9b\": rpc error: code = NotFound desc = could not find container \"5cc2cb61655c86e1525924a51d3453184a83078dae78fdcdaaf3b02a96d2db9b\": container with ID starting with 5cc2cb61655c86e1525924a51d3453184a83078dae78fdcdaaf3b02a96d2db9b not found: ID does not exist" Dec 13 08:28:16 crc kubenswrapper[4971]: I1213 08:28:16.591188 4971 scope.go:117] "RemoveContainer" containerID="4f8adea8465d5f321bddf483086b4ade07f29919777ae960e7f22d9bcfeb61b1" Dec 13 08:28:16 crc kubenswrapper[4971]: E1213 08:28:16.591910 4971 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"4f8adea8465d5f321bddf483086b4ade07f29919777ae960e7f22d9bcfeb61b1\": container with ID starting with 4f8adea8465d5f321bddf483086b4ade07f29919777ae960e7f22d9bcfeb61b1 not found: ID does not exist" containerID="4f8adea8465d5f321bddf483086b4ade07f29919777ae960e7f22d9bcfeb61b1" Dec 13 08:28:16 crc kubenswrapper[4971]: I1213 08:28:16.592012 4971 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"4f8adea8465d5f321bddf483086b4ade07f29919777ae960e7f22d9bcfeb61b1"} err="failed to get container status \"4f8adea8465d5f321bddf483086b4ade07f29919777ae960e7f22d9bcfeb61b1\": rpc error: code = NotFound desc = could not find container \"4f8adea8465d5f321bddf483086b4ade07f29919777ae960e7f22d9bcfeb61b1\": container with ID starting with 4f8adea8465d5f321bddf483086b4ade07f29919777ae960e7f22d9bcfeb61b1 not found: ID does not exist" Dec 13 08:28:17 crc kubenswrapper[4971]: I1213 08:28:17.783944 4971 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="9ede661a-912c-4701-8c62-aa1bac23c7c6" path="/var/lib/kubelet/pods/9ede661a-912c-4701-8c62-aa1bac23c7c6/volumes" Dec 13 08:28:19 crc kubenswrapper[4971]: I1213 08:28:19.768938 4971 scope.go:117] "RemoveContainer" containerID="bc1c06c3bcebf9c00aee813c7819c279356be670c515e1ea302b50b8618bc001" Dec 13 08:28:19 crc kubenswrapper[4971]: E1213 08:28:19.769774 4971 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-82xjz_openshift-machine-config-operator(e0c4af12-edda-41d7-9b44-f87396174ef2)\"" pod="openshift-machine-config-operator/machine-config-daemon-82xjz" podUID="e0c4af12-edda-41d7-9b44-f87396174ef2" Dec 13 08:28:32 crc kubenswrapper[4971]: I1213 08:28:32.769699 4971 scope.go:117] "RemoveContainer" containerID="bc1c06c3bcebf9c00aee813c7819c279356be670c515e1ea302b50b8618bc001" Dec 13 08:28:32 crc kubenswrapper[4971]: E1213 08:28:32.770841 4971 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-82xjz_openshift-machine-config-operator(e0c4af12-edda-41d7-9b44-f87396174ef2)\"" pod="openshift-machine-config-operator/machine-config-daemon-82xjz" podUID="e0c4af12-edda-41d7-9b44-f87396174ef2" Dec 13 08:28:44 crc kubenswrapper[4971]: I1213 08:28:44.769373 4971 scope.go:117] "RemoveContainer" containerID="bc1c06c3bcebf9c00aee813c7819c279356be670c515e1ea302b50b8618bc001" Dec 13 08:28:44 crc kubenswrapper[4971]: E1213 08:28:44.770622 4971 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-82xjz_openshift-machine-config-operator(e0c4af12-edda-41d7-9b44-f87396174ef2)\"" pod="openshift-machine-config-operator/machine-config-daemon-82xjz" podUID="e0c4af12-edda-41d7-9b44-f87396174ef2" Dec 13 08:28:56 crc kubenswrapper[4971]: I1213 08:28:56.769171 4971 scope.go:117] "RemoveContainer" containerID="bc1c06c3bcebf9c00aee813c7819c279356be670c515e1ea302b50b8618bc001" Dec 13 08:28:56 crc kubenswrapper[4971]: E1213 08:28:56.770922 4971 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-82xjz_openshift-machine-config-operator(e0c4af12-edda-41d7-9b44-f87396174ef2)\"" pod="openshift-machine-config-operator/machine-config-daemon-82xjz" podUID="e0c4af12-edda-41d7-9b44-f87396174ef2" Dec 13 08:29:07 crc kubenswrapper[4971]: I1213 08:29:07.770067 4971 scope.go:117] "RemoveContainer" containerID="bc1c06c3bcebf9c00aee813c7819c279356be670c515e1ea302b50b8618bc001" Dec 13 08:29:07 crc kubenswrapper[4971]: E1213 08:29:07.771321 4971 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-82xjz_openshift-machine-config-operator(e0c4af12-edda-41d7-9b44-f87396174ef2)\"" pod="openshift-machine-config-operator/machine-config-daemon-82xjz" podUID="e0c4af12-edda-41d7-9b44-f87396174ef2" Dec 13 08:29:22 crc kubenswrapper[4971]: I1213 08:29:22.769334 4971 scope.go:117] "RemoveContainer" containerID="bc1c06c3bcebf9c00aee813c7819c279356be670c515e1ea302b50b8618bc001" Dec 13 08:29:22 crc kubenswrapper[4971]: E1213 08:29:22.770598 4971 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-82xjz_openshift-machine-config-operator(e0c4af12-edda-41d7-9b44-f87396174ef2)\"" pod="openshift-machine-config-operator/machine-config-daemon-82xjz" podUID="e0c4af12-edda-41d7-9b44-f87396174ef2" Dec 13 08:29:35 crc kubenswrapper[4971]: I1213 08:29:35.769389 4971 scope.go:117] "RemoveContainer" containerID="bc1c06c3bcebf9c00aee813c7819c279356be670c515e1ea302b50b8618bc001" Dec 13 08:29:35 crc kubenswrapper[4971]: E1213 08:29:35.770497 4971 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-82xjz_openshift-machine-config-operator(e0c4af12-edda-41d7-9b44-f87396174ef2)\"" pod="openshift-machine-config-operator/machine-config-daemon-82xjz" podUID="e0c4af12-edda-41d7-9b44-f87396174ef2" Dec 13 08:29:48 crc kubenswrapper[4971]: I1213 08:29:48.769037 4971 scope.go:117] "RemoveContainer" containerID="bc1c06c3bcebf9c00aee813c7819c279356be670c515e1ea302b50b8618bc001" Dec 13 08:29:48 crc kubenswrapper[4971]: E1213 08:29:48.770296 4971 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-82xjz_openshift-machine-config-operator(e0c4af12-edda-41d7-9b44-f87396174ef2)\"" pod="openshift-machine-config-operator/machine-config-daemon-82xjz" podUID="e0c4af12-edda-41d7-9b44-f87396174ef2" Dec 13 08:30:00 crc kubenswrapper[4971]: I1213 08:30:00.159563 4971 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29426910-x72qh"] Dec 13 08:30:00 crc kubenswrapper[4971]: E1213 08:30:00.161207 4971 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="9ede661a-912c-4701-8c62-aa1bac23c7c6" containerName="registry-server" Dec 13 08:30:00 crc kubenswrapper[4971]: I1213 08:30:00.161229 4971 state_mem.go:107] "Deleted CPUSet assignment" podUID="9ede661a-912c-4701-8c62-aa1bac23c7c6" containerName="registry-server" Dec 13 08:30:00 crc kubenswrapper[4971]: E1213 08:30:00.161278 4971 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="9ede661a-912c-4701-8c62-aa1bac23c7c6" containerName="extract-content" Dec 13 08:30:00 crc kubenswrapper[4971]: I1213 08:30:00.161286 4971 state_mem.go:107] "Deleted CPUSet assignment" podUID="9ede661a-912c-4701-8c62-aa1bac23c7c6" containerName="extract-content" Dec 13 08:30:00 crc kubenswrapper[4971]: E1213 08:30:00.161320 4971 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="9ede661a-912c-4701-8c62-aa1bac23c7c6" containerName="extract-utilities" Dec 13 08:30:00 crc kubenswrapper[4971]: I1213 08:30:00.161330 4971 state_mem.go:107] "Deleted CPUSet assignment" podUID="9ede661a-912c-4701-8c62-aa1bac23c7c6" containerName="extract-utilities" Dec 13 08:30:00 crc kubenswrapper[4971]: I1213 08:30:00.161633 4971 memory_manager.go:354] "RemoveStaleState removing state" podUID="9ede661a-912c-4701-8c62-aa1bac23c7c6" containerName="registry-server" Dec 13 08:30:00 crc kubenswrapper[4971]: I1213 08:30:00.162811 4971 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29426910-x72qh" Dec 13 08:30:00 crc kubenswrapper[4971]: I1213 08:30:00.165248 4971 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-operator-lifecycle-manager"/"collect-profiles-config" Dec 13 08:30:00 crc kubenswrapper[4971]: I1213 08:30:00.166194 4971 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"collect-profiles-dockercfg-kzf4t" Dec 13 08:30:00 crc kubenswrapper[4971]: I1213 08:30:00.186717 4971 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29426910-x72qh"] Dec 13 08:30:00 crc kubenswrapper[4971]: I1213 08:30:00.261111 4971 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/d0ddcbe1-f21e-4b5d-b4a1-8aa5c45d41f7-config-volume\") pod \"collect-profiles-29426910-x72qh\" (UID: \"d0ddcbe1-f21e-4b5d-b4a1-8aa5c45d41f7\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29426910-x72qh" Dec 13 08:30:00 crc kubenswrapper[4971]: I1213 08:30:00.261272 4971 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-h2f6t\" (UniqueName: \"kubernetes.io/projected/d0ddcbe1-f21e-4b5d-b4a1-8aa5c45d41f7-kube-api-access-h2f6t\") pod \"collect-profiles-29426910-x72qh\" (UID: \"d0ddcbe1-f21e-4b5d-b4a1-8aa5c45d41f7\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29426910-x72qh" Dec 13 08:30:00 crc kubenswrapper[4971]: I1213 08:30:00.261373 4971 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/d0ddcbe1-f21e-4b5d-b4a1-8aa5c45d41f7-secret-volume\") pod \"collect-profiles-29426910-x72qh\" (UID: \"d0ddcbe1-f21e-4b5d-b4a1-8aa5c45d41f7\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29426910-x72qh" Dec 13 08:30:00 crc kubenswrapper[4971]: I1213 08:30:00.363683 4971 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/d0ddcbe1-f21e-4b5d-b4a1-8aa5c45d41f7-config-volume\") pod \"collect-profiles-29426910-x72qh\" (UID: \"d0ddcbe1-f21e-4b5d-b4a1-8aa5c45d41f7\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29426910-x72qh" Dec 13 08:30:00 crc kubenswrapper[4971]: I1213 08:30:00.364235 4971 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-h2f6t\" (UniqueName: \"kubernetes.io/projected/d0ddcbe1-f21e-4b5d-b4a1-8aa5c45d41f7-kube-api-access-h2f6t\") pod \"collect-profiles-29426910-x72qh\" (UID: \"d0ddcbe1-f21e-4b5d-b4a1-8aa5c45d41f7\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29426910-x72qh" Dec 13 08:30:00 crc kubenswrapper[4971]: I1213 08:30:00.365296 4971 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/d0ddcbe1-f21e-4b5d-b4a1-8aa5c45d41f7-config-volume\") pod \"collect-profiles-29426910-x72qh\" (UID: \"d0ddcbe1-f21e-4b5d-b4a1-8aa5c45d41f7\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29426910-x72qh" Dec 13 08:30:00 crc kubenswrapper[4971]: I1213 08:30:00.364509 4971 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/d0ddcbe1-f21e-4b5d-b4a1-8aa5c45d41f7-secret-volume\") pod \"collect-profiles-29426910-x72qh\" (UID: \"d0ddcbe1-f21e-4b5d-b4a1-8aa5c45d41f7\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29426910-x72qh" Dec 13 08:30:00 crc kubenswrapper[4971]: I1213 08:30:00.386219 4971 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/d0ddcbe1-f21e-4b5d-b4a1-8aa5c45d41f7-secret-volume\") pod \"collect-profiles-29426910-x72qh\" (UID: \"d0ddcbe1-f21e-4b5d-b4a1-8aa5c45d41f7\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29426910-x72qh" Dec 13 08:30:00 crc kubenswrapper[4971]: I1213 08:30:00.386418 4971 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-h2f6t\" (UniqueName: \"kubernetes.io/projected/d0ddcbe1-f21e-4b5d-b4a1-8aa5c45d41f7-kube-api-access-h2f6t\") pod \"collect-profiles-29426910-x72qh\" (UID: \"d0ddcbe1-f21e-4b5d-b4a1-8aa5c45d41f7\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29426910-x72qh" Dec 13 08:30:00 crc kubenswrapper[4971]: I1213 08:30:00.508684 4971 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29426910-x72qh" Dec 13 08:30:01 crc kubenswrapper[4971]: I1213 08:30:01.041230 4971 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29426910-x72qh"] Dec 13 08:30:01 crc kubenswrapper[4971]: I1213 08:30:01.728435 4971 generic.go:334] "Generic (PLEG): container finished" podID="d0ddcbe1-f21e-4b5d-b4a1-8aa5c45d41f7" containerID="227abad82435cbc9348715e19bc56b3b34cbbcb3ae0118da06c5fe5cb0a701ac" exitCode=0 Dec 13 08:30:01 crc kubenswrapper[4971]: I1213 08:30:01.729110 4971 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29426910-x72qh" event={"ID":"d0ddcbe1-f21e-4b5d-b4a1-8aa5c45d41f7","Type":"ContainerDied","Data":"227abad82435cbc9348715e19bc56b3b34cbbcb3ae0118da06c5fe5cb0a701ac"} Dec 13 08:30:01 crc kubenswrapper[4971]: I1213 08:30:01.729175 4971 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29426910-x72qh" event={"ID":"d0ddcbe1-f21e-4b5d-b4a1-8aa5c45d41f7","Type":"ContainerStarted","Data":"f140e61c6ac30093126af149057ad2c5565bb19a20cbd23ba57612a634893ad0"} Dec 13 08:30:01 crc kubenswrapper[4971]: E1213 08:30:01.811902 4971 cadvisor_stats_provider.go:516] "Partial failure issuing cadvisor.ContainerInfoV2" err="partial failures: [\"/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podd0ddcbe1_f21e_4b5d_b4a1_8aa5c45d41f7.slice/crio-conmon-227abad82435cbc9348715e19bc56b3b34cbbcb3ae0118da06c5fe5cb0a701ac.scope\": RecentStats: unable to find data in memory cache]" Dec 13 08:30:03 crc kubenswrapper[4971]: I1213 08:30:03.354828 4971 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29426910-x72qh" Dec 13 08:30:03 crc kubenswrapper[4971]: I1213 08:30:03.448609 4971 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-h2f6t\" (UniqueName: \"kubernetes.io/projected/d0ddcbe1-f21e-4b5d-b4a1-8aa5c45d41f7-kube-api-access-h2f6t\") pod \"d0ddcbe1-f21e-4b5d-b4a1-8aa5c45d41f7\" (UID: \"d0ddcbe1-f21e-4b5d-b4a1-8aa5c45d41f7\") " Dec 13 08:30:03 crc kubenswrapper[4971]: I1213 08:30:03.448715 4971 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/d0ddcbe1-f21e-4b5d-b4a1-8aa5c45d41f7-config-volume\") pod \"d0ddcbe1-f21e-4b5d-b4a1-8aa5c45d41f7\" (UID: \"d0ddcbe1-f21e-4b5d-b4a1-8aa5c45d41f7\") " Dec 13 08:30:03 crc kubenswrapper[4971]: I1213 08:30:03.449037 4971 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/d0ddcbe1-f21e-4b5d-b4a1-8aa5c45d41f7-secret-volume\") pod \"d0ddcbe1-f21e-4b5d-b4a1-8aa5c45d41f7\" (UID: \"d0ddcbe1-f21e-4b5d-b4a1-8aa5c45d41f7\") " Dec 13 08:30:03 crc kubenswrapper[4971]: I1213 08:30:03.450010 4971 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/d0ddcbe1-f21e-4b5d-b4a1-8aa5c45d41f7-config-volume" (OuterVolumeSpecName: "config-volume") pod "d0ddcbe1-f21e-4b5d-b4a1-8aa5c45d41f7" (UID: "d0ddcbe1-f21e-4b5d-b4a1-8aa5c45d41f7"). InnerVolumeSpecName "config-volume". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 13 08:30:03 crc kubenswrapper[4971]: I1213 08:30:03.456587 4971 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/d0ddcbe1-f21e-4b5d-b4a1-8aa5c45d41f7-secret-volume" (OuterVolumeSpecName: "secret-volume") pod "d0ddcbe1-f21e-4b5d-b4a1-8aa5c45d41f7" (UID: "d0ddcbe1-f21e-4b5d-b4a1-8aa5c45d41f7"). InnerVolumeSpecName "secret-volume". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 13 08:30:03 crc kubenswrapper[4971]: I1213 08:30:03.456809 4971 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/d0ddcbe1-f21e-4b5d-b4a1-8aa5c45d41f7-kube-api-access-h2f6t" (OuterVolumeSpecName: "kube-api-access-h2f6t") pod "d0ddcbe1-f21e-4b5d-b4a1-8aa5c45d41f7" (UID: "d0ddcbe1-f21e-4b5d-b4a1-8aa5c45d41f7"). InnerVolumeSpecName "kube-api-access-h2f6t". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 13 08:30:03 crc kubenswrapper[4971]: I1213 08:30:03.552133 4971 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-h2f6t\" (UniqueName: \"kubernetes.io/projected/d0ddcbe1-f21e-4b5d-b4a1-8aa5c45d41f7-kube-api-access-h2f6t\") on node \"crc\" DevicePath \"\"" Dec 13 08:30:03 crc kubenswrapper[4971]: I1213 08:30:03.552181 4971 reconciler_common.go:293] "Volume detached for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/d0ddcbe1-f21e-4b5d-b4a1-8aa5c45d41f7-config-volume\") on node \"crc\" DevicePath \"\"" Dec 13 08:30:03 crc kubenswrapper[4971]: I1213 08:30:03.552194 4971 reconciler_common.go:293] "Volume detached for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/d0ddcbe1-f21e-4b5d-b4a1-8aa5c45d41f7-secret-volume\") on node \"crc\" DevicePath \"\"" Dec 13 08:30:03 crc kubenswrapper[4971]: I1213 08:30:03.753329 4971 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29426910-x72qh" event={"ID":"d0ddcbe1-f21e-4b5d-b4a1-8aa5c45d41f7","Type":"ContainerDied","Data":"f140e61c6ac30093126af149057ad2c5565bb19a20cbd23ba57612a634893ad0"} Dec 13 08:30:03 crc kubenswrapper[4971]: I1213 08:30:03.753828 4971 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="f140e61c6ac30093126af149057ad2c5565bb19a20cbd23ba57612a634893ad0" Dec 13 08:30:03 crc kubenswrapper[4971]: I1213 08:30:03.753409 4971 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29426910-x72qh" Dec 13 08:30:03 crc kubenswrapper[4971]: I1213 08:30:03.793297 4971 scope.go:117] "RemoveContainer" containerID="bc1c06c3bcebf9c00aee813c7819c279356be670c515e1ea302b50b8618bc001" Dec 13 08:30:03 crc kubenswrapper[4971]: E1213 08:30:03.795977 4971 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-82xjz_openshift-machine-config-operator(e0c4af12-edda-41d7-9b44-f87396174ef2)\"" pod="openshift-machine-config-operator/machine-config-daemon-82xjz" podUID="e0c4af12-edda-41d7-9b44-f87396174ef2" Dec 13 08:30:04 crc kubenswrapper[4971]: I1213 08:30:04.443719 4971 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29426865-mqtkj"] Dec 13 08:30:04 crc kubenswrapper[4971]: I1213 08:30:04.452558 4971 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29426865-mqtkj"] Dec 13 08:30:05 crc kubenswrapper[4971]: I1213 08:30:05.786701 4971 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="099e03d5-1210-4976-ae03-e984b101d681" path="/var/lib/kubelet/pods/099e03d5-1210-4976-ae03-e984b101d681/volumes" Dec 13 08:30:14 crc kubenswrapper[4971]: I1213 08:30:14.768308 4971 scope.go:117] "RemoveContainer" containerID="bc1c06c3bcebf9c00aee813c7819c279356be670c515e1ea302b50b8618bc001" Dec 13 08:30:14 crc kubenswrapper[4971]: E1213 08:30:14.769102 4971 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-82xjz_openshift-machine-config-operator(e0c4af12-edda-41d7-9b44-f87396174ef2)\"" pod="openshift-machine-config-operator/machine-config-daemon-82xjz" podUID="e0c4af12-edda-41d7-9b44-f87396174ef2" Dec 13 08:30:27 crc kubenswrapper[4971]: I1213 08:30:27.551578 4971 scope.go:117] "RemoveContainer" containerID="6c76911f291cffb765fc22afc90bcd5afabdd2c106785593da6f0d3f650d878f" Dec 13 08:30:27 crc kubenswrapper[4971]: I1213 08:30:27.769208 4971 scope.go:117] "RemoveContainer" containerID="bc1c06c3bcebf9c00aee813c7819c279356be670c515e1ea302b50b8618bc001" Dec 13 08:30:28 crc kubenswrapper[4971]: I1213 08:30:28.042072 4971 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-82xjz" event={"ID":"e0c4af12-edda-41d7-9b44-f87396174ef2","Type":"ContainerStarted","Data":"87e58ec20b03b7a861da80d90f34f635ea8cc5825e7d209bb6d579dced58891c"} Dec 13 08:32:46 crc kubenswrapper[4971]: I1213 08:32:46.154251 4971 patch_prober.go:28] interesting pod/machine-config-daemon-82xjz container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 13 08:32:46 crc kubenswrapper[4971]: I1213 08:32:46.155201 4971 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-82xjz" podUID="e0c4af12-edda-41d7-9b44-f87396174ef2" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 13 08:33:06 crc kubenswrapper[4971]: I1213 08:33:06.455462 4971 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-operators-h48z2"] Dec 13 08:33:06 crc kubenswrapper[4971]: E1213 08:33:06.456933 4971 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="d0ddcbe1-f21e-4b5d-b4a1-8aa5c45d41f7" containerName="collect-profiles" Dec 13 08:33:06 crc kubenswrapper[4971]: I1213 08:33:06.456952 4971 state_mem.go:107] "Deleted CPUSet assignment" podUID="d0ddcbe1-f21e-4b5d-b4a1-8aa5c45d41f7" containerName="collect-profiles" Dec 13 08:33:06 crc kubenswrapper[4971]: I1213 08:33:06.457178 4971 memory_manager.go:354] "RemoveStaleState removing state" podUID="d0ddcbe1-f21e-4b5d-b4a1-8aa5c45d41f7" containerName="collect-profiles" Dec 13 08:33:06 crc kubenswrapper[4971]: I1213 08:33:06.458867 4971 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-h48z2" Dec 13 08:33:06 crc kubenswrapper[4971]: I1213 08:33:06.473025 4971 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-h48z2"] Dec 13 08:33:06 crc kubenswrapper[4971]: I1213 08:33:06.590409 4971 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-lqw44\" (UniqueName: \"kubernetes.io/projected/cea3f937-1bc0-4dfe-bb7d-a74fe6218006-kube-api-access-lqw44\") pod \"redhat-operators-h48z2\" (UID: \"cea3f937-1bc0-4dfe-bb7d-a74fe6218006\") " pod="openshift-marketplace/redhat-operators-h48z2" Dec 13 08:33:06 crc kubenswrapper[4971]: I1213 08:33:06.591207 4971 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/cea3f937-1bc0-4dfe-bb7d-a74fe6218006-utilities\") pod \"redhat-operators-h48z2\" (UID: \"cea3f937-1bc0-4dfe-bb7d-a74fe6218006\") " pod="openshift-marketplace/redhat-operators-h48z2" Dec 13 08:33:06 crc kubenswrapper[4971]: I1213 08:33:06.591271 4971 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/cea3f937-1bc0-4dfe-bb7d-a74fe6218006-catalog-content\") pod \"redhat-operators-h48z2\" (UID: \"cea3f937-1bc0-4dfe-bb7d-a74fe6218006\") " pod="openshift-marketplace/redhat-operators-h48z2" Dec 13 08:33:06 crc kubenswrapper[4971]: I1213 08:33:06.694677 4971 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-lqw44\" (UniqueName: \"kubernetes.io/projected/cea3f937-1bc0-4dfe-bb7d-a74fe6218006-kube-api-access-lqw44\") pod \"redhat-operators-h48z2\" (UID: \"cea3f937-1bc0-4dfe-bb7d-a74fe6218006\") " pod="openshift-marketplace/redhat-operators-h48z2" Dec 13 08:33:06 crc kubenswrapper[4971]: I1213 08:33:06.694854 4971 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/cea3f937-1bc0-4dfe-bb7d-a74fe6218006-utilities\") pod \"redhat-operators-h48z2\" (UID: \"cea3f937-1bc0-4dfe-bb7d-a74fe6218006\") " pod="openshift-marketplace/redhat-operators-h48z2" Dec 13 08:33:06 crc kubenswrapper[4971]: I1213 08:33:06.694924 4971 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/cea3f937-1bc0-4dfe-bb7d-a74fe6218006-catalog-content\") pod \"redhat-operators-h48z2\" (UID: \"cea3f937-1bc0-4dfe-bb7d-a74fe6218006\") " pod="openshift-marketplace/redhat-operators-h48z2" Dec 13 08:33:06 crc kubenswrapper[4971]: I1213 08:33:06.695617 4971 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/cea3f937-1bc0-4dfe-bb7d-a74fe6218006-catalog-content\") pod \"redhat-operators-h48z2\" (UID: \"cea3f937-1bc0-4dfe-bb7d-a74fe6218006\") " pod="openshift-marketplace/redhat-operators-h48z2" Dec 13 08:33:06 crc kubenswrapper[4971]: I1213 08:33:06.695662 4971 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/cea3f937-1bc0-4dfe-bb7d-a74fe6218006-utilities\") pod \"redhat-operators-h48z2\" (UID: \"cea3f937-1bc0-4dfe-bb7d-a74fe6218006\") " pod="openshift-marketplace/redhat-operators-h48z2" Dec 13 08:33:06 crc kubenswrapper[4971]: I1213 08:33:06.726936 4971 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-lqw44\" (UniqueName: \"kubernetes.io/projected/cea3f937-1bc0-4dfe-bb7d-a74fe6218006-kube-api-access-lqw44\") pod \"redhat-operators-h48z2\" (UID: \"cea3f937-1bc0-4dfe-bb7d-a74fe6218006\") " pod="openshift-marketplace/redhat-operators-h48z2" Dec 13 08:33:06 crc kubenswrapper[4971]: I1213 08:33:06.786664 4971 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-h48z2" Dec 13 08:33:07 crc kubenswrapper[4971]: I1213 08:33:07.415430 4971 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-h48z2"] Dec 13 08:33:08 crc kubenswrapper[4971]: I1213 08:33:08.388298 4971 generic.go:334] "Generic (PLEG): container finished" podID="cea3f937-1bc0-4dfe-bb7d-a74fe6218006" containerID="64a6e9263fe1529f02ae02241f800cb4df1132e31e4c8f2c0c138d650646ad67" exitCode=0 Dec 13 08:33:08 crc kubenswrapper[4971]: I1213 08:33:08.388383 4971 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-h48z2" event={"ID":"cea3f937-1bc0-4dfe-bb7d-a74fe6218006","Type":"ContainerDied","Data":"64a6e9263fe1529f02ae02241f800cb4df1132e31e4c8f2c0c138d650646ad67"} Dec 13 08:33:08 crc kubenswrapper[4971]: I1213 08:33:08.389273 4971 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-h48z2" event={"ID":"cea3f937-1bc0-4dfe-bb7d-a74fe6218006","Type":"ContainerStarted","Data":"66a9c0bb58de759545077f8d233677894e7c66eddcdc5db896f628648f6929eb"} Dec 13 08:33:08 crc kubenswrapper[4971]: I1213 08:33:08.392181 4971 provider.go:102] Refreshing cache for provider: *credentialprovider.defaultDockerConfigProvider Dec 13 08:33:10 crc kubenswrapper[4971]: I1213 08:33:10.233263 4971 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-marketplace-66qbp"] Dec 13 08:33:10 crc kubenswrapper[4971]: I1213 08:33:10.237830 4971 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-66qbp" Dec 13 08:33:10 crc kubenswrapper[4971]: I1213 08:33:10.261462 4971 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-66qbp"] Dec 13 08:33:10 crc kubenswrapper[4971]: I1213 08:33:10.397343 4971 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-b5mgh\" (UniqueName: \"kubernetes.io/projected/6eaf2f74-3685-4128-a19c-1284823b9b64-kube-api-access-b5mgh\") pod \"redhat-marketplace-66qbp\" (UID: \"6eaf2f74-3685-4128-a19c-1284823b9b64\") " pod="openshift-marketplace/redhat-marketplace-66qbp" Dec 13 08:33:10 crc kubenswrapper[4971]: I1213 08:33:10.397974 4971 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/6eaf2f74-3685-4128-a19c-1284823b9b64-utilities\") pod \"redhat-marketplace-66qbp\" (UID: \"6eaf2f74-3685-4128-a19c-1284823b9b64\") " pod="openshift-marketplace/redhat-marketplace-66qbp" Dec 13 08:33:10 crc kubenswrapper[4971]: I1213 08:33:10.398132 4971 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/6eaf2f74-3685-4128-a19c-1284823b9b64-catalog-content\") pod \"redhat-marketplace-66qbp\" (UID: \"6eaf2f74-3685-4128-a19c-1284823b9b64\") " pod="openshift-marketplace/redhat-marketplace-66qbp" Dec 13 08:33:10 crc kubenswrapper[4971]: I1213 08:33:10.415179 4971 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-h48z2" event={"ID":"cea3f937-1bc0-4dfe-bb7d-a74fe6218006","Type":"ContainerStarted","Data":"2df9843215147a593371cb1e1574e6a00a2b4c1b962f473f8e5e6002b1a112b7"} Dec 13 08:33:10 crc kubenswrapper[4971]: I1213 08:33:10.501144 4971 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/6eaf2f74-3685-4128-a19c-1284823b9b64-utilities\") pod \"redhat-marketplace-66qbp\" (UID: \"6eaf2f74-3685-4128-a19c-1284823b9b64\") " pod="openshift-marketplace/redhat-marketplace-66qbp" Dec 13 08:33:10 crc kubenswrapper[4971]: I1213 08:33:10.501258 4971 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/6eaf2f74-3685-4128-a19c-1284823b9b64-catalog-content\") pod \"redhat-marketplace-66qbp\" (UID: \"6eaf2f74-3685-4128-a19c-1284823b9b64\") " pod="openshift-marketplace/redhat-marketplace-66qbp" Dec 13 08:33:10 crc kubenswrapper[4971]: I1213 08:33:10.501336 4971 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-b5mgh\" (UniqueName: \"kubernetes.io/projected/6eaf2f74-3685-4128-a19c-1284823b9b64-kube-api-access-b5mgh\") pod \"redhat-marketplace-66qbp\" (UID: \"6eaf2f74-3685-4128-a19c-1284823b9b64\") " pod="openshift-marketplace/redhat-marketplace-66qbp" Dec 13 08:33:10 crc kubenswrapper[4971]: I1213 08:33:10.501984 4971 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/6eaf2f74-3685-4128-a19c-1284823b9b64-utilities\") pod \"redhat-marketplace-66qbp\" (UID: \"6eaf2f74-3685-4128-a19c-1284823b9b64\") " pod="openshift-marketplace/redhat-marketplace-66qbp" Dec 13 08:33:10 crc kubenswrapper[4971]: I1213 08:33:10.502031 4971 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/6eaf2f74-3685-4128-a19c-1284823b9b64-catalog-content\") pod \"redhat-marketplace-66qbp\" (UID: \"6eaf2f74-3685-4128-a19c-1284823b9b64\") " pod="openshift-marketplace/redhat-marketplace-66qbp" Dec 13 08:33:10 crc kubenswrapper[4971]: I1213 08:33:10.531496 4971 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-b5mgh\" (UniqueName: \"kubernetes.io/projected/6eaf2f74-3685-4128-a19c-1284823b9b64-kube-api-access-b5mgh\") pod \"redhat-marketplace-66qbp\" (UID: \"6eaf2f74-3685-4128-a19c-1284823b9b64\") " pod="openshift-marketplace/redhat-marketplace-66qbp" Dec 13 08:33:10 crc kubenswrapper[4971]: I1213 08:33:10.559355 4971 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-66qbp" Dec 13 08:33:11 crc kubenswrapper[4971]: I1213 08:33:11.067293 4971 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-66qbp"] Dec 13 08:33:11 crc kubenswrapper[4971]: W1213 08:33:11.073386 4971 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod6eaf2f74_3685_4128_a19c_1284823b9b64.slice/crio-930f73c13fd0b671982247b301c6251ca79919495211684fdfd4060b5cc5598e WatchSource:0}: Error finding container 930f73c13fd0b671982247b301c6251ca79919495211684fdfd4060b5cc5598e: Status 404 returned error can't find the container with id 930f73c13fd0b671982247b301c6251ca79919495211684fdfd4060b5cc5598e Dec 13 08:33:11 crc kubenswrapper[4971]: I1213 08:33:11.427750 4971 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-66qbp" event={"ID":"6eaf2f74-3685-4128-a19c-1284823b9b64","Type":"ContainerStarted","Data":"930f73c13fd0b671982247b301c6251ca79919495211684fdfd4060b5cc5598e"} Dec 13 08:33:12 crc kubenswrapper[4971]: I1213 08:33:12.440214 4971 generic.go:334] "Generic (PLEG): container finished" podID="6eaf2f74-3685-4128-a19c-1284823b9b64" containerID="8ad6e249660f055a5f83346bb5119e90e34c81d937903f96346bc7a5cfe6ae1f" exitCode=0 Dec 13 08:33:12 crc kubenswrapper[4971]: I1213 08:33:12.440397 4971 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-66qbp" event={"ID":"6eaf2f74-3685-4128-a19c-1284823b9b64","Type":"ContainerDied","Data":"8ad6e249660f055a5f83346bb5119e90e34c81d937903f96346bc7a5cfe6ae1f"} Dec 13 08:33:14 crc kubenswrapper[4971]: I1213 08:33:14.475168 4971 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-66qbp" event={"ID":"6eaf2f74-3685-4128-a19c-1284823b9b64","Type":"ContainerStarted","Data":"9ee2be33d92c34ef697368341f606303bf5684dfcbabf4efafa6406555f385d4"} Dec 13 08:33:15 crc kubenswrapper[4971]: I1213 08:33:15.487585 4971 generic.go:334] "Generic (PLEG): container finished" podID="6eaf2f74-3685-4128-a19c-1284823b9b64" containerID="9ee2be33d92c34ef697368341f606303bf5684dfcbabf4efafa6406555f385d4" exitCode=0 Dec 13 08:33:15 crc kubenswrapper[4971]: I1213 08:33:15.487732 4971 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-66qbp" event={"ID":"6eaf2f74-3685-4128-a19c-1284823b9b64","Type":"ContainerDied","Data":"9ee2be33d92c34ef697368341f606303bf5684dfcbabf4efafa6406555f385d4"} Dec 13 08:33:15 crc kubenswrapper[4971]: I1213 08:33:15.491431 4971 generic.go:334] "Generic (PLEG): container finished" podID="cea3f937-1bc0-4dfe-bb7d-a74fe6218006" containerID="2df9843215147a593371cb1e1574e6a00a2b4c1b962f473f8e5e6002b1a112b7" exitCode=0 Dec 13 08:33:15 crc kubenswrapper[4971]: I1213 08:33:15.491500 4971 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-h48z2" event={"ID":"cea3f937-1bc0-4dfe-bb7d-a74fe6218006","Type":"ContainerDied","Data":"2df9843215147a593371cb1e1574e6a00a2b4c1b962f473f8e5e6002b1a112b7"} Dec 13 08:33:16 crc kubenswrapper[4971]: I1213 08:33:16.153388 4971 patch_prober.go:28] interesting pod/machine-config-daemon-82xjz container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 13 08:33:16 crc kubenswrapper[4971]: I1213 08:33:16.154490 4971 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-82xjz" podUID="e0c4af12-edda-41d7-9b44-f87396174ef2" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 13 08:33:16 crc kubenswrapper[4971]: I1213 08:33:16.507285 4971 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-66qbp" event={"ID":"6eaf2f74-3685-4128-a19c-1284823b9b64","Type":"ContainerStarted","Data":"5b26f8105897e8f5cd4bc5170b89c09aa67e5c87be6915fdadfeffa77f0ce4cd"} Dec 13 08:33:16 crc kubenswrapper[4971]: I1213 08:33:16.513542 4971 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-h48z2" event={"ID":"cea3f937-1bc0-4dfe-bb7d-a74fe6218006","Type":"ContainerStarted","Data":"5d3bf4bd59b494a3b9725ac40fd9e72573afe95c3b614ff7ca2adb7fb7decfe1"} Dec 13 08:33:16 crc kubenswrapper[4971]: I1213 08:33:16.530060 4971 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-marketplace-66qbp" podStartSLOduration=3.065644405 podStartE2EDuration="6.530031669s" podCreationTimestamp="2025-12-13 08:33:10 +0000 UTC" firstStartedPulling="2025-12-13 08:33:12.444849613 +0000 UTC m=+6249.049259071" lastFinishedPulling="2025-12-13 08:33:15.909236877 +0000 UTC m=+6252.513646335" observedRunningTime="2025-12-13 08:33:16.528361539 +0000 UTC m=+6253.132770987" watchObservedRunningTime="2025-12-13 08:33:16.530031669 +0000 UTC m=+6253.134441107" Dec 13 08:33:16 crc kubenswrapper[4971]: I1213 08:33:16.556691 4971 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-operators-h48z2" podStartSLOduration=2.851714603 podStartE2EDuration="10.556658692s" podCreationTimestamp="2025-12-13 08:33:06 +0000 UTC" firstStartedPulling="2025-12-13 08:33:08.391494977 +0000 UTC m=+6244.995904465" lastFinishedPulling="2025-12-13 08:33:16.096439106 +0000 UTC m=+6252.700848554" observedRunningTime="2025-12-13 08:33:16.55291985 +0000 UTC m=+6253.157329308" watchObservedRunningTime="2025-12-13 08:33:16.556658692 +0000 UTC m=+6253.161068140" Dec 13 08:33:16 crc kubenswrapper[4971]: I1213 08:33:16.787122 4971 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-operators-h48z2" Dec 13 08:33:16 crc kubenswrapper[4971]: I1213 08:33:16.787211 4971 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-operators-h48z2" Dec 13 08:33:17 crc kubenswrapper[4971]: I1213 08:33:17.840627 4971 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-marketplace/redhat-operators-h48z2" podUID="cea3f937-1bc0-4dfe-bb7d-a74fe6218006" containerName="registry-server" probeResult="failure" output=< Dec 13 08:33:17 crc kubenswrapper[4971]: timeout: failed to connect service ":50051" within 1s Dec 13 08:33:17 crc kubenswrapper[4971]: > Dec 13 08:33:20 crc kubenswrapper[4971]: I1213 08:33:20.559780 4971 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-marketplace-66qbp" Dec 13 08:33:20 crc kubenswrapper[4971]: I1213 08:33:20.560852 4971 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-marketplace-66qbp" Dec 13 08:33:20 crc kubenswrapper[4971]: I1213 08:33:20.619438 4971 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-marketplace-66qbp" Dec 13 08:33:21 crc kubenswrapper[4971]: I1213 08:33:21.621890 4971 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-marketplace-66qbp" Dec 13 08:33:22 crc kubenswrapper[4971]: I1213 08:33:22.019150 4971 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-66qbp"] Dec 13 08:33:23 crc kubenswrapper[4971]: I1213 08:33:23.591446 4971 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-marketplace-66qbp" podUID="6eaf2f74-3685-4128-a19c-1284823b9b64" containerName="registry-server" containerID="cri-o://5b26f8105897e8f5cd4bc5170b89c09aa67e5c87be6915fdadfeffa77f0ce4cd" gracePeriod=2 Dec 13 08:33:24 crc kubenswrapper[4971]: I1213 08:33:24.230065 4971 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-66qbp" Dec 13 08:33:24 crc kubenswrapper[4971]: I1213 08:33:24.298365 4971 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/6eaf2f74-3685-4128-a19c-1284823b9b64-utilities\") pod \"6eaf2f74-3685-4128-a19c-1284823b9b64\" (UID: \"6eaf2f74-3685-4128-a19c-1284823b9b64\") " Dec 13 08:33:24 crc kubenswrapper[4971]: I1213 08:33:24.298766 4971 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/6eaf2f74-3685-4128-a19c-1284823b9b64-catalog-content\") pod \"6eaf2f74-3685-4128-a19c-1284823b9b64\" (UID: \"6eaf2f74-3685-4128-a19c-1284823b9b64\") " Dec 13 08:33:24 crc kubenswrapper[4971]: I1213 08:33:24.298918 4971 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-b5mgh\" (UniqueName: \"kubernetes.io/projected/6eaf2f74-3685-4128-a19c-1284823b9b64-kube-api-access-b5mgh\") pod \"6eaf2f74-3685-4128-a19c-1284823b9b64\" (UID: \"6eaf2f74-3685-4128-a19c-1284823b9b64\") " Dec 13 08:33:24 crc kubenswrapper[4971]: I1213 08:33:24.299843 4971 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/6eaf2f74-3685-4128-a19c-1284823b9b64-utilities" (OuterVolumeSpecName: "utilities") pod "6eaf2f74-3685-4128-a19c-1284823b9b64" (UID: "6eaf2f74-3685-4128-a19c-1284823b9b64"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 13 08:33:24 crc kubenswrapper[4971]: I1213 08:33:24.308580 4971 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/6eaf2f74-3685-4128-a19c-1284823b9b64-kube-api-access-b5mgh" (OuterVolumeSpecName: "kube-api-access-b5mgh") pod "6eaf2f74-3685-4128-a19c-1284823b9b64" (UID: "6eaf2f74-3685-4128-a19c-1284823b9b64"). InnerVolumeSpecName "kube-api-access-b5mgh". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 13 08:33:24 crc kubenswrapper[4971]: I1213 08:33:24.328394 4971 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/6eaf2f74-3685-4128-a19c-1284823b9b64-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "6eaf2f74-3685-4128-a19c-1284823b9b64" (UID: "6eaf2f74-3685-4128-a19c-1284823b9b64"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 13 08:33:24 crc kubenswrapper[4971]: I1213 08:33:24.402498 4971 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/6eaf2f74-3685-4128-a19c-1284823b9b64-utilities\") on node \"crc\" DevicePath \"\"" Dec 13 08:33:24 crc kubenswrapper[4971]: I1213 08:33:24.402629 4971 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/6eaf2f74-3685-4128-a19c-1284823b9b64-catalog-content\") on node \"crc\" DevicePath \"\"" Dec 13 08:33:24 crc kubenswrapper[4971]: I1213 08:33:24.402644 4971 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-b5mgh\" (UniqueName: \"kubernetes.io/projected/6eaf2f74-3685-4128-a19c-1284823b9b64-kube-api-access-b5mgh\") on node \"crc\" DevicePath \"\"" Dec 13 08:33:24 crc kubenswrapper[4971]: I1213 08:33:24.609703 4971 generic.go:334] "Generic (PLEG): container finished" podID="6eaf2f74-3685-4128-a19c-1284823b9b64" containerID="5b26f8105897e8f5cd4bc5170b89c09aa67e5c87be6915fdadfeffa77f0ce4cd" exitCode=0 Dec 13 08:33:24 crc kubenswrapper[4971]: I1213 08:33:24.609775 4971 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-66qbp" event={"ID":"6eaf2f74-3685-4128-a19c-1284823b9b64","Type":"ContainerDied","Data":"5b26f8105897e8f5cd4bc5170b89c09aa67e5c87be6915fdadfeffa77f0ce4cd"} Dec 13 08:33:24 crc kubenswrapper[4971]: I1213 08:33:24.609818 4971 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-66qbp" event={"ID":"6eaf2f74-3685-4128-a19c-1284823b9b64","Type":"ContainerDied","Data":"930f73c13fd0b671982247b301c6251ca79919495211684fdfd4060b5cc5598e"} Dec 13 08:33:24 crc kubenswrapper[4971]: I1213 08:33:24.609848 4971 scope.go:117] "RemoveContainer" containerID="5b26f8105897e8f5cd4bc5170b89c09aa67e5c87be6915fdadfeffa77f0ce4cd" Dec 13 08:33:24 crc kubenswrapper[4971]: I1213 08:33:24.610078 4971 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-66qbp" Dec 13 08:33:24 crc kubenswrapper[4971]: I1213 08:33:24.646014 4971 scope.go:117] "RemoveContainer" containerID="9ee2be33d92c34ef697368341f606303bf5684dfcbabf4efafa6406555f385d4" Dec 13 08:33:24 crc kubenswrapper[4971]: I1213 08:33:24.658319 4971 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-66qbp"] Dec 13 08:33:24 crc kubenswrapper[4971]: I1213 08:33:24.674708 4971 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-marketplace-66qbp"] Dec 13 08:33:24 crc kubenswrapper[4971]: I1213 08:33:24.700283 4971 scope.go:117] "RemoveContainer" containerID="8ad6e249660f055a5f83346bb5119e90e34c81d937903f96346bc7a5cfe6ae1f" Dec 13 08:33:24 crc kubenswrapper[4971]: I1213 08:33:24.733984 4971 scope.go:117] "RemoveContainer" containerID="5b26f8105897e8f5cd4bc5170b89c09aa67e5c87be6915fdadfeffa77f0ce4cd" Dec 13 08:33:24 crc kubenswrapper[4971]: E1213 08:33:24.735760 4971 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"5b26f8105897e8f5cd4bc5170b89c09aa67e5c87be6915fdadfeffa77f0ce4cd\": container with ID starting with 5b26f8105897e8f5cd4bc5170b89c09aa67e5c87be6915fdadfeffa77f0ce4cd not found: ID does not exist" containerID="5b26f8105897e8f5cd4bc5170b89c09aa67e5c87be6915fdadfeffa77f0ce4cd" Dec 13 08:33:24 crc kubenswrapper[4971]: I1213 08:33:24.735800 4971 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"5b26f8105897e8f5cd4bc5170b89c09aa67e5c87be6915fdadfeffa77f0ce4cd"} err="failed to get container status \"5b26f8105897e8f5cd4bc5170b89c09aa67e5c87be6915fdadfeffa77f0ce4cd\": rpc error: code = NotFound desc = could not find container \"5b26f8105897e8f5cd4bc5170b89c09aa67e5c87be6915fdadfeffa77f0ce4cd\": container with ID starting with 5b26f8105897e8f5cd4bc5170b89c09aa67e5c87be6915fdadfeffa77f0ce4cd not found: ID does not exist" Dec 13 08:33:24 crc kubenswrapper[4971]: I1213 08:33:24.735827 4971 scope.go:117] "RemoveContainer" containerID="9ee2be33d92c34ef697368341f606303bf5684dfcbabf4efafa6406555f385d4" Dec 13 08:33:24 crc kubenswrapper[4971]: E1213 08:33:24.736553 4971 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"9ee2be33d92c34ef697368341f606303bf5684dfcbabf4efafa6406555f385d4\": container with ID starting with 9ee2be33d92c34ef697368341f606303bf5684dfcbabf4efafa6406555f385d4 not found: ID does not exist" containerID="9ee2be33d92c34ef697368341f606303bf5684dfcbabf4efafa6406555f385d4" Dec 13 08:33:24 crc kubenswrapper[4971]: I1213 08:33:24.736586 4971 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"9ee2be33d92c34ef697368341f606303bf5684dfcbabf4efafa6406555f385d4"} err="failed to get container status \"9ee2be33d92c34ef697368341f606303bf5684dfcbabf4efafa6406555f385d4\": rpc error: code = NotFound desc = could not find container \"9ee2be33d92c34ef697368341f606303bf5684dfcbabf4efafa6406555f385d4\": container with ID starting with 9ee2be33d92c34ef697368341f606303bf5684dfcbabf4efafa6406555f385d4 not found: ID does not exist" Dec 13 08:33:24 crc kubenswrapper[4971]: I1213 08:33:24.736607 4971 scope.go:117] "RemoveContainer" containerID="8ad6e249660f055a5f83346bb5119e90e34c81d937903f96346bc7a5cfe6ae1f" Dec 13 08:33:24 crc kubenswrapper[4971]: E1213 08:33:24.736985 4971 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"8ad6e249660f055a5f83346bb5119e90e34c81d937903f96346bc7a5cfe6ae1f\": container with ID starting with 8ad6e249660f055a5f83346bb5119e90e34c81d937903f96346bc7a5cfe6ae1f not found: ID does not exist" containerID="8ad6e249660f055a5f83346bb5119e90e34c81d937903f96346bc7a5cfe6ae1f" Dec 13 08:33:24 crc kubenswrapper[4971]: I1213 08:33:24.737015 4971 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"8ad6e249660f055a5f83346bb5119e90e34c81d937903f96346bc7a5cfe6ae1f"} err="failed to get container status \"8ad6e249660f055a5f83346bb5119e90e34c81d937903f96346bc7a5cfe6ae1f\": rpc error: code = NotFound desc = could not find container \"8ad6e249660f055a5f83346bb5119e90e34c81d937903f96346bc7a5cfe6ae1f\": container with ID starting with 8ad6e249660f055a5f83346bb5119e90e34c81d937903f96346bc7a5cfe6ae1f not found: ID does not exist" Dec 13 08:33:25 crc kubenswrapper[4971]: I1213 08:33:25.783185 4971 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="6eaf2f74-3685-4128-a19c-1284823b9b64" path="/var/lib/kubelet/pods/6eaf2f74-3685-4128-a19c-1284823b9b64/volumes" Dec 13 08:33:26 crc kubenswrapper[4971]: I1213 08:33:26.853852 4971 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-operators-h48z2" Dec 13 08:33:26 crc kubenswrapper[4971]: I1213 08:33:26.923656 4971 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-operators-h48z2" Dec 13 08:33:27 crc kubenswrapper[4971]: I1213 08:33:27.417159 4971 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-h48z2"] Dec 13 08:33:28 crc kubenswrapper[4971]: I1213 08:33:28.680673 4971 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-operators-h48z2" podUID="cea3f937-1bc0-4dfe-bb7d-a74fe6218006" containerName="registry-server" containerID="cri-o://5d3bf4bd59b494a3b9725ac40fd9e72573afe95c3b614ff7ca2adb7fb7decfe1" gracePeriod=2 Dec 13 08:33:29 crc kubenswrapper[4971]: I1213 08:33:29.413649 4971 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-h48z2" Dec 13 08:33:29 crc kubenswrapper[4971]: I1213 08:33:29.567852 4971 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-lqw44\" (UniqueName: \"kubernetes.io/projected/cea3f937-1bc0-4dfe-bb7d-a74fe6218006-kube-api-access-lqw44\") pod \"cea3f937-1bc0-4dfe-bb7d-a74fe6218006\" (UID: \"cea3f937-1bc0-4dfe-bb7d-a74fe6218006\") " Dec 13 08:33:29 crc kubenswrapper[4971]: I1213 08:33:29.568037 4971 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/cea3f937-1bc0-4dfe-bb7d-a74fe6218006-catalog-content\") pod \"cea3f937-1bc0-4dfe-bb7d-a74fe6218006\" (UID: \"cea3f937-1bc0-4dfe-bb7d-a74fe6218006\") " Dec 13 08:33:29 crc kubenswrapper[4971]: I1213 08:33:29.568135 4971 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/cea3f937-1bc0-4dfe-bb7d-a74fe6218006-utilities\") pod \"cea3f937-1bc0-4dfe-bb7d-a74fe6218006\" (UID: \"cea3f937-1bc0-4dfe-bb7d-a74fe6218006\") " Dec 13 08:33:29 crc kubenswrapper[4971]: I1213 08:33:29.569348 4971 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/cea3f937-1bc0-4dfe-bb7d-a74fe6218006-utilities" (OuterVolumeSpecName: "utilities") pod "cea3f937-1bc0-4dfe-bb7d-a74fe6218006" (UID: "cea3f937-1bc0-4dfe-bb7d-a74fe6218006"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 13 08:33:29 crc kubenswrapper[4971]: I1213 08:33:29.580760 4971 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/cea3f937-1bc0-4dfe-bb7d-a74fe6218006-kube-api-access-lqw44" (OuterVolumeSpecName: "kube-api-access-lqw44") pod "cea3f937-1bc0-4dfe-bb7d-a74fe6218006" (UID: "cea3f937-1bc0-4dfe-bb7d-a74fe6218006"). InnerVolumeSpecName "kube-api-access-lqw44". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 13 08:33:29 crc kubenswrapper[4971]: I1213 08:33:29.679887 4971 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-lqw44\" (UniqueName: \"kubernetes.io/projected/cea3f937-1bc0-4dfe-bb7d-a74fe6218006-kube-api-access-lqw44\") on node \"crc\" DevicePath \"\"" Dec 13 08:33:29 crc kubenswrapper[4971]: I1213 08:33:29.679944 4971 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/cea3f937-1bc0-4dfe-bb7d-a74fe6218006-utilities\") on node \"crc\" DevicePath \"\"" Dec 13 08:33:29 crc kubenswrapper[4971]: I1213 08:33:29.708979 4971 generic.go:334] "Generic (PLEG): container finished" podID="cea3f937-1bc0-4dfe-bb7d-a74fe6218006" containerID="5d3bf4bd59b494a3b9725ac40fd9e72573afe95c3b614ff7ca2adb7fb7decfe1" exitCode=0 Dec 13 08:33:29 crc kubenswrapper[4971]: I1213 08:33:29.709043 4971 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-h48z2" event={"ID":"cea3f937-1bc0-4dfe-bb7d-a74fe6218006","Type":"ContainerDied","Data":"5d3bf4bd59b494a3b9725ac40fd9e72573afe95c3b614ff7ca2adb7fb7decfe1"} Dec 13 08:33:29 crc kubenswrapper[4971]: I1213 08:33:29.710098 4971 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-h48z2" event={"ID":"cea3f937-1bc0-4dfe-bb7d-a74fe6218006","Type":"ContainerDied","Data":"66a9c0bb58de759545077f8d233677894e7c66eddcdc5db896f628648f6929eb"} Dec 13 08:33:29 crc kubenswrapper[4971]: I1213 08:33:29.710236 4971 scope.go:117] "RemoveContainer" containerID="5d3bf4bd59b494a3b9725ac40fd9e72573afe95c3b614ff7ca2adb7fb7decfe1" Dec 13 08:33:29 crc kubenswrapper[4971]: I1213 08:33:29.709102 4971 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-h48z2" Dec 13 08:33:29 crc kubenswrapper[4971]: I1213 08:33:29.728340 4971 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/cea3f937-1bc0-4dfe-bb7d-a74fe6218006-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "cea3f937-1bc0-4dfe-bb7d-a74fe6218006" (UID: "cea3f937-1bc0-4dfe-bb7d-a74fe6218006"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 13 08:33:29 crc kubenswrapper[4971]: I1213 08:33:29.739231 4971 scope.go:117] "RemoveContainer" containerID="2df9843215147a593371cb1e1574e6a00a2b4c1b962f473f8e5e6002b1a112b7" Dec 13 08:33:29 crc kubenswrapper[4971]: I1213 08:33:29.762892 4971 scope.go:117] "RemoveContainer" containerID="64a6e9263fe1529f02ae02241f800cb4df1132e31e4c8f2c0c138d650646ad67" Dec 13 08:33:29 crc kubenswrapper[4971]: I1213 08:33:29.782451 4971 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/cea3f937-1bc0-4dfe-bb7d-a74fe6218006-catalog-content\") on node \"crc\" DevicePath \"\"" Dec 13 08:33:29 crc kubenswrapper[4971]: I1213 08:33:29.812794 4971 scope.go:117] "RemoveContainer" containerID="5d3bf4bd59b494a3b9725ac40fd9e72573afe95c3b614ff7ca2adb7fb7decfe1" Dec 13 08:33:29 crc kubenswrapper[4971]: E1213 08:33:29.813582 4971 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"5d3bf4bd59b494a3b9725ac40fd9e72573afe95c3b614ff7ca2adb7fb7decfe1\": container with ID starting with 5d3bf4bd59b494a3b9725ac40fd9e72573afe95c3b614ff7ca2adb7fb7decfe1 not found: ID does not exist" containerID="5d3bf4bd59b494a3b9725ac40fd9e72573afe95c3b614ff7ca2adb7fb7decfe1" Dec 13 08:33:29 crc kubenswrapper[4971]: I1213 08:33:29.813633 4971 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"5d3bf4bd59b494a3b9725ac40fd9e72573afe95c3b614ff7ca2adb7fb7decfe1"} err="failed to get container status \"5d3bf4bd59b494a3b9725ac40fd9e72573afe95c3b614ff7ca2adb7fb7decfe1\": rpc error: code = NotFound desc = could not find container \"5d3bf4bd59b494a3b9725ac40fd9e72573afe95c3b614ff7ca2adb7fb7decfe1\": container with ID starting with 5d3bf4bd59b494a3b9725ac40fd9e72573afe95c3b614ff7ca2adb7fb7decfe1 not found: ID does not exist" Dec 13 08:33:29 crc kubenswrapper[4971]: I1213 08:33:29.813669 4971 scope.go:117] "RemoveContainer" containerID="2df9843215147a593371cb1e1574e6a00a2b4c1b962f473f8e5e6002b1a112b7" Dec 13 08:33:29 crc kubenswrapper[4971]: E1213 08:33:29.817173 4971 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"2df9843215147a593371cb1e1574e6a00a2b4c1b962f473f8e5e6002b1a112b7\": container with ID starting with 2df9843215147a593371cb1e1574e6a00a2b4c1b962f473f8e5e6002b1a112b7 not found: ID does not exist" containerID="2df9843215147a593371cb1e1574e6a00a2b4c1b962f473f8e5e6002b1a112b7" Dec 13 08:33:29 crc kubenswrapper[4971]: I1213 08:33:29.817318 4971 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"2df9843215147a593371cb1e1574e6a00a2b4c1b962f473f8e5e6002b1a112b7"} err="failed to get container status \"2df9843215147a593371cb1e1574e6a00a2b4c1b962f473f8e5e6002b1a112b7\": rpc error: code = NotFound desc = could not find container \"2df9843215147a593371cb1e1574e6a00a2b4c1b962f473f8e5e6002b1a112b7\": container with ID starting with 2df9843215147a593371cb1e1574e6a00a2b4c1b962f473f8e5e6002b1a112b7 not found: ID does not exist" Dec 13 08:33:29 crc kubenswrapper[4971]: I1213 08:33:29.817358 4971 scope.go:117] "RemoveContainer" containerID="64a6e9263fe1529f02ae02241f800cb4df1132e31e4c8f2c0c138d650646ad67" Dec 13 08:33:29 crc kubenswrapper[4971]: E1213 08:33:29.817886 4971 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"64a6e9263fe1529f02ae02241f800cb4df1132e31e4c8f2c0c138d650646ad67\": container with ID starting with 64a6e9263fe1529f02ae02241f800cb4df1132e31e4c8f2c0c138d650646ad67 not found: ID does not exist" containerID="64a6e9263fe1529f02ae02241f800cb4df1132e31e4c8f2c0c138d650646ad67" Dec 13 08:33:29 crc kubenswrapper[4971]: I1213 08:33:29.817943 4971 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"64a6e9263fe1529f02ae02241f800cb4df1132e31e4c8f2c0c138d650646ad67"} err="failed to get container status \"64a6e9263fe1529f02ae02241f800cb4df1132e31e4c8f2c0c138d650646ad67\": rpc error: code = NotFound desc = could not find container \"64a6e9263fe1529f02ae02241f800cb4df1132e31e4c8f2c0c138d650646ad67\": container with ID starting with 64a6e9263fe1529f02ae02241f800cb4df1132e31e4c8f2c0c138d650646ad67 not found: ID does not exist" Dec 13 08:33:30 crc kubenswrapper[4971]: I1213 08:33:30.042283 4971 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-h48z2"] Dec 13 08:33:30 crc kubenswrapper[4971]: I1213 08:33:30.052812 4971 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-operators-h48z2"] Dec 13 08:33:31 crc kubenswrapper[4971]: I1213 08:33:31.780231 4971 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="cea3f937-1bc0-4dfe-bb7d-a74fe6218006" path="/var/lib/kubelet/pods/cea3f937-1bc0-4dfe-bb7d-a74fe6218006/volumes" Dec 13 08:33:46 crc kubenswrapper[4971]: I1213 08:33:46.153326 4971 patch_prober.go:28] interesting pod/machine-config-daemon-82xjz container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 13 08:33:46 crc kubenswrapper[4971]: I1213 08:33:46.154282 4971 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-82xjz" podUID="e0c4af12-edda-41d7-9b44-f87396174ef2" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 13 08:33:46 crc kubenswrapper[4971]: I1213 08:33:46.154337 4971 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-82xjz" Dec 13 08:33:46 crc kubenswrapper[4971]: I1213 08:33:46.155921 4971 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"87e58ec20b03b7a861da80d90f34f635ea8cc5825e7d209bb6d579dced58891c"} pod="openshift-machine-config-operator/machine-config-daemon-82xjz" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Dec 13 08:33:46 crc kubenswrapper[4971]: I1213 08:33:46.155998 4971 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-82xjz" podUID="e0c4af12-edda-41d7-9b44-f87396174ef2" containerName="machine-config-daemon" containerID="cri-o://87e58ec20b03b7a861da80d90f34f635ea8cc5825e7d209bb6d579dced58891c" gracePeriod=600 Dec 13 08:33:46 crc kubenswrapper[4971]: I1213 08:33:46.915531 4971 generic.go:334] "Generic (PLEG): container finished" podID="e0c4af12-edda-41d7-9b44-f87396174ef2" containerID="87e58ec20b03b7a861da80d90f34f635ea8cc5825e7d209bb6d579dced58891c" exitCode=0 Dec 13 08:33:46 crc kubenswrapper[4971]: I1213 08:33:46.915631 4971 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-82xjz" event={"ID":"e0c4af12-edda-41d7-9b44-f87396174ef2","Type":"ContainerDied","Data":"87e58ec20b03b7a861da80d90f34f635ea8cc5825e7d209bb6d579dced58891c"} Dec 13 08:33:46 crc kubenswrapper[4971]: I1213 08:33:46.916541 4971 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-82xjz" event={"ID":"e0c4af12-edda-41d7-9b44-f87396174ef2","Type":"ContainerStarted","Data":"f397fe912a74e38413c0149a7e7ca0ed1cff4acbea98dcbfa69153e27d3fdaa0"} Dec 13 08:33:46 crc kubenswrapper[4971]: I1213 08:33:46.916565 4971 scope.go:117] "RemoveContainer" containerID="bc1c06c3bcebf9c00aee813c7819c279356be670c515e1ea302b50b8618bc001" Dec 13 08:34:41 crc kubenswrapper[4971]: I1213 08:34:41.437144 4971 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/community-operators-fg8q2"] Dec 13 08:34:41 crc kubenswrapper[4971]: E1213 08:34:41.438817 4971 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="cea3f937-1bc0-4dfe-bb7d-a74fe6218006" containerName="registry-server" Dec 13 08:34:41 crc kubenswrapper[4971]: I1213 08:34:41.438834 4971 state_mem.go:107] "Deleted CPUSet assignment" podUID="cea3f937-1bc0-4dfe-bb7d-a74fe6218006" containerName="registry-server" Dec 13 08:34:41 crc kubenswrapper[4971]: E1213 08:34:41.438851 4971 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="cea3f937-1bc0-4dfe-bb7d-a74fe6218006" containerName="extract-utilities" Dec 13 08:34:41 crc kubenswrapper[4971]: I1213 08:34:41.438858 4971 state_mem.go:107] "Deleted CPUSet assignment" podUID="cea3f937-1bc0-4dfe-bb7d-a74fe6218006" containerName="extract-utilities" Dec 13 08:34:41 crc kubenswrapper[4971]: E1213 08:34:41.438866 4971 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="6eaf2f74-3685-4128-a19c-1284823b9b64" containerName="extract-content" Dec 13 08:34:41 crc kubenswrapper[4971]: I1213 08:34:41.438879 4971 state_mem.go:107] "Deleted CPUSet assignment" podUID="6eaf2f74-3685-4128-a19c-1284823b9b64" containerName="extract-content" Dec 13 08:34:41 crc kubenswrapper[4971]: E1213 08:34:41.438905 4971 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="6eaf2f74-3685-4128-a19c-1284823b9b64" containerName="extract-utilities" Dec 13 08:34:41 crc kubenswrapper[4971]: I1213 08:34:41.438911 4971 state_mem.go:107] "Deleted CPUSet assignment" podUID="6eaf2f74-3685-4128-a19c-1284823b9b64" containerName="extract-utilities" Dec 13 08:34:41 crc kubenswrapper[4971]: E1213 08:34:41.438928 4971 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="6eaf2f74-3685-4128-a19c-1284823b9b64" containerName="registry-server" Dec 13 08:34:41 crc kubenswrapper[4971]: I1213 08:34:41.438935 4971 state_mem.go:107] "Deleted CPUSet assignment" podUID="6eaf2f74-3685-4128-a19c-1284823b9b64" containerName="registry-server" Dec 13 08:34:41 crc kubenswrapper[4971]: E1213 08:34:41.438946 4971 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="cea3f937-1bc0-4dfe-bb7d-a74fe6218006" containerName="extract-content" Dec 13 08:34:41 crc kubenswrapper[4971]: I1213 08:34:41.438952 4971 state_mem.go:107] "Deleted CPUSet assignment" podUID="cea3f937-1bc0-4dfe-bb7d-a74fe6218006" containerName="extract-content" Dec 13 08:34:41 crc kubenswrapper[4971]: I1213 08:34:41.439143 4971 memory_manager.go:354] "RemoveStaleState removing state" podUID="6eaf2f74-3685-4128-a19c-1284823b9b64" containerName="registry-server" Dec 13 08:34:41 crc kubenswrapper[4971]: I1213 08:34:41.439163 4971 memory_manager.go:354] "RemoveStaleState removing state" podUID="cea3f937-1bc0-4dfe-bb7d-a74fe6218006" containerName="registry-server" Dec 13 08:34:41 crc kubenswrapper[4971]: I1213 08:34:41.442355 4971 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-fg8q2" Dec 13 08:34:41 crc kubenswrapper[4971]: I1213 08:34:41.456408 4971 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-fg8q2"] Dec 13 08:34:41 crc kubenswrapper[4971]: I1213 08:34:41.492568 4971 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/80462a80-331f-45fb-8980-d54dc6fcecbc-catalog-content\") pod \"community-operators-fg8q2\" (UID: \"80462a80-331f-45fb-8980-d54dc6fcecbc\") " pod="openshift-marketplace/community-operators-fg8q2" Dec 13 08:34:41 crc kubenswrapper[4971]: I1213 08:34:41.492745 4971 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-h927h\" (UniqueName: \"kubernetes.io/projected/80462a80-331f-45fb-8980-d54dc6fcecbc-kube-api-access-h927h\") pod \"community-operators-fg8q2\" (UID: \"80462a80-331f-45fb-8980-d54dc6fcecbc\") " pod="openshift-marketplace/community-operators-fg8q2" Dec 13 08:34:41 crc kubenswrapper[4971]: I1213 08:34:41.492901 4971 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/80462a80-331f-45fb-8980-d54dc6fcecbc-utilities\") pod \"community-operators-fg8q2\" (UID: \"80462a80-331f-45fb-8980-d54dc6fcecbc\") " pod="openshift-marketplace/community-operators-fg8q2" Dec 13 08:34:41 crc kubenswrapper[4971]: I1213 08:34:41.593920 4971 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/80462a80-331f-45fb-8980-d54dc6fcecbc-catalog-content\") pod \"community-operators-fg8q2\" (UID: \"80462a80-331f-45fb-8980-d54dc6fcecbc\") " pod="openshift-marketplace/community-operators-fg8q2" Dec 13 08:34:41 crc kubenswrapper[4971]: I1213 08:34:41.594021 4971 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-h927h\" (UniqueName: \"kubernetes.io/projected/80462a80-331f-45fb-8980-d54dc6fcecbc-kube-api-access-h927h\") pod \"community-operators-fg8q2\" (UID: \"80462a80-331f-45fb-8980-d54dc6fcecbc\") " pod="openshift-marketplace/community-operators-fg8q2" Dec 13 08:34:41 crc kubenswrapper[4971]: I1213 08:34:41.594072 4971 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/80462a80-331f-45fb-8980-d54dc6fcecbc-utilities\") pod \"community-operators-fg8q2\" (UID: \"80462a80-331f-45fb-8980-d54dc6fcecbc\") " pod="openshift-marketplace/community-operators-fg8q2" Dec 13 08:34:41 crc kubenswrapper[4971]: I1213 08:34:41.595434 4971 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/80462a80-331f-45fb-8980-d54dc6fcecbc-catalog-content\") pod \"community-operators-fg8q2\" (UID: \"80462a80-331f-45fb-8980-d54dc6fcecbc\") " pod="openshift-marketplace/community-operators-fg8q2" Dec 13 08:34:41 crc kubenswrapper[4971]: I1213 08:34:41.595459 4971 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/80462a80-331f-45fb-8980-d54dc6fcecbc-utilities\") pod \"community-operators-fg8q2\" (UID: \"80462a80-331f-45fb-8980-d54dc6fcecbc\") " pod="openshift-marketplace/community-operators-fg8q2" Dec 13 08:34:41 crc kubenswrapper[4971]: I1213 08:34:41.626090 4971 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-h927h\" (UniqueName: \"kubernetes.io/projected/80462a80-331f-45fb-8980-d54dc6fcecbc-kube-api-access-h927h\") pod \"community-operators-fg8q2\" (UID: \"80462a80-331f-45fb-8980-d54dc6fcecbc\") " pod="openshift-marketplace/community-operators-fg8q2" Dec 13 08:34:41 crc kubenswrapper[4971]: I1213 08:34:41.782378 4971 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-fg8q2" Dec 13 08:34:42 crc kubenswrapper[4971]: I1213 08:34:42.424923 4971 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-fg8q2"] Dec 13 08:34:42 crc kubenswrapper[4971]: I1213 08:34:42.513983 4971 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-fg8q2" event={"ID":"80462a80-331f-45fb-8980-d54dc6fcecbc","Type":"ContainerStarted","Data":"68d398bb8425dd2bdbaaf0ab39b6a879bcecd05b79dcd8dc8b10655bf1ae95ff"} Dec 13 08:34:43 crc kubenswrapper[4971]: I1213 08:34:43.527218 4971 generic.go:334] "Generic (PLEG): container finished" podID="80462a80-331f-45fb-8980-d54dc6fcecbc" containerID="885dd6ed00fa362f1b820f238ed944f5f441a629760b53b0f571dfca431e68ae" exitCode=0 Dec 13 08:34:43 crc kubenswrapper[4971]: I1213 08:34:43.528070 4971 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-fg8q2" event={"ID":"80462a80-331f-45fb-8980-d54dc6fcecbc","Type":"ContainerDied","Data":"885dd6ed00fa362f1b820f238ed944f5f441a629760b53b0f571dfca431e68ae"} Dec 13 08:34:45 crc kubenswrapper[4971]: I1213 08:34:45.556066 4971 generic.go:334] "Generic (PLEG): container finished" podID="80462a80-331f-45fb-8980-d54dc6fcecbc" containerID="d3a58563bb6466a27f26d9a9f77eed24eb3363edb0c6f49bd5767619e3fb7d9a" exitCode=0 Dec 13 08:34:45 crc kubenswrapper[4971]: I1213 08:34:45.556173 4971 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-fg8q2" event={"ID":"80462a80-331f-45fb-8980-d54dc6fcecbc","Type":"ContainerDied","Data":"d3a58563bb6466a27f26d9a9f77eed24eb3363edb0c6f49bd5767619e3fb7d9a"} Dec 13 08:34:47 crc kubenswrapper[4971]: I1213 08:34:47.583569 4971 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-fg8q2" event={"ID":"80462a80-331f-45fb-8980-d54dc6fcecbc","Type":"ContainerStarted","Data":"b73d2a9c79f0d4cac3273f3c194c625261b337fd586e33c669c3b07304a67b30"} Dec 13 08:34:47 crc kubenswrapper[4971]: I1213 08:34:47.617943 4971 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/community-operators-fg8q2" podStartSLOduration=3.723604542 podStartE2EDuration="6.617913395s" podCreationTimestamp="2025-12-13 08:34:41 +0000 UTC" firstStartedPulling="2025-12-13 08:34:43.531850417 +0000 UTC m=+6340.136259865" lastFinishedPulling="2025-12-13 08:34:46.42615927 +0000 UTC m=+6343.030568718" observedRunningTime="2025-12-13 08:34:47.608643728 +0000 UTC m=+6344.213053176" watchObservedRunningTime="2025-12-13 08:34:47.617913395 +0000 UTC m=+6344.222322843" Dec 13 08:34:51 crc kubenswrapper[4971]: I1213 08:34:51.783249 4971 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/community-operators-fg8q2" Dec 13 08:34:51 crc kubenswrapper[4971]: I1213 08:34:51.784158 4971 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/community-operators-fg8q2" Dec 13 08:34:51 crc kubenswrapper[4971]: I1213 08:34:51.836428 4971 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/community-operators-fg8q2" Dec 13 08:34:52 crc kubenswrapper[4971]: I1213 08:34:52.701285 4971 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/community-operators-fg8q2" Dec 13 08:34:52 crc kubenswrapper[4971]: I1213 08:34:52.768505 4971 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-fg8q2"] Dec 13 08:34:54 crc kubenswrapper[4971]: I1213 08:34:54.691342 4971 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/community-operators-fg8q2" podUID="80462a80-331f-45fb-8980-d54dc6fcecbc" containerName="registry-server" containerID="cri-o://b73d2a9c79f0d4cac3273f3c194c625261b337fd586e33c669c3b07304a67b30" gracePeriod=2 Dec 13 08:34:55 crc kubenswrapper[4971]: I1213 08:34:55.705713 4971 generic.go:334] "Generic (PLEG): container finished" podID="80462a80-331f-45fb-8980-d54dc6fcecbc" containerID="b73d2a9c79f0d4cac3273f3c194c625261b337fd586e33c669c3b07304a67b30" exitCode=0 Dec 13 08:34:55 crc kubenswrapper[4971]: I1213 08:34:55.705787 4971 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-fg8q2" event={"ID":"80462a80-331f-45fb-8980-d54dc6fcecbc","Type":"ContainerDied","Data":"b73d2a9c79f0d4cac3273f3c194c625261b337fd586e33c669c3b07304a67b30"} Dec 13 08:34:57 crc kubenswrapper[4971]: I1213 08:34:57.480150 4971 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-fg8q2" Dec 13 08:34:57 crc kubenswrapper[4971]: I1213 08:34:57.621302 4971 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-h927h\" (UniqueName: \"kubernetes.io/projected/80462a80-331f-45fb-8980-d54dc6fcecbc-kube-api-access-h927h\") pod \"80462a80-331f-45fb-8980-d54dc6fcecbc\" (UID: \"80462a80-331f-45fb-8980-d54dc6fcecbc\") " Dec 13 08:34:57 crc kubenswrapper[4971]: I1213 08:34:57.621768 4971 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/80462a80-331f-45fb-8980-d54dc6fcecbc-utilities\") pod \"80462a80-331f-45fb-8980-d54dc6fcecbc\" (UID: \"80462a80-331f-45fb-8980-d54dc6fcecbc\") " Dec 13 08:34:57 crc kubenswrapper[4971]: I1213 08:34:57.621894 4971 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/80462a80-331f-45fb-8980-d54dc6fcecbc-catalog-content\") pod \"80462a80-331f-45fb-8980-d54dc6fcecbc\" (UID: \"80462a80-331f-45fb-8980-d54dc6fcecbc\") " Dec 13 08:34:57 crc kubenswrapper[4971]: I1213 08:34:57.623081 4971 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/80462a80-331f-45fb-8980-d54dc6fcecbc-utilities" (OuterVolumeSpecName: "utilities") pod "80462a80-331f-45fb-8980-d54dc6fcecbc" (UID: "80462a80-331f-45fb-8980-d54dc6fcecbc"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 13 08:34:57 crc kubenswrapper[4971]: I1213 08:34:57.638816 4971 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/80462a80-331f-45fb-8980-d54dc6fcecbc-kube-api-access-h927h" (OuterVolumeSpecName: "kube-api-access-h927h") pod "80462a80-331f-45fb-8980-d54dc6fcecbc" (UID: "80462a80-331f-45fb-8980-d54dc6fcecbc"). InnerVolumeSpecName "kube-api-access-h927h". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 13 08:34:57 crc kubenswrapper[4971]: I1213 08:34:57.698622 4971 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/80462a80-331f-45fb-8980-d54dc6fcecbc-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "80462a80-331f-45fb-8980-d54dc6fcecbc" (UID: "80462a80-331f-45fb-8980-d54dc6fcecbc"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 13 08:34:57 crc kubenswrapper[4971]: I1213 08:34:57.725390 4971 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-h927h\" (UniqueName: \"kubernetes.io/projected/80462a80-331f-45fb-8980-d54dc6fcecbc-kube-api-access-h927h\") on node \"crc\" DevicePath \"\"" Dec 13 08:34:57 crc kubenswrapper[4971]: I1213 08:34:57.725741 4971 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/80462a80-331f-45fb-8980-d54dc6fcecbc-utilities\") on node \"crc\" DevicePath \"\"" Dec 13 08:34:57 crc kubenswrapper[4971]: I1213 08:34:57.725861 4971 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/80462a80-331f-45fb-8980-d54dc6fcecbc-catalog-content\") on node \"crc\" DevicePath \"\"" Dec 13 08:34:57 crc kubenswrapper[4971]: I1213 08:34:57.731552 4971 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-fg8q2" event={"ID":"80462a80-331f-45fb-8980-d54dc6fcecbc","Type":"ContainerDied","Data":"68d398bb8425dd2bdbaaf0ab39b6a879bcecd05b79dcd8dc8b10655bf1ae95ff"} Dec 13 08:34:57 crc kubenswrapper[4971]: I1213 08:34:57.732046 4971 scope.go:117] "RemoveContainer" containerID="b73d2a9c79f0d4cac3273f3c194c625261b337fd586e33c669c3b07304a67b30" Dec 13 08:34:57 crc kubenswrapper[4971]: I1213 08:34:57.731745 4971 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-fg8q2" Dec 13 08:34:57 crc kubenswrapper[4971]: I1213 08:34:57.825723 4971 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-fg8q2"] Dec 13 08:34:57 crc kubenswrapper[4971]: I1213 08:34:57.839014 4971 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/community-operators-fg8q2"] Dec 13 08:34:57 crc kubenswrapper[4971]: I1213 08:34:57.880129 4971 scope.go:117] "RemoveContainer" containerID="d3a58563bb6466a27f26d9a9f77eed24eb3363edb0c6f49bd5767619e3fb7d9a" Dec 13 08:34:57 crc kubenswrapper[4971]: I1213 08:34:57.918612 4971 scope.go:117] "RemoveContainer" containerID="885dd6ed00fa362f1b820f238ed944f5f441a629760b53b0f571dfca431e68ae" Dec 13 08:34:59 crc kubenswrapper[4971]: I1213 08:34:59.782193 4971 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="80462a80-331f-45fb-8980-d54dc6fcecbc" path="/var/lib/kubelet/pods/80462a80-331f-45fb-8980-d54dc6fcecbc/volumes" Dec 13 08:35:46 crc kubenswrapper[4971]: I1213 08:35:46.153398 4971 patch_prober.go:28] interesting pod/machine-config-daemon-82xjz container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 13 08:35:46 crc kubenswrapper[4971]: I1213 08:35:46.154023 4971 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-82xjz" podUID="e0c4af12-edda-41d7-9b44-f87396174ef2" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 13 08:36:16 crc kubenswrapper[4971]: I1213 08:36:16.153888 4971 patch_prober.go:28] interesting pod/machine-config-daemon-82xjz container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 13 08:36:16 crc kubenswrapper[4971]: I1213 08:36:16.155098 4971 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-82xjz" podUID="e0c4af12-edda-41d7-9b44-f87396174ef2" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 13 08:36:46 crc kubenswrapper[4971]: I1213 08:36:46.154030 4971 patch_prober.go:28] interesting pod/machine-config-daemon-82xjz container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 13 08:36:46 crc kubenswrapper[4971]: I1213 08:36:46.154997 4971 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-82xjz" podUID="e0c4af12-edda-41d7-9b44-f87396174ef2" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 13 08:36:46 crc kubenswrapper[4971]: I1213 08:36:46.155048 4971 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-82xjz" Dec 13 08:36:46 crc kubenswrapper[4971]: I1213 08:36:46.156220 4971 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"f397fe912a74e38413c0149a7e7ca0ed1cff4acbea98dcbfa69153e27d3fdaa0"} pod="openshift-machine-config-operator/machine-config-daemon-82xjz" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Dec 13 08:36:46 crc kubenswrapper[4971]: I1213 08:36:46.156294 4971 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-82xjz" podUID="e0c4af12-edda-41d7-9b44-f87396174ef2" containerName="machine-config-daemon" containerID="cri-o://f397fe912a74e38413c0149a7e7ca0ed1cff4acbea98dcbfa69153e27d3fdaa0" gracePeriod=600 Dec 13 08:36:46 crc kubenswrapper[4971]: E1213 08:36:46.300049 4971 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-82xjz_openshift-machine-config-operator(e0c4af12-edda-41d7-9b44-f87396174ef2)\"" pod="openshift-machine-config-operator/machine-config-daemon-82xjz" podUID="e0c4af12-edda-41d7-9b44-f87396174ef2" Dec 13 08:36:46 crc kubenswrapper[4971]: I1213 08:36:46.562273 4971 generic.go:334] "Generic (PLEG): container finished" podID="e0c4af12-edda-41d7-9b44-f87396174ef2" containerID="f397fe912a74e38413c0149a7e7ca0ed1cff4acbea98dcbfa69153e27d3fdaa0" exitCode=0 Dec 13 08:36:46 crc kubenswrapper[4971]: I1213 08:36:46.562354 4971 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-82xjz" event={"ID":"e0c4af12-edda-41d7-9b44-f87396174ef2","Type":"ContainerDied","Data":"f397fe912a74e38413c0149a7e7ca0ed1cff4acbea98dcbfa69153e27d3fdaa0"} Dec 13 08:36:46 crc kubenswrapper[4971]: I1213 08:36:46.562426 4971 scope.go:117] "RemoveContainer" containerID="87e58ec20b03b7a861da80d90f34f635ea8cc5825e7d209bb6d579dced58891c" Dec 13 08:36:46 crc kubenswrapper[4971]: I1213 08:36:46.568099 4971 scope.go:117] "RemoveContainer" containerID="f397fe912a74e38413c0149a7e7ca0ed1cff4acbea98dcbfa69153e27d3fdaa0" Dec 13 08:36:46 crc kubenswrapper[4971]: E1213 08:36:46.570724 4971 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-82xjz_openshift-machine-config-operator(e0c4af12-edda-41d7-9b44-f87396174ef2)\"" pod="openshift-machine-config-operator/machine-config-daemon-82xjz" podUID="e0c4af12-edda-41d7-9b44-f87396174ef2" Dec 13 08:36:59 crc kubenswrapper[4971]: I1213 08:36:59.802316 4971 scope.go:117] "RemoveContainer" containerID="f397fe912a74e38413c0149a7e7ca0ed1cff4acbea98dcbfa69153e27d3fdaa0" Dec 13 08:36:59 crc kubenswrapper[4971]: E1213 08:36:59.803492 4971 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-82xjz_openshift-machine-config-operator(e0c4af12-edda-41d7-9b44-f87396174ef2)\"" pod="openshift-machine-config-operator/machine-config-daemon-82xjz" podUID="e0c4af12-edda-41d7-9b44-f87396174ef2" Dec 13 08:37:12 crc kubenswrapper[4971]: I1213 08:37:12.770088 4971 scope.go:117] "RemoveContainer" containerID="f397fe912a74e38413c0149a7e7ca0ed1cff4acbea98dcbfa69153e27d3fdaa0" Dec 13 08:37:12 crc kubenswrapper[4971]: E1213 08:37:12.771659 4971 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-82xjz_openshift-machine-config-operator(e0c4af12-edda-41d7-9b44-f87396174ef2)\"" pod="openshift-machine-config-operator/machine-config-daemon-82xjz" podUID="e0c4af12-edda-41d7-9b44-f87396174ef2" Dec 13 08:37:25 crc kubenswrapper[4971]: I1213 08:37:25.768732 4971 scope.go:117] "RemoveContainer" containerID="f397fe912a74e38413c0149a7e7ca0ed1cff4acbea98dcbfa69153e27d3fdaa0" Dec 13 08:37:25 crc kubenswrapper[4971]: E1213 08:37:25.769941 4971 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-82xjz_openshift-machine-config-operator(e0c4af12-edda-41d7-9b44-f87396174ef2)\"" pod="openshift-machine-config-operator/machine-config-daemon-82xjz" podUID="e0c4af12-edda-41d7-9b44-f87396174ef2" Dec 13 08:37:39 crc kubenswrapper[4971]: I1213 08:37:39.769298 4971 scope.go:117] "RemoveContainer" containerID="f397fe912a74e38413c0149a7e7ca0ed1cff4acbea98dcbfa69153e27d3fdaa0" Dec 13 08:37:39 crc kubenswrapper[4971]: E1213 08:37:39.770608 4971 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-82xjz_openshift-machine-config-operator(e0c4af12-edda-41d7-9b44-f87396174ef2)\"" pod="openshift-machine-config-operator/machine-config-daemon-82xjz" podUID="e0c4af12-edda-41d7-9b44-f87396174ef2" Dec 13 08:37:53 crc kubenswrapper[4971]: I1213 08:37:53.778726 4971 scope.go:117] "RemoveContainer" containerID="f397fe912a74e38413c0149a7e7ca0ed1cff4acbea98dcbfa69153e27d3fdaa0" Dec 13 08:37:53 crc kubenswrapper[4971]: E1213 08:37:53.780046 4971 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-82xjz_openshift-machine-config-operator(e0c4af12-edda-41d7-9b44-f87396174ef2)\"" pod="openshift-machine-config-operator/machine-config-daemon-82xjz" podUID="e0c4af12-edda-41d7-9b44-f87396174ef2" Dec 13 08:38:06 crc kubenswrapper[4971]: I1213 08:38:06.769822 4971 scope.go:117] "RemoveContainer" containerID="f397fe912a74e38413c0149a7e7ca0ed1cff4acbea98dcbfa69153e27d3fdaa0" Dec 13 08:38:06 crc kubenswrapper[4971]: E1213 08:38:06.770908 4971 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-82xjz_openshift-machine-config-operator(e0c4af12-edda-41d7-9b44-f87396174ef2)\"" pod="openshift-machine-config-operator/machine-config-daemon-82xjz" podUID="e0c4af12-edda-41d7-9b44-f87396174ef2" Dec 13 08:38:14 crc kubenswrapper[4971]: I1213 08:38:14.902872 4971 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/certified-operators-r4l8j"] Dec 13 08:38:14 crc kubenswrapper[4971]: E1213 08:38:14.904289 4971 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="80462a80-331f-45fb-8980-d54dc6fcecbc" containerName="registry-server" Dec 13 08:38:14 crc kubenswrapper[4971]: I1213 08:38:14.904312 4971 state_mem.go:107] "Deleted CPUSet assignment" podUID="80462a80-331f-45fb-8980-d54dc6fcecbc" containerName="registry-server" Dec 13 08:38:14 crc kubenswrapper[4971]: E1213 08:38:14.904328 4971 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="80462a80-331f-45fb-8980-d54dc6fcecbc" containerName="extract-utilities" Dec 13 08:38:14 crc kubenswrapper[4971]: I1213 08:38:14.904335 4971 state_mem.go:107] "Deleted CPUSet assignment" podUID="80462a80-331f-45fb-8980-d54dc6fcecbc" containerName="extract-utilities" Dec 13 08:38:14 crc kubenswrapper[4971]: E1213 08:38:14.904385 4971 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="80462a80-331f-45fb-8980-d54dc6fcecbc" containerName="extract-content" Dec 13 08:38:14 crc kubenswrapper[4971]: I1213 08:38:14.904391 4971 state_mem.go:107] "Deleted CPUSet assignment" podUID="80462a80-331f-45fb-8980-d54dc6fcecbc" containerName="extract-content" Dec 13 08:38:14 crc kubenswrapper[4971]: I1213 08:38:14.908889 4971 memory_manager.go:354] "RemoveStaleState removing state" podUID="80462a80-331f-45fb-8980-d54dc6fcecbc" containerName="registry-server" Dec 13 08:38:14 crc kubenswrapper[4971]: I1213 08:38:14.910931 4971 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-r4l8j" Dec 13 08:38:14 crc kubenswrapper[4971]: I1213 08:38:14.936648 4971 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-r4l8j"] Dec 13 08:38:15 crc kubenswrapper[4971]: I1213 08:38:15.032921 4971 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-dfnkz\" (UniqueName: \"kubernetes.io/projected/102659f9-0a07-48c9-b12b-5aa22f2a5fda-kube-api-access-dfnkz\") pod \"certified-operators-r4l8j\" (UID: \"102659f9-0a07-48c9-b12b-5aa22f2a5fda\") " pod="openshift-marketplace/certified-operators-r4l8j" Dec 13 08:38:15 crc kubenswrapper[4971]: I1213 08:38:15.033465 4971 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/102659f9-0a07-48c9-b12b-5aa22f2a5fda-utilities\") pod \"certified-operators-r4l8j\" (UID: \"102659f9-0a07-48c9-b12b-5aa22f2a5fda\") " pod="openshift-marketplace/certified-operators-r4l8j" Dec 13 08:38:15 crc kubenswrapper[4971]: I1213 08:38:15.033634 4971 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/102659f9-0a07-48c9-b12b-5aa22f2a5fda-catalog-content\") pod \"certified-operators-r4l8j\" (UID: \"102659f9-0a07-48c9-b12b-5aa22f2a5fda\") " pod="openshift-marketplace/certified-operators-r4l8j" Dec 13 08:38:15 crc kubenswrapper[4971]: I1213 08:38:15.136192 4971 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-dfnkz\" (UniqueName: \"kubernetes.io/projected/102659f9-0a07-48c9-b12b-5aa22f2a5fda-kube-api-access-dfnkz\") pod \"certified-operators-r4l8j\" (UID: \"102659f9-0a07-48c9-b12b-5aa22f2a5fda\") " pod="openshift-marketplace/certified-operators-r4l8j" Dec 13 08:38:15 crc kubenswrapper[4971]: I1213 08:38:15.136377 4971 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/102659f9-0a07-48c9-b12b-5aa22f2a5fda-utilities\") pod \"certified-operators-r4l8j\" (UID: \"102659f9-0a07-48c9-b12b-5aa22f2a5fda\") " pod="openshift-marketplace/certified-operators-r4l8j" Dec 13 08:38:15 crc kubenswrapper[4971]: I1213 08:38:15.136406 4971 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/102659f9-0a07-48c9-b12b-5aa22f2a5fda-catalog-content\") pod \"certified-operators-r4l8j\" (UID: \"102659f9-0a07-48c9-b12b-5aa22f2a5fda\") " pod="openshift-marketplace/certified-operators-r4l8j" Dec 13 08:38:15 crc kubenswrapper[4971]: I1213 08:38:15.137023 4971 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/102659f9-0a07-48c9-b12b-5aa22f2a5fda-utilities\") pod \"certified-operators-r4l8j\" (UID: \"102659f9-0a07-48c9-b12b-5aa22f2a5fda\") " pod="openshift-marketplace/certified-operators-r4l8j" Dec 13 08:38:15 crc kubenswrapper[4971]: I1213 08:38:15.137046 4971 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/102659f9-0a07-48c9-b12b-5aa22f2a5fda-catalog-content\") pod \"certified-operators-r4l8j\" (UID: \"102659f9-0a07-48c9-b12b-5aa22f2a5fda\") " pod="openshift-marketplace/certified-operators-r4l8j" Dec 13 08:38:15 crc kubenswrapper[4971]: I1213 08:38:15.173345 4971 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-dfnkz\" (UniqueName: \"kubernetes.io/projected/102659f9-0a07-48c9-b12b-5aa22f2a5fda-kube-api-access-dfnkz\") pod \"certified-operators-r4l8j\" (UID: \"102659f9-0a07-48c9-b12b-5aa22f2a5fda\") " pod="openshift-marketplace/certified-operators-r4l8j" Dec 13 08:38:15 crc kubenswrapper[4971]: I1213 08:38:15.237062 4971 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-r4l8j" Dec 13 08:38:15 crc kubenswrapper[4971]: I1213 08:38:15.895586 4971 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-r4l8j"] Dec 13 08:38:16 crc kubenswrapper[4971]: I1213 08:38:16.864221 4971 generic.go:334] "Generic (PLEG): container finished" podID="102659f9-0a07-48c9-b12b-5aa22f2a5fda" containerID="543bc66edf0571d184c1217ca56eee753f966026a79aae3417a73406551c2838" exitCode=0 Dec 13 08:38:16 crc kubenswrapper[4971]: I1213 08:38:16.864349 4971 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-r4l8j" event={"ID":"102659f9-0a07-48c9-b12b-5aa22f2a5fda","Type":"ContainerDied","Data":"543bc66edf0571d184c1217ca56eee753f966026a79aae3417a73406551c2838"} Dec 13 08:38:16 crc kubenswrapper[4971]: I1213 08:38:16.865226 4971 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-r4l8j" event={"ID":"102659f9-0a07-48c9-b12b-5aa22f2a5fda","Type":"ContainerStarted","Data":"252700a554baa1706085d4bb44071ce2a9a0aae3014d4a662d8b1ba0b776fbb5"} Dec 13 08:38:16 crc kubenswrapper[4971]: I1213 08:38:16.870545 4971 provider.go:102] Refreshing cache for provider: *credentialprovider.defaultDockerConfigProvider Dec 13 08:38:18 crc kubenswrapper[4971]: I1213 08:38:18.896673 4971 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-r4l8j" event={"ID":"102659f9-0a07-48c9-b12b-5aa22f2a5fda","Type":"ContainerStarted","Data":"4ceb7a1738625662dd6cb11e27f6e1e36b671c7d3ae7bfb9489cfdde98ee7ed9"} Dec 13 08:38:20 crc kubenswrapper[4971]: I1213 08:38:20.922504 4971 generic.go:334] "Generic (PLEG): container finished" podID="102659f9-0a07-48c9-b12b-5aa22f2a5fda" containerID="4ceb7a1738625662dd6cb11e27f6e1e36b671c7d3ae7bfb9489cfdde98ee7ed9" exitCode=0 Dec 13 08:38:20 crc kubenswrapper[4971]: I1213 08:38:20.922593 4971 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-r4l8j" event={"ID":"102659f9-0a07-48c9-b12b-5aa22f2a5fda","Type":"ContainerDied","Data":"4ceb7a1738625662dd6cb11e27f6e1e36b671c7d3ae7bfb9489cfdde98ee7ed9"} Dec 13 08:38:21 crc kubenswrapper[4971]: I1213 08:38:21.769976 4971 scope.go:117] "RemoveContainer" containerID="f397fe912a74e38413c0149a7e7ca0ed1cff4acbea98dcbfa69153e27d3fdaa0" Dec 13 08:38:21 crc kubenswrapper[4971]: E1213 08:38:21.771285 4971 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-82xjz_openshift-machine-config-operator(e0c4af12-edda-41d7-9b44-f87396174ef2)\"" pod="openshift-machine-config-operator/machine-config-daemon-82xjz" podUID="e0c4af12-edda-41d7-9b44-f87396174ef2" Dec 13 08:38:21 crc kubenswrapper[4971]: I1213 08:38:21.976151 4971 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-r4l8j" event={"ID":"102659f9-0a07-48c9-b12b-5aa22f2a5fda","Type":"ContainerStarted","Data":"394641f16af20fe37971ac44c182491851a403642cca041c9f4624f56a8f098d"} Dec 13 08:38:22 crc kubenswrapper[4971]: I1213 08:38:22.007282 4971 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/certified-operators-r4l8j" podStartSLOduration=3.357504259 podStartE2EDuration="8.007252495s" podCreationTimestamp="2025-12-13 08:38:14 +0000 UTC" firstStartedPulling="2025-12-13 08:38:16.869858322 +0000 UTC m=+6553.474267770" lastFinishedPulling="2025-12-13 08:38:21.519606558 +0000 UTC m=+6558.124016006" observedRunningTime="2025-12-13 08:38:22.003408671 +0000 UTC m=+6558.607818119" watchObservedRunningTime="2025-12-13 08:38:22.007252495 +0000 UTC m=+6558.611661943" Dec 13 08:38:25 crc kubenswrapper[4971]: I1213 08:38:25.237590 4971 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/certified-operators-r4l8j" Dec 13 08:38:25 crc kubenswrapper[4971]: I1213 08:38:25.238622 4971 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/certified-operators-r4l8j" Dec 13 08:38:25 crc kubenswrapper[4971]: I1213 08:38:25.308145 4971 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/certified-operators-r4l8j" Dec 13 08:38:26 crc kubenswrapper[4971]: I1213 08:38:26.086650 4971 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/certified-operators-r4l8j" Dec 13 08:38:26 crc kubenswrapper[4971]: I1213 08:38:26.168886 4971 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-r4l8j"] Dec 13 08:38:28 crc kubenswrapper[4971]: I1213 08:38:28.067158 4971 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/certified-operators-r4l8j" podUID="102659f9-0a07-48c9-b12b-5aa22f2a5fda" containerName="registry-server" containerID="cri-o://394641f16af20fe37971ac44c182491851a403642cca041c9f4624f56a8f098d" gracePeriod=2 Dec 13 08:38:28 crc kubenswrapper[4971]: I1213 08:38:28.773330 4971 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-r4l8j" Dec 13 08:38:28 crc kubenswrapper[4971]: I1213 08:38:28.883781 4971 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/102659f9-0a07-48c9-b12b-5aa22f2a5fda-utilities\") pod \"102659f9-0a07-48c9-b12b-5aa22f2a5fda\" (UID: \"102659f9-0a07-48c9-b12b-5aa22f2a5fda\") " Dec 13 08:38:28 crc kubenswrapper[4971]: I1213 08:38:28.883965 4971 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-dfnkz\" (UniqueName: \"kubernetes.io/projected/102659f9-0a07-48c9-b12b-5aa22f2a5fda-kube-api-access-dfnkz\") pod \"102659f9-0a07-48c9-b12b-5aa22f2a5fda\" (UID: \"102659f9-0a07-48c9-b12b-5aa22f2a5fda\") " Dec 13 08:38:28 crc kubenswrapper[4971]: I1213 08:38:28.884028 4971 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/102659f9-0a07-48c9-b12b-5aa22f2a5fda-catalog-content\") pod \"102659f9-0a07-48c9-b12b-5aa22f2a5fda\" (UID: \"102659f9-0a07-48c9-b12b-5aa22f2a5fda\") " Dec 13 08:38:28 crc kubenswrapper[4971]: I1213 08:38:28.890641 4971 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/102659f9-0a07-48c9-b12b-5aa22f2a5fda-utilities" (OuterVolumeSpecName: "utilities") pod "102659f9-0a07-48c9-b12b-5aa22f2a5fda" (UID: "102659f9-0a07-48c9-b12b-5aa22f2a5fda"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 13 08:38:28 crc kubenswrapper[4971]: I1213 08:38:28.910027 4971 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/102659f9-0a07-48c9-b12b-5aa22f2a5fda-kube-api-access-dfnkz" (OuterVolumeSpecName: "kube-api-access-dfnkz") pod "102659f9-0a07-48c9-b12b-5aa22f2a5fda" (UID: "102659f9-0a07-48c9-b12b-5aa22f2a5fda"). InnerVolumeSpecName "kube-api-access-dfnkz". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 13 08:38:28 crc kubenswrapper[4971]: I1213 08:38:28.969480 4971 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/102659f9-0a07-48c9-b12b-5aa22f2a5fda-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "102659f9-0a07-48c9-b12b-5aa22f2a5fda" (UID: "102659f9-0a07-48c9-b12b-5aa22f2a5fda"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 13 08:38:28 crc kubenswrapper[4971]: I1213 08:38:28.988271 4971 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-dfnkz\" (UniqueName: \"kubernetes.io/projected/102659f9-0a07-48c9-b12b-5aa22f2a5fda-kube-api-access-dfnkz\") on node \"crc\" DevicePath \"\"" Dec 13 08:38:28 crc kubenswrapper[4971]: I1213 08:38:28.988926 4971 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/102659f9-0a07-48c9-b12b-5aa22f2a5fda-catalog-content\") on node \"crc\" DevicePath \"\"" Dec 13 08:38:28 crc kubenswrapper[4971]: I1213 08:38:28.988945 4971 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/102659f9-0a07-48c9-b12b-5aa22f2a5fda-utilities\") on node \"crc\" DevicePath \"\"" Dec 13 08:38:29 crc kubenswrapper[4971]: I1213 08:38:29.086435 4971 generic.go:334] "Generic (PLEG): container finished" podID="102659f9-0a07-48c9-b12b-5aa22f2a5fda" containerID="394641f16af20fe37971ac44c182491851a403642cca041c9f4624f56a8f098d" exitCode=0 Dec 13 08:38:29 crc kubenswrapper[4971]: I1213 08:38:29.086493 4971 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-r4l8j" event={"ID":"102659f9-0a07-48c9-b12b-5aa22f2a5fda","Type":"ContainerDied","Data":"394641f16af20fe37971ac44c182491851a403642cca041c9f4624f56a8f098d"} Dec 13 08:38:29 crc kubenswrapper[4971]: I1213 08:38:29.086613 4971 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-r4l8j" event={"ID":"102659f9-0a07-48c9-b12b-5aa22f2a5fda","Type":"ContainerDied","Data":"252700a554baa1706085d4bb44071ce2a9a0aae3014d4a662d8b1ba0b776fbb5"} Dec 13 08:38:29 crc kubenswrapper[4971]: I1213 08:38:29.086643 4971 scope.go:117] "RemoveContainer" containerID="394641f16af20fe37971ac44c182491851a403642cca041c9f4624f56a8f098d" Dec 13 08:38:29 crc kubenswrapper[4971]: I1213 08:38:29.087570 4971 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-r4l8j" Dec 13 08:38:29 crc kubenswrapper[4971]: I1213 08:38:29.131422 4971 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-r4l8j"] Dec 13 08:38:29 crc kubenswrapper[4971]: I1213 08:38:29.139475 4971 scope.go:117] "RemoveContainer" containerID="4ceb7a1738625662dd6cb11e27f6e1e36b671c7d3ae7bfb9489cfdde98ee7ed9" Dec 13 08:38:29 crc kubenswrapper[4971]: I1213 08:38:29.145243 4971 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/certified-operators-r4l8j"] Dec 13 08:38:29 crc kubenswrapper[4971]: I1213 08:38:29.169206 4971 scope.go:117] "RemoveContainer" containerID="543bc66edf0571d184c1217ca56eee753f966026a79aae3417a73406551c2838" Dec 13 08:38:29 crc kubenswrapper[4971]: I1213 08:38:29.222073 4971 scope.go:117] "RemoveContainer" containerID="394641f16af20fe37971ac44c182491851a403642cca041c9f4624f56a8f098d" Dec 13 08:38:29 crc kubenswrapper[4971]: E1213 08:38:29.223023 4971 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"394641f16af20fe37971ac44c182491851a403642cca041c9f4624f56a8f098d\": container with ID starting with 394641f16af20fe37971ac44c182491851a403642cca041c9f4624f56a8f098d not found: ID does not exist" containerID="394641f16af20fe37971ac44c182491851a403642cca041c9f4624f56a8f098d" Dec 13 08:38:29 crc kubenswrapper[4971]: I1213 08:38:29.223102 4971 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"394641f16af20fe37971ac44c182491851a403642cca041c9f4624f56a8f098d"} err="failed to get container status \"394641f16af20fe37971ac44c182491851a403642cca041c9f4624f56a8f098d\": rpc error: code = NotFound desc = could not find container \"394641f16af20fe37971ac44c182491851a403642cca041c9f4624f56a8f098d\": container with ID starting with 394641f16af20fe37971ac44c182491851a403642cca041c9f4624f56a8f098d not found: ID does not exist" Dec 13 08:38:29 crc kubenswrapper[4971]: I1213 08:38:29.223149 4971 scope.go:117] "RemoveContainer" containerID="4ceb7a1738625662dd6cb11e27f6e1e36b671c7d3ae7bfb9489cfdde98ee7ed9" Dec 13 08:38:29 crc kubenswrapper[4971]: E1213 08:38:29.223769 4971 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"4ceb7a1738625662dd6cb11e27f6e1e36b671c7d3ae7bfb9489cfdde98ee7ed9\": container with ID starting with 4ceb7a1738625662dd6cb11e27f6e1e36b671c7d3ae7bfb9489cfdde98ee7ed9 not found: ID does not exist" containerID="4ceb7a1738625662dd6cb11e27f6e1e36b671c7d3ae7bfb9489cfdde98ee7ed9" Dec 13 08:38:29 crc kubenswrapper[4971]: I1213 08:38:29.223813 4971 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"4ceb7a1738625662dd6cb11e27f6e1e36b671c7d3ae7bfb9489cfdde98ee7ed9"} err="failed to get container status \"4ceb7a1738625662dd6cb11e27f6e1e36b671c7d3ae7bfb9489cfdde98ee7ed9\": rpc error: code = NotFound desc = could not find container \"4ceb7a1738625662dd6cb11e27f6e1e36b671c7d3ae7bfb9489cfdde98ee7ed9\": container with ID starting with 4ceb7a1738625662dd6cb11e27f6e1e36b671c7d3ae7bfb9489cfdde98ee7ed9 not found: ID does not exist" Dec 13 08:38:29 crc kubenswrapper[4971]: I1213 08:38:29.223846 4971 scope.go:117] "RemoveContainer" containerID="543bc66edf0571d184c1217ca56eee753f966026a79aae3417a73406551c2838" Dec 13 08:38:29 crc kubenswrapper[4971]: E1213 08:38:29.224139 4971 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"543bc66edf0571d184c1217ca56eee753f966026a79aae3417a73406551c2838\": container with ID starting with 543bc66edf0571d184c1217ca56eee753f966026a79aae3417a73406551c2838 not found: ID does not exist" containerID="543bc66edf0571d184c1217ca56eee753f966026a79aae3417a73406551c2838" Dec 13 08:38:29 crc kubenswrapper[4971]: I1213 08:38:29.224172 4971 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"543bc66edf0571d184c1217ca56eee753f966026a79aae3417a73406551c2838"} err="failed to get container status \"543bc66edf0571d184c1217ca56eee753f966026a79aae3417a73406551c2838\": rpc error: code = NotFound desc = could not find container \"543bc66edf0571d184c1217ca56eee753f966026a79aae3417a73406551c2838\": container with ID starting with 543bc66edf0571d184c1217ca56eee753f966026a79aae3417a73406551c2838 not found: ID does not exist" Dec 13 08:38:29 crc kubenswrapper[4971]: I1213 08:38:29.783858 4971 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="102659f9-0a07-48c9-b12b-5aa22f2a5fda" path="/var/lib/kubelet/pods/102659f9-0a07-48c9-b12b-5aa22f2a5fda/volumes" Dec 13 08:38:32 crc kubenswrapper[4971]: I1213 08:38:32.769268 4971 scope.go:117] "RemoveContainer" containerID="f397fe912a74e38413c0149a7e7ca0ed1cff4acbea98dcbfa69153e27d3fdaa0" Dec 13 08:38:32 crc kubenswrapper[4971]: E1213 08:38:32.770505 4971 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-82xjz_openshift-machine-config-operator(e0c4af12-edda-41d7-9b44-f87396174ef2)\"" pod="openshift-machine-config-operator/machine-config-daemon-82xjz" podUID="e0c4af12-edda-41d7-9b44-f87396174ef2" Dec 13 08:38:46 crc kubenswrapper[4971]: I1213 08:38:46.769178 4971 scope.go:117] "RemoveContainer" containerID="f397fe912a74e38413c0149a7e7ca0ed1cff4acbea98dcbfa69153e27d3fdaa0" Dec 13 08:38:46 crc kubenswrapper[4971]: E1213 08:38:46.770687 4971 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-82xjz_openshift-machine-config-operator(e0c4af12-edda-41d7-9b44-f87396174ef2)\"" pod="openshift-machine-config-operator/machine-config-daemon-82xjz" podUID="e0c4af12-edda-41d7-9b44-f87396174ef2" Dec 13 08:38:58 crc kubenswrapper[4971]: I1213 08:38:58.768757 4971 scope.go:117] "RemoveContainer" containerID="f397fe912a74e38413c0149a7e7ca0ed1cff4acbea98dcbfa69153e27d3fdaa0" Dec 13 08:38:58 crc kubenswrapper[4971]: E1213 08:38:58.769998 4971 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-82xjz_openshift-machine-config-operator(e0c4af12-edda-41d7-9b44-f87396174ef2)\"" pod="openshift-machine-config-operator/machine-config-daemon-82xjz" podUID="e0c4af12-edda-41d7-9b44-f87396174ef2" Dec 13 08:39:09 crc kubenswrapper[4971]: I1213 08:39:09.769588 4971 scope.go:117] "RemoveContainer" containerID="f397fe912a74e38413c0149a7e7ca0ed1cff4acbea98dcbfa69153e27d3fdaa0" Dec 13 08:39:09 crc kubenswrapper[4971]: E1213 08:39:09.771236 4971 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-82xjz_openshift-machine-config-operator(e0c4af12-edda-41d7-9b44-f87396174ef2)\"" pod="openshift-machine-config-operator/machine-config-daemon-82xjz" podUID="e0c4af12-edda-41d7-9b44-f87396174ef2" Dec 13 08:39:23 crc kubenswrapper[4971]: I1213 08:39:23.776616 4971 scope.go:117] "RemoveContainer" containerID="f397fe912a74e38413c0149a7e7ca0ed1cff4acbea98dcbfa69153e27d3fdaa0" Dec 13 08:39:23 crc kubenswrapper[4971]: E1213 08:39:23.777941 4971 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-82xjz_openshift-machine-config-operator(e0c4af12-edda-41d7-9b44-f87396174ef2)\"" pod="openshift-machine-config-operator/machine-config-daemon-82xjz" podUID="e0c4af12-edda-41d7-9b44-f87396174ef2" Dec 13 08:39:34 crc kubenswrapper[4971]: I1213 08:39:34.768806 4971 scope.go:117] "RemoveContainer" containerID="f397fe912a74e38413c0149a7e7ca0ed1cff4acbea98dcbfa69153e27d3fdaa0" Dec 13 08:39:34 crc kubenswrapper[4971]: E1213 08:39:34.770165 4971 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-82xjz_openshift-machine-config-operator(e0c4af12-edda-41d7-9b44-f87396174ef2)\"" pod="openshift-machine-config-operator/machine-config-daemon-82xjz" podUID="e0c4af12-edda-41d7-9b44-f87396174ef2" Dec 13 08:39:48 crc kubenswrapper[4971]: I1213 08:39:48.769658 4971 scope.go:117] "RemoveContainer" containerID="f397fe912a74e38413c0149a7e7ca0ed1cff4acbea98dcbfa69153e27d3fdaa0" Dec 13 08:39:48 crc kubenswrapper[4971]: E1213 08:39:48.771092 4971 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-82xjz_openshift-machine-config-operator(e0c4af12-edda-41d7-9b44-f87396174ef2)\"" pod="openshift-machine-config-operator/machine-config-daemon-82xjz" podUID="e0c4af12-edda-41d7-9b44-f87396174ef2" Dec 13 08:39:59 crc kubenswrapper[4971]: I1213 08:39:59.772762 4971 scope.go:117] "RemoveContainer" containerID="f397fe912a74e38413c0149a7e7ca0ed1cff4acbea98dcbfa69153e27d3fdaa0" Dec 13 08:39:59 crc kubenswrapper[4971]: E1213 08:39:59.775734 4971 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-82xjz_openshift-machine-config-operator(e0c4af12-edda-41d7-9b44-f87396174ef2)\"" pod="openshift-machine-config-operator/machine-config-daemon-82xjz" podUID="e0c4af12-edda-41d7-9b44-f87396174ef2" Dec 13 08:40:11 crc kubenswrapper[4971]: I1213 08:40:11.768875 4971 scope.go:117] "RemoveContainer" containerID="f397fe912a74e38413c0149a7e7ca0ed1cff4acbea98dcbfa69153e27d3fdaa0" Dec 13 08:40:11 crc kubenswrapper[4971]: E1213 08:40:11.770102 4971 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-82xjz_openshift-machine-config-operator(e0c4af12-edda-41d7-9b44-f87396174ef2)\"" pod="openshift-machine-config-operator/machine-config-daemon-82xjz" podUID="e0c4af12-edda-41d7-9b44-f87396174ef2" Dec 13 08:40:25 crc kubenswrapper[4971]: I1213 08:40:25.786099 4971 scope.go:117] "RemoveContainer" containerID="f397fe912a74e38413c0149a7e7ca0ed1cff4acbea98dcbfa69153e27d3fdaa0" Dec 13 08:40:25 crc kubenswrapper[4971]: E1213 08:40:25.787430 4971 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-82xjz_openshift-machine-config-operator(e0c4af12-edda-41d7-9b44-f87396174ef2)\"" pod="openshift-machine-config-operator/machine-config-daemon-82xjz" podUID="e0c4af12-edda-41d7-9b44-f87396174ef2" Dec 13 08:40:38 crc kubenswrapper[4971]: I1213 08:40:38.768057 4971 scope.go:117] "RemoveContainer" containerID="f397fe912a74e38413c0149a7e7ca0ed1cff4acbea98dcbfa69153e27d3fdaa0" Dec 13 08:40:38 crc kubenswrapper[4971]: E1213 08:40:38.769314 4971 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-82xjz_openshift-machine-config-operator(e0c4af12-edda-41d7-9b44-f87396174ef2)\"" pod="openshift-machine-config-operator/machine-config-daemon-82xjz" podUID="e0c4af12-edda-41d7-9b44-f87396174ef2" Dec 13 08:40:49 crc kubenswrapper[4971]: I1213 08:40:49.776494 4971 scope.go:117] "RemoveContainer" containerID="f397fe912a74e38413c0149a7e7ca0ed1cff4acbea98dcbfa69153e27d3fdaa0" Dec 13 08:40:49 crc kubenswrapper[4971]: E1213 08:40:49.777672 4971 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-82xjz_openshift-machine-config-operator(e0c4af12-edda-41d7-9b44-f87396174ef2)\"" pod="openshift-machine-config-operator/machine-config-daemon-82xjz" podUID="e0c4af12-edda-41d7-9b44-f87396174ef2" Dec 13 08:41:03 crc kubenswrapper[4971]: I1213 08:41:03.776940 4971 scope.go:117] "RemoveContainer" containerID="f397fe912a74e38413c0149a7e7ca0ed1cff4acbea98dcbfa69153e27d3fdaa0" Dec 13 08:41:03 crc kubenswrapper[4971]: E1213 08:41:03.780344 4971 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-82xjz_openshift-machine-config-operator(e0c4af12-edda-41d7-9b44-f87396174ef2)\"" pod="openshift-machine-config-operator/machine-config-daemon-82xjz" podUID="e0c4af12-edda-41d7-9b44-f87396174ef2" Dec 13 08:41:14 crc kubenswrapper[4971]: I1213 08:41:14.065664 4971 scope.go:117] "RemoveContainer" containerID="f397fe912a74e38413c0149a7e7ca0ed1cff4acbea98dcbfa69153e27d3fdaa0" Dec 13 08:41:14 crc kubenswrapper[4971]: E1213 08:41:14.067097 4971 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-82xjz_openshift-machine-config-operator(e0c4af12-edda-41d7-9b44-f87396174ef2)\"" pod="openshift-machine-config-operator/machine-config-daemon-82xjz" podUID="e0c4af12-edda-41d7-9b44-f87396174ef2" Dec 13 08:41:29 crc kubenswrapper[4971]: I1213 08:41:29.769025 4971 scope.go:117] "RemoveContainer" containerID="f397fe912a74e38413c0149a7e7ca0ed1cff4acbea98dcbfa69153e27d3fdaa0" Dec 13 08:41:29 crc kubenswrapper[4971]: E1213 08:41:29.770458 4971 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-82xjz_openshift-machine-config-operator(e0c4af12-edda-41d7-9b44-f87396174ef2)\"" pod="openshift-machine-config-operator/machine-config-daemon-82xjz" podUID="e0c4af12-edda-41d7-9b44-f87396174ef2" Dec 13 08:41:44 crc kubenswrapper[4971]: I1213 08:41:44.769151 4971 scope.go:117] "RemoveContainer" containerID="f397fe912a74e38413c0149a7e7ca0ed1cff4acbea98dcbfa69153e27d3fdaa0" Dec 13 08:41:44 crc kubenswrapper[4971]: E1213 08:41:44.770410 4971 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-82xjz_openshift-machine-config-operator(e0c4af12-edda-41d7-9b44-f87396174ef2)\"" pod="openshift-machine-config-operator/machine-config-daemon-82xjz" podUID="e0c4af12-edda-41d7-9b44-f87396174ef2" Dec 13 08:41:59 crc kubenswrapper[4971]: I1213 08:41:59.769068 4971 scope.go:117] "RemoveContainer" containerID="f397fe912a74e38413c0149a7e7ca0ed1cff4acbea98dcbfa69153e27d3fdaa0" Dec 13 08:42:00 crc kubenswrapper[4971]: I1213 08:42:00.626603 4971 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-82xjz" event={"ID":"e0c4af12-edda-41d7-9b44-f87396174ef2","Type":"ContainerStarted","Data":"513b55f2bb24361c2480216b9b7a4b3b6659eda1ac7c17acc12836e086669008"} Dec 13 08:43:31 crc kubenswrapper[4971]: I1213 08:43:31.373747 4971 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-operators-hfhjv"] Dec 13 08:43:31 crc kubenswrapper[4971]: E1213 08:43:31.375315 4971 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="102659f9-0a07-48c9-b12b-5aa22f2a5fda" containerName="extract-utilities" Dec 13 08:43:31 crc kubenswrapper[4971]: I1213 08:43:31.375331 4971 state_mem.go:107] "Deleted CPUSet assignment" podUID="102659f9-0a07-48c9-b12b-5aa22f2a5fda" containerName="extract-utilities" Dec 13 08:43:31 crc kubenswrapper[4971]: E1213 08:43:31.375344 4971 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="102659f9-0a07-48c9-b12b-5aa22f2a5fda" containerName="extract-content" Dec 13 08:43:31 crc kubenswrapper[4971]: I1213 08:43:31.375351 4971 state_mem.go:107] "Deleted CPUSet assignment" podUID="102659f9-0a07-48c9-b12b-5aa22f2a5fda" containerName="extract-content" Dec 13 08:43:31 crc kubenswrapper[4971]: E1213 08:43:31.375373 4971 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="102659f9-0a07-48c9-b12b-5aa22f2a5fda" containerName="registry-server" Dec 13 08:43:31 crc kubenswrapper[4971]: I1213 08:43:31.375382 4971 state_mem.go:107] "Deleted CPUSet assignment" podUID="102659f9-0a07-48c9-b12b-5aa22f2a5fda" containerName="registry-server" Dec 13 08:43:31 crc kubenswrapper[4971]: I1213 08:43:31.375631 4971 memory_manager.go:354] "RemoveStaleState removing state" podUID="102659f9-0a07-48c9-b12b-5aa22f2a5fda" containerName="registry-server" Dec 13 08:43:31 crc kubenswrapper[4971]: I1213 08:43:31.377565 4971 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-hfhjv" Dec 13 08:43:31 crc kubenswrapper[4971]: I1213 08:43:31.391669 4971 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-hfhjv"] Dec 13 08:43:31 crc kubenswrapper[4971]: I1213 08:43:31.547558 4971 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/59d1fcc5-abdc-47ae-9686-bb7eddf8ded5-utilities\") pod \"redhat-operators-hfhjv\" (UID: \"59d1fcc5-abdc-47ae-9686-bb7eddf8ded5\") " pod="openshift-marketplace/redhat-operators-hfhjv" Dec 13 08:43:31 crc kubenswrapper[4971]: I1213 08:43:31.548297 4971 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/59d1fcc5-abdc-47ae-9686-bb7eddf8ded5-catalog-content\") pod \"redhat-operators-hfhjv\" (UID: \"59d1fcc5-abdc-47ae-9686-bb7eddf8ded5\") " pod="openshift-marketplace/redhat-operators-hfhjv" Dec 13 08:43:31 crc kubenswrapper[4971]: I1213 08:43:31.548369 4971 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-zgjwc\" (UniqueName: \"kubernetes.io/projected/59d1fcc5-abdc-47ae-9686-bb7eddf8ded5-kube-api-access-zgjwc\") pod \"redhat-operators-hfhjv\" (UID: \"59d1fcc5-abdc-47ae-9686-bb7eddf8ded5\") " pod="openshift-marketplace/redhat-operators-hfhjv" Dec 13 08:43:31 crc kubenswrapper[4971]: I1213 08:43:31.571669 4971 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-marketplace-dxrnq"] Dec 13 08:43:31 crc kubenswrapper[4971]: I1213 08:43:31.574556 4971 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-dxrnq" Dec 13 08:43:31 crc kubenswrapper[4971]: I1213 08:43:31.589210 4971 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-dxrnq"] Dec 13 08:43:31 crc kubenswrapper[4971]: I1213 08:43:31.651589 4971 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/59d1fcc5-abdc-47ae-9686-bb7eddf8ded5-utilities\") pod \"redhat-operators-hfhjv\" (UID: \"59d1fcc5-abdc-47ae-9686-bb7eddf8ded5\") " pod="openshift-marketplace/redhat-operators-hfhjv" Dec 13 08:43:31 crc kubenswrapper[4971]: I1213 08:43:31.651736 4971 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/59d1fcc5-abdc-47ae-9686-bb7eddf8ded5-catalog-content\") pod \"redhat-operators-hfhjv\" (UID: \"59d1fcc5-abdc-47ae-9686-bb7eddf8ded5\") " pod="openshift-marketplace/redhat-operators-hfhjv" Dec 13 08:43:31 crc kubenswrapper[4971]: I1213 08:43:31.651790 4971 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-zgjwc\" (UniqueName: \"kubernetes.io/projected/59d1fcc5-abdc-47ae-9686-bb7eddf8ded5-kube-api-access-zgjwc\") pod \"redhat-operators-hfhjv\" (UID: \"59d1fcc5-abdc-47ae-9686-bb7eddf8ded5\") " pod="openshift-marketplace/redhat-operators-hfhjv" Dec 13 08:43:31 crc kubenswrapper[4971]: I1213 08:43:31.652563 4971 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/59d1fcc5-abdc-47ae-9686-bb7eddf8ded5-catalog-content\") pod \"redhat-operators-hfhjv\" (UID: \"59d1fcc5-abdc-47ae-9686-bb7eddf8ded5\") " pod="openshift-marketplace/redhat-operators-hfhjv" Dec 13 08:43:31 crc kubenswrapper[4971]: I1213 08:43:31.653872 4971 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/59d1fcc5-abdc-47ae-9686-bb7eddf8ded5-utilities\") pod \"redhat-operators-hfhjv\" (UID: \"59d1fcc5-abdc-47ae-9686-bb7eddf8ded5\") " pod="openshift-marketplace/redhat-operators-hfhjv" Dec 13 08:43:31 crc kubenswrapper[4971]: I1213 08:43:31.694137 4971 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-zgjwc\" (UniqueName: \"kubernetes.io/projected/59d1fcc5-abdc-47ae-9686-bb7eddf8ded5-kube-api-access-zgjwc\") pod \"redhat-operators-hfhjv\" (UID: \"59d1fcc5-abdc-47ae-9686-bb7eddf8ded5\") " pod="openshift-marketplace/redhat-operators-hfhjv" Dec 13 08:43:31 crc kubenswrapper[4971]: I1213 08:43:31.706833 4971 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-hfhjv" Dec 13 08:43:31 crc kubenswrapper[4971]: I1213 08:43:31.757683 4971 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/d9cc4da9-21c9-441d-8a83-4e49ac416fed-catalog-content\") pod \"redhat-marketplace-dxrnq\" (UID: \"d9cc4da9-21c9-441d-8a83-4e49ac416fed\") " pod="openshift-marketplace/redhat-marketplace-dxrnq" Dec 13 08:43:31 crc kubenswrapper[4971]: I1213 08:43:31.757838 4971 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/d9cc4da9-21c9-441d-8a83-4e49ac416fed-utilities\") pod \"redhat-marketplace-dxrnq\" (UID: \"d9cc4da9-21c9-441d-8a83-4e49ac416fed\") " pod="openshift-marketplace/redhat-marketplace-dxrnq" Dec 13 08:43:31 crc kubenswrapper[4971]: I1213 08:43:31.757947 4971 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-25pdh\" (UniqueName: \"kubernetes.io/projected/d9cc4da9-21c9-441d-8a83-4e49ac416fed-kube-api-access-25pdh\") pod \"redhat-marketplace-dxrnq\" (UID: \"d9cc4da9-21c9-441d-8a83-4e49ac416fed\") " pod="openshift-marketplace/redhat-marketplace-dxrnq" Dec 13 08:43:31 crc kubenswrapper[4971]: I1213 08:43:31.860473 4971 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-25pdh\" (UniqueName: \"kubernetes.io/projected/d9cc4da9-21c9-441d-8a83-4e49ac416fed-kube-api-access-25pdh\") pod \"redhat-marketplace-dxrnq\" (UID: \"d9cc4da9-21c9-441d-8a83-4e49ac416fed\") " pod="openshift-marketplace/redhat-marketplace-dxrnq" Dec 13 08:43:31 crc kubenswrapper[4971]: I1213 08:43:31.860745 4971 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/d9cc4da9-21c9-441d-8a83-4e49ac416fed-catalog-content\") pod \"redhat-marketplace-dxrnq\" (UID: \"d9cc4da9-21c9-441d-8a83-4e49ac416fed\") " pod="openshift-marketplace/redhat-marketplace-dxrnq" Dec 13 08:43:31 crc kubenswrapper[4971]: I1213 08:43:31.861621 4971 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/d9cc4da9-21c9-441d-8a83-4e49ac416fed-catalog-content\") pod \"redhat-marketplace-dxrnq\" (UID: \"d9cc4da9-21c9-441d-8a83-4e49ac416fed\") " pod="openshift-marketplace/redhat-marketplace-dxrnq" Dec 13 08:43:31 crc kubenswrapper[4971]: I1213 08:43:31.861883 4971 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/d9cc4da9-21c9-441d-8a83-4e49ac416fed-utilities\") pod \"redhat-marketplace-dxrnq\" (UID: \"d9cc4da9-21c9-441d-8a83-4e49ac416fed\") " pod="openshift-marketplace/redhat-marketplace-dxrnq" Dec 13 08:43:31 crc kubenswrapper[4971]: I1213 08:43:31.862225 4971 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/d9cc4da9-21c9-441d-8a83-4e49ac416fed-utilities\") pod \"redhat-marketplace-dxrnq\" (UID: \"d9cc4da9-21c9-441d-8a83-4e49ac416fed\") " pod="openshift-marketplace/redhat-marketplace-dxrnq" Dec 13 08:43:31 crc kubenswrapper[4971]: I1213 08:43:31.911687 4971 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-25pdh\" (UniqueName: \"kubernetes.io/projected/d9cc4da9-21c9-441d-8a83-4e49ac416fed-kube-api-access-25pdh\") pod \"redhat-marketplace-dxrnq\" (UID: \"d9cc4da9-21c9-441d-8a83-4e49ac416fed\") " pod="openshift-marketplace/redhat-marketplace-dxrnq" Dec 13 08:43:32 crc kubenswrapper[4971]: I1213 08:43:32.197747 4971 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-dxrnq" Dec 13 08:43:32 crc kubenswrapper[4971]: I1213 08:43:32.372686 4971 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-hfhjv"] Dec 13 08:43:32 crc kubenswrapper[4971]: I1213 08:43:32.653742 4971 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-hfhjv" event={"ID":"59d1fcc5-abdc-47ae-9686-bb7eddf8ded5","Type":"ContainerStarted","Data":"731f745f651853fdac843940b62b3fcd7934c7c12fab22bbb715cb486a4cf386"} Dec 13 08:43:32 crc kubenswrapper[4971]: I1213 08:43:32.816281 4971 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-dxrnq"] Dec 13 08:43:33 crc kubenswrapper[4971]: I1213 08:43:33.666105 4971 generic.go:334] "Generic (PLEG): container finished" podID="59d1fcc5-abdc-47ae-9686-bb7eddf8ded5" containerID="4a2bd0723479da99039e652e4ba3104d19fd162773cacc199c90e1c1449c3f49" exitCode=0 Dec 13 08:43:33 crc kubenswrapper[4971]: I1213 08:43:33.666196 4971 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-hfhjv" event={"ID":"59d1fcc5-abdc-47ae-9686-bb7eddf8ded5","Type":"ContainerDied","Data":"4a2bd0723479da99039e652e4ba3104d19fd162773cacc199c90e1c1449c3f49"} Dec 13 08:43:33 crc kubenswrapper[4971]: I1213 08:43:33.669425 4971 generic.go:334] "Generic (PLEG): container finished" podID="d9cc4da9-21c9-441d-8a83-4e49ac416fed" containerID="fbd1ab6daa33e2828726ea648009f864c764f20ef18d98e7cea389768e77e9d0" exitCode=0 Dec 13 08:43:33 crc kubenswrapper[4971]: I1213 08:43:33.669489 4971 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-dxrnq" event={"ID":"d9cc4da9-21c9-441d-8a83-4e49ac416fed","Type":"ContainerDied","Data":"fbd1ab6daa33e2828726ea648009f864c764f20ef18d98e7cea389768e77e9d0"} Dec 13 08:43:33 crc kubenswrapper[4971]: I1213 08:43:33.669559 4971 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-dxrnq" event={"ID":"d9cc4da9-21c9-441d-8a83-4e49ac416fed","Type":"ContainerStarted","Data":"28f30b4c5ca77fadb2b4a9ef5b34a1456f89844dab555d8bf315652020241e3c"} Dec 13 08:43:33 crc kubenswrapper[4971]: I1213 08:43:33.669655 4971 provider.go:102] Refreshing cache for provider: *credentialprovider.defaultDockerConfigProvider Dec 13 08:43:35 crc kubenswrapper[4971]: I1213 08:43:35.711081 4971 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-hfhjv" event={"ID":"59d1fcc5-abdc-47ae-9686-bb7eddf8ded5","Type":"ContainerStarted","Data":"25f28620b60992886c96aedf9743e1ba9279ce237a5c9377c391e366ae2cddb6"} Dec 13 08:43:35 crc kubenswrapper[4971]: I1213 08:43:35.716787 4971 generic.go:334] "Generic (PLEG): container finished" podID="d9cc4da9-21c9-441d-8a83-4e49ac416fed" containerID="3488e3ead0a246e8284cda57ff1190a35b1eebed65f2f01018af2e554c566bdb" exitCode=0 Dec 13 08:43:35 crc kubenswrapper[4971]: I1213 08:43:35.716850 4971 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-dxrnq" event={"ID":"d9cc4da9-21c9-441d-8a83-4e49ac416fed","Type":"ContainerDied","Data":"3488e3ead0a246e8284cda57ff1190a35b1eebed65f2f01018af2e554c566bdb"} Dec 13 08:43:38 crc kubenswrapper[4971]: I1213 08:43:38.762310 4971 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-dxrnq" event={"ID":"d9cc4da9-21c9-441d-8a83-4e49ac416fed","Type":"ContainerStarted","Data":"06d47f4781d33377772464259a60f13f60e40914bd75c0443130682b470f3d1c"} Dec 13 08:43:38 crc kubenswrapper[4971]: I1213 08:43:38.827950 4971 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-marketplace-dxrnq" podStartSLOduration=4.526660893 podStartE2EDuration="7.827917044s" podCreationTimestamp="2025-12-13 08:43:31 +0000 UTC" firstStartedPulling="2025-12-13 08:43:33.67199646 +0000 UTC m=+6870.276405898" lastFinishedPulling="2025-12-13 08:43:36.973252601 +0000 UTC m=+6873.577662049" observedRunningTime="2025-12-13 08:43:38.803929573 +0000 UTC m=+6875.408339031" watchObservedRunningTime="2025-12-13 08:43:38.827917044 +0000 UTC m=+6875.432326492" Dec 13 08:43:39 crc kubenswrapper[4971]: I1213 08:43:39.777363 4971 generic.go:334] "Generic (PLEG): container finished" podID="59d1fcc5-abdc-47ae-9686-bb7eddf8ded5" containerID="25f28620b60992886c96aedf9743e1ba9279ce237a5c9377c391e366ae2cddb6" exitCode=0 Dec 13 08:43:39 crc kubenswrapper[4971]: I1213 08:43:39.780170 4971 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-hfhjv" event={"ID":"59d1fcc5-abdc-47ae-9686-bb7eddf8ded5","Type":"ContainerDied","Data":"25f28620b60992886c96aedf9743e1ba9279ce237a5c9377c391e366ae2cddb6"} Dec 13 08:43:40 crc kubenswrapper[4971]: I1213 08:43:40.793367 4971 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-hfhjv" event={"ID":"59d1fcc5-abdc-47ae-9686-bb7eddf8ded5","Type":"ContainerStarted","Data":"31f8c8c0103efb8be3b782e31cc9f1b6eb68a1198360267eb7ff7b2732b57b92"} Dec 13 08:43:40 crc kubenswrapper[4971]: I1213 08:43:40.825064 4971 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-operators-hfhjv" podStartSLOduration=3.216280859 podStartE2EDuration="9.825042915s" podCreationTimestamp="2025-12-13 08:43:31 +0000 UTC" firstStartedPulling="2025-12-13 08:43:33.669272543 +0000 UTC m=+6870.273682001" lastFinishedPulling="2025-12-13 08:43:40.278034609 +0000 UTC m=+6876.882444057" observedRunningTime="2025-12-13 08:43:40.820546025 +0000 UTC m=+6877.424955473" watchObservedRunningTime="2025-12-13 08:43:40.825042915 +0000 UTC m=+6877.429452363" Dec 13 08:43:41 crc kubenswrapper[4971]: I1213 08:43:41.707574 4971 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-operators-hfhjv" Dec 13 08:43:41 crc kubenswrapper[4971]: I1213 08:43:41.708120 4971 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-operators-hfhjv" Dec 13 08:43:42 crc kubenswrapper[4971]: I1213 08:43:42.198067 4971 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-marketplace-dxrnq" Dec 13 08:43:42 crc kubenswrapper[4971]: I1213 08:43:42.199576 4971 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-marketplace-dxrnq" Dec 13 08:43:42 crc kubenswrapper[4971]: I1213 08:43:42.252282 4971 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-marketplace-dxrnq" Dec 13 08:43:42 crc kubenswrapper[4971]: I1213 08:43:42.785711 4971 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-marketplace/redhat-operators-hfhjv" podUID="59d1fcc5-abdc-47ae-9686-bb7eddf8ded5" containerName="registry-server" probeResult="failure" output=< Dec 13 08:43:42 crc kubenswrapper[4971]: timeout: failed to connect service ":50051" within 1s Dec 13 08:43:42 crc kubenswrapper[4971]: > Dec 13 08:43:51 crc kubenswrapper[4971]: I1213 08:43:51.766114 4971 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-operators-hfhjv" Dec 13 08:43:51 crc kubenswrapper[4971]: I1213 08:43:51.824235 4971 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-operators-hfhjv" Dec 13 08:43:52 crc kubenswrapper[4971]: I1213 08:43:52.014806 4971 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-hfhjv"] Dec 13 08:43:52 crc kubenswrapper[4971]: I1213 08:43:52.256730 4971 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-marketplace-dxrnq" Dec 13 08:43:52 crc kubenswrapper[4971]: I1213 08:43:52.935297 4971 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-operators-hfhjv" podUID="59d1fcc5-abdc-47ae-9686-bb7eddf8ded5" containerName="registry-server" containerID="cri-o://31f8c8c0103efb8be3b782e31cc9f1b6eb68a1198360267eb7ff7b2732b57b92" gracePeriod=2 Dec 13 08:43:53 crc kubenswrapper[4971]: I1213 08:43:53.956221 4971 generic.go:334] "Generic (PLEG): container finished" podID="59d1fcc5-abdc-47ae-9686-bb7eddf8ded5" containerID="31f8c8c0103efb8be3b782e31cc9f1b6eb68a1198360267eb7ff7b2732b57b92" exitCode=0 Dec 13 08:43:53 crc kubenswrapper[4971]: I1213 08:43:53.956300 4971 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-hfhjv" event={"ID":"59d1fcc5-abdc-47ae-9686-bb7eddf8ded5","Type":"ContainerDied","Data":"31f8c8c0103efb8be3b782e31cc9f1b6eb68a1198360267eb7ff7b2732b57b92"} Dec 13 08:43:54 crc kubenswrapper[4971]: I1213 08:43:54.152246 4971 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-hfhjv" Dec 13 08:43:54 crc kubenswrapper[4971]: I1213 08:43:54.354848 4971 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/59d1fcc5-abdc-47ae-9686-bb7eddf8ded5-utilities\") pod \"59d1fcc5-abdc-47ae-9686-bb7eddf8ded5\" (UID: \"59d1fcc5-abdc-47ae-9686-bb7eddf8ded5\") " Dec 13 08:43:54 crc kubenswrapper[4971]: I1213 08:43:54.355040 4971 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/59d1fcc5-abdc-47ae-9686-bb7eddf8ded5-catalog-content\") pod \"59d1fcc5-abdc-47ae-9686-bb7eddf8ded5\" (UID: \"59d1fcc5-abdc-47ae-9686-bb7eddf8ded5\") " Dec 13 08:43:54 crc kubenswrapper[4971]: I1213 08:43:54.355118 4971 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-zgjwc\" (UniqueName: \"kubernetes.io/projected/59d1fcc5-abdc-47ae-9686-bb7eddf8ded5-kube-api-access-zgjwc\") pod \"59d1fcc5-abdc-47ae-9686-bb7eddf8ded5\" (UID: \"59d1fcc5-abdc-47ae-9686-bb7eddf8ded5\") " Dec 13 08:43:54 crc kubenswrapper[4971]: I1213 08:43:54.356056 4971 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/59d1fcc5-abdc-47ae-9686-bb7eddf8ded5-utilities" (OuterVolumeSpecName: "utilities") pod "59d1fcc5-abdc-47ae-9686-bb7eddf8ded5" (UID: "59d1fcc5-abdc-47ae-9686-bb7eddf8ded5"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 13 08:43:54 crc kubenswrapper[4971]: I1213 08:43:54.370104 4971 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/59d1fcc5-abdc-47ae-9686-bb7eddf8ded5-kube-api-access-zgjwc" (OuterVolumeSpecName: "kube-api-access-zgjwc") pod "59d1fcc5-abdc-47ae-9686-bb7eddf8ded5" (UID: "59d1fcc5-abdc-47ae-9686-bb7eddf8ded5"). InnerVolumeSpecName "kube-api-access-zgjwc". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 13 08:43:54 crc kubenswrapper[4971]: I1213 08:43:54.458251 4971 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/59d1fcc5-abdc-47ae-9686-bb7eddf8ded5-utilities\") on node \"crc\" DevicePath \"\"" Dec 13 08:43:54 crc kubenswrapper[4971]: I1213 08:43:54.458296 4971 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-zgjwc\" (UniqueName: \"kubernetes.io/projected/59d1fcc5-abdc-47ae-9686-bb7eddf8ded5-kube-api-access-zgjwc\") on node \"crc\" DevicePath \"\"" Dec 13 08:43:54 crc kubenswrapper[4971]: I1213 08:43:54.495283 4971 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/59d1fcc5-abdc-47ae-9686-bb7eddf8ded5-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "59d1fcc5-abdc-47ae-9686-bb7eddf8ded5" (UID: "59d1fcc5-abdc-47ae-9686-bb7eddf8ded5"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 13 08:43:54 crc kubenswrapper[4971]: I1213 08:43:54.559768 4971 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/59d1fcc5-abdc-47ae-9686-bb7eddf8ded5-catalog-content\") on node \"crc\" DevicePath \"\"" Dec 13 08:43:54 crc kubenswrapper[4971]: I1213 08:43:54.615147 4971 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-dxrnq"] Dec 13 08:43:54 crc kubenswrapper[4971]: I1213 08:43:54.615408 4971 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-marketplace-dxrnq" podUID="d9cc4da9-21c9-441d-8a83-4e49ac416fed" containerName="registry-server" containerID="cri-o://06d47f4781d33377772464259a60f13f60e40914bd75c0443130682b470f3d1c" gracePeriod=2 Dec 13 08:43:54 crc kubenswrapper[4971]: I1213 08:43:54.971191 4971 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-hfhjv" event={"ID":"59d1fcc5-abdc-47ae-9686-bb7eddf8ded5","Type":"ContainerDied","Data":"731f745f651853fdac843940b62b3fcd7934c7c12fab22bbb715cb486a4cf386"} Dec 13 08:43:54 crc kubenswrapper[4971]: I1213 08:43:54.972271 4971 scope.go:117] "RemoveContainer" containerID="31f8c8c0103efb8be3b782e31cc9f1b6eb68a1198360267eb7ff7b2732b57b92" Dec 13 08:43:54 crc kubenswrapper[4971]: I1213 08:43:54.971551 4971 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-hfhjv" Dec 13 08:43:54 crc kubenswrapper[4971]: I1213 08:43:54.980234 4971 generic.go:334] "Generic (PLEG): container finished" podID="d9cc4da9-21c9-441d-8a83-4e49ac416fed" containerID="06d47f4781d33377772464259a60f13f60e40914bd75c0443130682b470f3d1c" exitCode=0 Dec 13 08:43:54 crc kubenswrapper[4971]: I1213 08:43:54.980295 4971 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-dxrnq" event={"ID":"d9cc4da9-21c9-441d-8a83-4e49ac416fed","Type":"ContainerDied","Data":"06d47f4781d33377772464259a60f13f60e40914bd75c0443130682b470f3d1c"} Dec 13 08:43:55 crc kubenswrapper[4971]: I1213 08:43:55.015999 4971 scope.go:117] "RemoveContainer" containerID="25f28620b60992886c96aedf9743e1ba9279ce237a5c9377c391e366ae2cddb6" Dec 13 08:43:55 crc kubenswrapper[4971]: I1213 08:43:55.037954 4971 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-hfhjv"] Dec 13 08:43:55 crc kubenswrapper[4971]: I1213 08:43:55.056431 4971 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-operators-hfhjv"] Dec 13 08:43:55 crc kubenswrapper[4971]: I1213 08:43:55.088667 4971 scope.go:117] "RemoveContainer" containerID="4a2bd0723479da99039e652e4ba3104d19fd162773cacc199c90e1c1449c3f49" Dec 13 08:43:55 crc kubenswrapper[4971]: I1213 08:43:55.283246 4971 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-dxrnq" Dec 13 08:43:55 crc kubenswrapper[4971]: I1213 08:43:55.380872 4971 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/d9cc4da9-21c9-441d-8a83-4e49ac416fed-catalog-content\") pod \"d9cc4da9-21c9-441d-8a83-4e49ac416fed\" (UID: \"d9cc4da9-21c9-441d-8a83-4e49ac416fed\") " Dec 13 08:43:55 crc kubenswrapper[4971]: I1213 08:43:55.380971 4971 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-25pdh\" (UniqueName: \"kubernetes.io/projected/d9cc4da9-21c9-441d-8a83-4e49ac416fed-kube-api-access-25pdh\") pod \"d9cc4da9-21c9-441d-8a83-4e49ac416fed\" (UID: \"d9cc4da9-21c9-441d-8a83-4e49ac416fed\") " Dec 13 08:43:55 crc kubenswrapper[4971]: I1213 08:43:55.381087 4971 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/d9cc4da9-21c9-441d-8a83-4e49ac416fed-utilities\") pod \"d9cc4da9-21c9-441d-8a83-4e49ac416fed\" (UID: \"d9cc4da9-21c9-441d-8a83-4e49ac416fed\") " Dec 13 08:43:55 crc kubenswrapper[4971]: I1213 08:43:55.383572 4971 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/d9cc4da9-21c9-441d-8a83-4e49ac416fed-utilities" (OuterVolumeSpecName: "utilities") pod "d9cc4da9-21c9-441d-8a83-4e49ac416fed" (UID: "d9cc4da9-21c9-441d-8a83-4e49ac416fed"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 13 08:43:55 crc kubenswrapper[4971]: I1213 08:43:55.397067 4971 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/d9cc4da9-21c9-441d-8a83-4e49ac416fed-kube-api-access-25pdh" (OuterVolumeSpecName: "kube-api-access-25pdh") pod "d9cc4da9-21c9-441d-8a83-4e49ac416fed" (UID: "d9cc4da9-21c9-441d-8a83-4e49ac416fed"). InnerVolumeSpecName "kube-api-access-25pdh". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 13 08:43:55 crc kubenswrapper[4971]: I1213 08:43:55.407661 4971 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/d9cc4da9-21c9-441d-8a83-4e49ac416fed-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "d9cc4da9-21c9-441d-8a83-4e49ac416fed" (UID: "d9cc4da9-21c9-441d-8a83-4e49ac416fed"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 13 08:43:55 crc kubenswrapper[4971]: I1213 08:43:55.484064 4971 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/d9cc4da9-21c9-441d-8a83-4e49ac416fed-catalog-content\") on node \"crc\" DevicePath \"\"" Dec 13 08:43:55 crc kubenswrapper[4971]: I1213 08:43:55.484115 4971 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-25pdh\" (UniqueName: \"kubernetes.io/projected/d9cc4da9-21c9-441d-8a83-4e49ac416fed-kube-api-access-25pdh\") on node \"crc\" DevicePath \"\"" Dec 13 08:43:55 crc kubenswrapper[4971]: I1213 08:43:55.484127 4971 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/d9cc4da9-21c9-441d-8a83-4e49ac416fed-utilities\") on node \"crc\" DevicePath \"\"" Dec 13 08:43:55 crc kubenswrapper[4971]: I1213 08:43:55.784044 4971 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="59d1fcc5-abdc-47ae-9686-bb7eddf8ded5" path="/var/lib/kubelet/pods/59d1fcc5-abdc-47ae-9686-bb7eddf8ded5/volumes" Dec 13 08:43:55 crc kubenswrapper[4971]: E1213 08:43:55.927146 4971 cadvisor_stats_provider.go:516] "Partial failure issuing cadvisor.ContainerInfoV2" err="partial failures: [\"/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podd9cc4da9_21c9_441d_8a83_4e49ac416fed.slice\": RecentStats: unable to find data in memory cache]" Dec 13 08:43:56 crc kubenswrapper[4971]: I1213 08:43:56.000968 4971 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-dxrnq" event={"ID":"d9cc4da9-21c9-441d-8a83-4e49ac416fed","Type":"ContainerDied","Data":"28f30b4c5ca77fadb2b4a9ef5b34a1456f89844dab555d8bf315652020241e3c"} Dec 13 08:43:56 crc kubenswrapper[4971]: I1213 08:43:56.001642 4971 scope.go:117] "RemoveContainer" containerID="06d47f4781d33377772464259a60f13f60e40914bd75c0443130682b470f3d1c" Dec 13 08:43:56 crc kubenswrapper[4971]: I1213 08:43:56.001221 4971 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-dxrnq" Dec 13 08:43:56 crc kubenswrapper[4971]: I1213 08:43:56.028148 4971 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-dxrnq"] Dec 13 08:43:56 crc kubenswrapper[4971]: I1213 08:43:56.037886 4971 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-marketplace-dxrnq"] Dec 13 08:43:56 crc kubenswrapper[4971]: I1213 08:43:56.038210 4971 scope.go:117] "RemoveContainer" containerID="3488e3ead0a246e8284cda57ff1190a35b1eebed65f2f01018af2e554c566bdb" Dec 13 08:43:56 crc kubenswrapper[4971]: I1213 08:43:56.069775 4971 scope.go:117] "RemoveContainer" containerID="fbd1ab6daa33e2828726ea648009f864c764f20ef18d98e7cea389768e77e9d0" Dec 13 08:43:57 crc kubenswrapper[4971]: I1213 08:43:57.782668 4971 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="d9cc4da9-21c9-441d-8a83-4e49ac416fed" path="/var/lib/kubelet/pods/d9cc4da9-21c9-441d-8a83-4e49ac416fed/volumes" Dec 13 08:44:16 crc kubenswrapper[4971]: I1213 08:44:16.153729 4971 patch_prober.go:28] interesting pod/machine-config-daemon-82xjz container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 13 08:44:16 crc kubenswrapper[4971]: I1213 08:44:16.154792 4971 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-82xjz" podUID="e0c4af12-edda-41d7-9b44-f87396174ef2" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 13 08:44:46 crc kubenswrapper[4971]: I1213 08:44:46.155049 4971 patch_prober.go:28] interesting pod/machine-config-daemon-82xjz container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 13 08:44:46 crc kubenswrapper[4971]: I1213 08:44:46.156113 4971 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-82xjz" podUID="e0c4af12-edda-41d7-9b44-f87396174ef2" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 13 08:45:00 crc kubenswrapper[4971]: I1213 08:45:00.177588 4971 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29426925-rlgh8"] Dec 13 08:45:00 crc kubenswrapper[4971]: E1213 08:45:00.179290 4971 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="d9cc4da9-21c9-441d-8a83-4e49ac416fed" containerName="extract-utilities" Dec 13 08:45:00 crc kubenswrapper[4971]: I1213 08:45:00.179309 4971 state_mem.go:107] "Deleted CPUSet assignment" podUID="d9cc4da9-21c9-441d-8a83-4e49ac416fed" containerName="extract-utilities" Dec 13 08:45:00 crc kubenswrapper[4971]: E1213 08:45:00.179328 4971 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="d9cc4da9-21c9-441d-8a83-4e49ac416fed" containerName="extract-content" Dec 13 08:45:00 crc kubenswrapper[4971]: I1213 08:45:00.179336 4971 state_mem.go:107] "Deleted CPUSet assignment" podUID="d9cc4da9-21c9-441d-8a83-4e49ac416fed" containerName="extract-content" Dec 13 08:45:00 crc kubenswrapper[4971]: E1213 08:45:00.179347 4971 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="59d1fcc5-abdc-47ae-9686-bb7eddf8ded5" containerName="extract-content" Dec 13 08:45:00 crc kubenswrapper[4971]: I1213 08:45:00.179354 4971 state_mem.go:107] "Deleted CPUSet assignment" podUID="59d1fcc5-abdc-47ae-9686-bb7eddf8ded5" containerName="extract-content" Dec 13 08:45:00 crc kubenswrapper[4971]: E1213 08:45:00.179364 4971 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="d9cc4da9-21c9-441d-8a83-4e49ac416fed" containerName="registry-server" Dec 13 08:45:00 crc kubenswrapper[4971]: I1213 08:45:00.179370 4971 state_mem.go:107] "Deleted CPUSet assignment" podUID="d9cc4da9-21c9-441d-8a83-4e49ac416fed" containerName="registry-server" Dec 13 08:45:00 crc kubenswrapper[4971]: E1213 08:45:00.179387 4971 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="59d1fcc5-abdc-47ae-9686-bb7eddf8ded5" containerName="extract-utilities" Dec 13 08:45:00 crc kubenswrapper[4971]: I1213 08:45:00.179393 4971 state_mem.go:107] "Deleted CPUSet assignment" podUID="59d1fcc5-abdc-47ae-9686-bb7eddf8ded5" containerName="extract-utilities" Dec 13 08:45:00 crc kubenswrapper[4971]: E1213 08:45:00.179409 4971 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="59d1fcc5-abdc-47ae-9686-bb7eddf8ded5" containerName="registry-server" Dec 13 08:45:00 crc kubenswrapper[4971]: I1213 08:45:00.179418 4971 state_mem.go:107] "Deleted CPUSet assignment" podUID="59d1fcc5-abdc-47ae-9686-bb7eddf8ded5" containerName="registry-server" Dec 13 08:45:00 crc kubenswrapper[4971]: I1213 08:45:00.179655 4971 memory_manager.go:354] "RemoveStaleState removing state" podUID="59d1fcc5-abdc-47ae-9686-bb7eddf8ded5" containerName="registry-server" Dec 13 08:45:00 crc kubenswrapper[4971]: I1213 08:45:00.179673 4971 memory_manager.go:354] "RemoveStaleState removing state" podUID="d9cc4da9-21c9-441d-8a83-4e49ac416fed" containerName="registry-server" Dec 13 08:45:00 crc kubenswrapper[4971]: I1213 08:45:00.180618 4971 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29426925-rlgh8" Dec 13 08:45:00 crc kubenswrapper[4971]: I1213 08:45:00.183173 4971 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"collect-profiles-dockercfg-kzf4t" Dec 13 08:45:00 crc kubenswrapper[4971]: I1213 08:45:00.185482 4971 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-operator-lifecycle-manager"/"collect-profiles-config" Dec 13 08:45:00 crc kubenswrapper[4971]: I1213 08:45:00.187931 4971 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29426925-rlgh8"] Dec 13 08:45:00 crc kubenswrapper[4971]: I1213 08:45:00.281001 4971 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/83f18279-1f12-4157-8aad-1e1de8bdbbdf-config-volume\") pod \"collect-profiles-29426925-rlgh8\" (UID: \"83f18279-1f12-4157-8aad-1e1de8bdbbdf\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29426925-rlgh8" Dec 13 08:45:00 crc kubenswrapper[4971]: I1213 08:45:00.281145 4971 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/83f18279-1f12-4157-8aad-1e1de8bdbbdf-secret-volume\") pod \"collect-profiles-29426925-rlgh8\" (UID: \"83f18279-1f12-4157-8aad-1e1de8bdbbdf\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29426925-rlgh8" Dec 13 08:45:00 crc kubenswrapper[4971]: I1213 08:45:00.281227 4971 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-w684p\" (UniqueName: \"kubernetes.io/projected/83f18279-1f12-4157-8aad-1e1de8bdbbdf-kube-api-access-w684p\") pod \"collect-profiles-29426925-rlgh8\" (UID: \"83f18279-1f12-4157-8aad-1e1de8bdbbdf\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29426925-rlgh8" Dec 13 08:45:00 crc kubenswrapper[4971]: I1213 08:45:00.383802 4971 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/83f18279-1f12-4157-8aad-1e1de8bdbbdf-secret-volume\") pod \"collect-profiles-29426925-rlgh8\" (UID: \"83f18279-1f12-4157-8aad-1e1de8bdbbdf\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29426925-rlgh8" Dec 13 08:45:00 crc kubenswrapper[4971]: I1213 08:45:00.383936 4971 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-w684p\" (UniqueName: \"kubernetes.io/projected/83f18279-1f12-4157-8aad-1e1de8bdbbdf-kube-api-access-w684p\") pod \"collect-profiles-29426925-rlgh8\" (UID: \"83f18279-1f12-4157-8aad-1e1de8bdbbdf\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29426925-rlgh8" Dec 13 08:45:00 crc kubenswrapper[4971]: I1213 08:45:00.384016 4971 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/83f18279-1f12-4157-8aad-1e1de8bdbbdf-config-volume\") pod \"collect-profiles-29426925-rlgh8\" (UID: \"83f18279-1f12-4157-8aad-1e1de8bdbbdf\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29426925-rlgh8" Dec 13 08:45:00 crc kubenswrapper[4971]: I1213 08:45:00.385284 4971 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/83f18279-1f12-4157-8aad-1e1de8bdbbdf-config-volume\") pod \"collect-profiles-29426925-rlgh8\" (UID: \"83f18279-1f12-4157-8aad-1e1de8bdbbdf\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29426925-rlgh8" Dec 13 08:45:00 crc kubenswrapper[4971]: I1213 08:45:00.398492 4971 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/83f18279-1f12-4157-8aad-1e1de8bdbbdf-secret-volume\") pod \"collect-profiles-29426925-rlgh8\" (UID: \"83f18279-1f12-4157-8aad-1e1de8bdbbdf\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29426925-rlgh8" Dec 13 08:45:00 crc kubenswrapper[4971]: I1213 08:45:00.407333 4971 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-w684p\" (UniqueName: \"kubernetes.io/projected/83f18279-1f12-4157-8aad-1e1de8bdbbdf-kube-api-access-w684p\") pod \"collect-profiles-29426925-rlgh8\" (UID: \"83f18279-1f12-4157-8aad-1e1de8bdbbdf\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29426925-rlgh8" Dec 13 08:45:00 crc kubenswrapper[4971]: I1213 08:45:00.507685 4971 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29426925-rlgh8" Dec 13 08:45:01 crc kubenswrapper[4971]: I1213 08:45:01.374856 4971 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29426925-rlgh8"] Dec 13 08:45:01 crc kubenswrapper[4971]: I1213 08:45:01.897084 4971 generic.go:334] "Generic (PLEG): container finished" podID="83f18279-1f12-4157-8aad-1e1de8bdbbdf" containerID="db5a314515777dab5dfa0631e2f128f2f80f9846c80a9d3343d7959ea5da7a4a" exitCode=0 Dec 13 08:45:01 crc kubenswrapper[4971]: I1213 08:45:01.897156 4971 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29426925-rlgh8" event={"ID":"83f18279-1f12-4157-8aad-1e1de8bdbbdf","Type":"ContainerDied","Data":"db5a314515777dab5dfa0631e2f128f2f80f9846c80a9d3343d7959ea5da7a4a"} Dec 13 08:45:01 crc kubenswrapper[4971]: I1213 08:45:01.897230 4971 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29426925-rlgh8" event={"ID":"83f18279-1f12-4157-8aad-1e1de8bdbbdf","Type":"ContainerStarted","Data":"9f02ea070d997ca97e1947c6f3d7caa0e1a4f2df2e76a5fd2a3e23cd7c507f0d"} Dec 13 08:45:03 crc kubenswrapper[4971]: I1213 08:45:03.525929 4971 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29426925-rlgh8" Dec 13 08:45:03 crc kubenswrapper[4971]: I1213 08:45:03.621982 4971 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/83f18279-1f12-4157-8aad-1e1de8bdbbdf-config-volume\") pod \"83f18279-1f12-4157-8aad-1e1de8bdbbdf\" (UID: \"83f18279-1f12-4157-8aad-1e1de8bdbbdf\") " Dec 13 08:45:03 crc kubenswrapper[4971]: I1213 08:45:03.622080 4971 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-w684p\" (UniqueName: \"kubernetes.io/projected/83f18279-1f12-4157-8aad-1e1de8bdbbdf-kube-api-access-w684p\") pod \"83f18279-1f12-4157-8aad-1e1de8bdbbdf\" (UID: \"83f18279-1f12-4157-8aad-1e1de8bdbbdf\") " Dec 13 08:45:03 crc kubenswrapper[4971]: I1213 08:45:03.622116 4971 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/83f18279-1f12-4157-8aad-1e1de8bdbbdf-secret-volume\") pod \"83f18279-1f12-4157-8aad-1e1de8bdbbdf\" (UID: \"83f18279-1f12-4157-8aad-1e1de8bdbbdf\") " Dec 13 08:45:03 crc kubenswrapper[4971]: I1213 08:45:03.622776 4971 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/83f18279-1f12-4157-8aad-1e1de8bdbbdf-config-volume" (OuterVolumeSpecName: "config-volume") pod "83f18279-1f12-4157-8aad-1e1de8bdbbdf" (UID: "83f18279-1f12-4157-8aad-1e1de8bdbbdf"). InnerVolumeSpecName "config-volume". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 13 08:45:03 crc kubenswrapper[4971]: I1213 08:45:03.630733 4971 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/83f18279-1f12-4157-8aad-1e1de8bdbbdf-secret-volume" (OuterVolumeSpecName: "secret-volume") pod "83f18279-1f12-4157-8aad-1e1de8bdbbdf" (UID: "83f18279-1f12-4157-8aad-1e1de8bdbbdf"). InnerVolumeSpecName "secret-volume". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 13 08:45:03 crc kubenswrapper[4971]: I1213 08:45:03.637096 4971 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/83f18279-1f12-4157-8aad-1e1de8bdbbdf-kube-api-access-w684p" (OuterVolumeSpecName: "kube-api-access-w684p") pod "83f18279-1f12-4157-8aad-1e1de8bdbbdf" (UID: "83f18279-1f12-4157-8aad-1e1de8bdbbdf"). InnerVolumeSpecName "kube-api-access-w684p". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 13 08:45:03 crc kubenswrapper[4971]: I1213 08:45:03.727474 4971 reconciler_common.go:293] "Volume detached for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/83f18279-1f12-4157-8aad-1e1de8bdbbdf-config-volume\") on node \"crc\" DevicePath \"\"" Dec 13 08:45:03 crc kubenswrapper[4971]: I1213 08:45:03.727796 4971 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-w684p\" (UniqueName: \"kubernetes.io/projected/83f18279-1f12-4157-8aad-1e1de8bdbbdf-kube-api-access-w684p\") on node \"crc\" DevicePath \"\"" Dec 13 08:45:03 crc kubenswrapper[4971]: I1213 08:45:03.727818 4971 reconciler_common.go:293] "Volume detached for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/83f18279-1f12-4157-8aad-1e1de8bdbbdf-secret-volume\") on node \"crc\" DevicePath \"\"" Dec 13 08:45:03 crc kubenswrapper[4971]: I1213 08:45:03.927235 4971 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29426925-rlgh8" event={"ID":"83f18279-1f12-4157-8aad-1e1de8bdbbdf","Type":"ContainerDied","Data":"9f02ea070d997ca97e1947c6f3d7caa0e1a4f2df2e76a5fd2a3e23cd7c507f0d"} Dec 13 08:45:03 crc kubenswrapper[4971]: I1213 08:45:03.927290 4971 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="9f02ea070d997ca97e1947c6f3d7caa0e1a4f2df2e76a5fd2a3e23cd7c507f0d" Dec 13 08:45:03 crc kubenswrapper[4971]: I1213 08:45:03.927344 4971 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29426925-rlgh8" Dec 13 08:45:04 crc kubenswrapper[4971]: I1213 08:45:04.625109 4971 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29426880-2x7zw"] Dec 13 08:45:04 crc kubenswrapper[4971]: I1213 08:45:04.643151 4971 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29426880-2x7zw"] Dec 13 08:45:05 crc kubenswrapper[4971]: I1213 08:45:05.783904 4971 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="b9898935-5232-4bdb-846d-27084d2a8be6" path="/var/lib/kubelet/pods/b9898935-5232-4bdb-846d-27084d2a8be6/volumes" Dec 13 08:45:16 crc kubenswrapper[4971]: I1213 08:45:16.153920 4971 patch_prober.go:28] interesting pod/machine-config-daemon-82xjz container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 13 08:45:16 crc kubenswrapper[4971]: I1213 08:45:16.155007 4971 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-82xjz" podUID="e0c4af12-edda-41d7-9b44-f87396174ef2" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 13 08:45:16 crc kubenswrapper[4971]: I1213 08:45:16.155348 4971 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-82xjz" Dec 13 08:45:16 crc kubenswrapper[4971]: I1213 08:45:16.156429 4971 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"513b55f2bb24361c2480216b9b7a4b3b6659eda1ac7c17acc12836e086669008"} pod="openshift-machine-config-operator/machine-config-daemon-82xjz" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Dec 13 08:45:16 crc kubenswrapper[4971]: I1213 08:45:16.156507 4971 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-82xjz" podUID="e0c4af12-edda-41d7-9b44-f87396174ef2" containerName="machine-config-daemon" containerID="cri-o://513b55f2bb24361c2480216b9b7a4b3b6659eda1ac7c17acc12836e086669008" gracePeriod=600 Dec 13 08:45:17 crc kubenswrapper[4971]: I1213 08:45:17.232239 4971 generic.go:334] "Generic (PLEG): container finished" podID="e0c4af12-edda-41d7-9b44-f87396174ef2" containerID="513b55f2bb24361c2480216b9b7a4b3b6659eda1ac7c17acc12836e086669008" exitCode=0 Dec 13 08:45:17 crc kubenswrapper[4971]: I1213 08:45:17.232544 4971 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-82xjz" event={"ID":"e0c4af12-edda-41d7-9b44-f87396174ef2","Type":"ContainerDied","Data":"513b55f2bb24361c2480216b9b7a4b3b6659eda1ac7c17acc12836e086669008"} Dec 13 08:45:17 crc kubenswrapper[4971]: I1213 08:45:17.233362 4971 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-82xjz" event={"ID":"e0c4af12-edda-41d7-9b44-f87396174ef2","Type":"ContainerStarted","Data":"b875432558a41dd98c673e3df9c78c3a2bcd69eb63408b435826e3d82347d2a6"} Dec 13 08:45:17 crc kubenswrapper[4971]: I1213 08:45:17.233415 4971 scope.go:117] "RemoveContainer" containerID="f397fe912a74e38413c0149a7e7ca0ed1cff4acbea98dcbfa69153e27d3fdaa0" Dec 13 08:45:28 crc kubenswrapper[4971]: I1213 08:45:28.110398 4971 scope.go:117] "RemoveContainer" containerID="802b1f844ed17148a197d5f8e8c87f118de71e681fde826e33cd782b93ef770a" Dec 13 08:47:16 crc kubenswrapper[4971]: I1213 08:47:16.153807 4971 patch_prober.go:28] interesting pod/machine-config-daemon-82xjz container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 13 08:47:16 crc kubenswrapper[4971]: I1213 08:47:16.155745 4971 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-82xjz" podUID="e0c4af12-edda-41d7-9b44-f87396174ef2" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 13 08:47:46 crc kubenswrapper[4971]: I1213 08:47:46.154447 4971 patch_prober.go:28] interesting pod/machine-config-daemon-82xjz container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 13 08:47:46 crc kubenswrapper[4971]: I1213 08:47:46.155842 4971 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-82xjz" podUID="e0c4af12-edda-41d7-9b44-f87396174ef2" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 13 08:48:16 crc kubenswrapper[4971]: I1213 08:48:16.154057 4971 patch_prober.go:28] interesting pod/machine-config-daemon-82xjz container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 13 08:48:16 crc kubenswrapper[4971]: I1213 08:48:16.154896 4971 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-82xjz" podUID="e0c4af12-edda-41d7-9b44-f87396174ef2" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 13 08:48:16 crc kubenswrapper[4971]: I1213 08:48:16.154962 4971 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-82xjz" Dec 13 08:48:16 crc kubenswrapper[4971]: I1213 08:48:16.155966 4971 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"b875432558a41dd98c673e3df9c78c3a2bcd69eb63408b435826e3d82347d2a6"} pod="openshift-machine-config-operator/machine-config-daemon-82xjz" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Dec 13 08:48:16 crc kubenswrapper[4971]: I1213 08:48:16.156043 4971 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-82xjz" podUID="e0c4af12-edda-41d7-9b44-f87396174ef2" containerName="machine-config-daemon" containerID="cri-o://b875432558a41dd98c673e3df9c78c3a2bcd69eb63408b435826e3d82347d2a6" gracePeriod=600 Dec 13 08:48:16 crc kubenswrapper[4971]: E1213 08:48:16.289917 4971 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-82xjz_openshift-machine-config-operator(e0c4af12-edda-41d7-9b44-f87396174ef2)\"" pod="openshift-machine-config-operator/machine-config-daemon-82xjz" podUID="e0c4af12-edda-41d7-9b44-f87396174ef2" Dec 13 08:48:16 crc kubenswrapper[4971]: I1213 08:48:16.782953 4971 generic.go:334] "Generic (PLEG): container finished" podID="e0c4af12-edda-41d7-9b44-f87396174ef2" containerID="b875432558a41dd98c673e3df9c78c3a2bcd69eb63408b435826e3d82347d2a6" exitCode=0 Dec 13 08:48:16 crc kubenswrapper[4971]: I1213 08:48:16.783027 4971 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-82xjz" event={"ID":"e0c4af12-edda-41d7-9b44-f87396174ef2","Type":"ContainerDied","Data":"b875432558a41dd98c673e3df9c78c3a2bcd69eb63408b435826e3d82347d2a6"} Dec 13 08:48:16 crc kubenswrapper[4971]: I1213 08:48:16.783078 4971 scope.go:117] "RemoveContainer" containerID="513b55f2bb24361c2480216b9b7a4b3b6659eda1ac7c17acc12836e086669008" Dec 13 08:48:16 crc kubenswrapper[4971]: I1213 08:48:16.784104 4971 scope.go:117] "RemoveContainer" containerID="b875432558a41dd98c673e3df9c78c3a2bcd69eb63408b435826e3d82347d2a6" Dec 13 08:48:16 crc kubenswrapper[4971]: E1213 08:48:16.784393 4971 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-82xjz_openshift-machine-config-operator(e0c4af12-edda-41d7-9b44-f87396174ef2)\"" pod="openshift-machine-config-operator/machine-config-daemon-82xjz" podUID="e0c4af12-edda-41d7-9b44-f87396174ef2" Dec 13 08:48:30 crc kubenswrapper[4971]: I1213 08:48:30.770159 4971 scope.go:117] "RemoveContainer" containerID="b875432558a41dd98c673e3df9c78c3a2bcd69eb63408b435826e3d82347d2a6" Dec 13 08:48:30 crc kubenswrapper[4971]: E1213 08:48:30.771664 4971 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-82xjz_openshift-machine-config-operator(e0c4af12-edda-41d7-9b44-f87396174ef2)\"" pod="openshift-machine-config-operator/machine-config-daemon-82xjz" podUID="e0c4af12-edda-41d7-9b44-f87396174ef2" Dec 13 08:48:43 crc kubenswrapper[4971]: I1213 08:48:43.776716 4971 scope.go:117] "RemoveContainer" containerID="b875432558a41dd98c673e3df9c78c3a2bcd69eb63408b435826e3d82347d2a6" Dec 13 08:48:43 crc kubenswrapper[4971]: E1213 08:48:43.778009 4971 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-82xjz_openshift-machine-config-operator(e0c4af12-edda-41d7-9b44-f87396174ef2)\"" pod="openshift-machine-config-operator/machine-config-daemon-82xjz" podUID="e0c4af12-edda-41d7-9b44-f87396174ef2" Dec 13 08:48:53 crc kubenswrapper[4971]: I1213 08:48:53.694616 4971 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/certified-operators-nljkd"] Dec 13 08:48:53 crc kubenswrapper[4971]: E1213 08:48:53.695834 4971 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="83f18279-1f12-4157-8aad-1e1de8bdbbdf" containerName="collect-profiles" Dec 13 08:48:53 crc kubenswrapper[4971]: I1213 08:48:53.695856 4971 state_mem.go:107] "Deleted CPUSet assignment" podUID="83f18279-1f12-4157-8aad-1e1de8bdbbdf" containerName="collect-profiles" Dec 13 08:48:53 crc kubenswrapper[4971]: I1213 08:48:53.696180 4971 memory_manager.go:354] "RemoveStaleState removing state" podUID="83f18279-1f12-4157-8aad-1e1de8bdbbdf" containerName="collect-profiles" Dec 13 08:48:53 crc kubenswrapper[4971]: I1213 08:48:53.698132 4971 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-nljkd" Dec 13 08:48:53 crc kubenswrapper[4971]: I1213 08:48:53.711170 4971 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-nljkd"] Dec 13 08:48:53 crc kubenswrapper[4971]: I1213 08:48:53.829098 4971 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-6xtsq\" (UniqueName: \"kubernetes.io/projected/d4a6371a-9ede-4784-ab66-99e3908defd2-kube-api-access-6xtsq\") pod \"certified-operators-nljkd\" (UID: \"d4a6371a-9ede-4784-ab66-99e3908defd2\") " pod="openshift-marketplace/certified-operators-nljkd" Dec 13 08:48:53 crc kubenswrapper[4971]: I1213 08:48:53.829260 4971 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/d4a6371a-9ede-4784-ab66-99e3908defd2-catalog-content\") pod \"certified-operators-nljkd\" (UID: \"d4a6371a-9ede-4784-ab66-99e3908defd2\") " pod="openshift-marketplace/certified-operators-nljkd" Dec 13 08:48:53 crc kubenswrapper[4971]: I1213 08:48:53.829396 4971 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/d4a6371a-9ede-4784-ab66-99e3908defd2-utilities\") pod \"certified-operators-nljkd\" (UID: \"d4a6371a-9ede-4784-ab66-99e3908defd2\") " pod="openshift-marketplace/certified-operators-nljkd" Dec 13 08:48:53 crc kubenswrapper[4971]: I1213 08:48:53.899015 4971 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/community-operators-9d94b"] Dec 13 08:48:53 crc kubenswrapper[4971]: I1213 08:48:53.901569 4971 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-9d94b" Dec 13 08:48:53 crc kubenswrapper[4971]: I1213 08:48:53.912576 4971 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-9d94b"] Dec 13 08:48:53 crc kubenswrapper[4971]: I1213 08:48:53.932925 4971 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-6xtsq\" (UniqueName: \"kubernetes.io/projected/d4a6371a-9ede-4784-ab66-99e3908defd2-kube-api-access-6xtsq\") pod \"certified-operators-nljkd\" (UID: \"d4a6371a-9ede-4784-ab66-99e3908defd2\") " pod="openshift-marketplace/certified-operators-nljkd" Dec 13 08:48:53 crc kubenswrapper[4971]: I1213 08:48:53.933181 4971 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/d4a6371a-9ede-4784-ab66-99e3908defd2-catalog-content\") pod \"certified-operators-nljkd\" (UID: \"d4a6371a-9ede-4784-ab66-99e3908defd2\") " pod="openshift-marketplace/certified-operators-nljkd" Dec 13 08:48:53 crc kubenswrapper[4971]: I1213 08:48:53.933387 4971 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/d4a6371a-9ede-4784-ab66-99e3908defd2-utilities\") pod \"certified-operators-nljkd\" (UID: \"d4a6371a-9ede-4784-ab66-99e3908defd2\") " pod="openshift-marketplace/certified-operators-nljkd" Dec 13 08:48:53 crc kubenswrapper[4971]: I1213 08:48:53.935544 4971 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/d4a6371a-9ede-4784-ab66-99e3908defd2-catalog-content\") pod \"certified-operators-nljkd\" (UID: \"d4a6371a-9ede-4784-ab66-99e3908defd2\") " pod="openshift-marketplace/certified-operators-nljkd" Dec 13 08:48:53 crc kubenswrapper[4971]: I1213 08:48:53.937006 4971 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/d4a6371a-9ede-4784-ab66-99e3908defd2-utilities\") pod \"certified-operators-nljkd\" (UID: \"d4a6371a-9ede-4784-ab66-99e3908defd2\") " pod="openshift-marketplace/certified-operators-nljkd" Dec 13 08:48:53 crc kubenswrapper[4971]: I1213 08:48:53.976244 4971 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-6xtsq\" (UniqueName: \"kubernetes.io/projected/d4a6371a-9ede-4784-ab66-99e3908defd2-kube-api-access-6xtsq\") pod \"certified-operators-nljkd\" (UID: \"d4a6371a-9ede-4784-ab66-99e3908defd2\") " pod="openshift-marketplace/certified-operators-nljkd" Dec 13 08:48:54 crc kubenswrapper[4971]: I1213 08:48:54.022580 4971 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-nljkd" Dec 13 08:48:54 crc kubenswrapper[4971]: I1213 08:48:54.035891 4971 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/76ade42c-67ee-45b2-9b31-62002698a753-utilities\") pod \"community-operators-9d94b\" (UID: \"76ade42c-67ee-45b2-9b31-62002698a753\") " pod="openshift-marketplace/community-operators-9d94b" Dec 13 08:48:54 crc kubenswrapper[4971]: I1213 08:48:54.036335 4971 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/76ade42c-67ee-45b2-9b31-62002698a753-catalog-content\") pod \"community-operators-9d94b\" (UID: \"76ade42c-67ee-45b2-9b31-62002698a753\") " pod="openshift-marketplace/community-operators-9d94b" Dec 13 08:48:54 crc kubenswrapper[4971]: I1213 08:48:54.036496 4971 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-br82l\" (UniqueName: \"kubernetes.io/projected/76ade42c-67ee-45b2-9b31-62002698a753-kube-api-access-br82l\") pod \"community-operators-9d94b\" (UID: \"76ade42c-67ee-45b2-9b31-62002698a753\") " pod="openshift-marketplace/community-operators-9d94b" Dec 13 08:48:54 crc kubenswrapper[4971]: I1213 08:48:54.139896 4971 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-br82l\" (UniqueName: \"kubernetes.io/projected/76ade42c-67ee-45b2-9b31-62002698a753-kube-api-access-br82l\") pod \"community-operators-9d94b\" (UID: \"76ade42c-67ee-45b2-9b31-62002698a753\") " pod="openshift-marketplace/community-operators-9d94b" Dec 13 08:48:54 crc kubenswrapper[4971]: I1213 08:48:54.140085 4971 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/76ade42c-67ee-45b2-9b31-62002698a753-utilities\") pod \"community-operators-9d94b\" (UID: \"76ade42c-67ee-45b2-9b31-62002698a753\") " pod="openshift-marketplace/community-operators-9d94b" Dec 13 08:48:54 crc kubenswrapper[4971]: I1213 08:48:54.140114 4971 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/76ade42c-67ee-45b2-9b31-62002698a753-catalog-content\") pod \"community-operators-9d94b\" (UID: \"76ade42c-67ee-45b2-9b31-62002698a753\") " pod="openshift-marketplace/community-operators-9d94b" Dec 13 08:48:54 crc kubenswrapper[4971]: I1213 08:48:54.140821 4971 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/76ade42c-67ee-45b2-9b31-62002698a753-catalog-content\") pod \"community-operators-9d94b\" (UID: \"76ade42c-67ee-45b2-9b31-62002698a753\") " pod="openshift-marketplace/community-operators-9d94b" Dec 13 08:48:54 crc kubenswrapper[4971]: I1213 08:48:54.141504 4971 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/76ade42c-67ee-45b2-9b31-62002698a753-utilities\") pod \"community-operators-9d94b\" (UID: \"76ade42c-67ee-45b2-9b31-62002698a753\") " pod="openshift-marketplace/community-operators-9d94b" Dec 13 08:48:54 crc kubenswrapper[4971]: I1213 08:48:54.170083 4971 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-br82l\" (UniqueName: \"kubernetes.io/projected/76ade42c-67ee-45b2-9b31-62002698a753-kube-api-access-br82l\") pod \"community-operators-9d94b\" (UID: \"76ade42c-67ee-45b2-9b31-62002698a753\") " pod="openshift-marketplace/community-operators-9d94b" Dec 13 08:48:54 crc kubenswrapper[4971]: I1213 08:48:54.232422 4971 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-9d94b" Dec 13 08:48:54 crc kubenswrapper[4971]: I1213 08:48:54.889538 4971 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-nljkd"] Dec 13 08:48:55 crc kubenswrapper[4971]: I1213 08:48:55.085428 4971 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-9d94b"] Dec 13 08:48:55 crc kubenswrapper[4971]: W1213 08:48:55.087638 4971 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod76ade42c_67ee_45b2_9b31_62002698a753.slice/crio-4902985fb7b5c038c08ffaf044c7b5bde772daee626da519af65cbf68e090d61 WatchSource:0}: Error finding container 4902985fb7b5c038c08ffaf044c7b5bde772daee626da519af65cbf68e090d61: Status 404 returned error can't find the container with id 4902985fb7b5c038c08ffaf044c7b5bde772daee626da519af65cbf68e090d61 Dec 13 08:48:55 crc kubenswrapper[4971]: I1213 08:48:55.235032 4971 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-nljkd" event={"ID":"d4a6371a-9ede-4784-ab66-99e3908defd2","Type":"ContainerStarted","Data":"63ab678671f2b267c76588b6549b097973f4637f9c088006ff8d4cd1fbb2681b"} Dec 13 08:48:55 crc kubenswrapper[4971]: I1213 08:48:55.235787 4971 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-nljkd" event={"ID":"d4a6371a-9ede-4784-ab66-99e3908defd2","Type":"ContainerStarted","Data":"eba88caf9068b1976ed0d0cd30598f962d83d3d4f2a46e5b149ef2e3752789c0"} Dec 13 08:48:55 crc kubenswrapper[4971]: I1213 08:48:55.239577 4971 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-9d94b" event={"ID":"76ade42c-67ee-45b2-9b31-62002698a753","Type":"ContainerStarted","Data":"4902985fb7b5c038c08ffaf044c7b5bde772daee626da519af65cbf68e090d61"} Dec 13 08:48:56 crc kubenswrapper[4971]: I1213 08:48:56.253724 4971 generic.go:334] "Generic (PLEG): container finished" podID="d4a6371a-9ede-4784-ab66-99e3908defd2" containerID="63ab678671f2b267c76588b6549b097973f4637f9c088006ff8d4cd1fbb2681b" exitCode=0 Dec 13 08:48:56 crc kubenswrapper[4971]: I1213 08:48:56.253861 4971 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-nljkd" event={"ID":"d4a6371a-9ede-4784-ab66-99e3908defd2","Type":"ContainerDied","Data":"63ab678671f2b267c76588b6549b097973f4637f9c088006ff8d4cd1fbb2681b"} Dec 13 08:48:56 crc kubenswrapper[4971]: I1213 08:48:56.258222 4971 provider.go:102] Refreshing cache for provider: *credentialprovider.defaultDockerConfigProvider Dec 13 08:48:56 crc kubenswrapper[4971]: I1213 08:48:56.264898 4971 generic.go:334] "Generic (PLEG): container finished" podID="76ade42c-67ee-45b2-9b31-62002698a753" containerID="745e0a204fb6e0429a43f609a4054aa36cf6ec651600a27da9971129b1406d73" exitCode=0 Dec 13 08:48:56 crc kubenswrapper[4971]: I1213 08:48:56.264970 4971 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-9d94b" event={"ID":"76ade42c-67ee-45b2-9b31-62002698a753","Type":"ContainerDied","Data":"745e0a204fb6e0429a43f609a4054aa36cf6ec651600a27da9971129b1406d73"} Dec 13 08:48:57 crc kubenswrapper[4971]: I1213 08:48:57.282614 4971 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-nljkd" event={"ID":"d4a6371a-9ede-4784-ab66-99e3908defd2","Type":"ContainerStarted","Data":"f9c88220f3b23e03919489ce5b5fd5827a691e8b8645526e7efb99e5cd1da38e"} Dec 13 08:48:57 crc kubenswrapper[4971]: I1213 08:48:57.770039 4971 scope.go:117] "RemoveContainer" containerID="b875432558a41dd98c673e3df9c78c3a2bcd69eb63408b435826e3d82347d2a6" Dec 13 08:48:57 crc kubenswrapper[4971]: E1213 08:48:57.770780 4971 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-82xjz_openshift-machine-config-operator(e0c4af12-edda-41d7-9b44-f87396174ef2)\"" pod="openshift-machine-config-operator/machine-config-daemon-82xjz" podUID="e0c4af12-edda-41d7-9b44-f87396174ef2" Dec 13 08:48:58 crc kubenswrapper[4971]: I1213 08:48:58.295975 4971 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-9d94b" event={"ID":"76ade42c-67ee-45b2-9b31-62002698a753","Type":"ContainerStarted","Data":"d0297777d8c1e6e8ec1156e50c3a545c6243654391295cdc41ef7daa7f5f2db8"} Dec 13 08:48:59 crc kubenswrapper[4971]: I1213 08:48:59.311297 4971 generic.go:334] "Generic (PLEG): container finished" podID="76ade42c-67ee-45b2-9b31-62002698a753" containerID="d0297777d8c1e6e8ec1156e50c3a545c6243654391295cdc41ef7daa7f5f2db8" exitCode=0 Dec 13 08:48:59 crc kubenswrapper[4971]: I1213 08:48:59.311355 4971 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-9d94b" event={"ID":"76ade42c-67ee-45b2-9b31-62002698a753","Type":"ContainerDied","Data":"d0297777d8c1e6e8ec1156e50c3a545c6243654391295cdc41ef7daa7f5f2db8"} Dec 13 08:49:00 crc kubenswrapper[4971]: I1213 08:49:00.326194 4971 generic.go:334] "Generic (PLEG): container finished" podID="d4a6371a-9ede-4784-ab66-99e3908defd2" containerID="f9c88220f3b23e03919489ce5b5fd5827a691e8b8645526e7efb99e5cd1da38e" exitCode=0 Dec 13 08:49:00 crc kubenswrapper[4971]: I1213 08:49:00.326294 4971 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-nljkd" event={"ID":"d4a6371a-9ede-4784-ab66-99e3908defd2","Type":"ContainerDied","Data":"f9c88220f3b23e03919489ce5b5fd5827a691e8b8645526e7efb99e5cd1da38e"} Dec 13 08:49:00 crc kubenswrapper[4971]: I1213 08:49:00.330791 4971 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-9d94b" event={"ID":"76ade42c-67ee-45b2-9b31-62002698a753","Type":"ContainerStarted","Data":"6c07057c003fbf99efd497ab319fa3cc558337216cdcb6287d48db74213f3d13"} Dec 13 08:49:00 crc kubenswrapper[4971]: I1213 08:49:00.382933 4971 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/community-operators-9d94b" podStartSLOduration=3.865356665 podStartE2EDuration="7.382895524s" podCreationTimestamp="2025-12-13 08:48:53 +0000 UTC" firstStartedPulling="2025-12-13 08:48:56.267454154 +0000 UTC m=+7192.871863602" lastFinishedPulling="2025-12-13 08:48:59.784993013 +0000 UTC m=+7196.389402461" observedRunningTime="2025-12-13 08:49:00.376872693 +0000 UTC m=+7196.981282141" watchObservedRunningTime="2025-12-13 08:49:00.382895524 +0000 UTC m=+7196.987305132" Dec 13 08:49:01 crc kubenswrapper[4971]: I1213 08:49:01.344536 4971 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-nljkd" event={"ID":"d4a6371a-9ede-4784-ab66-99e3908defd2","Type":"ContainerStarted","Data":"1ee0d99fcc93a549a2dae368555cccdd0e75eb42dcc04be8146bb742a9c94e9f"} Dec 13 08:49:01 crc kubenswrapper[4971]: I1213 08:49:01.378623 4971 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/certified-operators-nljkd" podStartSLOduration=3.818498535 podStartE2EDuration="8.378592218s" podCreationTimestamp="2025-12-13 08:48:53 +0000 UTC" firstStartedPulling="2025-12-13 08:48:56.257753982 +0000 UTC m=+7192.862163430" lastFinishedPulling="2025-12-13 08:49:00.817847665 +0000 UTC m=+7197.422257113" observedRunningTime="2025-12-13 08:49:01.366152347 +0000 UTC m=+7197.970561805" watchObservedRunningTime="2025-12-13 08:49:01.378592218 +0000 UTC m=+7197.983001666" Dec 13 08:49:04 crc kubenswrapper[4971]: I1213 08:49:04.023551 4971 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/certified-operators-nljkd" Dec 13 08:49:04 crc kubenswrapper[4971]: I1213 08:49:04.023970 4971 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/certified-operators-nljkd" Dec 13 08:49:04 crc kubenswrapper[4971]: I1213 08:49:04.233827 4971 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/community-operators-9d94b" Dec 13 08:49:04 crc kubenswrapper[4971]: I1213 08:49:04.233892 4971 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/community-operators-9d94b" Dec 13 08:49:04 crc kubenswrapper[4971]: I1213 08:49:04.291615 4971 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/community-operators-9d94b" Dec 13 08:49:05 crc kubenswrapper[4971]: I1213 08:49:05.073819 4971 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-marketplace/certified-operators-nljkd" podUID="d4a6371a-9ede-4784-ab66-99e3908defd2" containerName="registry-server" probeResult="failure" output=< Dec 13 08:49:05 crc kubenswrapper[4971]: timeout: failed to connect service ":50051" within 1s Dec 13 08:49:05 crc kubenswrapper[4971]: > Dec 13 08:49:08 crc kubenswrapper[4971]: I1213 08:49:08.769375 4971 scope.go:117] "RemoveContainer" containerID="b875432558a41dd98c673e3df9c78c3a2bcd69eb63408b435826e3d82347d2a6" Dec 13 08:49:08 crc kubenswrapper[4971]: E1213 08:49:08.773136 4971 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-82xjz_openshift-machine-config-operator(e0c4af12-edda-41d7-9b44-f87396174ef2)\"" pod="openshift-machine-config-operator/machine-config-daemon-82xjz" podUID="e0c4af12-edda-41d7-9b44-f87396174ef2" Dec 13 08:49:14 crc kubenswrapper[4971]: I1213 08:49:14.075622 4971 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/certified-operators-nljkd" Dec 13 08:49:14 crc kubenswrapper[4971]: I1213 08:49:14.141628 4971 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/certified-operators-nljkd" Dec 13 08:49:14 crc kubenswrapper[4971]: I1213 08:49:14.313721 4971 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/community-operators-9d94b" Dec 13 08:49:14 crc kubenswrapper[4971]: I1213 08:49:14.333496 4971 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-nljkd"] Dec 13 08:49:15 crc kubenswrapper[4971]: I1213 08:49:15.493637 4971 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/certified-operators-nljkd" podUID="d4a6371a-9ede-4784-ab66-99e3908defd2" containerName="registry-server" containerID="cri-o://1ee0d99fcc93a549a2dae368555cccdd0e75eb42dcc04be8146bb742a9c94e9f" gracePeriod=2 Dec 13 08:49:16 crc kubenswrapper[4971]: I1213 08:49:16.136003 4971 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-nljkd" Dec 13 08:49:16 crc kubenswrapper[4971]: I1213 08:49:16.187880 4971 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/d4a6371a-9ede-4784-ab66-99e3908defd2-catalog-content\") pod \"d4a6371a-9ede-4784-ab66-99e3908defd2\" (UID: \"d4a6371a-9ede-4784-ab66-99e3908defd2\") " Dec 13 08:49:16 crc kubenswrapper[4971]: I1213 08:49:16.188753 4971 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/d4a6371a-9ede-4784-ab66-99e3908defd2-utilities\") pod \"d4a6371a-9ede-4784-ab66-99e3908defd2\" (UID: \"d4a6371a-9ede-4784-ab66-99e3908defd2\") " Dec 13 08:49:16 crc kubenswrapper[4971]: I1213 08:49:16.189018 4971 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-6xtsq\" (UniqueName: \"kubernetes.io/projected/d4a6371a-9ede-4784-ab66-99e3908defd2-kube-api-access-6xtsq\") pod \"d4a6371a-9ede-4784-ab66-99e3908defd2\" (UID: \"d4a6371a-9ede-4784-ab66-99e3908defd2\") " Dec 13 08:49:16 crc kubenswrapper[4971]: I1213 08:49:16.190620 4971 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/d4a6371a-9ede-4784-ab66-99e3908defd2-utilities" (OuterVolumeSpecName: "utilities") pod "d4a6371a-9ede-4784-ab66-99e3908defd2" (UID: "d4a6371a-9ede-4784-ab66-99e3908defd2"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 13 08:49:16 crc kubenswrapper[4971]: I1213 08:49:16.201153 4971 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/d4a6371a-9ede-4784-ab66-99e3908defd2-kube-api-access-6xtsq" (OuterVolumeSpecName: "kube-api-access-6xtsq") pod "d4a6371a-9ede-4784-ab66-99e3908defd2" (UID: "d4a6371a-9ede-4784-ab66-99e3908defd2"). InnerVolumeSpecName "kube-api-access-6xtsq". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 13 08:49:16 crc kubenswrapper[4971]: I1213 08:49:16.249806 4971 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/d4a6371a-9ede-4784-ab66-99e3908defd2-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "d4a6371a-9ede-4784-ab66-99e3908defd2" (UID: "d4a6371a-9ede-4784-ab66-99e3908defd2"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 13 08:49:16 crc kubenswrapper[4971]: I1213 08:49:16.292864 4971 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/d4a6371a-9ede-4784-ab66-99e3908defd2-utilities\") on node \"crc\" DevicePath \"\"" Dec 13 08:49:16 crc kubenswrapper[4971]: I1213 08:49:16.293206 4971 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-6xtsq\" (UniqueName: \"kubernetes.io/projected/d4a6371a-9ede-4784-ab66-99e3908defd2-kube-api-access-6xtsq\") on node \"crc\" DevicePath \"\"" Dec 13 08:49:16 crc kubenswrapper[4971]: I1213 08:49:16.293333 4971 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/d4a6371a-9ede-4784-ab66-99e3908defd2-catalog-content\") on node \"crc\" DevicePath \"\"" Dec 13 08:49:16 crc kubenswrapper[4971]: I1213 08:49:16.508488 4971 generic.go:334] "Generic (PLEG): container finished" podID="d4a6371a-9ede-4784-ab66-99e3908defd2" containerID="1ee0d99fcc93a549a2dae368555cccdd0e75eb42dcc04be8146bb742a9c94e9f" exitCode=0 Dec 13 08:49:16 crc kubenswrapper[4971]: I1213 08:49:16.508573 4971 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-nljkd" event={"ID":"d4a6371a-9ede-4784-ab66-99e3908defd2","Type":"ContainerDied","Data":"1ee0d99fcc93a549a2dae368555cccdd0e75eb42dcc04be8146bb742a9c94e9f"} Dec 13 08:49:16 crc kubenswrapper[4971]: I1213 08:49:16.508621 4971 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-nljkd" event={"ID":"d4a6371a-9ede-4784-ab66-99e3908defd2","Type":"ContainerDied","Data":"eba88caf9068b1976ed0d0cd30598f962d83d3d4f2a46e5b149ef2e3752789c0"} Dec 13 08:49:16 crc kubenswrapper[4971]: I1213 08:49:16.508618 4971 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-nljkd" Dec 13 08:49:16 crc kubenswrapper[4971]: I1213 08:49:16.508646 4971 scope.go:117] "RemoveContainer" containerID="1ee0d99fcc93a549a2dae368555cccdd0e75eb42dcc04be8146bb742a9c94e9f" Dec 13 08:49:16 crc kubenswrapper[4971]: I1213 08:49:16.549026 4971 scope.go:117] "RemoveContainer" containerID="f9c88220f3b23e03919489ce5b5fd5827a691e8b8645526e7efb99e5cd1da38e" Dec 13 08:49:16 crc kubenswrapper[4971]: I1213 08:49:16.562035 4971 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-nljkd"] Dec 13 08:49:16 crc kubenswrapper[4971]: I1213 08:49:16.580832 4971 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/certified-operators-nljkd"] Dec 13 08:49:16 crc kubenswrapper[4971]: I1213 08:49:16.661053 4971 scope.go:117] "RemoveContainer" containerID="63ab678671f2b267c76588b6549b097973f4637f9c088006ff8d4cd1fbb2681b" Dec 13 08:49:16 crc kubenswrapper[4971]: I1213 08:49:16.692642 4971 scope.go:117] "RemoveContainer" containerID="1ee0d99fcc93a549a2dae368555cccdd0e75eb42dcc04be8146bb742a9c94e9f" Dec 13 08:49:16 crc kubenswrapper[4971]: E1213 08:49:16.696873 4971 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"1ee0d99fcc93a549a2dae368555cccdd0e75eb42dcc04be8146bb742a9c94e9f\": container with ID starting with 1ee0d99fcc93a549a2dae368555cccdd0e75eb42dcc04be8146bb742a9c94e9f not found: ID does not exist" containerID="1ee0d99fcc93a549a2dae368555cccdd0e75eb42dcc04be8146bb742a9c94e9f" Dec 13 08:49:16 crc kubenswrapper[4971]: I1213 08:49:16.696930 4971 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"1ee0d99fcc93a549a2dae368555cccdd0e75eb42dcc04be8146bb742a9c94e9f"} err="failed to get container status \"1ee0d99fcc93a549a2dae368555cccdd0e75eb42dcc04be8146bb742a9c94e9f\": rpc error: code = NotFound desc = could not find container \"1ee0d99fcc93a549a2dae368555cccdd0e75eb42dcc04be8146bb742a9c94e9f\": container with ID starting with 1ee0d99fcc93a549a2dae368555cccdd0e75eb42dcc04be8146bb742a9c94e9f not found: ID does not exist" Dec 13 08:49:16 crc kubenswrapper[4971]: I1213 08:49:16.696964 4971 scope.go:117] "RemoveContainer" containerID="f9c88220f3b23e03919489ce5b5fd5827a691e8b8645526e7efb99e5cd1da38e" Dec 13 08:49:16 crc kubenswrapper[4971]: E1213 08:49:16.704852 4971 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"f9c88220f3b23e03919489ce5b5fd5827a691e8b8645526e7efb99e5cd1da38e\": container with ID starting with f9c88220f3b23e03919489ce5b5fd5827a691e8b8645526e7efb99e5cd1da38e not found: ID does not exist" containerID="f9c88220f3b23e03919489ce5b5fd5827a691e8b8645526e7efb99e5cd1da38e" Dec 13 08:49:16 crc kubenswrapper[4971]: I1213 08:49:16.704927 4971 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"f9c88220f3b23e03919489ce5b5fd5827a691e8b8645526e7efb99e5cd1da38e"} err="failed to get container status \"f9c88220f3b23e03919489ce5b5fd5827a691e8b8645526e7efb99e5cd1da38e\": rpc error: code = NotFound desc = could not find container \"f9c88220f3b23e03919489ce5b5fd5827a691e8b8645526e7efb99e5cd1da38e\": container with ID starting with f9c88220f3b23e03919489ce5b5fd5827a691e8b8645526e7efb99e5cd1da38e not found: ID does not exist" Dec 13 08:49:16 crc kubenswrapper[4971]: I1213 08:49:16.704978 4971 scope.go:117] "RemoveContainer" containerID="63ab678671f2b267c76588b6549b097973f4637f9c088006ff8d4cd1fbb2681b" Dec 13 08:49:16 crc kubenswrapper[4971]: E1213 08:49:16.708354 4971 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"63ab678671f2b267c76588b6549b097973f4637f9c088006ff8d4cd1fbb2681b\": container with ID starting with 63ab678671f2b267c76588b6549b097973f4637f9c088006ff8d4cd1fbb2681b not found: ID does not exist" containerID="63ab678671f2b267c76588b6549b097973f4637f9c088006ff8d4cd1fbb2681b" Dec 13 08:49:16 crc kubenswrapper[4971]: I1213 08:49:16.708421 4971 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"63ab678671f2b267c76588b6549b097973f4637f9c088006ff8d4cd1fbb2681b"} err="failed to get container status \"63ab678671f2b267c76588b6549b097973f4637f9c088006ff8d4cd1fbb2681b\": rpc error: code = NotFound desc = could not find container \"63ab678671f2b267c76588b6549b097973f4637f9c088006ff8d4cd1fbb2681b\": container with ID starting with 63ab678671f2b267c76588b6549b097973f4637f9c088006ff8d4cd1fbb2681b not found: ID does not exist" Dec 13 08:49:16 crc kubenswrapper[4971]: I1213 08:49:16.728058 4971 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-9d94b"] Dec 13 08:49:16 crc kubenswrapper[4971]: I1213 08:49:16.728372 4971 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/community-operators-9d94b" podUID="76ade42c-67ee-45b2-9b31-62002698a753" containerName="registry-server" containerID="cri-o://6c07057c003fbf99efd497ab319fa3cc558337216cdcb6287d48db74213f3d13" gracePeriod=2 Dec 13 08:49:17 crc kubenswrapper[4971]: I1213 08:49:17.366228 4971 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-9d94b" Dec 13 08:49:17 crc kubenswrapper[4971]: I1213 08:49:17.473464 4971 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/76ade42c-67ee-45b2-9b31-62002698a753-utilities\") pod \"76ade42c-67ee-45b2-9b31-62002698a753\" (UID: \"76ade42c-67ee-45b2-9b31-62002698a753\") " Dec 13 08:49:17 crc kubenswrapper[4971]: I1213 08:49:17.474101 4971 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-br82l\" (UniqueName: \"kubernetes.io/projected/76ade42c-67ee-45b2-9b31-62002698a753-kube-api-access-br82l\") pod \"76ade42c-67ee-45b2-9b31-62002698a753\" (UID: \"76ade42c-67ee-45b2-9b31-62002698a753\") " Dec 13 08:49:17 crc kubenswrapper[4971]: I1213 08:49:17.474341 4971 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/76ade42c-67ee-45b2-9b31-62002698a753-catalog-content\") pod \"76ade42c-67ee-45b2-9b31-62002698a753\" (UID: \"76ade42c-67ee-45b2-9b31-62002698a753\") " Dec 13 08:49:17 crc kubenswrapper[4971]: I1213 08:49:17.474943 4971 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/76ade42c-67ee-45b2-9b31-62002698a753-utilities" (OuterVolumeSpecName: "utilities") pod "76ade42c-67ee-45b2-9b31-62002698a753" (UID: "76ade42c-67ee-45b2-9b31-62002698a753"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 13 08:49:17 crc kubenswrapper[4971]: I1213 08:49:17.475846 4971 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/76ade42c-67ee-45b2-9b31-62002698a753-utilities\") on node \"crc\" DevicePath \"\"" Dec 13 08:49:17 crc kubenswrapper[4971]: I1213 08:49:17.482106 4971 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/76ade42c-67ee-45b2-9b31-62002698a753-kube-api-access-br82l" (OuterVolumeSpecName: "kube-api-access-br82l") pod "76ade42c-67ee-45b2-9b31-62002698a753" (UID: "76ade42c-67ee-45b2-9b31-62002698a753"). InnerVolumeSpecName "kube-api-access-br82l". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 13 08:49:17 crc kubenswrapper[4971]: I1213 08:49:17.528483 4971 generic.go:334] "Generic (PLEG): container finished" podID="76ade42c-67ee-45b2-9b31-62002698a753" containerID="6c07057c003fbf99efd497ab319fa3cc558337216cdcb6287d48db74213f3d13" exitCode=0 Dec 13 08:49:17 crc kubenswrapper[4971]: I1213 08:49:17.530768 4971 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-9d94b" event={"ID":"76ade42c-67ee-45b2-9b31-62002698a753","Type":"ContainerDied","Data":"6c07057c003fbf99efd497ab319fa3cc558337216cdcb6287d48db74213f3d13"} Dec 13 08:49:17 crc kubenswrapper[4971]: I1213 08:49:17.533075 4971 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-9d94b" event={"ID":"76ade42c-67ee-45b2-9b31-62002698a753","Type":"ContainerDied","Data":"4902985fb7b5c038c08ffaf044c7b5bde772daee626da519af65cbf68e090d61"} Dec 13 08:49:17 crc kubenswrapper[4971]: I1213 08:49:17.533281 4971 scope.go:117] "RemoveContainer" containerID="6c07057c003fbf99efd497ab319fa3cc558337216cdcb6287d48db74213f3d13" Dec 13 08:49:17 crc kubenswrapper[4971]: I1213 08:49:17.531567 4971 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-9d94b" Dec 13 08:49:17 crc kubenswrapper[4971]: I1213 08:49:17.547903 4971 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/76ade42c-67ee-45b2-9b31-62002698a753-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "76ade42c-67ee-45b2-9b31-62002698a753" (UID: "76ade42c-67ee-45b2-9b31-62002698a753"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 13 08:49:17 crc kubenswrapper[4971]: I1213 08:49:17.568783 4971 scope.go:117] "RemoveContainer" containerID="d0297777d8c1e6e8ec1156e50c3a545c6243654391295cdc41ef7daa7f5f2db8" Dec 13 08:49:17 crc kubenswrapper[4971]: I1213 08:49:17.579548 4971 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-br82l\" (UniqueName: \"kubernetes.io/projected/76ade42c-67ee-45b2-9b31-62002698a753-kube-api-access-br82l\") on node \"crc\" DevicePath \"\"" Dec 13 08:49:17 crc kubenswrapper[4971]: I1213 08:49:17.579629 4971 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/76ade42c-67ee-45b2-9b31-62002698a753-catalog-content\") on node \"crc\" DevicePath \"\"" Dec 13 08:49:17 crc kubenswrapper[4971]: I1213 08:49:17.597480 4971 scope.go:117] "RemoveContainer" containerID="745e0a204fb6e0429a43f609a4054aa36cf6ec651600a27da9971129b1406d73" Dec 13 08:49:17 crc kubenswrapper[4971]: I1213 08:49:17.623176 4971 scope.go:117] "RemoveContainer" containerID="6c07057c003fbf99efd497ab319fa3cc558337216cdcb6287d48db74213f3d13" Dec 13 08:49:17 crc kubenswrapper[4971]: E1213 08:49:17.623984 4971 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"6c07057c003fbf99efd497ab319fa3cc558337216cdcb6287d48db74213f3d13\": container with ID starting with 6c07057c003fbf99efd497ab319fa3cc558337216cdcb6287d48db74213f3d13 not found: ID does not exist" containerID="6c07057c003fbf99efd497ab319fa3cc558337216cdcb6287d48db74213f3d13" Dec 13 08:49:17 crc kubenswrapper[4971]: I1213 08:49:17.624077 4971 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"6c07057c003fbf99efd497ab319fa3cc558337216cdcb6287d48db74213f3d13"} err="failed to get container status \"6c07057c003fbf99efd497ab319fa3cc558337216cdcb6287d48db74213f3d13\": rpc error: code = NotFound desc = could not find container \"6c07057c003fbf99efd497ab319fa3cc558337216cdcb6287d48db74213f3d13\": container with ID starting with 6c07057c003fbf99efd497ab319fa3cc558337216cdcb6287d48db74213f3d13 not found: ID does not exist" Dec 13 08:49:17 crc kubenswrapper[4971]: I1213 08:49:17.624124 4971 scope.go:117] "RemoveContainer" containerID="d0297777d8c1e6e8ec1156e50c3a545c6243654391295cdc41ef7daa7f5f2db8" Dec 13 08:49:17 crc kubenswrapper[4971]: E1213 08:49:17.624648 4971 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"d0297777d8c1e6e8ec1156e50c3a545c6243654391295cdc41ef7daa7f5f2db8\": container with ID starting with d0297777d8c1e6e8ec1156e50c3a545c6243654391295cdc41ef7daa7f5f2db8 not found: ID does not exist" containerID="d0297777d8c1e6e8ec1156e50c3a545c6243654391295cdc41ef7daa7f5f2db8" Dec 13 08:49:17 crc kubenswrapper[4971]: I1213 08:49:17.624725 4971 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"d0297777d8c1e6e8ec1156e50c3a545c6243654391295cdc41ef7daa7f5f2db8"} err="failed to get container status \"d0297777d8c1e6e8ec1156e50c3a545c6243654391295cdc41ef7daa7f5f2db8\": rpc error: code = NotFound desc = could not find container \"d0297777d8c1e6e8ec1156e50c3a545c6243654391295cdc41ef7daa7f5f2db8\": container with ID starting with d0297777d8c1e6e8ec1156e50c3a545c6243654391295cdc41ef7daa7f5f2db8 not found: ID does not exist" Dec 13 08:49:17 crc kubenswrapper[4971]: I1213 08:49:17.624779 4971 scope.go:117] "RemoveContainer" containerID="745e0a204fb6e0429a43f609a4054aa36cf6ec651600a27da9971129b1406d73" Dec 13 08:49:17 crc kubenswrapper[4971]: E1213 08:49:17.625184 4971 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"745e0a204fb6e0429a43f609a4054aa36cf6ec651600a27da9971129b1406d73\": container with ID starting with 745e0a204fb6e0429a43f609a4054aa36cf6ec651600a27da9971129b1406d73 not found: ID does not exist" containerID="745e0a204fb6e0429a43f609a4054aa36cf6ec651600a27da9971129b1406d73" Dec 13 08:49:17 crc kubenswrapper[4971]: I1213 08:49:17.625219 4971 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"745e0a204fb6e0429a43f609a4054aa36cf6ec651600a27da9971129b1406d73"} err="failed to get container status \"745e0a204fb6e0429a43f609a4054aa36cf6ec651600a27da9971129b1406d73\": rpc error: code = NotFound desc = could not find container \"745e0a204fb6e0429a43f609a4054aa36cf6ec651600a27da9971129b1406d73\": container with ID starting with 745e0a204fb6e0429a43f609a4054aa36cf6ec651600a27da9971129b1406d73 not found: ID does not exist" Dec 13 08:49:17 crc kubenswrapper[4971]: I1213 08:49:17.782215 4971 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="d4a6371a-9ede-4784-ab66-99e3908defd2" path="/var/lib/kubelet/pods/d4a6371a-9ede-4784-ab66-99e3908defd2/volumes" Dec 13 08:49:17 crc kubenswrapper[4971]: I1213 08:49:17.878501 4971 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-9d94b"] Dec 13 08:49:17 crc kubenswrapper[4971]: I1213 08:49:17.897812 4971 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/community-operators-9d94b"] Dec 13 08:49:19 crc kubenswrapper[4971]: I1213 08:49:19.784549 4971 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="76ade42c-67ee-45b2-9b31-62002698a753" path="/var/lib/kubelet/pods/76ade42c-67ee-45b2-9b31-62002698a753/volumes" Dec 13 08:49:22 crc kubenswrapper[4971]: I1213 08:49:22.770178 4971 scope.go:117] "RemoveContainer" containerID="b875432558a41dd98c673e3df9c78c3a2bcd69eb63408b435826e3d82347d2a6" Dec 13 08:49:22 crc kubenswrapper[4971]: E1213 08:49:22.771099 4971 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-82xjz_openshift-machine-config-operator(e0c4af12-edda-41d7-9b44-f87396174ef2)\"" pod="openshift-machine-config-operator/machine-config-daemon-82xjz" podUID="e0c4af12-edda-41d7-9b44-f87396174ef2" Dec 13 08:49:35 crc kubenswrapper[4971]: I1213 08:49:35.768630 4971 scope.go:117] "RemoveContainer" containerID="b875432558a41dd98c673e3df9c78c3a2bcd69eb63408b435826e3d82347d2a6" Dec 13 08:49:35 crc kubenswrapper[4971]: E1213 08:49:35.769919 4971 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-82xjz_openshift-machine-config-operator(e0c4af12-edda-41d7-9b44-f87396174ef2)\"" pod="openshift-machine-config-operator/machine-config-daemon-82xjz" podUID="e0c4af12-edda-41d7-9b44-f87396174ef2" Dec 13 08:49:46 crc kubenswrapper[4971]: I1213 08:49:46.769450 4971 scope.go:117] "RemoveContainer" containerID="b875432558a41dd98c673e3df9c78c3a2bcd69eb63408b435826e3d82347d2a6" Dec 13 08:49:46 crc kubenswrapper[4971]: E1213 08:49:46.772491 4971 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-82xjz_openshift-machine-config-operator(e0c4af12-edda-41d7-9b44-f87396174ef2)\"" pod="openshift-machine-config-operator/machine-config-daemon-82xjz" podUID="e0c4af12-edda-41d7-9b44-f87396174ef2" Dec 13 08:50:00 crc kubenswrapper[4971]: I1213 08:50:00.769188 4971 scope.go:117] "RemoveContainer" containerID="b875432558a41dd98c673e3df9c78c3a2bcd69eb63408b435826e3d82347d2a6" Dec 13 08:50:00 crc kubenswrapper[4971]: E1213 08:50:00.770588 4971 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-82xjz_openshift-machine-config-operator(e0c4af12-edda-41d7-9b44-f87396174ef2)\"" pod="openshift-machine-config-operator/machine-config-daemon-82xjz" podUID="e0c4af12-edda-41d7-9b44-f87396174ef2" Dec 13 08:50:13 crc kubenswrapper[4971]: I1213 08:50:13.779004 4971 scope.go:117] "RemoveContainer" containerID="b875432558a41dd98c673e3df9c78c3a2bcd69eb63408b435826e3d82347d2a6" Dec 13 08:50:13 crc kubenswrapper[4971]: E1213 08:50:13.785194 4971 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-82xjz_openshift-machine-config-operator(e0c4af12-edda-41d7-9b44-f87396174ef2)\"" pod="openshift-machine-config-operator/machine-config-daemon-82xjz" podUID="e0c4af12-edda-41d7-9b44-f87396174ef2" Dec 13 08:50:25 crc kubenswrapper[4971]: I1213 08:50:25.771444 4971 scope.go:117] "RemoveContainer" containerID="b875432558a41dd98c673e3df9c78c3a2bcd69eb63408b435826e3d82347d2a6" Dec 13 08:50:25 crc kubenswrapper[4971]: E1213 08:50:25.772494 4971 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-82xjz_openshift-machine-config-operator(e0c4af12-edda-41d7-9b44-f87396174ef2)\"" pod="openshift-machine-config-operator/machine-config-daemon-82xjz" podUID="e0c4af12-edda-41d7-9b44-f87396174ef2" Dec 13 08:50:37 crc kubenswrapper[4971]: I1213 08:50:37.770253 4971 scope.go:117] "RemoveContainer" containerID="b875432558a41dd98c673e3df9c78c3a2bcd69eb63408b435826e3d82347d2a6" Dec 13 08:50:37 crc kubenswrapper[4971]: E1213 08:50:37.774068 4971 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-82xjz_openshift-machine-config-operator(e0c4af12-edda-41d7-9b44-f87396174ef2)\"" pod="openshift-machine-config-operator/machine-config-daemon-82xjz" podUID="e0c4af12-edda-41d7-9b44-f87396174ef2" Dec 13 08:50:49 crc kubenswrapper[4971]: I1213 08:50:49.769552 4971 scope.go:117] "RemoveContainer" containerID="b875432558a41dd98c673e3df9c78c3a2bcd69eb63408b435826e3d82347d2a6" Dec 13 08:50:49 crc kubenswrapper[4971]: E1213 08:50:49.771054 4971 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-82xjz_openshift-machine-config-operator(e0c4af12-edda-41d7-9b44-f87396174ef2)\"" pod="openshift-machine-config-operator/machine-config-daemon-82xjz" podUID="e0c4af12-edda-41d7-9b44-f87396174ef2" Dec 13 08:51:01 crc kubenswrapper[4971]: I1213 08:51:01.770079 4971 scope.go:117] "RemoveContainer" containerID="b875432558a41dd98c673e3df9c78c3a2bcd69eb63408b435826e3d82347d2a6" Dec 13 08:51:01 crc kubenswrapper[4971]: E1213 08:51:01.771336 4971 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-82xjz_openshift-machine-config-operator(e0c4af12-edda-41d7-9b44-f87396174ef2)\"" pod="openshift-machine-config-operator/machine-config-daemon-82xjz" podUID="e0c4af12-edda-41d7-9b44-f87396174ef2" Dec 13 08:51:12 crc kubenswrapper[4971]: I1213 08:51:12.769576 4971 scope.go:117] "RemoveContainer" containerID="b875432558a41dd98c673e3df9c78c3a2bcd69eb63408b435826e3d82347d2a6" Dec 13 08:51:12 crc kubenswrapper[4971]: E1213 08:51:12.770873 4971 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-82xjz_openshift-machine-config-operator(e0c4af12-edda-41d7-9b44-f87396174ef2)\"" pod="openshift-machine-config-operator/machine-config-daemon-82xjz" podUID="e0c4af12-edda-41d7-9b44-f87396174ef2" Dec 13 08:51:25 crc kubenswrapper[4971]: I1213 08:51:25.769032 4971 scope.go:117] "RemoveContainer" containerID="b875432558a41dd98c673e3df9c78c3a2bcd69eb63408b435826e3d82347d2a6" Dec 13 08:51:25 crc kubenswrapper[4971]: E1213 08:51:25.770134 4971 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-82xjz_openshift-machine-config-operator(e0c4af12-edda-41d7-9b44-f87396174ef2)\"" pod="openshift-machine-config-operator/machine-config-daemon-82xjz" podUID="e0c4af12-edda-41d7-9b44-f87396174ef2" Dec 13 08:51:38 crc kubenswrapper[4971]: I1213 08:51:38.770277 4971 scope.go:117] "RemoveContainer" containerID="b875432558a41dd98c673e3df9c78c3a2bcd69eb63408b435826e3d82347d2a6" Dec 13 08:51:38 crc kubenswrapper[4971]: E1213 08:51:38.773669 4971 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-82xjz_openshift-machine-config-operator(e0c4af12-edda-41d7-9b44-f87396174ef2)\"" pod="openshift-machine-config-operator/machine-config-daemon-82xjz" podUID="e0c4af12-edda-41d7-9b44-f87396174ef2" Dec 13 08:51:50 crc kubenswrapper[4971]: I1213 08:51:50.768630 4971 scope.go:117] "RemoveContainer" containerID="b875432558a41dd98c673e3df9c78c3a2bcd69eb63408b435826e3d82347d2a6" Dec 13 08:51:50 crc kubenswrapper[4971]: E1213 08:51:50.769745 4971 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-82xjz_openshift-machine-config-operator(e0c4af12-edda-41d7-9b44-f87396174ef2)\"" pod="openshift-machine-config-operator/machine-config-daemon-82xjz" podUID="e0c4af12-edda-41d7-9b44-f87396174ef2" Dec 13 08:52:01 crc kubenswrapper[4971]: I1213 08:52:01.770513 4971 scope.go:117] "RemoveContainer" containerID="b875432558a41dd98c673e3df9c78c3a2bcd69eb63408b435826e3d82347d2a6" Dec 13 08:52:01 crc kubenswrapper[4971]: E1213 08:52:01.773463 4971 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-82xjz_openshift-machine-config-operator(e0c4af12-edda-41d7-9b44-f87396174ef2)\"" pod="openshift-machine-config-operator/machine-config-daemon-82xjz" podUID="e0c4af12-edda-41d7-9b44-f87396174ef2" Dec 13 08:52:15 crc kubenswrapper[4971]: I1213 08:52:15.769018 4971 scope.go:117] "RemoveContainer" containerID="b875432558a41dd98c673e3df9c78c3a2bcd69eb63408b435826e3d82347d2a6" Dec 13 08:52:15 crc kubenswrapper[4971]: E1213 08:52:15.770043 4971 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-82xjz_openshift-machine-config-operator(e0c4af12-edda-41d7-9b44-f87396174ef2)\"" pod="openshift-machine-config-operator/machine-config-daemon-82xjz" podUID="e0c4af12-edda-41d7-9b44-f87396174ef2" Dec 13 08:52:30 crc kubenswrapper[4971]: I1213 08:52:30.769621 4971 scope.go:117] "RemoveContainer" containerID="b875432558a41dd98c673e3df9c78c3a2bcd69eb63408b435826e3d82347d2a6" Dec 13 08:52:30 crc kubenswrapper[4971]: E1213 08:52:30.770930 4971 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-82xjz_openshift-machine-config-operator(e0c4af12-edda-41d7-9b44-f87396174ef2)\"" pod="openshift-machine-config-operator/machine-config-daemon-82xjz" podUID="e0c4af12-edda-41d7-9b44-f87396174ef2" Dec 13 08:52:45 crc kubenswrapper[4971]: I1213 08:52:45.770983 4971 scope.go:117] "RemoveContainer" containerID="b875432558a41dd98c673e3df9c78c3a2bcd69eb63408b435826e3d82347d2a6" Dec 13 08:52:45 crc kubenswrapper[4971]: E1213 08:52:45.771830 4971 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-82xjz_openshift-machine-config-operator(e0c4af12-edda-41d7-9b44-f87396174ef2)\"" pod="openshift-machine-config-operator/machine-config-daemon-82xjz" podUID="e0c4af12-edda-41d7-9b44-f87396174ef2" Dec 13 08:52:56 crc kubenswrapper[4971]: I1213 08:52:56.769627 4971 scope.go:117] "RemoveContainer" containerID="b875432558a41dd98c673e3df9c78c3a2bcd69eb63408b435826e3d82347d2a6" Dec 13 08:52:56 crc kubenswrapper[4971]: E1213 08:52:56.771078 4971 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-82xjz_openshift-machine-config-operator(e0c4af12-edda-41d7-9b44-f87396174ef2)\"" pod="openshift-machine-config-operator/machine-config-daemon-82xjz" podUID="e0c4af12-edda-41d7-9b44-f87396174ef2" Dec 13 08:53:07 crc kubenswrapper[4971]: I1213 08:53:07.768894 4971 scope.go:117] "RemoveContainer" containerID="b875432558a41dd98c673e3df9c78c3a2bcd69eb63408b435826e3d82347d2a6" Dec 13 08:53:07 crc kubenswrapper[4971]: E1213 08:53:07.770229 4971 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-82xjz_openshift-machine-config-operator(e0c4af12-edda-41d7-9b44-f87396174ef2)\"" pod="openshift-machine-config-operator/machine-config-daemon-82xjz" podUID="e0c4af12-edda-41d7-9b44-f87396174ef2" Dec 13 08:53:22 crc kubenswrapper[4971]: I1213 08:53:22.769707 4971 scope.go:117] "RemoveContainer" containerID="b875432558a41dd98c673e3df9c78c3a2bcd69eb63408b435826e3d82347d2a6" Dec 13 08:53:23 crc kubenswrapper[4971]: I1213 08:53:23.572830 4971 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-82xjz" event={"ID":"e0c4af12-edda-41d7-9b44-f87396174ef2","Type":"ContainerStarted","Data":"3f281a9adda16a9b53c490949e166088ac22b2347e901a634eebb681b7729981"} Dec 13 08:54:26 crc kubenswrapper[4971]: I1213 08:54:26.833766 4971 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-marketplace-rvv8f"] Dec 13 08:54:26 crc kubenswrapper[4971]: E1213 08:54:26.834947 4971 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="76ade42c-67ee-45b2-9b31-62002698a753" containerName="registry-server" Dec 13 08:54:26 crc kubenswrapper[4971]: I1213 08:54:26.834968 4971 state_mem.go:107] "Deleted CPUSet assignment" podUID="76ade42c-67ee-45b2-9b31-62002698a753" containerName="registry-server" Dec 13 08:54:26 crc kubenswrapper[4971]: E1213 08:54:26.834996 4971 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="76ade42c-67ee-45b2-9b31-62002698a753" containerName="extract-utilities" Dec 13 08:54:26 crc kubenswrapper[4971]: I1213 08:54:26.835007 4971 state_mem.go:107] "Deleted CPUSet assignment" podUID="76ade42c-67ee-45b2-9b31-62002698a753" containerName="extract-utilities" Dec 13 08:54:26 crc kubenswrapper[4971]: E1213 08:54:26.835018 4971 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="d4a6371a-9ede-4784-ab66-99e3908defd2" containerName="registry-server" Dec 13 08:54:26 crc kubenswrapper[4971]: I1213 08:54:26.835028 4971 state_mem.go:107] "Deleted CPUSet assignment" podUID="d4a6371a-9ede-4784-ab66-99e3908defd2" containerName="registry-server" Dec 13 08:54:26 crc kubenswrapper[4971]: E1213 08:54:26.835040 4971 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="76ade42c-67ee-45b2-9b31-62002698a753" containerName="extract-content" Dec 13 08:54:26 crc kubenswrapper[4971]: I1213 08:54:26.835048 4971 state_mem.go:107] "Deleted CPUSet assignment" podUID="76ade42c-67ee-45b2-9b31-62002698a753" containerName="extract-content" Dec 13 08:54:26 crc kubenswrapper[4971]: E1213 08:54:26.835066 4971 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="d4a6371a-9ede-4784-ab66-99e3908defd2" containerName="extract-content" Dec 13 08:54:26 crc kubenswrapper[4971]: I1213 08:54:26.835074 4971 state_mem.go:107] "Deleted CPUSet assignment" podUID="d4a6371a-9ede-4784-ab66-99e3908defd2" containerName="extract-content" Dec 13 08:54:26 crc kubenswrapper[4971]: E1213 08:54:26.835087 4971 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="d4a6371a-9ede-4784-ab66-99e3908defd2" containerName="extract-utilities" Dec 13 08:54:26 crc kubenswrapper[4971]: I1213 08:54:26.835095 4971 state_mem.go:107] "Deleted CPUSet assignment" podUID="d4a6371a-9ede-4784-ab66-99e3908defd2" containerName="extract-utilities" Dec 13 08:54:26 crc kubenswrapper[4971]: I1213 08:54:26.835364 4971 memory_manager.go:354] "RemoveStaleState removing state" podUID="d4a6371a-9ede-4784-ab66-99e3908defd2" containerName="registry-server" Dec 13 08:54:26 crc kubenswrapper[4971]: I1213 08:54:26.835399 4971 memory_manager.go:354] "RemoveStaleState removing state" podUID="76ade42c-67ee-45b2-9b31-62002698a753" containerName="registry-server" Dec 13 08:54:26 crc kubenswrapper[4971]: I1213 08:54:26.837323 4971 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-rvv8f" Dec 13 08:54:26 crc kubenswrapper[4971]: I1213 08:54:26.848573 4971 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-rvv8f"] Dec 13 08:54:26 crc kubenswrapper[4971]: I1213 08:54:26.975476 4971 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/dda9fe48-7439-4f74-bd2a-7733a4697c13-utilities\") pod \"redhat-marketplace-rvv8f\" (UID: \"dda9fe48-7439-4f74-bd2a-7733a4697c13\") " pod="openshift-marketplace/redhat-marketplace-rvv8f" Dec 13 08:54:26 crc kubenswrapper[4971]: I1213 08:54:26.975702 4971 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/dda9fe48-7439-4f74-bd2a-7733a4697c13-catalog-content\") pod \"redhat-marketplace-rvv8f\" (UID: \"dda9fe48-7439-4f74-bd2a-7733a4697c13\") " pod="openshift-marketplace/redhat-marketplace-rvv8f" Dec 13 08:54:26 crc kubenswrapper[4971]: I1213 08:54:26.976375 4971 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-nnl98\" (UniqueName: \"kubernetes.io/projected/dda9fe48-7439-4f74-bd2a-7733a4697c13-kube-api-access-nnl98\") pod \"redhat-marketplace-rvv8f\" (UID: \"dda9fe48-7439-4f74-bd2a-7733a4697c13\") " pod="openshift-marketplace/redhat-marketplace-rvv8f" Dec 13 08:54:27 crc kubenswrapper[4971]: I1213 08:54:27.078656 4971 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-nnl98\" (UniqueName: \"kubernetes.io/projected/dda9fe48-7439-4f74-bd2a-7733a4697c13-kube-api-access-nnl98\") pod \"redhat-marketplace-rvv8f\" (UID: \"dda9fe48-7439-4f74-bd2a-7733a4697c13\") " pod="openshift-marketplace/redhat-marketplace-rvv8f" Dec 13 08:54:27 crc kubenswrapper[4971]: I1213 08:54:27.079076 4971 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/dda9fe48-7439-4f74-bd2a-7733a4697c13-utilities\") pod \"redhat-marketplace-rvv8f\" (UID: \"dda9fe48-7439-4f74-bd2a-7733a4697c13\") " pod="openshift-marketplace/redhat-marketplace-rvv8f" Dec 13 08:54:27 crc kubenswrapper[4971]: I1213 08:54:27.079201 4971 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/dda9fe48-7439-4f74-bd2a-7733a4697c13-catalog-content\") pod \"redhat-marketplace-rvv8f\" (UID: \"dda9fe48-7439-4f74-bd2a-7733a4697c13\") " pod="openshift-marketplace/redhat-marketplace-rvv8f" Dec 13 08:54:27 crc kubenswrapper[4971]: I1213 08:54:27.079978 4971 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/dda9fe48-7439-4f74-bd2a-7733a4697c13-utilities\") pod \"redhat-marketplace-rvv8f\" (UID: \"dda9fe48-7439-4f74-bd2a-7733a4697c13\") " pod="openshift-marketplace/redhat-marketplace-rvv8f" Dec 13 08:54:27 crc kubenswrapper[4971]: I1213 08:54:27.080005 4971 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/dda9fe48-7439-4f74-bd2a-7733a4697c13-catalog-content\") pod \"redhat-marketplace-rvv8f\" (UID: \"dda9fe48-7439-4f74-bd2a-7733a4697c13\") " pod="openshift-marketplace/redhat-marketplace-rvv8f" Dec 13 08:54:27 crc kubenswrapper[4971]: I1213 08:54:27.108939 4971 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-nnl98\" (UniqueName: \"kubernetes.io/projected/dda9fe48-7439-4f74-bd2a-7733a4697c13-kube-api-access-nnl98\") pod \"redhat-marketplace-rvv8f\" (UID: \"dda9fe48-7439-4f74-bd2a-7733a4697c13\") " pod="openshift-marketplace/redhat-marketplace-rvv8f" Dec 13 08:54:27 crc kubenswrapper[4971]: I1213 08:54:27.168970 4971 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-rvv8f" Dec 13 08:54:27 crc kubenswrapper[4971]: I1213 08:54:27.735199 4971 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-rvv8f"] Dec 13 08:54:28 crc kubenswrapper[4971]: I1213 08:54:28.299050 4971 generic.go:334] "Generic (PLEG): container finished" podID="dda9fe48-7439-4f74-bd2a-7733a4697c13" containerID="e92dd69eacd8e2a969e769f412345cd8cf13bb597ee235f6d07a1dff61186a38" exitCode=0 Dec 13 08:54:28 crc kubenswrapper[4971]: I1213 08:54:28.299875 4971 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-rvv8f" event={"ID":"dda9fe48-7439-4f74-bd2a-7733a4697c13","Type":"ContainerDied","Data":"e92dd69eacd8e2a969e769f412345cd8cf13bb597ee235f6d07a1dff61186a38"} Dec 13 08:54:28 crc kubenswrapper[4971]: I1213 08:54:28.299946 4971 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-rvv8f" event={"ID":"dda9fe48-7439-4f74-bd2a-7733a4697c13","Type":"ContainerStarted","Data":"2f75d1f92ed6fd222b29a53ac97720695766c4c34254ce060c73b93674ca63eb"} Dec 13 08:54:28 crc kubenswrapper[4971]: I1213 08:54:28.301342 4971 provider.go:102] Refreshing cache for provider: *credentialprovider.defaultDockerConfigProvider Dec 13 08:54:30 crc kubenswrapper[4971]: I1213 08:54:30.323745 4971 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-rvv8f" event={"ID":"dda9fe48-7439-4f74-bd2a-7733a4697c13","Type":"ContainerStarted","Data":"d78a25965a2bf65fbfe1c349506cf8f9006f48da37ccd1e3bd62f51cad684d11"} Dec 13 08:54:31 crc kubenswrapper[4971]: I1213 08:54:31.339134 4971 generic.go:334] "Generic (PLEG): container finished" podID="dda9fe48-7439-4f74-bd2a-7733a4697c13" containerID="d78a25965a2bf65fbfe1c349506cf8f9006f48da37ccd1e3bd62f51cad684d11" exitCode=0 Dec 13 08:54:31 crc kubenswrapper[4971]: I1213 08:54:31.339237 4971 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-rvv8f" event={"ID":"dda9fe48-7439-4f74-bd2a-7733a4697c13","Type":"ContainerDied","Data":"d78a25965a2bf65fbfe1c349506cf8f9006f48da37ccd1e3bd62f51cad684d11"} Dec 13 08:54:33 crc kubenswrapper[4971]: I1213 08:54:33.363244 4971 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-rvv8f" event={"ID":"dda9fe48-7439-4f74-bd2a-7733a4697c13","Type":"ContainerStarted","Data":"efb01dd66a626d4fbfdbd45508fda55e2c6ea432f6a3f70927f7a9ce03957d94"} Dec 13 08:54:33 crc kubenswrapper[4971]: I1213 08:54:33.385185 4971 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-marketplace-rvv8f" podStartSLOduration=3.588297667 podStartE2EDuration="7.38516236s" podCreationTimestamp="2025-12-13 08:54:26 +0000 UTC" firstStartedPulling="2025-12-13 08:54:28.300998291 +0000 UTC m=+7524.905407729" lastFinishedPulling="2025-12-13 08:54:32.097862974 +0000 UTC m=+7528.702272422" observedRunningTime="2025-12-13 08:54:33.383543629 +0000 UTC m=+7529.987953077" watchObservedRunningTime="2025-12-13 08:54:33.38516236 +0000 UTC m=+7529.989571808" Dec 13 08:54:37 crc kubenswrapper[4971]: I1213 08:54:37.169417 4971 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-marketplace-rvv8f" Dec 13 08:54:37 crc kubenswrapper[4971]: I1213 08:54:37.170483 4971 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-marketplace-rvv8f" Dec 13 08:54:37 crc kubenswrapper[4971]: I1213 08:54:37.228324 4971 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-marketplace-rvv8f" Dec 13 08:54:37 crc kubenswrapper[4971]: I1213 08:54:37.461359 4971 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-marketplace-rvv8f" Dec 13 08:54:37 crc kubenswrapper[4971]: I1213 08:54:37.535357 4971 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-rvv8f"] Dec 13 08:54:39 crc kubenswrapper[4971]: I1213 08:54:39.423577 4971 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-marketplace-rvv8f" podUID="dda9fe48-7439-4f74-bd2a-7733a4697c13" containerName="registry-server" containerID="cri-o://efb01dd66a626d4fbfdbd45508fda55e2c6ea432f6a3f70927f7a9ce03957d94" gracePeriod=2 Dec 13 08:54:40 crc kubenswrapper[4971]: I1213 08:54:40.309059 4971 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-rvv8f" Dec 13 08:54:40 crc kubenswrapper[4971]: I1213 08:54:40.433500 4971 generic.go:334] "Generic (PLEG): container finished" podID="dda9fe48-7439-4f74-bd2a-7733a4697c13" containerID="efb01dd66a626d4fbfdbd45508fda55e2c6ea432f6a3f70927f7a9ce03957d94" exitCode=0 Dec 13 08:54:40 crc kubenswrapper[4971]: I1213 08:54:40.433615 4971 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-rvv8f" event={"ID":"dda9fe48-7439-4f74-bd2a-7733a4697c13","Type":"ContainerDied","Data":"efb01dd66a626d4fbfdbd45508fda55e2c6ea432f6a3f70927f7a9ce03957d94"} Dec 13 08:54:40 crc kubenswrapper[4971]: I1213 08:54:40.433649 4971 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-rvv8f" event={"ID":"dda9fe48-7439-4f74-bd2a-7733a4697c13","Type":"ContainerDied","Data":"2f75d1f92ed6fd222b29a53ac97720695766c4c34254ce060c73b93674ca63eb"} Dec 13 08:54:40 crc kubenswrapper[4971]: I1213 08:54:40.433857 4971 scope.go:117] "RemoveContainer" containerID="efb01dd66a626d4fbfdbd45508fda55e2c6ea432f6a3f70927f7a9ce03957d94" Dec 13 08:54:40 crc kubenswrapper[4971]: I1213 08:54:40.434005 4971 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-rvv8f" Dec 13 08:54:40 crc kubenswrapper[4971]: I1213 08:54:40.462074 4971 scope.go:117] "RemoveContainer" containerID="d78a25965a2bf65fbfe1c349506cf8f9006f48da37ccd1e3bd62f51cad684d11" Dec 13 08:54:40 crc kubenswrapper[4971]: I1213 08:54:40.480217 4971 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-nnl98\" (UniqueName: \"kubernetes.io/projected/dda9fe48-7439-4f74-bd2a-7733a4697c13-kube-api-access-nnl98\") pod \"dda9fe48-7439-4f74-bd2a-7733a4697c13\" (UID: \"dda9fe48-7439-4f74-bd2a-7733a4697c13\") " Dec 13 08:54:40 crc kubenswrapper[4971]: I1213 08:54:40.480840 4971 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/dda9fe48-7439-4f74-bd2a-7733a4697c13-catalog-content\") pod \"dda9fe48-7439-4f74-bd2a-7733a4697c13\" (UID: \"dda9fe48-7439-4f74-bd2a-7733a4697c13\") " Dec 13 08:54:40 crc kubenswrapper[4971]: I1213 08:54:40.481020 4971 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/dda9fe48-7439-4f74-bd2a-7733a4697c13-utilities\") pod \"dda9fe48-7439-4f74-bd2a-7733a4697c13\" (UID: \"dda9fe48-7439-4f74-bd2a-7733a4697c13\") " Dec 13 08:54:40 crc kubenswrapper[4971]: I1213 08:54:40.481740 4971 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/dda9fe48-7439-4f74-bd2a-7733a4697c13-utilities" (OuterVolumeSpecName: "utilities") pod "dda9fe48-7439-4f74-bd2a-7733a4697c13" (UID: "dda9fe48-7439-4f74-bd2a-7733a4697c13"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 13 08:54:40 crc kubenswrapper[4971]: I1213 08:54:40.487810 4971 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/dda9fe48-7439-4f74-bd2a-7733a4697c13-kube-api-access-nnl98" (OuterVolumeSpecName: "kube-api-access-nnl98") pod "dda9fe48-7439-4f74-bd2a-7733a4697c13" (UID: "dda9fe48-7439-4f74-bd2a-7733a4697c13"). InnerVolumeSpecName "kube-api-access-nnl98". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 13 08:54:40 crc kubenswrapper[4971]: I1213 08:54:40.491062 4971 scope.go:117] "RemoveContainer" containerID="e92dd69eacd8e2a969e769f412345cd8cf13bb597ee235f6d07a1dff61186a38" Dec 13 08:54:40 crc kubenswrapper[4971]: I1213 08:54:40.502743 4971 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/dda9fe48-7439-4f74-bd2a-7733a4697c13-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "dda9fe48-7439-4f74-bd2a-7733a4697c13" (UID: "dda9fe48-7439-4f74-bd2a-7733a4697c13"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 13 08:54:40 crc kubenswrapper[4971]: I1213 08:54:40.587976 4971 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-nnl98\" (UniqueName: \"kubernetes.io/projected/dda9fe48-7439-4f74-bd2a-7733a4697c13-kube-api-access-nnl98\") on node \"crc\" DevicePath \"\"" Dec 13 08:54:40 crc kubenswrapper[4971]: I1213 08:54:40.588603 4971 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/dda9fe48-7439-4f74-bd2a-7733a4697c13-catalog-content\") on node \"crc\" DevicePath \"\"" Dec 13 08:54:40 crc kubenswrapper[4971]: I1213 08:54:40.588706 4971 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/dda9fe48-7439-4f74-bd2a-7733a4697c13-utilities\") on node \"crc\" DevicePath \"\"" Dec 13 08:54:40 crc kubenswrapper[4971]: I1213 08:54:40.590546 4971 scope.go:117] "RemoveContainer" containerID="efb01dd66a626d4fbfdbd45508fda55e2c6ea432f6a3f70927f7a9ce03957d94" Dec 13 08:54:40 crc kubenswrapper[4971]: E1213 08:54:40.591753 4971 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"efb01dd66a626d4fbfdbd45508fda55e2c6ea432f6a3f70927f7a9ce03957d94\": container with ID starting with efb01dd66a626d4fbfdbd45508fda55e2c6ea432f6a3f70927f7a9ce03957d94 not found: ID does not exist" containerID="efb01dd66a626d4fbfdbd45508fda55e2c6ea432f6a3f70927f7a9ce03957d94" Dec 13 08:54:40 crc kubenswrapper[4971]: I1213 08:54:40.591808 4971 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"efb01dd66a626d4fbfdbd45508fda55e2c6ea432f6a3f70927f7a9ce03957d94"} err="failed to get container status \"efb01dd66a626d4fbfdbd45508fda55e2c6ea432f6a3f70927f7a9ce03957d94\": rpc error: code = NotFound desc = could not find container \"efb01dd66a626d4fbfdbd45508fda55e2c6ea432f6a3f70927f7a9ce03957d94\": container with ID starting with efb01dd66a626d4fbfdbd45508fda55e2c6ea432f6a3f70927f7a9ce03957d94 not found: ID does not exist" Dec 13 08:54:40 crc kubenswrapper[4971]: I1213 08:54:40.591842 4971 scope.go:117] "RemoveContainer" containerID="d78a25965a2bf65fbfe1c349506cf8f9006f48da37ccd1e3bd62f51cad684d11" Dec 13 08:54:40 crc kubenswrapper[4971]: E1213 08:54:40.593012 4971 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"d78a25965a2bf65fbfe1c349506cf8f9006f48da37ccd1e3bd62f51cad684d11\": container with ID starting with d78a25965a2bf65fbfe1c349506cf8f9006f48da37ccd1e3bd62f51cad684d11 not found: ID does not exist" containerID="d78a25965a2bf65fbfe1c349506cf8f9006f48da37ccd1e3bd62f51cad684d11" Dec 13 08:54:40 crc kubenswrapper[4971]: I1213 08:54:40.593093 4971 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"d78a25965a2bf65fbfe1c349506cf8f9006f48da37ccd1e3bd62f51cad684d11"} err="failed to get container status \"d78a25965a2bf65fbfe1c349506cf8f9006f48da37ccd1e3bd62f51cad684d11\": rpc error: code = NotFound desc = could not find container \"d78a25965a2bf65fbfe1c349506cf8f9006f48da37ccd1e3bd62f51cad684d11\": container with ID starting with d78a25965a2bf65fbfe1c349506cf8f9006f48da37ccd1e3bd62f51cad684d11 not found: ID does not exist" Dec 13 08:54:40 crc kubenswrapper[4971]: I1213 08:54:40.593142 4971 scope.go:117] "RemoveContainer" containerID="e92dd69eacd8e2a969e769f412345cd8cf13bb597ee235f6d07a1dff61186a38" Dec 13 08:54:40 crc kubenswrapper[4971]: E1213 08:54:40.594399 4971 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"e92dd69eacd8e2a969e769f412345cd8cf13bb597ee235f6d07a1dff61186a38\": container with ID starting with e92dd69eacd8e2a969e769f412345cd8cf13bb597ee235f6d07a1dff61186a38 not found: ID does not exist" containerID="e92dd69eacd8e2a969e769f412345cd8cf13bb597ee235f6d07a1dff61186a38" Dec 13 08:54:40 crc kubenswrapper[4971]: I1213 08:54:40.594434 4971 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"e92dd69eacd8e2a969e769f412345cd8cf13bb597ee235f6d07a1dff61186a38"} err="failed to get container status \"e92dd69eacd8e2a969e769f412345cd8cf13bb597ee235f6d07a1dff61186a38\": rpc error: code = NotFound desc = could not find container \"e92dd69eacd8e2a969e769f412345cd8cf13bb597ee235f6d07a1dff61186a38\": container with ID starting with e92dd69eacd8e2a969e769f412345cd8cf13bb597ee235f6d07a1dff61186a38 not found: ID does not exist" Dec 13 08:54:40 crc kubenswrapper[4971]: I1213 08:54:40.781318 4971 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-rvv8f"] Dec 13 08:54:40 crc kubenswrapper[4971]: I1213 08:54:40.796667 4971 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-marketplace-rvv8f"] Dec 13 08:54:41 crc kubenswrapper[4971]: I1213 08:54:41.783438 4971 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="dda9fe48-7439-4f74-bd2a-7733a4697c13" path="/var/lib/kubelet/pods/dda9fe48-7439-4f74-bd2a-7733a4697c13/volumes" Dec 13 08:55:46 crc kubenswrapper[4971]: I1213 08:55:46.161403 4971 patch_prober.go:28] interesting pod/machine-config-daemon-82xjz container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 13 08:55:46 crc kubenswrapper[4971]: I1213 08:55:46.162106 4971 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-82xjz" podUID="e0c4af12-edda-41d7-9b44-f87396174ef2" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 13 08:56:16 crc kubenswrapper[4971]: I1213 08:56:16.153536 4971 patch_prober.go:28] interesting pod/machine-config-daemon-82xjz container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 13 08:56:16 crc kubenswrapper[4971]: I1213 08:56:16.154331 4971 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-82xjz" podUID="e0c4af12-edda-41d7-9b44-f87396174ef2" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 13 08:56:46 crc kubenswrapper[4971]: I1213 08:56:46.153438 4971 patch_prober.go:28] interesting pod/machine-config-daemon-82xjz container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 13 08:56:46 crc kubenswrapper[4971]: I1213 08:56:46.154267 4971 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-82xjz" podUID="e0c4af12-edda-41d7-9b44-f87396174ef2" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 13 08:56:46 crc kubenswrapper[4971]: I1213 08:56:46.154330 4971 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-82xjz" Dec 13 08:56:46 crc kubenswrapper[4971]: I1213 08:56:46.155393 4971 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"3f281a9adda16a9b53c490949e166088ac22b2347e901a634eebb681b7729981"} pod="openshift-machine-config-operator/machine-config-daemon-82xjz" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Dec 13 08:56:46 crc kubenswrapper[4971]: I1213 08:56:46.155456 4971 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-82xjz" podUID="e0c4af12-edda-41d7-9b44-f87396174ef2" containerName="machine-config-daemon" containerID="cri-o://3f281a9adda16a9b53c490949e166088ac22b2347e901a634eebb681b7729981" gracePeriod=600 Dec 13 08:56:46 crc kubenswrapper[4971]: I1213 08:56:46.502877 4971 generic.go:334] "Generic (PLEG): container finished" podID="e0c4af12-edda-41d7-9b44-f87396174ef2" containerID="3f281a9adda16a9b53c490949e166088ac22b2347e901a634eebb681b7729981" exitCode=0 Dec 13 08:56:46 crc kubenswrapper[4971]: I1213 08:56:46.502958 4971 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-82xjz" event={"ID":"e0c4af12-edda-41d7-9b44-f87396174ef2","Type":"ContainerDied","Data":"3f281a9adda16a9b53c490949e166088ac22b2347e901a634eebb681b7729981"} Dec 13 08:56:46 crc kubenswrapper[4971]: I1213 08:56:46.503511 4971 scope.go:117] "RemoveContainer" containerID="b875432558a41dd98c673e3df9c78c3a2bcd69eb63408b435826e3d82347d2a6" Dec 13 08:56:47 crc kubenswrapper[4971]: I1213 08:56:47.515164 4971 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-82xjz" event={"ID":"e0c4af12-edda-41d7-9b44-f87396174ef2","Type":"ContainerStarted","Data":"95d975a85829af1f5477ac9667504f9ab79ab97ddaabd918334d5b4218e3d164"} Dec 13 08:58:46 crc kubenswrapper[4971]: I1213 08:58:46.153419 4971 patch_prober.go:28] interesting pod/machine-config-daemon-82xjz container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 13 08:58:46 crc kubenswrapper[4971]: I1213 08:58:46.154484 4971 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-82xjz" podUID="e0c4af12-edda-41d7-9b44-f87396174ef2" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 13 08:59:04 crc kubenswrapper[4971]: I1213 08:59:04.873347 4971 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/community-operators-69gmv"] Dec 13 08:59:04 crc kubenswrapper[4971]: E1213 08:59:04.874709 4971 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="dda9fe48-7439-4f74-bd2a-7733a4697c13" containerName="extract-utilities" Dec 13 08:59:04 crc kubenswrapper[4971]: I1213 08:59:04.874730 4971 state_mem.go:107] "Deleted CPUSet assignment" podUID="dda9fe48-7439-4f74-bd2a-7733a4697c13" containerName="extract-utilities" Dec 13 08:59:04 crc kubenswrapper[4971]: E1213 08:59:04.874756 4971 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="dda9fe48-7439-4f74-bd2a-7733a4697c13" containerName="extract-content" Dec 13 08:59:04 crc kubenswrapper[4971]: I1213 08:59:04.874764 4971 state_mem.go:107] "Deleted CPUSet assignment" podUID="dda9fe48-7439-4f74-bd2a-7733a4697c13" containerName="extract-content" Dec 13 08:59:04 crc kubenswrapper[4971]: E1213 08:59:04.874784 4971 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="dda9fe48-7439-4f74-bd2a-7733a4697c13" containerName="registry-server" Dec 13 08:59:04 crc kubenswrapper[4971]: I1213 08:59:04.874790 4971 state_mem.go:107] "Deleted CPUSet assignment" podUID="dda9fe48-7439-4f74-bd2a-7733a4697c13" containerName="registry-server" Dec 13 08:59:04 crc kubenswrapper[4971]: I1213 08:59:04.874989 4971 memory_manager.go:354] "RemoveStaleState removing state" podUID="dda9fe48-7439-4f74-bd2a-7733a4697c13" containerName="registry-server" Dec 13 08:59:04 crc kubenswrapper[4971]: I1213 08:59:04.876746 4971 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-69gmv" Dec 13 08:59:04 crc kubenswrapper[4971]: I1213 08:59:04.907218 4971 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-69gmv"] Dec 13 08:59:04 crc kubenswrapper[4971]: I1213 08:59:04.982388 4971 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-jxfb4\" (UniqueName: \"kubernetes.io/projected/bf2757ef-bc9b-4a59-bfaa-083d5781203a-kube-api-access-jxfb4\") pod \"community-operators-69gmv\" (UID: \"bf2757ef-bc9b-4a59-bfaa-083d5781203a\") " pod="openshift-marketplace/community-operators-69gmv" Dec 13 08:59:04 crc kubenswrapper[4971]: I1213 08:59:04.982484 4971 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/bf2757ef-bc9b-4a59-bfaa-083d5781203a-catalog-content\") pod \"community-operators-69gmv\" (UID: \"bf2757ef-bc9b-4a59-bfaa-083d5781203a\") " pod="openshift-marketplace/community-operators-69gmv" Dec 13 08:59:04 crc kubenswrapper[4971]: I1213 08:59:04.982590 4971 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/bf2757ef-bc9b-4a59-bfaa-083d5781203a-utilities\") pod \"community-operators-69gmv\" (UID: \"bf2757ef-bc9b-4a59-bfaa-083d5781203a\") " pod="openshift-marketplace/community-operators-69gmv" Dec 13 08:59:05 crc kubenswrapper[4971]: I1213 08:59:05.083858 4971 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/bf2757ef-bc9b-4a59-bfaa-083d5781203a-utilities\") pod \"community-operators-69gmv\" (UID: \"bf2757ef-bc9b-4a59-bfaa-083d5781203a\") " pod="openshift-marketplace/community-operators-69gmv" Dec 13 08:59:05 crc kubenswrapper[4971]: I1213 08:59:05.084109 4971 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-jxfb4\" (UniqueName: \"kubernetes.io/projected/bf2757ef-bc9b-4a59-bfaa-083d5781203a-kube-api-access-jxfb4\") pod \"community-operators-69gmv\" (UID: \"bf2757ef-bc9b-4a59-bfaa-083d5781203a\") " pod="openshift-marketplace/community-operators-69gmv" Dec 13 08:59:05 crc kubenswrapper[4971]: I1213 08:59:05.084163 4971 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/bf2757ef-bc9b-4a59-bfaa-083d5781203a-catalog-content\") pod \"community-operators-69gmv\" (UID: \"bf2757ef-bc9b-4a59-bfaa-083d5781203a\") " pod="openshift-marketplace/community-operators-69gmv" Dec 13 08:59:05 crc kubenswrapper[4971]: I1213 08:59:05.084849 4971 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/bf2757ef-bc9b-4a59-bfaa-083d5781203a-utilities\") pod \"community-operators-69gmv\" (UID: \"bf2757ef-bc9b-4a59-bfaa-083d5781203a\") " pod="openshift-marketplace/community-operators-69gmv" Dec 13 08:59:05 crc kubenswrapper[4971]: I1213 08:59:05.084905 4971 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/bf2757ef-bc9b-4a59-bfaa-083d5781203a-catalog-content\") pod \"community-operators-69gmv\" (UID: \"bf2757ef-bc9b-4a59-bfaa-083d5781203a\") " pod="openshift-marketplace/community-operators-69gmv" Dec 13 08:59:05 crc kubenswrapper[4971]: I1213 08:59:05.118543 4971 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-jxfb4\" (UniqueName: \"kubernetes.io/projected/bf2757ef-bc9b-4a59-bfaa-083d5781203a-kube-api-access-jxfb4\") pod \"community-operators-69gmv\" (UID: \"bf2757ef-bc9b-4a59-bfaa-083d5781203a\") " pod="openshift-marketplace/community-operators-69gmv" Dec 13 08:59:05 crc kubenswrapper[4971]: I1213 08:59:05.208473 4971 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-69gmv" Dec 13 08:59:05 crc kubenswrapper[4971]: I1213 08:59:05.805455 4971 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-69gmv"] Dec 13 08:59:06 crc kubenswrapper[4971]: I1213 08:59:06.144309 4971 generic.go:334] "Generic (PLEG): container finished" podID="bf2757ef-bc9b-4a59-bfaa-083d5781203a" containerID="a0dd1f0cd93895d710aeec91d464622168735514ea7f8fcb6481bb1523e735b5" exitCode=0 Dec 13 08:59:06 crc kubenswrapper[4971]: I1213 08:59:06.146111 4971 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-69gmv" event={"ID":"bf2757ef-bc9b-4a59-bfaa-083d5781203a","Type":"ContainerDied","Data":"a0dd1f0cd93895d710aeec91d464622168735514ea7f8fcb6481bb1523e735b5"} Dec 13 08:59:06 crc kubenswrapper[4971]: I1213 08:59:06.146297 4971 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-69gmv" event={"ID":"bf2757ef-bc9b-4a59-bfaa-083d5781203a","Type":"ContainerStarted","Data":"0be31fdee4dfca02c5177ebe82e0bba8f9439617da9fdc037b5067ba74b489a0"} Dec 13 08:59:07 crc kubenswrapper[4971]: I1213 08:59:07.245857 4971 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-operators-f4tdk"] Dec 13 08:59:07 crc kubenswrapper[4971]: I1213 08:59:07.249285 4971 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-f4tdk" Dec 13 08:59:07 crc kubenswrapper[4971]: I1213 08:59:07.264595 4971 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-f4tdk"] Dec 13 08:59:07 crc kubenswrapper[4971]: I1213 08:59:07.360369 4971 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-mpbgz\" (UniqueName: \"kubernetes.io/projected/5b22775d-1694-4361-be84-144aa19aa2d1-kube-api-access-mpbgz\") pod \"redhat-operators-f4tdk\" (UID: \"5b22775d-1694-4361-be84-144aa19aa2d1\") " pod="openshift-marketplace/redhat-operators-f4tdk" Dec 13 08:59:07 crc kubenswrapper[4971]: I1213 08:59:07.360418 4971 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/5b22775d-1694-4361-be84-144aa19aa2d1-utilities\") pod \"redhat-operators-f4tdk\" (UID: \"5b22775d-1694-4361-be84-144aa19aa2d1\") " pod="openshift-marketplace/redhat-operators-f4tdk" Dec 13 08:59:07 crc kubenswrapper[4971]: I1213 08:59:07.360471 4971 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/5b22775d-1694-4361-be84-144aa19aa2d1-catalog-content\") pod \"redhat-operators-f4tdk\" (UID: \"5b22775d-1694-4361-be84-144aa19aa2d1\") " pod="openshift-marketplace/redhat-operators-f4tdk" Dec 13 08:59:07 crc kubenswrapper[4971]: I1213 08:59:07.436243 4971 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/certified-operators-tjln9"] Dec 13 08:59:07 crc kubenswrapper[4971]: I1213 08:59:07.438729 4971 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-tjln9" Dec 13 08:59:07 crc kubenswrapper[4971]: I1213 08:59:07.462644 4971 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/5b22775d-1694-4361-be84-144aa19aa2d1-utilities\") pod \"redhat-operators-f4tdk\" (UID: \"5b22775d-1694-4361-be84-144aa19aa2d1\") " pod="openshift-marketplace/redhat-operators-f4tdk" Dec 13 08:59:07 crc kubenswrapper[4971]: I1213 08:59:07.462758 4971 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/5b22775d-1694-4361-be84-144aa19aa2d1-catalog-content\") pod \"redhat-operators-f4tdk\" (UID: \"5b22775d-1694-4361-be84-144aa19aa2d1\") " pod="openshift-marketplace/redhat-operators-f4tdk" Dec 13 08:59:07 crc kubenswrapper[4971]: I1213 08:59:07.462990 4971 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-mpbgz\" (UniqueName: \"kubernetes.io/projected/5b22775d-1694-4361-be84-144aa19aa2d1-kube-api-access-mpbgz\") pod \"redhat-operators-f4tdk\" (UID: \"5b22775d-1694-4361-be84-144aa19aa2d1\") " pod="openshift-marketplace/redhat-operators-f4tdk" Dec 13 08:59:07 crc kubenswrapper[4971]: I1213 08:59:07.463416 4971 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/5b22775d-1694-4361-be84-144aa19aa2d1-utilities\") pod \"redhat-operators-f4tdk\" (UID: \"5b22775d-1694-4361-be84-144aa19aa2d1\") " pod="openshift-marketplace/redhat-operators-f4tdk" Dec 13 08:59:07 crc kubenswrapper[4971]: I1213 08:59:07.463466 4971 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/5b22775d-1694-4361-be84-144aa19aa2d1-catalog-content\") pod \"redhat-operators-f4tdk\" (UID: \"5b22775d-1694-4361-be84-144aa19aa2d1\") " pod="openshift-marketplace/redhat-operators-f4tdk" Dec 13 08:59:07 crc kubenswrapper[4971]: I1213 08:59:07.463601 4971 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-tjln9"] Dec 13 08:59:07 crc kubenswrapper[4971]: I1213 08:59:07.491551 4971 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-mpbgz\" (UniqueName: \"kubernetes.io/projected/5b22775d-1694-4361-be84-144aa19aa2d1-kube-api-access-mpbgz\") pod \"redhat-operators-f4tdk\" (UID: \"5b22775d-1694-4361-be84-144aa19aa2d1\") " pod="openshift-marketplace/redhat-operators-f4tdk" Dec 13 08:59:07 crc kubenswrapper[4971]: I1213 08:59:07.565651 4971 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/7cad90cc-49d3-455b-ac98-8f95077388b3-catalog-content\") pod \"certified-operators-tjln9\" (UID: \"7cad90cc-49d3-455b-ac98-8f95077388b3\") " pod="openshift-marketplace/certified-operators-tjln9" Dec 13 08:59:07 crc kubenswrapper[4971]: I1213 08:59:07.565720 4971 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-zztfd\" (UniqueName: \"kubernetes.io/projected/7cad90cc-49d3-455b-ac98-8f95077388b3-kube-api-access-zztfd\") pod \"certified-operators-tjln9\" (UID: \"7cad90cc-49d3-455b-ac98-8f95077388b3\") " pod="openshift-marketplace/certified-operators-tjln9" Dec 13 08:59:07 crc kubenswrapper[4971]: I1213 08:59:07.565902 4971 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/7cad90cc-49d3-455b-ac98-8f95077388b3-utilities\") pod \"certified-operators-tjln9\" (UID: \"7cad90cc-49d3-455b-ac98-8f95077388b3\") " pod="openshift-marketplace/certified-operators-tjln9" Dec 13 08:59:07 crc kubenswrapper[4971]: I1213 08:59:07.575637 4971 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-f4tdk" Dec 13 08:59:07 crc kubenswrapper[4971]: I1213 08:59:07.667572 4971 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/7cad90cc-49d3-455b-ac98-8f95077388b3-utilities\") pod \"certified-operators-tjln9\" (UID: \"7cad90cc-49d3-455b-ac98-8f95077388b3\") " pod="openshift-marketplace/certified-operators-tjln9" Dec 13 08:59:07 crc kubenswrapper[4971]: I1213 08:59:07.667704 4971 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-zztfd\" (UniqueName: \"kubernetes.io/projected/7cad90cc-49d3-455b-ac98-8f95077388b3-kube-api-access-zztfd\") pod \"certified-operators-tjln9\" (UID: \"7cad90cc-49d3-455b-ac98-8f95077388b3\") " pod="openshift-marketplace/certified-operators-tjln9" Dec 13 08:59:07 crc kubenswrapper[4971]: I1213 08:59:07.667727 4971 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/7cad90cc-49d3-455b-ac98-8f95077388b3-catalog-content\") pod \"certified-operators-tjln9\" (UID: \"7cad90cc-49d3-455b-ac98-8f95077388b3\") " pod="openshift-marketplace/certified-operators-tjln9" Dec 13 08:59:07 crc kubenswrapper[4971]: I1213 08:59:07.668167 4971 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/7cad90cc-49d3-455b-ac98-8f95077388b3-catalog-content\") pod \"certified-operators-tjln9\" (UID: \"7cad90cc-49d3-455b-ac98-8f95077388b3\") " pod="openshift-marketplace/certified-operators-tjln9" Dec 13 08:59:07 crc kubenswrapper[4971]: I1213 08:59:07.668393 4971 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/7cad90cc-49d3-455b-ac98-8f95077388b3-utilities\") pod \"certified-operators-tjln9\" (UID: \"7cad90cc-49d3-455b-ac98-8f95077388b3\") " pod="openshift-marketplace/certified-operators-tjln9" Dec 13 08:59:08 crc kubenswrapper[4971]: I1213 08:59:08.175938 4971 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-69gmv" event={"ID":"bf2757ef-bc9b-4a59-bfaa-083d5781203a","Type":"ContainerStarted","Data":"ce2128a389279a36602c454e6df1abaa3621b4884608cd1b42e3030e48f700cc"} Dec 13 08:59:08 crc kubenswrapper[4971]: I1213 08:59:08.240700 4971 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-zztfd\" (UniqueName: \"kubernetes.io/projected/7cad90cc-49d3-455b-ac98-8f95077388b3-kube-api-access-zztfd\") pod \"certified-operators-tjln9\" (UID: \"7cad90cc-49d3-455b-ac98-8f95077388b3\") " pod="openshift-marketplace/certified-operators-tjln9" Dec 13 08:59:08 crc kubenswrapper[4971]: I1213 08:59:08.359637 4971 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-tjln9" Dec 13 08:59:08 crc kubenswrapper[4971]: I1213 08:59:08.955428 4971 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-f4tdk"] Dec 13 08:59:08 crc kubenswrapper[4971]: W1213 08:59:08.960782 4971 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod5b22775d_1694_4361_be84_144aa19aa2d1.slice/crio-2cef81624a4f2296e31f3350369d31ad0f6b25e8ed22f4a0fd481123006a8260 WatchSource:0}: Error finding container 2cef81624a4f2296e31f3350369d31ad0f6b25e8ed22f4a0fd481123006a8260: Status 404 returned error can't find the container with id 2cef81624a4f2296e31f3350369d31ad0f6b25e8ed22f4a0fd481123006a8260 Dec 13 08:59:09 crc kubenswrapper[4971]: W1213 08:59:09.091957 4971 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod7cad90cc_49d3_455b_ac98_8f95077388b3.slice/crio-3753c74dd1adeb6c9fc8ce3e723e57f115d80f1ed211b8ba60cfb0c274a3b314 WatchSource:0}: Error finding container 3753c74dd1adeb6c9fc8ce3e723e57f115d80f1ed211b8ba60cfb0c274a3b314: Status 404 returned error can't find the container with id 3753c74dd1adeb6c9fc8ce3e723e57f115d80f1ed211b8ba60cfb0c274a3b314 Dec 13 08:59:09 crc kubenswrapper[4971]: I1213 08:59:09.098051 4971 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-tjln9"] Dec 13 08:59:09 crc kubenswrapper[4971]: I1213 08:59:09.187666 4971 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-tjln9" event={"ID":"7cad90cc-49d3-455b-ac98-8f95077388b3","Type":"ContainerStarted","Data":"3753c74dd1adeb6c9fc8ce3e723e57f115d80f1ed211b8ba60cfb0c274a3b314"} Dec 13 08:59:09 crc kubenswrapper[4971]: I1213 08:59:09.190218 4971 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-f4tdk" event={"ID":"5b22775d-1694-4361-be84-144aa19aa2d1","Type":"ContainerStarted","Data":"2cef81624a4f2296e31f3350369d31ad0f6b25e8ed22f4a0fd481123006a8260"} Dec 13 08:59:09 crc kubenswrapper[4971]: I1213 08:59:09.192241 4971 generic.go:334] "Generic (PLEG): container finished" podID="bf2757ef-bc9b-4a59-bfaa-083d5781203a" containerID="ce2128a389279a36602c454e6df1abaa3621b4884608cd1b42e3030e48f700cc" exitCode=0 Dec 13 08:59:09 crc kubenswrapper[4971]: I1213 08:59:09.192365 4971 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-69gmv" event={"ID":"bf2757ef-bc9b-4a59-bfaa-083d5781203a","Type":"ContainerDied","Data":"ce2128a389279a36602c454e6df1abaa3621b4884608cd1b42e3030e48f700cc"} Dec 13 08:59:10 crc kubenswrapper[4971]: I1213 08:59:10.218154 4971 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-69gmv" event={"ID":"bf2757ef-bc9b-4a59-bfaa-083d5781203a","Type":"ContainerStarted","Data":"719581aa2deef81126070199a771d37eee67b98bcfc5d2f7d62473e0cba05371"} Dec 13 08:59:10 crc kubenswrapper[4971]: I1213 08:59:10.226458 4971 generic.go:334] "Generic (PLEG): container finished" podID="7cad90cc-49d3-455b-ac98-8f95077388b3" containerID="d43d1b4fb8986d897439cb8e7b1d8624f35a7dd98b8129f1066754c2adfd0ead" exitCode=0 Dec 13 08:59:10 crc kubenswrapper[4971]: I1213 08:59:10.226699 4971 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-tjln9" event={"ID":"7cad90cc-49d3-455b-ac98-8f95077388b3","Type":"ContainerDied","Data":"d43d1b4fb8986d897439cb8e7b1d8624f35a7dd98b8129f1066754c2adfd0ead"} Dec 13 08:59:10 crc kubenswrapper[4971]: I1213 08:59:10.231658 4971 generic.go:334] "Generic (PLEG): container finished" podID="5b22775d-1694-4361-be84-144aa19aa2d1" containerID="06c02335597f23f902be3c13f3da3948e9cb1b31dff52479770d821db84fe94d" exitCode=0 Dec 13 08:59:10 crc kubenswrapper[4971]: I1213 08:59:10.231726 4971 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-f4tdk" event={"ID":"5b22775d-1694-4361-be84-144aa19aa2d1","Type":"ContainerDied","Data":"06c02335597f23f902be3c13f3da3948e9cb1b31dff52479770d821db84fe94d"} Dec 13 08:59:10 crc kubenswrapper[4971]: I1213 08:59:10.263228 4971 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/community-operators-69gmv" podStartSLOduration=2.784803002 podStartE2EDuration="6.263195411s" podCreationTimestamp="2025-12-13 08:59:04 +0000 UTC" firstStartedPulling="2025-12-13 08:59:06.151176713 +0000 UTC m=+7802.755586161" lastFinishedPulling="2025-12-13 08:59:09.629569122 +0000 UTC m=+7806.233978570" observedRunningTime="2025-12-13 08:59:10.246192501 +0000 UTC m=+7806.850601949" watchObservedRunningTime="2025-12-13 08:59:10.263195411 +0000 UTC m=+7806.867604859" Dec 13 08:59:11 crc kubenswrapper[4971]: I1213 08:59:11.244482 4971 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-f4tdk" event={"ID":"5b22775d-1694-4361-be84-144aa19aa2d1","Type":"ContainerStarted","Data":"211a0a1f5a37a1590dddaac9baec10ed76239734bbf636c98faba7e9dcd4fd63"} Dec 13 08:59:12 crc kubenswrapper[4971]: I1213 08:59:12.256662 4971 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-tjln9" event={"ID":"7cad90cc-49d3-455b-ac98-8f95077388b3","Type":"ContainerStarted","Data":"81d175237e71407aeb5af8443277901378ec6e513bff7968d53b9359a9f1bd46"} Dec 13 08:59:14 crc kubenswrapper[4971]: I1213 08:59:14.276704 4971 generic.go:334] "Generic (PLEG): container finished" podID="7cad90cc-49d3-455b-ac98-8f95077388b3" containerID="81d175237e71407aeb5af8443277901378ec6e513bff7968d53b9359a9f1bd46" exitCode=0 Dec 13 08:59:14 crc kubenswrapper[4971]: I1213 08:59:14.276752 4971 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-tjln9" event={"ID":"7cad90cc-49d3-455b-ac98-8f95077388b3","Type":"ContainerDied","Data":"81d175237e71407aeb5af8443277901378ec6e513bff7968d53b9359a9f1bd46"} Dec 13 08:59:15 crc kubenswrapper[4971]: I1213 08:59:15.208913 4971 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/community-operators-69gmv" Dec 13 08:59:15 crc kubenswrapper[4971]: I1213 08:59:15.209440 4971 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/community-operators-69gmv" Dec 13 08:59:15 crc kubenswrapper[4971]: I1213 08:59:15.290631 4971 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/community-operators-69gmv" Dec 13 08:59:15 crc kubenswrapper[4971]: I1213 08:59:15.358975 4971 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/community-operators-69gmv" Dec 13 08:59:16 crc kubenswrapper[4971]: I1213 08:59:16.153591 4971 patch_prober.go:28] interesting pod/machine-config-daemon-82xjz container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 13 08:59:16 crc kubenswrapper[4971]: I1213 08:59:16.153964 4971 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-82xjz" podUID="e0c4af12-edda-41d7-9b44-f87396174ef2" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 13 08:59:16 crc kubenswrapper[4971]: I1213 08:59:16.303878 4971 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-tjln9" event={"ID":"7cad90cc-49d3-455b-ac98-8f95077388b3","Type":"ContainerStarted","Data":"92b67aca617672326265647a15a521181507af70709717a835edf609337dceb2"} Dec 13 08:59:16 crc kubenswrapper[4971]: I1213 08:59:16.307930 4971 generic.go:334] "Generic (PLEG): container finished" podID="5b22775d-1694-4361-be84-144aa19aa2d1" containerID="211a0a1f5a37a1590dddaac9baec10ed76239734bbf636c98faba7e9dcd4fd63" exitCode=0 Dec 13 08:59:16 crc kubenswrapper[4971]: I1213 08:59:16.307975 4971 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-f4tdk" event={"ID":"5b22775d-1694-4361-be84-144aa19aa2d1","Type":"ContainerDied","Data":"211a0a1f5a37a1590dddaac9baec10ed76239734bbf636c98faba7e9dcd4fd63"} Dec 13 08:59:16 crc kubenswrapper[4971]: I1213 08:59:16.332767 4971 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/certified-operators-tjln9" podStartSLOduration=4.411314698 podStartE2EDuration="9.332743297s" podCreationTimestamp="2025-12-13 08:59:07 +0000 UTC" firstStartedPulling="2025-12-13 08:59:10.230078764 +0000 UTC m=+7806.834488262" lastFinishedPulling="2025-12-13 08:59:15.151507403 +0000 UTC m=+7811.755916861" observedRunningTime="2025-12-13 08:59:16.331803184 +0000 UTC m=+7812.936212662" watchObservedRunningTime="2025-12-13 08:59:16.332743297 +0000 UTC m=+7812.937152745" Dec 13 08:59:18 crc kubenswrapper[4971]: I1213 08:59:18.330046 4971 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-f4tdk" event={"ID":"5b22775d-1694-4361-be84-144aa19aa2d1","Type":"ContainerStarted","Data":"30a5ae8ed75d3ac7a83e7695d485b88f1f707159e224509c8405a85633a645a3"} Dec 13 08:59:18 crc kubenswrapper[4971]: I1213 08:59:18.357095 4971 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-operators-f4tdk" podStartSLOduration=4.478360183 podStartE2EDuration="11.357075174s" podCreationTimestamp="2025-12-13 08:59:07 +0000 UTC" firstStartedPulling="2025-12-13 08:59:10.239447124 +0000 UTC m=+7806.843856572" lastFinishedPulling="2025-12-13 08:59:17.118162115 +0000 UTC m=+7813.722571563" observedRunningTime="2025-12-13 08:59:18.35612832 +0000 UTC m=+7814.960537778" watchObservedRunningTime="2025-12-13 08:59:18.357075174 +0000 UTC m=+7814.961484612" Dec 13 08:59:18 crc kubenswrapper[4971]: I1213 08:59:18.360156 4971 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/certified-operators-tjln9" Dec 13 08:59:18 crc kubenswrapper[4971]: I1213 08:59:18.360223 4971 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/certified-operators-tjln9" Dec 13 08:59:18 crc kubenswrapper[4971]: I1213 08:59:18.415561 4971 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/certified-operators-tjln9" Dec 13 08:59:19 crc kubenswrapper[4971]: I1213 08:59:19.828750 4971 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-69gmv"] Dec 13 08:59:19 crc kubenswrapper[4971]: I1213 08:59:19.829608 4971 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/community-operators-69gmv" podUID="bf2757ef-bc9b-4a59-bfaa-083d5781203a" containerName="registry-server" containerID="cri-o://719581aa2deef81126070199a771d37eee67b98bcfc5d2f7d62473e0cba05371" gracePeriod=2 Dec 13 08:59:20 crc kubenswrapper[4971]: I1213 08:59:20.353863 4971 generic.go:334] "Generic (PLEG): container finished" podID="bf2757ef-bc9b-4a59-bfaa-083d5781203a" containerID="719581aa2deef81126070199a771d37eee67b98bcfc5d2f7d62473e0cba05371" exitCode=0 Dec 13 08:59:20 crc kubenswrapper[4971]: I1213 08:59:20.354066 4971 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-69gmv" event={"ID":"bf2757ef-bc9b-4a59-bfaa-083d5781203a","Type":"ContainerDied","Data":"719581aa2deef81126070199a771d37eee67b98bcfc5d2f7d62473e0cba05371"} Dec 13 08:59:20 crc kubenswrapper[4971]: I1213 08:59:20.487127 4971 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-69gmv" Dec 13 08:59:20 crc kubenswrapper[4971]: I1213 08:59:20.605259 4971 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/bf2757ef-bc9b-4a59-bfaa-083d5781203a-utilities\") pod \"bf2757ef-bc9b-4a59-bfaa-083d5781203a\" (UID: \"bf2757ef-bc9b-4a59-bfaa-083d5781203a\") " Dec 13 08:59:20 crc kubenswrapper[4971]: I1213 08:59:20.605633 4971 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-jxfb4\" (UniqueName: \"kubernetes.io/projected/bf2757ef-bc9b-4a59-bfaa-083d5781203a-kube-api-access-jxfb4\") pod \"bf2757ef-bc9b-4a59-bfaa-083d5781203a\" (UID: \"bf2757ef-bc9b-4a59-bfaa-083d5781203a\") " Dec 13 08:59:20 crc kubenswrapper[4971]: I1213 08:59:20.605890 4971 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/bf2757ef-bc9b-4a59-bfaa-083d5781203a-catalog-content\") pod \"bf2757ef-bc9b-4a59-bfaa-083d5781203a\" (UID: \"bf2757ef-bc9b-4a59-bfaa-083d5781203a\") " Dec 13 08:59:20 crc kubenswrapper[4971]: I1213 08:59:20.607470 4971 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/bf2757ef-bc9b-4a59-bfaa-083d5781203a-utilities" (OuterVolumeSpecName: "utilities") pod "bf2757ef-bc9b-4a59-bfaa-083d5781203a" (UID: "bf2757ef-bc9b-4a59-bfaa-083d5781203a"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 13 08:59:20 crc kubenswrapper[4971]: I1213 08:59:20.625866 4971 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/bf2757ef-bc9b-4a59-bfaa-083d5781203a-kube-api-access-jxfb4" (OuterVolumeSpecName: "kube-api-access-jxfb4") pod "bf2757ef-bc9b-4a59-bfaa-083d5781203a" (UID: "bf2757ef-bc9b-4a59-bfaa-083d5781203a"). InnerVolumeSpecName "kube-api-access-jxfb4". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 13 08:59:20 crc kubenswrapper[4971]: I1213 08:59:20.664525 4971 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/bf2757ef-bc9b-4a59-bfaa-083d5781203a-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "bf2757ef-bc9b-4a59-bfaa-083d5781203a" (UID: "bf2757ef-bc9b-4a59-bfaa-083d5781203a"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 13 08:59:20 crc kubenswrapper[4971]: I1213 08:59:20.708483 4971 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/bf2757ef-bc9b-4a59-bfaa-083d5781203a-catalog-content\") on node \"crc\" DevicePath \"\"" Dec 13 08:59:20 crc kubenswrapper[4971]: I1213 08:59:20.708519 4971 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/bf2757ef-bc9b-4a59-bfaa-083d5781203a-utilities\") on node \"crc\" DevicePath \"\"" Dec 13 08:59:20 crc kubenswrapper[4971]: I1213 08:59:20.708557 4971 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-jxfb4\" (UniqueName: \"kubernetes.io/projected/bf2757ef-bc9b-4a59-bfaa-083d5781203a-kube-api-access-jxfb4\") on node \"crc\" DevicePath \"\"" Dec 13 08:59:21 crc kubenswrapper[4971]: I1213 08:59:21.369005 4971 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-69gmv" event={"ID":"bf2757ef-bc9b-4a59-bfaa-083d5781203a","Type":"ContainerDied","Data":"0be31fdee4dfca02c5177ebe82e0bba8f9439617da9fdc037b5067ba74b489a0"} Dec 13 08:59:21 crc kubenswrapper[4971]: I1213 08:59:21.369061 4971 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-69gmv" Dec 13 08:59:21 crc kubenswrapper[4971]: I1213 08:59:21.369573 4971 scope.go:117] "RemoveContainer" containerID="719581aa2deef81126070199a771d37eee67b98bcfc5d2f7d62473e0cba05371" Dec 13 08:59:21 crc kubenswrapper[4971]: I1213 08:59:21.410413 4971 scope.go:117] "RemoveContainer" containerID="ce2128a389279a36602c454e6df1abaa3621b4884608cd1b42e3030e48f700cc" Dec 13 08:59:21 crc kubenswrapper[4971]: I1213 08:59:21.418134 4971 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-69gmv"] Dec 13 08:59:21 crc kubenswrapper[4971]: I1213 08:59:21.449830 4971 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/community-operators-69gmv"] Dec 13 08:59:21 crc kubenswrapper[4971]: I1213 08:59:21.452924 4971 scope.go:117] "RemoveContainer" containerID="a0dd1f0cd93895d710aeec91d464622168735514ea7f8fcb6481bb1523e735b5" Dec 13 08:59:21 crc kubenswrapper[4971]: I1213 08:59:21.785088 4971 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="bf2757ef-bc9b-4a59-bfaa-083d5781203a" path="/var/lib/kubelet/pods/bf2757ef-bc9b-4a59-bfaa-083d5781203a/volumes" Dec 13 08:59:27 crc kubenswrapper[4971]: I1213 08:59:27.576724 4971 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-operators-f4tdk" Dec 13 08:59:27 crc kubenswrapper[4971]: I1213 08:59:27.577428 4971 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-operators-f4tdk" Dec 13 08:59:27 crc kubenswrapper[4971]: I1213 08:59:27.633136 4971 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-operators-f4tdk" Dec 13 08:59:28 crc kubenswrapper[4971]: I1213 08:59:28.419067 4971 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/certified-operators-tjln9" Dec 13 08:59:28 crc kubenswrapper[4971]: I1213 08:59:28.487961 4971 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-tjln9"] Dec 13 08:59:28 crc kubenswrapper[4971]: I1213 08:59:28.488241 4971 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/certified-operators-tjln9" podUID="7cad90cc-49d3-455b-ac98-8f95077388b3" containerName="registry-server" containerID="cri-o://92b67aca617672326265647a15a521181507af70709717a835edf609337dceb2" gracePeriod=2 Dec 13 08:59:28 crc kubenswrapper[4971]: I1213 08:59:28.531694 4971 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-operators-f4tdk" Dec 13 08:59:29 crc kubenswrapper[4971]: I1213 08:59:29.076092 4971 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-tjln9" Dec 13 08:59:29 crc kubenswrapper[4971]: I1213 08:59:29.175647 4971 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-zztfd\" (UniqueName: \"kubernetes.io/projected/7cad90cc-49d3-455b-ac98-8f95077388b3-kube-api-access-zztfd\") pod \"7cad90cc-49d3-455b-ac98-8f95077388b3\" (UID: \"7cad90cc-49d3-455b-ac98-8f95077388b3\") " Dec 13 08:59:29 crc kubenswrapper[4971]: I1213 08:59:29.175964 4971 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/7cad90cc-49d3-455b-ac98-8f95077388b3-utilities\") pod \"7cad90cc-49d3-455b-ac98-8f95077388b3\" (UID: \"7cad90cc-49d3-455b-ac98-8f95077388b3\") " Dec 13 08:59:29 crc kubenswrapper[4971]: I1213 08:59:29.176105 4971 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/7cad90cc-49d3-455b-ac98-8f95077388b3-catalog-content\") pod \"7cad90cc-49d3-455b-ac98-8f95077388b3\" (UID: \"7cad90cc-49d3-455b-ac98-8f95077388b3\") " Dec 13 08:59:29 crc kubenswrapper[4971]: I1213 08:59:29.178238 4971 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/7cad90cc-49d3-455b-ac98-8f95077388b3-utilities" (OuterVolumeSpecName: "utilities") pod "7cad90cc-49d3-455b-ac98-8f95077388b3" (UID: "7cad90cc-49d3-455b-ac98-8f95077388b3"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 13 08:59:29 crc kubenswrapper[4971]: I1213 08:59:29.210237 4971 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/7cad90cc-49d3-455b-ac98-8f95077388b3-kube-api-access-zztfd" (OuterVolumeSpecName: "kube-api-access-zztfd") pod "7cad90cc-49d3-455b-ac98-8f95077388b3" (UID: "7cad90cc-49d3-455b-ac98-8f95077388b3"). InnerVolumeSpecName "kube-api-access-zztfd". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 13 08:59:29 crc kubenswrapper[4971]: I1213 08:59:29.280233 4971 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/7cad90cc-49d3-455b-ac98-8f95077388b3-utilities\") on node \"crc\" DevicePath \"\"" Dec 13 08:59:29 crc kubenswrapper[4971]: I1213 08:59:29.280732 4971 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-zztfd\" (UniqueName: \"kubernetes.io/projected/7cad90cc-49d3-455b-ac98-8f95077388b3-kube-api-access-zztfd\") on node \"crc\" DevicePath \"\"" Dec 13 08:59:29 crc kubenswrapper[4971]: I1213 08:59:29.345865 4971 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/7cad90cc-49d3-455b-ac98-8f95077388b3-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "7cad90cc-49d3-455b-ac98-8f95077388b3" (UID: "7cad90cc-49d3-455b-ac98-8f95077388b3"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 13 08:59:29 crc kubenswrapper[4971]: I1213 08:59:29.384169 4971 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/7cad90cc-49d3-455b-ac98-8f95077388b3-catalog-content\") on node \"crc\" DevicePath \"\"" Dec 13 08:59:29 crc kubenswrapper[4971]: I1213 08:59:29.470590 4971 generic.go:334] "Generic (PLEG): container finished" podID="7cad90cc-49d3-455b-ac98-8f95077388b3" containerID="92b67aca617672326265647a15a521181507af70709717a835edf609337dceb2" exitCode=0 Dec 13 08:59:29 crc kubenswrapper[4971]: I1213 08:59:29.470648 4971 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-tjln9" event={"ID":"7cad90cc-49d3-455b-ac98-8f95077388b3","Type":"ContainerDied","Data":"92b67aca617672326265647a15a521181507af70709717a835edf609337dceb2"} Dec 13 08:59:29 crc kubenswrapper[4971]: I1213 08:59:29.470707 4971 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-tjln9" Dec 13 08:59:29 crc kubenswrapper[4971]: I1213 08:59:29.470737 4971 scope.go:117] "RemoveContainer" containerID="92b67aca617672326265647a15a521181507af70709717a835edf609337dceb2" Dec 13 08:59:29 crc kubenswrapper[4971]: I1213 08:59:29.470714 4971 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-tjln9" event={"ID":"7cad90cc-49d3-455b-ac98-8f95077388b3","Type":"ContainerDied","Data":"3753c74dd1adeb6c9fc8ce3e723e57f115d80f1ed211b8ba60cfb0c274a3b314"} Dec 13 08:59:29 crc kubenswrapper[4971]: I1213 08:59:29.504478 4971 scope.go:117] "RemoveContainer" containerID="81d175237e71407aeb5af8443277901378ec6e513bff7968d53b9359a9f1bd46" Dec 13 08:59:29 crc kubenswrapper[4971]: I1213 08:59:29.532146 4971 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-tjln9"] Dec 13 08:59:29 crc kubenswrapper[4971]: I1213 08:59:29.550038 4971 scope.go:117] "RemoveContainer" containerID="d43d1b4fb8986d897439cb8e7b1d8624f35a7dd98b8129f1066754c2adfd0ead" Dec 13 08:59:29 crc kubenswrapper[4971]: I1213 08:59:29.563786 4971 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/certified-operators-tjln9"] Dec 13 08:59:29 crc kubenswrapper[4971]: I1213 08:59:29.611679 4971 scope.go:117] "RemoveContainer" containerID="92b67aca617672326265647a15a521181507af70709717a835edf609337dceb2" Dec 13 08:59:29 crc kubenswrapper[4971]: E1213 08:59:29.614038 4971 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"92b67aca617672326265647a15a521181507af70709717a835edf609337dceb2\": container with ID starting with 92b67aca617672326265647a15a521181507af70709717a835edf609337dceb2 not found: ID does not exist" containerID="92b67aca617672326265647a15a521181507af70709717a835edf609337dceb2" Dec 13 08:59:29 crc kubenswrapper[4971]: I1213 08:59:29.614105 4971 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"92b67aca617672326265647a15a521181507af70709717a835edf609337dceb2"} err="failed to get container status \"92b67aca617672326265647a15a521181507af70709717a835edf609337dceb2\": rpc error: code = NotFound desc = could not find container \"92b67aca617672326265647a15a521181507af70709717a835edf609337dceb2\": container with ID starting with 92b67aca617672326265647a15a521181507af70709717a835edf609337dceb2 not found: ID does not exist" Dec 13 08:59:29 crc kubenswrapper[4971]: I1213 08:59:29.614149 4971 scope.go:117] "RemoveContainer" containerID="81d175237e71407aeb5af8443277901378ec6e513bff7968d53b9359a9f1bd46" Dec 13 08:59:29 crc kubenswrapper[4971]: E1213 08:59:29.614476 4971 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"81d175237e71407aeb5af8443277901378ec6e513bff7968d53b9359a9f1bd46\": container with ID starting with 81d175237e71407aeb5af8443277901378ec6e513bff7968d53b9359a9f1bd46 not found: ID does not exist" containerID="81d175237e71407aeb5af8443277901378ec6e513bff7968d53b9359a9f1bd46" Dec 13 08:59:29 crc kubenswrapper[4971]: I1213 08:59:29.614519 4971 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"81d175237e71407aeb5af8443277901378ec6e513bff7968d53b9359a9f1bd46"} err="failed to get container status \"81d175237e71407aeb5af8443277901378ec6e513bff7968d53b9359a9f1bd46\": rpc error: code = NotFound desc = could not find container \"81d175237e71407aeb5af8443277901378ec6e513bff7968d53b9359a9f1bd46\": container with ID starting with 81d175237e71407aeb5af8443277901378ec6e513bff7968d53b9359a9f1bd46 not found: ID does not exist" Dec 13 08:59:29 crc kubenswrapper[4971]: I1213 08:59:29.614553 4971 scope.go:117] "RemoveContainer" containerID="d43d1b4fb8986d897439cb8e7b1d8624f35a7dd98b8129f1066754c2adfd0ead" Dec 13 08:59:29 crc kubenswrapper[4971]: E1213 08:59:29.614902 4971 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"d43d1b4fb8986d897439cb8e7b1d8624f35a7dd98b8129f1066754c2adfd0ead\": container with ID starting with d43d1b4fb8986d897439cb8e7b1d8624f35a7dd98b8129f1066754c2adfd0ead not found: ID does not exist" containerID="d43d1b4fb8986d897439cb8e7b1d8624f35a7dd98b8129f1066754c2adfd0ead" Dec 13 08:59:29 crc kubenswrapper[4971]: I1213 08:59:29.614925 4971 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"d43d1b4fb8986d897439cb8e7b1d8624f35a7dd98b8129f1066754c2adfd0ead"} err="failed to get container status \"d43d1b4fb8986d897439cb8e7b1d8624f35a7dd98b8129f1066754c2adfd0ead\": rpc error: code = NotFound desc = could not find container \"d43d1b4fb8986d897439cb8e7b1d8624f35a7dd98b8129f1066754c2adfd0ead\": container with ID starting with d43d1b4fb8986d897439cb8e7b1d8624f35a7dd98b8129f1066754c2adfd0ead not found: ID does not exist" Dec 13 08:59:29 crc kubenswrapper[4971]: I1213 08:59:29.785562 4971 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="7cad90cc-49d3-455b-ac98-8f95077388b3" path="/var/lib/kubelet/pods/7cad90cc-49d3-455b-ac98-8f95077388b3/volumes" Dec 13 08:59:30 crc kubenswrapper[4971]: I1213 08:59:30.876061 4971 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-f4tdk"] Dec 13 08:59:30 crc kubenswrapper[4971]: I1213 08:59:30.876648 4971 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-operators-f4tdk" podUID="5b22775d-1694-4361-be84-144aa19aa2d1" containerName="registry-server" containerID="cri-o://30a5ae8ed75d3ac7a83e7695d485b88f1f707159e224509c8405a85633a645a3" gracePeriod=2 Dec 13 08:59:31 crc kubenswrapper[4971]: I1213 08:59:31.496414 4971 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-f4tdk" Dec 13 08:59:31 crc kubenswrapper[4971]: I1213 08:59:31.498115 4971 generic.go:334] "Generic (PLEG): container finished" podID="5b22775d-1694-4361-be84-144aa19aa2d1" containerID="30a5ae8ed75d3ac7a83e7695d485b88f1f707159e224509c8405a85633a645a3" exitCode=0 Dec 13 08:59:31 crc kubenswrapper[4971]: I1213 08:59:31.498216 4971 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-f4tdk" event={"ID":"5b22775d-1694-4361-be84-144aa19aa2d1","Type":"ContainerDied","Data":"30a5ae8ed75d3ac7a83e7695d485b88f1f707159e224509c8405a85633a645a3"} Dec 13 08:59:31 crc kubenswrapper[4971]: I1213 08:59:31.498270 4971 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-f4tdk" event={"ID":"5b22775d-1694-4361-be84-144aa19aa2d1","Type":"ContainerDied","Data":"2cef81624a4f2296e31f3350369d31ad0f6b25e8ed22f4a0fd481123006a8260"} Dec 13 08:59:31 crc kubenswrapper[4971]: I1213 08:59:31.498290 4971 scope.go:117] "RemoveContainer" containerID="30a5ae8ed75d3ac7a83e7695d485b88f1f707159e224509c8405a85633a645a3" Dec 13 08:59:31 crc kubenswrapper[4971]: I1213 08:59:31.524172 4971 scope.go:117] "RemoveContainer" containerID="211a0a1f5a37a1590dddaac9baec10ed76239734bbf636c98faba7e9dcd4fd63" Dec 13 08:59:31 crc kubenswrapper[4971]: I1213 08:59:31.556587 4971 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/5b22775d-1694-4361-be84-144aa19aa2d1-catalog-content\") pod \"5b22775d-1694-4361-be84-144aa19aa2d1\" (UID: \"5b22775d-1694-4361-be84-144aa19aa2d1\") " Dec 13 08:59:31 crc kubenswrapper[4971]: I1213 08:59:31.556708 4971 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/5b22775d-1694-4361-be84-144aa19aa2d1-utilities\") pod \"5b22775d-1694-4361-be84-144aa19aa2d1\" (UID: \"5b22775d-1694-4361-be84-144aa19aa2d1\") " Dec 13 08:59:31 crc kubenswrapper[4971]: I1213 08:59:31.556945 4971 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-mpbgz\" (UniqueName: \"kubernetes.io/projected/5b22775d-1694-4361-be84-144aa19aa2d1-kube-api-access-mpbgz\") pod \"5b22775d-1694-4361-be84-144aa19aa2d1\" (UID: \"5b22775d-1694-4361-be84-144aa19aa2d1\") " Dec 13 08:59:31 crc kubenswrapper[4971]: I1213 08:59:31.559635 4971 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/5b22775d-1694-4361-be84-144aa19aa2d1-utilities" (OuterVolumeSpecName: "utilities") pod "5b22775d-1694-4361-be84-144aa19aa2d1" (UID: "5b22775d-1694-4361-be84-144aa19aa2d1"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 13 08:59:31 crc kubenswrapper[4971]: I1213 08:59:31.561978 4971 scope.go:117] "RemoveContainer" containerID="06c02335597f23f902be3c13f3da3948e9cb1b31dff52479770d821db84fe94d" Dec 13 08:59:31 crc kubenswrapper[4971]: I1213 08:59:31.571856 4971 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/5b22775d-1694-4361-be84-144aa19aa2d1-kube-api-access-mpbgz" (OuterVolumeSpecName: "kube-api-access-mpbgz") pod "5b22775d-1694-4361-be84-144aa19aa2d1" (UID: "5b22775d-1694-4361-be84-144aa19aa2d1"). InnerVolumeSpecName "kube-api-access-mpbgz". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 13 08:59:31 crc kubenswrapper[4971]: I1213 08:59:31.646967 4971 scope.go:117] "RemoveContainer" containerID="30a5ae8ed75d3ac7a83e7695d485b88f1f707159e224509c8405a85633a645a3" Dec 13 08:59:31 crc kubenswrapper[4971]: E1213 08:59:31.648297 4971 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"30a5ae8ed75d3ac7a83e7695d485b88f1f707159e224509c8405a85633a645a3\": container with ID starting with 30a5ae8ed75d3ac7a83e7695d485b88f1f707159e224509c8405a85633a645a3 not found: ID does not exist" containerID="30a5ae8ed75d3ac7a83e7695d485b88f1f707159e224509c8405a85633a645a3" Dec 13 08:59:31 crc kubenswrapper[4971]: I1213 08:59:31.648742 4971 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"30a5ae8ed75d3ac7a83e7695d485b88f1f707159e224509c8405a85633a645a3"} err="failed to get container status \"30a5ae8ed75d3ac7a83e7695d485b88f1f707159e224509c8405a85633a645a3\": rpc error: code = NotFound desc = could not find container \"30a5ae8ed75d3ac7a83e7695d485b88f1f707159e224509c8405a85633a645a3\": container with ID starting with 30a5ae8ed75d3ac7a83e7695d485b88f1f707159e224509c8405a85633a645a3 not found: ID does not exist" Dec 13 08:59:31 crc kubenswrapper[4971]: I1213 08:59:31.648778 4971 scope.go:117] "RemoveContainer" containerID="211a0a1f5a37a1590dddaac9baec10ed76239734bbf636c98faba7e9dcd4fd63" Dec 13 08:59:31 crc kubenswrapper[4971]: E1213 08:59:31.649097 4971 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"211a0a1f5a37a1590dddaac9baec10ed76239734bbf636c98faba7e9dcd4fd63\": container with ID starting with 211a0a1f5a37a1590dddaac9baec10ed76239734bbf636c98faba7e9dcd4fd63 not found: ID does not exist" containerID="211a0a1f5a37a1590dddaac9baec10ed76239734bbf636c98faba7e9dcd4fd63" Dec 13 08:59:31 crc kubenswrapper[4971]: I1213 08:59:31.649126 4971 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"211a0a1f5a37a1590dddaac9baec10ed76239734bbf636c98faba7e9dcd4fd63"} err="failed to get container status \"211a0a1f5a37a1590dddaac9baec10ed76239734bbf636c98faba7e9dcd4fd63\": rpc error: code = NotFound desc = could not find container \"211a0a1f5a37a1590dddaac9baec10ed76239734bbf636c98faba7e9dcd4fd63\": container with ID starting with 211a0a1f5a37a1590dddaac9baec10ed76239734bbf636c98faba7e9dcd4fd63 not found: ID does not exist" Dec 13 08:59:31 crc kubenswrapper[4971]: I1213 08:59:31.649141 4971 scope.go:117] "RemoveContainer" containerID="06c02335597f23f902be3c13f3da3948e9cb1b31dff52479770d821db84fe94d" Dec 13 08:59:31 crc kubenswrapper[4971]: E1213 08:59:31.651804 4971 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"06c02335597f23f902be3c13f3da3948e9cb1b31dff52479770d821db84fe94d\": container with ID starting with 06c02335597f23f902be3c13f3da3948e9cb1b31dff52479770d821db84fe94d not found: ID does not exist" containerID="06c02335597f23f902be3c13f3da3948e9cb1b31dff52479770d821db84fe94d" Dec 13 08:59:31 crc kubenswrapper[4971]: I1213 08:59:31.651864 4971 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"06c02335597f23f902be3c13f3da3948e9cb1b31dff52479770d821db84fe94d"} err="failed to get container status \"06c02335597f23f902be3c13f3da3948e9cb1b31dff52479770d821db84fe94d\": rpc error: code = NotFound desc = could not find container \"06c02335597f23f902be3c13f3da3948e9cb1b31dff52479770d821db84fe94d\": container with ID starting with 06c02335597f23f902be3c13f3da3948e9cb1b31dff52479770d821db84fe94d not found: ID does not exist" Dec 13 08:59:31 crc kubenswrapper[4971]: I1213 08:59:31.660356 4971 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-mpbgz\" (UniqueName: \"kubernetes.io/projected/5b22775d-1694-4361-be84-144aa19aa2d1-kube-api-access-mpbgz\") on node \"crc\" DevicePath \"\"" Dec 13 08:59:31 crc kubenswrapper[4971]: I1213 08:59:31.660392 4971 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/5b22775d-1694-4361-be84-144aa19aa2d1-utilities\") on node \"crc\" DevicePath \"\"" Dec 13 08:59:31 crc kubenswrapper[4971]: I1213 08:59:31.710109 4971 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/5b22775d-1694-4361-be84-144aa19aa2d1-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "5b22775d-1694-4361-be84-144aa19aa2d1" (UID: "5b22775d-1694-4361-be84-144aa19aa2d1"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 13 08:59:31 crc kubenswrapper[4971]: I1213 08:59:31.763063 4971 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/5b22775d-1694-4361-be84-144aa19aa2d1-catalog-content\") on node \"crc\" DevicePath \"\"" Dec 13 08:59:32 crc kubenswrapper[4971]: I1213 08:59:32.512069 4971 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-f4tdk" Dec 13 08:59:32 crc kubenswrapper[4971]: I1213 08:59:32.550083 4971 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-f4tdk"] Dec 13 08:59:32 crc kubenswrapper[4971]: I1213 08:59:32.561419 4971 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-operators-f4tdk"] Dec 13 08:59:33 crc kubenswrapper[4971]: I1213 08:59:33.783289 4971 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="5b22775d-1694-4361-be84-144aa19aa2d1" path="/var/lib/kubelet/pods/5b22775d-1694-4361-be84-144aa19aa2d1/volumes" Dec 13 08:59:46 crc kubenswrapper[4971]: I1213 08:59:46.153845 4971 patch_prober.go:28] interesting pod/machine-config-daemon-82xjz container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 13 08:59:46 crc kubenswrapper[4971]: I1213 08:59:46.154840 4971 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-82xjz" podUID="e0c4af12-edda-41d7-9b44-f87396174ef2" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 13 08:59:46 crc kubenswrapper[4971]: I1213 08:59:46.154916 4971 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-82xjz" Dec 13 08:59:46 crc kubenswrapper[4971]: I1213 08:59:46.156313 4971 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"95d975a85829af1f5477ac9667504f9ab79ab97ddaabd918334d5b4218e3d164"} pod="openshift-machine-config-operator/machine-config-daemon-82xjz" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Dec 13 08:59:46 crc kubenswrapper[4971]: I1213 08:59:46.156408 4971 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-82xjz" podUID="e0c4af12-edda-41d7-9b44-f87396174ef2" containerName="machine-config-daemon" containerID="cri-o://95d975a85829af1f5477ac9667504f9ab79ab97ddaabd918334d5b4218e3d164" gracePeriod=600 Dec 13 08:59:46 crc kubenswrapper[4971]: E1213 08:59:46.313289 4971 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-82xjz_openshift-machine-config-operator(e0c4af12-edda-41d7-9b44-f87396174ef2)\"" pod="openshift-machine-config-operator/machine-config-daemon-82xjz" podUID="e0c4af12-edda-41d7-9b44-f87396174ef2" Dec 13 08:59:46 crc kubenswrapper[4971]: I1213 08:59:46.674016 4971 generic.go:334] "Generic (PLEG): container finished" podID="e0c4af12-edda-41d7-9b44-f87396174ef2" containerID="95d975a85829af1f5477ac9667504f9ab79ab97ddaabd918334d5b4218e3d164" exitCode=0 Dec 13 08:59:46 crc kubenswrapper[4971]: I1213 08:59:46.674111 4971 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-82xjz" event={"ID":"e0c4af12-edda-41d7-9b44-f87396174ef2","Type":"ContainerDied","Data":"95d975a85829af1f5477ac9667504f9ab79ab97ddaabd918334d5b4218e3d164"} Dec 13 08:59:46 crc kubenswrapper[4971]: I1213 08:59:46.674247 4971 scope.go:117] "RemoveContainer" containerID="3f281a9adda16a9b53c490949e166088ac22b2347e901a634eebb681b7729981" Dec 13 08:59:46 crc kubenswrapper[4971]: I1213 08:59:46.674924 4971 scope.go:117] "RemoveContainer" containerID="95d975a85829af1f5477ac9667504f9ab79ab97ddaabd918334d5b4218e3d164" Dec 13 08:59:46 crc kubenswrapper[4971]: E1213 08:59:46.675305 4971 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-82xjz_openshift-machine-config-operator(e0c4af12-edda-41d7-9b44-f87396174ef2)\"" pod="openshift-machine-config-operator/machine-config-daemon-82xjz" podUID="e0c4af12-edda-41d7-9b44-f87396174ef2" Dec 13 08:59:57 crc kubenswrapper[4971]: I1213 08:59:57.769681 4971 scope.go:117] "RemoveContainer" containerID="95d975a85829af1f5477ac9667504f9ab79ab97ddaabd918334d5b4218e3d164" Dec 13 08:59:57 crc kubenswrapper[4971]: E1213 08:59:57.770727 4971 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-82xjz_openshift-machine-config-operator(e0c4af12-edda-41d7-9b44-f87396174ef2)\"" pod="openshift-machine-config-operator/machine-config-daemon-82xjz" podUID="e0c4af12-edda-41d7-9b44-f87396174ef2" Dec 13 09:00:00 crc kubenswrapper[4971]: I1213 09:00:00.199197 4971 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29426940-m6wck"] Dec 13 09:00:00 crc kubenswrapper[4971]: E1213 09:00:00.200370 4971 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="bf2757ef-bc9b-4a59-bfaa-083d5781203a" containerName="extract-content" Dec 13 09:00:00 crc kubenswrapper[4971]: I1213 09:00:00.200393 4971 state_mem.go:107] "Deleted CPUSet assignment" podUID="bf2757ef-bc9b-4a59-bfaa-083d5781203a" containerName="extract-content" Dec 13 09:00:00 crc kubenswrapper[4971]: E1213 09:00:00.200419 4971 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="7cad90cc-49d3-455b-ac98-8f95077388b3" containerName="registry-server" Dec 13 09:00:00 crc kubenswrapper[4971]: I1213 09:00:00.200427 4971 state_mem.go:107] "Deleted CPUSet assignment" podUID="7cad90cc-49d3-455b-ac98-8f95077388b3" containerName="registry-server" Dec 13 09:00:00 crc kubenswrapper[4971]: E1213 09:00:00.200443 4971 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="5b22775d-1694-4361-be84-144aa19aa2d1" containerName="extract-content" Dec 13 09:00:00 crc kubenswrapper[4971]: I1213 09:00:00.200452 4971 state_mem.go:107] "Deleted CPUSet assignment" podUID="5b22775d-1694-4361-be84-144aa19aa2d1" containerName="extract-content" Dec 13 09:00:00 crc kubenswrapper[4971]: E1213 09:00:00.200462 4971 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="7cad90cc-49d3-455b-ac98-8f95077388b3" containerName="extract-content" Dec 13 09:00:00 crc kubenswrapper[4971]: I1213 09:00:00.200469 4971 state_mem.go:107] "Deleted CPUSet assignment" podUID="7cad90cc-49d3-455b-ac98-8f95077388b3" containerName="extract-content" Dec 13 09:00:00 crc kubenswrapper[4971]: E1213 09:00:00.200480 4971 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="5b22775d-1694-4361-be84-144aa19aa2d1" containerName="extract-utilities" Dec 13 09:00:00 crc kubenswrapper[4971]: I1213 09:00:00.200489 4971 state_mem.go:107] "Deleted CPUSet assignment" podUID="5b22775d-1694-4361-be84-144aa19aa2d1" containerName="extract-utilities" Dec 13 09:00:00 crc kubenswrapper[4971]: E1213 09:00:00.200517 4971 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="bf2757ef-bc9b-4a59-bfaa-083d5781203a" containerName="registry-server" Dec 13 09:00:00 crc kubenswrapper[4971]: I1213 09:00:00.200627 4971 state_mem.go:107] "Deleted CPUSet assignment" podUID="bf2757ef-bc9b-4a59-bfaa-083d5781203a" containerName="registry-server" Dec 13 09:00:00 crc kubenswrapper[4971]: E1213 09:00:00.200646 4971 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="7cad90cc-49d3-455b-ac98-8f95077388b3" containerName="extract-utilities" Dec 13 09:00:00 crc kubenswrapper[4971]: I1213 09:00:00.200656 4971 state_mem.go:107] "Deleted CPUSet assignment" podUID="7cad90cc-49d3-455b-ac98-8f95077388b3" containerName="extract-utilities" Dec 13 09:00:00 crc kubenswrapper[4971]: E1213 09:00:00.200685 4971 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="bf2757ef-bc9b-4a59-bfaa-083d5781203a" containerName="extract-utilities" Dec 13 09:00:00 crc kubenswrapper[4971]: I1213 09:00:00.200695 4971 state_mem.go:107] "Deleted CPUSet assignment" podUID="bf2757ef-bc9b-4a59-bfaa-083d5781203a" containerName="extract-utilities" Dec 13 09:00:00 crc kubenswrapper[4971]: E1213 09:00:00.200717 4971 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="5b22775d-1694-4361-be84-144aa19aa2d1" containerName="registry-server" Dec 13 09:00:00 crc kubenswrapper[4971]: I1213 09:00:00.200725 4971 state_mem.go:107] "Deleted CPUSet assignment" podUID="5b22775d-1694-4361-be84-144aa19aa2d1" containerName="registry-server" Dec 13 09:00:00 crc kubenswrapper[4971]: I1213 09:00:00.201165 4971 memory_manager.go:354] "RemoveStaleState removing state" podUID="7cad90cc-49d3-455b-ac98-8f95077388b3" containerName="registry-server" Dec 13 09:00:00 crc kubenswrapper[4971]: I1213 09:00:00.201195 4971 memory_manager.go:354] "RemoveStaleState removing state" podUID="5b22775d-1694-4361-be84-144aa19aa2d1" containerName="registry-server" Dec 13 09:00:00 crc kubenswrapper[4971]: I1213 09:00:00.201236 4971 memory_manager.go:354] "RemoveStaleState removing state" podUID="bf2757ef-bc9b-4a59-bfaa-083d5781203a" containerName="registry-server" Dec 13 09:00:00 crc kubenswrapper[4971]: I1213 09:00:00.202336 4971 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29426940-m6wck"] Dec 13 09:00:00 crc kubenswrapper[4971]: I1213 09:00:00.202461 4971 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29426940-m6wck" Dec 13 09:00:00 crc kubenswrapper[4971]: I1213 09:00:00.205666 4971 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-operator-lifecycle-manager"/"collect-profiles-config" Dec 13 09:00:00 crc kubenswrapper[4971]: I1213 09:00:00.205993 4971 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"collect-profiles-dockercfg-kzf4t" Dec 13 09:00:00 crc kubenswrapper[4971]: I1213 09:00:00.324087 4971 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/a57afece-3e8b-4b98-80bb-437bc66f2469-secret-volume\") pod \"collect-profiles-29426940-m6wck\" (UID: \"a57afece-3e8b-4b98-80bb-437bc66f2469\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29426940-m6wck" Dec 13 09:00:00 crc kubenswrapper[4971]: I1213 09:00:00.324269 4971 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-9zhmj\" (UniqueName: \"kubernetes.io/projected/a57afece-3e8b-4b98-80bb-437bc66f2469-kube-api-access-9zhmj\") pod \"collect-profiles-29426940-m6wck\" (UID: \"a57afece-3e8b-4b98-80bb-437bc66f2469\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29426940-m6wck" Dec 13 09:00:00 crc kubenswrapper[4971]: I1213 09:00:00.324324 4971 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/a57afece-3e8b-4b98-80bb-437bc66f2469-config-volume\") pod \"collect-profiles-29426940-m6wck\" (UID: \"a57afece-3e8b-4b98-80bb-437bc66f2469\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29426940-m6wck" Dec 13 09:00:00 crc kubenswrapper[4971]: I1213 09:00:00.426715 4971 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-9zhmj\" (UniqueName: \"kubernetes.io/projected/a57afece-3e8b-4b98-80bb-437bc66f2469-kube-api-access-9zhmj\") pod \"collect-profiles-29426940-m6wck\" (UID: \"a57afece-3e8b-4b98-80bb-437bc66f2469\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29426940-m6wck" Dec 13 09:00:00 crc kubenswrapper[4971]: I1213 09:00:00.426813 4971 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/a57afece-3e8b-4b98-80bb-437bc66f2469-config-volume\") pod \"collect-profiles-29426940-m6wck\" (UID: \"a57afece-3e8b-4b98-80bb-437bc66f2469\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29426940-m6wck" Dec 13 09:00:00 crc kubenswrapper[4971]: I1213 09:00:00.426933 4971 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/a57afece-3e8b-4b98-80bb-437bc66f2469-secret-volume\") pod \"collect-profiles-29426940-m6wck\" (UID: \"a57afece-3e8b-4b98-80bb-437bc66f2469\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29426940-m6wck" Dec 13 09:00:00 crc kubenswrapper[4971]: I1213 09:00:00.428060 4971 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/a57afece-3e8b-4b98-80bb-437bc66f2469-config-volume\") pod \"collect-profiles-29426940-m6wck\" (UID: \"a57afece-3e8b-4b98-80bb-437bc66f2469\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29426940-m6wck" Dec 13 09:00:00 crc kubenswrapper[4971]: I1213 09:00:00.442198 4971 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/a57afece-3e8b-4b98-80bb-437bc66f2469-secret-volume\") pod \"collect-profiles-29426940-m6wck\" (UID: \"a57afece-3e8b-4b98-80bb-437bc66f2469\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29426940-m6wck" Dec 13 09:00:00 crc kubenswrapper[4971]: I1213 09:00:00.448506 4971 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-9zhmj\" (UniqueName: \"kubernetes.io/projected/a57afece-3e8b-4b98-80bb-437bc66f2469-kube-api-access-9zhmj\") pod \"collect-profiles-29426940-m6wck\" (UID: \"a57afece-3e8b-4b98-80bb-437bc66f2469\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29426940-m6wck" Dec 13 09:00:00 crc kubenswrapper[4971]: I1213 09:00:00.552929 4971 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29426940-m6wck" Dec 13 09:00:01 crc kubenswrapper[4971]: I1213 09:00:01.028176 4971 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29426940-m6wck"] Dec 13 09:00:01 crc kubenswrapper[4971]: I1213 09:00:01.367734 4971 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29426940-m6wck" event={"ID":"a57afece-3e8b-4b98-80bb-437bc66f2469","Type":"ContainerStarted","Data":"58d28af04ead826f4260e6960d316420735a77176a9509a70a342d7e593bc519"} Dec 13 09:00:01 crc kubenswrapper[4971]: I1213 09:00:01.368184 4971 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29426940-m6wck" event={"ID":"a57afece-3e8b-4b98-80bb-437bc66f2469","Type":"ContainerStarted","Data":"3ec3d4e265428e6fa4d635ab743a72fb3380ba7103796bebc109263c9dfd9d05"} Dec 13 09:00:01 crc kubenswrapper[4971]: I1213 09:00:01.394702 4971 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-operator-lifecycle-manager/collect-profiles-29426940-m6wck" podStartSLOduration=1.394665126 podStartE2EDuration="1.394665126s" podCreationTimestamp="2025-12-13 09:00:00 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-13 09:00:01.388947195 +0000 UTC m=+7857.993356643" watchObservedRunningTime="2025-12-13 09:00:01.394665126 +0000 UTC m=+7857.999074574" Dec 13 09:00:02 crc kubenswrapper[4971]: I1213 09:00:02.383649 4971 generic.go:334] "Generic (PLEG): container finished" podID="a57afece-3e8b-4b98-80bb-437bc66f2469" containerID="58d28af04ead826f4260e6960d316420735a77176a9509a70a342d7e593bc519" exitCode=0 Dec 13 09:00:02 crc kubenswrapper[4971]: I1213 09:00:02.383770 4971 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29426940-m6wck" event={"ID":"a57afece-3e8b-4b98-80bb-437bc66f2469","Type":"ContainerDied","Data":"58d28af04ead826f4260e6960d316420735a77176a9509a70a342d7e593bc519"} Dec 13 09:00:03 crc kubenswrapper[4971]: I1213 09:00:03.901339 4971 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29426940-m6wck" Dec 13 09:00:04 crc kubenswrapper[4971]: I1213 09:00:04.017034 4971 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/a57afece-3e8b-4b98-80bb-437bc66f2469-secret-volume\") pod \"a57afece-3e8b-4b98-80bb-437bc66f2469\" (UID: \"a57afece-3e8b-4b98-80bb-437bc66f2469\") " Dec 13 09:00:04 crc kubenswrapper[4971]: I1213 09:00:04.017152 4971 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/a57afece-3e8b-4b98-80bb-437bc66f2469-config-volume\") pod \"a57afece-3e8b-4b98-80bb-437bc66f2469\" (UID: \"a57afece-3e8b-4b98-80bb-437bc66f2469\") " Dec 13 09:00:04 crc kubenswrapper[4971]: I1213 09:00:04.017364 4971 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-9zhmj\" (UniqueName: \"kubernetes.io/projected/a57afece-3e8b-4b98-80bb-437bc66f2469-kube-api-access-9zhmj\") pod \"a57afece-3e8b-4b98-80bb-437bc66f2469\" (UID: \"a57afece-3e8b-4b98-80bb-437bc66f2469\") " Dec 13 09:00:04 crc kubenswrapper[4971]: I1213 09:00:04.018290 4971 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/a57afece-3e8b-4b98-80bb-437bc66f2469-config-volume" (OuterVolumeSpecName: "config-volume") pod "a57afece-3e8b-4b98-80bb-437bc66f2469" (UID: "a57afece-3e8b-4b98-80bb-437bc66f2469"). InnerVolumeSpecName "config-volume". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 13 09:00:04 crc kubenswrapper[4971]: I1213 09:00:04.024078 4971 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/a57afece-3e8b-4b98-80bb-437bc66f2469-kube-api-access-9zhmj" (OuterVolumeSpecName: "kube-api-access-9zhmj") pod "a57afece-3e8b-4b98-80bb-437bc66f2469" (UID: "a57afece-3e8b-4b98-80bb-437bc66f2469"). InnerVolumeSpecName "kube-api-access-9zhmj". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 13 09:00:04 crc kubenswrapper[4971]: I1213 09:00:04.026445 4971 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/a57afece-3e8b-4b98-80bb-437bc66f2469-secret-volume" (OuterVolumeSpecName: "secret-volume") pod "a57afece-3e8b-4b98-80bb-437bc66f2469" (UID: "a57afece-3e8b-4b98-80bb-437bc66f2469"). InnerVolumeSpecName "secret-volume". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 13 09:00:04 crc kubenswrapper[4971]: I1213 09:00:04.121321 4971 reconciler_common.go:293] "Volume detached for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/a57afece-3e8b-4b98-80bb-437bc66f2469-secret-volume\") on node \"crc\" DevicePath \"\"" Dec 13 09:00:04 crc kubenswrapper[4971]: I1213 09:00:04.121383 4971 reconciler_common.go:293] "Volume detached for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/a57afece-3e8b-4b98-80bb-437bc66f2469-config-volume\") on node \"crc\" DevicePath \"\"" Dec 13 09:00:04 crc kubenswrapper[4971]: I1213 09:00:04.121400 4971 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-9zhmj\" (UniqueName: \"kubernetes.io/projected/a57afece-3e8b-4b98-80bb-437bc66f2469-kube-api-access-9zhmj\") on node \"crc\" DevicePath \"\"" Dec 13 09:00:04 crc kubenswrapper[4971]: I1213 09:00:04.411477 4971 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29426940-m6wck" event={"ID":"a57afece-3e8b-4b98-80bb-437bc66f2469","Type":"ContainerDied","Data":"3ec3d4e265428e6fa4d635ab743a72fb3380ba7103796bebc109263c9dfd9d05"} Dec 13 09:00:04 crc kubenswrapper[4971]: I1213 09:00:04.411547 4971 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="3ec3d4e265428e6fa4d635ab743a72fb3380ba7103796bebc109263c9dfd9d05" Dec 13 09:00:04 crc kubenswrapper[4971]: I1213 09:00:04.411631 4971 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29426940-m6wck" Dec 13 09:00:04 crc kubenswrapper[4971]: I1213 09:00:04.997529 4971 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29426895-2sc8g"] Dec 13 09:00:05 crc kubenswrapper[4971]: I1213 09:00:05.012673 4971 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29426895-2sc8g"] Dec 13 09:00:05 crc kubenswrapper[4971]: I1213 09:00:05.788063 4971 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="ab8b4298-8945-4f7f-aa70-1d08d1a7b91c" path="/var/lib/kubelet/pods/ab8b4298-8945-4f7f-aa70-1d08d1a7b91c/volumes" Dec 13 09:00:12 crc kubenswrapper[4971]: I1213 09:00:12.769179 4971 scope.go:117] "RemoveContainer" containerID="95d975a85829af1f5477ac9667504f9ab79ab97ddaabd918334d5b4218e3d164" Dec 13 09:00:12 crc kubenswrapper[4971]: E1213 09:00:12.770923 4971 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-82xjz_openshift-machine-config-operator(e0c4af12-edda-41d7-9b44-f87396174ef2)\"" pod="openshift-machine-config-operator/machine-config-daemon-82xjz" podUID="e0c4af12-edda-41d7-9b44-f87396174ef2" Dec 13 09:00:25 crc kubenswrapper[4971]: I1213 09:00:25.769663 4971 scope.go:117] "RemoveContainer" containerID="95d975a85829af1f5477ac9667504f9ab79ab97ddaabd918334d5b4218e3d164" Dec 13 09:00:25 crc kubenswrapper[4971]: E1213 09:00:25.771879 4971 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-82xjz_openshift-machine-config-operator(e0c4af12-edda-41d7-9b44-f87396174ef2)\"" pod="openshift-machine-config-operator/machine-config-daemon-82xjz" podUID="e0c4af12-edda-41d7-9b44-f87396174ef2" Dec 13 09:00:28 crc kubenswrapper[4971]: I1213 09:00:28.764498 4971 scope.go:117] "RemoveContainer" containerID="eda1293f48d70f7f701282611259ef794a7a56c71da9779a2716decce6d7d261" Dec 13 09:00:36 crc kubenswrapper[4971]: I1213 09:00:36.770046 4971 scope.go:117] "RemoveContainer" containerID="95d975a85829af1f5477ac9667504f9ab79ab97ddaabd918334d5b4218e3d164" Dec 13 09:00:36 crc kubenswrapper[4971]: E1213 09:00:36.770935 4971 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-82xjz_openshift-machine-config-operator(e0c4af12-edda-41d7-9b44-f87396174ef2)\"" pod="openshift-machine-config-operator/machine-config-daemon-82xjz" podUID="e0c4af12-edda-41d7-9b44-f87396174ef2" Dec 13 09:00:49 crc kubenswrapper[4971]: I1213 09:00:49.769566 4971 scope.go:117] "RemoveContainer" containerID="95d975a85829af1f5477ac9667504f9ab79ab97ddaabd918334d5b4218e3d164" Dec 13 09:00:49 crc kubenswrapper[4971]: E1213 09:00:49.771853 4971 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-82xjz_openshift-machine-config-operator(e0c4af12-edda-41d7-9b44-f87396174ef2)\"" pod="openshift-machine-config-operator/machine-config-daemon-82xjz" podUID="e0c4af12-edda-41d7-9b44-f87396174ef2" Dec 13 09:01:00 crc kubenswrapper[4971]: I1213 09:01:00.180200 4971 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/keystone-cron-29426941-c6mj5"] Dec 13 09:01:00 crc kubenswrapper[4971]: E1213 09:01:00.183276 4971 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="a57afece-3e8b-4b98-80bb-437bc66f2469" containerName="collect-profiles" Dec 13 09:01:00 crc kubenswrapper[4971]: I1213 09:01:00.183304 4971 state_mem.go:107] "Deleted CPUSet assignment" podUID="a57afece-3e8b-4b98-80bb-437bc66f2469" containerName="collect-profiles" Dec 13 09:01:00 crc kubenswrapper[4971]: I1213 09:01:00.183578 4971 memory_manager.go:354] "RemoveStaleState removing state" podUID="a57afece-3e8b-4b98-80bb-437bc66f2469" containerName="collect-profiles" Dec 13 09:01:00 crc kubenswrapper[4971]: I1213 09:01:00.184475 4971 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-cron-29426941-c6mj5" Dec 13 09:01:00 crc kubenswrapper[4971]: I1213 09:01:00.194363 4971 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/keystone-cron-29426941-c6mj5"] Dec 13 09:01:00 crc kubenswrapper[4971]: I1213 09:01:00.365238 4971 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/a7d15eb0-4473-488c-8355-db1a4d139e7c-combined-ca-bundle\") pod \"keystone-cron-29426941-c6mj5\" (UID: \"a7d15eb0-4473-488c-8355-db1a4d139e7c\") " pod="openstack/keystone-cron-29426941-c6mj5" Dec 13 09:01:00 crc kubenswrapper[4971]: I1213 09:01:00.365361 4971 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-ls5qg\" (UniqueName: \"kubernetes.io/projected/a7d15eb0-4473-488c-8355-db1a4d139e7c-kube-api-access-ls5qg\") pod \"keystone-cron-29426941-c6mj5\" (UID: \"a7d15eb0-4473-488c-8355-db1a4d139e7c\") " pod="openstack/keystone-cron-29426941-c6mj5" Dec 13 09:01:00 crc kubenswrapper[4971]: I1213 09:01:00.365415 4971 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/a7d15eb0-4473-488c-8355-db1a4d139e7c-config-data\") pod \"keystone-cron-29426941-c6mj5\" (UID: \"a7d15eb0-4473-488c-8355-db1a4d139e7c\") " pod="openstack/keystone-cron-29426941-c6mj5" Dec 13 09:01:00 crc kubenswrapper[4971]: I1213 09:01:00.365447 4971 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/a7d15eb0-4473-488c-8355-db1a4d139e7c-fernet-keys\") pod \"keystone-cron-29426941-c6mj5\" (UID: \"a7d15eb0-4473-488c-8355-db1a4d139e7c\") " pod="openstack/keystone-cron-29426941-c6mj5" Dec 13 09:01:00 crc kubenswrapper[4971]: I1213 09:01:00.468151 4971 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/a7d15eb0-4473-488c-8355-db1a4d139e7c-combined-ca-bundle\") pod \"keystone-cron-29426941-c6mj5\" (UID: \"a7d15eb0-4473-488c-8355-db1a4d139e7c\") " pod="openstack/keystone-cron-29426941-c6mj5" Dec 13 09:01:00 crc kubenswrapper[4971]: I1213 09:01:00.468242 4971 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-ls5qg\" (UniqueName: \"kubernetes.io/projected/a7d15eb0-4473-488c-8355-db1a4d139e7c-kube-api-access-ls5qg\") pod \"keystone-cron-29426941-c6mj5\" (UID: \"a7d15eb0-4473-488c-8355-db1a4d139e7c\") " pod="openstack/keystone-cron-29426941-c6mj5" Dec 13 09:01:00 crc kubenswrapper[4971]: I1213 09:01:00.468288 4971 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/a7d15eb0-4473-488c-8355-db1a4d139e7c-config-data\") pod \"keystone-cron-29426941-c6mj5\" (UID: \"a7d15eb0-4473-488c-8355-db1a4d139e7c\") " pod="openstack/keystone-cron-29426941-c6mj5" Dec 13 09:01:00 crc kubenswrapper[4971]: I1213 09:01:00.468319 4971 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/a7d15eb0-4473-488c-8355-db1a4d139e7c-fernet-keys\") pod \"keystone-cron-29426941-c6mj5\" (UID: \"a7d15eb0-4473-488c-8355-db1a4d139e7c\") " pod="openstack/keystone-cron-29426941-c6mj5" Dec 13 09:01:00 crc kubenswrapper[4971]: I1213 09:01:00.478507 4971 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/a7d15eb0-4473-488c-8355-db1a4d139e7c-fernet-keys\") pod \"keystone-cron-29426941-c6mj5\" (UID: \"a7d15eb0-4473-488c-8355-db1a4d139e7c\") " pod="openstack/keystone-cron-29426941-c6mj5" Dec 13 09:01:00 crc kubenswrapper[4971]: I1213 09:01:00.478758 4971 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/a7d15eb0-4473-488c-8355-db1a4d139e7c-config-data\") pod \"keystone-cron-29426941-c6mj5\" (UID: \"a7d15eb0-4473-488c-8355-db1a4d139e7c\") " pod="openstack/keystone-cron-29426941-c6mj5" Dec 13 09:01:00 crc kubenswrapper[4971]: I1213 09:01:00.480209 4971 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/a7d15eb0-4473-488c-8355-db1a4d139e7c-combined-ca-bundle\") pod \"keystone-cron-29426941-c6mj5\" (UID: \"a7d15eb0-4473-488c-8355-db1a4d139e7c\") " pod="openstack/keystone-cron-29426941-c6mj5" Dec 13 09:01:00 crc kubenswrapper[4971]: I1213 09:01:00.489431 4971 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-ls5qg\" (UniqueName: \"kubernetes.io/projected/a7d15eb0-4473-488c-8355-db1a4d139e7c-kube-api-access-ls5qg\") pod \"keystone-cron-29426941-c6mj5\" (UID: \"a7d15eb0-4473-488c-8355-db1a4d139e7c\") " pod="openstack/keystone-cron-29426941-c6mj5" Dec 13 09:01:00 crc kubenswrapper[4971]: I1213 09:01:00.552460 4971 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-cron-29426941-c6mj5" Dec 13 09:01:01 crc kubenswrapper[4971]: I1213 09:01:01.055114 4971 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/keystone-cron-29426941-c6mj5"] Dec 13 09:01:02 crc kubenswrapper[4971]: I1213 09:01:02.038383 4971 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-cron-29426941-c6mj5" event={"ID":"a7d15eb0-4473-488c-8355-db1a4d139e7c","Type":"ContainerStarted","Data":"067ba17a361c61ba44f9618cf353ad06a82f0d357fe51956dd1699aab6716a33"} Dec 13 09:01:02 crc kubenswrapper[4971]: I1213 09:01:02.038899 4971 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-cron-29426941-c6mj5" event={"ID":"a7d15eb0-4473-488c-8355-db1a4d139e7c","Type":"ContainerStarted","Data":"dc4462cbde184ad280d1c9993122e790ced7178e9ef7557f4b311829ad5250e7"} Dec 13 09:01:02 crc kubenswrapper[4971]: I1213 09:01:02.067161 4971 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/keystone-cron-29426941-c6mj5" podStartSLOduration=2.067091172 podStartE2EDuration="2.067091172s" podCreationTimestamp="2025-12-13 09:01:00 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-13 09:01:02.061969496 +0000 UTC m=+7918.666378944" watchObservedRunningTime="2025-12-13 09:01:02.067091172 +0000 UTC m=+7918.671500620" Dec 13 09:01:04 crc kubenswrapper[4971]: I1213 09:01:04.769309 4971 scope.go:117] "RemoveContainer" containerID="95d975a85829af1f5477ac9667504f9ab79ab97ddaabd918334d5b4218e3d164" Dec 13 09:01:04 crc kubenswrapper[4971]: E1213 09:01:04.770377 4971 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-82xjz_openshift-machine-config-operator(e0c4af12-edda-41d7-9b44-f87396174ef2)\"" pod="openshift-machine-config-operator/machine-config-daemon-82xjz" podUID="e0c4af12-edda-41d7-9b44-f87396174ef2" Dec 13 09:01:06 crc kubenswrapper[4971]: I1213 09:01:06.086070 4971 generic.go:334] "Generic (PLEG): container finished" podID="a7d15eb0-4473-488c-8355-db1a4d139e7c" containerID="067ba17a361c61ba44f9618cf353ad06a82f0d357fe51956dd1699aab6716a33" exitCode=0 Dec 13 09:01:06 crc kubenswrapper[4971]: I1213 09:01:06.086560 4971 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-cron-29426941-c6mj5" event={"ID":"a7d15eb0-4473-488c-8355-db1a4d139e7c","Type":"ContainerDied","Data":"067ba17a361c61ba44f9618cf353ad06a82f0d357fe51956dd1699aab6716a33"} Dec 13 09:01:07 crc kubenswrapper[4971]: I1213 09:01:07.925443 4971 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-cron-29426941-c6mj5" Dec 13 09:01:08 crc kubenswrapper[4971]: I1213 09:01:08.021229 4971 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-ls5qg\" (UniqueName: \"kubernetes.io/projected/a7d15eb0-4473-488c-8355-db1a4d139e7c-kube-api-access-ls5qg\") pod \"a7d15eb0-4473-488c-8355-db1a4d139e7c\" (UID: \"a7d15eb0-4473-488c-8355-db1a4d139e7c\") " Dec 13 09:01:08 crc kubenswrapper[4971]: I1213 09:01:08.021553 4971 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/a7d15eb0-4473-488c-8355-db1a4d139e7c-combined-ca-bundle\") pod \"a7d15eb0-4473-488c-8355-db1a4d139e7c\" (UID: \"a7d15eb0-4473-488c-8355-db1a4d139e7c\") " Dec 13 09:01:08 crc kubenswrapper[4971]: I1213 09:01:08.021611 4971 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/a7d15eb0-4473-488c-8355-db1a4d139e7c-config-data\") pod \"a7d15eb0-4473-488c-8355-db1a4d139e7c\" (UID: \"a7d15eb0-4473-488c-8355-db1a4d139e7c\") " Dec 13 09:01:08 crc kubenswrapper[4971]: I1213 09:01:08.021642 4971 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/a7d15eb0-4473-488c-8355-db1a4d139e7c-fernet-keys\") pod \"a7d15eb0-4473-488c-8355-db1a4d139e7c\" (UID: \"a7d15eb0-4473-488c-8355-db1a4d139e7c\") " Dec 13 09:01:08 crc kubenswrapper[4971]: I1213 09:01:08.030061 4971 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/a7d15eb0-4473-488c-8355-db1a4d139e7c-fernet-keys" (OuterVolumeSpecName: "fernet-keys") pod "a7d15eb0-4473-488c-8355-db1a4d139e7c" (UID: "a7d15eb0-4473-488c-8355-db1a4d139e7c"). InnerVolumeSpecName "fernet-keys". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 13 09:01:08 crc kubenswrapper[4971]: I1213 09:01:08.035008 4971 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/a7d15eb0-4473-488c-8355-db1a4d139e7c-kube-api-access-ls5qg" (OuterVolumeSpecName: "kube-api-access-ls5qg") pod "a7d15eb0-4473-488c-8355-db1a4d139e7c" (UID: "a7d15eb0-4473-488c-8355-db1a4d139e7c"). InnerVolumeSpecName "kube-api-access-ls5qg". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 13 09:01:08 crc kubenswrapper[4971]: I1213 09:01:08.056173 4971 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/a7d15eb0-4473-488c-8355-db1a4d139e7c-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "a7d15eb0-4473-488c-8355-db1a4d139e7c" (UID: "a7d15eb0-4473-488c-8355-db1a4d139e7c"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 13 09:01:08 crc kubenswrapper[4971]: I1213 09:01:08.097797 4971 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/a7d15eb0-4473-488c-8355-db1a4d139e7c-config-data" (OuterVolumeSpecName: "config-data") pod "a7d15eb0-4473-488c-8355-db1a4d139e7c" (UID: "a7d15eb0-4473-488c-8355-db1a4d139e7c"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 13 09:01:08 crc kubenswrapper[4971]: I1213 09:01:08.125292 4971 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/a7d15eb0-4473-488c-8355-db1a4d139e7c-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 13 09:01:08 crc kubenswrapper[4971]: I1213 09:01:08.125350 4971 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/a7d15eb0-4473-488c-8355-db1a4d139e7c-config-data\") on node \"crc\" DevicePath \"\"" Dec 13 09:01:08 crc kubenswrapper[4971]: I1213 09:01:08.125365 4971 reconciler_common.go:293] "Volume detached for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/a7d15eb0-4473-488c-8355-db1a4d139e7c-fernet-keys\") on node \"crc\" DevicePath \"\"" Dec 13 09:01:08 crc kubenswrapper[4971]: I1213 09:01:08.125377 4971 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-ls5qg\" (UniqueName: \"kubernetes.io/projected/a7d15eb0-4473-488c-8355-db1a4d139e7c-kube-api-access-ls5qg\") on node \"crc\" DevicePath \"\"" Dec 13 09:01:08 crc kubenswrapper[4971]: I1213 09:01:08.363155 4971 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-cron-29426941-c6mj5" event={"ID":"a7d15eb0-4473-488c-8355-db1a4d139e7c","Type":"ContainerDied","Data":"dc4462cbde184ad280d1c9993122e790ced7178e9ef7557f4b311829ad5250e7"} Dec 13 09:01:08 crc kubenswrapper[4971]: I1213 09:01:08.363216 4971 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="dc4462cbde184ad280d1c9993122e790ced7178e9ef7557f4b311829ad5250e7" Dec 13 09:01:08 crc kubenswrapper[4971]: I1213 09:01:08.363270 4971 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-cron-29426941-c6mj5" Dec 13 09:01:18 crc kubenswrapper[4971]: I1213 09:01:18.768814 4971 scope.go:117] "RemoveContainer" containerID="95d975a85829af1f5477ac9667504f9ab79ab97ddaabd918334d5b4218e3d164" Dec 13 09:01:18 crc kubenswrapper[4971]: E1213 09:01:18.769762 4971 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-82xjz_openshift-machine-config-operator(e0c4af12-edda-41d7-9b44-f87396174ef2)\"" pod="openshift-machine-config-operator/machine-config-daemon-82xjz" podUID="e0c4af12-edda-41d7-9b44-f87396174ef2" Dec 13 09:01:29 crc kubenswrapper[4971]: I1213 09:01:29.770001 4971 scope.go:117] "RemoveContainer" containerID="95d975a85829af1f5477ac9667504f9ab79ab97ddaabd918334d5b4218e3d164" Dec 13 09:01:29 crc kubenswrapper[4971]: E1213 09:01:29.770864 4971 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-82xjz_openshift-machine-config-operator(e0c4af12-edda-41d7-9b44-f87396174ef2)\"" pod="openshift-machine-config-operator/machine-config-daemon-82xjz" podUID="e0c4af12-edda-41d7-9b44-f87396174ef2" Dec 13 09:01:44 crc kubenswrapper[4971]: I1213 09:01:44.769085 4971 scope.go:117] "RemoveContainer" containerID="95d975a85829af1f5477ac9667504f9ab79ab97ddaabd918334d5b4218e3d164" Dec 13 09:01:44 crc kubenswrapper[4971]: E1213 09:01:44.770170 4971 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-82xjz_openshift-machine-config-operator(e0c4af12-edda-41d7-9b44-f87396174ef2)\"" pod="openshift-machine-config-operator/machine-config-daemon-82xjz" podUID="e0c4af12-edda-41d7-9b44-f87396174ef2" Dec 13 09:01:57 crc kubenswrapper[4971]: I1213 09:01:57.779414 4971 scope.go:117] "RemoveContainer" containerID="95d975a85829af1f5477ac9667504f9ab79ab97ddaabd918334d5b4218e3d164" Dec 13 09:01:57 crc kubenswrapper[4971]: E1213 09:01:57.780848 4971 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-82xjz_openshift-machine-config-operator(e0c4af12-edda-41d7-9b44-f87396174ef2)\"" pod="openshift-machine-config-operator/machine-config-daemon-82xjz" podUID="e0c4af12-edda-41d7-9b44-f87396174ef2" Dec 13 09:02:08 crc kubenswrapper[4971]: I1213 09:02:08.769235 4971 scope.go:117] "RemoveContainer" containerID="95d975a85829af1f5477ac9667504f9ab79ab97ddaabd918334d5b4218e3d164" Dec 13 09:02:08 crc kubenswrapper[4971]: E1213 09:02:08.770599 4971 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-82xjz_openshift-machine-config-operator(e0c4af12-edda-41d7-9b44-f87396174ef2)\"" pod="openshift-machine-config-operator/machine-config-daemon-82xjz" podUID="e0c4af12-edda-41d7-9b44-f87396174ef2" Dec 13 09:02:23 crc kubenswrapper[4971]: I1213 09:02:23.777289 4971 scope.go:117] "RemoveContainer" containerID="95d975a85829af1f5477ac9667504f9ab79ab97ddaabd918334d5b4218e3d164" Dec 13 09:02:23 crc kubenswrapper[4971]: E1213 09:02:23.778264 4971 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-82xjz_openshift-machine-config-operator(e0c4af12-edda-41d7-9b44-f87396174ef2)\"" pod="openshift-machine-config-operator/machine-config-daemon-82xjz" podUID="e0c4af12-edda-41d7-9b44-f87396174ef2" Dec 13 09:02:35 crc kubenswrapper[4971]: I1213 09:02:35.768987 4971 scope.go:117] "RemoveContainer" containerID="95d975a85829af1f5477ac9667504f9ab79ab97ddaabd918334d5b4218e3d164" Dec 13 09:02:35 crc kubenswrapper[4971]: E1213 09:02:35.770013 4971 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-82xjz_openshift-machine-config-operator(e0c4af12-edda-41d7-9b44-f87396174ef2)\"" pod="openshift-machine-config-operator/machine-config-daemon-82xjz" podUID="e0c4af12-edda-41d7-9b44-f87396174ef2" Dec 13 09:02:50 crc kubenswrapper[4971]: I1213 09:02:50.768866 4971 scope.go:117] "RemoveContainer" containerID="95d975a85829af1f5477ac9667504f9ab79ab97ddaabd918334d5b4218e3d164" Dec 13 09:02:50 crc kubenswrapper[4971]: E1213 09:02:50.769948 4971 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-82xjz_openshift-machine-config-operator(e0c4af12-edda-41d7-9b44-f87396174ef2)\"" pod="openshift-machine-config-operator/machine-config-daemon-82xjz" podUID="e0c4af12-edda-41d7-9b44-f87396174ef2" Dec 13 09:03:04 crc kubenswrapper[4971]: I1213 09:03:04.768953 4971 scope.go:117] "RemoveContainer" containerID="95d975a85829af1f5477ac9667504f9ab79ab97ddaabd918334d5b4218e3d164" Dec 13 09:03:04 crc kubenswrapper[4971]: E1213 09:03:04.770371 4971 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-82xjz_openshift-machine-config-operator(e0c4af12-edda-41d7-9b44-f87396174ef2)\"" pod="openshift-machine-config-operator/machine-config-daemon-82xjz" podUID="e0c4af12-edda-41d7-9b44-f87396174ef2" Dec 13 09:03:16 crc kubenswrapper[4971]: I1213 09:03:16.769008 4971 scope.go:117] "RemoveContainer" containerID="95d975a85829af1f5477ac9667504f9ab79ab97ddaabd918334d5b4218e3d164" Dec 13 09:03:16 crc kubenswrapper[4971]: E1213 09:03:16.770626 4971 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-82xjz_openshift-machine-config-operator(e0c4af12-edda-41d7-9b44-f87396174ef2)\"" pod="openshift-machine-config-operator/machine-config-daemon-82xjz" podUID="e0c4af12-edda-41d7-9b44-f87396174ef2" Dec 13 09:03:29 crc kubenswrapper[4971]: I1213 09:03:29.769501 4971 scope.go:117] "RemoveContainer" containerID="95d975a85829af1f5477ac9667504f9ab79ab97ddaabd918334d5b4218e3d164" Dec 13 09:03:29 crc kubenswrapper[4971]: E1213 09:03:29.770718 4971 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-82xjz_openshift-machine-config-operator(e0c4af12-edda-41d7-9b44-f87396174ef2)\"" pod="openshift-machine-config-operator/machine-config-daemon-82xjz" podUID="e0c4af12-edda-41d7-9b44-f87396174ef2" Dec 13 09:03:40 crc kubenswrapper[4971]: I1213 09:03:40.768288 4971 scope.go:117] "RemoveContainer" containerID="95d975a85829af1f5477ac9667504f9ab79ab97ddaabd918334d5b4218e3d164" Dec 13 09:03:40 crc kubenswrapper[4971]: E1213 09:03:40.769381 4971 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-82xjz_openshift-machine-config-operator(e0c4af12-edda-41d7-9b44-f87396174ef2)\"" pod="openshift-machine-config-operator/machine-config-daemon-82xjz" podUID="e0c4af12-edda-41d7-9b44-f87396174ef2" Dec 13 09:03:54 crc kubenswrapper[4971]: I1213 09:03:54.769477 4971 scope.go:117] "RemoveContainer" containerID="95d975a85829af1f5477ac9667504f9ab79ab97ddaabd918334d5b4218e3d164" Dec 13 09:03:54 crc kubenswrapper[4971]: E1213 09:03:54.770690 4971 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-82xjz_openshift-machine-config-operator(e0c4af12-edda-41d7-9b44-f87396174ef2)\"" pod="openshift-machine-config-operator/machine-config-daemon-82xjz" podUID="e0c4af12-edda-41d7-9b44-f87396174ef2" Dec 13 09:04:09 crc kubenswrapper[4971]: I1213 09:04:09.770041 4971 scope.go:117] "RemoveContainer" containerID="95d975a85829af1f5477ac9667504f9ab79ab97ddaabd918334d5b4218e3d164" Dec 13 09:04:09 crc kubenswrapper[4971]: E1213 09:04:09.771842 4971 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-82xjz_openshift-machine-config-operator(e0c4af12-edda-41d7-9b44-f87396174ef2)\"" pod="openshift-machine-config-operator/machine-config-daemon-82xjz" podUID="e0c4af12-edda-41d7-9b44-f87396174ef2" Dec 13 09:04:23 crc kubenswrapper[4971]: I1213 09:04:23.778083 4971 scope.go:117] "RemoveContainer" containerID="95d975a85829af1f5477ac9667504f9ab79ab97ddaabd918334d5b4218e3d164" Dec 13 09:04:23 crc kubenswrapper[4971]: E1213 09:04:23.781127 4971 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-82xjz_openshift-machine-config-operator(e0c4af12-edda-41d7-9b44-f87396174ef2)\"" pod="openshift-machine-config-operator/machine-config-daemon-82xjz" podUID="e0c4af12-edda-41d7-9b44-f87396174ef2" Dec 13 09:04:36 crc kubenswrapper[4971]: I1213 09:04:36.769218 4971 scope.go:117] "RemoveContainer" containerID="95d975a85829af1f5477ac9667504f9ab79ab97ddaabd918334d5b4218e3d164" Dec 13 09:04:36 crc kubenswrapper[4971]: E1213 09:04:36.770450 4971 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-82xjz_openshift-machine-config-operator(e0c4af12-edda-41d7-9b44-f87396174ef2)\"" pod="openshift-machine-config-operator/machine-config-daemon-82xjz" podUID="e0c4af12-edda-41d7-9b44-f87396174ef2" Dec 13 09:04:51 crc kubenswrapper[4971]: I1213 09:04:51.769810 4971 scope.go:117] "RemoveContainer" containerID="95d975a85829af1f5477ac9667504f9ab79ab97ddaabd918334d5b4218e3d164" Dec 13 09:04:52 crc kubenswrapper[4971]: I1213 09:04:52.986242 4971 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-82xjz" event={"ID":"e0c4af12-edda-41d7-9b44-f87396174ef2","Type":"ContainerStarted","Data":"278085db9b74f17ff244df1fcc488bae1e210f4d39317c392d734386b6cc185c"} Dec 13 09:05:14 crc kubenswrapper[4971]: I1213 09:05:14.169977 4971 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-marketplace-b8jqn"] Dec 13 09:05:14 crc kubenswrapper[4971]: E1213 09:05:14.171485 4971 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="a7d15eb0-4473-488c-8355-db1a4d139e7c" containerName="keystone-cron" Dec 13 09:05:14 crc kubenswrapper[4971]: I1213 09:05:14.171506 4971 state_mem.go:107] "Deleted CPUSet assignment" podUID="a7d15eb0-4473-488c-8355-db1a4d139e7c" containerName="keystone-cron" Dec 13 09:05:14 crc kubenswrapper[4971]: I1213 09:05:14.172637 4971 memory_manager.go:354] "RemoveStaleState removing state" podUID="a7d15eb0-4473-488c-8355-db1a4d139e7c" containerName="keystone-cron" Dec 13 09:05:14 crc kubenswrapper[4971]: I1213 09:05:14.174637 4971 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-b8jqn" Dec 13 09:05:14 crc kubenswrapper[4971]: I1213 09:05:14.207888 4971 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-b8jqn"] Dec 13 09:05:14 crc kubenswrapper[4971]: I1213 09:05:14.319264 4971 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-spdvz\" (UniqueName: \"kubernetes.io/projected/8d0ab13c-c16e-4216-9373-833df278eaa9-kube-api-access-spdvz\") pod \"redhat-marketplace-b8jqn\" (UID: \"8d0ab13c-c16e-4216-9373-833df278eaa9\") " pod="openshift-marketplace/redhat-marketplace-b8jqn" Dec 13 09:05:14 crc kubenswrapper[4971]: I1213 09:05:14.319457 4971 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/8d0ab13c-c16e-4216-9373-833df278eaa9-catalog-content\") pod \"redhat-marketplace-b8jqn\" (UID: \"8d0ab13c-c16e-4216-9373-833df278eaa9\") " pod="openshift-marketplace/redhat-marketplace-b8jqn" Dec 13 09:05:14 crc kubenswrapper[4971]: I1213 09:05:14.319999 4971 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/8d0ab13c-c16e-4216-9373-833df278eaa9-utilities\") pod \"redhat-marketplace-b8jqn\" (UID: \"8d0ab13c-c16e-4216-9373-833df278eaa9\") " pod="openshift-marketplace/redhat-marketplace-b8jqn" Dec 13 09:05:14 crc kubenswrapper[4971]: I1213 09:05:14.439827 4971 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-spdvz\" (UniqueName: \"kubernetes.io/projected/8d0ab13c-c16e-4216-9373-833df278eaa9-kube-api-access-spdvz\") pod \"redhat-marketplace-b8jqn\" (UID: \"8d0ab13c-c16e-4216-9373-833df278eaa9\") " pod="openshift-marketplace/redhat-marketplace-b8jqn" Dec 13 09:05:14 crc kubenswrapper[4971]: I1213 09:05:14.440296 4971 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/8d0ab13c-c16e-4216-9373-833df278eaa9-catalog-content\") pod \"redhat-marketplace-b8jqn\" (UID: \"8d0ab13c-c16e-4216-9373-833df278eaa9\") " pod="openshift-marketplace/redhat-marketplace-b8jqn" Dec 13 09:05:14 crc kubenswrapper[4971]: I1213 09:05:14.440583 4971 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/8d0ab13c-c16e-4216-9373-833df278eaa9-utilities\") pod \"redhat-marketplace-b8jqn\" (UID: \"8d0ab13c-c16e-4216-9373-833df278eaa9\") " pod="openshift-marketplace/redhat-marketplace-b8jqn" Dec 13 09:05:14 crc kubenswrapper[4971]: I1213 09:05:14.442782 4971 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/8d0ab13c-c16e-4216-9373-833df278eaa9-utilities\") pod \"redhat-marketplace-b8jqn\" (UID: \"8d0ab13c-c16e-4216-9373-833df278eaa9\") " pod="openshift-marketplace/redhat-marketplace-b8jqn" Dec 13 09:05:14 crc kubenswrapper[4971]: I1213 09:05:14.443960 4971 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/8d0ab13c-c16e-4216-9373-833df278eaa9-catalog-content\") pod \"redhat-marketplace-b8jqn\" (UID: \"8d0ab13c-c16e-4216-9373-833df278eaa9\") " pod="openshift-marketplace/redhat-marketplace-b8jqn" Dec 13 09:05:14 crc kubenswrapper[4971]: I1213 09:05:14.469805 4971 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-spdvz\" (UniqueName: \"kubernetes.io/projected/8d0ab13c-c16e-4216-9373-833df278eaa9-kube-api-access-spdvz\") pod \"redhat-marketplace-b8jqn\" (UID: \"8d0ab13c-c16e-4216-9373-833df278eaa9\") " pod="openshift-marketplace/redhat-marketplace-b8jqn" Dec 13 09:05:14 crc kubenswrapper[4971]: I1213 09:05:14.506922 4971 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-b8jqn" Dec 13 09:05:15 crc kubenswrapper[4971]: I1213 09:05:15.121416 4971 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-b8jqn"] Dec 13 09:05:15 crc kubenswrapper[4971]: I1213 09:05:15.242612 4971 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-b8jqn" event={"ID":"8d0ab13c-c16e-4216-9373-833df278eaa9","Type":"ContainerStarted","Data":"233b2d584177ede3f1ecd0ba73b6bb613d9b35578aa8b756bf9a1e866ebb86b7"} Dec 13 09:05:16 crc kubenswrapper[4971]: I1213 09:05:16.256275 4971 generic.go:334] "Generic (PLEG): container finished" podID="8d0ab13c-c16e-4216-9373-833df278eaa9" containerID="5c661228d48f644a9d42b041191f6c518055b19d60cacc7c6f3463d719152c6d" exitCode=0 Dec 13 09:05:16 crc kubenswrapper[4971]: I1213 09:05:16.256393 4971 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-b8jqn" event={"ID":"8d0ab13c-c16e-4216-9373-833df278eaa9","Type":"ContainerDied","Data":"5c661228d48f644a9d42b041191f6c518055b19d60cacc7c6f3463d719152c6d"} Dec 13 09:05:16 crc kubenswrapper[4971]: I1213 09:05:16.260002 4971 provider.go:102] Refreshing cache for provider: *credentialprovider.defaultDockerConfigProvider Dec 13 09:05:18 crc kubenswrapper[4971]: I1213 09:05:18.279926 4971 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-b8jqn" event={"ID":"8d0ab13c-c16e-4216-9373-833df278eaa9","Type":"ContainerStarted","Data":"9d1adb7c465687f0d27c4edbfe01db2091ca4f44d16da48fef925a52e13d57d8"} Dec 13 09:05:19 crc kubenswrapper[4971]: I1213 09:05:19.293885 4971 generic.go:334] "Generic (PLEG): container finished" podID="8d0ab13c-c16e-4216-9373-833df278eaa9" containerID="9d1adb7c465687f0d27c4edbfe01db2091ca4f44d16da48fef925a52e13d57d8" exitCode=0 Dec 13 09:05:19 crc kubenswrapper[4971]: I1213 09:05:19.293973 4971 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-b8jqn" event={"ID":"8d0ab13c-c16e-4216-9373-833df278eaa9","Type":"ContainerDied","Data":"9d1adb7c465687f0d27c4edbfe01db2091ca4f44d16da48fef925a52e13d57d8"} Dec 13 09:05:20 crc kubenswrapper[4971]: I1213 09:05:20.311247 4971 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-b8jqn" event={"ID":"8d0ab13c-c16e-4216-9373-833df278eaa9","Type":"ContainerStarted","Data":"7c496069e28fdf6a179a715c1f78d7bb49518ac9e794cbd97b3e570c2445f00f"} Dec 13 09:05:20 crc kubenswrapper[4971]: I1213 09:05:20.337314 4971 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-marketplace-b8jqn" podStartSLOduration=2.841606353 podStartE2EDuration="6.337281463s" podCreationTimestamp="2025-12-13 09:05:14 +0000 UTC" firstStartedPulling="2025-12-13 09:05:16.259709559 +0000 UTC m=+8172.864119007" lastFinishedPulling="2025-12-13 09:05:19.755384669 +0000 UTC m=+8176.359794117" observedRunningTime="2025-12-13 09:05:20.334265708 +0000 UTC m=+8176.938675176" watchObservedRunningTime="2025-12-13 09:05:20.337281463 +0000 UTC m=+8176.941690921" Dec 13 09:05:24 crc kubenswrapper[4971]: I1213 09:05:24.507580 4971 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-marketplace-b8jqn" Dec 13 09:05:24 crc kubenswrapper[4971]: I1213 09:05:24.508638 4971 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-marketplace-b8jqn" Dec 13 09:05:24 crc kubenswrapper[4971]: I1213 09:05:24.559212 4971 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-marketplace-b8jqn" Dec 13 09:05:25 crc kubenswrapper[4971]: I1213 09:05:25.412579 4971 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-marketplace-b8jqn" Dec 13 09:05:25 crc kubenswrapper[4971]: I1213 09:05:25.471053 4971 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-b8jqn"] Dec 13 09:05:27 crc kubenswrapper[4971]: I1213 09:05:27.384466 4971 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-marketplace-b8jqn" podUID="8d0ab13c-c16e-4216-9373-833df278eaa9" containerName="registry-server" containerID="cri-o://7c496069e28fdf6a179a715c1f78d7bb49518ac9e794cbd97b3e570c2445f00f" gracePeriod=2 Dec 13 09:05:28 crc kubenswrapper[4971]: I1213 09:05:28.049400 4971 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-b8jqn" Dec 13 09:05:28 crc kubenswrapper[4971]: I1213 09:05:28.212974 4971 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/8d0ab13c-c16e-4216-9373-833df278eaa9-utilities\") pod \"8d0ab13c-c16e-4216-9373-833df278eaa9\" (UID: \"8d0ab13c-c16e-4216-9373-833df278eaa9\") " Dec 13 09:05:28 crc kubenswrapper[4971]: I1213 09:05:28.213247 4971 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-spdvz\" (UniqueName: \"kubernetes.io/projected/8d0ab13c-c16e-4216-9373-833df278eaa9-kube-api-access-spdvz\") pod \"8d0ab13c-c16e-4216-9373-833df278eaa9\" (UID: \"8d0ab13c-c16e-4216-9373-833df278eaa9\") " Dec 13 09:05:28 crc kubenswrapper[4971]: I1213 09:05:28.213401 4971 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/8d0ab13c-c16e-4216-9373-833df278eaa9-catalog-content\") pod \"8d0ab13c-c16e-4216-9373-833df278eaa9\" (UID: \"8d0ab13c-c16e-4216-9373-833df278eaa9\") " Dec 13 09:05:28 crc kubenswrapper[4971]: I1213 09:05:28.214035 4971 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/8d0ab13c-c16e-4216-9373-833df278eaa9-utilities" (OuterVolumeSpecName: "utilities") pod "8d0ab13c-c16e-4216-9373-833df278eaa9" (UID: "8d0ab13c-c16e-4216-9373-833df278eaa9"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 13 09:05:28 crc kubenswrapper[4971]: I1213 09:05:28.214129 4971 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/8d0ab13c-c16e-4216-9373-833df278eaa9-utilities\") on node \"crc\" DevicePath \"\"" Dec 13 09:05:28 crc kubenswrapper[4971]: I1213 09:05:28.231278 4971 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/8d0ab13c-c16e-4216-9373-833df278eaa9-kube-api-access-spdvz" (OuterVolumeSpecName: "kube-api-access-spdvz") pod "8d0ab13c-c16e-4216-9373-833df278eaa9" (UID: "8d0ab13c-c16e-4216-9373-833df278eaa9"). InnerVolumeSpecName "kube-api-access-spdvz". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 13 09:05:28 crc kubenswrapper[4971]: I1213 09:05:28.238298 4971 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/8d0ab13c-c16e-4216-9373-833df278eaa9-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "8d0ab13c-c16e-4216-9373-833df278eaa9" (UID: "8d0ab13c-c16e-4216-9373-833df278eaa9"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 13 09:05:28 crc kubenswrapper[4971]: I1213 09:05:28.316806 4971 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/8d0ab13c-c16e-4216-9373-833df278eaa9-catalog-content\") on node \"crc\" DevicePath \"\"" Dec 13 09:05:28 crc kubenswrapper[4971]: I1213 09:05:28.316853 4971 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-spdvz\" (UniqueName: \"kubernetes.io/projected/8d0ab13c-c16e-4216-9373-833df278eaa9-kube-api-access-spdvz\") on node \"crc\" DevicePath \"\"" Dec 13 09:05:28 crc kubenswrapper[4971]: I1213 09:05:28.397413 4971 generic.go:334] "Generic (PLEG): container finished" podID="8d0ab13c-c16e-4216-9373-833df278eaa9" containerID="7c496069e28fdf6a179a715c1f78d7bb49518ac9e794cbd97b3e570c2445f00f" exitCode=0 Dec 13 09:05:28 crc kubenswrapper[4971]: I1213 09:05:28.397483 4971 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-b8jqn" event={"ID":"8d0ab13c-c16e-4216-9373-833df278eaa9","Type":"ContainerDied","Data":"7c496069e28fdf6a179a715c1f78d7bb49518ac9e794cbd97b3e570c2445f00f"} Dec 13 09:05:28 crc kubenswrapper[4971]: I1213 09:05:28.397501 4971 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-b8jqn" Dec 13 09:05:28 crc kubenswrapper[4971]: I1213 09:05:28.397548 4971 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-b8jqn" event={"ID":"8d0ab13c-c16e-4216-9373-833df278eaa9","Type":"ContainerDied","Data":"233b2d584177ede3f1ecd0ba73b6bb613d9b35578aa8b756bf9a1e866ebb86b7"} Dec 13 09:05:28 crc kubenswrapper[4971]: I1213 09:05:28.397570 4971 scope.go:117] "RemoveContainer" containerID="7c496069e28fdf6a179a715c1f78d7bb49518ac9e794cbd97b3e570c2445f00f" Dec 13 09:05:28 crc kubenswrapper[4971]: I1213 09:05:28.427343 4971 scope.go:117] "RemoveContainer" containerID="9d1adb7c465687f0d27c4edbfe01db2091ca4f44d16da48fef925a52e13d57d8" Dec 13 09:05:28 crc kubenswrapper[4971]: I1213 09:05:28.465297 4971 scope.go:117] "RemoveContainer" containerID="5c661228d48f644a9d42b041191f6c518055b19d60cacc7c6f3463d719152c6d" Dec 13 09:05:28 crc kubenswrapper[4971]: I1213 09:05:28.482745 4971 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-b8jqn"] Dec 13 09:05:28 crc kubenswrapper[4971]: I1213 09:05:28.500165 4971 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-marketplace-b8jqn"] Dec 13 09:05:28 crc kubenswrapper[4971]: I1213 09:05:28.511678 4971 scope.go:117] "RemoveContainer" containerID="7c496069e28fdf6a179a715c1f78d7bb49518ac9e794cbd97b3e570c2445f00f" Dec 13 09:05:28 crc kubenswrapper[4971]: E1213 09:05:28.512373 4971 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"7c496069e28fdf6a179a715c1f78d7bb49518ac9e794cbd97b3e570c2445f00f\": container with ID starting with 7c496069e28fdf6a179a715c1f78d7bb49518ac9e794cbd97b3e570c2445f00f not found: ID does not exist" containerID="7c496069e28fdf6a179a715c1f78d7bb49518ac9e794cbd97b3e570c2445f00f" Dec 13 09:05:28 crc kubenswrapper[4971]: I1213 09:05:28.512422 4971 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"7c496069e28fdf6a179a715c1f78d7bb49518ac9e794cbd97b3e570c2445f00f"} err="failed to get container status \"7c496069e28fdf6a179a715c1f78d7bb49518ac9e794cbd97b3e570c2445f00f\": rpc error: code = NotFound desc = could not find container \"7c496069e28fdf6a179a715c1f78d7bb49518ac9e794cbd97b3e570c2445f00f\": container with ID starting with 7c496069e28fdf6a179a715c1f78d7bb49518ac9e794cbd97b3e570c2445f00f not found: ID does not exist" Dec 13 09:05:28 crc kubenswrapper[4971]: I1213 09:05:28.512457 4971 scope.go:117] "RemoveContainer" containerID="9d1adb7c465687f0d27c4edbfe01db2091ca4f44d16da48fef925a52e13d57d8" Dec 13 09:05:28 crc kubenswrapper[4971]: E1213 09:05:28.513043 4971 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"9d1adb7c465687f0d27c4edbfe01db2091ca4f44d16da48fef925a52e13d57d8\": container with ID starting with 9d1adb7c465687f0d27c4edbfe01db2091ca4f44d16da48fef925a52e13d57d8 not found: ID does not exist" containerID="9d1adb7c465687f0d27c4edbfe01db2091ca4f44d16da48fef925a52e13d57d8" Dec 13 09:05:28 crc kubenswrapper[4971]: I1213 09:05:28.513098 4971 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"9d1adb7c465687f0d27c4edbfe01db2091ca4f44d16da48fef925a52e13d57d8"} err="failed to get container status \"9d1adb7c465687f0d27c4edbfe01db2091ca4f44d16da48fef925a52e13d57d8\": rpc error: code = NotFound desc = could not find container \"9d1adb7c465687f0d27c4edbfe01db2091ca4f44d16da48fef925a52e13d57d8\": container with ID starting with 9d1adb7c465687f0d27c4edbfe01db2091ca4f44d16da48fef925a52e13d57d8 not found: ID does not exist" Dec 13 09:05:28 crc kubenswrapper[4971]: I1213 09:05:28.513135 4971 scope.go:117] "RemoveContainer" containerID="5c661228d48f644a9d42b041191f6c518055b19d60cacc7c6f3463d719152c6d" Dec 13 09:05:28 crc kubenswrapper[4971]: E1213 09:05:28.513608 4971 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"5c661228d48f644a9d42b041191f6c518055b19d60cacc7c6f3463d719152c6d\": container with ID starting with 5c661228d48f644a9d42b041191f6c518055b19d60cacc7c6f3463d719152c6d not found: ID does not exist" containerID="5c661228d48f644a9d42b041191f6c518055b19d60cacc7c6f3463d719152c6d" Dec 13 09:05:28 crc kubenswrapper[4971]: I1213 09:05:28.513644 4971 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"5c661228d48f644a9d42b041191f6c518055b19d60cacc7c6f3463d719152c6d"} err="failed to get container status \"5c661228d48f644a9d42b041191f6c518055b19d60cacc7c6f3463d719152c6d\": rpc error: code = NotFound desc = could not find container \"5c661228d48f644a9d42b041191f6c518055b19d60cacc7c6f3463d719152c6d\": container with ID starting with 5c661228d48f644a9d42b041191f6c518055b19d60cacc7c6f3463d719152c6d not found: ID does not exist" Dec 13 09:05:29 crc kubenswrapper[4971]: I1213 09:05:29.782893 4971 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="8d0ab13c-c16e-4216-9373-833df278eaa9" path="/var/lib/kubelet/pods/8d0ab13c-c16e-4216-9373-833df278eaa9/volumes" Dec 13 09:07:16 crc kubenswrapper[4971]: I1213 09:07:16.156831 4971 patch_prober.go:28] interesting pod/machine-config-daemon-82xjz container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 13 09:07:16 crc kubenswrapper[4971]: I1213 09:07:16.157855 4971 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-82xjz" podUID="e0c4af12-edda-41d7-9b44-f87396174ef2" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 13 09:07:46 crc kubenswrapper[4971]: I1213 09:07:46.153464 4971 patch_prober.go:28] interesting pod/machine-config-daemon-82xjz container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 13 09:07:46 crc kubenswrapper[4971]: I1213 09:07:46.154454 4971 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-82xjz" podUID="e0c4af12-edda-41d7-9b44-f87396174ef2" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 13 09:08:16 crc kubenswrapper[4971]: I1213 09:08:16.207418 4971 patch_prober.go:28] interesting pod/machine-config-daemon-82xjz container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 13 09:08:16 crc kubenswrapper[4971]: I1213 09:08:16.208131 4971 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-82xjz" podUID="e0c4af12-edda-41d7-9b44-f87396174ef2" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 13 09:08:16 crc kubenswrapper[4971]: I1213 09:08:16.208212 4971 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-82xjz" Dec 13 09:08:16 crc kubenswrapper[4971]: I1213 09:08:16.210498 4971 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"278085db9b74f17ff244df1fcc488bae1e210f4d39317c392d734386b6cc185c"} pod="openshift-machine-config-operator/machine-config-daemon-82xjz" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Dec 13 09:08:16 crc kubenswrapper[4971]: I1213 09:08:16.212237 4971 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-82xjz" podUID="e0c4af12-edda-41d7-9b44-f87396174ef2" containerName="machine-config-daemon" containerID="cri-o://278085db9b74f17ff244df1fcc488bae1e210f4d39317c392d734386b6cc185c" gracePeriod=600 Dec 13 09:08:17 crc kubenswrapper[4971]: I1213 09:08:17.299955 4971 generic.go:334] "Generic (PLEG): container finished" podID="e0c4af12-edda-41d7-9b44-f87396174ef2" containerID="278085db9b74f17ff244df1fcc488bae1e210f4d39317c392d734386b6cc185c" exitCode=0 Dec 13 09:08:17 crc kubenswrapper[4971]: I1213 09:08:17.300058 4971 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-82xjz" event={"ID":"e0c4af12-edda-41d7-9b44-f87396174ef2","Type":"ContainerDied","Data":"278085db9b74f17ff244df1fcc488bae1e210f4d39317c392d734386b6cc185c"} Dec 13 09:08:17 crc kubenswrapper[4971]: I1213 09:08:17.301813 4971 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-82xjz" event={"ID":"e0c4af12-edda-41d7-9b44-f87396174ef2","Type":"ContainerStarted","Data":"0d8863d17d13d9fdfa7767d29d08739bc5a2b2755f517451ebb079161a29c2fb"} Dec 13 09:08:17 crc kubenswrapper[4971]: I1213 09:08:17.301889 4971 scope.go:117] "RemoveContainer" containerID="95d975a85829af1f5477ac9667504f9ab79ab97ddaabd918334d5b4218e3d164" Dec 13 09:09:13 crc kubenswrapper[4971]: I1213 09:09:13.242692 4971 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/community-operators-pb4bz"] Dec 13 09:09:13 crc kubenswrapper[4971]: E1213 09:09:13.244161 4971 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="8d0ab13c-c16e-4216-9373-833df278eaa9" containerName="extract-utilities" Dec 13 09:09:13 crc kubenswrapper[4971]: I1213 09:09:13.244180 4971 state_mem.go:107] "Deleted CPUSet assignment" podUID="8d0ab13c-c16e-4216-9373-833df278eaa9" containerName="extract-utilities" Dec 13 09:09:13 crc kubenswrapper[4971]: E1213 09:09:13.244217 4971 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="8d0ab13c-c16e-4216-9373-833df278eaa9" containerName="registry-server" Dec 13 09:09:13 crc kubenswrapper[4971]: I1213 09:09:13.244223 4971 state_mem.go:107] "Deleted CPUSet assignment" podUID="8d0ab13c-c16e-4216-9373-833df278eaa9" containerName="registry-server" Dec 13 09:09:13 crc kubenswrapper[4971]: E1213 09:09:13.244234 4971 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="8d0ab13c-c16e-4216-9373-833df278eaa9" containerName="extract-content" Dec 13 09:09:13 crc kubenswrapper[4971]: I1213 09:09:13.244240 4971 state_mem.go:107] "Deleted CPUSet assignment" podUID="8d0ab13c-c16e-4216-9373-833df278eaa9" containerName="extract-content" Dec 13 09:09:13 crc kubenswrapper[4971]: I1213 09:09:13.244428 4971 memory_manager.go:354] "RemoveStaleState removing state" podUID="8d0ab13c-c16e-4216-9373-833df278eaa9" containerName="registry-server" Dec 13 09:09:13 crc kubenswrapper[4971]: I1213 09:09:13.246023 4971 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-pb4bz" Dec 13 09:09:13 crc kubenswrapper[4971]: I1213 09:09:13.259663 4971 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-nphpk\" (UniqueName: \"kubernetes.io/projected/bf4de8f1-dc89-4f24-9b65-3c3183521987-kube-api-access-nphpk\") pod \"community-operators-pb4bz\" (UID: \"bf4de8f1-dc89-4f24-9b65-3c3183521987\") " pod="openshift-marketplace/community-operators-pb4bz" Dec 13 09:09:13 crc kubenswrapper[4971]: I1213 09:09:13.260009 4971 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/bf4de8f1-dc89-4f24-9b65-3c3183521987-utilities\") pod \"community-operators-pb4bz\" (UID: \"bf4de8f1-dc89-4f24-9b65-3c3183521987\") " pod="openshift-marketplace/community-operators-pb4bz" Dec 13 09:09:13 crc kubenswrapper[4971]: I1213 09:09:13.260079 4971 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/bf4de8f1-dc89-4f24-9b65-3c3183521987-catalog-content\") pod \"community-operators-pb4bz\" (UID: \"bf4de8f1-dc89-4f24-9b65-3c3183521987\") " pod="openshift-marketplace/community-operators-pb4bz" Dec 13 09:09:13 crc kubenswrapper[4971]: I1213 09:09:13.263364 4971 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-pb4bz"] Dec 13 09:09:13 crc kubenswrapper[4971]: I1213 09:09:13.367886 4971 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/bf4de8f1-dc89-4f24-9b65-3c3183521987-utilities\") pod \"community-operators-pb4bz\" (UID: \"bf4de8f1-dc89-4f24-9b65-3c3183521987\") " pod="openshift-marketplace/community-operators-pb4bz" Dec 13 09:09:13 crc kubenswrapper[4971]: I1213 09:09:13.368264 4971 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/bf4de8f1-dc89-4f24-9b65-3c3183521987-catalog-content\") pod \"community-operators-pb4bz\" (UID: \"bf4de8f1-dc89-4f24-9b65-3c3183521987\") " pod="openshift-marketplace/community-operators-pb4bz" Dec 13 09:09:13 crc kubenswrapper[4971]: I1213 09:09:13.368347 4971 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-nphpk\" (UniqueName: \"kubernetes.io/projected/bf4de8f1-dc89-4f24-9b65-3c3183521987-kube-api-access-nphpk\") pod \"community-operators-pb4bz\" (UID: \"bf4de8f1-dc89-4f24-9b65-3c3183521987\") " pod="openshift-marketplace/community-operators-pb4bz" Dec 13 09:09:13 crc kubenswrapper[4971]: I1213 09:09:13.369629 4971 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/bf4de8f1-dc89-4f24-9b65-3c3183521987-utilities\") pod \"community-operators-pb4bz\" (UID: \"bf4de8f1-dc89-4f24-9b65-3c3183521987\") " pod="openshift-marketplace/community-operators-pb4bz" Dec 13 09:09:13 crc kubenswrapper[4971]: I1213 09:09:13.369629 4971 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/bf4de8f1-dc89-4f24-9b65-3c3183521987-catalog-content\") pod \"community-operators-pb4bz\" (UID: \"bf4de8f1-dc89-4f24-9b65-3c3183521987\") " pod="openshift-marketplace/community-operators-pb4bz" Dec 13 09:09:13 crc kubenswrapper[4971]: I1213 09:09:13.410310 4971 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-nphpk\" (UniqueName: \"kubernetes.io/projected/bf4de8f1-dc89-4f24-9b65-3c3183521987-kube-api-access-nphpk\") pod \"community-operators-pb4bz\" (UID: \"bf4de8f1-dc89-4f24-9b65-3c3183521987\") " pod="openshift-marketplace/community-operators-pb4bz" Dec 13 09:09:13 crc kubenswrapper[4971]: I1213 09:09:13.569241 4971 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-pb4bz" Dec 13 09:09:14 crc kubenswrapper[4971]: I1213 09:09:14.190900 4971 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-pb4bz"] Dec 13 09:09:14 crc kubenswrapper[4971]: I1213 09:09:14.472789 4971 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-pb4bz" event={"ID":"bf4de8f1-dc89-4f24-9b65-3c3183521987","Type":"ContainerStarted","Data":"e687c22a7daad245bb4711af1f7f8cc7c1178afddeddc856c020b1394689c63f"} Dec 13 09:09:14 crc kubenswrapper[4971]: I1213 09:09:14.478047 4971 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-pb4bz" event={"ID":"bf4de8f1-dc89-4f24-9b65-3c3183521987","Type":"ContainerStarted","Data":"f18a1e31e894d0bd7754432a56081ed63550145098b1febbf87182d17443aac2"} Dec 13 09:09:15 crc kubenswrapper[4971]: I1213 09:09:15.486724 4971 generic.go:334] "Generic (PLEG): container finished" podID="bf4de8f1-dc89-4f24-9b65-3c3183521987" containerID="e687c22a7daad245bb4711af1f7f8cc7c1178afddeddc856c020b1394689c63f" exitCode=0 Dec 13 09:09:15 crc kubenswrapper[4971]: I1213 09:09:15.486784 4971 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-pb4bz" event={"ID":"bf4de8f1-dc89-4f24-9b65-3c3183521987","Type":"ContainerDied","Data":"e687c22a7daad245bb4711af1f7f8cc7c1178afddeddc856c020b1394689c63f"} Dec 13 09:09:17 crc kubenswrapper[4971]: I1213 09:09:17.527812 4971 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-pb4bz" event={"ID":"bf4de8f1-dc89-4f24-9b65-3c3183521987","Type":"ContainerStarted","Data":"0603710a22695b2dc750a59cacbeb310392c10014d9d6c7b5ff71b165182e4fe"} Dec 13 09:09:18 crc kubenswrapper[4971]: I1213 09:09:18.542399 4971 generic.go:334] "Generic (PLEG): container finished" podID="bf4de8f1-dc89-4f24-9b65-3c3183521987" containerID="0603710a22695b2dc750a59cacbeb310392c10014d9d6c7b5ff71b165182e4fe" exitCode=0 Dec 13 09:09:18 crc kubenswrapper[4971]: I1213 09:09:18.542505 4971 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-pb4bz" event={"ID":"bf4de8f1-dc89-4f24-9b65-3c3183521987","Type":"ContainerDied","Data":"0603710a22695b2dc750a59cacbeb310392c10014d9d6c7b5ff71b165182e4fe"} Dec 13 09:09:22 crc kubenswrapper[4971]: I1213 09:09:22.593482 4971 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-pb4bz" event={"ID":"bf4de8f1-dc89-4f24-9b65-3c3183521987","Type":"ContainerStarted","Data":"681847c573b031e6605dc194272f7e92dbf93d1a1e052858b43365c1a342a9a4"} Dec 13 09:09:22 crc kubenswrapper[4971]: I1213 09:09:22.624170 4971 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/community-operators-pb4bz" podStartSLOduration=3.580376869 podStartE2EDuration="9.624146564s" podCreationTimestamp="2025-12-13 09:09:13 +0000 UTC" firstStartedPulling="2025-12-13 09:09:15.490698939 +0000 UTC m=+8412.095108387" lastFinishedPulling="2025-12-13 09:09:21.534468634 +0000 UTC m=+8418.138878082" observedRunningTime="2025-12-13 09:09:22.61710908 +0000 UTC m=+8419.221518548" watchObservedRunningTime="2025-12-13 09:09:22.624146564 +0000 UTC m=+8419.228556032" Dec 13 09:09:23 crc kubenswrapper[4971]: I1213 09:09:23.569567 4971 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/community-operators-pb4bz" Dec 13 09:09:23 crc kubenswrapper[4971]: I1213 09:09:23.570119 4971 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/community-operators-pb4bz" Dec 13 09:09:24 crc kubenswrapper[4971]: I1213 09:09:24.623727 4971 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-marketplace/community-operators-pb4bz" podUID="bf4de8f1-dc89-4f24-9b65-3c3183521987" containerName="registry-server" probeResult="failure" output=< Dec 13 09:09:24 crc kubenswrapper[4971]: timeout: failed to connect service ":50051" within 1s Dec 13 09:09:24 crc kubenswrapper[4971]: > Dec 13 09:09:29 crc kubenswrapper[4971]: I1213 09:09:29.043760 4971 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/certified-operators-s6hq7"] Dec 13 09:09:29 crc kubenswrapper[4971]: I1213 09:09:29.046620 4971 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-s6hq7" Dec 13 09:09:29 crc kubenswrapper[4971]: I1213 09:09:29.095843 4971 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-s6hq7"] Dec 13 09:09:29 crc kubenswrapper[4971]: I1213 09:09:29.125583 4971 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/d85b6745-4bbf-4a91-bca1-eeafc17dc13a-catalog-content\") pod \"certified-operators-s6hq7\" (UID: \"d85b6745-4bbf-4a91-bca1-eeafc17dc13a\") " pod="openshift-marketplace/certified-operators-s6hq7" Dec 13 09:09:29 crc kubenswrapper[4971]: I1213 09:09:29.126067 4971 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-xzd8b\" (UniqueName: \"kubernetes.io/projected/d85b6745-4bbf-4a91-bca1-eeafc17dc13a-kube-api-access-xzd8b\") pod \"certified-operators-s6hq7\" (UID: \"d85b6745-4bbf-4a91-bca1-eeafc17dc13a\") " pod="openshift-marketplace/certified-operators-s6hq7" Dec 13 09:09:29 crc kubenswrapper[4971]: I1213 09:09:29.126329 4971 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/d85b6745-4bbf-4a91-bca1-eeafc17dc13a-utilities\") pod \"certified-operators-s6hq7\" (UID: \"d85b6745-4bbf-4a91-bca1-eeafc17dc13a\") " pod="openshift-marketplace/certified-operators-s6hq7" Dec 13 09:09:29 crc kubenswrapper[4971]: I1213 09:09:29.228771 4971 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-xzd8b\" (UniqueName: \"kubernetes.io/projected/d85b6745-4bbf-4a91-bca1-eeafc17dc13a-kube-api-access-xzd8b\") pod \"certified-operators-s6hq7\" (UID: \"d85b6745-4bbf-4a91-bca1-eeafc17dc13a\") " pod="openshift-marketplace/certified-operators-s6hq7" Dec 13 09:09:29 crc kubenswrapper[4971]: I1213 09:09:29.228941 4971 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/d85b6745-4bbf-4a91-bca1-eeafc17dc13a-utilities\") pod \"certified-operators-s6hq7\" (UID: \"d85b6745-4bbf-4a91-bca1-eeafc17dc13a\") " pod="openshift-marketplace/certified-operators-s6hq7" Dec 13 09:09:29 crc kubenswrapper[4971]: I1213 09:09:29.229015 4971 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/d85b6745-4bbf-4a91-bca1-eeafc17dc13a-catalog-content\") pod \"certified-operators-s6hq7\" (UID: \"d85b6745-4bbf-4a91-bca1-eeafc17dc13a\") " pod="openshift-marketplace/certified-operators-s6hq7" Dec 13 09:09:29 crc kubenswrapper[4971]: I1213 09:09:29.229804 4971 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/d85b6745-4bbf-4a91-bca1-eeafc17dc13a-catalog-content\") pod \"certified-operators-s6hq7\" (UID: \"d85b6745-4bbf-4a91-bca1-eeafc17dc13a\") " pod="openshift-marketplace/certified-operators-s6hq7" Dec 13 09:09:29 crc kubenswrapper[4971]: I1213 09:09:29.231943 4971 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/d85b6745-4bbf-4a91-bca1-eeafc17dc13a-utilities\") pod \"certified-operators-s6hq7\" (UID: \"d85b6745-4bbf-4a91-bca1-eeafc17dc13a\") " pod="openshift-marketplace/certified-operators-s6hq7" Dec 13 09:09:29 crc kubenswrapper[4971]: I1213 09:09:29.253811 4971 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-xzd8b\" (UniqueName: \"kubernetes.io/projected/d85b6745-4bbf-4a91-bca1-eeafc17dc13a-kube-api-access-xzd8b\") pod \"certified-operators-s6hq7\" (UID: \"d85b6745-4bbf-4a91-bca1-eeafc17dc13a\") " pod="openshift-marketplace/certified-operators-s6hq7" Dec 13 09:09:29 crc kubenswrapper[4971]: I1213 09:09:29.382091 4971 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-s6hq7" Dec 13 09:09:30 crc kubenswrapper[4971]: I1213 09:09:30.022653 4971 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-s6hq7"] Dec 13 09:09:30 crc kubenswrapper[4971]: I1213 09:09:30.694243 4971 generic.go:334] "Generic (PLEG): container finished" podID="d85b6745-4bbf-4a91-bca1-eeafc17dc13a" containerID="fec787be3ebd3273629e8dd0ebb8ee75d86301867c8102b563b7f635720f8e0f" exitCode=0 Dec 13 09:09:30 crc kubenswrapper[4971]: I1213 09:09:30.694299 4971 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-s6hq7" event={"ID":"d85b6745-4bbf-4a91-bca1-eeafc17dc13a","Type":"ContainerDied","Data":"fec787be3ebd3273629e8dd0ebb8ee75d86301867c8102b563b7f635720f8e0f"} Dec 13 09:09:30 crc kubenswrapper[4971]: I1213 09:09:30.694339 4971 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-s6hq7" event={"ID":"d85b6745-4bbf-4a91-bca1-eeafc17dc13a","Type":"ContainerStarted","Data":"aa0eb6c7fc3b42f8920742f3397f68bd218a367c991b68d8253101225e005d94"} Dec 13 09:09:32 crc kubenswrapper[4971]: I1213 09:09:32.734879 4971 generic.go:334] "Generic (PLEG): container finished" podID="d85b6745-4bbf-4a91-bca1-eeafc17dc13a" containerID="41839c165b211fc2657214c2cd2184cf6de783bc415119cc1205e6c282bdabe6" exitCode=0 Dec 13 09:09:32 crc kubenswrapper[4971]: I1213 09:09:32.734952 4971 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-s6hq7" event={"ID":"d85b6745-4bbf-4a91-bca1-eeafc17dc13a","Type":"ContainerDied","Data":"41839c165b211fc2657214c2cd2184cf6de783bc415119cc1205e6c282bdabe6"} Dec 13 09:09:33 crc kubenswrapper[4971]: I1213 09:09:33.634820 4971 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/community-operators-pb4bz" Dec 13 09:09:33 crc kubenswrapper[4971]: I1213 09:09:33.704496 4971 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/community-operators-pb4bz" Dec 13 09:09:33 crc kubenswrapper[4971]: I1213 09:09:33.750503 4971 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-s6hq7" event={"ID":"d85b6745-4bbf-4a91-bca1-eeafc17dc13a","Type":"ContainerStarted","Data":"fa47fd06e45a49eca1404b7dceb112812a52257076e3ebc3fbe6f8a28403ff10"} Dec 13 09:09:33 crc kubenswrapper[4971]: I1213 09:09:33.784927 4971 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/certified-operators-s6hq7" podStartSLOduration=2.102912205 podStartE2EDuration="4.784899992s" podCreationTimestamp="2025-12-13 09:09:29 +0000 UTC" firstStartedPulling="2025-12-13 09:09:30.697275508 +0000 UTC m=+8427.301684956" lastFinishedPulling="2025-12-13 09:09:33.379263265 +0000 UTC m=+8429.983672743" observedRunningTime="2025-12-13 09:09:33.77467057 +0000 UTC m=+8430.379080028" watchObservedRunningTime="2025-12-13 09:09:33.784899992 +0000 UTC m=+8430.389309440" Dec 13 09:09:35 crc kubenswrapper[4971]: I1213 09:09:35.029484 4971 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-pb4bz"] Dec 13 09:09:35 crc kubenswrapper[4971]: I1213 09:09:35.030376 4971 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/community-operators-pb4bz" podUID="bf4de8f1-dc89-4f24-9b65-3c3183521987" containerName="registry-server" containerID="cri-o://681847c573b031e6605dc194272f7e92dbf93d1a1e052858b43365c1a342a9a4" gracePeriod=2 Dec 13 09:09:35 crc kubenswrapper[4971]: I1213 09:09:35.584602 4971 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-pb4bz" Dec 13 09:09:35 crc kubenswrapper[4971]: I1213 09:09:35.692422 4971 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/bf4de8f1-dc89-4f24-9b65-3c3183521987-catalog-content\") pod \"bf4de8f1-dc89-4f24-9b65-3c3183521987\" (UID: \"bf4de8f1-dc89-4f24-9b65-3c3183521987\") " Dec 13 09:09:35 crc kubenswrapper[4971]: I1213 09:09:35.692506 4971 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-nphpk\" (UniqueName: \"kubernetes.io/projected/bf4de8f1-dc89-4f24-9b65-3c3183521987-kube-api-access-nphpk\") pod \"bf4de8f1-dc89-4f24-9b65-3c3183521987\" (UID: \"bf4de8f1-dc89-4f24-9b65-3c3183521987\") " Dec 13 09:09:35 crc kubenswrapper[4971]: I1213 09:09:35.692613 4971 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/bf4de8f1-dc89-4f24-9b65-3c3183521987-utilities\") pod \"bf4de8f1-dc89-4f24-9b65-3c3183521987\" (UID: \"bf4de8f1-dc89-4f24-9b65-3c3183521987\") " Dec 13 09:09:35 crc kubenswrapper[4971]: I1213 09:09:35.694110 4971 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/bf4de8f1-dc89-4f24-9b65-3c3183521987-utilities" (OuterVolumeSpecName: "utilities") pod "bf4de8f1-dc89-4f24-9b65-3c3183521987" (UID: "bf4de8f1-dc89-4f24-9b65-3c3183521987"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 13 09:09:35 crc kubenswrapper[4971]: I1213 09:09:35.700980 4971 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/bf4de8f1-dc89-4f24-9b65-3c3183521987-kube-api-access-nphpk" (OuterVolumeSpecName: "kube-api-access-nphpk") pod "bf4de8f1-dc89-4f24-9b65-3c3183521987" (UID: "bf4de8f1-dc89-4f24-9b65-3c3183521987"). InnerVolumeSpecName "kube-api-access-nphpk". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 13 09:09:35 crc kubenswrapper[4971]: I1213 09:09:35.749933 4971 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/bf4de8f1-dc89-4f24-9b65-3c3183521987-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "bf4de8f1-dc89-4f24-9b65-3c3183521987" (UID: "bf4de8f1-dc89-4f24-9b65-3c3183521987"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 13 09:09:35 crc kubenswrapper[4971]: I1213 09:09:35.780159 4971 generic.go:334] "Generic (PLEG): container finished" podID="bf4de8f1-dc89-4f24-9b65-3c3183521987" containerID="681847c573b031e6605dc194272f7e92dbf93d1a1e052858b43365c1a342a9a4" exitCode=0 Dec 13 09:09:35 crc kubenswrapper[4971]: I1213 09:09:35.780405 4971 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-pb4bz" Dec 13 09:09:35 crc kubenswrapper[4971]: I1213 09:09:35.783962 4971 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-pb4bz" event={"ID":"bf4de8f1-dc89-4f24-9b65-3c3183521987","Type":"ContainerDied","Data":"681847c573b031e6605dc194272f7e92dbf93d1a1e052858b43365c1a342a9a4"} Dec 13 09:09:35 crc kubenswrapper[4971]: I1213 09:09:35.784024 4971 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-pb4bz" event={"ID":"bf4de8f1-dc89-4f24-9b65-3c3183521987","Type":"ContainerDied","Data":"f18a1e31e894d0bd7754432a56081ed63550145098b1febbf87182d17443aac2"} Dec 13 09:09:35 crc kubenswrapper[4971]: I1213 09:09:35.784049 4971 scope.go:117] "RemoveContainer" containerID="681847c573b031e6605dc194272f7e92dbf93d1a1e052858b43365c1a342a9a4" Dec 13 09:09:35 crc kubenswrapper[4971]: I1213 09:09:35.796343 4971 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/bf4de8f1-dc89-4f24-9b65-3c3183521987-catalog-content\") on node \"crc\" DevicePath \"\"" Dec 13 09:09:35 crc kubenswrapper[4971]: I1213 09:09:35.796380 4971 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-nphpk\" (UniqueName: \"kubernetes.io/projected/bf4de8f1-dc89-4f24-9b65-3c3183521987-kube-api-access-nphpk\") on node \"crc\" DevicePath \"\"" Dec 13 09:09:35 crc kubenswrapper[4971]: I1213 09:09:35.796392 4971 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/bf4de8f1-dc89-4f24-9b65-3c3183521987-utilities\") on node \"crc\" DevicePath \"\"" Dec 13 09:09:35 crc kubenswrapper[4971]: I1213 09:09:35.828152 4971 scope.go:117] "RemoveContainer" containerID="0603710a22695b2dc750a59cacbeb310392c10014d9d6c7b5ff71b165182e4fe" Dec 13 09:09:35 crc kubenswrapper[4971]: I1213 09:09:35.836834 4971 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-pb4bz"] Dec 13 09:09:35 crc kubenswrapper[4971]: I1213 09:09:35.854689 4971 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/community-operators-pb4bz"] Dec 13 09:09:35 crc kubenswrapper[4971]: I1213 09:09:35.870330 4971 scope.go:117] "RemoveContainer" containerID="e687c22a7daad245bb4711af1f7f8cc7c1178afddeddc856c020b1394689c63f" Dec 13 09:09:35 crc kubenswrapper[4971]: I1213 09:09:35.909431 4971 scope.go:117] "RemoveContainer" containerID="681847c573b031e6605dc194272f7e92dbf93d1a1e052858b43365c1a342a9a4" Dec 13 09:09:35 crc kubenswrapper[4971]: E1213 09:09:35.910018 4971 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"681847c573b031e6605dc194272f7e92dbf93d1a1e052858b43365c1a342a9a4\": container with ID starting with 681847c573b031e6605dc194272f7e92dbf93d1a1e052858b43365c1a342a9a4 not found: ID does not exist" containerID="681847c573b031e6605dc194272f7e92dbf93d1a1e052858b43365c1a342a9a4" Dec 13 09:09:35 crc kubenswrapper[4971]: I1213 09:09:35.910072 4971 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"681847c573b031e6605dc194272f7e92dbf93d1a1e052858b43365c1a342a9a4"} err="failed to get container status \"681847c573b031e6605dc194272f7e92dbf93d1a1e052858b43365c1a342a9a4\": rpc error: code = NotFound desc = could not find container \"681847c573b031e6605dc194272f7e92dbf93d1a1e052858b43365c1a342a9a4\": container with ID starting with 681847c573b031e6605dc194272f7e92dbf93d1a1e052858b43365c1a342a9a4 not found: ID does not exist" Dec 13 09:09:35 crc kubenswrapper[4971]: I1213 09:09:35.910101 4971 scope.go:117] "RemoveContainer" containerID="0603710a22695b2dc750a59cacbeb310392c10014d9d6c7b5ff71b165182e4fe" Dec 13 09:09:35 crc kubenswrapper[4971]: E1213 09:09:35.910497 4971 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"0603710a22695b2dc750a59cacbeb310392c10014d9d6c7b5ff71b165182e4fe\": container with ID starting with 0603710a22695b2dc750a59cacbeb310392c10014d9d6c7b5ff71b165182e4fe not found: ID does not exist" containerID="0603710a22695b2dc750a59cacbeb310392c10014d9d6c7b5ff71b165182e4fe" Dec 13 09:09:35 crc kubenswrapper[4971]: I1213 09:09:35.910554 4971 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"0603710a22695b2dc750a59cacbeb310392c10014d9d6c7b5ff71b165182e4fe"} err="failed to get container status \"0603710a22695b2dc750a59cacbeb310392c10014d9d6c7b5ff71b165182e4fe\": rpc error: code = NotFound desc = could not find container \"0603710a22695b2dc750a59cacbeb310392c10014d9d6c7b5ff71b165182e4fe\": container with ID starting with 0603710a22695b2dc750a59cacbeb310392c10014d9d6c7b5ff71b165182e4fe not found: ID does not exist" Dec 13 09:09:35 crc kubenswrapper[4971]: I1213 09:09:35.910576 4971 scope.go:117] "RemoveContainer" containerID="e687c22a7daad245bb4711af1f7f8cc7c1178afddeddc856c020b1394689c63f" Dec 13 09:09:35 crc kubenswrapper[4971]: E1213 09:09:35.910944 4971 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"e687c22a7daad245bb4711af1f7f8cc7c1178afddeddc856c020b1394689c63f\": container with ID starting with e687c22a7daad245bb4711af1f7f8cc7c1178afddeddc856c020b1394689c63f not found: ID does not exist" containerID="e687c22a7daad245bb4711af1f7f8cc7c1178afddeddc856c020b1394689c63f" Dec 13 09:09:35 crc kubenswrapper[4971]: I1213 09:09:35.910991 4971 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"e687c22a7daad245bb4711af1f7f8cc7c1178afddeddc856c020b1394689c63f"} err="failed to get container status \"e687c22a7daad245bb4711af1f7f8cc7c1178afddeddc856c020b1394689c63f\": rpc error: code = NotFound desc = could not find container \"e687c22a7daad245bb4711af1f7f8cc7c1178afddeddc856c020b1394689c63f\": container with ID starting with e687c22a7daad245bb4711af1f7f8cc7c1178afddeddc856c020b1394689c63f not found: ID does not exist" Dec 13 09:09:37 crc kubenswrapper[4971]: I1213 09:09:37.781090 4971 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="bf4de8f1-dc89-4f24-9b65-3c3183521987" path="/var/lib/kubelet/pods/bf4de8f1-dc89-4f24-9b65-3c3183521987/volumes" Dec 13 09:09:39 crc kubenswrapper[4971]: I1213 09:09:39.382645 4971 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/certified-operators-s6hq7" Dec 13 09:09:39 crc kubenswrapper[4971]: I1213 09:09:39.383480 4971 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/certified-operators-s6hq7" Dec 13 09:09:39 crc kubenswrapper[4971]: I1213 09:09:39.567696 4971 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/certified-operators-s6hq7" Dec 13 09:09:39 crc kubenswrapper[4971]: I1213 09:09:39.902245 4971 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/certified-operators-s6hq7" Dec 13 09:09:39 crc kubenswrapper[4971]: I1213 09:09:39.970978 4971 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-s6hq7"] Dec 13 09:09:41 crc kubenswrapper[4971]: I1213 09:09:41.860739 4971 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/certified-operators-s6hq7" podUID="d85b6745-4bbf-4a91-bca1-eeafc17dc13a" containerName="registry-server" containerID="cri-o://fa47fd06e45a49eca1404b7dceb112812a52257076e3ebc3fbe6f8a28403ff10" gracePeriod=2 Dec 13 09:09:42 crc kubenswrapper[4971]: I1213 09:09:42.431615 4971 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-s6hq7" Dec 13 09:09:42 crc kubenswrapper[4971]: I1213 09:09:42.508918 4971 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/d85b6745-4bbf-4a91-bca1-eeafc17dc13a-catalog-content\") pod \"d85b6745-4bbf-4a91-bca1-eeafc17dc13a\" (UID: \"d85b6745-4bbf-4a91-bca1-eeafc17dc13a\") " Dec 13 09:09:42 crc kubenswrapper[4971]: I1213 09:09:42.509224 4971 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-xzd8b\" (UniqueName: \"kubernetes.io/projected/d85b6745-4bbf-4a91-bca1-eeafc17dc13a-kube-api-access-xzd8b\") pod \"d85b6745-4bbf-4a91-bca1-eeafc17dc13a\" (UID: \"d85b6745-4bbf-4a91-bca1-eeafc17dc13a\") " Dec 13 09:09:42 crc kubenswrapper[4971]: I1213 09:09:42.509463 4971 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/d85b6745-4bbf-4a91-bca1-eeafc17dc13a-utilities\") pod \"d85b6745-4bbf-4a91-bca1-eeafc17dc13a\" (UID: \"d85b6745-4bbf-4a91-bca1-eeafc17dc13a\") " Dec 13 09:09:42 crc kubenswrapper[4971]: I1213 09:09:42.510668 4971 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/d85b6745-4bbf-4a91-bca1-eeafc17dc13a-utilities" (OuterVolumeSpecName: "utilities") pod "d85b6745-4bbf-4a91-bca1-eeafc17dc13a" (UID: "d85b6745-4bbf-4a91-bca1-eeafc17dc13a"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 13 09:09:42 crc kubenswrapper[4971]: I1213 09:09:42.518414 4971 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/d85b6745-4bbf-4a91-bca1-eeafc17dc13a-kube-api-access-xzd8b" (OuterVolumeSpecName: "kube-api-access-xzd8b") pod "d85b6745-4bbf-4a91-bca1-eeafc17dc13a" (UID: "d85b6745-4bbf-4a91-bca1-eeafc17dc13a"). InnerVolumeSpecName "kube-api-access-xzd8b". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 13 09:09:42 crc kubenswrapper[4971]: I1213 09:09:42.569375 4971 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/d85b6745-4bbf-4a91-bca1-eeafc17dc13a-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "d85b6745-4bbf-4a91-bca1-eeafc17dc13a" (UID: "d85b6745-4bbf-4a91-bca1-eeafc17dc13a"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 13 09:09:42 crc kubenswrapper[4971]: I1213 09:09:42.614480 4971 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/d85b6745-4bbf-4a91-bca1-eeafc17dc13a-catalog-content\") on node \"crc\" DevicePath \"\"" Dec 13 09:09:42 crc kubenswrapper[4971]: I1213 09:09:42.614548 4971 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-xzd8b\" (UniqueName: \"kubernetes.io/projected/d85b6745-4bbf-4a91-bca1-eeafc17dc13a-kube-api-access-xzd8b\") on node \"crc\" DevicePath \"\"" Dec 13 09:09:42 crc kubenswrapper[4971]: I1213 09:09:42.614568 4971 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/d85b6745-4bbf-4a91-bca1-eeafc17dc13a-utilities\") on node \"crc\" DevicePath \"\"" Dec 13 09:09:42 crc kubenswrapper[4971]: I1213 09:09:42.878579 4971 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-s6hq7" Dec 13 09:09:42 crc kubenswrapper[4971]: I1213 09:09:42.878559 4971 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-s6hq7" event={"ID":"d85b6745-4bbf-4a91-bca1-eeafc17dc13a","Type":"ContainerDied","Data":"fa47fd06e45a49eca1404b7dceb112812a52257076e3ebc3fbe6f8a28403ff10"} Dec 13 09:09:42 crc kubenswrapper[4971]: I1213 09:09:42.878677 4971 scope.go:117] "RemoveContainer" containerID="fa47fd06e45a49eca1404b7dceb112812a52257076e3ebc3fbe6f8a28403ff10" Dec 13 09:09:42 crc kubenswrapper[4971]: I1213 09:09:42.878501 4971 generic.go:334] "Generic (PLEG): container finished" podID="d85b6745-4bbf-4a91-bca1-eeafc17dc13a" containerID="fa47fd06e45a49eca1404b7dceb112812a52257076e3ebc3fbe6f8a28403ff10" exitCode=0 Dec 13 09:09:42 crc kubenswrapper[4971]: I1213 09:09:42.878757 4971 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-s6hq7" event={"ID":"d85b6745-4bbf-4a91-bca1-eeafc17dc13a","Type":"ContainerDied","Data":"aa0eb6c7fc3b42f8920742f3397f68bd218a367c991b68d8253101225e005d94"} Dec 13 09:09:42 crc kubenswrapper[4971]: I1213 09:09:42.922594 4971 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-s6hq7"] Dec 13 09:09:42 crc kubenswrapper[4971]: I1213 09:09:42.924632 4971 scope.go:117] "RemoveContainer" containerID="41839c165b211fc2657214c2cd2184cf6de783bc415119cc1205e6c282bdabe6" Dec 13 09:09:42 crc kubenswrapper[4971]: I1213 09:09:42.939325 4971 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/certified-operators-s6hq7"] Dec 13 09:09:42 crc kubenswrapper[4971]: I1213 09:09:42.965352 4971 scope.go:117] "RemoveContainer" containerID="fec787be3ebd3273629e8dd0ebb8ee75d86301867c8102b563b7f635720f8e0f" Dec 13 09:09:43 crc kubenswrapper[4971]: I1213 09:09:43.006147 4971 scope.go:117] "RemoveContainer" containerID="fa47fd06e45a49eca1404b7dceb112812a52257076e3ebc3fbe6f8a28403ff10" Dec 13 09:09:43 crc kubenswrapper[4971]: E1213 09:09:43.006790 4971 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"fa47fd06e45a49eca1404b7dceb112812a52257076e3ebc3fbe6f8a28403ff10\": container with ID starting with fa47fd06e45a49eca1404b7dceb112812a52257076e3ebc3fbe6f8a28403ff10 not found: ID does not exist" containerID="fa47fd06e45a49eca1404b7dceb112812a52257076e3ebc3fbe6f8a28403ff10" Dec 13 09:09:43 crc kubenswrapper[4971]: I1213 09:09:43.006830 4971 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"fa47fd06e45a49eca1404b7dceb112812a52257076e3ebc3fbe6f8a28403ff10"} err="failed to get container status \"fa47fd06e45a49eca1404b7dceb112812a52257076e3ebc3fbe6f8a28403ff10\": rpc error: code = NotFound desc = could not find container \"fa47fd06e45a49eca1404b7dceb112812a52257076e3ebc3fbe6f8a28403ff10\": container with ID starting with fa47fd06e45a49eca1404b7dceb112812a52257076e3ebc3fbe6f8a28403ff10 not found: ID does not exist" Dec 13 09:09:43 crc kubenswrapper[4971]: I1213 09:09:43.006855 4971 scope.go:117] "RemoveContainer" containerID="41839c165b211fc2657214c2cd2184cf6de783bc415119cc1205e6c282bdabe6" Dec 13 09:09:43 crc kubenswrapper[4971]: E1213 09:09:43.007169 4971 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"41839c165b211fc2657214c2cd2184cf6de783bc415119cc1205e6c282bdabe6\": container with ID starting with 41839c165b211fc2657214c2cd2184cf6de783bc415119cc1205e6c282bdabe6 not found: ID does not exist" containerID="41839c165b211fc2657214c2cd2184cf6de783bc415119cc1205e6c282bdabe6" Dec 13 09:09:43 crc kubenswrapper[4971]: I1213 09:09:43.007206 4971 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"41839c165b211fc2657214c2cd2184cf6de783bc415119cc1205e6c282bdabe6"} err="failed to get container status \"41839c165b211fc2657214c2cd2184cf6de783bc415119cc1205e6c282bdabe6\": rpc error: code = NotFound desc = could not find container \"41839c165b211fc2657214c2cd2184cf6de783bc415119cc1205e6c282bdabe6\": container with ID starting with 41839c165b211fc2657214c2cd2184cf6de783bc415119cc1205e6c282bdabe6 not found: ID does not exist" Dec 13 09:09:43 crc kubenswrapper[4971]: I1213 09:09:43.007229 4971 scope.go:117] "RemoveContainer" containerID="fec787be3ebd3273629e8dd0ebb8ee75d86301867c8102b563b7f635720f8e0f" Dec 13 09:09:43 crc kubenswrapper[4971]: E1213 09:09:43.007517 4971 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"fec787be3ebd3273629e8dd0ebb8ee75d86301867c8102b563b7f635720f8e0f\": container with ID starting with fec787be3ebd3273629e8dd0ebb8ee75d86301867c8102b563b7f635720f8e0f not found: ID does not exist" containerID="fec787be3ebd3273629e8dd0ebb8ee75d86301867c8102b563b7f635720f8e0f" Dec 13 09:09:43 crc kubenswrapper[4971]: I1213 09:09:43.007565 4971 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"fec787be3ebd3273629e8dd0ebb8ee75d86301867c8102b563b7f635720f8e0f"} err="failed to get container status \"fec787be3ebd3273629e8dd0ebb8ee75d86301867c8102b563b7f635720f8e0f\": rpc error: code = NotFound desc = could not find container \"fec787be3ebd3273629e8dd0ebb8ee75d86301867c8102b563b7f635720f8e0f\": container with ID starting with fec787be3ebd3273629e8dd0ebb8ee75d86301867c8102b563b7f635720f8e0f not found: ID does not exist" Dec 13 09:09:43 crc kubenswrapper[4971]: I1213 09:09:43.791320 4971 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="d85b6745-4bbf-4a91-bca1-eeafc17dc13a" path="/var/lib/kubelet/pods/d85b6745-4bbf-4a91-bca1-eeafc17dc13a/volumes" Dec 13 09:10:16 crc kubenswrapper[4971]: I1213 09:10:16.154306 4971 patch_prober.go:28] interesting pod/machine-config-daemon-82xjz container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 13 09:10:16 crc kubenswrapper[4971]: I1213 09:10:16.155188 4971 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-82xjz" podUID="e0c4af12-edda-41d7-9b44-f87396174ef2" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 13 09:10:46 crc kubenswrapper[4971]: I1213 09:10:46.153967 4971 patch_prober.go:28] interesting pod/machine-config-daemon-82xjz container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 13 09:10:46 crc kubenswrapper[4971]: I1213 09:10:46.154894 4971 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-82xjz" podUID="e0c4af12-edda-41d7-9b44-f87396174ef2" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 13 09:11:16 crc kubenswrapper[4971]: I1213 09:11:16.153493 4971 patch_prober.go:28] interesting pod/machine-config-daemon-82xjz container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 13 09:11:16 crc kubenswrapper[4971]: I1213 09:11:16.154229 4971 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-82xjz" podUID="e0c4af12-edda-41d7-9b44-f87396174ef2" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 13 09:11:16 crc kubenswrapper[4971]: I1213 09:11:16.154287 4971 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-82xjz" Dec 13 09:11:16 crc kubenswrapper[4971]: I1213 09:11:16.155366 4971 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"0d8863d17d13d9fdfa7767d29d08739bc5a2b2755f517451ebb079161a29c2fb"} pod="openshift-machine-config-operator/machine-config-daemon-82xjz" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Dec 13 09:11:16 crc kubenswrapper[4971]: I1213 09:11:16.155443 4971 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-82xjz" podUID="e0c4af12-edda-41d7-9b44-f87396174ef2" containerName="machine-config-daemon" containerID="cri-o://0d8863d17d13d9fdfa7767d29d08739bc5a2b2755f517451ebb079161a29c2fb" gracePeriod=600 Dec 13 09:11:16 crc kubenswrapper[4971]: E1213 09:11:16.280617 4971 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-82xjz_openshift-machine-config-operator(e0c4af12-edda-41d7-9b44-f87396174ef2)\"" pod="openshift-machine-config-operator/machine-config-daemon-82xjz" podUID="e0c4af12-edda-41d7-9b44-f87396174ef2" Dec 13 09:11:16 crc kubenswrapper[4971]: I1213 09:11:16.931940 4971 generic.go:334] "Generic (PLEG): container finished" podID="e0c4af12-edda-41d7-9b44-f87396174ef2" containerID="0d8863d17d13d9fdfa7767d29d08739bc5a2b2755f517451ebb079161a29c2fb" exitCode=0 Dec 13 09:11:16 crc kubenswrapper[4971]: I1213 09:11:16.932009 4971 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-82xjz" event={"ID":"e0c4af12-edda-41d7-9b44-f87396174ef2","Type":"ContainerDied","Data":"0d8863d17d13d9fdfa7767d29d08739bc5a2b2755f517451ebb079161a29c2fb"} Dec 13 09:11:16 crc kubenswrapper[4971]: I1213 09:11:16.932106 4971 scope.go:117] "RemoveContainer" containerID="278085db9b74f17ff244df1fcc488bae1e210f4d39317c392d734386b6cc185c" Dec 13 09:11:16 crc kubenswrapper[4971]: I1213 09:11:16.933039 4971 scope.go:117] "RemoveContainer" containerID="0d8863d17d13d9fdfa7767d29d08739bc5a2b2755f517451ebb079161a29c2fb" Dec 13 09:11:16 crc kubenswrapper[4971]: E1213 09:11:16.933550 4971 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-82xjz_openshift-machine-config-operator(e0c4af12-edda-41d7-9b44-f87396174ef2)\"" pod="openshift-machine-config-operator/machine-config-daemon-82xjz" podUID="e0c4af12-edda-41d7-9b44-f87396174ef2" Dec 13 09:11:31 crc kubenswrapper[4971]: I1213 09:11:31.769687 4971 scope.go:117] "RemoveContainer" containerID="0d8863d17d13d9fdfa7767d29d08739bc5a2b2755f517451ebb079161a29c2fb" Dec 13 09:11:31 crc kubenswrapper[4971]: E1213 09:11:31.770831 4971 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-82xjz_openshift-machine-config-operator(e0c4af12-edda-41d7-9b44-f87396174ef2)\"" pod="openshift-machine-config-operator/machine-config-daemon-82xjz" podUID="e0c4af12-edda-41d7-9b44-f87396174ef2" Dec 13 09:11:44 crc kubenswrapper[4971]: I1213 09:11:44.768696 4971 scope.go:117] "RemoveContainer" containerID="0d8863d17d13d9fdfa7767d29d08739bc5a2b2755f517451ebb079161a29c2fb" Dec 13 09:11:44 crc kubenswrapper[4971]: E1213 09:11:44.769660 4971 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-82xjz_openshift-machine-config-operator(e0c4af12-edda-41d7-9b44-f87396174ef2)\"" pod="openshift-machine-config-operator/machine-config-daemon-82xjz" podUID="e0c4af12-edda-41d7-9b44-f87396174ef2" Dec 13 09:11:57 crc kubenswrapper[4971]: I1213 09:11:57.770418 4971 scope.go:117] "RemoveContainer" containerID="0d8863d17d13d9fdfa7767d29d08739bc5a2b2755f517451ebb079161a29c2fb" Dec 13 09:11:57 crc kubenswrapper[4971]: E1213 09:11:57.771376 4971 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-82xjz_openshift-machine-config-operator(e0c4af12-edda-41d7-9b44-f87396174ef2)\"" pod="openshift-machine-config-operator/machine-config-daemon-82xjz" podUID="e0c4af12-edda-41d7-9b44-f87396174ef2" Dec 13 09:12:09 crc kubenswrapper[4971]: I1213 09:12:09.770288 4971 scope.go:117] "RemoveContainer" containerID="0d8863d17d13d9fdfa7767d29d08739bc5a2b2755f517451ebb079161a29c2fb" Dec 13 09:12:09 crc kubenswrapper[4971]: E1213 09:12:09.771397 4971 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-82xjz_openshift-machine-config-operator(e0c4af12-edda-41d7-9b44-f87396174ef2)\"" pod="openshift-machine-config-operator/machine-config-daemon-82xjz" podUID="e0c4af12-edda-41d7-9b44-f87396174ef2" Dec 13 09:12:23 crc kubenswrapper[4971]: I1213 09:12:23.775881 4971 scope.go:117] "RemoveContainer" containerID="0d8863d17d13d9fdfa7767d29d08739bc5a2b2755f517451ebb079161a29c2fb" Dec 13 09:12:23 crc kubenswrapper[4971]: E1213 09:12:23.777063 4971 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-82xjz_openshift-machine-config-operator(e0c4af12-edda-41d7-9b44-f87396174ef2)\"" pod="openshift-machine-config-operator/machine-config-daemon-82xjz" podUID="e0c4af12-edda-41d7-9b44-f87396174ef2" Dec 13 09:12:38 crc kubenswrapper[4971]: I1213 09:12:38.768464 4971 scope.go:117] "RemoveContainer" containerID="0d8863d17d13d9fdfa7767d29d08739bc5a2b2755f517451ebb079161a29c2fb" Dec 13 09:12:38 crc kubenswrapper[4971]: E1213 09:12:38.769493 4971 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-82xjz_openshift-machine-config-operator(e0c4af12-edda-41d7-9b44-f87396174ef2)\"" pod="openshift-machine-config-operator/machine-config-daemon-82xjz" podUID="e0c4af12-edda-41d7-9b44-f87396174ef2" Dec 13 09:12:50 crc kubenswrapper[4971]: I1213 09:12:50.770097 4971 scope.go:117] "RemoveContainer" containerID="0d8863d17d13d9fdfa7767d29d08739bc5a2b2755f517451ebb079161a29c2fb" Dec 13 09:12:50 crc kubenswrapper[4971]: E1213 09:12:50.771742 4971 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-82xjz_openshift-machine-config-operator(e0c4af12-edda-41d7-9b44-f87396174ef2)\"" pod="openshift-machine-config-operator/machine-config-daemon-82xjz" podUID="e0c4af12-edda-41d7-9b44-f87396174ef2" Dec 13 09:13:03 crc kubenswrapper[4971]: I1213 09:13:03.769118 4971 scope.go:117] "RemoveContainer" containerID="0d8863d17d13d9fdfa7767d29d08739bc5a2b2755f517451ebb079161a29c2fb" Dec 13 09:13:03 crc kubenswrapper[4971]: E1213 09:13:03.770287 4971 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-82xjz_openshift-machine-config-operator(e0c4af12-edda-41d7-9b44-f87396174ef2)\"" pod="openshift-machine-config-operator/machine-config-daemon-82xjz" podUID="e0c4af12-edda-41d7-9b44-f87396174ef2" Dec 13 09:13:18 crc kubenswrapper[4971]: I1213 09:13:18.774105 4971 scope.go:117] "RemoveContainer" containerID="0d8863d17d13d9fdfa7767d29d08739bc5a2b2755f517451ebb079161a29c2fb" Dec 13 09:13:18 crc kubenswrapper[4971]: E1213 09:13:18.777723 4971 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-82xjz_openshift-machine-config-operator(e0c4af12-edda-41d7-9b44-f87396174ef2)\"" pod="openshift-machine-config-operator/machine-config-daemon-82xjz" podUID="e0c4af12-edda-41d7-9b44-f87396174ef2" Dec 13 09:13:33 crc kubenswrapper[4971]: I1213 09:13:33.777161 4971 scope.go:117] "RemoveContainer" containerID="0d8863d17d13d9fdfa7767d29d08739bc5a2b2755f517451ebb079161a29c2fb" Dec 13 09:13:33 crc kubenswrapper[4971]: E1213 09:13:33.781408 4971 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-82xjz_openshift-machine-config-operator(e0c4af12-edda-41d7-9b44-f87396174ef2)\"" pod="openshift-machine-config-operator/machine-config-daemon-82xjz" podUID="e0c4af12-edda-41d7-9b44-f87396174ef2" Dec 13 09:13:45 crc kubenswrapper[4971]: I1213 09:13:45.769845 4971 scope.go:117] "RemoveContainer" containerID="0d8863d17d13d9fdfa7767d29d08739bc5a2b2755f517451ebb079161a29c2fb" Dec 13 09:13:45 crc kubenswrapper[4971]: E1213 09:13:45.795065 4971 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-82xjz_openshift-machine-config-operator(e0c4af12-edda-41d7-9b44-f87396174ef2)\"" pod="openshift-machine-config-operator/machine-config-daemon-82xjz" podUID="e0c4af12-edda-41d7-9b44-f87396174ef2" Dec 13 09:13:56 crc kubenswrapper[4971]: I1213 09:13:56.769337 4971 scope.go:117] "RemoveContainer" containerID="0d8863d17d13d9fdfa7767d29d08739bc5a2b2755f517451ebb079161a29c2fb" Dec 13 09:13:56 crc kubenswrapper[4971]: E1213 09:13:56.770480 4971 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-82xjz_openshift-machine-config-operator(e0c4af12-edda-41d7-9b44-f87396174ef2)\"" pod="openshift-machine-config-operator/machine-config-daemon-82xjz" podUID="e0c4af12-edda-41d7-9b44-f87396174ef2" Dec 13 09:14:08 crc kubenswrapper[4971]: I1213 09:14:08.768870 4971 scope.go:117] "RemoveContainer" containerID="0d8863d17d13d9fdfa7767d29d08739bc5a2b2755f517451ebb079161a29c2fb" Dec 13 09:14:08 crc kubenswrapper[4971]: E1213 09:14:08.769667 4971 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-82xjz_openshift-machine-config-operator(e0c4af12-edda-41d7-9b44-f87396174ef2)\"" pod="openshift-machine-config-operator/machine-config-daemon-82xjz" podUID="e0c4af12-edda-41d7-9b44-f87396174ef2" Dec 13 09:14:23 crc kubenswrapper[4971]: I1213 09:14:23.779160 4971 scope.go:117] "RemoveContainer" containerID="0d8863d17d13d9fdfa7767d29d08739bc5a2b2755f517451ebb079161a29c2fb" Dec 13 09:14:23 crc kubenswrapper[4971]: E1213 09:14:23.780318 4971 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-82xjz_openshift-machine-config-operator(e0c4af12-edda-41d7-9b44-f87396174ef2)\"" pod="openshift-machine-config-operator/machine-config-daemon-82xjz" podUID="e0c4af12-edda-41d7-9b44-f87396174ef2" Dec 13 09:14:36 crc kubenswrapper[4971]: I1213 09:14:36.768831 4971 scope.go:117] "RemoveContainer" containerID="0d8863d17d13d9fdfa7767d29d08739bc5a2b2755f517451ebb079161a29c2fb" Dec 13 09:14:36 crc kubenswrapper[4971]: E1213 09:14:36.770080 4971 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-82xjz_openshift-machine-config-operator(e0c4af12-edda-41d7-9b44-f87396174ef2)\"" pod="openshift-machine-config-operator/machine-config-daemon-82xjz" podUID="e0c4af12-edda-41d7-9b44-f87396174ef2" Dec 13 09:14:42 crc kubenswrapper[4971]: I1213 09:14:42.714198 4971 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-operators-6cl4h"] Dec 13 09:14:42 crc kubenswrapper[4971]: E1213 09:14:42.715664 4971 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="d85b6745-4bbf-4a91-bca1-eeafc17dc13a" containerName="registry-server" Dec 13 09:14:42 crc kubenswrapper[4971]: I1213 09:14:42.715687 4971 state_mem.go:107] "Deleted CPUSet assignment" podUID="d85b6745-4bbf-4a91-bca1-eeafc17dc13a" containerName="registry-server" Dec 13 09:14:42 crc kubenswrapper[4971]: E1213 09:14:42.715727 4971 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="bf4de8f1-dc89-4f24-9b65-3c3183521987" containerName="extract-content" Dec 13 09:14:42 crc kubenswrapper[4971]: I1213 09:14:42.715736 4971 state_mem.go:107] "Deleted CPUSet assignment" podUID="bf4de8f1-dc89-4f24-9b65-3c3183521987" containerName="extract-content" Dec 13 09:14:42 crc kubenswrapper[4971]: E1213 09:14:42.715751 4971 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="d85b6745-4bbf-4a91-bca1-eeafc17dc13a" containerName="extract-utilities" Dec 13 09:14:42 crc kubenswrapper[4971]: I1213 09:14:42.715761 4971 state_mem.go:107] "Deleted CPUSet assignment" podUID="d85b6745-4bbf-4a91-bca1-eeafc17dc13a" containerName="extract-utilities" Dec 13 09:14:42 crc kubenswrapper[4971]: E1213 09:14:42.715783 4971 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="d85b6745-4bbf-4a91-bca1-eeafc17dc13a" containerName="extract-content" Dec 13 09:14:42 crc kubenswrapper[4971]: I1213 09:14:42.715791 4971 state_mem.go:107] "Deleted CPUSet assignment" podUID="d85b6745-4bbf-4a91-bca1-eeafc17dc13a" containerName="extract-content" Dec 13 09:14:42 crc kubenswrapper[4971]: E1213 09:14:42.715820 4971 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="bf4de8f1-dc89-4f24-9b65-3c3183521987" containerName="registry-server" Dec 13 09:14:42 crc kubenswrapper[4971]: I1213 09:14:42.715829 4971 state_mem.go:107] "Deleted CPUSet assignment" podUID="bf4de8f1-dc89-4f24-9b65-3c3183521987" containerName="registry-server" Dec 13 09:14:42 crc kubenswrapper[4971]: E1213 09:14:42.715841 4971 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="bf4de8f1-dc89-4f24-9b65-3c3183521987" containerName="extract-utilities" Dec 13 09:14:42 crc kubenswrapper[4971]: I1213 09:14:42.715849 4971 state_mem.go:107] "Deleted CPUSet assignment" podUID="bf4de8f1-dc89-4f24-9b65-3c3183521987" containerName="extract-utilities" Dec 13 09:14:42 crc kubenswrapper[4971]: I1213 09:14:42.716095 4971 memory_manager.go:354] "RemoveStaleState removing state" podUID="bf4de8f1-dc89-4f24-9b65-3c3183521987" containerName="registry-server" Dec 13 09:14:42 crc kubenswrapper[4971]: I1213 09:14:42.716117 4971 memory_manager.go:354] "RemoveStaleState removing state" podUID="d85b6745-4bbf-4a91-bca1-eeafc17dc13a" containerName="registry-server" Dec 13 09:14:42 crc kubenswrapper[4971]: I1213 09:14:42.717740 4971 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-6cl4h" Dec 13 09:14:42 crc kubenswrapper[4971]: I1213 09:14:42.729753 4971 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-6cl4h"] Dec 13 09:14:42 crc kubenswrapper[4971]: I1213 09:14:42.779496 4971 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-zwn2q\" (UniqueName: \"kubernetes.io/projected/1ae30452-bee6-43ce-845a-cb32e1c4c5a9-kube-api-access-zwn2q\") pod \"redhat-operators-6cl4h\" (UID: \"1ae30452-bee6-43ce-845a-cb32e1c4c5a9\") " pod="openshift-marketplace/redhat-operators-6cl4h" Dec 13 09:14:42 crc kubenswrapper[4971]: I1213 09:14:42.779866 4971 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/1ae30452-bee6-43ce-845a-cb32e1c4c5a9-utilities\") pod \"redhat-operators-6cl4h\" (UID: \"1ae30452-bee6-43ce-845a-cb32e1c4c5a9\") " pod="openshift-marketplace/redhat-operators-6cl4h" Dec 13 09:14:42 crc kubenswrapper[4971]: I1213 09:14:42.779914 4971 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/1ae30452-bee6-43ce-845a-cb32e1c4c5a9-catalog-content\") pod \"redhat-operators-6cl4h\" (UID: \"1ae30452-bee6-43ce-845a-cb32e1c4c5a9\") " pod="openshift-marketplace/redhat-operators-6cl4h" Dec 13 09:14:42 crc kubenswrapper[4971]: I1213 09:14:42.882903 4971 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/1ae30452-bee6-43ce-845a-cb32e1c4c5a9-utilities\") pod \"redhat-operators-6cl4h\" (UID: \"1ae30452-bee6-43ce-845a-cb32e1c4c5a9\") " pod="openshift-marketplace/redhat-operators-6cl4h" Dec 13 09:14:42 crc kubenswrapper[4971]: I1213 09:14:42.890198 4971 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/1ae30452-bee6-43ce-845a-cb32e1c4c5a9-utilities\") pod \"redhat-operators-6cl4h\" (UID: \"1ae30452-bee6-43ce-845a-cb32e1c4c5a9\") " pod="openshift-marketplace/redhat-operators-6cl4h" Dec 13 09:14:42 crc kubenswrapper[4971]: I1213 09:14:42.890493 4971 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/1ae30452-bee6-43ce-845a-cb32e1c4c5a9-catalog-content\") pod \"redhat-operators-6cl4h\" (UID: \"1ae30452-bee6-43ce-845a-cb32e1c4c5a9\") " pod="openshift-marketplace/redhat-operators-6cl4h" Dec 13 09:14:42 crc kubenswrapper[4971]: I1213 09:14:42.894366 4971 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/1ae30452-bee6-43ce-845a-cb32e1c4c5a9-catalog-content\") pod \"redhat-operators-6cl4h\" (UID: \"1ae30452-bee6-43ce-845a-cb32e1c4c5a9\") " pod="openshift-marketplace/redhat-operators-6cl4h" Dec 13 09:14:42 crc kubenswrapper[4971]: I1213 09:14:42.895955 4971 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-zwn2q\" (UniqueName: \"kubernetes.io/projected/1ae30452-bee6-43ce-845a-cb32e1c4c5a9-kube-api-access-zwn2q\") pod \"redhat-operators-6cl4h\" (UID: \"1ae30452-bee6-43ce-845a-cb32e1c4c5a9\") " pod="openshift-marketplace/redhat-operators-6cl4h" Dec 13 09:14:42 crc kubenswrapper[4971]: I1213 09:14:42.934863 4971 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-zwn2q\" (UniqueName: \"kubernetes.io/projected/1ae30452-bee6-43ce-845a-cb32e1c4c5a9-kube-api-access-zwn2q\") pod \"redhat-operators-6cl4h\" (UID: \"1ae30452-bee6-43ce-845a-cb32e1c4c5a9\") " pod="openshift-marketplace/redhat-operators-6cl4h" Dec 13 09:14:43 crc kubenswrapper[4971]: I1213 09:14:43.076884 4971 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-6cl4h" Dec 13 09:14:43 crc kubenswrapper[4971]: I1213 09:14:43.651656 4971 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-6cl4h"] Dec 13 09:14:44 crc kubenswrapper[4971]: I1213 09:14:44.528834 4971 generic.go:334] "Generic (PLEG): container finished" podID="1ae30452-bee6-43ce-845a-cb32e1c4c5a9" containerID="84914ffd0bc27d5642e6ff188ea3744e8cde51fdd3836079ba928ba9e88af078" exitCode=0 Dec 13 09:14:44 crc kubenswrapper[4971]: I1213 09:14:44.528933 4971 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-6cl4h" event={"ID":"1ae30452-bee6-43ce-845a-cb32e1c4c5a9","Type":"ContainerDied","Data":"84914ffd0bc27d5642e6ff188ea3744e8cde51fdd3836079ba928ba9e88af078"} Dec 13 09:14:44 crc kubenswrapper[4971]: I1213 09:14:44.529340 4971 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-6cl4h" event={"ID":"1ae30452-bee6-43ce-845a-cb32e1c4c5a9","Type":"ContainerStarted","Data":"5b49e89154ee75e92190e3e91575838785f9365d664f829878c469b977d39a70"} Dec 13 09:14:44 crc kubenswrapper[4971]: I1213 09:14:44.531458 4971 provider.go:102] Refreshing cache for provider: *credentialprovider.defaultDockerConfigProvider Dec 13 09:14:46 crc kubenswrapper[4971]: I1213 09:14:46.564589 4971 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-6cl4h" event={"ID":"1ae30452-bee6-43ce-845a-cb32e1c4c5a9","Type":"ContainerStarted","Data":"ca0bf2b3e5994eda4c3f14b2380682dea8076e363b1fcef2d91325bb746aeacf"} Dec 13 09:14:47 crc kubenswrapper[4971]: I1213 09:14:47.596832 4971 generic.go:334] "Generic (PLEG): container finished" podID="1ae30452-bee6-43ce-845a-cb32e1c4c5a9" containerID="ca0bf2b3e5994eda4c3f14b2380682dea8076e363b1fcef2d91325bb746aeacf" exitCode=0 Dec 13 09:14:47 crc kubenswrapper[4971]: I1213 09:14:47.597069 4971 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-6cl4h" event={"ID":"1ae30452-bee6-43ce-845a-cb32e1c4c5a9","Type":"ContainerDied","Data":"ca0bf2b3e5994eda4c3f14b2380682dea8076e363b1fcef2d91325bb746aeacf"} Dec 13 09:14:47 crc kubenswrapper[4971]: I1213 09:14:47.769779 4971 scope.go:117] "RemoveContainer" containerID="0d8863d17d13d9fdfa7767d29d08739bc5a2b2755f517451ebb079161a29c2fb" Dec 13 09:14:47 crc kubenswrapper[4971]: E1213 09:14:47.770028 4971 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-82xjz_openshift-machine-config-operator(e0c4af12-edda-41d7-9b44-f87396174ef2)\"" pod="openshift-machine-config-operator/machine-config-daemon-82xjz" podUID="e0c4af12-edda-41d7-9b44-f87396174ef2" Dec 13 09:14:48 crc kubenswrapper[4971]: I1213 09:14:48.613065 4971 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-6cl4h" event={"ID":"1ae30452-bee6-43ce-845a-cb32e1c4c5a9","Type":"ContainerStarted","Data":"c84eb6fb0ff2fd892d6743861479609881cb5f1febf67c190319ff342d5680ad"} Dec 13 09:14:48 crc kubenswrapper[4971]: I1213 09:14:48.642740 4971 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-operators-6cl4h" podStartSLOduration=3.148777397 podStartE2EDuration="6.642710396s" podCreationTimestamp="2025-12-13 09:14:42 +0000 UTC" firstStartedPulling="2025-12-13 09:14:44.531180769 +0000 UTC m=+8741.135590217" lastFinishedPulling="2025-12-13 09:14:48.025113768 +0000 UTC m=+8744.629523216" observedRunningTime="2025-12-13 09:14:48.641388903 +0000 UTC m=+8745.245798361" watchObservedRunningTime="2025-12-13 09:14:48.642710396 +0000 UTC m=+8745.247119844" Dec 13 09:14:53 crc kubenswrapper[4971]: I1213 09:14:53.076394 4971 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-operators-6cl4h" Dec 13 09:14:53 crc kubenswrapper[4971]: I1213 09:14:53.077493 4971 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-operators-6cl4h" Dec 13 09:14:54 crc kubenswrapper[4971]: I1213 09:14:54.415183 4971 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-marketplace/redhat-operators-6cl4h" podUID="1ae30452-bee6-43ce-845a-cb32e1c4c5a9" containerName="registry-server" probeResult="failure" output=< Dec 13 09:14:54 crc kubenswrapper[4971]: timeout: failed to connect service ":50051" within 1s Dec 13 09:14:54 crc kubenswrapper[4971]: > Dec 13 09:14:58 crc kubenswrapper[4971]: I1213 09:14:58.770061 4971 scope.go:117] "RemoveContainer" containerID="0d8863d17d13d9fdfa7767d29d08739bc5a2b2755f517451ebb079161a29c2fb" Dec 13 09:14:58 crc kubenswrapper[4971]: E1213 09:14:58.771480 4971 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-82xjz_openshift-machine-config-operator(e0c4af12-edda-41d7-9b44-f87396174ef2)\"" pod="openshift-machine-config-operator/machine-config-daemon-82xjz" podUID="e0c4af12-edda-41d7-9b44-f87396174ef2" Dec 13 09:15:00 crc kubenswrapper[4971]: I1213 09:15:00.171450 4971 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29426955-r7l5j"] Dec 13 09:15:00 crc kubenswrapper[4971]: I1213 09:15:00.173191 4971 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29426955-r7l5j" Dec 13 09:15:00 crc kubenswrapper[4971]: I1213 09:15:00.175716 4971 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"collect-profiles-dockercfg-kzf4t" Dec 13 09:15:00 crc kubenswrapper[4971]: I1213 09:15:00.180478 4971 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-operator-lifecycle-manager"/"collect-profiles-config" Dec 13 09:15:00 crc kubenswrapper[4971]: I1213 09:15:00.188793 4971 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29426955-r7l5j"] Dec 13 09:15:00 crc kubenswrapper[4971]: I1213 09:15:00.234430 4971 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-g44bx\" (UniqueName: \"kubernetes.io/projected/cfdc3509-4b86-4dbd-b1c9-5415caeba0af-kube-api-access-g44bx\") pod \"collect-profiles-29426955-r7l5j\" (UID: \"cfdc3509-4b86-4dbd-b1c9-5415caeba0af\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29426955-r7l5j" Dec 13 09:15:00 crc kubenswrapper[4971]: I1213 09:15:00.234730 4971 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/cfdc3509-4b86-4dbd-b1c9-5415caeba0af-secret-volume\") pod \"collect-profiles-29426955-r7l5j\" (UID: \"cfdc3509-4b86-4dbd-b1c9-5415caeba0af\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29426955-r7l5j" Dec 13 09:15:00 crc kubenswrapper[4971]: I1213 09:15:00.234766 4971 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/cfdc3509-4b86-4dbd-b1c9-5415caeba0af-config-volume\") pod \"collect-profiles-29426955-r7l5j\" (UID: \"cfdc3509-4b86-4dbd-b1c9-5415caeba0af\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29426955-r7l5j" Dec 13 09:15:00 crc kubenswrapper[4971]: I1213 09:15:00.337260 4971 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/cfdc3509-4b86-4dbd-b1c9-5415caeba0af-secret-volume\") pod \"collect-profiles-29426955-r7l5j\" (UID: \"cfdc3509-4b86-4dbd-b1c9-5415caeba0af\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29426955-r7l5j" Dec 13 09:15:00 crc kubenswrapper[4971]: I1213 09:15:00.337319 4971 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/cfdc3509-4b86-4dbd-b1c9-5415caeba0af-config-volume\") pod \"collect-profiles-29426955-r7l5j\" (UID: \"cfdc3509-4b86-4dbd-b1c9-5415caeba0af\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29426955-r7l5j" Dec 13 09:15:00 crc kubenswrapper[4971]: I1213 09:15:00.337403 4971 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-g44bx\" (UniqueName: \"kubernetes.io/projected/cfdc3509-4b86-4dbd-b1c9-5415caeba0af-kube-api-access-g44bx\") pod \"collect-profiles-29426955-r7l5j\" (UID: \"cfdc3509-4b86-4dbd-b1c9-5415caeba0af\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29426955-r7l5j" Dec 13 09:15:00 crc kubenswrapper[4971]: I1213 09:15:00.339486 4971 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/cfdc3509-4b86-4dbd-b1c9-5415caeba0af-config-volume\") pod \"collect-profiles-29426955-r7l5j\" (UID: \"cfdc3509-4b86-4dbd-b1c9-5415caeba0af\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29426955-r7l5j" Dec 13 09:15:00 crc kubenswrapper[4971]: I1213 09:15:00.344983 4971 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/cfdc3509-4b86-4dbd-b1c9-5415caeba0af-secret-volume\") pod \"collect-profiles-29426955-r7l5j\" (UID: \"cfdc3509-4b86-4dbd-b1c9-5415caeba0af\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29426955-r7l5j" Dec 13 09:15:00 crc kubenswrapper[4971]: I1213 09:15:00.361906 4971 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-g44bx\" (UniqueName: \"kubernetes.io/projected/cfdc3509-4b86-4dbd-b1c9-5415caeba0af-kube-api-access-g44bx\") pod \"collect-profiles-29426955-r7l5j\" (UID: \"cfdc3509-4b86-4dbd-b1c9-5415caeba0af\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29426955-r7l5j" Dec 13 09:15:00 crc kubenswrapper[4971]: I1213 09:15:00.712406 4971 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29426955-r7l5j" Dec 13 09:15:01 crc kubenswrapper[4971]: I1213 09:15:01.343755 4971 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29426955-r7l5j"] Dec 13 09:15:01 crc kubenswrapper[4971]: I1213 09:15:01.837767 4971 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29426955-r7l5j" event={"ID":"cfdc3509-4b86-4dbd-b1c9-5415caeba0af","Type":"ContainerStarted","Data":"307d51742e36e442484f196286c5c4b9e94f330d69038cbd9a5ae6fe9acbec4c"} Dec 13 09:15:01 crc kubenswrapper[4971]: I1213 09:15:01.837843 4971 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29426955-r7l5j" event={"ID":"cfdc3509-4b86-4dbd-b1c9-5415caeba0af","Type":"ContainerStarted","Data":"15c72d885ca7b37be803fe253592739b375074e1295ad94f5c53012718acc184"} Dec 13 09:15:01 crc kubenswrapper[4971]: I1213 09:15:01.871683 4971 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-operator-lifecycle-manager/collect-profiles-29426955-r7l5j" podStartSLOduration=1.871658577 podStartE2EDuration="1.871658577s" podCreationTimestamp="2025-12-13 09:15:00 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-13 09:15:01.860778429 +0000 UTC m=+8758.465187897" watchObservedRunningTime="2025-12-13 09:15:01.871658577 +0000 UTC m=+8758.476068025" Dec 13 09:15:02 crc kubenswrapper[4971]: I1213 09:15:02.860475 4971 generic.go:334] "Generic (PLEG): container finished" podID="cfdc3509-4b86-4dbd-b1c9-5415caeba0af" containerID="307d51742e36e442484f196286c5c4b9e94f330d69038cbd9a5ae6fe9acbec4c" exitCode=0 Dec 13 09:15:02 crc kubenswrapper[4971]: I1213 09:15:02.860781 4971 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29426955-r7l5j" event={"ID":"cfdc3509-4b86-4dbd-b1c9-5415caeba0af","Type":"ContainerDied","Data":"307d51742e36e442484f196286c5c4b9e94f330d69038cbd9a5ae6fe9acbec4c"} Dec 13 09:15:03 crc kubenswrapper[4971]: I1213 09:15:03.157180 4971 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-operators-6cl4h" Dec 13 09:15:03 crc kubenswrapper[4971]: I1213 09:15:03.276322 4971 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-operators-6cl4h" Dec 13 09:15:03 crc kubenswrapper[4971]: I1213 09:15:03.649469 4971 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-6cl4h"] Dec 13 09:15:04 crc kubenswrapper[4971]: I1213 09:15:04.296621 4971 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29426955-r7l5j" Dec 13 09:15:04 crc kubenswrapper[4971]: I1213 09:15:04.445555 4971 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/cfdc3509-4b86-4dbd-b1c9-5415caeba0af-secret-volume\") pod \"cfdc3509-4b86-4dbd-b1c9-5415caeba0af\" (UID: \"cfdc3509-4b86-4dbd-b1c9-5415caeba0af\") " Dec 13 09:15:04 crc kubenswrapper[4971]: I1213 09:15:04.445673 4971 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-g44bx\" (UniqueName: \"kubernetes.io/projected/cfdc3509-4b86-4dbd-b1c9-5415caeba0af-kube-api-access-g44bx\") pod \"cfdc3509-4b86-4dbd-b1c9-5415caeba0af\" (UID: \"cfdc3509-4b86-4dbd-b1c9-5415caeba0af\") " Dec 13 09:15:04 crc kubenswrapper[4971]: I1213 09:15:04.445713 4971 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/cfdc3509-4b86-4dbd-b1c9-5415caeba0af-config-volume\") pod \"cfdc3509-4b86-4dbd-b1c9-5415caeba0af\" (UID: \"cfdc3509-4b86-4dbd-b1c9-5415caeba0af\") " Dec 13 09:15:04 crc kubenswrapper[4971]: I1213 09:15:04.446477 4971 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/cfdc3509-4b86-4dbd-b1c9-5415caeba0af-config-volume" (OuterVolumeSpecName: "config-volume") pod "cfdc3509-4b86-4dbd-b1c9-5415caeba0af" (UID: "cfdc3509-4b86-4dbd-b1c9-5415caeba0af"). InnerVolumeSpecName "config-volume". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 13 09:15:04 crc kubenswrapper[4971]: I1213 09:15:04.447748 4971 reconciler_common.go:293] "Volume detached for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/cfdc3509-4b86-4dbd-b1c9-5415caeba0af-config-volume\") on node \"crc\" DevicePath \"\"" Dec 13 09:15:04 crc kubenswrapper[4971]: I1213 09:15:04.454292 4971 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/cfdc3509-4b86-4dbd-b1c9-5415caeba0af-secret-volume" (OuterVolumeSpecName: "secret-volume") pod "cfdc3509-4b86-4dbd-b1c9-5415caeba0af" (UID: "cfdc3509-4b86-4dbd-b1c9-5415caeba0af"). InnerVolumeSpecName "secret-volume". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 13 09:15:04 crc kubenswrapper[4971]: I1213 09:15:04.457054 4971 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/cfdc3509-4b86-4dbd-b1c9-5415caeba0af-kube-api-access-g44bx" (OuterVolumeSpecName: "kube-api-access-g44bx") pod "cfdc3509-4b86-4dbd-b1c9-5415caeba0af" (UID: "cfdc3509-4b86-4dbd-b1c9-5415caeba0af"). InnerVolumeSpecName "kube-api-access-g44bx". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 13 09:15:04 crc kubenswrapper[4971]: I1213 09:15:04.551259 4971 reconciler_common.go:293] "Volume detached for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/cfdc3509-4b86-4dbd-b1c9-5415caeba0af-secret-volume\") on node \"crc\" DevicePath \"\"" Dec 13 09:15:04 crc kubenswrapper[4971]: I1213 09:15:04.551310 4971 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-g44bx\" (UniqueName: \"kubernetes.io/projected/cfdc3509-4b86-4dbd-b1c9-5415caeba0af-kube-api-access-g44bx\") on node \"crc\" DevicePath \"\"" Dec 13 09:15:04 crc kubenswrapper[4971]: I1213 09:15:04.631487 4971 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29426910-x72qh"] Dec 13 09:15:04 crc kubenswrapper[4971]: I1213 09:15:04.675687 4971 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29426910-x72qh"] Dec 13 09:15:04 crc kubenswrapper[4971]: I1213 09:15:04.885562 4971 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29426955-r7l5j" event={"ID":"cfdc3509-4b86-4dbd-b1c9-5415caeba0af","Type":"ContainerDied","Data":"15c72d885ca7b37be803fe253592739b375074e1295ad94f5c53012718acc184"} Dec 13 09:15:04 crc kubenswrapper[4971]: I1213 09:15:04.885614 4971 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29426955-r7l5j" Dec 13 09:15:04 crc kubenswrapper[4971]: I1213 09:15:04.885635 4971 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="15c72d885ca7b37be803fe253592739b375074e1295ad94f5c53012718acc184" Dec 13 09:15:04 crc kubenswrapper[4971]: I1213 09:15:04.885787 4971 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-operators-6cl4h" podUID="1ae30452-bee6-43ce-845a-cb32e1c4c5a9" containerName="registry-server" containerID="cri-o://c84eb6fb0ff2fd892d6743861479609881cb5f1febf67c190319ff342d5680ad" gracePeriod=2 Dec 13 09:15:05 crc kubenswrapper[4971]: I1213 09:15:05.469856 4971 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-6cl4h" Dec 13 09:15:05 crc kubenswrapper[4971]: I1213 09:15:05.587539 4971 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/1ae30452-bee6-43ce-845a-cb32e1c4c5a9-catalog-content\") pod \"1ae30452-bee6-43ce-845a-cb32e1c4c5a9\" (UID: \"1ae30452-bee6-43ce-845a-cb32e1c4c5a9\") " Dec 13 09:15:05 crc kubenswrapper[4971]: I1213 09:15:05.588211 4971 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-zwn2q\" (UniqueName: \"kubernetes.io/projected/1ae30452-bee6-43ce-845a-cb32e1c4c5a9-kube-api-access-zwn2q\") pod \"1ae30452-bee6-43ce-845a-cb32e1c4c5a9\" (UID: \"1ae30452-bee6-43ce-845a-cb32e1c4c5a9\") " Dec 13 09:15:05 crc kubenswrapper[4971]: I1213 09:15:05.588391 4971 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/1ae30452-bee6-43ce-845a-cb32e1c4c5a9-utilities\") pod \"1ae30452-bee6-43ce-845a-cb32e1c4c5a9\" (UID: \"1ae30452-bee6-43ce-845a-cb32e1c4c5a9\") " Dec 13 09:15:05 crc kubenswrapper[4971]: I1213 09:15:05.590417 4971 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/1ae30452-bee6-43ce-845a-cb32e1c4c5a9-utilities" (OuterVolumeSpecName: "utilities") pod "1ae30452-bee6-43ce-845a-cb32e1c4c5a9" (UID: "1ae30452-bee6-43ce-845a-cb32e1c4c5a9"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 13 09:15:05 crc kubenswrapper[4971]: I1213 09:15:05.601825 4971 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/1ae30452-bee6-43ce-845a-cb32e1c4c5a9-kube-api-access-zwn2q" (OuterVolumeSpecName: "kube-api-access-zwn2q") pod "1ae30452-bee6-43ce-845a-cb32e1c4c5a9" (UID: "1ae30452-bee6-43ce-845a-cb32e1c4c5a9"). InnerVolumeSpecName "kube-api-access-zwn2q". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 13 09:15:05 crc kubenswrapper[4971]: I1213 09:15:05.691658 4971 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-zwn2q\" (UniqueName: \"kubernetes.io/projected/1ae30452-bee6-43ce-845a-cb32e1c4c5a9-kube-api-access-zwn2q\") on node \"crc\" DevicePath \"\"" Dec 13 09:15:05 crc kubenswrapper[4971]: I1213 09:15:05.691703 4971 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/1ae30452-bee6-43ce-845a-cb32e1c4c5a9-utilities\") on node \"crc\" DevicePath \"\"" Dec 13 09:15:05 crc kubenswrapper[4971]: I1213 09:15:05.716330 4971 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/1ae30452-bee6-43ce-845a-cb32e1c4c5a9-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "1ae30452-bee6-43ce-845a-cb32e1c4c5a9" (UID: "1ae30452-bee6-43ce-845a-cb32e1c4c5a9"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 13 09:15:05 crc kubenswrapper[4971]: I1213 09:15:05.782967 4971 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="d0ddcbe1-f21e-4b5d-b4a1-8aa5c45d41f7" path="/var/lib/kubelet/pods/d0ddcbe1-f21e-4b5d-b4a1-8aa5c45d41f7/volumes" Dec 13 09:15:05 crc kubenswrapper[4971]: I1213 09:15:05.795618 4971 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/1ae30452-bee6-43ce-845a-cb32e1c4c5a9-catalog-content\") on node \"crc\" DevicePath \"\"" Dec 13 09:15:05 crc kubenswrapper[4971]: I1213 09:15:05.900359 4971 generic.go:334] "Generic (PLEG): container finished" podID="1ae30452-bee6-43ce-845a-cb32e1c4c5a9" containerID="c84eb6fb0ff2fd892d6743861479609881cb5f1febf67c190319ff342d5680ad" exitCode=0 Dec 13 09:15:05 crc kubenswrapper[4971]: I1213 09:15:05.900426 4971 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-6cl4h" event={"ID":"1ae30452-bee6-43ce-845a-cb32e1c4c5a9","Type":"ContainerDied","Data":"c84eb6fb0ff2fd892d6743861479609881cb5f1febf67c190319ff342d5680ad"} Dec 13 09:15:05 crc kubenswrapper[4971]: I1213 09:15:05.900472 4971 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-6cl4h" event={"ID":"1ae30452-bee6-43ce-845a-cb32e1c4c5a9","Type":"ContainerDied","Data":"5b49e89154ee75e92190e3e91575838785f9365d664f829878c469b977d39a70"} Dec 13 09:15:05 crc kubenswrapper[4971]: I1213 09:15:05.900503 4971 scope.go:117] "RemoveContainer" containerID="c84eb6fb0ff2fd892d6743861479609881cb5f1febf67c190319ff342d5680ad" Dec 13 09:15:05 crc kubenswrapper[4971]: I1213 09:15:05.900739 4971 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-6cl4h" Dec 13 09:15:05 crc kubenswrapper[4971]: I1213 09:15:05.933115 4971 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-6cl4h"] Dec 13 09:15:05 crc kubenswrapper[4971]: I1213 09:15:05.939946 4971 scope.go:117] "RemoveContainer" containerID="ca0bf2b3e5994eda4c3f14b2380682dea8076e363b1fcef2d91325bb746aeacf" Dec 13 09:15:05 crc kubenswrapper[4971]: I1213 09:15:05.945970 4971 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-operators-6cl4h"] Dec 13 09:15:05 crc kubenswrapper[4971]: I1213 09:15:05.995083 4971 scope.go:117] "RemoveContainer" containerID="84914ffd0bc27d5642e6ff188ea3744e8cde51fdd3836079ba928ba9e88af078" Dec 13 09:15:06 crc kubenswrapper[4971]: I1213 09:15:06.034642 4971 scope.go:117] "RemoveContainer" containerID="c84eb6fb0ff2fd892d6743861479609881cb5f1febf67c190319ff342d5680ad" Dec 13 09:15:06 crc kubenswrapper[4971]: E1213 09:15:06.035335 4971 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"c84eb6fb0ff2fd892d6743861479609881cb5f1febf67c190319ff342d5680ad\": container with ID starting with c84eb6fb0ff2fd892d6743861479609881cb5f1febf67c190319ff342d5680ad not found: ID does not exist" containerID="c84eb6fb0ff2fd892d6743861479609881cb5f1febf67c190319ff342d5680ad" Dec 13 09:15:06 crc kubenswrapper[4971]: I1213 09:15:06.035380 4971 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"c84eb6fb0ff2fd892d6743861479609881cb5f1febf67c190319ff342d5680ad"} err="failed to get container status \"c84eb6fb0ff2fd892d6743861479609881cb5f1febf67c190319ff342d5680ad\": rpc error: code = NotFound desc = could not find container \"c84eb6fb0ff2fd892d6743861479609881cb5f1febf67c190319ff342d5680ad\": container with ID starting with c84eb6fb0ff2fd892d6743861479609881cb5f1febf67c190319ff342d5680ad not found: ID does not exist" Dec 13 09:15:06 crc kubenswrapper[4971]: I1213 09:15:06.035407 4971 scope.go:117] "RemoveContainer" containerID="ca0bf2b3e5994eda4c3f14b2380682dea8076e363b1fcef2d91325bb746aeacf" Dec 13 09:15:06 crc kubenswrapper[4971]: E1213 09:15:06.035966 4971 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"ca0bf2b3e5994eda4c3f14b2380682dea8076e363b1fcef2d91325bb746aeacf\": container with ID starting with ca0bf2b3e5994eda4c3f14b2380682dea8076e363b1fcef2d91325bb746aeacf not found: ID does not exist" containerID="ca0bf2b3e5994eda4c3f14b2380682dea8076e363b1fcef2d91325bb746aeacf" Dec 13 09:15:06 crc kubenswrapper[4971]: I1213 09:15:06.035989 4971 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"ca0bf2b3e5994eda4c3f14b2380682dea8076e363b1fcef2d91325bb746aeacf"} err="failed to get container status \"ca0bf2b3e5994eda4c3f14b2380682dea8076e363b1fcef2d91325bb746aeacf\": rpc error: code = NotFound desc = could not find container \"ca0bf2b3e5994eda4c3f14b2380682dea8076e363b1fcef2d91325bb746aeacf\": container with ID starting with ca0bf2b3e5994eda4c3f14b2380682dea8076e363b1fcef2d91325bb746aeacf not found: ID does not exist" Dec 13 09:15:06 crc kubenswrapper[4971]: I1213 09:15:06.036005 4971 scope.go:117] "RemoveContainer" containerID="84914ffd0bc27d5642e6ff188ea3744e8cde51fdd3836079ba928ba9e88af078" Dec 13 09:15:06 crc kubenswrapper[4971]: E1213 09:15:06.036549 4971 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"84914ffd0bc27d5642e6ff188ea3744e8cde51fdd3836079ba928ba9e88af078\": container with ID starting with 84914ffd0bc27d5642e6ff188ea3744e8cde51fdd3836079ba928ba9e88af078 not found: ID does not exist" containerID="84914ffd0bc27d5642e6ff188ea3744e8cde51fdd3836079ba928ba9e88af078" Dec 13 09:15:06 crc kubenswrapper[4971]: I1213 09:15:06.036625 4971 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"84914ffd0bc27d5642e6ff188ea3744e8cde51fdd3836079ba928ba9e88af078"} err="failed to get container status \"84914ffd0bc27d5642e6ff188ea3744e8cde51fdd3836079ba928ba9e88af078\": rpc error: code = NotFound desc = could not find container \"84914ffd0bc27d5642e6ff188ea3744e8cde51fdd3836079ba928ba9e88af078\": container with ID starting with 84914ffd0bc27d5642e6ff188ea3744e8cde51fdd3836079ba928ba9e88af078 not found: ID does not exist" Dec 13 09:15:07 crc kubenswrapper[4971]: I1213 09:15:07.779040 4971 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="1ae30452-bee6-43ce-845a-cb32e1c4c5a9" path="/var/lib/kubelet/pods/1ae30452-bee6-43ce-845a-cb32e1c4c5a9/volumes" Dec 13 09:15:09 crc kubenswrapper[4971]: I1213 09:15:09.769196 4971 scope.go:117] "RemoveContainer" containerID="0d8863d17d13d9fdfa7767d29d08739bc5a2b2755f517451ebb079161a29c2fb" Dec 13 09:15:09 crc kubenswrapper[4971]: E1213 09:15:09.770057 4971 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-82xjz_openshift-machine-config-operator(e0c4af12-edda-41d7-9b44-f87396174ef2)\"" pod="openshift-machine-config-operator/machine-config-daemon-82xjz" podUID="e0c4af12-edda-41d7-9b44-f87396174ef2" Dec 13 09:15:24 crc kubenswrapper[4971]: I1213 09:15:24.769064 4971 scope.go:117] "RemoveContainer" containerID="0d8863d17d13d9fdfa7767d29d08739bc5a2b2755f517451ebb079161a29c2fb" Dec 13 09:15:24 crc kubenswrapper[4971]: E1213 09:15:24.770314 4971 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-82xjz_openshift-machine-config-operator(e0c4af12-edda-41d7-9b44-f87396174ef2)\"" pod="openshift-machine-config-operator/machine-config-daemon-82xjz" podUID="e0c4af12-edda-41d7-9b44-f87396174ef2" Dec 13 09:15:27 crc kubenswrapper[4971]: I1213 09:15:27.015603 4971 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-marketplace-5xl8m"] Dec 13 09:15:27 crc kubenswrapper[4971]: E1213 09:15:27.017026 4971 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="1ae30452-bee6-43ce-845a-cb32e1c4c5a9" containerName="extract-utilities" Dec 13 09:15:27 crc kubenswrapper[4971]: I1213 09:15:27.017052 4971 state_mem.go:107] "Deleted CPUSet assignment" podUID="1ae30452-bee6-43ce-845a-cb32e1c4c5a9" containerName="extract-utilities" Dec 13 09:15:27 crc kubenswrapper[4971]: E1213 09:15:27.017072 4971 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="cfdc3509-4b86-4dbd-b1c9-5415caeba0af" containerName="collect-profiles" Dec 13 09:15:27 crc kubenswrapper[4971]: I1213 09:15:27.017080 4971 state_mem.go:107] "Deleted CPUSet assignment" podUID="cfdc3509-4b86-4dbd-b1c9-5415caeba0af" containerName="collect-profiles" Dec 13 09:15:27 crc kubenswrapper[4971]: E1213 09:15:27.017107 4971 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="1ae30452-bee6-43ce-845a-cb32e1c4c5a9" containerName="extract-content" Dec 13 09:15:27 crc kubenswrapper[4971]: I1213 09:15:27.017116 4971 state_mem.go:107] "Deleted CPUSet assignment" podUID="1ae30452-bee6-43ce-845a-cb32e1c4c5a9" containerName="extract-content" Dec 13 09:15:27 crc kubenswrapper[4971]: E1213 09:15:27.017131 4971 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="1ae30452-bee6-43ce-845a-cb32e1c4c5a9" containerName="registry-server" Dec 13 09:15:27 crc kubenswrapper[4971]: I1213 09:15:27.017139 4971 state_mem.go:107] "Deleted CPUSet assignment" podUID="1ae30452-bee6-43ce-845a-cb32e1c4c5a9" containerName="registry-server" Dec 13 09:15:27 crc kubenswrapper[4971]: I1213 09:15:27.017428 4971 memory_manager.go:354] "RemoveStaleState removing state" podUID="cfdc3509-4b86-4dbd-b1c9-5415caeba0af" containerName="collect-profiles" Dec 13 09:15:27 crc kubenswrapper[4971]: I1213 09:15:27.017441 4971 memory_manager.go:354] "RemoveStaleState removing state" podUID="1ae30452-bee6-43ce-845a-cb32e1c4c5a9" containerName="registry-server" Dec 13 09:15:27 crc kubenswrapper[4971]: I1213 09:15:27.019428 4971 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-5xl8m" Dec 13 09:15:27 crc kubenswrapper[4971]: I1213 09:15:27.041182 4971 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-5xl8m"] Dec 13 09:15:27 crc kubenswrapper[4971]: I1213 09:15:27.236000 4971 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-4cqnv\" (UniqueName: \"kubernetes.io/projected/24ea1b93-4980-4fc2-820a-09663fbb6efa-kube-api-access-4cqnv\") pod \"redhat-marketplace-5xl8m\" (UID: \"24ea1b93-4980-4fc2-820a-09663fbb6efa\") " pod="openshift-marketplace/redhat-marketplace-5xl8m" Dec 13 09:15:27 crc kubenswrapper[4971]: I1213 09:15:27.236156 4971 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/24ea1b93-4980-4fc2-820a-09663fbb6efa-catalog-content\") pod \"redhat-marketplace-5xl8m\" (UID: \"24ea1b93-4980-4fc2-820a-09663fbb6efa\") " pod="openshift-marketplace/redhat-marketplace-5xl8m" Dec 13 09:15:27 crc kubenswrapper[4971]: I1213 09:15:27.236656 4971 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/24ea1b93-4980-4fc2-820a-09663fbb6efa-utilities\") pod \"redhat-marketplace-5xl8m\" (UID: \"24ea1b93-4980-4fc2-820a-09663fbb6efa\") " pod="openshift-marketplace/redhat-marketplace-5xl8m" Dec 13 09:15:27 crc kubenswrapper[4971]: I1213 09:15:27.339179 4971 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-4cqnv\" (UniqueName: \"kubernetes.io/projected/24ea1b93-4980-4fc2-820a-09663fbb6efa-kube-api-access-4cqnv\") pod \"redhat-marketplace-5xl8m\" (UID: \"24ea1b93-4980-4fc2-820a-09663fbb6efa\") " pod="openshift-marketplace/redhat-marketplace-5xl8m" Dec 13 09:15:27 crc kubenswrapper[4971]: I1213 09:15:27.339249 4971 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/24ea1b93-4980-4fc2-820a-09663fbb6efa-catalog-content\") pod \"redhat-marketplace-5xl8m\" (UID: \"24ea1b93-4980-4fc2-820a-09663fbb6efa\") " pod="openshift-marketplace/redhat-marketplace-5xl8m" Dec 13 09:15:27 crc kubenswrapper[4971]: I1213 09:15:27.339351 4971 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/24ea1b93-4980-4fc2-820a-09663fbb6efa-utilities\") pod \"redhat-marketplace-5xl8m\" (UID: \"24ea1b93-4980-4fc2-820a-09663fbb6efa\") " pod="openshift-marketplace/redhat-marketplace-5xl8m" Dec 13 09:15:27 crc kubenswrapper[4971]: I1213 09:15:27.339889 4971 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/24ea1b93-4980-4fc2-820a-09663fbb6efa-utilities\") pod \"redhat-marketplace-5xl8m\" (UID: \"24ea1b93-4980-4fc2-820a-09663fbb6efa\") " pod="openshift-marketplace/redhat-marketplace-5xl8m" Dec 13 09:15:27 crc kubenswrapper[4971]: I1213 09:15:27.340136 4971 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/24ea1b93-4980-4fc2-820a-09663fbb6efa-catalog-content\") pod \"redhat-marketplace-5xl8m\" (UID: \"24ea1b93-4980-4fc2-820a-09663fbb6efa\") " pod="openshift-marketplace/redhat-marketplace-5xl8m" Dec 13 09:15:27 crc kubenswrapper[4971]: I1213 09:15:27.368668 4971 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-4cqnv\" (UniqueName: \"kubernetes.io/projected/24ea1b93-4980-4fc2-820a-09663fbb6efa-kube-api-access-4cqnv\") pod \"redhat-marketplace-5xl8m\" (UID: \"24ea1b93-4980-4fc2-820a-09663fbb6efa\") " pod="openshift-marketplace/redhat-marketplace-5xl8m" Dec 13 09:15:27 crc kubenswrapper[4971]: I1213 09:15:27.566896 4971 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-5xl8m" Dec 13 09:15:28 crc kubenswrapper[4971]: I1213 09:15:28.270041 4971 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-5xl8m"] Dec 13 09:15:28 crc kubenswrapper[4971]: I1213 09:15:28.450290 4971 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-5xl8m" event={"ID":"24ea1b93-4980-4fc2-820a-09663fbb6efa","Type":"ContainerStarted","Data":"18baf1cda8cd0ab32f6d1ce7a97cbc1e345c02573170d124e8f7c178f5281aaa"} Dec 13 09:15:29 crc kubenswrapper[4971]: I1213 09:15:29.300312 4971 scope.go:117] "RemoveContainer" containerID="227abad82435cbc9348715e19bc56b3b34cbbcb3ae0118da06c5fe5cb0a701ac" Dec 13 09:15:30 crc kubenswrapper[4971]: I1213 09:15:30.484191 4971 generic.go:334] "Generic (PLEG): container finished" podID="24ea1b93-4980-4fc2-820a-09663fbb6efa" containerID="f3a3262d8de3c42bac2cdd2846e9d88ea1ca29c6918a3d262bdea507c51cb1f4" exitCode=0 Dec 13 09:15:30 crc kubenswrapper[4971]: I1213 09:15:30.484337 4971 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-5xl8m" event={"ID":"24ea1b93-4980-4fc2-820a-09663fbb6efa","Type":"ContainerDied","Data":"f3a3262d8de3c42bac2cdd2846e9d88ea1ca29c6918a3d262bdea507c51cb1f4"} Dec 13 09:15:32 crc kubenswrapper[4971]: I1213 09:15:32.508961 4971 generic.go:334] "Generic (PLEG): container finished" podID="24ea1b93-4980-4fc2-820a-09663fbb6efa" containerID="abbfa33f22c6e44cae5da928a6620e83023b9c309cc8494ed8d1a24cd2cdac83" exitCode=0 Dec 13 09:15:32 crc kubenswrapper[4971]: I1213 09:15:32.509080 4971 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-5xl8m" event={"ID":"24ea1b93-4980-4fc2-820a-09663fbb6efa","Type":"ContainerDied","Data":"abbfa33f22c6e44cae5da928a6620e83023b9c309cc8494ed8d1a24cd2cdac83"} Dec 13 09:15:33 crc kubenswrapper[4971]: I1213 09:15:33.524456 4971 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-5xl8m" event={"ID":"24ea1b93-4980-4fc2-820a-09663fbb6efa","Type":"ContainerStarted","Data":"17254125dd87b4c22f2eeabb53c968dd31e4f26909c6ef9656a235538aa7c64b"} Dec 13 09:15:33 crc kubenswrapper[4971]: I1213 09:15:33.555909 4971 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-marketplace-5xl8m" podStartSLOduration=5.07392092 podStartE2EDuration="7.555874458s" podCreationTimestamp="2025-12-13 09:15:26 +0000 UTC" firstStartedPulling="2025-12-13 09:15:30.487470319 +0000 UTC m=+8787.091879767" lastFinishedPulling="2025-12-13 09:15:32.969423857 +0000 UTC m=+8789.573833305" observedRunningTime="2025-12-13 09:15:33.544874078 +0000 UTC m=+8790.149283546" watchObservedRunningTime="2025-12-13 09:15:33.555874458 +0000 UTC m=+8790.160283916" Dec 13 09:15:37 crc kubenswrapper[4971]: I1213 09:15:37.573891 4971 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-marketplace-5xl8m" Dec 13 09:15:37 crc kubenswrapper[4971]: I1213 09:15:37.574745 4971 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-marketplace-5xl8m" Dec 13 09:15:37 crc kubenswrapper[4971]: I1213 09:15:37.646745 4971 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-marketplace-5xl8m" Dec 13 09:15:37 crc kubenswrapper[4971]: I1213 09:15:37.769147 4971 scope.go:117] "RemoveContainer" containerID="0d8863d17d13d9fdfa7767d29d08739bc5a2b2755f517451ebb079161a29c2fb" Dec 13 09:15:37 crc kubenswrapper[4971]: E1213 09:15:37.769478 4971 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-82xjz_openshift-machine-config-operator(e0c4af12-edda-41d7-9b44-f87396174ef2)\"" pod="openshift-machine-config-operator/machine-config-daemon-82xjz" podUID="e0c4af12-edda-41d7-9b44-f87396174ef2" Dec 13 09:15:38 crc kubenswrapper[4971]: I1213 09:15:38.693670 4971 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-marketplace-5xl8m" Dec 13 09:15:38 crc kubenswrapper[4971]: I1213 09:15:38.792257 4971 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-5xl8m"] Dec 13 09:15:40 crc kubenswrapper[4971]: I1213 09:15:40.615102 4971 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-marketplace-5xl8m" podUID="24ea1b93-4980-4fc2-820a-09663fbb6efa" containerName="registry-server" containerID="cri-o://17254125dd87b4c22f2eeabb53c968dd31e4f26909c6ef9656a235538aa7c64b" gracePeriod=2 Dec 13 09:15:41 crc kubenswrapper[4971]: I1213 09:15:41.247779 4971 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-5xl8m" Dec 13 09:15:41 crc kubenswrapper[4971]: I1213 09:15:41.625617 4971 generic.go:334] "Generic (PLEG): container finished" podID="24ea1b93-4980-4fc2-820a-09663fbb6efa" containerID="17254125dd87b4c22f2eeabb53c968dd31e4f26909c6ef9656a235538aa7c64b" exitCode=0 Dec 13 09:15:41 crc kubenswrapper[4971]: I1213 09:15:41.625675 4971 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-5xl8m" event={"ID":"24ea1b93-4980-4fc2-820a-09663fbb6efa","Type":"ContainerDied","Data":"17254125dd87b4c22f2eeabb53c968dd31e4f26909c6ef9656a235538aa7c64b"} Dec 13 09:15:41 crc kubenswrapper[4971]: I1213 09:15:41.625709 4971 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-5xl8m" Dec 13 09:15:41 crc kubenswrapper[4971]: I1213 09:15:41.625758 4971 scope.go:117] "RemoveContainer" containerID="17254125dd87b4c22f2eeabb53c968dd31e4f26909c6ef9656a235538aa7c64b" Dec 13 09:15:41 crc kubenswrapper[4971]: I1213 09:15:41.625728 4971 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-5xl8m" event={"ID":"24ea1b93-4980-4fc2-820a-09663fbb6efa","Type":"ContainerDied","Data":"18baf1cda8cd0ab32f6d1ce7a97cbc1e345c02573170d124e8f7c178f5281aaa"} Dec 13 09:15:41 crc kubenswrapper[4971]: I1213 09:15:41.644993 4971 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/24ea1b93-4980-4fc2-820a-09663fbb6efa-utilities\") pod \"24ea1b93-4980-4fc2-820a-09663fbb6efa\" (UID: \"24ea1b93-4980-4fc2-820a-09663fbb6efa\") " Dec 13 09:15:41 crc kubenswrapper[4971]: I1213 09:15:41.645127 4971 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-4cqnv\" (UniqueName: \"kubernetes.io/projected/24ea1b93-4980-4fc2-820a-09663fbb6efa-kube-api-access-4cqnv\") pod \"24ea1b93-4980-4fc2-820a-09663fbb6efa\" (UID: \"24ea1b93-4980-4fc2-820a-09663fbb6efa\") " Dec 13 09:15:41 crc kubenswrapper[4971]: I1213 09:15:41.645168 4971 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/24ea1b93-4980-4fc2-820a-09663fbb6efa-catalog-content\") pod \"24ea1b93-4980-4fc2-820a-09663fbb6efa\" (UID: \"24ea1b93-4980-4fc2-820a-09663fbb6efa\") " Dec 13 09:15:41 crc kubenswrapper[4971]: I1213 09:15:41.646358 4971 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/24ea1b93-4980-4fc2-820a-09663fbb6efa-utilities" (OuterVolumeSpecName: "utilities") pod "24ea1b93-4980-4fc2-820a-09663fbb6efa" (UID: "24ea1b93-4980-4fc2-820a-09663fbb6efa"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 13 09:15:41 crc kubenswrapper[4971]: I1213 09:15:41.668853 4971 scope.go:117] "RemoveContainer" containerID="abbfa33f22c6e44cae5da928a6620e83023b9c309cc8494ed8d1a24cd2cdac83" Dec 13 09:15:41 crc kubenswrapper[4971]: I1213 09:15:41.669185 4971 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/24ea1b93-4980-4fc2-820a-09663fbb6efa-kube-api-access-4cqnv" (OuterVolumeSpecName: "kube-api-access-4cqnv") pod "24ea1b93-4980-4fc2-820a-09663fbb6efa" (UID: "24ea1b93-4980-4fc2-820a-09663fbb6efa"). InnerVolumeSpecName "kube-api-access-4cqnv". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 13 09:15:41 crc kubenswrapper[4971]: I1213 09:15:41.669776 4971 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/24ea1b93-4980-4fc2-820a-09663fbb6efa-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "24ea1b93-4980-4fc2-820a-09663fbb6efa" (UID: "24ea1b93-4980-4fc2-820a-09663fbb6efa"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 13 09:15:41 crc kubenswrapper[4971]: I1213 09:15:41.725349 4971 scope.go:117] "RemoveContainer" containerID="f3a3262d8de3c42bac2cdd2846e9d88ea1ca29c6918a3d262bdea507c51cb1f4" Dec 13 09:15:41 crc kubenswrapper[4971]: I1213 09:15:41.748620 4971 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/24ea1b93-4980-4fc2-820a-09663fbb6efa-utilities\") on node \"crc\" DevicePath \"\"" Dec 13 09:15:41 crc kubenswrapper[4971]: I1213 09:15:41.749905 4971 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-4cqnv\" (UniqueName: \"kubernetes.io/projected/24ea1b93-4980-4fc2-820a-09663fbb6efa-kube-api-access-4cqnv\") on node \"crc\" DevicePath \"\"" Dec 13 09:15:41 crc kubenswrapper[4971]: I1213 09:15:41.749964 4971 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/24ea1b93-4980-4fc2-820a-09663fbb6efa-catalog-content\") on node \"crc\" DevicePath \"\"" Dec 13 09:15:41 crc kubenswrapper[4971]: I1213 09:15:41.774438 4971 scope.go:117] "RemoveContainer" containerID="17254125dd87b4c22f2eeabb53c968dd31e4f26909c6ef9656a235538aa7c64b" Dec 13 09:15:41 crc kubenswrapper[4971]: E1213 09:15:41.775510 4971 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"17254125dd87b4c22f2eeabb53c968dd31e4f26909c6ef9656a235538aa7c64b\": container with ID starting with 17254125dd87b4c22f2eeabb53c968dd31e4f26909c6ef9656a235538aa7c64b not found: ID does not exist" containerID="17254125dd87b4c22f2eeabb53c968dd31e4f26909c6ef9656a235538aa7c64b" Dec 13 09:15:41 crc kubenswrapper[4971]: I1213 09:15:41.775585 4971 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"17254125dd87b4c22f2eeabb53c968dd31e4f26909c6ef9656a235538aa7c64b"} err="failed to get container status \"17254125dd87b4c22f2eeabb53c968dd31e4f26909c6ef9656a235538aa7c64b\": rpc error: code = NotFound desc = could not find container \"17254125dd87b4c22f2eeabb53c968dd31e4f26909c6ef9656a235538aa7c64b\": container with ID starting with 17254125dd87b4c22f2eeabb53c968dd31e4f26909c6ef9656a235538aa7c64b not found: ID does not exist" Dec 13 09:15:41 crc kubenswrapper[4971]: I1213 09:15:41.775617 4971 scope.go:117] "RemoveContainer" containerID="abbfa33f22c6e44cae5da928a6620e83023b9c309cc8494ed8d1a24cd2cdac83" Dec 13 09:15:41 crc kubenswrapper[4971]: E1213 09:15:41.776046 4971 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"abbfa33f22c6e44cae5da928a6620e83023b9c309cc8494ed8d1a24cd2cdac83\": container with ID starting with abbfa33f22c6e44cae5da928a6620e83023b9c309cc8494ed8d1a24cd2cdac83 not found: ID does not exist" containerID="abbfa33f22c6e44cae5da928a6620e83023b9c309cc8494ed8d1a24cd2cdac83" Dec 13 09:15:41 crc kubenswrapper[4971]: I1213 09:15:41.776089 4971 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"abbfa33f22c6e44cae5da928a6620e83023b9c309cc8494ed8d1a24cd2cdac83"} err="failed to get container status \"abbfa33f22c6e44cae5da928a6620e83023b9c309cc8494ed8d1a24cd2cdac83\": rpc error: code = NotFound desc = could not find container \"abbfa33f22c6e44cae5da928a6620e83023b9c309cc8494ed8d1a24cd2cdac83\": container with ID starting with abbfa33f22c6e44cae5da928a6620e83023b9c309cc8494ed8d1a24cd2cdac83 not found: ID does not exist" Dec 13 09:15:41 crc kubenswrapper[4971]: I1213 09:15:41.776118 4971 scope.go:117] "RemoveContainer" containerID="f3a3262d8de3c42bac2cdd2846e9d88ea1ca29c6918a3d262bdea507c51cb1f4" Dec 13 09:15:41 crc kubenswrapper[4971]: E1213 09:15:41.776360 4971 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"f3a3262d8de3c42bac2cdd2846e9d88ea1ca29c6918a3d262bdea507c51cb1f4\": container with ID starting with f3a3262d8de3c42bac2cdd2846e9d88ea1ca29c6918a3d262bdea507c51cb1f4 not found: ID does not exist" containerID="f3a3262d8de3c42bac2cdd2846e9d88ea1ca29c6918a3d262bdea507c51cb1f4" Dec 13 09:15:41 crc kubenswrapper[4971]: I1213 09:15:41.776392 4971 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"f3a3262d8de3c42bac2cdd2846e9d88ea1ca29c6918a3d262bdea507c51cb1f4"} err="failed to get container status \"f3a3262d8de3c42bac2cdd2846e9d88ea1ca29c6918a3d262bdea507c51cb1f4\": rpc error: code = NotFound desc = could not find container \"f3a3262d8de3c42bac2cdd2846e9d88ea1ca29c6918a3d262bdea507c51cb1f4\": container with ID starting with f3a3262d8de3c42bac2cdd2846e9d88ea1ca29c6918a3d262bdea507c51cb1f4 not found: ID does not exist" Dec 13 09:15:41 crc kubenswrapper[4971]: I1213 09:15:41.958581 4971 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-5xl8m"] Dec 13 09:15:41 crc kubenswrapper[4971]: I1213 09:15:41.971984 4971 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-marketplace-5xl8m"] Dec 13 09:15:43 crc kubenswrapper[4971]: I1213 09:15:43.780841 4971 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="24ea1b93-4980-4fc2-820a-09663fbb6efa" path="/var/lib/kubelet/pods/24ea1b93-4980-4fc2-820a-09663fbb6efa/volumes" Dec 13 09:15:49 crc kubenswrapper[4971]: I1213 09:15:49.770304 4971 scope.go:117] "RemoveContainer" containerID="0d8863d17d13d9fdfa7767d29d08739bc5a2b2755f517451ebb079161a29c2fb" Dec 13 09:15:49 crc kubenswrapper[4971]: E1213 09:15:49.771106 4971 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-82xjz_openshift-machine-config-operator(e0c4af12-edda-41d7-9b44-f87396174ef2)\"" pod="openshift-machine-config-operator/machine-config-daemon-82xjz" podUID="e0c4af12-edda-41d7-9b44-f87396174ef2" Dec 13 09:16:02 crc kubenswrapper[4971]: I1213 09:16:02.769970 4971 scope.go:117] "RemoveContainer" containerID="0d8863d17d13d9fdfa7767d29d08739bc5a2b2755f517451ebb079161a29c2fb" Dec 13 09:16:02 crc kubenswrapper[4971]: E1213 09:16:02.772314 4971 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-82xjz_openshift-machine-config-operator(e0c4af12-edda-41d7-9b44-f87396174ef2)\"" pod="openshift-machine-config-operator/machine-config-daemon-82xjz" podUID="e0c4af12-edda-41d7-9b44-f87396174ef2" Dec 13 09:16:15 crc kubenswrapper[4971]: I1213 09:16:15.768969 4971 scope.go:117] "RemoveContainer" containerID="0d8863d17d13d9fdfa7767d29d08739bc5a2b2755f517451ebb079161a29c2fb" Dec 13 09:16:15 crc kubenswrapper[4971]: E1213 09:16:15.770271 4971 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-82xjz_openshift-machine-config-operator(e0c4af12-edda-41d7-9b44-f87396174ef2)\"" pod="openshift-machine-config-operator/machine-config-daemon-82xjz" podUID="e0c4af12-edda-41d7-9b44-f87396174ef2" Dec 13 09:16:27 crc kubenswrapper[4971]: I1213 09:16:27.774315 4971 scope.go:117] "RemoveContainer" containerID="0d8863d17d13d9fdfa7767d29d08739bc5a2b2755f517451ebb079161a29c2fb" Dec 13 09:16:28 crc kubenswrapper[4971]: I1213 09:16:28.273022 4971 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-82xjz" event={"ID":"e0c4af12-edda-41d7-9b44-f87396174ef2","Type":"ContainerStarted","Data":"05ac64a4f7ae8e1342c3b29cc16132aac9d985c5e8acf6af1eb069f94d71e38d"} Dec 13 09:18:46 crc kubenswrapper[4971]: I1213 09:18:46.153748 4971 patch_prober.go:28] interesting pod/machine-config-daemon-82xjz container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 13 09:18:46 crc kubenswrapper[4971]: I1213 09:18:46.154917 4971 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-82xjz" podUID="e0c4af12-edda-41d7-9b44-f87396174ef2" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 13 09:19:16 crc kubenswrapper[4971]: I1213 09:19:16.153870 4971 patch_prober.go:28] interesting pod/machine-config-daemon-82xjz container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 13 09:19:16 crc kubenswrapper[4971]: I1213 09:19:16.154804 4971 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-82xjz" podUID="e0c4af12-edda-41d7-9b44-f87396174ef2" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 13 09:19:46 crc kubenswrapper[4971]: I1213 09:19:46.153547 4971 patch_prober.go:28] interesting pod/machine-config-daemon-82xjz container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 13 09:19:46 crc kubenswrapper[4971]: I1213 09:19:46.154340 4971 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-82xjz" podUID="e0c4af12-edda-41d7-9b44-f87396174ef2" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 13 09:19:46 crc kubenswrapper[4971]: I1213 09:19:46.154397 4971 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-82xjz" Dec 13 09:19:46 crc kubenswrapper[4971]: I1213 09:19:46.155424 4971 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"05ac64a4f7ae8e1342c3b29cc16132aac9d985c5e8acf6af1eb069f94d71e38d"} pod="openshift-machine-config-operator/machine-config-daemon-82xjz" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Dec 13 09:19:46 crc kubenswrapper[4971]: I1213 09:19:46.155504 4971 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-82xjz" podUID="e0c4af12-edda-41d7-9b44-f87396174ef2" containerName="machine-config-daemon" containerID="cri-o://05ac64a4f7ae8e1342c3b29cc16132aac9d985c5e8acf6af1eb069f94d71e38d" gracePeriod=600 Dec 13 09:19:46 crc kubenswrapper[4971]: I1213 09:19:46.656921 4971 generic.go:334] "Generic (PLEG): container finished" podID="e0c4af12-edda-41d7-9b44-f87396174ef2" containerID="05ac64a4f7ae8e1342c3b29cc16132aac9d985c5e8acf6af1eb069f94d71e38d" exitCode=0 Dec 13 09:19:46 crc kubenswrapper[4971]: I1213 09:19:46.657251 4971 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-82xjz" event={"ID":"e0c4af12-edda-41d7-9b44-f87396174ef2","Type":"ContainerDied","Data":"05ac64a4f7ae8e1342c3b29cc16132aac9d985c5e8acf6af1eb069f94d71e38d"} Dec 13 09:19:46 crc kubenswrapper[4971]: I1213 09:19:46.657927 4971 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-82xjz" event={"ID":"e0c4af12-edda-41d7-9b44-f87396174ef2","Type":"ContainerStarted","Data":"e616f2406e276879960fe4d0372c2d2a34f8c21a80e5202a5083f7852225bab9"} Dec 13 09:19:46 crc kubenswrapper[4971]: I1213 09:19:46.657962 4971 scope.go:117] "RemoveContainer" containerID="0d8863d17d13d9fdfa7767d29d08739bc5a2b2755f517451ebb079161a29c2fb" Dec 13 09:19:53 crc kubenswrapper[4971]: I1213 09:19:53.250582 4971 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/certified-operators-5dcqp"] Dec 13 09:19:53 crc kubenswrapper[4971]: E1213 09:19:53.251956 4971 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="24ea1b93-4980-4fc2-820a-09663fbb6efa" containerName="registry-server" Dec 13 09:19:53 crc kubenswrapper[4971]: I1213 09:19:53.251977 4971 state_mem.go:107] "Deleted CPUSet assignment" podUID="24ea1b93-4980-4fc2-820a-09663fbb6efa" containerName="registry-server" Dec 13 09:19:53 crc kubenswrapper[4971]: E1213 09:19:53.252022 4971 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="24ea1b93-4980-4fc2-820a-09663fbb6efa" containerName="extract-content" Dec 13 09:19:53 crc kubenswrapper[4971]: I1213 09:19:53.252031 4971 state_mem.go:107] "Deleted CPUSet assignment" podUID="24ea1b93-4980-4fc2-820a-09663fbb6efa" containerName="extract-content" Dec 13 09:19:53 crc kubenswrapper[4971]: E1213 09:19:53.252062 4971 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="24ea1b93-4980-4fc2-820a-09663fbb6efa" containerName="extract-utilities" Dec 13 09:19:53 crc kubenswrapper[4971]: I1213 09:19:53.252073 4971 state_mem.go:107] "Deleted CPUSet assignment" podUID="24ea1b93-4980-4fc2-820a-09663fbb6efa" containerName="extract-utilities" Dec 13 09:19:53 crc kubenswrapper[4971]: I1213 09:19:53.252320 4971 memory_manager.go:354] "RemoveStaleState removing state" podUID="24ea1b93-4980-4fc2-820a-09663fbb6efa" containerName="registry-server" Dec 13 09:19:53 crc kubenswrapper[4971]: I1213 09:19:53.254287 4971 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-5dcqp" Dec 13 09:19:53 crc kubenswrapper[4971]: I1213 09:19:53.265601 4971 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-5dcqp"] Dec 13 09:19:53 crc kubenswrapper[4971]: I1213 09:19:53.350457 4971 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/e175976e-1561-43ce-9236-043e1a7b8d09-utilities\") pod \"certified-operators-5dcqp\" (UID: \"e175976e-1561-43ce-9236-043e1a7b8d09\") " pod="openshift-marketplace/certified-operators-5dcqp" Dec 13 09:19:53 crc kubenswrapper[4971]: I1213 09:19:53.350577 4971 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-6gqxb\" (UniqueName: \"kubernetes.io/projected/e175976e-1561-43ce-9236-043e1a7b8d09-kube-api-access-6gqxb\") pod \"certified-operators-5dcqp\" (UID: \"e175976e-1561-43ce-9236-043e1a7b8d09\") " pod="openshift-marketplace/certified-operators-5dcqp" Dec 13 09:19:53 crc kubenswrapper[4971]: I1213 09:19:53.350623 4971 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/e175976e-1561-43ce-9236-043e1a7b8d09-catalog-content\") pod \"certified-operators-5dcqp\" (UID: \"e175976e-1561-43ce-9236-043e1a7b8d09\") " pod="openshift-marketplace/certified-operators-5dcqp" Dec 13 09:19:53 crc kubenswrapper[4971]: I1213 09:19:53.453221 4971 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-6gqxb\" (UniqueName: \"kubernetes.io/projected/e175976e-1561-43ce-9236-043e1a7b8d09-kube-api-access-6gqxb\") pod \"certified-operators-5dcqp\" (UID: \"e175976e-1561-43ce-9236-043e1a7b8d09\") " pod="openshift-marketplace/certified-operators-5dcqp" Dec 13 09:19:53 crc kubenswrapper[4971]: I1213 09:19:53.453606 4971 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/e175976e-1561-43ce-9236-043e1a7b8d09-catalog-content\") pod \"certified-operators-5dcqp\" (UID: \"e175976e-1561-43ce-9236-043e1a7b8d09\") " pod="openshift-marketplace/certified-operators-5dcqp" Dec 13 09:19:53 crc kubenswrapper[4971]: I1213 09:19:53.453805 4971 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/e175976e-1561-43ce-9236-043e1a7b8d09-utilities\") pod \"certified-operators-5dcqp\" (UID: \"e175976e-1561-43ce-9236-043e1a7b8d09\") " pod="openshift-marketplace/certified-operators-5dcqp" Dec 13 09:19:53 crc kubenswrapper[4971]: I1213 09:19:53.454399 4971 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/e175976e-1561-43ce-9236-043e1a7b8d09-catalog-content\") pod \"certified-operators-5dcqp\" (UID: \"e175976e-1561-43ce-9236-043e1a7b8d09\") " pod="openshift-marketplace/certified-operators-5dcqp" Dec 13 09:19:53 crc kubenswrapper[4971]: I1213 09:19:53.454481 4971 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/e175976e-1561-43ce-9236-043e1a7b8d09-utilities\") pod \"certified-operators-5dcqp\" (UID: \"e175976e-1561-43ce-9236-043e1a7b8d09\") " pod="openshift-marketplace/certified-operators-5dcqp" Dec 13 09:19:53 crc kubenswrapper[4971]: I1213 09:19:53.479565 4971 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-6gqxb\" (UniqueName: \"kubernetes.io/projected/e175976e-1561-43ce-9236-043e1a7b8d09-kube-api-access-6gqxb\") pod \"certified-operators-5dcqp\" (UID: \"e175976e-1561-43ce-9236-043e1a7b8d09\") " pod="openshift-marketplace/certified-operators-5dcqp" Dec 13 09:19:53 crc kubenswrapper[4971]: I1213 09:19:53.601755 4971 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-5dcqp" Dec 13 09:19:54 crc kubenswrapper[4971]: I1213 09:19:54.135202 4971 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-5dcqp"] Dec 13 09:19:54 crc kubenswrapper[4971]: I1213 09:19:54.772049 4971 generic.go:334] "Generic (PLEG): container finished" podID="e175976e-1561-43ce-9236-043e1a7b8d09" containerID="7b452fd2f6a994de6221c963943d54f3f2b3632c846c1440b9502c4d1a1e3980" exitCode=0 Dec 13 09:19:54 crc kubenswrapper[4971]: I1213 09:19:54.772340 4971 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-5dcqp" event={"ID":"e175976e-1561-43ce-9236-043e1a7b8d09","Type":"ContainerDied","Data":"7b452fd2f6a994de6221c963943d54f3f2b3632c846c1440b9502c4d1a1e3980"} Dec 13 09:19:54 crc kubenswrapper[4971]: I1213 09:19:54.772421 4971 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-5dcqp" event={"ID":"e175976e-1561-43ce-9236-043e1a7b8d09","Type":"ContainerStarted","Data":"4f2541cdb1f85422318cc832c445c1474e91befd12732ad2a33e7e5c7f53cae5"} Dec 13 09:19:54 crc kubenswrapper[4971]: I1213 09:19:54.774410 4971 provider.go:102] Refreshing cache for provider: *credentialprovider.defaultDockerConfigProvider Dec 13 09:19:56 crc kubenswrapper[4971]: I1213 09:19:56.798667 4971 generic.go:334] "Generic (PLEG): container finished" podID="e175976e-1561-43ce-9236-043e1a7b8d09" containerID="dd228866c7cbe7b3e09320792e49b920776c795f34a5cf8560280c428f0ebba4" exitCode=0 Dec 13 09:19:56 crc kubenswrapper[4971]: I1213 09:19:56.798752 4971 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-5dcqp" event={"ID":"e175976e-1561-43ce-9236-043e1a7b8d09","Type":"ContainerDied","Data":"dd228866c7cbe7b3e09320792e49b920776c795f34a5cf8560280c428f0ebba4"} Dec 13 09:19:57 crc kubenswrapper[4971]: I1213 09:19:57.813089 4971 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-5dcqp" event={"ID":"e175976e-1561-43ce-9236-043e1a7b8d09","Type":"ContainerStarted","Data":"9a4dd6cac8337836463da1b12e70f1c8119591086b4daf5b7de99c9d2e9d1253"} Dec 13 09:19:57 crc kubenswrapper[4971]: I1213 09:19:57.850249 4971 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/certified-operators-5dcqp" podStartSLOduration=2.297556127 podStartE2EDuration="4.850215806s" podCreationTimestamp="2025-12-13 09:19:53 +0000 UTC" firstStartedPulling="2025-12-13 09:19:54.774118788 +0000 UTC m=+9051.378528236" lastFinishedPulling="2025-12-13 09:19:57.326778467 +0000 UTC m=+9053.931187915" observedRunningTime="2025-12-13 09:19:57.835938295 +0000 UTC m=+9054.440347753" watchObservedRunningTime="2025-12-13 09:19:57.850215806 +0000 UTC m=+9054.454625254" Dec 13 09:20:03 crc kubenswrapper[4971]: I1213 09:20:03.602143 4971 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/certified-operators-5dcqp" Dec 13 09:20:03 crc kubenswrapper[4971]: I1213 09:20:03.602906 4971 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/certified-operators-5dcqp" Dec 13 09:20:03 crc kubenswrapper[4971]: I1213 09:20:03.673959 4971 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/certified-operators-5dcqp" Dec 13 09:20:03 crc kubenswrapper[4971]: I1213 09:20:03.944567 4971 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/certified-operators-5dcqp" Dec 13 09:20:04 crc kubenswrapper[4971]: I1213 09:20:04.008579 4971 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-5dcqp"] Dec 13 09:20:05 crc kubenswrapper[4971]: I1213 09:20:05.911017 4971 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/certified-operators-5dcqp" podUID="e175976e-1561-43ce-9236-043e1a7b8d09" containerName="registry-server" containerID="cri-o://9a4dd6cac8337836463da1b12e70f1c8119591086b4daf5b7de99c9d2e9d1253" gracePeriod=2 Dec 13 09:20:06 crc kubenswrapper[4971]: E1213 09:20:06.166318 4971 cadvisor_stats_provider.go:516] "Partial failure issuing cadvisor.ContainerInfoV2" err="partial failures: [\"/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pode175976e_1561_43ce_9236_043e1a7b8d09.slice/crio-conmon-9a4dd6cac8337836463da1b12e70f1c8119591086b4daf5b7de99c9d2e9d1253.scope\": RecentStats: unable to find data in memory cache]" Dec 13 09:20:06 crc kubenswrapper[4971]: I1213 09:20:06.558050 4971 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-5dcqp" Dec 13 09:20:06 crc kubenswrapper[4971]: I1213 09:20:06.704890 4971 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/e175976e-1561-43ce-9236-043e1a7b8d09-utilities\") pod \"e175976e-1561-43ce-9236-043e1a7b8d09\" (UID: \"e175976e-1561-43ce-9236-043e1a7b8d09\") " Dec 13 09:20:06 crc kubenswrapper[4971]: I1213 09:20:06.704983 4971 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/e175976e-1561-43ce-9236-043e1a7b8d09-catalog-content\") pod \"e175976e-1561-43ce-9236-043e1a7b8d09\" (UID: \"e175976e-1561-43ce-9236-043e1a7b8d09\") " Dec 13 09:20:06 crc kubenswrapper[4971]: I1213 09:20:06.705116 4971 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-6gqxb\" (UniqueName: \"kubernetes.io/projected/e175976e-1561-43ce-9236-043e1a7b8d09-kube-api-access-6gqxb\") pod \"e175976e-1561-43ce-9236-043e1a7b8d09\" (UID: \"e175976e-1561-43ce-9236-043e1a7b8d09\") " Dec 13 09:20:06 crc kubenswrapper[4971]: I1213 09:20:06.706164 4971 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/e175976e-1561-43ce-9236-043e1a7b8d09-utilities" (OuterVolumeSpecName: "utilities") pod "e175976e-1561-43ce-9236-043e1a7b8d09" (UID: "e175976e-1561-43ce-9236-043e1a7b8d09"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 13 09:20:06 crc kubenswrapper[4971]: I1213 09:20:06.716729 4971 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/e175976e-1561-43ce-9236-043e1a7b8d09-kube-api-access-6gqxb" (OuterVolumeSpecName: "kube-api-access-6gqxb") pod "e175976e-1561-43ce-9236-043e1a7b8d09" (UID: "e175976e-1561-43ce-9236-043e1a7b8d09"). InnerVolumeSpecName "kube-api-access-6gqxb". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 13 09:20:06 crc kubenswrapper[4971]: I1213 09:20:06.809423 4971 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-6gqxb\" (UniqueName: \"kubernetes.io/projected/e175976e-1561-43ce-9236-043e1a7b8d09-kube-api-access-6gqxb\") on node \"crc\" DevicePath \"\"" Dec 13 09:20:06 crc kubenswrapper[4971]: I1213 09:20:06.809487 4971 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/e175976e-1561-43ce-9236-043e1a7b8d09-utilities\") on node \"crc\" DevicePath \"\"" Dec 13 09:20:06 crc kubenswrapper[4971]: I1213 09:20:06.923766 4971 generic.go:334] "Generic (PLEG): container finished" podID="e175976e-1561-43ce-9236-043e1a7b8d09" containerID="9a4dd6cac8337836463da1b12e70f1c8119591086b4daf5b7de99c9d2e9d1253" exitCode=0 Dec 13 09:20:06 crc kubenswrapper[4971]: I1213 09:20:06.923831 4971 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-5dcqp" event={"ID":"e175976e-1561-43ce-9236-043e1a7b8d09","Type":"ContainerDied","Data":"9a4dd6cac8337836463da1b12e70f1c8119591086b4daf5b7de99c9d2e9d1253"} Dec 13 09:20:06 crc kubenswrapper[4971]: I1213 09:20:06.923856 4971 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-5dcqp" Dec 13 09:20:06 crc kubenswrapper[4971]: I1213 09:20:06.923874 4971 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-5dcqp" event={"ID":"e175976e-1561-43ce-9236-043e1a7b8d09","Type":"ContainerDied","Data":"4f2541cdb1f85422318cc832c445c1474e91befd12732ad2a33e7e5c7f53cae5"} Dec 13 09:20:06 crc kubenswrapper[4971]: I1213 09:20:06.923903 4971 scope.go:117] "RemoveContainer" containerID="9a4dd6cac8337836463da1b12e70f1c8119591086b4daf5b7de99c9d2e9d1253" Dec 13 09:20:06 crc kubenswrapper[4971]: I1213 09:20:06.948107 4971 scope.go:117] "RemoveContainer" containerID="dd228866c7cbe7b3e09320792e49b920776c795f34a5cf8560280c428f0ebba4" Dec 13 09:20:06 crc kubenswrapper[4971]: I1213 09:20:06.988650 4971 scope.go:117] "RemoveContainer" containerID="7b452fd2f6a994de6221c963943d54f3f2b3632c846c1440b9502c4d1a1e3980" Dec 13 09:20:07 crc kubenswrapper[4971]: I1213 09:20:07.022312 4971 scope.go:117] "RemoveContainer" containerID="9a4dd6cac8337836463da1b12e70f1c8119591086b4daf5b7de99c9d2e9d1253" Dec 13 09:20:07 crc kubenswrapper[4971]: E1213 09:20:07.023231 4971 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"9a4dd6cac8337836463da1b12e70f1c8119591086b4daf5b7de99c9d2e9d1253\": container with ID starting with 9a4dd6cac8337836463da1b12e70f1c8119591086b4daf5b7de99c9d2e9d1253 not found: ID does not exist" containerID="9a4dd6cac8337836463da1b12e70f1c8119591086b4daf5b7de99c9d2e9d1253" Dec 13 09:20:07 crc kubenswrapper[4971]: I1213 09:20:07.023312 4971 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"9a4dd6cac8337836463da1b12e70f1c8119591086b4daf5b7de99c9d2e9d1253"} err="failed to get container status \"9a4dd6cac8337836463da1b12e70f1c8119591086b4daf5b7de99c9d2e9d1253\": rpc error: code = NotFound desc = could not find container \"9a4dd6cac8337836463da1b12e70f1c8119591086b4daf5b7de99c9d2e9d1253\": container with ID starting with 9a4dd6cac8337836463da1b12e70f1c8119591086b4daf5b7de99c9d2e9d1253 not found: ID does not exist" Dec 13 09:20:07 crc kubenswrapper[4971]: I1213 09:20:07.023358 4971 scope.go:117] "RemoveContainer" containerID="dd228866c7cbe7b3e09320792e49b920776c795f34a5cf8560280c428f0ebba4" Dec 13 09:20:07 crc kubenswrapper[4971]: E1213 09:20:07.023982 4971 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"dd228866c7cbe7b3e09320792e49b920776c795f34a5cf8560280c428f0ebba4\": container with ID starting with dd228866c7cbe7b3e09320792e49b920776c795f34a5cf8560280c428f0ebba4 not found: ID does not exist" containerID="dd228866c7cbe7b3e09320792e49b920776c795f34a5cf8560280c428f0ebba4" Dec 13 09:20:07 crc kubenswrapper[4971]: I1213 09:20:07.024015 4971 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"dd228866c7cbe7b3e09320792e49b920776c795f34a5cf8560280c428f0ebba4"} err="failed to get container status \"dd228866c7cbe7b3e09320792e49b920776c795f34a5cf8560280c428f0ebba4\": rpc error: code = NotFound desc = could not find container \"dd228866c7cbe7b3e09320792e49b920776c795f34a5cf8560280c428f0ebba4\": container with ID starting with dd228866c7cbe7b3e09320792e49b920776c795f34a5cf8560280c428f0ebba4 not found: ID does not exist" Dec 13 09:20:07 crc kubenswrapper[4971]: I1213 09:20:07.024036 4971 scope.go:117] "RemoveContainer" containerID="7b452fd2f6a994de6221c963943d54f3f2b3632c846c1440b9502c4d1a1e3980" Dec 13 09:20:07 crc kubenswrapper[4971]: E1213 09:20:07.024383 4971 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"7b452fd2f6a994de6221c963943d54f3f2b3632c846c1440b9502c4d1a1e3980\": container with ID starting with 7b452fd2f6a994de6221c963943d54f3f2b3632c846c1440b9502c4d1a1e3980 not found: ID does not exist" containerID="7b452fd2f6a994de6221c963943d54f3f2b3632c846c1440b9502c4d1a1e3980" Dec 13 09:20:07 crc kubenswrapper[4971]: I1213 09:20:07.024420 4971 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"7b452fd2f6a994de6221c963943d54f3f2b3632c846c1440b9502c4d1a1e3980"} err="failed to get container status \"7b452fd2f6a994de6221c963943d54f3f2b3632c846c1440b9502c4d1a1e3980\": rpc error: code = NotFound desc = could not find container \"7b452fd2f6a994de6221c963943d54f3f2b3632c846c1440b9502c4d1a1e3980\": container with ID starting with 7b452fd2f6a994de6221c963943d54f3f2b3632c846c1440b9502c4d1a1e3980 not found: ID does not exist" Dec 13 09:20:07 crc kubenswrapper[4971]: I1213 09:20:07.325431 4971 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/community-operators-z6fvq"] Dec 13 09:20:07 crc kubenswrapper[4971]: E1213 09:20:07.325952 4971 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="e175976e-1561-43ce-9236-043e1a7b8d09" containerName="registry-server" Dec 13 09:20:07 crc kubenswrapper[4971]: I1213 09:20:07.325976 4971 state_mem.go:107] "Deleted CPUSet assignment" podUID="e175976e-1561-43ce-9236-043e1a7b8d09" containerName="registry-server" Dec 13 09:20:07 crc kubenswrapper[4971]: E1213 09:20:07.326003 4971 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="e175976e-1561-43ce-9236-043e1a7b8d09" containerName="extract-content" Dec 13 09:20:07 crc kubenswrapper[4971]: I1213 09:20:07.326011 4971 state_mem.go:107] "Deleted CPUSet assignment" podUID="e175976e-1561-43ce-9236-043e1a7b8d09" containerName="extract-content" Dec 13 09:20:07 crc kubenswrapper[4971]: E1213 09:20:07.326028 4971 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="e175976e-1561-43ce-9236-043e1a7b8d09" containerName="extract-utilities" Dec 13 09:20:07 crc kubenswrapper[4971]: I1213 09:20:07.326035 4971 state_mem.go:107] "Deleted CPUSet assignment" podUID="e175976e-1561-43ce-9236-043e1a7b8d09" containerName="extract-utilities" Dec 13 09:20:07 crc kubenswrapper[4971]: I1213 09:20:07.326229 4971 memory_manager.go:354] "RemoveStaleState removing state" podUID="e175976e-1561-43ce-9236-043e1a7b8d09" containerName="registry-server" Dec 13 09:20:07 crc kubenswrapper[4971]: I1213 09:20:07.328501 4971 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-z6fvq" Dec 13 09:20:07 crc kubenswrapper[4971]: I1213 09:20:07.366897 4971 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-z6fvq"] Dec 13 09:20:07 crc kubenswrapper[4971]: I1213 09:20:07.398157 4971 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/e175976e-1561-43ce-9236-043e1a7b8d09-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "e175976e-1561-43ce-9236-043e1a7b8d09" (UID: "e175976e-1561-43ce-9236-043e1a7b8d09"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 13 09:20:07 crc kubenswrapper[4971]: I1213 09:20:07.424741 4971 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/c63d9282-272e-4eeb-a93a-917099b3b147-catalog-content\") pod \"community-operators-z6fvq\" (UID: \"c63d9282-272e-4eeb-a93a-917099b3b147\") " pod="openshift-marketplace/community-operators-z6fvq" Dec 13 09:20:07 crc kubenswrapper[4971]: I1213 09:20:07.424893 4971 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-dwrbf\" (UniqueName: \"kubernetes.io/projected/c63d9282-272e-4eeb-a93a-917099b3b147-kube-api-access-dwrbf\") pod \"community-operators-z6fvq\" (UID: \"c63d9282-272e-4eeb-a93a-917099b3b147\") " pod="openshift-marketplace/community-operators-z6fvq" Dec 13 09:20:07 crc kubenswrapper[4971]: I1213 09:20:07.425039 4971 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/c63d9282-272e-4eeb-a93a-917099b3b147-utilities\") pod \"community-operators-z6fvq\" (UID: \"c63d9282-272e-4eeb-a93a-917099b3b147\") " pod="openshift-marketplace/community-operators-z6fvq" Dec 13 09:20:07 crc kubenswrapper[4971]: I1213 09:20:07.425219 4971 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/e175976e-1561-43ce-9236-043e1a7b8d09-catalog-content\") on node \"crc\" DevicePath \"\"" Dec 13 09:20:07 crc kubenswrapper[4971]: I1213 09:20:07.527701 4971 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/c63d9282-272e-4eeb-a93a-917099b3b147-catalog-content\") pod \"community-operators-z6fvq\" (UID: \"c63d9282-272e-4eeb-a93a-917099b3b147\") " pod="openshift-marketplace/community-operators-z6fvq" Dec 13 09:20:07 crc kubenswrapper[4971]: I1213 09:20:07.527820 4971 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-dwrbf\" (UniqueName: \"kubernetes.io/projected/c63d9282-272e-4eeb-a93a-917099b3b147-kube-api-access-dwrbf\") pod \"community-operators-z6fvq\" (UID: \"c63d9282-272e-4eeb-a93a-917099b3b147\") " pod="openshift-marketplace/community-operators-z6fvq" Dec 13 09:20:07 crc kubenswrapper[4971]: I1213 09:20:07.527911 4971 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/c63d9282-272e-4eeb-a93a-917099b3b147-utilities\") pod \"community-operators-z6fvq\" (UID: \"c63d9282-272e-4eeb-a93a-917099b3b147\") " pod="openshift-marketplace/community-operators-z6fvq" Dec 13 09:20:07 crc kubenswrapper[4971]: I1213 09:20:07.528610 4971 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/c63d9282-272e-4eeb-a93a-917099b3b147-utilities\") pod \"community-operators-z6fvq\" (UID: \"c63d9282-272e-4eeb-a93a-917099b3b147\") " pod="openshift-marketplace/community-operators-z6fvq" Dec 13 09:20:07 crc kubenswrapper[4971]: I1213 09:20:07.528616 4971 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/c63d9282-272e-4eeb-a93a-917099b3b147-catalog-content\") pod \"community-operators-z6fvq\" (UID: \"c63d9282-272e-4eeb-a93a-917099b3b147\") " pod="openshift-marketplace/community-operators-z6fvq" Dec 13 09:20:07 crc kubenswrapper[4971]: I1213 09:20:07.557589 4971 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-dwrbf\" (UniqueName: \"kubernetes.io/projected/c63d9282-272e-4eeb-a93a-917099b3b147-kube-api-access-dwrbf\") pod \"community-operators-z6fvq\" (UID: \"c63d9282-272e-4eeb-a93a-917099b3b147\") " pod="openshift-marketplace/community-operators-z6fvq" Dec 13 09:20:07 crc kubenswrapper[4971]: I1213 09:20:07.579316 4971 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-5dcqp"] Dec 13 09:20:07 crc kubenswrapper[4971]: I1213 09:20:07.587721 4971 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/certified-operators-5dcqp"] Dec 13 09:20:07 crc kubenswrapper[4971]: I1213 09:20:07.653889 4971 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-z6fvq" Dec 13 09:20:07 crc kubenswrapper[4971]: I1213 09:20:07.791907 4971 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="e175976e-1561-43ce-9236-043e1a7b8d09" path="/var/lib/kubelet/pods/e175976e-1561-43ce-9236-043e1a7b8d09/volumes" Dec 13 09:20:08 crc kubenswrapper[4971]: I1213 09:20:08.269468 4971 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-z6fvq"] Dec 13 09:20:08 crc kubenswrapper[4971]: I1213 09:20:08.964370 4971 generic.go:334] "Generic (PLEG): container finished" podID="c63d9282-272e-4eeb-a93a-917099b3b147" containerID="536e85b06c73e15fea3c1112f8691e6b03045fc873bda6682b7cb3523815768f" exitCode=0 Dec 13 09:20:08 crc kubenswrapper[4971]: I1213 09:20:08.964635 4971 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-z6fvq" event={"ID":"c63d9282-272e-4eeb-a93a-917099b3b147","Type":"ContainerDied","Data":"536e85b06c73e15fea3c1112f8691e6b03045fc873bda6682b7cb3523815768f"} Dec 13 09:20:08 crc kubenswrapper[4971]: I1213 09:20:08.964916 4971 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-z6fvq" event={"ID":"c63d9282-272e-4eeb-a93a-917099b3b147","Type":"ContainerStarted","Data":"57b6f9e4e18ba5ce89f67f6f69fc53c4231b126c307bf40f23dcfaeaa286eebe"} Dec 13 09:20:10 crc kubenswrapper[4971]: I1213 09:20:10.989306 4971 generic.go:334] "Generic (PLEG): container finished" podID="c63d9282-272e-4eeb-a93a-917099b3b147" containerID="9d36fdb8e1768035881665d77bc2aad9f0c071478b3dda852dc0327dd113cb04" exitCode=0 Dec 13 09:20:10 crc kubenswrapper[4971]: I1213 09:20:10.989986 4971 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-z6fvq" event={"ID":"c63d9282-272e-4eeb-a93a-917099b3b147","Type":"ContainerDied","Data":"9d36fdb8e1768035881665d77bc2aad9f0c071478b3dda852dc0327dd113cb04"} Dec 13 09:20:13 crc kubenswrapper[4971]: I1213 09:20:13.015374 4971 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-z6fvq" event={"ID":"c63d9282-272e-4eeb-a93a-917099b3b147","Type":"ContainerStarted","Data":"e7990a009480688f73edd59d7d87fdc9516d70335ced893695b2140992733a8a"} Dec 13 09:20:13 crc kubenswrapper[4971]: I1213 09:20:13.045741 4971 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/community-operators-z6fvq" podStartSLOduration=2.850572933 podStartE2EDuration="6.045718903s" podCreationTimestamp="2025-12-13 09:20:07 +0000 UTC" firstStartedPulling="2025-12-13 09:20:08.967329623 +0000 UTC m=+9065.571739071" lastFinishedPulling="2025-12-13 09:20:12.162475593 +0000 UTC m=+9068.766885041" observedRunningTime="2025-12-13 09:20:13.042160435 +0000 UTC m=+9069.646569893" watchObservedRunningTime="2025-12-13 09:20:13.045718903 +0000 UTC m=+9069.650128351" Dec 13 09:20:17 crc kubenswrapper[4971]: I1213 09:20:17.654361 4971 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/community-operators-z6fvq" Dec 13 09:20:17 crc kubenswrapper[4971]: I1213 09:20:17.656338 4971 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/community-operators-z6fvq" Dec 13 09:20:17 crc kubenswrapper[4971]: I1213 09:20:17.759922 4971 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/community-operators-z6fvq" Dec 13 09:20:18 crc kubenswrapper[4971]: I1213 09:20:18.138772 4971 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/community-operators-z6fvq" Dec 13 09:20:18 crc kubenswrapper[4971]: I1213 09:20:18.209366 4971 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-z6fvq"] Dec 13 09:20:20 crc kubenswrapper[4971]: I1213 09:20:20.091092 4971 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/community-operators-z6fvq" podUID="c63d9282-272e-4eeb-a93a-917099b3b147" containerName="registry-server" containerID="cri-o://e7990a009480688f73edd59d7d87fdc9516d70335ced893695b2140992733a8a" gracePeriod=2 Dec 13 09:20:20 crc kubenswrapper[4971]: I1213 09:20:20.629571 4971 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-z6fvq" Dec 13 09:20:20 crc kubenswrapper[4971]: I1213 09:20:20.787271 4971 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/c63d9282-272e-4eeb-a93a-917099b3b147-catalog-content\") pod \"c63d9282-272e-4eeb-a93a-917099b3b147\" (UID: \"c63d9282-272e-4eeb-a93a-917099b3b147\") " Dec 13 09:20:20 crc kubenswrapper[4971]: I1213 09:20:20.787373 4971 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/c63d9282-272e-4eeb-a93a-917099b3b147-utilities\") pod \"c63d9282-272e-4eeb-a93a-917099b3b147\" (UID: \"c63d9282-272e-4eeb-a93a-917099b3b147\") " Dec 13 09:20:20 crc kubenswrapper[4971]: I1213 09:20:20.787834 4971 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-dwrbf\" (UniqueName: \"kubernetes.io/projected/c63d9282-272e-4eeb-a93a-917099b3b147-kube-api-access-dwrbf\") pod \"c63d9282-272e-4eeb-a93a-917099b3b147\" (UID: \"c63d9282-272e-4eeb-a93a-917099b3b147\") " Dec 13 09:20:20 crc kubenswrapper[4971]: I1213 09:20:20.788652 4971 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/c63d9282-272e-4eeb-a93a-917099b3b147-utilities" (OuterVolumeSpecName: "utilities") pod "c63d9282-272e-4eeb-a93a-917099b3b147" (UID: "c63d9282-272e-4eeb-a93a-917099b3b147"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 13 09:20:20 crc kubenswrapper[4971]: I1213 09:20:20.790942 4971 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/c63d9282-272e-4eeb-a93a-917099b3b147-utilities\") on node \"crc\" DevicePath \"\"" Dec 13 09:20:20 crc kubenswrapper[4971]: I1213 09:20:20.796080 4971 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/c63d9282-272e-4eeb-a93a-917099b3b147-kube-api-access-dwrbf" (OuterVolumeSpecName: "kube-api-access-dwrbf") pod "c63d9282-272e-4eeb-a93a-917099b3b147" (UID: "c63d9282-272e-4eeb-a93a-917099b3b147"). InnerVolumeSpecName "kube-api-access-dwrbf". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 13 09:20:20 crc kubenswrapper[4971]: I1213 09:20:20.843237 4971 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/c63d9282-272e-4eeb-a93a-917099b3b147-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "c63d9282-272e-4eeb-a93a-917099b3b147" (UID: "c63d9282-272e-4eeb-a93a-917099b3b147"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 13 09:20:20 crc kubenswrapper[4971]: I1213 09:20:20.891763 4971 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/c63d9282-272e-4eeb-a93a-917099b3b147-catalog-content\") on node \"crc\" DevicePath \"\"" Dec 13 09:20:20 crc kubenswrapper[4971]: I1213 09:20:20.891815 4971 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-dwrbf\" (UniqueName: \"kubernetes.io/projected/c63d9282-272e-4eeb-a93a-917099b3b147-kube-api-access-dwrbf\") on node \"crc\" DevicePath \"\"" Dec 13 09:20:21 crc kubenswrapper[4971]: I1213 09:20:21.104688 4971 generic.go:334] "Generic (PLEG): container finished" podID="c63d9282-272e-4eeb-a93a-917099b3b147" containerID="e7990a009480688f73edd59d7d87fdc9516d70335ced893695b2140992733a8a" exitCode=0 Dec 13 09:20:21 crc kubenswrapper[4971]: I1213 09:20:21.104742 4971 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-z6fvq" event={"ID":"c63d9282-272e-4eeb-a93a-917099b3b147","Type":"ContainerDied","Data":"e7990a009480688f73edd59d7d87fdc9516d70335ced893695b2140992733a8a"} Dec 13 09:20:21 crc kubenswrapper[4971]: I1213 09:20:21.104779 4971 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-z6fvq" event={"ID":"c63d9282-272e-4eeb-a93a-917099b3b147","Type":"ContainerDied","Data":"57b6f9e4e18ba5ce89f67f6f69fc53c4231b126c307bf40f23dcfaeaa286eebe"} Dec 13 09:20:21 crc kubenswrapper[4971]: I1213 09:20:21.104800 4971 scope.go:117] "RemoveContainer" containerID="e7990a009480688f73edd59d7d87fdc9516d70335ced893695b2140992733a8a" Dec 13 09:20:21 crc kubenswrapper[4971]: I1213 09:20:21.105311 4971 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-z6fvq" Dec 13 09:20:21 crc kubenswrapper[4971]: I1213 09:20:21.133111 4971 scope.go:117] "RemoveContainer" containerID="9d36fdb8e1768035881665d77bc2aad9f0c071478b3dda852dc0327dd113cb04" Dec 13 09:20:21 crc kubenswrapper[4971]: I1213 09:20:21.164039 4971 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-z6fvq"] Dec 13 09:20:21 crc kubenswrapper[4971]: I1213 09:20:21.174202 4971 scope.go:117] "RemoveContainer" containerID="536e85b06c73e15fea3c1112f8691e6b03045fc873bda6682b7cb3523815768f" Dec 13 09:20:21 crc kubenswrapper[4971]: I1213 09:20:21.179676 4971 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/community-operators-z6fvq"] Dec 13 09:20:21 crc kubenswrapper[4971]: I1213 09:20:21.217647 4971 scope.go:117] "RemoveContainer" containerID="e7990a009480688f73edd59d7d87fdc9516d70335ced893695b2140992733a8a" Dec 13 09:20:21 crc kubenswrapper[4971]: E1213 09:20:21.218614 4971 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"e7990a009480688f73edd59d7d87fdc9516d70335ced893695b2140992733a8a\": container with ID starting with e7990a009480688f73edd59d7d87fdc9516d70335ced893695b2140992733a8a not found: ID does not exist" containerID="e7990a009480688f73edd59d7d87fdc9516d70335ced893695b2140992733a8a" Dec 13 09:20:21 crc kubenswrapper[4971]: I1213 09:20:21.218663 4971 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"e7990a009480688f73edd59d7d87fdc9516d70335ced893695b2140992733a8a"} err="failed to get container status \"e7990a009480688f73edd59d7d87fdc9516d70335ced893695b2140992733a8a\": rpc error: code = NotFound desc = could not find container \"e7990a009480688f73edd59d7d87fdc9516d70335ced893695b2140992733a8a\": container with ID starting with e7990a009480688f73edd59d7d87fdc9516d70335ced893695b2140992733a8a not found: ID does not exist" Dec 13 09:20:21 crc kubenswrapper[4971]: I1213 09:20:21.218697 4971 scope.go:117] "RemoveContainer" containerID="9d36fdb8e1768035881665d77bc2aad9f0c071478b3dda852dc0327dd113cb04" Dec 13 09:20:21 crc kubenswrapper[4971]: E1213 09:20:21.219272 4971 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"9d36fdb8e1768035881665d77bc2aad9f0c071478b3dda852dc0327dd113cb04\": container with ID starting with 9d36fdb8e1768035881665d77bc2aad9f0c071478b3dda852dc0327dd113cb04 not found: ID does not exist" containerID="9d36fdb8e1768035881665d77bc2aad9f0c071478b3dda852dc0327dd113cb04" Dec 13 09:20:21 crc kubenswrapper[4971]: I1213 09:20:21.219307 4971 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"9d36fdb8e1768035881665d77bc2aad9f0c071478b3dda852dc0327dd113cb04"} err="failed to get container status \"9d36fdb8e1768035881665d77bc2aad9f0c071478b3dda852dc0327dd113cb04\": rpc error: code = NotFound desc = could not find container \"9d36fdb8e1768035881665d77bc2aad9f0c071478b3dda852dc0327dd113cb04\": container with ID starting with 9d36fdb8e1768035881665d77bc2aad9f0c071478b3dda852dc0327dd113cb04 not found: ID does not exist" Dec 13 09:20:21 crc kubenswrapper[4971]: I1213 09:20:21.219329 4971 scope.go:117] "RemoveContainer" containerID="536e85b06c73e15fea3c1112f8691e6b03045fc873bda6682b7cb3523815768f" Dec 13 09:20:21 crc kubenswrapper[4971]: E1213 09:20:21.219713 4971 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"536e85b06c73e15fea3c1112f8691e6b03045fc873bda6682b7cb3523815768f\": container with ID starting with 536e85b06c73e15fea3c1112f8691e6b03045fc873bda6682b7cb3523815768f not found: ID does not exist" containerID="536e85b06c73e15fea3c1112f8691e6b03045fc873bda6682b7cb3523815768f" Dec 13 09:20:21 crc kubenswrapper[4971]: I1213 09:20:21.219742 4971 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"536e85b06c73e15fea3c1112f8691e6b03045fc873bda6682b7cb3523815768f"} err="failed to get container status \"536e85b06c73e15fea3c1112f8691e6b03045fc873bda6682b7cb3523815768f\": rpc error: code = NotFound desc = could not find container \"536e85b06c73e15fea3c1112f8691e6b03045fc873bda6682b7cb3523815768f\": container with ID starting with 536e85b06c73e15fea3c1112f8691e6b03045fc873bda6682b7cb3523815768f not found: ID does not exist" Dec 13 09:20:21 crc kubenswrapper[4971]: I1213 09:20:21.792157 4971 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="c63d9282-272e-4eeb-a93a-917099b3b147" path="/var/lib/kubelet/pods/c63d9282-272e-4eeb-a93a-917099b3b147/volumes" Dec 13 09:21:46 crc kubenswrapper[4971]: I1213 09:21:46.153615 4971 patch_prober.go:28] interesting pod/machine-config-daemon-82xjz container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 13 09:21:46 crc kubenswrapper[4971]: I1213 09:21:46.154477 4971 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-82xjz" podUID="e0c4af12-edda-41d7-9b44-f87396174ef2" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 13 09:22:16 crc kubenswrapper[4971]: I1213 09:22:16.153287 4971 patch_prober.go:28] interesting pod/machine-config-daemon-82xjz container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 13 09:22:16 crc kubenswrapper[4971]: I1213 09:22:16.154207 4971 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-82xjz" podUID="e0c4af12-edda-41d7-9b44-f87396174ef2" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 13 09:22:46 crc kubenswrapper[4971]: I1213 09:22:46.154256 4971 patch_prober.go:28] interesting pod/machine-config-daemon-82xjz container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 13 09:22:46 crc kubenswrapper[4971]: I1213 09:22:46.155074 4971 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-82xjz" podUID="e0c4af12-edda-41d7-9b44-f87396174ef2" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 13 09:22:46 crc kubenswrapper[4971]: I1213 09:22:46.155133 4971 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-82xjz" Dec 13 09:22:46 crc kubenswrapper[4971]: I1213 09:22:46.156191 4971 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"e616f2406e276879960fe4d0372c2d2a34f8c21a80e5202a5083f7852225bab9"} pod="openshift-machine-config-operator/machine-config-daemon-82xjz" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Dec 13 09:22:46 crc kubenswrapper[4971]: I1213 09:22:46.156275 4971 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-82xjz" podUID="e0c4af12-edda-41d7-9b44-f87396174ef2" containerName="machine-config-daemon" containerID="cri-o://e616f2406e276879960fe4d0372c2d2a34f8c21a80e5202a5083f7852225bab9" gracePeriod=600 Dec 13 09:22:46 crc kubenswrapper[4971]: E1213 09:22:46.280574 4971 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-82xjz_openshift-machine-config-operator(e0c4af12-edda-41d7-9b44-f87396174ef2)\"" pod="openshift-machine-config-operator/machine-config-daemon-82xjz" podUID="e0c4af12-edda-41d7-9b44-f87396174ef2" Dec 13 09:22:46 crc kubenswrapper[4971]: I1213 09:22:46.821335 4971 generic.go:334] "Generic (PLEG): container finished" podID="e0c4af12-edda-41d7-9b44-f87396174ef2" containerID="e616f2406e276879960fe4d0372c2d2a34f8c21a80e5202a5083f7852225bab9" exitCode=0 Dec 13 09:22:46 crc kubenswrapper[4971]: I1213 09:22:46.821447 4971 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-82xjz" event={"ID":"e0c4af12-edda-41d7-9b44-f87396174ef2","Type":"ContainerDied","Data":"e616f2406e276879960fe4d0372c2d2a34f8c21a80e5202a5083f7852225bab9"} Dec 13 09:22:46 crc kubenswrapper[4971]: I1213 09:22:46.821806 4971 scope.go:117] "RemoveContainer" containerID="05ac64a4f7ae8e1342c3b29cc16132aac9d985c5e8acf6af1eb069f94d71e38d" Dec 13 09:22:46 crc kubenswrapper[4971]: I1213 09:22:46.823015 4971 scope.go:117] "RemoveContainer" containerID="e616f2406e276879960fe4d0372c2d2a34f8c21a80e5202a5083f7852225bab9" Dec 13 09:22:46 crc kubenswrapper[4971]: E1213 09:22:46.823488 4971 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-82xjz_openshift-machine-config-operator(e0c4af12-edda-41d7-9b44-f87396174ef2)\"" pod="openshift-machine-config-operator/machine-config-daemon-82xjz" podUID="e0c4af12-edda-41d7-9b44-f87396174ef2" Dec 13 09:22:59 crc kubenswrapper[4971]: I1213 09:22:59.768794 4971 scope.go:117] "RemoveContainer" containerID="e616f2406e276879960fe4d0372c2d2a34f8c21a80e5202a5083f7852225bab9" Dec 13 09:22:59 crc kubenswrapper[4971]: E1213 09:22:59.769663 4971 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-82xjz_openshift-machine-config-operator(e0c4af12-edda-41d7-9b44-f87396174ef2)\"" pod="openshift-machine-config-operator/machine-config-daemon-82xjz" podUID="e0c4af12-edda-41d7-9b44-f87396174ef2" Dec 13 09:23:10 crc kubenswrapper[4971]: I1213 09:23:10.769265 4971 scope.go:117] "RemoveContainer" containerID="e616f2406e276879960fe4d0372c2d2a34f8c21a80e5202a5083f7852225bab9" Dec 13 09:23:10 crc kubenswrapper[4971]: E1213 09:23:10.770444 4971 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-82xjz_openshift-machine-config-operator(e0c4af12-edda-41d7-9b44-f87396174ef2)\"" pod="openshift-machine-config-operator/machine-config-daemon-82xjz" podUID="e0c4af12-edda-41d7-9b44-f87396174ef2" Dec 13 09:23:23 crc kubenswrapper[4971]: I1213 09:23:23.777246 4971 scope.go:117] "RemoveContainer" containerID="e616f2406e276879960fe4d0372c2d2a34f8c21a80e5202a5083f7852225bab9" Dec 13 09:23:23 crc kubenswrapper[4971]: E1213 09:23:23.778443 4971 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-82xjz_openshift-machine-config-operator(e0c4af12-edda-41d7-9b44-f87396174ef2)\"" pod="openshift-machine-config-operator/machine-config-daemon-82xjz" podUID="e0c4af12-edda-41d7-9b44-f87396174ef2" Dec 13 09:23:36 crc kubenswrapper[4971]: I1213 09:23:36.769488 4971 scope.go:117] "RemoveContainer" containerID="e616f2406e276879960fe4d0372c2d2a34f8c21a80e5202a5083f7852225bab9" Dec 13 09:23:36 crc kubenswrapper[4971]: E1213 09:23:36.770301 4971 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-82xjz_openshift-machine-config-operator(e0c4af12-edda-41d7-9b44-f87396174ef2)\"" pod="openshift-machine-config-operator/machine-config-daemon-82xjz" podUID="e0c4af12-edda-41d7-9b44-f87396174ef2" Dec 13 09:23:41 crc kubenswrapper[4971]: E1213 09:23:41.886785 4971 cadvisor_stats_provider.go:516] "Partial failure issuing cadvisor.ContainerInfoV2" err="partial failures: [\"/system.slice/rpm-ostreed.service\": RecentStats: unable to find data in memory cache]" Dec 13 09:23:49 crc kubenswrapper[4971]: I1213 09:23:49.768833 4971 scope.go:117] "RemoveContainer" containerID="e616f2406e276879960fe4d0372c2d2a34f8c21a80e5202a5083f7852225bab9" Dec 13 09:23:49 crc kubenswrapper[4971]: E1213 09:23:49.769818 4971 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-82xjz_openshift-machine-config-operator(e0c4af12-edda-41d7-9b44-f87396174ef2)\"" pod="openshift-machine-config-operator/machine-config-daemon-82xjz" podUID="e0c4af12-edda-41d7-9b44-f87396174ef2" Dec 13 09:24:02 crc kubenswrapper[4971]: I1213 09:24:02.768908 4971 scope.go:117] "RemoveContainer" containerID="e616f2406e276879960fe4d0372c2d2a34f8c21a80e5202a5083f7852225bab9" Dec 13 09:24:02 crc kubenswrapper[4971]: E1213 09:24:02.770187 4971 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-82xjz_openshift-machine-config-operator(e0c4af12-edda-41d7-9b44-f87396174ef2)\"" pod="openshift-machine-config-operator/machine-config-daemon-82xjz" podUID="e0c4af12-edda-41d7-9b44-f87396174ef2" Dec 13 09:24:16 crc kubenswrapper[4971]: I1213 09:24:16.770364 4971 scope.go:117] "RemoveContainer" containerID="e616f2406e276879960fe4d0372c2d2a34f8c21a80e5202a5083f7852225bab9" Dec 13 09:24:16 crc kubenswrapper[4971]: E1213 09:24:16.771372 4971 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-82xjz_openshift-machine-config-operator(e0c4af12-edda-41d7-9b44-f87396174ef2)\"" pod="openshift-machine-config-operator/machine-config-daemon-82xjz" podUID="e0c4af12-edda-41d7-9b44-f87396174ef2" Dec 13 09:24:29 crc kubenswrapper[4971]: I1213 09:24:29.769648 4971 scope.go:117] "RemoveContainer" containerID="e616f2406e276879960fe4d0372c2d2a34f8c21a80e5202a5083f7852225bab9" Dec 13 09:24:29 crc kubenswrapper[4971]: E1213 09:24:29.770919 4971 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-82xjz_openshift-machine-config-operator(e0c4af12-edda-41d7-9b44-f87396174ef2)\"" pod="openshift-machine-config-operator/machine-config-daemon-82xjz" podUID="e0c4af12-edda-41d7-9b44-f87396174ef2" Dec 13 09:24:43 crc kubenswrapper[4971]: I1213 09:24:43.777229 4971 scope.go:117] "RemoveContainer" containerID="e616f2406e276879960fe4d0372c2d2a34f8c21a80e5202a5083f7852225bab9" Dec 13 09:24:43 crc kubenswrapper[4971]: E1213 09:24:43.778476 4971 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-82xjz_openshift-machine-config-operator(e0c4af12-edda-41d7-9b44-f87396174ef2)\"" pod="openshift-machine-config-operator/machine-config-daemon-82xjz" podUID="e0c4af12-edda-41d7-9b44-f87396174ef2" Dec 13 09:24:54 crc kubenswrapper[4971]: I1213 09:24:54.768461 4971 scope.go:117] "RemoveContainer" containerID="e616f2406e276879960fe4d0372c2d2a34f8c21a80e5202a5083f7852225bab9" Dec 13 09:24:54 crc kubenswrapper[4971]: E1213 09:24:54.771243 4971 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-82xjz_openshift-machine-config-operator(e0c4af12-edda-41d7-9b44-f87396174ef2)\"" pod="openshift-machine-config-operator/machine-config-daemon-82xjz" podUID="e0c4af12-edda-41d7-9b44-f87396174ef2" Dec 13 09:25:07 crc kubenswrapper[4971]: I1213 09:25:07.770729 4971 scope.go:117] "RemoveContainer" containerID="e616f2406e276879960fe4d0372c2d2a34f8c21a80e5202a5083f7852225bab9" Dec 13 09:25:07 crc kubenswrapper[4971]: E1213 09:25:07.771986 4971 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-82xjz_openshift-machine-config-operator(e0c4af12-edda-41d7-9b44-f87396174ef2)\"" pod="openshift-machine-config-operator/machine-config-daemon-82xjz" podUID="e0c4af12-edda-41d7-9b44-f87396174ef2" Dec 13 09:25:19 crc kubenswrapper[4971]: I1213 09:25:19.769509 4971 scope.go:117] "RemoveContainer" containerID="e616f2406e276879960fe4d0372c2d2a34f8c21a80e5202a5083f7852225bab9" Dec 13 09:25:19 crc kubenswrapper[4971]: E1213 09:25:19.770649 4971 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-82xjz_openshift-machine-config-operator(e0c4af12-edda-41d7-9b44-f87396174ef2)\"" pod="openshift-machine-config-operator/machine-config-daemon-82xjz" podUID="e0c4af12-edda-41d7-9b44-f87396174ef2" Dec 13 09:25:30 crc kubenswrapper[4971]: I1213 09:25:30.769104 4971 scope.go:117] "RemoveContainer" containerID="e616f2406e276879960fe4d0372c2d2a34f8c21a80e5202a5083f7852225bab9" Dec 13 09:25:30 crc kubenswrapper[4971]: E1213 09:25:30.770039 4971 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-82xjz_openshift-machine-config-operator(e0c4af12-edda-41d7-9b44-f87396174ef2)\"" pod="openshift-machine-config-operator/machine-config-daemon-82xjz" podUID="e0c4af12-edda-41d7-9b44-f87396174ef2" Dec 13 09:25:43 crc kubenswrapper[4971]: I1213 09:25:43.985098 4971 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-marketplace-mlqrx"] Dec 13 09:25:43 crc kubenswrapper[4971]: E1213 09:25:43.986552 4971 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="c63d9282-272e-4eeb-a93a-917099b3b147" containerName="extract-content" Dec 13 09:25:43 crc kubenswrapper[4971]: I1213 09:25:43.986571 4971 state_mem.go:107] "Deleted CPUSet assignment" podUID="c63d9282-272e-4eeb-a93a-917099b3b147" containerName="extract-content" Dec 13 09:25:43 crc kubenswrapper[4971]: E1213 09:25:43.986642 4971 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="c63d9282-272e-4eeb-a93a-917099b3b147" containerName="registry-server" Dec 13 09:25:43 crc kubenswrapper[4971]: I1213 09:25:43.986650 4971 state_mem.go:107] "Deleted CPUSet assignment" podUID="c63d9282-272e-4eeb-a93a-917099b3b147" containerName="registry-server" Dec 13 09:25:43 crc kubenswrapper[4971]: E1213 09:25:43.986668 4971 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="c63d9282-272e-4eeb-a93a-917099b3b147" containerName="extract-utilities" Dec 13 09:25:43 crc kubenswrapper[4971]: I1213 09:25:43.986677 4971 state_mem.go:107] "Deleted CPUSet assignment" podUID="c63d9282-272e-4eeb-a93a-917099b3b147" containerName="extract-utilities" Dec 13 09:25:43 crc kubenswrapper[4971]: I1213 09:25:43.986897 4971 memory_manager.go:354] "RemoveStaleState removing state" podUID="c63d9282-272e-4eeb-a93a-917099b3b147" containerName="registry-server" Dec 13 09:25:43 crc kubenswrapper[4971]: I1213 09:25:43.988602 4971 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-mlqrx" Dec 13 09:25:43 crc kubenswrapper[4971]: I1213 09:25:43.995956 4971 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-mlqrx"] Dec 13 09:25:44 crc kubenswrapper[4971]: I1213 09:25:44.136273 4971 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/c1dff35b-030d-4152-8f9b-6c4fa311dc0c-catalog-content\") pod \"redhat-marketplace-mlqrx\" (UID: \"c1dff35b-030d-4152-8f9b-6c4fa311dc0c\") " pod="openshift-marketplace/redhat-marketplace-mlqrx" Dec 13 09:25:44 crc kubenswrapper[4971]: I1213 09:25:44.136763 4971 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-9cdvm\" (UniqueName: \"kubernetes.io/projected/c1dff35b-030d-4152-8f9b-6c4fa311dc0c-kube-api-access-9cdvm\") pod \"redhat-marketplace-mlqrx\" (UID: \"c1dff35b-030d-4152-8f9b-6c4fa311dc0c\") " pod="openshift-marketplace/redhat-marketplace-mlqrx" Dec 13 09:25:44 crc kubenswrapper[4971]: I1213 09:25:44.136839 4971 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/c1dff35b-030d-4152-8f9b-6c4fa311dc0c-utilities\") pod \"redhat-marketplace-mlqrx\" (UID: \"c1dff35b-030d-4152-8f9b-6c4fa311dc0c\") " pod="openshift-marketplace/redhat-marketplace-mlqrx" Dec 13 09:25:44 crc kubenswrapper[4971]: I1213 09:25:44.239937 4971 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-9cdvm\" (UniqueName: \"kubernetes.io/projected/c1dff35b-030d-4152-8f9b-6c4fa311dc0c-kube-api-access-9cdvm\") pod \"redhat-marketplace-mlqrx\" (UID: \"c1dff35b-030d-4152-8f9b-6c4fa311dc0c\") " pod="openshift-marketplace/redhat-marketplace-mlqrx" Dec 13 09:25:44 crc kubenswrapper[4971]: I1213 09:25:44.239990 4971 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/c1dff35b-030d-4152-8f9b-6c4fa311dc0c-utilities\") pod \"redhat-marketplace-mlqrx\" (UID: \"c1dff35b-030d-4152-8f9b-6c4fa311dc0c\") " pod="openshift-marketplace/redhat-marketplace-mlqrx" Dec 13 09:25:44 crc kubenswrapper[4971]: I1213 09:25:44.240098 4971 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/c1dff35b-030d-4152-8f9b-6c4fa311dc0c-catalog-content\") pod \"redhat-marketplace-mlqrx\" (UID: \"c1dff35b-030d-4152-8f9b-6c4fa311dc0c\") " pod="openshift-marketplace/redhat-marketplace-mlqrx" Dec 13 09:25:44 crc kubenswrapper[4971]: I1213 09:25:44.240806 4971 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/c1dff35b-030d-4152-8f9b-6c4fa311dc0c-catalog-content\") pod \"redhat-marketplace-mlqrx\" (UID: \"c1dff35b-030d-4152-8f9b-6c4fa311dc0c\") " pod="openshift-marketplace/redhat-marketplace-mlqrx" Dec 13 09:25:44 crc kubenswrapper[4971]: I1213 09:25:44.240929 4971 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/c1dff35b-030d-4152-8f9b-6c4fa311dc0c-utilities\") pod \"redhat-marketplace-mlqrx\" (UID: \"c1dff35b-030d-4152-8f9b-6c4fa311dc0c\") " pod="openshift-marketplace/redhat-marketplace-mlqrx" Dec 13 09:25:44 crc kubenswrapper[4971]: I1213 09:25:44.266207 4971 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-9cdvm\" (UniqueName: \"kubernetes.io/projected/c1dff35b-030d-4152-8f9b-6c4fa311dc0c-kube-api-access-9cdvm\") pod \"redhat-marketplace-mlqrx\" (UID: \"c1dff35b-030d-4152-8f9b-6c4fa311dc0c\") " pod="openshift-marketplace/redhat-marketplace-mlqrx" Dec 13 09:25:44 crc kubenswrapper[4971]: I1213 09:25:44.310199 4971 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-mlqrx" Dec 13 09:25:44 crc kubenswrapper[4971]: I1213 09:25:44.896240 4971 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-mlqrx"] Dec 13 09:25:44 crc kubenswrapper[4971]: I1213 09:25:44.962443 4971 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-mlqrx" event={"ID":"c1dff35b-030d-4152-8f9b-6c4fa311dc0c","Type":"ContainerStarted","Data":"c64429a7d0171a046442014992b7933666ba37be333340d715d3f73322e6c981"} Dec 13 09:25:45 crc kubenswrapper[4971]: I1213 09:25:45.769557 4971 scope.go:117] "RemoveContainer" containerID="e616f2406e276879960fe4d0372c2d2a34f8c21a80e5202a5083f7852225bab9" Dec 13 09:25:45 crc kubenswrapper[4971]: E1213 09:25:45.771282 4971 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-82xjz_openshift-machine-config-operator(e0c4af12-edda-41d7-9b44-f87396174ef2)\"" pod="openshift-machine-config-operator/machine-config-daemon-82xjz" podUID="e0c4af12-edda-41d7-9b44-f87396174ef2" Dec 13 09:25:45 crc kubenswrapper[4971]: I1213 09:25:45.981004 4971 generic.go:334] "Generic (PLEG): container finished" podID="c1dff35b-030d-4152-8f9b-6c4fa311dc0c" containerID="76b172d79bab12d0e417d7fda476321f019b16f5bdae8b7c456a139408348f8f" exitCode=0 Dec 13 09:25:45 crc kubenswrapper[4971]: I1213 09:25:45.981070 4971 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-mlqrx" event={"ID":"c1dff35b-030d-4152-8f9b-6c4fa311dc0c","Type":"ContainerDied","Data":"76b172d79bab12d0e417d7fda476321f019b16f5bdae8b7c456a139408348f8f"} Dec 13 09:25:45 crc kubenswrapper[4971]: I1213 09:25:45.984085 4971 provider.go:102] Refreshing cache for provider: *credentialprovider.defaultDockerConfigProvider Dec 13 09:25:48 crc kubenswrapper[4971]: I1213 09:25:48.010074 4971 generic.go:334] "Generic (PLEG): container finished" podID="c1dff35b-030d-4152-8f9b-6c4fa311dc0c" containerID="cb11e635f1d2945aac96354ecf8cd33417a5fe414d5a4802eb723e8dd1928d8c" exitCode=0 Dec 13 09:25:48 crc kubenswrapper[4971]: I1213 09:25:48.010147 4971 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-mlqrx" event={"ID":"c1dff35b-030d-4152-8f9b-6c4fa311dc0c","Type":"ContainerDied","Data":"cb11e635f1d2945aac96354ecf8cd33417a5fe414d5a4802eb723e8dd1928d8c"} Dec 13 09:25:50 crc kubenswrapper[4971]: I1213 09:25:50.036369 4971 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-mlqrx" event={"ID":"c1dff35b-030d-4152-8f9b-6c4fa311dc0c","Type":"ContainerStarted","Data":"95bf37338fcbd4884d9b94f40b8db0f96c7420f254cb4771b80eb2cd0d9db7cf"} Dec 13 09:25:50 crc kubenswrapper[4971]: I1213 09:25:50.060007 4971 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-marketplace-mlqrx" podStartSLOduration=3.37505127 podStartE2EDuration="7.05998716s" podCreationTimestamp="2025-12-13 09:25:43 +0000 UTC" firstStartedPulling="2025-12-13 09:25:45.983751283 +0000 UTC m=+9402.588160731" lastFinishedPulling="2025-12-13 09:25:49.668687173 +0000 UTC m=+9406.273096621" observedRunningTime="2025-12-13 09:25:50.059718954 +0000 UTC m=+9406.664128412" watchObservedRunningTime="2025-12-13 09:25:50.05998716 +0000 UTC m=+9406.664396608" Dec 13 09:25:54 crc kubenswrapper[4971]: I1213 09:25:54.310820 4971 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-marketplace-mlqrx" Dec 13 09:25:54 crc kubenswrapper[4971]: I1213 09:25:54.311721 4971 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-marketplace-mlqrx" Dec 13 09:25:54 crc kubenswrapper[4971]: I1213 09:25:54.371817 4971 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-marketplace-mlqrx" Dec 13 09:25:55 crc kubenswrapper[4971]: I1213 09:25:55.144507 4971 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-marketplace-mlqrx" Dec 13 09:25:55 crc kubenswrapper[4971]: I1213 09:25:55.222429 4971 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-mlqrx"] Dec 13 09:25:57 crc kubenswrapper[4971]: I1213 09:25:57.102752 4971 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-marketplace-mlqrx" podUID="c1dff35b-030d-4152-8f9b-6c4fa311dc0c" containerName="registry-server" containerID="cri-o://95bf37338fcbd4884d9b94f40b8db0f96c7420f254cb4771b80eb2cd0d9db7cf" gracePeriod=2 Dec 13 09:25:57 crc kubenswrapper[4971]: I1213 09:25:57.654401 4971 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-mlqrx" Dec 13 09:25:57 crc kubenswrapper[4971]: I1213 09:25:57.792977 4971 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-9cdvm\" (UniqueName: \"kubernetes.io/projected/c1dff35b-030d-4152-8f9b-6c4fa311dc0c-kube-api-access-9cdvm\") pod \"c1dff35b-030d-4152-8f9b-6c4fa311dc0c\" (UID: \"c1dff35b-030d-4152-8f9b-6c4fa311dc0c\") " Dec 13 09:25:57 crc kubenswrapper[4971]: I1213 09:25:57.793103 4971 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/c1dff35b-030d-4152-8f9b-6c4fa311dc0c-utilities\") pod \"c1dff35b-030d-4152-8f9b-6c4fa311dc0c\" (UID: \"c1dff35b-030d-4152-8f9b-6c4fa311dc0c\") " Dec 13 09:25:57 crc kubenswrapper[4971]: I1213 09:25:57.793509 4971 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/c1dff35b-030d-4152-8f9b-6c4fa311dc0c-catalog-content\") pod \"c1dff35b-030d-4152-8f9b-6c4fa311dc0c\" (UID: \"c1dff35b-030d-4152-8f9b-6c4fa311dc0c\") " Dec 13 09:25:57 crc kubenswrapper[4971]: I1213 09:25:57.794859 4971 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/c1dff35b-030d-4152-8f9b-6c4fa311dc0c-utilities" (OuterVolumeSpecName: "utilities") pod "c1dff35b-030d-4152-8f9b-6c4fa311dc0c" (UID: "c1dff35b-030d-4152-8f9b-6c4fa311dc0c"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 13 09:25:57 crc kubenswrapper[4971]: I1213 09:25:57.800155 4971 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/c1dff35b-030d-4152-8f9b-6c4fa311dc0c-kube-api-access-9cdvm" (OuterVolumeSpecName: "kube-api-access-9cdvm") pod "c1dff35b-030d-4152-8f9b-6c4fa311dc0c" (UID: "c1dff35b-030d-4152-8f9b-6c4fa311dc0c"). InnerVolumeSpecName "kube-api-access-9cdvm". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 13 09:25:57 crc kubenswrapper[4971]: I1213 09:25:57.831469 4971 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/c1dff35b-030d-4152-8f9b-6c4fa311dc0c-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "c1dff35b-030d-4152-8f9b-6c4fa311dc0c" (UID: "c1dff35b-030d-4152-8f9b-6c4fa311dc0c"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 13 09:25:57 crc kubenswrapper[4971]: I1213 09:25:57.896639 4971 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-9cdvm\" (UniqueName: \"kubernetes.io/projected/c1dff35b-030d-4152-8f9b-6c4fa311dc0c-kube-api-access-9cdvm\") on node \"crc\" DevicePath \"\"" Dec 13 09:25:57 crc kubenswrapper[4971]: I1213 09:25:57.896686 4971 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/c1dff35b-030d-4152-8f9b-6c4fa311dc0c-utilities\") on node \"crc\" DevicePath \"\"" Dec 13 09:25:57 crc kubenswrapper[4971]: I1213 09:25:57.896700 4971 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/c1dff35b-030d-4152-8f9b-6c4fa311dc0c-catalog-content\") on node \"crc\" DevicePath \"\"" Dec 13 09:25:58 crc kubenswrapper[4971]: I1213 09:25:58.117502 4971 generic.go:334] "Generic (PLEG): container finished" podID="c1dff35b-030d-4152-8f9b-6c4fa311dc0c" containerID="95bf37338fcbd4884d9b94f40b8db0f96c7420f254cb4771b80eb2cd0d9db7cf" exitCode=0 Dec 13 09:25:58 crc kubenswrapper[4971]: I1213 09:25:58.117574 4971 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-mlqrx" event={"ID":"c1dff35b-030d-4152-8f9b-6c4fa311dc0c","Type":"ContainerDied","Data":"95bf37338fcbd4884d9b94f40b8db0f96c7420f254cb4771b80eb2cd0d9db7cf"} Dec 13 09:25:58 crc kubenswrapper[4971]: I1213 09:25:58.117608 4971 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-mlqrx" event={"ID":"c1dff35b-030d-4152-8f9b-6c4fa311dc0c","Type":"ContainerDied","Data":"c64429a7d0171a046442014992b7933666ba37be333340d715d3f73322e6c981"} Dec 13 09:25:58 crc kubenswrapper[4971]: I1213 09:25:58.117630 4971 scope.go:117] "RemoveContainer" containerID="95bf37338fcbd4884d9b94f40b8db0f96c7420f254cb4771b80eb2cd0d9db7cf" Dec 13 09:25:58 crc kubenswrapper[4971]: I1213 09:25:58.117751 4971 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-mlqrx" Dec 13 09:25:58 crc kubenswrapper[4971]: I1213 09:25:58.147855 4971 scope.go:117] "RemoveContainer" containerID="cb11e635f1d2945aac96354ecf8cd33417a5fe414d5a4802eb723e8dd1928d8c" Dec 13 09:25:58 crc kubenswrapper[4971]: I1213 09:25:58.165580 4971 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-mlqrx"] Dec 13 09:25:58 crc kubenswrapper[4971]: I1213 09:25:58.174377 4971 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-marketplace-mlqrx"] Dec 13 09:25:58 crc kubenswrapper[4971]: I1213 09:25:58.175306 4971 scope.go:117] "RemoveContainer" containerID="76b172d79bab12d0e417d7fda476321f019b16f5bdae8b7c456a139408348f8f" Dec 13 09:25:58 crc kubenswrapper[4971]: I1213 09:25:58.233887 4971 scope.go:117] "RemoveContainer" containerID="95bf37338fcbd4884d9b94f40b8db0f96c7420f254cb4771b80eb2cd0d9db7cf" Dec 13 09:25:58 crc kubenswrapper[4971]: E1213 09:25:58.235167 4971 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"95bf37338fcbd4884d9b94f40b8db0f96c7420f254cb4771b80eb2cd0d9db7cf\": container with ID starting with 95bf37338fcbd4884d9b94f40b8db0f96c7420f254cb4771b80eb2cd0d9db7cf not found: ID does not exist" containerID="95bf37338fcbd4884d9b94f40b8db0f96c7420f254cb4771b80eb2cd0d9db7cf" Dec 13 09:25:58 crc kubenswrapper[4971]: I1213 09:25:58.235234 4971 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"95bf37338fcbd4884d9b94f40b8db0f96c7420f254cb4771b80eb2cd0d9db7cf"} err="failed to get container status \"95bf37338fcbd4884d9b94f40b8db0f96c7420f254cb4771b80eb2cd0d9db7cf\": rpc error: code = NotFound desc = could not find container \"95bf37338fcbd4884d9b94f40b8db0f96c7420f254cb4771b80eb2cd0d9db7cf\": container with ID starting with 95bf37338fcbd4884d9b94f40b8db0f96c7420f254cb4771b80eb2cd0d9db7cf not found: ID does not exist" Dec 13 09:25:58 crc kubenswrapper[4971]: I1213 09:25:58.235272 4971 scope.go:117] "RemoveContainer" containerID="cb11e635f1d2945aac96354ecf8cd33417a5fe414d5a4802eb723e8dd1928d8c" Dec 13 09:25:58 crc kubenswrapper[4971]: E1213 09:25:58.236181 4971 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"cb11e635f1d2945aac96354ecf8cd33417a5fe414d5a4802eb723e8dd1928d8c\": container with ID starting with cb11e635f1d2945aac96354ecf8cd33417a5fe414d5a4802eb723e8dd1928d8c not found: ID does not exist" containerID="cb11e635f1d2945aac96354ecf8cd33417a5fe414d5a4802eb723e8dd1928d8c" Dec 13 09:25:58 crc kubenswrapper[4971]: I1213 09:25:58.236236 4971 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"cb11e635f1d2945aac96354ecf8cd33417a5fe414d5a4802eb723e8dd1928d8c"} err="failed to get container status \"cb11e635f1d2945aac96354ecf8cd33417a5fe414d5a4802eb723e8dd1928d8c\": rpc error: code = NotFound desc = could not find container \"cb11e635f1d2945aac96354ecf8cd33417a5fe414d5a4802eb723e8dd1928d8c\": container with ID starting with cb11e635f1d2945aac96354ecf8cd33417a5fe414d5a4802eb723e8dd1928d8c not found: ID does not exist" Dec 13 09:25:58 crc kubenswrapper[4971]: I1213 09:25:58.236270 4971 scope.go:117] "RemoveContainer" containerID="76b172d79bab12d0e417d7fda476321f019b16f5bdae8b7c456a139408348f8f" Dec 13 09:25:58 crc kubenswrapper[4971]: E1213 09:25:58.236569 4971 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"76b172d79bab12d0e417d7fda476321f019b16f5bdae8b7c456a139408348f8f\": container with ID starting with 76b172d79bab12d0e417d7fda476321f019b16f5bdae8b7c456a139408348f8f not found: ID does not exist" containerID="76b172d79bab12d0e417d7fda476321f019b16f5bdae8b7c456a139408348f8f" Dec 13 09:25:58 crc kubenswrapper[4971]: I1213 09:25:58.236596 4971 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"76b172d79bab12d0e417d7fda476321f019b16f5bdae8b7c456a139408348f8f"} err="failed to get container status \"76b172d79bab12d0e417d7fda476321f019b16f5bdae8b7c456a139408348f8f\": rpc error: code = NotFound desc = could not find container \"76b172d79bab12d0e417d7fda476321f019b16f5bdae8b7c456a139408348f8f\": container with ID starting with 76b172d79bab12d0e417d7fda476321f019b16f5bdae8b7c456a139408348f8f not found: ID does not exist" Dec 13 09:25:59 crc kubenswrapper[4971]: I1213 09:25:59.780431 4971 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="c1dff35b-030d-4152-8f9b-6c4fa311dc0c" path="/var/lib/kubelet/pods/c1dff35b-030d-4152-8f9b-6c4fa311dc0c/volumes" Dec 13 09:26:00 crc kubenswrapper[4971]: I1213 09:26:00.769501 4971 scope.go:117] "RemoveContainer" containerID="e616f2406e276879960fe4d0372c2d2a34f8c21a80e5202a5083f7852225bab9" Dec 13 09:26:00 crc kubenswrapper[4971]: E1213 09:26:00.770266 4971 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-82xjz_openshift-machine-config-operator(e0c4af12-edda-41d7-9b44-f87396174ef2)\"" pod="openshift-machine-config-operator/machine-config-daemon-82xjz" podUID="e0c4af12-edda-41d7-9b44-f87396174ef2" Dec 13 09:26:11 crc kubenswrapper[4971]: I1213 09:26:11.769734 4971 scope.go:117] "RemoveContainer" containerID="e616f2406e276879960fe4d0372c2d2a34f8c21a80e5202a5083f7852225bab9" Dec 13 09:26:11 crc kubenswrapper[4971]: E1213 09:26:11.770880 4971 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-82xjz_openshift-machine-config-operator(e0c4af12-edda-41d7-9b44-f87396174ef2)\"" pod="openshift-machine-config-operator/machine-config-daemon-82xjz" podUID="e0c4af12-edda-41d7-9b44-f87396174ef2" Dec 13 09:26:24 crc kubenswrapper[4971]: I1213 09:26:24.769752 4971 scope.go:117] "RemoveContainer" containerID="e616f2406e276879960fe4d0372c2d2a34f8c21a80e5202a5083f7852225bab9" Dec 13 09:26:24 crc kubenswrapper[4971]: E1213 09:26:24.770667 4971 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-82xjz_openshift-machine-config-operator(e0c4af12-edda-41d7-9b44-f87396174ef2)\"" pod="openshift-machine-config-operator/machine-config-daemon-82xjz" podUID="e0c4af12-edda-41d7-9b44-f87396174ef2" Dec 13 09:26:35 crc kubenswrapper[4971]: I1213 09:26:35.773358 4971 scope.go:117] "RemoveContainer" containerID="e616f2406e276879960fe4d0372c2d2a34f8c21a80e5202a5083f7852225bab9" Dec 13 09:26:35 crc kubenswrapper[4971]: E1213 09:26:35.774717 4971 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-82xjz_openshift-machine-config-operator(e0c4af12-edda-41d7-9b44-f87396174ef2)\"" pod="openshift-machine-config-operator/machine-config-daemon-82xjz" podUID="e0c4af12-edda-41d7-9b44-f87396174ef2" Dec 13 09:26:46 crc kubenswrapper[4971]: I1213 09:26:46.768942 4971 scope.go:117] "RemoveContainer" containerID="e616f2406e276879960fe4d0372c2d2a34f8c21a80e5202a5083f7852225bab9" Dec 13 09:26:46 crc kubenswrapper[4971]: E1213 09:26:46.770352 4971 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-82xjz_openshift-machine-config-operator(e0c4af12-edda-41d7-9b44-f87396174ef2)\"" pod="openshift-machine-config-operator/machine-config-daemon-82xjz" podUID="e0c4af12-edda-41d7-9b44-f87396174ef2" Dec 13 09:26:51 crc kubenswrapper[4971]: I1213 09:26:51.695302 4971 generic.go:334] "Generic (PLEG): container finished" podID="879e5242-5aa8-44fb-b759-6babf9c6559d" containerID="9a7931b4191657a8f36dd12bd8870834b1b27a528abc6bce6a58dea36f6fa1df" exitCode=0 Dec 13 09:26:51 crc kubenswrapper[4971]: I1213 09:26:51.696002 4971 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/tempest-tests-tempest-s00-full" event={"ID":"879e5242-5aa8-44fb-b759-6babf9c6559d","Type":"ContainerDied","Data":"9a7931b4191657a8f36dd12bd8870834b1b27a528abc6bce6a58dea36f6fa1df"} Dec 13 09:26:53 crc kubenswrapper[4971]: I1213 09:26:53.395903 4971 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/tempest-tests-tempest-s00-full" Dec 13 09:26:53 crc kubenswrapper[4971]: I1213 09:26:53.514049 4971 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/tempest-tests-tempest-s01-single-test"] Dec 13 09:26:53 crc kubenswrapper[4971]: E1213 09:26:53.514646 4971 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="879e5242-5aa8-44fb-b759-6babf9c6559d" containerName="tempest-tests-tempest-tests-runner" Dec 13 09:26:53 crc kubenswrapper[4971]: I1213 09:26:53.514851 4971 state_mem.go:107] "Deleted CPUSet assignment" podUID="879e5242-5aa8-44fb-b759-6babf9c6559d" containerName="tempest-tests-tempest-tests-runner" Dec 13 09:26:53 crc kubenswrapper[4971]: E1213 09:26:53.514888 4971 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="c1dff35b-030d-4152-8f9b-6c4fa311dc0c" containerName="extract-utilities" Dec 13 09:26:53 crc kubenswrapper[4971]: I1213 09:26:53.514895 4971 state_mem.go:107] "Deleted CPUSet assignment" podUID="c1dff35b-030d-4152-8f9b-6c4fa311dc0c" containerName="extract-utilities" Dec 13 09:26:53 crc kubenswrapper[4971]: E1213 09:26:53.514906 4971 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="c1dff35b-030d-4152-8f9b-6c4fa311dc0c" containerName="extract-content" Dec 13 09:26:53 crc kubenswrapper[4971]: I1213 09:26:53.514913 4971 state_mem.go:107] "Deleted CPUSet assignment" podUID="c1dff35b-030d-4152-8f9b-6c4fa311dc0c" containerName="extract-content" Dec 13 09:26:53 crc kubenswrapper[4971]: E1213 09:26:53.514927 4971 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="c1dff35b-030d-4152-8f9b-6c4fa311dc0c" containerName="registry-server" Dec 13 09:26:53 crc kubenswrapper[4971]: I1213 09:26:53.514935 4971 state_mem.go:107] "Deleted CPUSet assignment" podUID="c1dff35b-030d-4152-8f9b-6c4fa311dc0c" containerName="registry-server" Dec 13 09:26:53 crc kubenswrapper[4971]: I1213 09:26:53.515145 4971 memory_manager.go:354] "RemoveStaleState removing state" podUID="c1dff35b-030d-4152-8f9b-6c4fa311dc0c" containerName="registry-server" Dec 13 09:26:53 crc kubenswrapper[4971]: I1213 09:26:53.515174 4971 memory_manager.go:354] "RemoveStaleState removing state" podUID="879e5242-5aa8-44fb-b759-6babf9c6559d" containerName="tempest-tests-tempest-tests-runner" Dec 13 09:26:53 crc kubenswrapper[4971]: I1213 09:26:53.516124 4971 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/tempest-tests-tempest-s01-single-test" Dec 13 09:26:53 crc kubenswrapper[4971]: I1213 09:26:53.521911 4971 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"tempest-tests-tempest-custom-data-s1" Dec 13 09:26:53 crc kubenswrapper[4971]: I1213 09:26:53.522295 4971 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"tempest-tests-tempest-env-vars-s1" Dec 13 09:26:53 crc kubenswrapper[4971]: I1213 09:26:53.523720 4971 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/tempest-tests-tempest-s01-single-test"] Dec 13 09:26:53 crc kubenswrapper[4971]: I1213 09:26:53.541981 4971 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ca-certs\" (UniqueName: \"kubernetes.io/secret/879e5242-5aa8-44fb-b759-6babf9c6559d-ca-certs\") pod \"879e5242-5aa8-44fb-b759-6babf9c6559d\" (UID: \"879e5242-5aa8-44fb-b759-6babf9c6559d\") " Dec 13 09:26:53 crc kubenswrapper[4971]: I1213 09:26:53.542082 4971 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"openstack-config\" (UniqueName: \"kubernetes.io/configmap/879e5242-5aa8-44fb-b759-6babf9c6559d-openstack-config\") pod \"879e5242-5aa8-44fb-b759-6babf9c6559d\" (UID: \"879e5242-5aa8-44fb-b759-6babf9c6559d\") " Dec 13 09:26:53 crc kubenswrapper[4971]: I1213 09:26:53.542119 4971 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"openstack-config-secret\" (UniqueName: \"kubernetes.io/secret/879e5242-5aa8-44fb-b759-6babf9c6559d-openstack-config-secret\") pod \"879e5242-5aa8-44fb-b759-6babf9c6559d\" (UID: \"879e5242-5aa8-44fb-b759-6babf9c6559d\") " Dec 13 09:26:53 crc kubenswrapper[4971]: I1213 09:26:53.542217 4971 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"test-operator-logs\" (UniqueName: \"kubernetes.io/local-volume/local-storage12-crc\") pod \"879e5242-5aa8-44fb-b759-6babf9c6559d\" (UID: \"879e5242-5aa8-44fb-b759-6babf9c6559d\") " Dec 13 09:26:53 crc kubenswrapper[4971]: I1213 09:26:53.542254 4971 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-n5hn4\" (UniqueName: \"kubernetes.io/projected/879e5242-5aa8-44fb-b759-6babf9c6559d-kube-api-access-n5hn4\") pod \"879e5242-5aa8-44fb-b759-6babf9c6559d\" (UID: \"879e5242-5aa8-44fb-b759-6babf9c6559d\") " Dec 13 09:26:53 crc kubenswrapper[4971]: I1213 09:26:53.542346 4971 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/879e5242-5aa8-44fb-b759-6babf9c6559d-ceph\") pod \"879e5242-5aa8-44fb-b759-6babf9c6559d\" (UID: \"879e5242-5aa8-44fb-b759-6babf9c6559d\") " Dec 13 09:26:53 crc kubenswrapper[4971]: I1213 09:26:53.542453 4971 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"test-operator-ephemeral-workdir\" (UniqueName: \"kubernetes.io/empty-dir/879e5242-5aa8-44fb-b759-6babf9c6559d-test-operator-ephemeral-workdir\") pod \"879e5242-5aa8-44fb-b759-6babf9c6559d\" (UID: \"879e5242-5aa8-44fb-b759-6babf9c6559d\") " Dec 13 09:26:53 crc kubenswrapper[4971]: I1213 09:26:53.542481 4971 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"test-operator-ephemeral-temporary\" (UniqueName: \"kubernetes.io/empty-dir/879e5242-5aa8-44fb-b759-6babf9c6559d-test-operator-ephemeral-temporary\") pod \"879e5242-5aa8-44fb-b759-6babf9c6559d\" (UID: \"879e5242-5aa8-44fb-b759-6babf9c6559d\") " Dec 13 09:26:53 crc kubenswrapper[4971]: I1213 09:26:53.542564 4971 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/879e5242-5aa8-44fb-b759-6babf9c6559d-config-data\") pod \"879e5242-5aa8-44fb-b759-6babf9c6559d\" (UID: \"879e5242-5aa8-44fb-b759-6babf9c6559d\") " Dec 13 09:26:53 crc kubenswrapper[4971]: I1213 09:26:53.542624 4971 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/879e5242-5aa8-44fb-b759-6babf9c6559d-ssh-key\") pod \"879e5242-5aa8-44fb-b759-6babf9c6559d\" (UID: \"879e5242-5aa8-44fb-b759-6babf9c6559d\") " Dec 13 09:26:53 crc kubenswrapper[4971]: I1213 09:26:53.544627 4971 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/879e5242-5aa8-44fb-b759-6babf9c6559d-test-operator-ephemeral-temporary" (OuterVolumeSpecName: "test-operator-ephemeral-temporary") pod "879e5242-5aa8-44fb-b759-6babf9c6559d" (UID: "879e5242-5aa8-44fb-b759-6babf9c6559d"). InnerVolumeSpecName "test-operator-ephemeral-temporary". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 13 09:26:53 crc kubenswrapper[4971]: I1213 09:26:53.545165 4971 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/879e5242-5aa8-44fb-b759-6babf9c6559d-config-data" (OuterVolumeSpecName: "config-data") pod "879e5242-5aa8-44fb-b759-6babf9c6559d" (UID: "879e5242-5aa8-44fb-b759-6babf9c6559d"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 13 09:26:53 crc kubenswrapper[4971]: I1213 09:26:53.557315 4971 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/879e5242-5aa8-44fb-b759-6babf9c6559d-test-operator-ephemeral-workdir" (OuterVolumeSpecName: "test-operator-ephemeral-workdir") pod "879e5242-5aa8-44fb-b759-6babf9c6559d" (UID: "879e5242-5aa8-44fb-b759-6babf9c6559d"). InnerVolumeSpecName "test-operator-ephemeral-workdir". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 13 09:26:53 crc kubenswrapper[4971]: I1213 09:26:53.558694 4971 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/879e5242-5aa8-44fb-b759-6babf9c6559d-ceph" (OuterVolumeSpecName: "ceph") pod "879e5242-5aa8-44fb-b759-6babf9c6559d" (UID: "879e5242-5aa8-44fb-b759-6babf9c6559d"). InnerVolumeSpecName "ceph". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 13 09:26:53 crc kubenswrapper[4971]: I1213 09:26:53.565478 4971 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/879e5242-5aa8-44fb-b759-6babf9c6559d-kube-api-access-n5hn4" (OuterVolumeSpecName: "kube-api-access-n5hn4") pod "879e5242-5aa8-44fb-b759-6babf9c6559d" (UID: "879e5242-5aa8-44fb-b759-6babf9c6559d"). InnerVolumeSpecName "kube-api-access-n5hn4". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 13 09:26:53 crc kubenswrapper[4971]: I1213 09:26:53.567348 4971 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/local-volume/local-storage12-crc" (OuterVolumeSpecName: "test-operator-logs") pod "879e5242-5aa8-44fb-b759-6babf9c6559d" (UID: "879e5242-5aa8-44fb-b759-6babf9c6559d"). InnerVolumeSpecName "local-storage12-crc". PluginName "kubernetes.io/local-volume", VolumeGidValue "" Dec 13 09:26:53 crc kubenswrapper[4971]: I1213 09:26:53.583738 4971 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/879e5242-5aa8-44fb-b759-6babf9c6559d-ca-certs" (OuterVolumeSpecName: "ca-certs") pod "879e5242-5aa8-44fb-b759-6babf9c6559d" (UID: "879e5242-5aa8-44fb-b759-6babf9c6559d"). InnerVolumeSpecName "ca-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 13 09:26:53 crc kubenswrapper[4971]: I1213 09:26:53.586596 4971 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/879e5242-5aa8-44fb-b759-6babf9c6559d-openstack-config-secret" (OuterVolumeSpecName: "openstack-config-secret") pod "879e5242-5aa8-44fb-b759-6babf9c6559d" (UID: "879e5242-5aa8-44fb-b759-6babf9c6559d"). InnerVolumeSpecName "openstack-config-secret". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 13 09:26:53 crc kubenswrapper[4971]: I1213 09:26:53.590899 4971 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/879e5242-5aa8-44fb-b759-6babf9c6559d-ssh-key" (OuterVolumeSpecName: "ssh-key") pod "879e5242-5aa8-44fb-b759-6babf9c6559d" (UID: "879e5242-5aa8-44fb-b759-6babf9c6559d"). InnerVolumeSpecName "ssh-key". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 13 09:26:53 crc kubenswrapper[4971]: I1213 09:26:53.606244 4971 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/879e5242-5aa8-44fb-b759-6babf9c6559d-openstack-config" (OuterVolumeSpecName: "openstack-config") pod "879e5242-5aa8-44fb-b759-6babf9c6559d" (UID: "879e5242-5aa8-44fb-b759-6babf9c6559d"). InnerVolumeSpecName "openstack-config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 13 09:26:53 crc kubenswrapper[4971]: I1213 09:26:53.645046 4971 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/b0fcccc8-3a38-42b6-a10c-57a0a44b48c3-ssh-key\") pod \"tempest-tests-tempest-s01-single-test\" (UID: \"b0fcccc8-3a38-42b6-a10c-57a0a44b48c3\") " pod="openstack/tempest-tests-tempest-s01-single-test" Dec 13 09:26:53 crc kubenswrapper[4971]: I1213 09:26:53.645146 4971 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"openstack-config\" (UniqueName: \"kubernetes.io/configmap/b0fcccc8-3a38-42b6-a10c-57a0a44b48c3-openstack-config\") pod \"tempest-tests-tempest-s01-single-test\" (UID: \"b0fcccc8-3a38-42b6-a10c-57a0a44b48c3\") " pod="openstack/tempest-tests-tempest-s01-single-test" Dec 13 09:26:53 crc kubenswrapper[4971]: I1213 09:26:53.645244 4971 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"openstack-config-secret\" (UniqueName: \"kubernetes.io/secret/b0fcccc8-3a38-42b6-a10c-57a0a44b48c3-openstack-config-secret\") pod \"tempest-tests-tempest-s01-single-test\" (UID: \"b0fcccc8-3a38-42b6-a10c-57a0a44b48c3\") " pod="openstack/tempest-tests-tempest-s01-single-test" Dec 13 09:26:53 crc kubenswrapper[4971]: I1213 09:26:53.645285 4971 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"test-operator-ephemeral-workdir\" (UniqueName: \"kubernetes.io/empty-dir/b0fcccc8-3a38-42b6-a10c-57a0a44b48c3-test-operator-ephemeral-workdir\") pod \"tempest-tests-tempest-s01-single-test\" (UID: \"b0fcccc8-3a38-42b6-a10c-57a0a44b48c3\") " pod="openstack/tempest-tests-tempest-s01-single-test" Dec 13 09:26:53 crc kubenswrapper[4971]: I1213 09:26:53.645306 4971 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-pwptn\" (UniqueName: \"kubernetes.io/projected/b0fcccc8-3a38-42b6-a10c-57a0a44b48c3-kube-api-access-pwptn\") pod \"tempest-tests-tempest-s01-single-test\" (UID: \"b0fcccc8-3a38-42b6-a10c-57a0a44b48c3\") " pod="openstack/tempest-tests-tempest-s01-single-test" Dec 13 09:26:53 crc kubenswrapper[4971]: I1213 09:26:53.645354 4971 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/b0fcccc8-3a38-42b6-a10c-57a0a44b48c3-config-data\") pod \"tempest-tests-tempest-s01-single-test\" (UID: \"b0fcccc8-3a38-42b6-a10c-57a0a44b48c3\") " pod="openstack/tempest-tests-tempest-s01-single-test" Dec 13 09:26:53 crc kubenswrapper[4971]: I1213 09:26:53.645377 4971 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ca-certs\" (UniqueName: \"kubernetes.io/secret/b0fcccc8-3a38-42b6-a10c-57a0a44b48c3-ca-certs\") pod \"tempest-tests-tempest-s01-single-test\" (UID: \"b0fcccc8-3a38-42b6-a10c-57a0a44b48c3\") " pod="openstack/tempest-tests-tempest-s01-single-test" Dec 13 09:26:53 crc kubenswrapper[4971]: I1213 09:26:53.645433 4971 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"local-storage12-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage12-crc\") pod \"tempest-tests-tempest-s01-single-test\" (UID: \"b0fcccc8-3a38-42b6-a10c-57a0a44b48c3\") " pod="openstack/tempest-tests-tempest-s01-single-test" Dec 13 09:26:53 crc kubenswrapper[4971]: I1213 09:26:53.645476 4971 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"test-operator-ephemeral-temporary\" (UniqueName: \"kubernetes.io/empty-dir/b0fcccc8-3a38-42b6-a10c-57a0a44b48c3-test-operator-ephemeral-temporary\") pod \"tempest-tests-tempest-s01-single-test\" (UID: \"b0fcccc8-3a38-42b6-a10c-57a0a44b48c3\") " pod="openstack/tempest-tests-tempest-s01-single-test" Dec 13 09:26:53 crc kubenswrapper[4971]: I1213 09:26:53.645503 4971 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/b0fcccc8-3a38-42b6-a10c-57a0a44b48c3-ceph\") pod \"tempest-tests-tempest-s01-single-test\" (UID: \"b0fcccc8-3a38-42b6-a10c-57a0a44b48c3\") " pod="openstack/tempest-tests-tempest-s01-single-test" Dec 13 09:26:53 crc kubenswrapper[4971]: I1213 09:26:53.645618 4971 reconciler_common.go:293] "Volume detached for volume \"test-operator-ephemeral-workdir\" (UniqueName: \"kubernetes.io/empty-dir/879e5242-5aa8-44fb-b759-6babf9c6559d-test-operator-ephemeral-workdir\") on node \"crc\" DevicePath \"\"" Dec 13 09:26:53 crc kubenswrapper[4971]: I1213 09:26:53.645635 4971 reconciler_common.go:293] "Volume detached for volume \"test-operator-ephemeral-temporary\" (UniqueName: \"kubernetes.io/empty-dir/879e5242-5aa8-44fb-b759-6babf9c6559d-test-operator-ephemeral-temporary\") on node \"crc\" DevicePath \"\"" Dec 13 09:26:53 crc kubenswrapper[4971]: I1213 09:26:53.645646 4971 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/879e5242-5aa8-44fb-b759-6babf9c6559d-config-data\") on node \"crc\" DevicePath \"\"" Dec 13 09:26:53 crc kubenswrapper[4971]: I1213 09:26:53.645655 4971 reconciler_common.go:293] "Volume detached for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/879e5242-5aa8-44fb-b759-6babf9c6559d-ssh-key\") on node \"crc\" DevicePath \"\"" Dec 13 09:26:53 crc kubenswrapper[4971]: I1213 09:26:53.645665 4971 reconciler_common.go:293] "Volume detached for volume \"ca-certs\" (UniqueName: \"kubernetes.io/secret/879e5242-5aa8-44fb-b759-6babf9c6559d-ca-certs\") on node \"crc\" DevicePath \"\"" Dec 13 09:26:53 crc kubenswrapper[4971]: I1213 09:26:53.645673 4971 reconciler_common.go:293] "Volume detached for volume \"openstack-config\" (UniqueName: \"kubernetes.io/configmap/879e5242-5aa8-44fb-b759-6babf9c6559d-openstack-config\") on node \"crc\" DevicePath \"\"" Dec 13 09:26:53 crc kubenswrapper[4971]: I1213 09:26:53.645684 4971 reconciler_common.go:293] "Volume detached for volume \"openstack-config-secret\" (UniqueName: \"kubernetes.io/secret/879e5242-5aa8-44fb-b759-6babf9c6559d-openstack-config-secret\") on node \"crc\" DevicePath \"\"" Dec 13 09:26:53 crc kubenswrapper[4971]: I1213 09:26:53.645693 4971 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-n5hn4\" (UniqueName: \"kubernetes.io/projected/879e5242-5aa8-44fb-b759-6babf9c6559d-kube-api-access-n5hn4\") on node \"crc\" DevicePath \"\"" Dec 13 09:26:53 crc kubenswrapper[4971]: I1213 09:26:53.645703 4971 reconciler_common.go:293] "Volume detached for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/879e5242-5aa8-44fb-b759-6babf9c6559d-ceph\") on node \"crc\" DevicePath \"\"" Dec 13 09:26:53 crc kubenswrapper[4971]: I1213 09:26:53.678388 4971 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"local-storage12-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage12-crc\") pod \"tempest-tests-tempest-s01-single-test\" (UID: \"b0fcccc8-3a38-42b6-a10c-57a0a44b48c3\") " pod="openstack/tempest-tests-tempest-s01-single-test" Dec 13 09:26:53 crc kubenswrapper[4971]: I1213 09:26:53.720270 4971 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/tempest-tests-tempest-s00-full" Dec 13 09:26:53 crc kubenswrapper[4971]: I1213 09:26:53.720176 4971 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/tempest-tests-tempest-s00-full" event={"ID":"879e5242-5aa8-44fb-b759-6babf9c6559d","Type":"ContainerDied","Data":"9135ef510bc8b400a7dbe31c87cd94bee1f16e0554efd8a7aaf5693eb3e181ef"} Dec 13 09:26:53 crc kubenswrapper[4971]: I1213 09:26:53.721893 4971 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="9135ef510bc8b400a7dbe31c87cd94bee1f16e0554efd8a7aaf5693eb3e181ef" Dec 13 09:26:53 crc kubenswrapper[4971]: I1213 09:26:53.748461 4971 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"openstack-config\" (UniqueName: \"kubernetes.io/configmap/b0fcccc8-3a38-42b6-a10c-57a0a44b48c3-openstack-config\") pod \"tempest-tests-tempest-s01-single-test\" (UID: \"b0fcccc8-3a38-42b6-a10c-57a0a44b48c3\") " pod="openstack/tempest-tests-tempest-s01-single-test" Dec 13 09:26:53 crc kubenswrapper[4971]: I1213 09:26:53.748589 4971 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"openstack-config-secret\" (UniqueName: \"kubernetes.io/secret/b0fcccc8-3a38-42b6-a10c-57a0a44b48c3-openstack-config-secret\") pod \"tempest-tests-tempest-s01-single-test\" (UID: \"b0fcccc8-3a38-42b6-a10c-57a0a44b48c3\") " pod="openstack/tempest-tests-tempest-s01-single-test" Dec 13 09:26:53 crc kubenswrapper[4971]: I1213 09:26:53.748632 4971 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"test-operator-ephemeral-workdir\" (UniqueName: \"kubernetes.io/empty-dir/b0fcccc8-3a38-42b6-a10c-57a0a44b48c3-test-operator-ephemeral-workdir\") pod \"tempest-tests-tempest-s01-single-test\" (UID: \"b0fcccc8-3a38-42b6-a10c-57a0a44b48c3\") " pod="openstack/tempest-tests-tempest-s01-single-test" Dec 13 09:26:53 crc kubenswrapper[4971]: I1213 09:26:53.748666 4971 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-pwptn\" (UniqueName: \"kubernetes.io/projected/b0fcccc8-3a38-42b6-a10c-57a0a44b48c3-kube-api-access-pwptn\") pod \"tempest-tests-tempest-s01-single-test\" (UID: \"b0fcccc8-3a38-42b6-a10c-57a0a44b48c3\") " pod="openstack/tempest-tests-tempest-s01-single-test" Dec 13 09:26:53 crc kubenswrapper[4971]: I1213 09:26:53.748743 4971 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/b0fcccc8-3a38-42b6-a10c-57a0a44b48c3-config-data\") pod \"tempest-tests-tempest-s01-single-test\" (UID: \"b0fcccc8-3a38-42b6-a10c-57a0a44b48c3\") " pod="openstack/tempest-tests-tempest-s01-single-test" Dec 13 09:26:53 crc kubenswrapper[4971]: I1213 09:26:53.748783 4971 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ca-certs\" (UniqueName: \"kubernetes.io/secret/b0fcccc8-3a38-42b6-a10c-57a0a44b48c3-ca-certs\") pod \"tempest-tests-tempest-s01-single-test\" (UID: \"b0fcccc8-3a38-42b6-a10c-57a0a44b48c3\") " pod="openstack/tempest-tests-tempest-s01-single-test" Dec 13 09:26:53 crc kubenswrapper[4971]: I1213 09:26:53.748841 4971 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"test-operator-ephemeral-temporary\" (UniqueName: \"kubernetes.io/empty-dir/b0fcccc8-3a38-42b6-a10c-57a0a44b48c3-test-operator-ephemeral-temporary\") pod \"tempest-tests-tempest-s01-single-test\" (UID: \"b0fcccc8-3a38-42b6-a10c-57a0a44b48c3\") " pod="openstack/tempest-tests-tempest-s01-single-test" Dec 13 09:26:53 crc kubenswrapper[4971]: I1213 09:26:53.748877 4971 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/b0fcccc8-3a38-42b6-a10c-57a0a44b48c3-ceph\") pod \"tempest-tests-tempest-s01-single-test\" (UID: \"b0fcccc8-3a38-42b6-a10c-57a0a44b48c3\") " pod="openstack/tempest-tests-tempest-s01-single-test" Dec 13 09:26:53 crc kubenswrapper[4971]: I1213 09:26:53.748965 4971 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/b0fcccc8-3a38-42b6-a10c-57a0a44b48c3-ssh-key\") pod \"tempest-tests-tempest-s01-single-test\" (UID: \"b0fcccc8-3a38-42b6-a10c-57a0a44b48c3\") " pod="openstack/tempest-tests-tempest-s01-single-test" Dec 13 09:26:53 crc kubenswrapper[4971]: I1213 09:26:53.749564 4971 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"test-operator-ephemeral-workdir\" (UniqueName: \"kubernetes.io/empty-dir/b0fcccc8-3a38-42b6-a10c-57a0a44b48c3-test-operator-ephemeral-workdir\") pod \"tempest-tests-tempest-s01-single-test\" (UID: \"b0fcccc8-3a38-42b6-a10c-57a0a44b48c3\") " pod="openstack/tempest-tests-tempest-s01-single-test" Dec 13 09:26:53 crc kubenswrapper[4971]: I1213 09:26:53.750131 4971 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"openstack-config\" (UniqueName: \"kubernetes.io/configmap/b0fcccc8-3a38-42b6-a10c-57a0a44b48c3-openstack-config\") pod \"tempest-tests-tempest-s01-single-test\" (UID: \"b0fcccc8-3a38-42b6-a10c-57a0a44b48c3\") " pod="openstack/tempest-tests-tempest-s01-single-test" Dec 13 09:26:53 crc kubenswrapper[4971]: I1213 09:26:53.750596 4971 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"test-operator-ephemeral-temporary\" (UniqueName: \"kubernetes.io/empty-dir/b0fcccc8-3a38-42b6-a10c-57a0a44b48c3-test-operator-ephemeral-temporary\") pod \"tempest-tests-tempest-s01-single-test\" (UID: \"b0fcccc8-3a38-42b6-a10c-57a0a44b48c3\") " pod="openstack/tempest-tests-tempest-s01-single-test" Dec 13 09:26:53 crc kubenswrapper[4971]: I1213 09:26:53.751316 4971 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/b0fcccc8-3a38-42b6-a10c-57a0a44b48c3-config-data\") pod \"tempest-tests-tempest-s01-single-test\" (UID: \"b0fcccc8-3a38-42b6-a10c-57a0a44b48c3\") " pod="openstack/tempest-tests-tempest-s01-single-test" Dec 13 09:26:53 crc kubenswrapper[4971]: I1213 09:26:53.753255 4971 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"openstack-config-secret\" (UniqueName: \"kubernetes.io/secret/b0fcccc8-3a38-42b6-a10c-57a0a44b48c3-openstack-config-secret\") pod \"tempest-tests-tempest-s01-single-test\" (UID: \"b0fcccc8-3a38-42b6-a10c-57a0a44b48c3\") " pod="openstack/tempest-tests-tempest-s01-single-test" Dec 13 09:26:53 crc kubenswrapper[4971]: I1213 09:26:53.753456 4971 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/b0fcccc8-3a38-42b6-a10c-57a0a44b48c3-ssh-key\") pod \"tempest-tests-tempest-s01-single-test\" (UID: \"b0fcccc8-3a38-42b6-a10c-57a0a44b48c3\") " pod="openstack/tempest-tests-tempest-s01-single-test" Dec 13 09:26:53 crc kubenswrapper[4971]: I1213 09:26:53.753785 4971 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ca-certs\" (UniqueName: \"kubernetes.io/secret/b0fcccc8-3a38-42b6-a10c-57a0a44b48c3-ca-certs\") pod \"tempest-tests-tempest-s01-single-test\" (UID: \"b0fcccc8-3a38-42b6-a10c-57a0a44b48c3\") " pod="openstack/tempest-tests-tempest-s01-single-test" Dec 13 09:26:53 crc kubenswrapper[4971]: I1213 09:26:53.755721 4971 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/b0fcccc8-3a38-42b6-a10c-57a0a44b48c3-ceph\") pod \"tempest-tests-tempest-s01-single-test\" (UID: \"b0fcccc8-3a38-42b6-a10c-57a0a44b48c3\") " pod="openstack/tempest-tests-tempest-s01-single-test" Dec 13 09:26:53 crc kubenswrapper[4971]: I1213 09:26:53.769032 4971 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-pwptn\" (UniqueName: \"kubernetes.io/projected/b0fcccc8-3a38-42b6-a10c-57a0a44b48c3-kube-api-access-pwptn\") pod \"tempest-tests-tempest-s01-single-test\" (UID: \"b0fcccc8-3a38-42b6-a10c-57a0a44b48c3\") " pod="openstack/tempest-tests-tempest-s01-single-test" Dec 13 09:26:53 crc kubenswrapper[4971]: I1213 09:26:53.837231 4971 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/tempest-tests-tempest-s01-single-test" Dec 13 09:26:54 crc kubenswrapper[4971]: I1213 09:26:54.444203 4971 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/tempest-tests-tempest-s01-single-test"] Dec 13 09:26:54 crc kubenswrapper[4971]: I1213 09:26:54.737692 4971 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/tempest-tests-tempest-s01-single-test" event={"ID":"b0fcccc8-3a38-42b6-a10c-57a0a44b48c3","Type":"ContainerStarted","Data":"8fd093e74db72443fef79670a241b3b07bbb1e2b8867709469b691a472b3d640"} Dec 13 09:26:56 crc kubenswrapper[4971]: I1213 09:26:56.766508 4971 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/tempest-tests-tempest-s01-single-test" event={"ID":"b0fcccc8-3a38-42b6-a10c-57a0a44b48c3","Type":"ContainerStarted","Data":"39128e7c82af935994f5b9c56e004fa0f4ce9d8e22275a1983048044d97c2684"} Dec 13 09:26:56 crc kubenswrapper[4971]: I1213 09:26:56.807033 4971 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/tempest-tests-tempest-s01-single-test" podStartSLOduration=3.807001999 podStartE2EDuration="3.807001999s" podCreationTimestamp="2025-12-13 09:26:53 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-13 09:26:56.788754591 +0000 UTC m=+9473.393164039" watchObservedRunningTime="2025-12-13 09:26:56.807001999 +0000 UTC m=+9473.411411447" Dec 13 09:26:57 crc kubenswrapper[4971]: I1213 09:26:57.769170 4971 scope.go:117] "RemoveContainer" containerID="e616f2406e276879960fe4d0372c2d2a34f8c21a80e5202a5083f7852225bab9" Dec 13 09:26:57 crc kubenswrapper[4971]: E1213 09:26:57.769944 4971 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-82xjz_openshift-machine-config-operator(e0c4af12-edda-41d7-9b44-f87396174ef2)\"" pod="openshift-machine-config-operator/machine-config-daemon-82xjz" podUID="e0c4af12-edda-41d7-9b44-f87396174ef2" Dec 13 09:27:09 crc kubenswrapper[4971]: I1213 09:27:09.769607 4971 scope.go:117] "RemoveContainer" containerID="e616f2406e276879960fe4d0372c2d2a34f8c21a80e5202a5083f7852225bab9" Dec 13 09:27:09 crc kubenswrapper[4971]: E1213 09:27:09.770738 4971 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-82xjz_openshift-machine-config-operator(e0c4af12-edda-41d7-9b44-f87396174ef2)\"" pod="openshift-machine-config-operator/machine-config-daemon-82xjz" podUID="e0c4af12-edda-41d7-9b44-f87396174ef2" Dec 13 09:27:21 crc kubenswrapper[4971]: I1213 09:27:21.768878 4971 scope.go:117] "RemoveContainer" containerID="e616f2406e276879960fe4d0372c2d2a34f8c21a80e5202a5083f7852225bab9" Dec 13 09:27:21 crc kubenswrapper[4971]: E1213 09:27:21.770700 4971 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-82xjz_openshift-machine-config-operator(e0c4af12-edda-41d7-9b44-f87396174ef2)\"" pod="openshift-machine-config-operator/machine-config-daemon-82xjz" podUID="e0c4af12-edda-41d7-9b44-f87396174ef2" Dec 13 09:27:33 crc kubenswrapper[4971]: I1213 09:27:33.775626 4971 scope.go:117] "RemoveContainer" containerID="e616f2406e276879960fe4d0372c2d2a34f8c21a80e5202a5083f7852225bab9" Dec 13 09:27:33 crc kubenswrapper[4971]: E1213 09:27:33.776780 4971 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-82xjz_openshift-machine-config-operator(e0c4af12-edda-41d7-9b44-f87396174ef2)\"" pod="openshift-machine-config-operator/machine-config-daemon-82xjz" podUID="e0c4af12-edda-41d7-9b44-f87396174ef2" Dec 13 09:27:44 crc kubenswrapper[4971]: I1213 09:27:44.769548 4971 scope.go:117] "RemoveContainer" containerID="e616f2406e276879960fe4d0372c2d2a34f8c21a80e5202a5083f7852225bab9" Dec 13 09:27:44 crc kubenswrapper[4971]: E1213 09:27:44.770807 4971 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-82xjz_openshift-machine-config-operator(e0c4af12-edda-41d7-9b44-f87396174ef2)\"" pod="openshift-machine-config-operator/machine-config-daemon-82xjz" podUID="e0c4af12-edda-41d7-9b44-f87396174ef2" Dec 13 09:27:55 crc kubenswrapper[4971]: I1213 09:27:55.770220 4971 scope.go:117] "RemoveContainer" containerID="e616f2406e276879960fe4d0372c2d2a34f8c21a80e5202a5083f7852225bab9" Dec 13 09:27:56 crc kubenswrapper[4971]: I1213 09:27:56.379720 4971 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-82xjz" event={"ID":"e0c4af12-edda-41d7-9b44-f87396174ef2","Type":"ContainerStarted","Data":"541c0610036fe5d23e59f44d9df0207efecabde0efb5dc61dac954401fb98fc1"} Dec 13 09:30:00 crc kubenswrapper[4971]: I1213 09:30:00.154675 4971 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29426970-z6zsn"] Dec 13 09:30:00 crc kubenswrapper[4971]: I1213 09:30:00.159078 4971 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29426970-z6zsn" Dec 13 09:30:00 crc kubenswrapper[4971]: I1213 09:30:00.163779 4971 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"collect-profiles-dockercfg-kzf4t" Dec 13 09:30:00 crc kubenswrapper[4971]: I1213 09:30:00.164020 4971 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-operator-lifecycle-manager"/"collect-profiles-config" Dec 13 09:30:00 crc kubenswrapper[4971]: I1213 09:30:00.171322 4971 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29426970-z6zsn"] Dec 13 09:30:00 crc kubenswrapper[4971]: I1213 09:30:00.246130 4971 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/672f66fb-60bb-4813-bf6e-ffc00328d258-secret-volume\") pod \"collect-profiles-29426970-z6zsn\" (UID: \"672f66fb-60bb-4813-bf6e-ffc00328d258\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29426970-z6zsn" Dec 13 09:30:00 crc kubenswrapper[4971]: I1213 09:30:00.246263 4971 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-4h6vc\" (UniqueName: \"kubernetes.io/projected/672f66fb-60bb-4813-bf6e-ffc00328d258-kube-api-access-4h6vc\") pod \"collect-profiles-29426970-z6zsn\" (UID: \"672f66fb-60bb-4813-bf6e-ffc00328d258\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29426970-z6zsn" Dec 13 09:30:00 crc kubenswrapper[4971]: I1213 09:30:00.246306 4971 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/672f66fb-60bb-4813-bf6e-ffc00328d258-config-volume\") pod \"collect-profiles-29426970-z6zsn\" (UID: \"672f66fb-60bb-4813-bf6e-ffc00328d258\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29426970-z6zsn" Dec 13 09:30:00 crc kubenswrapper[4971]: I1213 09:30:00.348768 4971 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-4h6vc\" (UniqueName: \"kubernetes.io/projected/672f66fb-60bb-4813-bf6e-ffc00328d258-kube-api-access-4h6vc\") pod \"collect-profiles-29426970-z6zsn\" (UID: \"672f66fb-60bb-4813-bf6e-ffc00328d258\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29426970-z6zsn" Dec 13 09:30:00 crc kubenswrapper[4971]: I1213 09:30:00.348868 4971 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/672f66fb-60bb-4813-bf6e-ffc00328d258-config-volume\") pod \"collect-profiles-29426970-z6zsn\" (UID: \"672f66fb-60bb-4813-bf6e-ffc00328d258\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29426970-z6zsn" Dec 13 09:30:00 crc kubenswrapper[4971]: I1213 09:30:00.348955 4971 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/672f66fb-60bb-4813-bf6e-ffc00328d258-secret-volume\") pod \"collect-profiles-29426970-z6zsn\" (UID: \"672f66fb-60bb-4813-bf6e-ffc00328d258\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29426970-z6zsn" Dec 13 09:30:00 crc kubenswrapper[4971]: I1213 09:30:00.350189 4971 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/672f66fb-60bb-4813-bf6e-ffc00328d258-config-volume\") pod \"collect-profiles-29426970-z6zsn\" (UID: \"672f66fb-60bb-4813-bf6e-ffc00328d258\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29426970-z6zsn" Dec 13 09:30:00 crc kubenswrapper[4971]: I1213 09:30:00.356782 4971 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/672f66fb-60bb-4813-bf6e-ffc00328d258-secret-volume\") pod \"collect-profiles-29426970-z6zsn\" (UID: \"672f66fb-60bb-4813-bf6e-ffc00328d258\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29426970-z6zsn" Dec 13 09:30:00 crc kubenswrapper[4971]: I1213 09:30:00.367291 4971 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-4h6vc\" (UniqueName: \"kubernetes.io/projected/672f66fb-60bb-4813-bf6e-ffc00328d258-kube-api-access-4h6vc\") pod \"collect-profiles-29426970-z6zsn\" (UID: \"672f66fb-60bb-4813-bf6e-ffc00328d258\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29426970-z6zsn" Dec 13 09:30:00 crc kubenswrapper[4971]: I1213 09:30:00.481241 4971 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29426970-z6zsn" Dec 13 09:30:00 crc kubenswrapper[4971]: I1213 09:30:00.954251 4971 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29426970-z6zsn"] Dec 13 09:30:01 crc kubenswrapper[4971]: E1213 09:30:01.763642 4971 cadvisor_stats_provider.go:516] "Partial failure issuing cadvisor.ContainerInfoV2" err="partial failures: [\"/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod672f66fb_60bb_4813_bf6e_ffc00328d258.slice/crio-bb2779949576056b857501c439010961eb622eed629231965cbafd39bbd6c0f7.scope\": RecentStats: unable to find data in memory cache]" Dec 13 09:30:01 crc kubenswrapper[4971]: I1213 09:30:01.788873 4971 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29426970-z6zsn" event={"ID":"672f66fb-60bb-4813-bf6e-ffc00328d258","Type":"ContainerStarted","Data":"bb2779949576056b857501c439010961eb622eed629231965cbafd39bbd6c0f7"} Dec 13 09:30:01 crc kubenswrapper[4971]: I1213 09:30:01.788947 4971 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29426970-z6zsn" event={"ID":"672f66fb-60bb-4813-bf6e-ffc00328d258","Type":"ContainerStarted","Data":"977f570a292a835fe5b6620ab18e221e4edefc60189e9d92560b51a3bf9f232f"} Dec 13 09:30:01 crc kubenswrapper[4971]: I1213 09:30:01.817932 4971 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-operator-lifecycle-manager/collect-profiles-29426970-z6zsn" podStartSLOduration=1.8179011630000002 podStartE2EDuration="1.817901163s" podCreationTimestamp="2025-12-13 09:30:00 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-13 09:30:01.80745782 +0000 UTC m=+9658.411867278" watchObservedRunningTime="2025-12-13 09:30:01.817901163 +0000 UTC m=+9658.422310631" Dec 13 09:30:02 crc kubenswrapper[4971]: I1213 09:30:02.790988 4971 generic.go:334] "Generic (PLEG): container finished" podID="672f66fb-60bb-4813-bf6e-ffc00328d258" containerID="bb2779949576056b857501c439010961eb622eed629231965cbafd39bbd6c0f7" exitCode=0 Dec 13 09:30:02 crc kubenswrapper[4971]: I1213 09:30:02.791405 4971 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29426970-z6zsn" event={"ID":"672f66fb-60bb-4813-bf6e-ffc00328d258","Type":"ContainerDied","Data":"bb2779949576056b857501c439010961eb622eed629231965cbafd39bbd6c0f7"} Dec 13 09:30:04 crc kubenswrapper[4971]: I1213 09:30:04.813708 4971 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29426970-z6zsn" event={"ID":"672f66fb-60bb-4813-bf6e-ffc00328d258","Type":"ContainerDied","Data":"977f570a292a835fe5b6620ab18e221e4edefc60189e9d92560b51a3bf9f232f"} Dec 13 09:30:04 crc kubenswrapper[4971]: I1213 09:30:04.814083 4971 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="977f570a292a835fe5b6620ab18e221e4edefc60189e9d92560b51a3bf9f232f" Dec 13 09:30:04 crc kubenswrapper[4971]: I1213 09:30:04.842943 4971 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29426970-z6zsn" Dec 13 09:30:04 crc kubenswrapper[4971]: I1213 09:30:04.940359 4971 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/672f66fb-60bb-4813-bf6e-ffc00328d258-config-volume\") pod \"672f66fb-60bb-4813-bf6e-ffc00328d258\" (UID: \"672f66fb-60bb-4813-bf6e-ffc00328d258\") " Dec 13 09:30:04 crc kubenswrapper[4971]: I1213 09:30:04.940960 4971 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/672f66fb-60bb-4813-bf6e-ffc00328d258-secret-volume\") pod \"672f66fb-60bb-4813-bf6e-ffc00328d258\" (UID: \"672f66fb-60bb-4813-bf6e-ffc00328d258\") " Dec 13 09:30:04 crc kubenswrapper[4971]: I1213 09:30:04.940946 4971 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/672f66fb-60bb-4813-bf6e-ffc00328d258-config-volume" (OuterVolumeSpecName: "config-volume") pod "672f66fb-60bb-4813-bf6e-ffc00328d258" (UID: "672f66fb-60bb-4813-bf6e-ffc00328d258"). InnerVolumeSpecName "config-volume". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 13 09:30:04 crc kubenswrapper[4971]: I1213 09:30:04.941135 4971 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-4h6vc\" (UniqueName: \"kubernetes.io/projected/672f66fb-60bb-4813-bf6e-ffc00328d258-kube-api-access-4h6vc\") pod \"672f66fb-60bb-4813-bf6e-ffc00328d258\" (UID: \"672f66fb-60bb-4813-bf6e-ffc00328d258\") " Dec 13 09:30:04 crc kubenswrapper[4971]: I1213 09:30:04.941944 4971 reconciler_common.go:293] "Volume detached for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/672f66fb-60bb-4813-bf6e-ffc00328d258-config-volume\") on node \"crc\" DevicePath \"\"" Dec 13 09:30:04 crc kubenswrapper[4971]: I1213 09:30:04.948498 4971 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/672f66fb-60bb-4813-bf6e-ffc00328d258-kube-api-access-4h6vc" (OuterVolumeSpecName: "kube-api-access-4h6vc") pod "672f66fb-60bb-4813-bf6e-ffc00328d258" (UID: "672f66fb-60bb-4813-bf6e-ffc00328d258"). InnerVolumeSpecName "kube-api-access-4h6vc". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 13 09:30:04 crc kubenswrapper[4971]: I1213 09:30:04.952725 4971 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/672f66fb-60bb-4813-bf6e-ffc00328d258-secret-volume" (OuterVolumeSpecName: "secret-volume") pod "672f66fb-60bb-4813-bf6e-ffc00328d258" (UID: "672f66fb-60bb-4813-bf6e-ffc00328d258"). InnerVolumeSpecName "secret-volume". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 13 09:30:05 crc kubenswrapper[4971]: I1213 09:30:05.043420 4971 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-4h6vc\" (UniqueName: \"kubernetes.io/projected/672f66fb-60bb-4813-bf6e-ffc00328d258-kube-api-access-4h6vc\") on node \"crc\" DevicePath \"\"" Dec 13 09:30:05 crc kubenswrapper[4971]: I1213 09:30:05.043478 4971 reconciler_common.go:293] "Volume detached for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/672f66fb-60bb-4813-bf6e-ffc00328d258-secret-volume\") on node \"crc\" DevicePath \"\"" Dec 13 09:30:05 crc kubenswrapper[4971]: I1213 09:30:05.362510 4971 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/certified-operators-7dhvp"] Dec 13 09:30:05 crc kubenswrapper[4971]: E1213 09:30:05.363827 4971 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="672f66fb-60bb-4813-bf6e-ffc00328d258" containerName="collect-profiles" Dec 13 09:30:05 crc kubenswrapper[4971]: I1213 09:30:05.363896 4971 state_mem.go:107] "Deleted CPUSet assignment" podUID="672f66fb-60bb-4813-bf6e-ffc00328d258" containerName="collect-profiles" Dec 13 09:30:05 crc kubenswrapper[4971]: I1213 09:30:05.364351 4971 memory_manager.go:354] "RemoveStaleState removing state" podUID="672f66fb-60bb-4813-bf6e-ffc00328d258" containerName="collect-profiles" Dec 13 09:30:05 crc kubenswrapper[4971]: I1213 09:30:05.366817 4971 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-7dhvp" Dec 13 09:30:05 crc kubenswrapper[4971]: I1213 09:30:05.382276 4971 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-7dhvp"] Dec 13 09:30:05 crc kubenswrapper[4971]: I1213 09:30:05.452254 4971 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-rczt4\" (UniqueName: \"kubernetes.io/projected/743ee964-c553-483c-9130-a20e9463c12c-kube-api-access-rczt4\") pod \"certified-operators-7dhvp\" (UID: \"743ee964-c553-483c-9130-a20e9463c12c\") " pod="openshift-marketplace/certified-operators-7dhvp" Dec 13 09:30:05 crc kubenswrapper[4971]: I1213 09:30:05.452470 4971 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/743ee964-c553-483c-9130-a20e9463c12c-catalog-content\") pod \"certified-operators-7dhvp\" (UID: \"743ee964-c553-483c-9130-a20e9463c12c\") " pod="openshift-marketplace/certified-operators-7dhvp" Dec 13 09:30:05 crc kubenswrapper[4971]: I1213 09:30:05.452727 4971 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/743ee964-c553-483c-9130-a20e9463c12c-utilities\") pod \"certified-operators-7dhvp\" (UID: \"743ee964-c553-483c-9130-a20e9463c12c\") " pod="openshift-marketplace/certified-operators-7dhvp" Dec 13 09:30:05 crc kubenswrapper[4971]: I1213 09:30:05.554472 4971 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/743ee964-c553-483c-9130-a20e9463c12c-catalog-content\") pod \"certified-operators-7dhvp\" (UID: \"743ee964-c553-483c-9130-a20e9463c12c\") " pod="openshift-marketplace/certified-operators-7dhvp" Dec 13 09:30:05 crc kubenswrapper[4971]: I1213 09:30:05.554628 4971 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/743ee964-c553-483c-9130-a20e9463c12c-utilities\") pod \"certified-operators-7dhvp\" (UID: \"743ee964-c553-483c-9130-a20e9463c12c\") " pod="openshift-marketplace/certified-operators-7dhvp" Dec 13 09:30:05 crc kubenswrapper[4971]: I1213 09:30:05.554777 4971 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-rczt4\" (UniqueName: \"kubernetes.io/projected/743ee964-c553-483c-9130-a20e9463c12c-kube-api-access-rczt4\") pod \"certified-operators-7dhvp\" (UID: \"743ee964-c553-483c-9130-a20e9463c12c\") " pod="openshift-marketplace/certified-operators-7dhvp" Dec 13 09:30:05 crc kubenswrapper[4971]: I1213 09:30:05.555151 4971 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/743ee964-c553-483c-9130-a20e9463c12c-catalog-content\") pod \"certified-operators-7dhvp\" (UID: \"743ee964-c553-483c-9130-a20e9463c12c\") " pod="openshift-marketplace/certified-operators-7dhvp" Dec 13 09:30:05 crc kubenswrapper[4971]: I1213 09:30:05.555315 4971 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/743ee964-c553-483c-9130-a20e9463c12c-utilities\") pod \"certified-operators-7dhvp\" (UID: \"743ee964-c553-483c-9130-a20e9463c12c\") " pod="openshift-marketplace/certified-operators-7dhvp" Dec 13 09:30:05 crc kubenswrapper[4971]: I1213 09:30:05.584275 4971 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-rczt4\" (UniqueName: \"kubernetes.io/projected/743ee964-c553-483c-9130-a20e9463c12c-kube-api-access-rczt4\") pod \"certified-operators-7dhvp\" (UID: \"743ee964-c553-483c-9130-a20e9463c12c\") " pod="openshift-marketplace/certified-operators-7dhvp" Dec 13 09:30:05 crc kubenswrapper[4971]: I1213 09:30:05.696589 4971 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-7dhvp" Dec 13 09:30:05 crc kubenswrapper[4971]: I1213 09:30:05.833842 4971 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29426970-z6zsn" Dec 13 09:30:05 crc kubenswrapper[4971]: I1213 09:30:05.986107 4971 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29426925-rlgh8"] Dec 13 09:30:06 crc kubenswrapper[4971]: I1213 09:30:06.010676 4971 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29426925-rlgh8"] Dec 13 09:30:06 crc kubenswrapper[4971]: I1213 09:30:06.296285 4971 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-7dhvp"] Dec 13 09:30:06 crc kubenswrapper[4971]: I1213 09:30:06.852815 4971 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-7dhvp" event={"ID":"743ee964-c553-483c-9130-a20e9463c12c","Type":"ContainerStarted","Data":"9bfe20470383563a134b595bab0481630b1feeabaf008f2bd7e444b157ddc8cb"} Dec 13 09:30:06 crc kubenswrapper[4971]: I1213 09:30:06.853280 4971 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-7dhvp" event={"ID":"743ee964-c553-483c-9130-a20e9463c12c","Type":"ContainerStarted","Data":"24e09cf1a5240e9260a1bddde349a1aef1ae41017eada292a2fcf097b31a1a7b"} Dec 13 09:30:07 crc kubenswrapper[4971]: I1213 09:30:07.783422 4971 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="83f18279-1f12-4157-8aad-1e1de8bdbbdf" path="/var/lib/kubelet/pods/83f18279-1f12-4157-8aad-1e1de8bdbbdf/volumes" Dec 13 09:30:07 crc kubenswrapper[4971]: I1213 09:30:07.864768 4971 generic.go:334] "Generic (PLEG): container finished" podID="743ee964-c553-483c-9130-a20e9463c12c" containerID="9bfe20470383563a134b595bab0481630b1feeabaf008f2bd7e444b157ddc8cb" exitCode=0 Dec 13 09:30:07 crc kubenswrapper[4971]: I1213 09:30:07.864825 4971 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-7dhvp" event={"ID":"743ee964-c553-483c-9130-a20e9463c12c","Type":"ContainerDied","Data":"9bfe20470383563a134b595bab0481630b1feeabaf008f2bd7e444b157ddc8cb"} Dec 13 09:30:09 crc kubenswrapper[4971]: I1213 09:30:09.888281 4971 generic.go:334] "Generic (PLEG): container finished" podID="743ee964-c553-483c-9130-a20e9463c12c" containerID="d6d6676638e3a87034cc6b432c993cb06a4951f3b6c3ea8f970723f7d7b998ba" exitCode=0 Dec 13 09:30:09 crc kubenswrapper[4971]: I1213 09:30:09.888356 4971 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-7dhvp" event={"ID":"743ee964-c553-483c-9130-a20e9463c12c","Type":"ContainerDied","Data":"d6d6676638e3a87034cc6b432c993cb06a4951f3b6c3ea8f970723f7d7b998ba"} Dec 13 09:30:10 crc kubenswrapper[4971]: I1213 09:30:10.905252 4971 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-7dhvp" event={"ID":"743ee964-c553-483c-9130-a20e9463c12c","Type":"ContainerStarted","Data":"8ea69372e1d03c1fb85bb1e5d2f6323acf2fa1dfd5f99c506b74495a2fd09f04"} Dec 13 09:30:10 crc kubenswrapper[4971]: I1213 09:30:10.940274 4971 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/certified-operators-7dhvp" podStartSLOduration=3.435671359 podStartE2EDuration="5.940236751s" podCreationTimestamp="2025-12-13 09:30:05 +0000 UTC" firstStartedPulling="2025-12-13 09:30:07.867991227 +0000 UTC m=+9664.472400675" lastFinishedPulling="2025-12-13 09:30:10.372556619 +0000 UTC m=+9666.976966067" observedRunningTime="2025-12-13 09:30:10.923398714 +0000 UTC m=+9667.527808202" watchObservedRunningTime="2025-12-13 09:30:10.940236751 +0000 UTC m=+9667.544646199" Dec 13 09:30:15 crc kubenswrapper[4971]: I1213 09:30:15.697230 4971 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/certified-operators-7dhvp" Dec 13 09:30:15 crc kubenswrapper[4971]: I1213 09:30:15.698063 4971 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/certified-operators-7dhvp" Dec 13 09:30:15 crc kubenswrapper[4971]: I1213 09:30:15.744893 4971 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/certified-operators-7dhvp" Dec 13 09:30:16 crc kubenswrapper[4971]: I1213 09:30:16.013696 4971 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/certified-operators-7dhvp" Dec 13 09:30:16 crc kubenswrapper[4971]: I1213 09:30:16.082879 4971 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-7dhvp"] Dec 13 09:30:16 crc kubenswrapper[4971]: I1213 09:30:16.156184 4971 patch_prober.go:28] interesting pod/machine-config-daemon-82xjz container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 13 09:30:16 crc kubenswrapper[4971]: I1213 09:30:16.156261 4971 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-82xjz" podUID="e0c4af12-edda-41d7-9b44-f87396174ef2" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 13 09:30:17 crc kubenswrapper[4971]: I1213 09:30:17.976878 4971 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/certified-operators-7dhvp" podUID="743ee964-c553-483c-9130-a20e9463c12c" containerName="registry-server" containerID="cri-o://8ea69372e1d03c1fb85bb1e5d2f6323acf2fa1dfd5f99c506b74495a2fd09f04" gracePeriod=2 Dec 13 09:30:18 crc kubenswrapper[4971]: I1213 09:30:18.994055 4971 generic.go:334] "Generic (PLEG): container finished" podID="743ee964-c553-483c-9130-a20e9463c12c" containerID="8ea69372e1d03c1fb85bb1e5d2f6323acf2fa1dfd5f99c506b74495a2fd09f04" exitCode=0 Dec 13 09:30:18 crc kubenswrapper[4971]: I1213 09:30:18.994117 4971 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-7dhvp" event={"ID":"743ee964-c553-483c-9130-a20e9463c12c","Type":"ContainerDied","Data":"8ea69372e1d03c1fb85bb1e5d2f6323acf2fa1dfd5f99c506b74495a2fd09f04"} Dec 13 09:30:19 crc kubenswrapper[4971]: I1213 09:30:19.316095 4971 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-7dhvp" Dec 13 09:30:19 crc kubenswrapper[4971]: I1213 09:30:19.484805 4971 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/743ee964-c553-483c-9130-a20e9463c12c-utilities\") pod \"743ee964-c553-483c-9130-a20e9463c12c\" (UID: \"743ee964-c553-483c-9130-a20e9463c12c\") " Dec 13 09:30:19 crc kubenswrapper[4971]: I1213 09:30:19.485003 4971 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/743ee964-c553-483c-9130-a20e9463c12c-catalog-content\") pod \"743ee964-c553-483c-9130-a20e9463c12c\" (UID: \"743ee964-c553-483c-9130-a20e9463c12c\") " Dec 13 09:30:19 crc kubenswrapper[4971]: I1213 09:30:19.485261 4971 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-rczt4\" (UniqueName: \"kubernetes.io/projected/743ee964-c553-483c-9130-a20e9463c12c-kube-api-access-rczt4\") pod \"743ee964-c553-483c-9130-a20e9463c12c\" (UID: \"743ee964-c553-483c-9130-a20e9463c12c\") " Dec 13 09:30:19 crc kubenswrapper[4971]: I1213 09:30:19.486563 4971 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/743ee964-c553-483c-9130-a20e9463c12c-utilities" (OuterVolumeSpecName: "utilities") pod "743ee964-c553-483c-9130-a20e9463c12c" (UID: "743ee964-c553-483c-9130-a20e9463c12c"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 13 09:30:19 crc kubenswrapper[4971]: I1213 09:30:19.498573 4971 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/743ee964-c553-483c-9130-a20e9463c12c-kube-api-access-rczt4" (OuterVolumeSpecName: "kube-api-access-rczt4") pod "743ee964-c553-483c-9130-a20e9463c12c" (UID: "743ee964-c553-483c-9130-a20e9463c12c"). InnerVolumeSpecName "kube-api-access-rczt4". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 13 09:30:19 crc kubenswrapper[4971]: I1213 09:30:19.560785 4971 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/743ee964-c553-483c-9130-a20e9463c12c-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "743ee964-c553-483c-9130-a20e9463c12c" (UID: "743ee964-c553-483c-9130-a20e9463c12c"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 13 09:30:19 crc kubenswrapper[4971]: I1213 09:30:19.589200 4971 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-rczt4\" (UniqueName: \"kubernetes.io/projected/743ee964-c553-483c-9130-a20e9463c12c-kube-api-access-rczt4\") on node \"crc\" DevicePath \"\"" Dec 13 09:30:19 crc kubenswrapper[4971]: I1213 09:30:19.589248 4971 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/743ee964-c553-483c-9130-a20e9463c12c-utilities\") on node \"crc\" DevicePath \"\"" Dec 13 09:30:19 crc kubenswrapper[4971]: I1213 09:30:19.589260 4971 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/743ee964-c553-483c-9130-a20e9463c12c-catalog-content\") on node \"crc\" DevicePath \"\"" Dec 13 09:30:20 crc kubenswrapper[4971]: I1213 09:30:20.006432 4971 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-7dhvp" event={"ID":"743ee964-c553-483c-9130-a20e9463c12c","Type":"ContainerDied","Data":"24e09cf1a5240e9260a1bddde349a1aef1ae41017eada292a2fcf097b31a1a7b"} Dec 13 09:30:20 crc kubenswrapper[4971]: I1213 09:30:20.006496 4971 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-7dhvp" Dec 13 09:30:20 crc kubenswrapper[4971]: I1213 09:30:20.006816 4971 scope.go:117] "RemoveContainer" containerID="8ea69372e1d03c1fb85bb1e5d2f6323acf2fa1dfd5f99c506b74495a2fd09f04" Dec 13 09:30:20 crc kubenswrapper[4971]: I1213 09:30:20.043418 4971 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-7dhvp"] Dec 13 09:30:20 crc kubenswrapper[4971]: I1213 09:30:20.044340 4971 scope.go:117] "RemoveContainer" containerID="d6d6676638e3a87034cc6b432c993cb06a4951f3b6c3ea8f970723f7d7b998ba" Dec 13 09:30:20 crc kubenswrapper[4971]: I1213 09:30:20.056539 4971 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/certified-operators-7dhvp"] Dec 13 09:30:20 crc kubenswrapper[4971]: I1213 09:30:20.072002 4971 scope.go:117] "RemoveContainer" containerID="9bfe20470383563a134b595bab0481630b1feeabaf008f2bd7e444b157ddc8cb" Dec 13 09:30:20 crc kubenswrapper[4971]: I1213 09:30:20.614116 4971 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/community-operators-6hzwh"] Dec 13 09:30:20 crc kubenswrapper[4971]: E1213 09:30:20.614776 4971 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="743ee964-c553-483c-9130-a20e9463c12c" containerName="extract-utilities" Dec 13 09:30:20 crc kubenswrapper[4971]: I1213 09:30:20.614793 4971 state_mem.go:107] "Deleted CPUSet assignment" podUID="743ee964-c553-483c-9130-a20e9463c12c" containerName="extract-utilities" Dec 13 09:30:20 crc kubenswrapper[4971]: E1213 09:30:20.614809 4971 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="743ee964-c553-483c-9130-a20e9463c12c" containerName="extract-content" Dec 13 09:30:20 crc kubenswrapper[4971]: I1213 09:30:20.614817 4971 state_mem.go:107] "Deleted CPUSet assignment" podUID="743ee964-c553-483c-9130-a20e9463c12c" containerName="extract-content" Dec 13 09:30:20 crc kubenswrapper[4971]: E1213 09:30:20.614833 4971 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="743ee964-c553-483c-9130-a20e9463c12c" containerName="registry-server" Dec 13 09:30:20 crc kubenswrapper[4971]: I1213 09:30:20.614841 4971 state_mem.go:107] "Deleted CPUSet assignment" podUID="743ee964-c553-483c-9130-a20e9463c12c" containerName="registry-server" Dec 13 09:30:20 crc kubenswrapper[4971]: I1213 09:30:20.615132 4971 memory_manager.go:354] "RemoveStaleState removing state" podUID="743ee964-c553-483c-9130-a20e9463c12c" containerName="registry-server" Dec 13 09:30:20 crc kubenswrapper[4971]: I1213 09:30:20.616828 4971 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-6hzwh" Dec 13 09:30:20 crc kubenswrapper[4971]: I1213 09:30:20.629182 4971 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-6hzwh"] Dec 13 09:30:20 crc kubenswrapper[4971]: I1213 09:30:20.717101 4971 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-ns929\" (UniqueName: \"kubernetes.io/projected/f18bc810-f35c-4dc5-a2f9-7ea5f1594371-kube-api-access-ns929\") pod \"community-operators-6hzwh\" (UID: \"f18bc810-f35c-4dc5-a2f9-7ea5f1594371\") " pod="openshift-marketplace/community-operators-6hzwh" Dec 13 09:30:20 crc kubenswrapper[4971]: I1213 09:30:20.717178 4971 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/f18bc810-f35c-4dc5-a2f9-7ea5f1594371-catalog-content\") pod \"community-operators-6hzwh\" (UID: \"f18bc810-f35c-4dc5-a2f9-7ea5f1594371\") " pod="openshift-marketplace/community-operators-6hzwh" Dec 13 09:30:20 crc kubenswrapper[4971]: I1213 09:30:20.717945 4971 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/f18bc810-f35c-4dc5-a2f9-7ea5f1594371-utilities\") pod \"community-operators-6hzwh\" (UID: \"f18bc810-f35c-4dc5-a2f9-7ea5f1594371\") " pod="openshift-marketplace/community-operators-6hzwh" Dec 13 09:30:20 crc kubenswrapper[4971]: I1213 09:30:20.821260 4971 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/f18bc810-f35c-4dc5-a2f9-7ea5f1594371-utilities\") pod \"community-operators-6hzwh\" (UID: \"f18bc810-f35c-4dc5-a2f9-7ea5f1594371\") " pod="openshift-marketplace/community-operators-6hzwh" Dec 13 09:30:20 crc kubenswrapper[4971]: I1213 09:30:20.821481 4971 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-ns929\" (UniqueName: \"kubernetes.io/projected/f18bc810-f35c-4dc5-a2f9-7ea5f1594371-kube-api-access-ns929\") pod \"community-operators-6hzwh\" (UID: \"f18bc810-f35c-4dc5-a2f9-7ea5f1594371\") " pod="openshift-marketplace/community-operators-6hzwh" Dec 13 09:30:20 crc kubenswrapper[4971]: I1213 09:30:20.821632 4971 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/f18bc810-f35c-4dc5-a2f9-7ea5f1594371-catalog-content\") pod \"community-operators-6hzwh\" (UID: \"f18bc810-f35c-4dc5-a2f9-7ea5f1594371\") " pod="openshift-marketplace/community-operators-6hzwh" Dec 13 09:30:20 crc kubenswrapper[4971]: I1213 09:30:20.822295 4971 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/f18bc810-f35c-4dc5-a2f9-7ea5f1594371-catalog-content\") pod \"community-operators-6hzwh\" (UID: \"f18bc810-f35c-4dc5-a2f9-7ea5f1594371\") " pod="openshift-marketplace/community-operators-6hzwh" Dec 13 09:30:20 crc kubenswrapper[4971]: I1213 09:30:20.822431 4971 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/f18bc810-f35c-4dc5-a2f9-7ea5f1594371-utilities\") pod \"community-operators-6hzwh\" (UID: \"f18bc810-f35c-4dc5-a2f9-7ea5f1594371\") " pod="openshift-marketplace/community-operators-6hzwh" Dec 13 09:30:20 crc kubenswrapper[4971]: I1213 09:30:20.845319 4971 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-ns929\" (UniqueName: \"kubernetes.io/projected/f18bc810-f35c-4dc5-a2f9-7ea5f1594371-kube-api-access-ns929\") pod \"community-operators-6hzwh\" (UID: \"f18bc810-f35c-4dc5-a2f9-7ea5f1594371\") " pod="openshift-marketplace/community-operators-6hzwh" Dec 13 09:30:20 crc kubenswrapper[4971]: I1213 09:30:20.938797 4971 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-6hzwh" Dec 13 09:30:21 crc kubenswrapper[4971]: I1213 09:30:21.294694 4971 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-operators-zpdtr"] Dec 13 09:30:21 crc kubenswrapper[4971]: I1213 09:30:21.303368 4971 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-zpdtr" Dec 13 09:30:21 crc kubenswrapper[4971]: I1213 09:30:21.316742 4971 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-zpdtr"] Dec 13 09:30:21 crc kubenswrapper[4971]: I1213 09:30:21.453594 4971 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-4jjtp\" (UniqueName: \"kubernetes.io/projected/134aaef2-706e-41b6-a16f-05aba393a4b7-kube-api-access-4jjtp\") pod \"redhat-operators-zpdtr\" (UID: \"134aaef2-706e-41b6-a16f-05aba393a4b7\") " pod="openshift-marketplace/redhat-operators-zpdtr" Dec 13 09:30:21 crc kubenswrapper[4971]: I1213 09:30:21.454145 4971 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/134aaef2-706e-41b6-a16f-05aba393a4b7-catalog-content\") pod \"redhat-operators-zpdtr\" (UID: \"134aaef2-706e-41b6-a16f-05aba393a4b7\") " pod="openshift-marketplace/redhat-operators-zpdtr" Dec 13 09:30:21 crc kubenswrapper[4971]: I1213 09:30:21.454234 4971 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/134aaef2-706e-41b6-a16f-05aba393a4b7-utilities\") pod \"redhat-operators-zpdtr\" (UID: \"134aaef2-706e-41b6-a16f-05aba393a4b7\") " pod="openshift-marketplace/redhat-operators-zpdtr" Dec 13 09:30:21 crc kubenswrapper[4971]: I1213 09:30:21.556143 4971 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/134aaef2-706e-41b6-a16f-05aba393a4b7-utilities\") pod \"redhat-operators-zpdtr\" (UID: \"134aaef2-706e-41b6-a16f-05aba393a4b7\") " pod="openshift-marketplace/redhat-operators-zpdtr" Dec 13 09:30:21 crc kubenswrapper[4971]: I1213 09:30:21.556311 4971 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-4jjtp\" (UniqueName: \"kubernetes.io/projected/134aaef2-706e-41b6-a16f-05aba393a4b7-kube-api-access-4jjtp\") pod \"redhat-operators-zpdtr\" (UID: \"134aaef2-706e-41b6-a16f-05aba393a4b7\") " pod="openshift-marketplace/redhat-operators-zpdtr" Dec 13 09:30:21 crc kubenswrapper[4971]: I1213 09:30:21.556442 4971 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/134aaef2-706e-41b6-a16f-05aba393a4b7-catalog-content\") pod \"redhat-operators-zpdtr\" (UID: \"134aaef2-706e-41b6-a16f-05aba393a4b7\") " pod="openshift-marketplace/redhat-operators-zpdtr" Dec 13 09:30:21 crc kubenswrapper[4971]: I1213 09:30:21.556957 4971 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/134aaef2-706e-41b6-a16f-05aba393a4b7-utilities\") pod \"redhat-operators-zpdtr\" (UID: \"134aaef2-706e-41b6-a16f-05aba393a4b7\") " pod="openshift-marketplace/redhat-operators-zpdtr" Dec 13 09:30:21 crc kubenswrapper[4971]: I1213 09:30:21.557003 4971 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/134aaef2-706e-41b6-a16f-05aba393a4b7-catalog-content\") pod \"redhat-operators-zpdtr\" (UID: \"134aaef2-706e-41b6-a16f-05aba393a4b7\") " pod="openshift-marketplace/redhat-operators-zpdtr" Dec 13 09:30:21 crc kubenswrapper[4971]: I1213 09:30:21.594147 4971 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-4jjtp\" (UniqueName: \"kubernetes.io/projected/134aaef2-706e-41b6-a16f-05aba393a4b7-kube-api-access-4jjtp\") pod \"redhat-operators-zpdtr\" (UID: \"134aaef2-706e-41b6-a16f-05aba393a4b7\") " pod="openshift-marketplace/redhat-operators-zpdtr" Dec 13 09:30:21 crc kubenswrapper[4971]: I1213 09:30:21.650755 4971 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-zpdtr" Dec 13 09:30:21 crc kubenswrapper[4971]: I1213 09:30:21.786001 4971 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="743ee964-c553-483c-9130-a20e9463c12c" path="/var/lib/kubelet/pods/743ee964-c553-483c-9130-a20e9463c12c/volumes" Dec 13 09:30:21 crc kubenswrapper[4971]: I1213 09:30:21.838509 4971 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-6hzwh"] Dec 13 09:30:22 crc kubenswrapper[4971]: I1213 09:30:22.027568 4971 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-6hzwh" event={"ID":"f18bc810-f35c-4dc5-a2f9-7ea5f1594371","Type":"ContainerStarted","Data":"902840b2be1cc063542b4454ca864ee0f9e24c7bf98853504991fe2a9e2cc544"} Dec 13 09:30:22 crc kubenswrapper[4971]: I1213 09:30:22.212634 4971 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-zpdtr"] Dec 13 09:30:22 crc kubenswrapper[4971]: W1213 09:30:22.215284 4971 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod134aaef2_706e_41b6_a16f_05aba393a4b7.slice/crio-b5c9ed6af3370d402703c06686bbc6f7b31066bd94556c6f1da6374751a41a53 WatchSource:0}: Error finding container b5c9ed6af3370d402703c06686bbc6f7b31066bd94556c6f1da6374751a41a53: Status 404 returned error can't find the container with id b5c9ed6af3370d402703c06686bbc6f7b31066bd94556c6f1da6374751a41a53 Dec 13 09:30:22 crc kubenswrapper[4971]: E1213 09:30:22.278756 4971 cadvisor_stats_provider.go:516] "Partial failure issuing cadvisor.ContainerInfoV2" err="partial failures: [\"/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podf18bc810_f35c_4dc5_a2f9_7ea5f1594371.slice/crio-conmon-cc86a53e32c2914bf906f356959a530cbefff1c9bf57ff202deb3fc5d945791c.scope\": RecentStats: unable to find data in memory cache], [\"/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podf18bc810_f35c_4dc5_a2f9_7ea5f1594371.slice/crio-cc86a53e32c2914bf906f356959a530cbefff1c9bf57ff202deb3fc5d945791c.scope\": RecentStats: unable to find data in memory cache]" Dec 13 09:30:23 crc kubenswrapper[4971]: I1213 09:30:23.042255 4971 generic.go:334] "Generic (PLEG): container finished" podID="f18bc810-f35c-4dc5-a2f9-7ea5f1594371" containerID="cc86a53e32c2914bf906f356959a530cbefff1c9bf57ff202deb3fc5d945791c" exitCode=0 Dec 13 09:30:23 crc kubenswrapper[4971]: I1213 09:30:23.042391 4971 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-6hzwh" event={"ID":"f18bc810-f35c-4dc5-a2f9-7ea5f1594371","Type":"ContainerDied","Data":"cc86a53e32c2914bf906f356959a530cbefff1c9bf57ff202deb3fc5d945791c"} Dec 13 09:30:23 crc kubenswrapper[4971]: I1213 09:30:23.046010 4971 generic.go:334] "Generic (PLEG): container finished" podID="134aaef2-706e-41b6-a16f-05aba393a4b7" containerID="f4d0df53e38ccde7e16d16696d036abbd48b77ebeabff6bba701b33c47b22a2f" exitCode=0 Dec 13 09:30:23 crc kubenswrapper[4971]: I1213 09:30:23.046075 4971 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-zpdtr" event={"ID":"134aaef2-706e-41b6-a16f-05aba393a4b7","Type":"ContainerDied","Data":"f4d0df53e38ccde7e16d16696d036abbd48b77ebeabff6bba701b33c47b22a2f"} Dec 13 09:30:23 crc kubenswrapper[4971]: I1213 09:30:23.046115 4971 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-zpdtr" event={"ID":"134aaef2-706e-41b6-a16f-05aba393a4b7","Type":"ContainerStarted","Data":"b5c9ed6af3370d402703c06686bbc6f7b31066bd94556c6f1da6374751a41a53"} Dec 13 09:30:26 crc kubenswrapper[4971]: I1213 09:30:26.108438 4971 generic.go:334] "Generic (PLEG): container finished" podID="f18bc810-f35c-4dc5-a2f9-7ea5f1594371" containerID="4a11761b1f295ff6eabacef438bd250e6f30cf74eb29cf762135f04ca49acb00" exitCode=0 Dec 13 09:30:26 crc kubenswrapper[4971]: I1213 09:30:26.108615 4971 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-6hzwh" event={"ID":"f18bc810-f35c-4dc5-a2f9-7ea5f1594371","Type":"ContainerDied","Data":"4a11761b1f295ff6eabacef438bd250e6f30cf74eb29cf762135f04ca49acb00"} Dec 13 09:30:26 crc kubenswrapper[4971]: I1213 09:30:26.116424 4971 generic.go:334] "Generic (PLEG): container finished" podID="134aaef2-706e-41b6-a16f-05aba393a4b7" containerID="67d045063d92fb28a571cce33040fddfa029bd5a8e17ee775f7a57e0a5285560" exitCode=0 Dec 13 09:30:26 crc kubenswrapper[4971]: I1213 09:30:26.116469 4971 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-zpdtr" event={"ID":"134aaef2-706e-41b6-a16f-05aba393a4b7","Type":"ContainerDied","Data":"67d045063d92fb28a571cce33040fddfa029bd5a8e17ee775f7a57e0a5285560"} Dec 13 09:30:28 crc kubenswrapper[4971]: I1213 09:30:28.143367 4971 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-zpdtr" event={"ID":"134aaef2-706e-41b6-a16f-05aba393a4b7","Type":"ContainerStarted","Data":"cc246f33190b446205a7fc29a9b3ef9ce057cb996b447fe0a3d9ddba24e052a4"} Dec 13 09:30:28 crc kubenswrapper[4971]: I1213 09:30:28.151612 4971 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-6hzwh" event={"ID":"f18bc810-f35c-4dc5-a2f9-7ea5f1594371","Type":"ContainerStarted","Data":"b4b20fa697a6ae02101018f66af436b00c15c45e779ad7b669b6c27b3bfdb99e"} Dec 13 09:30:28 crc kubenswrapper[4971]: I1213 09:30:28.170716 4971 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-operators-zpdtr" podStartSLOduration=2.988854131 podStartE2EDuration="7.170687647s" podCreationTimestamp="2025-12-13 09:30:21 +0000 UTC" firstStartedPulling="2025-12-13 09:30:23.048626819 +0000 UTC m=+9679.653036267" lastFinishedPulling="2025-12-13 09:30:27.230460335 +0000 UTC m=+9683.834869783" observedRunningTime="2025-12-13 09:30:28.169222552 +0000 UTC m=+9684.773632030" watchObservedRunningTime="2025-12-13 09:30:28.170687647 +0000 UTC m=+9684.775097095" Dec 13 09:30:28 crc kubenswrapper[4971]: I1213 09:30:28.198582 4971 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/community-operators-6hzwh" podStartSLOduration=4.676724484 podStartE2EDuration="8.198526641s" podCreationTimestamp="2025-12-13 09:30:20 +0000 UTC" firstStartedPulling="2025-12-13 09:30:23.045133195 +0000 UTC m=+9679.649542643" lastFinishedPulling="2025-12-13 09:30:26.566935352 +0000 UTC m=+9683.171344800" observedRunningTime="2025-12-13 09:30:28.188281043 +0000 UTC m=+9684.792690501" watchObservedRunningTime="2025-12-13 09:30:28.198526641 +0000 UTC m=+9684.802936089" Dec 13 09:30:29 crc kubenswrapper[4971]: I1213 09:30:29.880778 4971 scope.go:117] "RemoveContainer" containerID="db5a314515777dab5dfa0631e2f128f2f80f9846c80a9d3343d7959ea5da7a4a" Dec 13 09:30:31 crc kubenswrapper[4971]: I1213 09:30:31.274565 4971 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/community-operators-6hzwh" Dec 13 09:30:31 crc kubenswrapper[4971]: I1213 09:30:31.274621 4971 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/community-operators-6hzwh" Dec 13 09:30:31 crc kubenswrapper[4971]: I1213 09:30:31.367486 4971 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/community-operators-6hzwh" Dec 13 09:30:31 crc kubenswrapper[4971]: I1213 09:30:31.652828 4971 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-operators-zpdtr" Dec 13 09:30:31 crc kubenswrapper[4971]: I1213 09:30:31.652901 4971 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-operators-zpdtr" Dec 13 09:30:32 crc kubenswrapper[4971]: I1213 09:30:32.384379 4971 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/community-operators-6hzwh" Dec 13 09:30:32 crc kubenswrapper[4971]: I1213 09:30:32.447316 4971 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-6hzwh"] Dec 13 09:30:32 crc kubenswrapper[4971]: I1213 09:30:32.702371 4971 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-marketplace/redhat-operators-zpdtr" podUID="134aaef2-706e-41b6-a16f-05aba393a4b7" containerName="registry-server" probeResult="failure" output=< Dec 13 09:30:32 crc kubenswrapper[4971]: timeout: failed to connect service ":50051" within 1s Dec 13 09:30:32 crc kubenswrapper[4971]: > Dec 13 09:30:34 crc kubenswrapper[4971]: I1213 09:30:34.351596 4971 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/community-operators-6hzwh" podUID="f18bc810-f35c-4dc5-a2f9-7ea5f1594371" containerName="registry-server" containerID="cri-o://b4b20fa697a6ae02101018f66af436b00c15c45e779ad7b669b6c27b3bfdb99e" gracePeriod=2 Dec 13 09:30:36 crc kubenswrapper[4971]: I1213 09:30:36.375619 4971 generic.go:334] "Generic (PLEG): container finished" podID="f18bc810-f35c-4dc5-a2f9-7ea5f1594371" containerID="b4b20fa697a6ae02101018f66af436b00c15c45e779ad7b669b6c27b3bfdb99e" exitCode=0 Dec 13 09:30:36 crc kubenswrapper[4971]: I1213 09:30:36.375676 4971 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-6hzwh" event={"ID":"f18bc810-f35c-4dc5-a2f9-7ea5f1594371","Type":"ContainerDied","Data":"b4b20fa697a6ae02101018f66af436b00c15c45e779ad7b669b6c27b3bfdb99e"} Dec 13 09:30:37 crc kubenswrapper[4971]: I1213 09:30:37.396466 4971 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-6hzwh" event={"ID":"f18bc810-f35c-4dc5-a2f9-7ea5f1594371","Type":"ContainerDied","Data":"902840b2be1cc063542b4454ca864ee0f9e24c7bf98853504991fe2a9e2cc544"} Dec 13 09:30:37 crc kubenswrapper[4971]: I1213 09:30:37.396954 4971 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="902840b2be1cc063542b4454ca864ee0f9e24c7bf98853504991fe2a9e2cc544" Dec 13 09:30:37 crc kubenswrapper[4971]: I1213 09:30:37.424443 4971 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-6hzwh" Dec 13 09:30:37 crc kubenswrapper[4971]: I1213 09:30:37.464652 4971 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-ns929\" (UniqueName: \"kubernetes.io/projected/f18bc810-f35c-4dc5-a2f9-7ea5f1594371-kube-api-access-ns929\") pod \"f18bc810-f35c-4dc5-a2f9-7ea5f1594371\" (UID: \"f18bc810-f35c-4dc5-a2f9-7ea5f1594371\") " Dec 13 09:30:37 crc kubenswrapper[4971]: I1213 09:30:37.472959 4971 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/f18bc810-f35c-4dc5-a2f9-7ea5f1594371-kube-api-access-ns929" (OuterVolumeSpecName: "kube-api-access-ns929") pod "f18bc810-f35c-4dc5-a2f9-7ea5f1594371" (UID: "f18bc810-f35c-4dc5-a2f9-7ea5f1594371"). InnerVolumeSpecName "kube-api-access-ns929". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 13 09:30:37 crc kubenswrapper[4971]: I1213 09:30:37.566380 4971 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/f18bc810-f35c-4dc5-a2f9-7ea5f1594371-utilities\") pod \"f18bc810-f35c-4dc5-a2f9-7ea5f1594371\" (UID: \"f18bc810-f35c-4dc5-a2f9-7ea5f1594371\") " Dec 13 09:30:37 crc kubenswrapper[4971]: I1213 09:30:37.566496 4971 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/f18bc810-f35c-4dc5-a2f9-7ea5f1594371-catalog-content\") pod \"f18bc810-f35c-4dc5-a2f9-7ea5f1594371\" (UID: \"f18bc810-f35c-4dc5-a2f9-7ea5f1594371\") " Dec 13 09:30:37 crc kubenswrapper[4971]: I1213 09:30:37.566818 4971 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-ns929\" (UniqueName: \"kubernetes.io/projected/f18bc810-f35c-4dc5-a2f9-7ea5f1594371-kube-api-access-ns929\") on node \"crc\" DevicePath \"\"" Dec 13 09:30:37 crc kubenswrapper[4971]: I1213 09:30:37.567780 4971 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/f18bc810-f35c-4dc5-a2f9-7ea5f1594371-utilities" (OuterVolumeSpecName: "utilities") pod "f18bc810-f35c-4dc5-a2f9-7ea5f1594371" (UID: "f18bc810-f35c-4dc5-a2f9-7ea5f1594371"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 13 09:30:37 crc kubenswrapper[4971]: I1213 09:30:37.641596 4971 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/f18bc810-f35c-4dc5-a2f9-7ea5f1594371-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "f18bc810-f35c-4dc5-a2f9-7ea5f1594371" (UID: "f18bc810-f35c-4dc5-a2f9-7ea5f1594371"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 13 09:30:37 crc kubenswrapper[4971]: I1213 09:30:37.671736 4971 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/f18bc810-f35c-4dc5-a2f9-7ea5f1594371-utilities\") on node \"crc\" DevicePath \"\"" Dec 13 09:30:37 crc kubenswrapper[4971]: I1213 09:30:37.671817 4971 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/f18bc810-f35c-4dc5-a2f9-7ea5f1594371-catalog-content\") on node \"crc\" DevicePath \"\"" Dec 13 09:30:38 crc kubenswrapper[4971]: I1213 09:30:38.405668 4971 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-6hzwh" Dec 13 09:30:38 crc kubenswrapper[4971]: I1213 09:30:38.433101 4971 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-6hzwh"] Dec 13 09:30:38 crc kubenswrapper[4971]: I1213 09:30:38.443339 4971 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/community-operators-6hzwh"] Dec 13 09:30:39 crc kubenswrapper[4971]: I1213 09:30:39.781104 4971 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="f18bc810-f35c-4dc5-a2f9-7ea5f1594371" path="/var/lib/kubelet/pods/f18bc810-f35c-4dc5-a2f9-7ea5f1594371/volumes" Dec 13 09:30:41 crc kubenswrapper[4971]: I1213 09:30:41.712773 4971 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-operators-zpdtr" Dec 13 09:30:41 crc kubenswrapper[4971]: I1213 09:30:41.782878 4971 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-operators-zpdtr" Dec 13 09:30:41 crc kubenswrapper[4971]: I1213 09:30:41.954271 4971 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-zpdtr"] Dec 13 09:30:43 crc kubenswrapper[4971]: I1213 09:30:43.452491 4971 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-operators-zpdtr" podUID="134aaef2-706e-41b6-a16f-05aba393a4b7" containerName="registry-server" containerID="cri-o://cc246f33190b446205a7fc29a9b3ef9ce057cb996b447fe0a3d9ddba24e052a4" gracePeriod=2 Dec 13 09:30:43 crc kubenswrapper[4971]: I1213 09:30:43.949198 4971 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-zpdtr" Dec 13 09:30:44 crc kubenswrapper[4971]: I1213 09:30:44.132264 4971 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-4jjtp\" (UniqueName: \"kubernetes.io/projected/134aaef2-706e-41b6-a16f-05aba393a4b7-kube-api-access-4jjtp\") pod \"134aaef2-706e-41b6-a16f-05aba393a4b7\" (UID: \"134aaef2-706e-41b6-a16f-05aba393a4b7\") " Dec 13 09:30:44 crc kubenswrapper[4971]: I1213 09:30:44.132599 4971 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/134aaef2-706e-41b6-a16f-05aba393a4b7-catalog-content\") pod \"134aaef2-706e-41b6-a16f-05aba393a4b7\" (UID: \"134aaef2-706e-41b6-a16f-05aba393a4b7\") " Dec 13 09:30:44 crc kubenswrapper[4971]: I1213 09:30:44.132778 4971 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/134aaef2-706e-41b6-a16f-05aba393a4b7-utilities\") pod \"134aaef2-706e-41b6-a16f-05aba393a4b7\" (UID: \"134aaef2-706e-41b6-a16f-05aba393a4b7\") " Dec 13 09:30:44 crc kubenswrapper[4971]: I1213 09:30:44.134121 4971 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/134aaef2-706e-41b6-a16f-05aba393a4b7-utilities" (OuterVolumeSpecName: "utilities") pod "134aaef2-706e-41b6-a16f-05aba393a4b7" (UID: "134aaef2-706e-41b6-a16f-05aba393a4b7"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 13 09:30:44 crc kubenswrapper[4971]: I1213 09:30:44.140715 4971 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/134aaef2-706e-41b6-a16f-05aba393a4b7-kube-api-access-4jjtp" (OuterVolumeSpecName: "kube-api-access-4jjtp") pod "134aaef2-706e-41b6-a16f-05aba393a4b7" (UID: "134aaef2-706e-41b6-a16f-05aba393a4b7"). InnerVolumeSpecName "kube-api-access-4jjtp". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 13 09:30:44 crc kubenswrapper[4971]: I1213 09:30:44.236075 4971 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/134aaef2-706e-41b6-a16f-05aba393a4b7-utilities\") on node \"crc\" DevicePath \"\"" Dec 13 09:30:44 crc kubenswrapper[4971]: I1213 09:30:44.237628 4971 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-4jjtp\" (UniqueName: \"kubernetes.io/projected/134aaef2-706e-41b6-a16f-05aba393a4b7-kube-api-access-4jjtp\") on node \"crc\" DevicePath \"\"" Dec 13 09:30:44 crc kubenswrapper[4971]: I1213 09:30:44.268691 4971 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/134aaef2-706e-41b6-a16f-05aba393a4b7-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "134aaef2-706e-41b6-a16f-05aba393a4b7" (UID: "134aaef2-706e-41b6-a16f-05aba393a4b7"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 13 09:30:44 crc kubenswrapper[4971]: I1213 09:30:44.340392 4971 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/134aaef2-706e-41b6-a16f-05aba393a4b7-catalog-content\") on node \"crc\" DevicePath \"\"" Dec 13 09:30:44 crc kubenswrapper[4971]: I1213 09:30:44.466640 4971 generic.go:334] "Generic (PLEG): container finished" podID="134aaef2-706e-41b6-a16f-05aba393a4b7" containerID="cc246f33190b446205a7fc29a9b3ef9ce057cb996b447fe0a3d9ddba24e052a4" exitCode=0 Dec 13 09:30:44 crc kubenswrapper[4971]: I1213 09:30:44.466716 4971 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-zpdtr" Dec 13 09:30:44 crc kubenswrapper[4971]: I1213 09:30:44.468172 4971 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-zpdtr" event={"ID":"134aaef2-706e-41b6-a16f-05aba393a4b7","Type":"ContainerDied","Data":"cc246f33190b446205a7fc29a9b3ef9ce057cb996b447fe0a3d9ddba24e052a4"} Dec 13 09:30:44 crc kubenswrapper[4971]: I1213 09:30:44.468365 4971 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-zpdtr" event={"ID":"134aaef2-706e-41b6-a16f-05aba393a4b7","Type":"ContainerDied","Data":"b5c9ed6af3370d402703c06686bbc6f7b31066bd94556c6f1da6374751a41a53"} Dec 13 09:30:44 crc kubenswrapper[4971]: I1213 09:30:44.468414 4971 scope.go:117] "RemoveContainer" containerID="cc246f33190b446205a7fc29a9b3ef9ce057cb996b447fe0a3d9ddba24e052a4" Dec 13 09:30:44 crc kubenswrapper[4971]: I1213 09:30:44.718946 4971 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-zpdtr"] Dec 13 09:30:44 crc kubenswrapper[4971]: I1213 09:30:44.728553 4971 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-operators-zpdtr"] Dec 13 09:30:44 crc kubenswrapper[4971]: I1213 09:30:44.735271 4971 scope.go:117] "RemoveContainer" containerID="67d045063d92fb28a571cce33040fddfa029bd5a8e17ee775f7a57e0a5285560" Dec 13 09:30:44 crc kubenswrapper[4971]: I1213 09:30:44.759125 4971 scope.go:117] "RemoveContainer" containerID="f4d0df53e38ccde7e16d16696d036abbd48b77ebeabff6bba701b33c47b22a2f" Dec 13 09:30:44 crc kubenswrapper[4971]: I1213 09:30:44.810906 4971 scope.go:117] "RemoveContainer" containerID="cc246f33190b446205a7fc29a9b3ef9ce057cb996b447fe0a3d9ddba24e052a4" Dec 13 09:30:44 crc kubenswrapper[4971]: E1213 09:30:44.811614 4971 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"cc246f33190b446205a7fc29a9b3ef9ce057cb996b447fe0a3d9ddba24e052a4\": container with ID starting with cc246f33190b446205a7fc29a9b3ef9ce057cb996b447fe0a3d9ddba24e052a4 not found: ID does not exist" containerID="cc246f33190b446205a7fc29a9b3ef9ce057cb996b447fe0a3d9ddba24e052a4" Dec 13 09:30:44 crc kubenswrapper[4971]: I1213 09:30:44.811694 4971 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"cc246f33190b446205a7fc29a9b3ef9ce057cb996b447fe0a3d9ddba24e052a4"} err="failed to get container status \"cc246f33190b446205a7fc29a9b3ef9ce057cb996b447fe0a3d9ddba24e052a4\": rpc error: code = NotFound desc = could not find container \"cc246f33190b446205a7fc29a9b3ef9ce057cb996b447fe0a3d9ddba24e052a4\": container with ID starting with cc246f33190b446205a7fc29a9b3ef9ce057cb996b447fe0a3d9ddba24e052a4 not found: ID does not exist" Dec 13 09:30:44 crc kubenswrapper[4971]: I1213 09:30:44.811747 4971 scope.go:117] "RemoveContainer" containerID="67d045063d92fb28a571cce33040fddfa029bd5a8e17ee775f7a57e0a5285560" Dec 13 09:30:44 crc kubenswrapper[4971]: E1213 09:30:44.812196 4971 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"67d045063d92fb28a571cce33040fddfa029bd5a8e17ee775f7a57e0a5285560\": container with ID starting with 67d045063d92fb28a571cce33040fddfa029bd5a8e17ee775f7a57e0a5285560 not found: ID does not exist" containerID="67d045063d92fb28a571cce33040fddfa029bd5a8e17ee775f7a57e0a5285560" Dec 13 09:30:44 crc kubenswrapper[4971]: I1213 09:30:44.812245 4971 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"67d045063d92fb28a571cce33040fddfa029bd5a8e17ee775f7a57e0a5285560"} err="failed to get container status \"67d045063d92fb28a571cce33040fddfa029bd5a8e17ee775f7a57e0a5285560\": rpc error: code = NotFound desc = could not find container \"67d045063d92fb28a571cce33040fddfa029bd5a8e17ee775f7a57e0a5285560\": container with ID starting with 67d045063d92fb28a571cce33040fddfa029bd5a8e17ee775f7a57e0a5285560 not found: ID does not exist" Dec 13 09:30:44 crc kubenswrapper[4971]: I1213 09:30:44.812293 4971 scope.go:117] "RemoveContainer" containerID="f4d0df53e38ccde7e16d16696d036abbd48b77ebeabff6bba701b33c47b22a2f" Dec 13 09:30:44 crc kubenswrapper[4971]: E1213 09:30:44.812679 4971 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"f4d0df53e38ccde7e16d16696d036abbd48b77ebeabff6bba701b33c47b22a2f\": container with ID starting with f4d0df53e38ccde7e16d16696d036abbd48b77ebeabff6bba701b33c47b22a2f not found: ID does not exist" containerID="f4d0df53e38ccde7e16d16696d036abbd48b77ebeabff6bba701b33c47b22a2f" Dec 13 09:30:44 crc kubenswrapper[4971]: I1213 09:30:44.812727 4971 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"f4d0df53e38ccde7e16d16696d036abbd48b77ebeabff6bba701b33c47b22a2f"} err="failed to get container status \"f4d0df53e38ccde7e16d16696d036abbd48b77ebeabff6bba701b33c47b22a2f\": rpc error: code = NotFound desc = could not find container \"f4d0df53e38ccde7e16d16696d036abbd48b77ebeabff6bba701b33c47b22a2f\": container with ID starting with f4d0df53e38ccde7e16d16696d036abbd48b77ebeabff6bba701b33c47b22a2f not found: ID does not exist" Dec 13 09:30:45 crc kubenswrapper[4971]: I1213 09:30:45.785196 4971 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="134aaef2-706e-41b6-a16f-05aba393a4b7" path="/var/lib/kubelet/pods/134aaef2-706e-41b6-a16f-05aba393a4b7/volumes" Dec 13 09:30:46 crc kubenswrapper[4971]: I1213 09:30:46.153798 4971 patch_prober.go:28] interesting pod/machine-config-daemon-82xjz container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 13 09:30:46 crc kubenswrapper[4971]: I1213 09:30:46.153856 4971 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-82xjz" podUID="e0c4af12-edda-41d7-9b44-f87396174ef2" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 13 09:30:57 crc kubenswrapper[4971]: I1213 09:30:57.612722 4971 patch_prober.go:28] interesting pod/olm-operator-6b444d44fb-pxjqw container/olm-operator namespace/openshift-operator-lifecycle-manager: Readiness probe status=failure output="Get \"https://10.217.0.30:8443/healthz\": context deadline exceeded" start-of-body= Dec 13 09:30:57 crc kubenswrapper[4971]: I1213 09:30:57.613379 4971 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-pxjqw" podUID="ea24f72e-df9d-418f-aaef-dfc0573eface" containerName="olm-operator" probeResult="failure" output="Get \"https://10.217.0.30:8443/healthz\": context deadline exceeded" Dec 13 09:30:57 crc kubenswrapper[4971]: I1213 09:30:57.616664 4971 patch_prober.go:28] interesting pod/authentication-operator-69f744f599-wjdd5 container/authentication-operator namespace/openshift-authentication-operator: Liveness probe status=failure output="Get \"https://10.217.0.42:8443/healthz\": context deadline exceeded (Client.Timeout exceeded while awaiting headers)" start-of-body= Dec 13 09:30:57 crc kubenswrapper[4971]: I1213 09:30:57.617067 4971 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-authentication-operator/authentication-operator-69f744f599-wjdd5" podUID="c0ca0ce3-929c-48d6-a3f7-6d334e3c3709" containerName="authentication-operator" probeResult="failure" output="Get \"https://10.217.0.42:8443/healthz\": context deadline exceeded (Client.Timeout exceeded while awaiting headers)" Dec 13 09:30:57 crc kubenswrapper[4971]: I1213 09:30:57.616996 4971 patch_prober.go:28] interesting pod/olm-operator-6b444d44fb-pxjqw container/olm-operator namespace/openshift-operator-lifecycle-manager: Liveness probe status=failure output="Get \"https://10.217.0.30:8443/healthz\": context deadline exceeded (Client.Timeout exceeded while awaiting headers)" start-of-body= Dec 13 09:30:57 crc kubenswrapper[4971]: I1213 09:30:57.617172 4971 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-pxjqw" podUID="ea24f72e-df9d-418f-aaef-dfc0573eface" containerName="olm-operator" probeResult="failure" output="Get \"https://10.217.0.30:8443/healthz\": context deadline exceeded (Client.Timeout exceeded while awaiting headers)" Dec 13 09:31:16 crc kubenswrapper[4971]: I1213 09:31:16.153800 4971 patch_prober.go:28] interesting pod/machine-config-daemon-82xjz container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 13 09:31:16 crc kubenswrapper[4971]: I1213 09:31:16.154597 4971 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-82xjz" podUID="e0c4af12-edda-41d7-9b44-f87396174ef2" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 13 09:31:16 crc kubenswrapper[4971]: I1213 09:31:16.154662 4971 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-82xjz" Dec 13 09:31:16 crc kubenswrapper[4971]: I1213 09:31:16.155750 4971 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"541c0610036fe5d23e59f44d9df0207efecabde0efb5dc61dac954401fb98fc1"} pod="openshift-machine-config-operator/machine-config-daemon-82xjz" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Dec 13 09:31:16 crc kubenswrapper[4971]: I1213 09:31:16.155825 4971 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-82xjz" podUID="e0c4af12-edda-41d7-9b44-f87396174ef2" containerName="machine-config-daemon" containerID="cri-o://541c0610036fe5d23e59f44d9df0207efecabde0efb5dc61dac954401fb98fc1" gracePeriod=600 Dec 13 09:31:17 crc kubenswrapper[4971]: I1213 09:31:17.010079 4971 generic.go:334] "Generic (PLEG): container finished" podID="e0c4af12-edda-41d7-9b44-f87396174ef2" containerID="541c0610036fe5d23e59f44d9df0207efecabde0efb5dc61dac954401fb98fc1" exitCode=0 Dec 13 09:31:17 crc kubenswrapper[4971]: I1213 09:31:17.011017 4971 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-82xjz" event={"ID":"e0c4af12-edda-41d7-9b44-f87396174ef2","Type":"ContainerDied","Data":"541c0610036fe5d23e59f44d9df0207efecabde0efb5dc61dac954401fb98fc1"} Dec 13 09:31:17 crc kubenswrapper[4971]: I1213 09:31:17.011188 4971 scope.go:117] "RemoveContainer" containerID="e616f2406e276879960fe4d0372c2d2a34f8c21a80e5202a5083f7852225bab9" Dec 13 09:31:18 crc kubenswrapper[4971]: I1213 09:31:18.028128 4971 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-82xjz" event={"ID":"e0c4af12-edda-41d7-9b44-f87396174ef2","Type":"ContainerStarted","Data":"7c80479d6aedc6e1a5cced1df9e55738d4a726b6ddd5adfe17363d9c5d097d6a"} Dec 13 09:31:45 crc kubenswrapper[4971]: I1213 09:31:45.559283 4971 generic.go:334] "Generic (PLEG): container finished" podID="b0fcccc8-3a38-42b6-a10c-57a0a44b48c3" containerID="39128e7c82af935994f5b9c56e004fa0f4ce9d8e22275a1983048044d97c2684" exitCode=0 Dec 13 09:31:45 crc kubenswrapper[4971]: I1213 09:31:45.559393 4971 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/tempest-tests-tempest-s01-single-test" event={"ID":"b0fcccc8-3a38-42b6-a10c-57a0a44b48c3","Type":"ContainerDied","Data":"39128e7c82af935994f5b9c56e004fa0f4ce9d8e22275a1983048044d97c2684"} Dec 13 09:31:47 crc kubenswrapper[4971]: I1213 09:31:47.048404 4971 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/tempest-tests-tempest-s01-single-test" Dec 13 09:31:47 crc kubenswrapper[4971]: I1213 09:31:47.142216 4971 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"test-operator-ephemeral-temporary\" (UniqueName: \"kubernetes.io/empty-dir/b0fcccc8-3a38-42b6-a10c-57a0a44b48c3-test-operator-ephemeral-temporary\") pod \"b0fcccc8-3a38-42b6-a10c-57a0a44b48c3\" (UID: \"b0fcccc8-3a38-42b6-a10c-57a0a44b48c3\") " Dec 13 09:31:47 crc kubenswrapper[4971]: I1213 09:31:47.142344 4971 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"test-operator-logs\" (UniqueName: \"kubernetes.io/local-volume/local-storage12-crc\") pod \"b0fcccc8-3a38-42b6-a10c-57a0a44b48c3\" (UID: \"b0fcccc8-3a38-42b6-a10c-57a0a44b48c3\") " Dec 13 09:31:47 crc kubenswrapper[4971]: I1213 09:31:47.142388 4971 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"test-operator-ephemeral-workdir\" (UniqueName: \"kubernetes.io/empty-dir/b0fcccc8-3a38-42b6-a10c-57a0a44b48c3-test-operator-ephemeral-workdir\") pod \"b0fcccc8-3a38-42b6-a10c-57a0a44b48c3\" (UID: \"b0fcccc8-3a38-42b6-a10c-57a0a44b48c3\") " Dec 13 09:31:47 crc kubenswrapper[4971]: I1213 09:31:47.142455 4971 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/b0fcccc8-3a38-42b6-a10c-57a0a44b48c3-ssh-key\") pod \"b0fcccc8-3a38-42b6-a10c-57a0a44b48c3\" (UID: \"b0fcccc8-3a38-42b6-a10c-57a0a44b48c3\") " Dec 13 09:31:47 crc kubenswrapper[4971]: I1213 09:31:47.142478 4971 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/b0fcccc8-3a38-42b6-a10c-57a0a44b48c3-config-data\") pod \"b0fcccc8-3a38-42b6-a10c-57a0a44b48c3\" (UID: \"b0fcccc8-3a38-42b6-a10c-57a0a44b48c3\") " Dec 13 09:31:47 crc kubenswrapper[4971]: I1213 09:31:47.142540 4971 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"openstack-config\" (UniqueName: \"kubernetes.io/configmap/b0fcccc8-3a38-42b6-a10c-57a0a44b48c3-openstack-config\") pod \"b0fcccc8-3a38-42b6-a10c-57a0a44b48c3\" (UID: \"b0fcccc8-3a38-42b6-a10c-57a0a44b48c3\") " Dec 13 09:31:47 crc kubenswrapper[4971]: I1213 09:31:47.142562 4971 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ca-certs\" (UniqueName: \"kubernetes.io/secret/b0fcccc8-3a38-42b6-a10c-57a0a44b48c3-ca-certs\") pod \"b0fcccc8-3a38-42b6-a10c-57a0a44b48c3\" (UID: \"b0fcccc8-3a38-42b6-a10c-57a0a44b48c3\") " Dec 13 09:31:47 crc kubenswrapper[4971]: I1213 09:31:47.142592 4971 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/b0fcccc8-3a38-42b6-a10c-57a0a44b48c3-ceph\") pod \"b0fcccc8-3a38-42b6-a10c-57a0a44b48c3\" (UID: \"b0fcccc8-3a38-42b6-a10c-57a0a44b48c3\") " Dec 13 09:31:47 crc kubenswrapper[4971]: I1213 09:31:47.142644 4971 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"openstack-config-secret\" (UniqueName: \"kubernetes.io/secret/b0fcccc8-3a38-42b6-a10c-57a0a44b48c3-openstack-config-secret\") pod \"b0fcccc8-3a38-42b6-a10c-57a0a44b48c3\" (UID: \"b0fcccc8-3a38-42b6-a10c-57a0a44b48c3\") " Dec 13 09:31:47 crc kubenswrapper[4971]: I1213 09:31:47.142806 4971 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-pwptn\" (UniqueName: \"kubernetes.io/projected/b0fcccc8-3a38-42b6-a10c-57a0a44b48c3-kube-api-access-pwptn\") pod \"b0fcccc8-3a38-42b6-a10c-57a0a44b48c3\" (UID: \"b0fcccc8-3a38-42b6-a10c-57a0a44b48c3\") " Dec 13 09:31:47 crc kubenswrapper[4971]: I1213 09:31:47.143056 4971 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/b0fcccc8-3a38-42b6-a10c-57a0a44b48c3-test-operator-ephemeral-temporary" (OuterVolumeSpecName: "test-operator-ephemeral-temporary") pod "b0fcccc8-3a38-42b6-a10c-57a0a44b48c3" (UID: "b0fcccc8-3a38-42b6-a10c-57a0a44b48c3"). InnerVolumeSpecName "test-operator-ephemeral-temporary". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 13 09:31:47 crc kubenswrapper[4971]: I1213 09:31:47.143887 4971 reconciler_common.go:293] "Volume detached for volume \"test-operator-ephemeral-temporary\" (UniqueName: \"kubernetes.io/empty-dir/b0fcccc8-3a38-42b6-a10c-57a0a44b48c3-test-operator-ephemeral-temporary\") on node \"crc\" DevicePath \"\"" Dec 13 09:31:47 crc kubenswrapper[4971]: I1213 09:31:47.144661 4971 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/b0fcccc8-3a38-42b6-a10c-57a0a44b48c3-config-data" (OuterVolumeSpecName: "config-data") pod "b0fcccc8-3a38-42b6-a10c-57a0a44b48c3" (UID: "b0fcccc8-3a38-42b6-a10c-57a0a44b48c3"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 13 09:31:47 crc kubenswrapper[4971]: I1213 09:31:47.149902 4971 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/b0fcccc8-3a38-42b6-a10c-57a0a44b48c3-test-operator-ephemeral-workdir" (OuterVolumeSpecName: "test-operator-ephemeral-workdir") pod "b0fcccc8-3a38-42b6-a10c-57a0a44b48c3" (UID: "b0fcccc8-3a38-42b6-a10c-57a0a44b48c3"). InnerVolumeSpecName "test-operator-ephemeral-workdir". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 13 09:31:47 crc kubenswrapper[4971]: I1213 09:31:47.246939 4971 reconciler_common.go:293] "Volume detached for volume \"test-operator-ephemeral-workdir\" (UniqueName: \"kubernetes.io/empty-dir/b0fcccc8-3a38-42b6-a10c-57a0a44b48c3-test-operator-ephemeral-workdir\") on node \"crc\" DevicePath \"\"" Dec 13 09:31:47 crc kubenswrapper[4971]: I1213 09:31:47.247256 4971 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/b0fcccc8-3a38-42b6-a10c-57a0a44b48c3-config-data\") on node \"crc\" DevicePath \"\"" Dec 13 09:31:47 crc kubenswrapper[4971]: I1213 09:31:47.582098 4971 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/tempest-tests-tempest-s01-single-test" event={"ID":"b0fcccc8-3a38-42b6-a10c-57a0a44b48c3","Type":"ContainerDied","Data":"8fd093e74db72443fef79670a241b3b07bbb1e2b8867709469b691a472b3d640"} Dec 13 09:31:47 crc kubenswrapper[4971]: I1213 09:31:47.582158 4971 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="8fd093e74db72443fef79670a241b3b07bbb1e2b8867709469b691a472b3d640" Dec 13 09:31:47 crc kubenswrapper[4971]: I1213 09:31:47.582248 4971 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/tempest-tests-tempest-s01-single-test" Dec 13 09:31:47 crc kubenswrapper[4971]: I1213 09:31:47.837513 4971 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/b0fcccc8-3a38-42b6-a10c-57a0a44b48c3-kube-api-access-pwptn" (OuterVolumeSpecName: "kube-api-access-pwptn") pod "b0fcccc8-3a38-42b6-a10c-57a0a44b48c3" (UID: "b0fcccc8-3a38-42b6-a10c-57a0a44b48c3"). InnerVolumeSpecName "kube-api-access-pwptn". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 13 09:31:47 crc kubenswrapper[4971]: I1213 09:31:47.837588 4971 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/b0fcccc8-3a38-42b6-a10c-57a0a44b48c3-ceph" (OuterVolumeSpecName: "ceph") pod "b0fcccc8-3a38-42b6-a10c-57a0a44b48c3" (UID: "b0fcccc8-3a38-42b6-a10c-57a0a44b48c3"). InnerVolumeSpecName "ceph". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 13 09:31:47 crc kubenswrapper[4971]: I1213 09:31:47.843364 4971 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/local-volume/local-storage12-crc" (OuterVolumeSpecName: "test-operator-logs") pod "b0fcccc8-3a38-42b6-a10c-57a0a44b48c3" (UID: "b0fcccc8-3a38-42b6-a10c-57a0a44b48c3"). InnerVolumeSpecName "local-storage12-crc". PluginName "kubernetes.io/local-volume", VolumeGidValue "" Dec 13 09:31:47 crc kubenswrapper[4971]: I1213 09:31:47.867409 4971 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-pwptn\" (UniqueName: \"kubernetes.io/projected/b0fcccc8-3a38-42b6-a10c-57a0a44b48c3-kube-api-access-pwptn\") on node \"crc\" DevicePath \"\"" Dec 13 09:31:47 crc kubenswrapper[4971]: I1213 09:31:47.867504 4971 reconciler_common.go:286] "operationExecutor.UnmountDevice started for volume \"local-storage12-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage12-crc\") on node \"crc\" " Dec 13 09:31:47 crc kubenswrapper[4971]: I1213 09:31:47.867571 4971 reconciler_common.go:293] "Volume detached for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/b0fcccc8-3a38-42b6-a10c-57a0a44b48c3-ceph\") on node \"crc\" DevicePath \"\"" Dec 13 09:31:47 crc kubenswrapper[4971]: I1213 09:31:47.877272 4971 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/b0fcccc8-3a38-42b6-a10c-57a0a44b48c3-ssh-key" (OuterVolumeSpecName: "ssh-key") pod "b0fcccc8-3a38-42b6-a10c-57a0a44b48c3" (UID: "b0fcccc8-3a38-42b6-a10c-57a0a44b48c3"). InnerVolumeSpecName "ssh-key". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 13 09:31:47 crc kubenswrapper[4971]: I1213 09:31:47.894740 4971 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/b0fcccc8-3a38-42b6-a10c-57a0a44b48c3-openstack-config-secret" (OuterVolumeSpecName: "openstack-config-secret") pod "b0fcccc8-3a38-42b6-a10c-57a0a44b48c3" (UID: "b0fcccc8-3a38-42b6-a10c-57a0a44b48c3"). InnerVolumeSpecName "openstack-config-secret". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 13 09:31:47 crc kubenswrapper[4971]: I1213 09:31:47.909166 4971 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/b0fcccc8-3a38-42b6-a10c-57a0a44b48c3-ca-certs" (OuterVolumeSpecName: "ca-certs") pod "b0fcccc8-3a38-42b6-a10c-57a0a44b48c3" (UID: "b0fcccc8-3a38-42b6-a10c-57a0a44b48c3"). InnerVolumeSpecName "ca-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 13 09:31:47 crc kubenswrapper[4971]: I1213 09:31:47.915377 4971 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/b0fcccc8-3a38-42b6-a10c-57a0a44b48c3-openstack-config" (OuterVolumeSpecName: "openstack-config") pod "b0fcccc8-3a38-42b6-a10c-57a0a44b48c3" (UID: "b0fcccc8-3a38-42b6-a10c-57a0a44b48c3"). InnerVolumeSpecName "openstack-config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 13 09:31:47 crc kubenswrapper[4971]: I1213 09:31:47.921685 4971 operation_generator.go:917] UnmountDevice succeeded for volume "local-storage12-crc" (UniqueName: "kubernetes.io/local-volume/local-storage12-crc") on node "crc" Dec 13 09:31:47 crc kubenswrapper[4971]: I1213 09:31:47.970174 4971 reconciler_common.go:293] "Volume detached for volume \"local-storage12-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage12-crc\") on node \"crc\" DevicePath \"\"" Dec 13 09:31:47 crc kubenswrapper[4971]: I1213 09:31:47.970213 4971 reconciler_common.go:293] "Volume detached for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/b0fcccc8-3a38-42b6-a10c-57a0a44b48c3-ssh-key\") on node \"crc\" DevicePath \"\"" Dec 13 09:31:47 crc kubenswrapper[4971]: I1213 09:31:47.970228 4971 reconciler_common.go:293] "Volume detached for volume \"openstack-config\" (UniqueName: \"kubernetes.io/configmap/b0fcccc8-3a38-42b6-a10c-57a0a44b48c3-openstack-config\") on node \"crc\" DevicePath \"\"" Dec 13 09:31:47 crc kubenswrapper[4971]: I1213 09:31:47.970243 4971 reconciler_common.go:293] "Volume detached for volume \"ca-certs\" (UniqueName: \"kubernetes.io/secret/b0fcccc8-3a38-42b6-a10c-57a0a44b48c3-ca-certs\") on node \"crc\" DevicePath \"\"" Dec 13 09:31:47 crc kubenswrapper[4971]: I1213 09:31:47.970253 4971 reconciler_common.go:293] "Volume detached for volume \"openstack-config-secret\" (UniqueName: \"kubernetes.io/secret/b0fcccc8-3a38-42b6-a10c-57a0a44b48c3-openstack-config-secret\") on node \"crc\" DevicePath \"\"" Dec 13 09:31:56 crc kubenswrapper[4971]: I1213 09:31:56.863275 4971 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/test-operator-logs-pod-tempest-tempest-tests-tempest"] Dec 13 09:31:56 crc kubenswrapper[4971]: E1213 09:31:56.864585 4971 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="b0fcccc8-3a38-42b6-a10c-57a0a44b48c3" containerName="tempest-tests-tempest-tests-runner" Dec 13 09:31:56 crc kubenswrapper[4971]: I1213 09:31:56.864601 4971 state_mem.go:107] "Deleted CPUSet assignment" podUID="b0fcccc8-3a38-42b6-a10c-57a0a44b48c3" containerName="tempest-tests-tempest-tests-runner" Dec 13 09:31:56 crc kubenswrapper[4971]: E1213 09:31:56.864610 4971 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="134aaef2-706e-41b6-a16f-05aba393a4b7" containerName="extract-utilities" Dec 13 09:31:56 crc kubenswrapper[4971]: I1213 09:31:56.864616 4971 state_mem.go:107] "Deleted CPUSet assignment" podUID="134aaef2-706e-41b6-a16f-05aba393a4b7" containerName="extract-utilities" Dec 13 09:31:56 crc kubenswrapper[4971]: E1213 09:31:56.864626 4971 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f18bc810-f35c-4dc5-a2f9-7ea5f1594371" containerName="extract-utilities" Dec 13 09:31:56 crc kubenswrapper[4971]: I1213 09:31:56.864633 4971 state_mem.go:107] "Deleted CPUSet assignment" podUID="f18bc810-f35c-4dc5-a2f9-7ea5f1594371" containerName="extract-utilities" Dec 13 09:31:56 crc kubenswrapper[4971]: E1213 09:31:56.864645 4971 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="134aaef2-706e-41b6-a16f-05aba393a4b7" containerName="registry-server" Dec 13 09:31:56 crc kubenswrapper[4971]: I1213 09:31:56.864651 4971 state_mem.go:107] "Deleted CPUSet assignment" podUID="134aaef2-706e-41b6-a16f-05aba393a4b7" containerName="registry-server" Dec 13 09:31:56 crc kubenswrapper[4971]: E1213 09:31:56.864667 4971 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f18bc810-f35c-4dc5-a2f9-7ea5f1594371" containerName="registry-server" Dec 13 09:31:56 crc kubenswrapper[4971]: I1213 09:31:56.864673 4971 state_mem.go:107] "Deleted CPUSet assignment" podUID="f18bc810-f35c-4dc5-a2f9-7ea5f1594371" containerName="registry-server" Dec 13 09:31:56 crc kubenswrapper[4971]: E1213 09:31:56.864685 4971 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f18bc810-f35c-4dc5-a2f9-7ea5f1594371" containerName="extract-content" Dec 13 09:31:56 crc kubenswrapper[4971]: I1213 09:31:56.864690 4971 state_mem.go:107] "Deleted CPUSet assignment" podUID="f18bc810-f35c-4dc5-a2f9-7ea5f1594371" containerName="extract-content" Dec 13 09:31:56 crc kubenswrapper[4971]: E1213 09:31:56.864703 4971 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="134aaef2-706e-41b6-a16f-05aba393a4b7" containerName="extract-content" Dec 13 09:31:56 crc kubenswrapper[4971]: I1213 09:31:56.864710 4971 state_mem.go:107] "Deleted CPUSet assignment" podUID="134aaef2-706e-41b6-a16f-05aba393a4b7" containerName="extract-content" Dec 13 09:31:56 crc kubenswrapper[4971]: I1213 09:31:56.864922 4971 memory_manager.go:354] "RemoveStaleState removing state" podUID="b0fcccc8-3a38-42b6-a10c-57a0a44b48c3" containerName="tempest-tests-tempest-tests-runner" Dec 13 09:31:56 crc kubenswrapper[4971]: I1213 09:31:56.864946 4971 memory_manager.go:354] "RemoveStaleState removing state" podUID="134aaef2-706e-41b6-a16f-05aba393a4b7" containerName="registry-server" Dec 13 09:31:56 crc kubenswrapper[4971]: I1213 09:31:56.864959 4971 memory_manager.go:354] "RemoveStaleState removing state" podUID="f18bc810-f35c-4dc5-a2f9-7ea5f1594371" containerName="registry-server" Dec 13 09:31:56 crc kubenswrapper[4971]: I1213 09:31:56.865709 4971 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/test-operator-logs-pod-tempest-tempest-tests-tempest" Dec 13 09:31:56 crc kubenswrapper[4971]: I1213 09:31:56.868826 4971 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"default-dockercfg-jxsqw" Dec 13 09:31:56 crc kubenswrapper[4971]: I1213 09:31:56.886410 4971 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/test-operator-logs-pod-tempest-tempest-tests-tempest"] Dec 13 09:31:56 crc kubenswrapper[4971]: I1213 09:31:56.981312 4971 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"local-storage12-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage12-crc\") pod \"test-operator-logs-pod-tempest-tempest-tests-tempest\" (UID: \"b5af123d-e343-4e82-861d-3eaed6d9ca0c\") " pod="openstack/test-operator-logs-pod-tempest-tempest-tests-tempest" Dec 13 09:31:56 crc kubenswrapper[4971]: I1213 09:31:56.981898 4971 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-bcjnb\" (UniqueName: \"kubernetes.io/projected/b5af123d-e343-4e82-861d-3eaed6d9ca0c-kube-api-access-bcjnb\") pod \"test-operator-logs-pod-tempest-tempest-tests-tempest\" (UID: \"b5af123d-e343-4e82-861d-3eaed6d9ca0c\") " pod="openstack/test-operator-logs-pod-tempest-tempest-tests-tempest" Dec 13 09:31:57 crc kubenswrapper[4971]: I1213 09:31:57.084397 4971 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"local-storage12-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage12-crc\") pod \"test-operator-logs-pod-tempest-tempest-tests-tempest\" (UID: \"b5af123d-e343-4e82-861d-3eaed6d9ca0c\") " pod="openstack/test-operator-logs-pod-tempest-tempest-tests-tempest" Dec 13 09:31:57 crc kubenswrapper[4971]: I1213 09:31:57.084484 4971 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-bcjnb\" (UniqueName: \"kubernetes.io/projected/b5af123d-e343-4e82-861d-3eaed6d9ca0c-kube-api-access-bcjnb\") pod \"test-operator-logs-pod-tempest-tempest-tests-tempest\" (UID: \"b5af123d-e343-4e82-861d-3eaed6d9ca0c\") " pod="openstack/test-operator-logs-pod-tempest-tempest-tests-tempest" Dec 13 09:31:57 crc kubenswrapper[4971]: I1213 09:31:57.085099 4971 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"local-storage12-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage12-crc\") pod \"test-operator-logs-pod-tempest-tempest-tests-tempest\" (UID: \"b5af123d-e343-4e82-861d-3eaed6d9ca0c\") device mount path \"/mnt/openstack/pv12\"" pod="openstack/test-operator-logs-pod-tempest-tempest-tests-tempest" Dec 13 09:31:57 crc kubenswrapper[4971]: I1213 09:31:57.107791 4971 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-bcjnb\" (UniqueName: \"kubernetes.io/projected/b5af123d-e343-4e82-861d-3eaed6d9ca0c-kube-api-access-bcjnb\") pod \"test-operator-logs-pod-tempest-tempest-tests-tempest\" (UID: \"b5af123d-e343-4e82-861d-3eaed6d9ca0c\") " pod="openstack/test-operator-logs-pod-tempest-tempest-tests-tempest" Dec 13 09:31:57 crc kubenswrapper[4971]: I1213 09:31:57.115221 4971 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"local-storage12-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage12-crc\") pod \"test-operator-logs-pod-tempest-tempest-tests-tempest\" (UID: \"b5af123d-e343-4e82-861d-3eaed6d9ca0c\") " pod="openstack/test-operator-logs-pod-tempest-tempest-tests-tempest" Dec 13 09:31:57 crc kubenswrapper[4971]: I1213 09:31:57.192497 4971 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/test-operator-logs-pod-tempest-tempest-tests-tempest" Dec 13 09:31:57 crc kubenswrapper[4971]: I1213 09:31:57.708392 4971 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/test-operator-logs-pod-tempest-tempest-tests-tempest"] Dec 13 09:31:57 crc kubenswrapper[4971]: I1213 09:31:57.713730 4971 provider.go:102] Refreshing cache for provider: *credentialprovider.defaultDockerConfigProvider Dec 13 09:31:58 crc kubenswrapper[4971]: I1213 09:31:58.719894 4971 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/test-operator-logs-pod-tempest-tempest-tests-tempest" event={"ID":"b5af123d-e343-4e82-861d-3eaed6d9ca0c","Type":"ContainerStarted","Data":"23dd55003b5ce28792bd6a9328cf1d00f1bd94be0b1eb707a43e25d32943f2c9"} Dec 13 09:31:59 crc kubenswrapper[4971]: I1213 09:31:59.738630 4971 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/test-operator-logs-pod-tempest-tempest-tests-tempest" event={"ID":"b5af123d-e343-4e82-861d-3eaed6d9ca0c","Type":"ContainerStarted","Data":"d802c8e9d622b4cfd98bdd7d2ca29cad2e6bb024f5c42c7c623ffc6859ab6337"} Dec 13 09:31:59 crc kubenswrapper[4971]: I1213 09:31:59.763799 4971 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/test-operator-logs-pod-tempest-tempest-tests-tempest" podStartSLOduration=2.711796846 podStartE2EDuration="3.763774703s" podCreationTimestamp="2025-12-13 09:31:56 +0000 UTC" firstStartedPulling="2025-12-13 09:31:57.71232978 +0000 UTC m=+9774.316739218" lastFinishedPulling="2025-12-13 09:31:58.764307627 +0000 UTC m=+9775.368717075" observedRunningTime="2025-12-13 09:31:59.753974906 +0000 UTC m=+9776.358384344" watchObservedRunningTime="2025-12-13 09:31:59.763774703 +0000 UTC m=+9776.368184151" Dec 13 09:32:16 crc kubenswrapper[4971]: I1213 09:32:16.733110 4971 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/tobiko-tests-tobiko-s00-podified-functional"] Dec 13 09:32:16 crc kubenswrapper[4971]: I1213 09:32:16.736070 4971 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/tobiko-tests-tobiko-s00-podified-functional" Dec 13 09:32:16 crc kubenswrapper[4971]: I1213 09:32:16.738731 4971 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"tobiko-secret" Dec 13 09:32:16 crc kubenswrapper[4971]: I1213 09:32:16.739348 4971 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"tobiko-tests-tobikotobiko-public-key" Dec 13 09:32:16 crc kubenswrapper[4971]: I1213 09:32:16.739560 4971 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"tobiko-tests-tobikotobiko-private-key" Dec 13 09:32:16 crc kubenswrapper[4971]: I1213 09:32:16.739667 4971 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"tobiko-tests-tobikotobiko-config" Dec 13 09:32:16 crc kubenswrapper[4971]: I1213 09:32:16.740041 4971 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"test-operator-clouds-config" Dec 13 09:32:16 crc kubenswrapper[4971]: I1213 09:32:16.754242 4971 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/tobiko-tests-tobiko-s00-podified-functional"] Dec 13 09:32:16 crc kubenswrapper[4971]: I1213 09:32:16.875191 4971 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"local-storage03-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage03-crc\") pod \"tobiko-tests-tobiko-s00-podified-functional\" (UID: \"0a604d8e-de8c-4ae6-b13a-aafd2adee17a\") " pod="openstack/tobiko-tests-tobiko-s00-podified-functional" Dec 13 09:32:16 crc kubenswrapper[4971]: I1213 09:32:16.875257 4971 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"tobiko-config\" (UniqueName: \"kubernetes.io/configmap/0a604d8e-de8c-4ae6-b13a-aafd2adee17a-tobiko-config\") pod \"tobiko-tests-tobiko-s00-podified-functional\" (UID: \"0a604d8e-de8c-4ae6-b13a-aafd2adee17a\") " pod="openstack/tobiko-tests-tobiko-s00-podified-functional" Dec 13 09:32:16 crc kubenswrapper[4971]: I1213 09:32:16.875344 4971 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-twpcl\" (UniqueName: \"kubernetes.io/projected/0a604d8e-de8c-4ae6-b13a-aafd2adee17a-kube-api-access-twpcl\") pod \"tobiko-tests-tobiko-s00-podified-functional\" (UID: \"0a604d8e-de8c-4ae6-b13a-aafd2adee17a\") " pod="openstack/tobiko-tests-tobiko-s00-podified-functional" Dec 13 09:32:16 crc kubenswrapper[4971]: I1213 09:32:16.875599 4971 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kubeconfig\" (UniqueName: \"kubernetes.io/secret/0a604d8e-de8c-4ae6-b13a-aafd2adee17a-kubeconfig\") pod \"tobiko-tests-tobiko-s00-podified-functional\" (UID: \"0a604d8e-de8c-4ae6-b13a-aafd2adee17a\") " pod="openstack/tobiko-tests-tobiko-s00-podified-functional" Dec 13 09:32:16 crc kubenswrapper[4971]: I1213 09:32:16.875671 4971 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ca-certs\" (UniqueName: \"kubernetes.io/secret/0a604d8e-de8c-4ae6-b13a-aafd2adee17a-ca-certs\") pod \"tobiko-tests-tobiko-s00-podified-functional\" (UID: \"0a604d8e-de8c-4ae6-b13a-aafd2adee17a\") " pod="openstack/tobiko-tests-tobiko-s00-podified-functional" Dec 13 09:32:16 crc kubenswrapper[4971]: I1213 09:32:16.875700 4971 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"test-operator-clouds-config\" (UniqueName: \"kubernetes.io/configmap/0a604d8e-de8c-4ae6-b13a-aafd2adee17a-test-operator-clouds-config\") pod \"tobiko-tests-tobiko-s00-podified-functional\" (UID: \"0a604d8e-de8c-4ae6-b13a-aafd2adee17a\") " pod="openstack/tobiko-tests-tobiko-s00-podified-functional" Dec 13 09:32:16 crc kubenswrapper[4971]: I1213 09:32:16.875737 4971 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"test-operator-ephemeral-workdir\" (UniqueName: \"kubernetes.io/empty-dir/0a604d8e-de8c-4ae6-b13a-aafd2adee17a-test-operator-ephemeral-workdir\") pod \"tobiko-tests-tobiko-s00-podified-functional\" (UID: \"0a604d8e-de8c-4ae6-b13a-aafd2adee17a\") " pod="openstack/tobiko-tests-tobiko-s00-podified-functional" Dec 13 09:32:16 crc kubenswrapper[4971]: I1213 09:32:16.875764 4971 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"test-operator-ephemeral-temporary\" (UniqueName: \"kubernetes.io/empty-dir/0a604d8e-de8c-4ae6-b13a-aafd2adee17a-test-operator-ephemeral-temporary\") pod \"tobiko-tests-tobiko-s00-podified-functional\" (UID: \"0a604d8e-de8c-4ae6-b13a-aafd2adee17a\") " pod="openstack/tobiko-tests-tobiko-s00-podified-functional" Dec 13 09:32:16 crc kubenswrapper[4971]: I1213 09:32:16.875857 4971 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"openstack-config-secret\" (UniqueName: \"kubernetes.io/secret/0a604d8e-de8c-4ae6-b13a-aafd2adee17a-openstack-config-secret\") pod \"tobiko-tests-tobiko-s00-podified-functional\" (UID: \"0a604d8e-de8c-4ae6-b13a-aafd2adee17a\") " pod="openstack/tobiko-tests-tobiko-s00-podified-functional" Dec 13 09:32:16 crc kubenswrapper[4971]: I1213 09:32:16.875894 4971 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"tobiko-public-key\" (UniqueName: \"kubernetes.io/configmap/0a604d8e-de8c-4ae6-b13a-aafd2adee17a-tobiko-public-key\") pod \"tobiko-tests-tobiko-s00-podified-functional\" (UID: \"0a604d8e-de8c-4ae6-b13a-aafd2adee17a\") " pod="openstack/tobiko-tests-tobiko-s00-podified-functional" Dec 13 09:32:16 crc kubenswrapper[4971]: I1213 09:32:16.875958 4971 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"tobiko-private-key\" (UniqueName: \"kubernetes.io/configmap/0a604d8e-de8c-4ae6-b13a-aafd2adee17a-tobiko-private-key\") pod \"tobiko-tests-tobiko-s00-podified-functional\" (UID: \"0a604d8e-de8c-4ae6-b13a-aafd2adee17a\") " pod="openstack/tobiko-tests-tobiko-s00-podified-functional" Dec 13 09:32:16 crc kubenswrapper[4971]: I1213 09:32:16.875993 4971 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/0a604d8e-de8c-4ae6-b13a-aafd2adee17a-ceph\") pod \"tobiko-tests-tobiko-s00-podified-functional\" (UID: \"0a604d8e-de8c-4ae6-b13a-aafd2adee17a\") " pod="openstack/tobiko-tests-tobiko-s00-podified-functional" Dec 13 09:32:16 crc kubenswrapper[4971]: I1213 09:32:16.978649 4971 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"local-storage03-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage03-crc\") pod \"tobiko-tests-tobiko-s00-podified-functional\" (UID: \"0a604d8e-de8c-4ae6-b13a-aafd2adee17a\") " pod="openstack/tobiko-tests-tobiko-s00-podified-functional" Dec 13 09:32:16 crc kubenswrapper[4971]: I1213 09:32:16.978730 4971 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"tobiko-config\" (UniqueName: \"kubernetes.io/configmap/0a604d8e-de8c-4ae6-b13a-aafd2adee17a-tobiko-config\") pod \"tobiko-tests-tobiko-s00-podified-functional\" (UID: \"0a604d8e-de8c-4ae6-b13a-aafd2adee17a\") " pod="openstack/tobiko-tests-tobiko-s00-podified-functional" Dec 13 09:32:16 crc kubenswrapper[4971]: I1213 09:32:16.978804 4971 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-twpcl\" (UniqueName: \"kubernetes.io/projected/0a604d8e-de8c-4ae6-b13a-aafd2adee17a-kube-api-access-twpcl\") pod \"tobiko-tests-tobiko-s00-podified-functional\" (UID: \"0a604d8e-de8c-4ae6-b13a-aafd2adee17a\") " pod="openstack/tobiko-tests-tobiko-s00-podified-functional" Dec 13 09:32:16 crc kubenswrapper[4971]: I1213 09:32:16.978862 4971 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kubeconfig\" (UniqueName: \"kubernetes.io/secret/0a604d8e-de8c-4ae6-b13a-aafd2adee17a-kubeconfig\") pod \"tobiko-tests-tobiko-s00-podified-functional\" (UID: \"0a604d8e-de8c-4ae6-b13a-aafd2adee17a\") " pod="openstack/tobiko-tests-tobiko-s00-podified-functional" Dec 13 09:32:16 crc kubenswrapper[4971]: I1213 09:32:16.978918 4971 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ca-certs\" (UniqueName: \"kubernetes.io/secret/0a604d8e-de8c-4ae6-b13a-aafd2adee17a-ca-certs\") pod \"tobiko-tests-tobiko-s00-podified-functional\" (UID: \"0a604d8e-de8c-4ae6-b13a-aafd2adee17a\") " pod="openstack/tobiko-tests-tobiko-s00-podified-functional" Dec 13 09:32:16 crc kubenswrapper[4971]: I1213 09:32:16.978945 4971 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"test-operator-clouds-config\" (UniqueName: \"kubernetes.io/configmap/0a604d8e-de8c-4ae6-b13a-aafd2adee17a-test-operator-clouds-config\") pod \"tobiko-tests-tobiko-s00-podified-functional\" (UID: \"0a604d8e-de8c-4ae6-b13a-aafd2adee17a\") " pod="openstack/tobiko-tests-tobiko-s00-podified-functional" Dec 13 09:32:16 crc kubenswrapper[4971]: I1213 09:32:16.978987 4971 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"test-operator-ephemeral-workdir\" (UniqueName: \"kubernetes.io/empty-dir/0a604d8e-de8c-4ae6-b13a-aafd2adee17a-test-operator-ephemeral-workdir\") pod \"tobiko-tests-tobiko-s00-podified-functional\" (UID: \"0a604d8e-de8c-4ae6-b13a-aafd2adee17a\") " pod="openstack/tobiko-tests-tobiko-s00-podified-functional" Dec 13 09:32:16 crc kubenswrapper[4971]: I1213 09:32:16.979010 4971 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"test-operator-ephemeral-temporary\" (UniqueName: \"kubernetes.io/empty-dir/0a604d8e-de8c-4ae6-b13a-aafd2adee17a-test-operator-ephemeral-temporary\") pod \"tobiko-tests-tobiko-s00-podified-functional\" (UID: \"0a604d8e-de8c-4ae6-b13a-aafd2adee17a\") " pod="openstack/tobiko-tests-tobiko-s00-podified-functional" Dec 13 09:32:16 crc kubenswrapper[4971]: I1213 09:32:16.979064 4971 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"openstack-config-secret\" (UniqueName: \"kubernetes.io/secret/0a604d8e-de8c-4ae6-b13a-aafd2adee17a-openstack-config-secret\") pod \"tobiko-tests-tobiko-s00-podified-functional\" (UID: \"0a604d8e-de8c-4ae6-b13a-aafd2adee17a\") " pod="openstack/tobiko-tests-tobiko-s00-podified-functional" Dec 13 09:32:16 crc kubenswrapper[4971]: I1213 09:32:16.979086 4971 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"tobiko-public-key\" (UniqueName: \"kubernetes.io/configmap/0a604d8e-de8c-4ae6-b13a-aafd2adee17a-tobiko-public-key\") pod \"tobiko-tests-tobiko-s00-podified-functional\" (UID: \"0a604d8e-de8c-4ae6-b13a-aafd2adee17a\") " pod="openstack/tobiko-tests-tobiko-s00-podified-functional" Dec 13 09:32:16 crc kubenswrapper[4971]: I1213 09:32:16.979134 4971 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"tobiko-private-key\" (UniqueName: \"kubernetes.io/configmap/0a604d8e-de8c-4ae6-b13a-aafd2adee17a-tobiko-private-key\") pod \"tobiko-tests-tobiko-s00-podified-functional\" (UID: \"0a604d8e-de8c-4ae6-b13a-aafd2adee17a\") " pod="openstack/tobiko-tests-tobiko-s00-podified-functional" Dec 13 09:32:16 crc kubenswrapper[4971]: I1213 09:32:16.979157 4971 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/0a604d8e-de8c-4ae6-b13a-aafd2adee17a-ceph\") pod \"tobiko-tests-tobiko-s00-podified-functional\" (UID: \"0a604d8e-de8c-4ae6-b13a-aafd2adee17a\") " pod="openstack/tobiko-tests-tobiko-s00-podified-functional" Dec 13 09:32:16 crc kubenswrapper[4971]: I1213 09:32:16.980421 4971 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"local-storage03-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage03-crc\") pod \"tobiko-tests-tobiko-s00-podified-functional\" (UID: \"0a604d8e-de8c-4ae6-b13a-aafd2adee17a\") device mount path \"/mnt/openstack/pv03\"" pod="openstack/tobiko-tests-tobiko-s00-podified-functional" Dec 13 09:32:16 crc kubenswrapper[4971]: I1213 09:32:16.981223 4971 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"tobiko-config\" (UniqueName: \"kubernetes.io/configmap/0a604d8e-de8c-4ae6-b13a-aafd2adee17a-tobiko-config\") pod \"tobiko-tests-tobiko-s00-podified-functional\" (UID: \"0a604d8e-de8c-4ae6-b13a-aafd2adee17a\") " pod="openstack/tobiko-tests-tobiko-s00-podified-functional" Dec 13 09:32:16 crc kubenswrapper[4971]: I1213 09:32:16.984670 4971 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"test-operator-ephemeral-temporary\" (UniqueName: \"kubernetes.io/empty-dir/0a604d8e-de8c-4ae6-b13a-aafd2adee17a-test-operator-ephemeral-temporary\") pod \"tobiko-tests-tobiko-s00-podified-functional\" (UID: \"0a604d8e-de8c-4ae6-b13a-aafd2adee17a\") " pod="openstack/tobiko-tests-tobiko-s00-podified-functional" Dec 13 09:32:16 crc kubenswrapper[4971]: I1213 09:32:16.985251 4971 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"test-operator-ephemeral-workdir\" (UniqueName: \"kubernetes.io/empty-dir/0a604d8e-de8c-4ae6-b13a-aafd2adee17a-test-operator-ephemeral-workdir\") pod \"tobiko-tests-tobiko-s00-podified-functional\" (UID: \"0a604d8e-de8c-4ae6-b13a-aafd2adee17a\") " pod="openstack/tobiko-tests-tobiko-s00-podified-functional" Dec 13 09:32:16 crc kubenswrapper[4971]: I1213 09:32:16.985486 4971 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"tobiko-public-key\" (UniqueName: \"kubernetes.io/configmap/0a604d8e-de8c-4ae6-b13a-aafd2adee17a-tobiko-public-key\") pod \"tobiko-tests-tobiko-s00-podified-functional\" (UID: \"0a604d8e-de8c-4ae6-b13a-aafd2adee17a\") " pod="openstack/tobiko-tests-tobiko-s00-podified-functional" Dec 13 09:32:16 crc kubenswrapper[4971]: I1213 09:32:16.986298 4971 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"tobiko-private-key\" (UniqueName: \"kubernetes.io/configmap/0a604d8e-de8c-4ae6-b13a-aafd2adee17a-tobiko-private-key\") pod \"tobiko-tests-tobiko-s00-podified-functional\" (UID: \"0a604d8e-de8c-4ae6-b13a-aafd2adee17a\") " pod="openstack/tobiko-tests-tobiko-s00-podified-functional" Dec 13 09:32:16 crc kubenswrapper[4971]: I1213 09:32:16.986983 4971 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"test-operator-clouds-config\" (UniqueName: \"kubernetes.io/configmap/0a604d8e-de8c-4ae6-b13a-aafd2adee17a-test-operator-clouds-config\") pod \"tobiko-tests-tobiko-s00-podified-functional\" (UID: \"0a604d8e-de8c-4ae6-b13a-aafd2adee17a\") " pod="openstack/tobiko-tests-tobiko-s00-podified-functional" Dec 13 09:32:16 crc kubenswrapper[4971]: I1213 09:32:16.991384 4971 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ca-certs\" (UniqueName: \"kubernetes.io/secret/0a604d8e-de8c-4ae6-b13a-aafd2adee17a-ca-certs\") pod \"tobiko-tests-tobiko-s00-podified-functional\" (UID: \"0a604d8e-de8c-4ae6-b13a-aafd2adee17a\") " pod="openstack/tobiko-tests-tobiko-s00-podified-functional" Dec 13 09:32:16 crc kubenswrapper[4971]: I1213 09:32:16.991685 4971 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/0a604d8e-de8c-4ae6-b13a-aafd2adee17a-ceph\") pod \"tobiko-tests-tobiko-s00-podified-functional\" (UID: \"0a604d8e-de8c-4ae6-b13a-aafd2adee17a\") " pod="openstack/tobiko-tests-tobiko-s00-podified-functional" Dec 13 09:32:17 crc kubenswrapper[4971]: I1213 09:32:17.004280 4971 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"openstack-config-secret\" (UniqueName: \"kubernetes.io/secret/0a604d8e-de8c-4ae6-b13a-aafd2adee17a-openstack-config-secret\") pod \"tobiko-tests-tobiko-s00-podified-functional\" (UID: \"0a604d8e-de8c-4ae6-b13a-aafd2adee17a\") " pod="openstack/tobiko-tests-tobiko-s00-podified-functional" Dec 13 09:32:17 crc kubenswrapper[4971]: I1213 09:32:17.004648 4971 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kubeconfig\" (UniqueName: \"kubernetes.io/secret/0a604d8e-de8c-4ae6-b13a-aafd2adee17a-kubeconfig\") pod \"tobiko-tests-tobiko-s00-podified-functional\" (UID: \"0a604d8e-de8c-4ae6-b13a-aafd2adee17a\") " pod="openstack/tobiko-tests-tobiko-s00-podified-functional" Dec 13 09:32:17 crc kubenswrapper[4971]: I1213 09:32:17.005943 4971 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-twpcl\" (UniqueName: \"kubernetes.io/projected/0a604d8e-de8c-4ae6-b13a-aafd2adee17a-kube-api-access-twpcl\") pod \"tobiko-tests-tobiko-s00-podified-functional\" (UID: \"0a604d8e-de8c-4ae6-b13a-aafd2adee17a\") " pod="openstack/tobiko-tests-tobiko-s00-podified-functional" Dec 13 09:32:17 crc kubenswrapper[4971]: I1213 09:32:17.018900 4971 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"local-storage03-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage03-crc\") pod \"tobiko-tests-tobiko-s00-podified-functional\" (UID: \"0a604d8e-de8c-4ae6-b13a-aafd2adee17a\") " pod="openstack/tobiko-tests-tobiko-s00-podified-functional" Dec 13 09:32:17 crc kubenswrapper[4971]: I1213 09:32:17.072192 4971 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/tobiko-tests-tobiko-s00-podified-functional" Dec 13 09:32:18 crc kubenswrapper[4971]: I1213 09:32:18.404867 4971 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/tobiko-tests-tobiko-s00-podified-functional"] Dec 13 09:32:18 crc kubenswrapper[4971]: W1213 09:32:18.408068 4971 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod0a604d8e_de8c_4ae6_b13a_aafd2adee17a.slice/crio-6951f82af407c23ab66239b37e5b23977d60f8228392b5bcb3c13d5bc6737c07 WatchSource:0}: Error finding container 6951f82af407c23ab66239b37e5b23977d60f8228392b5bcb3c13d5bc6737c07: Status 404 returned error can't find the container with id 6951f82af407c23ab66239b37e5b23977d60f8228392b5bcb3c13d5bc6737c07 Dec 13 09:32:18 crc kubenswrapper[4971]: I1213 09:32:18.997855 4971 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/tobiko-tests-tobiko-s00-podified-functional" event={"ID":"0a604d8e-de8c-4ae6-b13a-aafd2adee17a","Type":"ContainerStarted","Data":"6951f82af407c23ab66239b37e5b23977d60f8228392b5bcb3c13d5bc6737c07"} Dec 13 09:32:38 crc kubenswrapper[4971]: E1213 09:32:38.570086 4971 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying config: context canceled" image="quay.io/podified-antelope-centos9/openstack-tobiko:current-podified" Dec 13 09:32:38 crc kubenswrapper[4971]: E1213 09:32:38.571309 4971 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:tobiko-tests-tobiko,Image:quay.io/podified-antelope-centos9/openstack-tobiko:current-podified,Command:[],Args:[],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:TOBIKO_DEBUG_MODE,Value:false,ValueFrom:nil,},EnvVar{Name:TOBIKO_KEYS_FOLDER,Value:/etc/test_operator,ValueFrom:nil,},EnvVar{Name:TOBIKO_LOGS_DIR_NAME,Value:tobiko-tests-tobiko-s00-podified-functional,ValueFrom:nil,},EnvVar{Name:TOBIKO_PYTEST_ADDOPTS,Value:,ValueFrom:nil,},EnvVar{Name:TOBIKO_TESTENV,Value:functional -- tobiko/tests/functional/podified/test_topology.py,ValueFrom:nil,},EnvVar{Name:TOBIKO_VERSION,Value:master,ValueFrom:nil,},EnvVar{Name:TOX_NUM_PROCESSES,Value:2,ValueFrom:nil,},EnvVar{Name:USE_EXTERNAL_FILES,Value:True,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{cpu: {{8 0} {} 8 DecimalSI},memory: {{8589934592 0} {} BinarySI},},Requests:ResourceList{cpu: {{4 0} {} 4 DecimalSI},memory: {{4294967296 0} {} 4Gi BinarySI},},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:test-operator-ephemeral-workdir,ReadOnly:false,MountPath:/var/lib/tobiko,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:test-operator-ephemeral-temporary,ReadOnly:false,MountPath:/tmp,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:test-operator-logs,ReadOnly:false,MountPath:/var/lib/tobiko/external_files,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:test-operator-clouds-config,ReadOnly:true,MountPath:/var/lib/tobiko/.config/openstack/clouds.yaml,SubPath:clouds.yaml,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:test-operator-clouds-config,ReadOnly:true,MountPath:/etc/openstack/clouds.yaml,SubPath:clouds.yaml,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:openstack-config-secret,ReadOnly:false,MountPath:/etc/openstack/secure.yaml,SubPath:secure.yaml,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:tobiko-config,ReadOnly:false,MountPath:/etc/tobiko/tobiko.conf,SubPath:tobiko.conf,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:ca-certs,ReadOnly:true,MountPath:/etc/pki/ca-trust/extracted/pem/tls-ca-bundle.pem,SubPath:tls-ca-bundle.pem,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:ca-certs,ReadOnly:true,MountPath:/etc/pki/tls/certs/ca-bundle.trust.crt,SubPath:tls-ca-bundle.pem,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:tobiko-private-key,ReadOnly:true,MountPath:/etc/test_operator/id_ecdsa,SubPath:id_ecdsa,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:tobiko-public-key,ReadOnly:true,MountPath:/etc/test_operator/id_ecdsa.pub,SubPath:id_ecdsa.pub,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kubeconfig,ReadOnly:true,MountPath:/var/lib/tobiko/.kube/config,SubPath:config,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:ceph,ReadOnly:true,MountPath:/etc/ceph,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-twpcl,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[NET_ADMIN NET_RAW],Drop:[],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*42495,RunAsNonRoot:*false,ReadOnlyRootFilesystem:*false,AllowPrivilegeEscalation:*true,RunAsGroup:*42495,ProcMount:nil,WindowsOptions:nil,SeccompProfile:&SeccompProfile{Type:RuntimeDefault,LocalhostProfile:nil,},AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod tobiko-tests-tobiko-s00-podified-functional_openstack(0a604d8e-de8c-4ae6-b13a-aafd2adee17a): ErrImagePull: rpc error: code = Canceled desc = copying config: context canceled" logger="UnhandledError" Dec 13 09:32:38 crc kubenswrapper[4971]: E1213 09:32:38.572596 4971 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"tobiko-tests-tobiko\" with ErrImagePull: \"rpc error: code = Canceled desc = copying config: context canceled\"" pod="openstack/tobiko-tests-tobiko-s00-podified-functional" podUID="0a604d8e-de8c-4ae6-b13a-aafd2adee17a" Dec 13 09:32:39 crc kubenswrapper[4971]: E1213 09:32:39.279545 4971 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"tobiko-tests-tobiko\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/podified-antelope-centos9/openstack-tobiko:current-podified\\\"\"" pod="openstack/tobiko-tests-tobiko-s00-podified-functional" podUID="0a604d8e-de8c-4ae6-b13a-aafd2adee17a" Dec 13 09:32:58 crc kubenswrapper[4971]: I1213 09:32:58.483585 4971 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/tobiko-tests-tobiko-s00-podified-functional" event={"ID":"0a604d8e-de8c-4ae6-b13a-aafd2adee17a","Type":"ContainerStarted","Data":"6eb067ac6f27912ea3d398eba5ddfd9918af9977e7bd3562ef3ccd1a5d55ef92"} Dec 13 09:32:58 crc kubenswrapper[4971]: I1213 09:32:58.531070 4971 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/tobiko-tests-tobiko-s00-podified-functional" podStartSLOduration=4.2405639 podStartE2EDuration="43.5310436s" podCreationTimestamp="2025-12-13 09:32:15 +0000 UTC" firstStartedPulling="2025-12-13 09:32:18.414207625 +0000 UTC m=+9795.018617073" lastFinishedPulling="2025-12-13 09:32:57.704687325 +0000 UTC m=+9834.309096773" observedRunningTime="2025-12-13 09:32:58.525936956 +0000 UTC m=+9835.130346404" watchObservedRunningTime="2025-12-13 09:32:58.5310436 +0000 UTC m=+9835.135453048" Dec 13 09:33:46 crc kubenswrapper[4971]: I1213 09:33:46.153256 4971 patch_prober.go:28] interesting pod/machine-config-daemon-82xjz container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 13 09:33:46 crc kubenswrapper[4971]: I1213 09:33:46.153863 4971 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-82xjz" podUID="e0c4af12-edda-41d7-9b44-f87396174ef2" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 13 09:34:10 crc kubenswrapper[4971]: I1213 09:34:10.002865 4971 generic.go:334] "Generic (PLEG): container finished" podID="0a604d8e-de8c-4ae6-b13a-aafd2adee17a" containerID="6eb067ac6f27912ea3d398eba5ddfd9918af9977e7bd3562ef3ccd1a5d55ef92" exitCode=0 Dec 13 09:34:10 crc kubenswrapper[4971]: I1213 09:34:10.003031 4971 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/tobiko-tests-tobiko-s00-podified-functional" event={"ID":"0a604d8e-de8c-4ae6-b13a-aafd2adee17a","Type":"ContainerDied","Data":"6eb067ac6f27912ea3d398eba5ddfd9918af9977e7bd3562ef3ccd1a5d55ef92"} Dec 13 09:34:11 crc kubenswrapper[4971]: I1213 09:34:11.459062 4971 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/tobiko-tests-tobiko-s00-podified-functional" Dec 13 09:34:11 crc kubenswrapper[4971]: I1213 09:34:11.628480 4971 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"test-operator-clouds-config\" (UniqueName: \"kubernetes.io/configmap/0a604d8e-de8c-4ae6-b13a-aafd2adee17a-test-operator-clouds-config\") pod \"0a604d8e-de8c-4ae6-b13a-aafd2adee17a\" (UID: \"0a604d8e-de8c-4ae6-b13a-aafd2adee17a\") " Dec 13 09:34:11 crc kubenswrapper[4971]: I1213 09:34:11.628581 4971 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"test-operator-logs\" (UniqueName: \"kubernetes.io/local-volume/local-storage03-crc\") pod \"0a604d8e-de8c-4ae6-b13a-aafd2adee17a\" (UID: \"0a604d8e-de8c-4ae6-b13a-aafd2adee17a\") " Dec 13 09:34:11 crc kubenswrapper[4971]: I1213 09:34:11.628635 4971 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"tobiko-config\" (UniqueName: \"kubernetes.io/configmap/0a604d8e-de8c-4ae6-b13a-aafd2adee17a-tobiko-config\") pod \"0a604d8e-de8c-4ae6-b13a-aafd2adee17a\" (UID: \"0a604d8e-de8c-4ae6-b13a-aafd2adee17a\") " Dec 13 09:34:11 crc kubenswrapper[4971]: I1213 09:34:11.628872 4971 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"openstack-config-secret\" (UniqueName: \"kubernetes.io/secret/0a604d8e-de8c-4ae6-b13a-aafd2adee17a-openstack-config-secret\") pod \"0a604d8e-de8c-4ae6-b13a-aafd2adee17a\" (UID: \"0a604d8e-de8c-4ae6-b13a-aafd2adee17a\") " Dec 13 09:34:11 crc kubenswrapper[4971]: I1213 09:34:11.628899 4971 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"tobiko-private-key\" (UniqueName: \"kubernetes.io/configmap/0a604d8e-de8c-4ae6-b13a-aafd2adee17a-tobiko-private-key\") pod \"0a604d8e-de8c-4ae6-b13a-aafd2adee17a\" (UID: \"0a604d8e-de8c-4ae6-b13a-aafd2adee17a\") " Dec 13 09:34:11 crc kubenswrapper[4971]: I1213 09:34:11.628957 4971 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-twpcl\" (UniqueName: \"kubernetes.io/projected/0a604d8e-de8c-4ae6-b13a-aafd2adee17a-kube-api-access-twpcl\") pod \"0a604d8e-de8c-4ae6-b13a-aafd2adee17a\" (UID: \"0a604d8e-de8c-4ae6-b13a-aafd2adee17a\") " Dec 13 09:34:11 crc kubenswrapper[4971]: I1213 09:34:11.628983 4971 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"tobiko-public-key\" (UniqueName: \"kubernetes.io/configmap/0a604d8e-de8c-4ae6-b13a-aafd2adee17a-tobiko-public-key\") pod \"0a604d8e-de8c-4ae6-b13a-aafd2adee17a\" (UID: \"0a604d8e-de8c-4ae6-b13a-aafd2adee17a\") " Dec 13 09:34:11 crc kubenswrapper[4971]: I1213 09:34:11.629010 4971 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kubeconfig\" (UniqueName: \"kubernetes.io/secret/0a604d8e-de8c-4ae6-b13a-aafd2adee17a-kubeconfig\") pod \"0a604d8e-de8c-4ae6-b13a-aafd2adee17a\" (UID: \"0a604d8e-de8c-4ae6-b13a-aafd2adee17a\") " Dec 13 09:34:11 crc kubenswrapper[4971]: I1213 09:34:11.629072 4971 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ca-certs\" (UniqueName: \"kubernetes.io/secret/0a604d8e-de8c-4ae6-b13a-aafd2adee17a-ca-certs\") pod \"0a604d8e-de8c-4ae6-b13a-aafd2adee17a\" (UID: \"0a604d8e-de8c-4ae6-b13a-aafd2adee17a\") " Dec 13 09:34:11 crc kubenswrapper[4971]: I1213 09:34:11.629134 4971 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"test-operator-ephemeral-workdir\" (UniqueName: \"kubernetes.io/empty-dir/0a604d8e-de8c-4ae6-b13a-aafd2adee17a-test-operator-ephemeral-workdir\") pod \"0a604d8e-de8c-4ae6-b13a-aafd2adee17a\" (UID: \"0a604d8e-de8c-4ae6-b13a-aafd2adee17a\") " Dec 13 09:34:11 crc kubenswrapper[4971]: I1213 09:34:11.629181 4971 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"test-operator-ephemeral-temporary\" (UniqueName: \"kubernetes.io/empty-dir/0a604d8e-de8c-4ae6-b13a-aafd2adee17a-test-operator-ephemeral-temporary\") pod \"0a604d8e-de8c-4ae6-b13a-aafd2adee17a\" (UID: \"0a604d8e-de8c-4ae6-b13a-aafd2adee17a\") " Dec 13 09:34:11 crc kubenswrapper[4971]: I1213 09:34:11.629198 4971 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/0a604d8e-de8c-4ae6-b13a-aafd2adee17a-ceph\") pod \"0a604d8e-de8c-4ae6-b13a-aafd2adee17a\" (UID: \"0a604d8e-de8c-4ae6-b13a-aafd2adee17a\") " Dec 13 09:34:11 crc kubenswrapper[4971]: I1213 09:34:11.631147 4971 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/0a604d8e-de8c-4ae6-b13a-aafd2adee17a-test-operator-ephemeral-temporary" (OuterVolumeSpecName: "test-operator-ephemeral-temporary") pod "0a604d8e-de8c-4ae6-b13a-aafd2adee17a" (UID: "0a604d8e-de8c-4ae6-b13a-aafd2adee17a"). InnerVolumeSpecName "test-operator-ephemeral-temporary". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 13 09:34:11 crc kubenswrapper[4971]: I1213 09:34:11.732612 4971 reconciler_common.go:293] "Volume detached for volume \"test-operator-ephemeral-temporary\" (UniqueName: \"kubernetes.io/empty-dir/0a604d8e-de8c-4ae6-b13a-aafd2adee17a-test-operator-ephemeral-temporary\") on node \"crc\" DevicePath \"\"" Dec 13 09:34:11 crc kubenswrapper[4971]: I1213 09:34:11.997790 4971 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/tobiko-tests-tobiko-s01-sanity"] Dec 13 09:34:11 crc kubenswrapper[4971]: E1213 09:34:11.998404 4971 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="0a604d8e-de8c-4ae6-b13a-aafd2adee17a" containerName="tobiko-tests-tobiko" Dec 13 09:34:11 crc kubenswrapper[4971]: I1213 09:34:11.998428 4971 state_mem.go:107] "Deleted CPUSet assignment" podUID="0a604d8e-de8c-4ae6-b13a-aafd2adee17a" containerName="tobiko-tests-tobiko" Dec 13 09:34:11 crc kubenswrapper[4971]: I1213 09:34:11.998689 4971 memory_manager.go:354] "RemoveStaleState removing state" podUID="0a604d8e-de8c-4ae6-b13a-aafd2adee17a" containerName="tobiko-tests-tobiko" Dec 13 09:34:12 crc kubenswrapper[4971]: I1213 09:34:12.002182 4971 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/tobiko-tests-tobiko-s01-sanity" Dec 13 09:34:12 crc kubenswrapper[4971]: I1213 09:34:12.043574 4971 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/tobiko-tests-tobiko-s01-sanity"] Dec 13 09:34:12 crc kubenswrapper[4971]: I1213 09:34:12.051798 4971 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/tobiko-tests-tobiko-s00-podified-functional" event={"ID":"0a604d8e-de8c-4ae6-b13a-aafd2adee17a","Type":"ContainerDied","Data":"6951f82af407c23ab66239b37e5b23977d60f8228392b5bcb3c13d5bc6737c07"} Dec 13 09:34:12 crc kubenswrapper[4971]: I1213 09:34:12.051873 4971 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="6951f82af407c23ab66239b37e5b23977d60f8228392b5bcb3c13d5bc6737c07" Dec 13 09:34:12 crc kubenswrapper[4971]: I1213 09:34:12.051965 4971 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/tobiko-tests-tobiko-s00-podified-functional" Dec 13 09:34:12 crc kubenswrapper[4971]: I1213 09:34:12.143781 4971 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/local-volume/local-storage03-crc" (OuterVolumeSpecName: "test-operator-logs") pod "0a604d8e-de8c-4ae6-b13a-aafd2adee17a" (UID: "0a604d8e-de8c-4ae6-b13a-aafd2adee17a"). InnerVolumeSpecName "local-storage03-crc". PluginName "kubernetes.io/local-volume", VolumeGidValue "" Dec 13 09:34:12 crc kubenswrapper[4971]: I1213 09:34:12.143848 4971 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/0a604d8e-de8c-4ae6-b13a-aafd2adee17a-ceph" (OuterVolumeSpecName: "ceph") pod "0a604d8e-de8c-4ae6-b13a-aafd2adee17a" (UID: "0a604d8e-de8c-4ae6-b13a-aafd2adee17a"). InnerVolumeSpecName "ceph". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 13 09:34:12 crc kubenswrapper[4971]: I1213 09:34:12.143848 4971 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/0a604d8e-de8c-4ae6-b13a-aafd2adee17a-kube-api-access-twpcl" (OuterVolumeSpecName: "kube-api-access-twpcl") pod "0a604d8e-de8c-4ae6-b13a-aafd2adee17a" (UID: "0a604d8e-de8c-4ae6-b13a-aafd2adee17a"). InnerVolumeSpecName "kube-api-access-twpcl". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 13 09:34:12 crc kubenswrapper[4971]: I1213 09:34:12.145568 4971 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"test-operator-ephemeral-workdir\" (UniqueName: \"kubernetes.io/empty-dir/32d4b615-70d7-4af1-9280-8fb2452feb76-test-operator-ephemeral-workdir\") pod \"tobiko-tests-tobiko-s01-sanity\" (UID: \"32d4b615-70d7-4af1-9280-8fb2452feb76\") " pod="openstack/tobiko-tests-tobiko-s01-sanity" Dec 13 09:34:12 crc kubenswrapper[4971]: I1213 09:34:12.145657 4971 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"test-operator-ephemeral-temporary\" (UniqueName: \"kubernetes.io/empty-dir/32d4b615-70d7-4af1-9280-8fb2452feb76-test-operator-ephemeral-temporary\") pod \"tobiko-tests-tobiko-s01-sanity\" (UID: \"32d4b615-70d7-4af1-9280-8fb2452feb76\") " pod="openstack/tobiko-tests-tobiko-s01-sanity" Dec 13 09:34:12 crc kubenswrapper[4971]: I1213 09:34:12.145685 4971 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-9x8rp\" (UniqueName: \"kubernetes.io/projected/32d4b615-70d7-4af1-9280-8fb2452feb76-kube-api-access-9x8rp\") pod \"tobiko-tests-tobiko-s01-sanity\" (UID: \"32d4b615-70d7-4af1-9280-8fb2452feb76\") " pod="openstack/tobiko-tests-tobiko-s01-sanity" Dec 13 09:34:12 crc kubenswrapper[4971]: I1213 09:34:12.145716 4971 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"local-storage03-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage03-crc\") pod \"tobiko-tests-tobiko-s01-sanity\" (UID: \"32d4b615-70d7-4af1-9280-8fb2452feb76\") " pod="openstack/tobiko-tests-tobiko-s01-sanity" Dec 13 09:34:12 crc kubenswrapper[4971]: I1213 09:34:12.145781 4971 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"tobiko-public-key\" (UniqueName: \"kubernetes.io/configmap/32d4b615-70d7-4af1-9280-8fb2452feb76-tobiko-public-key\") pod \"tobiko-tests-tobiko-s01-sanity\" (UID: \"32d4b615-70d7-4af1-9280-8fb2452feb76\") " pod="openstack/tobiko-tests-tobiko-s01-sanity" Dec 13 09:34:12 crc kubenswrapper[4971]: I1213 09:34:12.145807 4971 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"tobiko-private-key\" (UniqueName: \"kubernetes.io/configmap/32d4b615-70d7-4af1-9280-8fb2452feb76-tobiko-private-key\") pod \"tobiko-tests-tobiko-s01-sanity\" (UID: \"32d4b615-70d7-4af1-9280-8fb2452feb76\") " pod="openstack/tobiko-tests-tobiko-s01-sanity" Dec 13 09:34:12 crc kubenswrapper[4971]: I1213 09:34:12.145870 4971 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"tobiko-config\" (UniqueName: \"kubernetes.io/configmap/32d4b615-70d7-4af1-9280-8fb2452feb76-tobiko-config\") pod \"tobiko-tests-tobiko-s01-sanity\" (UID: \"32d4b615-70d7-4af1-9280-8fb2452feb76\") " pod="openstack/tobiko-tests-tobiko-s01-sanity" Dec 13 09:34:12 crc kubenswrapper[4971]: I1213 09:34:12.145909 4971 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"openstack-config-secret\" (UniqueName: \"kubernetes.io/secret/32d4b615-70d7-4af1-9280-8fb2452feb76-openstack-config-secret\") pod \"tobiko-tests-tobiko-s01-sanity\" (UID: \"32d4b615-70d7-4af1-9280-8fb2452feb76\") " pod="openstack/tobiko-tests-tobiko-s01-sanity" Dec 13 09:34:12 crc kubenswrapper[4971]: I1213 09:34:12.145925 4971 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"test-operator-clouds-config\" (UniqueName: \"kubernetes.io/configmap/32d4b615-70d7-4af1-9280-8fb2452feb76-test-operator-clouds-config\") pod \"tobiko-tests-tobiko-s01-sanity\" (UID: \"32d4b615-70d7-4af1-9280-8fb2452feb76\") " pod="openstack/tobiko-tests-tobiko-s01-sanity" Dec 13 09:34:12 crc kubenswrapper[4971]: I1213 09:34:12.145944 4971 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kubeconfig\" (UniqueName: \"kubernetes.io/secret/32d4b615-70d7-4af1-9280-8fb2452feb76-kubeconfig\") pod \"tobiko-tests-tobiko-s01-sanity\" (UID: \"32d4b615-70d7-4af1-9280-8fb2452feb76\") " pod="openstack/tobiko-tests-tobiko-s01-sanity" Dec 13 09:34:12 crc kubenswrapper[4971]: I1213 09:34:12.145960 4971 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/32d4b615-70d7-4af1-9280-8fb2452feb76-ceph\") pod \"tobiko-tests-tobiko-s01-sanity\" (UID: \"32d4b615-70d7-4af1-9280-8fb2452feb76\") " pod="openstack/tobiko-tests-tobiko-s01-sanity" Dec 13 09:34:12 crc kubenswrapper[4971]: I1213 09:34:12.146006 4971 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ca-certs\" (UniqueName: \"kubernetes.io/secret/32d4b615-70d7-4af1-9280-8fb2452feb76-ca-certs\") pod \"tobiko-tests-tobiko-s01-sanity\" (UID: \"32d4b615-70d7-4af1-9280-8fb2452feb76\") " pod="openstack/tobiko-tests-tobiko-s01-sanity" Dec 13 09:34:12 crc kubenswrapper[4971]: I1213 09:34:12.146067 4971 reconciler_common.go:293] "Volume detached for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/0a604d8e-de8c-4ae6-b13a-aafd2adee17a-ceph\") on node \"crc\" DevicePath \"\"" Dec 13 09:34:12 crc kubenswrapper[4971]: I1213 09:34:12.146079 4971 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-twpcl\" (UniqueName: \"kubernetes.io/projected/0a604d8e-de8c-4ae6-b13a-aafd2adee17a-kube-api-access-twpcl\") on node \"crc\" DevicePath \"\"" Dec 13 09:34:12 crc kubenswrapper[4971]: I1213 09:34:12.161438 4971 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/0a604d8e-de8c-4ae6-b13a-aafd2adee17a-tobiko-private-key" (OuterVolumeSpecName: "tobiko-private-key") pod "0a604d8e-de8c-4ae6-b13a-aafd2adee17a" (UID: "0a604d8e-de8c-4ae6-b13a-aafd2adee17a"). InnerVolumeSpecName "tobiko-private-key". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 13 09:34:12 crc kubenswrapper[4971]: I1213 09:34:12.165397 4971 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/0a604d8e-de8c-4ae6-b13a-aafd2adee17a-kubeconfig" (OuterVolumeSpecName: "kubeconfig") pod "0a604d8e-de8c-4ae6-b13a-aafd2adee17a" (UID: "0a604d8e-de8c-4ae6-b13a-aafd2adee17a"). InnerVolumeSpecName "kubeconfig". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 13 09:34:12 crc kubenswrapper[4971]: I1213 09:34:12.169472 4971 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/0a604d8e-de8c-4ae6-b13a-aafd2adee17a-tobiko-public-key" (OuterVolumeSpecName: "tobiko-public-key") pod "0a604d8e-de8c-4ae6-b13a-aafd2adee17a" (UID: "0a604d8e-de8c-4ae6-b13a-aafd2adee17a"). InnerVolumeSpecName "tobiko-public-key". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 13 09:34:12 crc kubenswrapper[4971]: I1213 09:34:12.171622 4971 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/0a604d8e-de8c-4ae6-b13a-aafd2adee17a-tobiko-config" (OuterVolumeSpecName: "tobiko-config") pod "0a604d8e-de8c-4ae6-b13a-aafd2adee17a" (UID: "0a604d8e-de8c-4ae6-b13a-aafd2adee17a"). InnerVolumeSpecName "tobiko-config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 13 09:34:12 crc kubenswrapper[4971]: I1213 09:34:12.174919 4971 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/0a604d8e-de8c-4ae6-b13a-aafd2adee17a-openstack-config-secret" (OuterVolumeSpecName: "openstack-config-secret") pod "0a604d8e-de8c-4ae6-b13a-aafd2adee17a" (UID: "0a604d8e-de8c-4ae6-b13a-aafd2adee17a"). InnerVolumeSpecName "openstack-config-secret". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 13 09:34:12 crc kubenswrapper[4971]: I1213 09:34:12.182452 4971 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"local-storage03-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage03-crc\") pod \"tobiko-tests-tobiko-s01-sanity\" (UID: \"32d4b615-70d7-4af1-9280-8fb2452feb76\") " pod="openstack/tobiko-tests-tobiko-s01-sanity" Dec 13 09:34:12 crc kubenswrapper[4971]: I1213 09:34:12.195170 4971 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/0a604d8e-de8c-4ae6-b13a-aafd2adee17a-test-operator-clouds-config" (OuterVolumeSpecName: "test-operator-clouds-config") pod "0a604d8e-de8c-4ae6-b13a-aafd2adee17a" (UID: "0a604d8e-de8c-4ae6-b13a-aafd2adee17a"). InnerVolumeSpecName "test-operator-clouds-config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 13 09:34:12 crc kubenswrapper[4971]: I1213 09:34:12.204950 4971 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/0a604d8e-de8c-4ae6-b13a-aafd2adee17a-ca-certs" (OuterVolumeSpecName: "ca-certs") pod "0a604d8e-de8c-4ae6-b13a-aafd2adee17a" (UID: "0a604d8e-de8c-4ae6-b13a-aafd2adee17a"). InnerVolumeSpecName "ca-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 13 09:34:12 crc kubenswrapper[4971]: I1213 09:34:12.248160 4971 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-9x8rp\" (UniqueName: \"kubernetes.io/projected/32d4b615-70d7-4af1-9280-8fb2452feb76-kube-api-access-9x8rp\") pod \"tobiko-tests-tobiko-s01-sanity\" (UID: \"32d4b615-70d7-4af1-9280-8fb2452feb76\") " pod="openstack/tobiko-tests-tobiko-s01-sanity" Dec 13 09:34:12 crc kubenswrapper[4971]: I1213 09:34:12.248340 4971 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"tobiko-public-key\" (UniqueName: \"kubernetes.io/configmap/32d4b615-70d7-4af1-9280-8fb2452feb76-tobiko-public-key\") pod \"tobiko-tests-tobiko-s01-sanity\" (UID: \"32d4b615-70d7-4af1-9280-8fb2452feb76\") " pod="openstack/tobiko-tests-tobiko-s01-sanity" Dec 13 09:34:12 crc kubenswrapper[4971]: I1213 09:34:12.248404 4971 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"tobiko-private-key\" (UniqueName: \"kubernetes.io/configmap/32d4b615-70d7-4af1-9280-8fb2452feb76-tobiko-private-key\") pod \"tobiko-tests-tobiko-s01-sanity\" (UID: \"32d4b615-70d7-4af1-9280-8fb2452feb76\") " pod="openstack/tobiko-tests-tobiko-s01-sanity" Dec 13 09:34:12 crc kubenswrapper[4971]: I1213 09:34:12.248500 4971 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"tobiko-config\" (UniqueName: \"kubernetes.io/configmap/32d4b615-70d7-4af1-9280-8fb2452feb76-tobiko-config\") pod \"tobiko-tests-tobiko-s01-sanity\" (UID: \"32d4b615-70d7-4af1-9280-8fb2452feb76\") " pod="openstack/tobiko-tests-tobiko-s01-sanity" Dec 13 09:34:12 crc kubenswrapper[4971]: I1213 09:34:12.248580 4971 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"openstack-config-secret\" (UniqueName: \"kubernetes.io/secret/32d4b615-70d7-4af1-9280-8fb2452feb76-openstack-config-secret\") pod \"tobiko-tests-tobiko-s01-sanity\" (UID: \"32d4b615-70d7-4af1-9280-8fb2452feb76\") " pod="openstack/tobiko-tests-tobiko-s01-sanity" Dec 13 09:34:12 crc kubenswrapper[4971]: I1213 09:34:12.248629 4971 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"test-operator-clouds-config\" (UniqueName: \"kubernetes.io/configmap/32d4b615-70d7-4af1-9280-8fb2452feb76-test-operator-clouds-config\") pod \"tobiko-tests-tobiko-s01-sanity\" (UID: \"32d4b615-70d7-4af1-9280-8fb2452feb76\") " pod="openstack/tobiko-tests-tobiko-s01-sanity" Dec 13 09:34:12 crc kubenswrapper[4971]: I1213 09:34:12.248662 4971 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kubeconfig\" (UniqueName: \"kubernetes.io/secret/32d4b615-70d7-4af1-9280-8fb2452feb76-kubeconfig\") pod \"tobiko-tests-tobiko-s01-sanity\" (UID: \"32d4b615-70d7-4af1-9280-8fb2452feb76\") " pod="openstack/tobiko-tests-tobiko-s01-sanity" Dec 13 09:34:12 crc kubenswrapper[4971]: I1213 09:34:12.248687 4971 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/32d4b615-70d7-4af1-9280-8fb2452feb76-ceph\") pod \"tobiko-tests-tobiko-s01-sanity\" (UID: \"32d4b615-70d7-4af1-9280-8fb2452feb76\") " pod="openstack/tobiko-tests-tobiko-s01-sanity" Dec 13 09:34:12 crc kubenswrapper[4971]: I1213 09:34:12.248799 4971 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ca-certs\" (UniqueName: \"kubernetes.io/secret/32d4b615-70d7-4af1-9280-8fb2452feb76-ca-certs\") pod \"tobiko-tests-tobiko-s01-sanity\" (UID: \"32d4b615-70d7-4af1-9280-8fb2452feb76\") " pod="openstack/tobiko-tests-tobiko-s01-sanity" Dec 13 09:34:12 crc kubenswrapper[4971]: I1213 09:34:12.248834 4971 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"test-operator-ephemeral-workdir\" (UniqueName: \"kubernetes.io/empty-dir/32d4b615-70d7-4af1-9280-8fb2452feb76-test-operator-ephemeral-workdir\") pod \"tobiko-tests-tobiko-s01-sanity\" (UID: \"32d4b615-70d7-4af1-9280-8fb2452feb76\") " pod="openstack/tobiko-tests-tobiko-s01-sanity" Dec 13 09:34:12 crc kubenswrapper[4971]: I1213 09:34:12.248927 4971 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"test-operator-ephemeral-temporary\" (UniqueName: \"kubernetes.io/empty-dir/32d4b615-70d7-4af1-9280-8fb2452feb76-test-operator-ephemeral-temporary\") pod \"tobiko-tests-tobiko-s01-sanity\" (UID: \"32d4b615-70d7-4af1-9280-8fb2452feb76\") " pod="openstack/tobiko-tests-tobiko-s01-sanity" Dec 13 09:34:12 crc kubenswrapper[4971]: I1213 09:34:12.249055 4971 reconciler_common.go:293] "Volume detached for volume \"test-operator-clouds-config\" (UniqueName: \"kubernetes.io/configmap/0a604d8e-de8c-4ae6-b13a-aafd2adee17a-test-operator-clouds-config\") on node \"crc\" DevicePath \"\"" Dec 13 09:34:12 crc kubenswrapper[4971]: I1213 09:34:12.249076 4971 reconciler_common.go:293] "Volume detached for volume \"tobiko-config\" (UniqueName: \"kubernetes.io/configmap/0a604d8e-de8c-4ae6-b13a-aafd2adee17a-tobiko-config\") on node \"crc\" DevicePath \"\"" Dec 13 09:34:12 crc kubenswrapper[4971]: I1213 09:34:12.249090 4971 reconciler_common.go:293] "Volume detached for volume \"tobiko-private-key\" (UniqueName: \"kubernetes.io/configmap/0a604d8e-de8c-4ae6-b13a-aafd2adee17a-tobiko-private-key\") on node \"crc\" DevicePath \"\"" Dec 13 09:34:12 crc kubenswrapper[4971]: I1213 09:34:12.249104 4971 reconciler_common.go:293] "Volume detached for volume \"openstack-config-secret\" (UniqueName: \"kubernetes.io/secret/0a604d8e-de8c-4ae6-b13a-aafd2adee17a-openstack-config-secret\") on node \"crc\" DevicePath \"\"" Dec 13 09:34:12 crc kubenswrapper[4971]: I1213 09:34:12.249172 4971 reconciler_common.go:293] "Volume detached for volume \"tobiko-public-key\" (UniqueName: \"kubernetes.io/configmap/0a604d8e-de8c-4ae6-b13a-aafd2adee17a-tobiko-public-key\") on node \"crc\" DevicePath \"\"" Dec 13 09:34:12 crc kubenswrapper[4971]: I1213 09:34:12.249188 4971 reconciler_common.go:293] "Volume detached for volume \"kubeconfig\" (UniqueName: \"kubernetes.io/secret/0a604d8e-de8c-4ae6-b13a-aafd2adee17a-kubeconfig\") on node \"crc\" DevicePath \"\"" Dec 13 09:34:12 crc kubenswrapper[4971]: I1213 09:34:12.249198 4971 reconciler_common.go:293] "Volume detached for volume \"ca-certs\" (UniqueName: \"kubernetes.io/secret/0a604d8e-de8c-4ae6-b13a-aafd2adee17a-ca-certs\") on node \"crc\" DevicePath \"\"" Dec 13 09:34:12 crc kubenswrapper[4971]: I1213 09:34:12.250106 4971 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"tobiko-private-key\" (UniqueName: \"kubernetes.io/configmap/32d4b615-70d7-4af1-9280-8fb2452feb76-tobiko-private-key\") pod \"tobiko-tests-tobiko-s01-sanity\" (UID: \"32d4b615-70d7-4af1-9280-8fb2452feb76\") " pod="openstack/tobiko-tests-tobiko-s01-sanity" Dec 13 09:34:12 crc kubenswrapper[4971]: I1213 09:34:12.250261 4971 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"tobiko-public-key\" (UniqueName: \"kubernetes.io/configmap/32d4b615-70d7-4af1-9280-8fb2452feb76-tobiko-public-key\") pod \"tobiko-tests-tobiko-s01-sanity\" (UID: \"32d4b615-70d7-4af1-9280-8fb2452feb76\") " pod="openstack/tobiko-tests-tobiko-s01-sanity" Dec 13 09:34:12 crc kubenswrapper[4971]: I1213 09:34:12.251223 4971 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"test-operator-ephemeral-temporary\" (UniqueName: \"kubernetes.io/empty-dir/32d4b615-70d7-4af1-9280-8fb2452feb76-test-operator-ephemeral-temporary\") pod \"tobiko-tests-tobiko-s01-sanity\" (UID: \"32d4b615-70d7-4af1-9280-8fb2452feb76\") " pod="openstack/tobiko-tests-tobiko-s01-sanity" Dec 13 09:34:12 crc kubenswrapper[4971]: I1213 09:34:12.251406 4971 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"test-operator-ephemeral-workdir\" (UniqueName: \"kubernetes.io/empty-dir/32d4b615-70d7-4af1-9280-8fb2452feb76-test-operator-ephemeral-workdir\") pod \"tobiko-tests-tobiko-s01-sanity\" (UID: \"32d4b615-70d7-4af1-9280-8fb2452feb76\") " pod="openstack/tobiko-tests-tobiko-s01-sanity" Dec 13 09:34:12 crc kubenswrapper[4971]: I1213 09:34:12.251661 4971 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"test-operator-clouds-config\" (UniqueName: \"kubernetes.io/configmap/32d4b615-70d7-4af1-9280-8fb2452feb76-test-operator-clouds-config\") pod \"tobiko-tests-tobiko-s01-sanity\" (UID: \"32d4b615-70d7-4af1-9280-8fb2452feb76\") " pod="openstack/tobiko-tests-tobiko-s01-sanity" Dec 13 09:34:12 crc kubenswrapper[4971]: I1213 09:34:12.252138 4971 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"tobiko-config\" (UniqueName: \"kubernetes.io/configmap/32d4b615-70d7-4af1-9280-8fb2452feb76-tobiko-config\") pod \"tobiko-tests-tobiko-s01-sanity\" (UID: \"32d4b615-70d7-4af1-9280-8fb2452feb76\") " pod="openstack/tobiko-tests-tobiko-s01-sanity" Dec 13 09:34:12 crc kubenswrapper[4971]: I1213 09:34:12.255044 4971 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ca-certs\" (UniqueName: \"kubernetes.io/secret/32d4b615-70d7-4af1-9280-8fb2452feb76-ca-certs\") pod \"tobiko-tests-tobiko-s01-sanity\" (UID: \"32d4b615-70d7-4af1-9280-8fb2452feb76\") " pod="openstack/tobiko-tests-tobiko-s01-sanity" Dec 13 09:34:12 crc kubenswrapper[4971]: I1213 09:34:12.255320 4971 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/32d4b615-70d7-4af1-9280-8fb2452feb76-ceph\") pod \"tobiko-tests-tobiko-s01-sanity\" (UID: \"32d4b615-70d7-4af1-9280-8fb2452feb76\") " pod="openstack/tobiko-tests-tobiko-s01-sanity" Dec 13 09:34:12 crc kubenswrapper[4971]: I1213 09:34:12.256556 4971 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"openstack-config-secret\" (UniqueName: \"kubernetes.io/secret/32d4b615-70d7-4af1-9280-8fb2452feb76-openstack-config-secret\") pod \"tobiko-tests-tobiko-s01-sanity\" (UID: \"32d4b615-70d7-4af1-9280-8fb2452feb76\") " pod="openstack/tobiko-tests-tobiko-s01-sanity" Dec 13 09:34:12 crc kubenswrapper[4971]: I1213 09:34:12.262296 4971 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kubeconfig\" (UniqueName: \"kubernetes.io/secret/32d4b615-70d7-4af1-9280-8fb2452feb76-kubeconfig\") pod \"tobiko-tests-tobiko-s01-sanity\" (UID: \"32d4b615-70d7-4af1-9280-8fb2452feb76\") " pod="openstack/tobiko-tests-tobiko-s01-sanity" Dec 13 09:34:12 crc kubenswrapper[4971]: I1213 09:34:12.266996 4971 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-9x8rp\" (UniqueName: \"kubernetes.io/projected/32d4b615-70d7-4af1-9280-8fb2452feb76-kube-api-access-9x8rp\") pod \"tobiko-tests-tobiko-s01-sanity\" (UID: \"32d4b615-70d7-4af1-9280-8fb2452feb76\") " pod="openstack/tobiko-tests-tobiko-s01-sanity" Dec 13 09:34:12 crc kubenswrapper[4971]: I1213 09:34:12.333755 4971 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/tobiko-tests-tobiko-s01-sanity" Dec 13 09:34:12 crc kubenswrapper[4971]: I1213 09:34:12.902712 4971 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/tobiko-tests-tobiko-s01-sanity"] Dec 13 09:34:12 crc kubenswrapper[4971]: I1213 09:34:12.940218 4971 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/0a604d8e-de8c-4ae6-b13a-aafd2adee17a-test-operator-ephemeral-workdir" (OuterVolumeSpecName: "test-operator-ephemeral-workdir") pod "0a604d8e-de8c-4ae6-b13a-aafd2adee17a" (UID: "0a604d8e-de8c-4ae6-b13a-aafd2adee17a"). InnerVolumeSpecName "test-operator-ephemeral-workdir". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 13 09:34:12 crc kubenswrapper[4971]: I1213 09:34:12.987844 4971 reconciler_common.go:293] "Volume detached for volume \"test-operator-ephemeral-workdir\" (UniqueName: \"kubernetes.io/empty-dir/0a604d8e-de8c-4ae6-b13a-aafd2adee17a-test-operator-ephemeral-workdir\") on node \"crc\" DevicePath \"\"" Dec 13 09:34:13 crc kubenswrapper[4971]: I1213 09:34:13.067743 4971 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/tobiko-tests-tobiko-s01-sanity" event={"ID":"32d4b615-70d7-4af1-9280-8fb2452feb76","Type":"ContainerStarted","Data":"6168ab37f128f5736f76f3c675293327f98060acd9ef8c294b6a2e1674398a7d"} Dec 13 09:34:14 crc kubenswrapper[4971]: I1213 09:34:14.079395 4971 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/tobiko-tests-tobiko-s01-sanity" event={"ID":"32d4b615-70d7-4af1-9280-8fb2452feb76","Type":"ContainerStarted","Data":"c9798a3ae835d1d5f6a2a1d7e78c43bfae2af00c805b08d49843ee3a8c0aff82"} Dec 13 09:34:14 crc kubenswrapper[4971]: I1213 09:34:14.112595 4971 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/tobiko-tests-tobiko-s01-sanity" podStartSLOduration=3.112565798 podStartE2EDuration="3.112565798s" podCreationTimestamp="2025-12-13 09:34:11 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-13 09:34:14.102333139 +0000 UTC m=+9910.706742597" watchObservedRunningTime="2025-12-13 09:34:14.112565798 +0000 UTC m=+9910.716975276" Dec 13 09:34:16 crc kubenswrapper[4971]: I1213 09:34:16.153983 4971 patch_prober.go:28] interesting pod/machine-config-daemon-82xjz container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 13 09:34:16 crc kubenswrapper[4971]: I1213 09:34:16.154573 4971 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-82xjz" podUID="e0c4af12-edda-41d7-9b44-f87396174ef2" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 13 09:34:46 crc kubenswrapper[4971]: I1213 09:34:46.153442 4971 patch_prober.go:28] interesting pod/machine-config-daemon-82xjz container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 13 09:34:46 crc kubenswrapper[4971]: I1213 09:34:46.154211 4971 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-82xjz" podUID="e0c4af12-edda-41d7-9b44-f87396174ef2" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 13 09:34:46 crc kubenswrapper[4971]: I1213 09:34:46.154270 4971 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-82xjz" Dec 13 09:34:46 crc kubenswrapper[4971]: I1213 09:34:46.155287 4971 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"7c80479d6aedc6e1a5cced1df9e55738d4a726b6ddd5adfe17363d9c5d097d6a"} pod="openshift-machine-config-operator/machine-config-daemon-82xjz" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Dec 13 09:34:46 crc kubenswrapper[4971]: I1213 09:34:46.155350 4971 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-82xjz" podUID="e0c4af12-edda-41d7-9b44-f87396174ef2" containerName="machine-config-daemon" containerID="cri-o://7c80479d6aedc6e1a5cced1df9e55738d4a726b6ddd5adfe17363d9c5d097d6a" gracePeriod=600 Dec 13 09:34:46 crc kubenswrapper[4971]: E1213 09:34:46.313912 4971 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-82xjz_openshift-machine-config-operator(e0c4af12-edda-41d7-9b44-f87396174ef2)\"" pod="openshift-machine-config-operator/machine-config-daemon-82xjz" podUID="e0c4af12-edda-41d7-9b44-f87396174ef2" Dec 13 09:34:46 crc kubenswrapper[4971]: I1213 09:34:46.406254 4971 generic.go:334] "Generic (PLEG): container finished" podID="e0c4af12-edda-41d7-9b44-f87396174ef2" containerID="7c80479d6aedc6e1a5cced1df9e55738d4a726b6ddd5adfe17363d9c5d097d6a" exitCode=0 Dec 13 09:34:46 crc kubenswrapper[4971]: I1213 09:34:46.406316 4971 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-82xjz" event={"ID":"e0c4af12-edda-41d7-9b44-f87396174ef2","Type":"ContainerDied","Data":"7c80479d6aedc6e1a5cced1df9e55738d4a726b6ddd5adfe17363d9c5d097d6a"} Dec 13 09:34:46 crc kubenswrapper[4971]: I1213 09:34:46.406368 4971 scope.go:117] "RemoveContainer" containerID="541c0610036fe5d23e59f44d9df0207efecabde0efb5dc61dac954401fb98fc1" Dec 13 09:34:46 crc kubenswrapper[4971]: I1213 09:34:46.407442 4971 scope.go:117] "RemoveContainer" containerID="7c80479d6aedc6e1a5cced1df9e55738d4a726b6ddd5adfe17363d9c5d097d6a" Dec 13 09:34:46 crc kubenswrapper[4971]: E1213 09:34:46.407864 4971 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-82xjz_openshift-machine-config-operator(e0c4af12-edda-41d7-9b44-f87396174ef2)\"" pod="openshift-machine-config-operator/machine-config-daemon-82xjz" podUID="e0c4af12-edda-41d7-9b44-f87396174ef2" Dec 13 09:35:01 crc kubenswrapper[4971]: I1213 09:35:01.769477 4971 scope.go:117] "RemoveContainer" containerID="7c80479d6aedc6e1a5cced1df9e55738d4a726b6ddd5adfe17363d9c5d097d6a" Dec 13 09:35:01 crc kubenswrapper[4971]: E1213 09:35:01.770614 4971 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-82xjz_openshift-machine-config-operator(e0c4af12-edda-41d7-9b44-f87396174ef2)\"" pod="openshift-machine-config-operator/machine-config-daemon-82xjz" podUID="e0c4af12-edda-41d7-9b44-f87396174ef2" Dec 13 09:35:15 crc kubenswrapper[4971]: I1213 09:35:15.769037 4971 scope.go:117] "RemoveContainer" containerID="7c80479d6aedc6e1a5cced1df9e55738d4a726b6ddd5adfe17363d9c5d097d6a" Dec 13 09:35:15 crc kubenswrapper[4971]: E1213 09:35:15.770127 4971 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-82xjz_openshift-machine-config-operator(e0c4af12-edda-41d7-9b44-f87396174ef2)\"" pod="openshift-machine-config-operator/machine-config-daemon-82xjz" podUID="e0c4af12-edda-41d7-9b44-f87396174ef2" Dec 13 09:35:26 crc kubenswrapper[4971]: I1213 09:35:26.770224 4971 scope.go:117] "RemoveContainer" containerID="7c80479d6aedc6e1a5cced1df9e55738d4a726b6ddd5adfe17363d9c5d097d6a" Dec 13 09:35:26 crc kubenswrapper[4971]: E1213 09:35:26.771068 4971 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-82xjz_openshift-machine-config-operator(e0c4af12-edda-41d7-9b44-f87396174ef2)\"" pod="openshift-machine-config-operator/machine-config-daemon-82xjz" podUID="e0c4af12-edda-41d7-9b44-f87396174ef2" Dec 13 09:35:38 crc kubenswrapper[4971]: I1213 09:35:38.769022 4971 scope.go:117] "RemoveContainer" containerID="7c80479d6aedc6e1a5cced1df9e55738d4a726b6ddd5adfe17363d9c5d097d6a" Dec 13 09:35:38 crc kubenswrapper[4971]: E1213 09:35:38.770113 4971 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-82xjz_openshift-machine-config-operator(e0c4af12-edda-41d7-9b44-f87396174ef2)\"" pod="openshift-machine-config-operator/machine-config-daemon-82xjz" podUID="e0c4af12-edda-41d7-9b44-f87396174ef2" Dec 13 09:35:48 crc kubenswrapper[4971]: I1213 09:35:48.081599 4971 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-marketplace-gnllm"] Dec 13 09:35:48 crc kubenswrapper[4971]: I1213 09:35:48.086230 4971 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-gnllm" Dec 13 09:35:48 crc kubenswrapper[4971]: I1213 09:35:48.103703 4971 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-gnllm"] Dec 13 09:35:48 crc kubenswrapper[4971]: I1213 09:35:48.223047 4971 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-qwl24\" (UniqueName: \"kubernetes.io/projected/db50ff88-63bb-48b5-ad4f-0a40dc6236a9-kube-api-access-qwl24\") pod \"redhat-marketplace-gnllm\" (UID: \"db50ff88-63bb-48b5-ad4f-0a40dc6236a9\") " pod="openshift-marketplace/redhat-marketplace-gnllm" Dec 13 09:35:48 crc kubenswrapper[4971]: I1213 09:35:48.223164 4971 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/db50ff88-63bb-48b5-ad4f-0a40dc6236a9-utilities\") pod \"redhat-marketplace-gnllm\" (UID: \"db50ff88-63bb-48b5-ad4f-0a40dc6236a9\") " pod="openshift-marketplace/redhat-marketplace-gnllm" Dec 13 09:35:48 crc kubenswrapper[4971]: I1213 09:35:48.223195 4971 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/db50ff88-63bb-48b5-ad4f-0a40dc6236a9-catalog-content\") pod \"redhat-marketplace-gnllm\" (UID: \"db50ff88-63bb-48b5-ad4f-0a40dc6236a9\") " pod="openshift-marketplace/redhat-marketplace-gnllm" Dec 13 09:35:48 crc kubenswrapper[4971]: I1213 09:35:48.325254 4971 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/db50ff88-63bb-48b5-ad4f-0a40dc6236a9-utilities\") pod \"redhat-marketplace-gnllm\" (UID: \"db50ff88-63bb-48b5-ad4f-0a40dc6236a9\") " pod="openshift-marketplace/redhat-marketplace-gnllm" Dec 13 09:35:48 crc kubenswrapper[4971]: I1213 09:35:48.325655 4971 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/db50ff88-63bb-48b5-ad4f-0a40dc6236a9-catalog-content\") pod \"redhat-marketplace-gnllm\" (UID: \"db50ff88-63bb-48b5-ad4f-0a40dc6236a9\") " pod="openshift-marketplace/redhat-marketplace-gnllm" Dec 13 09:35:48 crc kubenswrapper[4971]: I1213 09:35:48.325860 4971 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-qwl24\" (UniqueName: \"kubernetes.io/projected/db50ff88-63bb-48b5-ad4f-0a40dc6236a9-kube-api-access-qwl24\") pod \"redhat-marketplace-gnllm\" (UID: \"db50ff88-63bb-48b5-ad4f-0a40dc6236a9\") " pod="openshift-marketplace/redhat-marketplace-gnllm" Dec 13 09:35:48 crc kubenswrapper[4971]: I1213 09:35:48.326215 4971 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/db50ff88-63bb-48b5-ad4f-0a40dc6236a9-utilities\") pod \"redhat-marketplace-gnllm\" (UID: \"db50ff88-63bb-48b5-ad4f-0a40dc6236a9\") " pod="openshift-marketplace/redhat-marketplace-gnllm" Dec 13 09:35:48 crc kubenswrapper[4971]: I1213 09:35:48.326712 4971 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/db50ff88-63bb-48b5-ad4f-0a40dc6236a9-catalog-content\") pod \"redhat-marketplace-gnllm\" (UID: \"db50ff88-63bb-48b5-ad4f-0a40dc6236a9\") " pod="openshift-marketplace/redhat-marketplace-gnllm" Dec 13 09:35:48 crc kubenswrapper[4971]: I1213 09:35:48.346398 4971 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-qwl24\" (UniqueName: \"kubernetes.io/projected/db50ff88-63bb-48b5-ad4f-0a40dc6236a9-kube-api-access-qwl24\") pod \"redhat-marketplace-gnllm\" (UID: \"db50ff88-63bb-48b5-ad4f-0a40dc6236a9\") " pod="openshift-marketplace/redhat-marketplace-gnllm" Dec 13 09:35:48 crc kubenswrapper[4971]: I1213 09:35:48.410245 4971 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-gnllm" Dec 13 09:35:49 crc kubenswrapper[4971]: I1213 09:35:49.008353 4971 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-gnllm"] Dec 13 09:35:49 crc kubenswrapper[4971]: I1213 09:35:49.093861 4971 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-gnllm" event={"ID":"db50ff88-63bb-48b5-ad4f-0a40dc6236a9","Type":"ContainerStarted","Data":"5b38325a711d59f442d48f28c17a6ff0bfe6e3aa4a7135f55ed9a1cbc5645c6d"} Dec 13 09:35:50 crc kubenswrapper[4971]: I1213 09:35:50.107594 4971 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-gnllm" event={"ID":"db50ff88-63bb-48b5-ad4f-0a40dc6236a9","Type":"ContainerStarted","Data":"3ef1dc6262992fe79d042727c97ab26fb8e038af4b3be406e8557b68f30f2368"} Dec 13 09:35:51 crc kubenswrapper[4971]: I1213 09:35:51.123074 4971 generic.go:334] "Generic (PLEG): container finished" podID="db50ff88-63bb-48b5-ad4f-0a40dc6236a9" containerID="3ef1dc6262992fe79d042727c97ab26fb8e038af4b3be406e8557b68f30f2368" exitCode=0 Dec 13 09:35:51 crc kubenswrapper[4971]: I1213 09:35:51.123146 4971 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-gnllm" event={"ID":"db50ff88-63bb-48b5-ad4f-0a40dc6236a9","Type":"ContainerDied","Data":"3ef1dc6262992fe79d042727c97ab26fb8e038af4b3be406e8557b68f30f2368"} Dec 13 09:35:53 crc kubenswrapper[4971]: I1213 09:35:53.149829 4971 generic.go:334] "Generic (PLEG): container finished" podID="db50ff88-63bb-48b5-ad4f-0a40dc6236a9" containerID="b1b0dade81cfcfd895e36aaea784a5b9def43c7dcce1a834c6bd02562ac4aaae" exitCode=0 Dec 13 09:35:53 crc kubenswrapper[4971]: I1213 09:35:53.149930 4971 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-gnllm" event={"ID":"db50ff88-63bb-48b5-ad4f-0a40dc6236a9","Type":"ContainerDied","Data":"b1b0dade81cfcfd895e36aaea784a5b9def43c7dcce1a834c6bd02562ac4aaae"} Dec 13 09:35:53 crc kubenswrapper[4971]: I1213 09:35:53.775222 4971 scope.go:117] "RemoveContainer" containerID="7c80479d6aedc6e1a5cced1df9e55738d4a726b6ddd5adfe17363d9c5d097d6a" Dec 13 09:35:53 crc kubenswrapper[4971]: E1213 09:35:53.775985 4971 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-82xjz_openshift-machine-config-operator(e0c4af12-edda-41d7-9b44-f87396174ef2)\"" pod="openshift-machine-config-operator/machine-config-daemon-82xjz" podUID="e0c4af12-edda-41d7-9b44-f87396174ef2" Dec 13 09:35:54 crc kubenswrapper[4971]: I1213 09:35:54.163872 4971 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-gnllm" event={"ID":"db50ff88-63bb-48b5-ad4f-0a40dc6236a9","Type":"ContainerStarted","Data":"76736fcbc3c3fb2be0c657d724ceba66d41b2a29b35f6e27829b9ffea818bc22"} Dec 13 09:35:54 crc kubenswrapper[4971]: I1213 09:35:54.192156 4971 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-marketplace-gnllm" podStartSLOduration=3.652500019 podStartE2EDuration="6.19213439s" podCreationTimestamp="2025-12-13 09:35:48 +0000 UTC" firstStartedPulling="2025-12-13 09:35:51.126662739 +0000 UTC m=+10007.731072187" lastFinishedPulling="2025-12-13 09:35:53.66629711 +0000 UTC m=+10010.270706558" observedRunningTime="2025-12-13 09:35:54.190126951 +0000 UTC m=+10010.794536409" watchObservedRunningTime="2025-12-13 09:35:54.19213439 +0000 UTC m=+10010.796543838" Dec 13 09:35:58 crc kubenswrapper[4971]: I1213 09:35:58.411231 4971 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-marketplace-gnllm" Dec 13 09:35:58 crc kubenswrapper[4971]: I1213 09:35:58.411885 4971 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-marketplace-gnllm" Dec 13 09:35:58 crc kubenswrapper[4971]: I1213 09:35:58.479768 4971 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-marketplace-gnllm" Dec 13 09:35:59 crc kubenswrapper[4971]: I1213 09:35:59.272428 4971 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-marketplace-gnllm" Dec 13 09:35:59 crc kubenswrapper[4971]: I1213 09:35:59.340118 4971 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-gnllm"] Dec 13 09:36:01 crc kubenswrapper[4971]: I1213 09:36:01.234801 4971 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-marketplace-gnllm" podUID="db50ff88-63bb-48b5-ad4f-0a40dc6236a9" containerName="registry-server" containerID="cri-o://76736fcbc3c3fb2be0c657d724ceba66d41b2a29b35f6e27829b9ffea818bc22" gracePeriod=2 Dec 13 09:36:01 crc kubenswrapper[4971]: I1213 09:36:01.784476 4971 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-gnllm" Dec 13 09:36:01 crc kubenswrapper[4971]: I1213 09:36:01.820352 4971 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-qwl24\" (UniqueName: \"kubernetes.io/projected/db50ff88-63bb-48b5-ad4f-0a40dc6236a9-kube-api-access-qwl24\") pod \"db50ff88-63bb-48b5-ad4f-0a40dc6236a9\" (UID: \"db50ff88-63bb-48b5-ad4f-0a40dc6236a9\") " Dec 13 09:36:01 crc kubenswrapper[4971]: I1213 09:36:01.820551 4971 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/db50ff88-63bb-48b5-ad4f-0a40dc6236a9-catalog-content\") pod \"db50ff88-63bb-48b5-ad4f-0a40dc6236a9\" (UID: \"db50ff88-63bb-48b5-ad4f-0a40dc6236a9\") " Dec 13 09:36:01 crc kubenswrapper[4971]: I1213 09:36:01.820711 4971 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/db50ff88-63bb-48b5-ad4f-0a40dc6236a9-utilities\") pod \"db50ff88-63bb-48b5-ad4f-0a40dc6236a9\" (UID: \"db50ff88-63bb-48b5-ad4f-0a40dc6236a9\") " Dec 13 09:36:01 crc kubenswrapper[4971]: I1213 09:36:01.822710 4971 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/db50ff88-63bb-48b5-ad4f-0a40dc6236a9-utilities" (OuterVolumeSpecName: "utilities") pod "db50ff88-63bb-48b5-ad4f-0a40dc6236a9" (UID: "db50ff88-63bb-48b5-ad4f-0a40dc6236a9"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 13 09:36:01 crc kubenswrapper[4971]: I1213 09:36:01.829124 4971 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/db50ff88-63bb-48b5-ad4f-0a40dc6236a9-kube-api-access-qwl24" (OuterVolumeSpecName: "kube-api-access-qwl24") pod "db50ff88-63bb-48b5-ad4f-0a40dc6236a9" (UID: "db50ff88-63bb-48b5-ad4f-0a40dc6236a9"). InnerVolumeSpecName "kube-api-access-qwl24". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 13 09:36:01 crc kubenswrapper[4971]: I1213 09:36:01.859489 4971 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/db50ff88-63bb-48b5-ad4f-0a40dc6236a9-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "db50ff88-63bb-48b5-ad4f-0a40dc6236a9" (UID: "db50ff88-63bb-48b5-ad4f-0a40dc6236a9"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 13 09:36:01 crc kubenswrapper[4971]: I1213 09:36:01.923688 4971 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-qwl24\" (UniqueName: \"kubernetes.io/projected/db50ff88-63bb-48b5-ad4f-0a40dc6236a9-kube-api-access-qwl24\") on node \"crc\" DevicePath \"\"" Dec 13 09:36:01 crc kubenswrapper[4971]: I1213 09:36:01.923735 4971 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/db50ff88-63bb-48b5-ad4f-0a40dc6236a9-catalog-content\") on node \"crc\" DevicePath \"\"" Dec 13 09:36:01 crc kubenswrapper[4971]: I1213 09:36:01.923747 4971 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/db50ff88-63bb-48b5-ad4f-0a40dc6236a9-utilities\") on node \"crc\" DevicePath \"\"" Dec 13 09:36:02 crc kubenswrapper[4971]: I1213 09:36:02.247872 4971 generic.go:334] "Generic (PLEG): container finished" podID="db50ff88-63bb-48b5-ad4f-0a40dc6236a9" containerID="76736fcbc3c3fb2be0c657d724ceba66d41b2a29b35f6e27829b9ffea818bc22" exitCode=0 Dec 13 09:36:02 crc kubenswrapper[4971]: I1213 09:36:02.247983 4971 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-gnllm" Dec 13 09:36:02 crc kubenswrapper[4971]: I1213 09:36:02.247988 4971 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-gnllm" event={"ID":"db50ff88-63bb-48b5-ad4f-0a40dc6236a9","Type":"ContainerDied","Data":"76736fcbc3c3fb2be0c657d724ceba66d41b2a29b35f6e27829b9ffea818bc22"} Dec 13 09:36:02 crc kubenswrapper[4971]: I1213 09:36:02.248139 4971 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-gnllm" event={"ID":"db50ff88-63bb-48b5-ad4f-0a40dc6236a9","Type":"ContainerDied","Data":"5b38325a711d59f442d48f28c17a6ff0bfe6e3aa4a7135f55ed9a1cbc5645c6d"} Dec 13 09:36:02 crc kubenswrapper[4971]: I1213 09:36:02.248172 4971 scope.go:117] "RemoveContainer" containerID="76736fcbc3c3fb2be0c657d724ceba66d41b2a29b35f6e27829b9ffea818bc22" Dec 13 09:36:02 crc kubenswrapper[4971]: I1213 09:36:02.298165 4971 scope.go:117] "RemoveContainer" containerID="b1b0dade81cfcfd895e36aaea784a5b9def43c7dcce1a834c6bd02562ac4aaae" Dec 13 09:36:02 crc kubenswrapper[4971]: I1213 09:36:02.301538 4971 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-gnllm"] Dec 13 09:36:02 crc kubenswrapper[4971]: I1213 09:36:02.313624 4971 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-marketplace-gnllm"] Dec 13 09:36:02 crc kubenswrapper[4971]: I1213 09:36:02.320918 4971 scope.go:117] "RemoveContainer" containerID="3ef1dc6262992fe79d042727c97ab26fb8e038af4b3be406e8557b68f30f2368" Dec 13 09:36:02 crc kubenswrapper[4971]: I1213 09:36:02.373054 4971 scope.go:117] "RemoveContainer" containerID="76736fcbc3c3fb2be0c657d724ceba66d41b2a29b35f6e27829b9ffea818bc22" Dec 13 09:36:02 crc kubenswrapper[4971]: E1213 09:36:02.374145 4971 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"76736fcbc3c3fb2be0c657d724ceba66d41b2a29b35f6e27829b9ffea818bc22\": container with ID starting with 76736fcbc3c3fb2be0c657d724ceba66d41b2a29b35f6e27829b9ffea818bc22 not found: ID does not exist" containerID="76736fcbc3c3fb2be0c657d724ceba66d41b2a29b35f6e27829b9ffea818bc22" Dec 13 09:36:02 crc kubenswrapper[4971]: I1213 09:36:02.374225 4971 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"76736fcbc3c3fb2be0c657d724ceba66d41b2a29b35f6e27829b9ffea818bc22"} err="failed to get container status \"76736fcbc3c3fb2be0c657d724ceba66d41b2a29b35f6e27829b9ffea818bc22\": rpc error: code = NotFound desc = could not find container \"76736fcbc3c3fb2be0c657d724ceba66d41b2a29b35f6e27829b9ffea818bc22\": container with ID starting with 76736fcbc3c3fb2be0c657d724ceba66d41b2a29b35f6e27829b9ffea818bc22 not found: ID does not exist" Dec 13 09:36:02 crc kubenswrapper[4971]: I1213 09:36:02.374272 4971 scope.go:117] "RemoveContainer" containerID="b1b0dade81cfcfd895e36aaea784a5b9def43c7dcce1a834c6bd02562ac4aaae" Dec 13 09:36:02 crc kubenswrapper[4971]: E1213 09:36:02.374914 4971 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"b1b0dade81cfcfd895e36aaea784a5b9def43c7dcce1a834c6bd02562ac4aaae\": container with ID starting with b1b0dade81cfcfd895e36aaea784a5b9def43c7dcce1a834c6bd02562ac4aaae not found: ID does not exist" containerID="b1b0dade81cfcfd895e36aaea784a5b9def43c7dcce1a834c6bd02562ac4aaae" Dec 13 09:36:02 crc kubenswrapper[4971]: I1213 09:36:02.374954 4971 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"b1b0dade81cfcfd895e36aaea784a5b9def43c7dcce1a834c6bd02562ac4aaae"} err="failed to get container status \"b1b0dade81cfcfd895e36aaea784a5b9def43c7dcce1a834c6bd02562ac4aaae\": rpc error: code = NotFound desc = could not find container \"b1b0dade81cfcfd895e36aaea784a5b9def43c7dcce1a834c6bd02562ac4aaae\": container with ID starting with b1b0dade81cfcfd895e36aaea784a5b9def43c7dcce1a834c6bd02562ac4aaae not found: ID does not exist" Dec 13 09:36:02 crc kubenswrapper[4971]: I1213 09:36:02.374987 4971 scope.go:117] "RemoveContainer" containerID="3ef1dc6262992fe79d042727c97ab26fb8e038af4b3be406e8557b68f30f2368" Dec 13 09:36:02 crc kubenswrapper[4971]: E1213 09:36:02.375581 4971 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"3ef1dc6262992fe79d042727c97ab26fb8e038af4b3be406e8557b68f30f2368\": container with ID starting with 3ef1dc6262992fe79d042727c97ab26fb8e038af4b3be406e8557b68f30f2368 not found: ID does not exist" containerID="3ef1dc6262992fe79d042727c97ab26fb8e038af4b3be406e8557b68f30f2368" Dec 13 09:36:02 crc kubenswrapper[4971]: I1213 09:36:02.375636 4971 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"3ef1dc6262992fe79d042727c97ab26fb8e038af4b3be406e8557b68f30f2368"} err="failed to get container status \"3ef1dc6262992fe79d042727c97ab26fb8e038af4b3be406e8557b68f30f2368\": rpc error: code = NotFound desc = could not find container \"3ef1dc6262992fe79d042727c97ab26fb8e038af4b3be406e8557b68f30f2368\": container with ID starting with 3ef1dc6262992fe79d042727c97ab26fb8e038af4b3be406e8557b68f30f2368 not found: ID does not exist" Dec 13 09:36:03 crc kubenswrapper[4971]: I1213 09:36:03.787941 4971 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="db50ff88-63bb-48b5-ad4f-0a40dc6236a9" path="/var/lib/kubelet/pods/db50ff88-63bb-48b5-ad4f-0a40dc6236a9/volumes" Dec 13 09:36:05 crc kubenswrapper[4971]: I1213 09:36:05.776183 4971 scope.go:117] "RemoveContainer" containerID="7c80479d6aedc6e1a5cced1df9e55738d4a726b6ddd5adfe17363d9c5d097d6a" Dec 13 09:36:05 crc kubenswrapper[4971]: E1213 09:36:05.776878 4971 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-82xjz_openshift-machine-config-operator(e0c4af12-edda-41d7-9b44-f87396174ef2)\"" pod="openshift-machine-config-operator/machine-config-daemon-82xjz" podUID="e0c4af12-edda-41d7-9b44-f87396174ef2" Dec 13 09:36:13 crc kubenswrapper[4971]: I1213 09:36:13.363834 4971 generic.go:334] "Generic (PLEG): container finished" podID="32d4b615-70d7-4af1-9280-8fb2452feb76" containerID="c9798a3ae835d1d5f6a2a1d7e78c43bfae2af00c805b08d49843ee3a8c0aff82" exitCode=0 Dec 13 09:36:13 crc kubenswrapper[4971]: I1213 09:36:13.364056 4971 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/tobiko-tests-tobiko-s01-sanity" event={"ID":"32d4b615-70d7-4af1-9280-8fb2452feb76","Type":"ContainerDied","Data":"c9798a3ae835d1d5f6a2a1d7e78c43bfae2af00c805b08d49843ee3a8c0aff82"} Dec 13 09:36:14 crc kubenswrapper[4971]: I1213 09:36:14.878494 4971 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/tobiko-tests-tobiko-s01-sanity" Dec 13 09:36:14 crc kubenswrapper[4971]: I1213 09:36:14.948782 4971 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ca-certs\" (UniqueName: \"kubernetes.io/secret/32d4b615-70d7-4af1-9280-8fb2452feb76-ca-certs\") pod \"32d4b615-70d7-4af1-9280-8fb2452feb76\" (UID: \"32d4b615-70d7-4af1-9280-8fb2452feb76\") " Dec 13 09:36:14 crc kubenswrapper[4971]: I1213 09:36:14.949496 4971 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"openstack-config-secret\" (UniqueName: \"kubernetes.io/secret/32d4b615-70d7-4af1-9280-8fb2452feb76-openstack-config-secret\") pod \"32d4b615-70d7-4af1-9280-8fb2452feb76\" (UID: \"32d4b615-70d7-4af1-9280-8fb2452feb76\") " Dec 13 09:36:14 crc kubenswrapper[4971]: I1213 09:36:14.949716 4971 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-9x8rp\" (UniqueName: \"kubernetes.io/projected/32d4b615-70d7-4af1-9280-8fb2452feb76-kube-api-access-9x8rp\") pod \"32d4b615-70d7-4af1-9280-8fb2452feb76\" (UID: \"32d4b615-70d7-4af1-9280-8fb2452feb76\") " Dec 13 09:36:14 crc kubenswrapper[4971]: I1213 09:36:14.949810 4971 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"tobiko-public-key\" (UniqueName: \"kubernetes.io/configmap/32d4b615-70d7-4af1-9280-8fb2452feb76-tobiko-public-key\") pod \"32d4b615-70d7-4af1-9280-8fb2452feb76\" (UID: \"32d4b615-70d7-4af1-9280-8fb2452feb76\") " Dec 13 09:36:14 crc kubenswrapper[4971]: I1213 09:36:14.949848 4971 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"test-operator-logs\" (UniqueName: \"kubernetes.io/local-volume/local-storage03-crc\") pod \"32d4b615-70d7-4af1-9280-8fb2452feb76\" (UID: \"32d4b615-70d7-4af1-9280-8fb2452feb76\") " Dec 13 09:36:14 crc kubenswrapper[4971]: I1213 09:36:14.949942 4971 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"test-operator-ephemeral-workdir\" (UniqueName: \"kubernetes.io/empty-dir/32d4b615-70d7-4af1-9280-8fb2452feb76-test-operator-ephemeral-workdir\") pod \"32d4b615-70d7-4af1-9280-8fb2452feb76\" (UID: \"32d4b615-70d7-4af1-9280-8fb2452feb76\") " Dec 13 09:36:14 crc kubenswrapper[4971]: I1213 09:36:14.949980 4971 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"tobiko-config\" (UniqueName: \"kubernetes.io/configmap/32d4b615-70d7-4af1-9280-8fb2452feb76-tobiko-config\") pod \"32d4b615-70d7-4af1-9280-8fb2452feb76\" (UID: \"32d4b615-70d7-4af1-9280-8fb2452feb76\") " Dec 13 09:36:14 crc kubenswrapper[4971]: I1213 09:36:14.950023 4971 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/32d4b615-70d7-4af1-9280-8fb2452feb76-ceph\") pod \"32d4b615-70d7-4af1-9280-8fb2452feb76\" (UID: \"32d4b615-70d7-4af1-9280-8fb2452feb76\") " Dec 13 09:36:14 crc kubenswrapper[4971]: I1213 09:36:14.950082 4971 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"test-operator-ephemeral-temporary\" (UniqueName: \"kubernetes.io/empty-dir/32d4b615-70d7-4af1-9280-8fb2452feb76-test-operator-ephemeral-temporary\") pod \"32d4b615-70d7-4af1-9280-8fb2452feb76\" (UID: \"32d4b615-70d7-4af1-9280-8fb2452feb76\") " Dec 13 09:36:14 crc kubenswrapper[4971]: I1213 09:36:14.950116 4971 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"tobiko-private-key\" (UniqueName: \"kubernetes.io/configmap/32d4b615-70d7-4af1-9280-8fb2452feb76-tobiko-private-key\") pod \"32d4b615-70d7-4af1-9280-8fb2452feb76\" (UID: \"32d4b615-70d7-4af1-9280-8fb2452feb76\") " Dec 13 09:36:14 crc kubenswrapper[4971]: I1213 09:36:14.950192 4971 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"test-operator-clouds-config\" (UniqueName: \"kubernetes.io/configmap/32d4b615-70d7-4af1-9280-8fb2452feb76-test-operator-clouds-config\") pod \"32d4b615-70d7-4af1-9280-8fb2452feb76\" (UID: \"32d4b615-70d7-4af1-9280-8fb2452feb76\") " Dec 13 09:36:14 crc kubenswrapper[4971]: I1213 09:36:14.950233 4971 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kubeconfig\" (UniqueName: \"kubernetes.io/secret/32d4b615-70d7-4af1-9280-8fb2452feb76-kubeconfig\") pod \"32d4b615-70d7-4af1-9280-8fb2452feb76\" (UID: \"32d4b615-70d7-4af1-9280-8fb2452feb76\") " Dec 13 09:36:14 crc kubenswrapper[4971]: I1213 09:36:14.958441 4971 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/local-volume/local-storage03-crc" (OuterVolumeSpecName: "test-operator-logs") pod "32d4b615-70d7-4af1-9280-8fb2452feb76" (UID: "32d4b615-70d7-4af1-9280-8fb2452feb76"). InnerVolumeSpecName "local-storage03-crc". PluginName "kubernetes.io/local-volume", VolumeGidValue "" Dec 13 09:36:14 crc kubenswrapper[4971]: I1213 09:36:14.958859 4971 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/32d4b615-70d7-4af1-9280-8fb2452feb76-kube-api-access-9x8rp" (OuterVolumeSpecName: "kube-api-access-9x8rp") pod "32d4b615-70d7-4af1-9280-8fb2452feb76" (UID: "32d4b615-70d7-4af1-9280-8fb2452feb76"). InnerVolumeSpecName "kube-api-access-9x8rp". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 13 09:36:14 crc kubenswrapper[4971]: I1213 09:36:14.963363 4971 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/32d4b615-70d7-4af1-9280-8fb2452feb76-ceph" (OuterVolumeSpecName: "ceph") pod "32d4b615-70d7-4af1-9280-8fb2452feb76" (UID: "32d4b615-70d7-4af1-9280-8fb2452feb76"). InnerVolumeSpecName "ceph". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 13 09:36:14 crc kubenswrapper[4971]: I1213 09:36:14.964428 4971 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/32d4b615-70d7-4af1-9280-8fb2452feb76-test-operator-ephemeral-temporary" (OuterVolumeSpecName: "test-operator-ephemeral-temporary") pod "32d4b615-70d7-4af1-9280-8fb2452feb76" (UID: "32d4b615-70d7-4af1-9280-8fb2452feb76"). InnerVolumeSpecName "test-operator-ephemeral-temporary". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 13 09:36:14 crc kubenswrapper[4971]: I1213 09:36:14.986040 4971 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/32d4b615-70d7-4af1-9280-8fb2452feb76-tobiko-config" (OuterVolumeSpecName: "tobiko-config") pod "32d4b615-70d7-4af1-9280-8fb2452feb76" (UID: "32d4b615-70d7-4af1-9280-8fb2452feb76"). InnerVolumeSpecName "tobiko-config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 13 09:36:14 crc kubenswrapper[4971]: I1213 09:36:14.987763 4971 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/32d4b615-70d7-4af1-9280-8fb2452feb76-tobiko-public-key" (OuterVolumeSpecName: "tobiko-public-key") pod "32d4b615-70d7-4af1-9280-8fb2452feb76" (UID: "32d4b615-70d7-4af1-9280-8fb2452feb76"). InnerVolumeSpecName "tobiko-public-key". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 13 09:36:14 crc kubenswrapper[4971]: I1213 09:36:14.989004 4971 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/32d4b615-70d7-4af1-9280-8fb2452feb76-openstack-config-secret" (OuterVolumeSpecName: "openstack-config-secret") pod "32d4b615-70d7-4af1-9280-8fb2452feb76" (UID: "32d4b615-70d7-4af1-9280-8fb2452feb76"). InnerVolumeSpecName "openstack-config-secret". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 13 09:36:14 crc kubenswrapper[4971]: I1213 09:36:14.993623 4971 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/32d4b615-70d7-4af1-9280-8fb2452feb76-kubeconfig" (OuterVolumeSpecName: "kubeconfig") pod "32d4b615-70d7-4af1-9280-8fb2452feb76" (UID: "32d4b615-70d7-4af1-9280-8fb2452feb76"). InnerVolumeSpecName "kubeconfig". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 13 09:36:14 crc kubenswrapper[4971]: I1213 09:36:14.997687 4971 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/32d4b615-70d7-4af1-9280-8fb2452feb76-tobiko-private-key" (OuterVolumeSpecName: "tobiko-private-key") pod "32d4b615-70d7-4af1-9280-8fb2452feb76" (UID: "32d4b615-70d7-4af1-9280-8fb2452feb76"). InnerVolumeSpecName "tobiko-private-key". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 13 09:36:15 crc kubenswrapper[4971]: I1213 09:36:15.014444 4971 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/32d4b615-70d7-4af1-9280-8fb2452feb76-test-operator-clouds-config" (OuterVolumeSpecName: "test-operator-clouds-config") pod "32d4b615-70d7-4af1-9280-8fb2452feb76" (UID: "32d4b615-70d7-4af1-9280-8fb2452feb76"). InnerVolumeSpecName "test-operator-clouds-config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 13 09:36:15 crc kubenswrapper[4971]: I1213 09:36:15.014973 4971 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/32d4b615-70d7-4af1-9280-8fb2452feb76-ca-certs" (OuterVolumeSpecName: "ca-certs") pod "32d4b615-70d7-4af1-9280-8fb2452feb76" (UID: "32d4b615-70d7-4af1-9280-8fb2452feb76"). InnerVolumeSpecName "ca-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 13 09:36:15 crc kubenswrapper[4971]: I1213 09:36:15.053126 4971 reconciler_common.go:293] "Volume detached for volume \"tobiko-config\" (UniqueName: \"kubernetes.io/configmap/32d4b615-70d7-4af1-9280-8fb2452feb76-tobiko-config\") on node \"crc\" DevicePath \"\"" Dec 13 09:36:15 crc kubenswrapper[4971]: I1213 09:36:15.053172 4971 reconciler_common.go:293] "Volume detached for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/32d4b615-70d7-4af1-9280-8fb2452feb76-ceph\") on node \"crc\" DevicePath \"\"" Dec 13 09:36:15 crc kubenswrapper[4971]: I1213 09:36:15.053185 4971 reconciler_common.go:293] "Volume detached for volume \"test-operator-ephemeral-temporary\" (UniqueName: \"kubernetes.io/empty-dir/32d4b615-70d7-4af1-9280-8fb2452feb76-test-operator-ephemeral-temporary\") on node \"crc\" DevicePath \"\"" Dec 13 09:36:15 crc kubenswrapper[4971]: I1213 09:36:15.053197 4971 reconciler_common.go:293] "Volume detached for volume \"tobiko-private-key\" (UniqueName: \"kubernetes.io/configmap/32d4b615-70d7-4af1-9280-8fb2452feb76-tobiko-private-key\") on node \"crc\" DevicePath \"\"" Dec 13 09:36:15 crc kubenswrapper[4971]: I1213 09:36:15.053208 4971 reconciler_common.go:293] "Volume detached for volume \"test-operator-clouds-config\" (UniqueName: \"kubernetes.io/configmap/32d4b615-70d7-4af1-9280-8fb2452feb76-test-operator-clouds-config\") on node \"crc\" DevicePath \"\"" Dec 13 09:36:15 crc kubenswrapper[4971]: I1213 09:36:15.053217 4971 reconciler_common.go:293] "Volume detached for volume \"kubeconfig\" (UniqueName: \"kubernetes.io/secret/32d4b615-70d7-4af1-9280-8fb2452feb76-kubeconfig\") on node \"crc\" DevicePath \"\"" Dec 13 09:36:15 crc kubenswrapper[4971]: I1213 09:36:15.053226 4971 reconciler_common.go:293] "Volume detached for volume \"ca-certs\" (UniqueName: \"kubernetes.io/secret/32d4b615-70d7-4af1-9280-8fb2452feb76-ca-certs\") on node \"crc\" DevicePath \"\"" Dec 13 09:36:15 crc kubenswrapper[4971]: I1213 09:36:15.053234 4971 reconciler_common.go:293] "Volume detached for volume \"openstack-config-secret\" (UniqueName: \"kubernetes.io/secret/32d4b615-70d7-4af1-9280-8fb2452feb76-openstack-config-secret\") on node \"crc\" DevicePath \"\"" Dec 13 09:36:15 crc kubenswrapper[4971]: I1213 09:36:15.053246 4971 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-9x8rp\" (UniqueName: \"kubernetes.io/projected/32d4b615-70d7-4af1-9280-8fb2452feb76-kube-api-access-9x8rp\") on node \"crc\" DevicePath \"\"" Dec 13 09:36:15 crc kubenswrapper[4971]: I1213 09:36:15.053254 4971 reconciler_common.go:293] "Volume detached for volume \"tobiko-public-key\" (UniqueName: \"kubernetes.io/configmap/32d4b615-70d7-4af1-9280-8fb2452feb76-tobiko-public-key\") on node \"crc\" DevicePath \"\"" Dec 13 09:36:15 crc kubenswrapper[4971]: I1213 09:36:15.053280 4971 reconciler_common.go:286] "operationExecutor.UnmountDevice started for volume \"local-storage03-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage03-crc\") on node \"crc\" " Dec 13 09:36:15 crc kubenswrapper[4971]: I1213 09:36:15.078891 4971 operation_generator.go:917] UnmountDevice succeeded for volume "local-storage03-crc" (UniqueName: "kubernetes.io/local-volume/local-storage03-crc") on node "crc" Dec 13 09:36:15 crc kubenswrapper[4971]: I1213 09:36:15.155414 4971 reconciler_common.go:293] "Volume detached for volume \"local-storage03-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage03-crc\") on node \"crc\" DevicePath \"\"" Dec 13 09:36:15 crc kubenswrapper[4971]: I1213 09:36:15.391230 4971 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/tobiko-tests-tobiko-s01-sanity" event={"ID":"32d4b615-70d7-4af1-9280-8fb2452feb76","Type":"ContainerDied","Data":"6168ab37f128f5736f76f3c675293327f98060acd9ef8c294b6a2e1674398a7d"} Dec 13 09:36:15 crc kubenswrapper[4971]: I1213 09:36:15.391282 4971 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="6168ab37f128f5736f76f3c675293327f98060acd9ef8c294b6a2e1674398a7d" Dec 13 09:36:15 crc kubenswrapper[4971]: I1213 09:36:15.391356 4971 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/tobiko-tests-tobiko-s01-sanity" Dec 13 09:36:16 crc kubenswrapper[4971]: I1213 09:36:16.269070 4971 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/32d4b615-70d7-4af1-9280-8fb2452feb76-test-operator-ephemeral-workdir" (OuterVolumeSpecName: "test-operator-ephemeral-workdir") pod "32d4b615-70d7-4af1-9280-8fb2452feb76" (UID: "32d4b615-70d7-4af1-9280-8fb2452feb76"). InnerVolumeSpecName "test-operator-ephemeral-workdir". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 13 09:36:16 crc kubenswrapper[4971]: I1213 09:36:16.283989 4971 reconciler_common.go:293] "Volume detached for volume \"test-operator-ephemeral-workdir\" (UniqueName: \"kubernetes.io/empty-dir/32d4b615-70d7-4af1-9280-8fb2452feb76-test-operator-ephemeral-workdir\") on node \"crc\" DevicePath \"\"" Dec 13 09:36:19 crc kubenswrapper[4971]: I1213 09:36:19.770032 4971 scope.go:117] "RemoveContainer" containerID="7c80479d6aedc6e1a5cced1df9e55738d4a726b6ddd5adfe17363d9c5d097d6a" Dec 13 09:36:19 crc kubenswrapper[4971]: E1213 09:36:19.771041 4971 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-82xjz_openshift-machine-config-operator(e0c4af12-edda-41d7-9b44-f87396174ef2)\"" pod="openshift-machine-config-operator/machine-config-daemon-82xjz" podUID="e0c4af12-edda-41d7-9b44-f87396174ef2" Dec 13 09:36:26 crc kubenswrapper[4971]: I1213 09:36:26.273499 4971 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/test-operator-logs-pod-tobiko-tobiko-tests-tobiko"] Dec 13 09:36:26 crc kubenswrapper[4971]: E1213 09:36:26.276306 4971 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="db50ff88-63bb-48b5-ad4f-0a40dc6236a9" containerName="registry-server" Dec 13 09:36:26 crc kubenswrapper[4971]: I1213 09:36:26.276335 4971 state_mem.go:107] "Deleted CPUSet assignment" podUID="db50ff88-63bb-48b5-ad4f-0a40dc6236a9" containerName="registry-server" Dec 13 09:36:26 crc kubenswrapper[4971]: E1213 09:36:26.276353 4971 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="db50ff88-63bb-48b5-ad4f-0a40dc6236a9" containerName="extract-utilities" Dec 13 09:36:26 crc kubenswrapper[4971]: I1213 09:36:26.276360 4971 state_mem.go:107] "Deleted CPUSet assignment" podUID="db50ff88-63bb-48b5-ad4f-0a40dc6236a9" containerName="extract-utilities" Dec 13 09:36:26 crc kubenswrapper[4971]: E1213 09:36:26.276374 4971 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="db50ff88-63bb-48b5-ad4f-0a40dc6236a9" containerName="extract-content" Dec 13 09:36:26 crc kubenswrapper[4971]: I1213 09:36:26.276379 4971 state_mem.go:107] "Deleted CPUSet assignment" podUID="db50ff88-63bb-48b5-ad4f-0a40dc6236a9" containerName="extract-content" Dec 13 09:36:26 crc kubenswrapper[4971]: E1213 09:36:26.276409 4971 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="32d4b615-70d7-4af1-9280-8fb2452feb76" containerName="tobiko-tests-tobiko" Dec 13 09:36:26 crc kubenswrapper[4971]: I1213 09:36:26.276414 4971 state_mem.go:107] "Deleted CPUSet assignment" podUID="32d4b615-70d7-4af1-9280-8fb2452feb76" containerName="tobiko-tests-tobiko" Dec 13 09:36:26 crc kubenswrapper[4971]: I1213 09:36:26.276703 4971 memory_manager.go:354] "RemoveStaleState removing state" podUID="db50ff88-63bb-48b5-ad4f-0a40dc6236a9" containerName="registry-server" Dec 13 09:36:26 crc kubenswrapper[4971]: I1213 09:36:26.276727 4971 memory_manager.go:354] "RemoveStaleState removing state" podUID="32d4b615-70d7-4af1-9280-8fb2452feb76" containerName="tobiko-tests-tobiko" Dec 13 09:36:26 crc kubenswrapper[4971]: I1213 09:36:26.295780 4971 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/test-operator-logs-pod-tobiko-tobiko-tests-tobiko" Dec 13 09:36:26 crc kubenswrapper[4971]: I1213 09:36:26.301255 4971 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/test-operator-logs-pod-tobiko-tobiko-tests-tobiko"] Dec 13 09:36:26 crc kubenswrapper[4971]: I1213 09:36:26.435891 4971 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"local-storage03-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage03-crc\") pod \"test-operator-logs-pod-tobiko-tobiko-tests-tobiko\" (UID: \"d608d411-d575-467c-8738-17284ac309d7\") " pod="openstack/test-operator-logs-pod-tobiko-tobiko-tests-tobiko" Dec 13 09:36:26 crc kubenswrapper[4971]: I1213 09:36:26.436317 4971 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-b8v45\" (UniqueName: \"kubernetes.io/projected/d608d411-d575-467c-8738-17284ac309d7-kube-api-access-b8v45\") pod \"test-operator-logs-pod-tobiko-tobiko-tests-tobiko\" (UID: \"d608d411-d575-467c-8738-17284ac309d7\") " pod="openstack/test-operator-logs-pod-tobiko-tobiko-tests-tobiko" Dec 13 09:36:26 crc kubenswrapper[4971]: I1213 09:36:26.539676 4971 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"local-storage03-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage03-crc\") pod \"test-operator-logs-pod-tobiko-tobiko-tests-tobiko\" (UID: \"d608d411-d575-467c-8738-17284ac309d7\") " pod="openstack/test-operator-logs-pod-tobiko-tobiko-tests-tobiko" Dec 13 09:36:26 crc kubenswrapper[4971]: I1213 09:36:26.539737 4971 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-b8v45\" (UniqueName: \"kubernetes.io/projected/d608d411-d575-467c-8738-17284ac309d7-kube-api-access-b8v45\") pod \"test-operator-logs-pod-tobiko-tobiko-tests-tobiko\" (UID: \"d608d411-d575-467c-8738-17284ac309d7\") " pod="openstack/test-operator-logs-pod-tobiko-tobiko-tests-tobiko" Dec 13 09:36:26 crc kubenswrapper[4971]: I1213 09:36:26.543250 4971 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"local-storage03-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage03-crc\") pod \"test-operator-logs-pod-tobiko-tobiko-tests-tobiko\" (UID: \"d608d411-d575-467c-8738-17284ac309d7\") device mount path \"/mnt/openstack/pv03\"" pod="openstack/test-operator-logs-pod-tobiko-tobiko-tests-tobiko" Dec 13 09:36:26 crc kubenswrapper[4971]: I1213 09:36:26.576915 4971 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"local-storage03-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage03-crc\") pod \"test-operator-logs-pod-tobiko-tobiko-tests-tobiko\" (UID: \"d608d411-d575-467c-8738-17284ac309d7\") " pod="openstack/test-operator-logs-pod-tobiko-tobiko-tests-tobiko" Dec 13 09:36:26 crc kubenswrapper[4971]: I1213 09:36:26.584366 4971 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-b8v45\" (UniqueName: \"kubernetes.io/projected/d608d411-d575-467c-8738-17284ac309d7-kube-api-access-b8v45\") pod \"test-operator-logs-pod-tobiko-tobiko-tests-tobiko\" (UID: \"d608d411-d575-467c-8738-17284ac309d7\") " pod="openstack/test-operator-logs-pod-tobiko-tobiko-tests-tobiko" Dec 13 09:36:26 crc kubenswrapper[4971]: I1213 09:36:26.627707 4971 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/test-operator-logs-pod-tobiko-tobiko-tests-tobiko" Dec 13 09:36:27 crc kubenswrapper[4971]: I1213 09:36:27.144819 4971 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/test-operator-logs-pod-tobiko-tobiko-tests-tobiko"] Dec 13 09:36:27 crc kubenswrapper[4971]: I1213 09:36:27.550257 4971 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/test-operator-logs-pod-tobiko-tobiko-tests-tobiko" event={"ID":"d608d411-d575-467c-8738-17284ac309d7","Type":"ContainerStarted","Data":"e91c00ecf30f5dfa9c4c0a802aff876773828f7c8d31e95703ed9f543bb5e611"} Dec 13 09:36:28 crc kubenswrapper[4971]: I1213 09:36:28.563500 4971 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/test-operator-logs-pod-tobiko-tobiko-tests-tobiko" event={"ID":"d608d411-d575-467c-8738-17284ac309d7","Type":"ContainerStarted","Data":"c04b03769fbe6413f2d108dc77c528bcb4174bfea92dbad05a5bb570ce94cef2"} Dec 13 09:36:28 crc kubenswrapper[4971]: I1213 09:36:28.580612 4971 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/test-operator-logs-pod-tobiko-tobiko-tests-tobiko" podStartSLOduration=2.01270953 podStartE2EDuration="2.580589728s" podCreationTimestamp="2025-12-13 09:36:26 +0000 UTC" firstStartedPulling="2025-12-13 09:36:27.13941235 +0000 UTC m=+10043.743821788" lastFinishedPulling="2025-12-13 09:36:27.707292538 +0000 UTC m=+10044.311701986" observedRunningTime="2025-12-13 09:36:28.57902376 +0000 UTC m=+10045.183433228" watchObservedRunningTime="2025-12-13 09:36:28.580589728 +0000 UTC m=+10045.184999176" Dec 13 09:36:30 crc kubenswrapper[4971]: I1213 09:36:30.082151 4971 scope.go:117] "RemoveContainer" containerID="cc86a53e32c2914bf906f356959a530cbefff1c9bf57ff202deb3fc5d945791c" Dec 13 09:36:30 crc kubenswrapper[4971]: I1213 09:36:30.112004 4971 scope.go:117] "RemoveContainer" containerID="b4b20fa697a6ae02101018f66af436b00c15c45e779ad7b669b6c27b3bfdb99e" Dec 13 09:36:30 crc kubenswrapper[4971]: I1213 09:36:30.161210 4971 scope.go:117] "RemoveContainer" containerID="4a11761b1f295ff6eabacef438bd250e6f30cf74eb29cf762135f04ca49acb00" Dec 13 09:36:31 crc kubenswrapper[4971]: I1213 09:36:31.769646 4971 scope.go:117] "RemoveContainer" containerID="7c80479d6aedc6e1a5cced1df9e55738d4a726b6ddd5adfe17363d9c5d097d6a" Dec 13 09:36:31 crc kubenswrapper[4971]: E1213 09:36:31.770436 4971 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-82xjz_openshift-machine-config-operator(e0c4af12-edda-41d7-9b44-f87396174ef2)\"" pod="openshift-machine-config-operator/machine-config-daemon-82xjz" podUID="e0c4af12-edda-41d7-9b44-f87396174ef2" Dec 13 09:36:41 crc kubenswrapper[4971]: I1213 09:36:41.578889 4971 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ansibletest-ansibletest"] Dec 13 09:36:41 crc kubenswrapper[4971]: I1213 09:36:41.581584 4971 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ansibletest-ansibletest" Dec 13 09:36:41 crc kubenswrapper[4971]: I1213 09:36:41.584896 4971 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplane-ansible-ssh-private-key-secret" Dec 13 09:36:41 crc kubenswrapper[4971]: I1213 09:36:41.584927 4971 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"test-operator-controller-priv-key" Dec 13 09:36:41 crc kubenswrapper[4971]: I1213 09:36:41.615773 4971 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ansibletest-ansibletest"] Dec 13 09:36:41 crc kubenswrapper[4971]: I1213 09:36:41.709954 4971 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/a020591e-029f-464f-8d6c-947acba2bf13-ceph\") pod \"ansibletest-ansibletest\" (UID: \"a020591e-029f-464f-8d6c-947acba2bf13\") " pod="openstack/ansibletest-ansibletest" Dec 13 09:36:41 crc kubenswrapper[4971]: I1213 09:36:41.710014 4971 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"test-operator-ephemeral-temporary\" (UniqueName: \"kubernetes.io/empty-dir/a020591e-029f-464f-8d6c-947acba2bf13-test-operator-ephemeral-temporary\") pod \"ansibletest-ansibletest\" (UID: \"a020591e-029f-464f-8d6c-947acba2bf13\") " pod="openstack/ansibletest-ansibletest" Dec 13 09:36:41 crc kubenswrapper[4971]: I1213 09:36:41.710040 4971 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-bnkck\" (UniqueName: \"kubernetes.io/projected/a020591e-029f-464f-8d6c-947acba2bf13-kube-api-access-bnkck\") pod \"ansibletest-ansibletest\" (UID: \"a020591e-029f-464f-8d6c-947acba2bf13\") " pod="openstack/ansibletest-ansibletest" Dec 13 09:36:41 crc kubenswrapper[4971]: I1213 09:36:41.710208 4971 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"test-operator-ephemeral-workdir\" (UniqueName: \"kubernetes.io/empty-dir/a020591e-029f-464f-8d6c-947acba2bf13-test-operator-ephemeral-workdir\") pod \"ansibletest-ansibletest\" (UID: \"a020591e-029f-464f-8d6c-947acba2bf13\") " pod="openstack/ansibletest-ansibletest" Dec 13 09:36:41 crc kubenswrapper[4971]: I1213 09:36:41.710310 4971 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"openstack-config-secret\" (UniqueName: \"kubernetes.io/secret/a020591e-029f-464f-8d6c-947acba2bf13-openstack-config-secret\") pod \"ansibletest-ansibletest\" (UID: \"a020591e-029f-464f-8d6c-947acba2bf13\") " pod="openstack/ansibletest-ansibletest" Dec 13 09:36:41 crc kubenswrapper[4971]: I1213 09:36:41.710379 4971 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"compute-ssh-secret\" (UniqueName: \"kubernetes.io/secret/a020591e-029f-464f-8d6c-947acba2bf13-compute-ssh-secret\") pod \"ansibletest-ansibletest\" (UID: \"a020591e-029f-464f-8d6c-947acba2bf13\") " pod="openstack/ansibletest-ansibletest" Dec 13 09:36:41 crc kubenswrapper[4971]: I1213 09:36:41.710698 4971 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"workload-ssh-secret\" (UniqueName: \"kubernetes.io/secret/a020591e-029f-464f-8d6c-947acba2bf13-workload-ssh-secret\") pod \"ansibletest-ansibletest\" (UID: \"a020591e-029f-464f-8d6c-947acba2bf13\") " pod="openstack/ansibletest-ansibletest" Dec 13 09:36:41 crc kubenswrapper[4971]: I1213 09:36:41.710795 4971 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ca-certs\" (UniqueName: \"kubernetes.io/secret/a020591e-029f-464f-8d6c-947acba2bf13-ca-certs\") pod \"ansibletest-ansibletest\" (UID: \"a020591e-029f-464f-8d6c-947acba2bf13\") " pod="openstack/ansibletest-ansibletest" Dec 13 09:36:41 crc kubenswrapper[4971]: I1213 09:36:41.710982 4971 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"local-storage04-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage04-crc\") pod \"ansibletest-ansibletest\" (UID: \"a020591e-029f-464f-8d6c-947acba2bf13\") " pod="openstack/ansibletest-ansibletest" Dec 13 09:36:41 crc kubenswrapper[4971]: I1213 09:36:41.711027 4971 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"openstack-config\" (UniqueName: \"kubernetes.io/configmap/a020591e-029f-464f-8d6c-947acba2bf13-openstack-config\") pod \"ansibletest-ansibletest\" (UID: \"a020591e-029f-464f-8d6c-947acba2bf13\") " pod="openstack/ansibletest-ansibletest" Dec 13 09:36:41 crc kubenswrapper[4971]: I1213 09:36:41.813060 4971 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"workload-ssh-secret\" (UniqueName: \"kubernetes.io/secret/a020591e-029f-464f-8d6c-947acba2bf13-workload-ssh-secret\") pod \"ansibletest-ansibletest\" (UID: \"a020591e-029f-464f-8d6c-947acba2bf13\") " pod="openstack/ansibletest-ansibletest" Dec 13 09:36:41 crc kubenswrapper[4971]: I1213 09:36:41.813496 4971 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ca-certs\" (UniqueName: \"kubernetes.io/secret/a020591e-029f-464f-8d6c-947acba2bf13-ca-certs\") pod \"ansibletest-ansibletest\" (UID: \"a020591e-029f-464f-8d6c-947acba2bf13\") " pod="openstack/ansibletest-ansibletest" Dec 13 09:36:41 crc kubenswrapper[4971]: I1213 09:36:41.813683 4971 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"local-storage04-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage04-crc\") pod \"ansibletest-ansibletest\" (UID: \"a020591e-029f-464f-8d6c-947acba2bf13\") " pod="openstack/ansibletest-ansibletest" Dec 13 09:36:41 crc kubenswrapper[4971]: I1213 09:36:41.813790 4971 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"openstack-config\" (UniqueName: \"kubernetes.io/configmap/a020591e-029f-464f-8d6c-947acba2bf13-openstack-config\") pod \"ansibletest-ansibletest\" (UID: \"a020591e-029f-464f-8d6c-947acba2bf13\") " pod="openstack/ansibletest-ansibletest" Dec 13 09:36:41 crc kubenswrapper[4971]: I1213 09:36:41.813967 4971 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/a020591e-029f-464f-8d6c-947acba2bf13-ceph\") pod \"ansibletest-ansibletest\" (UID: \"a020591e-029f-464f-8d6c-947acba2bf13\") " pod="openstack/ansibletest-ansibletest" Dec 13 09:36:41 crc kubenswrapper[4971]: I1213 09:36:41.814411 4971 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"test-operator-ephemeral-temporary\" (UniqueName: \"kubernetes.io/empty-dir/a020591e-029f-464f-8d6c-947acba2bf13-test-operator-ephemeral-temporary\") pod \"ansibletest-ansibletest\" (UID: \"a020591e-029f-464f-8d6c-947acba2bf13\") " pod="openstack/ansibletest-ansibletest" Dec 13 09:36:41 crc kubenswrapper[4971]: I1213 09:36:41.814546 4971 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-bnkck\" (UniqueName: \"kubernetes.io/projected/a020591e-029f-464f-8d6c-947acba2bf13-kube-api-access-bnkck\") pod \"ansibletest-ansibletest\" (UID: \"a020591e-029f-464f-8d6c-947acba2bf13\") " pod="openstack/ansibletest-ansibletest" Dec 13 09:36:41 crc kubenswrapper[4971]: I1213 09:36:41.814632 4971 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"local-storage04-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage04-crc\") pod \"ansibletest-ansibletest\" (UID: \"a020591e-029f-464f-8d6c-947acba2bf13\") device mount path \"/mnt/openstack/pv04\"" pod="openstack/ansibletest-ansibletest" Dec 13 09:36:41 crc kubenswrapper[4971]: I1213 09:36:41.814752 4971 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"test-operator-ephemeral-workdir\" (UniqueName: \"kubernetes.io/empty-dir/a020591e-029f-464f-8d6c-947acba2bf13-test-operator-ephemeral-workdir\") pod \"ansibletest-ansibletest\" (UID: \"a020591e-029f-464f-8d6c-947acba2bf13\") " pod="openstack/ansibletest-ansibletest" Dec 13 09:36:41 crc kubenswrapper[4971]: I1213 09:36:41.814831 4971 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"test-operator-ephemeral-temporary\" (UniqueName: \"kubernetes.io/empty-dir/a020591e-029f-464f-8d6c-947acba2bf13-test-operator-ephemeral-temporary\") pod \"ansibletest-ansibletest\" (UID: \"a020591e-029f-464f-8d6c-947acba2bf13\") " pod="openstack/ansibletest-ansibletest" Dec 13 09:36:41 crc kubenswrapper[4971]: I1213 09:36:41.814958 4971 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"openstack-config-secret\" (UniqueName: \"kubernetes.io/secret/a020591e-029f-464f-8d6c-947acba2bf13-openstack-config-secret\") pod \"ansibletest-ansibletest\" (UID: \"a020591e-029f-464f-8d6c-947acba2bf13\") " pod="openstack/ansibletest-ansibletest" Dec 13 09:36:41 crc kubenswrapper[4971]: I1213 09:36:41.815047 4971 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"openstack-config\" (UniqueName: \"kubernetes.io/configmap/a020591e-029f-464f-8d6c-947acba2bf13-openstack-config\") pod \"ansibletest-ansibletest\" (UID: \"a020591e-029f-464f-8d6c-947acba2bf13\") " pod="openstack/ansibletest-ansibletest" Dec 13 09:36:41 crc kubenswrapper[4971]: I1213 09:36:41.815196 4971 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"compute-ssh-secret\" (UniqueName: \"kubernetes.io/secret/a020591e-029f-464f-8d6c-947acba2bf13-compute-ssh-secret\") pod \"ansibletest-ansibletest\" (UID: \"a020591e-029f-464f-8d6c-947acba2bf13\") " pod="openstack/ansibletest-ansibletest" Dec 13 09:36:41 crc kubenswrapper[4971]: I1213 09:36:41.815573 4971 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"test-operator-ephemeral-workdir\" (UniqueName: \"kubernetes.io/empty-dir/a020591e-029f-464f-8d6c-947acba2bf13-test-operator-ephemeral-workdir\") pod \"ansibletest-ansibletest\" (UID: \"a020591e-029f-464f-8d6c-947acba2bf13\") " pod="openstack/ansibletest-ansibletest" Dec 13 09:36:41 crc kubenswrapper[4971]: I1213 09:36:41.822126 4971 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/a020591e-029f-464f-8d6c-947acba2bf13-ceph\") pod \"ansibletest-ansibletest\" (UID: \"a020591e-029f-464f-8d6c-947acba2bf13\") " pod="openstack/ansibletest-ansibletest" Dec 13 09:36:41 crc kubenswrapper[4971]: I1213 09:36:41.822145 4971 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ca-certs\" (UniqueName: \"kubernetes.io/secret/a020591e-029f-464f-8d6c-947acba2bf13-ca-certs\") pod \"ansibletest-ansibletest\" (UID: \"a020591e-029f-464f-8d6c-947acba2bf13\") " pod="openstack/ansibletest-ansibletest" Dec 13 09:36:41 crc kubenswrapper[4971]: I1213 09:36:41.822551 4971 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"compute-ssh-secret\" (UniqueName: \"kubernetes.io/secret/a020591e-029f-464f-8d6c-947acba2bf13-compute-ssh-secret\") pod \"ansibletest-ansibletest\" (UID: \"a020591e-029f-464f-8d6c-947acba2bf13\") " pod="openstack/ansibletest-ansibletest" Dec 13 09:36:41 crc kubenswrapper[4971]: I1213 09:36:41.823340 4971 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"openstack-config-secret\" (UniqueName: \"kubernetes.io/secret/a020591e-029f-464f-8d6c-947acba2bf13-openstack-config-secret\") pod \"ansibletest-ansibletest\" (UID: \"a020591e-029f-464f-8d6c-947acba2bf13\") " pod="openstack/ansibletest-ansibletest" Dec 13 09:36:41 crc kubenswrapper[4971]: I1213 09:36:41.827010 4971 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"workload-ssh-secret\" (UniqueName: \"kubernetes.io/secret/a020591e-029f-464f-8d6c-947acba2bf13-workload-ssh-secret\") pod \"ansibletest-ansibletest\" (UID: \"a020591e-029f-464f-8d6c-947acba2bf13\") " pod="openstack/ansibletest-ansibletest" Dec 13 09:36:41 crc kubenswrapper[4971]: I1213 09:36:41.839768 4971 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-bnkck\" (UniqueName: \"kubernetes.io/projected/a020591e-029f-464f-8d6c-947acba2bf13-kube-api-access-bnkck\") pod \"ansibletest-ansibletest\" (UID: \"a020591e-029f-464f-8d6c-947acba2bf13\") " pod="openstack/ansibletest-ansibletest" Dec 13 09:36:41 crc kubenswrapper[4971]: I1213 09:36:41.844318 4971 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"local-storage04-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage04-crc\") pod \"ansibletest-ansibletest\" (UID: \"a020591e-029f-464f-8d6c-947acba2bf13\") " pod="openstack/ansibletest-ansibletest" Dec 13 09:36:41 crc kubenswrapper[4971]: I1213 09:36:41.912044 4971 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ansibletest-ansibletest" Dec 13 09:36:42 crc kubenswrapper[4971]: I1213 09:36:42.368572 4971 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ansibletest-ansibletest"] Dec 13 09:36:42 crc kubenswrapper[4971]: I1213 09:36:42.697982 4971 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ansibletest-ansibletest" event={"ID":"a020591e-029f-464f-8d6c-947acba2bf13","Type":"ContainerStarted","Data":"1406752040d8a0316e85e47c78a3181cefbdb6f371fd4a179ba913a20d6620e2"} Dec 13 09:36:46 crc kubenswrapper[4971]: I1213 09:36:46.769456 4971 scope.go:117] "RemoveContainer" containerID="7c80479d6aedc6e1a5cced1df9e55738d4a726b6ddd5adfe17363d9c5d097d6a" Dec 13 09:36:46 crc kubenswrapper[4971]: E1213 09:36:46.770809 4971 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-82xjz_openshift-machine-config-operator(e0c4af12-edda-41d7-9b44-f87396174ef2)\"" pod="openshift-machine-config-operator/machine-config-daemon-82xjz" podUID="e0c4af12-edda-41d7-9b44-f87396174ef2" Dec 13 09:37:01 crc kubenswrapper[4971]: I1213 09:37:01.768443 4971 scope.go:117] "RemoveContainer" containerID="7c80479d6aedc6e1a5cced1df9e55738d4a726b6ddd5adfe17363d9c5d097d6a" Dec 13 09:37:01 crc kubenswrapper[4971]: E1213 09:37:01.769422 4971 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-82xjz_openshift-machine-config-operator(e0c4af12-edda-41d7-9b44-f87396174ef2)\"" pod="openshift-machine-config-operator/machine-config-daemon-82xjz" podUID="e0c4af12-edda-41d7-9b44-f87396174ef2" Dec 13 09:37:03 crc kubenswrapper[4971]: E1213 09:37:03.179288 4971 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying config: context canceled" image="quay.io/podified-antelope-centos9/openstack-ansible-tests:current-podified" Dec 13 09:37:03 crc kubenswrapper[4971]: E1213 09:37:03.179954 4971 kuberuntime_manager.go:1274] "Unhandled Error" err=< Dec 13 09:37:03 crc kubenswrapper[4971]: container &Container{Name:ansibletest-ansibletest,Image:quay.io/podified-antelope-centos9/openstack-ansible-tests:current-podified,Command:[],Args:[],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:POD_ANSIBLE_EXTRA_VARS,Value:-e manual_run=false,ValueFrom:nil,},EnvVar{Name:POD_ANSIBLE_FILE_EXTRA_VARS,Value:--- Dec 13 09:37:03 crc kubenswrapper[4971]: foo: bar Dec 13 09:37:03 crc kubenswrapper[4971]: ,ValueFrom:nil,},EnvVar{Name:POD_ANSIBLE_GIT_BRANCH,Value:,ValueFrom:nil,},EnvVar{Name:POD_ANSIBLE_GIT_REPO,Value:https://github.com/ansible/test-playbooks,ValueFrom:nil,},EnvVar{Name:POD_ANSIBLE_INVENTORY,Value:localhost ansible_connection=local ansible_python_interpreter=python3 Dec 13 09:37:03 crc kubenswrapper[4971]: ,ValueFrom:nil,},EnvVar{Name:POD_ANSIBLE_PLAYBOOK,Value:./debug.yml,ValueFrom:nil,},EnvVar{Name:POD_DEBUG,Value:false,ValueFrom:nil,},EnvVar{Name:POD_INSTALL_COLLECTIONS,Value:,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{cpu: {{4 0} {} 4 DecimalSI},memory: {{4294967296 0} {} 4Gi BinarySI},},Requests:ResourceList{cpu: {{2 0} {} 2 DecimalSI},memory: {{2147483648 0} {} 2Gi BinarySI},},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:test-operator-ephemeral-workdir,ReadOnly:false,MountPath:/var/lib/ansible,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:test-operator-ephemeral-temporary,ReadOnly:false,MountPath:/tmp,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:test-operator-logs,ReadOnly:false,MountPath:/var/lib/AnsibleTests/external_files,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:openstack-config,ReadOnly:true,MountPath:/etc/openstack/clouds.yaml,SubPath:clouds.yaml,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:openstack-config,ReadOnly:true,MountPath:/var/lib/ansible/.config/openstack/clouds.yaml,SubPath:clouds.yaml,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:openstack-config-secret,ReadOnly:false,MountPath:/var/lib/ansible/.config/openstack/secure.yaml,SubPath:secure.yaml,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:ca-certs,ReadOnly:true,MountPath:/etc/pki/ca-trust/extracted/pem/tls-ca-bundle.pem,SubPath:tls-ca-bundle.pem,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:ca-certs,ReadOnly:true,MountPath:/etc/pki/tls/certs/ca-bundle.trust.crt,SubPath:tls-ca-bundle.pem,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:compute-ssh-secret,ReadOnly:true,MountPath:/var/lib/ansible/.ssh/compute_id,SubPath:ssh-privatekey,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:workload-ssh-secret,ReadOnly:true,MountPath:/var/lib/ansible/test_keypair.key,SubPath:ssh-privatekey,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:ceph,ReadOnly:true,MountPath:/etc/ceph,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-bnkck,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[NET_ADMIN NET_RAW],Drop:[],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*227,RunAsNonRoot:*false,ReadOnlyRootFilesystem:*false,AllowPrivilegeEscalation:*true,RunAsGroup:*227,ProcMount:nil,WindowsOptions:nil,SeccompProfile:&SeccompProfile{Type:RuntimeDefault,LocalhostProfile:nil,},AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod ansibletest-ansibletest_openstack(a020591e-029f-464f-8d6c-947acba2bf13): ErrImagePull: rpc error: code = Canceled desc = copying config: context canceled Dec 13 09:37:03 crc kubenswrapper[4971]: > logger="UnhandledError" Dec 13 09:37:03 crc kubenswrapper[4971]: E1213 09:37:03.181126 4971 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"ansibletest-ansibletest\" with ErrImagePull: \"rpc error: code = Canceled desc = copying config: context canceled\"" pod="openstack/ansibletest-ansibletest" podUID="a020591e-029f-464f-8d6c-947acba2bf13" Dec 13 09:37:03 crc kubenswrapper[4971]: E1213 09:37:03.932267 4971 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"ansibletest-ansibletest\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/podified-antelope-centos9/openstack-ansible-tests:current-podified\\\"\"" pod="openstack/ansibletest-ansibletest" podUID="a020591e-029f-464f-8d6c-947acba2bf13" Dec 13 09:37:15 crc kubenswrapper[4971]: I1213 09:37:15.769590 4971 scope.go:117] "RemoveContainer" containerID="7c80479d6aedc6e1a5cced1df9e55738d4a726b6ddd5adfe17363d9c5d097d6a" Dec 13 09:37:15 crc kubenswrapper[4971]: E1213 09:37:15.771843 4971 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-82xjz_openshift-machine-config-operator(e0c4af12-edda-41d7-9b44-f87396174ef2)\"" pod="openshift-machine-config-operator/machine-config-daemon-82xjz" podUID="e0c4af12-edda-41d7-9b44-f87396174ef2" Dec 13 09:37:17 crc kubenswrapper[4971]: I1213 09:37:17.772742 4971 provider.go:102] Refreshing cache for provider: *credentialprovider.defaultDockerConfigProvider Dec 13 09:37:21 crc kubenswrapper[4971]: I1213 09:37:21.112997 4971 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ansibletest-ansibletest" event={"ID":"a020591e-029f-464f-8d6c-947acba2bf13","Type":"ContainerStarted","Data":"bacd73d5ac85a6239bded1aadf72f77c20ec5206cff800813e04a7e1b69bacbe"} Dec 13 09:37:23 crc kubenswrapper[4971]: I1213 09:37:23.135569 4971 generic.go:334] "Generic (PLEG): container finished" podID="a020591e-029f-464f-8d6c-947acba2bf13" containerID="bacd73d5ac85a6239bded1aadf72f77c20ec5206cff800813e04a7e1b69bacbe" exitCode=0 Dec 13 09:37:23 crc kubenswrapper[4971]: I1213 09:37:23.135689 4971 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ansibletest-ansibletest" event={"ID":"a020591e-029f-464f-8d6c-947acba2bf13","Type":"ContainerDied","Data":"bacd73d5ac85a6239bded1aadf72f77c20ec5206cff800813e04a7e1b69bacbe"} Dec 13 09:37:25 crc kubenswrapper[4971]: I1213 09:37:25.147883 4971 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ansibletest-ansibletest" Dec 13 09:37:25 crc kubenswrapper[4971]: I1213 09:37:25.163809 4971 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ansibletest-ansibletest" event={"ID":"a020591e-029f-464f-8d6c-947acba2bf13","Type":"ContainerDied","Data":"1406752040d8a0316e85e47c78a3181cefbdb6f371fd4a179ba913a20d6620e2"} Dec 13 09:37:25 crc kubenswrapper[4971]: I1213 09:37:25.163869 4971 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="1406752040d8a0316e85e47c78a3181cefbdb6f371fd4a179ba913a20d6620e2" Dec 13 09:37:25 crc kubenswrapper[4971]: I1213 09:37:25.163911 4971 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ansibletest-ansibletest" Dec 13 09:37:25 crc kubenswrapper[4971]: I1213 09:37:25.293892 4971 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ca-certs\" (UniqueName: \"kubernetes.io/secret/a020591e-029f-464f-8d6c-947acba2bf13-ca-certs\") pod \"a020591e-029f-464f-8d6c-947acba2bf13\" (UID: \"a020591e-029f-464f-8d6c-947acba2bf13\") " Dec 13 09:37:25 crc kubenswrapper[4971]: I1213 09:37:25.293979 4971 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-bnkck\" (UniqueName: \"kubernetes.io/projected/a020591e-029f-464f-8d6c-947acba2bf13-kube-api-access-bnkck\") pod \"a020591e-029f-464f-8d6c-947acba2bf13\" (UID: \"a020591e-029f-464f-8d6c-947acba2bf13\") " Dec 13 09:37:25 crc kubenswrapper[4971]: I1213 09:37:25.294134 4971 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"workload-ssh-secret\" (UniqueName: \"kubernetes.io/secret/a020591e-029f-464f-8d6c-947acba2bf13-workload-ssh-secret\") pod \"a020591e-029f-464f-8d6c-947acba2bf13\" (UID: \"a020591e-029f-464f-8d6c-947acba2bf13\") " Dec 13 09:37:25 crc kubenswrapper[4971]: I1213 09:37:25.294305 4971 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/a020591e-029f-464f-8d6c-947acba2bf13-ceph\") pod \"a020591e-029f-464f-8d6c-947acba2bf13\" (UID: \"a020591e-029f-464f-8d6c-947acba2bf13\") " Dec 13 09:37:25 crc kubenswrapper[4971]: I1213 09:37:25.294809 4971 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"test-operator-logs\" (UniqueName: \"kubernetes.io/local-volume/local-storage04-crc\") pod \"a020591e-029f-464f-8d6c-947acba2bf13\" (UID: \"a020591e-029f-464f-8d6c-947acba2bf13\") " Dec 13 09:37:25 crc kubenswrapper[4971]: I1213 09:37:25.294932 4971 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"compute-ssh-secret\" (UniqueName: \"kubernetes.io/secret/a020591e-029f-464f-8d6c-947acba2bf13-compute-ssh-secret\") pod \"a020591e-029f-464f-8d6c-947acba2bf13\" (UID: \"a020591e-029f-464f-8d6c-947acba2bf13\") " Dec 13 09:37:25 crc kubenswrapper[4971]: I1213 09:37:25.295087 4971 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"openstack-config-secret\" (UniqueName: \"kubernetes.io/secret/a020591e-029f-464f-8d6c-947acba2bf13-openstack-config-secret\") pod \"a020591e-029f-464f-8d6c-947acba2bf13\" (UID: \"a020591e-029f-464f-8d6c-947acba2bf13\") " Dec 13 09:37:25 crc kubenswrapper[4971]: I1213 09:37:25.295297 4971 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"test-operator-ephemeral-workdir\" (UniqueName: \"kubernetes.io/empty-dir/a020591e-029f-464f-8d6c-947acba2bf13-test-operator-ephemeral-workdir\") pod \"a020591e-029f-464f-8d6c-947acba2bf13\" (UID: \"a020591e-029f-464f-8d6c-947acba2bf13\") " Dec 13 09:37:25 crc kubenswrapper[4971]: I1213 09:37:25.295471 4971 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"openstack-config\" (UniqueName: \"kubernetes.io/configmap/a020591e-029f-464f-8d6c-947acba2bf13-openstack-config\") pod \"a020591e-029f-464f-8d6c-947acba2bf13\" (UID: \"a020591e-029f-464f-8d6c-947acba2bf13\") " Dec 13 09:37:25 crc kubenswrapper[4971]: I1213 09:37:25.295566 4971 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"test-operator-ephemeral-temporary\" (UniqueName: \"kubernetes.io/empty-dir/a020591e-029f-464f-8d6c-947acba2bf13-test-operator-ephemeral-temporary\") pod \"a020591e-029f-464f-8d6c-947acba2bf13\" (UID: \"a020591e-029f-464f-8d6c-947acba2bf13\") " Dec 13 09:37:25 crc kubenswrapper[4971]: I1213 09:37:25.297014 4971 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/a020591e-029f-464f-8d6c-947acba2bf13-test-operator-ephemeral-temporary" (OuterVolumeSpecName: "test-operator-ephemeral-temporary") pod "a020591e-029f-464f-8d6c-947acba2bf13" (UID: "a020591e-029f-464f-8d6c-947acba2bf13"). InnerVolumeSpecName "test-operator-ephemeral-temporary". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 13 09:37:25 crc kubenswrapper[4971]: I1213 09:37:25.303139 4971 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/a020591e-029f-464f-8d6c-947acba2bf13-kube-api-access-bnkck" (OuterVolumeSpecName: "kube-api-access-bnkck") pod "a020591e-029f-464f-8d6c-947acba2bf13" (UID: "a020591e-029f-464f-8d6c-947acba2bf13"). InnerVolumeSpecName "kube-api-access-bnkck". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 13 09:37:25 crc kubenswrapper[4971]: I1213 09:37:25.314138 4971 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/a020591e-029f-464f-8d6c-947acba2bf13-ceph" (OuterVolumeSpecName: "ceph") pod "a020591e-029f-464f-8d6c-947acba2bf13" (UID: "a020591e-029f-464f-8d6c-947acba2bf13"). InnerVolumeSpecName "ceph". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 13 09:37:25 crc kubenswrapper[4971]: I1213 09:37:25.314717 4971 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/a020591e-029f-464f-8d6c-947acba2bf13-test-operator-ephemeral-workdir" (OuterVolumeSpecName: "test-operator-ephemeral-workdir") pod "a020591e-029f-464f-8d6c-947acba2bf13" (UID: "a020591e-029f-464f-8d6c-947acba2bf13"). InnerVolumeSpecName "test-operator-ephemeral-workdir". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 13 09:37:25 crc kubenswrapper[4971]: I1213 09:37:25.321970 4971 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/local-volume/local-storage04-crc" (OuterVolumeSpecName: "test-operator-logs") pod "a020591e-029f-464f-8d6c-947acba2bf13" (UID: "a020591e-029f-464f-8d6c-947acba2bf13"). InnerVolumeSpecName "local-storage04-crc". PluginName "kubernetes.io/local-volume", VolumeGidValue "" Dec 13 09:37:25 crc kubenswrapper[4971]: I1213 09:37:25.329954 4971 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/a020591e-029f-464f-8d6c-947acba2bf13-workload-ssh-secret" (OuterVolumeSpecName: "workload-ssh-secret") pod "a020591e-029f-464f-8d6c-947acba2bf13" (UID: "a020591e-029f-464f-8d6c-947acba2bf13"). InnerVolumeSpecName "workload-ssh-secret". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 13 09:37:25 crc kubenswrapper[4971]: I1213 09:37:25.330681 4971 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/a020591e-029f-464f-8d6c-947acba2bf13-openstack-config-secret" (OuterVolumeSpecName: "openstack-config-secret") pod "a020591e-029f-464f-8d6c-947acba2bf13" (UID: "a020591e-029f-464f-8d6c-947acba2bf13"). InnerVolumeSpecName "openstack-config-secret". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 13 09:37:25 crc kubenswrapper[4971]: I1213 09:37:25.331114 4971 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/a020591e-029f-464f-8d6c-947acba2bf13-compute-ssh-secret" (OuterVolumeSpecName: "compute-ssh-secret") pod "a020591e-029f-464f-8d6c-947acba2bf13" (UID: "a020591e-029f-464f-8d6c-947acba2bf13"). InnerVolumeSpecName "compute-ssh-secret". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 13 09:37:25 crc kubenswrapper[4971]: I1213 09:37:25.355150 4971 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/a020591e-029f-464f-8d6c-947acba2bf13-ca-certs" (OuterVolumeSpecName: "ca-certs") pod "a020591e-029f-464f-8d6c-947acba2bf13" (UID: "a020591e-029f-464f-8d6c-947acba2bf13"). InnerVolumeSpecName "ca-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 13 09:37:25 crc kubenswrapper[4971]: I1213 09:37:25.364403 4971 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/a020591e-029f-464f-8d6c-947acba2bf13-openstack-config" (OuterVolumeSpecName: "openstack-config") pod "a020591e-029f-464f-8d6c-947acba2bf13" (UID: "a020591e-029f-464f-8d6c-947acba2bf13"). InnerVolumeSpecName "openstack-config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 13 09:37:25 crc kubenswrapper[4971]: I1213 09:37:25.402802 4971 reconciler_common.go:293] "Volume detached for volume \"openstack-config-secret\" (UniqueName: \"kubernetes.io/secret/a020591e-029f-464f-8d6c-947acba2bf13-openstack-config-secret\") on node \"crc\" DevicePath \"\"" Dec 13 09:37:25 crc kubenswrapper[4971]: I1213 09:37:25.402868 4971 reconciler_common.go:293] "Volume detached for volume \"test-operator-ephemeral-workdir\" (UniqueName: \"kubernetes.io/empty-dir/a020591e-029f-464f-8d6c-947acba2bf13-test-operator-ephemeral-workdir\") on node \"crc\" DevicePath \"\"" Dec 13 09:37:25 crc kubenswrapper[4971]: I1213 09:37:25.402886 4971 reconciler_common.go:293] "Volume detached for volume \"openstack-config\" (UniqueName: \"kubernetes.io/configmap/a020591e-029f-464f-8d6c-947acba2bf13-openstack-config\") on node \"crc\" DevicePath \"\"" Dec 13 09:37:25 crc kubenswrapper[4971]: I1213 09:37:25.402905 4971 reconciler_common.go:293] "Volume detached for volume \"test-operator-ephemeral-temporary\" (UniqueName: \"kubernetes.io/empty-dir/a020591e-029f-464f-8d6c-947acba2bf13-test-operator-ephemeral-temporary\") on node \"crc\" DevicePath \"\"" Dec 13 09:37:25 crc kubenswrapper[4971]: I1213 09:37:25.402920 4971 reconciler_common.go:293] "Volume detached for volume \"ca-certs\" (UniqueName: \"kubernetes.io/secret/a020591e-029f-464f-8d6c-947acba2bf13-ca-certs\") on node \"crc\" DevicePath \"\"" Dec 13 09:37:25 crc kubenswrapper[4971]: I1213 09:37:25.402931 4971 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-bnkck\" (UniqueName: \"kubernetes.io/projected/a020591e-029f-464f-8d6c-947acba2bf13-kube-api-access-bnkck\") on node \"crc\" DevicePath \"\"" Dec 13 09:37:25 crc kubenswrapper[4971]: I1213 09:37:25.402944 4971 reconciler_common.go:293] "Volume detached for volume \"workload-ssh-secret\" (UniqueName: \"kubernetes.io/secret/a020591e-029f-464f-8d6c-947acba2bf13-workload-ssh-secret\") on node \"crc\" DevicePath \"\"" Dec 13 09:37:25 crc kubenswrapper[4971]: I1213 09:37:25.402955 4971 reconciler_common.go:293] "Volume detached for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/a020591e-029f-464f-8d6c-947acba2bf13-ceph\") on node \"crc\" DevicePath \"\"" Dec 13 09:37:25 crc kubenswrapper[4971]: I1213 09:37:25.402987 4971 reconciler_common.go:286] "operationExecutor.UnmountDevice started for volume \"local-storage04-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage04-crc\") on node \"crc\" " Dec 13 09:37:25 crc kubenswrapper[4971]: I1213 09:37:25.402996 4971 reconciler_common.go:293] "Volume detached for volume \"compute-ssh-secret\" (UniqueName: \"kubernetes.io/secret/a020591e-029f-464f-8d6c-947acba2bf13-compute-ssh-secret\") on node \"crc\" DevicePath \"\"" Dec 13 09:37:25 crc kubenswrapper[4971]: I1213 09:37:25.425820 4971 operation_generator.go:917] UnmountDevice succeeded for volume "local-storage04-crc" (UniqueName: "kubernetes.io/local-volume/local-storage04-crc") on node "crc" Dec 13 09:37:25 crc kubenswrapper[4971]: I1213 09:37:25.504599 4971 reconciler_common.go:293] "Volume detached for volume \"local-storage04-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage04-crc\") on node \"crc\" DevicePath \"\"" Dec 13 09:37:29 crc kubenswrapper[4971]: I1213 09:37:29.769237 4971 scope.go:117] "RemoveContainer" containerID="7c80479d6aedc6e1a5cced1df9e55738d4a726b6ddd5adfe17363d9c5d097d6a" Dec 13 09:37:29 crc kubenswrapper[4971]: E1213 09:37:29.770927 4971 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-82xjz_openshift-machine-config-operator(e0c4af12-edda-41d7-9b44-f87396174ef2)\"" pod="openshift-machine-config-operator/machine-config-daemon-82xjz" podUID="e0c4af12-edda-41d7-9b44-f87396174ef2" Dec 13 09:37:37 crc kubenswrapper[4971]: I1213 09:37:37.390404 4971 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/test-operator-logs-pod-ansibletest-ansibletest-ansibletest"] Dec 13 09:37:37 crc kubenswrapper[4971]: E1213 09:37:37.391621 4971 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="a020591e-029f-464f-8d6c-947acba2bf13" containerName="ansibletest-ansibletest" Dec 13 09:37:37 crc kubenswrapper[4971]: I1213 09:37:37.391641 4971 state_mem.go:107] "Deleted CPUSet assignment" podUID="a020591e-029f-464f-8d6c-947acba2bf13" containerName="ansibletest-ansibletest" Dec 13 09:37:37 crc kubenswrapper[4971]: I1213 09:37:37.391878 4971 memory_manager.go:354] "RemoveStaleState removing state" podUID="a020591e-029f-464f-8d6c-947acba2bf13" containerName="ansibletest-ansibletest" Dec 13 09:37:37 crc kubenswrapper[4971]: I1213 09:37:37.392776 4971 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/test-operator-logs-pod-ansibletest-ansibletest-ansibletest" Dec 13 09:37:37 crc kubenswrapper[4971]: I1213 09:37:37.404666 4971 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/test-operator-logs-pod-ansibletest-ansibletest-ansibletest"] Dec 13 09:37:37 crc kubenswrapper[4971]: I1213 09:37:37.407725 4971 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"local-storage04-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage04-crc\") pod \"test-operator-logs-pod-ansibletest-ansibletest-ansibletest\" (UID: \"4cf376d3-529c-41ee-8da5-9aa3bb37cfed\") " pod="openstack/test-operator-logs-pod-ansibletest-ansibletest-ansibletest" Dec 13 09:37:37 crc kubenswrapper[4971]: I1213 09:37:37.407775 4971 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-4qp7m\" (UniqueName: \"kubernetes.io/projected/4cf376d3-529c-41ee-8da5-9aa3bb37cfed-kube-api-access-4qp7m\") pod \"test-operator-logs-pod-ansibletest-ansibletest-ansibletest\" (UID: \"4cf376d3-529c-41ee-8da5-9aa3bb37cfed\") " pod="openstack/test-operator-logs-pod-ansibletest-ansibletest-ansibletest" Dec 13 09:37:37 crc kubenswrapper[4971]: I1213 09:37:37.514387 4971 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"local-storage04-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage04-crc\") pod \"test-operator-logs-pod-ansibletest-ansibletest-ansibletest\" (UID: \"4cf376d3-529c-41ee-8da5-9aa3bb37cfed\") " pod="openstack/test-operator-logs-pod-ansibletest-ansibletest-ansibletest" Dec 13 09:37:37 crc kubenswrapper[4971]: I1213 09:37:37.514436 4971 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-4qp7m\" (UniqueName: \"kubernetes.io/projected/4cf376d3-529c-41ee-8da5-9aa3bb37cfed-kube-api-access-4qp7m\") pod \"test-operator-logs-pod-ansibletest-ansibletest-ansibletest\" (UID: \"4cf376d3-529c-41ee-8da5-9aa3bb37cfed\") " pod="openstack/test-operator-logs-pod-ansibletest-ansibletest-ansibletest" Dec 13 09:37:37 crc kubenswrapper[4971]: I1213 09:37:37.514918 4971 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"local-storage04-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage04-crc\") pod \"test-operator-logs-pod-ansibletest-ansibletest-ansibletest\" (UID: \"4cf376d3-529c-41ee-8da5-9aa3bb37cfed\") device mount path \"/mnt/openstack/pv04\"" pod="openstack/test-operator-logs-pod-ansibletest-ansibletest-ansibletest" Dec 13 09:37:37 crc kubenswrapper[4971]: I1213 09:37:37.645443 4971 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-4qp7m\" (UniqueName: \"kubernetes.io/projected/4cf376d3-529c-41ee-8da5-9aa3bb37cfed-kube-api-access-4qp7m\") pod \"test-operator-logs-pod-ansibletest-ansibletest-ansibletest\" (UID: \"4cf376d3-529c-41ee-8da5-9aa3bb37cfed\") " pod="openstack/test-operator-logs-pod-ansibletest-ansibletest-ansibletest" Dec 13 09:37:37 crc kubenswrapper[4971]: I1213 09:37:37.663639 4971 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"local-storage04-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage04-crc\") pod \"test-operator-logs-pod-ansibletest-ansibletest-ansibletest\" (UID: \"4cf376d3-529c-41ee-8da5-9aa3bb37cfed\") " pod="openstack/test-operator-logs-pod-ansibletest-ansibletest-ansibletest" Dec 13 09:37:37 crc kubenswrapper[4971]: I1213 09:37:37.726016 4971 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/test-operator-logs-pod-ansibletest-ansibletest-ansibletest" Dec 13 09:37:38 crc kubenswrapper[4971]: I1213 09:37:38.502091 4971 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/test-operator-logs-pod-ansibletest-ansibletest-ansibletest"] Dec 13 09:37:39 crc kubenswrapper[4971]: I1213 09:37:39.307249 4971 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/test-operator-logs-pod-ansibletest-ansibletest-ansibletest" event={"ID":"4cf376d3-529c-41ee-8da5-9aa3bb37cfed","Type":"ContainerStarted","Data":"6c31b608282681947a676adb8c2426c4ad9437713e60de6524007b05031c1b24"} Dec 13 09:37:41 crc kubenswrapper[4971]: I1213 09:37:41.328263 4971 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/test-operator-logs-pod-ansibletest-ansibletest-ansibletest" event={"ID":"4cf376d3-529c-41ee-8da5-9aa3bb37cfed","Type":"ContainerStarted","Data":"37262dcb76a3397783b70dabb9f2dc4072c13314f26fc08dfbf85beae9bfbcb8"} Dec 13 09:37:41 crc kubenswrapper[4971]: I1213 09:37:41.352354 4971 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/test-operator-logs-pod-ansibletest-ansibletest-ansibletest" podStartSLOduration=2.308566038 podStartE2EDuration="4.352325495s" podCreationTimestamp="2025-12-13 09:37:37 +0000 UTC" firstStartedPulling="2025-12-13 09:37:38.508966441 +0000 UTC m=+10115.113375879" lastFinishedPulling="2025-12-13 09:37:40.552725888 +0000 UTC m=+10117.157135336" observedRunningTime="2025-12-13 09:37:41.342061927 +0000 UTC m=+10117.946471385" watchObservedRunningTime="2025-12-13 09:37:41.352325495 +0000 UTC m=+10117.956734943" Dec 13 09:37:44 crc kubenswrapper[4971]: I1213 09:37:44.769768 4971 scope.go:117] "RemoveContainer" containerID="7c80479d6aedc6e1a5cced1df9e55738d4a726b6ddd5adfe17363d9c5d097d6a" Dec 13 09:37:44 crc kubenswrapper[4971]: E1213 09:37:44.770662 4971 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-82xjz_openshift-machine-config-operator(e0c4af12-edda-41d7-9b44-f87396174ef2)\"" pod="openshift-machine-config-operator/machine-config-daemon-82xjz" podUID="e0c4af12-edda-41d7-9b44-f87396174ef2" Dec 13 09:37:52 crc kubenswrapper[4971]: I1213 09:37:52.431142 4971 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/horizontest-tests-horizontest"] Dec 13 09:37:52 crc kubenswrapper[4971]: I1213 09:37:52.433832 4971 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/horizontest-tests-horizontest" Dec 13 09:37:52 crc kubenswrapper[4971]: I1213 09:37:52.436961 4971 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"horizontest-tests-horizontesthorizontest-config" Dec 13 09:37:52 crc kubenswrapper[4971]: I1213 09:37:52.436975 4971 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"test-operator-clouds-config" Dec 13 09:37:52 crc kubenswrapper[4971]: I1213 09:37:52.446547 4971 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/horizontest-tests-horizontest"] Dec 13 09:37:52 crc kubenswrapper[4971]: I1213 09:37:52.516174 4971 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"test-operator-clouds-config\" (UniqueName: \"kubernetes.io/configmap/abaeac3e-11a1-40b0-b694-26822d0c02bf-test-operator-clouds-config\") pod \"horizontest-tests-horizontest\" (UID: \"abaeac3e-11a1-40b0-b694-26822d0c02bf\") " pod="openstack/horizontest-tests-horizontest" Dec 13 09:37:52 crc kubenswrapper[4971]: I1213 09:37:52.516259 4971 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"openstack-config-secret\" (UniqueName: \"kubernetes.io/secret/abaeac3e-11a1-40b0-b694-26822d0c02bf-openstack-config-secret\") pod \"horizontest-tests-horizontest\" (UID: \"abaeac3e-11a1-40b0-b694-26822d0c02bf\") " pod="openstack/horizontest-tests-horizontest" Dec 13 09:37:52 crc kubenswrapper[4971]: I1213 09:37:52.618624 4971 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"test-operator-ephemeral-temporary\" (UniqueName: \"kubernetes.io/empty-dir/abaeac3e-11a1-40b0-b694-26822d0c02bf-test-operator-ephemeral-temporary\") pod \"horizontest-tests-horizontest\" (UID: \"abaeac3e-11a1-40b0-b694-26822d0c02bf\") " pod="openstack/horizontest-tests-horizontest" Dec 13 09:37:52 crc kubenswrapper[4971]: I1213 09:37:52.618709 4971 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"local-storage05-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage05-crc\") pod \"horizontest-tests-horizontest\" (UID: \"abaeac3e-11a1-40b0-b694-26822d0c02bf\") " pod="openstack/horizontest-tests-horizontest" Dec 13 09:37:52 crc kubenswrapper[4971]: I1213 09:37:52.619021 4971 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-czd2t\" (UniqueName: \"kubernetes.io/projected/abaeac3e-11a1-40b0-b694-26822d0c02bf-kube-api-access-czd2t\") pod \"horizontest-tests-horizontest\" (UID: \"abaeac3e-11a1-40b0-b694-26822d0c02bf\") " pod="openstack/horizontest-tests-horizontest" Dec 13 09:37:52 crc kubenswrapper[4971]: I1213 09:37:52.619135 4971 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/abaeac3e-11a1-40b0-b694-26822d0c02bf-ceph\") pod \"horizontest-tests-horizontest\" (UID: \"abaeac3e-11a1-40b0-b694-26822d0c02bf\") " pod="openstack/horizontest-tests-horizontest" Dec 13 09:37:52 crc kubenswrapper[4971]: I1213 09:37:52.619388 4971 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"test-operator-ephemeral-workdir\" (UniqueName: \"kubernetes.io/empty-dir/abaeac3e-11a1-40b0-b694-26822d0c02bf-test-operator-ephemeral-workdir\") pod \"horizontest-tests-horizontest\" (UID: \"abaeac3e-11a1-40b0-b694-26822d0c02bf\") " pod="openstack/horizontest-tests-horizontest" Dec 13 09:37:52 crc kubenswrapper[4971]: I1213 09:37:52.619544 4971 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ca-certs\" (UniqueName: \"kubernetes.io/secret/abaeac3e-11a1-40b0-b694-26822d0c02bf-ca-certs\") pod \"horizontest-tests-horizontest\" (UID: \"abaeac3e-11a1-40b0-b694-26822d0c02bf\") " pod="openstack/horizontest-tests-horizontest" Dec 13 09:37:52 crc kubenswrapper[4971]: I1213 09:37:52.619615 4971 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"test-operator-clouds-config\" (UniqueName: \"kubernetes.io/configmap/abaeac3e-11a1-40b0-b694-26822d0c02bf-test-operator-clouds-config\") pod \"horizontest-tests-horizontest\" (UID: \"abaeac3e-11a1-40b0-b694-26822d0c02bf\") " pod="openstack/horizontest-tests-horizontest" Dec 13 09:37:52 crc kubenswrapper[4971]: I1213 09:37:52.619652 4971 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"openstack-config-secret\" (UniqueName: \"kubernetes.io/secret/abaeac3e-11a1-40b0-b694-26822d0c02bf-openstack-config-secret\") pod \"horizontest-tests-horizontest\" (UID: \"abaeac3e-11a1-40b0-b694-26822d0c02bf\") " pod="openstack/horizontest-tests-horizontest" Dec 13 09:37:52 crc kubenswrapper[4971]: I1213 09:37:52.620952 4971 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"test-operator-clouds-config\" (UniqueName: \"kubernetes.io/configmap/abaeac3e-11a1-40b0-b694-26822d0c02bf-test-operator-clouds-config\") pod \"horizontest-tests-horizontest\" (UID: \"abaeac3e-11a1-40b0-b694-26822d0c02bf\") " pod="openstack/horizontest-tests-horizontest" Dec 13 09:37:52 crc kubenswrapper[4971]: I1213 09:37:52.634305 4971 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"openstack-config-secret\" (UniqueName: \"kubernetes.io/secret/abaeac3e-11a1-40b0-b694-26822d0c02bf-openstack-config-secret\") pod \"horizontest-tests-horizontest\" (UID: \"abaeac3e-11a1-40b0-b694-26822d0c02bf\") " pod="openstack/horizontest-tests-horizontest" Dec 13 09:37:52 crc kubenswrapper[4971]: I1213 09:37:52.722410 4971 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"local-storage05-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage05-crc\") pod \"horizontest-tests-horizontest\" (UID: \"abaeac3e-11a1-40b0-b694-26822d0c02bf\") " pod="openstack/horizontest-tests-horizontest" Dec 13 09:37:52 crc kubenswrapper[4971]: I1213 09:37:52.722885 4971 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-czd2t\" (UniqueName: \"kubernetes.io/projected/abaeac3e-11a1-40b0-b694-26822d0c02bf-kube-api-access-czd2t\") pod \"horizontest-tests-horizontest\" (UID: \"abaeac3e-11a1-40b0-b694-26822d0c02bf\") " pod="openstack/horizontest-tests-horizontest" Dec 13 09:37:52 crc kubenswrapper[4971]: I1213 09:37:52.722920 4971 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/abaeac3e-11a1-40b0-b694-26822d0c02bf-ceph\") pod \"horizontest-tests-horizontest\" (UID: \"abaeac3e-11a1-40b0-b694-26822d0c02bf\") " pod="openstack/horizontest-tests-horizontest" Dec 13 09:37:52 crc kubenswrapper[4971]: I1213 09:37:52.722981 4971 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"test-operator-ephemeral-workdir\" (UniqueName: \"kubernetes.io/empty-dir/abaeac3e-11a1-40b0-b694-26822d0c02bf-test-operator-ephemeral-workdir\") pod \"horizontest-tests-horizontest\" (UID: \"abaeac3e-11a1-40b0-b694-26822d0c02bf\") " pod="openstack/horizontest-tests-horizontest" Dec 13 09:37:52 crc kubenswrapper[4971]: I1213 09:37:52.723020 4971 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ca-certs\" (UniqueName: \"kubernetes.io/secret/abaeac3e-11a1-40b0-b694-26822d0c02bf-ca-certs\") pod \"horizontest-tests-horizontest\" (UID: \"abaeac3e-11a1-40b0-b694-26822d0c02bf\") " pod="openstack/horizontest-tests-horizontest" Dec 13 09:37:52 crc kubenswrapper[4971]: I1213 09:37:52.723138 4971 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"test-operator-ephemeral-temporary\" (UniqueName: \"kubernetes.io/empty-dir/abaeac3e-11a1-40b0-b694-26822d0c02bf-test-operator-ephemeral-temporary\") pod \"horizontest-tests-horizontest\" (UID: \"abaeac3e-11a1-40b0-b694-26822d0c02bf\") " pod="openstack/horizontest-tests-horizontest" Dec 13 09:37:52 crc kubenswrapper[4971]: I1213 09:37:52.723700 4971 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"test-operator-ephemeral-temporary\" (UniqueName: \"kubernetes.io/empty-dir/abaeac3e-11a1-40b0-b694-26822d0c02bf-test-operator-ephemeral-temporary\") pod \"horizontest-tests-horizontest\" (UID: \"abaeac3e-11a1-40b0-b694-26822d0c02bf\") " pod="openstack/horizontest-tests-horizontest" Dec 13 09:37:52 crc kubenswrapper[4971]: I1213 09:37:52.723699 4971 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"local-storage05-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage05-crc\") pod \"horizontest-tests-horizontest\" (UID: \"abaeac3e-11a1-40b0-b694-26822d0c02bf\") device mount path \"/mnt/openstack/pv05\"" pod="openstack/horizontest-tests-horizontest" Dec 13 09:37:52 crc kubenswrapper[4971]: I1213 09:37:52.724329 4971 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"test-operator-ephemeral-workdir\" (UniqueName: \"kubernetes.io/empty-dir/abaeac3e-11a1-40b0-b694-26822d0c02bf-test-operator-ephemeral-workdir\") pod \"horizontest-tests-horizontest\" (UID: \"abaeac3e-11a1-40b0-b694-26822d0c02bf\") " pod="openstack/horizontest-tests-horizontest" Dec 13 09:37:52 crc kubenswrapper[4971]: I1213 09:37:52.728146 4971 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ca-certs\" (UniqueName: \"kubernetes.io/secret/abaeac3e-11a1-40b0-b694-26822d0c02bf-ca-certs\") pod \"horizontest-tests-horizontest\" (UID: \"abaeac3e-11a1-40b0-b694-26822d0c02bf\") " pod="openstack/horizontest-tests-horizontest" Dec 13 09:37:52 crc kubenswrapper[4971]: I1213 09:37:52.729022 4971 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/abaeac3e-11a1-40b0-b694-26822d0c02bf-ceph\") pod \"horizontest-tests-horizontest\" (UID: \"abaeac3e-11a1-40b0-b694-26822d0c02bf\") " pod="openstack/horizontest-tests-horizontest" Dec 13 09:37:52 crc kubenswrapper[4971]: I1213 09:37:52.743243 4971 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-czd2t\" (UniqueName: \"kubernetes.io/projected/abaeac3e-11a1-40b0-b694-26822d0c02bf-kube-api-access-czd2t\") pod \"horizontest-tests-horizontest\" (UID: \"abaeac3e-11a1-40b0-b694-26822d0c02bf\") " pod="openstack/horizontest-tests-horizontest" Dec 13 09:37:52 crc kubenswrapper[4971]: I1213 09:37:52.756729 4971 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"local-storage05-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage05-crc\") pod \"horizontest-tests-horizontest\" (UID: \"abaeac3e-11a1-40b0-b694-26822d0c02bf\") " pod="openstack/horizontest-tests-horizontest" Dec 13 09:37:52 crc kubenswrapper[4971]: I1213 09:37:52.772308 4971 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/horizontest-tests-horizontest" Dec 13 09:37:53 crc kubenswrapper[4971]: I1213 09:37:53.310446 4971 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/horizontest-tests-horizontest"] Dec 13 09:37:54 crc kubenswrapper[4971]: I1213 09:37:54.451324 4971 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/horizontest-tests-horizontest" event={"ID":"abaeac3e-11a1-40b0-b694-26822d0c02bf","Type":"ContainerStarted","Data":"0fb2450c7e73fc3a279482d7e35206a08454c790da3cf44dec1a7b8359244714"} Dec 13 09:37:58 crc kubenswrapper[4971]: I1213 09:37:58.776104 4971 scope.go:117] "RemoveContainer" containerID="7c80479d6aedc6e1a5cced1df9e55738d4a726b6ddd5adfe17363d9c5d097d6a" Dec 13 09:37:58 crc kubenswrapper[4971]: E1213 09:37:58.784367 4971 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-82xjz_openshift-machine-config-operator(e0c4af12-edda-41d7-9b44-f87396174ef2)\"" pod="openshift-machine-config-operator/machine-config-daemon-82xjz" podUID="e0c4af12-edda-41d7-9b44-f87396174ef2" Dec 13 09:38:11 crc kubenswrapper[4971]: I1213 09:38:11.769393 4971 scope.go:117] "RemoveContainer" containerID="7c80479d6aedc6e1a5cced1df9e55738d4a726b6ddd5adfe17363d9c5d097d6a" Dec 13 09:38:11 crc kubenswrapper[4971]: E1213 09:38:11.770587 4971 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-82xjz_openshift-machine-config-operator(e0c4af12-edda-41d7-9b44-f87396174ef2)\"" pod="openshift-machine-config-operator/machine-config-daemon-82xjz" podUID="e0c4af12-edda-41d7-9b44-f87396174ef2" Dec 13 09:38:22 crc kubenswrapper[4971]: E1213 09:38:22.300001 4971 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying config: context canceled" image="quay.io/podified-antelope-centos9/openstack-horizontest:current-podified" Dec 13 09:38:22 crc kubenswrapper[4971]: E1213 09:38:22.301155 4971 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:horizontest-tests-horizontest,Image:quay.io/podified-antelope-centos9/openstack-horizontest:current-podified,Command:[],Args:[],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:ADMIN_PASSWORD,Value:12345678,ValueFrom:nil,},EnvVar{Name:ADMIN_USERNAME,Value:admin,ValueFrom:nil,},EnvVar{Name:AUTH_URL,Value:https://keystone-public-openstack.apps-crc.testing,ValueFrom:nil,},EnvVar{Name:DASHBOARD_URL,Value:https://horizon-openstack.apps-crc.testing/,ValueFrom:nil,},EnvVar{Name:EXTRA_FLAG,Value:not pagination and test_users.py,ValueFrom:nil,},EnvVar{Name:FLAVOR_NAME,Value:m1.tiny,ValueFrom:nil,},EnvVar{Name:HORIZONTEST_DEBUG_MODE,Value:false,ValueFrom:nil,},EnvVar{Name:HORIZON_KEYS_FOLDER,Value:/etc/test_operator,ValueFrom:nil,},EnvVar{Name:HORIZON_LOGS_DIR_NAME,Value:horizon,ValueFrom:nil,},EnvVar{Name:HORIZON_REPO_BRANCH,Value:master,ValueFrom:nil,},EnvVar{Name:IMAGE_FILE,Value:/var/lib/horizontest/cirros-0.6.2-x86_64-disk.img,ValueFrom:nil,},EnvVar{Name:IMAGE_FILE_NAME,Value:cirros-0.6.2-x86_64-disk,ValueFrom:nil,},EnvVar{Name:IMAGE_URL,Value:http://download.cirros-cloud.net/0.6.2/cirros-0.6.2-x86_64-disk.img,ValueFrom:nil,},EnvVar{Name:PASSWORD,Value:horizontest,ValueFrom:nil,},EnvVar{Name:PROJECT_NAME,Value:horizontest,ValueFrom:nil,},EnvVar{Name:PROJECT_NAME_XPATH,Value://*[@class=\"context-project\"]//ancestor::ul,ValueFrom:nil,},EnvVar{Name:REPO_URL,Value:https://review.opendev.org/openstack/horizon,ValueFrom:nil,},EnvVar{Name:USER_NAME,Value:horizontest,ValueFrom:nil,},EnvVar{Name:USE_EXTERNAL_FILES,Value:True,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{cpu: {{2 0} {} 2 DecimalSI},memory: {{4294967296 0} {} 4Gi BinarySI},},Requests:ResourceList{cpu: {{1 0} {} 1 DecimalSI},memory: {{2147483648 0} {} 2Gi BinarySI},},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:test-operator-ephemeral-workdir,ReadOnly:false,MountPath:/var/lib/horizontest,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:test-operator-ephemeral-temporary,ReadOnly:false,MountPath:/tmp,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:test-operator-logs,ReadOnly:false,MountPath:/var/lib/horizontest/external_files,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:test-operator-clouds-config,ReadOnly:true,MountPath:/var/lib/horizontest/.config/openstack/clouds.yaml,SubPath:clouds.yaml,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:test-operator-clouds-config,ReadOnly:true,MountPath:/etc/openstack/clouds.yaml,SubPath:clouds.yaml,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:openstack-config-secret,ReadOnly:false,MountPath:/etc/openstack/secure.yaml,SubPath:secure.yaml,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:ca-certs,ReadOnly:true,MountPath:/etc/pki/ca-trust/extracted/pem/tls-ca-bundle.pem,SubPath:tls-ca-bundle.pem,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:ca-certs,ReadOnly:true,MountPath:/etc/pki/tls/certs/ca-bundle.trust.crt,SubPath:tls-ca-bundle.pem,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:ceph,ReadOnly:true,MountPath:/etc/ceph,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-czd2t,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[NET_ADMIN NET_RAW],Drop:[],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*42455,RunAsNonRoot:*false,ReadOnlyRootFilesystem:*false,AllowPrivilegeEscalation:*true,RunAsGroup:*42455,ProcMount:nil,WindowsOptions:nil,SeccompProfile:&SeccompProfile{Type:RuntimeDefault,LocalhostProfile:nil,},AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod horizontest-tests-horizontest_openstack(abaeac3e-11a1-40b0-b694-26822d0c02bf): ErrImagePull: rpc error: code = Canceled desc = copying config: context canceled" logger="UnhandledError" Dec 13 09:38:22 crc kubenswrapper[4971]: E1213 09:38:22.302828 4971 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"horizontest-tests-horizontest\" with ErrImagePull: \"rpc error: code = Canceled desc = copying config: context canceled\"" pod="openstack/horizontest-tests-horizontest" podUID="abaeac3e-11a1-40b0-b694-26822d0c02bf" Dec 13 09:38:22 crc kubenswrapper[4971]: E1213 09:38:22.510790 4971 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"horizontest-tests-horizontest\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/podified-antelope-centos9/openstack-horizontest:current-podified\\\"\"" pod="openstack/horizontest-tests-horizontest" podUID="abaeac3e-11a1-40b0-b694-26822d0c02bf" Dec 13 09:38:26 crc kubenswrapper[4971]: I1213 09:38:26.768860 4971 scope.go:117] "RemoveContainer" containerID="7c80479d6aedc6e1a5cced1df9e55738d4a726b6ddd5adfe17363d9c5d097d6a" Dec 13 09:38:26 crc kubenswrapper[4971]: E1213 09:38:26.769976 4971 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-82xjz_openshift-machine-config-operator(e0c4af12-edda-41d7-9b44-f87396174ef2)\"" pod="openshift-machine-config-operator/machine-config-daemon-82xjz" podUID="e0c4af12-edda-41d7-9b44-f87396174ef2" Dec 13 09:38:36 crc kubenswrapper[4971]: I1213 09:38:36.658505 4971 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/horizontest-tests-horizontest" event={"ID":"abaeac3e-11a1-40b0-b694-26822d0c02bf","Type":"ContainerStarted","Data":"fec73e3f45b80e5a35f856d1e51ff9c601d94e21be4639194920cd60922fec06"} Dec 13 09:38:36 crc kubenswrapper[4971]: I1213 09:38:36.692292 4971 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/horizontest-tests-horizontest" podStartSLOduration=4.044414274 podStartE2EDuration="45.692251318s" podCreationTimestamp="2025-12-13 09:37:51 +0000 UTC" firstStartedPulling="2025-12-13 09:37:53.744936674 +0000 UTC m=+10130.349346122" lastFinishedPulling="2025-12-13 09:38:35.392773718 +0000 UTC m=+10171.997183166" observedRunningTime="2025-12-13 09:38:36.682106699 +0000 UTC m=+10173.286516157" watchObservedRunningTime="2025-12-13 09:38:36.692251318 +0000 UTC m=+10173.296660766" Dec 13 09:38:39 crc kubenswrapper[4971]: I1213 09:38:39.769409 4971 scope.go:117] "RemoveContainer" containerID="7c80479d6aedc6e1a5cced1df9e55738d4a726b6ddd5adfe17363d9c5d097d6a" Dec 13 09:38:39 crc kubenswrapper[4971]: E1213 09:38:39.770448 4971 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-82xjz_openshift-machine-config-operator(e0c4af12-edda-41d7-9b44-f87396174ef2)\"" pod="openshift-machine-config-operator/machine-config-daemon-82xjz" podUID="e0c4af12-edda-41d7-9b44-f87396174ef2" Dec 13 09:38:48 crc kubenswrapper[4971]: I1213 09:38:48.740365 4971 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/openstack-galera-0" podUID="91831fd5-fc52-46ec-ac49-63859378116b" containerName="galera" probeResult="failure" output="command timed out" Dec 13 09:38:52 crc kubenswrapper[4971]: I1213 09:38:52.768565 4971 scope.go:117] "RemoveContainer" containerID="7c80479d6aedc6e1a5cced1df9e55738d4a726b6ddd5adfe17363d9c5d097d6a" Dec 13 09:38:52 crc kubenswrapper[4971]: E1213 09:38:52.769377 4971 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-82xjz_openshift-machine-config-operator(e0c4af12-edda-41d7-9b44-f87396174ef2)\"" pod="openshift-machine-config-operator/machine-config-daemon-82xjz" podUID="e0c4af12-edda-41d7-9b44-f87396174ef2" Dec 13 09:39:06 crc kubenswrapper[4971]: I1213 09:39:06.768528 4971 scope.go:117] "RemoveContainer" containerID="7c80479d6aedc6e1a5cced1df9e55738d4a726b6ddd5adfe17363d9c5d097d6a" Dec 13 09:39:06 crc kubenswrapper[4971]: E1213 09:39:06.769422 4971 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-82xjz_openshift-machine-config-operator(e0c4af12-edda-41d7-9b44-f87396174ef2)\"" pod="openshift-machine-config-operator/machine-config-daemon-82xjz" podUID="e0c4af12-edda-41d7-9b44-f87396174ef2" Dec 13 09:39:20 crc kubenswrapper[4971]: I1213 09:39:20.769235 4971 scope.go:117] "RemoveContainer" containerID="7c80479d6aedc6e1a5cced1df9e55738d4a726b6ddd5adfe17363d9c5d097d6a" Dec 13 09:39:20 crc kubenswrapper[4971]: E1213 09:39:20.770057 4971 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-82xjz_openshift-machine-config-operator(e0c4af12-edda-41d7-9b44-f87396174ef2)\"" pod="openshift-machine-config-operator/machine-config-daemon-82xjz" podUID="e0c4af12-edda-41d7-9b44-f87396174ef2" Dec 13 09:39:35 crc kubenswrapper[4971]: I1213 09:39:35.770384 4971 scope.go:117] "RemoveContainer" containerID="7c80479d6aedc6e1a5cced1df9e55738d4a726b6ddd5adfe17363d9c5d097d6a" Dec 13 09:39:35 crc kubenswrapper[4971]: E1213 09:39:35.771139 4971 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-82xjz_openshift-machine-config-operator(e0c4af12-edda-41d7-9b44-f87396174ef2)\"" pod="openshift-machine-config-operator/machine-config-daemon-82xjz" podUID="e0c4af12-edda-41d7-9b44-f87396174ef2" Dec 13 09:39:48 crc kubenswrapper[4971]: I1213 09:39:48.768273 4971 scope.go:117] "RemoveContainer" containerID="7c80479d6aedc6e1a5cced1df9e55738d4a726b6ddd5adfe17363d9c5d097d6a" Dec 13 09:39:49 crc kubenswrapper[4971]: I1213 09:39:49.465246 4971 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-82xjz" event={"ID":"e0c4af12-edda-41d7-9b44-f87396174ef2","Type":"ContainerStarted","Data":"b2e549456db49c5609edf5288a8e1e209f21e8f767afaa3d5510b083ba33ae70"} Dec 13 09:40:09 crc kubenswrapper[4971]: I1213 09:40:09.814798 4971 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/certified-operators-dcfd2"] Dec 13 09:40:09 crc kubenswrapper[4971]: I1213 09:40:09.819007 4971 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-dcfd2" Dec 13 09:40:09 crc kubenswrapper[4971]: I1213 09:40:09.827571 4971 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-dcfd2"] Dec 13 09:40:09 crc kubenswrapper[4971]: I1213 09:40:09.942791 4971 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/232e5703-7181-44ae-963a-53255085e5da-utilities\") pod \"certified-operators-dcfd2\" (UID: \"232e5703-7181-44ae-963a-53255085e5da\") " pod="openshift-marketplace/certified-operators-dcfd2" Dec 13 09:40:09 crc kubenswrapper[4971]: I1213 09:40:09.942862 4971 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-224sm\" (UniqueName: \"kubernetes.io/projected/232e5703-7181-44ae-963a-53255085e5da-kube-api-access-224sm\") pod \"certified-operators-dcfd2\" (UID: \"232e5703-7181-44ae-963a-53255085e5da\") " pod="openshift-marketplace/certified-operators-dcfd2" Dec 13 09:40:09 crc kubenswrapper[4971]: I1213 09:40:09.942991 4971 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/232e5703-7181-44ae-963a-53255085e5da-catalog-content\") pod \"certified-operators-dcfd2\" (UID: \"232e5703-7181-44ae-963a-53255085e5da\") " pod="openshift-marketplace/certified-operators-dcfd2" Dec 13 09:40:10 crc kubenswrapper[4971]: I1213 09:40:10.045936 4971 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/232e5703-7181-44ae-963a-53255085e5da-utilities\") pod \"certified-operators-dcfd2\" (UID: \"232e5703-7181-44ae-963a-53255085e5da\") " pod="openshift-marketplace/certified-operators-dcfd2" Dec 13 09:40:10 crc kubenswrapper[4971]: I1213 09:40:10.046045 4971 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-224sm\" (UniqueName: \"kubernetes.io/projected/232e5703-7181-44ae-963a-53255085e5da-kube-api-access-224sm\") pod \"certified-operators-dcfd2\" (UID: \"232e5703-7181-44ae-963a-53255085e5da\") " pod="openshift-marketplace/certified-operators-dcfd2" Dec 13 09:40:10 crc kubenswrapper[4971]: I1213 09:40:10.046210 4971 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/232e5703-7181-44ae-963a-53255085e5da-catalog-content\") pod \"certified-operators-dcfd2\" (UID: \"232e5703-7181-44ae-963a-53255085e5da\") " pod="openshift-marketplace/certified-operators-dcfd2" Dec 13 09:40:10 crc kubenswrapper[4971]: I1213 09:40:10.046666 4971 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/232e5703-7181-44ae-963a-53255085e5da-utilities\") pod \"certified-operators-dcfd2\" (UID: \"232e5703-7181-44ae-963a-53255085e5da\") " pod="openshift-marketplace/certified-operators-dcfd2" Dec 13 09:40:10 crc kubenswrapper[4971]: I1213 09:40:10.046831 4971 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/232e5703-7181-44ae-963a-53255085e5da-catalog-content\") pod \"certified-operators-dcfd2\" (UID: \"232e5703-7181-44ae-963a-53255085e5da\") " pod="openshift-marketplace/certified-operators-dcfd2" Dec 13 09:40:10 crc kubenswrapper[4971]: I1213 09:40:10.070191 4971 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-224sm\" (UniqueName: \"kubernetes.io/projected/232e5703-7181-44ae-963a-53255085e5da-kube-api-access-224sm\") pod \"certified-operators-dcfd2\" (UID: \"232e5703-7181-44ae-963a-53255085e5da\") " pod="openshift-marketplace/certified-operators-dcfd2" Dec 13 09:40:10 crc kubenswrapper[4971]: I1213 09:40:10.152448 4971 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-dcfd2" Dec 13 09:40:10 crc kubenswrapper[4971]: I1213 09:40:10.689040 4971 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-dcfd2"] Dec 13 09:40:11 crc kubenswrapper[4971]: I1213 09:40:11.711376 4971 generic.go:334] "Generic (PLEG): container finished" podID="232e5703-7181-44ae-963a-53255085e5da" containerID="19f66b05910639c642ecb0a9c860cb7f2eb1786bf998218a28db290bf840515a" exitCode=0 Dec 13 09:40:11 crc kubenswrapper[4971]: I1213 09:40:11.711460 4971 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-dcfd2" event={"ID":"232e5703-7181-44ae-963a-53255085e5da","Type":"ContainerDied","Data":"19f66b05910639c642ecb0a9c860cb7f2eb1786bf998218a28db290bf840515a"} Dec 13 09:40:11 crc kubenswrapper[4971]: I1213 09:40:11.713137 4971 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-dcfd2" event={"ID":"232e5703-7181-44ae-963a-53255085e5da","Type":"ContainerStarted","Data":"334a2599586ef7441158b9d06ac84251065093a88742ccbb89f90e77933015cb"} Dec 13 09:40:12 crc kubenswrapper[4971]: I1213 09:40:12.725841 4971 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-dcfd2" event={"ID":"232e5703-7181-44ae-963a-53255085e5da","Type":"ContainerStarted","Data":"d6c2afd36c0a04a6c9d276c1301b60f18e56f80f1ebae9f3fc95aed5931a4323"} Dec 13 09:40:13 crc kubenswrapper[4971]: I1213 09:40:13.737129 4971 generic.go:334] "Generic (PLEG): container finished" podID="232e5703-7181-44ae-963a-53255085e5da" containerID="d6c2afd36c0a04a6c9d276c1301b60f18e56f80f1ebae9f3fc95aed5931a4323" exitCode=0 Dec 13 09:40:13 crc kubenswrapper[4971]: I1213 09:40:13.737171 4971 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-dcfd2" event={"ID":"232e5703-7181-44ae-963a-53255085e5da","Type":"ContainerDied","Data":"d6c2afd36c0a04a6c9d276c1301b60f18e56f80f1ebae9f3fc95aed5931a4323"} Dec 13 09:40:14 crc kubenswrapper[4971]: I1213 09:40:14.751027 4971 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-dcfd2" event={"ID":"232e5703-7181-44ae-963a-53255085e5da","Type":"ContainerStarted","Data":"dbf01d2a63c5d509878d416b54acb3ccfb747cfd5aeac060ac94f29119437fac"} Dec 13 09:40:14 crc kubenswrapper[4971]: I1213 09:40:14.780015 4971 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/certified-operators-dcfd2" podStartSLOduration=3.328354494 podStartE2EDuration="5.779987716s" podCreationTimestamp="2025-12-13 09:40:09 +0000 UTC" firstStartedPulling="2025-12-13 09:40:11.714216655 +0000 UTC m=+10268.318626103" lastFinishedPulling="2025-12-13 09:40:14.165849877 +0000 UTC m=+10270.770259325" observedRunningTime="2025-12-13 09:40:14.774622475 +0000 UTC m=+10271.379031933" watchObservedRunningTime="2025-12-13 09:40:14.779987716 +0000 UTC m=+10271.384397174" Dec 13 09:40:20 crc kubenswrapper[4971]: I1213 09:40:20.152923 4971 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/certified-operators-dcfd2" Dec 13 09:40:20 crc kubenswrapper[4971]: I1213 09:40:20.153722 4971 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/certified-operators-dcfd2" Dec 13 09:40:20 crc kubenswrapper[4971]: I1213 09:40:20.203993 4971 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/certified-operators-dcfd2" Dec 13 09:40:20 crc kubenswrapper[4971]: I1213 09:40:20.876262 4971 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/certified-operators-dcfd2" Dec 13 09:40:20 crc kubenswrapper[4971]: I1213 09:40:20.946158 4971 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-dcfd2"] Dec 13 09:40:22 crc kubenswrapper[4971]: I1213 09:40:22.841192 4971 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/certified-operators-dcfd2" podUID="232e5703-7181-44ae-963a-53255085e5da" containerName="registry-server" containerID="cri-o://dbf01d2a63c5d509878d416b54acb3ccfb747cfd5aeac060ac94f29119437fac" gracePeriod=2 Dec 13 09:40:23 crc kubenswrapper[4971]: I1213 09:40:23.854028 4971 generic.go:334] "Generic (PLEG): container finished" podID="232e5703-7181-44ae-963a-53255085e5da" containerID="dbf01d2a63c5d509878d416b54acb3ccfb747cfd5aeac060ac94f29119437fac" exitCode=0 Dec 13 09:40:23 crc kubenswrapper[4971]: I1213 09:40:23.855185 4971 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-dcfd2" event={"ID":"232e5703-7181-44ae-963a-53255085e5da","Type":"ContainerDied","Data":"dbf01d2a63c5d509878d416b54acb3ccfb747cfd5aeac060ac94f29119437fac"} Dec 13 09:40:24 crc kubenswrapper[4971]: I1213 09:40:24.488280 4971 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-dcfd2" Dec 13 09:40:24 crc kubenswrapper[4971]: I1213 09:40:24.659607 4971 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-224sm\" (UniqueName: \"kubernetes.io/projected/232e5703-7181-44ae-963a-53255085e5da-kube-api-access-224sm\") pod \"232e5703-7181-44ae-963a-53255085e5da\" (UID: \"232e5703-7181-44ae-963a-53255085e5da\") " Dec 13 09:40:24 crc kubenswrapper[4971]: I1213 09:40:24.660275 4971 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/232e5703-7181-44ae-963a-53255085e5da-utilities\") pod \"232e5703-7181-44ae-963a-53255085e5da\" (UID: \"232e5703-7181-44ae-963a-53255085e5da\") " Dec 13 09:40:24 crc kubenswrapper[4971]: I1213 09:40:24.660367 4971 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/232e5703-7181-44ae-963a-53255085e5da-catalog-content\") pod \"232e5703-7181-44ae-963a-53255085e5da\" (UID: \"232e5703-7181-44ae-963a-53255085e5da\") " Dec 13 09:40:24 crc kubenswrapper[4971]: I1213 09:40:24.660947 4971 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/232e5703-7181-44ae-963a-53255085e5da-utilities" (OuterVolumeSpecName: "utilities") pod "232e5703-7181-44ae-963a-53255085e5da" (UID: "232e5703-7181-44ae-963a-53255085e5da"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 13 09:40:24 crc kubenswrapper[4971]: I1213 09:40:24.666960 4971 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/232e5703-7181-44ae-963a-53255085e5da-kube-api-access-224sm" (OuterVolumeSpecName: "kube-api-access-224sm") pod "232e5703-7181-44ae-963a-53255085e5da" (UID: "232e5703-7181-44ae-963a-53255085e5da"). InnerVolumeSpecName "kube-api-access-224sm". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 13 09:40:24 crc kubenswrapper[4971]: I1213 09:40:24.713385 4971 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/232e5703-7181-44ae-963a-53255085e5da-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "232e5703-7181-44ae-963a-53255085e5da" (UID: "232e5703-7181-44ae-963a-53255085e5da"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 13 09:40:24 crc kubenswrapper[4971]: I1213 09:40:24.763389 4971 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-224sm\" (UniqueName: \"kubernetes.io/projected/232e5703-7181-44ae-963a-53255085e5da-kube-api-access-224sm\") on node \"crc\" DevicePath \"\"" Dec 13 09:40:24 crc kubenswrapper[4971]: I1213 09:40:24.763442 4971 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/232e5703-7181-44ae-963a-53255085e5da-utilities\") on node \"crc\" DevicePath \"\"" Dec 13 09:40:24 crc kubenswrapper[4971]: I1213 09:40:24.763455 4971 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/232e5703-7181-44ae-963a-53255085e5da-catalog-content\") on node \"crc\" DevicePath \"\"" Dec 13 09:40:24 crc kubenswrapper[4971]: I1213 09:40:24.869507 4971 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-dcfd2" event={"ID":"232e5703-7181-44ae-963a-53255085e5da","Type":"ContainerDied","Data":"334a2599586ef7441158b9d06ac84251065093a88742ccbb89f90e77933015cb"} Dec 13 09:40:24 crc kubenswrapper[4971]: I1213 09:40:24.869583 4971 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-dcfd2" Dec 13 09:40:24 crc kubenswrapper[4971]: I1213 09:40:24.869629 4971 scope.go:117] "RemoveContainer" containerID="dbf01d2a63c5d509878d416b54acb3ccfb747cfd5aeac060ac94f29119437fac" Dec 13 09:40:24 crc kubenswrapper[4971]: I1213 09:40:24.895682 4971 scope.go:117] "RemoveContainer" containerID="d6c2afd36c0a04a6c9d276c1301b60f18e56f80f1ebae9f3fc95aed5931a4323" Dec 13 09:40:24 crc kubenswrapper[4971]: I1213 09:40:24.912096 4971 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-dcfd2"] Dec 13 09:40:24 crc kubenswrapper[4971]: I1213 09:40:24.922798 4971 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/certified-operators-dcfd2"] Dec 13 09:40:24 crc kubenswrapper[4971]: I1213 09:40:24.925540 4971 scope.go:117] "RemoveContainer" containerID="19f66b05910639c642ecb0a9c860cb7f2eb1786bf998218a28db290bf840515a" Dec 13 09:40:25 crc kubenswrapper[4971]: I1213 09:40:25.784212 4971 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="232e5703-7181-44ae-963a-53255085e5da" path="/var/lib/kubelet/pods/232e5703-7181-44ae-963a-53255085e5da/volumes" Dec 13 09:40:34 crc kubenswrapper[4971]: I1213 09:40:34.915252 4971 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/community-operators-gtr2j"] Dec 13 09:40:34 crc kubenswrapper[4971]: E1213 09:40:34.916870 4971 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="232e5703-7181-44ae-963a-53255085e5da" containerName="extract-content" Dec 13 09:40:34 crc kubenswrapper[4971]: I1213 09:40:34.916893 4971 state_mem.go:107] "Deleted CPUSet assignment" podUID="232e5703-7181-44ae-963a-53255085e5da" containerName="extract-content" Dec 13 09:40:34 crc kubenswrapper[4971]: E1213 09:40:34.916922 4971 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="232e5703-7181-44ae-963a-53255085e5da" containerName="registry-server" Dec 13 09:40:34 crc kubenswrapper[4971]: I1213 09:40:34.916930 4971 state_mem.go:107] "Deleted CPUSet assignment" podUID="232e5703-7181-44ae-963a-53255085e5da" containerName="registry-server" Dec 13 09:40:34 crc kubenswrapper[4971]: E1213 09:40:34.916957 4971 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="232e5703-7181-44ae-963a-53255085e5da" containerName="extract-utilities" Dec 13 09:40:34 crc kubenswrapper[4971]: I1213 09:40:34.916965 4971 state_mem.go:107] "Deleted CPUSet assignment" podUID="232e5703-7181-44ae-963a-53255085e5da" containerName="extract-utilities" Dec 13 09:40:34 crc kubenswrapper[4971]: I1213 09:40:34.917203 4971 memory_manager.go:354] "RemoveStaleState removing state" podUID="232e5703-7181-44ae-963a-53255085e5da" containerName="registry-server" Dec 13 09:40:34 crc kubenswrapper[4971]: I1213 09:40:34.919136 4971 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-gtr2j" Dec 13 09:40:34 crc kubenswrapper[4971]: I1213 09:40:34.933863 4971 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-gtr2j"] Dec 13 09:40:35 crc kubenswrapper[4971]: I1213 09:40:35.032119 4971 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/ecd2e077-5a83-46c0-92af-d86b26f8411c-catalog-content\") pod \"community-operators-gtr2j\" (UID: \"ecd2e077-5a83-46c0-92af-d86b26f8411c\") " pod="openshift-marketplace/community-operators-gtr2j" Dec 13 09:40:35 crc kubenswrapper[4971]: I1213 09:40:35.032335 4971 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/ecd2e077-5a83-46c0-92af-d86b26f8411c-utilities\") pod \"community-operators-gtr2j\" (UID: \"ecd2e077-5a83-46c0-92af-d86b26f8411c\") " pod="openshift-marketplace/community-operators-gtr2j" Dec 13 09:40:35 crc kubenswrapper[4971]: I1213 09:40:35.032460 4971 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-mxbtj\" (UniqueName: \"kubernetes.io/projected/ecd2e077-5a83-46c0-92af-d86b26f8411c-kube-api-access-mxbtj\") pod \"community-operators-gtr2j\" (UID: \"ecd2e077-5a83-46c0-92af-d86b26f8411c\") " pod="openshift-marketplace/community-operators-gtr2j" Dec 13 09:40:35 crc kubenswrapper[4971]: I1213 09:40:35.134867 4971 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/ecd2e077-5a83-46c0-92af-d86b26f8411c-utilities\") pod \"community-operators-gtr2j\" (UID: \"ecd2e077-5a83-46c0-92af-d86b26f8411c\") " pod="openshift-marketplace/community-operators-gtr2j" Dec 13 09:40:35 crc kubenswrapper[4971]: I1213 09:40:35.134992 4971 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-mxbtj\" (UniqueName: \"kubernetes.io/projected/ecd2e077-5a83-46c0-92af-d86b26f8411c-kube-api-access-mxbtj\") pod \"community-operators-gtr2j\" (UID: \"ecd2e077-5a83-46c0-92af-d86b26f8411c\") " pod="openshift-marketplace/community-operators-gtr2j" Dec 13 09:40:35 crc kubenswrapper[4971]: I1213 09:40:35.135167 4971 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/ecd2e077-5a83-46c0-92af-d86b26f8411c-catalog-content\") pod \"community-operators-gtr2j\" (UID: \"ecd2e077-5a83-46c0-92af-d86b26f8411c\") " pod="openshift-marketplace/community-operators-gtr2j" Dec 13 09:40:35 crc kubenswrapper[4971]: I1213 09:40:35.135438 4971 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/ecd2e077-5a83-46c0-92af-d86b26f8411c-utilities\") pod \"community-operators-gtr2j\" (UID: \"ecd2e077-5a83-46c0-92af-d86b26f8411c\") " pod="openshift-marketplace/community-operators-gtr2j" Dec 13 09:40:35 crc kubenswrapper[4971]: I1213 09:40:35.135773 4971 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/ecd2e077-5a83-46c0-92af-d86b26f8411c-catalog-content\") pod \"community-operators-gtr2j\" (UID: \"ecd2e077-5a83-46c0-92af-d86b26f8411c\") " pod="openshift-marketplace/community-operators-gtr2j" Dec 13 09:40:35 crc kubenswrapper[4971]: I1213 09:40:35.161794 4971 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-mxbtj\" (UniqueName: \"kubernetes.io/projected/ecd2e077-5a83-46c0-92af-d86b26f8411c-kube-api-access-mxbtj\") pod \"community-operators-gtr2j\" (UID: \"ecd2e077-5a83-46c0-92af-d86b26f8411c\") " pod="openshift-marketplace/community-operators-gtr2j" Dec 13 09:40:35 crc kubenswrapper[4971]: I1213 09:40:35.244243 4971 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-gtr2j" Dec 13 09:40:35 crc kubenswrapper[4971]: I1213 09:40:35.936474 4971 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-gtr2j"] Dec 13 09:40:36 crc kubenswrapper[4971]: I1213 09:40:36.002426 4971 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-gtr2j" event={"ID":"ecd2e077-5a83-46c0-92af-d86b26f8411c","Type":"ContainerStarted","Data":"1ac971f0228aebc1511ac0cb171cf0e4fc944bc180a177c550b2ffb545f0714e"} Dec 13 09:40:37 crc kubenswrapper[4971]: I1213 09:40:37.018894 4971 generic.go:334] "Generic (PLEG): container finished" podID="ecd2e077-5a83-46c0-92af-d86b26f8411c" containerID="b9607774c79cac370e55a8111e8b56b9784a008bc90157f8aab3579aabada394" exitCode=0 Dec 13 09:40:37 crc kubenswrapper[4971]: I1213 09:40:37.018985 4971 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-gtr2j" event={"ID":"ecd2e077-5a83-46c0-92af-d86b26f8411c","Type":"ContainerDied","Data":"b9607774c79cac370e55a8111e8b56b9784a008bc90157f8aab3579aabada394"} Dec 13 09:40:39 crc kubenswrapper[4971]: I1213 09:40:39.044830 4971 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-gtr2j" event={"ID":"ecd2e077-5a83-46c0-92af-d86b26f8411c","Type":"ContainerStarted","Data":"7a04d5305c1c69e4c0835004cdc829ccfb1b82368610ca7e637d142456d11be1"} Dec 13 09:40:40 crc kubenswrapper[4971]: I1213 09:40:40.057168 4971 generic.go:334] "Generic (PLEG): container finished" podID="ecd2e077-5a83-46c0-92af-d86b26f8411c" containerID="7a04d5305c1c69e4c0835004cdc829ccfb1b82368610ca7e637d142456d11be1" exitCode=0 Dec 13 09:40:40 crc kubenswrapper[4971]: I1213 09:40:40.057218 4971 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-gtr2j" event={"ID":"ecd2e077-5a83-46c0-92af-d86b26f8411c","Type":"ContainerDied","Data":"7a04d5305c1c69e4c0835004cdc829ccfb1b82368610ca7e637d142456d11be1"} Dec 13 09:40:41 crc kubenswrapper[4971]: I1213 09:40:41.073002 4971 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-gtr2j" event={"ID":"ecd2e077-5a83-46c0-92af-d86b26f8411c","Type":"ContainerStarted","Data":"a1ff570c36c6e4d793ee3e43c1ee769eac621f90b23858ac6d557600e6854b67"} Dec 13 09:40:41 crc kubenswrapper[4971]: I1213 09:40:41.097352 4971 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/community-operators-gtr2j" podStartSLOduration=3.6509900330000002 podStartE2EDuration="7.097326924s" podCreationTimestamp="2025-12-13 09:40:34 +0000 UTC" firstStartedPulling="2025-12-13 09:40:37.021959139 +0000 UTC m=+10293.626368577" lastFinishedPulling="2025-12-13 09:40:40.46829602 +0000 UTC m=+10297.072705468" observedRunningTime="2025-12-13 09:40:41.092295651 +0000 UTC m=+10297.696705109" watchObservedRunningTime="2025-12-13 09:40:41.097326924 +0000 UTC m=+10297.701736372" Dec 13 09:40:45 crc kubenswrapper[4971]: I1213 09:40:45.246607 4971 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/community-operators-gtr2j" Dec 13 09:40:45 crc kubenswrapper[4971]: I1213 09:40:45.247714 4971 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/community-operators-gtr2j" Dec 13 09:40:45 crc kubenswrapper[4971]: I1213 09:40:45.294799 4971 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/community-operators-gtr2j" Dec 13 09:40:46 crc kubenswrapper[4971]: I1213 09:40:46.187039 4971 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/community-operators-gtr2j" Dec 13 09:40:46 crc kubenswrapper[4971]: I1213 09:40:46.265926 4971 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-gtr2j"] Dec 13 09:40:48 crc kubenswrapper[4971]: I1213 09:40:48.140100 4971 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/community-operators-gtr2j" podUID="ecd2e077-5a83-46c0-92af-d86b26f8411c" containerName="registry-server" containerID="cri-o://a1ff570c36c6e4d793ee3e43c1ee769eac621f90b23858ac6d557600e6854b67" gracePeriod=2 Dec 13 09:40:48 crc kubenswrapper[4971]: I1213 09:40:48.690515 4971 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-gtr2j" Dec 13 09:40:48 crc kubenswrapper[4971]: I1213 09:40:48.779303 4971 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/ecd2e077-5a83-46c0-92af-d86b26f8411c-catalog-content\") pod \"ecd2e077-5a83-46c0-92af-d86b26f8411c\" (UID: \"ecd2e077-5a83-46c0-92af-d86b26f8411c\") " Dec 13 09:40:48 crc kubenswrapper[4971]: I1213 09:40:48.779496 4971 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/ecd2e077-5a83-46c0-92af-d86b26f8411c-utilities\") pod \"ecd2e077-5a83-46c0-92af-d86b26f8411c\" (UID: \"ecd2e077-5a83-46c0-92af-d86b26f8411c\") " Dec 13 09:40:48 crc kubenswrapper[4971]: I1213 09:40:48.779563 4971 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-mxbtj\" (UniqueName: \"kubernetes.io/projected/ecd2e077-5a83-46c0-92af-d86b26f8411c-kube-api-access-mxbtj\") pod \"ecd2e077-5a83-46c0-92af-d86b26f8411c\" (UID: \"ecd2e077-5a83-46c0-92af-d86b26f8411c\") " Dec 13 09:40:48 crc kubenswrapper[4971]: I1213 09:40:48.780287 4971 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/ecd2e077-5a83-46c0-92af-d86b26f8411c-utilities" (OuterVolumeSpecName: "utilities") pod "ecd2e077-5a83-46c0-92af-d86b26f8411c" (UID: "ecd2e077-5a83-46c0-92af-d86b26f8411c"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 13 09:40:48 crc kubenswrapper[4971]: I1213 09:40:48.780937 4971 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/ecd2e077-5a83-46c0-92af-d86b26f8411c-utilities\") on node \"crc\" DevicePath \"\"" Dec 13 09:40:48 crc kubenswrapper[4971]: I1213 09:40:48.787024 4971 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/ecd2e077-5a83-46c0-92af-d86b26f8411c-kube-api-access-mxbtj" (OuterVolumeSpecName: "kube-api-access-mxbtj") pod "ecd2e077-5a83-46c0-92af-d86b26f8411c" (UID: "ecd2e077-5a83-46c0-92af-d86b26f8411c"). InnerVolumeSpecName "kube-api-access-mxbtj". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 13 09:40:48 crc kubenswrapper[4971]: I1213 09:40:48.836112 4971 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/ecd2e077-5a83-46c0-92af-d86b26f8411c-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "ecd2e077-5a83-46c0-92af-d86b26f8411c" (UID: "ecd2e077-5a83-46c0-92af-d86b26f8411c"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 13 09:40:48 crc kubenswrapper[4971]: I1213 09:40:48.884578 4971 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-mxbtj\" (UniqueName: \"kubernetes.io/projected/ecd2e077-5a83-46c0-92af-d86b26f8411c-kube-api-access-mxbtj\") on node \"crc\" DevicePath \"\"" Dec 13 09:40:48 crc kubenswrapper[4971]: I1213 09:40:48.884761 4971 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/ecd2e077-5a83-46c0-92af-d86b26f8411c-catalog-content\") on node \"crc\" DevicePath \"\"" Dec 13 09:40:49 crc kubenswrapper[4971]: I1213 09:40:49.153726 4971 generic.go:334] "Generic (PLEG): container finished" podID="ecd2e077-5a83-46c0-92af-d86b26f8411c" containerID="a1ff570c36c6e4d793ee3e43c1ee769eac621f90b23858ac6d557600e6854b67" exitCode=0 Dec 13 09:40:49 crc kubenswrapper[4971]: I1213 09:40:49.153812 4971 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-gtr2j" event={"ID":"ecd2e077-5a83-46c0-92af-d86b26f8411c","Type":"ContainerDied","Data":"a1ff570c36c6e4d793ee3e43c1ee769eac621f90b23858ac6d557600e6854b67"} Dec 13 09:40:49 crc kubenswrapper[4971]: I1213 09:40:49.154095 4971 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-gtr2j" event={"ID":"ecd2e077-5a83-46c0-92af-d86b26f8411c","Type":"ContainerDied","Data":"1ac971f0228aebc1511ac0cb171cf0e4fc944bc180a177c550b2ffb545f0714e"} Dec 13 09:40:49 crc kubenswrapper[4971]: I1213 09:40:49.154118 4971 scope.go:117] "RemoveContainer" containerID="a1ff570c36c6e4d793ee3e43c1ee769eac621f90b23858ac6d557600e6854b67" Dec 13 09:40:49 crc kubenswrapper[4971]: I1213 09:40:49.153817 4971 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-gtr2j" Dec 13 09:40:49 crc kubenswrapper[4971]: I1213 09:40:49.186484 4971 scope.go:117] "RemoveContainer" containerID="7a04d5305c1c69e4c0835004cdc829ccfb1b82368610ca7e637d142456d11be1" Dec 13 09:40:49 crc kubenswrapper[4971]: I1213 09:40:49.202728 4971 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-gtr2j"] Dec 13 09:40:49 crc kubenswrapper[4971]: I1213 09:40:49.215987 4971 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/community-operators-gtr2j"] Dec 13 09:40:49 crc kubenswrapper[4971]: I1213 09:40:49.219366 4971 scope.go:117] "RemoveContainer" containerID="b9607774c79cac370e55a8111e8b56b9784a008bc90157f8aab3579aabada394" Dec 13 09:40:49 crc kubenswrapper[4971]: I1213 09:40:49.263278 4971 scope.go:117] "RemoveContainer" containerID="a1ff570c36c6e4d793ee3e43c1ee769eac621f90b23858ac6d557600e6854b67" Dec 13 09:40:49 crc kubenswrapper[4971]: E1213 09:40:49.263889 4971 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"a1ff570c36c6e4d793ee3e43c1ee769eac621f90b23858ac6d557600e6854b67\": container with ID starting with a1ff570c36c6e4d793ee3e43c1ee769eac621f90b23858ac6d557600e6854b67 not found: ID does not exist" containerID="a1ff570c36c6e4d793ee3e43c1ee769eac621f90b23858ac6d557600e6854b67" Dec 13 09:40:49 crc kubenswrapper[4971]: I1213 09:40:49.263934 4971 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"a1ff570c36c6e4d793ee3e43c1ee769eac621f90b23858ac6d557600e6854b67"} err="failed to get container status \"a1ff570c36c6e4d793ee3e43c1ee769eac621f90b23858ac6d557600e6854b67\": rpc error: code = NotFound desc = could not find container \"a1ff570c36c6e4d793ee3e43c1ee769eac621f90b23858ac6d557600e6854b67\": container with ID starting with a1ff570c36c6e4d793ee3e43c1ee769eac621f90b23858ac6d557600e6854b67 not found: ID does not exist" Dec 13 09:40:49 crc kubenswrapper[4971]: I1213 09:40:49.263957 4971 scope.go:117] "RemoveContainer" containerID="7a04d5305c1c69e4c0835004cdc829ccfb1b82368610ca7e637d142456d11be1" Dec 13 09:40:49 crc kubenswrapper[4971]: E1213 09:40:49.264438 4971 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"7a04d5305c1c69e4c0835004cdc829ccfb1b82368610ca7e637d142456d11be1\": container with ID starting with 7a04d5305c1c69e4c0835004cdc829ccfb1b82368610ca7e637d142456d11be1 not found: ID does not exist" containerID="7a04d5305c1c69e4c0835004cdc829ccfb1b82368610ca7e637d142456d11be1" Dec 13 09:40:49 crc kubenswrapper[4971]: I1213 09:40:49.264491 4971 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"7a04d5305c1c69e4c0835004cdc829ccfb1b82368610ca7e637d142456d11be1"} err="failed to get container status \"7a04d5305c1c69e4c0835004cdc829ccfb1b82368610ca7e637d142456d11be1\": rpc error: code = NotFound desc = could not find container \"7a04d5305c1c69e4c0835004cdc829ccfb1b82368610ca7e637d142456d11be1\": container with ID starting with 7a04d5305c1c69e4c0835004cdc829ccfb1b82368610ca7e637d142456d11be1 not found: ID does not exist" Dec 13 09:40:49 crc kubenswrapper[4971]: I1213 09:40:49.264511 4971 scope.go:117] "RemoveContainer" containerID="b9607774c79cac370e55a8111e8b56b9784a008bc90157f8aab3579aabada394" Dec 13 09:40:49 crc kubenswrapper[4971]: E1213 09:40:49.264880 4971 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"b9607774c79cac370e55a8111e8b56b9784a008bc90157f8aab3579aabada394\": container with ID starting with b9607774c79cac370e55a8111e8b56b9784a008bc90157f8aab3579aabada394 not found: ID does not exist" containerID="b9607774c79cac370e55a8111e8b56b9784a008bc90157f8aab3579aabada394" Dec 13 09:40:49 crc kubenswrapper[4971]: I1213 09:40:49.264908 4971 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"b9607774c79cac370e55a8111e8b56b9784a008bc90157f8aab3579aabada394"} err="failed to get container status \"b9607774c79cac370e55a8111e8b56b9784a008bc90157f8aab3579aabada394\": rpc error: code = NotFound desc = could not find container \"b9607774c79cac370e55a8111e8b56b9784a008bc90157f8aab3579aabada394\": container with ID starting with b9607774c79cac370e55a8111e8b56b9784a008bc90157f8aab3579aabada394 not found: ID does not exist" Dec 13 09:40:49 crc kubenswrapper[4971]: I1213 09:40:49.781461 4971 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="ecd2e077-5a83-46c0-92af-d86b26f8411c" path="/var/lib/kubelet/pods/ecd2e077-5a83-46c0-92af-d86b26f8411c/volumes" Dec 13 09:41:16 crc kubenswrapper[4971]: I1213 09:41:16.426662 4971 generic.go:334] "Generic (PLEG): container finished" podID="abaeac3e-11a1-40b0-b694-26822d0c02bf" containerID="fec73e3f45b80e5a35f856d1e51ff9c601d94e21be4639194920cd60922fec06" exitCode=0 Dec 13 09:41:16 crc kubenswrapper[4971]: I1213 09:41:16.426834 4971 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/horizontest-tests-horizontest" event={"ID":"abaeac3e-11a1-40b0-b694-26822d0c02bf","Type":"ContainerDied","Data":"fec73e3f45b80e5a35f856d1e51ff9c601d94e21be4639194920cd60922fec06"} Dec 13 09:41:17 crc kubenswrapper[4971]: I1213 09:41:17.854900 4971 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/horizontest-tests-horizontest" Dec 13 09:41:17 crc kubenswrapper[4971]: I1213 09:41:17.878614 4971 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ca-certs\" (UniqueName: \"kubernetes.io/secret/abaeac3e-11a1-40b0-b694-26822d0c02bf-ca-certs\") pod \"abaeac3e-11a1-40b0-b694-26822d0c02bf\" (UID: \"abaeac3e-11a1-40b0-b694-26822d0c02bf\") " Dec 13 09:41:17 crc kubenswrapper[4971]: I1213 09:41:17.878706 4971 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"test-operator-clouds-config\" (UniqueName: \"kubernetes.io/configmap/abaeac3e-11a1-40b0-b694-26822d0c02bf-test-operator-clouds-config\") pod \"abaeac3e-11a1-40b0-b694-26822d0c02bf\" (UID: \"abaeac3e-11a1-40b0-b694-26822d0c02bf\") " Dec 13 09:41:17 crc kubenswrapper[4971]: I1213 09:41:17.878818 4971 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"test-operator-ephemeral-temporary\" (UniqueName: \"kubernetes.io/empty-dir/abaeac3e-11a1-40b0-b694-26822d0c02bf-test-operator-ephemeral-temporary\") pod \"abaeac3e-11a1-40b0-b694-26822d0c02bf\" (UID: \"abaeac3e-11a1-40b0-b694-26822d0c02bf\") " Dec 13 09:41:17 crc kubenswrapper[4971]: I1213 09:41:17.878921 4971 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-czd2t\" (UniqueName: \"kubernetes.io/projected/abaeac3e-11a1-40b0-b694-26822d0c02bf-kube-api-access-czd2t\") pod \"abaeac3e-11a1-40b0-b694-26822d0c02bf\" (UID: \"abaeac3e-11a1-40b0-b694-26822d0c02bf\") " Dec 13 09:41:17 crc kubenswrapper[4971]: I1213 09:41:17.878959 4971 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"test-operator-logs\" (UniqueName: \"kubernetes.io/local-volume/local-storage05-crc\") pod \"abaeac3e-11a1-40b0-b694-26822d0c02bf\" (UID: \"abaeac3e-11a1-40b0-b694-26822d0c02bf\") " Dec 13 09:41:17 crc kubenswrapper[4971]: I1213 09:41:17.878977 4971 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/abaeac3e-11a1-40b0-b694-26822d0c02bf-ceph\") pod \"abaeac3e-11a1-40b0-b694-26822d0c02bf\" (UID: \"abaeac3e-11a1-40b0-b694-26822d0c02bf\") " Dec 13 09:41:17 crc kubenswrapper[4971]: I1213 09:41:17.879034 4971 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"openstack-config-secret\" (UniqueName: \"kubernetes.io/secret/abaeac3e-11a1-40b0-b694-26822d0c02bf-openstack-config-secret\") pod \"abaeac3e-11a1-40b0-b694-26822d0c02bf\" (UID: \"abaeac3e-11a1-40b0-b694-26822d0c02bf\") " Dec 13 09:41:17 crc kubenswrapper[4971]: I1213 09:41:17.879103 4971 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"test-operator-ephemeral-workdir\" (UniqueName: \"kubernetes.io/empty-dir/abaeac3e-11a1-40b0-b694-26822d0c02bf-test-operator-ephemeral-workdir\") pod \"abaeac3e-11a1-40b0-b694-26822d0c02bf\" (UID: \"abaeac3e-11a1-40b0-b694-26822d0c02bf\") " Dec 13 09:41:17 crc kubenswrapper[4971]: I1213 09:41:17.879814 4971 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/abaeac3e-11a1-40b0-b694-26822d0c02bf-test-operator-ephemeral-temporary" (OuterVolumeSpecName: "test-operator-ephemeral-temporary") pod "abaeac3e-11a1-40b0-b694-26822d0c02bf" (UID: "abaeac3e-11a1-40b0-b694-26822d0c02bf"). InnerVolumeSpecName "test-operator-ephemeral-temporary". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 13 09:41:17 crc kubenswrapper[4971]: I1213 09:41:17.887906 4971 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/abaeac3e-11a1-40b0-b694-26822d0c02bf-kube-api-access-czd2t" (OuterVolumeSpecName: "kube-api-access-czd2t") pod "abaeac3e-11a1-40b0-b694-26822d0c02bf" (UID: "abaeac3e-11a1-40b0-b694-26822d0c02bf"). InnerVolumeSpecName "kube-api-access-czd2t". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 13 09:41:17 crc kubenswrapper[4971]: I1213 09:41:17.887930 4971 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/local-volume/local-storage05-crc" (OuterVolumeSpecName: "test-operator-logs") pod "abaeac3e-11a1-40b0-b694-26822d0c02bf" (UID: "abaeac3e-11a1-40b0-b694-26822d0c02bf"). InnerVolumeSpecName "local-storage05-crc". PluginName "kubernetes.io/local-volume", VolumeGidValue "" Dec 13 09:41:17 crc kubenswrapper[4971]: I1213 09:41:17.938058 4971 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/abaeac3e-11a1-40b0-b694-26822d0c02bf-ceph" (OuterVolumeSpecName: "ceph") pod "abaeac3e-11a1-40b0-b694-26822d0c02bf" (UID: "abaeac3e-11a1-40b0-b694-26822d0c02bf"). InnerVolumeSpecName "ceph". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 13 09:41:17 crc kubenswrapper[4971]: I1213 09:41:17.982578 4971 reconciler_common.go:293] "Volume detached for volume \"test-operator-ephemeral-temporary\" (UniqueName: \"kubernetes.io/empty-dir/abaeac3e-11a1-40b0-b694-26822d0c02bf-test-operator-ephemeral-temporary\") on node \"crc\" DevicePath \"\"" Dec 13 09:41:17 crc kubenswrapper[4971]: I1213 09:41:17.982655 4971 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-czd2t\" (UniqueName: \"kubernetes.io/projected/abaeac3e-11a1-40b0-b694-26822d0c02bf-kube-api-access-czd2t\") on node \"crc\" DevicePath \"\"" Dec 13 09:41:17 crc kubenswrapper[4971]: I1213 09:41:17.982817 4971 reconciler_common.go:286] "operationExecutor.UnmountDevice started for volume \"local-storage05-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage05-crc\") on node \"crc\" " Dec 13 09:41:17 crc kubenswrapper[4971]: I1213 09:41:17.982884 4971 reconciler_common.go:293] "Volume detached for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/abaeac3e-11a1-40b0-b694-26822d0c02bf-ceph\") on node \"crc\" DevicePath \"\"" Dec 13 09:41:17 crc kubenswrapper[4971]: I1213 09:41:17.986368 4971 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/abaeac3e-11a1-40b0-b694-26822d0c02bf-openstack-config-secret" (OuterVolumeSpecName: "openstack-config-secret") pod "abaeac3e-11a1-40b0-b694-26822d0c02bf" (UID: "abaeac3e-11a1-40b0-b694-26822d0c02bf"). InnerVolumeSpecName "openstack-config-secret". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 13 09:41:17 crc kubenswrapper[4971]: I1213 09:41:17.998643 4971 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/abaeac3e-11a1-40b0-b694-26822d0c02bf-test-operator-clouds-config" (OuterVolumeSpecName: "test-operator-clouds-config") pod "abaeac3e-11a1-40b0-b694-26822d0c02bf" (UID: "abaeac3e-11a1-40b0-b694-26822d0c02bf"). InnerVolumeSpecName "test-operator-clouds-config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 13 09:41:18 crc kubenswrapper[4971]: I1213 09:41:18.012153 4971 operation_generator.go:917] UnmountDevice succeeded for volume "local-storage05-crc" (UniqueName: "kubernetes.io/local-volume/local-storage05-crc") on node "crc" Dec 13 09:41:18 crc kubenswrapper[4971]: I1213 09:41:18.022091 4971 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/abaeac3e-11a1-40b0-b694-26822d0c02bf-ca-certs" (OuterVolumeSpecName: "ca-certs") pod "abaeac3e-11a1-40b0-b694-26822d0c02bf" (UID: "abaeac3e-11a1-40b0-b694-26822d0c02bf"). InnerVolumeSpecName "ca-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 13 09:41:18 crc kubenswrapper[4971]: I1213 09:41:18.084929 4971 reconciler_common.go:293] "Volume detached for volume \"ca-certs\" (UniqueName: \"kubernetes.io/secret/abaeac3e-11a1-40b0-b694-26822d0c02bf-ca-certs\") on node \"crc\" DevicePath \"\"" Dec 13 09:41:18 crc kubenswrapper[4971]: I1213 09:41:18.084968 4971 reconciler_common.go:293] "Volume detached for volume \"test-operator-clouds-config\" (UniqueName: \"kubernetes.io/configmap/abaeac3e-11a1-40b0-b694-26822d0c02bf-test-operator-clouds-config\") on node \"crc\" DevicePath \"\"" Dec 13 09:41:18 crc kubenswrapper[4971]: I1213 09:41:18.084983 4971 reconciler_common.go:293] "Volume detached for volume \"local-storage05-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage05-crc\") on node \"crc\" DevicePath \"\"" Dec 13 09:41:18 crc kubenswrapper[4971]: I1213 09:41:18.084993 4971 reconciler_common.go:293] "Volume detached for volume \"openstack-config-secret\" (UniqueName: \"kubernetes.io/secret/abaeac3e-11a1-40b0-b694-26822d0c02bf-openstack-config-secret\") on node \"crc\" DevicePath \"\"" Dec 13 09:41:18 crc kubenswrapper[4971]: I1213 09:41:18.151945 4971 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/abaeac3e-11a1-40b0-b694-26822d0c02bf-test-operator-ephemeral-workdir" (OuterVolumeSpecName: "test-operator-ephemeral-workdir") pod "abaeac3e-11a1-40b0-b694-26822d0c02bf" (UID: "abaeac3e-11a1-40b0-b694-26822d0c02bf"). InnerVolumeSpecName "test-operator-ephemeral-workdir". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 13 09:41:18 crc kubenswrapper[4971]: I1213 09:41:18.186948 4971 reconciler_common.go:293] "Volume detached for volume \"test-operator-ephemeral-workdir\" (UniqueName: \"kubernetes.io/empty-dir/abaeac3e-11a1-40b0-b694-26822d0c02bf-test-operator-ephemeral-workdir\") on node \"crc\" DevicePath \"\"" Dec 13 09:41:18 crc kubenswrapper[4971]: I1213 09:41:18.451020 4971 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/horizontest-tests-horizontest" event={"ID":"abaeac3e-11a1-40b0-b694-26822d0c02bf","Type":"ContainerDied","Data":"0fb2450c7e73fc3a279482d7e35206a08454c790da3cf44dec1a7b8359244714"} Dec 13 09:41:18 crc kubenswrapper[4971]: I1213 09:41:18.451070 4971 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="0fb2450c7e73fc3a279482d7e35206a08454c790da3cf44dec1a7b8359244714" Dec 13 09:41:18 crc kubenswrapper[4971]: I1213 09:41:18.451087 4971 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/horizontest-tests-horizontest" Dec 13 09:41:29 crc kubenswrapper[4971]: I1213 09:41:29.928362 4971 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/test-operator-logs-pod-horizontest-horizontest-tests-horizontest"] Dec 13 09:41:29 crc kubenswrapper[4971]: E1213 09:41:29.929556 4971 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="abaeac3e-11a1-40b0-b694-26822d0c02bf" containerName="horizontest-tests-horizontest" Dec 13 09:41:29 crc kubenswrapper[4971]: I1213 09:41:29.929571 4971 state_mem.go:107] "Deleted CPUSet assignment" podUID="abaeac3e-11a1-40b0-b694-26822d0c02bf" containerName="horizontest-tests-horizontest" Dec 13 09:41:29 crc kubenswrapper[4971]: E1213 09:41:29.929583 4971 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="ecd2e077-5a83-46c0-92af-d86b26f8411c" containerName="extract-utilities" Dec 13 09:41:29 crc kubenswrapper[4971]: I1213 09:41:29.929591 4971 state_mem.go:107] "Deleted CPUSet assignment" podUID="ecd2e077-5a83-46c0-92af-d86b26f8411c" containerName="extract-utilities" Dec 13 09:41:29 crc kubenswrapper[4971]: E1213 09:41:29.929643 4971 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="ecd2e077-5a83-46c0-92af-d86b26f8411c" containerName="registry-server" Dec 13 09:41:29 crc kubenswrapper[4971]: I1213 09:41:29.929653 4971 state_mem.go:107] "Deleted CPUSet assignment" podUID="ecd2e077-5a83-46c0-92af-d86b26f8411c" containerName="registry-server" Dec 13 09:41:29 crc kubenswrapper[4971]: E1213 09:41:29.929667 4971 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="ecd2e077-5a83-46c0-92af-d86b26f8411c" containerName="extract-content" Dec 13 09:41:29 crc kubenswrapper[4971]: I1213 09:41:29.929673 4971 state_mem.go:107] "Deleted CPUSet assignment" podUID="ecd2e077-5a83-46c0-92af-d86b26f8411c" containerName="extract-content" Dec 13 09:41:29 crc kubenswrapper[4971]: I1213 09:41:29.929874 4971 memory_manager.go:354] "RemoveStaleState removing state" podUID="ecd2e077-5a83-46c0-92af-d86b26f8411c" containerName="registry-server" Dec 13 09:41:29 crc kubenswrapper[4971]: I1213 09:41:29.929892 4971 memory_manager.go:354] "RemoveStaleState removing state" podUID="abaeac3e-11a1-40b0-b694-26822d0c02bf" containerName="horizontest-tests-horizontest" Dec 13 09:41:29 crc kubenswrapper[4971]: I1213 09:41:29.930794 4971 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/test-operator-logs-pod-horizontest-horizontest-tests-horizontest" Dec 13 09:41:29 crc kubenswrapper[4971]: I1213 09:41:29.955465 4971 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/test-operator-logs-pod-horizontest-horizontest-tests-horizontest"] Dec 13 09:41:30 crc kubenswrapper[4971]: I1213 09:41:30.087659 4971 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-4jzzl\" (UniqueName: \"kubernetes.io/projected/73a838a8-433b-4ddf-ab44-1527de29aa6f-kube-api-access-4jzzl\") pod \"test-operator-logs-pod-horizontest-horizontest-tests-horizontest\" (UID: \"73a838a8-433b-4ddf-ab44-1527de29aa6f\") " pod="openstack/test-operator-logs-pod-horizontest-horizontest-tests-horizontest" Dec 13 09:41:30 crc kubenswrapper[4971]: I1213 09:41:30.088150 4971 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"local-storage05-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage05-crc\") pod \"test-operator-logs-pod-horizontest-horizontest-tests-horizontest\" (UID: \"73a838a8-433b-4ddf-ab44-1527de29aa6f\") " pod="openstack/test-operator-logs-pod-horizontest-horizontest-tests-horizontest" Dec 13 09:41:30 crc kubenswrapper[4971]: I1213 09:41:30.195365 4971 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-4jzzl\" (UniqueName: \"kubernetes.io/projected/73a838a8-433b-4ddf-ab44-1527de29aa6f-kube-api-access-4jzzl\") pod \"test-operator-logs-pod-horizontest-horizontest-tests-horizontest\" (UID: \"73a838a8-433b-4ddf-ab44-1527de29aa6f\") " pod="openstack/test-operator-logs-pod-horizontest-horizontest-tests-horizontest" Dec 13 09:41:30 crc kubenswrapper[4971]: I1213 09:41:30.195573 4971 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"local-storage05-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage05-crc\") pod \"test-operator-logs-pod-horizontest-horizontest-tests-horizontest\" (UID: \"73a838a8-433b-4ddf-ab44-1527de29aa6f\") " pod="openstack/test-operator-logs-pod-horizontest-horizontest-tests-horizontest" Dec 13 09:41:30 crc kubenswrapper[4971]: I1213 09:41:30.196707 4971 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"local-storage05-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage05-crc\") pod \"test-operator-logs-pod-horizontest-horizontest-tests-horizontest\" (UID: \"73a838a8-433b-4ddf-ab44-1527de29aa6f\") device mount path \"/mnt/openstack/pv05\"" pod="openstack/test-operator-logs-pod-horizontest-horizontest-tests-horizontest" Dec 13 09:41:30 crc kubenswrapper[4971]: I1213 09:41:30.235680 4971 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"local-storage05-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage05-crc\") pod \"test-operator-logs-pod-horizontest-horizontest-tests-horizontest\" (UID: \"73a838a8-433b-4ddf-ab44-1527de29aa6f\") " pod="openstack/test-operator-logs-pod-horizontest-horizontest-tests-horizontest" Dec 13 09:41:30 crc kubenswrapper[4971]: I1213 09:41:30.238446 4971 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-4jzzl\" (UniqueName: \"kubernetes.io/projected/73a838a8-433b-4ddf-ab44-1527de29aa6f-kube-api-access-4jzzl\") pod \"test-operator-logs-pod-horizontest-horizontest-tests-horizontest\" (UID: \"73a838a8-433b-4ddf-ab44-1527de29aa6f\") " pod="openstack/test-operator-logs-pod-horizontest-horizontest-tests-horizontest" Dec 13 09:41:30 crc kubenswrapper[4971]: I1213 09:41:30.259824 4971 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/test-operator-logs-pod-horizontest-horizontest-tests-horizontest" Dec 13 09:41:30 crc kubenswrapper[4971]: E1213 09:41:30.260109 4971 kubelet_pods.go:538] "Hostname for pod was too long, truncated it" podName="test-operator-logs-pod-horizontest-horizontest-tests-horizontest" hostnameMaxLen=63 truncatedHostname="test-operator-logs-pod-horizontest-horizontest-tests-horizontes" Dec 13 09:41:30 crc kubenswrapper[4971]: I1213 09:41:30.731550 4971 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/test-operator-logs-pod-horizontest-horizontest-tests-horizontest"] Dec 13 09:41:30 crc kubenswrapper[4971]: E1213 09:41:30.739656 4971 kubelet_pods.go:538] "Hostname for pod was too long, truncated it" podName="test-operator-logs-pod-horizontest-horizontest-tests-horizontest" hostnameMaxLen=63 truncatedHostname="test-operator-logs-pod-horizontest-horizontest-tests-horizontes" Dec 13 09:41:31 crc kubenswrapper[4971]: E1213 09:41:31.247824 4971 kubelet_pods.go:538] "Hostname for pod was too long, truncated it" podName="test-operator-logs-pod-horizontest-horizontest-tests-horizontest" hostnameMaxLen=63 truncatedHostname="test-operator-logs-pod-horizontest-horizontest-tests-horizontes" Dec 13 09:41:31 crc kubenswrapper[4971]: I1213 09:41:31.591185 4971 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/test-operator-logs-pod-horizontest-horizontest-tests-horizontest" event={"ID":"73a838a8-433b-4ddf-ab44-1527de29aa6f","Type":"ContainerStarted","Data":"7848c5498788ce4d65757aa933b70db4e2878d7afba7a469e362ed4808683faf"} Dec 13 09:41:31 crc kubenswrapper[4971]: I1213 09:41:31.591311 4971 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/test-operator-logs-pod-horizontest-horizontest-tests-horizontest" event={"ID":"73a838a8-433b-4ddf-ab44-1527de29aa6f","Type":"ContainerStarted","Data":"cfbe42339a794433a97f25fdcf8180ec555d263eab026e86237d2da085febb01"} Dec 13 09:41:31 crc kubenswrapper[4971]: E1213 09:41:31.592492 4971 kubelet_pods.go:538] "Hostname for pod was too long, truncated it" podName="test-operator-logs-pod-horizontest-horizontest-tests-horizontest" hostnameMaxLen=63 truncatedHostname="test-operator-logs-pod-horizontest-horizontest-tests-horizontes" Dec 13 09:41:31 crc kubenswrapper[4971]: I1213 09:41:31.622773 4971 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/test-operator-logs-pod-horizontest-horizontest-tests-horizontest" podStartSLOduration=2.116337957 podStartE2EDuration="2.622749177s" podCreationTimestamp="2025-12-13 09:41:29 +0000 UTC" firstStartedPulling="2025-12-13 09:41:30.741170859 +0000 UTC m=+10347.345580317" lastFinishedPulling="2025-12-13 09:41:31.247582089 +0000 UTC m=+10347.851991537" observedRunningTime="2025-12-13 09:41:31.612580849 +0000 UTC m=+10348.216990307" watchObservedRunningTime="2025-12-13 09:41:31.622749177 +0000 UTC m=+10348.227158625" Dec 13 09:41:32 crc kubenswrapper[4971]: E1213 09:41:32.603341 4971 kubelet_pods.go:538] "Hostname for pod was too long, truncated it" podName="test-operator-logs-pod-horizontest-horizontest-tests-horizontest" hostnameMaxLen=63 truncatedHostname="test-operator-logs-pod-horizontest-horizontest-tests-horizontes" Dec 13 09:41:46 crc kubenswrapper[4971]: I1213 09:41:46.842332 4971 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-operators-2lknx"] Dec 13 09:41:46 crc kubenswrapper[4971]: I1213 09:41:46.845754 4971 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-2lknx" Dec 13 09:41:46 crc kubenswrapper[4971]: I1213 09:41:46.854376 4971 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-2lknx"] Dec 13 09:41:46 crc kubenswrapper[4971]: I1213 09:41:46.910936 4971 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/47c666f1-9904-43f0-8aa8-536cce82308a-catalog-content\") pod \"redhat-operators-2lknx\" (UID: \"47c666f1-9904-43f0-8aa8-536cce82308a\") " pod="openshift-marketplace/redhat-operators-2lknx" Dec 13 09:41:46 crc kubenswrapper[4971]: I1213 09:41:46.911034 4971 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-6x846\" (UniqueName: \"kubernetes.io/projected/47c666f1-9904-43f0-8aa8-536cce82308a-kube-api-access-6x846\") pod \"redhat-operators-2lknx\" (UID: \"47c666f1-9904-43f0-8aa8-536cce82308a\") " pod="openshift-marketplace/redhat-operators-2lknx" Dec 13 09:41:46 crc kubenswrapper[4971]: I1213 09:41:46.911081 4971 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/47c666f1-9904-43f0-8aa8-536cce82308a-utilities\") pod \"redhat-operators-2lknx\" (UID: \"47c666f1-9904-43f0-8aa8-536cce82308a\") " pod="openshift-marketplace/redhat-operators-2lknx" Dec 13 09:41:47 crc kubenswrapper[4971]: I1213 09:41:47.015069 4971 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/47c666f1-9904-43f0-8aa8-536cce82308a-catalog-content\") pod \"redhat-operators-2lknx\" (UID: \"47c666f1-9904-43f0-8aa8-536cce82308a\") " pod="openshift-marketplace/redhat-operators-2lknx" Dec 13 09:41:47 crc kubenswrapper[4971]: I1213 09:41:47.015131 4971 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-6x846\" (UniqueName: \"kubernetes.io/projected/47c666f1-9904-43f0-8aa8-536cce82308a-kube-api-access-6x846\") pod \"redhat-operators-2lknx\" (UID: \"47c666f1-9904-43f0-8aa8-536cce82308a\") " pod="openshift-marketplace/redhat-operators-2lknx" Dec 13 09:41:47 crc kubenswrapper[4971]: I1213 09:41:47.015160 4971 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/47c666f1-9904-43f0-8aa8-536cce82308a-utilities\") pod \"redhat-operators-2lknx\" (UID: \"47c666f1-9904-43f0-8aa8-536cce82308a\") " pod="openshift-marketplace/redhat-operators-2lknx" Dec 13 09:41:47 crc kubenswrapper[4971]: I1213 09:41:47.015878 4971 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/47c666f1-9904-43f0-8aa8-536cce82308a-catalog-content\") pod \"redhat-operators-2lknx\" (UID: \"47c666f1-9904-43f0-8aa8-536cce82308a\") " pod="openshift-marketplace/redhat-operators-2lknx" Dec 13 09:41:47 crc kubenswrapper[4971]: I1213 09:41:47.015898 4971 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/47c666f1-9904-43f0-8aa8-536cce82308a-utilities\") pod \"redhat-operators-2lknx\" (UID: \"47c666f1-9904-43f0-8aa8-536cce82308a\") " pod="openshift-marketplace/redhat-operators-2lknx" Dec 13 09:41:47 crc kubenswrapper[4971]: I1213 09:41:47.040647 4971 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-6x846\" (UniqueName: \"kubernetes.io/projected/47c666f1-9904-43f0-8aa8-536cce82308a-kube-api-access-6x846\") pod \"redhat-operators-2lknx\" (UID: \"47c666f1-9904-43f0-8aa8-536cce82308a\") " pod="openshift-marketplace/redhat-operators-2lknx" Dec 13 09:41:47 crc kubenswrapper[4971]: I1213 09:41:47.185955 4971 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-2lknx" Dec 13 09:41:47 crc kubenswrapper[4971]: I1213 09:41:47.720586 4971 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-2lknx"] Dec 13 09:41:47 crc kubenswrapper[4971]: I1213 09:41:47.759971 4971 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-2lknx" event={"ID":"47c666f1-9904-43f0-8aa8-536cce82308a","Type":"ContainerStarted","Data":"b927222f4da5fe0987f6a4aeb57732330fccf3265998e3a96968c80aa07e67c1"} Dec 13 09:41:48 crc kubenswrapper[4971]: I1213 09:41:48.774101 4971 generic.go:334] "Generic (PLEG): container finished" podID="47c666f1-9904-43f0-8aa8-536cce82308a" containerID="4d12722693ddc54d849660b8a61b16ec8562bb76ece3de5a956b12cec797b9fa" exitCode=0 Dec 13 09:41:48 crc kubenswrapper[4971]: I1213 09:41:48.774206 4971 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-2lknx" event={"ID":"47c666f1-9904-43f0-8aa8-536cce82308a","Type":"ContainerDied","Data":"4d12722693ddc54d849660b8a61b16ec8562bb76ece3de5a956b12cec797b9fa"} Dec 13 09:41:50 crc kubenswrapper[4971]: I1213 09:41:50.825414 4971 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-2lknx" event={"ID":"47c666f1-9904-43f0-8aa8-536cce82308a","Type":"ContainerStarted","Data":"d71d0a29980ea09d5878223236e99a6f18337bb728627214599443921da810ed"} Dec 13 09:41:51 crc kubenswrapper[4971]: I1213 09:41:51.838704 4971 generic.go:334] "Generic (PLEG): container finished" podID="47c666f1-9904-43f0-8aa8-536cce82308a" containerID="d71d0a29980ea09d5878223236e99a6f18337bb728627214599443921da810ed" exitCode=0 Dec 13 09:41:51 crc kubenswrapper[4971]: I1213 09:41:51.838761 4971 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-2lknx" event={"ID":"47c666f1-9904-43f0-8aa8-536cce82308a","Type":"ContainerDied","Data":"d71d0a29980ea09d5878223236e99a6f18337bb728627214599443921da810ed"} Dec 13 09:41:52 crc kubenswrapper[4971]: I1213 09:41:52.851794 4971 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-2lknx" event={"ID":"47c666f1-9904-43f0-8aa8-536cce82308a","Type":"ContainerStarted","Data":"742f85ce02e7ab3aa88c454ceead458f2543e0029207b61d5fdf5ac7cc7383e1"} Dec 13 09:41:52 crc kubenswrapper[4971]: I1213 09:41:52.881635 4971 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-operators-2lknx" podStartSLOduration=3.333906797 podStartE2EDuration="6.881608673s" podCreationTimestamp="2025-12-13 09:41:46 +0000 UTC" firstStartedPulling="2025-12-13 09:41:48.776934202 +0000 UTC m=+10365.381343650" lastFinishedPulling="2025-12-13 09:41:52.324636078 +0000 UTC m=+10368.929045526" observedRunningTime="2025-12-13 09:41:52.877113124 +0000 UTC m=+10369.481522582" watchObservedRunningTime="2025-12-13 09:41:52.881608673 +0000 UTC m=+10369.486018131" Dec 13 09:41:57 crc kubenswrapper[4971]: I1213 09:41:57.187534 4971 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-operators-2lknx" Dec 13 09:41:57 crc kubenswrapper[4971]: I1213 09:41:57.188151 4971 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-operators-2lknx" Dec 13 09:41:57 crc kubenswrapper[4971]: I1213 09:41:57.788473 4971 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-must-gather-69gc8/must-gather-rgtmv"] Dec 13 09:41:57 crc kubenswrapper[4971]: I1213 09:41:57.792659 4971 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-must-gather-69gc8/must-gather-rgtmv"] Dec 13 09:41:57 crc kubenswrapper[4971]: I1213 09:41:57.792841 4971 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-69gc8/must-gather-rgtmv" Dec 13 09:41:57 crc kubenswrapper[4971]: I1213 09:41:57.798292 4971 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-must-gather-69gc8"/"default-dockercfg-4zcmx" Dec 13 09:41:57 crc kubenswrapper[4971]: I1213 09:41:57.798656 4971 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-must-gather-69gc8"/"kube-root-ca.crt" Dec 13 09:41:57 crc kubenswrapper[4971]: I1213 09:41:57.798831 4971 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-must-gather-69gc8"/"openshift-service-ca.crt" Dec 13 09:41:57 crc kubenswrapper[4971]: I1213 09:41:57.909416 4971 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-mfbrn\" (UniqueName: \"kubernetes.io/projected/b2ac511e-e641-4ee8-aca7-6054195a75d8-kube-api-access-mfbrn\") pod \"must-gather-rgtmv\" (UID: \"b2ac511e-e641-4ee8-aca7-6054195a75d8\") " pod="openshift-must-gather-69gc8/must-gather-rgtmv" Dec 13 09:41:57 crc kubenswrapper[4971]: I1213 09:41:57.909963 4971 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"must-gather-output\" (UniqueName: \"kubernetes.io/empty-dir/b2ac511e-e641-4ee8-aca7-6054195a75d8-must-gather-output\") pod \"must-gather-rgtmv\" (UID: \"b2ac511e-e641-4ee8-aca7-6054195a75d8\") " pod="openshift-must-gather-69gc8/must-gather-rgtmv" Dec 13 09:41:58 crc kubenswrapper[4971]: I1213 09:41:58.013341 4971 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-mfbrn\" (UniqueName: \"kubernetes.io/projected/b2ac511e-e641-4ee8-aca7-6054195a75d8-kube-api-access-mfbrn\") pod \"must-gather-rgtmv\" (UID: \"b2ac511e-e641-4ee8-aca7-6054195a75d8\") " pod="openshift-must-gather-69gc8/must-gather-rgtmv" Dec 13 09:41:58 crc kubenswrapper[4971]: I1213 09:41:58.013445 4971 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"must-gather-output\" (UniqueName: \"kubernetes.io/empty-dir/b2ac511e-e641-4ee8-aca7-6054195a75d8-must-gather-output\") pod \"must-gather-rgtmv\" (UID: \"b2ac511e-e641-4ee8-aca7-6054195a75d8\") " pod="openshift-must-gather-69gc8/must-gather-rgtmv" Dec 13 09:41:58 crc kubenswrapper[4971]: I1213 09:41:58.014215 4971 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"must-gather-output\" (UniqueName: \"kubernetes.io/empty-dir/b2ac511e-e641-4ee8-aca7-6054195a75d8-must-gather-output\") pod \"must-gather-rgtmv\" (UID: \"b2ac511e-e641-4ee8-aca7-6054195a75d8\") " pod="openshift-must-gather-69gc8/must-gather-rgtmv" Dec 13 09:41:58 crc kubenswrapper[4971]: I1213 09:41:58.037175 4971 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-mfbrn\" (UniqueName: \"kubernetes.io/projected/b2ac511e-e641-4ee8-aca7-6054195a75d8-kube-api-access-mfbrn\") pod \"must-gather-rgtmv\" (UID: \"b2ac511e-e641-4ee8-aca7-6054195a75d8\") " pod="openshift-must-gather-69gc8/must-gather-rgtmv" Dec 13 09:41:58 crc kubenswrapper[4971]: I1213 09:41:58.120941 4971 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-69gc8/must-gather-rgtmv" Dec 13 09:41:58 crc kubenswrapper[4971]: I1213 09:41:58.267675 4971 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-marketplace/redhat-operators-2lknx" podUID="47c666f1-9904-43f0-8aa8-536cce82308a" containerName="registry-server" probeResult="failure" output=< Dec 13 09:41:58 crc kubenswrapper[4971]: timeout: failed to connect service ":50051" within 1s Dec 13 09:41:58 crc kubenswrapper[4971]: > Dec 13 09:41:58 crc kubenswrapper[4971]: I1213 09:41:58.705103 4971 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-must-gather-69gc8/must-gather-rgtmv"] Dec 13 09:41:58 crc kubenswrapper[4971]: I1213 09:41:58.913977 4971 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-69gc8/must-gather-rgtmv" event={"ID":"b2ac511e-e641-4ee8-aca7-6054195a75d8","Type":"ContainerStarted","Data":"7f598cd400397647c4d4296ba86129899a86a66b4b5800d75a9fb9ff307a1e20"} Dec 13 09:42:07 crc kubenswrapper[4971]: I1213 09:42:07.256147 4971 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-operators-2lknx" Dec 13 09:42:07 crc kubenswrapper[4971]: I1213 09:42:07.321648 4971 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-operators-2lknx" Dec 13 09:42:07 crc kubenswrapper[4971]: I1213 09:42:07.512323 4971 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-2lknx"] Dec 13 09:42:09 crc kubenswrapper[4971]: I1213 09:42:09.060777 4971 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-69gc8/must-gather-rgtmv" event={"ID":"b2ac511e-e641-4ee8-aca7-6054195a75d8","Type":"ContainerStarted","Data":"423189d833cf0d72b52f4d07c31c19ad439ce4ef0db37aa7024772e15119e377"} Dec 13 09:42:09 crc kubenswrapper[4971]: I1213 09:42:09.060877 4971 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-operators-2lknx" podUID="47c666f1-9904-43f0-8aa8-536cce82308a" containerName="registry-server" containerID="cri-o://742f85ce02e7ab3aa88c454ceead458f2543e0029207b61d5fdf5ac7cc7383e1" gracePeriod=2 Dec 13 09:42:09 crc kubenswrapper[4971]: I1213 09:42:09.646377 4971 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-2lknx" Dec 13 09:42:09 crc kubenswrapper[4971]: I1213 09:42:09.741342 4971 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/47c666f1-9904-43f0-8aa8-536cce82308a-utilities\") pod \"47c666f1-9904-43f0-8aa8-536cce82308a\" (UID: \"47c666f1-9904-43f0-8aa8-536cce82308a\") " Dec 13 09:42:09 crc kubenswrapper[4971]: I1213 09:42:09.741487 4971 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-6x846\" (UniqueName: \"kubernetes.io/projected/47c666f1-9904-43f0-8aa8-536cce82308a-kube-api-access-6x846\") pod \"47c666f1-9904-43f0-8aa8-536cce82308a\" (UID: \"47c666f1-9904-43f0-8aa8-536cce82308a\") " Dec 13 09:42:09 crc kubenswrapper[4971]: I1213 09:42:09.741564 4971 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/47c666f1-9904-43f0-8aa8-536cce82308a-catalog-content\") pod \"47c666f1-9904-43f0-8aa8-536cce82308a\" (UID: \"47c666f1-9904-43f0-8aa8-536cce82308a\") " Dec 13 09:42:09 crc kubenswrapper[4971]: I1213 09:42:09.742675 4971 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/47c666f1-9904-43f0-8aa8-536cce82308a-utilities" (OuterVolumeSpecName: "utilities") pod "47c666f1-9904-43f0-8aa8-536cce82308a" (UID: "47c666f1-9904-43f0-8aa8-536cce82308a"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 13 09:42:09 crc kubenswrapper[4971]: I1213 09:42:09.757454 4971 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/47c666f1-9904-43f0-8aa8-536cce82308a-kube-api-access-6x846" (OuterVolumeSpecName: "kube-api-access-6x846") pod "47c666f1-9904-43f0-8aa8-536cce82308a" (UID: "47c666f1-9904-43f0-8aa8-536cce82308a"). InnerVolumeSpecName "kube-api-access-6x846". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 13 09:42:09 crc kubenswrapper[4971]: I1213 09:42:09.840898 4971 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/47c666f1-9904-43f0-8aa8-536cce82308a-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "47c666f1-9904-43f0-8aa8-536cce82308a" (UID: "47c666f1-9904-43f0-8aa8-536cce82308a"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 13 09:42:09 crc kubenswrapper[4971]: I1213 09:42:09.844956 4971 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/47c666f1-9904-43f0-8aa8-536cce82308a-utilities\") on node \"crc\" DevicePath \"\"" Dec 13 09:42:09 crc kubenswrapper[4971]: I1213 09:42:09.845013 4971 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-6x846\" (UniqueName: \"kubernetes.io/projected/47c666f1-9904-43f0-8aa8-536cce82308a-kube-api-access-6x846\") on node \"crc\" DevicePath \"\"" Dec 13 09:42:09 crc kubenswrapper[4971]: I1213 09:42:09.845028 4971 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/47c666f1-9904-43f0-8aa8-536cce82308a-catalog-content\") on node \"crc\" DevicePath \"\"" Dec 13 09:42:10 crc kubenswrapper[4971]: I1213 09:42:10.076317 4971 generic.go:334] "Generic (PLEG): container finished" podID="47c666f1-9904-43f0-8aa8-536cce82308a" containerID="742f85ce02e7ab3aa88c454ceead458f2543e0029207b61d5fdf5ac7cc7383e1" exitCode=0 Dec 13 09:42:10 crc kubenswrapper[4971]: I1213 09:42:10.076392 4971 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-2lknx" event={"ID":"47c666f1-9904-43f0-8aa8-536cce82308a","Type":"ContainerDied","Data":"742f85ce02e7ab3aa88c454ceead458f2543e0029207b61d5fdf5ac7cc7383e1"} Dec 13 09:42:10 crc kubenswrapper[4971]: I1213 09:42:10.076427 4971 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-2lknx" event={"ID":"47c666f1-9904-43f0-8aa8-536cce82308a","Type":"ContainerDied","Data":"b927222f4da5fe0987f6a4aeb57732330fccf3265998e3a96968c80aa07e67c1"} Dec 13 09:42:10 crc kubenswrapper[4971]: I1213 09:42:10.076450 4971 scope.go:117] "RemoveContainer" containerID="742f85ce02e7ab3aa88c454ceead458f2543e0029207b61d5fdf5ac7cc7383e1" Dec 13 09:42:10 crc kubenswrapper[4971]: I1213 09:42:10.076652 4971 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-2lknx" Dec 13 09:42:10 crc kubenswrapper[4971]: I1213 09:42:10.081990 4971 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-69gc8/must-gather-rgtmv" event={"ID":"b2ac511e-e641-4ee8-aca7-6054195a75d8","Type":"ContainerStarted","Data":"244712079a41bb332b354715ba32466e0a7638204a7d9474213db4f5299ae1bc"} Dec 13 09:42:10 crc kubenswrapper[4971]: I1213 09:42:10.107530 4971 scope.go:117] "RemoveContainer" containerID="d71d0a29980ea09d5878223236e99a6f18337bb728627214599443921da810ed" Dec 13 09:42:10 crc kubenswrapper[4971]: I1213 09:42:10.127702 4971 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-must-gather-69gc8/must-gather-rgtmv" podStartSLOduration=3.473894832 podStartE2EDuration="13.127665521s" podCreationTimestamp="2025-12-13 09:41:57 +0000 UTC" firstStartedPulling="2025-12-13 09:41:58.712382866 +0000 UTC m=+10375.316792314" lastFinishedPulling="2025-12-13 09:42:08.366153565 +0000 UTC m=+10384.970563003" observedRunningTime="2025-12-13 09:42:10.106929685 +0000 UTC m=+10386.711339133" watchObservedRunningTime="2025-12-13 09:42:10.127665521 +0000 UTC m=+10386.732074969" Dec 13 09:42:10 crc kubenswrapper[4971]: I1213 09:42:10.148275 4971 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-2lknx"] Dec 13 09:42:10 crc kubenswrapper[4971]: I1213 09:42:10.154877 4971 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-operators-2lknx"] Dec 13 09:42:10 crc kubenswrapper[4971]: I1213 09:42:10.155108 4971 scope.go:117] "RemoveContainer" containerID="4d12722693ddc54d849660b8a61b16ec8562bb76ece3de5a956b12cec797b9fa" Dec 13 09:42:10 crc kubenswrapper[4971]: I1213 09:42:10.204603 4971 scope.go:117] "RemoveContainer" containerID="742f85ce02e7ab3aa88c454ceead458f2543e0029207b61d5fdf5ac7cc7383e1" Dec 13 09:42:10 crc kubenswrapper[4971]: E1213 09:42:10.205231 4971 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"742f85ce02e7ab3aa88c454ceead458f2543e0029207b61d5fdf5ac7cc7383e1\": container with ID starting with 742f85ce02e7ab3aa88c454ceead458f2543e0029207b61d5fdf5ac7cc7383e1 not found: ID does not exist" containerID="742f85ce02e7ab3aa88c454ceead458f2543e0029207b61d5fdf5ac7cc7383e1" Dec 13 09:42:10 crc kubenswrapper[4971]: I1213 09:42:10.205274 4971 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"742f85ce02e7ab3aa88c454ceead458f2543e0029207b61d5fdf5ac7cc7383e1"} err="failed to get container status \"742f85ce02e7ab3aa88c454ceead458f2543e0029207b61d5fdf5ac7cc7383e1\": rpc error: code = NotFound desc = could not find container \"742f85ce02e7ab3aa88c454ceead458f2543e0029207b61d5fdf5ac7cc7383e1\": container with ID starting with 742f85ce02e7ab3aa88c454ceead458f2543e0029207b61d5fdf5ac7cc7383e1 not found: ID does not exist" Dec 13 09:42:10 crc kubenswrapper[4971]: I1213 09:42:10.205299 4971 scope.go:117] "RemoveContainer" containerID="d71d0a29980ea09d5878223236e99a6f18337bb728627214599443921da810ed" Dec 13 09:42:10 crc kubenswrapper[4971]: E1213 09:42:10.205718 4971 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"d71d0a29980ea09d5878223236e99a6f18337bb728627214599443921da810ed\": container with ID starting with d71d0a29980ea09d5878223236e99a6f18337bb728627214599443921da810ed not found: ID does not exist" containerID="d71d0a29980ea09d5878223236e99a6f18337bb728627214599443921da810ed" Dec 13 09:42:10 crc kubenswrapper[4971]: I1213 09:42:10.205780 4971 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"d71d0a29980ea09d5878223236e99a6f18337bb728627214599443921da810ed"} err="failed to get container status \"d71d0a29980ea09d5878223236e99a6f18337bb728627214599443921da810ed\": rpc error: code = NotFound desc = could not find container \"d71d0a29980ea09d5878223236e99a6f18337bb728627214599443921da810ed\": container with ID starting with d71d0a29980ea09d5878223236e99a6f18337bb728627214599443921da810ed not found: ID does not exist" Dec 13 09:42:10 crc kubenswrapper[4971]: I1213 09:42:10.205827 4971 scope.go:117] "RemoveContainer" containerID="4d12722693ddc54d849660b8a61b16ec8562bb76ece3de5a956b12cec797b9fa" Dec 13 09:42:10 crc kubenswrapper[4971]: E1213 09:42:10.206242 4971 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"4d12722693ddc54d849660b8a61b16ec8562bb76ece3de5a956b12cec797b9fa\": container with ID starting with 4d12722693ddc54d849660b8a61b16ec8562bb76ece3de5a956b12cec797b9fa not found: ID does not exist" containerID="4d12722693ddc54d849660b8a61b16ec8562bb76ece3de5a956b12cec797b9fa" Dec 13 09:42:10 crc kubenswrapper[4971]: I1213 09:42:10.206268 4971 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"4d12722693ddc54d849660b8a61b16ec8562bb76ece3de5a956b12cec797b9fa"} err="failed to get container status \"4d12722693ddc54d849660b8a61b16ec8562bb76ece3de5a956b12cec797b9fa\": rpc error: code = NotFound desc = could not find container \"4d12722693ddc54d849660b8a61b16ec8562bb76ece3de5a956b12cec797b9fa\": container with ID starting with 4d12722693ddc54d849660b8a61b16ec8562bb76ece3de5a956b12cec797b9fa not found: ID does not exist" Dec 13 09:42:11 crc kubenswrapper[4971]: I1213 09:42:11.783598 4971 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="47c666f1-9904-43f0-8aa8-536cce82308a" path="/var/lib/kubelet/pods/47c666f1-9904-43f0-8aa8-536cce82308a/volumes" Dec 13 09:42:15 crc kubenswrapper[4971]: I1213 09:42:15.627736 4971 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-must-gather-69gc8/crc-debug-5t4zw"] Dec 13 09:42:15 crc kubenswrapper[4971]: E1213 09:42:15.633707 4971 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="47c666f1-9904-43f0-8aa8-536cce82308a" containerName="extract-utilities" Dec 13 09:42:15 crc kubenswrapper[4971]: I1213 09:42:15.633831 4971 state_mem.go:107] "Deleted CPUSet assignment" podUID="47c666f1-9904-43f0-8aa8-536cce82308a" containerName="extract-utilities" Dec 13 09:42:15 crc kubenswrapper[4971]: E1213 09:42:15.633991 4971 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="47c666f1-9904-43f0-8aa8-536cce82308a" containerName="extract-content" Dec 13 09:42:15 crc kubenswrapper[4971]: I1213 09:42:15.634129 4971 state_mem.go:107] "Deleted CPUSet assignment" podUID="47c666f1-9904-43f0-8aa8-536cce82308a" containerName="extract-content" Dec 13 09:42:15 crc kubenswrapper[4971]: E1213 09:42:15.634259 4971 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="47c666f1-9904-43f0-8aa8-536cce82308a" containerName="registry-server" Dec 13 09:42:15 crc kubenswrapper[4971]: I1213 09:42:15.634339 4971 state_mem.go:107] "Deleted CPUSet assignment" podUID="47c666f1-9904-43f0-8aa8-536cce82308a" containerName="registry-server" Dec 13 09:42:15 crc kubenswrapper[4971]: I1213 09:42:15.634771 4971 memory_manager.go:354] "RemoveStaleState removing state" podUID="47c666f1-9904-43f0-8aa8-536cce82308a" containerName="registry-server" Dec 13 09:42:15 crc kubenswrapper[4971]: I1213 09:42:15.635649 4971 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-69gc8/crc-debug-5t4zw" Dec 13 09:42:15 crc kubenswrapper[4971]: I1213 09:42:15.693390 4971 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host\" (UniqueName: \"kubernetes.io/host-path/eff630d7-c0fd-4cc7-8815-012305a6b3d8-host\") pod \"crc-debug-5t4zw\" (UID: \"eff630d7-c0fd-4cc7-8815-012305a6b3d8\") " pod="openshift-must-gather-69gc8/crc-debug-5t4zw" Dec 13 09:42:15 crc kubenswrapper[4971]: I1213 09:42:15.693749 4971 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-szq8q\" (UniqueName: \"kubernetes.io/projected/eff630d7-c0fd-4cc7-8815-012305a6b3d8-kube-api-access-szq8q\") pod \"crc-debug-5t4zw\" (UID: \"eff630d7-c0fd-4cc7-8815-012305a6b3d8\") " pod="openshift-must-gather-69gc8/crc-debug-5t4zw" Dec 13 09:42:15 crc kubenswrapper[4971]: I1213 09:42:15.795438 4971 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-szq8q\" (UniqueName: \"kubernetes.io/projected/eff630d7-c0fd-4cc7-8815-012305a6b3d8-kube-api-access-szq8q\") pod \"crc-debug-5t4zw\" (UID: \"eff630d7-c0fd-4cc7-8815-012305a6b3d8\") " pod="openshift-must-gather-69gc8/crc-debug-5t4zw" Dec 13 09:42:15 crc kubenswrapper[4971]: I1213 09:42:15.795535 4971 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host\" (UniqueName: \"kubernetes.io/host-path/eff630d7-c0fd-4cc7-8815-012305a6b3d8-host\") pod \"crc-debug-5t4zw\" (UID: \"eff630d7-c0fd-4cc7-8815-012305a6b3d8\") " pod="openshift-must-gather-69gc8/crc-debug-5t4zw" Dec 13 09:42:15 crc kubenswrapper[4971]: I1213 09:42:15.795684 4971 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host\" (UniqueName: \"kubernetes.io/host-path/eff630d7-c0fd-4cc7-8815-012305a6b3d8-host\") pod \"crc-debug-5t4zw\" (UID: \"eff630d7-c0fd-4cc7-8815-012305a6b3d8\") " pod="openshift-must-gather-69gc8/crc-debug-5t4zw" Dec 13 09:42:15 crc kubenswrapper[4971]: I1213 09:42:15.823489 4971 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-szq8q\" (UniqueName: \"kubernetes.io/projected/eff630d7-c0fd-4cc7-8815-012305a6b3d8-kube-api-access-szq8q\") pod \"crc-debug-5t4zw\" (UID: \"eff630d7-c0fd-4cc7-8815-012305a6b3d8\") " pod="openshift-must-gather-69gc8/crc-debug-5t4zw" Dec 13 09:42:15 crc kubenswrapper[4971]: I1213 09:42:15.961622 4971 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-69gc8/crc-debug-5t4zw" Dec 13 09:42:16 crc kubenswrapper[4971]: I1213 09:42:16.154023 4971 patch_prober.go:28] interesting pod/machine-config-daemon-82xjz container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 13 09:42:16 crc kubenswrapper[4971]: I1213 09:42:16.154090 4971 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-82xjz" podUID="e0c4af12-edda-41d7-9b44-f87396174ef2" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 13 09:42:16 crc kubenswrapper[4971]: I1213 09:42:16.163700 4971 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-69gc8/crc-debug-5t4zw" event={"ID":"eff630d7-c0fd-4cc7-8815-012305a6b3d8","Type":"ContainerStarted","Data":"9eddf556651d6c4d4f8c569e65895c5948161f7a41744cc4765fff74e23a2e20"} Dec 13 09:42:29 crc kubenswrapper[4971]: I1213 09:42:29.312688 4971 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-69gc8/crc-debug-5t4zw" event={"ID":"eff630d7-c0fd-4cc7-8815-012305a6b3d8","Type":"ContainerStarted","Data":"b24457bee748c4942f2344b7bcededdf6e9a8e90748496caea2610a8e50fbd39"} Dec 13 09:42:29 crc kubenswrapper[4971]: I1213 09:42:29.338609 4971 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-must-gather-69gc8/crc-debug-5t4zw" podStartSLOduration=1.7095556589999998 podStartE2EDuration="14.33858019s" podCreationTimestamp="2025-12-13 09:42:15 +0000 UTC" firstStartedPulling="2025-12-13 09:42:16.009249585 +0000 UTC m=+10392.613659043" lastFinishedPulling="2025-12-13 09:42:28.638274136 +0000 UTC m=+10405.242683574" observedRunningTime="2025-12-13 09:42:29.330062982 +0000 UTC m=+10405.934472430" watchObservedRunningTime="2025-12-13 09:42:29.33858019 +0000 UTC m=+10405.942989648" Dec 13 09:42:42 crc kubenswrapper[4971]: E1213 09:42:42.770119 4971 kubelet_pods.go:538] "Hostname for pod was too long, truncated it" podName="test-operator-logs-pod-horizontest-horizontest-tests-horizontest" hostnameMaxLen=63 truncatedHostname="test-operator-logs-pod-horizontest-horizontest-tests-horizontes" Dec 13 09:42:46 crc kubenswrapper[4971]: I1213 09:42:46.153424 4971 patch_prober.go:28] interesting pod/machine-config-daemon-82xjz container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 13 09:42:46 crc kubenswrapper[4971]: I1213 09:42:46.154310 4971 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-82xjz" podUID="e0c4af12-edda-41d7-9b44-f87396174ef2" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 13 09:43:16 crc kubenswrapper[4971]: I1213 09:43:16.153545 4971 patch_prober.go:28] interesting pod/machine-config-daemon-82xjz container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 13 09:43:16 crc kubenswrapper[4971]: I1213 09:43:16.154160 4971 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-82xjz" podUID="e0c4af12-edda-41d7-9b44-f87396174ef2" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 13 09:43:16 crc kubenswrapper[4971]: I1213 09:43:16.154228 4971 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-82xjz" Dec 13 09:43:16 crc kubenswrapper[4971]: I1213 09:43:16.155568 4971 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"b2e549456db49c5609edf5288a8e1e209f21e8f767afaa3d5510b083ba33ae70"} pod="openshift-machine-config-operator/machine-config-daemon-82xjz" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Dec 13 09:43:16 crc kubenswrapper[4971]: I1213 09:43:16.155657 4971 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-82xjz" podUID="e0c4af12-edda-41d7-9b44-f87396174ef2" containerName="machine-config-daemon" containerID="cri-o://b2e549456db49c5609edf5288a8e1e209f21e8f767afaa3d5510b083ba33ae70" gracePeriod=600 Dec 13 09:43:16 crc kubenswrapper[4971]: I1213 09:43:16.791702 4971 generic.go:334] "Generic (PLEG): container finished" podID="e0c4af12-edda-41d7-9b44-f87396174ef2" containerID="b2e549456db49c5609edf5288a8e1e209f21e8f767afaa3d5510b083ba33ae70" exitCode=0 Dec 13 09:43:16 crc kubenswrapper[4971]: I1213 09:43:16.792322 4971 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-82xjz" event={"ID":"e0c4af12-edda-41d7-9b44-f87396174ef2","Type":"ContainerDied","Data":"b2e549456db49c5609edf5288a8e1e209f21e8f767afaa3d5510b083ba33ae70"} Dec 13 09:43:16 crc kubenswrapper[4971]: I1213 09:43:16.792444 4971 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-82xjz" event={"ID":"e0c4af12-edda-41d7-9b44-f87396174ef2","Type":"ContainerStarted","Data":"80deb18ddf85834e136ab62ea93a1fa75315016d9f220eb217720a5e562f5e54"} Dec 13 09:43:16 crc kubenswrapper[4971]: I1213 09:43:16.792564 4971 scope.go:117] "RemoveContainer" containerID="7c80479d6aedc6e1a5cced1df9e55738d4a726b6ddd5adfe17363d9c5d097d6a" Dec 13 09:43:26 crc kubenswrapper[4971]: I1213 09:43:26.902027 4971 generic.go:334] "Generic (PLEG): container finished" podID="eff630d7-c0fd-4cc7-8815-012305a6b3d8" containerID="b24457bee748c4942f2344b7bcededdf6e9a8e90748496caea2610a8e50fbd39" exitCode=0 Dec 13 09:43:26 crc kubenswrapper[4971]: I1213 09:43:26.902130 4971 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-69gc8/crc-debug-5t4zw" event={"ID":"eff630d7-c0fd-4cc7-8815-012305a6b3d8","Type":"ContainerDied","Data":"b24457bee748c4942f2344b7bcededdf6e9a8e90748496caea2610a8e50fbd39"} Dec 13 09:43:28 crc kubenswrapper[4971]: I1213 09:43:28.027310 4971 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-69gc8/crc-debug-5t4zw" Dec 13 09:43:28 crc kubenswrapper[4971]: I1213 09:43:28.073138 4971 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-must-gather-69gc8/crc-debug-5t4zw"] Dec 13 09:43:28 crc kubenswrapper[4971]: I1213 09:43:28.085783 4971 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-must-gather-69gc8/crc-debug-5t4zw"] Dec 13 09:43:28 crc kubenswrapper[4971]: I1213 09:43:28.172209 4971 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-szq8q\" (UniqueName: \"kubernetes.io/projected/eff630d7-c0fd-4cc7-8815-012305a6b3d8-kube-api-access-szq8q\") pod \"eff630d7-c0fd-4cc7-8815-012305a6b3d8\" (UID: \"eff630d7-c0fd-4cc7-8815-012305a6b3d8\") " Dec 13 09:43:28 crc kubenswrapper[4971]: I1213 09:43:28.172404 4971 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"host\" (UniqueName: \"kubernetes.io/host-path/eff630d7-c0fd-4cc7-8815-012305a6b3d8-host\") pod \"eff630d7-c0fd-4cc7-8815-012305a6b3d8\" (UID: \"eff630d7-c0fd-4cc7-8815-012305a6b3d8\") " Dec 13 09:43:28 crc kubenswrapper[4971]: I1213 09:43:28.172911 4971 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/eff630d7-c0fd-4cc7-8815-012305a6b3d8-host" (OuterVolumeSpecName: "host") pod "eff630d7-c0fd-4cc7-8815-012305a6b3d8" (UID: "eff630d7-c0fd-4cc7-8815-012305a6b3d8"). InnerVolumeSpecName "host". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 13 09:43:28 crc kubenswrapper[4971]: I1213 09:43:28.173907 4971 reconciler_common.go:293] "Volume detached for volume \"host\" (UniqueName: \"kubernetes.io/host-path/eff630d7-c0fd-4cc7-8815-012305a6b3d8-host\") on node \"crc\" DevicePath \"\"" Dec 13 09:43:28 crc kubenswrapper[4971]: I1213 09:43:28.835983 4971 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/eff630d7-c0fd-4cc7-8815-012305a6b3d8-kube-api-access-szq8q" (OuterVolumeSpecName: "kube-api-access-szq8q") pod "eff630d7-c0fd-4cc7-8815-012305a6b3d8" (UID: "eff630d7-c0fd-4cc7-8815-012305a6b3d8"). InnerVolumeSpecName "kube-api-access-szq8q". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 13 09:43:28 crc kubenswrapper[4971]: I1213 09:43:28.892871 4971 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-szq8q\" (UniqueName: \"kubernetes.io/projected/eff630d7-c0fd-4cc7-8815-012305a6b3d8-kube-api-access-szq8q\") on node \"crc\" DevicePath \"\"" Dec 13 09:43:28 crc kubenswrapper[4971]: I1213 09:43:28.942473 4971 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="9eddf556651d6c4d4f8c569e65895c5948161f7a41744cc4765fff74e23a2e20" Dec 13 09:43:28 crc kubenswrapper[4971]: I1213 09:43:28.942584 4971 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-69gc8/crc-debug-5t4zw" Dec 13 09:43:29 crc kubenswrapper[4971]: I1213 09:43:29.291565 4971 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-must-gather-69gc8/crc-debug-bvkr7"] Dec 13 09:43:29 crc kubenswrapper[4971]: E1213 09:43:29.292687 4971 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="eff630d7-c0fd-4cc7-8815-012305a6b3d8" containerName="container-00" Dec 13 09:43:29 crc kubenswrapper[4971]: I1213 09:43:29.292708 4971 state_mem.go:107] "Deleted CPUSet assignment" podUID="eff630d7-c0fd-4cc7-8815-012305a6b3d8" containerName="container-00" Dec 13 09:43:29 crc kubenswrapper[4971]: I1213 09:43:29.292911 4971 memory_manager.go:354] "RemoveStaleState removing state" podUID="eff630d7-c0fd-4cc7-8815-012305a6b3d8" containerName="container-00" Dec 13 09:43:29 crc kubenswrapper[4971]: I1213 09:43:29.293844 4971 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-69gc8/crc-debug-bvkr7" Dec 13 09:43:29 crc kubenswrapper[4971]: I1213 09:43:29.404502 4971 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host\" (UniqueName: \"kubernetes.io/host-path/7f2e0cbb-3022-4cd2-987e-a6f5151fe8f9-host\") pod \"crc-debug-bvkr7\" (UID: \"7f2e0cbb-3022-4cd2-987e-a6f5151fe8f9\") " pod="openshift-must-gather-69gc8/crc-debug-bvkr7" Dec 13 09:43:29 crc kubenswrapper[4971]: I1213 09:43:29.404939 4971 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-4xkjz\" (UniqueName: \"kubernetes.io/projected/7f2e0cbb-3022-4cd2-987e-a6f5151fe8f9-kube-api-access-4xkjz\") pod \"crc-debug-bvkr7\" (UID: \"7f2e0cbb-3022-4cd2-987e-a6f5151fe8f9\") " pod="openshift-must-gather-69gc8/crc-debug-bvkr7" Dec 13 09:43:29 crc kubenswrapper[4971]: I1213 09:43:29.507741 4971 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host\" (UniqueName: \"kubernetes.io/host-path/7f2e0cbb-3022-4cd2-987e-a6f5151fe8f9-host\") pod \"crc-debug-bvkr7\" (UID: \"7f2e0cbb-3022-4cd2-987e-a6f5151fe8f9\") " pod="openshift-must-gather-69gc8/crc-debug-bvkr7" Dec 13 09:43:29 crc kubenswrapper[4971]: I1213 09:43:29.507878 4971 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-4xkjz\" (UniqueName: \"kubernetes.io/projected/7f2e0cbb-3022-4cd2-987e-a6f5151fe8f9-kube-api-access-4xkjz\") pod \"crc-debug-bvkr7\" (UID: \"7f2e0cbb-3022-4cd2-987e-a6f5151fe8f9\") " pod="openshift-must-gather-69gc8/crc-debug-bvkr7" Dec 13 09:43:29 crc kubenswrapper[4971]: I1213 09:43:29.507927 4971 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host\" (UniqueName: \"kubernetes.io/host-path/7f2e0cbb-3022-4cd2-987e-a6f5151fe8f9-host\") pod \"crc-debug-bvkr7\" (UID: \"7f2e0cbb-3022-4cd2-987e-a6f5151fe8f9\") " pod="openshift-must-gather-69gc8/crc-debug-bvkr7" Dec 13 09:43:29 crc kubenswrapper[4971]: I1213 09:43:29.528177 4971 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-4xkjz\" (UniqueName: \"kubernetes.io/projected/7f2e0cbb-3022-4cd2-987e-a6f5151fe8f9-kube-api-access-4xkjz\") pod \"crc-debug-bvkr7\" (UID: \"7f2e0cbb-3022-4cd2-987e-a6f5151fe8f9\") " pod="openshift-must-gather-69gc8/crc-debug-bvkr7" Dec 13 09:43:29 crc kubenswrapper[4971]: I1213 09:43:29.611947 4971 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-69gc8/crc-debug-bvkr7" Dec 13 09:43:29 crc kubenswrapper[4971]: W1213 09:43:29.642549 4971 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod7f2e0cbb_3022_4cd2_987e_a6f5151fe8f9.slice/crio-a078ced17d7f60f6a38cf299306753bfa9e8293ebc5dac358276abeb20d386df WatchSource:0}: Error finding container a078ced17d7f60f6a38cf299306753bfa9e8293ebc5dac358276abeb20d386df: Status 404 returned error can't find the container with id a078ced17d7f60f6a38cf299306753bfa9e8293ebc5dac358276abeb20d386df Dec 13 09:43:29 crc kubenswrapper[4971]: I1213 09:43:29.783294 4971 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="eff630d7-c0fd-4cc7-8815-012305a6b3d8" path="/var/lib/kubelet/pods/eff630d7-c0fd-4cc7-8815-012305a6b3d8/volumes" Dec 13 09:43:29 crc kubenswrapper[4971]: I1213 09:43:29.954109 4971 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-69gc8/crc-debug-bvkr7" event={"ID":"7f2e0cbb-3022-4cd2-987e-a6f5151fe8f9","Type":"ContainerStarted","Data":"2a73ca6e689a48e12b02f5a59bd2ef8c20b256cad668fe51b8e585606302a17c"} Dec 13 09:43:29 crc kubenswrapper[4971]: I1213 09:43:29.954461 4971 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-69gc8/crc-debug-bvkr7" event={"ID":"7f2e0cbb-3022-4cd2-987e-a6f5151fe8f9","Type":"ContainerStarted","Data":"a078ced17d7f60f6a38cf299306753bfa9e8293ebc5dac358276abeb20d386df"} Dec 13 09:43:31 crc kubenswrapper[4971]: I1213 09:43:31.005374 4971 generic.go:334] "Generic (PLEG): container finished" podID="7f2e0cbb-3022-4cd2-987e-a6f5151fe8f9" containerID="2a73ca6e689a48e12b02f5a59bd2ef8c20b256cad668fe51b8e585606302a17c" exitCode=0 Dec 13 09:43:31 crc kubenswrapper[4971]: I1213 09:43:31.005441 4971 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-69gc8/crc-debug-bvkr7" event={"ID":"7f2e0cbb-3022-4cd2-987e-a6f5151fe8f9","Type":"ContainerDied","Data":"2a73ca6e689a48e12b02f5a59bd2ef8c20b256cad668fe51b8e585606302a17c"} Dec 13 09:43:32 crc kubenswrapper[4971]: I1213 09:43:32.129804 4971 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-69gc8/crc-debug-bvkr7" Dec 13 09:43:32 crc kubenswrapper[4971]: I1213 09:43:32.175213 4971 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"host\" (UniqueName: \"kubernetes.io/host-path/7f2e0cbb-3022-4cd2-987e-a6f5151fe8f9-host\") pod \"7f2e0cbb-3022-4cd2-987e-a6f5151fe8f9\" (UID: \"7f2e0cbb-3022-4cd2-987e-a6f5151fe8f9\") " Dec 13 09:43:32 crc kubenswrapper[4971]: I1213 09:43:32.175294 4971 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/7f2e0cbb-3022-4cd2-987e-a6f5151fe8f9-host" (OuterVolumeSpecName: "host") pod "7f2e0cbb-3022-4cd2-987e-a6f5151fe8f9" (UID: "7f2e0cbb-3022-4cd2-987e-a6f5151fe8f9"). InnerVolumeSpecName "host". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 13 09:43:32 crc kubenswrapper[4971]: I1213 09:43:32.175478 4971 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-4xkjz\" (UniqueName: \"kubernetes.io/projected/7f2e0cbb-3022-4cd2-987e-a6f5151fe8f9-kube-api-access-4xkjz\") pod \"7f2e0cbb-3022-4cd2-987e-a6f5151fe8f9\" (UID: \"7f2e0cbb-3022-4cd2-987e-a6f5151fe8f9\") " Dec 13 09:43:32 crc kubenswrapper[4971]: I1213 09:43:32.176065 4971 reconciler_common.go:293] "Volume detached for volume \"host\" (UniqueName: \"kubernetes.io/host-path/7f2e0cbb-3022-4cd2-987e-a6f5151fe8f9-host\") on node \"crc\" DevicePath \"\"" Dec 13 09:43:32 crc kubenswrapper[4971]: I1213 09:43:32.182808 4971 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/7f2e0cbb-3022-4cd2-987e-a6f5151fe8f9-kube-api-access-4xkjz" (OuterVolumeSpecName: "kube-api-access-4xkjz") pod "7f2e0cbb-3022-4cd2-987e-a6f5151fe8f9" (UID: "7f2e0cbb-3022-4cd2-987e-a6f5151fe8f9"). InnerVolumeSpecName "kube-api-access-4xkjz". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 13 09:43:32 crc kubenswrapper[4971]: I1213 09:43:32.279275 4971 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-4xkjz\" (UniqueName: \"kubernetes.io/projected/7f2e0cbb-3022-4cd2-987e-a6f5151fe8f9-kube-api-access-4xkjz\") on node \"crc\" DevicePath \"\"" Dec 13 09:43:32 crc kubenswrapper[4971]: I1213 09:43:32.675213 4971 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-must-gather-69gc8/crc-debug-bvkr7"] Dec 13 09:43:32 crc kubenswrapper[4971]: I1213 09:43:32.685921 4971 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-must-gather-69gc8/crc-debug-bvkr7"] Dec 13 09:43:33 crc kubenswrapper[4971]: I1213 09:43:33.028986 4971 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="a078ced17d7f60f6a38cf299306753bfa9e8293ebc5dac358276abeb20d386df" Dec 13 09:43:33 crc kubenswrapper[4971]: I1213 09:43:33.029070 4971 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-69gc8/crc-debug-bvkr7" Dec 13 09:43:33 crc kubenswrapper[4971]: I1213 09:43:33.802408 4971 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="7f2e0cbb-3022-4cd2-987e-a6f5151fe8f9" path="/var/lib/kubelet/pods/7f2e0cbb-3022-4cd2-987e-a6f5151fe8f9/volumes" Dec 13 09:43:33 crc kubenswrapper[4971]: I1213 09:43:33.906179 4971 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-must-gather-69gc8/crc-debug-4v4lh"] Dec 13 09:43:33 crc kubenswrapper[4971]: E1213 09:43:33.906914 4971 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="7f2e0cbb-3022-4cd2-987e-a6f5151fe8f9" containerName="container-00" Dec 13 09:43:33 crc kubenswrapper[4971]: I1213 09:43:33.906942 4971 state_mem.go:107] "Deleted CPUSet assignment" podUID="7f2e0cbb-3022-4cd2-987e-a6f5151fe8f9" containerName="container-00" Dec 13 09:43:33 crc kubenswrapper[4971]: I1213 09:43:33.907201 4971 memory_manager.go:354] "RemoveStaleState removing state" podUID="7f2e0cbb-3022-4cd2-987e-a6f5151fe8f9" containerName="container-00" Dec 13 09:43:33 crc kubenswrapper[4971]: I1213 09:43:33.908168 4971 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-69gc8/crc-debug-4v4lh" Dec 13 09:43:34 crc kubenswrapper[4971]: I1213 09:43:34.024333 4971 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host\" (UniqueName: \"kubernetes.io/host-path/83363a89-be3c-45e5-b8af-816d5b3dd17c-host\") pod \"crc-debug-4v4lh\" (UID: \"83363a89-be3c-45e5-b8af-816d5b3dd17c\") " pod="openshift-must-gather-69gc8/crc-debug-4v4lh" Dec 13 09:43:34 crc kubenswrapper[4971]: I1213 09:43:34.024544 4971 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-wg8qn\" (UniqueName: \"kubernetes.io/projected/83363a89-be3c-45e5-b8af-816d5b3dd17c-kube-api-access-wg8qn\") pod \"crc-debug-4v4lh\" (UID: \"83363a89-be3c-45e5-b8af-816d5b3dd17c\") " pod="openshift-must-gather-69gc8/crc-debug-4v4lh" Dec 13 09:43:34 crc kubenswrapper[4971]: I1213 09:43:34.126787 4971 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-wg8qn\" (UniqueName: \"kubernetes.io/projected/83363a89-be3c-45e5-b8af-816d5b3dd17c-kube-api-access-wg8qn\") pod \"crc-debug-4v4lh\" (UID: \"83363a89-be3c-45e5-b8af-816d5b3dd17c\") " pod="openshift-must-gather-69gc8/crc-debug-4v4lh" Dec 13 09:43:34 crc kubenswrapper[4971]: I1213 09:43:34.127006 4971 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host\" (UniqueName: \"kubernetes.io/host-path/83363a89-be3c-45e5-b8af-816d5b3dd17c-host\") pod \"crc-debug-4v4lh\" (UID: \"83363a89-be3c-45e5-b8af-816d5b3dd17c\") " pod="openshift-must-gather-69gc8/crc-debug-4v4lh" Dec 13 09:43:34 crc kubenswrapper[4971]: I1213 09:43:34.127254 4971 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host\" (UniqueName: \"kubernetes.io/host-path/83363a89-be3c-45e5-b8af-816d5b3dd17c-host\") pod \"crc-debug-4v4lh\" (UID: \"83363a89-be3c-45e5-b8af-816d5b3dd17c\") " pod="openshift-must-gather-69gc8/crc-debug-4v4lh" Dec 13 09:43:34 crc kubenswrapper[4971]: I1213 09:43:34.147592 4971 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-wg8qn\" (UniqueName: \"kubernetes.io/projected/83363a89-be3c-45e5-b8af-816d5b3dd17c-kube-api-access-wg8qn\") pod \"crc-debug-4v4lh\" (UID: \"83363a89-be3c-45e5-b8af-816d5b3dd17c\") " pod="openshift-must-gather-69gc8/crc-debug-4v4lh" Dec 13 09:43:34 crc kubenswrapper[4971]: I1213 09:43:34.235276 4971 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-69gc8/crc-debug-4v4lh" Dec 13 09:43:34 crc kubenswrapper[4971]: W1213 09:43:34.263837 4971 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod83363a89_be3c_45e5_b8af_816d5b3dd17c.slice/crio-0d0f883827360c2ca983f9c1fc0300938c9e3144c8a876453ec705a4fb920f88 WatchSource:0}: Error finding container 0d0f883827360c2ca983f9c1fc0300938c9e3144c8a876453ec705a4fb920f88: Status 404 returned error can't find the container with id 0d0f883827360c2ca983f9c1fc0300938c9e3144c8a876453ec705a4fb920f88 Dec 13 09:43:35 crc kubenswrapper[4971]: I1213 09:43:35.050396 4971 generic.go:334] "Generic (PLEG): container finished" podID="83363a89-be3c-45e5-b8af-816d5b3dd17c" containerID="9e6eefc27e895abb38f58df4634cf66db588cd75322abd73c47f2b98bb041d64" exitCode=0 Dec 13 09:43:35 crc kubenswrapper[4971]: I1213 09:43:35.050670 4971 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-69gc8/crc-debug-4v4lh" event={"ID":"83363a89-be3c-45e5-b8af-816d5b3dd17c","Type":"ContainerDied","Data":"9e6eefc27e895abb38f58df4634cf66db588cd75322abd73c47f2b98bb041d64"} Dec 13 09:43:35 crc kubenswrapper[4971]: I1213 09:43:35.050705 4971 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-69gc8/crc-debug-4v4lh" event={"ID":"83363a89-be3c-45e5-b8af-816d5b3dd17c","Type":"ContainerStarted","Data":"0d0f883827360c2ca983f9c1fc0300938c9e3144c8a876453ec705a4fb920f88"} Dec 13 09:43:35 crc kubenswrapper[4971]: I1213 09:43:35.106178 4971 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-must-gather-69gc8/crc-debug-4v4lh"] Dec 13 09:43:35 crc kubenswrapper[4971]: I1213 09:43:35.120599 4971 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-must-gather-69gc8/crc-debug-4v4lh"] Dec 13 09:43:36 crc kubenswrapper[4971]: I1213 09:43:36.181940 4971 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-69gc8/crc-debug-4v4lh" Dec 13 09:43:36 crc kubenswrapper[4971]: I1213 09:43:36.276374 4971 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"host\" (UniqueName: \"kubernetes.io/host-path/83363a89-be3c-45e5-b8af-816d5b3dd17c-host\") pod \"83363a89-be3c-45e5-b8af-816d5b3dd17c\" (UID: \"83363a89-be3c-45e5-b8af-816d5b3dd17c\") " Dec 13 09:43:36 crc kubenswrapper[4971]: I1213 09:43:36.276582 4971 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/83363a89-be3c-45e5-b8af-816d5b3dd17c-host" (OuterVolumeSpecName: "host") pod "83363a89-be3c-45e5-b8af-816d5b3dd17c" (UID: "83363a89-be3c-45e5-b8af-816d5b3dd17c"). InnerVolumeSpecName "host". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 13 09:43:36 crc kubenswrapper[4971]: I1213 09:43:36.276740 4971 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-wg8qn\" (UniqueName: \"kubernetes.io/projected/83363a89-be3c-45e5-b8af-816d5b3dd17c-kube-api-access-wg8qn\") pod \"83363a89-be3c-45e5-b8af-816d5b3dd17c\" (UID: \"83363a89-be3c-45e5-b8af-816d5b3dd17c\") " Dec 13 09:43:36 crc kubenswrapper[4971]: I1213 09:43:36.277621 4971 reconciler_common.go:293] "Volume detached for volume \"host\" (UniqueName: \"kubernetes.io/host-path/83363a89-be3c-45e5-b8af-816d5b3dd17c-host\") on node \"crc\" DevicePath \"\"" Dec 13 09:43:36 crc kubenswrapper[4971]: I1213 09:43:36.282970 4971 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/83363a89-be3c-45e5-b8af-816d5b3dd17c-kube-api-access-wg8qn" (OuterVolumeSpecName: "kube-api-access-wg8qn") pod "83363a89-be3c-45e5-b8af-816d5b3dd17c" (UID: "83363a89-be3c-45e5-b8af-816d5b3dd17c"). InnerVolumeSpecName "kube-api-access-wg8qn". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 13 09:43:36 crc kubenswrapper[4971]: I1213 09:43:36.379961 4971 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-wg8qn\" (UniqueName: \"kubernetes.io/projected/83363a89-be3c-45e5-b8af-816d5b3dd17c-kube-api-access-wg8qn\") on node \"crc\" DevicePath \"\"" Dec 13 09:43:37 crc kubenswrapper[4971]: I1213 09:43:37.075366 4971 scope.go:117] "RemoveContainer" containerID="9e6eefc27e895abb38f58df4634cf66db588cd75322abd73c47f2b98bb041d64" Dec 13 09:43:37 crc kubenswrapper[4971]: I1213 09:43:37.075405 4971 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-69gc8/crc-debug-4v4lh" Dec 13 09:43:37 crc kubenswrapper[4971]: I1213 09:43:37.786564 4971 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="83363a89-be3c-45e5-b8af-816d5b3dd17c" path="/var/lib/kubelet/pods/83363a89-be3c-45e5-b8af-816d5b3dd17c/volumes" Dec 13 09:43:46 crc kubenswrapper[4971]: E1213 09:43:46.769210 4971 kubelet_pods.go:538] "Hostname for pod was too long, truncated it" podName="test-operator-logs-pod-horizontest-horizontest-tests-horizontest" hostnameMaxLen=63 truncatedHostname="test-operator-logs-pod-horizontest-horizontest-tests-horizontes" Dec 13 09:44:25 crc kubenswrapper[4971]: I1213 09:44:25.975724 4971 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ansibletest-ansibletest_a020591e-029f-464f-8d6c-947acba2bf13/ansibletest-ansibletest/0.log" Dec 13 09:44:26 crc kubenswrapper[4971]: I1213 09:44:26.128116 4971 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_barbican-api-55549fd796-4tztx_d553c6ea-f592-4890-bccd-f970616d23fe/barbican-api/0.log" Dec 13 09:44:26 crc kubenswrapper[4971]: I1213 09:44:26.231189 4971 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_barbican-api-55549fd796-4tztx_d553c6ea-f592-4890-bccd-f970616d23fe/barbican-api-log/0.log" Dec 13 09:44:26 crc kubenswrapper[4971]: I1213 09:44:26.358917 4971 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_barbican-keystone-listener-6959d6f94b-tq6hh_12a4a604-178e-4ad8-a0f7-e2e3d47590bd/barbican-keystone-listener/0.log" Dec 13 09:44:26 crc kubenswrapper[4971]: I1213 09:44:26.950603 4971 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_barbican-keystone-listener-6959d6f94b-tq6hh_12a4a604-178e-4ad8-a0f7-e2e3d47590bd/barbican-keystone-listener-log/0.log" Dec 13 09:44:26 crc kubenswrapper[4971]: I1213 09:44:26.959959 4971 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_barbican-worker-c7c6b44fc-8hd9n_fb068297-28dd-4a70-acbf-650b9edf380e/barbican-worker/0.log" Dec 13 09:44:27 crc kubenswrapper[4971]: I1213 09:44:27.049691 4971 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_barbican-worker-c7c6b44fc-8hd9n_fb068297-28dd-4a70-acbf-650b9edf380e/barbican-worker-log/0.log" Dec 13 09:44:27 crc kubenswrapper[4971]: I1213 09:44:27.275067 4971 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_bootstrap-edpm-deployment-openstack-edpm-ipam-jdcq5_27dafb6f-1f04-448c-97d2-0287c277c26a/bootstrap-edpm-deployment-openstack-edpm-ipam/0.log" Dec 13 09:44:27 crc kubenswrapper[4971]: I1213 09:44:27.408233 4971 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ceilometer-0_7ba95a4c-be84-460a-9dfc-b6e1474e1b81/ceilometer-central-agent/0.log" Dec 13 09:44:27 crc kubenswrapper[4971]: I1213 09:44:27.597620 4971 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ceilometer-0_7ba95a4c-be84-460a-9dfc-b6e1474e1b81/sg-core/0.log" Dec 13 09:44:27 crc kubenswrapper[4971]: I1213 09:44:27.603100 4971 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ceilometer-0_7ba95a4c-be84-460a-9dfc-b6e1474e1b81/proxy-httpd/0.log" Dec 13 09:44:27 crc kubenswrapper[4971]: I1213 09:44:27.613991 4971 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ceilometer-0_7ba95a4c-be84-460a-9dfc-b6e1474e1b81/ceilometer-notification-agent/0.log" Dec 13 09:44:27 crc kubenswrapper[4971]: I1213 09:44:27.892535 4971 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ceph-client-edpm-deployment-openstack-edpm-ipam-7grkn_42cff6d1-138e-4a16-a570-cdcbcbe6ab49/ceph-client-edpm-deployment-openstack-edpm-ipam/0.log" Dec 13 09:44:28 crc kubenswrapper[4971]: I1213 09:44:28.046105 4971 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ceph-hci-pre-edpm-deployment-openstack-edpm-ipam-4fh5s_bbbc7fc6-2323-4765-9160-a1c085de09d8/ceph-hci-pre-edpm-deployment-openstack-edpm-ipam/0.log" Dec 13 09:44:28 crc kubenswrapper[4971]: I1213 09:44:28.340027 4971 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_cinder-api-0_7cbba320-ebcc-40d7-88ea-ba668d549245/cinder-api-log/0.log" Dec 13 09:44:28 crc kubenswrapper[4971]: I1213 09:44:28.427472 4971 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_cinder-api-0_7cbba320-ebcc-40d7-88ea-ba668d549245/cinder-api/0.log" Dec 13 09:44:28 crc kubenswrapper[4971]: I1213 09:44:28.812663 4971 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_cinder-scheduler-0_4c666e54-0053-4ecc-b8bd-6a2771ebbad1/cinder-scheduler/0.log" Dec 13 09:44:28 crc kubenswrapper[4971]: I1213 09:44:28.832436 4971 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_cinder-backup-0_be3cc24a-8820-4884-8f26-e2885d933527/probe/0.log" Dec 13 09:44:28 crc kubenswrapper[4971]: I1213 09:44:28.840083 4971 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_cinder-backup-0_be3cc24a-8820-4884-8f26-e2885d933527/cinder-backup/0.log" Dec 13 09:44:29 crc kubenswrapper[4971]: I1213 09:44:29.119307 4971 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_cinder-volume-volume1-0_7bbf960d-93e9-49d6-8545-28c34e36229e/cinder-volume/0.log" Dec 13 09:44:29 crc kubenswrapper[4971]: I1213 09:44:29.218570 4971 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_cinder-scheduler-0_4c666e54-0053-4ecc-b8bd-6a2771ebbad1/probe/0.log" Dec 13 09:44:29 crc kubenswrapper[4971]: I1213 09:44:29.255621 4971 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_cinder-volume-volume1-0_7bbf960d-93e9-49d6-8545-28c34e36229e/probe/0.log" Dec 13 09:44:29 crc kubenswrapper[4971]: I1213 09:44:29.396748 4971 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_configure-network-edpm-deployment-openstack-edpm-ipam-95c87_ef08e99c-0bd8-4f27-ae28-2cfdeb4ec045/configure-network-edpm-deployment-openstack-edpm-ipam/0.log" Dec 13 09:44:29 crc kubenswrapper[4971]: I1213 09:44:29.531592 4971 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_configure-os-edpm-deployment-openstack-edpm-ipam-7qg86_b03f6448-21d3-417f-a1a9-6e7c733550ec/configure-os-edpm-deployment-openstack-edpm-ipam/0.log" Dec 13 09:44:29 crc kubenswrapper[4971]: I1213 09:44:29.698713 4971 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_dnsmasq-dns-ccc856b4f-zh8np_409ac970-9fc4-4b77-9548-1714f54c8d51/init/0.log" Dec 13 09:44:29 crc kubenswrapper[4971]: I1213 09:44:29.939843 4971 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_dnsmasq-dns-ccc856b4f-zh8np_409ac970-9fc4-4b77-9548-1714f54c8d51/init/0.log" Dec 13 09:44:29 crc kubenswrapper[4971]: I1213 09:44:29.997639 4971 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_glance-default-external-api-0_9fbf6500-bf43-4a4c-820c-14a60a27afb9/glance-log/0.log" Dec 13 09:44:30 crc kubenswrapper[4971]: I1213 09:44:30.150741 4971 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_glance-default-external-api-0_9fbf6500-bf43-4a4c-820c-14a60a27afb9/glance-httpd/0.log" Dec 13 09:44:30 crc kubenswrapper[4971]: I1213 09:44:30.253427 4971 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_dnsmasq-dns-ccc856b4f-zh8np_409ac970-9fc4-4b77-9548-1714f54c8d51/dnsmasq-dns/0.log" Dec 13 09:44:30 crc kubenswrapper[4971]: I1213 09:44:30.353414 4971 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_glance-default-internal-api-0_3158ebe6-4c8c-41d1-bf95-6b4492a7e0aa/glance-log/0.log" Dec 13 09:44:30 crc kubenswrapper[4971]: I1213 09:44:30.404086 4971 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_glance-default-internal-api-0_3158ebe6-4c8c-41d1-bf95-6b4492a7e0aa/glance-httpd/0.log" Dec 13 09:44:30 crc kubenswrapper[4971]: I1213 09:44:30.700336 4971 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_horizon-797dfcfc58-wlch7_9a4736be-5537-4638-a203-d36db1ae35cc/horizon/0.log" Dec 13 09:44:30 crc kubenswrapper[4971]: I1213 09:44:30.759642 4971 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_horizontest-tests-horizontest_abaeac3e-11a1-40b0-b694-26822d0c02bf/horizontest-tests-horizontest/0.log" Dec 13 09:44:31 crc kubenswrapper[4971]: I1213 09:44:31.021446 4971 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_install-certs-edpm-deployment-openstack-edpm-ipam-lkhcs_2f26ed2c-695e-4425-85cc-6e03e59b39db/install-certs-edpm-deployment-openstack-edpm-ipam/0.log" Dec 13 09:44:31 crc kubenswrapper[4971]: I1213 09:44:31.048312 4971 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_install-os-edpm-deployment-openstack-edpm-ipam-ffdjh_b9a8435d-184f-4451-82cc-3b88e2e19050/install-os-edpm-deployment-openstack-edpm-ipam/0.log" Dec 13 09:44:31 crc kubenswrapper[4971]: I1213 09:44:31.688659 4971 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_keystone-cron-29426881-xx4lc_13b568d7-cd02-480b-aa55-ef6074c71e60/keystone-cron/0.log" Dec 13 09:44:31 crc kubenswrapper[4971]: I1213 09:44:31.974429 4971 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_keystone-cron-29426941-c6mj5_a7d15eb0-4473-488c-8355-db1a4d139e7c/keystone-cron/0.log" Dec 13 09:44:32 crc kubenswrapper[4971]: I1213 09:44:32.052755 4971 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_kube-state-metrics-0_8314d6ed-4778-4709-a494-0642e0ce0617/kube-state-metrics/0.log" Dec 13 09:44:32 crc kubenswrapper[4971]: I1213 09:44:32.296878 4971 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_libvirt-edpm-deployment-openstack-edpm-ipam-47wpm_ce803e3c-be8a-4ee3-ac0b-42b035c043a5/libvirt-edpm-deployment-openstack-edpm-ipam/0.log" Dec 13 09:44:32 crc kubenswrapper[4971]: I1213 09:44:32.357242 4971 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_horizon-797dfcfc58-wlch7_9a4736be-5537-4638-a203-d36db1ae35cc/horizon-log/0.log" Dec 13 09:44:32 crc kubenswrapper[4971]: I1213 09:44:32.576894 4971 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_manila-api-0_e8e297bb-43e7-45b9-a82b-a46c104519c1/manila-api-log/0.log" Dec 13 09:44:32 crc kubenswrapper[4971]: I1213 09:44:32.908541 4971 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_manila-scheduler-0_13b76412-5c09-4398-b34f-f07adad9d270/probe/0.log" Dec 13 09:44:32 crc kubenswrapper[4971]: I1213 09:44:32.954305 4971 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_manila-scheduler-0_13b76412-5c09-4398-b34f-f07adad9d270/manila-scheduler/0.log" Dec 13 09:44:32 crc kubenswrapper[4971]: I1213 09:44:32.971434 4971 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_manila-api-0_e8e297bb-43e7-45b9-a82b-a46c104519c1/manila-api/0.log" Dec 13 09:44:33 crc kubenswrapper[4971]: I1213 09:44:33.247983 4971 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_manila-share-share1-0_7d1dae22-44f3-47a3-99fa-a281fcffcdd2/manila-share/0.log" Dec 13 09:44:33 crc kubenswrapper[4971]: I1213 09:44:33.291161 4971 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_manila-share-share1-0_7d1dae22-44f3-47a3-99fa-a281fcffcdd2/probe/0.log" Dec 13 09:44:34 crc kubenswrapper[4971]: I1213 09:44:34.117051 4971 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_neutron-metadata-edpm-deployment-openstack-edpm-ipam-tpz7w_3a9bc76d-7313-4304-ad56-d6c075837f50/neutron-metadata-edpm-deployment-openstack-edpm-ipam/0.log" Dec 13 09:44:35 crc kubenswrapper[4971]: I1213 09:44:35.030198 4971 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_neutron-7956479f45-8g7zz_4426fcd4-0b89-4e70-8f2f-43b056896953/neutron-httpd/0.log" Dec 13 09:44:36 crc kubenswrapper[4971]: I1213 09:44:36.057249 4971 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_keystone-7969997c86-5t8c8_8e1b68c0-d925-4c7c-8904-d71b9f6ef0e1/keystone-api/0.log" Dec 13 09:44:36 crc kubenswrapper[4971]: I1213 09:44:36.234757 4971 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_neutron-7956479f45-8g7zz_4426fcd4-0b89-4e70-8f2f-43b056896953/neutron-api/0.log" Dec 13 09:44:37 crc kubenswrapper[4971]: I1213 09:44:37.097500 4971 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_nova-cell1-conductor-0_c4b4d448-bf03-4d5c-9017-3b2257922796/nova-cell1-conductor-conductor/0.log" Dec 13 09:44:37 crc kubenswrapper[4971]: I1213 09:44:37.282293 4971 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_nova-cell0-conductor-0_5220d566-7310-4e2f-b5b0-4e0c5e826aa9/nova-cell0-conductor-conductor/0.log" Dec 13 09:44:37 crc kubenswrapper[4971]: I1213 09:44:37.854332 4971 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_nova-cell1-novncproxy-0_38d62929-7b21-4bf7-a2ac-f31047134d54/nova-cell1-novncproxy-novncproxy/0.log" Dec 13 09:44:38 crc kubenswrapper[4971]: I1213 09:44:38.486915 4971 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_nova-custom-ceph-edpm-deployment-openstack-edpm-ipam-4lrjf_e3e414b3-58e3-4da8-9542-7b6d2f229b8d/nova-custom-ceph-edpm-deployment-openstack-edpm-ipam/0.log" Dec 13 09:44:38 crc kubenswrapper[4971]: I1213 09:44:38.922781 4971 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_nova-metadata-0_d4bb47ae-3205-452b-b596-563d3833ee82/nova-metadata-log/0.log" Dec 13 09:44:40 crc kubenswrapper[4971]: I1213 09:44:40.390899 4971 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_nova-scheduler-0_14465eba-34c5-43d3-ad9d-06b454f79963/nova-scheduler-scheduler/0.log" Dec 13 09:44:40 crc kubenswrapper[4971]: I1213 09:44:40.463228 4971 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_nova-api-0_d66b7a63-bb42-46dc-8dbb-94452070e5b2/nova-api-log/0.log" Dec 13 09:44:40 crc kubenswrapper[4971]: I1213 09:44:40.730150 4971 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_openstack-cell1-galera-0_96571a02-70f1-46e7-bdd3-935b1f43ecc6/mysql-bootstrap/0.log" Dec 13 09:44:40 crc kubenswrapper[4971]: I1213 09:44:40.998712 4971 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_openstack-cell1-galera-0_96571a02-70f1-46e7-bdd3-935b1f43ecc6/mysql-bootstrap/0.log" Dec 13 09:44:41 crc kubenswrapper[4971]: I1213 09:44:41.036440 4971 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_openstack-cell1-galera-0_96571a02-70f1-46e7-bdd3-935b1f43ecc6/galera/0.log" Dec 13 09:44:41 crc kubenswrapper[4971]: I1213 09:44:41.291876 4971 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_openstack-galera-0_91831fd5-fc52-46ec-ac49-63859378116b/mysql-bootstrap/0.log" Dec 13 09:44:41 crc kubenswrapper[4971]: I1213 09:44:41.586458 4971 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_openstack-galera-0_91831fd5-fc52-46ec-ac49-63859378116b/mysql-bootstrap/0.log" Dec 13 09:44:41 crc kubenswrapper[4971]: I1213 09:44:41.717485 4971 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_openstack-galera-0_91831fd5-fc52-46ec-ac49-63859378116b/galera/0.log" Dec 13 09:44:42 crc kubenswrapper[4971]: I1213 09:44:42.053280 4971 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_openstackclient_b21860ea-6663-4bc0-ac29-b91d640fad5a/openstackclient/0.log" Dec 13 09:44:42 crc kubenswrapper[4971]: I1213 09:44:42.152539 4971 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_nova-api-0_d66b7a63-bb42-46dc-8dbb-94452070e5b2/nova-api-api/0.log" Dec 13 09:44:42 crc kubenswrapper[4971]: I1213 09:44:42.287470 4971 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovn-controller-k5fz9_c0bf2685-ca54-4508-a580-531b05473e76/ovn-controller/0.log" Dec 13 09:44:42 crc kubenswrapper[4971]: I1213 09:44:42.431240 4971 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovn-controller-metrics-r2cbj_fa04788e-0c35-4205-97ed-cb1e38ec9e69/openstack-network-exporter/0.log" Dec 13 09:44:42 crc kubenswrapper[4971]: I1213 09:44:42.691500 4971 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovn-controller-ovs-jwh28_1920db8d-2d52-46fa-a4cc-ad787e9886fc/ovsdb-server-init/0.log" Dec 13 09:44:42 crc kubenswrapper[4971]: I1213 09:44:42.991299 4971 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovn-controller-ovs-jwh28_1920db8d-2d52-46fa-a4cc-ad787e9886fc/ovsdb-server-init/0.log" Dec 13 09:44:43 crc kubenswrapper[4971]: I1213 09:44:43.024193 4971 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovn-controller-ovs-jwh28_1920db8d-2d52-46fa-a4cc-ad787e9886fc/ovs-vswitchd/0.log" Dec 13 09:44:43 crc kubenswrapper[4971]: I1213 09:44:43.108126 4971 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovn-controller-ovs-jwh28_1920db8d-2d52-46fa-a4cc-ad787e9886fc/ovsdb-server/0.log" Dec 13 09:44:43 crc kubenswrapper[4971]: I1213 09:44:43.351495 4971 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovn-edpm-deployment-openstack-edpm-ipam-lswgl_9f1e6cde-d76e-45f1-a6a6-8546b16e389b/ovn-edpm-deployment-openstack-edpm-ipam/0.log" Dec 13 09:44:43 crc kubenswrapper[4971]: I1213 09:44:43.568829 4971 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovn-northd-0_2bbf81f1-6733-4320-9bf5-c73a00d9051a/openstack-network-exporter/0.log" Dec 13 09:44:43 crc kubenswrapper[4971]: I1213 09:44:43.657844 4971 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovn-northd-0_2bbf81f1-6733-4320-9bf5-c73a00d9051a/ovn-northd/0.log" Dec 13 09:44:43 crc kubenswrapper[4971]: I1213 09:44:43.852997 4971 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovsdbserver-nb-0_3952f27c-fb18-4e8a-b3a4-a3fa077378c4/openstack-network-exporter/0.log" Dec 13 09:44:44 crc kubenswrapper[4971]: I1213 09:44:44.018299 4971 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovsdbserver-nb-0_3952f27c-fb18-4e8a-b3a4-a3fa077378c4/ovsdbserver-nb/0.log" Dec 13 09:44:44 crc kubenswrapper[4971]: I1213 09:44:44.225778 4971 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovsdbserver-sb-0_a85733fd-9af5-4fc4-8096-9b368a793fe7/openstack-network-exporter/0.log" Dec 13 09:44:44 crc kubenswrapper[4971]: I1213 09:44:44.286811 4971 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovsdbserver-sb-0_a85733fd-9af5-4fc4-8096-9b368a793fe7/ovsdbserver-sb/0.log" Dec 13 09:44:44 crc kubenswrapper[4971]: I1213 09:44:44.978469 4971 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_rabbitmq-cell1-server-0_299317ec-a652-40f2-b0c5-e25f099204c4/setup-container/0.log" Dec 13 09:44:45 crc kubenswrapper[4971]: I1213 09:44:45.192590 4971 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_rabbitmq-cell1-server-0_299317ec-a652-40f2-b0c5-e25f099204c4/setup-container/0.log" Dec 13 09:44:45 crc kubenswrapper[4971]: I1213 09:44:45.271564 4971 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_placement-569b6fd56b-w7xmm_abb2fa2a-6153-4b45-83b9-2c34265c5e7f/placement-api/0.log" Dec 13 09:44:45 crc kubenswrapper[4971]: I1213 09:44:45.428853 4971 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_rabbitmq-cell1-server-0_299317ec-a652-40f2-b0c5-e25f099204c4/rabbitmq/0.log" Dec 13 09:44:45 crc kubenswrapper[4971]: I1213 09:44:45.541148 4971 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_placement-569b6fd56b-w7xmm_abb2fa2a-6153-4b45-83b9-2c34265c5e7f/placement-log/0.log" Dec 13 09:44:45 crc kubenswrapper[4971]: I1213 09:44:45.682351 4971 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_nova-metadata-0_d4bb47ae-3205-452b-b596-563d3833ee82/nova-metadata-metadata/0.log" Dec 13 09:44:45 crc kubenswrapper[4971]: I1213 09:44:45.696282 4971 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_rabbitmq-server-0_42eb5011-6651-43cb-a562-c89cc6df2c53/setup-container/0.log" Dec 13 09:44:46 crc kubenswrapper[4971]: I1213 09:44:46.030056 4971 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_rabbitmq-server-0_42eb5011-6651-43cb-a562-c89cc6df2c53/setup-container/0.log" Dec 13 09:44:46 crc kubenswrapper[4971]: I1213 09:44:46.062992 4971 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_rabbitmq-server-0_42eb5011-6651-43cb-a562-c89cc6df2c53/rabbitmq/0.log" Dec 13 09:44:46 crc kubenswrapper[4971]: I1213 09:44:46.184460 4971 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_reboot-os-edpm-deployment-openstack-edpm-ipam-mmm7v_94cbb455-7651-4d43-ad24-495457496a2d/reboot-os-edpm-deployment-openstack-edpm-ipam/0.log" Dec 13 09:44:46 crc kubenswrapper[4971]: I1213 09:44:46.592072 4971 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_repo-setup-edpm-deployment-openstack-edpm-ipam-fpgpm_53d7daa1-9679-44ba-9023-808ce5c250a6/repo-setup-edpm-deployment-openstack-edpm-ipam/0.log" Dec 13 09:44:46 crc kubenswrapper[4971]: I1213 09:44:46.593726 4971 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_run-os-edpm-deployment-openstack-edpm-ipam-tn7h4_88c03dc6-31a0-4c8c-93b6-61a9c895dffe/run-os-edpm-deployment-openstack-edpm-ipam/0.log" Dec 13 09:44:46 crc kubenswrapper[4971]: I1213 09:44:46.767320 4971 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ssh-known-hosts-edpm-deployment-b4wgp_82daff50-eacf-45ef-b206-ffe0e1d2fe89/ssh-known-hosts-edpm-deployment/0.log" Dec 13 09:44:46 crc kubenswrapper[4971]: I1213 09:44:46.997609 4971 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_tempest-tests-tempest-s00-full_879e5242-5aa8-44fb-b759-6babf9c6559d/tempest-tests-tempest-tests-runner/0.log" Dec 13 09:44:47 crc kubenswrapper[4971]: I1213 09:44:47.080740 4971 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_tempest-tests-tempest-s01-single-test_b0fcccc8-3a38-42b6-a10c-57a0a44b48c3/tempest-tests-tempest-tests-runner/0.log" Dec 13 09:44:47 crc kubenswrapper[4971]: I1213 09:44:47.271478 4971 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_test-operator-logs-pod-ansibletest-ansibletest-ansibletest_4cf376d3-529c-41ee-8da5-9aa3bb37cfed/test-operator-logs-container/0.log" Dec 13 09:44:47 crc kubenswrapper[4971]: I1213 09:44:47.411819 4971 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_test-operator-logs-pod-horizontest-horizontest-tests-horizontest_73a838a8-433b-4ddf-ab44-1527de29aa6f/test-operator-logs-container/0.log" Dec 13 09:44:47 crc kubenswrapper[4971]: I1213 09:44:47.650658 4971 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_test-operator-logs-pod-tempest-tempest-tests-tempest_b5af123d-e343-4e82-861d-3eaed6d9ca0c/test-operator-logs-container/0.log" Dec 13 09:44:47 crc kubenswrapper[4971]: I1213 09:44:47.799385 4971 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_test-operator-logs-pod-tobiko-tobiko-tests-tobiko_d608d411-d575-467c-8738-17284ac309d7/test-operator-logs-container/0.log" Dec 13 09:44:48 crc kubenswrapper[4971]: I1213 09:44:48.091665 4971 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_tobiko-tests-tobiko-s01-sanity_32d4b615-70d7-4af1-9280-8fb2452feb76/tobiko-tests-tobiko/0.log" Dec 13 09:44:48 crc kubenswrapper[4971]: I1213 09:44:48.104218 4971 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_tobiko-tests-tobiko-s00-podified-functional_0a604d8e-de8c-4ae6-b13a-aafd2adee17a/tobiko-tests-tobiko/0.log" Dec 13 09:44:48 crc kubenswrapper[4971]: I1213 09:44:48.600313 4971 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_validate-network-edpm-deployment-openstack-edpm-ipam-4lhml_2185de1e-5198-4f0f-8918-1e90aadb7fb7/validate-network-edpm-deployment-openstack-edpm-ipam/0.log" Dec 13 09:44:50 crc kubenswrapper[4971]: I1213 09:44:50.217794 4971 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_memcached-0_e6adf8ca-c137-4b7d-bf8c-74c0766e57f0/memcached/0.log" Dec 13 09:45:00 crc kubenswrapper[4971]: I1213 09:45:00.183752 4971 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29426985-ppnbf"] Dec 13 09:45:00 crc kubenswrapper[4971]: E1213 09:45:00.185081 4971 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="83363a89-be3c-45e5-b8af-816d5b3dd17c" containerName="container-00" Dec 13 09:45:00 crc kubenswrapper[4971]: I1213 09:45:00.185099 4971 state_mem.go:107] "Deleted CPUSet assignment" podUID="83363a89-be3c-45e5-b8af-816d5b3dd17c" containerName="container-00" Dec 13 09:45:00 crc kubenswrapper[4971]: I1213 09:45:00.185316 4971 memory_manager.go:354] "RemoveStaleState removing state" podUID="83363a89-be3c-45e5-b8af-816d5b3dd17c" containerName="container-00" Dec 13 09:45:00 crc kubenswrapper[4971]: I1213 09:45:00.186150 4971 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29426985-ppnbf" Dec 13 09:45:00 crc kubenswrapper[4971]: I1213 09:45:00.193911 4971 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"collect-profiles-dockercfg-kzf4t" Dec 13 09:45:00 crc kubenswrapper[4971]: I1213 09:45:00.202032 4971 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29426985-ppnbf"] Dec 13 09:45:00 crc kubenswrapper[4971]: I1213 09:45:00.205882 4971 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-operator-lifecycle-manager"/"collect-profiles-config" Dec 13 09:45:00 crc kubenswrapper[4971]: I1213 09:45:00.211493 4971 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-hqbsn\" (UniqueName: \"kubernetes.io/projected/67b2dfa6-27b0-4dc4-bd58-0ac26186530f-kube-api-access-hqbsn\") pod \"collect-profiles-29426985-ppnbf\" (UID: \"67b2dfa6-27b0-4dc4-bd58-0ac26186530f\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29426985-ppnbf" Dec 13 09:45:00 crc kubenswrapper[4971]: I1213 09:45:00.211555 4971 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/67b2dfa6-27b0-4dc4-bd58-0ac26186530f-config-volume\") pod \"collect-profiles-29426985-ppnbf\" (UID: \"67b2dfa6-27b0-4dc4-bd58-0ac26186530f\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29426985-ppnbf" Dec 13 09:45:00 crc kubenswrapper[4971]: I1213 09:45:00.211627 4971 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/67b2dfa6-27b0-4dc4-bd58-0ac26186530f-secret-volume\") pod \"collect-profiles-29426985-ppnbf\" (UID: \"67b2dfa6-27b0-4dc4-bd58-0ac26186530f\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29426985-ppnbf" Dec 13 09:45:00 crc kubenswrapper[4971]: I1213 09:45:00.315744 4971 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-hqbsn\" (UniqueName: \"kubernetes.io/projected/67b2dfa6-27b0-4dc4-bd58-0ac26186530f-kube-api-access-hqbsn\") pod \"collect-profiles-29426985-ppnbf\" (UID: \"67b2dfa6-27b0-4dc4-bd58-0ac26186530f\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29426985-ppnbf" Dec 13 09:45:00 crc kubenswrapper[4971]: I1213 09:45:00.315814 4971 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/67b2dfa6-27b0-4dc4-bd58-0ac26186530f-config-volume\") pod \"collect-profiles-29426985-ppnbf\" (UID: \"67b2dfa6-27b0-4dc4-bd58-0ac26186530f\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29426985-ppnbf" Dec 13 09:45:00 crc kubenswrapper[4971]: I1213 09:45:00.315896 4971 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/67b2dfa6-27b0-4dc4-bd58-0ac26186530f-secret-volume\") pod \"collect-profiles-29426985-ppnbf\" (UID: \"67b2dfa6-27b0-4dc4-bd58-0ac26186530f\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29426985-ppnbf" Dec 13 09:45:00 crc kubenswrapper[4971]: I1213 09:45:00.316852 4971 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/67b2dfa6-27b0-4dc4-bd58-0ac26186530f-config-volume\") pod \"collect-profiles-29426985-ppnbf\" (UID: \"67b2dfa6-27b0-4dc4-bd58-0ac26186530f\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29426985-ppnbf" Dec 13 09:45:00 crc kubenswrapper[4971]: I1213 09:45:00.327948 4971 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/67b2dfa6-27b0-4dc4-bd58-0ac26186530f-secret-volume\") pod \"collect-profiles-29426985-ppnbf\" (UID: \"67b2dfa6-27b0-4dc4-bd58-0ac26186530f\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29426985-ppnbf" Dec 13 09:45:00 crc kubenswrapper[4971]: I1213 09:45:00.337275 4971 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-hqbsn\" (UniqueName: \"kubernetes.io/projected/67b2dfa6-27b0-4dc4-bd58-0ac26186530f-kube-api-access-hqbsn\") pod \"collect-profiles-29426985-ppnbf\" (UID: \"67b2dfa6-27b0-4dc4-bd58-0ac26186530f\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29426985-ppnbf" Dec 13 09:45:00 crc kubenswrapper[4971]: I1213 09:45:00.516795 4971 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29426985-ppnbf" Dec 13 09:45:01 crc kubenswrapper[4971]: I1213 09:45:01.066925 4971 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29426985-ppnbf"] Dec 13 09:45:02 crc kubenswrapper[4971]: I1213 09:45:02.021859 4971 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29426985-ppnbf" event={"ID":"67b2dfa6-27b0-4dc4-bd58-0ac26186530f","Type":"ContainerStarted","Data":"827fc19d5c035fc5320fbdedfc34b1defdb4a2bcbe6ca40cef7bc44370cd82e0"} Dec 13 09:45:03 crc kubenswrapper[4971]: I1213 09:45:03.035432 4971 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29426985-ppnbf" event={"ID":"67b2dfa6-27b0-4dc4-bd58-0ac26186530f","Type":"ContainerStarted","Data":"0bd0c20bc268f5b919cf45041265adde7829dc9c311a0482ad1b66216003b2a3"} Dec 13 09:45:03 crc kubenswrapper[4971]: I1213 09:45:03.065978 4971 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-operator-lifecycle-manager/collect-profiles-29426985-ppnbf" podStartSLOduration=3.065953939 podStartE2EDuration="3.065953939s" podCreationTimestamp="2025-12-13 09:45:00 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-13 09:45:03.055211326 +0000 UTC m=+10559.659620774" watchObservedRunningTime="2025-12-13 09:45:03.065953939 +0000 UTC m=+10559.670363387" Dec 13 09:45:04 crc kubenswrapper[4971]: I1213 09:45:04.056776 4971 generic.go:334] "Generic (PLEG): container finished" podID="67b2dfa6-27b0-4dc4-bd58-0ac26186530f" containerID="0bd0c20bc268f5b919cf45041265adde7829dc9c311a0482ad1b66216003b2a3" exitCode=0 Dec 13 09:45:04 crc kubenswrapper[4971]: I1213 09:45:04.056913 4971 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29426985-ppnbf" event={"ID":"67b2dfa6-27b0-4dc4-bd58-0ac26186530f","Type":"ContainerDied","Data":"0bd0c20bc268f5b919cf45041265adde7829dc9c311a0482ad1b66216003b2a3"} Dec 13 09:45:05 crc kubenswrapper[4971]: I1213 09:45:05.528686 4971 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29426985-ppnbf" Dec 13 09:45:05 crc kubenswrapper[4971]: I1213 09:45:05.676654 4971 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/67b2dfa6-27b0-4dc4-bd58-0ac26186530f-config-volume\") pod \"67b2dfa6-27b0-4dc4-bd58-0ac26186530f\" (UID: \"67b2dfa6-27b0-4dc4-bd58-0ac26186530f\") " Dec 13 09:45:05 crc kubenswrapper[4971]: I1213 09:45:05.677084 4971 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-hqbsn\" (UniqueName: \"kubernetes.io/projected/67b2dfa6-27b0-4dc4-bd58-0ac26186530f-kube-api-access-hqbsn\") pod \"67b2dfa6-27b0-4dc4-bd58-0ac26186530f\" (UID: \"67b2dfa6-27b0-4dc4-bd58-0ac26186530f\") " Dec 13 09:45:05 crc kubenswrapper[4971]: I1213 09:45:05.677167 4971 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/67b2dfa6-27b0-4dc4-bd58-0ac26186530f-secret-volume\") pod \"67b2dfa6-27b0-4dc4-bd58-0ac26186530f\" (UID: \"67b2dfa6-27b0-4dc4-bd58-0ac26186530f\") " Dec 13 09:45:05 crc kubenswrapper[4971]: I1213 09:45:05.680509 4971 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/67b2dfa6-27b0-4dc4-bd58-0ac26186530f-config-volume" (OuterVolumeSpecName: "config-volume") pod "67b2dfa6-27b0-4dc4-bd58-0ac26186530f" (UID: "67b2dfa6-27b0-4dc4-bd58-0ac26186530f"). InnerVolumeSpecName "config-volume". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 13 09:45:05 crc kubenswrapper[4971]: I1213 09:45:05.699867 4971 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/67b2dfa6-27b0-4dc4-bd58-0ac26186530f-secret-volume" (OuterVolumeSpecName: "secret-volume") pod "67b2dfa6-27b0-4dc4-bd58-0ac26186530f" (UID: "67b2dfa6-27b0-4dc4-bd58-0ac26186530f"). InnerVolumeSpecName "secret-volume". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 13 09:45:05 crc kubenswrapper[4971]: I1213 09:45:05.706948 4971 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/67b2dfa6-27b0-4dc4-bd58-0ac26186530f-kube-api-access-hqbsn" (OuterVolumeSpecName: "kube-api-access-hqbsn") pod "67b2dfa6-27b0-4dc4-bd58-0ac26186530f" (UID: "67b2dfa6-27b0-4dc4-bd58-0ac26186530f"). InnerVolumeSpecName "kube-api-access-hqbsn". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 13 09:45:05 crc kubenswrapper[4971]: E1213 09:45:05.742840 4971 cadvisor_stats_provider.go:516] "Partial failure issuing cadvisor.ContainerInfoV2" err="partial failures: [\"/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod67b2dfa6_27b0_4dc4_bd58_0ac26186530f.slice/crio-0bd0c20bc268f5b919cf45041265adde7829dc9c311a0482ad1b66216003b2a3.scope\": RecentStats: unable to find data in memory cache]" Dec 13 09:45:05 crc kubenswrapper[4971]: I1213 09:45:05.779592 4971 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-hqbsn\" (UniqueName: \"kubernetes.io/projected/67b2dfa6-27b0-4dc4-bd58-0ac26186530f-kube-api-access-hqbsn\") on node \"crc\" DevicePath \"\"" Dec 13 09:45:05 crc kubenswrapper[4971]: I1213 09:45:05.779661 4971 reconciler_common.go:293] "Volume detached for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/67b2dfa6-27b0-4dc4-bd58-0ac26186530f-secret-volume\") on node \"crc\" DevicePath \"\"" Dec 13 09:45:05 crc kubenswrapper[4971]: I1213 09:45:05.779671 4971 reconciler_common.go:293] "Volume detached for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/67b2dfa6-27b0-4dc4-bd58-0ac26186530f-config-volume\") on node \"crc\" DevicePath \"\"" Dec 13 09:45:06 crc kubenswrapper[4971]: I1213 09:45:06.080746 4971 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29426985-ppnbf" event={"ID":"67b2dfa6-27b0-4dc4-bd58-0ac26186530f","Type":"ContainerDied","Data":"827fc19d5c035fc5320fbdedfc34b1defdb4a2bcbe6ca40cef7bc44370cd82e0"} Dec 13 09:45:06 crc kubenswrapper[4971]: I1213 09:45:06.081153 4971 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="827fc19d5c035fc5320fbdedfc34b1defdb4a2bcbe6ca40cef7bc44370cd82e0" Dec 13 09:45:06 crc kubenswrapper[4971]: I1213 09:45:06.081102 4971 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29426985-ppnbf" Dec 13 09:45:06 crc kubenswrapper[4971]: I1213 09:45:06.223489 4971 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29426940-m6wck"] Dec 13 09:45:06 crc kubenswrapper[4971]: I1213 09:45:06.245335 4971 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29426940-m6wck"] Dec 13 09:45:07 crc kubenswrapper[4971]: I1213 09:45:07.786705 4971 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="a57afece-3e8b-4b98-80bb-437bc66f2469" path="/var/lib/kubelet/pods/a57afece-3e8b-4b98-80bb-437bc66f2469/volumes" Dec 13 09:45:11 crc kubenswrapper[4971]: I1213 09:45:11.730707 4971 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-must-gather-g74xz/must-gather-p9ltj"] Dec 13 09:45:11 crc kubenswrapper[4971]: E1213 09:45:11.731774 4971 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="67b2dfa6-27b0-4dc4-bd58-0ac26186530f" containerName="collect-profiles" Dec 13 09:45:11 crc kubenswrapper[4971]: I1213 09:45:11.731788 4971 state_mem.go:107] "Deleted CPUSet assignment" podUID="67b2dfa6-27b0-4dc4-bd58-0ac26186530f" containerName="collect-profiles" Dec 13 09:45:11 crc kubenswrapper[4971]: I1213 09:45:11.731986 4971 memory_manager.go:354] "RemoveStaleState removing state" podUID="67b2dfa6-27b0-4dc4-bd58-0ac26186530f" containerName="collect-profiles" Dec 13 09:45:11 crc kubenswrapper[4971]: I1213 09:45:11.733150 4971 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-g74xz/must-gather-p9ltj" Dec 13 09:45:11 crc kubenswrapper[4971]: I1213 09:45:11.735295 4971 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-must-gather-g74xz"/"kube-root-ca.crt" Dec 13 09:45:11 crc kubenswrapper[4971]: I1213 09:45:11.736266 4971 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-must-gather-g74xz"/"openshift-service-ca.crt" Dec 13 09:45:11 crc kubenswrapper[4971]: I1213 09:45:11.765758 4971 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-must-gather-g74xz/must-gather-p9ltj"] Dec 13 09:45:11 crc kubenswrapper[4971]: I1213 09:45:11.928612 4971 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-6b7m5\" (UniqueName: \"kubernetes.io/projected/d92c389a-6647-4873-8be9-53ff04c2d3fd-kube-api-access-6b7m5\") pod \"must-gather-p9ltj\" (UID: \"d92c389a-6647-4873-8be9-53ff04c2d3fd\") " pod="openshift-must-gather-g74xz/must-gather-p9ltj" Dec 13 09:45:11 crc kubenswrapper[4971]: I1213 09:45:11.928718 4971 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"must-gather-output\" (UniqueName: \"kubernetes.io/empty-dir/d92c389a-6647-4873-8be9-53ff04c2d3fd-must-gather-output\") pod \"must-gather-p9ltj\" (UID: \"d92c389a-6647-4873-8be9-53ff04c2d3fd\") " pod="openshift-must-gather-g74xz/must-gather-p9ltj" Dec 13 09:45:12 crc kubenswrapper[4971]: I1213 09:45:12.031250 4971 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-6b7m5\" (UniqueName: \"kubernetes.io/projected/d92c389a-6647-4873-8be9-53ff04c2d3fd-kube-api-access-6b7m5\") pod \"must-gather-p9ltj\" (UID: \"d92c389a-6647-4873-8be9-53ff04c2d3fd\") " pod="openshift-must-gather-g74xz/must-gather-p9ltj" Dec 13 09:45:12 crc kubenswrapper[4971]: I1213 09:45:12.031381 4971 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"must-gather-output\" (UniqueName: \"kubernetes.io/empty-dir/d92c389a-6647-4873-8be9-53ff04c2d3fd-must-gather-output\") pod \"must-gather-p9ltj\" (UID: \"d92c389a-6647-4873-8be9-53ff04c2d3fd\") " pod="openshift-must-gather-g74xz/must-gather-p9ltj" Dec 13 09:45:12 crc kubenswrapper[4971]: I1213 09:45:12.032093 4971 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"must-gather-output\" (UniqueName: \"kubernetes.io/empty-dir/d92c389a-6647-4873-8be9-53ff04c2d3fd-must-gather-output\") pod \"must-gather-p9ltj\" (UID: \"d92c389a-6647-4873-8be9-53ff04c2d3fd\") " pod="openshift-must-gather-g74xz/must-gather-p9ltj" Dec 13 09:45:12 crc kubenswrapper[4971]: I1213 09:45:12.066849 4971 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-6b7m5\" (UniqueName: \"kubernetes.io/projected/d92c389a-6647-4873-8be9-53ff04c2d3fd-kube-api-access-6b7m5\") pod \"must-gather-p9ltj\" (UID: \"d92c389a-6647-4873-8be9-53ff04c2d3fd\") " pod="openshift-must-gather-g74xz/must-gather-p9ltj" Dec 13 09:45:12 crc kubenswrapper[4971]: I1213 09:45:12.353558 4971 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-g74xz/must-gather-p9ltj" Dec 13 09:45:12 crc kubenswrapper[4971]: I1213 09:45:12.909779 4971 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-must-gather-g74xz/must-gather-p9ltj"] Dec 13 09:45:13 crc kubenswrapper[4971]: I1213 09:45:13.181089 4971 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-g74xz/must-gather-p9ltj" event={"ID":"d92c389a-6647-4873-8be9-53ff04c2d3fd","Type":"ContainerStarted","Data":"462b9f4f16974d39c235beb8e93f656b17fdf5e4f9eac2682c756a07df097d33"} Dec 13 09:45:13 crc kubenswrapper[4971]: E1213 09:45:13.784383 4971 kubelet_pods.go:538] "Hostname for pod was too long, truncated it" podName="test-operator-logs-pod-horizontest-horizontest-tests-horizontest" hostnameMaxLen=63 truncatedHostname="test-operator-logs-pod-horizontest-horizontest-tests-horizontes" Dec 13 09:45:14 crc kubenswrapper[4971]: I1213 09:45:14.197650 4971 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-g74xz/must-gather-p9ltj" event={"ID":"d92c389a-6647-4873-8be9-53ff04c2d3fd","Type":"ContainerStarted","Data":"66c643c4043229d4a1f17628345e97eff91da8ce8b6bbbc20d6e26c5f9a281e6"} Dec 13 09:45:14 crc kubenswrapper[4971]: I1213 09:45:14.197700 4971 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-g74xz/must-gather-p9ltj" event={"ID":"d92c389a-6647-4873-8be9-53ff04c2d3fd","Type":"ContainerStarted","Data":"1e352e561f17b984001acd985b962cbbb93c4fc02c09a78dd54852cca53bcaf6"} Dec 13 09:45:14 crc kubenswrapper[4971]: I1213 09:45:14.223981 4971 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-must-gather-g74xz/must-gather-p9ltj" podStartSLOduration=3.223955664 podStartE2EDuration="3.223955664s" podCreationTimestamp="2025-12-13 09:45:11 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-13 09:45:14.215132239 +0000 UTC m=+10570.819541697" watchObservedRunningTime="2025-12-13 09:45:14.223955664 +0000 UTC m=+10570.828365112" Dec 13 09:45:16 crc kubenswrapper[4971]: E1213 09:45:16.070396 4971 cadvisor_stats_provider.go:516] "Partial failure issuing cadvisor.ContainerInfoV2" err="partial failures: [\"/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod67b2dfa6_27b0_4dc4_bd58_0ac26186530f.slice/crio-0bd0c20bc268f5b919cf45041265adde7829dc9c311a0482ad1b66216003b2a3.scope\": RecentStats: unable to find data in memory cache]" Dec 13 09:45:16 crc kubenswrapper[4971]: I1213 09:45:16.153711 4971 patch_prober.go:28] interesting pod/machine-config-daemon-82xjz container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 13 09:45:16 crc kubenswrapper[4971]: I1213 09:45:16.153804 4971 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-82xjz" podUID="e0c4af12-edda-41d7-9b44-f87396174ef2" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 13 09:45:20 crc kubenswrapper[4971]: I1213 09:45:20.630951 4971 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-must-gather-g74xz/crc-debug-65smr"] Dec 13 09:45:20 crc kubenswrapper[4971]: I1213 09:45:20.637254 4971 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-g74xz/crc-debug-65smr" Dec 13 09:45:20 crc kubenswrapper[4971]: I1213 09:45:20.641718 4971 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-must-gather-g74xz"/"default-dockercfg-2tvhw" Dec 13 09:45:20 crc kubenswrapper[4971]: I1213 09:45:20.644750 4971 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-kfs55\" (UniqueName: \"kubernetes.io/projected/2c7c5219-d904-4610-805a-6324810e3760-kube-api-access-kfs55\") pod \"crc-debug-65smr\" (UID: \"2c7c5219-d904-4610-805a-6324810e3760\") " pod="openshift-must-gather-g74xz/crc-debug-65smr" Dec 13 09:45:20 crc kubenswrapper[4971]: I1213 09:45:20.644816 4971 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host\" (UniqueName: \"kubernetes.io/host-path/2c7c5219-d904-4610-805a-6324810e3760-host\") pod \"crc-debug-65smr\" (UID: \"2c7c5219-d904-4610-805a-6324810e3760\") " pod="openshift-must-gather-g74xz/crc-debug-65smr" Dec 13 09:45:20 crc kubenswrapper[4971]: I1213 09:45:20.747981 4971 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-kfs55\" (UniqueName: \"kubernetes.io/projected/2c7c5219-d904-4610-805a-6324810e3760-kube-api-access-kfs55\") pod \"crc-debug-65smr\" (UID: \"2c7c5219-d904-4610-805a-6324810e3760\") " pod="openshift-must-gather-g74xz/crc-debug-65smr" Dec 13 09:45:20 crc kubenswrapper[4971]: I1213 09:45:20.748075 4971 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host\" (UniqueName: \"kubernetes.io/host-path/2c7c5219-d904-4610-805a-6324810e3760-host\") pod \"crc-debug-65smr\" (UID: \"2c7c5219-d904-4610-805a-6324810e3760\") " pod="openshift-must-gather-g74xz/crc-debug-65smr" Dec 13 09:45:20 crc kubenswrapper[4971]: I1213 09:45:20.748458 4971 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host\" (UniqueName: \"kubernetes.io/host-path/2c7c5219-d904-4610-805a-6324810e3760-host\") pod \"crc-debug-65smr\" (UID: \"2c7c5219-d904-4610-805a-6324810e3760\") " pod="openshift-must-gather-g74xz/crc-debug-65smr" Dec 13 09:45:20 crc kubenswrapper[4971]: I1213 09:45:20.774179 4971 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-kfs55\" (UniqueName: \"kubernetes.io/projected/2c7c5219-d904-4610-805a-6324810e3760-kube-api-access-kfs55\") pod \"crc-debug-65smr\" (UID: \"2c7c5219-d904-4610-805a-6324810e3760\") " pod="openshift-must-gather-g74xz/crc-debug-65smr" Dec 13 09:45:20 crc kubenswrapper[4971]: I1213 09:45:20.963599 4971 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-g74xz/crc-debug-65smr" Dec 13 09:45:21 crc kubenswrapper[4971]: I1213 09:45:21.305737 4971 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-g74xz/crc-debug-65smr" event={"ID":"2c7c5219-d904-4610-805a-6324810e3760","Type":"ContainerStarted","Data":"82dfc97e6d7f85f46bc66f2ac62d963663bf8b0af56c97b98997c24b064fe9cf"} Dec 13 09:45:23 crc kubenswrapper[4971]: I1213 09:45:23.333211 4971 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-g74xz/crc-debug-65smr" event={"ID":"2c7c5219-d904-4610-805a-6324810e3760","Type":"ContainerStarted","Data":"92077d04a5c5f8ddbc19fb5ec1cda65d959a3d4835d1b05c4668b8853e148eb1"} Dec 13 09:45:23 crc kubenswrapper[4971]: I1213 09:45:23.358343 4971 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-must-gather-g74xz/crc-debug-65smr" podStartSLOduration=3.358320406 podStartE2EDuration="3.358320406s" podCreationTimestamp="2025-12-13 09:45:20 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-13 09:45:23.351011207 +0000 UTC m=+10579.955420665" watchObservedRunningTime="2025-12-13 09:45:23.358320406 +0000 UTC m=+10579.962729854" Dec 13 09:45:25 crc kubenswrapper[4971]: I1213 09:45:25.981743 4971 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_barbican-operator-controller-manager-95949466-6hkxl_3b74a6fd-0613-4f48-b585-64570abdb09f/manager/0.log" Dec 13 09:45:26 crc kubenswrapper[4971]: E1213 09:45:26.440044 4971 cadvisor_stats_provider.go:516] "Partial failure issuing cadvisor.ContainerInfoV2" err="partial failures: [\"/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod67b2dfa6_27b0_4dc4_bd58_0ac26186530f.slice/crio-0bd0c20bc268f5b919cf45041265adde7829dc9c311a0482ad1b66216003b2a3.scope\": RecentStats: unable to find data in memory cache]" Dec 13 09:45:26 crc kubenswrapper[4971]: I1213 09:45:26.508938 4971 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_cinder-operator-controller-manager-5cf45c46bd-8sg42_62c93900-e946-4684-9c40-36c1d8ff0468/manager/0.log" Dec 13 09:45:26 crc kubenswrapper[4971]: I1213 09:45:26.556414 4971 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_designate-operator-controller-manager-66f8b87655-qxr27_cfe5eaa1-4beb-49ff-a080-ab230292c8ce/manager/0.log" Dec 13 09:45:26 crc kubenswrapper[4971]: I1213 09:45:26.857494 4971 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_ea5d71784a81c95dff2031a02f0a0b3f756f86f14acad8f152d938f56fl9xnm_8fbc1cb9-af8d-4ab7-92e4-63035d0c1d74/util/0.log" Dec 13 09:45:27 crc kubenswrapper[4971]: I1213 09:45:27.420447 4971 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_ea5d71784a81c95dff2031a02f0a0b3f756f86f14acad8f152d938f56fl9xnm_8fbc1cb9-af8d-4ab7-92e4-63035d0c1d74/pull/0.log" Dec 13 09:45:27 crc kubenswrapper[4971]: I1213 09:45:27.430188 4971 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_ea5d71784a81c95dff2031a02f0a0b3f756f86f14acad8f152d938f56fl9xnm_8fbc1cb9-af8d-4ab7-92e4-63035d0c1d74/util/0.log" Dec 13 09:45:27 crc kubenswrapper[4971]: I1213 09:45:27.538701 4971 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_ea5d71784a81c95dff2031a02f0a0b3f756f86f14acad8f152d938f56fl9xnm_8fbc1cb9-af8d-4ab7-92e4-63035d0c1d74/pull/0.log" Dec 13 09:45:27 crc kubenswrapper[4971]: I1213 09:45:27.782502 4971 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_ea5d71784a81c95dff2031a02f0a0b3f756f86f14acad8f152d938f56fl9xnm_8fbc1cb9-af8d-4ab7-92e4-63035d0c1d74/pull/0.log" Dec 13 09:45:27 crc kubenswrapper[4971]: I1213 09:45:27.814125 4971 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_ea5d71784a81c95dff2031a02f0a0b3f756f86f14acad8f152d938f56fl9xnm_8fbc1cb9-af8d-4ab7-92e4-63035d0c1d74/util/0.log" Dec 13 09:45:27 crc kubenswrapper[4971]: I1213 09:45:27.822017 4971 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_ea5d71784a81c95dff2031a02f0a0b3f756f86f14acad8f152d938f56fl9xnm_8fbc1cb9-af8d-4ab7-92e4-63035d0c1d74/extract/0.log" Dec 13 09:45:28 crc kubenswrapper[4971]: I1213 09:45:28.221673 4971 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_heat-operator-controller-manager-59b8dcb766-4xg9v_808957a1-29e0-4991-bc8a-23f5a17ecb74/manager/0.log" Dec 13 09:45:28 crc kubenswrapper[4971]: I1213 09:45:28.242000 4971 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_glance-operator-controller-manager-767f9d7567-tmc6q_2062b86b-1412-4974-a153-802d9f390944/manager/0.log" Dec 13 09:45:28 crc kubenswrapper[4971]: I1213 09:45:28.850461 4971 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_horizon-operator-controller-manager-6ccf486b9-bb2ft_8e94aa39-05a2-4a63-8b45-a914021bccd7/manager/0.log" Dec 13 09:45:29 crc kubenswrapper[4971]: I1213 09:45:29.000330 4971 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_infra-operator-controller-manager-58944d7758-59mz9_846e13b5-58bb-4a12-a77a-0a4a9b53247f/manager/0.log" Dec 13 09:45:29 crc kubenswrapper[4971]: I1213 09:45:29.210911 4971 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_ironic-operator-controller-manager-f458558d7-m9grc_85d76572-e53b-420f-9ffe-d0534bdb1c96/manager/0.log" Dec 13 09:45:29 crc kubenswrapper[4971]: I1213 09:45:29.431216 4971 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_keystone-operator-controller-manager-5c7cbf548f-qv7p8_3e2847dd-dcdc-4740-a4a8-4a2417247dc5/manager/0.log" Dec 13 09:45:29 crc kubenswrapper[4971]: I1213 09:45:29.623329 4971 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_manila-operator-controller-manager-5fdd9786f7-jjw6m_2761a6c0-8c0d-4430-8976-378c2d6df9f7/manager/0.log" Dec 13 09:45:29 crc kubenswrapper[4971]: I1213 09:45:29.905726 4971 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_mariadb-operator-controller-manager-f76f4954c-cst7r_4bd9b122-00a8-4895-a44b-0773aae9f6a9/manager/0.log" Dec 13 09:45:30 crc kubenswrapper[4971]: I1213 09:45:30.068137 4971 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_neutron-operator-controller-manager-7cd87b778f-l6b8q_0c834237-c914-43bc-b312-b208ba303bb3/manager/0.log" Dec 13 09:45:30 crc kubenswrapper[4971]: I1213 09:45:30.276723 4971 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_nova-operator-controller-manager-5fbbf8b6cc-k4vtr_a8909dd9-9995-4694-9b5b-47106910656c/manager/0.log" Dec 13 09:45:30 crc kubenswrapper[4971]: I1213 09:45:30.472049 4971 scope.go:117] "RemoveContainer" containerID="58d28af04ead826f4260e6960d316420735a77176a9509a70a342d7e593bc519" Dec 13 09:45:30 crc kubenswrapper[4971]: I1213 09:45:30.889843 4971 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_openstack-baremetal-operator-controller-manager-689f887b54jnn8l_706451b8-644b-4151-b5b9-9975082e556f/manager/0.log" Dec 13 09:45:30 crc kubenswrapper[4971]: I1213 09:45:30.937667 4971 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_octavia-operator-controller-manager-68c649d9d-v9t9p_8a160d6b-ce31-4d73-918b-368fbc44eb2b/manager/0.log" Dec 13 09:45:31 crc kubenswrapper[4971]: I1213 09:45:31.310463 4971 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_openstack-operator-controller-operator-5bf59778d8-rrxtx_d0ece558-05e2-4e08-84e0-a314f68a453b/operator/0.log" Dec 13 09:45:31 crc kubenswrapper[4971]: I1213 09:45:31.941245 4971 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_openstack-operator-index-hbjwh_1a3059b7-5a8b-4357-8f6d-caa37f9adef2/registry-server/0.log" Dec 13 09:45:32 crc kubenswrapper[4971]: I1213 09:45:32.267067 4971 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_ovn-operator-controller-manager-bf6d4f946-g2r89_9ed7aa8c-f56d-477d-91f5-ef1944e66d57/manager/0.log" Dec 13 09:45:32 crc kubenswrapper[4971]: I1213 09:45:32.318503 4971 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_placement-operator-controller-manager-8665b56d78-hrflw_6b9f3caa-4280-43f3-8a05-14e72f09c384/manager/0.log" Dec 13 09:45:32 crc kubenswrapper[4971]: I1213 09:45:32.583942 4971 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_rabbitmq-cluster-operator-manager-668c99d594-pzcql_14e41caa-c89a-4ca3-98c3-d75114cd2c76/operator/0.log" Dec 13 09:45:32 crc kubenswrapper[4971]: I1213 09:45:32.638848 4971 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_swift-operator-controller-manager-5c6df8f9-rt5n5_bd48845c-5d55-4690-960b-ca319f596beb/manager/0.log" Dec 13 09:45:32 crc kubenswrapper[4971]: I1213 09:45:32.861223 4971 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_openstack-operator-controller-manager-56f6fbdf6-vzrkl_40f2727e-ef9d-4557-8619-b37174d00b67/manager/0.log" Dec 13 09:45:33 crc kubenswrapper[4971]: I1213 09:45:33.147943 4971 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_telemetry-operator-controller-manager-97d456b9-j4fpf_f111629c-eb0b-4294-8f85-d4b0b6500454/manager/0.log" Dec 13 09:45:33 crc kubenswrapper[4971]: I1213 09:45:33.548027 4971 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_test-operator-controller-manager-85c4449d86-s4v55_ccfc4136-eeb7-4fd8-a4d2-d09dd4a55802/manager/0.log" Dec 13 09:45:33 crc kubenswrapper[4971]: I1213 09:45:33.607100 4971 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_watcher-operator-controller-manager-55f78b7c4c-dwmgr_5bb0c1cd-9a8e-4c03-80c0-d9e66dae91ed/manager/0.log" Dec 13 09:45:36 crc kubenswrapper[4971]: E1213 09:45:36.763332 4971 cadvisor_stats_provider.go:516] "Partial failure issuing cadvisor.ContainerInfoV2" err="partial failures: [\"/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod67b2dfa6_27b0_4dc4_bd58_0ac26186530f.slice/crio-0bd0c20bc268f5b919cf45041265adde7829dc9c311a0482ad1b66216003b2a3.scope\": RecentStats: unable to find data in memory cache]" Dec 13 09:45:46 crc kubenswrapper[4971]: I1213 09:45:46.154250 4971 patch_prober.go:28] interesting pod/machine-config-daemon-82xjz container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 13 09:45:46 crc kubenswrapper[4971]: I1213 09:45:46.155045 4971 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-82xjz" podUID="e0c4af12-edda-41d7-9b44-f87396174ef2" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 13 09:45:47 crc kubenswrapper[4971]: E1213 09:45:47.065942 4971 cadvisor_stats_provider.go:516] "Partial failure issuing cadvisor.ContainerInfoV2" err="partial failures: [\"/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod67b2dfa6_27b0_4dc4_bd58_0ac26186530f.slice/crio-0bd0c20bc268f5b919cf45041265adde7829dc9c311a0482ad1b66216003b2a3.scope\": RecentStats: unable to find data in memory cache]" Dec 13 09:45:52 crc kubenswrapper[4971]: I1213 09:45:52.355380 4971 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-marketplace-qg9vl"] Dec 13 09:45:52 crc kubenswrapper[4971]: I1213 09:45:52.362207 4971 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-qg9vl" Dec 13 09:45:52 crc kubenswrapper[4971]: I1213 09:45:52.422275 4971 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-qg9vl"] Dec 13 09:45:52 crc kubenswrapper[4971]: I1213 09:45:52.498080 4971 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/e7937c8f-d8c3-470b-9b53-a5deae341280-catalog-content\") pod \"redhat-marketplace-qg9vl\" (UID: \"e7937c8f-d8c3-470b-9b53-a5deae341280\") " pod="openshift-marketplace/redhat-marketplace-qg9vl" Dec 13 09:45:52 crc kubenswrapper[4971]: I1213 09:45:52.498845 4971 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/e7937c8f-d8c3-470b-9b53-a5deae341280-utilities\") pod \"redhat-marketplace-qg9vl\" (UID: \"e7937c8f-d8c3-470b-9b53-a5deae341280\") " pod="openshift-marketplace/redhat-marketplace-qg9vl" Dec 13 09:45:52 crc kubenswrapper[4971]: I1213 09:45:52.499084 4971 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-x4qbf\" (UniqueName: \"kubernetes.io/projected/e7937c8f-d8c3-470b-9b53-a5deae341280-kube-api-access-x4qbf\") pod \"redhat-marketplace-qg9vl\" (UID: \"e7937c8f-d8c3-470b-9b53-a5deae341280\") " pod="openshift-marketplace/redhat-marketplace-qg9vl" Dec 13 09:45:52 crc kubenswrapper[4971]: I1213 09:45:52.601879 4971 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/e7937c8f-d8c3-470b-9b53-a5deae341280-catalog-content\") pod \"redhat-marketplace-qg9vl\" (UID: \"e7937c8f-d8c3-470b-9b53-a5deae341280\") " pod="openshift-marketplace/redhat-marketplace-qg9vl" Dec 13 09:45:52 crc kubenswrapper[4971]: I1213 09:45:52.602247 4971 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/e7937c8f-d8c3-470b-9b53-a5deae341280-utilities\") pod \"redhat-marketplace-qg9vl\" (UID: \"e7937c8f-d8c3-470b-9b53-a5deae341280\") " pod="openshift-marketplace/redhat-marketplace-qg9vl" Dec 13 09:45:52 crc kubenswrapper[4971]: I1213 09:45:52.602380 4971 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-x4qbf\" (UniqueName: \"kubernetes.io/projected/e7937c8f-d8c3-470b-9b53-a5deae341280-kube-api-access-x4qbf\") pod \"redhat-marketplace-qg9vl\" (UID: \"e7937c8f-d8c3-470b-9b53-a5deae341280\") " pod="openshift-marketplace/redhat-marketplace-qg9vl" Dec 13 09:45:52 crc kubenswrapper[4971]: I1213 09:45:52.602466 4971 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/e7937c8f-d8c3-470b-9b53-a5deae341280-catalog-content\") pod \"redhat-marketplace-qg9vl\" (UID: \"e7937c8f-d8c3-470b-9b53-a5deae341280\") " pod="openshift-marketplace/redhat-marketplace-qg9vl" Dec 13 09:45:52 crc kubenswrapper[4971]: I1213 09:45:52.602773 4971 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/e7937c8f-d8c3-470b-9b53-a5deae341280-utilities\") pod \"redhat-marketplace-qg9vl\" (UID: \"e7937c8f-d8c3-470b-9b53-a5deae341280\") " pod="openshift-marketplace/redhat-marketplace-qg9vl" Dec 13 09:45:52 crc kubenswrapper[4971]: I1213 09:45:52.621635 4971 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-x4qbf\" (UniqueName: \"kubernetes.io/projected/e7937c8f-d8c3-470b-9b53-a5deae341280-kube-api-access-x4qbf\") pod \"redhat-marketplace-qg9vl\" (UID: \"e7937c8f-d8c3-470b-9b53-a5deae341280\") " pod="openshift-marketplace/redhat-marketplace-qg9vl" Dec 13 09:45:52 crc kubenswrapper[4971]: I1213 09:45:52.712764 4971 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-qg9vl" Dec 13 09:45:53 crc kubenswrapper[4971]: I1213 09:45:53.254389 4971 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-qg9vl"] Dec 13 09:45:53 crc kubenswrapper[4971]: I1213 09:45:53.927066 4971 generic.go:334] "Generic (PLEG): container finished" podID="e7937c8f-d8c3-470b-9b53-a5deae341280" containerID="1031ff4c215278920476a9dce2f5d13239624d46dafb305b72d8e30a896edc11" exitCode=0 Dec 13 09:45:53 crc kubenswrapper[4971]: I1213 09:45:53.927300 4971 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-qg9vl" event={"ID":"e7937c8f-d8c3-470b-9b53-a5deae341280","Type":"ContainerDied","Data":"1031ff4c215278920476a9dce2f5d13239624d46dafb305b72d8e30a896edc11"} Dec 13 09:45:53 crc kubenswrapper[4971]: I1213 09:45:53.927598 4971 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-qg9vl" event={"ID":"e7937c8f-d8c3-470b-9b53-a5deae341280","Type":"ContainerStarted","Data":"98ecb58848d3d50e295d70f84ce5c98edc72af5f196c9e21eac3ea24c4ac9170"} Dec 13 09:45:53 crc kubenswrapper[4971]: I1213 09:45:53.930185 4971 provider.go:102] Refreshing cache for provider: *credentialprovider.defaultDockerConfigProvider Dec 13 09:45:55 crc kubenswrapper[4971]: I1213 09:45:55.968867 4971 generic.go:334] "Generic (PLEG): container finished" podID="e7937c8f-d8c3-470b-9b53-a5deae341280" containerID="adb834e662e29552ae0523f0798c2caaf353310eda5f6592c1aafaa9d75936df" exitCode=0 Dec 13 09:45:55 crc kubenswrapper[4971]: I1213 09:45:55.969596 4971 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-qg9vl" event={"ID":"e7937c8f-d8c3-470b-9b53-a5deae341280","Type":"ContainerDied","Data":"adb834e662e29552ae0523f0798c2caaf353310eda5f6592c1aafaa9d75936df"} Dec 13 09:45:56 crc kubenswrapper[4971]: I1213 09:45:56.989357 4971 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-qg9vl" event={"ID":"e7937c8f-d8c3-470b-9b53-a5deae341280","Type":"ContainerStarted","Data":"449d876cf2c1e6db3fb106c4ad82726bc5c5853cd744124d3843a5b1ebd13681"} Dec 13 09:45:57 crc kubenswrapper[4971]: I1213 09:45:57.026602 4971 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-marketplace-qg9vl" podStartSLOduration=2.456399488 podStartE2EDuration="5.026570962s" podCreationTimestamp="2025-12-13 09:45:52 +0000 UTC" firstStartedPulling="2025-12-13 09:45:53.929853015 +0000 UTC m=+10610.534262463" lastFinishedPulling="2025-12-13 09:45:56.500024449 +0000 UTC m=+10613.104433937" observedRunningTime="2025-12-13 09:45:57.01500974 +0000 UTC m=+10613.619419188" watchObservedRunningTime="2025-12-13 09:45:57.026570962 +0000 UTC m=+10613.630980410" Dec 13 09:45:57 crc kubenswrapper[4971]: E1213 09:45:57.389205 4971 cadvisor_stats_provider.go:516] "Partial failure issuing cadvisor.ContainerInfoV2" err="partial failures: [\"/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod67b2dfa6_27b0_4dc4_bd58_0ac26186530f.slice/crio-0bd0c20bc268f5b919cf45041265adde7829dc9c311a0482ad1b66216003b2a3.scope\": RecentStats: unable to find data in memory cache]" Dec 13 09:46:02 crc kubenswrapper[4971]: I1213 09:46:02.841621 4971 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-marketplace-qg9vl" Dec 13 09:46:02 crc kubenswrapper[4971]: I1213 09:46:02.842355 4971 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-marketplace-qg9vl" Dec 13 09:46:02 crc kubenswrapper[4971]: I1213 09:46:02.919958 4971 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-marketplace-qg9vl" Dec 13 09:46:03 crc kubenswrapper[4971]: I1213 09:46:03.012733 4971 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-marketplace-qg9vl" Dec 13 09:46:03 crc kubenswrapper[4971]: I1213 09:46:03.172685 4971 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-qg9vl"] Dec 13 09:46:04 crc kubenswrapper[4971]: I1213 09:46:04.926651 4971 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-marketplace-qg9vl" podUID="e7937c8f-d8c3-470b-9b53-a5deae341280" containerName="registry-server" containerID="cri-o://449d876cf2c1e6db3fb106c4ad82726bc5c5853cd744124d3843a5b1ebd13681" gracePeriod=2 Dec 13 09:46:05 crc kubenswrapper[4971]: I1213 09:46:05.481960 4971 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-qg9vl" Dec 13 09:46:05 crc kubenswrapper[4971]: I1213 09:46:05.615135 4971 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-x4qbf\" (UniqueName: \"kubernetes.io/projected/e7937c8f-d8c3-470b-9b53-a5deae341280-kube-api-access-x4qbf\") pod \"e7937c8f-d8c3-470b-9b53-a5deae341280\" (UID: \"e7937c8f-d8c3-470b-9b53-a5deae341280\") " Dec 13 09:46:05 crc kubenswrapper[4971]: I1213 09:46:05.615214 4971 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/e7937c8f-d8c3-470b-9b53-a5deae341280-utilities\") pod \"e7937c8f-d8c3-470b-9b53-a5deae341280\" (UID: \"e7937c8f-d8c3-470b-9b53-a5deae341280\") " Dec 13 09:46:05 crc kubenswrapper[4971]: I1213 09:46:05.615281 4971 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/e7937c8f-d8c3-470b-9b53-a5deae341280-catalog-content\") pod \"e7937c8f-d8c3-470b-9b53-a5deae341280\" (UID: \"e7937c8f-d8c3-470b-9b53-a5deae341280\") " Dec 13 09:46:05 crc kubenswrapper[4971]: I1213 09:46:05.617084 4971 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/e7937c8f-d8c3-470b-9b53-a5deae341280-utilities" (OuterVolumeSpecName: "utilities") pod "e7937c8f-d8c3-470b-9b53-a5deae341280" (UID: "e7937c8f-d8c3-470b-9b53-a5deae341280"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 13 09:46:05 crc kubenswrapper[4971]: I1213 09:46:05.623789 4971 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/e7937c8f-d8c3-470b-9b53-a5deae341280-kube-api-access-x4qbf" (OuterVolumeSpecName: "kube-api-access-x4qbf") pod "e7937c8f-d8c3-470b-9b53-a5deae341280" (UID: "e7937c8f-d8c3-470b-9b53-a5deae341280"). InnerVolumeSpecName "kube-api-access-x4qbf". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 13 09:46:05 crc kubenswrapper[4971]: I1213 09:46:05.646610 4971 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/e7937c8f-d8c3-470b-9b53-a5deae341280-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "e7937c8f-d8c3-470b-9b53-a5deae341280" (UID: "e7937c8f-d8c3-470b-9b53-a5deae341280"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 13 09:46:05 crc kubenswrapper[4971]: I1213 09:46:05.719232 4971 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-x4qbf\" (UniqueName: \"kubernetes.io/projected/e7937c8f-d8c3-470b-9b53-a5deae341280-kube-api-access-x4qbf\") on node \"crc\" DevicePath \"\"" Dec 13 09:46:05 crc kubenswrapper[4971]: I1213 09:46:05.719294 4971 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/e7937c8f-d8c3-470b-9b53-a5deae341280-utilities\") on node \"crc\" DevicePath \"\"" Dec 13 09:46:05 crc kubenswrapper[4971]: I1213 09:46:05.719304 4971 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/e7937c8f-d8c3-470b-9b53-a5deae341280-catalog-content\") on node \"crc\" DevicePath \"\"" Dec 13 09:46:05 crc kubenswrapper[4971]: I1213 09:46:05.945406 4971 generic.go:334] "Generic (PLEG): container finished" podID="e7937c8f-d8c3-470b-9b53-a5deae341280" containerID="449d876cf2c1e6db3fb106c4ad82726bc5c5853cd744124d3843a5b1ebd13681" exitCode=0 Dec 13 09:46:05 crc kubenswrapper[4971]: I1213 09:46:05.945479 4971 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-qg9vl" event={"ID":"e7937c8f-d8c3-470b-9b53-a5deae341280","Type":"ContainerDied","Data":"449d876cf2c1e6db3fb106c4ad82726bc5c5853cd744124d3843a5b1ebd13681"} Dec 13 09:46:05 crc kubenswrapper[4971]: I1213 09:46:05.945539 4971 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-qg9vl" event={"ID":"e7937c8f-d8c3-470b-9b53-a5deae341280","Type":"ContainerDied","Data":"98ecb58848d3d50e295d70f84ce5c98edc72af5f196c9e21eac3ea24c4ac9170"} Dec 13 09:46:05 crc kubenswrapper[4971]: I1213 09:46:05.945570 4971 scope.go:117] "RemoveContainer" containerID="449d876cf2c1e6db3fb106c4ad82726bc5c5853cd744124d3843a5b1ebd13681" Dec 13 09:46:05 crc kubenswrapper[4971]: I1213 09:46:05.945795 4971 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-qg9vl" Dec 13 09:46:06 crc kubenswrapper[4971]: I1213 09:46:06.011602 4971 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-qg9vl"] Dec 13 09:46:06 crc kubenswrapper[4971]: I1213 09:46:06.011731 4971 scope.go:117] "RemoveContainer" containerID="adb834e662e29552ae0523f0798c2caaf353310eda5f6592c1aafaa9d75936df" Dec 13 09:46:06 crc kubenswrapper[4971]: I1213 09:46:06.033724 4971 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-marketplace-qg9vl"] Dec 13 09:46:06 crc kubenswrapper[4971]: I1213 09:46:06.073027 4971 scope.go:117] "RemoveContainer" containerID="1031ff4c215278920476a9dce2f5d13239624d46dafb305b72d8e30a896edc11" Dec 13 09:46:06 crc kubenswrapper[4971]: I1213 09:46:06.442472 4971 scope.go:117] "RemoveContainer" containerID="449d876cf2c1e6db3fb106c4ad82726bc5c5853cd744124d3843a5b1ebd13681" Dec 13 09:46:06 crc kubenswrapper[4971]: E1213 09:46:06.443784 4971 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"449d876cf2c1e6db3fb106c4ad82726bc5c5853cd744124d3843a5b1ebd13681\": container with ID starting with 449d876cf2c1e6db3fb106c4ad82726bc5c5853cd744124d3843a5b1ebd13681 not found: ID does not exist" containerID="449d876cf2c1e6db3fb106c4ad82726bc5c5853cd744124d3843a5b1ebd13681" Dec 13 09:46:06 crc kubenswrapper[4971]: I1213 09:46:06.443870 4971 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"449d876cf2c1e6db3fb106c4ad82726bc5c5853cd744124d3843a5b1ebd13681"} err="failed to get container status \"449d876cf2c1e6db3fb106c4ad82726bc5c5853cd744124d3843a5b1ebd13681\": rpc error: code = NotFound desc = could not find container \"449d876cf2c1e6db3fb106c4ad82726bc5c5853cd744124d3843a5b1ebd13681\": container with ID starting with 449d876cf2c1e6db3fb106c4ad82726bc5c5853cd744124d3843a5b1ebd13681 not found: ID does not exist" Dec 13 09:46:06 crc kubenswrapper[4971]: I1213 09:46:06.443917 4971 scope.go:117] "RemoveContainer" containerID="adb834e662e29552ae0523f0798c2caaf353310eda5f6592c1aafaa9d75936df" Dec 13 09:46:06 crc kubenswrapper[4971]: E1213 09:46:06.444465 4971 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"adb834e662e29552ae0523f0798c2caaf353310eda5f6592c1aafaa9d75936df\": container with ID starting with adb834e662e29552ae0523f0798c2caaf353310eda5f6592c1aafaa9d75936df not found: ID does not exist" containerID="adb834e662e29552ae0523f0798c2caaf353310eda5f6592c1aafaa9d75936df" Dec 13 09:46:06 crc kubenswrapper[4971]: I1213 09:46:06.444685 4971 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"adb834e662e29552ae0523f0798c2caaf353310eda5f6592c1aafaa9d75936df"} err="failed to get container status \"adb834e662e29552ae0523f0798c2caaf353310eda5f6592c1aafaa9d75936df\": rpc error: code = NotFound desc = could not find container \"adb834e662e29552ae0523f0798c2caaf353310eda5f6592c1aafaa9d75936df\": container with ID starting with adb834e662e29552ae0523f0798c2caaf353310eda5f6592c1aafaa9d75936df not found: ID does not exist" Dec 13 09:46:06 crc kubenswrapper[4971]: I1213 09:46:06.444731 4971 scope.go:117] "RemoveContainer" containerID="1031ff4c215278920476a9dce2f5d13239624d46dafb305b72d8e30a896edc11" Dec 13 09:46:06 crc kubenswrapper[4971]: E1213 09:46:06.447152 4971 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"1031ff4c215278920476a9dce2f5d13239624d46dafb305b72d8e30a896edc11\": container with ID starting with 1031ff4c215278920476a9dce2f5d13239624d46dafb305b72d8e30a896edc11 not found: ID does not exist" containerID="1031ff4c215278920476a9dce2f5d13239624d46dafb305b72d8e30a896edc11" Dec 13 09:46:06 crc kubenswrapper[4971]: I1213 09:46:06.447206 4971 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"1031ff4c215278920476a9dce2f5d13239624d46dafb305b72d8e30a896edc11"} err="failed to get container status \"1031ff4c215278920476a9dce2f5d13239624d46dafb305b72d8e30a896edc11\": rpc error: code = NotFound desc = could not find container \"1031ff4c215278920476a9dce2f5d13239624d46dafb305b72d8e30a896edc11\": container with ID starting with 1031ff4c215278920476a9dce2f5d13239624d46dafb305b72d8e30a896edc11 not found: ID does not exist" Dec 13 09:46:07 crc kubenswrapper[4971]: I1213 09:46:07.783713 4971 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="e7937c8f-d8c3-470b-9b53-a5deae341280" path="/var/lib/kubelet/pods/e7937c8f-d8c3-470b-9b53-a5deae341280/volumes" Dec 13 09:46:09 crc kubenswrapper[4971]: I1213 09:46:09.937441 4971 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-machine-api_control-plane-machine-set-operator-78cbb6b69f-bgdbk_930fe38c-5ff7-45a8-a292-cd69d07e86f7/control-plane-machine-set-operator/0.log" Dec 13 09:46:10 crc kubenswrapper[4971]: I1213 09:46:10.282284 4971 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-machine-api_machine-api-operator-5694c8668f-t24dm_c45a314c-77da-43fc-b2ae-9de350540c2b/kube-rbac-proxy/0.log" Dec 13 09:46:10 crc kubenswrapper[4971]: I1213 09:46:10.410416 4971 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-machine-api_machine-api-operator-5694c8668f-t24dm_c45a314c-77da-43fc-b2ae-9de350540c2b/machine-api-operator/0.log" Dec 13 09:46:16 crc kubenswrapper[4971]: I1213 09:46:16.153663 4971 patch_prober.go:28] interesting pod/machine-config-daemon-82xjz container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 13 09:46:16 crc kubenswrapper[4971]: I1213 09:46:16.154645 4971 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-82xjz" podUID="e0c4af12-edda-41d7-9b44-f87396174ef2" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 13 09:46:16 crc kubenswrapper[4971]: I1213 09:46:16.154716 4971 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-82xjz" Dec 13 09:46:16 crc kubenswrapper[4971]: I1213 09:46:16.155992 4971 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"80deb18ddf85834e136ab62ea93a1fa75315016d9f220eb217720a5e562f5e54"} pod="openshift-machine-config-operator/machine-config-daemon-82xjz" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Dec 13 09:46:16 crc kubenswrapper[4971]: I1213 09:46:16.156071 4971 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-82xjz" podUID="e0c4af12-edda-41d7-9b44-f87396174ef2" containerName="machine-config-daemon" containerID="cri-o://80deb18ddf85834e136ab62ea93a1fa75315016d9f220eb217720a5e562f5e54" gracePeriod=600 Dec 13 09:46:16 crc kubenswrapper[4971]: E1213 09:46:16.429442 4971 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-82xjz_openshift-machine-config-operator(e0c4af12-edda-41d7-9b44-f87396174ef2)\"" pod="openshift-machine-config-operator/machine-config-daemon-82xjz" podUID="e0c4af12-edda-41d7-9b44-f87396174ef2" Dec 13 09:46:17 crc kubenswrapper[4971]: I1213 09:46:17.421278 4971 generic.go:334] "Generic (PLEG): container finished" podID="e0c4af12-edda-41d7-9b44-f87396174ef2" containerID="80deb18ddf85834e136ab62ea93a1fa75315016d9f220eb217720a5e562f5e54" exitCode=0 Dec 13 09:46:17 crc kubenswrapper[4971]: I1213 09:46:17.421811 4971 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-82xjz" event={"ID":"e0c4af12-edda-41d7-9b44-f87396174ef2","Type":"ContainerDied","Data":"80deb18ddf85834e136ab62ea93a1fa75315016d9f220eb217720a5e562f5e54"} Dec 13 09:46:17 crc kubenswrapper[4971]: I1213 09:46:17.421860 4971 scope.go:117] "RemoveContainer" containerID="b2e549456db49c5609edf5288a8e1e209f21e8f767afaa3d5510b083ba33ae70" Dec 13 09:46:17 crc kubenswrapper[4971]: I1213 09:46:17.423014 4971 scope.go:117] "RemoveContainer" containerID="80deb18ddf85834e136ab62ea93a1fa75315016d9f220eb217720a5e562f5e54" Dec 13 09:46:17 crc kubenswrapper[4971]: E1213 09:46:17.423460 4971 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-82xjz_openshift-machine-config-operator(e0c4af12-edda-41d7-9b44-f87396174ef2)\"" pod="openshift-machine-config-operator/machine-config-daemon-82xjz" podUID="e0c4af12-edda-41d7-9b44-f87396174ef2" Dec 13 09:46:30 crc kubenswrapper[4971]: I1213 09:46:30.770025 4971 scope.go:117] "RemoveContainer" containerID="80deb18ddf85834e136ab62ea93a1fa75315016d9f220eb217720a5e562f5e54" Dec 13 09:46:30 crc kubenswrapper[4971]: E1213 09:46:30.771233 4971 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-82xjz_openshift-machine-config-operator(e0c4af12-edda-41d7-9b44-f87396174ef2)\"" pod="openshift-machine-config-operator/machine-config-daemon-82xjz" podUID="e0c4af12-edda-41d7-9b44-f87396174ef2" Dec 13 09:46:38 crc kubenswrapper[4971]: I1213 09:46:38.626761 4971 trace.go:236] Trace[220009666]: "Calculate volume metrics of registry-storage for pod openshift-image-registry/image-registry-66df7c8f76-xjwwz" (13-Dec-2025 09:46:30.576) (total time: 8047ms): Dec 13 09:46:38 crc kubenswrapper[4971]: Trace[220009666]: [8.047482471s] [8.047482471s] END Dec 13 09:46:38 crc kubenswrapper[4971]: I1213 09:46:38.769007 4971 generic.go:334] "Generic (PLEG): container finished" podID="2c7c5219-d904-4610-805a-6324810e3760" containerID="92077d04a5c5f8ddbc19fb5ec1cda65d959a3d4835d1b05c4668b8853e148eb1" exitCode=0 Dec 13 09:46:38 crc kubenswrapper[4971]: I1213 09:46:38.770157 4971 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-g74xz/crc-debug-65smr" event={"ID":"2c7c5219-d904-4610-805a-6324810e3760","Type":"ContainerDied","Data":"92077d04a5c5f8ddbc19fb5ec1cda65d959a3d4835d1b05c4668b8853e148eb1"} Dec 13 09:46:39 crc kubenswrapper[4971]: E1213 09:46:39.769904 4971 kubelet_pods.go:538] "Hostname for pod was too long, truncated it" podName="test-operator-logs-pod-horizontest-horizontest-tests-horizontest" hostnameMaxLen=63 truncatedHostname="test-operator-logs-pod-horizontest-horizontest-tests-horizontes" Dec 13 09:46:39 crc kubenswrapper[4971]: I1213 09:46:39.925984 4971 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-g74xz/crc-debug-65smr" Dec 13 09:46:39 crc kubenswrapper[4971]: I1213 09:46:39.977740 4971 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-must-gather-g74xz/crc-debug-65smr"] Dec 13 09:46:39 crc kubenswrapper[4971]: I1213 09:46:39.991012 4971 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-must-gather-g74xz/crc-debug-65smr"] Dec 13 09:46:40 crc kubenswrapper[4971]: I1213 09:46:40.095252 4971 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"host\" (UniqueName: \"kubernetes.io/host-path/2c7c5219-d904-4610-805a-6324810e3760-host\") pod \"2c7c5219-d904-4610-805a-6324810e3760\" (UID: \"2c7c5219-d904-4610-805a-6324810e3760\") " Dec 13 09:46:40 crc kubenswrapper[4971]: I1213 09:46:40.095400 4971 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/2c7c5219-d904-4610-805a-6324810e3760-host" (OuterVolumeSpecName: "host") pod "2c7c5219-d904-4610-805a-6324810e3760" (UID: "2c7c5219-d904-4610-805a-6324810e3760"). InnerVolumeSpecName "host". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 13 09:46:40 crc kubenswrapper[4971]: I1213 09:46:40.095606 4971 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-kfs55\" (UniqueName: \"kubernetes.io/projected/2c7c5219-d904-4610-805a-6324810e3760-kube-api-access-kfs55\") pod \"2c7c5219-d904-4610-805a-6324810e3760\" (UID: \"2c7c5219-d904-4610-805a-6324810e3760\") " Dec 13 09:46:40 crc kubenswrapper[4971]: I1213 09:46:40.096337 4971 reconciler_common.go:293] "Volume detached for volume \"host\" (UniqueName: \"kubernetes.io/host-path/2c7c5219-d904-4610-805a-6324810e3760-host\") on node \"crc\" DevicePath \"\"" Dec 13 09:46:40 crc kubenswrapper[4971]: I1213 09:46:40.107184 4971 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/2c7c5219-d904-4610-805a-6324810e3760-kube-api-access-kfs55" (OuterVolumeSpecName: "kube-api-access-kfs55") pod "2c7c5219-d904-4610-805a-6324810e3760" (UID: "2c7c5219-d904-4610-805a-6324810e3760"). InnerVolumeSpecName "kube-api-access-kfs55". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 13 09:46:40 crc kubenswrapper[4971]: I1213 09:46:40.198622 4971 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-kfs55\" (UniqueName: \"kubernetes.io/projected/2c7c5219-d904-4610-805a-6324810e3760-kube-api-access-kfs55\") on node \"crc\" DevicePath \"\"" Dec 13 09:46:40 crc kubenswrapper[4971]: I1213 09:46:40.795751 4971 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="82dfc97e6d7f85f46bc66f2ac62d963663bf8b0af56c97b98997c24b064fe9cf" Dec 13 09:46:40 crc kubenswrapper[4971]: I1213 09:46:40.796587 4971 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-g74xz/crc-debug-65smr" Dec 13 09:46:41 crc kubenswrapper[4971]: I1213 09:46:41.748967 4971 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-must-gather-g74xz/crc-debug-jzcqj"] Dec 13 09:46:41 crc kubenswrapper[4971]: E1213 09:46:41.749676 4971 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="e7937c8f-d8c3-470b-9b53-a5deae341280" containerName="extract-utilities" Dec 13 09:46:41 crc kubenswrapper[4971]: I1213 09:46:41.749707 4971 state_mem.go:107] "Deleted CPUSet assignment" podUID="e7937c8f-d8c3-470b-9b53-a5deae341280" containerName="extract-utilities" Dec 13 09:46:41 crc kubenswrapper[4971]: E1213 09:46:41.749737 4971 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="e7937c8f-d8c3-470b-9b53-a5deae341280" containerName="extract-content" Dec 13 09:46:41 crc kubenswrapper[4971]: I1213 09:46:41.749748 4971 state_mem.go:107] "Deleted CPUSet assignment" podUID="e7937c8f-d8c3-470b-9b53-a5deae341280" containerName="extract-content" Dec 13 09:46:41 crc kubenswrapper[4971]: E1213 09:46:41.749798 4971 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="2c7c5219-d904-4610-805a-6324810e3760" containerName="container-00" Dec 13 09:46:41 crc kubenswrapper[4971]: I1213 09:46:41.749808 4971 state_mem.go:107] "Deleted CPUSet assignment" podUID="2c7c5219-d904-4610-805a-6324810e3760" containerName="container-00" Dec 13 09:46:41 crc kubenswrapper[4971]: E1213 09:46:41.749824 4971 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="e7937c8f-d8c3-470b-9b53-a5deae341280" containerName="registry-server" Dec 13 09:46:41 crc kubenswrapper[4971]: I1213 09:46:41.749833 4971 state_mem.go:107] "Deleted CPUSet assignment" podUID="e7937c8f-d8c3-470b-9b53-a5deae341280" containerName="registry-server" Dec 13 09:46:41 crc kubenswrapper[4971]: I1213 09:46:41.750112 4971 memory_manager.go:354] "RemoveStaleState removing state" podUID="e7937c8f-d8c3-470b-9b53-a5deae341280" containerName="registry-server" Dec 13 09:46:41 crc kubenswrapper[4971]: I1213 09:46:41.750167 4971 memory_manager.go:354] "RemoveStaleState removing state" podUID="2c7c5219-d904-4610-805a-6324810e3760" containerName="container-00" Dec 13 09:46:41 crc kubenswrapper[4971]: I1213 09:46:41.751325 4971 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-g74xz/crc-debug-jzcqj" Dec 13 09:46:41 crc kubenswrapper[4971]: I1213 09:46:41.753967 4971 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-must-gather-g74xz"/"default-dockercfg-2tvhw" Dec 13 09:46:41 crc kubenswrapper[4971]: I1213 09:46:41.783405 4971 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="2c7c5219-d904-4610-805a-6324810e3760" path="/var/lib/kubelet/pods/2c7c5219-d904-4610-805a-6324810e3760/volumes" Dec 13 09:46:41 crc kubenswrapper[4971]: I1213 09:46:41.846857 4971 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host\" (UniqueName: \"kubernetes.io/host-path/553dd05f-d4b1-4aa9-afc9-1ed89824955d-host\") pod \"crc-debug-jzcqj\" (UID: \"553dd05f-d4b1-4aa9-afc9-1ed89824955d\") " pod="openshift-must-gather-g74xz/crc-debug-jzcqj" Dec 13 09:46:41 crc kubenswrapper[4971]: I1213 09:46:41.846979 4971 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-6dndn\" (UniqueName: \"kubernetes.io/projected/553dd05f-d4b1-4aa9-afc9-1ed89824955d-kube-api-access-6dndn\") pod \"crc-debug-jzcqj\" (UID: \"553dd05f-d4b1-4aa9-afc9-1ed89824955d\") " pod="openshift-must-gather-g74xz/crc-debug-jzcqj" Dec 13 09:46:41 crc kubenswrapper[4971]: I1213 09:46:41.949935 4971 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host\" (UniqueName: \"kubernetes.io/host-path/553dd05f-d4b1-4aa9-afc9-1ed89824955d-host\") pod \"crc-debug-jzcqj\" (UID: \"553dd05f-d4b1-4aa9-afc9-1ed89824955d\") " pod="openshift-must-gather-g74xz/crc-debug-jzcqj" Dec 13 09:46:41 crc kubenswrapper[4971]: I1213 09:46:41.950027 4971 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-6dndn\" (UniqueName: \"kubernetes.io/projected/553dd05f-d4b1-4aa9-afc9-1ed89824955d-kube-api-access-6dndn\") pod \"crc-debug-jzcqj\" (UID: \"553dd05f-d4b1-4aa9-afc9-1ed89824955d\") " pod="openshift-must-gather-g74xz/crc-debug-jzcqj" Dec 13 09:46:41 crc kubenswrapper[4971]: I1213 09:46:41.950938 4971 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host\" (UniqueName: \"kubernetes.io/host-path/553dd05f-d4b1-4aa9-afc9-1ed89824955d-host\") pod \"crc-debug-jzcqj\" (UID: \"553dd05f-d4b1-4aa9-afc9-1ed89824955d\") " pod="openshift-must-gather-g74xz/crc-debug-jzcqj" Dec 13 09:46:41 crc kubenswrapper[4971]: I1213 09:46:41.977249 4971 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-6dndn\" (UniqueName: \"kubernetes.io/projected/553dd05f-d4b1-4aa9-afc9-1ed89824955d-kube-api-access-6dndn\") pod \"crc-debug-jzcqj\" (UID: \"553dd05f-d4b1-4aa9-afc9-1ed89824955d\") " pod="openshift-must-gather-g74xz/crc-debug-jzcqj" Dec 13 09:46:42 crc kubenswrapper[4971]: I1213 09:46:42.074195 4971 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-g74xz/crc-debug-jzcqj" Dec 13 09:46:42 crc kubenswrapper[4971]: I1213 09:46:42.819120 4971 generic.go:334] "Generic (PLEG): container finished" podID="553dd05f-d4b1-4aa9-afc9-1ed89824955d" containerID="474aee8c65e63d738d23dbf5b98c68b47526a7f328a39f7f07485285afc74dc4" exitCode=0 Dec 13 09:46:42 crc kubenswrapper[4971]: I1213 09:46:42.819261 4971 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-g74xz/crc-debug-jzcqj" event={"ID":"553dd05f-d4b1-4aa9-afc9-1ed89824955d","Type":"ContainerDied","Data":"474aee8c65e63d738d23dbf5b98c68b47526a7f328a39f7f07485285afc74dc4"} Dec 13 09:46:42 crc kubenswrapper[4971]: I1213 09:46:42.819505 4971 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-g74xz/crc-debug-jzcqj" event={"ID":"553dd05f-d4b1-4aa9-afc9-1ed89824955d","Type":"ContainerStarted","Data":"7fa1590164c3ad0ac6fe3340523ceb2818d18012ec13dea17c013e277da3d024"} Dec 13 09:46:43 crc kubenswrapper[4971]: I1213 09:46:43.972011 4971 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-g74xz/crc-debug-jzcqj" Dec 13 09:46:44 crc kubenswrapper[4971]: I1213 09:46:44.114285 4971 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"host\" (UniqueName: \"kubernetes.io/host-path/553dd05f-d4b1-4aa9-afc9-1ed89824955d-host\") pod \"553dd05f-d4b1-4aa9-afc9-1ed89824955d\" (UID: \"553dd05f-d4b1-4aa9-afc9-1ed89824955d\") " Dec 13 09:46:44 crc kubenswrapper[4971]: I1213 09:46:44.114536 4971 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-6dndn\" (UniqueName: \"kubernetes.io/projected/553dd05f-d4b1-4aa9-afc9-1ed89824955d-kube-api-access-6dndn\") pod \"553dd05f-d4b1-4aa9-afc9-1ed89824955d\" (UID: \"553dd05f-d4b1-4aa9-afc9-1ed89824955d\") " Dec 13 09:46:44 crc kubenswrapper[4971]: I1213 09:46:44.114718 4971 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/553dd05f-d4b1-4aa9-afc9-1ed89824955d-host" (OuterVolumeSpecName: "host") pod "553dd05f-d4b1-4aa9-afc9-1ed89824955d" (UID: "553dd05f-d4b1-4aa9-afc9-1ed89824955d"). InnerVolumeSpecName "host". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 13 09:46:44 crc kubenswrapper[4971]: I1213 09:46:44.115223 4971 reconciler_common.go:293] "Volume detached for volume \"host\" (UniqueName: \"kubernetes.io/host-path/553dd05f-d4b1-4aa9-afc9-1ed89824955d-host\") on node \"crc\" DevicePath \"\"" Dec 13 09:46:44 crc kubenswrapper[4971]: I1213 09:46:44.127786 4971 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/553dd05f-d4b1-4aa9-afc9-1ed89824955d-kube-api-access-6dndn" (OuterVolumeSpecName: "kube-api-access-6dndn") pod "553dd05f-d4b1-4aa9-afc9-1ed89824955d" (UID: "553dd05f-d4b1-4aa9-afc9-1ed89824955d"). InnerVolumeSpecName "kube-api-access-6dndn". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 13 09:46:44 crc kubenswrapper[4971]: I1213 09:46:44.216949 4971 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-6dndn\" (UniqueName: \"kubernetes.io/projected/553dd05f-d4b1-4aa9-afc9-1ed89824955d-kube-api-access-6dndn\") on node \"crc\" DevicePath \"\"" Dec 13 09:46:44 crc kubenswrapper[4971]: I1213 09:46:44.982559 4971 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-g74xz/crc-debug-jzcqj" event={"ID":"553dd05f-d4b1-4aa9-afc9-1ed89824955d","Type":"ContainerDied","Data":"7fa1590164c3ad0ac6fe3340523ceb2818d18012ec13dea17c013e277da3d024"} Dec 13 09:46:44 crc kubenswrapper[4971]: I1213 09:46:44.983172 4971 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="7fa1590164c3ad0ac6fe3340523ceb2818d18012ec13dea17c013e277da3d024" Dec 13 09:46:44 crc kubenswrapper[4971]: I1213 09:46:44.983275 4971 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-g74xz/crc-debug-jzcqj" Dec 13 09:46:45 crc kubenswrapper[4971]: I1213 09:46:45.581697 4971 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-must-gather-g74xz/crc-debug-jzcqj"] Dec 13 09:46:45 crc kubenswrapper[4971]: I1213 09:46:45.592141 4971 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-must-gather-g74xz/crc-debug-jzcqj"] Dec 13 09:46:45 crc kubenswrapper[4971]: I1213 09:46:45.769648 4971 scope.go:117] "RemoveContainer" containerID="80deb18ddf85834e136ab62ea93a1fa75315016d9f220eb217720a5e562f5e54" Dec 13 09:46:45 crc kubenswrapper[4971]: E1213 09:46:45.769950 4971 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-82xjz_openshift-machine-config-operator(e0c4af12-edda-41d7-9b44-f87396174ef2)\"" pod="openshift-machine-config-operator/machine-config-daemon-82xjz" podUID="e0c4af12-edda-41d7-9b44-f87396174ef2" Dec 13 09:46:45 crc kubenswrapper[4971]: I1213 09:46:45.782875 4971 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="553dd05f-d4b1-4aa9-afc9-1ed89824955d" path="/var/lib/kubelet/pods/553dd05f-d4b1-4aa9-afc9-1ed89824955d/volumes" Dec 13 09:46:46 crc kubenswrapper[4971]: I1213 09:46:46.801365 4971 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-must-gather-g74xz/crc-debug-ccpwn"] Dec 13 09:46:46 crc kubenswrapper[4971]: E1213 09:46:46.802700 4971 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="553dd05f-d4b1-4aa9-afc9-1ed89824955d" containerName="container-00" Dec 13 09:46:46 crc kubenswrapper[4971]: I1213 09:46:46.802723 4971 state_mem.go:107] "Deleted CPUSet assignment" podUID="553dd05f-d4b1-4aa9-afc9-1ed89824955d" containerName="container-00" Dec 13 09:46:46 crc kubenswrapper[4971]: I1213 09:46:46.803040 4971 memory_manager.go:354] "RemoveStaleState removing state" podUID="553dd05f-d4b1-4aa9-afc9-1ed89824955d" containerName="container-00" Dec 13 09:46:46 crc kubenswrapper[4971]: I1213 09:46:46.804092 4971 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-g74xz/crc-debug-ccpwn" Dec 13 09:46:46 crc kubenswrapper[4971]: I1213 09:46:46.811875 4971 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-must-gather-g74xz"/"default-dockercfg-2tvhw" Dec 13 09:46:46 crc kubenswrapper[4971]: I1213 09:46:46.909074 4971 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host\" (UniqueName: \"kubernetes.io/host-path/f6295795-bcc0-4966-b872-78a938ff1ec1-host\") pod \"crc-debug-ccpwn\" (UID: \"f6295795-bcc0-4966-b872-78a938ff1ec1\") " pod="openshift-must-gather-g74xz/crc-debug-ccpwn" Dec 13 09:46:46 crc kubenswrapper[4971]: I1213 09:46:46.909789 4971 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-sc8l6\" (UniqueName: \"kubernetes.io/projected/f6295795-bcc0-4966-b872-78a938ff1ec1-kube-api-access-sc8l6\") pod \"crc-debug-ccpwn\" (UID: \"f6295795-bcc0-4966-b872-78a938ff1ec1\") " pod="openshift-must-gather-g74xz/crc-debug-ccpwn" Dec 13 09:46:47 crc kubenswrapper[4971]: I1213 09:46:47.012813 4971 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host\" (UniqueName: \"kubernetes.io/host-path/f6295795-bcc0-4966-b872-78a938ff1ec1-host\") pod \"crc-debug-ccpwn\" (UID: \"f6295795-bcc0-4966-b872-78a938ff1ec1\") " pod="openshift-must-gather-g74xz/crc-debug-ccpwn" Dec 13 09:46:47 crc kubenswrapper[4971]: I1213 09:46:47.012888 4971 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-sc8l6\" (UniqueName: \"kubernetes.io/projected/f6295795-bcc0-4966-b872-78a938ff1ec1-kube-api-access-sc8l6\") pod \"crc-debug-ccpwn\" (UID: \"f6295795-bcc0-4966-b872-78a938ff1ec1\") " pod="openshift-must-gather-g74xz/crc-debug-ccpwn" Dec 13 09:46:47 crc kubenswrapper[4971]: I1213 09:46:47.013029 4971 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host\" (UniqueName: \"kubernetes.io/host-path/f6295795-bcc0-4966-b872-78a938ff1ec1-host\") pod \"crc-debug-ccpwn\" (UID: \"f6295795-bcc0-4966-b872-78a938ff1ec1\") " pod="openshift-must-gather-g74xz/crc-debug-ccpwn" Dec 13 09:46:47 crc kubenswrapper[4971]: I1213 09:46:47.039826 4971 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-sc8l6\" (UniqueName: \"kubernetes.io/projected/f6295795-bcc0-4966-b872-78a938ff1ec1-kube-api-access-sc8l6\") pod \"crc-debug-ccpwn\" (UID: \"f6295795-bcc0-4966-b872-78a938ff1ec1\") " pod="openshift-must-gather-g74xz/crc-debug-ccpwn" Dec 13 09:46:47 crc kubenswrapper[4971]: I1213 09:46:47.137419 4971 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-g74xz/crc-debug-ccpwn" Dec 13 09:46:48 crc kubenswrapper[4971]: I1213 09:46:48.015276 4971 generic.go:334] "Generic (PLEG): container finished" podID="f6295795-bcc0-4966-b872-78a938ff1ec1" containerID="8f91e4e4aa510a175f4794ebc0f1f5fa78c8696afc3be5f4a3ffbbc7a7960d72" exitCode=0 Dec 13 09:46:48 crc kubenswrapper[4971]: I1213 09:46:48.015387 4971 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-g74xz/crc-debug-ccpwn" event={"ID":"f6295795-bcc0-4966-b872-78a938ff1ec1","Type":"ContainerDied","Data":"8f91e4e4aa510a175f4794ebc0f1f5fa78c8696afc3be5f4a3ffbbc7a7960d72"} Dec 13 09:46:48 crc kubenswrapper[4971]: I1213 09:46:48.016469 4971 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-g74xz/crc-debug-ccpwn" event={"ID":"f6295795-bcc0-4966-b872-78a938ff1ec1","Type":"ContainerStarted","Data":"e8b6e62784914a2955a78c1e6309382d36198d3f68852eb9dece1e9517fb1111"} Dec 13 09:46:48 crc kubenswrapper[4971]: I1213 09:46:48.098666 4971 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-must-gather-g74xz/crc-debug-ccpwn"] Dec 13 09:46:48 crc kubenswrapper[4971]: I1213 09:46:48.113305 4971 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-must-gather-g74xz/crc-debug-ccpwn"] Dec 13 09:46:49 crc kubenswrapper[4971]: I1213 09:46:49.144819 4971 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-g74xz/crc-debug-ccpwn" Dec 13 09:46:49 crc kubenswrapper[4971]: I1213 09:46:49.327669 4971 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-sc8l6\" (UniqueName: \"kubernetes.io/projected/f6295795-bcc0-4966-b872-78a938ff1ec1-kube-api-access-sc8l6\") pod \"f6295795-bcc0-4966-b872-78a938ff1ec1\" (UID: \"f6295795-bcc0-4966-b872-78a938ff1ec1\") " Dec 13 09:46:49 crc kubenswrapper[4971]: I1213 09:46:49.328810 4971 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"host\" (UniqueName: \"kubernetes.io/host-path/f6295795-bcc0-4966-b872-78a938ff1ec1-host\") pod \"f6295795-bcc0-4966-b872-78a938ff1ec1\" (UID: \"f6295795-bcc0-4966-b872-78a938ff1ec1\") " Dec 13 09:46:49 crc kubenswrapper[4971]: I1213 09:46:49.328919 4971 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/f6295795-bcc0-4966-b872-78a938ff1ec1-host" (OuterVolumeSpecName: "host") pod "f6295795-bcc0-4966-b872-78a938ff1ec1" (UID: "f6295795-bcc0-4966-b872-78a938ff1ec1"). InnerVolumeSpecName "host". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 13 09:46:49 crc kubenswrapper[4971]: I1213 09:46:49.330449 4971 reconciler_common.go:293] "Volume detached for volume \"host\" (UniqueName: \"kubernetes.io/host-path/f6295795-bcc0-4966-b872-78a938ff1ec1-host\") on node \"crc\" DevicePath \"\"" Dec 13 09:46:49 crc kubenswrapper[4971]: I1213 09:46:49.345826 4971 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/f6295795-bcc0-4966-b872-78a938ff1ec1-kube-api-access-sc8l6" (OuterVolumeSpecName: "kube-api-access-sc8l6") pod "f6295795-bcc0-4966-b872-78a938ff1ec1" (UID: "f6295795-bcc0-4966-b872-78a938ff1ec1"). InnerVolumeSpecName "kube-api-access-sc8l6". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 13 09:46:49 crc kubenswrapper[4971]: I1213 09:46:49.432783 4971 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-sc8l6\" (UniqueName: \"kubernetes.io/projected/f6295795-bcc0-4966-b872-78a938ff1ec1-kube-api-access-sc8l6\") on node \"crc\" DevicePath \"\"" Dec 13 09:46:49 crc kubenswrapper[4971]: I1213 09:46:49.781368 4971 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="f6295795-bcc0-4966-b872-78a938ff1ec1" path="/var/lib/kubelet/pods/f6295795-bcc0-4966-b872-78a938ff1ec1/volumes" Dec 13 09:46:50 crc kubenswrapper[4971]: I1213 09:46:50.042450 4971 scope.go:117] "RemoveContainer" containerID="8f91e4e4aa510a175f4794ebc0f1f5fa78c8696afc3be5f4a3ffbbc7a7960d72" Dec 13 09:46:50 crc kubenswrapper[4971]: I1213 09:46:50.042553 4971 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-g74xz/crc-debug-ccpwn" Dec 13 09:46:54 crc kubenswrapper[4971]: I1213 09:46:54.599282 4971 log.go:25] "Finished parsing log file" path="/var/log/pods/cert-manager_cert-manager-5b446d88c5-bmjtf_65bdb0e4-18e8-4db6-af21-341a748772e9/cert-manager-controller/0.log" Dec 13 09:46:54 crc kubenswrapper[4971]: I1213 09:46:54.778113 4971 log.go:25] "Finished parsing log file" path="/var/log/pods/cert-manager_cert-manager-cainjector-7f985d654d-dp572_fb8b0fc5-85d4-4df8-9b43-8a567a888b20/cert-manager-cainjector/0.log" Dec 13 09:46:54 crc kubenswrapper[4971]: I1213 09:46:54.910853 4971 log.go:25] "Finished parsing log file" path="/var/log/pods/cert-manager_cert-manager-webhook-5655c58dd6-qbbjr_974beed7-18ac-4d6b-8192-d844b50b7ad3/cert-manager-webhook/0.log" Dec 13 09:47:00 crc kubenswrapper[4971]: I1213 09:47:00.768707 4971 scope.go:117] "RemoveContainer" containerID="80deb18ddf85834e136ab62ea93a1fa75315016d9f220eb217720a5e562f5e54" Dec 13 09:47:00 crc kubenswrapper[4971]: E1213 09:47:00.769643 4971 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-82xjz_openshift-machine-config-operator(e0c4af12-edda-41d7-9b44-f87396174ef2)\"" pod="openshift-machine-config-operator/machine-config-daemon-82xjz" podUID="e0c4af12-edda-41d7-9b44-f87396174ef2" Dec 13 09:47:12 crc kubenswrapper[4971]: I1213 09:47:12.768924 4971 scope.go:117] "RemoveContainer" containerID="80deb18ddf85834e136ab62ea93a1fa75315016d9f220eb217720a5e562f5e54" Dec 13 09:47:12 crc kubenswrapper[4971]: E1213 09:47:12.769853 4971 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-82xjz_openshift-machine-config-operator(e0c4af12-edda-41d7-9b44-f87396174ef2)\"" pod="openshift-machine-config-operator/machine-config-daemon-82xjz" podUID="e0c4af12-edda-41d7-9b44-f87396174ef2" Dec 13 09:47:16 crc kubenswrapper[4971]: I1213 09:47:16.226127 4971 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-nmstate_nmstate-console-plugin-6ff7998486-hx58s_dfa2c127-31ee-40a0-814a-5af53e135ae8/nmstate-console-plugin/0.log" Dec 13 09:47:16 crc kubenswrapper[4971]: I1213 09:47:16.856608 4971 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-nmstate_nmstate-handler-5b4bx_d8139542-2862-4e1b-8e84-62a77e5bb7e5/nmstate-handler/0.log" Dec 13 09:47:16 crc kubenswrapper[4971]: I1213 09:47:16.874186 4971 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-nmstate_nmstate-metrics-7f7f7578db-xsb6k_f2968080-99c8-4320-9b39-4bcab4dad759/kube-rbac-proxy/0.log" Dec 13 09:47:16 crc kubenswrapper[4971]: I1213 09:47:16.911066 4971 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-nmstate_nmstate-metrics-7f7f7578db-xsb6k_f2968080-99c8-4320-9b39-4bcab4dad759/nmstate-metrics/0.log" Dec 13 09:47:17 crc kubenswrapper[4971]: I1213 09:47:17.122901 4971 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-nmstate_nmstate-operator-6769fb99d-n9v79_658a55da-1b10-45a1-b449-d8d4b8916242/nmstate-operator/0.log" Dec 13 09:47:17 crc kubenswrapper[4971]: I1213 09:47:17.221996 4971 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-nmstate_nmstate-webhook-f8fb84555-pwxvn_9e887c99-7f51-4b00-9571-f4d29c191054/nmstate-webhook/0.log" Dec 13 09:47:25 crc kubenswrapper[4971]: I1213 09:47:25.769074 4971 scope.go:117] "RemoveContainer" containerID="80deb18ddf85834e136ab62ea93a1fa75315016d9f220eb217720a5e562f5e54" Dec 13 09:47:25 crc kubenswrapper[4971]: E1213 09:47:25.769920 4971 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-82xjz_openshift-machine-config-operator(e0c4af12-edda-41d7-9b44-f87396174ef2)\"" pod="openshift-machine-config-operator/machine-config-daemon-82xjz" podUID="e0c4af12-edda-41d7-9b44-f87396174ef2" Dec 13 09:47:30 crc kubenswrapper[4971]: I1213 09:47:30.805419 4971 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ansibletest-ansibletest_a020591e-029f-464f-8d6c-947acba2bf13/ansibletest-ansibletest/0.log" Dec 13 09:47:30 crc kubenswrapper[4971]: I1213 09:47:30.861485 4971 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_barbican-api-55549fd796-4tztx_d553c6ea-f592-4890-bccd-f970616d23fe/barbican-api/0.log" Dec 13 09:47:31 crc kubenswrapper[4971]: I1213 09:47:31.425182 4971 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_barbican-keystone-listener-6959d6f94b-tq6hh_12a4a604-178e-4ad8-a0f7-e2e3d47590bd/barbican-keystone-listener/0.log" Dec 13 09:47:31 crc kubenswrapper[4971]: I1213 09:47:31.469725 4971 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_barbican-api-55549fd796-4tztx_d553c6ea-f592-4890-bccd-f970616d23fe/barbican-api-log/0.log" Dec 13 09:47:31 crc kubenswrapper[4971]: I1213 09:47:31.823461 4971 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_barbican-worker-c7c6b44fc-8hd9n_fb068297-28dd-4a70-acbf-650b9edf380e/barbican-worker-log/0.log" Dec 13 09:47:31 crc kubenswrapper[4971]: I1213 09:47:31.832164 4971 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_barbican-worker-c7c6b44fc-8hd9n_fb068297-28dd-4a70-acbf-650b9edf380e/barbican-worker/0.log" Dec 13 09:47:32 crc kubenswrapper[4971]: I1213 09:47:32.056251 4971 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_barbican-keystone-listener-6959d6f94b-tq6hh_12a4a604-178e-4ad8-a0f7-e2e3d47590bd/barbican-keystone-listener-log/0.log" Dec 13 09:47:32 crc kubenswrapper[4971]: I1213 09:47:32.142700 4971 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_bootstrap-edpm-deployment-openstack-edpm-ipam-jdcq5_27dafb6f-1f04-448c-97d2-0287c277c26a/bootstrap-edpm-deployment-openstack-edpm-ipam/0.log" Dec 13 09:47:32 crc kubenswrapper[4971]: I1213 09:47:32.411181 4971 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ceilometer-0_7ba95a4c-be84-460a-9dfc-b6e1474e1b81/ceilometer-central-agent/0.log" Dec 13 09:47:32 crc kubenswrapper[4971]: I1213 09:47:32.471324 4971 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ceilometer-0_7ba95a4c-be84-460a-9dfc-b6e1474e1b81/ceilometer-notification-agent/0.log" Dec 13 09:47:32 crc kubenswrapper[4971]: I1213 09:47:32.472956 4971 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ceilometer-0_7ba95a4c-be84-460a-9dfc-b6e1474e1b81/proxy-httpd/0.log" Dec 13 09:47:32 crc kubenswrapper[4971]: I1213 09:47:32.545485 4971 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ceilometer-0_7ba95a4c-be84-460a-9dfc-b6e1474e1b81/sg-core/0.log" Dec 13 09:47:32 crc kubenswrapper[4971]: I1213 09:47:32.779481 4971 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ceph-client-edpm-deployment-openstack-edpm-ipam-7grkn_42cff6d1-138e-4a16-a570-cdcbcbe6ab49/ceph-client-edpm-deployment-openstack-edpm-ipam/0.log" Dec 13 09:47:32 crc kubenswrapper[4971]: I1213 09:47:32.826028 4971 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ceph-hci-pre-edpm-deployment-openstack-edpm-ipam-4fh5s_bbbc7fc6-2323-4765-9160-a1c085de09d8/ceph-hci-pre-edpm-deployment-openstack-edpm-ipam/0.log" Dec 13 09:47:33 crc kubenswrapper[4971]: I1213 09:47:33.177404 4971 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_cinder-api-0_7cbba320-ebcc-40d7-88ea-ba668d549245/cinder-api-log/0.log" Dec 13 09:47:33 crc kubenswrapper[4971]: I1213 09:47:33.219835 4971 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_cinder-api-0_7cbba320-ebcc-40d7-88ea-ba668d549245/cinder-api/0.log" Dec 13 09:47:33 crc kubenswrapper[4971]: I1213 09:47:33.581747 4971 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_cinder-backup-0_be3cc24a-8820-4884-8f26-e2885d933527/probe/0.log" Dec 13 09:47:33 crc kubenswrapper[4971]: I1213 09:47:33.702006 4971 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_cinder-scheduler-0_4c666e54-0053-4ecc-b8bd-6a2771ebbad1/cinder-scheduler/0.log" Dec 13 09:47:33 crc kubenswrapper[4971]: I1213 09:47:33.720578 4971 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_cinder-backup-0_be3cc24a-8820-4884-8f26-e2885d933527/cinder-backup/0.log" Dec 13 09:47:33 crc kubenswrapper[4971]: I1213 09:47:33.892309 4971 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_cinder-scheduler-0_4c666e54-0053-4ecc-b8bd-6a2771ebbad1/probe/0.log" Dec 13 09:47:34 crc kubenswrapper[4971]: I1213 09:47:34.092677 4971 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_cinder-volume-volume1-0_7bbf960d-93e9-49d6-8545-28c34e36229e/cinder-volume/0.log" Dec 13 09:47:34 crc kubenswrapper[4971]: I1213 09:47:34.202127 4971 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_cinder-volume-volume1-0_7bbf960d-93e9-49d6-8545-28c34e36229e/probe/0.log" Dec 13 09:47:34 crc kubenswrapper[4971]: I1213 09:47:34.303413 4971 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_configure-network-edpm-deployment-openstack-edpm-ipam-95c87_ef08e99c-0bd8-4f27-ae28-2cfdeb4ec045/configure-network-edpm-deployment-openstack-edpm-ipam/0.log" Dec 13 09:47:34 crc kubenswrapper[4971]: I1213 09:47:34.473431 4971 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_configure-os-edpm-deployment-openstack-edpm-ipam-7qg86_b03f6448-21d3-417f-a1a9-6e7c733550ec/configure-os-edpm-deployment-openstack-edpm-ipam/0.log" Dec 13 09:47:34 crc kubenswrapper[4971]: I1213 09:47:34.600476 4971 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_dnsmasq-dns-ccc856b4f-zh8np_409ac970-9fc4-4b77-9548-1714f54c8d51/init/0.log" Dec 13 09:47:34 crc kubenswrapper[4971]: I1213 09:47:34.823025 4971 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_dnsmasq-dns-ccc856b4f-zh8np_409ac970-9fc4-4b77-9548-1714f54c8d51/init/0.log" Dec 13 09:47:34 crc kubenswrapper[4971]: I1213 09:47:34.963387 4971 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_glance-default-external-api-0_9fbf6500-bf43-4a4c-820c-14a60a27afb9/glance-httpd/0.log" Dec 13 09:47:35 crc kubenswrapper[4971]: I1213 09:47:35.069809 4971 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_dnsmasq-dns-ccc856b4f-zh8np_409ac970-9fc4-4b77-9548-1714f54c8d51/dnsmasq-dns/0.log" Dec 13 09:47:35 crc kubenswrapper[4971]: I1213 09:47:35.168904 4971 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_glance-default-external-api-0_9fbf6500-bf43-4a4c-820c-14a60a27afb9/glance-log/0.log" Dec 13 09:47:35 crc kubenswrapper[4971]: I1213 09:47:35.352871 4971 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_glance-default-internal-api-0_3158ebe6-4c8c-41d1-bf95-6b4492a7e0aa/glance-log/0.log" Dec 13 09:47:35 crc kubenswrapper[4971]: I1213 09:47:35.396740 4971 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_glance-default-internal-api-0_3158ebe6-4c8c-41d1-bf95-6b4492a7e0aa/glance-httpd/0.log" Dec 13 09:47:35 crc kubenswrapper[4971]: I1213 09:47:35.662700 4971 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_horizon-797dfcfc58-wlch7_9a4736be-5537-4638-a203-d36db1ae35cc/horizon/0.log" Dec 13 09:47:35 crc kubenswrapper[4971]: I1213 09:47:35.772848 4971 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_horizontest-tests-horizontest_abaeac3e-11a1-40b0-b694-26822d0c02bf/horizontest-tests-horizontest/0.log" Dec 13 09:47:36 crc kubenswrapper[4971]: I1213 09:47:36.032563 4971 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_install-certs-edpm-deployment-openstack-edpm-ipam-lkhcs_2f26ed2c-695e-4425-85cc-6e03e59b39db/install-certs-edpm-deployment-openstack-edpm-ipam/0.log" Dec 13 09:47:36 crc kubenswrapper[4971]: I1213 09:47:36.161814 4971 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_install-os-edpm-deployment-openstack-edpm-ipam-ffdjh_b9a8435d-184f-4451-82cc-3b88e2e19050/install-os-edpm-deployment-openstack-edpm-ipam/0.log" Dec 13 09:47:36 crc kubenswrapper[4971]: I1213 09:47:36.488470 4971 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_keystone-cron-29426881-xx4lc_13b568d7-cd02-480b-aa55-ef6074c71e60/keystone-cron/0.log" Dec 13 09:47:36 crc kubenswrapper[4971]: I1213 09:47:36.728719 4971 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_keystone-cron-29426941-c6mj5_a7d15eb0-4473-488c-8355-db1a4d139e7c/keystone-cron/0.log" Dec 13 09:47:36 crc kubenswrapper[4971]: I1213 09:47:36.892172 4971 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_kube-state-metrics-0_8314d6ed-4778-4709-a494-0642e0ce0617/kube-state-metrics/0.log" Dec 13 09:47:37 crc kubenswrapper[4971]: I1213 09:47:37.445493 4971 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_libvirt-edpm-deployment-openstack-edpm-ipam-47wpm_ce803e3c-be8a-4ee3-ac0b-42b035c043a5/libvirt-edpm-deployment-openstack-edpm-ipam/0.log" Dec 13 09:47:37 crc kubenswrapper[4971]: I1213 09:47:37.734316 4971 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_manila-api-0_e8e297bb-43e7-45b9-a82b-a46c104519c1/manila-api-log/0.log" Dec 13 09:47:37 crc kubenswrapper[4971]: I1213 09:47:37.768932 4971 scope.go:117] "RemoveContainer" containerID="80deb18ddf85834e136ab62ea93a1fa75315016d9f220eb217720a5e562f5e54" Dec 13 09:47:37 crc kubenswrapper[4971]: E1213 09:47:37.769420 4971 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-82xjz_openshift-machine-config-operator(e0c4af12-edda-41d7-9b44-f87396174ef2)\"" pod="openshift-machine-config-operator/machine-config-daemon-82xjz" podUID="e0c4af12-edda-41d7-9b44-f87396174ef2" Dec 13 09:47:38 crc kubenswrapper[4971]: I1213 09:47:38.089085 4971 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_manila-api-0_e8e297bb-43e7-45b9-a82b-a46c104519c1/manila-api/0.log" Dec 13 09:47:38 crc kubenswrapper[4971]: I1213 09:47:38.314464 4971 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_manila-scheduler-0_13b76412-5c09-4398-b34f-f07adad9d270/manila-scheduler/0.log" Dec 13 09:47:38 crc kubenswrapper[4971]: I1213 09:47:38.336722 4971 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_horizon-797dfcfc58-wlch7_9a4736be-5537-4638-a203-d36db1ae35cc/horizon-log/0.log" Dec 13 09:47:38 crc kubenswrapper[4971]: I1213 09:47:38.347085 4971 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_manila-scheduler-0_13b76412-5c09-4398-b34f-f07adad9d270/probe/0.log" Dec 13 09:47:38 crc kubenswrapper[4971]: I1213 09:47:38.691542 4971 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_manila-share-share1-0_7d1dae22-44f3-47a3-99fa-a281fcffcdd2/probe/0.log" Dec 13 09:47:38 crc kubenswrapper[4971]: I1213 09:47:38.894450 4971 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_manila-share-share1-0_7d1dae22-44f3-47a3-99fa-a281fcffcdd2/manila-share/0.log" Dec 13 09:47:40 crc kubenswrapper[4971]: I1213 09:47:40.235029 4971 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_keystone-7969997c86-5t8c8_8e1b68c0-d925-4c7c-8904-d71b9f6ef0e1/keystone-api/0.log" Dec 13 09:47:40 crc kubenswrapper[4971]: I1213 09:47:40.275401 4971 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_neutron-metadata-edpm-deployment-openstack-edpm-ipam-tpz7w_3a9bc76d-7313-4304-ad56-d6c075837f50/neutron-metadata-edpm-deployment-openstack-edpm-ipam/0.log" Dec 13 09:47:40 crc kubenswrapper[4971]: I1213 09:47:40.394879 4971 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_controller-5bddd4b946-2kfxb_7cc4bd2b-7f1a-46d0-bae1-dcd35dc513ad/kube-rbac-proxy/0.log" Dec 13 09:47:40 crc kubenswrapper[4971]: I1213 09:47:40.728729 4971 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-szzcz_1f567f13-d990-4035-92e4-64835af4aac8/cp-frr-files/0.log" Dec 13 09:47:40 crc kubenswrapper[4971]: I1213 09:47:40.873003 4971 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_controller-5bddd4b946-2kfxb_7cc4bd2b-7f1a-46d0-bae1-dcd35dc513ad/controller/0.log" Dec 13 09:47:40 crc kubenswrapper[4971]: I1213 09:47:40.945976 4971 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-szzcz_1f567f13-d990-4035-92e4-64835af4aac8/cp-frr-files/0.log" Dec 13 09:47:41 crc kubenswrapper[4971]: I1213 09:47:41.036390 4971 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-szzcz_1f567f13-d990-4035-92e4-64835af4aac8/cp-reloader/0.log" Dec 13 09:47:41 crc kubenswrapper[4971]: I1213 09:47:41.044214 4971 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-szzcz_1f567f13-d990-4035-92e4-64835af4aac8/cp-metrics/0.log" Dec 13 09:47:41 crc kubenswrapper[4971]: I1213 09:47:41.123163 4971 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_neutron-7956479f45-8g7zz_4426fcd4-0b89-4e70-8f2f-43b056896953/neutron-httpd/0.log" Dec 13 09:47:41 crc kubenswrapper[4971]: I1213 09:47:41.126289 4971 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-szzcz_1f567f13-d990-4035-92e4-64835af4aac8/cp-reloader/0.log" Dec 13 09:47:41 crc kubenswrapper[4971]: I1213 09:47:41.492279 4971 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-szzcz_1f567f13-d990-4035-92e4-64835af4aac8/cp-frr-files/0.log" Dec 13 09:47:41 crc kubenswrapper[4971]: I1213 09:47:41.541050 4971 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-szzcz_1f567f13-d990-4035-92e4-64835af4aac8/cp-metrics/0.log" Dec 13 09:47:41 crc kubenswrapper[4971]: I1213 09:47:41.593991 4971 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-szzcz_1f567f13-d990-4035-92e4-64835af4aac8/cp-reloader/0.log" Dec 13 09:47:41 crc kubenswrapper[4971]: I1213 09:47:41.609828 4971 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-szzcz_1f567f13-d990-4035-92e4-64835af4aac8/cp-metrics/0.log" Dec 13 09:47:41 crc kubenswrapper[4971]: I1213 09:47:41.902805 4971 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-szzcz_1f567f13-d990-4035-92e4-64835af4aac8/cp-reloader/0.log" Dec 13 09:47:41 crc kubenswrapper[4971]: I1213 09:47:41.932548 4971 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-szzcz_1f567f13-d990-4035-92e4-64835af4aac8/cp-metrics/0.log" Dec 13 09:47:41 crc kubenswrapper[4971]: I1213 09:47:41.953021 4971 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-szzcz_1f567f13-d990-4035-92e4-64835af4aac8/cp-frr-files/0.log" Dec 13 09:47:42 crc kubenswrapper[4971]: I1213 09:47:42.059715 4971 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-szzcz_1f567f13-d990-4035-92e4-64835af4aac8/controller/0.log" Dec 13 09:47:42 crc kubenswrapper[4971]: I1213 09:47:42.276816 4971 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-szzcz_1f567f13-d990-4035-92e4-64835af4aac8/frr-metrics/0.log" Dec 13 09:47:42 crc kubenswrapper[4971]: I1213 09:47:42.330980 4971 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-szzcz_1f567f13-d990-4035-92e4-64835af4aac8/kube-rbac-proxy-frr/0.log" Dec 13 09:47:42 crc kubenswrapper[4971]: I1213 09:47:42.352480 4971 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-szzcz_1f567f13-d990-4035-92e4-64835af4aac8/kube-rbac-proxy/0.log" Dec 13 09:47:42 crc kubenswrapper[4971]: I1213 09:47:42.718883 4971 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-szzcz_1f567f13-d990-4035-92e4-64835af4aac8/reloader/0.log" Dec 13 09:47:42 crc kubenswrapper[4971]: I1213 09:47:42.821070 4971 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-webhook-server-7784b6fcf-tkg6z_f093984d-c941-4805-a0c1-0926aa7d7f38/frr-k8s-webhook-server/0.log" Dec 13 09:47:42 crc kubenswrapper[4971]: I1213 09:47:42.848051 4971 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_neutron-7956479f45-8g7zz_4426fcd4-0b89-4e70-8f2f-43b056896953/neutron-api/0.log" Dec 13 09:47:43 crc kubenswrapper[4971]: I1213 09:47:43.139908 4971 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_metallb-operator-controller-manager-9d86ffb46-9kqjk_1a9addf5-4d0c-447d-8a9a-e6c15b9cdb3a/manager/0.log" Dec 13 09:47:43 crc kubenswrapper[4971]: I1213 09:47:43.174402 4971 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_nova-cell0-conductor-0_5220d566-7310-4e2f-b5b0-4e0c5e826aa9/nova-cell0-conductor-conductor/0.log" Dec 13 09:47:43 crc kubenswrapper[4971]: I1213 09:47:43.276107 4971 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_metallb-operator-webhook-server-657bbb5cbf-njfvg_98e5a1fd-e920-4f5d-82c3-7450cfc0487a/webhook-server/0.log" Dec 13 09:47:43 crc kubenswrapper[4971]: I1213 09:47:43.575751 4971 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_speaker-v229x_c0ee70f0-8158-4fec-bae8-7cd72bdef461/kube-rbac-proxy/0.log" Dec 13 09:47:44 crc kubenswrapper[4971]: I1213 09:47:44.054890 4971 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_nova-cell1-conductor-0_c4b4d448-bf03-4d5c-9017-3b2257922796/nova-cell1-conductor-conductor/0.log" Dec 13 09:47:44 crc kubenswrapper[4971]: I1213 09:47:44.500625 4971 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_nova-cell1-novncproxy-0_38d62929-7b21-4bf7-a2ac-f31047134d54/nova-cell1-novncproxy-novncproxy/0.log" Dec 13 09:47:44 crc kubenswrapper[4971]: I1213 09:47:44.890797 4971 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_nova-custom-ceph-edpm-deployment-openstack-edpm-ipam-4lrjf_e3e414b3-58e3-4da8-9542-7b6d2f229b8d/nova-custom-ceph-edpm-deployment-openstack-edpm-ipam/0.log" Dec 13 09:47:45 crc kubenswrapper[4971]: I1213 09:47:45.246621 4971 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_speaker-v229x_c0ee70f0-8158-4fec-bae8-7cd72bdef461/speaker/0.log" Dec 13 09:47:45 crc kubenswrapper[4971]: I1213 09:47:45.308086 4971 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_nova-metadata-0_d4bb47ae-3205-452b-b596-563d3833ee82/nova-metadata-log/0.log" Dec 13 09:47:45 crc kubenswrapper[4971]: E1213 09:47:45.769665 4971 kubelet_pods.go:538] "Hostname for pod was too long, truncated it" podName="test-operator-logs-pod-horizontest-horizontest-tests-horizontest" hostnameMaxLen=63 truncatedHostname="test-operator-logs-pod-horizontest-horizontest-tests-horizontes" Dec 13 09:47:47 crc kubenswrapper[4971]: I1213 09:47:47.127079 4971 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_nova-scheduler-0_14465eba-34c5-43d3-ad9d-06b454f79963/nova-scheduler-scheduler/0.log" Dec 13 09:47:47 crc kubenswrapper[4971]: I1213 09:47:47.674624 4971 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_nova-api-0_d66b7a63-bb42-46dc-8dbb-94452070e5b2/nova-api-log/0.log" Dec 13 09:47:47 crc kubenswrapper[4971]: I1213 09:47:47.774146 4971 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_openstack-cell1-galera-0_96571a02-70f1-46e7-bdd3-935b1f43ecc6/mysql-bootstrap/0.log" Dec 13 09:47:48 crc kubenswrapper[4971]: I1213 09:47:48.052956 4971 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_openstack-cell1-galera-0_96571a02-70f1-46e7-bdd3-935b1f43ecc6/galera/0.log" Dec 13 09:47:48 crc kubenswrapper[4971]: I1213 09:47:48.059126 4971 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_openstack-cell1-galera-0_96571a02-70f1-46e7-bdd3-935b1f43ecc6/mysql-bootstrap/0.log" Dec 13 09:47:48 crc kubenswrapper[4971]: I1213 09:47:48.403508 4971 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_openstack-galera-0_91831fd5-fc52-46ec-ac49-63859378116b/mysql-bootstrap/0.log" Dec 13 09:47:48 crc kubenswrapper[4971]: I1213 09:47:48.655199 4971 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_openstack-galera-0_91831fd5-fc52-46ec-ac49-63859378116b/mysql-bootstrap/0.log" Dec 13 09:47:48 crc kubenswrapper[4971]: I1213 09:47:48.745737 4971 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_openstack-galera-0_91831fd5-fc52-46ec-ac49-63859378116b/galera/0.log" Dec 13 09:47:48 crc kubenswrapper[4971]: I1213 09:47:48.768844 4971 scope.go:117] "RemoveContainer" containerID="80deb18ddf85834e136ab62ea93a1fa75315016d9f220eb217720a5e562f5e54" Dec 13 09:47:48 crc kubenswrapper[4971]: E1213 09:47:48.769217 4971 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-82xjz_openshift-machine-config-operator(e0c4af12-edda-41d7-9b44-f87396174ef2)\"" pod="openshift-machine-config-operator/machine-config-daemon-82xjz" podUID="e0c4af12-edda-41d7-9b44-f87396174ef2" Dec 13 09:47:49 crc kubenswrapper[4971]: I1213 09:47:49.034091 4971 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_openstackclient_b21860ea-6663-4bc0-ac29-b91d640fad5a/openstackclient/0.log" Dec 13 09:47:49 crc kubenswrapper[4971]: I1213 09:47:49.294021 4971 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovn-controller-k5fz9_c0bf2685-ca54-4508-a580-531b05473e76/ovn-controller/0.log" Dec 13 09:47:49 crc kubenswrapper[4971]: I1213 09:47:49.512318 4971 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovn-controller-metrics-r2cbj_fa04788e-0c35-4205-97ed-cb1e38ec9e69/openstack-network-exporter/0.log" Dec 13 09:47:49 crc kubenswrapper[4971]: I1213 09:47:49.901910 4971 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-szzcz_1f567f13-d990-4035-92e4-64835af4aac8/frr/0.log" Dec 13 09:47:49 crc kubenswrapper[4971]: I1213 09:47:49.999104 4971 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovn-controller-ovs-jwh28_1920db8d-2d52-46fa-a4cc-ad787e9886fc/ovsdb-server-init/0.log" Dec 13 09:47:50 crc kubenswrapper[4971]: I1213 09:47:50.193799 4971 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovn-controller-ovs-jwh28_1920db8d-2d52-46fa-a4cc-ad787e9886fc/ovsdb-server-init/0.log" Dec 13 09:47:50 crc kubenswrapper[4971]: I1213 09:47:50.273312 4971 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovn-controller-ovs-jwh28_1920db8d-2d52-46fa-a4cc-ad787e9886fc/ovs-vswitchd/0.log" Dec 13 09:47:50 crc kubenswrapper[4971]: I1213 09:47:50.354170 4971 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_nova-api-0_d66b7a63-bb42-46dc-8dbb-94452070e5b2/nova-api-api/0.log" Dec 13 09:47:50 crc kubenswrapper[4971]: I1213 09:47:50.448855 4971 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovn-controller-ovs-jwh28_1920db8d-2d52-46fa-a4cc-ad787e9886fc/ovsdb-server/0.log" Dec 13 09:47:50 crc kubenswrapper[4971]: I1213 09:47:50.764293 4971 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovn-edpm-deployment-openstack-edpm-ipam-lswgl_9f1e6cde-d76e-45f1-a6a6-8546b16e389b/ovn-edpm-deployment-openstack-edpm-ipam/0.log" Dec 13 09:47:50 crc kubenswrapper[4971]: I1213 09:47:50.843435 4971 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovn-northd-0_2bbf81f1-6733-4320-9bf5-c73a00d9051a/openstack-network-exporter/0.log" Dec 13 09:47:51 crc kubenswrapper[4971]: I1213 09:47:51.012127 4971 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovn-northd-0_2bbf81f1-6733-4320-9bf5-c73a00d9051a/ovn-northd/0.log" Dec 13 09:47:51 crc kubenswrapper[4971]: I1213 09:47:51.109363 4971 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovsdbserver-nb-0_3952f27c-fb18-4e8a-b3a4-a3fa077378c4/openstack-network-exporter/0.log" Dec 13 09:47:51 crc kubenswrapper[4971]: I1213 09:47:51.279237 4971 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovsdbserver-nb-0_3952f27c-fb18-4e8a-b3a4-a3fa077378c4/ovsdbserver-nb/0.log" Dec 13 09:47:51 crc kubenswrapper[4971]: I1213 09:47:51.450795 4971 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovsdbserver-sb-0_a85733fd-9af5-4fc4-8096-9b368a793fe7/openstack-network-exporter/0.log" Dec 13 09:47:51 crc kubenswrapper[4971]: I1213 09:47:51.595140 4971 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovsdbserver-sb-0_a85733fd-9af5-4fc4-8096-9b368a793fe7/ovsdbserver-sb/0.log" Dec 13 09:47:52 crc kubenswrapper[4971]: I1213 09:47:52.407776 4971 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_rabbitmq-cell1-server-0_299317ec-a652-40f2-b0c5-e25f099204c4/setup-container/0.log" Dec 13 09:47:52 crc kubenswrapper[4971]: I1213 09:47:52.636372 4971 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_rabbitmq-cell1-server-0_299317ec-a652-40f2-b0c5-e25f099204c4/setup-container/0.log" Dec 13 09:47:52 crc kubenswrapper[4971]: I1213 09:47:52.657917 4971 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_placement-569b6fd56b-w7xmm_abb2fa2a-6153-4b45-83b9-2c34265c5e7f/placement-api/0.log" Dec 13 09:47:52 crc kubenswrapper[4971]: I1213 09:47:52.969448 4971 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_rabbitmq-cell1-server-0_299317ec-a652-40f2-b0c5-e25f099204c4/rabbitmq/0.log" Dec 13 09:47:53 crc kubenswrapper[4971]: I1213 09:47:53.026310 4971 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_placement-569b6fd56b-w7xmm_abb2fa2a-6153-4b45-83b9-2c34265c5e7f/placement-log/0.log" Dec 13 09:47:53 crc kubenswrapper[4971]: I1213 09:47:53.227611 4971 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_rabbitmq-server-0_42eb5011-6651-43cb-a562-c89cc6df2c53/setup-container/0.log" Dec 13 09:47:53 crc kubenswrapper[4971]: I1213 09:47:53.438911 4971 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_rabbitmq-server-0_42eb5011-6651-43cb-a562-c89cc6df2c53/setup-container/0.log" Dec 13 09:47:53 crc kubenswrapper[4971]: I1213 09:47:53.474444 4971 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_nova-metadata-0_d4bb47ae-3205-452b-b596-563d3833ee82/nova-metadata-metadata/0.log" Dec 13 09:47:53 crc kubenswrapper[4971]: I1213 09:47:53.572734 4971 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_rabbitmq-server-0_42eb5011-6651-43cb-a562-c89cc6df2c53/rabbitmq/0.log" Dec 13 09:47:53 crc kubenswrapper[4971]: I1213 09:47:53.710981 4971 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_reboot-os-edpm-deployment-openstack-edpm-ipam-mmm7v_94cbb455-7651-4d43-ad24-495457496a2d/reboot-os-edpm-deployment-openstack-edpm-ipam/0.log" Dec 13 09:47:53 crc kubenswrapper[4971]: I1213 09:47:53.880870 4971 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_repo-setup-edpm-deployment-openstack-edpm-ipam-fpgpm_53d7daa1-9679-44ba-9023-808ce5c250a6/repo-setup-edpm-deployment-openstack-edpm-ipam/0.log" Dec 13 09:47:54 crc kubenswrapper[4971]: I1213 09:47:54.000505 4971 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_run-os-edpm-deployment-openstack-edpm-ipam-tn7h4_88c03dc6-31a0-4c8c-93b6-61a9c895dffe/run-os-edpm-deployment-openstack-edpm-ipam/0.log" Dec 13 09:47:54 crc kubenswrapper[4971]: I1213 09:47:54.128707 4971 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ssh-known-hosts-edpm-deployment-b4wgp_82daff50-eacf-45ef-b206-ffe0e1d2fe89/ssh-known-hosts-edpm-deployment/0.log" Dec 13 09:47:54 crc kubenswrapper[4971]: I1213 09:47:54.375334 4971 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_tempest-tests-tempest-s00-full_879e5242-5aa8-44fb-b759-6babf9c6559d/tempest-tests-tempest-tests-runner/0.log" Dec 13 09:47:54 crc kubenswrapper[4971]: I1213 09:47:54.442461 4971 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_tempest-tests-tempest-s01-single-test_b0fcccc8-3a38-42b6-a10c-57a0a44b48c3/tempest-tests-tempest-tests-runner/0.log" Dec 13 09:47:54 crc kubenswrapper[4971]: I1213 09:47:54.686107 4971 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_test-operator-logs-pod-ansibletest-ansibletest-ansibletest_4cf376d3-529c-41ee-8da5-9aa3bb37cfed/test-operator-logs-container/0.log" Dec 13 09:47:54 crc kubenswrapper[4971]: I1213 09:47:54.757919 4971 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_test-operator-logs-pod-horizontest-horizontest-tests-horizontest_73a838a8-433b-4ddf-ab44-1527de29aa6f/test-operator-logs-container/0.log" Dec 13 09:47:54 crc kubenswrapper[4971]: I1213 09:47:54.958987 4971 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_test-operator-logs-pod-tempest-tempest-tests-tempest_b5af123d-e343-4e82-861d-3eaed6d9ca0c/test-operator-logs-container/0.log" Dec 13 09:47:55 crc kubenswrapper[4971]: I1213 09:47:55.067211 4971 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_test-operator-logs-pod-tobiko-tobiko-tests-tobiko_d608d411-d575-467c-8738-17284ac309d7/test-operator-logs-container/0.log" Dec 13 09:47:55 crc kubenswrapper[4971]: I1213 09:47:55.227614 4971 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_tobiko-tests-tobiko-s00-podified-functional_0a604d8e-de8c-4ae6-b13a-aafd2adee17a/tobiko-tests-tobiko/0.log" Dec 13 09:47:55 crc kubenswrapper[4971]: I1213 09:47:55.425345 4971 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_tobiko-tests-tobiko-s01-sanity_32d4b615-70d7-4af1-9280-8fb2452feb76/tobiko-tests-tobiko/0.log" Dec 13 09:47:55 crc kubenswrapper[4971]: I1213 09:47:55.575539 4971 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_validate-network-edpm-deployment-openstack-edpm-ipam-4lhml_2185de1e-5198-4f0f-8918-1e90aadb7fb7/validate-network-edpm-deployment-openstack-edpm-ipam/0.log" Dec 13 09:47:58 crc kubenswrapper[4971]: I1213 09:47:58.289250 4971 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_memcached-0_e6adf8ca-c137-4b7d-bf8c-74c0766e57f0/memcached/0.log" Dec 13 09:48:02 crc kubenswrapper[4971]: I1213 09:48:02.769918 4971 scope.go:117] "RemoveContainer" containerID="80deb18ddf85834e136ab62ea93a1fa75315016d9f220eb217720a5e562f5e54" Dec 13 09:48:02 crc kubenswrapper[4971]: E1213 09:48:02.771074 4971 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-82xjz_openshift-machine-config-operator(e0c4af12-edda-41d7-9b44-f87396174ef2)\"" pod="openshift-machine-config-operator/machine-config-daemon-82xjz" podUID="e0c4af12-edda-41d7-9b44-f87396174ef2" Dec 13 09:48:04 crc kubenswrapper[4971]: I1213 09:48:04.360841 4971 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_5b7fccbebf0e22d2dd769066fa7aaa90fd620c5db34f2af6c91e4319d4gv4b7_75227ae7-8658-447d-90c5-b0619fe75d8f/util/0.log" Dec 13 09:48:04 crc kubenswrapper[4971]: I1213 09:48:04.809619 4971 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_5b7fccbebf0e22d2dd769066fa7aaa90fd620c5db34f2af6c91e4319d4gv4b7_75227ae7-8658-447d-90c5-b0619fe75d8f/util/0.log" Dec 13 09:48:04 crc kubenswrapper[4971]: I1213 09:48:04.896444 4971 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_5b7fccbebf0e22d2dd769066fa7aaa90fd620c5db34f2af6c91e4319d4gv4b7_75227ae7-8658-447d-90c5-b0619fe75d8f/pull/0.log" Dec 13 09:48:04 crc kubenswrapper[4971]: I1213 09:48:04.962695 4971 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_5b7fccbebf0e22d2dd769066fa7aaa90fd620c5db34f2af6c91e4319d4gv4b7_75227ae7-8658-447d-90c5-b0619fe75d8f/pull/0.log" Dec 13 09:48:05 crc kubenswrapper[4971]: I1213 09:48:05.304371 4971 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_5b7fccbebf0e22d2dd769066fa7aaa90fd620c5db34f2af6c91e4319d4gv4b7_75227ae7-8658-447d-90c5-b0619fe75d8f/util/0.log" Dec 13 09:48:05 crc kubenswrapper[4971]: I1213 09:48:05.354782 4971 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_5b7fccbebf0e22d2dd769066fa7aaa90fd620c5db34f2af6c91e4319d4gv4b7_75227ae7-8658-447d-90c5-b0619fe75d8f/extract/0.log" Dec 13 09:48:05 crc kubenswrapper[4971]: I1213 09:48:05.355545 4971 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_5b7fccbebf0e22d2dd769066fa7aaa90fd620c5db34f2af6c91e4319d4gv4b7_75227ae7-8658-447d-90c5-b0619fe75d8f/pull/0.log" Dec 13 09:48:05 crc kubenswrapper[4971]: I1213 09:48:05.565791 4971 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_98085b0df3808ebec39f9f9529f737144fe2dbcdaa4f334014817c0fa89s2m9_c94817e4-313a-4e9a-a0d9-becb8044d4a8/util/0.log" Dec 13 09:48:05 crc kubenswrapper[4971]: I1213 09:48:05.848767 4971 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_98085b0df3808ebec39f9f9529f737144fe2dbcdaa4f334014817c0fa89s2m9_c94817e4-313a-4e9a-a0d9-becb8044d4a8/util/0.log" Dec 13 09:48:05 crc kubenswrapper[4971]: I1213 09:48:05.892078 4971 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_98085b0df3808ebec39f9f9529f737144fe2dbcdaa4f334014817c0fa89s2m9_c94817e4-313a-4e9a-a0d9-becb8044d4a8/pull/0.log" Dec 13 09:48:05 crc kubenswrapper[4971]: I1213 09:48:05.927915 4971 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_98085b0df3808ebec39f9f9529f737144fe2dbcdaa4f334014817c0fa89s2m9_c94817e4-313a-4e9a-a0d9-becb8044d4a8/pull/0.log" Dec 13 09:48:06 crc kubenswrapper[4971]: I1213 09:48:06.242917 4971 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_98085b0df3808ebec39f9f9529f737144fe2dbcdaa4f334014817c0fa89s2m9_c94817e4-313a-4e9a-a0d9-becb8044d4a8/util/0.log" Dec 13 09:48:06 crc kubenswrapper[4971]: I1213 09:48:06.363105 4971 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_98085b0df3808ebec39f9f9529f737144fe2dbcdaa4f334014817c0fa89s2m9_c94817e4-313a-4e9a-a0d9-becb8044d4a8/pull/0.log" Dec 13 09:48:06 crc kubenswrapper[4971]: I1213 09:48:06.370922 4971 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_98085b0df3808ebec39f9f9529f737144fe2dbcdaa4f334014817c0fa89s2m9_c94817e4-313a-4e9a-a0d9-becb8044d4a8/extract/0.log" Dec 13 09:48:06 crc kubenswrapper[4971]: I1213 09:48:06.791427 4971 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_certified-operators-szqz6_33212ad2-538d-4059-bc1f-e98a51c93b62/extract-utilities/0.log" Dec 13 09:48:07 crc kubenswrapper[4971]: I1213 09:48:07.560781 4971 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_certified-operators-szqz6_33212ad2-538d-4059-bc1f-e98a51c93b62/extract-content/0.log" Dec 13 09:48:07 crc kubenswrapper[4971]: I1213 09:48:07.564103 4971 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_certified-operators-szqz6_33212ad2-538d-4059-bc1f-e98a51c93b62/extract-utilities/0.log" Dec 13 09:48:07 crc kubenswrapper[4971]: I1213 09:48:07.722123 4971 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_certified-operators-szqz6_33212ad2-538d-4059-bc1f-e98a51c93b62/extract-content/0.log" Dec 13 09:48:07 crc kubenswrapper[4971]: I1213 09:48:07.877623 4971 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_certified-operators-szqz6_33212ad2-538d-4059-bc1f-e98a51c93b62/extract-utilities/0.log" Dec 13 09:48:07 crc kubenswrapper[4971]: I1213 09:48:07.931151 4971 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_certified-operators-szqz6_33212ad2-538d-4059-bc1f-e98a51c93b62/extract-content/0.log" Dec 13 09:48:08 crc kubenswrapper[4971]: I1213 09:48:08.258177 4971 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_community-operators-kzw5r_b2475f4e-bfe7-4678-8779-2b56c189b3b0/extract-utilities/0.log" Dec 13 09:48:08 crc kubenswrapper[4971]: I1213 09:48:08.678834 4971 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_community-operators-kzw5r_b2475f4e-bfe7-4678-8779-2b56c189b3b0/extract-utilities/0.log" Dec 13 09:48:08 crc kubenswrapper[4971]: I1213 09:48:08.799489 4971 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_community-operators-kzw5r_b2475f4e-bfe7-4678-8779-2b56c189b3b0/extract-content/0.log" Dec 13 09:48:08 crc kubenswrapper[4971]: I1213 09:48:08.868741 4971 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_community-operators-kzw5r_b2475f4e-bfe7-4678-8779-2b56c189b3b0/extract-content/0.log" Dec 13 09:48:09 crc kubenswrapper[4971]: I1213 09:48:09.254374 4971 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_community-operators-kzw5r_b2475f4e-bfe7-4678-8779-2b56c189b3b0/extract-content/0.log" Dec 13 09:48:09 crc kubenswrapper[4971]: I1213 09:48:09.350051 4971 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_community-operators-kzw5r_b2475f4e-bfe7-4678-8779-2b56c189b3b0/extract-utilities/0.log" Dec 13 09:48:09 crc kubenswrapper[4971]: I1213 09:48:09.978786 4971 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_certified-operators-szqz6_33212ad2-538d-4059-bc1f-e98a51c93b62/registry-server/0.log" Dec 13 09:48:10 crc kubenswrapper[4971]: I1213 09:48:10.352897 4971 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-marketplace-hvzhk_4b52f930-cf4c-4a97-b838-9afc9e3d2b31/extract-utilities/0.log" Dec 13 09:48:10 crc kubenswrapper[4971]: I1213 09:48:10.418814 4971 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_marketplace-operator-79b997595-w4r9w_f8802833-bb84-4479-be26-f0fae563f1b0/marketplace-operator/0.log" Dec 13 09:48:10 crc kubenswrapper[4971]: I1213 09:48:10.718016 4971 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-marketplace-hvzhk_4b52f930-cf4c-4a97-b838-9afc9e3d2b31/extract-content/0.log" Dec 13 09:48:10 crc kubenswrapper[4971]: I1213 09:48:10.730024 4971 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-marketplace-hvzhk_4b52f930-cf4c-4a97-b838-9afc9e3d2b31/extract-utilities/0.log" Dec 13 09:48:10 crc kubenswrapper[4971]: I1213 09:48:10.906443 4971 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-marketplace-hvzhk_4b52f930-cf4c-4a97-b838-9afc9e3d2b31/extract-content/0.log" Dec 13 09:48:11 crc kubenswrapper[4971]: I1213 09:48:11.117708 4971 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-marketplace-hvzhk_4b52f930-cf4c-4a97-b838-9afc9e3d2b31/extract-utilities/0.log" Dec 13 09:48:11 crc kubenswrapper[4971]: I1213 09:48:11.160280 4971 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_community-operators-kzw5r_b2475f4e-bfe7-4678-8779-2b56c189b3b0/registry-server/0.log" Dec 13 09:48:11 crc kubenswrapper[4971]: I1213 09:48:11.204535 4971 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-marketplace-hvzhk_4b52f930-cf4c-4a97-b838-9afc9e3d2b31/extract-content/0.log" Dec 13 09:48:11 crc kubenswrapper[4971]: I1213 09:48:11.477919 4971 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-operators-cvcsq_6fa06763-60e2-4bc5-91f7-58ded0345c6a/extract-utilities/0.log" Dec 13 09:48:11 crc kubenswrapper[4971]: I1213 09:48:11.539851 4971 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-marketplace-hvzhk_4b52f930-cf4c-4a97-b838-9afc9e3d2b31/registry-server/0.log" Dec 13 09:48:11 crc kubenswrapper[4971]: I1213 09:48:11.764731 4971 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-operators-cvcsq_6fa06763-60e2-4bc5-91f7-58ded0345c6a/extract-content/0.log" Dec 13 09:48:11 crc kubenswrapper[4971]: I1213 09:48:11.789434 4971 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-operators-cvcsq_6fa06763-60e2-4bc5-91f7-58ded0345c6a/extract-content/0.log" Dec 13 09:48:11 crc kubenswrapper[4971]: I1213 09:48:11.798833 4971 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-operators-cvcsq_6fa06763-60e2-4bc5-91f7-58ded0345c6a/extract-utilities/0.log" Dec 13 09:48:12 crc kubenswrapper[4971]: I1213 09:48:12.118183 4971 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-operators-cvcsq_6fa06763-60e2-4bc5-91f7-58ded0345c6a/extract-content/0.log" Dec 13 09:48:12 crc kubenswrapper[4971]: I1213 09:48:12.219243 4971 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-operators-cvcsq_6fa06763-60e2-4bc5-91f7-58ded0345c6a/extract-utilities/0.log" Dec 13 09:48:13 crc kubenswrapper[4971]: I1213 09:48:13.540830 4971 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-operators-cvcsq_6fa06763-60e2-4bc5-91f7-58ded0345c6a/registry-server/0.log" Dec 13 09:48:14 crc kubenswrapper[4971]: I1213 09:48:14.768891 4971 scope.go:117] "RemoveContainer" containerID="80deb18ddf85834e136ab62ea93a1fa75315016d9f220eb217720a5e562f5e54" Dec 13 09:48:14 crc kubenswrapper[4971]: E1213 09:48:14.769578 4971 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-82xjz_openshift-machine-config-operator(e0c4af12-edda-41d7-9b44-f87396174ef2)\"" pod="openshift-machine-config-operator/machine-config-daemon-82xjz" podUID="e0c4af12-edda-41d7-9b44-f87396174ef2" Dec 13 09:48:28 crc kubenswrapper[4971]: I1213 09:48:28.804352 4971 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_barbican-operator-controller-manager-95949466-6hkxl_3b74a6fd-0613-4f48-b585-64570abdb09f/manager/0.log" Dec 13 09:48:29 crc kubenswrapper[4971]: I1213 09:48:29.084748 4971 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_cinder-operator-controller-manager-5cf45c46bd-8sg42_62c93900-e946-4684-9c40-36c1d8ff0468/manager/0.log" Dec 13 09:48:29 crc kubenswrapper[4971]: I1213 09:48:29.192852 4971 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_designate-operator-controller-manager-66f8b87655-qxr27_cfe5eaa1-4beb-49ff-a080-ab230292c8ce/manager/0.log" Dec 13 09:48:29 crc kubenswrapper[4971]: I1213 09:48:29.348283 4971 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_ea5d71784a81c95dff2031a02f0a0b3f756f86f14acad8f152d938f56fl9xnm_8fbc1cb9-af8d-4ab7-92e4-63035d0c1d74/util/0.log" Dec 13 09:48:29 crc kubenswrapper[4971]: I1213 09:48:29.590661 4971 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_ea5d71784a81c95dff2031a02f0a0b3f756f86f14acad8f152d938f56fl9xnm_8fbc1cb9-af8d-4ab7-92e4-63035d0c1d74/util/0.log" Dec 13 09:48:29 crc kubenswrapper[4971]: I1213 09:48:29.616336 4971 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_ea5d71784a81c95dff2031a02f0a0b3f756f86f14acad8f152d938f56fl9xnm_8fbc1cb9-af8d-4ab7-92e4-63035d0c1d74/pull/0.log" Dec 13 09:48:29 crc kubenswrapper[4971]: I1213 09:48:29.636401 4971 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_ea5d71784a81c95dff2031a02f0a0b3f756f86f14acad8f152d938f56fl9xnm_8fbc1cb9-af8d-4ab7-92e4-63035d0c1d74/pull/0.log" Dec 13 09:48:29 crc kubenswrapper[4971]: I1213 09:48:29.770794 4971 scope.go:117] "RemoveContainer" containerID="80deb18ddf85834e136ab62ea93a1fa75315016d9f220eb217720a5e562f5e54" Dec 13 09:48:29 crc kubenswrapper[4971]: E1213 09:48:29.771084 4971 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-82xjz_openshift-machine-config-operator(e0c4af12-edda-41d7-9b44-f87396174ef2)\"" pod="openshift-machine-config-operator/machine-config-daemon-82xjz" podUID="e0c4af12-edda-41d7-9b44-f87396174ef2" Dec 13 09:48:29 crc kubenswrapper[4971]: I1213 09:48:29.829791 4971 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_ea5d71784a81c95dff2031a02f0a0b3f756f86f14acad8f152d938f56fl9xnm_8fbc1cb9-af8d-4ab7-92e4-63035d0c1d74/pull/0.log" Dec 13 09:48:29 crc kubenswrapper[4971]: I1213 09:48:29.922042 4971 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_ea5d71784a81c95dff2031a02f0a0b3f756f86f14acad8f152d938f56fl9xnm_8fbc1cb9-af8d-4ab7-92e4-63035d0c1d74/util/0.log" Dec 13 09:48:29 crc kubenswrapper[4971]: I1213 09:48:29.944304 4971 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_ea5d71784a81c95dff2031a02f0a0b3f756f86f14acad8f152d938f56fl9xnm_8fbc1cb9-af8d-4ab7-92e4-63035d0c1d74/extract/0.log" Dec 13 09:48:30 crc kubenswrapper[4971]: I1213 09:48:30.140943 4971 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_glance-operator-controller-manager-767f9d7567-tmc6q_2062b86b-1412-4974-a153-802d9f390944/manager/0.log" Dec 13 09:48:30 crc kubenswrapper[4971]: I1213 09:48:30.195391 4971 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_heat-operator-controller-manager-59b8dcb766-4xg9v_808957a1-29e0-4991-bc8a-23f5a17ecb74/manager/0.log" Dec 13 09:48:30 crc kubenswrapper[4971]: I1213 09:48:30.424672 4971 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_horizon-operator-controller-manager-6ccf486b9-bb2ft_8e94aa39-05a2-4a63-8b45-a914021bccd7/manager/0.log" Dec 13 09:48:30 crc kubenswrapper[4971]: I1213 09:48:30.758144 4971 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_infra-operator-controller-manager-58944d7758-59mz9_846e13b5-58bb-4a12-a77a-0a4a9b53247f/manager/0.log" Dec 13 09:48:30 crc kubenswrapper[4971]: I1213 09:48:30.773848 4971 scope.go:117] "RemoveContainer" containerID="b24457bee748c4942f2344b7bcededdf6e9a8e90748496caea2610a8e50fbd39" Dec 13 09:48:30 crc kubenswrapper[4971]: I1213 09:48:30.775844 4971 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_ironic-operator-controller-manager-f458558d7-m9grc_85d76572-e53b-420f-9ffe-d0534bdb1c96/manager/0.log" Dec 13 09:48:30 crc kubenswrapper[4971]: I1213 09:48:30.958855 4971 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_keystone-operator-controller-manager-5c7cbf548f-qv7p8_3e2847dd-dcdc-4740-a4a8-4a2417247dc5/manager/0.log" Dec 13 09:48:31 crc kubenswrapper[4971]: I1213 09:48:31.191170 4971 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_manila-operator-controller-manager-5fdd9786f7-jjw6m_2761a6c0-8c0d-4430-8976-378c2d6df9f7/manager/0.log" Dec 13 09:48:31 crc kubenswrapper[4971]: I1213 09:48:31.351177 4971 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_mariadb-operator-controller-manager-f76f4954c-cst7r_4bd9b122-00a8-4895-a44b-0773aae9f6a9/manager/0.log" Dec 13 09:48:31 crc kubenswrapper[4971]: I1213 09:48:31.502657 4971 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_neutron-operator-controller-manager-7cd87b778f-l6b8q_0c834237-c914-43bc-b312-b208ba303bb3/manager/0.log" Dec 13 09:48:31 crc kubenswrapper[4971]: I1213 09:48:31.720680 4971 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_nova-operator-controller-manager-5fbbf8b6cc-k4vtr_a8909dd9-9995-4694-9b5b-47106910656c/manager/0.log" Dec 13 09:48:31 crc kubenswrapper[4971]: I1213 09:48:31.820081 4971 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_octavia-operator-controller-manager-68c649d9d-v9t9p_8a160d6b-ce31-4d73-918b-368fbc44eb2b/manager/0.log" Dec 13 09:48:31 crc kubenswrapper[4971]: I1213 09:48:31.976242 4971 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_openstack-baremetal-operator-controller-manager-689f887b54jnn8l_706451b8-644b-4151-b5b9-9975082e556f/manager/0.log" Dec 13 09:48:32 crc kubenswrapper[4971]: I1213 09:48:32.455024 4971 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_openstack-operator-controller-operator-5bf59778d8-rrxtx_d0ece558-05e2-4e08-84e0-a314f68a453b/operator/0.log" Dec 13 09:48:32 crc kubenswrapper[4971]: I1213 09:48:32.997848 4971 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_ovn-operator-controller-manager-bf6d4f946-g2r89_9ed7aa8c-f56d-477d-91f5-ef1944e66d57/manager/0.log" Dec 13 09:48:33 crc kubenswrapper[4971]: I1213 09:48:33.006551 4971 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_openstack-operator-index-hbjwh_1a3059b7-5a8b-4357-8f6d-caa37f9adef2/registry-server/0.log" Dec 13 09:48:33 crc kubenswrapper[4971]: I1213 09:48:33.348661 4971 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_placement-operator-controller-manager-8665b56d78-hrflw_6b9f3caa-4280-43f3-8a05-14e72f09c384/manager/0.log" Dec 13 09:48:33 crc kubenswrapper[4971]: I1213 09:48:33.583982 4971 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_rabbitmq-cluster-operator-manager-668c99d594-pzcql_14e41caa-c89a-4ca3-98c3-d75114cd2c76/operator/0.log" Dec 13 09:48:33 crc kubenswrapper[4971]: I1213 09:48:33.714492 4971 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_swift-operator-controller-manager-5c6df8f9-rt5n5_bd48845c-5d55-4690-960b-ca319f596beb/manager/0.log" Dec 13 09:48:33 crc kubenswrapper[4971]: I1213 09:48:33.726757 4971 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_openstack-operator-controller-manager-56f6fbdf6-vzrkl_40f2727e-ef9d-4557-8619-b37174d00b67/manager/0.log" Dec 13 09:48:33 crc kubenswrapper[4971]: I1213 09:48:33.921607 4971 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_telemetry-operator-controller-manager-97d456b9-j4fpf_f111629c-eb0b-4294-8f85-d4b0b6500454/manager/0.log" Dec 13 09:48:33 crc kubenswrapper[4971]: I1213 09:48:33.992883 4971 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_test-operator-controller-manager-85c4449d86-s4v55_ccfc4136-eeb7-4fd8-a4d2-d09dd4a55802/manager/0.log" Dec 13 09:48:34 crc kubenswrapper[4971]: I1213 09:48:34.114241 4971 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_watcher-operator-controller-manager-55f78b7c4c-dwmgr_5bb0c1cd-9a8e-4c03-80c0-d9e66dae91ed/manager/0.log" Dec 13 09:48:43 crc kubenswrapper[4971]: I1213 09:48:43.778585 4971 scope.go:117] "RemoveContainer" containerID="80deb18ddf85834e136ab62ea93a1fa75315016d9f220eb217720a5e562f5e54" Dec 13 09:48:43 crc kubenswrapper[4971]: E1213 09:48:43.779408 4971 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-82xjz_openshift-machine-config-operator(e0c4af12-edda-41d7-9b44-f87396174ef2)\"" pod="openshift-machine-config-operator/machine-config-daemon-82xjz" podUID="e0c4af12-edda-41d7-9b44-f87396174ef2" Dec 13 09:48:56 crc kubenswrapper[4971]: I1213 09:48:56.769742 4971 scope.go:117] "RemoveContainer" containerID="80deb18ddf85834e136ab62ea93a1fa75315016d9f220eb217720a5e562f5e54" Dec 13 09:48:56 crc kubenswrapper[4971]: E1213 09:48:56.770724 4971 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-82xjz_openshift-machine-config-operator(e0c4af12-edda-41d7-9b44-f87396174ef2)\"" pod="openshift-machine-config-operator/machine-config-daemon-82xjz" podUID="e0c4af12-edda-41d7-9b44-f87396174ef2" Dec 13 09:49:07 crc kubenswrapper[4971]: I1213 09:49:07.191434 4971 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-machine-api_control-plane-machine-set-operator-78cbb6b69f-bgdbk_930fe38c-5ff7-45a8-a292-cd69d07e86f7/control-plane-machine-set-operator/0.log" Dec 13 09:49:07 crc kubenswrapper[4971]: I1213 09:49:07.466810 4971 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-machine-api_machine-api-operator-5694c8668f-t24dm_c45a314c-77da-43fc-b2ae-9de350540c2b/kube-rbac-proxy/0.log" Dec 13 09:49:07 crc kubenswrapper[4971]: I1213 09:49:07.467686 4971 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-machine-api_machine-api-operator-5694c8668f-t24dm_c45a314c-77da-43fc-b2ae-9de350540c2b/machine-api-operator/0.log" Dec 13 09:49:07 crc kubenswrapper[4971]: I1213 09:49:07.770198 4971 scope.go:117] "RemoveContainer" containerID="80deb18ddf85834e136ab62ea93a1fa75315016d9f220eb217720a5e562f5e54" Dec 13 09:49:07 crc kubenswrapper[4971]: E1213 09:49:07.770586 4971 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-82xjz_openshift-machine-config-operator(e0c4af12-edda-41d7-9b44-f87396174ef2)\"" pod="openshift-machine-config-operator/machine-config-daemon-82xjz" podUID="e0c4af12-edda-41d7-9b44-f87396174ef2" Dec 13 09:49:15 crc kubenswrapper[4971]: E1213 09:49:15.769025 4971 kubelet_pods.go:538] "Hostname for pod was too long, truncated it" podName="test-operator-logs-pod-horizontest-horizontest-tests-horizontest" hostnameMaxLen=63 truncatedHostname="test-operator-logs-pod-horizontest-horizontest-tests-horizontes" Dec 13 09:49:20 crc kubenswrapper[4971]: I1213 09:49:20.769613 4971 scope.go:117] "RemoveContainer" containerID="80deb18ddf85834e136ab62ea93a1fa75315016d9f220eb217720a5e562f5e54" Dec 13 09:49:20 crc kubenswrapper[4971]: E1213 09:49:20.770706 4971 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-82xjz_openshift-machine-config-operator(e0c4af12-edda-41d7-9b44-f87396174ef2)\"" pod="openshift-machine-config-operator/machine-config-daemon-82xjz" podUID="e0c4af12-edda-41d7-9b44-f87396174ef2" Dec 13 09:49:24 crc kubenswrapper[4971]: I1213 09:49:24.756902 4971 log.go:25] "Finished parsing log file" path="/var/log/pods/cert-manager_cert-manager-5b446d88c5-bmjtf_65bdb0e4-18e8-4db6-af21-341a748772e9/cert-manager-controller/0.log" Dec 13 09:49:24 crc kubenswrapper[4971]: I1213 09:49:24.926157 4971 log.go:25] "Finished parsing log file" path="/var/log/pods/cert-manager_cert-manager-cainjector-7f985d654d-dp572_fb8b0fc5-85d4-4df8-9b43-8a567a888b20/cert-manager-cainjector/0.log" Dec 13 09:49:24 crc kubenswrapper[4971]: I1213 09:49:24.998024 4971 log.go:25] "Finished parsing log file" path="/var/log/pods/cert-manager_cert-manager-webhook-5655c58dd6-qbbjr_974beed7-18ac-4d6b-8192-d844b50b7ad3/cert-manager-webhook/0.log" Dec 13 09:49:30 crc kubenswrapper[4971]: I1213 09:49:30.838580 4971 scope.go:117] "RemoveContainer" containerID="2a73ca6e689a48e12b02f5a59bd2ef8c20b256cad668fe51b8e585606302a17c" Dec 13 09:49:33 crc kubenswrapper[4971]: I1213 09:49:33.789216 4971 scope.go:117] "RemoveContainer" containerID="80deb18ddf85834e136ab62ea93a1fa75315016d9f220eb217720a5e562f5e54" Dec 13 09:49:33 crc kubenswrapper[4971]: E1213 09:49:33.790636 4971 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-82xjz_openshift-machine-config-operator(e0c4af12-edda-41d7-9b44-f87396174ef2)\"" pod="openshift-machine-config-operator/machine-config-daemon-82xjz" podUID="e0c4af12-edda-41d7-9b44-f87396174ef2" Dec 13 09:49:41 crc kubenswrapper[4971]: I1213 09:49:41.304144 4971 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-nmstate_nmstate-console-plugin-6ff7998486-hx58s_dfa2c127-31ee-40a0-814a-5af53e135ae8/nmstate-console-plugin/0.log" Dec 13 09:49:41 crc kubenswrapper[4971]: I1213 09:49:41.443098 4971 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-nmstate_nmstate-handler-5b4bx_d8139542-2862-4e1b-8e84-62a77e5bb7e5/nmstate-handler/0.log" Dec 13 09:49:41 crc kubenswrapper[4971]: I1213 09:49:41.585978 4971 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-nmstate_nmstate-metrics-7f7f7578db-xsb6k_f2968080-99c8-4320-9b39-4bcab4dad759/kube-rbac-proxy/0.log" Dec 13 09:49:41 crc kubenswrapper[4971]: I1213 09:49:41.612673 4971 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-nmstate_nmstate-metrics-7f7f7578db-xsb6k_f2968080-99c8-4320-9b39-4bcab4dad759/nmstate-metrics/0.log" Dec 13 09:49:41 crc kubenswrapper[4971]: I1213 09:49:41.818011 4971 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-nmstate_nmstate-operator-6769fb99d-n9v79_658a55da-1b10-45a1-b449-d8d4b8916242/nmstate-operator/0.log" Dec 13 09:49:41 crc kubenswrapper[4971]: I1213 09:49:41.841049 4971 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-nmstate_nmstate-webhook-f8fb84555-pwxvn_9e887c99-7f51-4b00-9571-f4d29c191054/nmstate-webhook/0.log" Dec 13 09:49:46 crc kubenswrapper[4971]: I1213 09:49:46.768620 4971 scope.go:117] "RemoveContainer" containerID="80deb18ddf85834e136ab62ea93a1fa75315016d9f220eb217720a5e562f5e54" Dec 13 09:49:46 crc kubenswrapper[4971]: E1213 09:49:46.769540 4971 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-82xjz_openshift-machine-config-operator(e0c4af12-edda-41d7-9b44-f87396174ef2)\"" pod="openshift-machine-config-operator/machine-config-daemon-82xjz" podUID="e0c4af12-edda-41d7-9b44-f87396174ef2" Dec 13 09:50:00 crc kubenswrapper[4971]: I1213 09:50:00.768759 4971 scope.go:117] "RemoveContainer" containerID="80deb18ddf85834e136ab62ea93a1fa75315016d9f220eb217720a5e562f5e54" Dec 13 09:50:00 crc kubenswrapper[4971]: E1213 09:50:00.769598 4971 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-82xjz_openshift-machine-config-operator(e0c4af12-edda-41d7-9b44-f87396174ef2)\"" pod="openshift-machine-config-operator/machine-config-daemon-82xjz" podUID="e0c4af12-edda-41d7-9b44-f87396174ef2" Dec 13 09:50:00 crc kubenswrapper[4971]: I1213 09:50:00.870748 4971 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_controller-5bddd4b946-2kfxb_7cc4bd2b-7f1a-46d0-bae1-dcd35dc513ad/kube-rbac-proxy/0.log" Dec 13 09:50:00 crc kubenswrapper[4971]: I1213 09:50:00.927596 4971 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_controller-5bddd4b946-2kfxb_7cc4bd2b-7f1a-46d0-bae1-dcd35dc513ad/controller/0.log" Dec 13 09:50:01 crc kubenswrapper[4971]: I1213 09:50:01.761383 4971 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-szzcz_1f567f13-d990-4035-92e4-64835af4aac8/cp-frr-files/0.log" Dec 13 09:50:01 crc kubenswrapper[4971]: I1213 09:50:01.934847 4971 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-szzcz_1f567f13-d990-4035-92e4-64835af4aac8/cp-frr-files/0.log" Dec 13 09:50:01 crc kubenswrapper[4971]: I1213 09:50:01.953479 4971 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-szzcz_1f567f13-d990-4035-92e4-64835af4aac8/cp-reloader/0.log" Dec 13 09:50:02 crc kubenswrapper[4971]: I1213 09:50:02.004369 4971 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-szzcz_1f567f13-d990-4035-92e4-64835af4aac8/cp-metrics/0.log" Dec 13 09:50:02 crc kubenswrapper[4971]: I1213 09:50:02.021959 4971 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-szzcz_1f567f13-d990-4035-92e4-64835af4aac8/cp-reloader/0.log" Dec 13 09:50:02 crc kubenswrapper[4971]: I1213 09:50:02.334845 4971 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-szzcz_1f567f13-d990-4035-92e4-64835af4aac8/cp-metrics/0.log" Dec 13 09:50:02 crc kubenswrapper[4971]: I1213 09:50:02.343365 4971 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-szzcz_1f567f13-d990-4035-92e4-64835af4aac8/cp-frr-files/0.log" Dec 13 09:50:02 crc kubenswrapper[4971]: I1213 09:50:02.352625 4971 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-szzcz_1f567f13-d990-4035-92e4-64835af4aac8/cp-metrics/0.log" Dec 13 09:50:02 crc kubenswrapper[4971]: I1213 09:50:02.611345 4971 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-szzcz_1f567f13-d990-4035-92e4-64835af4aac8/cp-reloader/0.log" Dec 13 09:50:02 crc kubenswrapper[4971]: I1213 09:50:02.614608 4971 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-szzcz_1f567f13-d990-4035-92e4-64835af4aac8/cp-frr-files/0.log" Dec 13 09:50:02 crc kubenswrapper[4971]: I1213 09:50:02.615266 4971 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-szzcz_1f567f13-d990-4035-92e4-64835af4aac8/cp-reloader/0.log" Dec 13 09:50:02 crc kubenswrapper[4971]: I1213 09:50:02.673426 4971 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-szzcz_1f567f13-d990-4035-92e4-64835af4aac8/cp-metrics/0.log" Dec 13 09:50:02 crc kubenswrapper[4971]: I1213 09:50:02.905423 4971 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-szzcz_1f567f13-d990-4035-92e4-64835af4aac8/frr-metrics/0.log" Dec 13 09:50:02 crc kubenswrapper[4971]: I1213 09:50:02.955086 4971 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-szzcz_1f567f13-d990-4035-92e4-64835af4aac8/controller/0.log" Dec 13 09:50:03 crc kubenswrapper[4971]: I1213 09:50:03.011801 4971 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-szzcz_1f567f13-d990-4035-92e4-64835af4aac8/kube-rbac-proxy/0.log" Dec 13 09:50:03 crc kubenswrapper[4971]: I1213 09:50:03.221721 4971 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-szzcz_1f567f13-d990-4035-92e4-64835af4aac8/kube-rbac-proxy-frr/0.log" Dec 13 09:50:03 crc kubenswrapper[4971]: I1213 09:50:03.329201 4971 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-szzcz_1f567f13-d990-4035-92e4-64835af4aac8/reloader/0.log" Dec 13 09:50:04 crc kubenswrapper[4971]: I1213 09:50:04.315284 4971 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-webhook-server-7784b6fcf-tkg6z_f093984d-c941-4805-a0c1-0926aa7d7f38/frr-k8s-webhook-server/0.log" Dec 13 09:50:04 crc kubenswrapper[4971]: I1213 09:50:04.480558 4971 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_metallb-operator-controller-manager-9d86ffb46-9kqjk_1a9addf5-4d0c-447d-8a9a-e6c15b9cdb3a/manager/0.log" Dec 13 09:50:04 crc kubenswrapper[4971]: I1213 09:50:04.614046 4971 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_metallb-operator-webhook-server-657bbb5cbf-njfvg_98e5a1fd-e920-4f5d-82c3-7450cfc0487a/webhook-server/0.log" Dec 13 09:50:04 crc kubenswrapper[4971]: I1213 09:50:04.919009 4971 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_speaker-v229x_c0ee70f0-8158-4fec-bae8-7cd72bdef461/kube-rbac-proxy/0.log" Dec 13 09:50:05 crc kubenswrapper[4971]: I1213 09:50:05.439951 4971 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_speaker-v229x_c0ee70f0-8158-4fec-bae8-7cd72bdef461/speaker/0.log" Dec 13 09:50:05 crc kubenswrapper[4971]: I1213 09:50:05.583485 4971 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-szzcz_1f567f13-d990-4035-92e4-64835af4aac8/frr/0.log" Dec 13 09:50:11 crc kubenswrapper[4971]: I1213 09:50:11.769645 4971 scope.go:117] "RemoveContainer" containerID="80deb18ddf85834e136ab62ea93a1fa75315016d9f220eb217720a5e562f5e54" Dec 13 09:50:11 crc kubenswrapper[4971]: E1213 09:50:11.770663 4971 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-82xjz_openshift-machine-config-operator(e0c4af12-edda-41d7-9b44-f87396174ef2)\"" pod="openshift-machine-config-operator/machine-config-daemon-82xjz" podUID="e0c4af12-edda-41d7-9b44-f87396174ef2" Dec 13 09:50:19 crc kubenswrapper[4971]: I1213 09:50:19.072039 4971 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_5b7fccbebf0e22d2dd769066fa7aaa90fd620c5db34f2af6c91e4319d4gv4b7_75227ae7-8658-447d-90c5-b0619fe75d8f/util/0.log" Dec 13 09:50:19 crc kubenswrapper[4971]: I1213 09:50:19.278651 4971 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_5b7fccbebf0e22d2dd769066fa7aaa90fd620c5db34f2af6c91e4319d4gv4b7_75227ae7-8658-447d-90c5-b0619fe75d8f/pull/0.log" Dec 13 09:50:19 crc kubenswrapper[4971]: I1213 09:50:19.283512 4971 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_5b7fccbebf0e22d2dd769066fa7aaa90fd620c5db34f2af6c91e4319d4gv4b7_75227ae7-8658-447d-90c5-b0619fe75d8f/util/0.log" Dec 13 09:50:19 crc kubenswrapper[4971]: I1213 09:50:19.319267 4971 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_5b7fccbebf0e22d2dd769066fa7aaa90fd620c5db34f2af6c91e4319d4gv4b7_75227ae7-8658-447d-90c5-b0619fe75d8f/pull/0.log" Dec 13 09:50:19 crc kubenswrapper[4971]: I1213 09:50:19.511593 4971 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_5b7fccbebf0e22d2dd769066fa7aaa90fd620c5db34f2af6c91e4319d4gv4b7_75227ae7-8658-447d-90c5-b0619fe75d8f/util/0.log" Dec 13 09:50:19 crc kubenswrapper[4971]: I1213 09:50:19.571008 4971 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_5b7fccbebf0e22d2dd769066fa7aaa90fd620c5db34f2af6c91e4319d4gv4b7_75227ae7-8658-447d-90c5-b0619fe75d8f/extract/0.log" Dec 13 09:50:19 crc kubenswrapper[4971]: I1213 09:50:19.593002 4971 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_5b7fccbebf0e22d2dd769066fa7aaa90fd620c5db34f2af6c91e4319d4gv4b7_75227ae7-8658-447d-90c5-b0619fe75d8f/pull/0.log" Dec 13 09:50:19 crc kubenswrapper[4971]: I1213 09:50:19.753790 4971 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_98085b0df3808ebec39f9f9529f737144fe2dbcdaa4f334014817c0fa89s2m9_c94817e4-313a-4e9a-a0d9-becb8044d4a8/util/0.log" Dec 13 09:50:19 crc kubenswrapper[4971]: I1213 09:50:19.944002 4971 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_98085b0df3808ebec39f9f9529f737144fe2dbcdaa4f334014817c0fa89s2m9_c94817e4-313a-4e9a-a0d9-becb8044d4a8/util/0.log" Dec 13 09:50:19 crc kubenswrapper[4971]: I1213 09:50:19.958654 4971 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_98085b0df3808ebec39f9f9529f737144fe2dbcdaa4f334014817c0fa89s2m9_c94817e4-313a-4e9a-a0d9-becb8044d4a8/pull/0.log" Dec 13 09:50:19 crc kubenswrapper[4971]: I1213 09:50:19.959167 4971 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_98085b0df3808ebec39f9f9529f737144fe2dbcdaa4f334014817c0fa89s2m9_c94817e4-313a-4e9a-a0d9-becb8044d4a8/pull/0.log" Dec 13 09:50:20 crc kubenswrapper[4971]: I1213 09:50:20.120683 4971 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_98085b0df3808ebec39f9f9529f737144fe2dbcdaa4f334014817c0fa89s2m9_c94817e4-313a-4e9a-a0d9-becb8044d4a8/util/0.log" Dec 13 09:50:20 crc kubenswrapper[4971]: I1213 09:50:20.135946 4971 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_98085b0df3808ebec39f9f9529f737144fe2dbcdaa4f334014817c0fa89s2m9_c94817e4-313a-4e9a-a0d9-becb8044d4a8/pull/0.log" Dec 13 09:50:20 crc kubenswrapper[4971]: I1213 09:50:20.175200 4971 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_98085b0df3808ebec39f9f9529f737144fe2dbcdaa4f334014817c0fa89s2m9_c94817e4-313a-4e9a-a0d9-becb8044d4a8/extract/0.log" Dec 13 09:50:20 crc kubenswrapper[4971]: I1213 09:50:20.417112 4971 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_certified-operators-szqz6_33212ad2-538d-4059-bc1f-e98a51c93b62/extract-utilities/0.log" Dec 13 09:50:20 crc kubenswrapper[4971]: I1213 09:50:20.567035 4971 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_certified-operators-szqz6_33212ad2-538d-4059-bc1f-e98a51c93b62/extract-utilities/0.log" Dec 13 09:50:20 crc kubenswrapper[4971]: I1213 09:50:20.584688 4971 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_certified-operators-szqz6_33212ad2-538d-4059-bc1f-e98a51c93b62/extract-content/0.log" Dec 13 09:50:20 crc kubenswrapper[4971]: I1213 09:50:20.644504 4971 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_certified-operators-szqz6_33212ad2-538d-4059-bc1f-e98a51c93b62/extract-content/0.log" Dec 13 09:50:20 crc kubenswrapper[4971]: I1213 09:50:20.810742 4971 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_certified-operators-szqz6_33212ad2-538d-4059-bc1f-e98a51c93b62/extract-utilities/0.log" Dec 13 09:50:20 crc kubenswrapper[4971]: I1213 09:50:20.821923 4971 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_certified-operators-szqz6_33212ad2-538d-4059-bc1f-e98a51c93b62/extract-content/0.log" Dec 13 09:50:21 crc kubenswrapper[4971]: I1213 09:50:21.095712 4971 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_community-operators-kzw5r_b2475f4e-bfe7-4678-8779-2b56c189b3b0/extract-utilities/0.log" Dec 13 09:50:21 crc kubenswrapper[4971]: I1213 09:50:21.311566 4971 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_community-operators-kzw5r_b2475f4e-bfe7-4678-8779-2b56c189b3b0/extract-utilities/0.log" Dec 13 09:50:21 crc kubenswrapper[4971]: I1213 09:50:21.424410 4971 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_community-operators-kzw5r_b2475f4e-bfe7-4678-8779-2b56c189b3b0/extract-content/0.log" Dec 13 09:50:21 crc kubenswrapper[4971]: I1213 09:50:21.479223 4971 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_community-operators-kzw5r_b2475f4e-bfe7-4678-8779-2b56c189b3b0/extract-content/0.log" Dec 13 09:50:21 crc kubenswrapper[4971]: I1213 09:50:21.684370 4971 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_community-operators-kzw5r_b2475f4e-bfe7-4678-8779-2b56c189b3b0/extract-utilities/0.log" Dec 13 09:50:21 crc kubenswrapper[4971]: I1213 09:50:21.765276 4971 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_community-operators-kzw5r_b2475f4e-bfe7-4678-8779-2b56c189b3b0/extract-content/0.log" Dec 13 09:50:22 crc kubenswrapper[4971]: I1213 09:50:22.061489 4971 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_marketplace-operator-79b997595-w4r9w_f8802833-bb84-4479-be26-f0fae563f1b0/marketplace-operator/0.log" Dec 13 09:50:22 crc kubenswrapper[4971]: I1213 09:50:22.345681 4971 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-marketplace-hvzhk_4b52f930-cf4c-4a97-b838-9afc9e3d2b31/extract-utilities/0.log" Dec 13 09:50:22 crc kubenswrapper[4971]: I1213 09:50:22.665318 4971 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-marketplace-hvzhk_4b52f930-cf4c-4a97-b838-9afc9e3d2b31/extract-utilities/0.log" Dec 13 09:50:22 crc kubenswrapper[4971]: I1213 09:50:22.670847 4971 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-marketplace-hvzhk_4b52f930-cf4c-4a97-b838-9afc9e3d2b31/extract-content/0.log" Dec 13 09:50:22 crc kubenswrapper[4971]: I1213 09:50:22.785934 4971 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_certified-operators-szqz6_33212ad2-538d-4059-bc1f-e98a51c93b62/registry-server/0.log" Dec 13 09:50:22 crc kubenswrapper[4971]: I1213 09:50:22.878770 4971 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-marketplace-hvzhk_4b52f930-cf4c-4a97-b838-9afc9e3d2b31/extract-content/0.log" Dec 13 09:50:23 crc kubenswrapper[4971]: I1213 09:50:23.170652 4971 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-marketplace-hvzhk_4b52f930-cf4c-4a97-b838-9afc9e3d2b31/extract-utilities/0.log" Dec 13 09:50:23 crc kubenswrapper[4971]: I1213 09:50:23.251236 4971 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-marketplace-hvzhk_4b52f930-cf4c-4a97-b838-9afc9e3d2b31/extract-content/0.log" Dec 13 09:50:23 crc kubenswrapper[4971]: I1213 09:50:23.498215 4971 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_community-operators-kzw5r_b2475f4e-bfe7-4678-8779-2b56c189b3b0/registry-server/0.log" Dec 13 09:50:23 crc kubenswrapper[4971]: I1213 09:50:23.511058 4971 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-operators-cvcsq_6fa06763-60e2-4bc5-91f7-58ded0345c6a/extract-utilities/0.log" Dec 13 09:50:23 crc kubenswrapper[4971]: I1213 09:50:23.754274 4971 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-marketplace-hvzhk_4b52f930-cf4c-4a97-b838-9afc9e3d2b31/registry-server/0.log" Dec 13 09:50:23 crc kubenswrapper[4971]: I1213 09:50:23.777558 4971 scope.go:117] "RemoveContainer" containerID="80deb18ddf85834e136ab62ea93a1fa75315016d9f220eb217720a5e562f5e54" Dec 13 09:50:23 crc kubenswrapper[4971]: E1213 09:50:23.778019 4971 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-82xjz_openshift-machine-config-operator(e0c4af12-edda-41d7-9b44-f87396174ef2)\"" pod="openshift-machine-config-operator/machine-config-daemon-82xjz" podUID="e0c4af12-edda-41d7-9b44-f87396174ef2" Dec 13 09:50:23 crc kubenswrapper[4971]: I1213 09:50:23.847014 4971 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-operators-cvcsq_6fa06763-60e2-4bc5-91f7-58ded0345c6a/extract-content/0.log" Dec 13 09:50:23 crc kubenswrapper[4971]: I1213 09:50:23.871183 4971 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-operators-cvcsq_6fa06763-60e2-4bc5-91f7-58ded0345c6a/extract-content/0.log" Dec 13 09:50:23 crc kubenswrapper[4971]: I1213 09:50:23.875987 4971 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-operators-cvcsq_6fa06763-60e2-4bc5-91f7-58ded0345c6a/extract-utilities/0.log" Dec 13 09:50:24 crc kubenswrapper[4971]: I1213 09:50:24.107772 4971 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-operators-cvcsq_6fa06763-60e2-4bc5-91f7-58ded0345c6a/extract-utilities/0.log" Dec 13 09:50:24 crc kubenswrapper[4971]: I1213 09:50:24.124819 4971 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-operators-cvcsq_6fa06763-60e2-4bc5-91f7-58ded0345c6a/extract-content/0.log" Dec 13 09:50:25 crc kubenswrapper[4971]: I1213 09:50:25.507367 4971 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-operators-cvcsq_6fa06763-60e2-4bc5-91f7-58ded0345c6a/registry-server/0.log" Dec 13 09:50:26 crc kubenswrapper[4971]: E1213 09:50:26.768946 4971 kubelet_pods.go:538] "Hostname for pod was too long, truncated it" podName="test-operator-logs-pod-horizontest-horizontest-tests-horizontest" hostnameMaxLen=63 truncatedHostname="test-operator-logs-pod-horizontest-horizontest-tests-horizontes" Dec 13 09:50:36 crc kubenswrapper[4971]: I1213 09:50:36.769035 4971 scope.go:117] "RemoveContainer" containerID="80deb18ddf85834e136ab62ea93a1fa75315016d9f220eb217720a5e562f5e54" Dec 13 09:50:36 crc kubenswrapper[4971]: E1213 09:50:36.770400 4971 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-82xjz_openshift-machine-config-operator(e0c4af12-edda-41d7-9b44-f87396174ef2)\"" pod="openshift-machine-config-operator/machine-config-daemon-82xjz" podUID="e0c4af12-edda-41d7-9b44-f87396174ef2" Dec 13 09:50:50 crc kubenswrapper[4971]: I1213 09:50:50.768920 4971 scope.go:117] "RemoveContainer" containerID="80deb18ddf85834e136ab62ea93a1fa75315016d9f220eb217720a5e562f5e54" Dec 13 09:50:50 crc kubenswrapper[4971]: E1213 09:50:50.769776 4971 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-82xjz_openshift-machine-config-operator(e0c4af12-edda-41d7-9b44-f87396174ef2)\"" pod="openshift-machine-config-operator/machine-config-daemon-82xjz" podUID="e0c4af12-edda-41d7-9b44-f87396174ef2" Dec 13 09:51:01 crc kubenswrapper[4971]: I1213 09:51:01.769923 4971 scope.go:117] "RemoveContainer" containerID="80deb18ddf85834e136ab62ea93a1fa75315016d9f220eb217720a5e562f5e54" Dec 13 09:51:01 crc kubenswrapper[4971]: E1213 09:51:01.771457 4971 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-82xjz_openshift-machine-config-operator(e0c4af12-edda-41d7-9b44-f87396174ef2)\"" pod="openshift-machine-config-operator/machine-config-daemon-82xjz" podUID="e0c4af12-edda-41d7-9b44-f87396174ef2" Dec 13 09:51:02 crc kubenswrapper[4971]: I1213 09:51:02.101105 4971 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/community-operators-gxp7t"] Dec 13 09:51:02 crc kubenswrapper[4971]: E1213 09:51:02.101750 4971 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f6295795-bcc0-4966-b872-78a938ff1ec1" containerName="container-00" Dec 13 09:51:02 crc kubenswrapper[4971]: I1213 09:51:02.101778 4971 state_mem.go:107] "Deleted CPUSet assignment" podUID="f6295795-bcc0-4966-b872-78a938ff1ec1" containerName="container-00" Dec 13 09:51:02 crc kubenswrapper[4971]: I1213 09:51:02.102062 4971 memory_manager.go:354] "RemoveStaleState removing state" podUID="f6295795-bcc0-4966-b872-78a938ff1ec1" containerName="container-00" Dec 13 09:51:02 crc kubenswrapper[4971]: I1213 09:51:02.104044 4971 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-gxp7t" Dec 13 09:51:02 crc kubenswrapper[4971]: I1213 09:51:02.114108 4971 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-gxp7t"] Dec 13 09:51:02 crc kubenswrapper[4971]: I1213 09:51:02.241380 4971 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/761521dd-cee4-4f2d-add4-8fc4fc966112-utilities\") pod \"community-operators-gxp7t\" (UID: \"761521dd-cee4-4f2d-add4-8fc4fc966112\") " pod="openshift-marketplace/community-operators-gxp7t" Dec 13 09:51:02 crc kubenswrapper[4971]: I1213 09:51:02.241482 4971 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/761521dd-cee4-4f2d-add4-8fc4fc966112-catalog-content\") pod \"community-operators-gxp7t\" (UID: \"761521dd-cee4-4f2d-add4-8fc4fc966112\") " pod="openshift-marketplace/community-operators-gxp7t" Dec 13 09:51:02 crc kubenswrapper[4971]: I1213 09:51:02.241586 4971 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-qfwjh\" (UniqueName: \"kubernetes.io/projected/761521dd-cee4-4f2d-add4-8fc4fc966112-kube-api-access-qfwjh\") pod \"community-operators-gxp7t\" (UID: \"761521dd-cee4-4f2d-add4-8fc4fc966112\") " pod="openshift-marketplace/community-operators-gxp7t" Dec 13 09:51:02 crc kubenswrapper[4971]: I1213 09:51:02.344459 4971 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-qfwjh\" (UniqueName: \"kubernetes.io/projected/761521dd-cee4-4f2d-add4-8fc4fc966112-kube-api-access-qfwjh\") pod \"community-operators-gxp7t\" (UID: \"761521dd-cee4-4f2d-add4-8fc4fc966112\") " pod="openshift-marketplace/community-operators-gxp7t" Dec 13 09:51:02 crc kubenswrapper[4971]: I1213 09:51:02.345477 4971 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/761521dd-cee4-4f2d-add4-8fc4fc966112-utilities\") pod \"community-operators-gxp7t\" (UID: \"761521dd-cee4-4f2d-add4-8fc4fc966112\") " pod="openshift-marketplace/community-operators-gxp7t" Dec 13 09:51:02 crc kubenswrapper[4971]: I1213 09:51:02.346084 4971 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/761521dd-cee4-4f2d-add4-8fc4fc966112-utilities\") pod \"community-operators-gxp7t\" (UID: \"761521dd-cee4-4f2d-add4-8fc4fc966112\") " pod="openshift-marketplace/community-operators-gxp7t" Dec 13 09:51:02 crc kubenswrapper[4971]: I1213 09:51:02.346285 4971 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/761521dd-cee4-4f2d-add4-8fc4fc966112-catalog-content\") pod \"community-operators-gxp7t\" (UID: \"761521dd-cee4-4f2d-add4-8fc4fc966112\") " pod="openshift-marketplace/community-operators-gxp7t" Dec 13 09:51:02 crc kubenswrapper[4971]: I1213 09:51:02.346724 4971 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/761521dd-cee4-4f2d-add4-8fc4fc966112-catalog-content\") pod \"community-operators-gxp7t\" (UID: \"761521dd-cee4-4f2d-add4-8fc4fc966112\") " pod="openshift-marketplace/community-operators-gxp7t" Dec 13 09:51:02 crc kubenswrapper[4971]: I1213 09:51:02.376828 4971 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-qfwjh\" (UniqueName: \"kubernetes.io/projected/761521dd-cee4-4f2d-add4-8fc4fc966112-kube-api-access-qfwjh\") pod \"community-operators-gxp7t\" (UID: \"761521dd-cee4-4f2d-add4-8fc4fc966112\") " pod="openshift-marketplace/community-operators-gxp7t" Dec 13 09:51:02 crc kubenswrapper[4971]: I1213 09:51:02.428402 4971 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-gxp7t" Dec 13 09:51:03 crc kubenswrapper[4971]: I1213 09:51:03.165839 4971 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-gxp7t"] Dec 13 09:51:03 crc kubenswrapper[4971]: E1213 09:51:03.178455 4971 upgradeaware.go:427] Error proxying data from client to backend: readfrom tcp 38.102.83.195:58350->38.102.83.195:40419: write tcp 38.102.83.195:58350->38.102.83.195:40419: write: broken pipe Dec 13 09:51:04 crc kubenswrapper[4971]: I1213 09:51:04.563452 4971 generic.go:334] "Generic (PLEG): container finished" podID="761521dd-cee4-4f2d-add4-8fc4fc966112" containerID="d89c8cbaa3310d58e42f34e288d11b68f75a40f551022733c4fcb3c0805c7448" exitCode=0 Dec 13 09:51:04 crc kubenswrapper[4971]: I1213 09:51:04.563535 4971 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-gxp7t" event={"ID":"761521dd-cee4-4f2d-add4-8fc4fc966112","Type":"ContainerDied","Data":"d89c8cbaa3310d58e42f34e288d11b68f75a40f551022733c4fcb3c0805c7448"} Dec 13 09:51:04 crc kubenswrapper[4971]: I1213 09:51:04.564156 4971 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-gxp7t" event={"ID":"761521dd-cee4-4f2d-add4-8fc4fc966112","Type":"ContainerStarted","Data":"92698845ee69feeaad008ddfad894e6996e275c9138139289ff7d668561f11d2"} Dec 13 09:51:04 crc kubenswrapper[4971]: I1213 09:51:04.570352 4971 provider.go:102] Refreshing cache for provider: *credentialprovider.defaultDockerConfigProvider Dec 13 09:51:06 crc kubenswrapper[4971]: I1213 09:51:06.595050 4971 generic.go:334] "Generic (PLEG): container finished" podID="761521dd-cee4-4f2d-add4-8fc4fc966112" containerID="1bddbe848df9a00cd96cd861ce925d609ad3e1c08492e17e75540673406b01e9" exitCode=0 Dec 13 09:51:06 crc kubenswrapper[4971]: I1213 09:51:06.596004 4971 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-gxp7t" event={"ID":"761521dd-cee4-4f2d-add4-8fc4fc966112","Type":"ContainerDied","Data":"1bddbe848df9a00cd96cd861ce925d609ad3e1c08492e17e75540673406b01e9"} Dec 13 09:51:07 crc kubenswrapper[4971]: I1213 09:51:07.610095 4971 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-gxp7t" event={"ID":"761521dd-cee4-4f2d-add4-8fc4fc966112","Type":"ContainerStarted","Data":"733fd5e3452e597574867132276373227029b4935604b337e657b268f6ee64d1"} Dec 13 09:51:09 crc kubenswrapper[4971]: E1213 09:51:09.406867 4971 upgradeaware.go:441] Error proxying data from backend to client: writeto tcp 38.102.83.195:39806->38.102.83.195:40419: read tcp 38.102.83.195:39806->38.102.83.195:40419: read: connection reset by peer Dec 13 09:51:12 crc kubenswrapper[4971]: I1213 09:51:12.445104 4971 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/community-operators-gxp7t" Dec 13 09:51:12 crc kubenswrapper[4971]: I1213 09:51:12.445738 4971 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/community-operators-gxp7t" Dec 13 09:51:12 crc kubenswrapper[4971]: I1213 09:51:12.499278 4971 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/community-operators-gxp7t" Dec 13 09:51:12 crc kubenswrapper[4971]: I1213 09:51:12.525163 4971 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/community-operators-gxp7t" podStartSLOduration=8.043468365 podStartE2EDuration="10.525144994s" podCreationTimestamp="2025-12-13 09:51:02 +0000 UTC" firstStartedPulling="2025-12-13 09:51:04.570024259 +0000 UTC m=+10921.174433707" lastFinishedPulling="2025-12-13 09:51:07.051700888 +0000 UTC m=+10923.656110336" observedRunningTime="2025-12-13 09:51:07.640827372 +0000 UTC m=+10924.245236820" watchObservedRunningTime="2025-12-13 09:51:12.525144994 +0000 UTC m=+10929.129554432" Dec 13 09:51:12 crc kubenswrapper[4971]: I1213 09:51:12.667947 4971 generic.go:334] "Generic (PLEG): container finished" podID="b2ac511e-e641-4ee8-aca7-6054195a75d8" containerID="423189d833cf0d72b52f4d07c31c19ad439ce4ef0db37aa7024772e15119e377" exitCode=0 Dec 13 09:51:12 crc kubenswrapper[4971]: I1213 09:51:12.668064 4971 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-69gc8/must-gather-rgtmv" event={"ID":"b2ac511e-e641-4ee8-aca7-6054195a75d8","Type":"ContainerDied","Data":"423189d833cf0d72b52f4d07c31c19ad439ce4ef0db37aa7024772e15119e377"} Dec 13 09:51:12 crc kubenswrapper[4971]: I1213 09:51:12.669080 4971 scope.go:117] "RemoveContainer" containerID="423189d833cf0d72b52f4d07c31c19ad439ce4ef0db37aa7024772e15119e377" Dec 13 09:51:12 crc kubenswrapper[4971]: I1213 09:51:12.714601 4971 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/community-operators-gxp7t" Dec 13 09:51:12 crc kubenswrapper[4971]: I1213 09:51:12.732465 4971 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-must-gather-69gc8_must-gather-rgtmv_b2ac511e-e641-4ee8-aca7-6054195a75d8/gather/0.log" Dec 13 09:51:12 crc kubenswrapper[4971]: I1213 09:51:12.773421 4971 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-gxp7t"] Dec 13 09:51:13 crc kubenswrapper[4971]: I1213 09:51:13.777149 4971 scope.go:117] "RemoveContainer" containerID="80deb18ddf85834e136ab62ea93a1fa75315016d9f220eb217720a5e562f5e54" Dec 13 09:51:13 crc kubenswrapper[4971]: E1213 09:51:13.778393 4971 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-82xjz_openshift-machine-config-operator(e0c4af12-edda-41d7-9b44-f87396174ef2)\"" pod="openshift-machine-config-operator/machine-config-daemon-82xjz" podUID="e0c4af12-edda-41d7-9b44-f87396174ef2" Dec 13 09:51:14 crc kubenswrapper[4971]: I1213 09:51:14.685736 4971 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/community-operators-gxp7t" podUID="761521dd-cee4-4f2d-add4-8fc4fc966112" containerName="registry-server" containerID="cri-o://733fd5e3452e597574867132276373227029b4935604b337e657b268f6ee64d1" gracePeriod=2 Dec 13 09:51:15 crc kubenswrapper[4971]: I1213 09:51:15.710600 4971 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-gxp7t" Dec 13 09:51:15 crc kubenswrapper[4971]: I1213 09:51:15.710831 4971 generic.go:334] "Generic (PLEG): container finished" podID="761521dd-cee4-4f2d-add4-8fc4fc966112" containerID="733fd5e3452e597574867132276373227029b4935604b337e657b268f6ee64d1" exitCode=0 Dec 13 09:51:15 crc kubenswrapper[4971]: I1213 09:51:15.710867 4971 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-gxp7t" event={"ID":"761521dd-cee4-4f2d-add4-8fc4fc966112","Type":"ContainerDied","Data":"733fd5e3452e597574867132276373227029b4935604b337e657b268f6ee64d1"} Dec 13 09:51:15 crc kubenswrapper[4971]: I1213 09:51:15.711212 4971 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-gxp7t" event={"ID":"761521dd-cee4-4f2d-add4-8fc4fc966112","Type":"ContainerDied","Data":"92698845ee69feeaad008ddfad894e6996e275c9138139289ff7d668561f11d2"} Dec 13 09:51:15 crc kubenswrapper[4971]: I1213 09:51:15.711251 4971 scope.go:117] "RemoveContainer" containerID="733fd5e3452e597574867132276373227029b4935604b337e657b268f6ee64d1" Dec 13 09:51:15 crc kubenswrapper[4971]: I1213 09:51:15.750995 4971 scope.go:117] "RemoveContainer" containerID="1bddbe848df9a00cd96cd861ce925d609ad3e1c08492e17e75540673406b01e9" Dec 13 09:51:15 crc kubenswrapper[4971]: I1213 09:51:15.754899 4971 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/761521dd-cee4-4f2d-add4-8fc4fc966112-catalog-content\") pod \"761521dd-cee4-4f2d-add4-8fc4fc966112\" (UID: \"761521dd-cee4-4f2d-add4-8fc4fc966112\") " Dec 13 09:51:15 crc kubenswrapper[4971]: I1213 09:51:15.754953 4971 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/761521dd-cee4-4f2d-add4-8fc4fc966112-utilities\") pod \"761521dd-cee4-4f2d-add4-8fc4fc966112\" (UID: \"761521dd-cee4-4f2d-add4-8fc4fc966112\") " Dec 13 09:51:15 crc kubenswrapper[4971]: I1213 09:51:15.755026 4971 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-qfwjh\" (UniqueName: \"kubernetes.io/projected/761521dd-cee4-4f2d-add4-8fc4fc966112-kube-api-access-qfwjh\") pod \"761521dd-cee4-4f2d-add4-8fc4fc966112\" (UID: \"761521dd-cee4-4f2d-add4-8fc4fc966112\") " Dec 13 09:51:15 crc kubenswrapper[4971]: I1213 09:51:15.756152 4971 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/761521dd-cee4-4f2d-add4-8fc4fc966112-utilities" (OuterVolumeSpecName: "utilities") pod "761521dd-cee4-4f2d-add4-8fc4fc966112" (UID: "761521dd-cee4-4f2d-add4-8fc4fc966112"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 13 09:51:15 crc kubenswrapper[4971]: I1213 09:51:15.781063 4971 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/761521dd-cee4-4f2d-add4-8fc4fc966112-kube-api-access-qfwjh" (OuterVolumeSpecName: "kube-api-access-qfwjh") pod "761521dd-cee4-4f2d-add4-8fc4fc966112" (UID: "761521dd-cee4-4f2d-add4-8fc4fc966112"). InnerVolumeSpecName "kube-api-access-qfwjh". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 13 09:51:15 crc kubenswrapper[4971]: I1213 09:51:15.786792 4971 scope.go:117] "RemoveContainer" containerID="d89c8cbaa3310d58e42f34e288d11b68f75a40f551022733c4fcb3c0805c7448" Dec 13 09:51:15 crc kubenswrapper[4971]: I1213 09:51:15.808539 4971 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/761521dd-cee4-4f2d-add4-8fc4fc966112-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "761521dd-cee4-4f2d-add4-8fc4fc966112" (UID: "761521dd-cee4-4f2d-add4-8fc4fc966112"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 13 09:51:15 crc kubenswrapper[4971]: I1213 09:51:15.858672 4971 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/761521dd-cee4-4f2d-add4-8fc4fc966112-utilities\") on node \"crc\" DevicePath \"\"" Dec 13 09:51:15 crc kubenswrapper[4971]: I1213 09:51:15.858714 4971 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-qfwjh\" (UniqueName: \"kubernetes.io/projected/761521dd-cee4-4f2d-add4-8fc4fc966112-kube-api-access-qfwjh\") on node \"crc\" DevicePath \"\"" Dec 13 09:51:15 crc kubenswrapper[4971]: I1213 09:51:15.858727 4971 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/761521dd-cee4-4f2d-add4-8fc4fc966112-catalog-content\") on node \"crc\" DevicePath \"\"" Dec 13 09:51:15 crc kubenswrapper[4971]: I1213 09:51:15.879465 4971 scope.go:117] "RemoveContainer" containerID="733fd5e3452e597574867132276373227029b4935604b337e657b268f6ee64d1" Dec 13 09:51:15 crc kubenswrapper[4971]: E1213 09:51:15.879924 4971 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"733fd5e3452e597574867132276373227029b4935604b337e657b268f6ee64d1\": container with ID starting with 733fd5e3452e597574867132276373227029b4935604b337e657b268f6ee64d1 not found: ID does not exist" containerID="733fd5e3452e597574867132276373227029b4935604b337e657b268f6ee64d1" Dec 13 09:51:15 crc kubenswrapper[4971]: I1213 09:51:15.879959 4971 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"733fd5e3452e597574867132276373227029b4935604b337e657b268f6ee64d1"} err="failed to get container status \"733fd5e3452e597574867132276373227029b4935604b337e657b268f6ee64d1\": rpc error: code = NotFound desc = could not find container \"733fd5e3452e597574867132276373227029b4935604b337e657b268f6ee64d1\": container with ID starting with 733fd5e3452e597574867132276373227029b4935604b337e657b268f6ee64d1 not found: ID does not exist" Dec 13 09:51:15 crc kubenswrapper[4971]: I1213 09:51:15.879979 4971 scope.go:117] "RemoveContainer" containerID="1bddbe848df9a00cd96cd861ce925d609ad3e1c08492e17e75540673406b01e9" Dec 13 09:51:15 crc kubenswrapper[4971]: E1213 09:51:15.880179 4971 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"1bddbe848df9a00cd96cd861ce925d609ad3e1c08492e17e75540673406b01e9\": container with ID starting with 1bddbe848df9a00cd96cd861ce925d609ad3e1c08492e17e75540673406b01e9 not found: ID does not exist" containerID="1bddbe848df9a00cd96cd861ce925d609ad3e1c08492e17e75540673406b01e9" Dec 13 09:51:15 crc kubenswrapper[4971]: I1213 09:51:15.880199 4971 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"1bddbe848df9a00cd96cd861ce925d609ad3e1c08492e17e75540673406b01e9"} err="failed to get container status \"1bddbe848df9a00cd96cd861ce925d609ad3e1c08492e17e75540673406b01e9\": rpc error: code = NotFound desc = could not find container \"1bddbe848df9a00cd96cd861ce925d609ad3e1c08492e17e75540673406b01e9\": container with ID starting with 1bddbe848df9a00cd96cd861ce925d609ad3e1c08492e17e75540673406b01e9 not found: ID does not exist" Dec 13 09:51:15 crc kubenswrapper[4971]: I1213 09:51:15.880210 4971 scope.go:117] "RemoveContainer" containerID="d89c8cbaa3310d58e42f34e288d11b68f75a40f551022733c4fcb3c0805c7448" Dec 13 09:51:15 crc kubenswrapper[4971]: E1213 09:51:15.880560 4971 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"d89c8cbaa3310d58e42f34e288d11b68f75a40f551022733c4fcb3c0805c7448\": container with ID starting with d89c8cbaa3310d58e42f34e288d11b68f75a40f551022733c4fcb3c0805c7448 not found: ID does not exist" containerID="d89c8cbaa3310d58e42f34e288d11b68f75a40f551022733c4fcb3c0805c7448" Dec 13 09:51:15 crc kubenswrapper[4971]: I1213 09:51:15.880622 4971 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"d89c8cbaa3310d58e42f34e288d11b68f75a40f551022733c4fcb3c0805c7448"} err="failed to get container status \"d89c8cbaa3310d58e42f34e288d11b68f75a40f551022733c4fcb3c0805c7448\": rpc error: code = NotFound desc = could not find container \"d89c8cbaa3310d58e42f34e288d11b68f75a40f551022733c4fcb3c0805c7448\": container with ID starting with d89c8cbaa3310d58e42f34e288d11b68f75a40f551022733c4fcb3c0805c7448 not found: ID does not exist" Dec 13 09:51:16 crc kubenswrapper[4971]: I1213 09:51:16.726228 4971 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-gxp7t" Dec 13 09:51:16 crc kubenswrapper[4971]: I1213 09:51:16.769271 4971 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-gxp7t"] Dec 13 09:51:16 crc kubenswrapper[4971]: I1213 09:51:16.788247 4971 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/community-operators-gxp7t"] Dec 13 09:51:17 crc kubenswrapper[4971]: I1213 09:51:17.782013 4971 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="761521dd-cee4-4f2d-add4-8fc4fc966112" path="/var/lib/kubelet/pods/761521dd-cee4-4f2d-add4-8fc4fc966112/volumes" Dec 13 09:51:22 crc kubenswrapper[4971]: I1213 09:51:22.077141 4971 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-must-gather-69gc8/must-gather-rgtmv"] Dec 13 09:51:22 crc kubenswrapper[4971]: I1213 09:51:22.079006 4971 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-must-gather-69gc8/must-gather-rgtmv" podUID="b2ac511e-e641-4ee8-aca7-6054195a75d8" containerName="copy" containerID="cri-o://244712079a41bb332b354715ba32466e0a7638204a7d9474213db4f5299ae1bc" gracePeriod=2 Dec 13 09:51:22 crc kubenswrapper[4971]: I1213 09:51:22.086855 4971 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-must-gather-69gc8/must-gather-rgtmv"] Dec 13 09:51:22 crc kubenswrapper[4971]: I1213 09:51:22.576692 4971 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-must-gather-69gc8_must-gather-rgtmv_b2ac511e-e641-4ee8-aca7-6054195a75d8/copy/0.log" Dec 13 09:51:22 crc kubenswrapper[4971]: I1213 09:51:22.577582 4971 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-69gc8/must-gather-rgtmv" Dec 13 09:51:22 crc kubenswrapper[4971]: I1213 09:51:22.665193 4971 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"must-gather-output\" (UniqueName: \"kubernetes.io/empty-dir/b2ac511e-e641-4ee8-aca7-6054195a75d8-must-gather-output\") pod \"b2ac511e-e641-4ee8-aca7-6054195a75d8\" (UID: \"b2ac511e-e641-4ee8-aca7-6054195a75d8\") " Dec 13 09:51:22 crc kubenswrapper[4971]: I1213 09:51:22.665460 4971 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-mfbrn\" (UniqueName: \"kubernetes.io/projected/b2ac511e-e641-4ee8-aca7-6054195a75d8-kube-api-access-mfbrn\") pod \"b2ac511e-e641-4ee8-aca7-6054195a75d8\" (UID: \"b2ac511e-e641-4ee8-aca7-6054195a75d8\") " Dec 13 09:51:22 crc kubenswrapper[4971]: I1213 09:51:22.674484 4971 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/b2ac511e-e641-4ee8-aca7-6054195a75d8-kube-api-access-mfbrn" (OuterVolumeSpecName: "kube-api-access-mfbrn") pod "b2ac511e-e641-4ee8-aca7-6054195a75d8" (UID: "b2ac511e-e641-4ee8-aca7-6054195a75d8"). InnerVolumeSpecName "kube-api-access-mfbrn". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 13 09:51:22 crc kubenswrapper[4971]: I1213 09:51:22.769382 4971 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-mfbrn\" (UniqueName: \"kubernetes.io/projected/b2ac511e-e641-4ee8-aca7-6054195a75d8-kube-api-access-mfbrn\") on node \"crc\" DevicePath \"\"" Dec 13 09:51:22 crc kubenswrapper[4971]: I1213 09:51:22.794250 4971 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-must-gather-69gc8_must-gather-rgtmv_b2ac511e-e641-4ee8-aca7-6054195a75d8/copy/0.log" Dec 13 09:51:22 crc kubenswrapper[4971]: I1213 09:51:22.794618 4971 generic.go:334] "Generic (PLEG): container finished" podID="b2ac511e-e641-4ee8-aca7-6054195a75d8" containerID="244712079a41bb332b354715ba32466e0a7638204a7d9474213db4f5299ae1bc" exitCode=143 Dec 13 09:51:22 crc kubenswrapper[4971]: I1213 09:51:22.794669 4971 scope.go:117] "RemoveContainer" containerID="244712079a41bb332b354715ba32466e0a7638204a7d9474213db4f5299ae1bc" Dec 13 09:51:22 crc kubenswrapper[4971]: I1213 09:51:22.794835 4971 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-69gc8/must-gather-rgtmv" Dec 13 09:51:22 crc kubenswrapper[4971]: I1213 09:51:22.831590 4971 scope.go:117] "RemoveContainer" containerID="423189d833cf0d72b52f4d07c31c19ad439ce4ef0db37aa7024772e15119e377" Dec 13 09:51:22 crc kubenswrapper[4971]: I1213 09:51:22.924495 4971 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/b2ac511e-e641-4ee8-aca7-6054195a75d8-must-gather-output" (OuterVolumeSpecName: "must-gather-output") pod "b2ac511e-e641-4ee8-aca7-6054195a75d8" (UID: "b2ac511e-e641-4ee8-aca7-6054195a75d8"). InnerVolumeSpecName "must-gather-output". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 13 09:51:22 crc kubenswrapper[4971]: I1213 09:51:22.944535 4971 scope.go:117] "RemoveContainer" containerID="244712079a41bb332b354715ba32466e0a7638204a7d9474213db4f5299ae1bc" Dec 13 09:51:22 crc kubenswrapper[4971]: E1213 09:51:22.945014 4971 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"244712079a41bb332b354715ba32466e0a7638204a7d9474213db4f5299ae1bc\": container with ID starting with 244712079a41bb332b354715ba32466e0a7638204a7d9474213db4f5299ae1bc not found: ID does not exist" containerID="244712079a41bb332b354715ba32466e0a7638204a7d9474213db4f5299ae1bc" Dec 13 09:51:22 crc kubenswrapper[4971]: I1213 09:51:22.945074 4971 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"244712079a41bb332b354715ba32466e0a7638204a7d9474213db4f5299ae1bc"} err="failed to get container status \"244712079a41bb332b354715ba32466e0a7638204a7d9474213db4f5299ae1bc\": rpc error: code = NotFound desc = could not find container \"244712079a41bb332b354715ba32466e0a7638204a7d9474213db4f5299ae1bc\": container with ID starting with 244712079a41bb332b354715ba32466e0a7638204a7d9474213db4f5299ae1bc not found: ID does not exist" Dec 13 09:51:22 crc kubenswrapper[4971]: I1213 09:51:22.945113 4971 scope.go:117] "RemoveContainer" containerID="423189d833cf0d72b52f4d07c31c19ad439ce4ef0db37aa7024772e15119e377" Dec 13 09:51:22 crc kubenswrapper[4971]: E1213 09:51:22.945910 4971 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"423189d833cf0d72b52f4d07c31c19ad439ce4ef0db37aa7024772e15119e377\": container with ID starting with 423189d833cf0d72b52f4d07c31c19ad439ce4ef0db37aa7024772e15119e377 not found: ID does not exist" containerID="423189d833cf0d72b52f4d07c31c19ad439ce4ef0db37aa7024772e15119e377" Dec 13 09:51:22 crc kubenswrapper[4971]: I1213 09:51:22.945975 4971 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"423189d833cf0d72b52f4d07c31c19ad439ce4ef0db37aa7024772e15119e377"} err="failed to get container status \"423189d833cf0d72b52f4d07c31c19ad439ce4ef0db37aa7024772e15119e377\": rpc error: code = NotFound desc = could not find container \"423189d833cf0d72b52f4d07c31c19ad439ce4ef0db37aa7024772e15119e377\": container with ID starting with 423189d833cf0d72b52f4d07c31c19ad439ce4ef0db37aa7024772e15119e377 not found: ID does not exist" Dec 13 09:51:22 crc kubenswrapper[4971]: I1213 09:51:22.979435 4971 reconciler_common.go:293] "Volume detached for volume \"must-gather-output\" (UniqueName: \"kubernetes.io/empty-dir/b2ac511e-e641-4ee8-aca7-6054195a75d8-must-gather-output\") on node \"crc\" DevicePath \"\"" Dec 13 09:51:23 crc kubenswrapper[4971]: I1213 09:51:23.846651 4971 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="b2ac511e-e641-4ee8-aca7-6054195a75d8" path="/var/lib/kubelet/pods/b2ac511e-e641-4ee8-aca7-6054195a75d8/volumes" Dec 13 09:51:27 crc kubenswrapper[4971]: I1213 09:51:27.769981 4971 scope.go:117] "RemoveContainer" containerID="80deb18ddf85834e136ab62ea93a1fa75315016d9f220eb217720a5e562f5e54" Dec 13 09:51:28 crc kubenswrapper[4971]: I1213 09:51:28.897616 4971 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-82xjz" event={"ID":"e0c4af12-edda-41d7-9b44-f87396174ef2","Type":"ContainerStarted","Data":"ba79b440c3c7f17627b56e5585d148b00ef9febdfb525490ad7ac2be28ca8ced"} Dec 13 09:51:30 crc kubenswrapper[4971]: I1213 09:51:30.928467 4971 scope.go:117] "RemoveContainer" containerID="92077d04a5c5f8ddbc19fb5ec1cda65d959a3d4835d1b05c4668b8853e148eb1" Dec 13 09:51:37 crc kubenswrapper[4971]: E1213 09:51:37.769425 4971 kubelet_pods.go:538] "Hostname for pod was too long, truncated it" podName="test-operator-logs-pod-horizontest-horizontest-tests-horizontest" hostnameMaxLen=63 truncatedHostname="test-operator-logs-pod-horizontest-horizontest-tests-horizontes" Dec 13 09:52:23 crc kubenswrapper[4971]: I1213 09:52:23.760764 4971 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-operators-rtj52"] Dec 13 09:52:23 crc kubenswrapper[4971]: E1213 09:52:23.762320 4971 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="b2ac511e-e641-4ee8-aca7-6054195a75d8" containerName="copy" Dec 13 09:52:23 crc kubenswrapper[4971]: I1213 09:52:23.762550 4971 state_mem.go:107] "Deleted CPUSet assignment" podUID="b2ac511e-e641-4ee8-aca7-6054195a75d8" containerName="copy" Dec 13 09:52:23 crc kubenswrapper[4971]: E1213 09:52:23.762578 4971 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="761521dd-cee4-4f2d-add4-8fc4fc966112" containerName="extract-content" Dec 13 09:52:23 crc kubenswrapper[4971]: I1213 09:52:23.762587 4971 state_mem.go:107] "Deleted CPUSet assignment" podUID="761521dd-cee4-4f2d-add4-8fc4fc966112" containerName="extract-content" Dec 13 09:52:23 crc kubenswrapper[4971]: E1213 09:52:23.762608 4971 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="b2ac511e-e641-4ee8-aca7-6054195a75d8" containerName="gather" Dec 13 09:52:23 crc kubenswrapper[4971]: I1213 09:52:23.762614 4971 state_mem.go:107] "Deleted CPUSet assignment" podUID="b2ac511e-e641-4ee8-aca7-6054195a75d8" containerName="gather" Dec 13 09:52:23 crc kubenswrapper[4971]: E1213 09:52:23.762623 4971 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="761521dd-cee4-4f2d-add4-8fc4fc966112" containerName="extract-utilities" Dec 13 09:52:23 crc kubenswrapper[4971]: I1213 09:52:23.762630 4971 state_mem.go:107] "Deleted CPUSet assignment" podUID="761521dd-cee4-4f2d-add4-8fc4fc966112" containerName="extract-utilities" Dec 13 09:52:23 crc kubenswrapper[4971]: E1213 09:52:23.762642 4971 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="761521dd-cee4-4f2d-add4-8fc4fc966112" containerName="registry-server" Dec 13 09:52:23 crc kubenswrapper[4971]: I1213 09:52:23.762648 4971 state_mem.go:107] "Deleted CPUSet assignment" podUID="761521dd-cee4-4f2d-add4-8fc4fc966112" containerName="registry-server" Dec 13 09:52:23 crc kubenswrapper[4971]: I1213 09:52:23.762858 4971 memory_manager.go:354] "RemoveStaleState removing state" podUID="b2ac511e-e641-4ee8-aca7-6054195a75d8" containerName="copy" Dec 13 09:52:23 crc kubenswrapper[4971]: I1213 09:52:23.762894 4971 memory_manager.go:354] "RemoveStaleState removing state" podUID="b2ac511e-e641-4ee8-aca7-6054195a75d8" containerName="gather" Dec 13 09:52:23 crc kubenswrapper[4971]: I1213 09:52:23.762903 4971 memory_manager.go:354] "RemoveStaleState removing state" podUID="761521dd-cee4-4f2d-add4-8fc4fc966112" containerName="registry-server" Dec 13 09:52:23 crc kubenswrapper[4971]: I1213 09:52:23.764505 4971 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-rtj52" Dec 13 09:52:23 crc kubenswrapper[4971]: I1213 09:52:23.781954 4971 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-rtj52"] Dec 13 09:52:23 crc kubenswrapper[4971]: I1213 09:52:23.816204 4971 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/cf26e1f0-ebbb-4196-8f83-961d61a1f4f6-catalog-content\") pod \"redhat-operators-rtj52\" (UID: \"cf26e1f0-ebbb-4196-8f83-961d61a1f4f6\") " pod="openshift-marketplace/redhat-operators-rtj52" Dec 13 09:52:23 crc kubenswrapper[4971]: I1213 09:52:23.816617 4971 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/cf26e1f0-ebbb-4196-8f83-961d61a1f4f6-utilities\") pod \"redhat-operators-rtj52\" (UID: \"cf26e1f0-ebbb-4196-8f83-961d61a1f4f6\") " pod="openshift-marketplace/redhat-operators-rtj52" Dec 13 09:52:23 crc kubenswrapper[4971]: I1213 09:52:23.816740 4971 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-dv26b\" (UniqueName: \"kubernetes.io/projected/cf26e1f0-ebbb-4196-8f83-961d61a1f4f6-kube-api-access-dv26b\") pod \"redhat-operators-rtj52\" (UID: \"cf26e1f0-ebbb-4196-8f83-961d61a1f4f6\") " pod="openshift-marketplace/redhat-operators-rtj52" Dec 13 09:52:23 crc kubenswrapper[4971]: I1213 09:52:23.920214 4971 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/cf26e1f0-ebbb-4196-8f83-961d61a1f4f6-catalog-content\") pod \"redhat-operators-rtj52\" (UID: \"cf26e1f0-ebbb-4196-8f83-961d61a1f4f6\") " pod="openshift-marketplace/redhat-operators-rtj52" Dec 13 09:52:23 crc kubenswrapper[4971]: I1213 09:52:23.920342 4971 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/cf26e1f0-ebbb-4196-8f83-961d61a1f4f6-utilities\") pod \"redhat-operators-rtj52\" (UID: \"cf26e1f0-ebbb-4196-8f83-961d61a1f4f6\") " pod="openshift-marketplace/redhat-operators-rtj52" Dec 13 09:52:23 crc kubenswrapper[4971]: I1213 09:52:23.920409 4971 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-dv26b\" (UniqueName: \"kubernetes.io/projected/cf26e1f0-ebbb-4196-8f83-961d61a1f4f6-kube-api-access-dv26b\") pod \"redhat-operators-rtj52\" (UID: \"cf26e1f0-ebbb-4196-8f83-961d61a1f4f6\") " pod="openshift-marketplace/redhat-operators-rtj52" Dec 13 09:52:23 crc kubenswrapper[4971]: I1213 09:52:23.921450 4971 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/cf26e1f0-ebbb-4196-8f83-961d61a1f4f6-catalog-content\") pod \"redhat-operators-rtj52\" (UID: \"cf26e1f0-ebbb-4196-8f83-961d61a1f4f6\") " pod="openshift-marketplace/redhat-operators-rtj52" Dec 13 09:52:23 crc kubenswrapper[4971]: I1213 09:52:23.921604 4971 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/cf26e1f0-ebbb-4196-8f83-961d61a1f4f6-utilities\") pod \"redhat-operators-rtj52\" (UID: \"cf26e1f0-ebbb-4196-8f83-961d61a1f4f6\") " pod="openshift-marketplace/redhat-operators-rtj52" Dec 13 09:52:23 crc kubenswrapper[4971]: I1213 09:52:23.947732 4971 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-dv26b\" (UniqueName: \"kubernetes.io/projected/cf26e1f0-ebbb-4196-8f83-961d61a1f4f6-kube-api-access-dv26b\") pod \"redhat-operators-rtj52\" (UID: \"cf26e1f0-ebbb-4196-8f83-961d61a1f4f6\") " pod="openshift-marketplace/redhat-operators-rtj52" Dec 13 09:52:24 crc kubenswrapper[4971]: I1213 09:52:24.096848 4971 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-rtj52" Dec 13 09:52:24 crc kubenswrapper[4971]: I1213 09:52:24.619019 4971 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-rtj52"] Dec 13 09:52:25 crc kubenswrapper[4971]: I1213 09:52:25.551801 4971 generic.go:334] "Generic (PLEG): container finished" podID="cf26e1f0-ebbb-4196-8f83-961d61a1f4f6" containerID="272f98012a308d207e1c817db6dd5ef5c2a9a1eb2c030a99845fa9b733c6a906" exitCode=0 Dec 13 09:52:25 crc kubenswrapper[4971]: I1213 09:52:25.551908 4971 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-rtj52" event={"ID":"cf26e1f0-ebbb-4196-8f83-961d61a1f4f6","Type":"ContainerDied","Data":"272f98012a308d207e1c817db6dd5ef5c2a9a1eb2c030a99845fa9b733c6a906"} Dec 13 09:52:25 crc kubenswrapper[4971]: I1213 09:52:25.552260 4971 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-rtj52" event={"ID":"cf26e1f0-ebbb-4196-8f83-961d61a1f4f6","Type":"ContainerStarted","Data":"2f3e78a3f0a4e563dbd1479abc1341e1cd4adda032493c20973d534a3ca0121c"} Dec 13 09:52:26 crc kubenswrapper[4971]: I1213 09:52:26.565440 4971 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-rtj52" event={"ID":"cf26e1f0-ebbb-4196-8f83-961d61a1f4f6","Type":"ContainerStarted","Data":"185dd6a589116807401a36235fb214660155e99d47167787b75b930f4a266cae"} Dec 13 09:52:28 crc kubenswrapper[4971]: I1213 09:52:28.585569 4971 generic.go:334] "Generic (PLEG): container finished" podID="cf26e1f0-ebbb-4196-8f83-961d61a1f4f6" containerID="185dd6a589116807401a36235fb214660155e99d47167787b75b930f4a266cae" exitCode=0 Dec 13 09:52:28 crc kubenswrapper[4971]: I1213 09:52:28.585814 4971 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-rtj52" event={"ID":"cf26e1f0-ebbb-4196-8f83-961d61a1f4f6","Type":"ContainerDied","Data":"185dd6a589116807401a36235fb214660155e99d47167787b75b930f4a266cae"} Dec 13 09:52:29 crc kubenswrapper[4971]: I1213 09:52:29.596364 4971 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-rtj52" event={"ID":"cf26e1f0-ebbb-4196-8f83-961d61a1f4f6","Type":"ContainerStarted","Data":"1f251eda4306549dd9fa4abaea85bc46e386a7366715f18774354fa4eab2903b"} Dec 13 09:52:29 crc kubenswrapper[4971]: I1213 09:52:29.631452 4971 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-operators-rtj52" podStartSLOduration=3.124475563 podStartE2EDuration="6.631426674s" podCreationTimestamp="2025-12-13 09:52:23 +0000 UTC" firstStartedPulling="2025-12-13 09:52:25.55380746 +0000 UTC m=+11002.158216908" lastFinishedPulling="2025-12-13 09:52:29.060758571 +0000 UTC m=+11005.665168019" observedRunningTime="2025-12-13 09:52:29.616933519 +0000 UTC m=+11006.221342967" watchObservedRunningTime="2025-12-13 09:52:29.631426674 +0000 UTC m=+11006.235836122" Dec 13 09:52:34 crc kubenswrapper[4971]: I1213 09:52:34.099141 4971 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-operators-rtj52" Dec 13 09:52:34 crc kubenswrapper[4971]: I1213 09:52:34.099587 4971 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-operators-rtj52" Dec 13 09:52:35 crc kubenswrapper[4971]: I1213 09:52:35.151892 4971 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-marketplace/redhat-operators-rtj52" podUID="cf26e1f0-ebbb-4196-8f83-961d61a1f4f6" containerName="registry-server" probeResult="failure" output=< Dec 13 09:52:35 crc kubenswrapper[4971]: timeout: failed to connect service ":50051" within 1s Dec 13 09:52:35 crc kubenswrapper[4971]: > Dec 13 09:52:44 crc kubenswrapper[4971]: I1213 09:52:44.162829 4971 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-operators-rtj52" Dec 13 09:52:44 crc kubenswrapper[4971]: I1213 09:52:44.216016 4971 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-operators-rtj52" Dec 13 09:52:44 crc kubenswrapper[4971]: I1213 09:52:44.414728 4971 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-rtj52"] Dec 13 09:52:45 crc kubenswrapper[4971]: I1213 09:52:45.790250 4971 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-operators-rtj52" podUID="cf26e1f0-ebbb-4196-8f83-961d61a1f4f6" containerName="registry-server" containerID="cri-o://1f251eda4306549dd9fa4abaea85bc46e386a7366715f18774354fa4eab2903b" gracePeriod=2 Dec 13 09:52:46 crc kubenswrapper[4971]: I1213 09:52:46.303283 4971 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-rtj52" Dec 13 09:52:46 crc kubenswrapper[4971]: I1213 09:52:46.436707 4971 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-dv26b\" (UniqueName: \"kubernetes.io/projected/cf26e1f0-ebbb-4196-8f83-961d61a1f4f6-kube-api-access-dv26b\") pod \"cf26e1f0-ebbb-4196-8f83-961d61a1f4f6\" (UID: \"cf26e1f0-ebbb-4196-8f83-961d61a1f4f6\") " Dec 13 09:52:46 crc kubenswrapper[4971]: I1213 09:52:46.436804 4971 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/cf26e1f0-ebbb-4196-8f83-961d61a1f4f6-utilities\") pod \"cf26e1f0-ebbb-4196-8f83-961d61a1f4f6\" (UID: \"cf26e1f0-ebbb-4196-8f83-961d61a1f4f6\") " Dec 13 09:52:46 crc kubenswrapper[4971]: I1213 09:52:46.436960 4971 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/cf26e1f0-ebbb-4196-8f83-961d61a1f4f6-catalog-content\") pod \"cf26e1f0-ebbb-4196-8f83-961d61a1f4f6\" (UID: \"cf26e1f0-ebbb-4196-8f83-961d61a1f4f6\") " Dec 13 09:52:46 crc kubenswrapper[4971]: I1213 09:52:46.448135 4971 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/cf26e1f0-ebbb-4196-8f83-961d61a1f4f6-utilities" (OuterVolumeSpecName: "utilities") pod "cf26e1f0-ebbb-4196-8f83-961d61a1f4f6" (UID: "cf26e1f0-ebbb-4196-8f83-961d61a1f4f6"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 13 09:52:46 crc kubenswrapper[4971]: I1213 09:52:46.448240 4971 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/cf26e1f0-ebbb-4196-8f83-961d61a1f4f6-kube-api-access-dv26b" (OuterVolumeSpecName: "kube-api-access-dv26b") pod "cf26e1f0-ebbb-4196-8f83-961d61a1f4f6" (UID: "cf26e1f0-ebbb-4196-8f83-961d61a1f4f6"). InnerVolumeSpecName "kube-api-access-dv26b". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 13 09:52:46 crc kubenswrapper[4971]: I1213 09:52:46.539246 4971 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-dv26b\" (UniqueName: \"kubernetes.io/projected/cf26e1f0-ebbb-4196-8f83-961d61a1f4f6-kube-api-access-dv26b\") on node \"crc\" DevicePath \"\"" Dec 13 09:52:46 crc kubenswrapper[4971]: I1213 09:52:46.539760 4971 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/cf26e1f0-ebbb-4196-8f83-961d61a1f4f6-utilities\") on node \"crc\" DevicePath \"\"" Dec 13 09:52:46 crc kubenswrapper[4971]: I1213 09:52:46.576123 4971 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/cf26e1f0-ebbb-4196-8f83-961d61a1f4f6-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "cf26e1f0-ebbb-4196-8f83-961d61a1f4f6" (UID: "cf26e1f0-ebbb-4196-8f83-961d61a1f4f6"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 13 09:52:46 crc kubenswrapper[4971]: I1213 09:52:46.642386 4971 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/cf26e1f0-ebbb-4196-8f83-961d61a1f4f6-catalog-content\") on node \"crc\" DevicePath \"\"" Dec 13 09:52:46 crc kubenswrapper[4971]: I1213 09:52:46.820841 4971 generic.go:334] "Generic (PLEG): container finished" podID="cf26e1f0-ebbb-4196-8f83-961d61a1f4f6" containerID="1f251eda4306549dd9fa4abaea85bc46e386a7366715f18774354fa4eab2903b" exitCode=0 Dec 13 09:52:46 crc kubenswrapper[4971]: I1213 09:52:46.820897 4971 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-rtj52" event={"ID":"cf26e1f0-ebbb-4196-8f83-961d61a1f4f6","Type":"ContainerDied","Data":"1f251eda4306549dd9fa4abaea85bc46e386a7366715f18774354fa4eab2903b"} Dec 13 09:52:46 crc kubenswrapper[4971]: I1213 09:52:46.820925 4971 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-rtj52" Dec 13 09:52:46 crc kubenswrapper[4971]: I1213 09:52:46.820939 4971 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-rtj52" event={"ID":"cf26e1f0-ebbb-4196-8f83-961d61a1f4f6","Type":"ContainerDied","Data":"2f3e78a3f0a4e563dbd1479abc1341e1cd4adda032493c20973d534a3ca0121c"} Dec 13 09:52:46 crc kubenswrapper[4971]: I1213 09:52:46.820960 4971 scope.go:117] "RemoveContainer" containerID="1f251eda4306549dd9fa4abaea85bc46e386a7366715f18774354fa4eab2903b" Dec 13 09:52:46 crc kubenswrapper[4971]: I1213 09:52:46.860069 4971 scope.go:117] "RemoveContainer" containerID="185dd6a589116807401a36235fb214660155e99d47167787b75b930f4a266cae" Dec 13 09:52:46 crc kubenswrapper[4971]: I1213 09:52:46.860896 4971 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-rtj52"] Dec 13 09:52:46 crc kubenswrapper[4971]: I1213 09:52:46.872418 4971 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-operators-rtj52"] Dec 13 09:52:46 crc kubenswrapper[4971]: I1213 09:52:46.895331 4971 scope.go:117] "RemoveContainer" containerID="272f98012a308d207e1c817db6dd5ef5c2a9a1eb2c030a99845fa9b733c6a906" Dec 13 09:52:46 crc kubenswrapper[4971]: I1213 09:52:46.953020 4971 scope.go:117] "RemoveContainer" containerID="1f251eda4306549dd9fa4abaea85bc46e386a7366715f18774354fa4eab2903b" Dec 13 09:52:46 crc kubenswrapper[4971]: E1213 09:52:46.953939 4971 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"1f251eda4306549dd9fa4abaea85bc46e386a7366715f18774354fa4eab2903b\": container with ID starting with 1f251eda4306549dd9fa4abaea85bc46e386a7366715f18774354fa4eab2903b not found: ID does not exist" containerID="1f251eda4306549dd9fa4abaea85bc46e386a7366715f18774354fa4eab2903b" Dec 13 09:52:46 crc kubenswrapper[4971]: I1213 09:52:46.953982 4971 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"1f251eda4306549dd9fa4abaea85bc46e386a7366715f18774354fa4eab2903b"} err="failed to get container status \"1f251eda4306549dd9fa4abaea85bc46e386a7366715f18774354fa4eab2903b\": rpc error: code = NotFound desc = could not find container \"1f251eda4306549dd9fa4abaea85bc46e386a7366715f18774354fa4eab2903b\": container with ID starting with 1f251eda4306549dd9fa4abaea85bc46e386a7366715f18774354fa4eab2903b not found: ID does not exist" Dec 13 09:52:46 crc kubenswrapper[4971]: I1213 09:52:46.954012 4971 scope.go:117] "RemoveContainer" containerID="185dd6a589116807401a36235fb214660155e99d47167787b75b930f4a266cae" Dec 13 09:52:46 crc kubenswrapper[4971]: E1213 09:52:46.954613 4971 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"185dd6a589116807401a36235fb214660155e99d47167787b75b930f4a266cae\": container with ID starting with 185dd6a589116807401a36235fb214660155e99d47167787b75b930f4a266cae not found: ID does not exist" containerID="185dd6a589116807401a36235fb214660155e99d47167787b75b930f4a266cae" Dec 13 09:52:46 crc kubenswrapper[4971]: I1213 09:52:46.954689 4971 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"185dd6a589116807401a36235fb214660155e99d47167787b75b930f4a266cae"} err="failed to get container status \"185dd6a589116807401a36235fb214660155e99d47167787b75b930f4a266cae\": rpc error: code = NotFound desc = could not find container \"185dd6a589116807401a36235fb214660155e99d47167787b75b930f4a266cae\": container with ID starting with 185dd6a589116807401a36235fb214660155e99d47167787b75b930f4a266cae not found: ID does not exist" Dec 13 09:52:46 crc kubenswrapper[4971]: I1213 09:52:46.954728 4971 scope.go:117] "RemoveContainer" containerID="272f98012a308d207e1c817db6dd5ef5c2a9a1eb2c030a99845fa9b733c6a906" Dec 13 09:52:46 crc kubenswrapper[4971]: E1213 09:52:46.955304 4971 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"272f98012a308d207e1c817db6dd5ef5c2a9a1eb2c030a99845fa9b733c6a906\": container with ID starting with 272f98012a308d207e1c817db6dd5ef5c2a9a1eb2c030a99845fa9b733c6a906 not found: ID does not exist" containerID="272f98012a308d207e1c817db6dd5ef5c2a9a1eb2c030a99845fa9b733c6a906" Dec 13 09:52:46 crc kubenswrapper[4971]: I1213 09:52:46.955414 4971 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"272f98012a308d207e1c817db6dd5ef5c2a9a1eb2c030a99845fa9b733c6a906"} err="failed to get container status \"272f98012a308d207e1c817db6dd5ef5c2a9a1eb2c030a99845fa9b733c6a906\": rpc error: code = NotFound desc = could not find container \"272f98012a308d207e1c817db6dd5ef5c2a9a1eb2c030a99845fa9b733c6a906\": container with ID starting with 272f98012a308d207e1c817db6dd5ef5c2a9a1eb2c030a99845fa9b733c6a906 not found: ID does not exist" Dec 13 09:52:47 crc kubenswrapper[4971]: I1213 09:52:47.786828 4971 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="cf26e1f0-ebbb-4196-8f83-961d61a1f4f6" path="/var/lib/kubelet/pods/cf26e1f0-ebbb-4196-8f83-961d61a1f4f6/volumes" Dec 13 09:52:59 crc kubenswrapper[4971]: E1213 09:52:59.770766 4971 kubelet_pods.go:538] "Hostname for pod was too long, truncated it" podName="test-operator-logs-pod-horizontest-horizontest-tests-horizontest" hostnameMaxLen=63 truncatedHostname="test-operator-logs-pod-horizontest-horizontest-tests-horizontes" Dec 13 09:53:07 crc kubenswrapper[4971]: I1213 09:53:07.303551 4971 generic.go:334] "Generic (PLEG): container finished" podID="d92c389a-6647-4873-8be9-53ff04c2d3fd" containerID="1e352e561f17b984001acd985b962cbbb93c4fc02c09a78dd54852cca53bcaf6" exitCode=0 Dec 13 09:53:07 crc kubenswrapper[4971]: I1213 09:53:07.303622 4971 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-g74xz/must-gather-p9ltj" event={"ID":"d92c389a-6647-4873-8be9-53ff04c2d3fd","Type":"ContainerDied","Data":"1e352e561f17b984001acd985b962cbbb93c4fc02c09a78dd54852cca53bcaf6"} Dec 13 09:53:07 crc kubenswrapper[4971]: I1213 09:53:07.306177 4971 scope.go:117] "RemoveContainer" containerID="1e352e561f17b984001acd985b962cbbb93c4fc02c09a78dd54852cca53bcaf6" Dec 13 09:53:07 crc kubenswrapper[4971]: I1213 09:53:07.892253 4971 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-must-gather-g74xz_must-gather-p9ltj_d92c389a-6647-4873-8be9-53ff04c2d3fd/gather/0.log" Dec 13 09:53:21 crc kubenswrapper[4971]: I1213 09:53:21.510713 4971 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-must-gather-g74xz/must-gather-p9ltj"] Dec 13 09:53:21 crc kubenswrapper[4971]: I1213 09:53:21.512216 4971 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-must-gather-g74xz/must-gather-p9ltj" podUID="d92c389a-6647-4873-8be9-53ff04c2d3fd" containerName="copy" containerID="cri-o://66c643c4043229d4a1f17628345e97eff91da8ce8b6bbbc20d6e26c5f9a281e6" gracePeriod=2 Dec 13 09:53:21 crc kubenswrapper[4971]: I1213 09:53:21.521870 4971 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-must-gather-g74xz/must-gather-p9ltj"] Dec 13 09:53:22 crc kubenswrapper[4971]: I1213 09:53:22.080357 4971 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-must-gather-g74xz_must-gather-p9ltj_d92c389a-6647-4873-8be9-53ff04c2d3fd/copy/0.log" Dec 13 09:53:22 crc kubenswrapper[4971]: I1213 09:53:22.081367 4971 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-g74xz/must-gather-p9ltj" Dec 13 09:53:22 crc kubenswrapper[4971]: I1213 09:53:22.148392 4971 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-6b7m5\" (UniqueName: \"kubernetes.io/projected/d92c389a-6647-4873-8be9-53ff04c2d3fd-kube-api-access-6b7m5\") pod \"d92c389a-6647-4873-8be9-53ff04c2d3fd\" (UID: \"d92c389a-6647-4873-8be9-53ff04c2d3fd\") " Dec 13 09:53:22 crc kubenswrapper[4971]: I1213 09:53:22.148801 4971 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"must-gather-output\" (UniqueName: \"kubernetes.io/empty-dir/d92c389a-6647-4873-8be9-53ff04c2d3fd-must-gather-output\") pod \"d92c389a-6647-4873-8be9-53ff04c2d3fd\" (UID: \"d92c389a-6647-4873-8be9-53ff04c2d3fd\") " Dec 13 09:53:22 crc kubenswrapper[4971]: I1213 09:53:22.156693 4971 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/d92c389a-6647-4873-8be9-53ff04c2d3fd-kube-api-access-6b7m5" (OuterVolumeSpecName: "kube-api-access-6b7m5") pod "d92c389a-6647-4873-8be9-53ff04c2d3fd" (UID: "d92c389a-6647-4873-8be9-53ff04c2d3fd"). InnerVolumeSpecName "kube-api-access-6b7m5". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 13 09:53:22 crc kubenswrapper[4971]: I1213 09:53:22.252396 4971 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-6b7m5\" (UniqueName: \"kubernetes.io/projected/d92c389a-6647-4873-8be9-53ff04c2d3fd-kube-api-access-6b7m5\") on node \"crc\" DevicePath \"\"" Dec 13 09:53:22 crc kubenswrapper[4971]: I1213 09:53:22.368470 4971 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/d92c389a-6647-4873-8be9-53ff04c2d3fd-must-gather-output" (OuterVolumeSpecName: "must-gather-output") pod "d92c389a-6647-4873-8be9-53ff04c2d3fd" (UID: "d92c389a-6647-4873-8be9-53ff04c2d3fd"). InnerVolumeSpecName "must-gather-output". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 13 09:53:22 crc kubenswrapper[4971]: I1213 09:53:22.458660 4971 reconciler_common.go:293] "Volume detached for volume \"must-gather-output\" (UniqueName: \"kubernetes.io/empty-dir/d92c389a-6647-4873-8be9-53ff04c2d3fd-must-gather-output\") on node \"crc\" DevicePath \"\"" Dec 13 09:53:22 crc kubenswrapper[4971]: I1213 09:53:22.514421 4971 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-must-gather-g74xz_must-gather-p9ltj_d92c389a-6647-4873-8be9-53ff04c2d3fd/copy/0.log" Dec 13 09:53:22 crc kubenswrapper[4971]: I1213 09:53:22.515003 4971 generic.go:334] "Generic (PLEG): container finished" podID="d92c389a-6647-4873-8be9-53ff04c2d3fd" containerID="66c643c4043229d4a1f17628345e97eff91da8ce8b6bbbc20d6e26c5f9a281e6" exitCode=143 Dec 13 09:53:22 crc kubenswrapper[4971]: I1213 09:53:22.515083 4971 scope.go:117] "RemoveContainer" containerID="66c643c4043229d4a1f17628345e97eff91da8ce8b6bbbc20d6e26c5f9a281e6" Dec 13 09:53:22 crc kubenswrapper[4971]: I1213 09:53:22.515120 4971 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-g74xz/must-gather-p9ltj" Dec 13 09:53:22 crc kubenswrapper[4971]: I1213 09:53:22.541146 4971 scope.go:117] "RemoveContainer" containerID="1e352e561f17b984001acd985b962cbbb93c4fc02c09a78dd54852cca53bcaf6" Dec 13 09:53:22 crc kubenswrapper[4971]: I1213 09:53:22.635845 4971 scope.go:117] "RemoveContainer" containerID="66c643c4043229d4a1f17628345e97eff91da8ce8b6bbbc20d6e26c5f9a281e6" Dec 13 09:53:22 crc kubenswrapper[4971]: E1213 09:53:22.636480 4971 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"66c643c4043229d4a1f17628345e97eff91da8ce8b6bbbc20d6e26c5f9a281e6\": container with ID starting with 66c643c4043229d4a1f17628345e97eff91da8ce8b6bbbc20d6e26c5f9a281e6 not found: ID does not exist" containerID="66c643c4043229d4a1f17628345e97eff91da8ce8b6bbbc20d6e26c5f9a281e6" Dec 13 09:53:22 crc kubenswrapper[4971]: I1213 09:53:22.636540 4971 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"66c643c4043229d4a1f17628345e97eff91da8ce8b6bbbc20d6e26c5f9a281e6"} err="failed to get container status \"66c643c4043229d4a1f17628345e97eff91da8ce8b6bbbc20d6e26c5f9a281e6\": rpc error: code = NotFound desc = could not find container \"66c643c4043229d4a1f17628345e97eff91da8ce8b6bbbc20d6e26c5f9a281e6\": container with ID starting with 66c643c4043229d4a1f17628345e97eff91da8ce8b6bbbc20d6e26c5f9a281e6 not found: ID does not exist" Dec 13 09:53:22 crc kubenswrapper[4971]: I1213 09:53:22.636560 4971 scope.go:117] "RemoveContainer" containerID="1e352e561f17b984001acd985b962cbbb93c4fc02c09a78dd54852cca53bcaf6" Dec 13 09:53:22 crc kubenswrapper[4971]: E1213 09:53:22.637198 4971 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"1e352e561f17b984001acd985b962cbbb93c4fc02c09a78dd54852cca53bcaf6\": container with ID starting with 1e352e561f17b984001acd985b962cbbb93c4fc02c09a78dd54852cca53bcaf6 not found: ID does not exist" containerID="1e352e561f17b984001acd985b962cbbb93c4fc02c09a78dd54852cca53bcaf6" Dec 13 09:53:22 crc kubenswrapper[4971]: I1213 09:53:22.637265 4971 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"1e352e561f17b984001acd985b962cbbb93c4fc02c09a78dd54852cca53bcaf6"} err="failed to get container status \"1e352e561f17b984001acd985b962cbbb93c4fc02c09a78dd54852cca53bcaf6\": rpc error: code = NotFound desc = could not find container \"1e352e561f17b984001acd985b962cbbb93c4fc02c09a78dd54852cca53bcaf6\": container with ID starting with 1e352e561f17b984001acd985b962cbbb93c4fc02c09a78dd54852cca53bcaf6 not found: ID does not exist" Dec 13 09:53:23 crc kubenswrapper[4971]: I1213 09:53:23.784890 4971 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="d92c389a-6647-4873-8be9-53ff04c2d3fd" path="/var/lib/kubelet/pods/d92c389a-6647-4873-8be9-53ff04c2d3fd/volumes" Dec 13 09:53:31 crc kubenswrapper[4971]: I1213 09:53:31.062405 4971 scope.go:117] "RemoveContainer" containerID="474aee8c65e63d738d23dbf5b98c68b47526a7f328a39f7f07485285afc74dc4" Dec 13 09:53:38 crc kubenswrapper[4971]: I1213 09:53:38.007497 4971 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/certified-operators-24blq"] Dec 13 09:53:38 crc kubenswrapper[4971]: E1213 09:53:38.009068 4971 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="cf26e1f0-ebbb-4196-8f83-961d61a1f4f6" containerName="extract-content" Dec 13 09:53:38 crc kubenswrapper[4971]: I1213 09:53:38.009088 4971 state_mem.go:107] "Deleted CPUSet assignment" podUID="cf26e1f0-ebbb-4196-8f83-961d61a1f4f6" containerName="extract-content" Dec 13 09:53:38 crc kubenswrapper[4971]: E1213 09:53:38.009104 4971 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="cf26e1f0-ebbb-4196-8f83-961d61a1f4f6" containerName="extract-utilities" Dec 13 09:53:38 crc kubenswrapper[4971]: I1213 09:53:38.009111 4971 state_mem.go:107] "Deleted CPUSet assignment" podUID="cf26e1f0-ebbb-4196-8f83-961d61a1f4f6" containerName="extract-utilities" Dec 13 09:53:38 crc kubenswrapper[4971]: E1213 09:53:38.009126 4971 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="d92c389a-6647-4873-8be9-53ff04c2d3fd" containerName="copy" Dec 13 09:53:38 crc kubenswrapper[4971]: I1213 09:53:38.009133 4971 state_mem.go:107] "Deleted CPUSet assignment" podUID="d92c389a-6647-4873-8be9-53ff04c2d3fd" containerName="copy" Dec 13 09:53:38 crc kubenswrapper[4971]: E1213 09:53:38.009140 4971 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="d92c389a-6647-4873-8be9-53ff04c2d3fd" containerName="gather" Dec 13 09:53:38 crc kubenswrapper[4971]: I1213 09:53:38.009146 4971 state_mem.go:107] "Deleted CPUSet assignment" podUID="d92c389a-6647-4873-8be9-53ff04c2d3fd" containerName="gather" Dec 13 09:53:38 crc kubenswrapper[4971]: E1213 09:53:38.009184 4971 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="cf26e1f0-ebbb-4196-8f83-961d61a1f4f6" containerName="registry-server" Dec 13 09:53:38 crc kubenswrapper[4971]: I1213 09:53:38.009192 4971 state_mem.go:107] "Deleted CPUSet assignment" podUID="cf26e1f0-ebbb-4196-8f83-961d61a1f4f6" containerName="registry-server" Dec 13 09:53:38 crc kubenswrapper[4971]: I1213 09:53:38.009420 4971 memory_manager.go:354] "RemoveStaleState removing state" podUID="cf26e1f0-ebbb-4196-8f83-961d61a1f4f6" containerName="registry-server" Dec 13 09:53:38 crc kubenswrapper[4971]: I1213 09:53:38.009430 4971 memory_manager.go:354] "RemoveStaleState removing state" podUID="d92c389a-6647-4873-8be9-53ff04c2d3fd" containerName="gather" Dec 13 09:53:38 crc kubenswrapper[4971]: I1213 09:53:38.009437 4971 memory_manager.go:354] "RemoveStaleState removing state" podUID="d92c389a-6647-4873-8be9-53ff04c2d3fd" containerName="copy" Dec 13 09:53:38 crc kubenswrapper[4971]: I1213 09:53:38.012854 4971 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-24blq" Dec 13 09:53:38 crc kubenswrapper[4971]: I1213 09:53:38.027075 4971 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-24blq"] Dec 13 09:53:38 crc kubenswrapper[4971]: I1213 09:53:38.077963 4971 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/2b6b1aa6-d683-46a9-a5ee-9ea29177d2e0-utilities\") pod \"certified-operators-24blq\" (UID: \"2b6b1aa6-d683-46a9-a5ee-9ea29177d2e0\") " pod="openshift-marketplace/certified-operators-24blq" Dec 13 09:53:38 crc kubenswrapper[4971]: I1213 09:53:38.078178 4971 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-9g4t5\" (UniqueName: \"kubernetes.io/projected/2b6b1aa6-d683-46a9-a5ee-9ea29177d2e0-kube-api-access-9g4t5\") pod \"certified-operators-24blq\" (UID: \"2b6b1aa6-d683-46a9-a5ee-9ea29177d2e0\") " pod="openshift-marketplace/certified-operators-24blq" Dec 13 09:53:38 crc kubenswrapper[4971]: I1213 09:53:38.078817 4971 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/2b6b1aa6-d683-46a9-a5ee-9ea29177d2e0-catalog-content\") pod \"certified-operators-24blq\" (UID: \"2b6b1aa6-d683-46a9-a5ee-9ea29177d2e0\") " pod="openshift-marketplace/certified-operators-24blq" Dec 13 09:53:38 crc kubenswrapper[4971]: I1213 09:53:38.183634 4971 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/2b6b1aa6-d683-46a9-a5ee-9ea29177d2e0-catalog-content\") pod \"certified-operators-24blq\" (UID: \"2b6b1aa6-d683-46a9-a5ee-9ea29177d2e0\") " pod="openshift-marketplace/certified-operators-24blq" Dec 13 09:53:38 crc kubenswrapper[4971]: I1213 09:53:38.183756 4971 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/2b6b1aa6-d683-46a9-a5ee-9ea29177d2e0-utilities\") pod \"certified-operators-24blq\" (UID: \"2b6b1aa6-d683-46a9-a5ee-9ea29177d2e0\") " pod="openshift-marketplace/certified-operators-24blq" Dec 13 09:53:38 crc kubenswrapper[4971]: I1213 09:53:38.183933 4971 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-9g4t5\" (UniqueName: \"kubernetes.io/projected/2b6b1aa6-d683-46a9-a5ee-9ea29177d2e0-kube-api-access-9g4t5\") pod \"certified-operators-24blq\" (UID: \"2b6b1aa6-d683-46a9-a5ee-9ea29177d2e0\") " pod="openshift-marketplace/certified-operators-24blq" Dec 13 09:53:38 crc kubenswrapper[4971]: I1213 09:53:38.184588 4971 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/2b6b1aa6-d683-46a9-a5ee-9ea29177d2e0-catalog-content\") pod \"certified-operators-24blq\" (UID: \"2b6b1aa6-d683-46a9-a5ee-9ea29177d2e0\") " pod="openshift-marketplace/certified-operators-24blq" Dec 13 09:53:38 crc kubenswrapper[4971]: I1213 09:53:38.184882 4971 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/2b6b1aa6-d683-46a9-a5ee-9ea29177d2e0-utilities\") pod \"certified-operators-24blq\" (UID: \"2b6b1aa6-d683-46a9-a5ee-9ea29177d2e0\") " pod="openshift-marketplace/certified-operators-24blq" Dec 13 09:53:38 crc kubenswrapper[4971]: I1213 09:53:38.214594 4971 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-9g4t5\" (UniqueName: \"kubernetes.io/projected/2b6b1aa6-d683-46a9-a5ee-9ea29177d2e0-kube-api-access-9g4t5\") pod \"certified-operators-24blq\" (UID: \"2b6b1aa6-d683-46a9-a5ee-9ea29177d2e0\") " pod="openshift-marketplace/certified-operators-24blq" Dec 13 09:53:38 crc kubenswrapper[4971]: I1213 09:53:38.346649 4971 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-24blq" Dec 13 09:53:38 crc kubenswrapper[4971]: I1213 09:53:38.923437 4971 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-24blq"] Dec 13 09:53:38 crc kubenswrapper[4971]: W1213 09:53:38.941937 4971 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod2b6b1aa6_d683_46a9_a5ee_9ea29177d2e0.slice/crio-5071993e6806966452aee14d53ca269c2b49344e960954e114779c872ee370e0 WatchSource:0}: Error finding container 5071993e6806966452aee14d53ca269c2b49344e960954e114779c872ee370e0: Status 404 returned error can't find the container with id 5071993e6806966452aee14d53ca269c2b49344e960954e114779c872ee370e0 Dec 13 09:53:39 crc kubenswrapper[4971]: I1213 09:53:39.725102 4971 generic.go:334] "Generic (PLEG): container finished" podID="2b6b1aa6-d683-46a9-a5ee-9ea29177d2e0" containerID="3cae7fc7d9448cfa8bb3b77761824ab43ffb45c44953cf731ac699086a5e3063" exitCode=0 Dec 13 09:53:39 crc kubenswrapper[4971]: I1213 09:53:39.725502 4971 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-24blq" event={"ID":"2b6b1aa6-d683-46a9-a5ee-9ea29177d2e0","Type":"ContainerDied","Data":"3cae7fc7d9448cfa8bb3b77761824ab43ffb45c44953cf731ac699086a5e3063"} Dec 13 09:53:39 crc kubenswrapper[4971]: I1213 09:53:39.726085 4971 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-24blq" event={"ID":"2b6b1aa6-d683-46a9-a5ee-9ea29177d2e0","Type":"ContainerStarted","Data":"5071993e6806966452aee14d53ca269c2b49344e960954e114779c872ee370e0"} Dec 13 09:53:41 crc kubenswrapper[4971]: I1213 09:53:41.749670 4971 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-24blq" event={"ID":"2b6b1aa6-d683-46a9-a5ee-9ea29177d2e0","Type":"ContainerStarted","Data":"77d99de6bd1aff39924e1c88ff696bd585fbd8290a22e844eae7e19c0210a928"} Dec 13 09:53:42 crc kubenswrapper[4971]: I1213 09:53:42.763722 4971 generic.go:334] "Generic (PLEG): container finished" podID="2b6b1aa6-d683-46a9-a5ee-9ea29177d2e0" containerID="77d99de6bd1aff39924e1c88ff696bd585fbd8290a22e844eae7e19c0210a928" exitCode=0 Dec 13 09:53:42 crc kubenswrapper[4971]: I1213 09:53:42.763783 4971 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-24blq" event={"ID":"2b6b1aa6-d683-46a9-a5ee-9ea29177d2e0","Type":"ContainerDied","Data":"77d99de6bd1aff39924e1c88ff696bd585fbd8290a22e844eae7e19c0210a928"} Dec 13 09:53:43 crc kubenswrapper[4971]: I1213 09:53:43.781497 4971 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-24blq" event={"ID":"2b6b1aa6-d683-46a9-a5ee-9ea29177d2e0","Type":"ContainerStarted","Data":"9ceb3af3a6f70c76503023efc53777524329655272c9e2baacd3516598db8127"} Dec 13 09:53:43 crc kubenswrapper[4971]: I1213 09:53:43.808160 4971 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/certified-operators-24blq" podStartSLOduration=3.239912071 podStartE2EDuration="6.808129272s" podCreationTimestamp="2025-12-13 09:53:37 +0000 UTC" firstStartedPulling="2025-12-13 09:53:39.728819427 +0000 UTC m=+11076.333228875" lastFinishedPulling="2025-12-13 09:53:43.297036608 +0000 UTC m=+11079.901446076" observedRunningTime="2025-12-13 09:53:43.803891838 +0000 UTC m=+11080.408301286" watchObservedRunningTime="2025-12-13 09:53:43.808129272 +0000 UTC m=+11080.412538720" Dec 13 09:53:46 crc kubenswrapper[4971]: I1213 09:53:46.154036 4971 patch_prober.go:28] interesting pod/machine-config-daemon-82xjz container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 13 09:53:46 crc kubenswrapper[4971]: I1213 09:53:46.154761 4971 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-82xjz" podUID="e0c4af12-edda-41d7-9b44-f87396174ef2" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 13 09:53:48 crc kubenswrapper[4971]: I1213 09:53:48.347182 4971 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/certified-operators-24blq" Dec 13 09:53:48 crc kubenswrapper[4971]: I1213 09:53:48.347685 4971 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/certified-operators-24blq" Dec 13 09:53:48 crc kubenswrapper[4971]: I1213 09:53:48.405370 4971 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/certified-operators-24blq" Dec 13 09:53:48 crc kubenswrapper[4971]: I1213 09:53:48.888606 4971 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/certified-operators-24blq" Dec 13 09:53:48 crc kubenswrapper[4971]: I1213 09:53:48.943597 4971 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-24blq"] Dec 13 09:53:50 crc kubenswrapper[4971]: I1213 09:53:50.863295 4971 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/certified-operators-24blq" podUID="2b6b1aa6-d683-46a9-a5ee-9ea29177d2e0" containerName="registry-server" containerID="cri-o://9ceb3af3a6f70c76503023efc53777524329655272c9e2baacd3516598db8127" gracePeriod=2 Dec 13 09:53:51 crc kubenswrapper[4971]: I1213 09:53:51.372390 4971 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-24blq" Dec 13 09:53:51 crc kubenswrapper[4971]: I1213 09:53:51.439063 4971 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/2b6b1aa6-d683-46a9-a5ee-9ea29177d2e0-catalog-content\") pod \"2b6b1aa6-d683-46a9-a5ee-9ea29177d2e0\" (UID: \"2b6b1aa6-d683-46a9-a5ee-9ea29177d2e0\") " Dec 13 09:53:51 crc kubenswrapper[4971]: I1213 09:53:51.439134 4971 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-9g4t5\" (UniqueName: \"kubernetes.io/projected/2b6b1aa6-d683-46a9-a5ee-9ea29177d2e0-kube-api-access-9g4t5\") pod \"2b6b1aa6-d683-46a9-a5ee-9ea29177d2e0\" (UID: \"2b6b1aa6-d683-46a9-a5ee-9ea29177d2e0\") " Dec 13 09:53:51 crc kubenswrapper[4971]: I1213 09:53:51.439221 4971 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/2b6b1aa6-d683-46a9-a5ee-9ea29177d2e0-utilities\") pod \"2b6b1aa6-d683-46a9-a5ee-9ea29177d2e0\" (UID: \"2b6b1aa6-d683-46a9-a5ee-9ea29177d2e0\") " Dec 13 09:53:51 crc kubenswrapper[4971]: I1213 09:53:51.440167 4971 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/2b6b1aa6-d683-46a9-a5ee-9ea29177d2e0-utilities" (OuterVolumeSpecName: "utilities") pod "2b6b1aa6-d683-46a9-a5ee-9ea29177d2e0" (UID: "2b6b1aa6-d683-46a9-a5ee-9ea29177d2e0"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 13 09:53:51 crc kubenswrapper[4971]: I1213 09:53:51.461929 4971 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/2b6b1aa6-d683-46a9-a5ee-9ea29177d2e0-kube-api-access-9g4t5" (OuterVolumeSpecName: "kube-api-access-9g4t5") pod "2b6b1aa6-d683-46a9-a5ee-9ea29177d2e0" (UID: "2b6b1aa6-d683-46a9-a5ee-9ea29177d2e0"). InnerVolumeSpecName "kube-api-access-9g4t5". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 13 09:53:51 crc kubenswrapper[4971]: I1213 09:53:51.510433 4971 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/2b6b1aa6-d683-46a9-a5ee-9ea29177d2e0-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "2b6b1aa6-d683-46a9-a5ee-9ea29177d2e0" (UID: "2b6b1aa6-d683-46a9-a5ee-9ea29177d2e0"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 13 09:53:51 crc kubenswrapper[4971]: I1213 09:53:51.544168 4971 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/2b6b1aa6-d683-46a9-a5ee-9ea29177d2e0-catalog-content\") on node \"crc\" DevicePath \"\"" Dec 13 09:53:51 crc kubenswrapper[4971]: I1213 09:53:51.544440 4971 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-9g4t5\" (UniqueName: \"kubernetes.io/projected/2b6b1aa6-d683-46a9-a5ee-9ea29177d2e0-kube-api-access-9g4t5\") on node \"crc\" DevicePath \"\"" Dec 13 09:53:51 crc kubenswrapper[4971]: I1213 09:53:51.544557 4971 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/2b6b1aa6-d683-46a9-a5ee-9ea29177d2e0-utilities\") on node \"crc\" DevicePath \"\"" Dec 13 09:53:51 crc kubenswrapper[4971]: I1213 09:53:51.880307 4971 generic.go:334] "Generic (PLEG): container finished" podID="2b6b1aa6-d683-46a9-a5ee-9ea29177d2e0" containerID="9ceb3af3a6f70c76503023efc53777524329655272c9e2baacd3516598db8127" exitCode=0 Dec 13 09:53:51 crc kubenswrapper[4971]: I1213 09:53:51.880417 4971 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-24blq" event={"ID":"2b6b1aa6-d683-46a9-a5ee-9ea29177d2e0","Type":"ContainerDied","Data":"9ceb3af3a6f70c76503023efc53777524329655272c9e2baacd3516598db8127"} Dec 13 09:53:51 crc kubenswrapper[4971]: I1213 09:53:51.880471 4971 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-24blq" Dec 13 09:53:51 crc kubenswrapper[4971]: I1213 09:53:51.880558 4971 scope.go:117] "RemoveContainer" containerID="9ceb3af3a6f70c76503023efc53777524329655272c9e2baacd3516598db8127" Dec 13 09:53:51 crc kubenswrapper[4971]: I1213 09:53:51.880502 4971 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-24blq" event={"ID":"2b6b1aa6-d683-46a9-a5ee-9ea29177d2e0","Type":"ContainerDied","Data":"5071993e6806966452aee14d53ca269c2b49344e960954e114779c872ee370e0"} Dec 13 09:53:51 crc kubenswrapper[4971]: I1213 09:53:51.910213 4971 scope.go:117] "RemoveContainer" containerID="77d99de6bd1aff39924e1c88ff696bd585fbd8290a22e844eae7e19c0210a928" Dec 13 09:53:51 crc kubenswrapper[4971]: I1213 09:53:51.925424 4971 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-24blq"] Dec 13 09:53:51 crc kubenswrapper[4971]: I1213 09:53:51.937861 4971 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/certified-operators-24blq"] Dec 13 09:53:52 crc kubenswrapper[4971]: I1213 09:53:52.003351 4971 scope.go:117] "RemoveContainer" containerID="3cae7fc7d9448cfa8bb3b77761824ab43ffb45c44953cf731ac699086a5e3063" Dec 13 09:53:52 crc kubenswrapper[4971]: I1213 09:53:52.048819 4971 scope.go:117] "RemoveContainer" containerID="9ceb3af3a6f70c76503023efc53777524329655272c9e2baacd3516598db8127" Dec 13 09:53:52 crc kubenswrapper[4971]: E1213 09:53:52.050147 4971 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"9ceb3af3a6f70c76503023efc53777524329655272c9e2baacd3516598db8127\": container with ID starting with 9ceb3af3a6f70c76503023efc53777524329655272c9e2baacd3516598db8127 not found: ID does not exist" containerID="9ceb3af3a6f70c76503023efc53777524329655272c9e2baacd3516598db8127" Dec 13 09:53:52 crc kubenswrapper[4971]: I1213 09:53:52.050190 4971 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"9ceb3af3a6f70c76503023efc53777524329655272c9e2baacd3516598db8127"} err="failed to get container status \"9ceb3af3a6f70c76503023efc53777524329655272c9e2baacd3516598db8127\": rpc error: code = NotFound desc = could not find container \"9ceb3af3a6f70c76503023efc53777524329655272c9e2baacd3516598db8127\": container with ID starting with 9ceb3af3a6f70c76503023efc53777524329655272c9e2baacd3516598db8127 not found: ID does not exist" Dec 13 09:53:52 crc kubenswrapper[4971]: I1213 09:53:52.050220 4971 scope.go:117] "RemoveContainer" containerID="77d99de6bd1aff39924e1c88ff696bd585fbd8290a22e844eae7e19c0210a928" Dec 13 09:53:52 crc kubenswrapper[4971]: E1213 09:53:52.050962 4971 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"77d99de6bd1aff39924e1c88ff696bd585fbd8290a22e844eae7e19c0210a928\": container with ID starting with 77d99de6bd1aff39924e1c88ff696bd585fbd8290a22e844eae7e19c0210a928 not found: ID does not exist" containerID="77d99de6bd1aff39924e1c88ff696bd585fbd8290a22e844eae7e19c0210a928" Dec 13 09:53:52 crc kubenswrapper[4971]: I1213 09:53:52.051094 4971 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"77d99de6bd1aff39924e1c88ff696bd585fbd8290a22e844eae7e19c0210a928"} err="failed to get container status \"77d99de6bd1aff39924e1c88ff696bd585fbd8290a22e844eae7e19c0210a928\": rpc error: code = NotFound desc = could not find container \"77d99de6bd1aff39924e1c88ff696bd585fbd8290a22e844eae7e19c0210a928\": container with ID starting with 77d99de6bd1aff39924e1c88ff696bd585fbd8290a22e844eae7e19c0210a928 not found: ID does not exist" Dec 13 09:53:52 crc kubenswrapper[4971]: I1213 09:53:52.051165 4971 scope.go:117] "RemoveContainer" containerID="3cae7fc7d9448cfa8bb3b77761824ab43ffb45c44953cf731ac699086a5e3063" Dec 13 09:53:52 crc kubenswrapper[4971]: E1213 09:53:52.055505 4971 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"3cae7fc7d9448cfa8bb3b77761824ab43ffb45c44953cf731ac699086a5e3063\": container with ID starting with 3cae7fc7d9448cfa8bb3b77761824ab43ffb45c44953cf731ac699086a5e3063 not found: ID does not exist" containerID="3cae7fc7d9448cfa8bb3b77761824ab43ffb45c44953cf731ac699086a5e3063" Dec 13 09:53:52 crc kubenswrapper[4971]: I1213 09:53:52.055642 4971 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"3cae7fc7d9448cfa8bb3b77761824ab43ffb45c44953cf731ac699086a5e3063"} err="failed to get container status \"3cae7fc7d9448cfa8bb3b77761824ab43ffb45c44953cf731ac699086a5e3063\": rpc error: code = NotFound desc = could not find container \"3cae7fc7d9448cfa8bb3b77761824ab43ffb45c44953cf731ac699086a5e3063\": container with ID starting with 3cae7fc7d9448cfa8bb3b77761824ab43ffb45c44953cf731ac699086a5e3063 not found: ID does not exist" Dec 13 09:53:53 crc kubenswrapper[4971]: I1213 09:53:53.783824 4971 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="2b6b1aa6-d683-46a9-a5ee-9ea29177d2e0" path="/var/lib/kubelet/pods/2b6b1aa6-d683-46a9-a5ee-9ea29177d2e0/volumes" Dec 13 09:53:58 crc kubenswrapper[4971]: I1213 09:53:58.736221 4971 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/openstack-galera-0" podUID="91831fd5-fc52-46ec-ac49-63859378116b" containerName="galera" probeResult="failure" output="command timed out" Dec 13 09:53:58 crc kubenswrapper[4971]: I1213 09:53:58.737064 4971 prober.go:107] "Probe failed" probeType="Liveness" pod="openstack/openstack-galera-0" podUID="91831fd5-fc52-46ec-ac49-63859378116b" containerName="galera" probeResult="failure" output="command timed out" Dec 13 09:54:11 crc kubenswrapper[4971]: E1213 09:54:11.771228 4971 kubelet_pods.go:538] "Hostname for pod was too long, truncated it" podName="test-operator-logs-pod-horizontest-horizontest-tests-horizontest" hostnameMaxLen=63 truncatedHostname="test-operator-logs-pod-horizontest-horizontest-tests-horizontes" Dec 13 09:54:16 crc kubenswrapper[4971]: I1213 09:54:16.154254 4971 patch_prober.go:28] interesting pod/machine-config-daemon-82xjz container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 13 09:54:16 crc kubenswrapper[4971]: I1213 09:54:16.156602 4971 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-82xjz" podUID="e0c4af12-edda-41d7-9b44-f87396174ef2" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 13 09:54:46 crc kubenswrapper[4971]: I1213 09:54:46.153362 4971 patch_prober.go:28] interesting pod/machine-config-daemon-82xjz container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 13 09:54:46 crc kubenswrapper[4971]: I1213 09:54:46.154164 4971 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-82xjz" podUID="e0c4af12-edda-41d7-9b44-f87396174ef2" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 13 09:54:46 crc kubenswrapper[4971]: I1213 09:54:46.154211 4971 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-82xjz" Dec 13 09:54:46 crc kubenswrapper[4971]: I1213 09:54:46.155267 4971 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"ba79b440c3c7f17627b56e5585d148b00ef9febdfb525490ad7ac2be28ca8ced"} pod="openshift-machine-config-operator/machine-config-daemon-82xjz" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Dec 13 09:54:46 crc kubenswrapper[4971]: I1213 09:54:46.155349 4971 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-82xjz" podUID="e0c4af12-edda-41d7-9b44-f87396174ef2" containerName="machine-config-daemon" containerID="cri-o://ba79b440c3c7f17627b56e5585d148b00ef9febdfb525490ad7ac2be28ca8ced" gracePeriod=600 Dec 13 09:54:46 crc kubenswrapper[4971]: I1213 09:54:46.596616 4971 generic.go:334] "Generic (PLEG): container finished" podID="e0c4af12-edda-41d7-9b44-f87396174ef2" containerID="ba79b440c3c7f17627b56e5585d148b00ef9febdfb525490ad7ac2be28ca8ced" exitCode=0 Dec 13 09:54:46 crc kubenswrapper[4971]: I1213 09:54:46.596676 4971 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-82xjz" event={"ID":"e0c4af12-edda-41d7-9b44-f87396174ef2","Type":"ContainerDied","Data":"ba79b440c3c7f17627b56e5585d148b00ef9febdfb525490ad7ac2be28ca8ced"} Dec 13 09:54:46 crc kubenswrapper[4971]: I1213 09:54:46.597059 4971 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-82xjz" event={"ID":"e0c4af12-edda-41d7-9b44-f87396174ef2","Type":"ContainerStarted","Data":"cda1648a520bdb005612441665dc1882952f5112d26585083f205622e00bff90"} Dec 13 09:54:46 crc kubenswrapper[4971]: I1213 09:54:46.597090 4971 scope.go:117] "RemoveContainer" containerID="80deb18ddf85834e136ab62ea93a1fa75315016d9f220eb217720a5e562f5e54" Dec 13 09:55:20 crc kubenswrapper[4971]: E1213 09:55:20.769548 4971 kubelet_pods.go:538] "Hostname for pod was too long, truncated it" podName="test-operator-logs-pod-horizontest-horizontest-tests-horizontest" hostnameMaxLen=63 truncatedHostname="test-operator-logs-pod-horizontest-horizontest-tests-horizontes" Dec 13 09:56:12 crc kubenswrapper[4971]: I1213 09:56:12.266309 4971 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-marketplace-tj9pm"] Dec 13 09:56:12 crc kubenswrapper[4971]: E1213 09:56:12.268054 4971 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="2b6b1aa6-d683-46a9-a5ee-9ea29177d2e0" containerName="extract-utilities" Dec 13 09:56:12 crc kubenswrapper[4971]: I1213 09:56:12.268080 4971 state_mem.go:107] "Deleted CPUSet assignment" podUID="2b6b1aa6-d683-46a9-a5ee-9ea29177d2e0" containerName="extract-utilities" Dec 13 09:56:12 crc kubenswrapper[4971]: E1213 09:56:12.268109 4971 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="2b6b1aa6-d683-46a9-a5ee-9ea29177d2e0" containerName="registry-server" Dec 13 09:56:12 crc kubenswrapper[4971]: I1213 09:56:12.268118 4971 state_mem.go:107] "Deleted CPUSet assignment" podUID="2b6b1aa6-d683-46a9-a5ee-9ea29177d2e0" containerName="registry-server" Dec 13 09:56:12 crc kubenswrapper[4971]: E1213 09:56:12.268150 4971 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="2b6b1aa6-d683-46a9-a5ee-9ea29177d2e0" containerName="extract-content" Dec 13 09:56:12 crc kubenswrapper[4971]: I1213 09:56:12.268163 4971 state_mem.go:107] "Deleted CPUSet assignment" podUID="2b6b1aa6-d683-46a9-a5ee-9ea29177d2e0" containerName="extract-content" Dec 13 09:56:12 crc kubenswrapper[4971]: I1213 09:56:12.268486 4971 memory_manager.go:354] "RemoveStaleState removing state" podUID="2b6b1aa6-d683-46a9-a5ee-9ea29177d2e0" containerName="registry-server" Dec 13 09:56:12 crc kubenswrapper[4971]: I1213 09:56:12.270434 4971 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-tj9pm" Dec 13 09:56:12 crc kubenswrapper[4971]: I1213 09:56:12.286819 4971 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-tj9pm"] Dec 13 09:56:12 crc kubenswrapper[4971]: I1213 09:56:12.354780 4971 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/322f4234-6d15-478b-90da-93577da668e8-catalog-content\") pod \"redhat-marketplace-tj9pm\" (UID: \"322f4234-6d15-478b-90da-93577da668e8\") " pod="openshift-marketplace/redhat-marketplace-tj9pm" Dec 13 09:56:12 crc kubenswrapper[4971]: I1213 09:56:12.354875 4971 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-z9jxc\" (UniqueName: \"kubernetes.io/projected/322f4234-6d15-478b-90da-93577da668e8-kube-api-access-z9jxc\") pod \"redhat-marketplace-tj9pm\" (UID: \"322f4234-6d15-478b-90da-93577da668e8\") " pod="openshift-marketplace/redhat-marketplace-tj9pm" Dec 13 09:56:12 crc kubenswrapper[4971]: I1213 09:56:12.354909 4971 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/322f4234-6d15-478b-90da-93577da668e8-utilities\") pod \"redhat-marketplace-tj9pm\" (UID: \"322f4234-6d15-478b-90da-93577da668e8\") " pod="openshift-marketplace/redhat-marketplace-tj9pm" Dec 13 09:56:12 crc kubenswrapper[4971]: I1213 09:56:12.458584 4971 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/322f4234-6d15-478b-90da-93577da668e8-catalog-content\") pod \"redhat-marketplace-tj9pm\" (UID: \"322f4234-6d15-478b-90da-93577da668e8\") " pod="openshift-marketplace/redhat-marketplace-tj9pm" Dec 13 09:56:12 crc kubenswrapper[4971]: I1213 09:56:12.458732 4971 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-z9jxc\" (UniqueName: \"kubernetes.io/projected/322f4234-6d15-478b-90da-93577da668e8-kube-api-access-z9jxc\") pod \"redhat-marketplace-tj9pm\" (UID: \"322f4234-6d15-478b-90da-93577da668e8\") " pod="openshift-marketplace/redhat-marketplace-tj9pm" Dec 13 09:56:12 crc kubenswrapper[4971]: I1213 09:56:12.458763 4971 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/322f4234-6d15-478b-90da-93577da668e8-utilities\") pod \"redhat-marketplace-tj9pm\" (UID: \"322f4234-6d15-478b-90da-93577da668e8\") " pod="openshift-marketplace/redhat-marketplace-tj9pm" Dec 13 09:56:12 crc kubenswrapper[4971]: I1213 09:56:12.459329 4971 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/322f4234-6d15-478b-90da-93577da668e8-catalog-content\") pod \"redhat-marketplace-tj9pm\" (UID: \"322f4234-6d15-478b-90da-93577da668e8\") " pod="openshift-marketplace/redhat-marketplace-tj9pm" Dec 13 09:56:12 crc kubenswrapper[4971]: I1213 09:56:12.459661 4971 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/322f4234-6d15-478b-90da-93577da668e8-utilities\") pod \"redhat-marketplace-tj9pm\" (UID: \"322f4234-6d15-478b-90da-93577da668e8\") " pod="openshift-marketplace/redhat-marketplace-tj9pm" Dec 13 09:56:13 crc kubenswrapper[4971]: I1213 09:56:13.145452 4971 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-z9jxc\" (UniqueName: \"kubernetes.io/projected/322f4234-6d15-478b-90da-93577da668e8-kube-api-access-z9jxc\") pod \"redhat-marketplace-tj9pm\" (UID: \"322f4234-6d15-478b-90da-93577da668e8\") " pod="openshift-marketplace/redhat-marketplace-tj9pm" Dec 13 09:56:13 crc kubenswrapper[4971]: I1213 09:56:13.221169 4971 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-tj9pm" Dec 13 09:56:13 crc kubenswrapper[4971]: I1213 09:56:13.792179 4971 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-tj9pm"] Dec 13 09:56:14 crc kubenswrapper[4971]: I1213 09:56:14.291382 4971 generic.go:334] "Generic (PLEG): container finished" podID="322f4234-6d15-478b-90da-93577da668e8" containerID="4403a0b32677b0364bb1950d14d5e4d77210250353e235a8e6fab7f9a91a3e49" exitCode=0 Dec 13 09:56:14 crc kubenswrapper[4971]: I1213 09:56:14.291551 4971 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-tj9pm" event={"ID":"322f4234-6d15-478b-90da-93577da668e8","Type":"ContainerDied","Data":"4403a0b32677b0364bb1950d14d5e4d77210250353e235a8e6fab7f9a91a3e49"} Dec 13 09:56:14 crc kubenswrapper[4971]: I1213 09:56:14.291976 4971 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-tj9pm" event={"ID":"322f4234-6d15-478b-90da-93577da668e8","Type":"ContainerStarted","Data":"330690f0faae350f7e9daa2462a22bf3028ec891404b10ff1d3c912c2266cca5"} Dec 13 09:56:14 crc kubenswrapper[4971]: I1213 09:56:14.296322 4971 provider.go:102] Refreshing cache for provider: *credentialprovider.defaultDockerConfigProvider Dec 13 09:56:15 crc kubenswrapper[4971]: I1213 09:56:15.307676 4971 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-tj9pm" event={"ID":"322f4234-6d15-478b-90da-93577da668e8","Type":"ContainerStarted","Data":"b2627a383743d528282fbf3e5d509595a2990af183c3776c4bf0541ba91a3daf"} Dec 13 09:56:16 crc kubenswrapper[4971]: I1213 09:56:16.325737 4971 generic.go:334] "Generic (PLEG): container finished" podID="322f4234-6d15-478b-90da-93577da668e8" containerID="b2627a383743d528282fbf3e5d509595a2990af183c3776c4bf0541ba91a3daf" exitCode=0 Dec 13 09:56:16 crc kubenswrapper[4971]: I1213 09:56:16.325818 4971 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-tj9pm" event={"ID":"322f4234-6d15-478b-90da-93577da668e8","Type":"ContainerDied","Data":"b2627a383743d528282fbf3e5d509595a2990af183c3776c4bf0541ba91a3daf"} Dec 13 09:56:17 crc kubenswrapper[4971]: I1213 09:56:17.342884 4971 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-tj9pm" event={"ID":"322f4234-6d15-478b-90da-93577da668e8","Type":"ContainerStarted","Data":"14308518cc47f43d753de07b399b2239829d313631db2fc3a10f5d9095a7c404"} Dec 13 09:56:17 crc kubenswrapper[4971]: I1213 09:56:17.372140 4971 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-marketplace-tj9pm" podStartSLOduration=2.785078531 podStartE2EDuration="5.372107175s" podCreationTimestamp="2025-12-13 09:56:12 +0000 UTC" firstStartedPulling="2025-12-13 09:56:14.295892561 +0000 UTC m=+11230.900302019" lastFinishedPulling="2025-12-13 09:56:16.882921205 +0000 UTC m=+11233.487330663" observedRunningTime="2025-12-13 09:56:17.365470162 +0000 UTC m=+11233.969879620" watchObservedRunningTime="2025-12-13 09:56:17.372107175 +0000 UTC m=+11233.976516623" Dec 13 09:56:23 crc kubenswrapper[4971]: I1213 09:56:23.221772 4971 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-marketplace-tj9pm" Dec 13 09:56:23 crc kubenswrapper[4971]: I1213 09:56:23.222792 4971 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-marketplace-tj9pm" Dec 13 09:56:23 crc kubenswrapper[4971]: I1213 09:56:23.285600 4971 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-marketplace-tj9pm" Dec 13 09:56:23 crc kubenswrapper[4971]: I1213 09:56:23.498521 4971 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-marketplace-tj9pm" Dec 13 09:56:23 crc kubenswrapper[4971]: I1213 09:56:23.570329 4971 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-tj9pm"] Dec 13 09:56:25 crc kubenswrapper[4971]: I1213 09:56:25.439149 4971 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-marketplace-tj9pm" podUID="322f4234-6d15-478b-90da-93577da668e8" containerName="registry-server" containerID="cri-o://14308518cc47f43d753de07b399b2239829d313631db2fc3a10f5d9095a7c404" gracePeriod=2 Dec 13 09:56:26 crc kubenswrapper[4971]: I1213 09:56:26.057153 4971 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-tj9pm" Dec 13 09:56:26 crc kubenswrapper[4971]: I1213 09:56:26.224623 4971 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/322f4234-6d15-478b-90da-93577da668e8-utilities\") pod \"322f4234-6d15-478b-90da-93577da668e8\" (UID: \"322f4234-6d15-478b-90da-93577da668e8\") " Dec 13 09:56:26 crc kubenswrapper[4971]: I1213 09:56:26.224751 4971 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/322f4234-6d15-478b-90da-93577da668e8-catalog-content\") pod \"322f4234-6d15-478b-90da-93577da668e8\" (UID: \"322f4234-6d15-478b-90da-93577da668e8\") " Dec 13 09:56:26 crc kubenswrapper[4971]: I1213 09:56:26.224852 4971 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-z9jxc\" (UniqueName: \"kubernetes.io/projected/322f4234-6d15-478b-90da-93577da668e8-kube-api-access-z9jxc\") pod \"322f4234-6d15-478b-90da-93577da668e8\" (UID: \"322f4234-6d15-478b-90da-93577da668e8\") " Dec 13 09:56:26 crc kubenswrapper[4971]: I1213 09:56:26.225716 4971 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/322f4234-6d15-478b-90da-93577da668e8-utilities" (OuterVolumeSpecName: "utilities") pod "322f4234-6d15-478b-90da-93577da668e8" (UID: "322f4234-6d15-478b-90da-93577da668e8"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 13 09:56:26 crc kubenswrapper[4971]: I1213 09:56:26.238785 4971 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/322f4234-6d15-478b-90da-93577da668e8-kube-api-access-z9jxc" (OuterVolumeSpecName: "kube-api-access-z9jxc") pod "322f4234-6d15-478b-90da-93577da668e8" (UID: "322f4234-6d15-478b-90da-93577da668e8"). InnerVolumeSpecName "kube-api-access-z9jxc". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 13 09:56:26 crc kubenswrapper[4971]: I1213 09:56:26.252919 4971 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/322f4234-6d15-478b-90da-93577da668e8-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "322f4234-6d15-478b-90da-93577da668e8" (UID: "322f4234-6d15-478b-90da-93577da668e8"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 13 09:56:26 crc kubenswrapper[4971]: I1213 09:56:26.327765 4971 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/322f4234-6d15-478b-90da-93577da668e8-catalog-content\") on node \"crc\" DevicePath \"\"" Dec 13 09:56:26 crc kubenswrapper[4971]: I1213 09:56:26.327814 4971 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-z9jxc\" (UniqueName: \"kubernetes.io/projected/322f4234-6d15-478b-90da-93577da668e8-kube-api-access-z9jxc\") on node \"crc\" DevicePath \"\"" Dec 13 09:56:26 crc kubenswrapper[4971]: I1213 09:56:26.327826 4971 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/322f4234-6d15-478b-90da-93577da668e8-utilities\") on node \"crc\" DevicePath \"\"" Dec 13 09:56:26 crc kubenswrapper[4971]: I1213 09:56:26.452800 4971 generic.go:334] "Generic (PLEG): container finished" podID="322f4234-6d15-478b-90da-93577da668e8" containerID="14308518cc47f43d753de07b399b2239829d313631db2fc3a10f5d9095a7c404" exitCode=0 Dec 13 09:56:26 crc kubenswrapper[4971]: I1213 09:56:26.452871 4971 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-tj9pm" Dec 13 09:56:26 crc kubenswrapper[4971]: I1213 09:56:26.452871 4971 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-tj9pm" event={"ID":"322f4234-6d15-478b-90da-93577da668e8","Type":"ContainerDied","Data":"14308518cc47f43d753de07b399b2239829d313631db2fc3a10f5d9095a7c404"} Dec 13 09:56:26 crc kubenswrapper[4971]: I1213 09:56:26.453013 4971 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-tj9pm" event={"ID":"322f4234-6d15-478b-90da-93577da668e8","Type":"ContainerDied","Data":"330690f0faae350f7e9daa2462a22bf3028ec891404b10ff1d3c912c2266cca5"} Dec 13 09:56:26 crc kubenswrapper[4971]: I1213 09:56:26.453042 4971 scope.go:117] "RemoveContainer" containerID="14308518cc47f43d753de07b399b2239829d313631db2fc3a10f5d9095a7c404" Dec 13 09:56:26 crc kubenswrapper[4971]: I1213 09:56:26.478067 4971 scope.go:117] "RemoveContainer" containerID="b2627a383743d528282fbf3e5d509595a2990af183c3776c4bf0541ba91a3daf" Dec 13 09:56:26 crc kubenswrapper[4971]: I1213 09:56:26.502831 4971 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-tj9pm"] Dec 13 09:56:26 crc kubenswrapper[4971]: I1213 09:56:26.527585 4971 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-marketplace-tj9pm"] Dec 13 09:56:26 crc kubenswrapper[4971]: I1213 09:56:26.535343 4971 scope.go:117] "RemoveContainer" containerID="4403a0b32677b0364bb1950d14d5e4d77210250353e235a8e6fab7f9a91a3e49" Dec 13 09:56:26 crc kubenswrapper[4971]: I1213 09:56:26.560203 4971 scope.go:117] "RemoveContainer" containerID="14308518cc47f43d753de07b399b2239829d313631db2fc3a10f5d9095a7c404" Dec 13 09:56:26 crc kubenswrapper[4971]: E1213 09:56:26.560898 4971 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"14308518cc47f43d753de07b399b2239829d313631db2fc3a10f5d9095a7c404\": container with ID starting with 14308518cc47f43d753de07b399b2239829d313631db2fc3a10f5d9095a7c404 not found: ID does not exist" containerID="14308518cc47f43d753de07b399b2239829d313631db2fc3a10f5d9095a7c404" Dec 13 09:56:26 crc kubenswrapper[4971]: I1213 09:56:26.560950 4971 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"14308518cc47f43d753de07b399b2239829d313631db2fc3a10f5d9095a7c404"} err="failed to get container status \"14308518cc47f43d753de07b399b2239829d313631db2fc3a10f5d9095a7c404\": rpc error: code = NotFound desc = could not find container \"14308518cc47f43d753de07b399b2239829d313631db2fc3a10f5d9095a7c404\": container with ID starting with 14308518cc47f43d753de07b399b2239829d313631db2fc3a10f5d9095a7c404 not found: ID does not exist" Dec 13 09:56:26 crc kubenswrapper[4971]: I1213 09:56:26.560982 4971 scope.go:117] "RemoveContainer" containerID="b2627a383743d528282fbf3e5d509595a2990af183c3776c4bf0541ba91a3daf" Dec 13 09:56:26 crc kubenswrapper[4971]: E1213 09:56:26.561661 4971 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"b2627a383743d528282fbf3e5d509595a2990af183c3776c4bf0541ba91a3daf\": container with ID starting with b2627a383743d528282fbf3e5d509595a2990af183c3776c4bf0541ba91a3daf not found: ID does not exist" containerID="b2627a383743d528282fbf3e5d509595a2990af183c3776c4bf0541ba91a3daf" Dec 13 09:56:26 crc kubenswrapper[4971]: I1213 09:56:26.561702 4971 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"b2627a383743d528282fbf3e5d509595a2990af183c3776c4bf0541ba91a3daf"} err="failed to get container status \"b2627a383743d528282fbf3e5d509595a2990af183c3776c4bf0541ba91a3daf\": rpc error: code = NotFound desc = could not find container \"b2627a383743d528282fbf3e5d509595a2990af183c3776c4bf0541ba91a3daf\": container with ID starting with b2627a383743d528282fbf3e5d509595a2990af183c3776c4bf0541ba91a3daf not found: ID does not exist" Dec 13 09:56:26 crc kubenswrapper[4971]: I1213 09:56:26.561723 4971 scope.go:117] "RemoveContainer" containerID="4403a0b32677b0364bb1950d14d5e4d77210250353e235a8e6fab7f9a91a3e49" Dec 13 09:56:26 crc kubenswrapper[4971]: E1213 09:56:26.562045 4971 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"4403a0b32677b0364bb1950d14d5e4d77210250353e235a8e6fab7f9a91a3e49\": container with ID starting with 4403a0b32677b0364bb1950d14d5e4d77210250353e235a8e6fab7f9a91a3e49 not found: ID does not exist" containerID="4403a0b32677b0364bb1950d14d5e4d77210250353e235a8e6fab7f9a91a3e49" Dec 13 09:56:26 crc kubenswrapper[4971]: I1213 09:56:26.562076 4971 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"4403a0b32677b0364bb1950d14d5e4d77210250353e235a8e6fab7f9a91a3e49"} err="failed to get container status \"4403a0b32677b0364bb1950d14d5e4d77210250353e235a8e6fab7f9a91a3e49\": rpc error: code = NotFound desc = could not find container \"4403a0b32677b0364bb1950d14d5e4d77210250353e235a8e6fab7f9a91a3e49\": container with ID starting with 4403a0b32677b0364bb1950d14d5e4d77210250353e235a8e6fab7f9a91a3e49 not found: ID does not exist" Dec 13 09:56:27 crc kubenswrapper[4971]: I1213 09:56:27.784147 4971 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="322f4234-6d15-478b-90da-93577da668e8" path="/var/lib/kubelet/pods/322f4234-6d15-478b-90da-93577da668e8/volumes" Dec 13 09:56:41 crc kubenswrapper[4971]: E1213 09:56:41.768240 4971 kubelet_pods.go:538] "Hostname for pod was too long, truncated it" podName="test-operator-logs-pod-horizontest-horizontest-tests-horizontest" hostnameMaxLen=63 truncatedHostname="test-operator-logs-pod-horizontest-horizontest-tests-horizontes" Dec 13 09:56:46 crc kubenswrapper[4971]: I1213 09:56:46.153448 4971 patch_prober.go:28] interesting pod/machine-config-daemon-82xjz container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 13 09:56:46 crc kubenswrapper[4971]: I1213 09:56:46.154564 4971 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-82xjz" podUID="e0c4af12-edda-41d7-9b44-f87396174ef2" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 13 09:57:16 crc kubenswrapper[4971]: I1213 09:57:16.155222 4971 patch_prober.go:28] interesting pod/machine-config-daemon-82xjz container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 13 09:57:16 crc kubenswrapper[4971]: I1213 09:57:16.156039 4971 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-82xjz" podUID="e0c4af12-edda-41d7-9b44-f87396174ef2" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" var/home/core/zuul-output/logs/crc-cloud-workdir-crc-all-logs.tar.gz0000644000175000000000000000005515117234236024451 0ustar coreroot‹íÁ  ÷Om7 €7šÞ'(var/home/core/zuul-output/logs/crc-cloud/0000755000175000000000000000000015117234236017366 5ustar corerootvar/home/core/zuul-output/artifacts/0000755000175000017500000000000015117205536016512 5ustar corecorevar/home/core/zuul-output/docs/0000755000175000017500000000000015117205536015462 5ustar corecore